{"adam_epsilon": 1e-08, "alphabet": "gene_prot", "append_eos": true, "asl_gamma_neg": 4.0, "asl_gamma_pos": 1.0, "best_metric_type": "sp_statistic", "beta1": 0.9, "beta2": 0.98, "buffer_size": 4096, "cache_dir": null, "classifier_activate_func": "gelu", "classifier_size": 128, "codes_file": null, "config_path": "../config/luca_base/luca_base_config.json", "cross_atten": false, "dataset_name": "DMS_Bind_Reps_Strain", "dataset_type": "protein", "delete_old": true, "dev_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/dev/", "device": "cuda", "do_eval": true, "do_lower_case": false, "do_metrics": true, "do_predict": true, "do_train": true, "dropout_prob": 0.1, "early_stop_epoch": -1, "emb_activate_func": "gelu", "embedding_complete": true, "embedding_complete_seg_overlap": true, "embedding_fixed_len_a_time": 3072, "embedding_input_size": 1152, "embedding_input_size_a": null, "embedding_input_size_b": null, "eval_all_checkpoints": false, "evaluate_during_training": true, "evaluate_steps": 2000, "evaluate_strategy": "epoch", "fc_activate_func": "gelu", "focal_loss_alpha": 0.7, "focal_loss_gamma": 2.0, "focal_loss_reduce": false, "fp16": false, "fp16_embedding": false, "fp16_opt_level": "O1", "fusion_type": "concat", "gradient_accumulation_steps": 1, "hidden_size": 1024, "ignore_index": -100, "input_mode": "single", "input_type": "matrix", "intermediate_size": 4096, "label_filepath": "../dataset/DMS_Bind_Reps_Strain/protein/regression/label.txt", "label_size": 1, "label_type": "DMS_Bind_Reps_Strain", "learning_rate": 0.0001, "llm_dir": "..", "llm_dirpath": null, "llm_step": null, "llm_task_level": "token_level,span_level,seq_level,structure_level", "llm_time_str": null, "llm_type": "esmc", "llm_version": "600M", "lmdb_path": null, "local_rank": -1, "log_dir": "../logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250409154052", "logging_steps": 10, "loss_reduction": "mean", "loss_type": "l2", "lr_decay_rate": 0.9, "lr_update_strategy": "step", "matrix_add_special_token": true, "matrix_dirpath": "../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/600M/esmc//", "matrix_embedding_exists": false, "matrix_encoder": false, "matrix_encoder_act": false, "matrix_fc_size": "128", "matrix_max_length": 100000, "matrix_max_length_a": null, "matrix_max_length_b": null, "matrix_pooling_type": "value_attention", "max_grad_norm": 1.0, "max_sentence_length": null, "max_sentences": null, "max_steps": -1, "model_dirpath": null, "model_type": "luca_base", "n_gpu": 1, "no_cuda": false, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "non_ignore": false, "not_append_eos": false, "not_matrix_encoder_shared": false, "not_prepend_bos": false, "not_save_emb_to_disk": false, "not_seq_encoder_shared": false, "num_attention_heads": 4, "num_hidden_layers": 2, "num_train_epochs": 50, "output_dir": "../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250409154052", "output_mode": "regression", "overwrite_cache": false, "overwrite_output_dir": true, "per_gpu_eval_batch_size": 16, "per_gpu_train_batch_size": 16, "pos_weight": 1.0, "position_embedding_type": "absolute", "prepend_bos": true, "save_all": true, "save_steps": -1, "seed": 1221, "self_atten": false, "seq_fc_size": "null", "seq_max_length": 100000, "seq_max_length_a": null, "seq_max_length_b": null, "seq_pooling_type": "value_attention", "seq_subword": false, "seq_vocab_path": "gene_prot", "sigmoid": false, "task_level_type": "seq_level", "task_type": "regression", "tb_log_dir": "../tb-logs/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250409154052", "test_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/test/", "time_str": "20250409154056", "train_data_dir": "../dataset/DMS_Bind_Reps_Strain/protein/regression/train/", "trunc_type": "right", "vector_dirpath": "../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/600M/esmc//", "vector_fc_size": "null", "vocab_size": 39, "warmup_steps": 1000, "weight": null, "weight_decay": 0.01, "worker_num": 0} ################################################## n_gpu: 1 ################################################## Inputs: Input Name List: protein,embedding_matrix ################################################## Encoder Config: {'llm_type': 'esmc', 'llm_version': '600M', 'llm_step': None, 'llm_dirpath': None, 'input_type': 'matrix', 'trunc_type': 'right', 'seq_max_length': 100000, 'atom_seq_max_length': None, 'vector_dirpath': '../vectors/DMS_Bind_Reps_Strain/protein/regression/luca_base/600M/esmc//', 'matrix_dirpath': '../matrices/DMS_Bind_Reps_Strain/protein/regression/luca_base/600M/esmc//', 'local_rank': -1, 'max_sentence_length': None, 'max_sentences': None, 'matrix_add_special_token': True, 'embedding_complete': True, 'embedding_complete_seg_overlap': True, 'embedding_fixed_len_a_time': 3072, 'matrix_embedding_exists': False, 'save_emb_to_disk': True, 'fp16_embedding': False} ################################################## Model Config: LucaConfig { "_attn_implementation_autoset": true, "alphabet": "gene_prot", "attention_probs_dropout_prob": 0.1, "classifier_activate_func": "gelu", "classifier_dropout_prob": 0.1, "classifier_size": 128, "cls_token_id": 2, "cross_atten": false, "directionality": "bidi", "emb_activate_func": "gelu", "embedding_input_size": 1152, "fc_activate_func": "gelu", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "LABEL_0" }, "ignore_index": -100, "initializer_range": 0.02, "intermediate_size": 4096, "kernel_size": 7, "label2id": { "LABEL_0": 0 }, "layer_norm_eps": 1e-12, "loss_reduction": "mean", "matrix_fc_size": [ 128 ], "matrix_max_length": 100000, "matrix_pooling_type": "value_attention", "max_position_embeddings": 100002, "no_position_embeddings": true, "no_token_embeddings": true, "no_token_type_embeddings": true, "num_attention_heads": 4, "num_hidden_layers": 2, "pad_token_id": 0, "pos_weight": 1.0, "position_embedding_type": "absolute", "self_atten": false, "sep_token_id": 3, "seq_fc_size": null, "seq_max_length": 100000, "seq_pooling_type": "value_attention", "token_dropout": null, "transformers_version": "4.46.3", "type_vocab_size": 2, "use_luca_layer_norm_v2": true, "vector_fc_size": null, "vocab_size": 39 } ################################################## Mode Architecture: LucaBase( (matrix_pooler): GlobalMaskValueAttentionPooling1D (1152 -> 1152) (linear): ModuleList( (0): ModuleList( (0): Linear(in_features=1152, out_features=128, bias=True) (1): GELU(approximate='none') ) ) (dropout): Dropout(p=0.1, inplace=False) (hidden_layer): Linear(in_features=128, out_features=128, bias=True) (hidden_act): GELU(approximate='none') (classifier): Linear(in_features=128, out_features=1, bias=True) (loss_fct): MaskedMSELoss( (criterion): MSELoss() ) ) ################################################## Model parameters: 4145537 ################################################## {"total_num": "3.950000M", "total_size": "15.810000MB", "param_sum": "3.950000M", "param_size": "15.810000MB", "buffer_sum": "0.000000M", "buffer_size": "0.000000MB", "trainable_num": "3.953492M", "trainable_size": "15.813969MB"} ################################################## Train dataset len: 34055, batch size: 16, batch num: 2129 Train dataset t_total: 106450, max_steps: -1 ***** Running training ***** Train Dataset Num examples = 34055 Train Dataset Num Epochs = 50 Logging Steps = 10 Saving Steps = -1 Evaluating Strategy = epoch Train Dataset Instantaneous batch size per GPU = 16 Train Dataset Total train batch size (w. parallel, distributed & accumulation) = 16 Train Dataset Gradient Accumulation steps = 1 Train Dataset Total optimization steps = 106450 ################################################## Training, Epoch: 0001, Batch: 000010, Sample Num: 160, Cur Loss: 70.83910370, Cur Avg Loss: 71.90415649, Log Avg loss: 71.90415649, Global Avg Loss: 71.90415649, Time: 0.0073 Steps: 10, Updated lr: 0.000001 Training, Epoch: 0001, Batch: 000020, Sample Num: 320, Cur Loss: 73.32386780, Cur Avg Loss: 71.38260860, Log Avg loss: 70.86106071, Global Avg Loss: 71.38260860, Time: 0.0072 Steps: 20, Updated lr: 0.000002 Training, Epoch: 0001, Batch: 000030, Sample Num: 480, Cur Loss: 73.00057983, Cur Avg Loss: 72.70027428, Log Avg loss: 75.33560562, Global Avg Loss: 72.70027428, Time: 0.0072 Steps: 30, Updated lr: 0.000003 Training, Epoch: 0001, Batch: 000040, Sample Num: 640, Cur Loss: 67.71291351, Cur Avg Loss: 72.18511343, Log Avg loss: 70.63963089, Global Avg Loss: 72.18511343, Time: 0.0072 Steps: 40, Updated lr: 0.000004 Training, Epoch: 0001, Batch: 000050, Sample Num: 800, Cur Loss: 63.23311234, Cur Avg Loss: 71.99548279, Log Avg loss: 71.23696022, Global Avg Loss: 71.99548279, Time: 0.0072 Steps: 50, Updated lr: 0.000005 Training, Epoch: 0001, Batch: 000060, Sample Num: 960, Cur Loss: 74.04693604, Cur Avg Loss: 71.93639501, Log Avg loss: 71.64095612, Global Avg Loss: 71.93639501, Time: 0.0073 Steps: 60, Updated lr: 0.000006 Training, Epoch: 0001, Batch: 000070, Sample Num: 1120, Cur Loss: 63.26525116, Cur Avg Loss: 71.60833468, Log Avg loss: 69.63997269, Global Avg Loss: 71.60833468, Time: 0.0072 Steps: 70, Updated lr: 0.000007 Training, Epoch: 0001, Batch: 000080, Sample Num: 1280, Cur Loss: 74.91941833, Cur Avg Loss: 71.23042455, Log Avg loss: 68.58505363, Global Avg Loss: 71.23042455, Time: 0.0073 Steps: 80, Updated lr: 0.000008 Training, Epoch: 0001, Batch: 000090, Sample Num: 1440, Cur Loss: 73.67050171, Cur Avg Loss: 70.89131449, Log Avg loss: 68.17843399, Global Avg Loss: 70.89131449, Time: 0.0072 Steps: 90, Updated lr: 0.000009 Training, Epoch: 0001, Batch: 000100, Sample Num: 1600, Cur Loss: 77.10635376, Cur Avg Loss: 71.12522587, Log Avg loss: 73.23042831, Global Avg Loss: 71.12522587, Time: 0.0073 Steps: 100, Updated lr: 0.000010 Training, Epoch: 0001, Batch: 000110, Sample Num: 1760, Cur Loss: 68.19125366, Cur Avg Loss: 71.14837483, Log Avg loss: 71.37986450, Global Avg Loss: 71.14837483, Time: 0.0072 Steps: 110, Updated lr: 0.000011 Training, Epoch: 0001, Batch: 000120, Sample Num: 1920, Cur Loss: 71.65132904, Cur Avg Loss: 71.30006259, Log Avg loss: 72.96862793, Global Avg Loss: 71.30006259, Time: 0.0073 Steps: 120, Updated lr: 0.000012 Training, Epoch: 0001, Batch: 000130, Sample Num: 2080, Cur Loss: 77.70171356, Cur Avg Loss: 71.45294069, Log Avg loss: 73.28747787, Global Avg Loss: 71.45294069, Time: 0.0073 Steps: 130, Updated lr: 0.000013 Training, Epoch: 0001, Batch: 000140, Sample Num: 2240, Cur Loss: 77.63897705, Cur Avg Loss: 71.58328190, Log Avg loss: 73.27771759, Global Avg Loss: 71.58328190, Time: 0.0072 Steps: 140, Updated lr: 0.000014 Training, Epoch: 0001, Batch: 000150, Sample Num: 2400, Cur Loss: 73.47779846, Cur Avg Loss: 71.85213153, Log Avg loss: 75.61602631, Global Avg Loss: 71.85213153, Time: 0.0072 Steps: 150, Updated lr: 0.000015 Training, Epoch: 0001, Batch: 000160, Sample Num: 2560, Cur Loss: 65.66584015, Cur Avg Loss: 71.76020415, Log Avg loss: 70.38129349, Global Avg Loss: 71.76020415, Time: 0.0072 Steps: 160, Updated lr: 0.000016 Training, Epoch: 0001, Batch: 000170, Sample Num: 2720, Cur Loss: 71.38087463, Cur Avg Loss: 71.67051470, Log Avg loss: 70.23548355, Global Avg Loss: 71.67051470, Time: 0.0072 Steps: 170, Updated lr: 0.000017 Training, Epoch: 0001, Batch: 000180, Sample Num: 2880, Cur Loss: 63.08214951, Cur Avg Loss: 71.66331613, Log Avg loss: 71.54094048, Global Avg Loss: 71.66331613, Time: 0.0073 Steps: 180, Updated lr: 0.000018 Training, Epoch: 0001, Batch: 000190, Sample Num: 3040, Cur Loss: 78.48107910, Cur Avg Loss: 71.73704324, Log Avg loss: 73.06413116, Global Avg Loss: 71.73704324, Time: 0.0072 Steps: 190, Updated lr: 0.000019 Training, Epoch: 0001, Batch: 000200, Sample Num: 3200, Cur Loss: 65.68239594, Cur Avg Loss: 71.76494787, Log Avg loss: 72.29513588, Global Avg Loss: 71.76494787, Time: 0.0072 Steps: 200, Updated lr: 0.000020 Training, Epoch: 0001, Batch: 000210, Sample Num: 3360, Cur Loss: 69.47570038, Cur Avg Loss: 71.73056286, Log Avg loss: 71.04286270, Global Avg Loss: 71.73056286, Time: 0.0073 Steps: 210, Updated lr: 0.000021 Training, Epoch: 0001, Batch: 000220, Sample Num: 3520, Cur Loss: 64.50194550, Cur Avg Loss: 71.65515931, Log Avg loss: 70.07168465, Global Avg Loss: 71.65515931, Time: 0.0073 Steps: 220, Updated lr: 0.000022 Training, Epoch: 0001, Batch: 000230, Sample Num: 3680, Cur Loss: 78.55580902, Cur Avg Loss: 71.54949994, Log Avg loss: 69.22499390, Global Avg Loss: 71.54949994, Time: 0.0073 Steps: 230, Updated lr: 0.000023 Training, Epoch: 0001, Batch: 000240, Sample Num: 3840, Cur Loss: 71.84241486, Cur Avg Loss: 71.66429529, Log Avg loss: 74.30458832, Global Avg Loss: 71.66429529, Time: 0.0072 Steps: 240, Updated lr: 0.000024 Training, Epoch: 0001, Batch: 000250, Sample Num: 4000, Cur Loss: 73.60624695, Cur Avg Loss: 71.66580092, Log Avg loss: 71.70193596, Global Avg Loss: 71.66580092, Time: 0.0073 Steps: 250, Updated lr: 0.000025 Training, Epoch: 0001, Batch: 000260, Sample Num: 4160, Cur Loss: 74.86742401, Cur Avg Loss: 71.65400298, Log Avg loss: 71.35905457, Global Avg Loss: 71.65400298, Time: 0.0073 Steps: 260, Updated lr: 0.000026 Training, Epoch: 0001, Batch: 000270, Sample Num: 4320, Cur Loss: 61.46968842, Cur Avg Loss: 71.59051295, Log Avg loss: 69.93977203, Global Avg Loss: 71.59051295, Time: 0.0072 Steps: 270, Updated lr: 0.000027 Training, Epoch: 0001, Batch: 000280, Sample Num: 4480, Cur Loss: 68.58329773, Cur Avg Loss: 71.59230201, Log Avg loss: 71.64060669, Global Avg Loss: 71.59230201, Time: 0.0073 Steps: 280, Updated lr: 0.000028 Training, Epoch: 0001, Batch: 000290, Sample Num: 4640, Cur Loss: 69.20315552, Cur Avg Loss: 71.60555769, Log Avg loss: 71.97671661, Global Avg Loss: 71.60555769, Time: 0.0072 Steps: 290, Updated lr: 0.000029 Training, Epoch: 0001, Batch: 000300, Sample Num: 4800, Cur Loss: 73.40220642, Cur Avg Loss: 71.59023406, Log Avg loss: 71.14584885, Global Avg Loss: 71.59023406, Time: 0.0072 Steps: 300, Updated lr: 0.000030 Training, Epoch: 0001, Batch: 000310, Sample Num: 4960, Cur Loss: 64.70092010, Cur Avg Loss: 71.52342302, Log Avg loss: 69.51909180, Global Avg Loss: 71.52342302, Time: 0.0072 Steps: 310, Updated lr: 0.000031 Training, Epoch: 0001, Batch: 000320, Sample Num: 5120, Cur Loss: 75.16917419, Cur Avg Loss: 71.47233020, Log Avg loss: 69.88845291, Global Avg Loss: 71.47233020, Time: 0.0072 Steps: 320, Updated lr: 0.000032 Training, Epoch: 0001, Batch: 000330, Sample Num: 5280, Cur Loss: 71.61771393, Cur Avg Loss: 71.36001890, Log Avg loss: 67.76605721, Global Avg Loss: 71.36001890, Time: 0.0072 Steps: 330, Updated lr: 0.000033 Training, Epoch: 0001, Batch: 000340, Sample Num: 5440, Cur Loss: 67.15634155, Cur Avg Loss: 71.27766878, Log Avg loss: 68.56011505, Global Avg Loss: 71.27766878, Time: 0.0073 Steps: 340, Updated lr: 0.000034 Training, Epoch: 0001, Batch: 000350, Sample Num: 5600, Cur Loss: 80.79173279, Cur Avg Loss: 71.30849322, Log Avg loss: 72.35652390, Global Avg Loss: 71.30849322, Time: 0.0073 Steps: 350, Updated lr: 0.000035 Training, Epoch: 0001, Batch: 000360, Sample Num: 5760, Cur Loss: 67.73081970, Cur Avg Loss: 71.28555766, Log Avg loss: 70.48281326, Global Avg Loss: 71.28555766, Time: 0.0072 Steps: 360, Updated lr: 0.000036 Training, Epoch: 0001, Batch: 000370, Sample Num: 5920, Cur Loss: 65.30223083, Cur Avg Loss: 71.17551241, Log Avg loss: 67.21388321, Global Avg Loss: 71.17551241, Time: 0.0073 Steps: 370, Updated lr: 0.000037 Training, Epoch: 0001, Batch: 000380, Sample Num: 6080, Cur Loss: 68.34089661, Cur Avg Loss: 71.08903519, Log Avg loss: 67.88937836, Global Avg Loss: 71.08903519, Time: 0.0073 Steps: 380, Updated lr: 0.000038 Training, Epoch: 0001, Batch: 000390, Sample Num: 6240, Cur Loss: 58.08480453, Cur Avg Loss: 70.99248183, Log Avg loss: 67.32345390, Global Avg Loss: 70.99248183, Time: 0.0073 Steps: 390, Updated lr: 0.000039 Training, Epoch: 0001, Batch: 000400, Sample Num: 6400, Cur Loss: 69.35942078, Cur Avg Loss: 70.92654924, Log Avg loss: 68.35517845, Global Avg Loss: 70.92654924, Time: 0.0073 Steps: 400, Updated lr: 0.000040 Training, Epoch: 0001, Batch: 000410, Sample Num: 6560, Cur Loss: 71.29988098, Cur Avg Loss: 70.87773733, Log Avg loss: 68.92526093, Global Avg Loss: 70.87773733, Time: 0.0072 Steps: 410, Updated lr: 0.000041 Training, Epoch: 0001, Batch: 000420, Sample Num: 6720, Cur Loss: 68.86959839, Cur Avg Loss: 70.81311557, Log Avg loss: 68.16362343, Global Avg Loss: 70.81311557, Time: 0.0072 Steps: 420, Updated lr: 0.000042 Training, Epoch: 0001, Batch: 000430, Sample Num: 6880, Cur Loss: 63.79372025, Cur Avg Loss: 70.72736618, Log Avg loss: 67.12589149, Global Avg Loss: 70.72736618, Time: 0.0072 Steps: 430, Updated lr: 0.000043 Training, Epoch: 0001, Batch: 000440, Sample Num: 7040, Cur Loss: 71.51623535, Cur Avg Loss: 70.63459199, Log Avg loss: 66.64530182, Global Avg Loss: 70.63459199, Time: 0.0072 Steps: 440, Updated lr: 0.000044 Training, Epoch: 0001, Batch: 000450, Sample Num: 7200, Cur Loss: 62.56367111, Cur Avg Loss: 70.55094771, Log Avg loss: 66.87059975, Global Avg Loss: 70.55094771, Time: 0.0073 Steps: 450, Updated lr: 0.000045 Training, Epoch: 0001, Batch: 000460, Sample Num: 7360, Cur Loss: 65.93601990, Cur Avg Loss: 70.42604082, Log Avg loss: 64.80523033, Global Avg Loss: 70.42604082, Time: 0.0073 Steps: 460, Updated lr: 0.000046 Training, Epoch: 0001, Batch: 000470, Sample Num: 7520, Cur Loss: 73.80090332, Cur Avg Loss: 70.30349592, Log Avg loss: 64.66643066, Global Avg Loss: 70.30349592, Time: 0.0073 Steps: 470, Updated lr: 0.000047 Training, Epoch: 0001, Batch: 000480, Sample Num: 7680, Cur Loss: 53.39826202, Cur Avg Loss: 70.13109879, Log Avg loss: 62.02843399, Global Avg Loss: 70.13109879, Time: 0.0072 Steps: 480, Updated lr: 0.000048 Training, Epoch: 0001, Batch: 000490, Sample Num: 7840, Cur Loss: 62.01364899, Cur Avg Loss: 69.97690339, Log Avg loss: 62.57552376, Global Avg Loss: 69.97690339, Time: 0.0073 Steps: 490, Updated lr: 0.000049 Training, Epoch: 0001, Batch: 000500, Sample Num: 8000, Cur Loss: 63.34084702, Cur Avg Loss: 69.79484232, Log Avg loss: 60.87384987, Global Avg Loss: 69.79484232, Time: 0.0072 Steps: 500, Updated lr: 0.000050 Training, Epoch: 0001, Batch: 000510, Sample Num: 8160, Cur Loss: 54.15161133, Cur Avg Loss: 69.54492748, Log Avg loss: 57.04918556, Global Avg Loss: 69.54492748, Time: 0.0073 Steps: 510, Updated lr: 0.000051 Training, Epoch: 0001, Batch: 000520, Sample Num: 8320, Cur Loss: 54.30599976, Cur Avg Loss: 69.34032745, Log Avg loss: 58.90572624, Global Avg Loss: 69.34032745, Time: 0.0067 Steps: 520, Updated lr: 0.000052 Training, Epoch: 0001, Batch: 000530, Sample Num: 8480, Cur Loss: 56.62905884, Cur Avg Loss: 69.09637022, Log Avg loss: 56.41059418, Global Avg Loss: 69.09637022, Time: 0.0067 Steps: 530, Updated lr: 0.000053 Training, Epoch: 0001, Batch: 000540, Sample Num: 8640, Cur Loss: 58.53425217, Cur Avg Loss: 68.90178831, Log Avg loss: 58.58894691, Global Avg Loss: 68.90178831, Time: 0.0064 Steps: 540, Updated lr: 0.000054 Training, Epoch: 0001, Batch: 000550, Sample Num: 8800, Cur Loss: 48.75948715, Cur Avg Loss: 68.59578943, Log Avg loss: 52.07185020, Global Avg Loss: 68.59578943, Time: 0.0063 Steps: 550, Updated lr: 0.000055 Training, Epoch: 0001, Batch: 000560, Sample Num: 8960, Cur Loss: 52.20358276, Cur Avg Loss: 68.32435842, Log Avg loss: 53.39565239, Global Avg Loss: 68.32435842, Time: 0.0063 Steps: 560, Updated lr: 0.000056 Training, Epoch: 0001, Batch: 000570, Sample Num: 9120, Cur Loss: 49.40600586, Cur Avg Loss: 67.99113155, Log Avg loss: 49.33042717, Global Avg Loss: 67.99113155, Time: 0.0064 Steps: 570, Updated lr: 0.000057 Training, Epoch: 0001, Batch: 000580, Sample Num: 9280, Cur Loss: 51.20043182, Cur Avg Loss: 67.69720539, Log Avg loss: 50.94341431, Global Avg Loss: 67.69720539, Time: 0.0063 Steps: 580, Updated lr: 0.000058 Training, Epoch: 0001, Batch: 000590, Sample Num: 9440, Cur Loss: 49.87834549, Cur Avg Loss: 67.32953891, Log Avg loss: 46.00488281, Global Avg Loss: 67.32953891, Time: 0.0064 Steps: 590, Updated lr: 0.000059 Training, Epoch: 0001, Batch: 000600, Sample Num: 9600, Cur Loss: 41.38011551, Cur Avg Loss: 66.94175835, Log Avg loss: 44.06270523, Global Avg Loss: 66.94175835, Time: 0.0063 Steps: 600, Updated lr: 0.000060 Training, Epoch: 0001, Batch: 000610, Sample Num: 9760, Cur Loss: 36.45760727, Cur Avg Loss: 66.51442827, Log Avg loss: 40.87462349, Global Avg Loss: 66.51442827, Time: 0.0073 Steps: 610, Updated lr: 0.000061 Training, Epoch: 0001, Batch: 000620, Sample Num: 9920, Cur Loss: 37.01134491, Cur Avg Loss: 66.03314085, Log Avg loss: 36.67460823, Global Avg Loss: 66.03314085, Time: 0.0073 Steps: 620, Updated lr: 0.000062 Training, Epoch: 0001, Batch: 000630, Sample Num: 10080, Cur Loss: 37.03860474, Cur Avg Loss: 65.52205072, Log Avg loss: 33.83446274, Global Avg Loss: 65.52205072, Time: 0.0073 Steps: 630, Updated lr: 0.000063 Training, Epoch: 0001, Batch: 000640, Sample Num: 10240, Cur Loss: 30.84494019, Cur Avg Loss: 64.96533217, Log Avg loss: 29.89206371, Global Avg Loss: 64.96533217, Time: 0.0073 Steps: 640, Updated lr: 0.000064 Training, Epoch: 0001, Batch: 000650, Sample Num: 10400, Cur Loss: 23.78942490, Cur Avg Loss: 64.39972181, Log Avg loss: 28.20065880, Global Avg Loss: 64.39972181, Time: 0.0073 Steps: 650, Updated lr: 0.000065 Training, Epoch: 0001, Batch: 000660, Sample Num: 10560, Cur Loss: 23.30581474, Cur Avg Loss: 63.79989535, Log Avg loss: 24.81117554, Global Avg Loss: 63.79989535, Time: 0.0073 Steps: 660, Updated lr: 0.000066 Training, Epoch: 0001, Batch: 000670, Sample Num: 10720, Cur Loss: 18.57476425, Cur Avg Loss: 63.19007241, Log Avg loss: 22.94175797, Global Avg Loss: 63.19007241, Time: 0.0073 Steps: 670, Updated lr: 0.000067 Training, Epoch: 0001, Batch: 000680, Sample Num: 10880, Cur Loss: 16.00554276, Cur Avg Loss: 62.54290356, Log Avg loss: 19.18259106, Global Avg Loss: 62.54290356, Time: 0.0073 Steps: 680, Updated lr: 0.000068 Training, Epoch: 0001, Batch: 000690, Sample Num: 11040, Cur Loss: 16.24849701, Cur Avg Loss: 61.87325829, Log Avg loss: 16.33737946, Global Avg Loss: 61.87325829, Time: 0.0072 Steps: 690, Updated lr: 0.000069 Training, Epoch: 0001, Batch: 000700, Sample Num: 11200, Cur Loss: 11.43223572, Cur Avg Loss: 61.16328264, Log Avg loss: 12.17496319, Global Avg Loss: 61.16328264, Time: 0.0072 Steps: 700, Updated lr: 0.000070 Training, Epoch: 0001, Batch: 000710, Sample Num: 11360, Cur Loss: 9.12982368, Cur Avg Loss: 60.43745128, Log Avg loss: 9.62925596, Global Avg Loss: 60.43745128, Time: 0.0072 Steps: 710, Updated lr: 0.000071 Training, Epoch: 0001, Batch: 000720, Sample Num: 11520, Cur Loss: 5.50712490, Cur Avg Loss: 59.69262396, Log Avg loss: 6.80988402, Global Avg Loss: 59.69262396, Time: 0.0072 Steps: 720, Updated lr: 0.000072 Training, Epoch: 0001, Batch: 000730, Sample Num: 11680, Cur Loss: 3.23025703, Cur Avg Loss: 58.94593487, Log Avg loss: 5.18432028, Global Avg Loss: 58.94593487, Time: 0.0073 Steps: 730, Updated lr: 0.000073 Training, Epoch: 0001, Batch: 000740, Sample Num: 11840, Cur Loss: 3.88057804, Cur Avg Loss: 58.20092974, Log Avg loss: 3.81555588, Global Avg Loss: 58.20092974, Time: 0.0074 Steps: 740, Updated lr: 0.000074 Training, Epoch: 0001, Batch: 000750, Sample Num: 12000, Cur Loss: 3.67555237, Cur Avg Loss: 57.46143220, Log Avg loss: 2.73861415, Global Avg Loss: 57.46143220, Time: 0.0076 Steps: 750, Updated lr: 0.000075 Training, Epoch: 0001, Batch: 000760, Sample Num: 12160, Cur Loss: 1.83925080, Cur Avg Loss: 56.73546334, Log Avg loss: 2.28779819, Global Avg Loss: 56.73546334, Time: 0.0074 Steps: 760, Updated lr: 0.000076 Training, Epoch: 0001, Batch: 000770, Sample Num: 12320, Cur Loss: 2.11679173, Cur Avg Loss: 56.03071795, Log Avg loss: 2.47006836, Global Avg Loss: 56.03071795, Time: 0.0141 Steps: 770, Updated lr: 0.000077 Training, Epoch: 0001, Batch: 000780, Sample Num: 12480, Cur Loss: 2.02943420, Cur Avg Loss: 55.34355790, Log Avg loss: 2.43223410, Global Avg Loss: 55.34355790, Time: 0.0067 Steps: 780, Updated lr: 0.000078 Training, Epoch: 0001, Batch: 000790, Sample Num: 12640, Cur Loss: 1.62880743, Cur Avg Loss: 54.66608705, Log Avg loss: 1.82336104, Global Avg Loss: 54.66608705, Time: 0.0114 Steps: 790, Updated lr: 0.000079 Training, Epoch: 0001, Batch: 000800, Sample Num: 12800, Cur Loss: 1.64141679, Cur Avg Loss: 54.00697777, Log Avg loss: 1.93734463, Global Avg Loss: 54.00697777, Time: 0.0095 Steps: 800, Updated lr: 0.000080 Training, Epoch: 0001, Batch: 000810, Sample Num: 12960, Cur Loss: 2.40942669, Cur Avg Loss: 53.36566578, Log Avg loss: 2.06070698, Global Avg Loss: 53.36566578, Time: 0.0095 Steps: 810, Updated lr: 0.000081 Training, Epoch: 0001, Batch: 000820, Sample Num: 13120, Cur Loss: 1.06334662, Cur Avg Loss: 52.73933468, Log Avg loss: 2.00651484, Global Avg Loss: 52.73933468, Time: 0.0151 Steps: 820, Updated lr: 0.000082 Training, Epoch: 0001, Batch: 000830, Sample Num: 13280, Cur Loss: 1.46831405, Cur Avg Loss: 52.13030857, Log Avg loss: 2.19016804, Global Avg Loss: 52.13030857, Time: 0.0064 Steps: 830, Updated lr: 0.000083 Training, Epoch: 0001, Batch: 000840, Sample Num: 13440, Cur Loss: 0.99921620, Cur Avg Loss: 51.53542496, Log Avg loss: 2.16008548, Global Avg Loss: 51.53542496, Time: 0.0112 Steps: 840, Updated lr: 0.000084 Training, Epoch: 0001, Batch: 000850, Sample Num: 13600, Cur Loss: 3.21066260, Cur Avg Loss: 50.95105154, Log Avg loss: 1.86368375, Global Avg Loss: 50.95105154, Time: 0.0066 Steps: 850, Updated lr: 0.000085 Training, Epoch: 0001, Batch: 000860, Sample Num: 13760, Cur Loss: 2.07099009, Cur Avg Loss: 50.38000028, Log Avg loss: 1.84064382, Global Avg Loss: 50.38000028, Time: 0.0095 Steps: 860, Updated lr: 0.000086 Training, Epoch: 0001, Batch: 000870, Sample Num: 13920, Cur Loss: 0.84608173, Cur Avg Loss: 49.82422477, Log Avg loss: 2.02753055, Global Avg Loss: 49.82422477, Time: 0.0236 Steps: 870, Updated lr: 0.000087 Training, Epoch: 0001, Batch: 000880, Sample Num: 14080, Cur Loss: 1.54964805, Cur Avg Loss: 49.28011310, Log Avg loss: 1.94239743, Global Avg Loss: 49.28011310, Time: 0.0112 Steps: 880, Updated lr: 0.000088 Training, Epoch: 0001, Batch: 000890, Sample Num: 14240, Cur Loss: 1.32802486, Cur Avg Loss: 48.74963032, Log Avg loss: 2.06714584, Global Avg Loss: 48.74963032, Time: 0.0115 Steps: 890, Updated lr: 0.000089 Training, Epoch: 0001, Batch: 000900, Sample Num: 14400, Cur Loss: 1.70602906, Cur Avg Loss: 48.22889519, Log Avg loss: 1.88346859, Global Avg Loss: 48.22889519, Time: 0.0195 Steps: 900, Updated lr: 0.000090 Training, Epoch: 0001, Batch: 000910, Sample Num: 14560, Cur Loss: 2.69575953, Cur Avg Loss: 47.72303963, Log Avg loss: 2.19603908, Global Avg Loss: 47.72303963, Time: 0.0112 Steps: 910, Updated lr: 0.000091 Training, Epoch: 0001, Batch: 000920, Sample Num: 14720, Cur Loss: 2.65892553, Cur Avg Loss: 47.22550758, Log Avg loss: 1.95009161, Global Avg Loss: 47.22550758, Time: 0.0066 Steps: 920, Updated lr: 0.000092 Training, Epoch: 0001, Batch: 000930, Sample Num: 14880, Cur Loss: 2.64195466, Cur Avg Loss: 46.74038356, Log Avg loss: 2.10897361, Global Avg Loss: 46.74038356, Time: 0.0125 Steps: 930, Updated lr: 0.000093 Training, Epoch: 0001, Batch: 000940, Sample Num: 15040, Cur Loss: 3.51569510, Cur Avg Loss: 46.26564406, Log Avg loss: 2.11487041, Global Avg Loss: 46.26564406, Time: 0.0064 Steps: 940, Updated lr: 0.000094 Training, Epoch: 0001, Batch: 000950, Sample Num: 15200, Cur Loss: 2.40997696, Cur Avg Loss: 45.80112430, Log Avg loss: 2.13626705, Global Avg Loss: 45.80112430, Time: 0.0118 Steps: 950, Updated lr: 0.000095 Training, Epoch: 0001, Batch: 000960, Sample Num: 15360, Cur Loss: 2.98941517, Cur Avg Loss: 45.34639453, Log Avg loss: 2.14706639, Global Avg Loss: 45.34639453, Time: 0.0076 Steps: 960, Updated lr: 0.000096 Training, Epoch: 0001, Batch: 000970, Sample Num: 15520, Cur Loss: 1.80744743, Cur Avg Loss: 44.90016527, Log Avg loss: 2.06215568, Global Avg Loss: 44.90016527, Time: 0.0096 Steps: 970, Updated lr: 0.000097 Training, Epoch: 0001, Batch: 000980, Sample Num: 15680, Cur Loss: 1.63093174, Cur Avg Loss: 44.46324095, Log Avg loss: 2.08158262, Global Avg Loss: 44.46324095, Time: 0.0142 Steps: 980, Updated lr: 0.000098 Training, Epoch: 0001, Batch: 000990, Sample Num: 15840, Cur Loss: 2.35898924, Cur Avg Loss: 44.03566146, Log Avg loss: 2.13287081, Global Avg Loss: 44.03566146, Time: 0.0127 Steps: 990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001000, Sample Num: 16000, Cur Loss: 1.00556231, Cur Avg Loss: 43.61361809, Log Avg loss: 1.83132434, Global Avg Loss: 43.61361809, Time: 0.0087 Steps: 1000, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001010, Sample Num: 16160, Cur Loss: 1.68823040, Cur Avg Loss: 43.20076105, Log Avg loss: 1.91505713, Global Avg Loss: 43.20076105, Time: 0.0108 Steps: 1010, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001020, Sample Num: 16320, Cur Loss: 1.72988653, Cur Avg Loss: 42.79334544, Log Avg loss: 1.64436888, Global Avg Loss: 42.79334544, Time: 0.0113 Steps: 1020, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001030, Sample Num: 16480, Cur Loss: 2.07036519, Cur Avg Loss: 42.39337943, Log Avg loss: 1.59684697, Global Avg Loss: 42.39337943, Time: 0.0116 Steps: 1030, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001040, Sample Num: 16640, Cur Loss: 1.27378273, Cur Avg Loss: 41.99882188, Log Avg loss: 1.35939451, Global Avg Loss: 41.99882188, Time: 0.0111 Steps: 1040, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001050, Sample Num: 16800, Cur Loss: 1.34227729, Cur Avg Loss: 41.61773390, Log Avg loss: 1.98458393, Global Avg Loss: 41.61773390, Time: 0.0153 Steps: 1050, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001060, Sample Num: 16960, Cur Loss: 2.27726078, Cur Avg Loss: 41.24367518, Log Avg loss: 1.96750966, Global Avg Loss: 41.24367518, Time: 0.0118 Steps: 1060, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001070, Sample Num: 17120, Cur Loss: 3.91559720, Cur Avg Loss: 40.88069708, Log Avg loss: 2.40501769, Global Avg Loss: 40.88069708, Time: 0.0064 Steps: 1070, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001080, Sample Num: 17280, Cur Loss: 3.46984386, Cur Avg Loss: 40.52111805, Log Avg loss: 2.04616238, Global Avg Loss: 40.52111805, Time: 0.0064 Steps: 1080, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001090, Sample Num: 17440, Cur Loss: 2.98502803, Cur Avg Loss: 40.16842005, Log Avg loss: 2.07703575, Global Avg Loss: 40.16842005, Time: 0.0064 Steps: 1090, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001100, Sample Num: 17600, Cur Loss: 2.38460350, Cur Avg Loss: 39.82173982, Log Avg loss: 2.03359452, Global Avg Loss: 39.82173982, Time: 0.0065 Steps: 1100, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001110, Sample Num: 17760, Cur Loss: 2.01167822, Cur Avg Loss: 39.47966797, Log Avg loss: 1.85176513, Global Avg Loss: 39.47966797, Time: 0.0072 Steps: 1110, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001120, Sample Num: 17920, Cur Loss: 1.31261015, Cur Avg Loss: 39.14413428, Log Avg loss: 1.89989407, Global Avg Loss: 39.14413428, Time: 0.0066 Steps: 1120, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001130, Sample Num: 18080, Cur Loss: 1.53363085, Cur Avg Loss: 38.81330899, Log Avg loss: 1.76087723, Global Avg Loss: 38.81330899, Time: 0.0075 Steps: 1130, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001140, Sample Num: 18240, Cur Loss: 2.56993294, Cur Avg Loss: 38.48734846, Log Avg loss: 1.65380821, Global Avg Loss: 38.48734846, Time: 0.0158 Steps: 1140, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001150, Sample Num: 18400, Cur Loss: 1.38882887, Cur Avg Loss: 38.16693322, Log Avg loss: 1.63959557, Global Avg Loss: 38.16693322, Time: 0.0241 Steps: 1150, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001160, Sample Num: 18560, Cur Loss: 2.52950215, Cur Avg Loss: 37.85409588, Log Avg loss: 1.87780174, Global Avg Loss: 37.85409588, Time: 0.0065 Steps: 1160, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001170, Sample Num: 18720, Cur Loss: 2.37785745, Cur Avg Loss: 37.54381273, Log Avg loss: 1.55096740, Global Avg Loss: 37.54381273, Time: 0.0066 Steps: 1170, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001180, Sample Num: 18880, Cur Loss: 1.11731315, Cur Avg Loss: 37.24325005, Log Avg loss: 2.07741709, Global Avg Loss: 37.24325005, Time: 0.0240 Steps: 1180, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001190, Sample Num: 19040, Cur Loss: 1.50492287, Cur Avg Loss: 36.94339966, Log Avg loss: 1.56105354, Global Avg Loss: 36.94339966, Time: 0.0235 Steps: 1190, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001200, Sample Num: 19200, Cur Loss: 1.78540671, Cur Avg Loss: 36.64871538, Log Avg loss: 1.58128566, Global Avg Loss: 36.64871538, Time: 0.0152 Steps: 1200, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001210, Sample Num: 19360, Cur Loss: 1.10066915, Cur Avg Loss: 36.35908101, Log Avg loss: 1.60295615, Global Avg Loss: 36.35908101, Time: 0.0205 Steps: 1210, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001220, Sample Num: 19520, Cur Loss: 1.17026401, Cur Avg Loss: 36.07536832, Log Avg loss: 1.74613297, Global Avg Loss: 36.07536832, Time: 0.0076 Steps: 1220, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001230, Sample Num: 19680, Cur Loss: 2.39103627, Cur Avg Loss: 35.79578966, Log Avg loss: 1.68719405, Global Avg Loss: 35.79578966, Time: 0.0073 Steps: 1230, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001240, Sample Num: 19840, Cur Loss: 1.75240016, Cur Avg Loss: 35.52269960, Log Avg loss: 1.93262228, Global Avg Loss: 35.52269960, Time: 0.0233 Steps: 1240, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001250, Sample Num: 20000, Cur Loss: 2.10134768, Cur Avg Loss: 35.25126223, Log Avg loss: 1.59302757, Global Avg Loss: 35.25126223, Time: 0.0198 Steps: 1250, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001260, Sample Num: 20160, Cur Loss: 1.23365927, Cur Avg Loss: 34.98039868, Log Avg loss: 1.12245512, Global Avg Loss: 34.98039868, Time: 0.0115 Steps: 1260, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001270, Sample Num: 20320, Cur Loss: 1.85369229, Cur Avg Loss: 34.71813244, Log Avg loss: 1.67258661, Global Avg Loss: 34.71813244, Time: 0.0064 Steps: 1270, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001280, Sample Num: 20480, Cur Loss: 1.56962657, Cur Avg Loss: 34.46400450, Log Avg loss: 2.18975555, Global Avg Loss: 34.46400450, Time: 0.0081 Steps: 1280, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001290, Sample Num: 20640, Cur Loss: 0.83197981, Cur Avg Loss: 34.20873504, Log Avg loss: 1.53424393, Global Avg Loss: 34.20873504, Time: 0.0212 Steps: 1290, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001300, Sample Num: 20800, Cur Loss: 2.64227009, Cur Avg Loss: 33.95894206, Log Avg loss: 1.73564829, Global Avg Loss: 33.95894206, Time: 0.0068 Steps: 1300, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001310, Sample Num: 20960, Cur Loss: 1.31213069, Cur Avg Loss: 33.70927297, Log Avg loss: 1.25229134, Global Avg Loss: 33.70927297, Time: 0.0106 Steps: 1310, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001320, Sample Num: 21120, Cur Loss: 1.60270214, Cur Avg Loss: 33.46667857, Log Avg loss: 1.68681205, Global Avg Loss: 33.46667857, Time: 0.0122 Steps: 1320, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001330, Sample Num: 21280, Cur Loss: 0.52709270, Cur Avg Loss: 33.22822095, Log Avg loss: 1.75181507, Global Avg Loss: 33.22822095, Time: 0.0127 Steps: 1330, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001340, Sample Num: 21440, Cur Loss: 2.09213376, Cur Avg Loss: 32.99163143, Log Avg loss: 1.52522459, Global Avg Loss: 32.99163143, Time: 0.0068 Steps: 1340, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001350, Sample Num: 21600, Cur Loss: 1.68380308, Cur Avg Loss: 32.76013399, Log Avg loss: 1.73947741, Global Avg Loss: 32.76013399, Time: 0.0067 Steps: 1350, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001360, Sample Num: 21760, Cur Loss: 1.31654012, Cur Avg Loss: 32.53259547, Log Avg loss: 1.81489584, Global Avg Loss: 32.53259547, Time: 0.0112 Steps: 1360, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001370, Sample Num: 21920, Cur Loss: 1.34475517, Cur Avg Loss: 32.30783011, Log Avg loss: 1.73974122, Global Avg Loss: 32.30783011, Time: 0.0136 Steps: 1370, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001380, Sample Num: 22080, Cur Loss: 1.25145614, Cur Avg Loss: 32.08792277, Log Avg loss: 1.96061661, Global Avg Loss: 32.08792277, Time: 0.0206 Steps: 1380, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001390, Sample Num: 22240, Cur Loss: 1.52185285, Cur Avg Loss: 31.87055996, Log Avg loss: 1.87449198, Global Avg Loss: 31.87055996, Time: 0.0122 Steps: 1390, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001400, Sample Num: 22400, Cur Loss: 2.44321012, Cur Avg Loss: 31.65740505, Log Avg loss: 2.02887276, Global Avg Loss: 31.65740505, Time: 0.0067 Steps: 1400, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001410, Sample Num: 22560, Cur Loss: 0.81589246, Cur Avg Loss: 31.44377165, Log Avg loss: 1.53509516, Global Avg Loss: 31.44377165, Time: 0.0117 Steps: 1410, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001420, Sample Num: 22720, Cur Loss: 1.32026911, Cur Avg Loss: 31.23347280, Log Avg loss: 1.58133572, Global Avg Loss: 31.23347280, Time: 0.0064 Steps: 1420, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001430, Sample Num: 22880, Cur Loss: 1.31160235, Cur Avg Loss: 31.02879280, Log Avg loss: 1.96423222, Global Avg Loss: 31.02879280, Time: 0.0142 Steps: 1430, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001440, Sample Num: 23040, Cur Loss: 1.01971436, Cur Avg Loss: 30.82385400, Log Avg loss: 1.51760632, Global Avg Loss: 30.82385400, Time: 0.0117 Steps: 1440, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001450, Sample Num: 23200, Cur Loss: 0.45723468, Cur Avg Loss: 30.62115092, Log Avg loss: 1.43190774, Global Avg Loss: 30.62115092, Time: 0.0163 Steps: 1450, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001460, Sample Num: 23360, Cur Loss: 1.64206839, Cur Avg Loss: 30.42211108, Log Avg loss: 1.56133422, Global Avg Loss: 30.42211108, Time: 0.0128 Steps: 1460, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001470, Sample Num: 23520, Cur Loss: 2.87241292, Cur Avg Loss: 30.22523513, Log Avg loss: 1.48134536, Global Avg Loss: 30.22523513, Time: 0.0114 Steps: 1470, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001480, Sample Num: 23680, Cur Loss: 1.25459683, Cur Avg Loss: 30.03190565, Log Avg loss: 1.61247314, Global Avg Loss: 30.03190565, Time: 0.0065 Steps: 1480, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001490, Sample Num: 23840, Cur Loss: 1.94925523, Cur Avg Loss: 29.83964829, Log Avg loss: 1.38555877, Global Avg Loss: 29.83964829, Time: 0.0152 Steps: 1490, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001500, Sample Num: 24000, Cur Loss: 1.75037897, Cur Avg Loss: 29.65227647, Log Avg loss: 1.73387455, Global Avg Loss: 29.65227647, Time: 0.0240 Steps: 1500, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001510, Sample Num: 24160, Cur Loss: 0.52172607, Cur Avg Loss: 29.46666702, Log Avg loss: 1.62524948, Global Avg Loss: 29.46666702, Time: 0.0067 Steps: 1510, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001520, Sample Num: 24320, Cur Loss: 1.23553836, Cur Avg Loss: 29.28293466, Log Avg loss: 1.53934921, Global Avg Loss: 29.28293466, Time: 0.0120 Steps: 1520, Updated lr: 0.000100 Training, Epoch: 0001, Batch: 001530, Sample Num: 24480, Cur Loss: 0.99540359, Cur Avg Loss: 29.10005538, Log Avg loss: 1.30240418, Global Avg Loss: 29.10005538, Time: 0.0111 Steps: 1530, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001540, Sample Num: 24640, Cur Loss: 2.09688044, Cur Avg Loss: 28.92502775, Log Avg loss: 2.14580069, Global Avg Loss: 28.92502775, Time: 0.0126 Steps: 1540, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001550, Sample Num: 24800, Cur Loss: 2.27352619, Cur Avg Loss: 28.74921221, Log Avg loss: 1.67361931, Global Avg Loss: 28.74921221, Time: 0.0074 Steps: 1550, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001560, Sample Num: 24960, Cur Loss: 0.97087371, Cur Avg Loss: 28.57399143, Log Avg loss: 1.41476937, Global Avg Loss: 28.57399143, Time: 0.0116 Steps: 1560, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001570, Sample Num: 25120, Cur Loss: 1.71783686, Cur Avg Loss: 28.40231177, Log Avg loss: 1.62028476, Global Avg Loss: 28.40231177, Time: 0.0064 Steps: 1570, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001580, Sample Num: 25280, Cur Loss: 1.55353606, Cur Avg Loss: 28.23222241, Log Avg loss: 1.52819342, Global Avg Loss: 28.23222241, Time: 0.0064 Steps: 1580, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001590, Sample Num: 25440, Cur Loss: 1.50997591, Cur Avg Loss: 28.06456077, Log Avg loss: 1.57402121, Global Avg Loss: 28.06456077, Time: 0.0065 Steps: 1590, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001600, Sample Num: 25600, Cur Loss: 1.37346196, Cur Avg Loss: 27.89807795, Log Avg loss: 1.42730945, Global Avg Loss: 27.89807795, Time: 0.0068 Steps: 1600, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001610, Sample Num: 25760, Cur Loss: 1.80176127, Cur Avg Loss: 27.73492117, Log Avg loss: 1.62983791, Global Avg Loss: 27.73492117, Time: 0.0067 Steps: 1610, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001620, Sample Num: 25920, Cur Loss: 1.24743652, Cur Avg Loss: 27.57385113, Log Avg loss: 1.64157373, Global Avg Loss: 27.57385113, Time: 0.0119 Steps: 1620, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001630, Sample Num: 26080, Cur Loss: 1.21752465, Cur Avg Loss: 27.41198616, Log Avg loss: 1.18986132, Global Avg Loss: 27.41198616, Time: 0.0119 Steps: 1630, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001640, Sample Num: 26240, Cur Loss: 2.59594870, Cur Avg Loss: 27.25413958, Log Avg loss: 1.52514766, Global Avg Loss: 27.25413958, Time: 0.0143 Steps: 1640, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001650, Sample Num: 26400, Cur Loss: 1.02093458, Cur Avg Loss: 27.09693135, Log Avg loss: 1.31478081, Global Avg Loss: 27.09693135, Time: 0.0129 Steps: 1650, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001660, Sample Num: 26560, Cur Loss: 2.08692884, Cur Avg Loss: 26.94474580, Log Avg loss: 1.83412962, Global Avg Loss: 26.94474580, Time: 0.0071 Steps: 1660, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001670, Sample Num: 26720, Cur Loss: 1.18558180, Cur Avg Loss: 26.79194541, Log Avg loss: 1.42708063, Global Avg Loss: 26.79194541, Time: 0.0121 Steps: 1670, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001680, Sample Num: 26880, Cur Loss: 0.68142331, Cur Avg Loss: 26.64229400, Log Avg loss: 1.65050983, Global Avg Loss: 26.64229400, Time: 0.0145 Steps: 1680, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001690, Sample Num: 27040, Cur Loss: 1.53842413, Cur Avg Loss: 26.49483235, Log Avg loss: 1.72127455, Global Avg Loss: 26.49483235, Time: 0.0169 Steps: 1690, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001700, Sample Num: 27200, Cur Loss: 1.39799607, Cur Avg Loss: 26.34937539, Log Avg loss: 1.76714818, Global Avg Loss: 26.34937539, Time: 0.0067 Steps: 1700, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001710, Sample Num: 27360, Cur Loss: 1.72717059, Cur Avg Loss: 26.20495078, Log Avg loss: 1.65276818, Global Avg Loss: 26.20495078, Time: 0.0184 Steps: 1710, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001720, Sample Num: 27520, Cur Loss: 0.89328766, Cur Avg Loss: 26.06062797, Log Avg loss: 1.38142646, Global Avg Loss: 26.06062797, Time: 0.0069 Steps: 1720, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001730, Sample Num: 27680, Cur Loss: 1.87614918, Cur Avg Loss: 25.92042366, Log Avg loss: 1.80528281, Global Avg Loss: 25.92042366, Time: 0.0072 Steps: 1730, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001740, Sample Num: 27840, Cur Loss: 1.22611260, Cur Avg Loss: 25.77922984, Log Avg loss: 1.35269895, Global Avg Loss: 25.77922984, Time: 0.0069 Steps: 1740, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001750, Sample Num: 28000, Cur Loss: 1.75985646, Cur Avg Loss: 25.64106807, Log Avg loss: 1.60091960, Global Avg Loss: 25.64106807, Time: 0.0074 Steps: 1750, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001760, Sample Num: 28160, Cur Loss: 2.05730891, Cur Avg Loss: 25.50507261, Log Avg loss: 1.70586784, Global Avg Loss: 25.50507261, Time: 0.0065 Steps: 1760, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001770, Sample Num: 28320, Cur Loss: 2.21754980, Cur Avg Loss: 25.37010143, Log Avg loss: 1.61517347, Global Avg Loss: 25.37010143, Time: 0.0218 Steps: 1770, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001780, Sample Num: 28480, Cur Loss: 1.58620393, Cur Avg Loss: 25.23568427, Log Avg loss: 1.44384815, Global Avg Loss: 25.23568427, Time: 0.0208 Steps: 1780, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001790, Sample Num: 28640, Cur Loss: 0.97943741, Cur Avg Loss: 25.10319143, Log Avg loss: 1.51946566, Global Avg Loss: 25.10319143, Time: 0.0068 Steps: 1790, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001800, Sample Num: 28800, Cur Loss: 2.26911163, Cur Avg Loss: 24.97121475, Log Avg loss: 1.34738796, Global Avg Loss: 24.97121475, Time: 0.0131 Steps: 1800, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001810, Sample Num: 28960, Cur Loss: 1.91402841, Cur Avg Loss: 24.84222765, Log Avg loss: 1.62455097, Global Avg Loss: 24.84222765, Time: 0.0117 Steps: 1810, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001820, Sample Num: 29120, Cur Loss: 1.13449454, Cur Avg Loss: 24.71409329, Log Avg loss: 1.52177379, Global Avg Loss: 24.71409329, Time: 0.0089 Steps: 1820, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001830, Sample Num: 29280, Cur Loss: 1.13655901, Cur Avg Loss: 24.58674848, Log Avg loss: 1.40999256, Global Avg Loss: 24.58674848, Time: 0.0105 Steps: 1830, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001840, Sample Num: 29440, Cur Loss: 1.54766333, Cur Avg Loss: 24.46180744, Log Avg loss: 1.59759760, Global Avg Loss: 24.46180744, Time: 0.0067 Steps: 1840, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001850, Sample Num: 29600, Cur Loss: 2.28504372, Cur Avg Loss: 24.33813321, Log Avg loss: 1.58207492, Global Avg Loss: 24.33813321, Time: 0.0140 Steps: 1850, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001860, Sample Num: 29760, Cur Loss: 2.01446724, Cur Avg Loss: 24.21540534, Log Avg loss: 1.51074805, Global Avg Loss: 24.21540534, Time: 0.0125 Steps: 1860, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001870, Sample Num: 29920, Cur Loss: 1.39669371, Cur Avg Loss: 24.09521183, Log Avg loss: 1.73921906, Global Avg Loss: 24.09521183, Time: 0.0127 Steps: 1870, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001880, Sample Num: 30080, Cur Loss: 1.95794010, Cur Avg Loss: 23.97478941, Log Avg loss: 1.45579768, Global Avg Loss: 23.97478941, Time: 0.0109 Steps: 1880, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001890, Sample Num: 30240, Cur Loss: 0.48287386, Cur Avg Loss: 23.85626714, Log Avg loss: 1.57408028, Global Avg Loss: 23.85626714, Time: 0.0120 Steps: 1890, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001900, Sample Num: 30400, Cur Loss: 1.53405237, Cur Avg Loss: 23.73810473, Log Avg loss: 1.40540832, Global Avg Loss: 23.73810473, Time: 0.0064 Steps: 1900, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001910, Sample Num: 30560, Cur Loss: 1.61090875, Cur Avg Loss: 23.62272051, Log Avg loss: 1.69971914, Global Avg Loss: 23.62272051, Time: 0.0120 Steps: 1910, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001920, Sample Num: 30720, Cur Loss: 1.13722908, Cur Avg Loss: 23.50951938, Log Avg loss: 1.88810325, Global Avg Loss: 23.50951938, Time: 0.0120 Steps: 1920, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001930, Sample Num: 30880, Cur Loss: 0.81878972, Cur Avg Loss: 23.39633438, Log Avg loss: 1.66481583, Global Avg Loss: 23.39633438, Time: 0.0144 Steps: 1930, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001940, Sample Num: 31040, Cur Loss: 0.80865538, Cur Avg Loss: 23.28331262, Log Avg loss: 1.47011321, Global Avg Loss: 23.28331262, Time: 0.0124 Steps: 1940, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001950, Sample Num: 31200, Cur Loss: 1.37271261, Cur Avg Loss: 23.17199926, Log Avg loss: 1.57720701, Global Avg Loss: 23.17199926, Time: 0.0096 Steps: 1950, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001960, Sample Num: 31360, Cur Loss: 1.63878489, Cur Avg Loss: 23.06305780, Log Avg loss: 1.81947204, Global Avg Loss: 23.06305780, Time: 0.0114 Steps: 1960, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001970, Sample Num: 31520, Cur Loss: 1.74422741, Cur Avg Loss: 22.95426082, Log Avg loss: 1.63005234, Global Avg Loss: 22.95426082, Time: 0.0066 Steps: 1970, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001980, Sample Num: 31680, Cur Loss: 1.93018675, Cur Avg Loss: 22.84713795, Log Avg loss: 1.74393382, Global Avg Loss: 22.84713795, Time: 0.0123 Steps: 1980, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 001990, Sample Num: 31840, Cur Loss: 1.01272917, Cur Avg Loss: 22.74081717, Log Avg loss: 1.68930278, Global Avg Loss: 22.74081717, Time: 0.0118 Steps: 1990, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002000, Sample Num: 32000, Cur Loss: 1.03745890, Cur Avg Loss: 22.63403214, Log Avg loss: 1.38381116, Global Avg Loss: 22.63403214, Time: 0.0131 Steps: 2000, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002010, Sample Num: 32160, Cur Loss: 1.61945736, Cur Avg Loss: 22.52935386, Log Avg loss: 1.59369712, Global Avg Loss: 22.52935386, Time: 0.0068 Steps: 2010, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002020, Sample Num: 32320, Cur Loss: 1.86752677, Cur Avg Loss: 22.42581730, Log Avg loss: 1.61496841, Global Avg Loss: 22.42581730, Time: 0.0069 Steps: 2020, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002030, Sample Num: 32480, Cur Loss: 1.21249092, Cur Avg Loss: 22.32260720, Log Avg loss: 1.47416764, Global Avg Loss: 22.32260720, Time: 0.0109 Steps: 2030, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002040, Sample Num: 32640, Cur Loss: 1.84272194, Cur Avg Loss: 22.22152771, Log Avg loss: 1.70239116, Global Avg Loss: 22.22152771, Time: 0.0108 Steps: 2040, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002050, Sample Num: 32800, Cur Loss: 2.34695911, Cur Avg Loss: 22.12124800, Log Avg loss: 1.66418704, Global Avg Loss: 22.12124800, Time: 0.0121 Steps: 2050, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002060, Sample Num: 32960, Cur Loss: 2.12054157, Cur Avg Loss: 22.02191110, Log Avg loss: 1.65784686, Global Avg Loss: 22.02191110, Time: 0.0203 Steps: 2060, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002070, Sample Num: 33120, Cur Loss: 1.69632018, Cur Avg Loss: 21.92344507, Log Avg loss: 1.63944240, Global Avg Loss: 21.92344507, Time: 0.0120 Steps: 2070, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002080, Sample Num: 33280, Cur Loss: 1.46666515, Cur Avg Loss: 21.82427389, Log Avg loss: 1.29583953, Global Avg Loss: 21.82427389, Time: 0.0217 Steps: 2080, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002090, Sample Num: 33440, Cur Loss: 1.68712866, Cur Avg Loss: 21.72614185, Log Avg loss: 1.31467850, Global Avg Loss: 21.72614185, Time: 0.0111 Steps: 2090, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002100, Sample Num: 33600, Cur Loss: 1.32057548, Cur Avg Loss: 21.62999212, Log Avg loss: 1.53469893, Global Avg Loss: 21.62999212, Time: 0.0123 Steps: 2100, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002110, Sample Num: 33760, Cur Loss: 1.23417199, Cur Avg Loss: 21.53413366, Log Avg loss: 1.40385603, Global Avg Loss: 21.53413366, Time: 0.0123 Steps: 2110, Updated lr: 0.000099 Training, Epoch: 0001, Batch: 002120, Sample Num: 33920, Cur Loss: 1.43903518, Cur Avg Loss: 21.44122177, Log Avg loss: 1.83681295, Global Avg Loss: 21.44122177, Time: 0.0067 Steps: 2120, Updated lr: 0.000099 ***** Running evaluation checkpoint-2129 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-2129 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 20.326639, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.513437, "eval_total_loss": 1063.946111, "eval_mae": 1.059972, "eval_mse": 1.513558, "eval_r2": 0.037883, "eval_sp_statistic": 0.25343, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.233571, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 2.152865, "test_total_loss": 1080.738426, "test_mae": 1.392895, "test_mse": 2.152537, "test_r2": -0.389266, "test_sp_statistic": 0.045012, "test_sp_pvalue": 5.5e-05, "test_ps_statistic": 0.147422, "test_ps_pvalue": 0.0, "lr": 9.892935040303462e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 21.357291536607782, "train_cur_epoch_loss": 45469.67368143797, "train_cur_epoch_avg_loss": 21.357291536607782, "train_cur_epoch_time": 20.32663869857788, "train_cur_epoch_avg_time": 0.009547505260017793, "epoch": 1, "step": 2129} ################################################## Training, Epoch: 0002, Batch: 000001, Sample Num: 16, Cur Loss: 1.16169107, Cur Avg Loss: 1.16169107, Log Avg loss: 1.54452231, Global Avg Loss: 21.34781003, Time: 0.0159 Steps: 2130, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000011, Sample Num: 176, Cur Loss: 1.74199164, Cur Avg Loss: 1.49025402, Log Avg loss: 1.52311032, Global Avg Loss: 21.25517125, Time: 0.0072 Steps: 2140, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000021, Sample Num: 336, Cur Loss: 3.43195939, Cur Avg Loss: 1.44697674, Log Avg loss: 1.39937173, Global Avg Loss: 21.16281869, Time: 0.0067 Steps: 2150, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000031, Sample Num: 496, Cur Loss: 2.19679117, Cur Avg Loss: 1.51851189, Log Avg loss: 1.66873572, Global Avg Loss: 21.07256831, Time: 0.0122 Steps: 2160, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000041, Sample Num: 656, Cur Loss: 1.17204988, Cur Avg Loss: 1.59111781, Log Avg loss: 1.81619616, Global Avg Loss: 20.98382927, Time: 0.0235 Steps: 2170, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000051, Sample Num: 816, Cur Loss: 0.52828079, Cur Avg Loss: 1.52872848, Log Avg loss: 1.27293222, Global Avg Loss: 20.89341231, Time: 0.0137 Steps: 2180, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000061, Sample Num: 976, Cur Loss: 1.28126216, Cur Avg Loss: 1.48736018, Log Avg loss: 1.27638184, Global Avg Loss: 20.80383683, Time: 0.0133 Steps: 2190, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000071, Sample Num: 1136, Cur Loss: 1.30215752, Cur Avg Loss: 1.51647875, Log Avg loss: 1.69410204, Global Avg Loss: 20.71697440, Time: 0.0137 Steps: 2200, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000081, Sample Num: 1296, Cur Loss: 1.80179811, Cur Avg Loss: 1.50323397, Log Avg loss: 1.40919603, Global Avg Loss: 20.62960888, Time: 0.0136 Steps: 2210, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000091, Sample Num: 1456, Cur Loss: 2.66614437, Cur Avg Loss: 1.53704012, Log Avg loss: 1.81086992, Global Avg Loss: 20.54483979, Time: 0.0111 Steps: 2220, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000101, Sample Num: 1616, Cur Loss: 2.77600193, Cur Avg Loss: 1.57219334, Log Avg loss: 1.89208764, Global Avg Loss: 20.46119516, Time: 0.0067 Steps: 2230, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000111, Sample Num: 1776, Cur Loss: 1.88220823, Cur Avg Loss: 1.58621639, Log Avg loss: 1.72784926, Global Avg Loss: 20.37756415, Time: 0.0067 Steps: 2240, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000121, Sample Num: 1936, Cur Loss: 1.01657271, Cur Avg Loss: 1.56524080, Log Avg loss: 1.33241177, Global Avg Loss: 20.29291903, Time: 0.0070 Steps: 2250, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000131, Sample Num: 2096, Cur Loss: 1.64618373, Cur Avg Loss: 1.56415355, Log Avg loss: 1.55099778, Global Avg Loss: 20.20999018, Time: 0.0159 Steps: 2260, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000141, Sample Num: 2256, Cur Loss: 1.57606852, Cur Avg Loss: 1.55282556, Log Avg loss: 1.40442885, Global Avg Loss: 20.12714629, Time: 0.0068 Steps: 2270, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000151, Sample Num: 2416, Cur Loss: 1.46946168, Cur Avg Loss: 1.55977064, Log Avg loss: 1.65769626, Global Avg Loss: 20.04613993, Time: 0.0067 Steps: 2280, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000161, Sample Num: 2576, Cur Loss: 2.45986176, Cur Avg Loss: 1.57649274, Log Avg loss: 1.82899644, Global Avg Loss: 19.96658909, Time: 0.0107 Steps: 2290, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000171, Sample Num: 2736, Cur Loss: 2.19410372, Cur Avg Loss: 1.58561953, Log Avg loss: 1.73256095, Global Avg Loss: 19.88731071, Time: 0.0067 Steps: 2300, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000181, Sample Num: 2896, Cur Loss: 1.00882590, Cur Avg Loss: 1.57369556, Log Avg loss: 1.36979572, Global Avg Loss: 19.80714830, Time: 0.0077 Steps: 2310, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000191, Sample Num: 3056, Cur Loss: 1.68590283, Cur Avg Loss: 1.56784735, Log Avg loss: 1.46199465, Global Avg Loss: 19.72807436, Time: 0.0069 Steps: 2320, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000201, Sample Num: 3216, Cur Loss: 2.64786530, Cur Avg Loss: 1.56806753, Log Avg loss: 1.57227296, Global Avg Loss: 19.65015247, Time: 0.0104 Steps: 2330, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000211, Sample Num: 3376, Cur Loss: 1.75331974, Cur Avg Loss: 1.55069622, Log Avg loss: 1.20153283, Global Avg Loss: 19.57131221, Time: 0.0067 Steps: 2340, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000221, Sample Num: 3536, Cur Loss: 0.90351659, Cur Avg Loss: 1.54508179, Log Avg loss: 1.42661740, Global Avg Loss: 19.49410075, Time: 0.0232 Steps: 2350, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000231, Sample Num: 3696, Cur Loss: 0.42284131, Cur Avg Loss: 1.54886868, Log Avg loss: 1.63255899, Global Avg Loss: 19.41841625, Time: 0.0067 Steps: 2360, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000241, Sample Num: 3856, Cur Loss: 2.09008527, Cur Avg Loss: 1.54944717, Log Avg loss: 1.56281016, Global Avg Loss: 19.34307614, Time: 0.0096 Steps: 2370, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000251, Sample Num: 4016, Cur Loss: 1.03776741, Cur Avg Loss: 1.54554015, Log Avg loss: 1.45138116, Global Avg Loss: 19.26790095, Time: 0.0090 Steps: 2380, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000261, Sample Num: 4176, Cur Loss: 1.64486337, Cur Avg Loss: 1.55034358, Log Avg loss: 1.67090967, Global Avg Loss: 19.19427337, Time: 0.0087 Steps: 2390, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000271, Sample Num: 4336, Cur Loss: 1.95445061, Cur Avg Loss: 1.55174154, Log Avg loss: 1.58822832, Global Avg Loss: 19.12091485, Time: 0.0113 Steps: 2400, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000281, Sample Num: 4496, Cur Loss: 1.12346005, Cur Avg Loss: 1.55003176, Log Avg loss: 1.50369647, Global Avg Loss: 19.04781436, Time: 0.0069 Steps: 2410, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000291, Sample Num: 4656, Cur Loss: 3.20257664, Cur Avg Loss: 1.56856428, Log Avg loss: 2.08932819, Global Avg Loss: 18.97773797, Time: 0.0084 Steps: 2420, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000301, Sample Num: 4816, Cur Loss: 0.84497458, Cur Avg Loss: 1.56373136, Log Avg loss: 1.42309335, Global Avg Loss: 18.90549663, Time: 0.0117 Steps: 2430, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000311, Sample Num: 4976, Cur Loss: 1.65017021, Cur Avg Loss: 1.56331816, Log Avg loss: 1.55088092, Global Avg Loss: 18.83437116, Time: 0.0069 Steps: 2440, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000321, Sample Num: 5136, Cur Loss: 1.42956531, Cur Avg Loss: 1.56542167, Log Avg loss: 1.63084097, Global Avg Loss: 18.76415267, Time: 0.0068 Steps: 2450, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000331, Sample Num: 5296, Cur Loss: 3.03163815, Cur Avg Loss: 1.57742981, Log Avg loss: 1.96289095, Global Avg Loss: 18.69585486, Time: 0.0223 Steps: 2460, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000341, Sample Num: 5456, Cur Loss: 1.44182432, Cur Avg Loss: 1.57041214, Log Avg loss: 1.33812739, Global Avg Loss: 18.62558066, Time: 0.0068 Steps: 2470, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000351, Sample Num: 5616, Cur Loss: 1.10404778, Cur Avg Loss: 1.57525845, Log Avg loss: 1.74051756, Global Avg Loss: 18.55749572, Time: 0.0120 Steps: 2480, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000361, Sample Num: 5776, Cur Loss: 1.09605896, Cur Avg Loss: 1.56480317, Log Avg loss: 1.19782289, Global Avg Loss: 18.48777816, Time: 0.0116 Steps: 2490, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000371, Sample Num: 5936, Cur Loss: 1.24504399, Cur Avg Loss: 1.56973752, Log Avg loss: 1.74786733, Global Avg Loss: 18.42081852, Time: 0.0131 Steps: 2500, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000381, Sample Num: 6096, Cur Loss: 1.02374840, Cur Avg Loss: 1.56523905, Log Avg loss: 1.39834587, Global Avg Loss: 18.35299990, Time: 0.0071 Steps: 2510, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000391, Sample Num: 6256, Cur Loss: 1.94208205, Cur Avg Loss: 1.56378770, Log Avg loss: 1.50849119, Global Avg Loss: 18.28615662, Time: 0.0065 Steps: 2520, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000401, Sample Num: 6416, Cur Loss: 1.09348893, Cur Avg Loss: 1.55797147, Log Avg loss: 1.33055682, Global Avg Loss: 18.21913843, Time: 0.0065 Steps: 2530, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000411, Sample Num: 6576, Cur Loss: 1.89883828, Cur Avg Loss: 1.55616702, Log Avg loss: 1.48380895, Global Avg Loss: 18.15325131, Time: 0.0219 Steps: 2540, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000421, Sample Num: 6736, Cur Loss: 0.84700662, Cur Avg Loss: 1.55743962, Log Avg loss: 1.60974348, Global Avg Loss: 18.08837481, Time: 0.0227 Steps: 2550, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000431, Sample Num: 6896, Cur Loss: 1.26991820, Cur Avg Loss: 1.55527154, Log Avg loss: 1.46399501, Global Avg Loss: 18.02343583, Time: 0.0140 Steps: 2560, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000441, Sample Num: 7056, Cur Loss: 2.15124512, Cur Avg Loss: 1.55937605, Log Avg loss: 1.73628076, Global Avg Loss: 17.96006168, Time: 0.0072 Steps: 2570, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000451, Sample Num: 7216, Cur Loss: 1.67734706, Cur Avg Loss: 1.54857407, Log Avg loss: 1.07220677, Global Avg Loss: 17.89460488, Time: 0.0170 Steps: 2580, Updated lr: 0.000099 Training, Epoch: 0002, Batch: 000461, Sample Num: 7376, Cur Loss: 0.42334503, Cur Avg Loss: 1.54292127, Log Avg loss: 1.28798002, Global Avg Loss: 17.83048664, Time: 0.0235 Steps: 2590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000471, Sample Num: 7536, Cur Loss: 1.56110930, Cur Avg Loss: 1.53792395, Log Avg loss: 1.30754752, Global Avg Loss: 17.76693687, Time: 0.0207 Steps: 2600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000481, Sample Num: 7696, Cur Loss: 2.50980926, Cur Avg Loss: 1.53626267, Log Avg loss: 1.45801617, Global Avg Loss: 17.70445058, Time: 0.0071 Steps: 2610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000491, Sample Num: 7856, Cur Loss: 1.12932873, Cur Avg Loss: 1.53475820, Log Avg loss: 1.46239337, Global Avg Loss: 17.64245800, Time: 0.0073 Steps: 2620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000501, Sample Num: 8016, Cur Loss: 1.39824116, Cur Avg Loss: 1.53454664, Log Avg loss: 1.52415903, Global Avg Loss: 17.58117169, Time: 0.0112 Steps: 2630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000511, Sample Num: 8176, Cur Loss: 1.11809683, Cur Avg Loss: 1.53032124, Log Avg loss: 1.31862843, Global Avg Loss: 17.51957115, Time: 0.0065 Steps: 2640, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000521, Sample Num: 8336, Cur Loss: 1.99765623, Cur Avg Loss: 1.53040021, Log Avg loss: 1.53443567, Global Avg Loss: 17.45924988, Time: 0.0068 Steps: 2650, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000531, Sample Num: 8496, Cur Loss: 2.62437654, Cur Avg Loss: 1.52504691, Log Avg loss: 1.24613994, Global Avg Loss: 17.39829834, Time: 0.0121 Steps: 2660, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000541, Sample Num: 8656, Cur Loss: 1.05657637, Cur Avg Loss: 1.52746492, Log Avg loss: 1.65586135, Global Avg Loss: 17.33933790, Time: 0.0072 Steps: 2670, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000551, Sample Num: 8816, Cur Loss: 1.07449019, Cur Avg Loss: 1.52785932, Log Avg loss: 1.54919659, Global Avg Loss: 17.28041947, Time: 0.0067 Steps: 2680, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000561, Sample Num: 8976, Cur Loss: 2.83372068, Cur Avg Loss: 1.52863228, Log Avg loss: 1.57122195, Global Avg Loss: 17.22202096, Time: 0.0072 Steps: 2690, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000571, Sample Num: 9136, Cur Loss: 1.19091868, Cur Avg Loss: 1.52936901, Log Avg loss: 1.57069967, Global Avg Loss: 17.16405311, Time: 0.0109 Steps: 2700, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000581, Sample Num: 9296, Cur Loss: 0.67489243, Cur Avg Loss: 1.52584495, Log Avg loss: 1.32462136, Global Avg Loss: 17.10560502, Time: 0.0122 Steps: 2710, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000591, Sample Num: 9456, Cur Loss: 1.86093700, Cur Avg Loss: 1.51989243, Log Avg loss: 1.17405066, Global Avg Loss: 17.04703313, Time: 0.0066 Steps: 2720, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000601, Sample Num: 9616, Cur Loss: 1.77736032, Cur Avg Loss: 1.52071419, Log Avg loss: 1.56928016, Global Avg Loss: 16.99033806, Time: 0.0112 Steps: 2730, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000611, Sample Num: 9776, Cur Loss: 1.07609618, Cur Avg Loss: 1.51599212, Log Avg loss: 1.23219575, Global Avg Loss: 16.93282659, Time: 0.0067 Steps: 2740, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000621, Sample Num: 9936, Cur Loss: 1.10781407, Cur Avg Loss: 1.51411562, Log Avg loss: 1.39946185, Global Avg Loss: 16.87634163, Time: 0.0158 Steps: 2750, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000631, Sample Num: 10096, Cur Loss: 2.09858990, Cur Avg Loss: 1.51955232, Log Avg loss: 1.85717137, Global Avg Loss: 16.82192435, Time: 0.0064 Steps: 2760, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000641, Sample Num: 10256, Cur Loss: 0.40450454, Cur Avg Loss: 1.52106966, Log Avg loss: 1.61681353, Global Avg Loss: 16.76703225, Time: 0.0124 Steps: 2770, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000651, Sample Num: 10416, Cur Loss: 2.02944517, Cur Avg Loss: 1.52002993, Log Avg loss: 1.45338336, Global Avg Loss: 16.71194718, Time: 0.0145 Steps: 2780, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000661, Sample Num: 10576, Cur Loss: 1.11039996, Cur Avg Loss: 1.51933924, Log Avg loss: 1.47437545, Global Avg Loss: 16.65733223, Time: 0.0065 Steps: 2790, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000671, Sample Num: 10736, Cur Loss: 2.27988291, Cur Avg Loss: 1.52215910, Log Avg loss: 1.70855157, Global Avg Loss: 16.60394373, Time: 0.0073 Steps: 2800, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000681, Sample Num: 10896, Cur Loss: 2.19681621, Cur Avg Loss: 1.52282249, Log Avg loss: 1.56733629, Global Avg Loss: 16.55043267, Time: 0.0067 Steps: 2810, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000691, Sample Num: 11056, Cur Loss: 1.19008982, Cur Avg Loss: 1.52212630, Log Avg loss: 1.47471524, Global Avg Loss: 16.49697268, Time: 0.0117 Steps: 2820, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000701, Sample Num: 11216, Cur Loss: 2.01841164, Cur Avg Loss: 1.52297025, Log Avg loss: 1.58128750, Global Avg Loss: 16.44426708, Time: 0.0138 Steps: 2830, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000711, Sample Num: 11376, Cur Loss: 0.83663833, Cur Avg Loss: 1.51982901, Log Avg loss: 1.29962838, Global Avg Loss: 16.39094088, Time: 0.0121 Steps: 2840, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000721, Sample Num: 11536, Cur Loss: 1.52566707, Cur Avg Loss: 1.51767976, Log Avg loss: 1.36486805, Global Avg Loss: 16.33821782, Time: 0.0072 Steps: 2850, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000731, Sample Num: 11696, Cur Loss: 1.49418032, Cur Avg Loss: 1.51497524, Log Avg loss: 1.31997908, Global Avg Loss: 16.28570650, Time: 0.0117 Steps: 2860, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000741, Sample Num: 11856, Cur Loss: 1.25390744, Cur Avg Loss: 1.51085894, Log Avg loss: 1.20995779, Global Avg Loss: 16.23317776, Time: 0.0066 Steps: 2870, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000751, Sample Num: 12016, Cur Loss: 1.37615514, Cur Avg Loss: 1.51018371, Log Avg loss: 1.46014867, Global Avg Loss: 16.18188252, Time: 0.0109 Steps: 2880, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000761, Sample Num: 12176, Cur Loss: 0.84169734, Cur Avg Loss: 1.50288370, Log Avg loss: 0.95465279, Global Avg Loss: 16.12919314, Time: 0.0113 Steps: 2890, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000771, Sample Num: 12336, Cur Loss: 1.58617640, Cur Avg Loss: 1.49988404, Log Avg loss: 1.27160995, Global Avg Loss: 16.07796009, Time: 0.0065 Steps: 2900, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000781, Sample Num: 12496, Cur Loss: 1.71765280, Cur Avg Loss: 1.50640159, Log Avg loss: 2.00890473, Global Avg Loss: 16.02961283, Time: 0.0106 Steps: 2910, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000791, Sample Num: 12656, Cur Loss: 2.04682493, Cur Avg Loss: 1.50764231, Log Avg loss: 1.60454264, Global Avg Loss: 15.98021190, Time: 0.0141 Steps: 2920, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000801, Sample Num: 12816, Cur Loss: 1.23203051, Cur Avg Loss: 1.51341440, Log Avg loss: 1.96998710, Global Avg Loss: 15.93239543, Time: 0.0113 Steps: 2930, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000811, Sample Num: 12976, Cur Loss: 1.88916075, Cur Avg Loss: 1.51362033, Log Avg loss: 1.53011532, Global Avg Loss: 15.88340809, Time: 0.0143 Steps: 2940, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000821, Sample Num: 13136, Cur Loss: 1.49686837, Cur Avg Loss: 1.51556322, Log Avg loss: 1.67313153, Global Avg Loss: 15.83523766, Time: 0.0106 Steps: 2950, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000831, Sample Num: 13296, Cur Loss: 2.27585268, Cur Avg Loss: 1.51537825, Log Avg loss: 1.50019202, Global Avg Loss: 15.78680845, Time: 0.0119 Steps: 2960, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000841, Sample Num: 13456, Cur Loss: 1.67952299, Cur Avg Loss: 1.51343888, Log Avg loss: 1.35227699, Global Avg Loss: 15.73820733, Time: 0.0118 Steps: 2970, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000851, Sample Num: 13616, Cur Loss: 1.63894176, Cur Avg Loss: 1.51163687, Log Avg loss: 1.36008816, Global Avg Loss: 15.68995861, Time: 0.0108 Steps: 2980, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000861, Sample Num: 13776, Cur Loss: 1.39752018, Cur Avg Loss: 1.51211087, Log Avg loss: 1.55244845, Global Avg Loss: 15.64267597, Time: 0.0069 Steps: 2990, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000871, Sample Num: 13936, Cur Loss: 1.03307056, Cur Avg Loss: 1.50659773, Log Avg loss: 1.03191652, Global Avg Loss: 15.59397344, Time: 0.0136 Steps: 3000, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000881, Sample Num: 14096, Cur Loss: 1.47970819, Cur Avg Loss: 1.50827195, Log Avg loss: 1.65409589, Global Avg Loss: 15.54766155, Time: 0.0123 Steps: 3010, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000891, Sample Num: 14256, Cur Loss: 1.65678000, Cur Avg Loss: 1.50768393, Log Avg loss: 1.45587952, Global Avg Loss: 15.50100002, Time: 0.0106 Steps: 3020, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000901, Sample Num: 14416, Cur Loss: 1.03388214, Cur Avg Loss: 1.50634523, Log Avg loss: 1.38706706, Global Avg Loss: 15.45441938, Time: 0.0121 Steps: 3030, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000911, Sample Num: 14576, Cur Loss: 1.11741972, Cur Avg Loss: 1.50787510, Log Avg loss: 1.64571661, Global Avg Loss: 15.40899602, Time: 0.0118 Steps: 3040, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000921, Sample Num: 14736, Cur Loss: 2.27829814, Cur Avg Loss: 1.51231445, Log Avg loss: 1.91673890, Global Avg Loss: 15.36475911, Time: 0.0084 Steps: 3050, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000931, Sample Num: 14896, Cur Loss: 2.75292635, Cur Avg Loss: 1.51007886, Log Avg loss: 1.30418146, Global Avg Loss: 15.31880951, Time: 0.0119 Steps: 3060, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000941, Sample Num: 15056, Cur Loss: 1.02621281, Cur Avg Loss: 1.51004298, Log Avg loss: 1.50670207, Global Avg Loss: 15.27381893, Time: 0.0183 Steps: 3070, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000951, Sample Num: 15216, Cur Loss: 1.23094845, Cur Avg Loss: 1.51425836, Log Avg loss: 1.91092567, Global Avg Loss: 15.23043292, Time: 0.0072 Steps: 3080, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000961, Sample Num: 15376, Cur Loss: 1.58257473, Cur Avg Loss: 1.51531668, Log Avg loss: 1.61596324, Global Avg Loss: 15.18637314, Time: 0.0115 Steps: 3090, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000971, Sample Num: 15536, Cur Loss: 1.63088989, Cur Avg Loss: 1.51526234, Log Avg loss: 1.51004008, Global Avg Loss: 15.14225594, Time: 0.0107 Steps: 3100, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000981, Sample Num: 15696, Cur Loss: 0.82255828, Cur Avg Loss: 1.51539049, Log Avg loss: 1.52783371, Global Avg Loss: 15.09847966, Time: 0.0070 Steps: 3110, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 000991, Sample Num: 15856, Cur Loss: 1.01347280, Cur Avg Loss: 1.51288546, Log Avg loss: 1.26714237, Global Avg Loss: 15.05414845, Time: 0.0120 Steps: 3120, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001001, Sample Num: 16016, Cur Loss: 0.93612772, Cur Avg Loss: 1.50929528, Log Avg loss: 1.15350838, Global Avg Loss: 15.00973746, Time: 0.0112 Steps: 3130, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001011, Sample Num: 16176, Cur Loss: 2.01879811, Cur Avg Loss: 1.51150665, Log Avg loss: 1.73286439, Global Avg Loss: 14.96745443, Time: 0.0118 Steps: 3140, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001021, Sample Num: 16336, Cur Loss: 0.94065166, Cur Avg Loss: 1.51158697, Log Avg loss: 1.51970707, Global Avg Loss: 14.92476317, Time: 0.0066 Steps: 3150, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001031, Sample Num: 16496, Cur Loss: 1.30035043, Cur Avg Loss: 1.51277887, Log Avg loss: 1.63447268, Global Avg Loss: 14.88270528, Time: 0.0124 Steps: 3160, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001041, Sample Num: 16656, Cur Loss: 1.67225492, Cur Avg Loss: 1.51217536, Log Avg loss: 1.44995344, Global Avg Loss: 14.84033067, Time: 0.0115 Steps: 3170, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001051, Sample Num: 16816, Cur Loss: 1.18253732, Cur Avg Loss: 1.51413643, Log Avg loss: 1.71828333, Global Avg Loss: 14.79906637, Time: 0.0135 Steps: 3180, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001061, Sample Num: 16976, Cur Loss: 0.93714720, Cur Avg Loss: 1.51411433, Log Avg loss: 1.51179218, Global Avg Loss: 14.75741348, Time: 0.0142 Steps: 3190, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001071, Sample Num: 17136, Cur Loss: 2.31666946, Cur Avg Loss: 1.51361169, Log Avg loss: 1.46028105, Global Avg Loss: 14.71585994, Time: 0.0154 Steps: 3200, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001081, Sample Num: 17296, Cur Loss: 1.70847511, Cur Avg Loss: 1.51504387, Log Avg loss: 1.66843091, Global Avg Loss: 14.67521374, Time: 0.0119 Steps: 3210, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001091, Sample Num: 17456, Cur Loss: 1.62092435, Cur Avg Loss: 1.51564217, Log Avg loss: 1.58031836, Global Avg Loss: 14.63454636, Time: 0.0067 Steps: 3220, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001101, Sample Num: 17616, Cur Loss: 1.34205461, Cur Avg Loss: 1.51463772, Log Avg loss: 1.40505233, Global Avg Loss: 14.59358818, Time: 0.0135 Steps: 3230, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001111, Sample Num: 17776, Cur Loss: 1.26085913, Cur Avg Loss: 1.51343842, Log Avg loss: 1.38139474, Global Avg Loss: 14.55280980, Time: 0.0073 Steps: 3240, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001121, Sample Num: 17936, Cur Loss: 1.25278449, Cur Avg Loss: 1.51276323, Log Avg loss: 1.43774984, Global Avg Loss: 14.51245577, Time: 0.0123 Steps: 3250, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001131, Sample Num: 18096, Cur Loss: 0.95879376, Cur Avg Loss: 1.51239843, Log Avg loss: 1.47150441, Global Avg Loss: 14.47245285, Time: 0.0066 Steps: 3260, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001141, Sample Num: 18256, Cur Loss: 1.53202105, Cur Avg Loss: 1.51067235, Log Avg loss: 1.31545305, Global Avg Loss: 14.43221738, Time: 0.0067 Steps: 3270, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001151, Sample Num: 18416, Cur Loss: 1.59432864, Cur Avg Loss: 1.51061661, Log Avg loss: 1.50425615, Global Avg Loss: 14.39280287, Time: 0.0084 Steps: 3280, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001161, Sample Num: 18576, Cur Loss: 1.40067279, Cur Avg Loss: 1.51240342, Log Avg loss: 1.71806545, Global Avg Loss: 14.35427783, Time: 0.0117 Steps: 3290, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001171, Sample Num: 18736, Cur Loss: 1.17621195, Cur Avg Loss: 1.51287319, Log Avg loss: 1.56741390, Global Avg Loss: 14.31552975, Time: 0.0156 Steps: 3300, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001181, Sample Num: 18896, Cur Loss: 1.85396600, Cur Avg Loss: 1.51314791, Log Avg loss: 1.54531743, Global Avg Loss: 14.27694905, Time: 0.0151 Steps: 3310, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001191, Sample Num: 19056, Cur Loss: 1.46609294, Cur Avg Loss: 1.51371593, Log Avg loss: 1.58079898, Global Avg Loss: 14.23870764, Time: 0.0161 Steps: 3320, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001201, Sample Num: 19216, Cur Loss: 2.21873736, Cur Avg Loss: 1.51313357, Log Avg loss: 1.44377456, Global Avg Loss: 14.20028441, Time: 0.0067 Steps: 3330, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001211, Sample Num: 19376, Cur Loss: 3.30505085, Cur Avg Loss: 1.51132938, Log Avg loss: 1.29464547, Global Avg Loss: 14.16164478, Time: 0.0097 Steps: 3340, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001221, Sample Num: 19536, Cur Loss: 0.83262724, Cur Avg Loss: 1.51325043, Log Avg loss: 1.74589009, Global Avg Loss: 14.12458282, Time: 0.0121 Steps: 3350, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001231, Sample Num: 19696, Cur Loss: 0.72543103, Cur Avg Loss: 1.51228229, Log Avg loss: 1.39407225, Global Avg Loss: 14.08669440, Time: 0.0158 Steps: 3360, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001241, Sample Num: 19856, Cur Loss: 2.52330804, Cur Avg Loss: 1.51050103, Log Avg loss: 1.29122765, Global Avg Loss: 14.04872565, Time: 0.0072 Steps: 3370, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001251, Sample Num: 20016, Cur Loss: 1.16668200, Cur Avg Loss: 1.51065739, Log Avg loss: 1.53006174, Global Avg Loss: 14.01168819, Time: 0.0118 Steps: 3380, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001261, Sample Num: 20176, Cur Loss: 1.74197793, Cur Avg Loss: 1.51290983, Log Avg loss: 1.79469043, Global Avg Loss: 13.97564985, Time: 0.0116 Steps: 3390, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001271, Sample Num: 20336, Cur Loss: 1.96417618, Cur Avg Loss: 1.51069363, Log Avg loss: 1.23123140, Global Avg Loss: 13.93816626, Time: 0.0094 Steps: 3400, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001281, Sample Num: 20496, Cur Loss: 2.31189823, Cur Avg Loss: 1.50992661, Log Avg loss: 1.41243839, Global Avg Loss: 13.90143392, Time: 0.0215 Steps: 3410, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001291, Sample Num: 20656, Cur Loss: 2.43306994, Cur Avg Loss: 1.51073828, Log Avg loss: 1.61471281, Global Avg Loss: 13.86550784, Time: 0.0066 Steps: 3420, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001301, Sample Num: 20816, Cur Loss: 1.17125392, Cur Avg Loss: 1.51248048, Log Avg loss: 1.73739887, Global Avg Loss: 13.83014892, Time: 0.0082 Steps: 3430, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001311, Sample Num: 20976, Cur Loss: 2.06909418, Cur Avg Loss: 1.51302929, Log Avg loss: 1.58442864, Global Avg Loss: 13.79455089, Time: 0.0097 Steps: 3440, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001321, Sample Num: 21136, Cur Loss: 1.16040587, Cur Avg Loss: 1.51325033, Log Avg loss: 1.54222946, Global Avg Loss: 13.75903692, Time: 0.0072 Steps: 3450, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001331, Sample Num: 21296, Cur Loss: 0.73066938, Cur Avg Loss: 1.51225344, Log Avg loss: 1.38056364, Global Avg Loss: 13.72326099, Time: 0.0114 Steps: 3460, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001341, Sample Num: 21456, Cur Loss: 1.71943653, Cur Avg Loss: 1.51253057, Log Avg loss: 1.54941638, Global Avg Loss: 13.68817786, Time: 0.0105 Steps: 3470, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001351, Sample Num: 21616, Cur Loss: 0.94442534, Cur Avg Loss: 1.51122549, Log Avg loss: 1.33621510, Global Avg Loss: 13.65268371, Time: 0.0066 Steps: 3480, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001361, Sample Num: 21776, Cur Loss: 1.68536198, Cur Avg Loss: 1.50906130, Log Avg loss: 1.21667843, Global Avg Loss: 13.61705046, Time: 0.0067 Steps: 3490, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001371, Sample Num: 21936, Cur Loss: 2.79774475, Cur Avg Loss: 1.51221781, Log Avg loss: 1.94181839, Global Avg Loss: 13.58369265, Time: 0.0067 Steps: 3500, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001381, Sample Num: 22096, Cur Loss: 3.17990923, Cur Avg Loss: 1.51310087, Log Avg loss: 1.63416846, Global Avg Loss: 13.54964843, Time: 0.0140 Steps: 3510, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001391, Sample Num: 22256, Cur Loss: 1.51985455, Cur Avg Loss: 1.51188685, Log Avg loss: 1.34423119, Global Avg Loss: 13.51497395, Time: 0.0233 Steps: 3520, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001401, Sample Num: 22416, Cur Loss: 2.49287891, Cur Avg Loss: 1.51493082, Log Avg loss: 1.93834672, Global Avg Loss: 13.48217897, Time: 0.0119 Steps: 3530, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001411, Sample Num: 22576, Cur Loss: 2.21705842, Cur Avg Loss: 1.51594575, Log Avg loss: 1.65813730, Global Avg Loss: 13.44877772, Time: 0.0121 Steps: 3540, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001421, Sample Num: 22736, Cur Loss: 0.80091214, Cur Avg Loss: 1.51706981, Log Avg loss: 1.67567480, Global Avg Loss: 13.41561405, Time: 0.0120 Steps: 3550, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001431, Sample Num: 22896, Cur Loss: 0.84705281, Cur Avg Loss: 1.51680302, Log Avg loss: 1.47889237, Global Avg Loss: 13.38208393, Time: 0.0070 Steps: 3560, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001441, Sample Num: 23056, Cur Loss: 0.74687898, Cur Avg Loss: 1.51682095, Log Avg loss: 1.51938651, Global Avg Loss: 13.34885509, Time: 0.0069 Steps: 3570, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001451, Sample Num: 23216, Cur Loss: 1.80127358, Cur Avg Loss: 1.51508042, Log Avg loss: 1.26427048, Global Avg Loss: 13.31509927, Time: 0.0069 Steps: 3580, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001461, Sample Num: 23376, Cur Loss: 1.59913075, Cur Avg Loss: 1.51578942, Log Avg loss: 1.61866514, Global Avg Loss: 13.28251867, Time: 0.0112 Steps: 3590, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001471, Sample Num: 23536, Cur Loss: 0.70038295, Cur Avg Loss: 1.51231854, Log Avg loss: 1.00522380, Global Avg Loss: 13.24841507, Time: 0.0067 Steps: 3600, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001481, Sample Num: 23696, Cur Loss: 2.75359011, Cur Avg Loss: 1.51117806, Log Avg loss: 1.34341254, Global Avg Loss: 13.21543723, Time: 0.0229 Steps: 3610, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001491, Sample Num: 23856, Cur Loss: 0.71094096, Cur Avg Loss: 1.51171003, Log Avg loss: 1.59049573, Global Avg Loss: 13.18332413, Time: 0.0120 Steps: 3620, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001501, Sample Num: 24016, Cur Loss: 1.97758698, Cur Avg Loss: 1.51048216, Log Avg loss: 1.32740541, Global Avg Loss: 13.15066319, Time: 0.0233 Steps: 3630, Updated lr: 0.000098 Training, Epoch: 0002, Batch: 001511, Sample Num: 24176, Cur Loss: 1.39445066, Cur Avg Loss: 1.50827395, Log Avg loss: 1.17682303, Global Avg Loss: 13.11776803, Time: 0.0064 Steps: 3640, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001521, Sample Num: 24336, Cur Loss: 0.39918691, Cur Avg Loss: 1.50769323, Log Avg loss: 1.41994563, Global Avg Loss: 13.08571920, Time: 0.0105 Steps: 3650, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001531, Sample Num: 24496, Cur Loss: 0.74161029, Cur Avg Loss: 1.50614165, Log Avg loss: 1.27014583, Global Avg Loss: 13.05343621, Time: 0.0067 Steps: 3660, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001541, Sample Num: 24656, Cur Loss: 0.79954624, Cur Avg Loss: 1.50395053, Log Avg loss: 1.16849093, Global Avg Loss: 13.02105217, Time: 0.0067 Steps: 3670, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001551, Sample Num: 24816, Cur Loss: 1.15455163, Cur Avg Loss: 1.50284609, Log Avg loss: 1.33265166, Global Avg Loss: 12.98929021, Time: 0.0083 Steps: 3680, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001561, Sample Num: 24976, Cur Loss: 1.05921853, Cur Avg Loss: 1.50042304, Log Avg loss: 1.12460856, Global Avg Loss: 12.95713660, Time: 0.0067 Steps: 3690, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001571, Sample Num: 25136, Cur Loss: 1.47641706, Cur Avg Loss: 1.50014868, Log Avg loss: 1.45732058, Global Avg Loss: 12.92605602, Time: 0.0113 Steps: 3700, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001581, Sample Num: 25296, Cur Loss: 0.85468674, Cur Avg Loss: 1.49766489, Log Avg loss: 1.10746206, Global Avg Loss: 12.89419997, Time: 0.0150 Steps: 3710, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001591, Sample Num: 25456, Cur Loss: 2.46897364, Cur Avg Loss: 1.49842759, Log Avg loss: 1.61900968, Global Avg Loss: 12.86389032, Time: 0.0109 Steps: 3720, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001601, Sample Num: 25616, Cur Loss: 0.72885442, Cur Avg Loss: 1.49776382, Log Avg loss: 1.39215816, Global Avg Loss: 12.83313500, Time: 0.0066 Steps: 3730, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001611, Sample Num: 25776, Cur Loss: 1.55009723, Cur Avg Loss: 1.49848956, Log Avg loss: 1.61467982, Global Avg Loss: 12.80313913, Time: 0.0068 Steps: 3740, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001621, Sample Num: 25936, Cur Loss: 1.31804621, Cur Avg Loss: 1.49758270, Log Avg loss: 1.35148798, Global Avg Loss: 12.77260140, Time: 0.0116 Steps: 3750, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001631, Sample Num: 26096, Cur Loss: 1.80444717, Cur Avg Loss: 1.49812936, Log Avg loss: 1.58674247, Global Avg Loss: 12.74285177, Time: 0.0094 Steps: 3760, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001641, Sample Num: 26256, Cur Loss: 0.77814937, Cur Avg Loss: 1.49565852, Log Avg loss: 1.09266521, Global Avg Loss: 12.71194942, Time: 0.0133 Steps: 3770, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001651, Sample Num: 26416, Cur Loss: 2.00325155, Cur Avg Loss: 1.49480321, Log Avg loss: 1.35444694, Global Avg Loss: 12.68190312, Time: 0.0112 Steps: 3780, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001661, Sample Num: 26576, Cur Loss: 0.93471014, Cur Avg Loss: 1.49253791, Log Avg loss: 1.11853607, Global Avg Loss: 12.65139291, Time: 0.0092 Steps: 3790, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001671, Sample Num: 26736, Cur Loss: 1.30138469, Cur Avg Loss: 1.49150887, Log Avg loss: 1.32058603, Global Avg Loss: 12.62157500, Time: 0.0114 Steps: 3800, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001681, Sample Num: 26896, Cur Loss: 1.19275165, Cur Avg Loss: 1.49013460, Log Avg loss: 1.26049347, Global Avg Loss: 12.59175589, Time: 0.0114 Steps: 3810, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001691, Sample Num: 27056, Cur Loss: 1.11606336, Cur Avg Loss: 1.49172421, Log Avg loss: 1.75893795, Global Avg Loss: 12.56339773, Time: 0.0067 Steps: 3820, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001701, Sample Num: 27216, Cur Loss: 0.88425398, Cur Avg Loss: 1.49166202, Log Avg loss: 1.48114519, Global Avg Loss: 12.53446234, Time: 0.0220 Steps: 3830, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001711, Sample Num: 27376, Cur Loss: 0.52298313, Cur Avg Loss: 1.49067345, Log Avg loss: 1.32251862, Global Avg Loss: 12.50526457, Time: 0.0067 Steps: 3840, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001721, Sample Num: 27536, Cur Loss: 1.32532215, Cur Avg Loss: 1.49205798, Log Avg loss: 1.72895038, Global Avg Loss: 12.47727415, Time: 0.0113 Steps: 3850, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001731, Sample Num: 27696, Cur Loss: 1.57796681, Cur Avg Loss: 1.49161631, Log Avg loss: 1.41560571, Global Avg Loss: 12.44861697, Time: 0.0065 Steps: 3860, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001741, Sample Num: 27856, Cur Loss: 0.72325599, Cur Avg Loss: 1.49150320, Log Avg loss: 1.47192301, Global Avg Loss: 12.42025342, Time: 0.0072 Steps: 3870, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001751, Sample Num: 28016, Cur Loss: 1.63808239, Cur Avg Loss: 1.49040744, Log Avg loss: 1.29963612, Global Avg Loss: 12.39159204, Time: 0.0139 Steps: 3880, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001761, Sample Num: 28176, Cur Loss: 1.01788974, Cur Avg Loss: 1.49186536, Log Avg loss: 1.74714753, Global Avg Loss: 12.36422843, Time: 0.0064 Steps: 3890, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001771, Sample Num: 28336, Cur Loss: 2.70043969, Cur Avg Loss: 1.49249255, Log Avg loss: 1.60293973, Global Avg Loss: 12.33663538, Time: 0.0066 Steps: 3900, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001781, Sample Num: 28496, Cur Loss: 2.41201305, Cur Avg Loss: 1.49297556, Log Avg loss: 1.57851708, Global Avg Loss: 12.30912101, Time: 0.0138 Steps: 3910, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001791, Sample Num: 28656, Cur Loss: 2.92205095, Cur Avg Loss: 1.49313838, Log Avg loss: 1.52213668, Global Avg Loss: 12.28160319, Time: 0.0109 Steps: 3920, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001801, Sample Num: 28816, Cur Loss: 0.59252536, Cur Avg Loss: 1.49152505, Log Avg loss: 1.20257792, Global Avg Loss: 12.25341229, Time: 0.0085 Steps: 3930, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001811, Sample Num: 28976, Cur Loss: 1.69902623, Cur Avg Loss: 1.49227836, Log Avg loss: 1.62794881, Global Avg Loss: 12.22644411, Time: 0.0065 Steps: 3940, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001821, Sample Num: 29136, Cur Loss: 2.40793753, Cur Avg Loss: 1.49301686, Log Avg loss: 1.62676051, Global Avg Loss: 12.19960947, Time: 0.0105 Steps: 3950, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001831, Sample Num: 29296, Cur Loss: 1.76612687, Cur Avg Loss: 1.49445261, Log Avg loss: 1.75590237, Global Avg Loss: 12.17323647, Time: 0.0072 Steps: 3960, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001841, Sample Num: 29456, Cur Loss: 1.28115869, Cur Avg Loss: 1.49313394, Log Avg loss: 1.25168601, Global Avg Loss: 12.14572627, Time: 0.0143 Steps: 3970, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001851, Sample Num: 29616, Cur Loss: 0.84840310, Cur Avg Loss: 1.49134330, Log Avg loss: 1.16168535, Global Avg Loss: 12.11812817, Time: 0.0066 Steps: 3980, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001861, Sample Num: 29776, Cur Loss: 1.26642072, Cur Avg Loss: 1.49131300, Log Avg loss: 1.48570488, Global Avg Loss: 12.09148050, Time: 0.0139 Steps: 3990, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001871, Sample Num: 29936, Cur Loss: 1.53321552, Cur Avg Loss: 1.49125016, Log Avg loss: 1.47955494, Global Avg Loss: 12.06495068, Time: 0.0114 Steps: 4000, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001881, Sample Num: 30096, Cur Loss: 1.82157314, Cur Avg Loss: 1.49202214, Log Avg loss: 1.63646019, Global Avg Loss: 12.03894447, Time: 0.0155 Steps: 4010, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001891, Sample Num: 30256, Cur Loss: 1.10390186, Cur Avg Loss: 1.49216872, Log Avg loss: 1.51973984, Global Avg Loss: 12.01277730, Time: 0.0067 Steps: 4020, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001901, Sample Num: 30416, Cur Loss: 1.55087781, Cur Avg Loss: 1.49227213, Log Avg loss: 1.51182834, Global Avg Loss: 11.98672035, Time: 0.0139 Steps: 4030, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001911, Sample Num: 30576, Cur Loss: 1.08552170, Cur Avg Loss: 1.49320044, Log Avg loss: 1.66967200, Global Avg Loss: 11.96118310, Time: 0.0111 Steps: 4040, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001921, Sample Num: 30736, Cur Loss: 0.70250845, Cur Avg Loss: 1.49251167, Log Avg loss: 1.36088643, Global Avg Loss: 11.93500953, Time: 0.0083 Steps: 4050, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001931, Sample Num: 30896, Cur Loss: 1.14441824, Cur Avg Loss: 1.49381118, Log Avg loss: 1.74344685, Global Avg Loss: 11.90990716, Time: 0.0065 Steps: 4060, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001941, Sample Num: 31056, Cur Loss: 1.42561305, Cur Avg Loss: 1.49428478, Log Avg loss: 1.58573747, Global Avg Loss: 11.88454065, Time: 0.0112 Steps: 4070, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001951, Sample Num: 31216, Cur Loss: 1.92944217, Cur Avg Loss: 1.49456049, Log Avg loss: 1.54807655, Global Avg Loss: 11.85920618, Time: 0.0079 Steps: 4080, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001961, Sample Num: 31376, Cur Loss: 1.36519003, Cur Avg Loss: 1.49416720, Log Avg loss: 1.41743628, Global Avg Loss: 11.83367618, Time: 0.0067 Steps: 4090, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001971, Sample Num: 31536, Cur Loss: 2.03447342, Cur Avg Loss: 1.49384246, Log Avg loss: 1.43016028, Global Avg Loss: 11.80830175, Time: 0.0191 Steps: 4100, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001981, Sample Num: 31696, Cur Loss: 2.59625340, Cur Avg Loss: 1.49373457, Log Avg loss: 1.47246999, Global Avg Loss: 11.78315374, Time: 0.0121 Steps: 4110, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 001991, Sample Num: 31856, Cur Loss: 2.22124672, Cur Avg Loss: 1.49195934, Log Avg loss: 1.14028678, Global Avg Loss: 11.75732154, Time: 0.0067 Steps: 4120, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002001, Sample Num: 32016, Cur Loss: 2.99440384, Cur Avg Loss: 1.49270144, Log Avg loss: 1.64045218, Global Avg Loss: 11.73282549, Time: 0.0132 Steps: 4130, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002011, Sample Num: 32176, Cur Loss: 0.85898125, Cur Avg Loss: 1.49153823, Log Avg loss: 1.25878015, Global Avg Loss: 11.70752586, Time: 0.0116 Steps: 4140, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002021, Sample Num: 32336, Cur Loss: 1.69169903, Cur Avg Loss: 1.49086717, Log Avg loss: 1.35591739, Global Avg Loss: 11.68258222, Time: 0.0230 Steps: 4150, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002031, Sample Num: 32496, Cur Loss: 1.77828670, Cur Avg Loss: 1.49025460, Log Avg loss: 1.36645378, Global Avg Loss: 11.65778384, Time: 0.0117 Steps: 4160, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002041, Sample Num: 32656, Cur Loss: 1.62504554, Cur Avg Loss: 1.49120136, Log Avg loss: 1.68348848, Global Avg Loss: 11.63386467, Time: 0.0082 Steps: 4170, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002051, Sample Num: 32816, Cur Loss: 1.11241984, Cur Avg Loss: 1.49014847, Log Avg loss: 1.27525306, Global Avg Loss: 11.60908330, Time: 0.0127 Steps: 4180, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002061, Sample Num: 32976, Cur Loss: 1.20001042, Cur Avg Loss: 1.48972754, Log Avg loss: 1.40339609, Global Avg Loss: 11.58472605, Time: 0.0071 Steps: 4190, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002071, Sample Num: 33136, Cur Loss: 0.81607771, Cur Avg Loss: 1.48815503, Log Avg loss: 1.16405967, Global Avg Loss: 11.55991494, Time: 0.0065 Steps: 4200, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002081, Sample Num: 33296, Cur Loss: 1.11202431, Cur Avg Loss: 1.48832420, Log Avg loss: 1.52335939, Global Avg Loss: 11.53607514, Time: 0.0188 Steps: 4210, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002091, Sample Num: 33456, Cur Loss: 0.84706128, Cur Avg Loss: 1.48855366, Log Avg loss: 1.53630371, Global Avg Loss: 11.51237900, Time: 0.0068 Steps: 4220, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002101, Sample Num: 33616, Cur Loss: 1.39025450, Cur Avg Loss: 1.48725687, Log Avg loss: 1.21609971, Global Avg Loss: 11.48803791, Time: 0.0072 Steps: 4230, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002111, Sample Num: 33776, Cur Loss: 1.20400751, Cur Avg Loss: 1.48690979, Log Avg loss: 1.41398662, Global Avg Loss: 11.46427836, Time: 0.0158 Steps: 4240, Updated lr: 0.000097 Training, Epoch: 0002, Batch: 002121, Sample Num: 33936, Cur Loss: 1.24721193, Cur Avg Loss: 1.48664417, Log Avg loss: 1.43057274, Global Avg Loss: 11.44066964, Time: 0.0196 Steps: 4250, Updated lr: 0.000097 ***** Running evaluation checkpoint-4258 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-4258 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.446041, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.462414, "eval_total_loss": 1028.077209, "eval_mae": 1.05813, "eval_mse": 1.462509, "eval_r2": 0.070333, "eval_sp_statistic": 0.350423, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.349789, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 3.185637, "test_total_loss": 1599.189702, "test_mae": 1.704293, "test_mse": 3.185016, "test_r2": -1.055637, "test_sp_statistic": 0.115557, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.180444, "test_ps_pvalue": 0.0, "lr": 9.691038406827881e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 11.421815915852996, "train_cur_epoch_loss": 3164.418488264084, "train_cur_epoch_avg_loss": 1.4863402950982076, "train_cur_epoch_time": 22.446040630340576, "train_cur_epoch_avg_time": 0.01054299700814494, "epoch": 2, "step": 4258} ################################################## Training, Epoch: 0003, Batch: 000002, Sample Num: 32, Cur Loss: 1.73394632, Cur Avg Loss: 1.71224684, Log Avg loss: 1.46706979, Global Avg Loss: 11.41725743, Time: 0.0068 Steps: 4260, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000012, Sample Num: 192, Cur Loss: 1.28565669, Cur Avg Loss: 1.52367138, Log Avg loss: 1.48595629, Global Avg Loss: 11.39399912, Time: 0.0109 Steps: 4270, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000022, Sample Num: 352, Cur Loss: 1.83823013, Cur Avg Loss: 1.42356858, Log Avg loss: 1.30344523, Global Avg Loss: 11.37042306, Time: 0.0075 Steps: 4280, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000032, Sample Num: 512, Cur Loss: 1.51924062, Cur Avg Loss: 1.39885304, Log Avg loss: 1.34447883, Global Avg Loss: 11.34705256, Time: 0.0116 Steps: 4290, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000042, Sample Num: 672, Cur Loss: 0.97920966, Cur Avg Loss: 1.40328734, Log Avg loss: 1.41747710, Global Avg Loss: 11.32396052, Time: 0.0113 Steps: 4300, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000052, Sample Num: 832, Cur Loss: 3.16802192, Cur Avg Loss: 1.42012507, Log Avg loss: 1.49084355, Global Avg Loss: 11.30114586, Time: 0.0085 Steps: 4310, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000062, Sample Num: 992, Cur Loss: 1.65257311, Cur Avg Loss: 1.40857564, Log Avg loss: 1.34851863, Global Avg Loss: 11.27810737, Time: 0.0066 Steps: 4320, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000072, Sample Num: 1152, Cur Loss: 1.93521786, Cur Avg Loss: 1.38328032, Log Avg loss: 1.22644930, Global Avg Loss: 11.25489338, Time: 0.0066 Steps: 4330, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000082, Sample Num: 1312, Cur Loss: 2.15156174, Cur Avg Loss: 1.37801968, Log Avg loss: 1.34014312, Global Avg Loss: 11.23204834, Time: 0.0069 Steps: 4340, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000092, Sample Num: 1472, Cur Loss: 0.67546093, Cur Avg Loss: 1.39556943, Log Avg loss: 1.53947734, Global Avg Loss: 11.20976656, Time: 0.0067 Steps: 4350, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000102, Sample Num: 1632, Cur Loss: 1.49853432, Cur Avg Loss: 1.40537795, Log Avg loss: 1.49561631, Global Avg Loss: 11.18748640, Time: 0.0090 Steps: 4360, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000112, Sample Num: 1792, Cur Loss: 1.37173128, Cur Avg Loss: 1.42165279, Log Avg loss: 1.58765616, Global Avg Loss: 11.16551883, Time: 0.0137 Steps: 4370, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000122, Sample Num: 1952, Cur Loss: 1.74980640, Cur Avg Loss: 1.41836935, Log Avg loss: 1.38159478, Global Avg Loss: 11.14318110, Time: 0.0067 Steps: 4380, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000132, Sample Num: 2112, Cur Loss: 2.01007915, Cur Avg Loss: 1.44287129, Log Avg loss: 1.74179503, Global Avg Loss: 11.12176564, Time: 0.0067 Steps: 4390, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000142, Sample Num: 2272, Cur Loss: 1.37649953, Cur Avg Loss: 1.45414708, Log Avg loss: 1.60298752, Global Avg Loss: 11.10013206, Time: 0.0114 Steps: 4400, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000152, Sample Num: 2432, Cur Loss: 1.44160640, Cur Avg Loss: 1.46399490, Log Avg loss: 1.60383384, Global Avg Loss: 11.07859850, Time: 0.0107 Steps: 4410, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000162, Sample Num: 2592, Cur Loss: 2.23223567, Cur Avg Loss: 1.48822812, Log Avg loss: 1.85657309, Global Avg Loss: 11.05773419, Time: 0.0113 Steps: 4420, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000172, Sample Num: 2752, Cur Loss: 1.21555996, Cur Avg Loss: 1.47556101, Log Avg loss: 1.27035384, Global Avg Loss: 11.03564078, Time: 0.0124 Steps: 4430, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000182, Sample Num: 2912, Cur Loss: 1.00948620, Cur Avg Loss: 1.46590705, Log Avg loss: 1.29985898, Global Avg Loss: 11.01371335, Time: 0.0135 Steps: 4440, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000192, Sample Num: 3072, Cur Loss: 0.96931100, Cur Avg Loss: 1.45423794, Log Avg loss: 1.24186012, Global Avg Loss: 10.99175412, Time: 0.0107 Steps: 4450, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000202, Sample Num: 3232, Cur Loss: 1.10286224, Cur Avg Loss: 1.47248691, Log Avg loss: 1.82286721, Global Avg Loss: 10.97119608, Time: 0.0107 Steps: 4460, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000212, Sample Num: 3392, Cur Loss: 0.56224442, Cur Avg Loss: 1.45014768, Log Avg loss: 0.99889523, Global Avg Loss: 10.94888668, Time: 0.0066 Steps: 4470, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000222, Sample Num: 3552, Cur Loss: 2.14016104, Cur Avg Loss: 1.45511426, Log Avg loss: 1.56040578, Global Avg Loss: 10.92793025, Time: 0.0105 Steps: 4480, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000232, Sample Num: 3712, Cur Loss: 2.76230478, Cur Avg Loss: 1.45609075, Log Avg loss: 1.47776884, Global Avg Loss: 10.90688312, Time: 0.0128 Steps: 4490, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000242, Sample Num: 3872, Cur Loss: 1.19304943, Cur Avg Loss: 1.44368437, Log Avg loss: 1.15585619, Global Avg Loss: 10.88521417, Time: 0.0067 Steps: 4500, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000252, Sample Num: 4032, Cur Loss: 0.70993441, Cur Avg Loss: 1.42708669, Log Avg loss: 1.02542279, Global Avg Loss: 10.86335211, Time: 0.0067 Steps: 4510, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000262, Sample Num: 4192, Cur Loss: 1.34946680, Cur Avg Loss: 1.42101198, Log Avg loss: 1.26792927, Global Avg Loss: 10.84212330, Time: 0.0167 Steps: 4520, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000272, Sample Num: 4352, Cur Loss: 0.78801394, Cur Avg Loss: 1.42382262, Log Avg loss: 1.49746139, Global Avg Loss: 10.82149491, Time: 0.0065 Steps: 4530, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000282, Sample Num: 4512, Cur Loss: 0.92248094, Cur Avg Loss: 1.42493100, Log Avg loss: 1.45507908, Global Avg Loss: 10.80086403, Time: 0.0109 Steps: 4540, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000292, Sample Num: 4672, Cur Loss: 0.79087496, Cur Avg Loss: 1.41910330, Log Avg loss: 1.25476215, Global Avg Loss: 10.77988359, Time: 0.0067 Steps: 4550, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000302, Sample Num: 4832, Cur Loss: 1.48356986, Cur Avg Loss: 1.41917169, Log Avg loss: 1.42116858, Global Avg Loss: 10.75936009, Time: 0.0066 Steps: 4560, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000312, Sample Num: 4992, Cur Loss: 1.27863693, Cur Avg Loss: 1.42334374, Log Avg loss: 1.54933971, Global Avg Loss: 10.73920687, Time: 0.0235 Steps: 4570, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000322, Sample Num: 5152, Cur Loss: 0.65494585, Cur Avg Loss: 1.41485702, Log Avg loss: 1.15007122, Global Avg Loss: 10.71826990, Time: 0.0070 Steps: 4580, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000332, Sample Num: 5312, Cur Loss: 2.70361733, Cur Avg Loss: 1.41423514, Log Avg loss: 1.39421089, Global Avg Loss: 10.69795604, Time: 0.0064 Steps: 4590, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000342, Sample Num: 5472, Cur Loss: 0.84615231, Cur Avg Loss: 1.42072955, Log Avg loss: 1.63634399, Global Avg Loss: 10.67825689, Time: 0.0065 Steps: 4600, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000352, Sample Num: 5632, Cur Loss: 1.64170814, Cur Avg Loss: 1.41932137, Log Avg loss: 1.37116161, Global Avg Loss: 10.65806796, Time: 0.0066 Steps: 4610, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000362, Sample Num: 5792, Cur Loss: 0.94891810, Cur Avg Loss: 1.41909568, Log Avg loss: 1.41115108, Global Avg Loss: 10.63805299, Time: 0.0066 Steps: 4620, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000372, Sample Num: 5952, Cur Loss: 1.13146234, Cur Avg Loss: 1.41900270, Log Avg loss: 1.41563714, Global Avg Loss: 10.61813416, Time: 0.0066 Steps: 4630, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000382, Sample Num: 6112, Cur Loss: 2.44004345, Cur Avg Loss: 1.42244203, Log Avg loss: 1.55038496, Global Avg Loss: 10.59859160, Time: 0.0069 Steps: 4640, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000392, Sample Num: 6272, Cur Loss: 1.62100279, Cur Avg Loss: 1.42785009, Log Avg loss: 1.63443795, Global Avg Loss: 10.57931385, Time: 0.0067 Steps: 4650, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000402, Sample Num: 6432, Cur Loss: 2.56413984, Cur Avg Loss: 1.42520552, Log Avg loss: 1.32153857, Global Avg Loss: 10.55944738, Time: 0.0067 Steps: 4660, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000412, Sample Num: 6592, Cur Loss: 1.18217134, Cur Avg Loss: 1.42954742, Log Avg loss: 1.60409172, Global Avg Loss: 10.54027103, Time: 0.0082 Steps: 4670, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000422, Sample Num: 6752, Cur Loss: 1.14582491, Cur Avg Loss: 1.42619265, Log Avg loss: 1.28797616, Global Avg Loss: 10.52050117, Time: 0.0066 Steps: 4680, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000432, Sample Num: 6912, Cur Loss: 1.11449277, Cur Avg Loss: 1.42915048, Log Avg loss: 1.55397080, Global Avg Loss: 10.50138277, Time: 0.0172 Steps: 4690, Updated lr: 0.000097 Training, Epoch: 0003, Batch: 000442, Sample Num: 7072, Cur Loss: 0.61623931, Cur Avg Loss: 1.43231885, Log Avg loss: 1.56919262, Global Avg Loss: 10.48237811, Time: 0.0069 Steps: 4700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000452, Sample Num: 7232, Cur Loss: 0.85122079, Cur Avg Loss: 1.43319421, Log Avg loss: 1.47188492, Global Avg Loss: 10.46324755, Time: 0.0071 Steps: 4710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000462, Sample Num: 7392, Cur Loss: 1.16700315, Cur Avg Loss: 1.43066849, Log Avg loss: 1.31650595, Global Avg Loss: 10.44386886, Time: 0.0069 Steps: 4720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000472, Sample Num: 7552, Cur Loss: 0.97541726, Cur Avg Loss: 1.42530782, Log Avg loss: 1.17764503, Global Avg Loss: 10.42427853, Time: 0.0069 Steps: 4730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000482, Sample Num: 7712, Cur Loss: 2.18028545, Cur Avg Loss: 1.43033104, Log Avg loss: 1.66742673, Global Avg Loss: 10.40580416, Time: 0.0196 Steps: 4740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000492, Sample Num: 7872, Cur Loss: 0.66732121, Cur Avg Loss: 1.42858123, Log Avg loss: 1.34424059, Global Avg Loss: 10.38672719, Time: 0.0066 Steps: 4750, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000502, Sample Num: 8032, Cur Loss: 1.49778807, Cur Avg Loss: 1.42052293, Log Avg loss: 1.02405472, Global Avg Loss: 10.36705771, Time: 0.0064 Steps: 4760, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000512, Sample Num: 8192, Cur Loss: 0.79471380, Cur Avg Loss: 1.42076114, Log Avg loss: 1.43271881, Global Avg Loss: 10.34832744, Time: 0.0168 Steps: 4770, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000522, Sample Num: 8352, Cur Loss: 1.25824928, Cur Avg Loss: 1.41758583, Log Avg loss: 1.25501019, Global Avg Loss: 10.32930376, Time: 0.0084 Steps: 4780, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000532, Sample Num: 8512, Cur Loss: 1.13234353, Cur Avg Loss: 1.41289635, Log Avg loss: 1.16810527, Global Avg Loss: 10.31017808, Time: 0.0109 Steps: 4790, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000542, Sample Num: 8672, Cur Loss: 2.40832853, Cur Avg Loss: 1.42347780, Log Avg loss: 1.98641090, Global Avg Loss: 10.29283690, Time: 0.0134 Steps: 4800, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000552, Sample Num: 8832, Cur Loss: 1.39475286, Cur Avg Loss: 1.42285050, Log Avg loss: 1.38885123, Global Avg Loss: 10.27432550, Time: 0.0135 Steps: 4810, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000562, Sample Num: 8992, Cur Loss: 1.03572392, Cur Avg Loss: 1.42128313, Log Avg loss: 1.33476421, Global Avg Loss: 10.25577869, Time: 0.0118 Steps: 4820, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000572, Sample Num: 9152, Cur Loss: 1.10665572, Cur Avg Loss: 1.41704734, Log Avg loss: 1.17899583, Global Avg Loss: 10.23698618, Time: 0.0067 Steps: 4830, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000582, Sample Num: 9312, Cur Loss: 2.34332657, Cur Avg Loss: 1.41711909, Log Avg loss: 1.42122331, Global Avg Loss: 10.21877179, Time: 0.0068 Steps: 4840, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000592, Sample Num: 9472, Cur Loss: 1.83092165, Cur Avg Loss: 1.41250179, Log Avg loss: 1.14377510, Global Avg Loss: 10.20006046, Time: 0.0070 Steps: 4850, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000602, Sample Num: 9632, Cur Loss: 0.78500295, Cur Avg Loss: 1.40943941, Log Avg loss: 1.22814626, Global Avg Loss: 10.18159973, Time: 0.0112 Steps: 4860, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000612, Sample Num: 9792, Cur Loss: 1.50893247, Cur Avg Loss: 1.40754789, Log Avg loss: 1.29367819, Global Avg Loss: 10.16334938, Time: 0.0082 Steps: 4870, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000622, Sample Num: 9952, Cur Loss: 0.78792465, Cur Avg Loss: 1.40965979, Log Avg loss: 1.53890843, Global Avg Loss: 10.14567634, Time: 0.0067 Steps: 4880, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000632, Sample Num: 10112, Cur Loss: 1.78464437, Cur Avg Loss: 1.40825593, Log Avg loss: 1.32093601, Global Avg Loss: 10.12762984, Time: 0.0112 Steps: 4890, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000642, Sample Num: 10272, Cur Loss: 1.26305962, Cur Avg Loss: 1.40420954, Log Avg loss: 1.14847736, Global Avg Loss: 10.10930504, Time: 0.0226 Steps: 4900, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000652, Sample Num: 10432, Cur Loss: 2.06239772, Cur Avg Loss: 1.40703150, Log Avg loss: 1.58820135, Global Avg Loss: 10.09195045, Time: 0.0088 Steps: 4910, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000662, Sample Num: 10592, Cur Loss: 1.98738039, Cur Avg Loss: 1.40732948, Log Avg loss: 1.42675779, Global Avg Loss: 10.07433827, Time: 0.0117 Steps: 4920, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000672, Sample Num: 10752, Cur Loss: 2.37474871, Cur Avg Loss: 1.41447898, Log Avg loss: 1.88777562, Global Avg Loss: 10.05773267, Time: 0.0067 Steps: 4930, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000682, Sample Num: 10912, Cur Loss: 1.04895020, Cur Avg Loss: 1.41439335, Log Avg loss: 1.40863925, Global Avg Loss: 10.04022438, Time: 0.0067 Steps: 4940, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000692, Sample Num: 11072, Cur Loss: 0.98065984, Cur Avg Loss: 1.40977138, Log Avg loss: 1.09455309, Global Avg Loss: 10.02215232, Time: 0.0068 Steps: 4950, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000702, Sample Num: 11232, Cur Loss: 0.84105778, Cur Avg Loss: 1.40342919, Log Avg loss: 0.96454949, Global Avg Loss: 10.00389102, Time: 0.0092 Steps: 4960, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000712, Sample Num: 11392, Cur Loss: 1.06707382, Cur Avg Loss: 1.40125929, Log Avg loss: 1.24893266, Global Avg Loss: 9.98627541, Time: 0.0116 Steps: 4970, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000722, Sample Num: 11552, Cur Loss: 1.79243076, Cur Avg Loss: 1.40101803, Log Avg loss: 1.38384044, Global Avg Loss: 9.96900144, Time: 0.0112 Steps: 4980, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000732, Sample Num: 11712, Cur Loss: 1.97066319, Cur Avg Loss: 1.40002184, Log Avg loss: 1.32809677, Global Avg Loss: 9.95168500, Time: 0.0072 Steps: 4990, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000742, Sample Num: 11872, Cur Loss: 1.31007075, Cur Avg Loss: 1.39686917, Log Avg loss: 1.16609349, Global Avg Loss: 9.93411382, Time: 0.0113 Steps: 5000, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000752, Sample Num: 12032, Cur Loss: 1.36891592, Cur Avg Loss: 1.39850696, Log Avg loss: 1.52003104, Global Avg Loss: 9.91731924, Time: 0.0085 Steps: 5010, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000762, Sample Num: 12192, Cur Loss: 1.65163803, Cur Avg Loss: 1.39224240, Log Avg loss: 0.92114731, Global Avg Loss: 9.89939858, Time: 0.0067 Steps: 5020, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000772, Sample Num: 12352, Cur Loss: 2.02197552, Cur Avg Loss: 1.39413151, Log Avg loss: 1.53808207, Global Avg Loss: 9.88277569, Time: 0.0135 Steps: 5030, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000782, Sample Num: 12512, Cur Loss: 1.48818982, Cur Avg Loss: 1.39274225, Log Avg loss: 1.28549122, Global Avg Loss: 9.86571758, Time: 0.0121 Steps: 5040, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000792, Sample Num: 12672, Cur Loss: 1.84374809, Cur Avg Loss: 1.39408171, Log Avg loss: 1.49882731, Global Avg Loss: 9.84914948, Time: 0.0109 Steps: 5050, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000802, Sample Num: 12832, Cur Loss: 0.77879858, Cur Avg Loss: 1.39229436, Log Avg loss: 1.25073605, Global Avg Loss: 9.83215657, Time: 0.0107 Steps: 5060, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000812, Sample Num: 12992, Cur Loss: 1.66577387, Cur Avg Loss: 1.39348899, Log Avg loss: 1.48929836, Global Avg Loss: 9.81570123, Time: 0.0111 Steps: 5070, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000822, Sample Num: 13152, Cur Loss: 1.42260993, Cur Avg Loss: 1.39507388, Log Avg loss: 1.52376682, Global Avg Loss: 9.79937852, Time: 0.0118 Steps: 5080, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000832, Sample Num: 13312, Cur Loss: 0.65000224, Cur Avg Loss: 1.39410794, Log Avg loss: 1.31470770, Global Avg Loss: 9.78270923, Time: 0.0155 Steps: 5090, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000842, Sample Num: 13472, Cur Loss: 1.56464326, Cur Avg Loss: 1.39744174, Log Avg loss: 1.67481402, Global Avg Loss: 9.76681139, Time: 0.0068 Steps: 5100, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000852, Sample Num: 13632, Cur Loss: 0.67251241, Cur Avg Loss: 1.39628415, Log Avg loss: 1.29881513, Global Avg Loss: 9.75023997, Time: 0.0065 Steps: 5110, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000862, Sample Num: 13792, Cur Loss: 0.65090919, Cur Avg Loss: 1.39655940, Log Avg loss: 1.42001100, Global Avg Loss: 9.73396999, Time: 0.0066 Steps: 5120, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000872, Sample Num: 13952, Cur Loss: 0.88430536, Cur Avg Loss: 1.39830022, Log Avg loss: 1.54835865, Global Avg Loss: 9.71801364, Time: 0.0079 Steps: 5130, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000882, Sample Num: 14112, Cur Loss: 0.86122513, Cur Avg Loss: 1.39815341, Log Avg loss: 1.38535177, Global Avg Loss: 9.70180223, Time: 0.0132 Steps: 5140, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000892, Sample Num: 14272, Cur Loss: 1.00828493, Cur Avg Loss: 1.39757222, Log Avg loss: 1.34631119, Global Avg Loss: 9.68557798, Time: 0.0085 Steps: 5150, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000902, Sample Num: 14432, Cur Loss: 1.34142303, Cur Avg Loss: 1.39930045, Log Avg loss: 1.55345849, Global Avg Loss: 9.66981806, Time: 0.0067 Steps: 5160, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000912, Sample Num: 14592, Cur Loss: 1.14625978, Cur Avg Loss: 1.40200625, Log Avg loss: 1.64606928, Global Avg Loss: 9.65429823, Time: 0.0122 Steps: 5170, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000922, Sample Num: 14752, Cur Loss: 1.27105939, Cur Avg Loss: 1.40504034, Log Avg loss: 1.68174963, Global Avg Loss: 9.63890721, Time: 0.0111 Steps: 5180, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000932, Sample Num: 14912, Cur Loss: 0.66837275, Cur Avg Loss: 1.40324007, Log Avg loss: 1.23725561, Global Avg Loss: 9.62271906, Time: 0.0073 Steps: 5190, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000942, Sample Num: 15072, Cur Loss: 1.87080336, Cur Avg Loss: 1.40407953, Log Avg loss: 1.48231642, Global Avg Loss: 9.60706444, Time: 0.0119 Steps: 5200, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000952, Sample Num: 15232, Cur Loss: 1.52180767, Cur Avg Loss: 1.40274626, Log Avg loss: 1.27715271, Global Avg Loss: 9.59107612, Time: 0.0195 Steps: 5210, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000962, Sample Num: 15392, Cur Loss: 0.85400748, Cur Avg Loss: 1.40218832, Log Avg loss: 1.34907198, Global Avg Loss: 9.57528684, Time: 0.0227 Steps: 5220, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000972, Sample Num: 15552, Cur Loss: 1.12955630, Cur Avg Loss: 1.40124952, Log Avg loss: 1.31093725, Global Avg Loss: 9.55948503, Time: 0.0210 Steps: 5230, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000982, Sample Num: 15712, Cur Loss: 0.57861161, Cur Avg Loss: 1.40196836, Log Avg loss: 1.47184007, Global Avg Loss: 9.54405059, Time: 0.0071 Steps: 5240, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 000992, Sample Num: 15872, Cur Loss: 1.76563394, Cur Avg Loss: 1.40456041, Log Avg loss: 1.65909974, Global Avg Loss: 9.52903164, Time: 0.0070 Steps: 5250, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001002, Sample Num: 16032, Cur Loss: 0.65780079, Cur Avg Loss: 1.40560724, Log Avg loss: 1.50945262, Global Avg Loss: 9.51378529, Time: 0.0067 Steps: 5260, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001012, Sample Num: 16192, Cur Loss: 1.19035780, Cur Avg Loss: 1.40607518, Log Avg loss: 1.45296230, Global Avg Loss: 9.49848961, Time: 0.0066 Steps: 5270, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001022, Sample Num: 16352, Cur Loss: 0.49858940, Cur Avg Loss: 1.40380869, Log Avg loss: 1.17443969, Global Avg Loss: 9.48272437, Time: 0.0202 Steps: 5280, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001032, Sample Num: 16512, Cur Loss: 2.01289511, Cur Avg Loss: 1.40624928, Log Avg loss: 1.65567805, Global Avg Loss: 9.46792844, Time: 0.0066 Steps: 5290, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001042, Sample Num: 16672, Cur Loss: 0.94858968, Cur Avg Loss: 1.40506099, Log Avg loss: 1.28242978, Global Avg Loss: 9.45248410, Time: 0.0140 Steps: 5300, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001052, Sample Num: 16832, Cur Loss: 1.62790501, Cur Avg Loss: 1.40529203, Log Avg loss: 1.42936638, Global Avg Loss: 9.43737465, Time: 0.0089 Steps: 5310, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001062, Sample Num: 16992, Cur Loss: 2.92347240, Cur Avg Loss: 1.40573198, Log Avg loss: 1.45201482, Global Avg Loss: 9.42236457, Time: 0.0104 Steps: 5320, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001072, Sample Num: 17152, Cur Loss: 0.72238481, Cur Avg Loss: 1.40539234, Log Avg loss: 1.36932166, Global Avg Loss: 9.40725568, Time: 0.0152 Steps: 5330, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001082, Sample Num: 17312, Cur Loss: 1.79387641, Cur Avg Loss: 1.40499762, Log Avg loss: 1.36268402, Global Avg Loss: 9.39219094, Time: 0.0069 Steps: 5340, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001092, Sample Num: 17472, Cur Loss: 1.13118899, Cur Avg Loss: 1.40457580, Log Avg loss: 1.35893539, Global Avg Loss: 9.37717550, Time: 0.0162 Steps: 5350, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001102, Sample Num: 17632, Cur Loss: 1.25015950, Cur Avg Loss: 1.40365826, Log Avg loss: 1.30346291, Global Avg Loss: 9.36211261, Time: 0.0123 Steps: 5360, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001112, Sample Num: 17792, Cur Loss: 2.16729760, Cur Avg Loss: 1.40117931, Log Avg loss: 1.12799849, Global Avg Loss: 9.34677906, Time: 0.0114 Steps: 5370, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001122, Sample Num: 17952, Cur Loss: 1.08620596, Cur Avg Loss: 1.40273324, Log Avg loss: 1.57552999, Global Avg Loss: 9.33233436, Time: 0.0107 Steps: 5380, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001132, Sample Num: 18112, Cur Loss: 0.55278921, Cur Avg Loss: 1.40572074, Log Avg loss: 1.74091878, Global Avg Loss: 9.31825010, Time: 0.0129 Steps: 5390, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001142, Sample Num: 18272, Cur Loss: 0.83179820, Cur Avg Loss: 1.40538369, Log Avg loss: 1.36722990, Global Avg Loss: 9.30352599, Time: 0.0118 Steps: 5400, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001152, Sample Num: 18432, Cur Loss: 1.69097710, Cur Avg Loss: 1.40384244, Log Avg loss: 1.22783093, Global Avg Loss: 9.28859864, Time: 0.0148 Steps: 5410, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001162, Sample Num: 18592, Cur Loss: 0.61797762, Cur Avg Loss: 1.40661971, Log Avg loss: 1.72656109, Global Avg Loss: 9.27464654, Time: 0.0087 Steps: 5420, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001172, Sample Num: 18752, Cur Loss: 0.85748708, Cur Avg Loss: 1.40353203, Log Avg loss: 1.04474427, Global Avg Loss: 9.25949019, Time: 0.0107 Steps: 5430, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001182, Sample Num: 18912, Cur Loss: 0.85803956, Cur Avg Loss: 1.40274089, Log Avg loss: 1.31001852, Global Avg Loss: 9.24487719, Time: 0.0116 Steps: 5440, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001192, Sample Num: 19072, Cur Loss: 1.34096813, Cur Avg Loss: 1.40310693, Log Avg loss: 1.44637274, Global Avg Loss: 9.23056800, Time: 0.0082 Steps: 5450, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001202, Sample Num: 19232, Cur Loss: 1.07193875, Cur Avg Loss: 1.40342823, Log Avg loss: 1.44172721, Global Avg Loss: 9.21630273, Time: 0.0156 Steps: 5460, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001212, Sample Num: 19392, Cur Loss: 2.38760138, Cur Avg Loss: 1.40393341, Log Avg loss: 1.46465716, Global Avg Loss: 9.20213153, Time: 0.0119 Steps: 5470, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001222, Sample Num: 19552, Cur Loss: 0.53821635, Cur Avg Loss: 1.40288950, Log Avg loss: 1.27636708, Global Avg Loss: 9.18766846, Time: 0.0081 Steps: 5480, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001232, Sample Num: 19712, Cur Loss: 1.04414427, Cur Avg Loss: 1.40593537, Log Avg loss: 1.77814012, Global Avg Loss: 9.17417205, Time: 0.0117 Steps: 5490, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001242, Sample Num: 19872, Cur Loss: 2.53530741, Cur Avg Loss: 1.40577345, Log Avg loss: 1.38582502, Global Avg Loss: 9.16001142, Time: 0.0066 Steps: 5500, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001252, Sample Num: 20032, Cur Loss: 1.38713145, Cur Avg Loss: 1.40391641, Log Avg loss: 1.17327192, Global Avg Loss: 9.14551643, Time: 0.0107 Steps: 5510, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001262, Sample Num: 20192, Cur Loss: 3.04629326, Cur Avg Loss: 1.40589830, Log Avg loss: 1.65403138, Global Avg Loss: 9.13194490, Time: 0.0105 Steps: 5520, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001272, Sample Num: 20352, Cur Loss: 2.24272037, Cur Avg Loss: 1.40471667, Log Avg loss: 1.25559551, Global Avg Loss: 9.11770195, Time: 0.0070 Steps: 5530, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001282, Sample Num: 20512, Cur Loss: 1.02488708, Cur Avg Loss: 1.40376925, Log Avg loss: 1.28325721, Global Avg Loss: 9.10356035, Time: 0.0068 Steps: 5540, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001292, Sample Num: 20672, Cur Loss: 0.42934960, Cur Avg Loss: 1.40142847, Log Avg loss: 1.10134071, Global Avg Loss: 9.08914194, Time: 0.0114 Steps: 5550, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001302, Sample Num: 20832, Cur Loss: 0.80093110, Cur Avg Loss: 1.39949683, Log Avg loss: 1.14992802, Global Avg Loss: 9.07486278, Time: 0.0069 Steps: 5560, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001312, Sample Num: 20992, Cur Loss: 0.71990478, Cur Avg Loss: 1.40169787, Log Avg loss: 1.68827437, Global Avg Loss: 9.06160140, Time: 0.0128 Steps: 5570, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001322, Sample Num: 21152, Cur Loss: 1.44032371, Cur Avg Loss: 1.40115262, Log Avg loss: 1.32961467, Global Avg Loss: 9.04774479, Time: 0.0118 Steps: 5580, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001332, Sample Num: 21312, Cur Loss: 2.71891546, Cur Avg Loss: 1.40014187, Log Avg loss: 1.26652169, Global Avg Loss: 9.03382489, Time: 0.0116 Steps: 5590, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001342, Sample Num: 21472, Cur Loss: 2.57671714, Cur Avg Loss: 1.40105486, Log Avg loss: 1.52266449, Global Avg Loss: 9.02041211, Time: 0.0065 Steps: 5600, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001352, Sample Num: 21632, Cur Loss: 2.38831186, Cur Avg Loss: 1.40422220, Log Avg loss: 1.82927900, Global Avg Loss: 9.00759369, Time: 0.0109 Steps: 5610, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001362, Sample Num: 21792, Cur Loss: 1.53303361, Cur Avg Loss: 1.40490335, Log Avg loss: 1.49699546, Global Avg Loss: 8.99422963, Time: 0.0119 Steps: 5620, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001372, Sample Num: 21952, Cur Loss: 0.89502388, Cur Avg Loss: 1.40342870, Log Avg loss: 1.20258083, Global Avg Loss: 8.98039011, Time: 0.0068 Steps: 5630, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001382, Sample Num: 22112, Cur Loss: 1.99967885, Cur Avg Loss: 1.40154412, Log Avg loss: 1.14298007, Global Avg Loss: 8.96649400, Time: 0.0125 Steps: 5640, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001392, Sample Num: 22272, Cur Loss: 2.01975036, Cur Avg Loss: 1.40066451, Log Avg loss: 1.27910242, Global Avg Loss: 8.95288799, Time: 0.0113 Steps: 5650, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001402, Sample Num: 22432, Cur Loss: 1.87822723, Cur Avg Loss: 1.39934857, Log Avg loss: 1.21617008, Global Avg Loss: 8.93921888, Time: 0.0105 Steps: 5660, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001412, Sample Num: 22592, Cur Loss: 1.43335688, Cur Avg Loss: 1.40091739, Log Avg loss: 1.62086624, Global Avg Loss: 8.92631173, Time: 0.0111 Steps: 5670, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001422, Sample Num: 22752, Cur Loss: 1.06087160, Cur Avg Loss: 1.40153083, Log Avg loss: 1.48814793, Global Avg Loss: 8.91321638, Time: 0.0119 Steps: 5680, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001432, Sample Num: 22912, Cur Loss: 0.74684328, Cur Avg Loss: 1.40105571, Log Avg loss: 1.33349359, Global Avg Loss: 8.89989525, Time: 0.0133 Steps: 5690, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001442, Sample Num: 23072, Cur Loss: 1.01806867, Cur Avg Loss: 1.40295075, Log Avg loss: 1.67431991, Global Avg Loss: 8.88721880, Time: 0.0117 Steps: 5700, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001452, Sample Num: 23232, Cur Loss: 0.67987347, Cur Avg Loss: 1.40345578, Log Avg loss: 1.47628201, Global Avg Loss: 8.87423992, Time: 0.0070 Steps: 5710, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001462, Sample Num: 23392, Cur Loss: 2.13236070, Cur Avg Loss: 1.40155297, Log Avg loss: 1.12526444, Global Avg Loss: 8.86069276, Time: 0.0105 Steps: 5720, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001472, Sample Num: 23552, Cur Loss: 2.02990770, Cur Avg Loss: 1.40299455, Log Avg loss: 1.61375342, Global Avg Loss: 8.84804540, Time: 0.0065 Steps: 5730, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001482, Sample Num: 23712, Cur Loss: 1.01560771, Cur Avg Loss: 1.40110598, Log Avg loss: 1.12310916, Global Avg Loss: 8.83458732, Time: 0.0069 Steps: 5740, Updated lr: 0.000096 Training, Epoch: 0003, Batch: 001492, Sample Num: 23872, Cur Loss: 1.92474759, Cur Avg Loss: 1.40074809, Log Avg loss: 1.34770845, Global Avg Loss: 8.82156666, Time: 0.0069 Steps: 5750, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001502, Sample Num: 24032, Cur Loss: 1.00454807, Cur Avg Loss: 1.40242241, Log Avg loss: 1.65223110, Global Avg Loss: 8.80911990, Time: 0.0064 Steps: 5760, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001512, Sample Num: 24192, Cur Loss: 0.53617895, Cur Avg Loss: 1.40310709, Log Avg loss: 1.50594648, Global Avg Loss: 8.79646275, Time: 0.0085 Steps: 5770, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001522, Sample Num: 24352, Cur Loss: 1.58271217, Cur Avg Loss: 1.40409384, Log Avg loss: 1.55329053, Global Avg Loss: 8.78393132, Time: 0.0106 Steps: 5780, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001532, Sample Num: 24512, Cur Loss: 1.47409213, Cur Avg Loss: 1.40279945, Log Avg loss: 1.20579272, Global Avg Loss: 8.77084299, Time: 0.0088 Steps: 5790, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001542, Sample Num: 24672, Cur Loss: 1.03142595, Cur Avg Loss: 1.40272906, Log Avg loss: 1.39194538, Global Avg Loss: 8.75812076, Time: 0.0100 Steps: 5800, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001552, Sample Num: 24832, Cur Loss: 0.85887581, Cur Avg Loss: 1.40029093, Log Avg loss: 1.02433137, Global Avg Loss: 8.74480959, Time: 0.0158 Steps: 5810, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001562, Sample Num: 24992, Cur Loss: 1.17283630, Cur Avg Loss: 1.40025080, Log Avg loss: 1.39402273, Global Avg Loss: 8.73217937, Time: 0.0069 Steps: 5820, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001572, Sample Num: 25152, Cur Loss: 0.85488039, Cur Avg Loss: 1.40031575, Log Avg loss: 1.41046089, Global Avg Loss: 8.71962067, Time: 0.0111 Steps: 5830, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001582, Sample Num: 25312, Cur Loss: 0.32428992, Cur Avg Loss: 1.39907632, Log Avg loss: 1.20423748, Global Avg Loss: 8.70675187, Time: 0.0069 Steps: 5840, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001592, Sample Num: 25472, Cur Loss: 1.94793439, Cur Avg Loss: 1.39966356, Log Avg loss: 1.49256487, Global Avg Loss: 8.69441992, Time: 0.0068 Steps: 5850, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001602, Sample Num: 25632, Cur Loss: 1.19569039, Cur Avg Loss: 1.40117046, Log Avg loss: 1.64106887, Global Avg Loss: 8.68238349, Time: 0.0108 Steps: 5860, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001612, Sample Num: 25792, Cur Loss: 2.18415427, Cur Avg Loss: 1.40016498, Log Avg loss: 1.23908799, Global Avg Loss: 8.66970326, Time: 0.0068 Steps: 5870, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001622, Sample Num: 25952, Cur Loss: 0.72070622, Cur Avg Loss: 1.40027315, Log Avg loss: 1.41771005, Global Avg Loss: 8.65736994, Time: 0.0073 Steps: 5880, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001632, Sample Num: 26112, Cur Loss: 0.84010237, Cur Avg Loss: 1.39883466, Log Avg loss: 1.16551100, Global Avg Loss: 8.64465031, Time: 0.0067 Steps: 5890, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001642, Sample Num: 26272, Cur Loss: 1.34333706, Cur Avg Loss: 1.39996508, Log Avg loss: 1.58444923, Global Avg Loss: 8.63268387, Time: 0.0088 Steps: 5900, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001652, Sample Num: 26432, Cur Loss: 1.00679016, Cur Avg Loss: 1.40148463, Log Avg loss: 1.65099432, Global Avg Loss: 8.62087052, Time: 0.0122 Steps: 5910, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001662, Sample Num: 26592, Cur Loss: 1.27864242, Cur Avg Loss: 1.40135828, Log Avg loss: 1.38048555, Global Avg Loss: 8.60864014, Time: 0.0093 Steps: 5920, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001672, Sample Num: 26752, Cur Loss: 1.01646698, Cur Avg Loss: 1.40071234, Log Avg loss: 1.29335790, Global Avg Loss: 8.59630408, Time: 0.0066 Steps: 5930, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001682, Sample Num: 26912, Cur Loss: 1.15970063, Cur Avg Loss: 1.39972638, Log Avg loss: 1.23487329, Global Avg Loss: 8.58391110, Time: 0.0119 Steps: 5940, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001692, Sample Num: 27072, Cur Loss: 1.29146504, Cur Avg Loss: 1.40079776, Log Avg loss: 1.58100486, Global Avg Loss: 8.57214151, Time: 0.0109 Steps: 5950, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001702, Sample Num: 27232, Cur Loss: 1.32538104, Cur Avg Loss: 1.40121774, Log Avg loss: 1.47227715, Global Avg Loss: 8.56022899, Time: 0.0070 Steps: 5960, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001712, Sample Num: 27392, Cur Loss: 1.58108425, Cur Avg Loss: 1.39925851, Log Avg loss: 1.06579879, Global Avg Loss: 8.54767550, Time: 0.0119 Steps: 5970, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001722, Sample Num: 27552, Cur Loss: 0.74939245, Cur Avg Loss: 1.39945004, Log Avg loss: 1.43223851, Global Avg Loss: 8.53577678, Time: 0.0097 Steps: 5980, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001732, Sample Num: 27712, Cur Loss: 1.77593017, Cur Avg Loss: 1.39841207, Log Avg loss: 1.21967373, Global Avg Loss: 8.52356292, Time: 0.0101 Steps: 5990, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001742, Sample Num: 27872, Cur Loss: 1.17809415, Cur Avg Loss: 1.39965597, Log Avg loss: 1.61510038, Global Avg Loss: 8.51204881, Time: 0.0119 Steps: 6000, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001752, Sample Num: 28032, Cur Loss: 1.34528732, Cur Avg Loss: 1.39779950, Log Avg loss: 1.07440228, Global Avg Loss: 8.49967336, Time: 0.0079 Steps: 6010, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001762, Sample Num: 28192, Cur Loss: 0.53541988, Cur Avg Loss: 1.39732086, Log Avg loss: 1.31346359, Global Avg Loss: 8.48773613, Time: 0.0074 Steps: 6020, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001772, Sample Num: 28352, Cur Loss: 2.81913948, Cur Avg Loss: 1.39665479, Log Avg loss: 1.27929183, Global Avg Loss: 8.47578183, Time: 0.0067 Steps: 6030, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001782, Sample Num: 28512, Cur Loss: 0.60244590, Cur Avg Loss: 1.39523110, Log Avg loss: 1.14295451, Global Avg Loss: 8.46364139, Time: 0.0117 Steps: 6040, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001792, Sample Num: 28672, Cur Loss: 1.53260219, Cur Avg Loss: 1.39381801, Log Avg loss: 1.14200460, Global Avg Loss: 8.45153951, Time: 0.0127 Steps: 6050, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001802, Sample Num: 28832, Cur Loss: 1.62122750, Cur Avg Loss: 1.39426876, Log Avg loss: 1.47504301, Global Avg Loss: 8.44002714, Time: 0.0087 Steps: 6060, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001812, Sample Num: 28992, Cur Loss: 1.14823151, Cur Avg Loss: 1.39456112, Log Avg loss: 1.44724498, Global Avg Loss: 8.42850691, Time: 0.0111 Steps: 6070, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001822, Sample Num: 29152, Cur Loss: 1.39094436, Cur Avg Loss: 1.39589138, Log Avg loss: 1.63693455, Global Avg Loss: 8.41733656, Time: 0.0068 Steps: 6080, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001832, Sample Num: 29312, Cur Loss: 1.01675725, Cur Avg Loss: 1.39541841, Log Avg loss: 1.30924245, Global Avg Loss: 8.40566481, Time: 0.0068 Steps: 6090, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001842, Sample Num: 29472, Cur Loss: 1.21694350, Cur Avg Loss: 1.39375974, Log Avg loss: 1.08989194, Global Avg Loss: 8.39367174, Time: 0.0131 Steps: 6100, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001852, Sample Num: 29632, Cur Loss: 1.68215656, Cur Avg Loss: 1.39362423, Log Avg loss: 1.36866278, Global Avg Loss: 8.38217418, Time: 0.0124 Steps: 6110, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001862, Sample Num: 29792, Cur Loss: 2.14225817, Cur Avg Loss: 1.39252383, Log Avg loss: 1.18873042, Global Avg Loss: 8.37042019, Time: 0.0067 Steps: 6120, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001872, Sample Num: 29952, Cur Loss: 0.44883069, Cur Avg Loss: 1.39180760, Log Avg loss: 1.25844471, Global Avg Loss: 8.35881827, Time: 0.0066 Steps: 6130, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001882, Sample Num: 30112, Cur Loss: 1.63196242, Cur Avg Loss: 1.39291851, Log Avg loss: 1.60088125, Global Avg Loss: 8.34781186, Time: 0.0068 Steps: 6140, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001892, Sample Num: 30272, Cur Loss: 2.46851444, Cur Avg Loss: 1.39406260, Log Avg loss: 1.60938076, Global Avg Loss: 8.33685506, Time: 0.0124 Steps: 6150, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001902, Sample Num: 30432, Cur Loss: 0.61467296, Cur Avg Loss: 1.39440287, Log Avg loss: 1.45878161, Global Avg Loss: 8.32568935, Time: 0.0141 Steps: 6160, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001912, Sample Num: 30592, Cur Loss: 1.75479054, Cur Avg Loss: 1.39444094, Log Avg loss: 1.40168252, Global Avg Loss: 8.31446730, Time: 0.0131 Steps: 6170, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001922, Sample Num: 30752, Cur Loss: 1.29027522, Cur Avg Loss: 1.39309815, Log Avg loss: 1.13635714, Global Avg Loss: 8.30285224, Time: 0.0068 Steps: 6180, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001932, Sample Num: 30912, Cur Loss: 0.95729232, Cur Avg Loss: 1.39287041, Log Avg loss: 1.34909795, Global Avg Loss: 8.29161838, Time: 0.0116 Steps: 6190, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001942, Sample Num: 31072, Cur Loss: 0.79174042, Cur Avg Loss: 1.39351746, Log Avg loss: 1.51852793, Global Avg Loss: 8.28069405, Time: 0.0121 Steps: 6200, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001952, Sample Num: 31232, Cur Loss: 1.92445338, Cur Avg Loss: 1.39348318, Log Avg loss: 1.38682624, Global Avg Loss: 8.26959281, Time: 0.0192 Steps: 6210, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001962, Sample Num: 31392, Cur Loss: 2.03320169, Cur Avg Loss: 1.39297006, Log Avg loss: 1.29280887, Global Avg Loss: 8.25837611, Time: 0.0121 Steps: 6220, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001972, Sample Num: 31552, Cur Loss: 0.81585288, Cur Avg Loss: 1.39216361, Log Avg loss: 1.23393772, Global Avg Loss: 8.24710093, Time: 0.0107 Steps: 6230, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001982, Sample Num: 31712, Cur Loss: 0.86030537, Cur Avg Loss: 1.39063193, Log Avg loss: 1.08858456, Global Avg Loss: 8.23562895, Time: 0.0113 Steps: 6240, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 001992, Sample Num: 31872, Cur Loss: 0.91406715, Cur Avg Loss: 1.39042630, Log Avg loss: 1.34966959, Global Avg Loss: 8.22461142, Time: 0.0069 Steps: 6250, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002002, Sample Num: 32032, Cur Loss: 1.47269261, Cur Avg Loss: 1.39008551, Log Avg loss: 1.32220051, Global Avg Loss: 8.21358520, Time: 0.0086 Steps: 6260, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002012, Sample Num: 32192, Cur Loss: 0.24679948, Cur Avg Loss: 1.39031891, Log Avg loss: 1.43704606, Global Avg Loss: 8.20277732, Time: 0.0117 Steps: 6270, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002022, Sample Num: 32352, Cur Loss: 0.97086716, Cur Avg Loss: 1.38847418, Log Avg loss: 1.01731546, Global Avg Loss: 8.19133550, Time: 0.0116 Steps: 6280, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002032, Sample Num: 32512, Cur Loss: 0.95870161, Cur Avg Loss: 1.38983966, Log Avg loss: 1.66593819, Global Avg Loss: 8.18096126, Time: 0.0111 Steps: 6290, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002042, Sample Num: 32672, Cur Loss: 1.47212780, Cur Avg Loss: 1.39077322, Log Avg loss: 1.58047284, Global Avg Loss: 8.17048430, Time: 0.0116 Steps: 6300, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002052, Sample Num: 32832, Cur Loss: 1.13888705, Cur Avg Loss: 1.39030146, Log Avg loss: 1.29396781, Global Avg Loss: 8.15958649, Time: 0.0068 Steps: 6310, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002062, Sample Num: 32992, Cur Loss: 1.85413480, Cur Avg Loss: 1.39047402, Log Avg loss: 1.42588350, Global Avg Loss: 8.14893190, Time: 0.0069 Steps: 6320, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002072, Sample Num: 33152, Cur Loss: 0.62811637, Cur Avg Loss: 1.39134251, Log Avg loss: 1.57042562, Global Avg Loss: 8.13853931, Time: 0.0068 Steps: 6330, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002082, Sample Num: 33312, Cur Loss: 0.68448782, Cur Avg Loss: 1.39137980, Log Avg loss: 1.39910703, Global Avg Loss: 8.12790929, Time: 0.0196 Steps: 6340, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002092, Sample Num: 33472, Cur Loss: 1.80118155, Cur Avg Loss: 1.39019429, Log Avg loss: 1.14337083, Global Avg Loss: 8.11691002, Time: 0.0064 Steps: 6350, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002102, Sample Num: 33632, Cur Loss: 0.92584342, Cur Avg Loss: 1.38757048, Log Avg loss: 0.83866826, Global Avg Loss: 8.10546624, Time: 0.0139 Steps: 6360, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002112, Sample Num: 33792, Cur Loss: 2.39313650, Cur Avg Loss: 1.38885761, Log Avg loss: 1.65941271, Global Avg Loss: 8.09534685, Time: 0.0064 Steps: 6370, Updated lr: 0.000095 Training, Epoch: 0003, Batch: 002122, Sample Num: 33952, Cur Loss: 0.84559751, Cur Avg Loss: 1.38868167, Log Avg loss: 1.35152363, Global Avg Loss: 8.08477659, Time: 0.0064 Steps: 6380, Updated lr: 0.000095 ***** Running evaluation checkpoint-6387 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-6387 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.212727, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.356361, "eval_total_loss": 953.522105, "eval_mae": 0.995283, "eval_mse": 1.356551, "eval_r2": 0.137687, "eval_sp_statistic": 0.408095, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.414036, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 3.703595, "test_total_loss": 1859.204835, "test_mae": 1.816508, "test_mse": 3.702929, "test_r2": -1.389903, "test_sp_statistic": 0.139903, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.196792, "test_ps_pvalue": 0.0, "lr": 9.489141773352301e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 8.076954282735068, "train_cur_epoch_loss": 2953.41483412683, "train_cur_epoch_avg_loss": 1.3872310164992157, "train_cur_epoch_time": 22.212727069854736, "train_cur_epoch_avg_time": 0.010433408675366245, "epoch": 3, "step": 6387} ################################################## Training, Epoch: 0004, Batch: 000003, Sample Num: 48, Cur Loss: 1.02919173, Cur Avg Loss: 1.85536305, Log Avg loss: 1.21984185, Global Avg Loss: 8.07403335, Time: 0.0067 Steps: 6390, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000013, Sample Num: 208, Cur Loss: 0.80309355, Cur Avg Loss: 1.48583441, Log Avg loss: 1.37497581, Global Avg Loss: 8.06356607, Time: 0.0120 Steps: 6400, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000023, Sample Num: 368, Cur Loss: 0.72879118, Cur Avg Loss: 1.32304623, Log Avg loss: 1.11142160, Global Avg Loss: 8.05272029, Time: 0.0067 Steps: 6410, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000033, Sample Num: 528, Cur Loss: 1.16262317, Cur Avg Loss: 1.37925276, Log Avg loss: 1.50852777, Global Avg Loss: 8.04252684, Time: 0.0106 Steps: 6420, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000043, Sample Num: 688, Cur Loss: 1.56319880, Cur Avg Loss: 1.35397496, Log Avg loss: 1.27055824, Global Avg Loss: 8.03199501, Time: 0.0067 Steps: 6430, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000053, Sample Num: 848, Cur Loss: 1.78981650, Cur Avg Loss: 1.34817266, Log Avg loss: 1.32322277, Global Avg Loss: 8.02157766, Time: 0.0066 Steps: 6440, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000063, Sample Num: 1008, Cur Loss: 1.74910259, Cur Avg Loss: 1.33535736, Log Avg loss: 1.26743628, Global Avg Loss: 8.01110613, Time: 0.0066 Steps: 6450, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000073, Sample Num: 1168, Cur Loss: 2.33294201, Cur Avg Loss: 1.35260020, Log Avg loss: 1.46123005, Global Avg Loss: 8.00096700, Time: 0.0117 Steps: 6460, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000083, Sample Num: 1328, Cur Loss: 1.78326702, Cur Avg Loss: 1.32783093, Log Avg loss: 1.14701524, Global Avg Loss: 7.99037357, Time: 0.0118 Steps: 6470, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000093, Sample Num: 1488, Cur Loss: 0.87440705, Cur Avg Loss: 1.31781064, Log Avg loss: 1.23464225, Global Avg Loss: 7.97994805, Time: 0.0078 Steps: 6480, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000103, Sample Num: 1648, Cur Loss: 1.82512283, Cur Avg Loss: 1.33499313, Log Avg loss: 1.49479034, Global Avg Loss: 7.96995552, Time: 0.0067 Steps: 6490, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000113, Sample Num: 1808, Cur Loss: 1.44515491, Cur Avg Loss: 1.33827037, Log Avg loss: 1.37202587, Global Avg Loss: 7.95980485, Time: 0.0066 Steps: 6500, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000123, Sample Num: 1968, Cur Loss: 1.35087538, Cur Avg Loss: 1.33691221, Log Avg loss: 1.32156501, Global Avg Loss: 7.94960787, Time: 0.0135 Steps: 6510, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000133, Sample Num: 2128, Cur Loss: 1.33308911, Cur Avg Loss: 1.33965239, Log Avg loss: 1.37335668, Global Avg Loss: 7.93952159, Time: 0.0066 Steps: 6520, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000143, Sample Num: 2288, Cur Loss: 0.78439802, Cur Avg Loss: 1.34525523, Log Avg loss: 1.41977295, Global Avg Loss: 7.92953729, Time: 0.0066 Steps: 6530, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000153, Sample Num: 2448, Cur Loss: 0.97601455, Cur Avg Loss: 1.33463849, Log Avg loss: 1.18281917, Global Avg Loss: 7.91922121, Time: 0.0137 Steps: 6540, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000163, Sample Num: 2608, Cur Loss: 1.04821944, Cur Avg Loss: 1.31837710, Log Avg loss: 1.06957786, Global Avg Loss: 7.90876374, Time: 0.0136 Steps: 6550, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000173, Sample Num: 2768, Cur Loss: 1.81777370, Cur Avg Loss: 1.32104702, Log Avg loss: 1.36456673, Global Avg Loss: 7.89878783, Time: 0.0139 Steps: 6560, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000183, Sample Num: 2928, Cur Loss: 1.38910890, Cur Avg Loss: 1.33511754, Log Avg loss: 1.57853741, Global Avg Loss: 7.88916796, Time: 0.0067 Steps: 6570, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000193, Sample Num: 3088, Cur Loss: 0.73559803, Cur Avg Loss: 1.32964699, Log Avg loss: 1.22953593, Global Avg Loss: 7.87904694, Time: 0.0067 Steps: 6580, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000203, Sample Num: 3248, Cur Loss: 0.95472944, Cur Avg Loss: 1.32169722, Log Avg loss: 1.16826661, Global Avg Loss: 7.86886366, Time: 0.0066 Steps: 6590, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000213, Sample Num: 3408, Cur Loss: 0.40153968, Cur Avg Loss: 1.30985139, Log Avg loss: 1.06938105, Global Avg Loss: 7.85856142, Time: 0.0134 Steps: 6600, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000223, Sample Num: 3568, Cur Loss: 0.42617708, Cur Avg Loss: 1.31635466, Log Avg loss: 1.45487439, Global Avg Loss: 7.84887354, Time: 0.0067 Steps: 6610, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000233, Sample Num: 3728, Cur Loss: 1.14240861, Cur Avg Loss: 1.32190140, Log Avg loss: 1.44559372, Global Avg Loss: 7.83920091, Time: 0.0067 Steps: 6620, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000243, Sample Num: 3888, Cur Loss: 1.12220621, Cur Avg Loss: 1.31795484, Log Avg loss: 1.22599999, Global Avg Loss: 7.82922625, Time: 0.0068 Steps: 6630, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000253, Sample Num: 4048, Cur Loss: 1.46027339, Cur Avg Loss: 1.33073672, Log Avg loss: 1.64133646, Global Avg Loss: 7.81990714, Time: 0.0146 Steps: 6640, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000263, Sample Num: 4208, Cur Loss: 2.05098319, Cur Avg Loss: 1.33068494, Log Avg loss: 1.32937471, Global Avg Loss: 7.81014694, Time: 0.0067 Steps: 6650, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000273, Sample Num: 4368, Cur Loss: 2.17468405, Cur Avg Loss: 1.34252904, Log Avg loss: 1.65402893, Global Avg Loss: 7.80090352, Time: 0.0186 Steps: 6660, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000283, Sample Num: 4528, Cur Loss: 0.96307397, Cur Avg Loss: 1.34315365, Log Avg loss: 1.36020564, Global Avg Loss: 7.79124730, Time: 0.0066 Steps: 6670, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000293, Sample Num: 4688, Cur Loss: 0.94857192, Cur Avg Loss: 1.33929839, Log Avg loss: 1.23019455, Global Avg Loss: 7.78142536, Time: 0.0070 Steps: 6680, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000303, Sample Num: 4848, Cur Loss: 1.00426030, Cur Avg Loss: 1.34087383, Log Avg loss: 1.38703418, Global Avg Loss: 7.77186723, Time: 0.0231 Steps: 6690, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000313, Sample Num: 5008, Cur Loss: 1.29356980, Cur Avg Loss: 1.33866834, Log Avg loss: 1.27184194, Global Avg Loss: 7.76216570, Time: 0.0072 Steps: 6700, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000323, Sample Num: 5168, Cur Loss: 0.83663028, Cur Avg Loss: 1.34243884, Log Avg loss: 1.46045549, Global Avg Loss: 7.75277418, Time: 0.0065 Steps: 6710, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000333, Sample Num: 5328, Cur Loss: 1.49515605, Cur Avg Loss: 1.34167046, Log Avg loss: 1.31685193, Global Avg Loss: 7.74319691, Time: 0.0067 Steps: 6720, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000343, Sample Num: 5488, Cur Loss: 1.48937738, Cur Avg Loss: 1.33706869, Log Avg loss: 1.18382950, Global Avg Loss: 7.73345046, Time: 0.0204 Steps: 6730, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000353, Sample Num: 5648, Cur Loss: 0.64064717, Cur Avg Loss: 1.32371814, Log Avg loss: 0.86579435, Global Avg Loss: 7.72326105, Time: 0.0064 Steps: 6740, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000363, Sample Num: 5808, Cur Loss: 0.75882828, Cur Avg Loss: 1.32101879, Log Avg loss: 1.22573170, Global Avg Loss: 7.71363509, Time: 0.0067 Steps: 6750, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000373, Sample Num: 5968, Cur Loss: 1.42405033, Cur Avg Loss: 1.32341536, Log Avg loss: 1.41041080, Global Avg Loss: 7.70431079, Time: 0.0070 Steps: 6760, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000383, Sample Num: 6128, Cur Loss: 0.94353312, Cur Avg Loss: 1.32073371, Log Avg loss: 1.22070814, Global Avg Loss: 7.69473383, Time: 0.0071 Steps: 6770, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000393, Sample Num: 6288, Cur Loss: 1.89984715, Cur Avg Loss: 1.31783605, Log Avg loss: 1.20685561, Global Avg Loss: 7.68516469, Time: 0.0194 Steps: 6780, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000403, Sample Num: 6448, Cur Loss: 1.53909683, Cur Avg Loss: 1.31434535, Log Avg loss: 1.17716120, Global Avg Loss: 7.67558000, Time: 0.0237 Steps: 6790, Updated lr: 0.000095 Training, Epoch: 0004, Batch: 000413, Sample Num: 6608, Cur Loss: 1.69264841, Cur Avg Loss: 1.31585458, Log Avg loss: 1.37667648, Global Avg Loss: 7.66631690, Time: 0.0067 Steps: 6800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000423, Sample Num: 6768, Cur Loss: 1.09496987, Cur Avg Loss: 1.32067165, Log Avg loss: 1.51961632, Global Avg Loss: 7.65729091, Time: 0.0066 Steps: 6810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000433, Sample Num: 6928, Cur Loss: 1.31454253, Cur Avg Loss: 1.31588911, Log Avg loss: 1.11358806, Global Avg Loss: 7.64769604, Time: 0.0123 Steps: 6820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000443, Sample Num: 7088, Cur Loss: 1.62264585, Cur Avg Loss: 1.31799029, Log Avg loss: 1.40897107, Global Avg Loss: 7.63856174, Time: 0.0065 Steps: 6830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000453, Sample Num: 7248, Cur Loss: 0.96849126, Cur Avg Loss: 1.31378069, Log Avg loss: 1.12729555, Global Avg Loss: 7.62904235, Time: 0.0069 Steps: 6840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000463, Sample Num: 7408, Cur Loss: 0.76197517, Cur Avg Loss: 1.31096239, Log Avg loss: 1.18329352, Global Avg Loss: 7.61963250, Time: 0.0205 Steps: 6850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000473, Sample Num: 7568, Cur Loss: 0.91920424, Cur Avg Loss: 1.31418987, Log Avg loss: 1.46362222, Global Avg Loss: 7.61065872, Time: 0.0126 Steps: 6860, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000483, Sample Num: 7728, Cur Loss: 0.76534224, Cur Avg Loss: 1.31358801, Log Avg loss: 1.28511991, Global Avg Loss: 7.60145124, Time: 0.0123 Steps: 6870, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000493, Sample Num: 7888, Cur Loss: 1.01012504, Cur Avg Loss: 1.31385748, Log Avg loss: 1.32687283, Global Avg Loss: 7.59233121, Time: 0.0091 Steps: 6880, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000503, Sample Num: 8048, Cur Loss: 1.99576652, Cur Avg Loss: 1.31343673, Log Avg loss: 1.29269387, Global Avg Loss: 7.58318805, Time: 0.0109 Steps: 6890, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000513, Sample Num: 8208, Cur Loss: 1.51124096, Cur Avg Loss: 1.31748196, Log Avg loss: 1.52095675, Global Avg Loss: 7.57440221, Time: 0.0150 Steps: 6900, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000523, Sample Num: 8368, Cur Loss: 0.79854870, Cur Avg Loss: 1.31900470, Log Avg loss: 1.39712162, Global Avg Loss: 7.56546259, Time: 0.0065 Steps: 6910, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000533, Sample Num: 8528, Cur Loss: 1.20690703, Cur Avg Loss: 1.32090826, Log Avg loss: 1.42046428, Global Avg Loss: 7.55658253, Time: 0.0112 Steps: 6920, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000543, Sample Num: 8688, Cur Loss: 1.21990204, Cur Avg Loss: 1.31638024, Log Avg loss: 1.07503660, Global Avg Loss: 7.54722965, Time: 0.0068 Steps: 6930, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000553, Sample Num: 8848, Cur Loss: 2.42408991, Cur Avg Loss: 1.31966397, Log Avg loss: 1.49797064, Global Avg Loss: 7.53851314, Time: 0.0068 Steps: 6940, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000563, Sample Num: 9008, Cur Loss: 0.68363774, Cur Avg Loss: 1.31527411, Log Avg loss: 1.07251506, Global Avg Loss: 7.52920954, Time: 0.0073 Steps: 6950, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000573, Sample Num: 9168, Cur Loss: 0.75670117, Cur Avg Loss: 1.31762537, Log Avg loss: 1.45000097, Global Avg Loss: 7.52047505, Time: 0.0109 Steps: 6960, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000583, Sample Num: 9328, Cur Loss: 1.55588210, Cur Avg Loss: 1.31732642, Log Avg loss: 1.30019657, Global Avg Loss: 7.51155069, Time: 0.0072 Steps: 6970, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000593, Sample Num: 9488, Cur Loss: 1.12617493, Cur Avg Loss: 1.31666168, Log Avg loss: 1.27790778, Global Avg Loss: 7.50261997, Time: 0.0118 Steps: 6980, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000603, Sample Num: 9648, Cur Loss: 0.74037397, Cur Avg Loss: 1.31211608, Log Avg loss: 1.04256186, Global Avg Loss: 7.49337811, Time: 0.0115 Steps: 6990, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000613, Sample Num: 9808, Cur Loss: 1.83046675, Cur Avg Loss: 1.31462641, Log Avg loss: 1.46599914, Global Avg Loss: 7.48476757, Time: 0.0069 Steps: 7000, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000623, Sample Num: 9968, Cur Loss: 0.86622584, Cur Avg Loss: 1.31334085, Log Avg loss: 1.23453595, Global Avg Loss: 7.47585141, Time: 0.0073 Steps: 7010, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000633, Sample Num: 10128, Cur Loss: 0.66567647, Cur Avg Loss: 1.31924255, Log Avg loss: 1.68691882, Global Avg Loss: 7.46760506, Time: 0.0122 Steps: 7020, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000643, Sample Num: 10288, Cur Loss: 1.48327339, Cur Avg Loss: 1.31823333, Log Avg loss: 1.25434954, Global Avg Loss: 7.45876686, Time: 0.0125 Steps: 7030, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000653, Sample Num: 10448, Cur Loss: 1.35145664, Cur Avg Loss: 1.31479487, Log Avg loss: 1.09370156, Global Avg Loss: 7.44972558, Time: 0.0068 Steps: 7040, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000663, Sample Num: 10608, Cur Loss: 2.83278346, Cur Avg Loss: 1.31509400, Log Avg loss: 1.33462750, Global Avg Loss: 7.44105168, Time: 0.0154 Steps: 7050, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000673, Sample Num: 10768, Cur Loss: 0.85329765, Cur Avg Loss: 1.31080329, Log Avg loss: 1.02632905, Global Avg Loss: 7.43196567, Time: 0.0081 Steps: 7060, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000683, Sample Num: 10928, Cur Loss: 0.96123081, Cur Avg Loss: 1.31130066, Log Avg loss: 1.34477376, Global Avg Loss: 7.42335578, Time: 0.0067 Steps: 7070, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000693, Sample Num: 11088, Cur Loss: 0.91269058, Cur Avg Loss: 1.30564675, Log Avg loss: 0.91948446, Global Avg Loss: 7.41416952, Time: 0.0114 Steps: 7080, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000703, Sample Num: 11248, Cur Loss: 0.74160141, Cur Avg Loss: 1.30438830, Log Avg loss: 1.21717798, Global Avg Loss: 7.40542905, Time: 0.0124 Steps: 7090, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000713, Sample Num: 11408, Cur Loss: 0.95573813, Cur Avg Loss: 1.30609992, Log Avg loss: 1.42642699, Global Avg Loss: 7.39700792, Time: 0.0065 Steps: 7100, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000723, Sample Num: 11568, Cur Loss: 0.30366904, Cur Avg Loss: 1.30014935, Log Avg loss: 0.87587336, Global Avg Loss: 7.38783614, Time: 0.0119 Steps: 7110, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000733, Sample Num: 11728, Cur Loss: 0.91066939, Cur Avg Loss: 1.29718189, Log Avg loss: 1.08263485, Global Avg Loss: 7.37898052, Time: 0.0155 Steps: 7120, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000743, Sample Num: 11888, Cur Loss: 1.44887710, Cur Avg Loss: 1.29619782, Log Avg loss: 1.22406560, Global Avg Loss: 7.37034810, Time: 0.0111 Steps: 7130, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000753, Sample Num: 12048, Cur Loss: 2.51580572, Cur Avg Loss: 1.29701437, Log Avg loss: 1.35768368, Global Avg Loss: 7.36192701, Time: 0.0086 Steps: 7140, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000763, Sample Num: 12208, Cur Loss: 0.78231782, Cur Avg Loss: 1.29387284, Log Avg loss: 1.05731589, Global Avg Loss: 7.35310937, Time: 0.0101 Steps: 7150, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000773, Sample Num: 12368, Cur Loss: 1.31220722, Cur Avg Loss: 1.29526690, Log Avg loss: 1.40163372, Global Avg Loss: 7.34479725, Time: 0.0186 Steps: 7160, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000783, Sample Num: 12528, Cur Loss: 1.17038095, Cur Avg Loss: 1.29384908, Log Avg loss: 1.18425124, Global Avg Loss: 7.33620514, Time: 0.0072 Steps: 7170, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000793, Sample Num: 12688, Cur Loss: 1.06728935, Cur Avg Loss: 1.29425708, Log Avg loss: 1.32620386, Global Avg Loss: 7.32783466, Time: 0.0121 Steps: 7180, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000803, Sample Num: 12848, Cur Loss: 1.42347860, Cur Avg Loss: 1.29494327, Log Avg loss: 1.34935758, Global Avg Loss: 7.31951967, Time: 0.0134 Steps: 7190, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000813, Sample Num: 13008, Cur Loss: 1.56768525, Cur Avg Loss: 1.29863296, Log Avg loss: 1.59491533, Global Avg Loss: 7.31156883, Time: 0.0119 Steps: 7200, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000823, Sample Num: 13168, Cur Loss: 0.71102959, Cur Avg Loss: 1.29705104, Log Avg loss: 1.16844064, Global Avg Loss: 7.30304854, Time: 0.0160 Steps: 7210, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000833, Sample Num: 13328, Cur Loss: 2.52727079, Cur Avg Loss: 1.29600687, Log Avg loss: 1.21007226, Global Avg Loss: 7.29460952, Time: 0.0065 Steps: 7220, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000843, Sample Num: 13488, Cur Loss: 1.29833770, Cur Avg Loss: 1.29824568, Log Avg loss: 1.48473798, Global Avg Loss: 7.28657374, Time: 0.0110 Steps: 7230, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000853, Sample Num: 13648, Cur Loss: 1.87043607, Cur Avg Loss: 1.30479473, Log Avg loss: 1.85687995, Global Avg Loss: 7.27907416, Time: 0.0131 Steps: 7240, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000863, Sample Num: 13808, Cur Loss: 1.60489607, Cur Avg Loss: 1.30708311, Log Avg loss: 1.50228163, Global Avg Loss: 7.27110617, Time: 0.0067 Steps: 7250, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000873, Sample Num: 13968, Cur Loss: 1.01333201, Cur Avg Loss: 1.30805861, Log Avg loss: 1.39224463, Global Avg Loss: 7.26300856, Time: 0.0072 Steps: 7260, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000883, Sample Num: 14128, Cur Loss: 0.67438865, Cur Avg Loss: 1.30699077, Log Avg loss: 1.21376873, Global Avg Loss: 7.25468774, Time: 0.0072 Steps: 7270, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000893, Sample Num: 14288, Cur Loss: 1.65970314, Cur Avg Loss: 1.30669751, Log Avg loss: 1.28080213, Global Avg Loss: 7.24648185, Time: 0.0098 Steps: 7280, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000903, Sample Num: 14448, Cur Loss: 1.87646401, Cur Avg Loss: 1.30812473, Log Avg loss: 1.43557521, Global Avg Loss: 7.23851079, Time: 0.0202 Steps: 7290, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000913, Sample Num: 14608, Cur Loss: 0.85854483, Cur Avg Loss: 1.30744412, Log Avg loss: 1.24598498, Global Avg Loss: 7.23030185, Time: 0.0069 Steps: 7300, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000923, Sample Num: 14768, Cur Loss: 1.87404835, Cur Avg Loss: 1.31017447, Log Avg loss: 1.55945610, Global Avg Loss: 7.22254419, Time: 0.0064 Steps: 7310, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000933, Sample Num: 14928, Cur Loss: 1.30267811, Cur Avg Loss: 1.30710233, Log Avg loss: 1.02354307, Global Avg Loss: 7.21407561, Time: 0.0178 Steps: 7320, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000943, Sample Num: 15088, Cur Loss: 1.06014895, Cur Avg Loss: 1.31075936, Log Avg loss: 1.65196069, Global Avg Loss: 7.20648746, Time: 0.0064 Steps: 7330, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000953, Sample Num: 15248, Cur Loss: 0.91817737, Cur Avg Loss: 1.30970439, Log Avg loss: 1.21022098, Global Avg Loss: 7.19831816, Time: 0.0065 Steps: 7340, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000963, Sample Num: 15408, Cur Loss: 1.30989933, Cur Avg Loss: 1.31054279, Log Avg loss: 1.39044178, Global Avg Loss: 7.19041629, Time: 0.0067 Steps: 7350, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000973, Sample Num: 15568, Cur Loss: 1.09061432, Cur Avg Loss: 1.31039812, Log Avg loss: 1.29646689, Global Avg Loss: 7.18240820, Time: 0.0064 Steps: 7360, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000983, Sample Num: 15728, Cur Loss: 1.02771783, Cur Avg Loss: 1.30795771, Log Avg loss: 1.07050557, Global Avg Loss: 7.17411526, Time: 0.0219 Steps: 7370, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 000993, Sample Num: 15888, Cur Loss: 0.84747332, Cur Avg Loss: 1.30741291, Log Avg loss: 1.25385936, Global Avg Loss: 7.16609323, Time: 0.0075 Steps: 7380, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001003, Sample Num: 16048, Cur Loss: 1.88436282, Cur Avg Loss: 1.30693614, Log Avg loss: 1.25959305, Global Avg Loss: 7.15810067, Time: 0.0063 Steps: 7390, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001013, Sample Num: 16208, Cur Loss: 0.48871708, Cur Avg Loss: 1.30628355, Log Avg loss: 1.24082825, Global Avg Loss: 7.15010436, Time: 0.0065 Steps: 7400, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001023, Sample Num: 16368, Cur Loss: 1.17407990, Cur Avg Loss: 1.30709065, Log Avg loss: 1.38885029, Global Avg Loss: 7.14232938, Time: 0.0069 Steps: 7410, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001033, Sample Num: 16528, Cur Loss: 1.47465181, Cur Avg Loss: 1.30563633, Log Avg loss: 1.15685950, Global Avg Loss: 7.13426271, Time: 0.0069 Steps: 7420, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001043, Sample Num: 16688, Cur Loss: 1.59458351, Cur Avg Loss: 1.30595244, Log Avg loss: 1.33860669, Global Avg Loss: 7.12646237, Time: 0.0135 Steps: 7430, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001053, Sample Num: 16848, Cur Loss: 0.96658546, Cur Avg Loss: 1.30506576, Log Avg loss: 1.21258448, Global Avg Loss: 7.11851361, Time: 0.0119 Steps: 7440, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001063, Sample Num: 17008, Cur Loss: 0.79690951, Cur Avg Loss: 1.30264485, Log Avg loss: 1.04772268, Global Avg Loss: 7.11036490, Time: 0.0115 Steps: 7450, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001073, Sample Num: 17168, Cur Loss: 1.15315604, Cur Avg Loss: 1.30105043, Log Avg loss: 1.13156404, Global Avg Loss: 7.10235042, Time: 0.0071 Steps: 7460, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001083, Sample Num: 17328, Cur Loss: 1.11768031, Cur Avg Loss: 1.30213510, Log Avg loss: 1.41851994, Global Avg Loss: 7.09474154, Time: 0.0122 Steps: 7470, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001093, Sample Num: 17488, Cur Loss: 1.90738535, Cur Avg Loss: 1.30216284, Log Avg loss: 1.30516763, Global Avg Loss: 7.08700147, Time: 0.0118 Steps: 7480, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001103, Sample Num: 17648, Cur Loss: 1.60034943, Cur Avg Loss: 1.30242121, Log Avg loss: 1.33066074, Global Avg Loss: 7.07931610, Time: 0.0114 Steps: 7490, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001113, Sample Num: 17808, Cur Loss: 1.28929830, Cur Avg Loss: 1.30194617, Log Avg loss: 1.24954962, Global Avg Loss: 7.07154308, Time: 0.0118 Steps: 7500, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001123, Sample Num: 17968, Cur Loss: 2.33597398, Cur Avg Loss: 1.30811688, Log Avg loss: 1.99491693, Global Avg Loss: 7.06478326, Time: 0.0108 Steps: 7510, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001133, Sample Num: 18128, Cur Loss: 1.75005281, Cur Avg Loss: 1.30884803, Log Avg loss: 1.39095624, Global Avg Loss: 7.05723827, Time: 0.0086 Steps: 7520, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001143, Sample Num: 18288, Cur Loss: 1.36076140, Cur Avg Loss: 1.30777842, Log Avg loss: 1.18659099, Global Avg Loss: 7.04944193, Time: 0.0111 Steps: 7530, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001153, Sample Num: 18448, Cur Loss: 1.83570004, Cur Avg Loss: 1.30492376, Log Avg loss: 0.97863664, Global Avg Loss: 7.04139046, Time: 0.0066 Steps: 7540, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001163, Sample Num: 18608, Cur Loss: 1.06515312, Cur Avg Loss: 1.30527329, Log Avg loss: 1.34557332, Global Avg Loss: 7.03384634, Time: 0.0107 Steps: 7550, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001173, Sample Num: 18768, Cur Loss: 1.55150390, Cur Avg Loss: 1.30379223, Log Avg loss: 1.13154489, Global Avg Loss: 7.02603906, Time: 0.0097 Steps: 7560, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001183, Sample Num: 18928, Cur Loss: 0.80319846, Cur Avg Loss: 1.30233455, Log Avg loss: 1.13134941, Global Avg Loss: 7.01825215, Time: 0.0071 Steps: 7570, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001193, Sample Num: 19088, Cur Loss: 1.54053271, Cur Avg Loss: 1.30389486, Log Avg loss: 1.48847867, Global Avg Loss: 7.01095693, Time: 0.0069 Steps: 7580, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001203, Sample Num: 19248, Cur Loss: 1.53476572, Cur Avg Loss: 1.30362485, Log Avg loss: 1.27141343, Global Avg Loss: 7.00339495, Time: 0.0068 Steps: 7590, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001213, Sample Num: 19408, Cur Loss: 0.96892208, Cur Avg Loss: 1.30227014, Log Avg loss: 1.13929872, Global Avg Loss: 6.99567904, Time: 0.0121 Steps: 7600, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001223, Sample Num: 19568, Cur Loss: 1.45755863, Cur Avg Loss: 1.29964336, Log Avg loss: 0.98101442, Global Avg Loss: 6.98777540, Time: 0.0118 Steps: 7610, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001233, Sample Num: 19728, Cur Loss: 3.45393205, Cur Avg Loss: 1.29907782, Log Avg loss: 1.22991225, Global Avg Loss: 6.98021915, Time: 0.0073 Steps: 7620, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001243, Sample Num: 19888, Cur Loss: 1.31703842, Cur Avg Loss: 1.30225112, Log Avg loss: 1.69351979, Global Avg Loss: 6.97329032, Time: 0.0068 Steps: 7630, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001253, Sample Num: 20048, Cur Loss: 0.90948892, Cur Avg Loss: 1.30234191, Log Avg loss: 1.31362604, Global Avg Loss: 6.96588238, Time: 0.0070 Steps: 7640, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001263, Sample Num: 20208, Cur Loss: 0.91257125, Cur Avg Loss: 1.30289448, Log Avg loss: 1.37213189, Global Avg Loss: 6.95857029, Time: 0.0109 Steps: 7650, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001273, Sample Num: 20368, Cur Loss: 0.81113648, Cur Avg Loss: 1.30363446, Log Avg loss: 1.39709413, Global Avg Loss: 6.95130988, Time: 0.0130 Steps: 7660, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001283, Sample Num: 20528, Cur Loss: 1.17078137, Cur Avg Loss: 1.30369129, Log Avg loss: 1.31092595, Global Avg Loss: 6.94395605, Time: 0.0068 Steps: 7670, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001293, Sample Num: 20688, Cur Loss: 1.04512501, Cur Avg Loss: 1.30338669, Log Avg loss: 1.26430644, Global Avg Loss: 6.93656068, Time: 0.0116 Steps: 7680, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001303, Sample Num: 20848, Cur Loss: 1.16518736, Cur Avg Loss: 1.30264822, Log Avg loss: 1.20716366, Global Avg Loss: 6.92911023, Time: 0.0115 Steps: 7690, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001313, Sample Num: 21008, Cur Loss: 0.68608773, Cur Avg Loss: 1.30216442, Log Avg loss: 1.23912576, Global Avg Loss: 6.92172064, Time: 0.0064 Steps: 7700, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001323, Sample Num: 21168, Cur Loss: 1.24780774, Cur Avg Loss: 1.30162470, Log Avg loss: 1.23075877, Global Avg Loss: 6.91433936, Time: 0.0108 Steps: 7710, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001333, Sample Num: 21328, Cur Loss: 1.30018818, Cur Avg Loss: 1.30271821, Log Avg loss: 1.44739024, Global Avg Loss: 6.90725782, Time: 0.0070 Steps: 7720, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001343, Sample Num: 21488, Cur Loss: 0.42891032, Cur Avg Loss: 1.30229435, Log Avg loss: 1.24579390, Global Avg Loss: 6.89993381, Time: 0.0117 Steps: 7730, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001353, Sample Num: 21648, Cur Loss: 0.69724226, Cur Avg Loss: 1.30243695, Log Avg loss: 1.32158836, Global Avg Loss: 6.89272664, Time: 0.0122 Steps: 7740, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001363, Sample Num: 21808, Cur Loss: 2.67474222, Cur Avg Loss: 1.30263752, Log Avg loss: 1.32977459, Global Avg Loss: 6.88554864, Time: 0.0121 Steps: 7750, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001373, Sample Num: 21968, Cur Loss: 0.61722404, Cur Avg Loss: 1.29945120, Log Avg loss: 0.86515497, Global Avg Loss: 6.87779040, Time: 0.0071 Steps: 7760, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001383, Sample Num: 22128, Cur Loss: 1.53427243, Cur Avg Loss: 1.29931670, Log Avg loss: 1.28084954, Global Avg Loss: 6.87058713, Time: 0.0118 Steps: 7770, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001393, Sample Num: 22288, Cur Loss: 1.16968441, Cur Avg Loss: 1.29734306, Log Avg loss: 1.02438964, Global Avg Loss: 6.86307274, Time: 0.0111 Steps: 7780, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001403, Sample Num: 22448, Cur Loss: 2.07598543, Cur Avg Loss: 1.29850477, Log Avg loss: 1.46033005, Global Avg Loss: 6.85613725, Time: 0.0129 Steps: 7790, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001413, Sample Num: 22608, Cur Loss: 0.89450347, Cur Avg Loss: 1.29916583, Log Avg loss: 1.39191294, Global Avg Loss: 6.84913184, Time: 0.0108 Steps: 7800, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001423, Sample Num: 22768, Cur Loss: 2.03157020, Cur Avg Loss: 1.29928188, Log Avg loss: 1.31567914, Global Avg Loss: 6.84204675, Time: 0.0148 Steps: 7810, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001433, Sample Num: 22928, Cur Loss: 1.27869070, Cur Avg Loss: 1.29846523, Log Avg loss: 1.18225675, Global Avg Loss: 6.83480917, Time: 0.0065 Steps: 7820, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001443, Sample Num: 23088, Cur Loss: 1.51106620, Cur Avg Loss: 1.29979071, Log Avg loss: 1.48973164, Global Avg Loss: 6.82798276, Time: 0.0120 Steps: 7830, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001453, Sample Num: 23248, Cur Loss: 1.88440895, Cur Avg Loss: 1.30224163, Log Avg loss: 1.65590975, Global Avg Loss: 6.82138573, Time: 0.0132 Steps: 7840, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001463, Sample Num: 23408, Cur Loss: 1.04722404, Cur Avg Loss: 1.30296745, Log Avg loss: 1.40842945, Global Avg Loss: 6.81449024, Time: 0.0065 Steps: 7850, Updated lr: 0.000094 Training, Epoch: 0004, Batch: 001473, Sample Num: 23568, Cur Loss: 1.30884862, Cur Avg Loss: 1.30028600, Log Avg loss: 0.90798889, Global Avg Loss: 6.80697561, Time: 0.0105 Steps: 7860, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001483, Sample Num: 23728, Cur Loss: 2.32359076, Cur Avg Loss: 1.30019310, Log Avg loss: 1.28650870, Global Avg Loss: 6.79996104, Time: 0.0118 Steps: 7870, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001493, Sample Num: 23888, Cur Loss: 0.95823574, Cur Avg Loss: 1.29859287, Log Avg loss: 1.06127912, Global Avg Loss: 6.79267845, Time: 0.0069 Steps: 7880, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001503, Sample Num: 24048, Cur Loss: 0.86656475, Cur Avg Loss: 1.29799340, Log Avg loss: 1.20849271, Global Avg Loss: 6.78560090, Time: 0.0142 Steps: 7890, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001513, Sample Num: 24208, Cur Loss: 1.37468958, Cur Avg Loss: 1.29738320, Log Avg loss: 1.20567023, Global Avg Loss: 6.77853769, Time: 0.0111 Steps: 7900, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001523, Sample Num: 24368, Cur Loss: 0.86040878, Cur Avg Loss: 1.29882574, Log Avg loss: 1.51708183, Global Avg Loss: 6.77188604, Time: 0.0072 Steps: 7910, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001533, Sample Num: 24528, Cur Loss: 1.07152748, Cur Avg Loss: 1.30002102, Log Avg loss: 1.48206236, Global Avg Loss: 6.76520697, Time: 0.0158 Steps: 7920, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001543, Sample Num: 24688, Cur Loss: 0.62937236, Cur Avg Loss: 1.29980632, Log Avg loss: 1.26689211, Global Avg Loss: 6.75827341, Time: 0.0067 Steps: 7930, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001553, Sample Num: 24848, Cur Loss: 0.71126854, Cur Avg Loss: 1.29989870, Log Avg loss: 1.31415427, Global Avg Loss: 6.75141684, Time: 0.0118 Steps: 7940, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001563, Sample Num: 25008, Cur Loss: 1.43456554, Cur Avg Loss: 1.30030313, Log Avg loss: 1.36311104, Global Avg Loss: 6.74463909, Time: 0.0070 Steps: 7950, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001573, Sample Num: 25168, Cur Loss: 1.37643707, Cur Avg Loss: 1.30040253, Log Avg loss: 1.31593834, Global Avg Loss: 6.73781912, Time: 0.0137 Steps: 7960, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001583, Sample Num: 25328, Cur Loss: 1.86347389, Cur Avg Loss: 1.30053598, Log Avg loss: 1.32152714, Global Avg Loss: 6.73102327, Time: 0.0112 Steps: 7970, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001593, Sample Num: 25488, Cur Loss: 1.28223181, Cur Avg Loss: 1.29939040, Log Avg loss: 1.11804566, Global Avg Loss: 6.72398946, Time: 0.0119 Steps: 7980, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001603, Sample Num: 25648, Cur Loss: 0.97835988, Cur Avg Loss: 1.29708599, Log Avg loss: 0.92999362, Global Avg Loss: 6.71673790, Time: 0.0072 Steps: 7990, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001613, Sample Num: 25808, Cur Loss: 1.00378633, Cur Avg Loss: 1.29728311, Log Avg loss: 1.32888180, Global Avg Loss: 6.71000308, Time: 0.0075 Steps: 8000, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001623, Sample Num: 25968, Cur Loss: 0.89046925, Cur Avg Loss: 1.29558725, Log Avg loss: 1.02204482, Global Avg Loss: 6.70290201, Time: 0.0114 Steps: 8010, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001633, Sample Num: 26128, Cur Loss: 1.43649149, Cur Avg Loss: 1.29447389, Log Avg loss: 1.11377532, Global Avg Loss: 6.69593303, Time: 0.0107 Steps: 8020, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001643, Sample Num: 26288, Cur Loss: 0.96165645, Cur Avg Loss: 1.29351947, Log Avg loss: 1.13766242, Global Avg Loss: 6.68901114, Time: 0.0137 Steps: 8030, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001653, Sample Num: 26448, Cur Loss: 2.01976395, Cur Avg Loss: 1.29389504, Log Avg loss: 1.35560164, Global Avg Loss: 6.68237755, Time: 0.0065 Steps: 8040, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001663, Sample Num: 26608, Cur Loss: 0.92261696, Cur Avg Loss: 1.29135842, Log Avg loss: 0.87205506, Global Avg Loss: 6.67515976, Time: 0.0114 Steps: 8050, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001673, Sample Num: 26768, Cur Loss: 1.64926565, Cur Avg Loss: 1.29198709, Log Avg loss: 1.39653504, Global Avg Loss: 6.66861060, Time: 0.0137 Steps: 8060, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001683, Sample Num: 26928, Cur Loss: 1.18002009, Cur Avg Loss: 1.29243041, Log Avg loss: 1.36659818, Global Avg Loss: 6.66204057, Time: 0.0068 Steps: 8070, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001693, Sample Num: 27088, Cur Loss: 1.70926154, Cur Avg Loss: 1.29328376, Log Avg loss: 1.43690131, Global Avg Loss: 6.65557381, Time: 0.0067 Steps: 8080, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001703, Sample Num: 27248, Cur Loss: 1.33516955, Cur Avg Loss: 1.29283476, Log Avg loss: 1.21682025, Global Avg Loss: 6.64885100, Time: 0.0073 Steps: 8090, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001713, Sample Num: 27408, Cur Loss: 1.51226568, Cur Avg Loss: 1.29242972, Log Avg loss: 1.22345074, Global Avg Loss: 6.64215298, Time: 0.0069 Steps: 8100, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001723, Sample Num: 27568, Cur Loss: 2.30553746, Cur Avg Loss: 1.29283239, Log Avg loss: 1.36181009, Global Avg Loss: 6.63564207, Time: 0.0111 Steps: 8110, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001733, Sample Num: 27728, Cur Loss: 1.03608680, Cur Avg Loss: 1.29246392, Log Avg loss: 1.22897642, Global Avg Loss: 6.62898362, Time: 0.0119 Steps: 8120, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001743, Sample Num: 27888, Cur Loss: 1.79653680, Cur Avg Loss: 1.29210920, Log Avg loss: 1.23063557, Global Avg Loss: 6.62234358, Time: 0.0067 Steps: 8130, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001753, Sample Num: 28048, Cur Loss: 0.94277048, Cur Avg Loss: 1.29299405, Log Avg loss: 1.44722465, Global Avg Loss: 6.61598594, Time: 0.0069 Steps: 8140, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001763, Sample Num: 28208, Cur Loss: 0.68142760, Cur Avg Loss: 1.29051870, Log Avg loss: 0.85658867, Global Avg Loss: 6.60891920, Time: 0.0074 Steps: 8150, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001773, Sample Num: 28368, Cur Loss: 1.28240013, Cur Avg Loss: 1.29286638, Log Avg loss: 1.70676307, Global Avg Loss: 6.60291165, Time: 0.0072 Steps: 8160, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001783, Sample Num: 28528, Cur Loss: 0.89469659, Cur Avg Loss: 1.29185750, Log Avg loss: 1.11298315, Global Avg Loss: 6.59619204, Time: 0.0107 Steps: 8170, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001793, Sample Num: 28688, Cur Loss: 1.81709552, Cur Avg Loss: 1.29226074, Log Avg loss: 1.36415752, Global Avg Loss: 6.58979591, Time: 0.0091 Steps: 8180, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001803, Sample Num: 28848, Cur Loss: 0.96946108, Cur Avg Loss: 1.29228435, Log Avg loss: 1.29651757, Global Avg Loss: 6.58333281, Time: 0.0115 Steps: 8190, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001813, Sample Num: 29008, Cur Loss: 0.62784564, Cur Avg Loss: 1.29154854, Log Avg loss: 1.15888281, Global Avg Loss: 6.57671762, Time: 0.0114 Steps: 8200, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001823, Sample Num: 29168, Cur Loss: 2.11976147, Cur Avg Loss: 1.29217015, Log Avg loss: 1.40486735, Global Avg Loss: 6.57041817, Time: 0.0107 Steps: 8210, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001833, Sample Num: 29328, Cur Loss: 0.92068267, Cur Avg Loss: 1.29191932, Log Avg loss: 1.24619366, Global Avg Loss: 6.56394101, Time: 0.0117 Steps: 8220, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001843, Sample Num: 29488, Cur Loss: 1.04510033, Cur Avg Loss: 1.29189898, Log Avg loss: 1.28816973, Global Avg Loss: 6.55753060, Time: 0.0107 Steps: 8230, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001853, Sample Num: 29648, Cur Loss: 0.81898582, Cur Avg Loss: 1.29126021, Log Avg loss: 1.17353647, Global Avg Loss: 6.55099662, Time: 0.0107 Steps: 8240, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001863, Sample Num: 29808, Cur Loss: 1.10584474, Cur Avg Loss: 1.28875744, Log Avg loss: 0.82499234, Global Avg Loss: 6.54405601, Time: 0.0066 Steps: 8250, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001873, Sample Num: 29968, Cur Loss: 1.28080463, Cur Avg Loss: 1.28978831, Log Avg loss: 1.48184047, Global Avg Loss: 6.53792742, Time: 0.0085 Steps: 8260, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001883, Sample Num: 30128, Cur Loss: 1.14049804, Cur Avg Loss: 1.28914739, Log Avg loss: 1.16910359, Global Avg Loss: 6.53143550, Time: 0.0066 Steps: 8270, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001893, Sample Num: 30288, Cur Loss: 0.99003792, Cur Avg Loss: 1.28964026, Log Avg loss: 1.38244627, Global Avg Loss: 6.52521691, Time: 0.0117 Steps: 8280, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001903, Sample Num: 30448, Cur Loss: 1.02820420, Cur Avg Loss: 1.28879333, Log Avg loss: 1.12847011, Global Avg Loss: 6.51870696, Time: 0.0127 Steps: 8290, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001913, Sample Num: 30608, Cur Loss: 1.41220498, Cur Avg Loss: 1.28989608, Log Avg loss: 1.49974878, Global Avg Loss: 6.51266002, Time: 0.0111 Steps: 8300, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001923, Sample Num: 30768, Cur Loss: 0.59812993, Cur Avg Loss: 1.28992470, Log Avg loss: 1.29540015, Global Avg Loss: 6.50638173, Time: 0.0129 Steps: 8310, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001933, Sample Num: 30928, Cur Loss: 1.23775530, Cur Avg Loss: 1.28917409, Log Avg loss: 1.14483160, Global Avg Loss: 6.49993756, Time: 0.0158 Steps: 8320, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001943, Sample Num: 31088, Cur Loss: 1.13299775, Cur Avg Loss: 1.28798552, Log Avg loss: 1.05823460, Global Avg Loss: 6.49340491, Time: 0.0120 Steps: 8330, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001953, Sample Num: 31248, Cur Loss: 0.45672521, Cur Avg Loss: 1.28812687, Log Avg loss: 1.31559247, Global Avg Loss: 6.48719650, Time: 0.0234 Steps: 8340, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001963, Sample Num: 31408, Cur Loss: 1.17548442, Cur Avg Loss: 1.28831710, Log Avg loss: 1.32546878, Global Avg Loss: 6.48101479, Time: 0.0063 Steps: 8350, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001973, Sample Num: 31568, Cur Loss: 0.96919233, Cur Avg Loss: 1.28707547, Log Avg loss: 1.04334248, Global Avg Loss: 6.47451039, Time: 0.0065 Steps: 8360, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001983, Sample Num: 31728, Cur Loss: 1.39036512, Cur Avg Loss: 1.28811316, Log Avg loss: 1.49285013, Global Avg Loss: 6.46855859, Time: 0.0065 Steps: 8370, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 001993, Sample Num: 31888, Cur Loss: 1.38571739, Cur Avg Loss: 1.28824168, Log Avg loss: 1.31372816, Global Avg Loss: 6.46240724, Time: 0.0063 Steps: 8380, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002003, Sample Num: 32048, Cur Loss: 2.31641746, Cur Avg Loss: 1.29056052, Log Avg loss: 1.75270523, Global Avg Loss: 6.45679377, Time: 0.0067 Steps: 8390, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002013, Sample Num: 32208, Cur Loss: 0.42488164, Cur Avg Loss: 1.28908442, Log Avg loss: 0.99342059, Global Avg Loss: 6.45028975, Time: 0.0125 Steps: 8400, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002023, Sample Num: 32368, Cur Loss: 0.63963103, Cur Avg Loss: 1.28932676, Log Avg loss: 1.33810979, Global Avg Loss: 6.44421106, Time: 0.0072 Steps: 8410, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002033, Sample Num: 32528, Cur Loss: 0.89074135, Cur Avg Loss: 1.28922651, Log Avg loss: 1.26894670, Global Avg Loss: 6.43806467, Time: 0.0140 Steps: 8420, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002043, Sample Num: 32688, Cur Loss: 0.96584129, Cur Avg Loss: 1.28952711, Log Avg loss: 1.35063924, Global Avg Loss: 6.43202976, Time: 0.0113 Steps: 8430, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002053, Sample Num: 32848, Cur Loss: 0.51221979, Cur Avg Loss: 1.28834156, Log Avg loss: 1.04613384, Global Avg Loss: 6.42564837, Time: 0.0193 Steps: 8440, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002063, Sample Num: 33008, Cur Loss: 0.40304735, Cur Avg Loss: 1.28903292, Log Avg loss: 1.43096872, Global Avg Loss: 6.41973751, Time: 0.0238 Steps: 8450, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002073, Sample Num: 33168, Cur Loss: 1.99826264, Cur Avg Loss: 1.28853959, Log Avg loss: 1.18676551, Global Avg Loss: 6.41355196, Time: 0.0071 Steps: 8460, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002083, Sample Num: 33328, Cur Loss: 0.94481343, Cur Avg Loss: 1.28684694, Log Avg loss: 0.93596107, Global Avg Loss: 6.40708491, Time: 0.0065 Steps: 8470, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002093, Sample Num: 33488, Cur Loss: 0.63355643, Cur Avg Loss: 1.28692380, Log Avg loss: 1.30293189, Global Avg Loss: 6.40106586, Time: 0.0235 Steps: 8480, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002103, Sample Num: 33648, Cur Loss: 0.55062246, Cur Avg Loss: 1.28668793, Log Avg loss: 1.23732054, Global Avg Loss: 6.39498371, Time: 0.0063 Steps: 8490, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002113, Sample Num: 33808, Cur Loss: 0.88290030, Cur Avg Loss: 1.28553162, Log Avg loss: 1.04236133, Global Avg Loss: 6.38868651, Time: 0.0066 Steps: 8500, Updated lr: 0.000093 Training, Epoch: 0004, Batch: 002123, Sample Num: 33968, Cur Loss: 1.89972281, Cur Avg Loss: 1.28502657, Log Avg loss: 1.17830893, Global Avg Loss: 6.38256386, Time: 0.0180 Steps: 8510, Updated lr: 0.000093 ***** Running evaluation checkpoint-8516 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-8516 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.076224, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.33554, "eval_total_loss": 938.884583, "eval_mae": 1.006987, "eval_mse": 1.335656, "eval_r2": 0.150969, "eval_sp_statistic": 0.471961, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.481933, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 5.042491, "test_total_loss": 2531.330635, "test_mae": 2.077435, "test_mse": 5.041592, "test_r2": -2.253888, "test_sp_statistic": 0.128703, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.212309, "test_ps_pvalue": 0.0, "lr": 9.287245139876719e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 6.378964456059861, "train_cur_epoch_loss": 2735.7543039768934, "train_cur_epoch_avg_loss": 1.2849949760342383, "train_cur_epoch_time": 22.076224327087402, "train_cur_epoch_avg_time": 0.010369292779280132, "epoch": 4, "step": 8516} ################################################## Training, Epoch: 0005, Batch: 000004, Sample Num: 64, Cur Loss: 1.30296576, Cur Avg Loss: 1.57623112, Log Avg loss: 1.39478175, Global Avg Loss: 6.37670965, Time: 0.0065 Steps: 8520, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000014, Sample Num: 224, Cur Loss: 1.47974062, Cur Avg Loss: 1.36303582, Log Avg loss: 1.27775769, Global Avg Loss: 6.37073198, Time: 0.0114 Steps: 8530, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000024, Sample Num: 384, Cur Loss: 1.83887422, Cur Avg Loss: 1.33107577, Log Avg loss: 1.28633170, Global Avg Loss: 6.36477835, Time: 0.0224 Steps: 8540, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000034, Sample Num: 544, Cur Loss: 0.94085288, Cur Avg Loss: 1.34508246, Log Avg loss: 1.37869852, Global Avg Loss: 6.35894668, Time: 0.0131 Steps: 8550, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000044, Sample Num: 704, Cur Loss: 0.72710770, Cur Avg Loss: 1.31845054, Log Avg loss: 1.22790204, Global Avg Loss: 6.35295247, Time: 0.0114 Steps: 8560, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000054, Sample Num: 864, Cur Loss: 1.00500536, Cur Avg Loss: 1.32221546, Log Avg loss: 1.33878109, Global Avg Loss: 6.34710163, Time: 0.0120 Steps: 8570, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000064, Sample Num: 1024, Cur Loss: 0.84668064, Cur Avg Loss: 1.32523106, Log Avg loss: 1.34151529, Global Avg Loss: 6.34126761, Time: 0.0113 Steps: 8580, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000074, Sample Num: 1184, Cur Loss: 0.62207866, Cur Avg Loss: 1.30420653, Log Avg loss: 1.16964954, Global Avg Loss: 6.33524710, Time: 0.0120 Steps: 8590, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000084, Sample Num: 1344, Cur Loss: 0.86356318, Cur Avg Loss: 1.29741856, Log Avg loss: 1.24718758, Global Avg Loss: 6.32933075, Time: 0.0226 Steps: 8600, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000094, Sample Num: 1504, Cur Loss: 1.07972932, Cur Avg Loss: 1.29145141, Log Avg loss: 1.24132740, Global Avg Loss: 6.32342134, Time: 0.0121 Steps: 8610, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000104, Sample Num: 1664, Cur Loss: 0.60662270, Cur Avg Loss: 1.25555187, Log Avg loss: 0.91809610, Global Avg Loss: 6.31715066, Time: 0.0071 Steps: 8620, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000114, Sample Num: 1824, Cur Loss: 0.49685150, Cur Avg Loss: 1.23599404, Log Avg loss: 1.03259266, Global Avg Loss: 6.31102719, Time: 0.0098 Steps: 8630, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000124, Sample Num: 1984, Cur Loss: 0.60807884, Cur Avg Loss: 1.22379986, Log Avg loss: 1.08478625, Global Avg Loss: 6.30497830, Time: 0.0108 Steps: 8640, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000134, Sample Num: 2144, Cur Loss: 1.03462768, Cur Avg Loss: 1.21487353, Log Avg loss: 1.10418699, Global Avg Loss: 6.29896582, Time: 0.0123 Steps: 8650, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000144, Sample Num: 2304, Cur Loss: 0.96209574, Cur Avg Loss: 1.23036404, Log Avg loss: 1.43793685, Global Avg Loss: 6.29335262, Time: 0.0138 Steps: 8660, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000154, Sample Num: 2464, Cur Loss: 1.13722944, Cur Avg Loss: 1.23494786, Log Avg loss: 1.30095481, Global Avg Loss: 6.28759438, Time: 0.0127 Steps: 8670, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000164, Sample Num: 2624, Cur Loss: 1.15479338, Cur Avg Loss: 1.23902098, Log Avg loss: 1.30174705, Global Avg Loss: 6.28185032, Time: 0.0067 Steps: 8680, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000174, Sample Num: 2784, Cur Loss: 1.37622964, Cur Avg Loss: 1.23670438, Log Avg loss: 1.19871227, Global Avg Loss: 6.27600091, Time: 0.0109 Steps: 8690, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000184, Sample Num: 2944, Cur Loss: 1.07671010, Cur Avg Loss: 1.23240498, Log Avg loss: 1.15759542, Global Avg Loss: 6.27011768, Time: 0.0097 Steps: 8700, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000194, Sample Num: 3104, Cur Loss: 0.59934437, Cur Avg Loss: 1.23645148, Log Avg loss: 1.31090692, Global Avg Loss: 6.26442398, Time: 0.0240 Steps: 8710, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000204, Sample Num: 3264, Cur Loss: 0.44751292, Cur Avg Loss: 1.22847123, Log Avg loss: 1.07365447, Global Avg Loss: 6.25847127, Time: 0.0128 Steps: 8720, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000214, Sample Num: 3424, Cur Loss: 0.42668617, Cur Avg Loss: 1.24702464, Log Avg loss: 1.62551410, Global Avg Loss: 6.25316433, Time: 0.0068 Steps: 8730, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000224, Sample Num: 3584, Cur Loss: 0.80160165, Cur Avg Loss: 1.23325791, Log Avg loss: 0.93865005, Global Avg Loss: 6.24708365, Time: 0.0106 Steps: 8740, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000234, Sample Num: 3744, Cur Loss: 2.20281458, Cur Avg Loss: 1.24236644, Log Avg loss: 1.44639752, Global Avg Loss: 6.24159715, Time: 0.0068 Steps: 8750, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000244, Sample Num: 3904, Cur Loss: 1.77250528, Cur Avg Loss: 1.23603877, Log Avg loss: 1.08797116, Global Avg Loss: 6.23571401, Time: 0.0109 Steps: 8760, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000254, Sample Num: 4064, Cur Loss: 1.09546959, Cur Avg Loss: 1.22308071, Log Avg loss: 0.90690421, Global Avg Loss: 6.22963783, Time: 0.0068 Steps: 8770, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000264, Sample Num: 4224, Cur Loss: 0.79436588, Cur Avg Loss: 1.23365700, Log Avg loss: 1.50229477, Global Avg Loss: 6.22425362, Time: 0.0065 Steps: 8780, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000274, Sample Num: 4384, Cur Loss: 0.52158606, Cur Avg Loss: 1.23273257, Log Avg loss: 1.20832748, Global Avg Loss: 6.21854722, Time: 0.0069 Steps: 8790, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000284, Sample Num: 4544, Cur Loss: 2.00540376, Cur Avg Loss: 1.24206859, Log Avg loss: 1.49787562, Global Avg Loss: 6.21318282, Time: 0.0230 Steps: 8800, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000294, Sample Num: 4704, Cur Loss: 0.44109583, Cur Avg Loss: 1.25019009, Log Avg loss: 1.48084080, Global Avg Loss: 6.20781126, Time: 0.0088 Steps: 8810, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000304, Sample Num: 4864, Cur Loss: 1.14660120, Cur Avg Loss: 1.24246665, Log Avg loss: 1.01539736, Global Avg Loss: 6.20192417, Time: 0.0072 Steps: 8820, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000314, Sample Num: 5024, Cur Loss: 1.95846188, Cur Avg Loss: 1.24574309, Log Avg loss: 1.34534678, Global Avg Loss: 6.19642408, Time: 0.0076 Steps: 8830, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000324, Sample Num: 5184, Cur Loss: 0.74780917, Cur Avg Loss: 1.23103429, Log Avg loss: 0.76917813, Global Avg Loss: 6.19028466, Time: 0.0227 Steps: 8840, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000334, Sample Num: 5344, Cur Loss: 1.51025319, Cur Avg Loss: 1.23312038, Log Avg loss: 1.30070974, Global Avg Loss: 6.18475972, Time: 0.0114 Steps: 8850, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000344, Sample Num: 5504, Cur Loss: 1.48952842, Cur Avg Loss: 1.23500156, Log Avg loss: 1.29783273, Global Avg Loss: 6.17924400, Time: 0.0123 Steps: 8860, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000354, Sample Num: 5664, Cur Loss: 1.41933024, Cur Avg Loss: 1.23975095, Log Avg loss: 1.40313000, Global Avg Loss: 6.17385943, Time: 0.0068 Steps: 8870, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000364, Sample Num: 5824, Cur Loss: 1.15328550, Cur Avg Loss: 1.23669312, Log Avg loss: 1.12844593, Global Avg Loss: 6.16817766, Time: 0.0067 Steps: 8880, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000374, Sample Num: 5984, Cur Loss: 1.40201581, Cur Avg Loss: 1.22949483, Log Avg loss: 0.96747729, Global Avg Loss: 6.16232760, Time: 0.0143 Steps: 8890, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000384, Sample Num: 6144, Cur Loss: 1.31634116, Cur Avg Loss: 1.22999571, Log Avg loss: 1.24872832, Global Avg Loss: 6.15680670, Time: 0.0064 Steps: 8900, Updated lr: 0.000093 Training, Epoch: 0005, Batch: 000394, Sample Num: 6304, Cur Loss: 0.79620522, Cur Avg Loss: 1.22931732, Log Avg loss: 1.20326747, Global Avg Loss: 6.15124718, Time: 0.0125 Steps: 8910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000404, Sample Num: 6464, Cur Loss: 0.71998227, Cur Avg Loss: 1.22169678, Log Avg loss: 0.92144741, Global Avg Loss: 6.14538417, Time: 0.0066 Steps: 8920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000414, Sample Num: 6624, Cur Loss: 0.98344183, Cur Avg Loss: 1.21803523, Log Avg loss: 1.07010863, Global Avg Loss: 6.13970077, Time: 0.0068 Steps: 8930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000424, Sample Num: 6784, Cur Loss: 1.59583044, Cur Avg Loss: 1.22066247, Log Avg loss: 1.32943003, Global Avg Loss: 6.13432016, Time: 0.0161 Steps: 8940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000434, Sample Num: 6944, Cur Loss: 1.19843459, Cur Avg Loss: 1.22259314, Log Avg loss: 1.30445353, Global Avg Loss: 6.12892366, Time: 0.0066 Steps: 8950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000444, Sample Num: 7104, Cur Loss: 0.73534876, Cur Avg Loss: 1.22029465, Log Avg loss: 1.12054024, Global Avg Loss: 6.12333394, Time: 0.0110 Steps: 8960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000454, Sample Num: 7264, Cur Loss: 0.62526953, Cur Avg Loss: 1.21570371, Log Avg loss: 1.01186619, Global Avg Loss: 6.11763554, Time: 0.0067 Steps: 8970, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000464, Sample Num: 7424, Cur Loss: 0.55110776, Cur Avg Loss: 1.21087593, Log Avg loss: 0.99169470, Global Avg Loss: 6.11192737, Time: 0.0129 Steps: 8980, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000474, Sample Num: 7584, Cur Loss: 1.53656626, Cur Avg Loss: 1.20922872, Log Avg loss: 1.13279782, Global Avg Loss: 6.10638885, Time: 0.0155 Steps: 8990, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000484, Sample Num: 7744, Cur Loss: 0.71268255, Cur Avg Loss: 1.20760709, Log Avg loss: 1.13074197, Global Avg Loss: 6.10086035, Time: 0.0140 Steps: 9000, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000494, Sample Num: 7904, Cur Loss: 1.03567863, Cur Avg Loss: 1.21081353, Log Avg loss: 1.36600549, Global Avg Loss: 6.09560524, Time: 0.0143 Steps: 9010, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000504, Sample Num: 8064, Cur Loss: 2.15694904, Cur Avg Loss: 1.21297601, Log Avg loss: 1.31980248, Global Avg Loss: 6.09031056, Time: 0.0137 Steps: 9020, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000514, Sample Num: 8224, Cur Loss: 1.13150513, Cur Avg Loss: 1.21203248, Log Avg loss: 1.16447855, Global Avg Loss: 6.08485559, Time: 0.0072 Steps: 9030, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000524, Sample Num: 8384, Cur Loss: 0.72976995, Cur Avg Loss: 1.21463140, Log Avg loss: 1.34821560, Global Avg Loss: 6.07961595, Time: 0.0117 Steps: 9040, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000534, Sample Num: 8544, Cur Loss: 1.44916821, Cur Avg Loss: 1.21467026, Log Avg loss: 1.21670656, Global Avg Loss: 6.07424257, Time: 0.0111 Steps: 9050, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000544, Sample Num: 8704, Cur Loss: 0.86103296, Cur Avg Loss: 1.21366869, Log Avg loss: 1.16018475, Global Avg Loss: 6.06881866, Time: 0.0095 Steps: 9060, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000554, Sample Num: 8864, Cur Loss: 1.11958146, Cur Avg Loss: 1.21306526, Log Avg loss: 1.18023867, Global Avg Loss: 6.06342883, Time: 0.0072 Steps: 9070, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000564, Sample Num: 9024, Cur Loss: 0.88110912, Cur Avg Loss: 1.20980937, Log Avg loss: 1.02943337, Global Avg Loss: 6.05788478, Time: 0.0143 Steps: 9080, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000574, Sample Num: 9184, Cur Loss: 1.15785766, Cur Avg Loss: 1.21281700, Log Avg loss: 1.38244734, Global Avg Loss: 6.05274128, Time: 0.0087 Steps: 9090, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000584, Sample Num: 9344, Cur Loss: 0.86227840, Cur Avg Loss: 1.21043721, Log Avg loss: 1.07383733, Global Avg Loss: 6.04726996, Time: 0.0143 Steps: 9100, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000594, Sample Num: 9504, Cur Loss: 0.59991157, Cur Avg Loss: 1.20816836, Log Avg loss: 1.07566712, Global Avg Loss: 6.04181266, Time: 0.0133 Steps: 9110, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000604, Sample Num: 9664, Cur Loss: 2.93328667, Cur Avg Loss: 1.20917979, Log Avg loss: 1.26925872, Global Avg Loss: 6.03657959, Time: 0.0143 Steps: 9120, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000614, Sample Num: 9824, Cur Loss: 0.91103446, Cur Avg Loss: 1.20839017, Log Avg loss: 1.16069763, Global Avg Loss: 6.03123909, Time: 0.0119 Steps: 9130, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000624, Sample Num: 9984, Cur Loss: 1.38868356, Cur Avg Loss: 1.20779755, Log Avg loss: 1.17141025, Global Avg Loss: 6.02592199, Time: 0.0072 Steps: 9140, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000634, Sample Num: 10144, Cur Loss: 1.36238492, Cur Avg Loss: 1.20820232, Log Avg loss: 1.23345991, Global Avg Loss: 6.02068433, Time: 0.0098 Steps: 9150, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000644, Sample Num: 10304, Cur Loss: 1.52234137, Cur Avg Loss: 1.20902172, Log Avg loss: 1.26097178, Global Avg Loss: 6.01548813, Time: 0.0072 Steps: 9160, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000654, Sample Num: 10464, Cur Loss: 0.86413658, Cur Avg Loss: 1.20836947, Log Avg loss: 1.16636444, Global Avg Loss: 6.01020010, Time: 0.0141 Steps: 9170, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000664, Sample Num: 10624, Cur Loss: 1.30945659, Cur Avg Loss: 1.20680401, Log Avg loss: 1.10442340, Global Avg Loss: 6.00485612, Time: 0.0164 Steps: 9180, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000674, Sample Num: 10784, Cur Loss: 0.23046878, Cur Avg Loss: 1.20443539, Log Avg loss: 1.04715875, Global Avg Loss: 5.99946145, Time: 0.0109 Steps: 9190, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000684, Sample Num: 10944, Cur Loss: 0.45000145, Cur Avg Loss: 1.19984635, Log Avg loss: 0.89054487, Global Avg Loss: 5.99390828, Time: 0.0065 Steps: 9200, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000694, Sample Num: 11104, Cur Loss: 1.57747960, Cur Avg Loss: 1.20018611, Log Avg loss: 1.22342606, Global Avg Loss: 5.98872861, Time: 0.0102 Steps: 9210, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000704, Sample Num: 11264, Cur Loss: 0.58822834, Cur Avg Loss: 1.20102842, Log Avg loss: 1.25948438, Global Avg Loss: 5.98359927, Time: 0.0108 Steps: 9220, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000714, Sample Num: 11424, Cur Loss: 0.67581964, Cur Avg Loss: 1.20041075, Log Avg loss: 1.15692704, Global Avg Loss: 5.97836994, Time: 0.0066 Steps: 9230, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000724, Sample Num: 11584, Cur Loss: 0.85438764, Cur Avg Loss: 1.19782132, Log Avg loss: 1.01293604, Global Avg Loss: 5.97299610, Time: 0.0084 Steps: 9240, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000734, Sample Num: 11744, Cur Loss: 1.64717126, Cur Avg Loss: 1.19837232, Log Avg loss: 1.23826503, Global Avg Loss: 5.96787747, Time: 0.0089 Steps: 9250, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000744, Sample Num: 11904, Cur Loss: 0.85295498, Cur Avg Loss: 1.20260638, Log Avg loss: 1.51338578, Global Avg Loss: 5.96306700, Time: 0.0067 Steps: 9260, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000754, Sample Num: 12064, Cur Loss: 0.56048375, Cur Avg Loss: 1.20046982, Log Avg loss: 1.04151010, Global Avg Loss: 5.95775788, Time: 0.0119 Steps: 9270, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000764, Sample Num: 12224, Cur Loss: 1.48076248, Cur Avg Loss: 1.20032500, Log Avg loss: 1.18940513, Global Avg Loss: 5.95261957, Time: 0.0188 Steps: 9280, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000774, Sample Num: 12384, Cur Loss: 1.18318474, Cur Avg Loss: 1.19727614, Log Avg loss: 0.96434382, Global Avg Loss: 5.94725006, Time: 0.0081 Steps: 9290, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000784, Sample Num: 12544, Cur Loss: 1.30602086, Cur Avg Loss: 1.19530912, Log Avg loss: 1.04306178, Global Avg Loss: 5.94197674, Time: 0.0067 Steps: 9300, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000794, Sample Num: 12704, Cur Loss: 1.42114067, Cur Avg Loss: 1.19544829, Log Avg loss: 1.20635920, Global Avg Loss: 5.93689015, Time: 0.0064 Steps: 9310, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000804, Sample Num: 12864, Cur Loss: 1.02605295, Cur Avg Loss: 1.19775371, Log Avg loss: 1.38080370, Global Avg Loss: 5.93200164, Time: 0.0219 Steps: 9320, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000814, Sample Num: 13024, Cur Loss: 1.22520423, Cur Avg Loss: 1.19891767, Log Avg loss: 1.29250014, Global Avg Loss: 5.92702897, Time: 0.0066 Steps: 9330, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000824, Sample Num: 13184, Cur Loss: 1.09090817, Cur Avg Loss: 1.19994072, Log Avg loss: 1.28321743, Global Avg Loss: 5.92205701, Time: 0.0067 Steps: 9340, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000834, Sample Num: 13344, Cur Loss: 1.73400736, Cur Avg Loss: 1.19884055, Log Avg loss: 1.10818583, Global Avg Loss: 5.91690848, Time: 0.0070 Steps: 9350, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000844, Sample Num: 13504, Cur Loss: 1.57443285, Cur Avg Loss: 1.19956583, Log Avg loss: 1.26005410, Global Avg Loss: 5.91193321, Time: 0.0074 Steps: 9360, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000854, Sample Num: 13664, Cur Loss: 0.78743148, Cur Avg Loss: 1.19793870, Log Avg loss: 1.06060945, Global Avg Loss: 5.90675571, Time: 0.0064 Steps: 9370, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000864, Sample Num: 13824, Cur Loss: 1.39043784, Cur Avg Loss: 1.19974864, Log Avg loss: 1.35431700, Global Avg Loss: 5.90190236, Time: 0.0072 Steps: 9380, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000874, Sample Num: 13984, Cur Loss: 1.51376915, Cur Avg Loss: 1.19897476, Log Avg loss: 1.13211199, Global Avg Loss: 5.89682271, Time: 0.0217 Steps: 9390, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000884, Sample Num: 14144, Cur Loss: 0.93599248, Cur Avg Loss: 1.19993080, Log Avg loss: 1.28348868, Global Avg Loss: 5.89191491, Time: 0.0072 Steps: 9400, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000894, Sample Num: 14304, Cur Loss: 2.05767298, Cur Avg Loss: 1.20170329, Log Avg loss: 1.35839103, Global Avg Loss: 5.88709714, Time: 0.0215 Steps: 9410, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000904, Sample Num: 14464, Cur Loss: 1.91530478, Cur Avg Loss: 1.19960325, Log Avg loss: 1.01186025, Global Avg Loss: 5.88192172, Time: 0.0067 Steps: 9420, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000914, Sample Num: 14624, Cur Loss: 1.24791789, Cur Avg Loss: 1.19932214, Log Avg loss: 1.17390999, Global Avg Loss: 5.87692914, Time: 0.0065 Steps: 9430, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000924, Sample Num: 14784, Cur Loss: 1.54478025, Cur Avg Loss: 1.20126365, Log Avg loss: 1.37871756, Global Avg Loss: 5.87216408, Time: 0.0076 Steps: 9440, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000934, Sample Num: 14944, Cur Loss: 2.16264391, Cur Avg Loss: 1.20020254, Log Avg loss: 1.10215542, Global Avg Loss: 5.86711645, Time: 0.0069 Steps: 9450, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000944, Sample Num: 15104, Cur Loss: 0.43889296, Cur Avg Loss: 1.19858006, Log Avg loss: 1.04704045, Global Avg Loss: 5.86202123, Time: 0.0173 Steps: 9460, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000954, Sample Num: 15264, Cur Loss: 1.55481756, Cur Avg Loss: 1.20267528, Log Avg loss: 1.58926404, Global Avg Loss: 5.85750935, Time: 0.0183 Steps: 9470, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000964, Sample Num: 15424, Cur Loss: 0.78912830, Cur Avg Loss: 1.20174104, Log Avg loss: 1.11261508, Global Avg Loss: 5.85250418, Time: 0.0067 Steps: 9480, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000974, Sample Num: 15584, Cur Loss: 0.76683080, Cur Avg Loss: 1.20218598, Log Avg loss: 1.24507806, Global Avg Loss: 5.84764915, Time: 0.0064 Steps: 9490, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000984, Sample Num: 15744, Cur Loss: 1.14039147, Cur Avg Loss: 1.20502995, Log Avg loss: 1.48203233, Global Avg Loss: 5.84305377, Time: 0.0068 Steps: 9500, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 000994, Sample Num: 15904, Cur Loss: 2.41206098, Cur Avg Loss: 1.20555614, Log Avg loss: 1.25733352, Global Avg Loss: 5.83823177, Time: 0.0065 Steps: 9510, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001004, Sample Num: 16064, Cur Loss: 1.42186141, Cur Avg Loss: 1.20611353, Log Avg loss: 1.26151824, Global Avg Loss: 5.83342430, Time: 0.0207 Steps: 9520, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001014, Sample Num: 16224, Cur Loss: 1.28614831, Cur Avg Loss: 1.20562654, Log Avg loss: 1.15673253, Global Avg Loss: 5.82851696, Time: 0.0113 Steps: 9530, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001024, Sample Num: 16384, Cur Loss: 1.45544744, Cur Avg Loss: 1.20847166, Log Avg loss: 1.49696680, Global Avg Loss: 5.82397655, Time: 0.0078 Steps: 9540, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001034, Sample Num: 16544, Cur Loss: 1.64413834, Cur Avg Loss: 1.20613779, Log Avg loss: 0.96714904, Global Avg Loss: 5.81889087, Time: 0.0140 Steps: 9550, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001044, Sample Num: 16704, Cur Loss: 1.31574512, Cur Avg Loss: 1.20848537, Log Avg loss: 1.45122560, Global Avg Loss: 5.81432218, Time: 0.0152 Steps: 9560, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001054, Sample Num: 16864, Cur Loss: 1.74388707, Cur Avg Loss: 1.20901845, Log Avg loss: 1.26467172, Global Avg Loss: 5.80956810, Time: 0.0119 Steps: 9570, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001064, Sample Num: 17024, Cur Loss: 0.96767652, Cur Avg Loss: 1.20896804, Log Avg loss: 1.20365489, Global Avg Loss: 5.80476026, Time: 0.0068 Steps: 9580, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001074, Sample Num: 17184, Cur Loss: 2.23735809, Cur Avg Loss: 1.21013846, Log Avg loss: 1.33467087, Global Avg Loss: 5.80009906, Time: 0.0122 Steps: 9590, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001084, Sample Num: 17344, Cur Loss: 0.61587548, Cur Avg Loss: 1.20756498, Log Avg loss: 0.93117348, Global Avg Loss: 5.79502727, Time: 0.0067 Steps: 9600, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001094, Sample Num: 17504, Cur Loss: 1.58409894, Cur Avg Loss: 1.21020589, Log Avg loss: 1.49648034, Global Avg Loss: 5.79055427, Time: 0.0067 Steps: 9610, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001104, Sample Num: 17664, Cur Loss: 0.64111745, Cur Avg Loss: 1.21228837, Log Avg loss: 1.44011150, Global Avg Loss: 5.78603198, Time: 0.0125 Steps: 9620, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001114, Sample Num: 17824, Cur Loss: 1.57804203, Cur Avg Loss: 1.21540524, Log Avg loss: 1.55950854, Global Avg Loss: 5.78164307, Time: 0.0072 Steps: 9630, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001124, Sample Num: 17984, Cur Loss: 1.02898765, Cur Avg Loss: 1.21490766, Log Avg loss: 1.15947748, Global Avg Loss: 5.77684829, Time: 0.0067 Steps: 9640, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001134, Sample Num: 18144, Cur Loss: 0.97160143, Cur Avg Loss: 1.21416636, Log Avg loss: 1.13084387, Global Avg Loss: 5.77203378, Time: 0.0129 Steps: 9650, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001144, Sample Num: 18304, Cur Loss: 1.07738590, Cur Avg Loss: 1.21346695, Log Avg loss: 1.13415411, Global Avg Loss: 5.76723266, Time: 0.0122 Steps: 9660, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001154, Sample Num: 18464, Cur Loss: 0.76055008, Cur Avg Loss: 1.21496281, Log Avg loss: 1.38608846, Global Avg Loss: 5.76270200, Time: 0.0226 Steps: 9670, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001164, Sample Num: 18624, Cur Loss: 0.92153031, Cur Avg Loss: 1.21260086, Log Avg loss: 0.94003245, Global Avg Loss: 5.75771991, Time: 0.0127 Steps: 9680, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001174, Sample Num: 18784, Cur Loss: 1.72016788, Cur Avg Loss: 1.21041195, Log Avg loss: 0.95562295, Global Avg Loss: 5.75276418, Time: 0.0099 Steps: 9690, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001184, Sample Num: 18944, Cur Loss: 2.00103569, Cur Avg Loss: 1.21294129, Log Avg loss: 1.50988572, Global Avg Loss: 5.74839008, Time: 0.0067 Steps: 9700, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001194, Sample Num: 19104, Cur Loss: 0.96771163, Cur Avg Loss: 1.21190294, Log Avg loss: 1.08896147, Global Avg Loss: 5.74359149, Time: 0.0108 Steps: 9710, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001204, Sample Num: 19264, Cur Loss: 0.78852761, Cur Avg Loss: 1.20978417, Log Avg loss: 0.95680323, Global Avg Loss: 5.73866682, Time: 0.0120 Steps: 9720, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001214, Sample Num: 19424, Cur Loss: 1.08019161, Cur Avg Loss: 1.21082977, Log Avg loss: 1.33671987, Global Avg Loss: 5.73414272, Time: 0.0073 Steps: 9730, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001224, Sample Num: 19584, Cur Loss: 0.97352129, Cur Avg Loss: 1.21227458, Log Avg loss: 1.38767470, Global Avg Loss: 5.72968022, Time: 0.0072 Steps: 9740, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001234, Sample Num: 19744, Cur Loss: 1.09843779, Cur Avg Loss: 1.21071768, Log Avg loss: 1.02015293, Global Avg Loss: 5.72484994, Time: 0.0068 Steps: 9750, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001244, Sample Num: 19904, Cur Loss: 1.13414407, Cur Avg Loss: 1.21122072, Log Avg loss: 1.27329612, Global Avg Loss: 5.72028892, Time: 0.0068 Steps: 9760, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001254, Sample Num: 20064, Cur Loss: 0.77163947, Cur Avg Loss: 1.21053888, Log Avg loss: 1.12571857, Global Avg Loss: 5.71558619, Time: 0.0098 Steps: 9770, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001264, Sample Num: 20224, Cur Loss: 0.85551268, Cur Avg Loss: 1.20956474, Log Avg loss: 1.08740723, Global Avg Loss: 5.71085390, Time: 0.0067 Steps: 9780, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001274, Sample Num: 20384, Cur Loss: 0.61753821, Cur Avg Loss: 1.21024994, Log Avg loss: 1.29685862, Global Avg Loss: 5.70634522, Time: 0.0067 Steps: 9790, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001284, Sample Num: 20544, Cur Loss: 0.96265697, Cur Avg Loss: 1.20934865, Log Avg loss: 1.09452528, Global Avg Loss: 5.70163928, Time: 0.0071 Steps: 9800, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001294, Sample Num: 20704, Cur Loss: 0.92645860, Cur Avg Loss: 1.20757850, Log Avg loss: 0.98029042, Global Avg Loss: 5.69682649, Time: 0.0108 Steps: 9810, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001304, Sample Num: 20864, Cur Loss: 0.93536991, Cur Avg Loss: 1.20841941, Log Avg loss: 1.31723384, Global Avg Loss: 5.69236662, Time: 0.0123 Steps: 9820, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001314, Sample Num: 21024, Cur Loss: 1.29145646, Cur Avg Loss: 1.20904674, Log Avg loss: 1.29084963, Global Avg Loss: 5.68788898, Time: 0.0116 Steps: 9830, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001324, Sample Num: 21184, Cur Loss: 0.74894273, Cur Avg Loss: 1.20757776, Log Avg loss: 1.01455401, Global Avg Loss: 5.68313966, Time: 0.0118 Steps: 9840, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001334, Sample Num: 21344, Cur Loss: 1.18563223, Cur Avg Loss: 1.20670668, Log Avg loss: 1.09137557, Global Avg Loss: 5.67847797, Time: 0.0067 Steps: 9850, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001344, Sample Num: 21504, Cur Loss: 2.21322918, Cur Avg Loss: 1.20796021, Log Avg loss: 1.37518165, Global Avg Loss: 5.67411357, Time: 0.0121 Steps: 9860, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001354, Sample Num: 21664, Cur Loss: 1.17580581, Cur Avg Loss: 1.20769060, Log Avg loss: 1.17145470, Global Avg Loss: 5.66955161, Time: 0.0067 Steps: 9870, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001364, Sample Num: 21824, Cur Loss: 0.63948530, Cur Avg Loss: 1.20743810, Log Avg loss: 1.17325014, Global Avg Loss: 5.66500070, Time: 0.0067 Steps: 9880, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001374, Sample Num: 21984, Cur Loss: 2.00136423, Cur Avg Loss: 1.20820933, Log Avg loss: 1.31340541, Global Avg Loss: 5.66060070, Time: 0.0105 Steps: 9890, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001384, Sample Num: 22144, Cur Loss: 1.95705891, Cur Avg Loss: 1.20773108, Log Avg loss: 1.14201912, Global Avg Loss: 5.65603648, Time: 0.0090 Steps: 9900, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001394, Sample Num: 22304, Cur Loss: 0.89131725, Cur Avg Loss: 1.20756598, Log Avg loss: 1.18471592, Global Avg Loss: 5.65152455, Time: 0.0068 Steps: 9910, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001404, Sample Num: 22464, Cur Loss: 0.94999808, Cur Avg Loss: 1.20642646, Log Avg loss: 1.04757789, Global Avg Loss: 5.64688347, Time: 0.0072 Steps: 9920, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001414, Sample Num: 22624, Cur Loss: 0.51375651, Cur Avg Loss: 1.20447549, Log Avg loss: 0.93055938, Global Avg Loss: 5.64213390, Time: 0.0122 Steps: 9930, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001424, Sample Num: 22784, Cur Loss: 1.17629576, Cur Avg Loss: 1.20527752, Log Avg loss: 1.31868394, Global Avg Loss: 5.63778436, Time: 0.0122 Steps: 9940, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001434, Sample Num: 22944, Cur Loss: 0.72533464, Cur Avg Loss: 1.20551268, Log Avg loss: 1.23899969, Global Avg Loss: 5.63336347, Time: 0.0071 Steps: 9950, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001444, Sample Num: 23104, Cur Loss: 2.27328110, Cur Avg Loss: 1.20638471, Log Avg loss: 1.33143403, Global Avg Loss: 5.62904426, Time: 0.0068 Steps: 9960, Updated lr: 0.000092 Training, Epoch: 0005, Batch: 001454, Sample Num: 23264, Cur Loss: 0.64153528, Cur Avg Loss: 1.20765676, Log Avg loss: 1.39134022, Global Avg Loss: 5.62479380, Time: 0.0067 Steps: 9970, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001464, Sample Num: 23424, Cur Loss: 1.43792212, Cur Avg Loss: 1.20842482, Log Avg loss: 1.32010134, Global Avg Loss: 5.62048049, Time: 0.0067 Steps: 9980, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001474, Sample Num: 23584, Cur Loss: 0.75898409, Cur Avg Loss: 1.20858097, Log Avg loss: 1.23144089, Global Avg Loss: 5.61608705, Time: 0.0227 Steps: 9990, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001484, Sample Num: 23744, Cur Loss: 1.96941376, Cur Avg Loss: 1.21083230, Log Avg loss: 1.54267809, Global Avg Loss: 5.61201364, Time: 0.0067 Steps: 10000, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001494, Sample Num: 23904, Cur Loss: 0.87412786, Cur Avg Loss: 1.20944815, Log Avg loss: 1.00404058, Global Avg Loss: 5.60741027, Time: 0.0069 Steps: 10010, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001504, Sample Num: 24064, Cur Loss: 0.87347770, Cur Avg Loss: 1.20884445, Log Avg loss: 1.11865230, Global Avg Loss: 5.60293048, Time: 0.0073 Steps: 10020, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001514, Sample Num: 24224, Cur Loss: 0.86359543, Cur Avg Loss: 1.20843746, Log Avg loss: 1.14722622, Global Avg Loss: 5.59848810, Time: 0.0075 Steps: 10030, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001524, Sample Num: 24384, Cur Loss: 1.30636060, Cur Avg Loss: 1.20883409, Log Avg loss: 1.26888369, Global Avg Loss: 5.59417574, Time: 0.0067 Steps: 10040, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001534, Sample Num: 24544, Cur Loss: 0.97467947, Cur Avg Loss: 1.20748830, Log Avg loss: 1.00239010, Global Avg Loss: 5.58960680, Time: 0.0070 Steps: 10050, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001544, Sample Num: 24704, Cur Loss: 0.43443614, Cur Avg Loss: 1.20655173, Log Avg loss: 1.06288070, Global Avg Loss: 5.58510707, Time: 0.0111 Steps: 10060, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001554, Sample Num: 24864, Cur Loss: 0.54563308, Cur Avg Loss: 1.20749782, Log Avg loss: 1.35357481, Global Avg Loss: 5.58090496, Time: 0.0068 Steps: 10070, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001564, Sample Num: 25024, Cur Loss: 1.07730889, Cur Avg Loss: 1.20683392, Log Avg loss: 1.10366314, Global Avg Loss: 5.57646325, Time: 0.0127 Steps: 10080, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001574, Sample Num: 25184, Cur Loss: 1.19941664, Cur Avg Loss: 1.20606016, Log Avg loss: 1.08504444, Global Avg Loss: 5.57201189, Time: 0.0072 Steps: 10090, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001584, Sample Num: 25344, Cur Loss: 0.50636971, Cur Avg Loss: 1.20425853, Log Avg loss: 0.92068181, Global Avg Loss: 5.56740662, Time: 0.0117 Steps: 10100, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001594, Sample Num: 25504, Cur Loss: 0.89508295, Cur Avg Loss: 1.20318362, Log Avg loss: 1.03291904, Global Avg Loss: 5.56292146, Time: 0.0111 Steps: 10110, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001604, Sample Num: 25664, Cur Loss: 1.15118861, Cur Avg Loss: 1.20302087, Log Avg loss: 1.17707829, Global Avg Loss: 5.55858763, Time: 0.0068 Steps: 10120, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001614, Sample Num: 25824, Cur Loss: 1.39265132, Cur Avg Loss: 1.20263388, Log Avg loss: 1.14056091, Global Avg Loss: 5.55422630, Time: 0.0237 Steps: 10130, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001624, Sample Num: 25984, Cur Loss: 1.34226131, Cur Avg Loss: 1.20264552, Log Avg loss: 1.20452302, Global Avg Loss: 5.54993665, Time: 0.0118 Steps: 10140, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001634, Sample Num: 26144, Cur Loss: 0.78096688, Cur Avg Loss: 1.20279342, Log Avg loss: 1.22681302, Global Avg Loss: 5.54567741, Time: 0.0135 Steps: 10150, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001644, Sample Num: 26304, Cur Loss: 0.97019517, Cur Avg Loss: 1.20212208, Log Avg loss: 1.09242538, Global Avg Loss: 5.54129429, Time: 0.0067 Steps: 10160, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001654, Sample Num: 26464, Cur Loss: 1.93866980, Cur Avg Loss: 1.20135272, Log Avg loss: 1.07487006, Global Avg Loss: 5.53690253, Time: 0.0118 Steps: 10170, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001664, Sample Num: 26624, Cur Loss: 0.74386990, Cur Avg Loss: 1.20208341, Log Avg loss: 1.32293943, Global Avg Loss: 5.53276308, Time: 0.0067 Steps: 10180, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001674, Sample Num: 26784, Cur Loss: 1.50658357, Cur Avg Loss: 1.20140748, Log Avg loss: 1.08893213, Global Avg Loss: 5.52840210, Time: 0.0113 Steps: 10190, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001684, Sample Num: 26944, Cur Loss: 2.39902878, Cur Avg Loss: 1.20168674, Log Avg loss: 1.24843513, Global Avg Loss: 5.52420606, Time: 0.0121 Steps: 10200, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001694, Sample Num: 27104, Cur Loss: 0.96142805, Cur Avg Loss: 1.20191024, Log Avg loss: 1.23954772, Global Avg Loss: 5.52000953, Time: 0.0083 Steps: 10210, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001704, Sample Num: 27264, Cur Loss: 1.25451565, Cur Avg Loss: 1.20058338, Log Avg loss: 0.97581382, Global Avg Loss: 5.51556315, Time: 0.0065 Steps: 10220, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001714, Sample Num: 27424, Cur Loss: 1.90047038, Cur Avg Loss: 1.20180085, Log Avg loss: 1.40925711, Global Avg Loss: 5.51154917, Time: 0.0110 Steps: 10230, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001724, Sample Num: 27584, Cur Loss: 1.55670536, Cur Avg Loss: 1.20080820, Log Avg loss: 1.03066707, Global Avg Loss: 5.50717330, Time: 0.0222 Steps: 10240, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001734, Sample Num: 27744, Cur Loss: 1.49366665, Cur Avg Loss: 1.19980818, Log Avg loss: 1.02740558, Global Avg Loss: 5.50280280, Time: 0.0226 Steps: 10250, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001744, Sample Num: 27904, Cur Loss: 0.99391919, Cur Avg Loss: 1.20008415, Log Avg loss: 1.24793772, Global Avg Loss: 5.49865576, Time: 0.0064 Steps: 10260, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001754, Sample Num: 28064, Cur Loss: 1.42440128, Cur Avg Loss: 1.20004338, Log Avg loss: 1.19293248, Global Avg Loss: 5.49446323, Time: 0.0139 Steps: 10270, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001764, Sample Num: 28224, Cur Loss: 0.41936547, Cur Avg Loss: 1.19865915, Log Avg loss: 0.95586540, Global Avg Loss: 5.49004825, Time: 0.0217 Steps: 10280, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001774, Sample Num: 28384, Cur Loss: 0.52631545, Cur Avg Loss: 1.19876764, Log Avg loss: 1.21790606, Global Avg Loss: 5.48589651, Time: 0.0067 Steps: 10290, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001784, Sample Num: 28544, Cur Loss: 1.46946502, Cur Avg Loss: 1.19940753, Log Avg loss: 1.31292297, Global Avg Loss: 5.48184508, Time: 0.0067 Steps: 10300, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001794, Sample Num: 28704, Cur Loss: 1.17387569, Cur Avg Loss: 1.19903119, Log Avg loss: 1.13189195, Global Avg Loss: 5.47762592, Time: 0.0140 Steps: 10310, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001804, Sample Num: 28864, Cur Loss: 1.74934804, Cur Avg Loss: 1.19967192, Log Avg loss: 1.31461891, Global Avg Loss: 5.47359200, Time: 0.0114 Steps: 10320, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001814, Sample Num: 29024, Cur Loss: 1.07355332, Cur Avg Loss: 1.20027256, Log Avg loss: 1.30862764, Global Avg Loss: 5.46956009, Time: 0.0104 Steps: 10330, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001824, Sample Num: 29184, Cur Loss: 3.04040003, Cur Avg Loss: 1.20097521, Log Avg loss: 1.32843683, Global Avg Loss: 5.46555513, Time: 0.0113 Steps: 10340, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001834, Sample Num: 29344, Cur Loss: 1.36159599, Cur Avg Loss: 1.20135665, Log Avg loss: 1.27093043, Global Avg Loss: 5.46150236, Time: 0.0068 Steps: 10350, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001844, Sample Num: 29504, Cur Loss: 1.45468187, Cur Avg Loss: 1.20075936, Log Avg loss: 1.09121744, Global Avg Loss: 5.45728394, Time: 0.0168 Steps: 10360, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001854, Sample Num: 29664, Cur Loss: 1.63376832, Cur Avg Loss: 1.20068319, Log Avg loss: 1.18663653, Global Avg Loss: 5.45316566, Time: 0.0117 Steps: 10370, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001864, Sample Num: 29824, Cur Loss: 1.13027728, Cur Avg Loss: 1.20156413, Log Avg loss: 1.36489137, Global Avg Loss: 5.44922706, Time: 0.0073 Steps: 10380, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001874, Sample Num: 29984, Cur Loss: 1.21213233, Cur Avg Loss: 1.20108034, Log Avg loss: 1.11090175, Global Avg Loss: 5.44505158, Time: 0.0067 Steps: 10390, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001884, Sample Num: 30144, Cur Loss: 0.72303879, Cur Avg Loss: 1.20093622, Log Avg loss: 1.17392864, Global Avg Loss: 5.44094473, Time: 0.0111 Steps: 10400, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001894, Sample Num: 30304, Cur Loss: 1.24876857, Cur Avg Loss: 1.20014840, Log Avg loss: 1.05172216, Global Avg Loss: 5.43672837, Time: 0.0067 Steps: 10410, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001904, Sample Num: 30464, Cur Loss: 1.60698771, Cur Avg Loss: 1.19953769, Log Avg loss: 1.08386883, Global Avg Loss: 5.43255097, Time: 0.0066 Steps: 10420, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001914, Sample Num: 30624, Cur Loss: 0.61854792, Cur Avg Loss: 1.19931785, Log Avg loss: 1.15746048, Global Avg Loss: 5.42845213, Time: 0.0066 Steps: 10430, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001924, Sample Num: 30784, Cur Loss: 0.59297734, Cur Avg Loss: 1.19839261, Log Avg loss: 1.02130138, Global Avg Loss: 5.42423072, Time: 0.0123 Steps: 10440, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001934, Sample Num: 30944, Cur Loss: 1.00935698, Cur Avg Loss: 1.19727838, Log Avg loss: 0.98290093, Global Avg Loss: 5.41998064, Time: 0.0119 Steps: 10450, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001944, Sample Num: 31104, Cur Loss: 1.30317998, Cur Avg Loss: 1.19764727, Log Avg loss: 1.26899180, Global Avg Loss: 5.41601220, Time: 0.0231 Steps: 10460, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001954, Sample Num: 31264, Cur Loss: 1.04718411, Cur Avg Loss: 1.19756074, Log Avg loss: 1.18073812, Global Avg Loss: 5.41196705, Time: 0.0110 Steps: 10470, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001964, Sample Num: 31424, Cur Loss: 1.20235026, Cur Avg Loss: 1.19737942, Log Avg loss: 1.16194965, Global Avg Loss: 5.40791169, Time: 0.0071 Steps: 10480, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001974, Sample Num: 31584, Cur Loss: 0.80143154, Cur Avg Loss: 1.19640162, Log Avg loss: 1.00436185, Global Avg Loss: 5.40371383, Time: 0.0123 Steps: 10490, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001984, Sample Num: 31744, Cur Loss: 1.07832789, Cur Avg Loss: 1.19719992, Log Avg loss: 1.35478416, Global Avg Loss: 5.39985771, Time: 0.0131 Steps: 10500, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 001994, Sample Num: 31904, Cur Loss: 0.60144871, Cur Avg Loss: 1.19623121, Log Avg loss: 1.00403925, Global Avg Loss: 5.39567520, Time: 0.0094 Steps: 10510, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002004, Sample Num: 32064, Cur Loss: 0.85382152, Cur Avg Loss: 1.19675341, Log Avg loss: 1.30088021, Global Avg Loss: 5.39178281, Time: 0.0118 Steps: 10520, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002014, Sample Num: 32224, Cur Loss: 0.88369542, Cur Avg Loss: 1.19825445, Log Avg loss: 1.49906265, Global Avg Loss: 5.38808602, Time: 0.0116 Steps: 10530, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002024, Sample Num: 32384, Cur Loss: 1.51106262, Cur Avg Loss: 1.19768842, Log Avg loss: 1.08368929, Global Avg Loss: 5.38400215, Time: 0.0066 Steps: 10540, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002034, Sample Num: 32544, Cur Loss: 2.21506262, Cur Avg Loss: 1.19821594, Log Avg loss: 1.30498637, Global Avg Loss: 5.38013578, Time: 0.0113 Steps: 10550, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002044, Sample Num: 32704, Cur Loss: 1.32028270, Cur Avg Loss: 1.19792694, Log Avg loss: 1.13914478, Global Avg Loss: 5.37611969, Time: 0.0105 Steps: 10560, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002054, Sample Num: 32864, Cur Loss: 1.32937396, Cur Avg Loss: 1.19679863, Log Avg loss: 0.96617258, Global Avg Loss: 5.37194756, Time: 0.0127 Steps: 10570, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002064, Sample Num: 33024, Cur Loss: 0.43058935, Cur Avg Loss: 1.19568032, Log Avg loss: 0.96597926, Global Avg Loss: 5.36778313, Time: 0.0066 Steps: 10580, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002074, Sample Num: 33184, Cur Loss: 0.54616070, Cur Avg Loss: 1.19535095, Log Avg loss: 1.12736847, Global Avg Loss: 5.36377896, Time: 0.0066 Steps: 10590, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002084, Sample Num: 33344, Cur Loss: 1.35696042, Cur Avg Loss: 1.19508000, Log Avg loss: 1.13888445, Global Avg Loss: 5.35979321, Time: 0.0070 Steps: 10600, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002094, Sample Num: 33504, Cur Loss: 0.63677806, Cur Avg Loss: 1.19391398, Log Avg loss: 0.95091698, Global Avg Loss: 5.35563781, Time: 0.0233 Steps: 10610, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002104, Sample Num: 33664, Cur Loss: 1.80764174, Cur Avg Loss: 1.19370414, Log Avg loss: 1.14976283, Global Avg Loss: 5.35167748, Time: 0.0116 Steps: 10620, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002114, Sample Num: 33824, Cur Loss: 1.67058253, Cur Avg Loss: 1.19388211, Log Avg loss: 1.23132593, Global Avg Loss: 5.34780132, Time: 0.0116 Steps: 10630, Updated lr: 0.000091 Training, Epoch: 0005, Batch: 002124, Sample Num: 33984, Cur Loss: 0.80045176, Cur Avg Loss: 1.19369338, Log Avg loss: 1.15379728, Global Avg Loss: 5.34385959, Time: 0.0186 Steps: 10640, Updated lr: 0.000091 ***** Running evaluation checkpoint-10645 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-10645 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.451749, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.195594, "eval_total_loss": 840.502274, "eval_mae": 0.923115, "eval_mse": 1.195713, "eval_r2": 0.239926, "eval_sp_statistic": 0.493739, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.525122, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 3.985929, "test_total_loss": 2000.936486, "test_mae": 1.853947, "test_mse": 3.985212, "test_r2": -1.572091, "test_sp_statistic": 0.182185, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.305377, "test_ps_pvalue": 0.0, "lr": 9.085348506401138e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 5.34179644760824, "train_cur_epoch_loss": 2540.1618769839406, "train_cur_epoch_avg_loss": 1.193124413801757, "train_cur_epoch_time": 22.451748609542847, "train_cur_epoch_avg_time": 0.010545678069301479, "epoch": 5, "step": 10645} ################################################## Training, Epoch: 0006, Batch: 000005, Sample Num: 80, Cur Loss: 0.46041411, Cur Avg Loss: 1.27659562, Log Avg loss: 1.11401117, Global Avg Loss: 5.33988790, Time: 0.0090 Steps: 10650, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000015, Sample Num: 240, Cur Loss: 0.82450414, Cur Avg Loss: 1.11670267, Log Avg loss: 1.03675620, Global Avg Loss: 5.33585119, Time: 0.0117 Steps: 10660, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000025, Sample Num: 400, Cur Loss: 1.11638880, Cur Avg Loss: 1.09148057, Log Avg loss: 1.05364742, Global Avg Loss: 5.33183788, Time: 0.0064 Steps: 10670, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000035, Sample Num: 560, Cur Loss: 0.75643110, Cur Avg Loss: 1.07223949, Log Avg loss: 1.02413677, Global Avg Loss: 5.32780445, Time: 0.0120 Steps: 10680, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000045, Sample Num: 720, Cur Loss: 1.25399792, Cur Avg Loss: 1.08090948, Log Avg loss: 1.11125446, Global Avg Loss: 5.32386007, Time: 0.0067 Steps: 10690, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000055, Sample Num: 880, Cur Loss: 1.78740501, Cur Avg Loss: 1.06957153, Log Avg loss: 1.01855077, Global Avg Loss: 5.31983641, Time: 0.0066 Steps: 10700, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000065, Sample Num: 1040, Cur Loss: 1.68713880, Cur Avg Loss: 1.10356486, Log Avg loss: 1.29052815, Global Avg Loss: 5.31607422, Time: 0.0114 Steps: 10710, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000075, Sample Num: 1200, Cur Loss: 1.11256099, Cur Avg Loss: 1.11345488, Log Avg loss: 1.17774001, Global Avg Loss: 5.31221383, Time: 0.0111 Steps: 10720, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000085, Sample Num: 1360, Cur Loss: 1.12889910, Cur Avg Loss: 1.12702596, Log Avg loss: 1.22880907, Global Avg Loss: 5.30840824, Time: 0.0066 Steps: 10730, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000095, Sample Num: 1520, Cur Loss: 0.90276068, Cur Avg Loss: 1.12923745, Log Avg loss: 1.14803512, Global Avg Loss: 5.30453452, Time: 0.0105 Steps: 10740, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000105, Sample Num: 1680, Cur Loss: 1.17294955, Cur Avg Loss: 1.12611230, Log Avg loss: 1.09642335, Global Avg Loss: 5.30062000, Time: 0.0118 Steps: 10750, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000115, Sample Num: 1840, Cur Loss: 1.12217259, Cur Avg Loss: 1.11436937, Log Avg loss: 0.99106866, Global Avg Loss: 5.29661484, Time: 0.0117 Steps: 10760, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000125, Sample Num: 2000, Cur Loss: 1.13024318, Cur Avg Loss: 1.13025386, Log Avg loss: 1.31292546, Global Avg Loss: 5.29291596, Time: 0.0104 Steps: 10770, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000135, Sample Num: 2160, Cur Loss: 0.64583594, Cur Avg Loss: 1.14285678, Log Avg loss: 1.30039323, Global Avg Loss: 5.28921232, Time: 0.0064 Steps: 10780, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000145, Sample Num: 2320, Cur Loss: 1.11386371, Cur Avg Loss: 1.15173303, Log Avg loss: 1.27156252, Global Avg Loss: 5.28548883, Time: 0.0107 Steps: 10790, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000155, Sample Num: 2480, Cur Loss: 1.54217792, Cur Avg Loss: 1.15978828, Log Avg loss: 1.27658941, Global Avg Loss: 5.28177689, Time: 0.0070 Steps: 10800, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000165, Sample Num: 2640, Cur Loss: 0.26900348, Cur Avg Loss: 1.14031644, Log Avg loss: 0.83850288, Global Avg Loss: 5.27766655, Time: 0.0129 Steps: 10810, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000175, Sample Num: 2800, Cur Loss: 0.74624401, Cur Avg Loss: 1.13837243, Log Avg loss: 1.10629621, Global Avg Loss: 5.27381131, Time: 0.0122 Steps: 10820, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000185, Sample Num: 2960, Cur Loss: 1.23004198, Cur Avg Loss: 1.14384646, Log Avg loss: 1.23964210, Global Avg Loss: 5.27008631, Time: 0.0065 Steps: 10830, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000195, Sample Num: 3120, Cur Loss: 0.43584371, Cur Avg Loss: 1.13067276, Log Avg loss: 0.88695914, Global Avg Loss: 5.26604284, Time: 0.0226 Steps: 10840, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000205, Sample Num: 3280, Cur Loss: 1.29446781, Cur Avg Loss: 1.13887881, Log Avg loss: 1.29889683, Global Avg Loss: 5.26238648, Time: 0.0066 Steps: 10850, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000215, Sample Num: 3440, Cur Loss: 1.12678313, Cur Avg Loss: 1.14056493, Log Avg loss: 1.17513042, Global Avg Loss: 5.25862290, Time: 0.0105 Steps: 10860, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000225, Sample Num: 3600, Cur Loss: 2.15573454, Cur Avg Loss: 1.14305882, Log Avg loss: 1.19667754, Global Avg Loss: 5.25488606, Time: 0.0163 Steps: 10870, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000235, Sample Num: 3760, Cur Loss: 0.52555341, Cur Avg Loss: 1.13810741, Log Avg loss: 1.02670060, Global Avg Loss: 5.25099986, Time: 0.0069 Steps: 10880, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000245, Sample Num: 3920, Cur Loss: 1.11425328, Cur Avg Loss: 1.13962185, Log Avg loss: 1.17521118, Global Avg Loss: 5.24725717, Time: 0.0133 Steps: 10890, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000255, Sample Num: 4080, Cur Loss: 1.18939257, Cur Avg Loss: 1.15511016, Log Avg loss: 1.53457370, Global Avg Loss: 5.24385103, Time: 0.0109 Steps: 10900, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000265, Sample Num: 4240, Cur Loss: 0.41895831, Cur Avg Loss: 1.15699573, Log Avg loss: 1.20507773, Global Avg Loss: 5.24014913, Time: 0.0067 Steps: 10910, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000275, Sample Num: 4400, Cur Loss: 0.80941534, Cur Avg Loss: 1.15402849, Log Avg loss: 1.07539666, Global Avg Loss: 5.23633526, Time: 0.0068 Steps: 10920, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000285, Sample Num: 4560, Cur Loss: 1.40366912, Cur Avg Loss: 1.17323914, Log Avg loss: 1.70153217, Global Avg Loss: 5.23310122, Time: 0.0064 Steps: 10930, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000295, Sample Num: 4720, Cur Loss: 2.16431117, Cur Avg Loss: 1.17668132, Log Avg loss: 1.27478349, Global Avg Loss: 5.22948301, Time: 0.0080 Steps: 10940, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000305, Sample Num: 4880, Cur Loss: 0.57429647, Cur Avg Loss: 1.17757838, Log Avg loss: 1.20404158, Global Avg Loss: 5.22580681, Time: 0.0109 Steps: 10950, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000315, Sample Num: 5040, Cur Loss: 1.24086237, Cur Avg Loss: 1.17569859, Log Avg loss: 1.11836506, Global Avg Loss: 5.22205915, Time: 0.0091 Steps: 10960, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000325, Sample Num: 5200, Cur Loss: 1.37471914, Cur Avg Loss: 1.17665035, Log Avg loss: 1.20663072, Global Avg Loss: 5.21839877, Time: 0.0104 Steps: 10970, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000335, Sample Num: 5360, Cur Loss: 1.23426437, Cur Avg Loss: 1.16890112, Log Avg loss: 0.91705126, Global Avg Loss: 5.21448134, Time: 0.0131 Steps: 10980, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000345, Sample Num: 5520, Cur Loss: 0.95388705, Cur Avg Loss: 1.16872615, Log Avg loss: 1.16286438, Global Avg Loss: 5.21079470, Time: 0.0065 Steps: 10990, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000355, Sample Num: 5680, Cur Loss: 0.30496892, Cur Avg Loss: 1.15882042, Log Avg loss: 0.81707280, Global Avg Loss: 5.20680040, Time: 0.0071 Steps: 11000, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000365, Sample Num: 5840, Cur Loss: 1.39328313, Cur Avg Loss: 1.16420938, Log Avg loss: 1.35551739, Global Avg Loss: 5.20330242, Time: 0.0114 Steps: 11010, Updated lr: 0.000091 Training, Epoch: 0006, Batch: 000375, Sample Num: 6000, Cur Loss: 1.08305097, Cur Avg Loss: 1.16700638, Log Avg loss: 1.26909705, Global Avg Loss: 5.19973236, Time: 0.0113 Steps: 11020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000385, Sample Num: 6160, Cur Loss: 1.69050705, Cur Avg Loss: 1.16590261, Log Avg loss: 1.12451110, Global Avg Loss: 5.19603769, Time: 0.0068 Steps: 11030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000395, Sample Num: 6320, Cur Loss: 1.14800799, Cur Avg Loss: 1.16186321, Log Avg loss: 1.00634624, Global Avg Loss: 5.19224268, Time: 0.0131 Steps: 11040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000405, Sample Num: 6480, Cur Loss: 1.10979867, Cur Avg Loss: 1.15555208, Log Avg loss: 0.90626259, Global Avg Loss: 5.18836396, Time: 0.0067 Steps: 11050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000415, Sample Num: 6640, Cur Loss: 1.09851980, Cur Avg Loss: 1.15261793, Log Avg loss: 1.03378496, Global Avg Loss: 5.18460756, Time: 0.0118 Steps: 11060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000425, Sample Num: 6800, Cur Loss: 0.46996966, Cur Avg Loss: 1.14936989, Log Avg loss: 1.01457617, Global Avg Loss: 5.18084060, Time: 0.0112 Steps: 11070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000435, Sample Num: 6960, Cur Loss: 1.79518557, Cur Avg Loss: 1.15320020, Log Avg loss: 1.31598828, Global Avg Loss: 5.17735246, Time: 0.0083 Steps: 11080, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000445, Sample Num: 7120, Cur Loss: 0.96401322, Cur Avg Loss: 1.15019092, Log Avg loss: 1.01928711, Global Avg Loss: 5.17360308, Time: 0.0090 Steps: 11090, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000455, Sample Num: 7280, Cur Loss: 0.96727526, Cur Avg Loss: 1.14715829, Log Avg loss: 1.01220637, Global Avg Loss: 5.16985407, Time: 0.0134 Steps: 11100, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000465, Sample Num: 7440, Cur Loss: 1.68088865, Cur Avg Loss: 1.14626556, Log Avg loss: 1.10564634, Global Avg Loss: 5.16619592, Time: 0.0128 Steps: 11110, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000475, Sample Num: 7600, Cur Loss: 2.04518723, Cur Avg Loss: 1.14502632, Log Avg loss: 1.08740195, Global Avg Loss: 5.16252794, Time: 0.0107 Steps: 11120, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000485, Sample Num: 7760, Cur Loss: 1.59944010, Cur Avg Loss: 1.14254617, Log Avg loss: 1.02473902, Global Avg Loss: 5.15881025, Time: 0.0120 Steps: 11130, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000495, Sample Num: 7920, Cur Loss: 1.24637115, Cur Avg Loss: 1.14639705, Log Avg loss: 1.33316466, Global Avg Loss: 5.15537610, Time: 0.0136 Steps: 11140, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000505, Sample Num: 8080, Cur Loss: 2.23206878, Cur Avg Loss: 1.14918985, Log Avg loss: 1.28743332, Global Avg Loss: 5.15190709, Time: 0.0123 Steps: 11150, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000515, Sample Num: 8240, Cur Loss: 1.42639494, Cur Avg Loss: 1.14968762, Log Avg loss: 1.17482497, Global Avg Loss: 5.14834340, Time: 0.0113 Steps: 11160, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000525, Sample Num: 8400, Cur Loss: 2.16354775, Cur Avg Loss: 1.15413110, Log Avg loss: 1.38297048, Global Avg Loss: 5.14497243, Time: 0.0122 Steps: 11170, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000535, Sample Num: 8560, Cur Loss: 1.12231505, Cur Avg Loss: 1.15087429, Log Avg loss: 0.97989162, Global Avg Loss: 5.14124695, Time: 0.0068 Steps: 11180, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000545, Sample Num: 8720, Cur Loss: 1.56790709, Cur Avg Loss: 1.14951976, Log Avg loss: 1.07705231, Global Avg Loss: 5.13761496, Time: 0.0073 Steps: 11190, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000555, Sample Num: 8880, Cur Loss: 1.08806098, Cur Avg Loss: 1.14731127, Log Avg loss: 1.02694865, Global Avg Loss: 5.13394473, Time: 0.0118 Steps: 11200, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000565, Sample Num: 9040, Cur Loss: 0.53040671, Cur Avg Loss: 1.14477555, Log Avg loss: 1.00404331, Global Avg Loss: 5.13026060, Time: 0.0112 Steps: 11210, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000575, Sample Num: 9200, Cur Loss: 0.51415157, Cur Avg Loss: 1.13696363, Log Avg loss: 0.69559014, Global Avg Loss: 5.12630813, Time: 0.0073 Steps: 11220, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000585, Sample Num: 9360, Cur Loss: 1.05379939, Cur Avg Loss: 1.14122374, Log Avg loss: 1.38617985, Global Avg Loss: 5.12297766, Time: 0.0071 Steps: 11230, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000595, Sample Num: 9520, Cur Loss: 1.13068175, Cur Avg Loss: 1.13768920, Log Avg loss: 0.93091864, Global Avg Loss: 5.11924807, Time: 0.0135 Steps: 11240, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000605, Sample Num: 9680, Cur Loss: 1.20435596, Cur Avg Loss: 1.14193997, Log Avg loss: 1.39486058, Global Avg Loss: 5.11593750, Time: 0.0153 Steps: 11250, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000615, Sample Num: 9840, Cur Loss: 0.95566916, Cur Avg Loss: 1.14187367, Log Avg loss: 1.13786246, Global Avg Loss: 5.11240457, Time: 0.0067 Steps: 11260, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000625, Sample Num: 10000, Cur Loss: 0.86041725, Cur Avg Loss: 1.14380933, Log Avg loss: 1.26285298, Global Avg Loss: 5.10898882, Time: 0.0069 Steps: 11270, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000635, Sample Num: 10160, Cur Loss: 1.07026505, Cur Avg Loss: 1.14553415, Log Avg loss: 1.25333515, Global Avg Loss: 5.10557069, Time: 0.0115 Steps: 11280, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000645, Sample Num: 10320, Cur Loss: 4.05571079, Cur Avg Loss: 1.15202350, Log Avg loss: 1.56409696, Global Avg Loss: 5.10243387, Time: 0.0113 Steps: 11290, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000655, Sample Num: 10480, Cur Loss: 1.64913797, Cur Avg Loss: 1.15336318, Log Avg loss: 1.23977249, Global Avg Loss: 5.09901558, Time: 0.0130 Steps: 11300, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000665, Sample Num: 10640, Cur Loss: 0.68612301, Cur Avg Loss: 1.15347064, Log Avg loss: 1.16050951, Global Avg Loss: 5.09553326, Time: 0.0068 Steps: 11310, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000675, Sample Num: 10800, Cur Loss: 1.03757095, Cur Avg Loss: 1.15262946, Log Avg loss: 1.09669079, Global Avg Loss: 5.09200071, Time: 0.0066 Steps: 11320, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000685, Sample Num: 10960, Cur Loss: 0.81500781, Cur Avg Loss: 1.14986365, Log Avg loss: 0.96317186, Global Avg Loss: 5.08835656, Time: 0.0103 Steps: 11330, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000695, Sample Num: 11120, Cur Loss: 1.25758672, Cur Avg Loss: 1.14628325, Log Avg loss: 0.90102564, Global Avg Loss: 5.08466402, Time: 0.0123 Steps: 11340, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000705, Sample Num: 11280, Cur Loss: 1.04528368, Cur Avg Loss: 1.14828368, Log Avg loss: 1.28731348, Global Avg Loss: 5.08131834, Time: 0.0112 Steps: 11350, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000715, Sample Num: 11440, Cur Loss: 1.19185519, Cur Avg Loss: 1.14822597, Log Avg loss: 1.14415739, Global Avg Loss: 5.07785253, Time: 0.0117 Steps: 11360, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000725, Sample Num: 11600, Cur Loss: 1.31065547, Cur Avg Loss: 1.14535447, Log Avg loss: 0.94004267, Global Avg Loss: 5.07421330, Time: 0.0066 Steps: 11370, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000735, Sample Num: 11760, Cur Loss: 1.10572553, Cur Avg Loss: 1.14168395, Log Avg loss: 0.87557113, Global Avg Loss: 5.07052380, Time: 0.0120 Steps: 11380, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000745, Sample Num: 11920, Cur Loss: 1.37629044, Cur Avg Loss: 1.13980244, Log Avg loss: 1.00151104, Global Avg Loss: 5.06695136, Time: 0.0068 Steps: 11390, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000755, Sample Num: 12080, Cur Loss: 0.34026408, Cur Avg Loss: 1.13896495, Log Avg loss: 1.07657198, Global Avg Loss: 5.06345103, Time: 0.0066 Steps: 11400, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000765, Sample Num: 12240, Cur Loss: 0.80811250, Cur Avg Loss: 1.13745561, Log Avg loss: 1.02350063, Global Avg Loss: 5.05991032, Time: 0.0155 Steps: 11410, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000775, Sample Num: 12400, Cur Loss: 0.97866440, Cur Avg Loss: 1.13566954, Log Avg loss: 0.99903496, Global Avg Loss: 5.05635438, Time: 0.0235 Steps: 11420, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000785, Sample Num: 12560, Cur Loss: 0.75211900, Cur Avg Loss: 1.13770904, Log Avg loss: 1.29577057, Global Avg Loss: 5.05306429, Time: 0.0240 Steps: 11430, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000795, Sample Num: 12720, Cur Loss: 0.93954164, Cur Avg Loss: 1.13723313, Log Avg loss: 1.09987381, Global Avg Loss: 5.04960870, Time: 0.0071 Steps: 11440, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000805, Sample Num: 12880, Cur Loss: 0.82786751, Cur Avg Loss: 1.14179635, Log Avg loss: 1.50457237, Global Avg Loss: 5.04651260, Time: 0.0199 Steps: 11450, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000815, Sample Num: 13040, Cur Loss: 1.02032566, Cur Avg Loss: 1.14136270, Log Avg loss: 1.10645383, Global Avg Loss: 5.04307450, Time: 0.0064 Steps: 11460, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000825, Sample Num: 13200, Cur Loss: 1.61543047, Cur Avg Loss: 1.13929320, Log Avg loss: 0.97062898, Global Avg Loss: 5.03952398, Time: 0.0065 Steps: 11470, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000835, Sample Num: 13360, Cur Loss: 2.86857295, Cur Avg Loss: 1.13882990, Log Avg loss: 1.10060804, Global Avg Loss: 5.03609287, Time: 0.0109 Steps: 11480, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000845, Sample Num: 13520, Cur Loss: 0.73700452, Cur Avg Loss: 1.13800184, Log Avg loss: 1.06885904, Global Avg Loss: 5.03264010, Time: 0.0134 Steps: 11490, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000855, Sample Num: 13680, Cur Loss: 1.36475897, Cur Avg Loss: 1.13880701, Log Avg loss: 1.20684349, Global Avg Loss: 5.02931332, Time: 0.0113 Steps: 11500, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000865, Sample Num: 13840, Cur Loss: 0.74724603, Cur Avg Loss: 1.13856298, Log Avg loss: 1.11769857, Global Avg Loss: 5.02591487, Time: 0.0066 Steps: 11510, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000875, Sample Num: 14000, Cur Loss: 1.66383553, Cur Avg Loss: 1.13924304, Log Avg loss: 1.19806822, Global Avg Loss: 5.02259209, Time: 0.0067 Steps: 11520, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000885, Sample Num: 14160, Cur Loss: 0.98495811, Cur Avg Loss: 1.13915431, Log Avg loss: 1.13139053, Global Avg Loss: 5.01921724, Time: 0.0119 Steps: 11530, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000895, Sample Num: 14320, Cur Loss: 1.13190269, Cur Avg Loss: 1.13799367, Log Avg loss: 1.03527737, Global Avg Loss: 5.01576495, Time: 0.0115 Steps: 11540, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000905, Sample Num: 14480, Cur Loss: 1.42161238, Cur Avg Loss: 1.13687265, Log Avg loss: 1.03654102, Global Avg Loss: 5.01231973, Time: 0.0066 Steps: 11550, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000915, Sample Num: 14640, Cur Loss: 2.86907816, Cur Avg Loss: 1.13981031, Log Avg loss: 1.40566830, Global Avg Loss: 5.00919979, Time: 0.0123 Steps: 11560, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000925, Sample Num: 14800, Cur Loss: 1.85521710, Cur Avg Loss: 1.14148775, Log Avg loss: 1.29497376, Global Avg Loss: 5.00598957, Time: 0.0064 Steps: 11570, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000935, Sample Num: 14960, Cur Loss: 1.86036706, Cur Avg Loss: 1.13923088, Log Avg loss: 0.93047018, Global Avg Loss: 5.00247013, Time: 0.0105 Steps: 11580, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000945, Sample Num: 15120, Cur Loss: 0.64131576, Cur Avg Loss: 1.14026495, Log Avg loss: 1.23695108, Global Avg Loss: 4.99922119, Time: 0.0064 Steps: 11590, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000955, Sample Num: 15280, Cur Loss: 1.87317133, Cur Avg Loss: 1.14063728, Log Avg loss: 1.17582170, Global Avg Loss: 4.99592515, Time: 0.0109 Steps: 11600, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000965, Sample Num: 15440, Cur Loss: 1.46937335, Cur Avg Loss: 1.13925359, Log Avg loss: 1.00711194, Global Avg Loss: 4.99248948, Time: 0.0139 Steps: 11610, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000975, Sample Num: 15600, Cur Loss: 0.81250858, Cur Avg Loss: 1.14233210, Log Avg loss: 1.43940804, Global Avg Loss: 4.98943175, Time: 0.0133 Steps: 11620, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000985, Sample Num: 15760, Cur Loss: 1.44164824, Cur Avg Loss: 1.14419506, Log Avg loss: 1.32583316, Global Avg Loss: 4.98628163, Time: 0.0116 Steps: 11630, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 000995, Sample Num: 15920, Cur Loss: 0.82883394, Cur Avg Loss: 1.14406483, Log Avg loss: 1.13123769, Global Avg Loss: 4.98296973, Time: 0.0137 Steps: 11640, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001005, Sample Num: 16080, Cur Loss: 1.13199902, Cur Avg Loss: 1.14581278, Log Avg loss: 1.31973368, Global Avg Loss: 4.97982532, Time: 0.0067 Steps: 11650, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001015, Sample Num: 16240, Cur Loss: 0.79682845, Cur Avg Loss: 1.14618382, Log Avg loss: 1.18347304, Global Avg Loss: 4.97656945, Time: 0.0223 Steps: 11660, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001025, Sample Num: 16400, Cur Loss: 0.86912978, Cur Avg Loss: 1.14479085, Log Avg loss: 1.00340420, Global Avg Loss: 4.97316485, Time: 0.0186 Steps: 11670, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001035, Sample Num: 16560, Cur Loss: 0.88932800, Cur Avg Loss: 1.14414745, Log Avg loss: 1.07819965, Global Avg Loss: 4.96983012, Time: 0.0103 Steps: 11680, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001045, Sample Num: 16720, Cur Loss: 0.99032927, Cur Avg Loss: 1.14808783, Log Avg loss: 1.55591661, Global Avg Loss: 4.96690975, Time: 0.0067 Steps: 11690, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001055, Sample Num: 16880, Cur Loss: 1.69179988, Cur Avg Loss: 1.14715985, Log Avg loss: 1.05018676, Global Avg Loss: 4.96356212, Time: 0.0122 Steps: 11700, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001065, Sample Num: 17040, Cur Loss: 1.14709449, Cur Avg Loss: 1.14766782, Log Avg loss: 1.20125788, Global Avg Loss: 4.96034922, Time: 0.0068 Steps: 11710, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001075, Sample Num: 17200, Cur Loss: 1.12316346, Cur Avg Loss: 1.14662570, Log Avg loss: 1.03564033, Global Avg Loss: 4.95700050, Time: 0.0072 Steps: 11720, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001085, Sample Num: 17360, Cur Loss: 1.94970322, Cur Avg Loss: 1.14726741, Log Avg loss: 1.21625142, Global Avg Loss: 4.95381145, Time: 0.0155 Steps: 11730, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001095, Sample Num: 17520, Cur Loss: 2.51295376, Cur Avg Loss: 1.14853007, Log Avg loss: 1.28552795, Global Avg Loss: 4.95068685, Time: 0.0066 Steps: 11740, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001105, Sample Num: 17680, Cur Loss: 1.11110830, Cur Avg Loss: 1.14701991, Log Avg loss: 0.98165770, Global Avg Loss: 4.94730895, Time: 0.0138 Steps: 11750, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001115, Sample Num: 17840, Cur Loss: 1.15744710, Cur Avg Loss: 1.14738612, Log Avg loss: 1.18785253, Global Avg Loss: 4.94411214, Time: 0.0107 Steps: 11760, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001125, Sample Num: 18000, Cur Loss: 1.27479744, Cur Avg Loss: 1.14751332, Log Avg loss: 1.16169651, Global Avg Loss: 4.94089853, Time: 0.0067 Steps: 11770, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001135, Sample Num: 18160, Cur Loss: 0.42951128, Cur Avg Loss: 1.14662786, Log Avg loss: 1.04701342, Global Avg Loss: 4.93759302, Time: 0.0072 Steps: 11780, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001145, Sample Num: 18320, Cur Loss: 1.14576221, Cur Avg Loss: 1.14813188, Log Avg loss: 1.31883764, Global Avg Loss: 4.93452368, Time: 0.0072 Steps: 11790, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001155, Sample Num: 18480, Cur Loss: 1.53473270, Cur Avg Loss: 1.14753068, Log Avg loss: 1.07869378, Global Avg Loss: 4.93125603, Time: 0.0067 Steps: 11800, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001165, Sample Num: 18640, Cur Loss: 1.15184712, Cur Avg Loss: 1.14851856, Log Avg loss: 1.26261825, Global Avg Loss: 4.92814964, Time: 0.0092 Steps: 11810, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001175, Sample Num: 18800, Cur Loss: 1.29287922, Cur Avg Loss: 1.14894552, Log Avg loss: 1.19868640, Global Avg Loss: 4.92499443, Time: 0.0122 Steps: 11820, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001185, Sample Num: 18960, Cur Loss: 1.05892193, Cur Avg Loss: 1.15017405, Log Avg loss: 1.29452611, Global Avg Loss: 4.92192556, Time: 0.0067 Steps: 11830, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001195, Sample Num: 19120, Cur Loss: 1.15080762, Cur Avg Loss: 1.14989966, Log Avg loss: 1.11738512, Global Avg Loss: 4.91871227, Time: 0.0126 Steps: 11840, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001205, Sample Num: 19280, Cur Loss: 0.79601049, Cur Avg Loss: 1.14933968, Log Avg loss: 1.08242136, Global Avg Loss: 4.91547489, Time: 0.0116 Steps: 11850, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001215, Sample Num: 19440, Cur Loss: 0.35710806, Cur Avg Loss: 1.14839173, Log Avg loss: 1.03416411, Global Avg Loss: 4.91220229, Time: 0.0138 Steps: 11860, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001225, Sample Num: 19600, Cur Loss: 0.51167607, Cur Avg Loss: 1.14781973, Log Avg loss: 1.07832118, Global Avg Loss: 4.90897240, Time: 0.0111 Steps: 11870, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001235, Sample Num: 19760, Cur Loss: 0.52778858, Cur Avg Loss: 1.14739596, Log Avg loss: 1.09548488, Global Avg Loss: 4.90576239, Time: 0.0067 Steps: 11880, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001245, Sample Num: 19920, Cur Loss: 1.23133516, Cur Avg Loss: 1.14743687, Log Avg loss: 1.15248883, Global Avg Loss: 4.90260573, Time: 0.0183 Steps: 11890, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001255, Sample Num: 20080, Cur Loss: 0.67936182, Cur Avg Loss: 1.14585504, Log Avg loss: 0.94891741, Global Avg Loss: 4.89928330, Time: 0.0068 Steps: 11900, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001265, Sample Num: 20240, Cur Loss: 1.37358153, Cur Avg Loss: 1.14351173, Log Avg loss: 0.84942671, Global Avg Loss: 4.89588292, Time: 0.0067 Steps: 11910, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001275, Sample Num: 20400, Cur Loss: 1.94615257, Cur Avg Loss: 1.14413994, Log Avg loss: 1.22360778, Global Avg Loss: 4.89280215, Time: 0.0131 Steps: 11920, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001285, Sample Num: 20560, Cur Loss: 0.94502568, Cur Avg Loss: 1.14353621, Log Avg loss: 1.06656077, Global Avg Loss: 4.88959490, Time: 0.0065 Steps: 11930, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001295, Sample Num: 20720, Cur Loss: 1.30918705, Cur Avg Loss: 1.14375783, Log Avg loss: 1.17223563, Global Avg Loss: 4.88648154, Time: 0.0068 Steps: 11940, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001305, Sample Num: 20880, Cur Loss: 1.58211732, Cur Avg Loss: 1.14566231, Log Avg loss: 1.39229332, Global Avg Loss: 4.88355753, Time: 0.0076 Steps: 11950, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001315, Sample Num: 21040, Cur Loss: 0.89774024, Cur Avg Loss: 1.14618787, Log Avg loss: 1.21477333, Global Avg Loss: 4.88048999, Time: 0.0226 Steps: 11960, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001325, Sample Num: 21200, Cur Loss: 1.68686366, Cur Avg Loss: 1.14541924, Log Avg loss: 1.04434448, Global Avg Loss: 4.87728519, Time: 0.0065 Steps: 11970, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001335, Sample Num: 21360, Cur Loss: 0.83315986, Cur Avg Loss: 1.14612499, Log Avg loss: 1.23963664, Global Avg Loss: 4.87424875, Time: 0.0070 Steps: 11980, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001345, Sample Num: 21520, Cur Loss: 2.10069919, Cur Avg Loss: 1.14624929, Log Avg loss: 1.16284271, Global Avg Loss: 4.87115333, Time: 0.0224 Steps: 11990, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001355, Sample Num: 21680, Cur Loss: 0.87057734, Cur Avg Loss: 1.14470909, Log Avg loss: 0.93755314, Global Avg Loss: 4.86787533, Time: 0.0078 Steps: 12000, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001365, Sample Num: 21840, Cur Loss: 0.88762438, Cur Avg Loss: 1.14262053, Log Avg loss: 0.85962085, Global Avg Loss: 4.86453790, Time: 0.0095 Steps: 12010, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001375, Sample Num: 22000, Cur Loss: 0.44988272, Cur Avg Loss: 1.14236229, Log Avg loss: 1.10711138, Global Avg Loss: 4.86141192, Time: 0.0067 Steps: 12020, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001385, Sample Num: 22160, Cur Loss: 1.04423404, Cur Avg Loss: 1.14242278, Log Avg loss: 1.15074131, Global Avg Loss: 4.85832741, Time: 0.0065 Steps: 12030, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001395, Sample Num: 22320, Cur Loss: 1.47093010, Cur Avg Loss: 1.14212921, Log Avg loss: 1.10146964, Global Avg Loss: 4.85520710, Time: 0.0076 Steps: 12040, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001405, Sample Num: 22480, Cur Loss: 0.57461786, Cur Avg Loss: 1.13961292, Log Avg loss: 0.78859072, Global Avg Loss: 4.85183231, Time: 0.0066 Steps: 12050, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001415, Sample Num: 22640, Cur Loss: 0.74883932, Cur Avg Loss: 1.13986190, Log Avg loss: 1.17484334, Global Avg Loss: 4.84878340, Time: 0.0078 Steps: 12060, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001425, Sample Num: 22800, Cur Loss: 0.92347306, Cur Avg Loss: 1.13905579, Log Avg loss: 1.02499080, Global Avg Loss: 4.84561538, Time: 0.0143 Steps: 12070, Updated lr: 0.000090 Training, Epoch: 0006, Batch: 001435, Sample Num: 22960, Cur Loss: 0.41342604, Cur Avg Loss: 1.13801365, Log Avg loss: 0.98950855, Global Avg Loss: 4.84242324, Time: 0.0112 Steps: 12080, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001445, Sample Num: 23120, Cur Loss: 0.43508565, Cur Avg Loss: 1.13790639, Log Avg loss: 1.12251437, Global Avg Loss: 4.83934639, Time: 0.0065 Steps: 12090, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001455, Sample Num: 23280, Cur Loss: 2.02828026, Cur Avg Loss: 1.13995697, Log Avg loss: 1.43626588, Global Avg Loss: 4.83653393, Time: 0.0110 Steps: 12100, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001465, Sample Num: 23440, Cur Loss: 1.36384082, Cur Avg Loss: 1.14137583, Log Avg loss: 1.34782078, Global Avg Loss: 4.83365308, Time: 0.0118 Steps: 12110, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001475, Sample Num: 23600, Cur Loss: 0.56571865, Cur Avg Loss: 1.14211208, Log Avg loss: 1.24997185, Global Avg Loss: 4.83069625, Time: 0.0127 Steps: 12120, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001485, Sample Num: 23760, Cur Loss: 3.62811852, Cur Avg Loss: 1.14323144, Log Avg loss: 1.30833748, Global Avg Loss: 4.82779241, Time: 0.0063 Steps: 12130, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001495, Sample Num: 23920, Cur Loss: 1.30886745, Cur Avg Loss: 1.14342164, Log Avg loss: 1.17166696, Global Avg Loss: 4.82478077, Time: 0.0105 Steps: 12140, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001505, Sample Num: 24080, Cur Loss: 1.23029089, Cur Avg Loss: 1.14448753, Log Avg loss: 1.30383793, Global Avg Loss: 4.82188287, Time: 0.0065 Steps: 12150, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001515, Sample Num: 24240, Cur Loss: 0.74334586, Cur Avg Loss: 1.14476735, Log Avg loss: 1.18687970, Global Avg Loss: 4.81889356, Time: 0.0103 Steps: 12160, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001525, Sample Num: 24400, Cur Loss: 0.74991655, Cur Avg Loss: 1.14268423, Log Avg loss: 0.82709086, Global Avg Loss: 4.81561353, Time: 0.0115 Steps: 12170, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001535, Sample Num: 24560, Cur Loss: 0.94168878, Cur Avg Loss: 1.14061374, Log Avg loss: 0.82486400, Global Avg Loss: 4.81233705, Time: 0.0093 Steps: 12180, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001545, Sample Num: 24720, Cur Loss: 1.07369423, Cur Avg Loss: 1.13962135, Log Avg loss: 0.98729021, Global Avg Loss: 4.80919919, Time: 0.0063 Steps: 12190, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001555, Sample Num: 24880, Cur Loss: 1.01612735, Cur Avg Loss: 1.13876458, Log Avg loss: 1.00639285, Global Avg Loss: 4.80608214, Time: 0.0064 Steps: 12200, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001565, Sample Num: 25040, Cur Loss: 0.30636954, Cur Avg Loss: 1.13678006, Log Avg loss: 0.82818848, Global Avg Loss: 4.80282424, Time: 0.0119 Steps: 12210, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001575, Sample Num: 25200, Cur Loss: 1.26987958, Cur Avg Loss: 1.13580564, Log Avg loss: 0.98330773, Global Avg Loss: 4.79969861, Time: 0.0119 Steps: 12220, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001585, Sample Num: 25360, Cur Loss: 0.22477962, Cur Avg Loss: 1.13296967, Log Avg loss: 0.68630455, Global Avg Loss: 4.79633525, Time: 0.0147 Steps: 12230, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001595, Sample Num: 25520, Cur Loss: 0.43063611, Cur Avg Loss: 1.13081922, Log Avg loss: 0.78997411, Global Avg Loss: 4.79306208, Time: 0.0066 Steps: 12240, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001605, Sample Num: 25680, Cur Loss: 0.58732033, Cur Avg Loss: 1.13125053, Log Avg loss: 1.20004427, Global Avg Loss: 4.79012900, Time: 0.0106 Steps: 12250, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001615, Sample Num: 25840, Cur Loss: 1.78712845, Cur Avg Loss: 1.13078594, Log Avg loss: 1.05621847, Global Avg Loss: 4.78708340, Time: 0.0117 Steps: 12260, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001625, Sample Num: 26000, Cur Loss: 0.57714742, Cur Avg Loss: 1.13178838, Log Avg loss: 1.29368277, Global Avg Loss: 4.78423629, Time: 0.0112 Steps: 12270, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001635, Sample Num: 26160, Cur Loss: 1.64766443, Cur Avg Loss: 1.13190315, Log Avg loss: 1.15055270, Global Avg Loss: 4.78127727, Time: 0.0077 Steps: 12280, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001645, Sample Num: 26320, Cur Loss: 0.98564482, Cur Avg Loss: 1.12962629, Log Avg loss: 0.75735949, Global Avg Loss: 4.77800313, Time: 0.0067 Steps: 12290, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001655, Sample Num: 26480, Cur Loss: 1.33536553, Cur Avg Loss: 1.12970536, Log Avg loss: 1.14271344, Global Avg Loss: 4.77504761, Time: 0.0067 Steps: 12300, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001665, Sample Num: 26640, Cur Loss: 1.62199593, Cur Avg Loss: 1.13194780, Log Avg loss: 1.50307191, Global Avg Loss: 4.77238962, Time: 0.0066 Steps: 12310, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001675, Sample Num: 26800, Cur Loss: 1.22616196, Cur Avg Loss: 1.13282375, Log Avg loss: 1.27866793, Global Avg Loss: 4.76955381, Time: 0.0108 Steps: 12320, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001685, Sample Num: 26960, Cur Loss: 0.83917558, Cur Avg Loss: 1.13207168, Log Avg loss: 1.00610038, Global Avg Loss: 4.76650154, Time: 0.0071 Steps: 12330, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001695, Sample Num: 27120, Cur Loss: 0.60543525, Cur Avg Loss: 1.13245662, Log Avg loss: 1.19731871, Global Avg Loss: 4.76360917, Time: 0.0128 Steps: 12340, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001705, Sample Num: 27280, Cur Loss: 0.87541366, Cur Avg Loss: 1.13248636, Log Avg loss: 1.13752767, Global Avg Loss: 4.76067307, Time: 0.0107 Steps: 12350, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001715, Sample Num: 27440, Cur Loss: 0.46771890, Cur Avg Loss: 1.13138906, Log Avg loss: 0.94429934, Global Avg Loss: 4.75758539, Time: 0.0064 Steps: 12360, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001725, Sample Num: 27600, Cur Loss: 1.68546879, Cur Avg Loss: 1.13335146, Log Avg loss: 1.46990409, Global Avg Loss: 4.75492760, Time: 0.0109 Steps: 12370, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001735, Sample Num: 27760, Cur Loss: 2.08860230, Cur Avg Loss: 1.13407513, Log Avg loss: 1.25890828, Global Avg Loss: 4.75210368, Time: 0.0135 Steps: 12380, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001745, Sample Num: 27920, Cur Loss: 0.78375828, Cur Avg Loss: 1.13372721, Log Avg loss: 1.07336198, Global Avg Loss: 4.74913456, Time: 0.0120 Steps: 12390, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001755, Sample Num: 28080, Cur Loss: 0.77579975, Cur Avg Loss: 1.13398972, Log Avg loss: 1.17979740, Global Avg Loss: 4.74625606, Time: 0.0066 Steps: 12400, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001765, Sample Num: 28240, Cur Loss: 0.95666611, Cur Avg Loss: 1.13516445, Log Avg loss: 1.34132992, Global Avg Loss: 4.74351236, Time: 0.0113 Steps: 12410, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001775, Sample Num: 28400, Cur Loss: 1.27137864, Cur Avg Loss: 1.13423858, Log Avg loss: 0.97082254, Global Avg Loss: 4.74047477, Time: 0.0107 Steps: 12420, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001785, Sample Num: 28560, Cur Loss: 1.26032662, Cur Avg Loss: 1.13409413, Log Avg loss: 1.10845430, Global Avg Loss: 4.73755279, Time: 0.0130 Steps: 12430, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001795, Sample Num: 28720, Cur Loss: 1.41266358, Cur Avg Loss: 1.13382922, Log Avg loss: 1.08654361, Global Avg Loss: 4.73461790, Time: 0.0068 Steps: 12440, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001805, Sample Num: 28880, Cur Loss: 1.07299685, Cur Avg Loss: 1.13490660, Log Avg loss: 1.32829513, Global Avg Loss: 4.73188189, Time: 0.0065 Steps: 12450, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001815, Sample Num: 29040, Cur Loss: 0.71755010, Cur Avg Loss: 1.13515869, Log Avg loss: 1.18066098, Global Avg Loss: 4.72903180, Time: 0.0237 Steps: 12460, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001825, Sample Num: 29200, Cur Loss: 1.85615861, Cur Avg Loss: 1.13628751, Log Avg loss: 1.34116832, Global Avg Loss: 4.72631499, Time: 0.0066 Steps: 12470, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001835, Sample Num: 29360, Cur Loss: 0.40867406, Cur Avg Loss: 1.13771331, Log Avg loss: 1.39792250, Global Avg Loss: 4.72364801, Time: 0.0070 Steps: 12480, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001845, Sample Num: 29520, Cur Loss: 1.11763632, Cur Avg Loss: 1.13725185, Log Avg loss: 1.05257325, Global Avg Loss: 4.72070879, Time: 0.0135 Steps: 12490, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001855, Sample Num: 29680, Cur Loss: 0.54621553, Cur Avg Loss: 1.13630691, Log Avg loss: 0.96196613, Global Avg Loss: 4.71770180, Time: 0.0067 Steps: 12500, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001865, Sample Num: 29840, Cur Loss: 0.81488335, Cur Avg Loss: 1.13685827, Log Avg loss: 1.23913519, Global Avg Loss: 4.71492117, Time: 0.0067 Steps: 12510, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001875, Sample Num: 30000, Cur Loss: 0.34071830, Cur Avg Loss: 1.13469773, Log Avg loss: 0.73175758, Global Avg Loss: 4.71173973, Time: 0.0117 Steps: 12520, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001885, Sample Num: 30160, Cur Loss: 1.75880170, Cur Avg Loss: 1.13618386, Log Avg loss: 1.41483274, Global Avg Loss: 4.70910852, Time: 0.0071 Steps: 12530, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001895, Sample Num: 30320, Cur Loss: 1.03076339, Cur Avg Loss: 1.13541864, Log Avg loss: 0.99117548, Global Avg Loss: 4.70614366, Time: 0.0197 Steps: 12540, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001905, Sample Num: 30480, Cur Loss: 0.68741375, Cur Avg Loss: 1.13409482, Log Avg loss: 0.88323001, Global Avg Loss: 4.70309752, Time: 0.0065 Steps: 12550, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001915, Sample Num: 30640, Cur Loss: 0.79934430, Cur Avg Loss: 1.13373227, Log Avg loss: 1.06466621, Global Avg Loss: 4.70020067, Time: 0.0075 Steps: 12560, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001925, Sample Num: 30800, Cur Loss: 0.93607664, Cur Avg Loss: 1.13369511, Log Avg loss: 1.12658016, Global Avg Loss: 4.69735770, Time: 0.0063 Steps: 12570, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001935, Sample Num: 30960, Cur Loss: 1.75093067, Cur Avg Loss: 1.13369019, Log Avg loss: 1.13274253, Global Avg Loss: 4.69452414, Time: 0.0064 Steps: 12580, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001945, Sample Num: 31120, Cur Loss: 1.38845468, Cur Avg Loss: 1.13416027, Log Avg loss: 1.22512165, Global Avg Loss: 4.69176846, Time: 0.0068 Steps: 12590, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001955, Sample Num: 31280, Cur Loss: 1.07131743, Cur Avg Loss: 1.13290205, Log Avg loss: 0.88817708, Global Avg Loss: 4.68874974, Time: 0.0226 Steps: 12600, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001965, Sample Num: 31440, Cur Loss: 0.98405153, Cur Avg Loss: 1.13380250, Log Avg loss: 1.30984045, Global Avg Loss: 4.68607019, Time: 0.0070 Steps: 12610, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001975, Sample Num: 31600, Cur Loss: 1.06020403, Cur Avg Loss: 1.13423570, Log Avg loss: 1.21936021, Global Avg Loss: 4.68332319, Time: 0.0198 Steps: 12620, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001985, Sample Num: 31760, Cur Loss: 1.63475132, Cur Avg Loss: 1.13626121, Log Avg loss: 1.53629929, Global Avg Loss: 4.68083149, Time: 0.0065 Steps: 12630, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 001995, Sample Num: 31920, Cur Loss: 1.25034666, Cur Avg Loss: 1.13691587, Log Avg loss: 1.26686582, Global Avg Loss: 4.67813057, Time: 0.0067 Steps: 12640, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002005, Sample Num: 32080, Cur Loss: 0.84041256, Cur Avg Loss: 1.13623060, Log Avg loss: 0.99951998, Global Avg Loss: 4.67522257, Time: 0.0065 Steps: 12650, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002015, Sample Num: 32240, Cur Loss: 0.97489119, Cur Avg Loss: 1.13552923, Log Avg loss: 0.99490367, Global Avg Loss: 4.67231553, Time: 0.0202 Steps: 12660, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002025, Sample Num: 32400, Cur Loss: 0.90031946, Cur Avg Loss: 1.13603529, Log Avg loss: 1.23800548, Global Avg Loss: 4.66960494, Time: 0.0071 Steps: 12670, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002035, Sample Num: 32560, Cur Loss: 0.87943697, Cur Avg Loss: 1.13616635, Log Avg loss: 1.16270716, Global Avg Loss: 4.66683925, Time: 0.0066 Steps: 12680, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002045, Sample Num: 32720, Cur Loss: 0.88559198, Cur Avg Loss: 1.13598125, Log Avg loss: 1.09831225, Global Avg Loss: 4.66402717, Time: 0.0185 Steps: 12690, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002055, Sample Num: 32880, Cur Loss: 1.70457435, Cur Avg Loss: 1.13441779, Log Avg loss: 0.81469008, Global Avg Loss: 4.66099620, Time: 0.0068 Steps: 12700, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002065, Sample Num: 33040, Cur Loss: 0.64493889, Cur Avg Loss: 1.13391181, Log Avg loss: 1.02993301, Global Avg Loss: 4.65813934, Time: 0.0118 Steps: 12710, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002075, Sample Num: 33200, Cur Loss: 1.09854293, Cur Avg Loss: 1.13340255, Log Avg loss: 1.02824140, Global Avg Loss: 4.65528565, Time: 0.0068 Steps: 12720, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002085, Sample Num: 33360, Cur Loss: 1.22720027, Cur Avg Loss: 1.13337170, Log Avg loss: 1.12697017, Global Avg Loss: 4.65251400, Time: 0.0120 Steps: 12730, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002095, Sample Num: 33520, Cur Loss: 0.68152523, Cur Avg Loss: 1.13211204, Log Avg loss: 0.86947279, Global Avg Loss: 4.64954458, Time: 0.0144 Steps: 12740, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002105, Sample Num: 33680, Cur Loss: 0.95991367, Cur Avg Loss: 1.13193782, Log Avg loss: 1.09543860, Global Avg Loss: 4.64675704, Time: 0.0116 Steps: 12750, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002115, Sample Num: 33840, Cur Loss: 0.85110182, Cur Avg Loss: 1.13154470, Log Avg loss: 1.04879403, Global Avg Loss: 4.64393732, Time: 0.0116 Steps: 12760, Updated lr: 0.000089 Training, Epoch: 0006, Batch: 002125, Sample Num: 34000, Cur Loss: 0.83290434, Cur Avg Loss: 1.13111051, Log Avg loss: 1.03927945, Global Avg Loss: 4.64111457, Time: 0.0115 Steps: 12770, Updated lr: 0.000089 ***** Running evaluation checkpoint-12774 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-12774 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.068266, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.306297, "eval_total_loss": 918.326803, "eval_mae": 0.999922, "eval_mse": 1.306272, "eval_r2": 0.169648, "eval_sp_statistic": 0.521676, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.549319, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 4.53562, "test_total_loss": 2276.881406, "test_mae": 1.967636, "test_mse": 4.534853, "test_r2": -1.926834, "test_sp_statistic": 0.20798, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.330498, "test_ps_pvalue": 0.0, "lr": 8.883451872925558e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 4.639952339870639, "train_cur_epoch_loss": 2407.328004717827, "train_cur_epoch_avg_loss": 1.1307318011826335, "train_cur_epoch_time": 22.06826615333557, "train_cur_epoch_avg_time": 0.010365554792548413, "epoch": 6, "step": 12774} ################################################## Training, Epoch: 0007, Batch: 000006, Sample Num: 96, Cur Loss: 1.18720961, Cur Avg Loss: 1.00392205, Log Avg loss: 0.97416938, Global Avg Loss: 4.63824528, Time: 0.0128 Steps: 12780, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000016, Sample Num: 256, Cur Loss: 0.94927275, Cur Avg Loss: 1.13930557, Log Avg loss: 1.22053568, Global Avg Loss: 4.63557311, Time: 0.0119 Steps: 12790, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000026, Sample Num: 416, Cur Loss: 1.12481654, Cur Avg Loss: 1.11753302, Log Avg loss: 1.08269694, Global Avg Loss: 4.63279743, Time: 0.0117 Steps: 12800, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000036, Sample Num: 576, Cur Loss: 0.83731556, Cur Avg Loss: 1.00877453, Log Avg loss: 0.72600245, Global Avg Loss: 4.62974762, Time: 0.0120 Steps: 12810, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000046, Sample Num: 736, Cur Loss: 1.56234014, Cur Avg Loss: 1.02392762, Log Avg loss: 1.07847873, Global Avg Loss: 4.62697752, Time: 0.0066 Steps: 12820, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000056, Sample Num: 896, Cur Loss: 2.21074796, Cur Avg Loss: 1.04131696, Log Avg loss: 1.12130794, Global Avg Loss: 4.62424512, Time: 0.0089 Steps: 12830, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000066, Sample Num: 1056, Cur Loss: 1.61570859, Cur Avg Loss: 1.04845175, Log Avg loss: 1.08840660, Global Avg Loss: 4.62149136, Time: 0.0216 Steps: 12840, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000076, Sample Num: 1216, Cur Loss: 1.45752048, Cur Avg Loss: 1.08148379, Log Avg loss: 1.29949525, Global Avg Loss: 4.61890614, Time: 0.0116 Steps: 12850, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000086, Sample Num: 1376, Cur Loss: 1.00016475, Cur Avg Loss: 1.09113534, Log Avg loss: 1.16448707, Global Avg Loss: 4.61621997, Time: 0.0067 Steps: 12860, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000096, Sample Num: 1536, Cur Loss: 0.54831159, Cur Avg Loss: 1.07634293, Log Avg loss: 0.94912819, Global Avg Loss: 4.61337064, Time: 0.0155 Steps: 12870, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000106, Sample Num: 1696, Cur Loss: 1.01071918, Cur Avg Loss: 1.08206360, Log Avg loss: 1.13698209, Global Avg Loss: 4.61067158, Time: 0.0111 Steps: 12880, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000116, Sample Num: 1856, Cur Loss: 1.39801335, Cur Avg Loss: 1.09230458, Log Avg loss: 1.20085897, Global Avg Loss: 4.60802626, Time: 0.0121 Steps: 12890, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000126, Sample Num: 2016, Cur Loss: 2.43803453, Cur Avg Loss: 1.10667272, Log Avg loss: 1.27334313, Global Avg Loss: 4.60544124, Time: 0.0067 Steps: 12900, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000136, Sample Num: 2176, Cur Loss: 0.80456406, Cur Avg Loss: 1.10374364, Log Avg loss: 1.06683723, Global Avg Loss: 4.60270026, Time: 0.0067 Steps: 12910, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000146, Sample Num: 2336, Cur Loss: 1.35987842, Cur Avg Loss: 1.11299388, Log Avg loss: 1.23879718, Global Avg Loss: 4.60009662, Time: 0.0069 Steps: 12920, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000156, Sample Num: 2496, Cur Loss: 0.87985039, Cur Avg Loss: 1.11195696, Log Avg loss: 1.09681782, Global Avg Loss: 4.59738720, Time: 0.0066 Steps: 12930, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000166, Sample Num: 2656, Cur Loss: 0.75848019, Cur Avg Loss: 1.11671567, Log Avg loss: 1.19095163, Global Avg Loss: 4.59475471, Time: 0.0068 Steps: 12940, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000176, Sample Num: 2816, Cur Loss: 1.99334049, Cur Avg Loss: 1.11713120, Log Avg loss: 1.12402903, Global Avg Loss: 4.59207462, Time: 0.0113 Steps: 12950, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000186, Sample Num: 2976, Cur Loss: 1.34887648, Cur Avg Loss: 1.10966748, Log Avg loss: 0.97830595, Global Avg Loss: 4.58928621, Time: 0.0067 Steps: 12960, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000196, Sample Num: 3136, Cur Loss: 1.75417328, Cur Avg Loss: 1.11911095, Log Avg loss: 1.29475947, Global Avg Loss: 4.58674610, Time: 0.0218 Steps: 12970, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000206, Sample Num: 3296, Cur Loss: 1.40716267, Cur Avg Loss: 1.11214130, Log Avg loss: 0.97553612, Global Avg Loss: 4.58396397, Time: 0.0086 Steps: 12980, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000216, Sample Num: 3456, Cur Loss: 1.11041498, Cur Avg Loss: 1.11260141, Log Avg loss: 1.12207982, Global Avg Loss: 4.58129893, Time: 0.0122 Steps: 12990, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000226, Sample Num: 3616, Cur Loss: 0.86624146, Cur Avg Loss: 1.10415973, Log Avg loss: 0.92181931, Global Avg Loss: 4.57848395, Time: 0.0116 Steps: 13000, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000236, Sample Num: 3776, Cur Loss: 0.82316387, Cur Avg Loss: 1.10004221, Log Avg loss: 1.00698639, Global Avg Loss: 4.57573875, Time: 0.0109 Steps: 13010, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000246, Sample Num: 3936, Cur Loss: 2.93238807, Cur Avg Loss: 1.10723518, Log Avg loss: 1.27698929, Global Avg Loss: 4.57320515, Time: 0.0067 Steps: 13020, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000256, Sample Num: 4096, Cur Loss: 2.29609275, Cur Avg Loss: 1.11134623, Log Avg loss: 1.21247784, Global Avg Loss: 4.57062593, Time: 0.0073 Steps: 13030, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000266, Sample Num: 4256, Cur Loss: 0.49669585, Cur Avg Loss: 1.11083552, Log Avg loss: 1.09776146, Global Avg Loss: 4.56796269, Time: 0.0066 Steps: 13040, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000276, Sample Num: 4416, Cur Loss: 0.40499091, Cur Avg Loss: 1.10640959, Log Avg loss: 0.98867996, Global Avg Loss: 4.56521994, Time: 0.0118 Steps: 13050, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000286, Sample Num: 4576, Cur Loss: 0.80476451, Cur Avg Loss: 1.11065614, Log Avg loss: 1.22786087, Global Avg Loss: 4.56266454, Time: 0.0093 Steps: 13060, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000296, Sample Num: 4736, Cur Loss: 1.99069011, Cur Avg Loss: 1.11722125, Log Avg loss: 1.30498328, Global Avg Loss: 4.56017205, Time: 0.0093 Steps: 13070, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000306, Sample Num: 4896, Cur Loss: 0.91813540, Cur Avg Loss: 1.11329174, Log Avg loss: 0.99697831, Global Avg Loss: 4.55744789, Time: 0.0118 Steps: 13080, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000316, Sample Num: 5056, Cur Loss: 0.39515865, Cur Avg Loss: 1.11177681, Log Avg loss: 1.06542006, Global Avg Loss: 4.55478019, Time: 0.0097 Steps: 13090, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000326, Sample Num: 5216, Cur Loss: 1.47357953, Cur Avg Loss: 1.11391438, Log Avg loss: 1.18146145, Global Avg Loss: 4.55220514, Time: 0.0071 Steps: 13100, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000336, Sample Num: 5376, Cur Loss: 0.97268826, Cur Avg Loss: 1.11105532, Log Avg loss: 1.01784986, Global Avg Loss: 4.54950921, Time: 0.0083 Steps: 13110, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000346, Sample Num: 5536, Cur Loss: 1.14349651, Cur Avg Loss: 1.10895093, Log Avg loss: 1.03824355, Global Avg Loss: 4.54683294, Time: 0.0107 Steps: 13120, Updated lr: 0.000089 Training, Epoch: 0007, Batch: 000356, Sample Num: 5696, Cur Loss: 1.74937797, Cur Avg Loss: 1.10799235, Log Avg loss: 1.07482540, Global Avg Loss: 4.54418861, Time: 0.0077 Steps: 13130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000366, Sample Num: 5856, Cur Loss: 1.12725043, Cur Avg Loss: 1.11036770, Log Avg loss: 1.19493040, Global Avg Loss: 4.54163971, Time: 0.0139 Steps: 13140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000376, Sample Num: 6016, Cur Loss: 0.91122508, Cur Avg Loss: 1.10753321, Log Avg loss: 1.00379080, Global Avg Loss: 4.53894933, Time: 0.0068 Steps: 13150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000386, Sample Num: 6176, Cur Loss: 0.80039501, Cur Avg Loss: 1.10166009, Log Avg loss: 0.88083055, Global Avg Loss: 4.53616960, Time: 0.0109 Steps: 13160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000396, Sample Num: 6336, Cur Loss: 1.25788450, Cur Avg Loss: 1.10348372, Log Avg loss: 1.17387604, Global Avg Loss: 4.53361661, Time: 0.0070 Steps: 13170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000406, Sample Num: 6496, Cur Loss: 0.71479201, Cur Avg Loss: 1.09782970, Log Avg loss: 0.87393065, Global Avg Loss: 4.53083991, Time: 0.0065 Steps: 13180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000416, Sample Num: 6656, Cur Loss: 0.53236878, Cur Avg Loss: 1.09248426, Log Avg loss: 0.87545940, Global Avg Loss: 4.52806859, Time: 0.0066 Steps: 13190, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000426, Sample Num: 6816, Cur Loss: 0.97525251, Cur Avg Loss: 1.09746521, Log Avg loss: 1.30467247, Global Avg Loss: 4.52562662, Time: 0.0066 Steps: 13200, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000436, Sample Num: 6976, Cur Loss: 1.34764433, Cur Avg Loss: 1.09584036, Log Avg loss: 1.02662202, Global Avg Loss: 4.52297786, Time: 0.0079 Steps: 13210, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000446, Sample Num: 7136, Cur Loss: 1.43002188, Cur Avg Loss: 1.09607582, Log Avg loss: 1.10634151, Global Avg Loss: 4.52039342, Time: 0.0121 Steps: 13220, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000456, Sample Num: 7296, Cur Loss: 0.38304168, Cur Avg Loss: 1.09478119, Log Avg loss: 1.03704090, Global Avg Loss: 4.51776050, Time: 0.0237 Steps: 13230, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000466, Sample Num: 7456, Cur Loss: 1.04906487, Cur Avg Loss: 1.10132221, Log Avg loss: 1.39959252, Global Avg Loss: 4.51540539, Time: 0.0067 Steps: 13240, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000476, Sample Num: 7616, Cur Loss: 0.85368669, Cur Avg Loss: 1.10324084, Log Avg loss: 1.19264898, Global Avg Loss: 4.51289765, Time: 0.0109 Steps: 13250, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000486, Sample Num: 7776, Cur Loss: 1.16814053, Cur Avg Loss: 1.10288563, Log Avg loss: 1.08597789, Global Avg Loss: 4.51031324, Time: 0.0105 Steps: 13260, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000496, Sample Num: 7936, Cur Loss: 0.69654095, Cur Avg Loss: 1.10622449, Log Avg loss: 1.26849277, Global Avg Loss: 4.50787027, Time: 0.0106 Steps: 13270, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000506, Sample Num: 8096, Cur Loss: 0.66384792, Cur Avg Loss: 1.10395919, Log Avg loss: 0.99160039, Global Avg Loss: 4.50522248, Time: 0.0066 Steps: 13280, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000516, Sample Num: 8256, Cur Loss: 0.93710196, Cur Avg Loss: 1.10360539, Log Avg loss: 1.08570320, Global Avg Loss: 4.50264948, Time: 0.0135 Steps: 13290, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000526, Sample Num: 8416, Cur Loss: 0.42401418, Cur Avg Loss: 1.09779067, Log Avg loss: 0.79775124, Global Avg Loss: 4.49986384, Time: 0.0097 Steps: 13300, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000536, Sample Num: 8576, Cur Loss: 0.77164692, Cur Avg Loss: 1.09754411, Log Avg loss: 1.08457476, Global Avg Loss: 4.49729788, Time: 0.0092 Steps: 13310, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000546, Sample Num: 8736, Cur Loss: 0.77844447, Cur Avg Loss: 1.09474451, Log Avg loss: 0.94468603, Global Avg Loss: 4.49463076, Time: 0.0112 Steps: 13320, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000556, Sample Num: 8896, Cur Loss: 0.82998168, Cur Avg Loss: 1.09236435, Log Avg loss: 0.96240790, Global Avg Loss: 4.49198093, Time: 0.0120 Steps: 13330, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000566, Sample Num: 9056, Cur Loss: 1.50247300, Cur Avg Loss: 1.09492102, Log Avg loss: 1.23707184, Global Avg Loss: 4.48954097, Time: 0.0118 Steps: 13340, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000576, Sample Num: 9216, Cur Loss: 0.55805361, Cur Avg Loss: 1.09399266, Log Avg loss: 1.04144745, Global Avg Loss: 4.48695812, Time: 0.0117 Steps: 13350, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000586, Sample Num: 9376, Cur Loss: 1.79709136, Cur Avg Loss: 1.09658952, Log Avg loss: 1.24616853, Global Avg Loss: 4.48453238, Time: 0.0135 Steps: 13360, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000596, Sample Num: 9536, Cur Loss: 1.90029585, Cur Avg Loss: 1.09501024, Log Avg loss: 1.00246463, Global Avg Loss: 4.48192800, Time: 0.0106 Steps: 13370, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000606, Sample Num: 9696, Cur Loss: 0.83233976, Cur Avg Loss: 1.09782121, Log Avg loss: 1.26535498, Global Avg Loss: 4.47952398, Time: 0.0105 Steps: 13380, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000616, Sample Num: 9856, Cur Loss: 0.49220282, Cur Avg Loss: 1.09700864, Log Avg loss: 1.04776663, Global Avg Loss: 4.47696105, Time: 0.0074 Steps: 13390, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000626, Sample Num: 10016, Cur Loss: 0.91404718, Cur Avg Loss: 1.09531842, Log Avg loss: 0.99120088, Global Avg Loss: 4.47435974, Time: 0.0123 Steps: 13400, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000636, Sample Num: 10176, Cur Loss: 1.91134834, Cur Avg Loss: 1.09595856, Log Avg loss: 1.13603131, Global Avg Loss: 4.47187031, Time: 0.0122 Steps: 13410, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000646, Sample Num: 10336, Cur Loss: 1.08358574, Cur Avg Loss: 1.09449589, Log Avg loss: 1.00147041, Global Avg Loss: 4.46928432, Time: 0.0120 Steps: 13420, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000656, Sample Num: 10496, Cur Loss: 1.09108853, Cur Avg Loss: 1.09766502, Log Avg loss: 1.30239064, Global Avg Loss: 4.46692624, Time: 0.0068 Steps: 13430, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000666, Sample Num: 10656, Cur Loss: 0.92095971, Cur Avg Loss: 1.09298763, Log Avg loss: 0.78615105, Global Avg Loss: 4.46418757, Time: 0.0076 Steps: 13440, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000676, Sample Num: 10816, Cur Loss: 0.97972858, Cur Avg Loss: 1.09151281, Log Avg loss: 0.99328988, Global Avg Loss: 4.46160698, Time: 0.0078 Steps: 13450, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000686, Sample Num: 10976, Cur Loss: 1.61893821, Cur Avg Loss: 1.09561420, Log Avg loss: 1.37286770, Global Avg Loss: 4.45931222, Time: 0.0116 Steps: 13460, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000696, Sample Num: 11136, Cur Loss: 0.31145489, Cur Avg Loss: 1.09030183, Log Avg loss: 0.72587346, Global Avg Loss: 4.45654055, Time: 0.0067 Steps: 13470, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000706, Sample Num: 11296, Cur Loss: 0.87490457, Cur Avg Loss: 1.08970000, Log Avg loss: 1.04781228, Global Avg Loss: 4.45401182, Time: 0.0122 Steps: 13480, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000716, Sample Num: 11456, Cur Loss: 0.39582083, Cur Avg Loss: 1.08788332, Log Avg loss: 0.95962624, Global Avg Loss: 4.45142147, Time: 0.0141 Steps: 13490, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000726, Sample Num: 11616, Cur Loss: 1.02465487, Cur Avg Loss: 1.08961448, Log Avg loss: 1.21356556, Global Avg Loss: 4.44902306, Time: 0.0067 Steps: 13500, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000736, Sample Num: 11776, Cur Loss: 1.13771188, Cur Avg Loss: 1.08988870, Log Avg loss: 1.10979712, Global Avg Loss: 4.44655139, Time: 0.0064 Steps: 13510, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000746, Sample Num: 11936, Cur Loss: 1.19052219, Cur Avg Loss: 1.08885095, Log Avg loss: 1.01247204, Global Avg Loss: 4.44401139, Time: 0.0131 Steps: 13520, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000756, Sample Num: 12096, Cur Loss: 1.33036184, Cur Avg Loss: 1.09151562, Log Avg loss: 1.29029988, Global Avg Loss: 4.44168049, Time: 0.0104 Steps: 13530, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000766, Sample Num: 12256, Cur Loss: 0.74872506, Cur Avg Loss: 1.08788618, Log Avg loss: 0.81350092, Global Avg Loss: 4.43900089, Time: 0.0121 Steps: 13540, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000776, Sample Num: 12416, Cur Loss: 1.42875767, Cur Avg Loss: 1.08643189, Log Avg loss: 0.97503303, Global Avg Loss: 4.43644445, Time: 0.0114 Steps: 13550, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000786, Sample Num: 12576, Cur Loss: 0.91731399, Cur Avg Loss: 1.08403823, Log Avg loss: 0.89829023, Global Avg Loss: 4.43383519, Time: 0.0067 Steps: 13560, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000796, Sample Num: 12736, Cur Loss: 1.90666461, Cur Avg Loss: 1.08389710, Log Avg loss: 1.07280459, Global Avg Loss: 4.43135838, Time: 0.0121 Steps: 13570, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000806, Sample Num: 12896, Cur Loss: 0.99169409, Cur Avg Loss: 1.08073916, Log Avg loss: 0.82936704, Global Avg Loss: 4.42870596, Time: 0.0066 Steps: 13580, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000816, Sample Num: 13056, Cur Loss: 1.09697556, Cur Avg Loss: 1.08075166, Log Avg loss: 1.08175890, Global Avg Loss: 4.42624316, Time: 0.0117 Steps: 13590, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000826, Sample Num: 13216, Cur Loss: 1.23341644, Cur Avg Loss: 1.08660417, Log Avg loss: 1.56416892, Global Avg Loss: 4.42413869, Time: 0.0069 Steps: 13600, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000836, Sample Num: 13376, Cur Loss: 1.02547789, Cur Avg Loss: 1.08663034, Log Avg loss: 1.08879209, Global Avg Loss: 4.42168803, Time: 0.0110 Steps: 13610, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000846, Sample Num: 13536, Cur Loss: 1.22035456, Cur Avg Loss: 1.08753988, Log Avg loss: 1.16357726, Global Avg Loss: 4.41929588, Time: 0.0112 Steps: 13620, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000856, Sample Num: 13696, Cur Loss: 0.77751553, Cur Avg Loss: 1.08551203, Log Avg loss: 0.91395627, Global Avg Loss: 4.41672410, Time: 0.0068 Steps: 13630, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000866, Sample Num: 13856, Cur Loss: 0.78071022, Cur Avg Loss: 1.08485530, Log Avg loss: 1.02863950, Global Avg Loss: 4.41424017, Time: 0.0089 Steps: 13640, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000876, Sample Num: 14016, Cur Loss: 1.04835248, Cur Avg Loss: 1.08682007, Log Avg loss: 1.25696856, Global Avg Loss: 4.41192715, Time: 0.0134 Steps: 13650, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000886, Sample Num: 14176, Cur Loss: 0.61996508, Cur Avg Loss: 1.08739800, Log Avg loss: 1.13802474, Global Avg Loss: 4.40953044, Time: 0.0069 Steps: 13660, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000896, Sample Num: 14336, Cur Loss: 0.81487799, Cur Avg Loss: 1.09164292, Log Avg loss: 1.46774343, Global Avg Loss: 4.40737844, Time: 0.0069 Steps: 13670, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000906, Sample Num: 14496, Cur Loss: 1.37242734, Cur Avg Loss: 1.09353779, Log Avg loss: 1.26331810, Global Avg Loss: 4.40508015, Time: 0.0116 Steps: 13680, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000916, Sample Num: 14656, Cur Loss: 0.44977844, Cur Avg Loss: 1.09323725, Log Avg loss: 1.06600779, Global Avg Loss: 4.40264109, Time: 0.0118 Steps: 13690, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000926, Sample Num: 14816, Cur Loss: 0.60230041, Cur Avg Loss: 1.09242653, Log Avg loss: 1.01816434, Global Avg Loss: 4.40017067, Time: 0.0113 Steps: 13700, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000936, Sample Num: 14976, Cur Loss: 1.05363905, Cur Avg Loss: 1.09120482, Log Avg loss: 0.97807536, Global Avg Loss: 4.39767461, Time: 0.0068 Steps: 13710, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000946, Sample Num: 15136, Cur Loss: 0.65365386, Cur Avg Loss: 1.08970131, Log Avg loss: 0.94897234, Global Avg Loss: 4.39516098, Time: 0.0114 Steps: 13720, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000956, Sample Num: 15296, Cur Loss: 1.67559695, Cur Avg Loss: 1.09120869, Log Avg loss: 1.23380712, Global Avg Loss: 4.39285846, Time: 0.0077 Steps: 13730, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000966, Sample Num: 15456, Cur Loss: 0.64067912, Cur Avg Loss: 1.09280319, Log Avg loss: 1.24523677, Global Avg Loss: 4.39056762, Time: 0.0106 Steps: 13740, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000976, Sample Num: 15616, Cur Loss: 0.56280077, Cur Avg Loss: 1.09164783, Log Avg loss: 0.98004053, Global Avg Loss: 4.38808723, Time: 0.0119 Steps: 13750, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000986, Sample Num: 15776, Cur Loss: 0.54854846, Cur Avg Loss: 1.09231212, Log Avg loss: 1.15714682, Global Avg Loss: 4.38573917, Time: 0.0102 Steps: 13760, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 000996, Sample Num: 15936, Cur Loss: 0.92010272, Cur Avg Loss: 1.09269015, Log Avg loss: 1.12996420, Global Avg Loss: 4.38337477, Time: 0.0128 Steps: 13770, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001006, Sample Num: 16096, Cur Loss: 1.31260598, Cur Avg Loss: 1.09283171, Log Avg loss: 1.10693067, Global Avg Loss: 4.38099709, Time: 0.0072 Steps: 13780, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001016, Sample Num: 16256, Cur Loss: 1.69220567, Cur Avg Loss: 1.09204793, Log Avg loss: 1.01320006, Global Avg Loss: 4.37855489, Time: 0.0067 Steps: 13790, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001026, Sample Num: 16416, Cur Loss: 1.32484245, Cur Avg Loss: 1.09184481, Log Avg loss: 1.07120739, Global Avg Loss: 4.37615826, Time: 0.0111 Steps: 13800, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001036, Sample Num: 16576, Cur Loss: 0.71244538, Cur Avg Loss: 1.09277578, Log Avg loss: 1.18829327, Global Avg Loss: 4.37384988, Time: 0.0065 Steps: 13810, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001046, Sample Num: 16736, Cur Loss: 1.41159225, Cur Avg Loss: 1.09223091, Log Avg loss: 1.03578211, Global Avg Loss: 4.37143449, Time: 0.0116 Steps: 13820, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001056, Sample Num: 16896, Cur Loss: 0.86733818, Cur Avg Loss: 1.09269680, Log Avg loss: 1.14142957, Global Avg Loss: 4.36909899, Time: 0.0074 Steps: 13830, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001066, Sample Num: 17056, Cur Loss: 1.18000567, Cur Avg Loss: 1.09461043, Log Avg loss: 1.29668973, Global Avg Loss: 4.36687904, Time: 0.0134 Steps: 13840, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001076, Sample Num: 17216, Cur Loss: 1.57255423, Cur Avg Loss: 1.09193123, Log Avg loss: 0.80632822, Global Avg Loss: 4.36430825, Time: 0.0111 Steps: 13850, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001086, Sample Num: 17376, Cur Loss: 1.02573490, Cur Avg Loss: 1.09531032, Log Avg loss: 1.45890024, Global Avg Loss: 4.36221199, Time: 0.0068 Steps: 13860, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001096, Sample Num: 17536, Cur Loss: 1.14048648, Cur Avg Loss: 1.09303031, Log Avg loss: 0.84542089, Global Avg Loss: 4.35967645, Time: 0.0088 Steps: 13870, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001106, Sample Num: 17696, Cur Loss: 0.59599972, Cur Avg Loss: 1.09162415, Log Avg loss: 0.93750942, Global Avg Loss: 4.35721091, Time: 0.0125 Steps: 13880, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001116, Sample Num: 17856, Cur Loss: 1.50345230, Cur Avg Loss: 1.09064291, Log Avg loss: 0.98211835, Global Avg Loss: 4.35478104, Time: 0.0108 Steps: 13890, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001126, Sample Num: 18016, Cur Loss: 1.87164068, Cur Avg Loss: 1.09248855, Log Avg loss: 1.29846122, Global Avg Loss: 4.35258225, Time: 0.0106 Steps: 13900, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001136, Sample Num: 18176, Cur Loss: 2.40148449, Cur Avg Loss: 1.09466021, Log Avg loss: 1.33918924, Global Avg Loss: 4.35041590, Time: 0.0119 Steps: 13910, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001146, Sample Num: 18336, Cur Loss: 0.50904143, Cur Avg Loss: 1.09402423, Log Avg loss: 1.02177709, Global Avg Loss: 4.34802464, Time: 0.0067 Steps: 13920, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001156, Sample Num: 18496, Cur Loss: 1.18689275, Cur Avg Loss: 1.09499419, Log Avg loss: 1.20615183, Global Avg Loss: 4.34576917, Time: 0.0109 Steps: 13930, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001166, Sample Num: 18656, Cur Loss: 1.23510170, Cur Avg Loss: 1.09560416, Log Avg loss: 1.16611695, Global Avg Loss: 4.34348821, Time: 0.0123 Steps: 13940, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001176, Sample Num: 18816, Cur Loss: 1.35332549, Cur Avg Loss: 1.09358760, Log Avg loss: 0.85845589, Global Avg Loss: 4.34098998, Time: 0.0126 Steps: 13950, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001186, Sample Num: 18976, Cur Loss: 1.49276900, Cur Avg Loss: 1.09204116, Log Avg loss: 0.91018053, Global Avg Loss: 4.33853238, Time: 0.0132 Steps: 13960, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001196, Sample Num: 19136, Cur Loss: 1.28026640, Cur Avg Loss: 1.09153151, Log Avg loss: 1.03108702, Global Avg Loss: 4.33616484, Time: 0.0111 Steps: 13970, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001206, Sample Num: 19296, Cur Loss: 1.08446538, Cur Avg Loss: 1.09101655, Log Avg loss: 1.02942719, Global Avg Loss: 4.33379951, Time: 0.0064 Steps: 13980, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001216, Sample Num: 19456, Cur Loss: 0.56837732, Cur Avg Loss: 1.09172542, Log Avg loss: 1.17721532, Global Avg Loss: 4.33154320, Time: 0.0106 Steps: 13990, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001226, Sample Num: 19616, Cur Loss: 0.77185333, Cur Avg Loss: 1.09124533, Log Avg loss: 1.03286593, Global Avg Loss: 4.32918700, Time: 0.0084 Steps: 14000, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001236, Sample Num: 19776, Cur Loss: 0.35629562, Cur Avg Loss: 1.08970694, Log Avg loss: 0.90110053, Global Avg Loss: 4.32674011, Time: 0.0086 Steps: 14010, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001246, Sample Num: 19936, Cur Loss: 1.30852270, Cur Avg Loss: 1.09316533, Log Avg loss: 1.52062269, Global Avg Loss: 4.32473860, Time: 0.0077 Steps: 14020, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001256, Sample Num: 20096, Cur Loss: 2.09370780, Cur Avg Loss: 1.09456771, Log Avg loss: 1.26930411, Global Avg Loss: 4.32256082, Time: 0.0068 Steps: 14030, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001266, Sample Num: 20256, Cur Loss: 0.62078017, Cur Avg Loss: 1.09490437, Log Avg loss: 1.13718832, Global Avg Loss: 4.32029203, Time: 0.0066 Steps: 14040, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001276, Sample Num: 20416, Cur Loss: 0.80941772, Cur Avg Loss: 1.09436412, Log Avg loss: 1.02596877, Global Avg Loss: 4.31794732, Time: 0.0111 Steps: 14050, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001286, Sample Num: 20576, Cur Loss: 0.53308076, Cur Avg Loss: 1.09222129, Log Avg loss: 0.81879537, Global Avg Loss: 4.31545859, Time: 0.0064 Steps: 14060, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001296, Sample Num: 20736, Cur Loss: 0.79276586, Cur Avg Loss: 1.09141294, Log Avg loss: 0.98745988, Global Avg Loss: 4.31309327, Time: 0.0187 Steps: 14070, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001306, Sample Num: 20896, Cur Loss: 0.63879639, Cur Avg Loss: 1.09161988, Log Avg loss: 1.11843923, Global Avg Loss: 4.31082434, Time: 0.0064 Steps: 14080, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001316, Sample Num: 21056, Cur Loss: 0.24642220, Cur Avg Loss: 1.09103085, Log Avg loss: 1.01410294, Global Avg Loss: 4.30848458, Time: 0.0066 Steps: 14090, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001326, Sample Num: 21216, Cur Loss: 1.62743211, Cur Avg Loss: 1.09400178, Log Avg loss: 1.48497727, Global Avg Loss: 4.30648210, Time: 0.0074 Steps: 14100, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001336, Sample Num: 21376, Cur Loss: 1.58126354, Cur Avg Loss: 1.09457824, Log Avg loss: 1.17101613, Global Avg Loss: 4.30425994, Time: 0.0066 Steps: 14110, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001346, Sample Num: 21536, Cur Loss: 0.92953980, Cur Avg Loss: 1.09341205, Log Avg loss: 0.93760915, Global Avg Loss: 4.30187562, Time: 0.0112 Steps: 14120, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001356, Sample Num: 21696, Cur Loss: 3.65594435, Cur Avg Loss: 1.09548101, Log Avg loss: 1.37396275, Global Avg Loss: 4.29980350, Time: 0.0090 Steps: 14130, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001366, Sample Num: 21856, Cur Loss: 0.84102976, Cur Avg Loss: 1.09491152, Log Avg loss: 1.01768857, Global Avg Loss: 4.29748234, Time: 0.0129 Steps: 14140, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001376, Sample Num: 22016, Cur Loss: 0.86218500, Cur Avg Loss: 1.09544650, Log Avg loss: 1.16852527, Global Avg Loss: 4.29527107, Time: 0.0115 Steps: 14150, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001386, Sample Num: 22176, Cur Loss: 1.39410758, Cur Avg Loss: 1.09485471, Log Avg loss: 1.01342399, Global Avg Loss: 4.29295338, Time: 0.0066 Steps: 14160, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001396, Sample Num: 22336, Cur Loss: 2.52306604, Cur Avg Loss: 1.09538191, Log Avg loss: 1.16845161, Global Avg Loss: 4.29074836, Time: 0.0067 Steps: 14170, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001406, Sample Num: 22496, Cur Loss: 0.99015713, Cur Avg Loss: 1.09695730, Log Avg loss: 1.31688184, Global Avg Loss: 4.28865114, Time: 0.0068 Steps: 14180, Updated lr: 0.000088 Training, Epoch: 0007, Batch: 001416, Sample Num: 22656, Cur Loss: 0.50441891, Cur Avg Loss: 1.09582763, Log Avg loss: 0.93699601, Global Avg Loss: 4.28628915, Time: 0.0068 Steps: 14190, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001426, Sample Num: 22816, Cur Loss: 0.84900564, Cur Avg Loss: 1.09715249, Log Avg loss: 1.28475331, Global Avg Loss: 4.28417540, Time: 0.0092 Steps: 14200, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001436, Sample Num: 22976, Cur Loss: 0.62347567, Cur Avg Loss: 1.09775635, Log Avg loss: 1.18386632, Global Avg Loss: 4.28199362, Time: 0.0073 Steps: 14210, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001446, Sample Num: 23136, Cur Loss: 1.68623853, Cur Avg Loss: 1.09587278, Log Avg loss: 0.82539263, Global Avg Loss: 4.27956282, Time: 0.0067 Steps: 14220, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001456, Sample Num: 23296, Cur Loss: 1.15578079, Cur Avg Loss: 1.09563654, Log Avg loss: 1.06147556, Global Avg Loss: 4.27730133, Time: 0.0234 Steps: 14230, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001466, Sample Num: 23456, Cur Loss: 0.79649949, Cur Avg Loss: 1.09487061, Log Avg loss: 0.98335220, Global Avg Loss: 4.27498817, Time: 0.0119 Steps: 14240, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001476, Sample Num: 23616, Cur Loss: 1.27778053, Cur Avg Loss: 1.09475245, Log Avg loss: 1.07742997, Global Avg Loss: 4.27274427, Time: 0.0234 Steps: 14250, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001486, Sample Num: 23776, Cur Loss: 0.43811098, Cur Avg Loss: 1.09237630, Log Avg loss: 0.74165620, Global Avg Loss: 4.27026805, Time: 0.0116 Steps: 14260, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001496, Sample Num: 23936, Cur Loss: 1.02094829, Cur Avg Loss: 1.09243528, Log Avg loss: 1.10119956, Global Avg Loss: 4.26804726, Time: 0.0112 Steps: 14270, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001506, Sample Num: 24096, Cur Loss: 1.91815162, Cur Avg Loss: 1.09171199, Log Avg loss: 0.98350780, Global Avg Loss: 4.26574716, Time: 0.0235 Steps: 14280, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001516, Sample Num: 24256, Cur Loss: 1.54830587, Cur Avg Loss: 1.09092841, Log Avg loss: 0.97292148, Global Avg Loss: 4.26344287, Time: 0.0066 Steps: 14290, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001526, Sample Num: 24416, Cur Loss: 1.86775863, Cur Avg Loss: 1.09108469, Log Avg loss: 1.11477632, Global Avg Loss: 4.26124101, Time: 0.0221 Steps: 14300, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001536, Sample Num: 24576, Cur Loss: 1.07101727, Cur Avg Loss: 1.09072587, Log Avg loss: 1.03597080, Global Avg Loss: 4.25898715, Time: 0.0130 Steps: 14310, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001546, Sample Num: 24736, Cur Loss: 0.99037731, Cur Avg Loss: 1.08897679, Log Avg loss: 0.82031784, Global Avg Loss: 4.25658585, Time: 0.0069 Steps: 14320, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001556, Sample Num: 24896, Cur Loss: 2.05723023, Cur Avg Loss: 1.09007134, Log Avg loss: 1.25928860, Global Avg Loss: 4.25449422, Time: 0.0108 Steps: 14330, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001566, Sample Num: 25056, Cur Loss: 1.02796125, Cur Avg Loss: 1.08842006, Log Avg loss: 0.83148033, Global Avg Loss: 4.25210718, Time: 0.0113 Steps: 14340, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001576, Sample Num: 25216, Cur Loss: 1.59623086, Cur Avg Loss: 1.08811416, Log Avg loss: 1.04021084, Global Avg Loss: 4.24986893, Time: 0.0115 Steps: 14350, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001586, Sample Num: 25376, Cur Loss: 0.89654487, Cur Avg Loss: 1.08821326, Log Avg loss: 1.10383184, Global Avg Loss: 4.24767809, Time: 0.0115 Steps: 14360, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001596, Sample Num: 25536, Cur Loss: 1.30208492, Cur Avg Loss: 1.08907013, Log Avg loss: 1.22496939, Global Avg Loss: 4.24557461, Time: 0.0113 Steps: 14370, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001606, Sample Num: 25696, Cur Loss: 1.24160683, Cur Avg Loss: 1.09131326, Log Avg loss: 1.44931620, Global Avg Loss: 4.24363006, Time: 0.0071 Steps: 14380, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001616, Sample Num: 25856, Cur Loss: 0.78376162, Cur Avg Loss: 1.09150579, Log Avg loss: 1.12242638, Global Avg Loss: 4.24146105, Time: 0.0107 Steps: 14390, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001626, Sample Num: 26016, Cur Loss: 1.68794072, Cur Avg Loss: 1.09203004, Log Avg loss: 1.17674877, Global Avg Loss: 4.23933278, Time: 0.0116 Steps: 14400, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001636, Sample Num: 26176, Cur Loss: 1.61063075, Cur Avg Loss: 1.09116512, Log Avg loss: 0.95053000, Global Avg Loss: 4.23705047, Time: 0.0151 Steps: 14410, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001646, Sample Num: 26336, Cur Loss: 0.36778748, Cur Avg Loss: 1.08872044, Log Avg loss: 0.68877047, Global Avg Loss: 4.23458981, Time: 0.0113 Steps: 14420, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001656, Sample Num: 26496, Cur Loss: 0.79919785, Cur Avg Loss: 1.08808888, Log Avg loss: 0.98413385, Global Avg Loss: 4.23233724, Time: 0.0132 Steps: 14430, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001666, Sample Num: 26656, Cur Loss: 0.79653466, Cur Avg Loss: 1.08830574, Log Avg loss: 1.12421795, Global Avg Loss: 4.23018480, Time: 0.0114 Steps: 14440, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001676, Sample Num: 26816, Cur Loss: 0.35543376, Cur Avg Loss: 1.08819933, Log Avg loss: 1.07047166, Global Avg Loss: 4.22799815, Time: 0.0103 Steps: 14450, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001686, Sample Num: 26976, Cur Loss: 1.49571872, Cur Avg Loss: 1.08673834, Log Avg loss: 0.84187573, Global Avg Loss: 4.22565643, Time: 0.0070 Steps: 14460, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001696, Sample Num: 27136, Cur Loss: 1.37843215, Cur Avg Loss: 1.08656555, Log Avg loss: 1.05743366, Global Avg Loss: 4.22346692, Time: 0.0067 Steps: 14470, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001706, Sample Num: 27296, Cur Loss: 0.80594933, Cur Avg Loss: 1.08679542, Log Avg loss: 1.12578060, Global Avg Loss: 4.22132764, Time: 0.0161 Steps: 14480, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001716, Sample Num: 27456, Cur Loss: 1.44309461, Cur Avg Loss: 1.08624886, Log Avg loss: 0.99300663, Global Avg Loss: 4.21909967, Time: 0.0110 Steps: 14490, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001726, Sample Num: 27616, Cur Loss: 0.60433662, Cur Avg Loss: 1.08540553, Log Avg loss: 0.94068972, Global Avg Loss: 4.21683870, Time: 0.0116 Steps: 14500, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001736, Sample Num: 27776, Cur Loss: 0.66644973, Cur Avg Loss: 1.08584170, Log Avg loss: 1.16112479, Global Avg Loss: 4.21473276, Time: 0.0117 Steps: 14510, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001746, Sample Num: 27936, Cur Loss: 0.83376324, Cur Avg Loss: 1.08506132, Log Avg loss: 0.94958771, Global Avg Loss: 4.21248404, Time: 0.0109 Steps: 14520, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001756, Sample Num: 28096, Cur Loss: 0.41734788, Cur Avg Loss: 1.08523930, Log Avg loss: 1.11631368, Global Avg Loss: 4.21035316, Time: 0.0067 Steps: 14530, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001766, Sample Num: 28256, Cur Loss: 1.71988285, Cur Avg Loss: 1.08685578, Log Avg loss: 1.37070963, Global Avg Loss: 4.20840017, Time: 0.0156 Steps: 14540, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001776, Sample Num: 28416, Cur Loss: 0.76398367, Cur Avg Loss: 1.08650703, Log Avg loss: 1.02491858, Global Avg Loss: 4.20621221, Time: 0.0069 Steps: 14550, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001786, Sample Num: 28576, Cur Loss: 0.33327261, Cur Avg Loss: 1.08623906, Log Avg loss: 1.03864725, Global Avg Loss: 4.20403669, Time: 0.0118 Steps: 14560, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001796, Sample Num: 28736, Cur Loss: 0.52115250, Cur Avg Loss: 1.08519540, Log Avg loss: 0.89879784, Global Avg Loss: 4.20176816, Time: 0.0067 Steps: 14570, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001806, Sample Num: 28896, Cur Loss: 1.03674960, Cur Avg Loss: 1.08445504, Log Avg loss: 0.95148702, Global Avg Loss: 4.19953889, Time: 0.0064 Steps: 14580, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001816, Sample Num: 29056, Cur Loss: 0.55141616, Cur Avg Loss: 1.08485323, Log Avg loss: 1.15676527, Global Avg Loss: 4.19745337, Time: 0.0096 Steps: 14590, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001826, Sample Num: 29216, Cur Loss: 0.63295460, Cur Avg Loss: 1.08458096, Log Avg loss: 1.03513709, Global Avg Loss: 4.19528740, Time: 0.0068 Steps: 14600, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001836, Sample Num: 29376, Cur Loss: 1.48658729, Cur Avg Loss: 1.08708728, Log Avg loss: 1.54474092, Global Avg Loss: 4.19347320, Time: 0.0063 Steps: 14610, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001846, Sample Num: 29536, Cur Loss: 0.42639720, Cur Avg Loss: 1.08722101, Log Avg loss: 1.11177498, Global Avg Loss: 4.19136533, Time: 0.0072 Steps: 14620, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001856, Sample Num: 29696, Cur Loss: 1.02596653, Cur Avg Loss: 1.08701523, Log Avg loss: 1.04902753, Global Avg Loss: 4.18921746, Time: 0.0096 Steps: 14630, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001866, Sample Num: 29856, Cur Loss: 0.65112031, Cur Avg Loss: 1.08489762, Log Avg loss: 0.69186898, Global Avg Loss: 4.18682856, Time: 0.0069 Steps: 14640, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001876, Sample Num: 30016, Cur Loss: 0.52296513, Cur Avg Loss: 1.08561926, Log Avg loss: 1.22027705, Global Avg Loss: 4.18480361, Time: 0.0065 Steps: 14650, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001886, Sample Num: 30176, Cur Loss: 0.83142781, Cur Avg Loss: 1.08438699, Log Avg loss: 0.85321320, Global Avg Loss: 4.18253104, Time: 0.0068 Steps: 14660, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001896, Sample Num: 30336, Cur Loss: 0.99117815, Cur Avg Loss: 1.08387873, Log Avg loss: 0.98802085, Global Avg Loss: 4.18035346, Time: 0.0184 Steps: 14670, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001906, Sample Num: 30496, Cur Loss: 1.75416207, Cur Avg Loss: 1.08472001, Log Avg loss: 1.24422770, Global Avg Loss: 4.17835337, Time: 0.0064 Steps: 14680, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001916, Sample Num: 30656, Cur Loss: 0.70095253, Cur Avg Loss: 1.08356984, Log Avg loss: 0.86434738, Global Avg Loss: 4.17609741, Time: 0.0117 Steps: 14690, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001926, Sample Num: 30816, Cur Loss: 1.09158576, Cur Avg Loss: 1.08289524, Log Avg loss: 0.95364193, Global Avg Loss: 4.17390527, Time: 0.0070 Steps: 14700, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001936, Sample Num: 30976, Cur Loss: 0.34435242, Cur Avg Loss: 1.08296670, Log Avg loss: 1.09672982, Global Avg Loss: 4.17181337, Time: 0.0115 Steps: 14710, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001946, Sample Num: 31136, Cur Loss: 1.39669263, Cur Avg Loss: 1.08281063, Log Avg loss: 1.05259394, Global Avg Loss: 4.16969434, Time: 0.0073 Steps: 14720, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001956, Sample Num: 31296, Cur Loss: 0.90388304, Cur Avg Loss: 1.08333786, Log Avg loss: 1.18593711, Global Avg Loss: 4.16766871, Time: 0.0065 Steps: 14730, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001966, Sample Num: 31456, Cur Loss: 0.88691932, Cur Avg Loss: 1.08385512, Log Avg loss: 1.18503134, Global Avg Loss: 4.16564521, Time: 0.0075 Steps: 14740, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001976, Sample Num: 31616, Cur Loss: 1.34958005, Cur Avg Loss: 1.08465771, Log Avg loss: 1.24244652, Global Avg Loss: 4.16366338, Time: 0.0176 Steps: 14750, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001986, Sample Num: 31776, Cur Loss: 0.55115438, Cur Avg Loss: 1.08463741, Log Avg loss: 1.08062795, Global Avg Loss: 4.16157460, Time: 0.0117 Steps: 14760, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 001996, Sample Num: 31936, Cur Loss: 0.94393218, Cur Avg Loss: 1.08489257, Log Avg loss: 1.13556587, Global Avg Loss: 4.15952585, Time: 0.0070 Steps: 14770, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002006, Sample Num: 32096, Cur Loss: 0.99158978, Cur Avg Loss: 1.08461418, Log Avg loss: 1.02904738, Global Avg Loss: 4.15740780, Time: 0.0065 Steps: 14780, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002016, Sample Num: 32256, Cur Loss: 1.62567735, Cur Avg Loss: 1.08591321, Log Avg loss: 1.34650010, Global Avg Loss: 4.15550725, Time: 0.0100 Steps: 14790, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002026, Sample Num: 32416, Cur Loss: 0.82874012, Cur Avg Loss: 1.08524325, Log Avg loss: 0.95017943, Global Avg Loss: 4.15334149, Time: 0.0068 Steps: 14800, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002036, Sample Num: 32576, Cur Loss: 1.16129613, Cur Avg Loss: 1.08458127, Log Avg loss: 0.95046260, Global Avg Loss: 4.15117884, Time: 0.0068 Steps: 14810, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002046, Sample Num: 32736, Cur Loss: 0.27075770, Cur Avg Loss: 1.08381091, Log Avg loss: 0.92696618, Global Avg Loss: 4.14900326, Time: 0.0067 Steps: 14820, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002056, Sample Num: 32896, Cur Loss: 0.77431691, Cur Avg Loss: 1.08236958, Log Avg loss: 0.78747422, Global Avg Loss: 4.14673655, Time: 0.0109 Steps: 14830, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002066, Sample Num: 33056, Cur Loss: 1.06066513, Cur Avg Loss: 1.08120757, Log Avg loss: 0.84229786, Global Avg Loss: 4.14450984, Time: 0.0126 Steps: 14840, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002076, Sample Num: 33216, Cur Loss: 0.91272795, Cur Avg Loss: 1.08021747, Log Avg loss: 0.87566267, Global Avg Loss: 4.14230860, Time: 0.0115 Steps: 14850, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002086, Sample Num: 33376, Cur Loss: 1.62117767, Cur Avg Loss: 1.07881266, Log Avg loss: 0.78717398, Global Avg Loss: 4.14005077, Time: 0.0111 Steps: 14860, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002096, Sample Num: 33536, Cur Loss: 1.18911862, Cur Avg Loss: 1.07894747, Log Avg loss: 1.10706927, Global Avg Loss: 4.13801110, Time: 0.0071 Steps: 14870, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002106, Sample Num: 33696, Cur Loss: 0.95992506, Cur Avg Loss: 1.07930842, Log Avg loss: 1.15496401, Global Avg Loss: 4.13600637, Time: 0.0226 Steps: 14880, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002116, Sample Num: 33856, Cur Loss: 1.31117988, Cur Avg Loss: 1.07875845, Log Avg loss: 0.96293447, Global Avg Loss: 4.13387536, Time: 0.0225 Steps: 14890, Updated lr: 0.000087 Training, Epoch: 0007, Batch: 002126, Sample Num: 34016, Cur Loss: 1.15768671, Cur Avg Loss: 1.07877744, Log Avg loss: 1.08279478, Global Avg Loss: 4.13182765, Time: 0.0069 Steps: 14900, Updated lr: 0.000087 ***** Running evaluation checkpoint-14903 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-14903 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.871292, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.063231, "eval_total_loss": 747.451696, "eval_mae": 0.816358, "eval_mse": 1.063462, "eval_r2": 0.323994, "eval_sp_statistic": 0.55907, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.573242, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 2.428888, "test_total_loss": 1219.301923, "test_mae": 1.445779, "test_mse": 2.428544, "test_r2": -0.567403, "test_sp_statistic": 0.26721, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.399673, "test_ps_pvalue": 0.0, "lr": 8.681555239449977e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 4.13131009699232, "train_cur_epoch_loss": 2298.163185968995, "train_cur_epoch_avg_loss": 1.0794566397224026, "train_cur_epoch_time": 21.871291875839233, "train_cur_epoch_avg_time": 0.010273035169487663, "epoch": 7, "step": 14903} ################################################## Training, Epoch: 0008, Batch: 000007, Sample Num: 112, Cur Loss: 1.47192693, Cur Avg Loss: 1.49879839, Log Avg loss: 1.51739405, Global Avg Loss: 4.13007418, Time: 0.0101 Steps: 14910, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000017, Sample Num: 272, Cur Loss: 0.48574924, Cur Avg Loss: 1.17556237, Log Avg loss: 0.94929716, Global Avg Loss: 4.12794229, Time: 0.0114 Steps: 14920, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000027, Sample Num: 432, Cur Loss: 0.92532897, Cur Avg Loss: 1.06684369, Log Avg loss: 0.88202194, Global Avg Loss: 4.12576820, Time: 0.0109 Steps: 14930, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000037, Sample Num: 592, Cur Loss: 0.49308288, Cur Avg Loss: 0.99756883, Log Avg loss: 0.81052669, Global Avg Loss: 4.12354916, Time: 0.0068 Steps: 14940, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000047, Sample Num: 752, Cur Loss: 0.98298860, Cur Avg Loss: 1.05861242, Log Avg loss: 1.28447372, Global Avg Loss: 4.12165011, Time: 0.0067 Steps: 14950, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000057, Sample Num: 912, Cur Loss: 1.89609456, Cur Avg Loss: 1.06351925, Log Avg loss: 1.08658133, Global Avg Loss: 4.11962132, Time: 0.0069 Steps: 14960, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000067, Sample Num: 1072, Cur Loss: 1.64411199, Cur Avg Loss: 1.06614054, Log Avg loss: 1.08108190, Global Avg Loss: 4.11759157, Time: 0.0161 Steps: 14970, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000077, Sample Num: 1232, Cur Loss: 0.27560049, Cur Avg Loss: 1.08018778, Log Avg loss: 1.17430429, Global Avg Loss: 4.11562676, Time: 0.0223 Steps: 14980, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000087, Sample Num: 1392, Cur Loss: 0.64728880, Cur Avg Loss: 1.12845749, Log Avg loss: 1.50013428, Global Avg Loss: 4.11388193, Time: 0.0123 Steps: 14990, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000097, Sample Num: 1552, Cur Loss: 0.38694477, Cur Avg Loss: 1.12189141, Log Avg loss: 1.06476646, Global Avg Loss: 4.11184919, Time: 0.0073 Steps: 15000, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000107, Sample Num: 1712, Cur Loss: 0.52326912, Cur Avg Loss: 1.11587694, Log Avg loss: 1.05753663, Global Avg Loss: 4.10981434, Time: 0.0068 Steps: 15010, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000117, Sample Num: 1872, Cur Loss: 0.92525685, Cur Avg Loss: 1.12385885, Log Avg loss: 1.20926526, Global Avg Loss: 4.10788321, Time: 0.0071 Steps: 15020, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000127, Sample Num: 2032, Cur Loss: 1.51419437, Cur Avg Loss: 1.13705975, Log Avg loss: 1.29151028, Global Avg Loss: 4.10600938, Time: 0.0067 Steps: 15030, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000137, Sample Num: 2192, Cur Loss: 0.81599963, Cur Avg Loss: 1.12767886, Log Avg loss: 1.00854161, Global Avg Loss: 4.10394989, Time: 0.0116 Steps: 15040, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000147, Sample Num: 2352, Cur Loss: 0.53309906, Cur Avg Loss: 1.11663220, Log Avg loss: 0.96529289, Global Avg Loss: 4.10186441, Time: 0.0117 Steps: 15050, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000157, Sample Num: 2512, Cur Loss: 1.33859634, Cur Avg Loss: 1.12149410, Log Avg loss: 1.19296402, Global Avg Loss: 4.09993287, Time: 0.0078 Steps: 15060, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000167, Sample Num: 2672, Cur Loss: 0.89735574, Cur Avg Loss: 1.10890759, Log Avg loss: 0.91129937, Global Avg Loss: 4.09781698, Time: 0.0108 Steps: 15070, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000177, Sample Num: 2832, Cur Loss: 0.44059190, Cur Avg Loss: 1.09886704, Log Avg loss: 0.93118995, Global Avg Loss: 4.09571710, Time: 0.0072 Steps: 15080, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000187, Sample Num: 2992, Cur Loss: 0.88794827, Cur Avg Loss: 1.10380738, Log Avg loss: 1.19125127, Global Avg Loss: 4.09379234, Time: 0.0067 Steps: 15090, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000197, Sample Num: 3152, Cur Loss: 0.47952557, Cur Avg Loss: 1.09878919, Log Avg loss: 1.00494920, Global Avg Loss: 4.09174674, Time: 0.0066 Steps: 15100, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000207, Sample Num: 3312, Cur Loss: 0.71935785, Cur Avg Loss: 1.09890097, Log Avg loss: 1.10110286, Global Avg Loss: 4.08976750, Time: 0.0068 Steps: 15110, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000217, Sample Num: 3472, Cur Loss: 1.12931979, Cur Avg Loss: 1.09086301, Log Avg loss: 0.92447728, Global Avg Loss: 4.08767405, Time: 0.0112 Steps: 15120, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000227, Sample Num: 3632, Cur Loss: 0.53790283, Cur Avg Loss: 1.09702518, Log Avg loss: 1.23074442, Global Avg Loss: 4.08578580, Time: 0.0119 Steps: 15130, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000237, Sample Num: 3792, Cur Loss: 0.77929652, Cur Avg Loss: 1.09044186, Log Avg loss: 0.94100030, Global Avg Loss: 4.08370866, Time: 0.0110 Steps: 15140, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000247, Sample Num: 3952, Cur Loss: 1.67249680, Cur Avg Loss: 1.08991179, Log Avg loss: 1.07734920, Global Avg Loss: 4.08172426, Time: 0.0070 Steps: 15150, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000257, Sample Num: 4112, Cur Loss: 1.53691232, Cur Avg Loss: 1.08911010, Log Avg loss: 1.06930832, Global Avg Loss: 4.07973718, Time: 0.0159 Steps: 15160, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000267, Sample Num: 4272, Cur Loss: 0.35219324, Cur Avg Loss: 1.07993267, Log Avg loss: 0.84407287, Global Avg Loss: 4.07760425, Time: 0.0066 Steps: 15170, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000277, Sample Num: 4432, Cur Loss: 1.49918830, Cur Avg Loss: 1.08007662, Log Avg loss: 1.08391998, Global Avg Loss: 4.07563212, Time: 0.0071 Steps: 15180, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000287, Sample Num: 4592, Cur Loss: 0.72594678, Cur Avg Loss: 1.07947660, Log Avg loss: 1.06285604, Global Avg Loss: 4.07364873, Time: 0.0092 Steps: 15190, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000297, Sample Num: 4752, Cur Loss: 0.75466180, Cur Avg Loss: 1.07686002, Log Avg loss: 1.00176415, Global Avg Loss: 4.07162775, Time: 0.0067 Steps: 15200, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000307, Sample Num: 4912, Cur Loss: 2.00072026, Cur Avg Loss: 1.08031168, Log Avg loss: 1.18282616, Global Avg Loss: 4.06972847, Time: 0.0136 Steps: 15210, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000317, Sample Num: 5072, Cur Loss: 0.37189338, Cur Avg Loss: 1.07770794, Log Avg loss: 0.99777300, Global Avg Loss: 4.06771010, Time: 0.0134 Steps: 15220, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000327, Sample Num: 5232, Cur Loss: 1.17850578, Cur Avg Loss: 1.08274182, Log Avg loss: 1.24231588, Global Avg Loss: 4.06585495, Time: 0.0085 Steps: 15230, Updated lr: 0.000087 Training, Epoch: 0008, Batch: 000337, Sample Num: 5392, Cur Loss: 0.94985771, Cur Avg Loss: 1.07988176, Log Avg loss: 0.98635782, Global Avg Loss: 4.06383429, Time: 0.0126 Steps: 15240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000347, Sample Num: 5552, Cur Loss: 2.10869527, Cur Avg Loss: 1.07892568, Log Avg loss: 1.04670563, Global Avg Loss: 4.06185584, Time: 0.0110 Steps: 15250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000357, Sample Num: 5712, Cur Loss: 1.25831914, Cur Avg Loss: 1.07448394, Log Avg loss: 0.92035561, Global Avg Loss: 4.05979719, Time: 0.0109 Steps: 15260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000367, Sample Num: 5872, Cur Loss: 1.54772854, Cur Avg Loss: 1.07165006, Log Avg loss: 0.97048050, Global Avg Loss: 4.05777406, Time: 0.0078 Steps: 15270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000377, Sample Num: 6032, Cur Loss: 0.26283297, Cur Avg Loss: 1.07227968, Log Avg loss: 1.09538662, Global Avg Loss: 4.05583533, Time: 0.0114 Steps: 15280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000387, Sample Num: 6192, Cur Loss: 1.14691234, Cur Avg Loss: 1.07167113, Log Avg loss: 1.04872878, Global Avg Loss: 4.05386861, Time: 0.0108 Steps: 15290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000397, Sample Num: 6352, Cur Loss: 0.51516271, Cur Avg Loss: 1.07024930, Log Avg loss: 1.01522454, Global Avg Loss: 4.05188257, Time: 0.0115 Steps: 15300, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000407, Sample Num: 6512, Cur Loss: 1.06174767, Cur Avg Loss: 1.07105087, Log Avg loss: 1.10287336, Global Avg Loss: 4.04995637, Time: 0.0106 Steps: 15310, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000417, Sample Num: 6672, Cur Loss: 0.68596083, Cur Avg Loss: 1.06977602, Log Avg loss: 1.01788962, Global Avg Loss: 4.04797722, Time: 0.0114 Steps: 15320, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000427, Sample Num: 6832, Cur Loss: 0.52193797, Cur Avg Loss: 1.06582029, Log Avg loss: 0.90086648, Global Avg Loss: 4.04592431, Time: 0.0110 Steps: 15330, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000437, Sample Num: 6992, Cur Loss: 1.40071547, Cur Avg Loss: 1.06552408, Log Avg loss: 1.05287554, Global Avg Loss: 4.04397317, Time: 0.0072 Steps: 15340, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000447, Sample Num: 7152, Cur Loss: 1.04796958, Cur Avg Loss: 1.06599976, Log Avg loss: 1.08678721, Global Avg Loss: 4.04204666, Time: 0.0066 Steps: 15350, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000457, Sample Num: 7312, Cur Loss: 0.68753386, Cur Avg Loss: 1.06484441, Log Avg loss: 1.01320007, Global Avg Loss: 4.04007476, Time: 0.0160 Steps: 15360, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000467, Sample Num: 7472, Cur Loss: 1.74298573, Cur Avg Loss: 1.06671132, Log Avg loss: 1.15202915, Global Avg Loss: 4.03819574, Time: 0.0103 Steps: 15370, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000477, Sample Num: 7632, Cur Loss: 1.99124277, Cur Avg Loss: 1.06937715, Log Avg loss: 1.19387175, Global Avg Loss: 4.03634638, Time: 0.0105 Steps: 15380, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000487, Sample Num: 7792, Cur Loss: 0.77712512, Cur Avg Loss: 1.06698774, Log Avg loss: 0.95301280, Global Avg Loss: 4.03434291, Time: 0.0074 Steps: 15390, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000497, Sample Num: 7952, Cur Loss: 1.70757353, Cur Avg Loss: 1.06643870, Log Avg loss: 1.03970041, Global Avg Loss: 4.03239834, Time: 0.0068 Steps: 15400, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000507, Sample Num: 8112, Cur Loss: 3.11113191, Cur Avg Loss: 1.06793816, Log Avg loss: 1.14246137, Global Avg Loss: 4.03052297, Time: 0.0067 Steps: 15410, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000517, Sample Num: 8272, Cur Loss: 0.37149173, Cur Avg Loss: 1.06322074, Log Avg loss: 0.82404745, Global Avg Loss: 4.02844355, Time: 0.0120 Steps: 15420, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000527, Sample Num: 8432, Cur Loss: 0.51776481, Cur Avg Loss: 1.06270084, Log Avg loss: 1.03582195, Global Avg Loss: 4.02650406, Time: 0.0154 Steps: 15430, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000537, Sample Num: 8592, Cur Loss: 0.33030647, Cur Avg Loss: 1.06306755, Log Avg loss: 1.08239331, Global Avg Loss: 4.02459726, Time: 0.0068 Steps: 15440, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000547, Sample Num: 8752, Cur Loss: 0.52857369, Cur Avg Loss: 1.06243923, Log Avg loss: 1.02869809, Global Avg Loss: 4.02265816, Time: 0.0115 Steps: 15450, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000557, Sample Num: 8912, Cur Loss: 1.16073775, Cur Avg Loss: 1.06202416, Log Avg loss: 1.03932008, Global Avg Loss: 4.02072845, Time: 0.0127 Steps: 15460, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000567, Sample Num: 9072, Cur Loss: 1.70975101, Cur Avg Loss: 1.06222820, Log Avg loss: 1.07359348, Global Avg Loss: 4.01882339, Time: 0.0073 Steps: 15470, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000577, Sample Num: 9232, Cur Loss: 1.16070747, Cur Avg Loss: 1.06212162, Log Avg loss: 1.05607831, Global Avg Loss: 4.01690947, Time: 0.0066 Steps: 15480, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000587, Sample Num: 9392, Cur Loss: 1.64267528, Cur Avg Loss: 1.06104618, Log Avg loss: 0.99899322, Global Avg Loss: 4.01496117, Time: 0.0140 Steps: 15490, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000597, Sample Num: 9552, Cur Loss: 1.21875441, Cur Avg Loss: 1.06074182, Log Avg loss: 1.04287583, Global Avg Loss: 4.01304369, Time: 0.0141 Steps: 15500, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000607, Sample Num: 9712, Cur Loss: 1.61827087, Cur Avg Loss: 1.06093341, Log Avg loss: 1.07237126, Global Avg Loss: 4.01114771, Time: 0.0133 Steps: 15510, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000617, Sample Num: 9872, Cur Loss: 0.98135585, Cur Avg Loss: 1.05841242, Log Avg loss: 0.90538872, Global Avg Loss: 4.00914657, Time: 0.0065 Steps: 15520, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000627, Sample Num: 10032, Cur Loss: 0.97584170, Cur Avg Loss: 1.05609860, Log Avg loss: 0.91333569, Global Avg Loss: 4.00715314, Time: 0.0123 Steps: 15530, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000637, Sample Num: 10192, Cur Loss: 0.58455563, Cur Avg Loss: 1.05613886, Log Avg loss: 1.05866342, Global Avg Loss: 4.00525578, Time: 0.0118 Steps: 15540, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000647, Sample Num: 10352, Cur Loss: 2.05976224, Cur Avg Loss: 1.05535433, Log Avg loss: 1.00537959, Global Avg Loss: 4.00332660, Time: 0.0138 Steps: 15550, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000657, Sample Num: 10512, Cur Loss: 1.38891840, Cur Avg Loss: 1.05476168, Log Avg loss: 1.01641732, Global Avg Loss: 4.00140699, Time: 0.0070 Steps: 15560, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000667, Sample Num: 10672, Cur Loss: 1.19716108, Cur Avg Loss: 1.05756365, Log Avg loss: 1.24165260, Global Avg Loss: 3.99963451, Time: 0.0071 Steps: 15570, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000677, Sample Num: 10832, Cur Loss: 1.63973725, Cur Avg Loss: 1.05704210, Log Avg loss: 1.02225531, Global Avg Loss: 3.99772348, Time: 0.0068 Steps: 15580, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000687, Sample Num: 10992, Cur Loss: 0.65032542, Cur Avg Loss: 1.05432766, Log Avg loss: 0.87055956, Global Avg Loss: 3.99571761, Time: 0.0112 Steps: 15590, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000697, Sample Num: 11152, Cur Loss: 0.58309424, Cur Avg Loss: 1.05125933, Log Avg loss: 0.84046499, Global Avg Loss: 3.99369501, Time: 0.0067 Steps: 15600, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000707, Sample Num: 11312, Cur Loss: 0.72769904, Cur Avg Loss: 1.05182827, Log Avg loss: 1.09148337, Global Avg Loss: 3.99183581, Time: 0.0109 Steps: 15610, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000717, Sample Num: 11472, Cur Loss: 1.54893708, Cur Avg Loss: 1.04905770, Log Avg loss: 0.85317893, Global Avg Loss: 3.98982642, Time: 0.0135 Steps: 15620, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000727, Sample Num: 11632, Cur Loss: 1.69916296, Cur Avg Loss: 1.05098875, Log Avg loss: 1.18944469, Global Avg Loss: 3.98803475, Time: 0.0112 Steps: 15630, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000737, Sample Num: 11792, Cur Loss: 3.62809753, Cur Avg Loss: 1.05143820, Log Avg loss: 1.08411306, Global Avg Loss: 3.98617803, Time: 0.0080 Steps: 15640, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000747, Sample Num: 11952, Cur Loss: 0.94782257, Cur Avg Loss: 1.05369401, Log Avg loss: 1.21994776, Global Avg Loss: 3.98441047, Time: 0.0064 Steps: 15650, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000757, Sample Num: 12112, Cur Loss: 1.91176677, Cur Avg Loss: 1.05488101, Log Avg loss: 1.14354938, Global Avg Loss: 3.98259638, Time: 0.0065 Steps: 15660, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000767, Sample Num: 12272, Cur Loss: 0.97485220, Cur Avg Loss: 1.05087216, Log Avg loss: 0.74740267, Global Avg Loss: 3.98053180, Time: 0.0080 Steps: 15670, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000777, Sample Num: 12432, Cur Loss: 0.59941816, Cur Avg Loss: 1.04808827, Log Avg loss: 0.83456391, Global Avg Loss: 3.97852544, Time: 0.0096 Steps: 15680, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000787, Sample Num: 12592, Cur Loss: 1.36469650, Cur Avg Loss: 1.04620336, Log Avg loss: 0.89974540, Global Avg Loss: 3.97656319, Time: 0.0067 Steps: 15690, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000797, Sample Num: 12752, Cur Loss: 0.63855553, Cur Avg Loss: 1.04698422, Log Avg loss: 1.10843784, Global Avg Loss: 3.97473636, Time: 0.0066 Steps: 15700, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000807, Sample Num: 12912, Cur Loss: 1.44386816, Cur Avg Loss: 1.04790275, Log Avg loss: 1.12110952, Global Avg Loss: 3.97291992, Time: 0.0066 Steps: 15710, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000817, Sample Num: 13072, Cur Loss: 0.95329726, Cur Avg Loss: 1.04783633, Log Avg loss: 1.04247655, Global Avg Loss: 3.97105577, Time: 0.0156 Steps: 15720, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000827, Sample Num: 13232, Cur Loss: 0.83881241, Cur Avg Loss: 1.04699797, Log Avg loss: 0.97850382, Global Avg Loss: 3.96915332, Time: 0.0131 Steps: 15730, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000837, Sample Num: 13392, Cur Loss: 0.86909324, Cur Avg Loss: 1.04726695, Log Avg loss: 1.06951190, Global Avg Loss: 3.96731111, Time: 0.0112 Steps: 15740, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000847, Sample Num: 13552, Cur Loss: 1.86150169, Cur Avg Loss: 1.04722224, Log Avg loss: 1.04347990, Global Avg Loss: 3.96545471, Time: 0.0120 Steps: 15750, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000857, Sample Num: 13712, Cur Loss: 0.84520167, Cur Avg Loss: 1.04750160, Log Avg loss: 1.07116353, Global Avg Loss: 3.96361823, Time: 0.0114 Steps: 15760, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000867, Sample Num: 13872, Cur Loss: 2.13502407, Cur Avg Loss: 1.04651147, Log Avg loss: 0.96165693, Global Avg Loss: 3.96171464, Time: 0.0146 Steps: 15770, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000877, Sample Num: 14032, Cur Loss: 0.79038000, Cur Avg Loss: 1.04688321, Log Avg loss: 1.07911295, Global Avg Loss: 3.95988789, Time: 0.0067 Steps: 15780, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000887, Sample Num: 14192, Cur Loss: 0.67355990, Cur Avg Loss: 1.04601100, Log Avg loss: 0.96951892, Global Avg Loss: 3.95799406, Time: 0.0128 Steps: 15790, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000897, Sample Num: 14352, Cur Loss: 1.50995874, Cur Avg Loss: 1.04793336, Log Avg loss: 1.21844652, Global Avg Loss: 3.95626016, Time: 0.0119 Steps: 15800, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000907, Sample Num: 14512, Cur Loss: 0.59121394, Cur Avg Loss: 1.04981461, Log Avg loss: 1.21856211, Global Avg Loss: 3.95452854, Time: 0.0118 Steps: 15810, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000917, Sample Num: 14672, Cur Loss: 0.80223382, Cur Avg Loss: 1.05022034, Log Avg loss: 1.08702093, Global Avg Loss: 3.95271596, Time: 0.0073 Steps: 15820, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000927, Sample Num: 14832, Cur Loss: 1.39467096, Cur Avg Loss: 1.04824818, Log Avg loss: 0.86740111, Global Avg Loss: 3.95076693, Time: 0.0115 Steps: 15830, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000937, Sample Num: 14992, Cur Loss: 0.55505276, Cur Avg Loss: 1.05058812, Log Avg loss: 1.26749980, Global Avg Loss: 3.94907294, Time: 0.0101 Steps: 15840, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000947, Sample Num: 15152, Cur Loss: 1.07962513, Cur Avg Loss: 1.05073151, Log Avg loss: 1.06416773, Global Avg Loss: 3.94725282, Time: 0.0116 Steps: 15850, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000957, Sample Num: 15312, Cur Loss: 1.17068338, Cur Avg Loss: 1.05085601, Log Avg loss: 1.06264600, Global Avg Loss: 3.94543402, Time: 0.0085 Steps: 15860, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000967, Sample Num: 15472, Cur Loss: 0.99317819, Cur Avg Loss: 1.05083526, Log Avg loss: 1.04884925, Global Avg Loss: 3.94360883, Time: 0.0131 Steps: 15870, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000977, Sample Num: 15632, Cur Loss: 1.17265689, Cur Avg Loss: 1.04952944, Log Avg loss: 0.92325678, Global Avg Loss: 3.94170684, Time: 0.0069 Steps: 15880, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000987, Sample Num: 15792, Cur Loss: 1.53801370, Cur Avg Loss: 1.04954146, Log Avg loss: 1.05071536, Global Avg Loss: 3.93988746, Time: 0.0095 Steps: 15890, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 000997, Sample Num: 15952, Cur Loss: 2.01654387, Cur Avg Loss: 1.04913357, Log Avg loss: 1.00887510, Global Avg Loss: 3.93804406, Time: 0.0083 Steps: 15900, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001007, Sample Num: 16112, Cur Loss: 0.73948544, Cur Avg Loss: 1.05046675, Log Avg loss: 1.18338523, Global Avg Loss: 3.93631266, Time: 0.0068 Steps: 15910, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001017, Sample Num: 16272, Cur Loss: 0.81318069, Cur Avg Loss: 1.04986304, Log Avg loss: 0.98906884, Global Avg Loss: 3.93446137, Time: 0.0082 Steps: 15920, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001027, Sample Num: 16432, Cur Loss: 0.99909514, Cur Avg Loss: 1.05304985, Log Avg loss: 1.37714828, Global Avg Loss: 3.93285603, Time: 0.0136 Steps: 15930, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001037, Sample Num: 16592, Cur Loss: 1.14318657, Cur Avg Loss: 1.05307194, Log Avg loss: 1.05534118, Global Avg Loss: 3.93105081, Time: 0.0067 Steps: 15940, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001047, Sample Num: 16752, Cur Loss: 0.54716766, Cur Avg Loss: 1.05261779, Log Avg loss: 1.00552209, Global Avg Loss: 3.92921663, Time: 0.0108 Steps: 15950, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001057, Sample Num: 16912, Cur Loss: 0.44504490, Cur Avg Loss: 1.05269301, Log Avg loss: 1.06056909, Global Avg Loss: 3.92741923, Time: 0.0092 Steps: 15960, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001067, Sample Num: 17072, Cur Loss: 0.97885203, Cur Avg Loss: 1.05344142, Log Avg loss: 1.13254856, Global Avg Loss: 3.92566915, Time: 0.0129 Steps: 15970, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001077, Sample Num: 17232, Cur Loss: 0.82038736, Cur Avg Loss: 1.05252363, Log Avg loss: 0.95459517, Global Avg Loss: 3.92380991, Time: 0.0120 Steps: 15980, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001087, Sample Num: 17392, Cur Loss: 0.85861468, Cur Avg Loss: 1.05189714, Log Avg loss: 0.98442423, Global Avg Loss: 3.92197164, Time: 0.0134 Steps: 15990, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001097, Sample Num: 17552, Cur Loss: 0.91640401, Cur Avg Loss: 1.05344307, Log Avg loss: 1.22148519, Global Avg Loss: 3.92028384, Time: 0.0084 Steps: 16000, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001107, Sample Num: 17712, Cur Loss: 1.34182048, Cur Avg Loss: 1.05445260, Log Avg loss: 1.16519805, Global Avg Loss: 3.91856299, Time: 0.0113 Steps: 16010, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001117, Sample Num: 17872, Cur Loss: 0.53539705, Cur Avg Loss: 1.05405274, Log Avg loss: 1.00978799, Global Avg Loss: 3.91674727, Time: 0.0156 Steps: 16020, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001127, Sample Num: 18032, Cur Loss: 1.32390881, Cur Avg Loss: 1.05242640, Log Avg loss: 0.87076444, Global Avg Loss: 3.91484709, Time: 0.0130 Steps: 16030, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001137, Sample Num: 18192, Cur Loss: 1.24329412, Cur Avg Loss: 1.05079689, Log Avg loss: 0.86715148, Global Avg Loss: 3.91294703, Time: 0.0070 Steps: 16040, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001147, Sample Num: 18352, Cur Loss: 1.40100813, Cur Avg Loss: 1.05098336, Log Avg loss: 1.07218523, Global Avg Loss: 3.91117709, Time: 0.0073 Steps: 16050, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001157, Sample Num: 18512, Cur Loss: 0.94661725, Cur Avg Loss: 1.05230367, Log Avg loss: 1.20374247, Global Avg Loss: 3.90949127, Time: 0.0064 Steps: 16060, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001167, Sample Num: 18672, Cur Loss: 0.89862156, Cur Avg Loss: 1.05239589, Log Avg loss: 1.06306551, Global Avg Loss: 3.90772000, Time: 0.0111 Steps: 16070, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001177, Sample Num: 18832, Cur Loss: 0.82046568, Cur Avg Loss: 1.05048161, Log Avg loss: 0.82708570, Global Avg Loss: 3.90580418, Time: 0.0112 Steps: 16080, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001187, Sample Num: 18992, Cur Loss: 1.98372221, Cur Avg Loss: 1.05344592, Log Avg loss: 1.40234512, Global Avg Loss: 3.90424827, Time: 0.0131 Steps: 16090, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001197, Sample Num: 19152, Cur Loss: 0.58837652, Cur Avg Loss: 1.05269147, Log Avg loss: 0.96313801, Global Avg Loss: 3.90242149, Time: 0.0064 Steps: 16100, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001207, Sample Num: 19312, Cur Loss: 1.41302872, Cur Avg Loss: 1.05335026, Log Avg loss: 1.13220726, Global Avg Loss: 3.90070193, Time: 0.0133 Steps: 16110, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001217, Sample Num: 19472, Cur Loss: 1.17030418, Cur Avg Loss: 1.05324409, Log Avg loss: 1.04042949, Global Avg Loss: 3.89892757, Time: 0.0107 Steps: 16120, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001227, Sample Num: 19632, Cur Loss: 1.29453528, Cur Avg Loss: 1.05524791, Log Avg loss: 1.29911264, Global Avg Loss: 3.89731578, Time: 0.0112 Steps: 16130, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001237, Sample Num: 19792, Cur Loss: 0.59377205, Cur Avg Loss: 1.05409574, Log Avg loss: 0.91272516, Global Avg Loss: 3.89546659, Time: 0.0111 Steps: 16140, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001247, Sample Num: 19952, Cur Loss: 0.94093204, Cur Avg Loss: 1.05369402, Log Avg loss: 1.00400099, Global Avg Loss: 3.89367621, Time: 0.0085 Steps: 16150, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001257, Sample Num: 20112, Cur Loss: 1.86716151, Cur Avg Loss: 1.05486614, Log Avg loss: 1.20102903, Global Avg Loss: 3.89200997, Time: 0.0122 Steps: 16160, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001267, Sample Num: 20272, Cur Loss: 1.08205783, Cur Avg Loss: 1.05431381, Log Avg loss: 0.98488713, Global Avg Loss: 3.89021212, Time: 0.0065 Steps: 16170, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001277, Sample Num: 20432, Cur Loss: 0.67683268, Cur Avg Loss: 1.05227678, Log Avg loss: 0.79418434, Global Avg Loss: 3.88829863, Time: 0.0111 Steps: 16180, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001287, Sample Num: 20592, Cur Loss: 0.80764467, Cur Avg Loss: 1.05243509, Log Avg loss: 1.07265132, Global Avg Loss: 3.88655950, Time: 0.0119 Steps: 16190, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001297, Sample Num: 20752, Cur Loss: 0.78911650, Cur Avg Loss: 1.05364534, Log Avg loss: 1.20940467, Global Avg Loss: 3.88490694, Time: 0.0228 Steps: 16200, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001307, Sample Num: 20912, Cur Loss: 1.40073442, Cur Avg Loss: 1.05414282, Log Avg loss: 1.11866594, Global Avg Loss: 3.88320043, Time: 0.0219 Steps: 16210, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001317, Sample Num: 21072, Cur Loss: 0.86597902, Cur Avg Loss: 1.05262070, Log Avg loss: 0.85368015, Global Avg Loss: 3.88133267, Time: 0.0097 Steps: 16220, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001327, Sample Num: 21232, Cur Loss: 0.33474523, Cur Avg Loss: 1.05158962, Log Avg loss: 0.91579609, Global Avg Loss: 3.87950547, Time: 0.0152 Steps: 16230, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001337, Sample Num: 21392, Cur Loss: 2.08832836, Cur Avg Loss: 1.05114126, Log Avg loss: 0.99164381, Global Avg Loss: 3.87772723, Time: 0.0067 Steps: 16240, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001347, Sample Num: 21552, Cur Loss: 0.91819346, Cur Avg Loss: 1.05273227, Log Avg loss: 1.26545045, Global Avg Loss: 3.87611968, Time: 0.0067 Steps: 16250, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001357, Sample Num: 21712, Cur Loss: 0.65039575, Cur Avg Loss: 1.05233272, Log Avg loss: 0.99851287, Global Avg Loss: 3.87434993, Time: 0.0120 Steps: 16260, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001367, Sample Num: 21872, Cur Loss: 0.92938864, Cur Avg Loss: 1.05261988, Log Avg loss: 1.09158811, Global Avg Loss: 3.87263957, Time: 0.0105 Steps: 16270, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001377, Sample Num: 22032, Cur Loss: 0.82469290, Cur Avg Loss: 1.05180426, Log Avg loss: 0.94030916, Global Avg Loss: 3.87083838, Time: 0.0067 Steps: 16280, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001387, Sample Num: 22192, Cur Loss: 0.60879791, Cur Avg Loss: 1.05203798, Log Avg loss: 1.08422102, Global Avg Loss: 3.86912775, Time: 0.0067 Steps: 16290, Updated lr: 0.000086 Training, Epoch: 0008, Batch: 001397, Sample Num: 22352, Cur Loss: 0.95435846, Cur Avg Loss: 1.05115698, Log Avg loss: 0.92896154, Global Avg Loss: 3.86732397, Time: 0.0091 Steps: 16300, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001407, Sample Num: 22512, Cur Loss: 0.86743259, Cur Avg Loss: 1.05150728, Log Avg loss: 1.10044412, Global Avg Loss: 3.86562754, Time: 0.0086 Steps: 16310, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001417, Sample Num: 22672, Cur Loss: 0.70097756, Cur Avg Loss: 1.05271911, Log Avg loss: 1.22322457, Global Avg Loss: 3.86400842, Time: 0.0092 Steps: 16320, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001427, Sample Num: 22832, Cur Loss: 1.21250057, Cur Avg Loss: 1.05093704, Log Avg loss: 0.79841735, Global Avg Loss: 3.86213114, Time: 0.0112 Steps: 16330, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001437, Sample Num: 22992, Cur Loss: 0.58940852, Cur Avg Loss: 1.05200007, Log Avg loss: 1.20369395, Global Avg Loss: 3.86050419, Time: 0.0110 Steps: 16340, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001447, Sample Num: 23152, Cur Loss: 1.09247494, Cur Avg Loss: 1.05140608, Log Avg loss: 0.96604961, Global Avg Loss: 3.85873388, Time: 0.0134 Steps: 16350, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001457, Sample Num: 23312, Cur Loss: 1.17267442, Cur Avg Loss: 1.04976644, Log Avg loss: 0.81251133, Global Avg Loss: 3.85687189, Time: 0.0073 Steps: 16360, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001467, Sample Num: 23472, Cur Loss: 1.67537427, Cur Avg Loss: 1.05183657, Log Avg loss: 1.35345359, Global Avg Loss: 3.85534262, Time: 0.0216 Steps: 16370, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001477, Sample Num: 23632, Cur Loss: 1.67719114, Cur Avg Loss: 1.05242478, Log Avg loss: 1.13871581, Global Avg Loss: 3.85368411, Time: 0.0112 Steps: 16380, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001487, Sample Num: 23792, Cur Loss: 0.89299679, Cur Avg Loss: 1.05043401, Log Avg loss: 0.75639772, Global Avg Loss: 3.85179437, Time: 0.0074 Steps: 16390, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001497, Sample Num: 23952, Cur Loss: 0.75801188, Cur Avg Loss: 1.05188441, Log Avg loss: 1.26755849, Global Avg Loss: 3.85021862, Time: 0.0074 Steps: 16400, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001507, Sample Num: 24112, Cur Loss: 0.65558147, Cur Avg Loss: 1.04996239, Log Avg loss: 0.76223621, Global Avg Loss: 3.84833685, Time: 0.0115 Steps: 16410, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001517, Sample Num: 24272, Cur Loss: 0.41611469, Cur Avg Loss: 1.04945773, Log Avg loss: 0.97340519, Global Avg Loss: 3.84658598, Time: 0.0071 Steps: 16420, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001527, Sample Num: 24432, Cur Loss: 2.11152697, Cur Avg Loss: 1.04946895, Log Avg loss: 1.05117071, Global Avg Loss: 3.84488457, Time: 0.0133 Steps: 16430, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001537, Sample Num: 24592, Cur Loss: 0.57795846, Cur Avg Loss: 1.04944826, Log Avg loss: 1.04628873, Global Avg Loss: 3.84318226, Time: 0.0234 Steps: 16440, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001547, Sample Num: 24752, Cur Loss: 0.93326443, Cur Avg Loss: 1.04989838, Log Avg loss: 1.11908151, Global Avg Loss: 3.84152627, Time: 0.0071 Steps: 16450, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001557, Sample Num: 24912, Cur Loss: 0.75930804, Cur Avg Loss: 1.04962319, Log Avg loss: 1.00705245, Global Avg Loss: 3.83980423, Time: 0.0067 Steps: 16460, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001567, Sample Num: 25072, Cur Loss: 1.64841390, Cur Avg Loss: 1.04893025, Log Avg loss: 0.94103885, Global Avg Loss: 3.83804421, Time: 0.0118 Steps: 16470, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001577, Sample Num: 25232, Cur Loss: 1.41177964, Cur Avg Loss: 1.05060983, Log Avg loss: 1.31380095, Global Avg Loss: 3.83651251, Time: 0.0233 Steps: 16480, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001587, Sample Num: 25392, Cur Loss: 0.63801312, Cur Avg Loss: 1.04967057, Log Avg loss: 0.90154917, Global Avg Loss: 3.83473266, Time: 0.0116 Steps: 16490, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001597, Sample Num: 25552, Cur Loss: 0.81286395, Cur Avg Loss: 1.04822963, Log Avg loss: 0.81955125, Global Avg Loss: 3.83290528, Time: 0.0102 Steps: 16500, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001607, Sample Num: 25712, Cur Loss: 0.46171916, Cur Avg Loss: 1.04759473, Log Avg loss: 0.94620229, Global Avg Loss: 3.83115682, Time: 0.0074 Steps: 16510, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001617, Sample Num: 25872, Cur Loss: 1.24669635, Cur Avg Loss: 1.04573217, Log Avg loss: 0.74641759, Global Avg Loss: 3.82928955, Time: 0.0068 Steps: 16520, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001627, Sample Num: 26032, Cur Loss: 0.69369394, Cur Avg Loss: 1.04542577, Log Avg loss: 0.99588118, Global Avg Loss: 3.82757544, Time: 0.0138 Steps: 16530, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001637, Sample Num: 26192, Cur Loss: 1.09066319, Cur Avg Loss: 1.04609847, Log Avg loss: 1.15554657, Global Avg Loss: 3.82595995, Time: 0.0066 Steps: 16540, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001647, Sample Num: 26352, Cur Loss: 0.65231651, Cur Avg Loss: 1.04624201, Log Avg loss: 1.06974005, Global Avg Loss: 3.82429456, Time: 0.0066 Steps: 16550, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001657, Sample Num: 26512, Cur Loss: 0.81977999, Cur Avg Loss: 1.04638805, Log Avg loss: 1.07044125, Global Avg Loss: 3.82263160, Time: 0.0110 Steps: 16560, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001667, Sample Num: 26672, Cur Loss: 0.87412024, Cur Avg Loss: 1.04594518, Log Avg loss: 0.97256191, Global Avg Loss: 3.82091159, Time: 0.0145 Steps: 16570, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001677, Sample Num: 26832, Cur Loss: 1.51824820, Cur Avg Loss: 1.04651696, Log Avg loss: 1.14183262, Global Avg Loss: 3.81929574, Time: 0.0112 Steps: 16580, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001687, Sample Num: 26992, Cur Loss: 1.84426999, Cur Avg Loss: 1.04699119, Log Avg loss: 1.12651895, Global Avg Loss: 3.81767260, Time: 0.0116 Steps: 16590, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001697, Sample Num: 27152, Cur Loss: 0.28690606, Cur Avg Loss: 1.04521701, Log Avg loss: 0.74591297, Global Avg Loss: 3.81582215, Time: 0.0064 Steps: 16600, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001707, Sample Num: 27312, Cur Loss: 0.47766298, Cur Avg Loss: 1.04609036, Log Avg loss: 1.19429738, Global Avg Loss: 3.81424387, Time: 0.0089 Steps: 16610, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001717, Sample Num: 27472, Cur Loss: 0.83308423, Cur Avg Loss: 1.04461699, Log Avg loss: 0.79311258, Global Avg Loss: 3.81242610, Time: 0.0111 Steps: 16620, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001727, Sample Num: 27632, Cur Loss: 0.29639608, Cur Avg Loss: 1.04441665, Log Avg loss: 1.01001800, Global Avg Loss: 3.81074095, Time: 0.0117 Steps: 16630, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001737, Sample Num: 27792, Cur Loss: 1.09632170, Cur Avg Loss: 1.04406164, Log Avg loss: 0.98275194, Global Avg Loss: 3.80904143, Time: 0.0099 Steps: 16640, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001747, Sample Num: 27952, Cur Loss: 0.72541475, Cur Avg Loss: 1.04312447, Log Avg loss: 0.88033910, Global Avg Loss: 3.80728245, Time: 0.0120 Steps: 16650, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001757, Sample Num: 28112, Cur Loss: 1.32014585, Cur Avg Loss: 1.04259757, Log Avg loss: 0.95054654, Global Avg Loss: 3.80556772, Time: 0.0067 Steps: 16660, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001767, Sample Num: 28272, Cur Loss: 0.67169559, Cur Avg Loss: 1.04328112, Log Avg loss: 1.16338158, Global Avg Loss: 3.80398273, Time: 0.0068 Steps: 16670, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001777, Sample Num: 28432, Cur Loss: 0.66734838, Cur Avg Loss: 1.04424151, Log Avg loss: 1.21394237, Global Avg Loss: 3.80242995, Time: 0.0067 Steps: 16680, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001787, Sample Num: 28592, Cur Loss: 1.39255047, Cur Avg Loss: 1.04499630, Log Avg loss: 1.17912216, Global Avg Loss: 3.80085816, Time: 0.0125 Steps: 16690, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001797, Sample Num: 28752, Cur Loss: 0.94634122, Cur Avg Loss: 1.04526109, Log Avg loss: 1.09257948, Global Avg Loss: 3.79923644, Time: 0.0065 Steps: 16700, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001807, Sample Num: 28912, Cur Loss: 0.97892839, Cur Avg Loss: 1.04588850, Log Avg loss: 1.15863342, Global Avg Loss: 3.79765619, Time: 0.0064 Steps: 16710, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001817, Sample Num: 29072, Cur Loss: 0.75884557, Cur Avg Loss: 1.04559097, Log Avg loss: 0.99182802, Global Avg Loss: 3.79597806, Time: 0.0081 Steps: 16720, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001827, Sample Num: 29232, Cur Loss: 1.21499896, Cur Avg Loss: 1.04625835, Log Avg loss: 1.16752197, Global Avg Loss: 3.79440696, Time: 0.0072 Steps: 16730, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001837, Sample Num: 29392, Cur Loss: 1.85354912, Cur Avg Loss: 1.04666636, Log Avg loss: 1.12120967, Global Avg Loss: 3.79281006, Time: 0.0077 Steps: 16740, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001847, Sample Num: 29552, Cur Loss: 1.06200719, Cur Avg Loss: 1.04658476, Log Avg loss: 1.03159416, Global Avg Loss: 3.79116158, Time: 0.0227 Steps: 16750, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001857, Sample Num: 29712, Cur Loss: 0.92416382, Cur Avg Loss: 1.04594366, Log Avg loss: 0.92753257, Global Avg Loss: 3.78945297, Time: 0.0072 Steps: 16760, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001867, Sample Num: 29872, Cur Loss: 0.74511850, Cur Avg Loss: 1.04507908, Log Avg loss: 0.88452608, Global Avg Loss: 3.78772075, Time: 0.0067 Steps: 16770, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001877, Sample Num: 30032, Cur Loss: 0.46627676, Cur Avg Loss: 1.04435616, Log Avg loss: 0.90938834, Global Avg Loss: 3.78600542, Time: 0.0065 Steps: 16780, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001887, Sample Num: 30192, Cur Loss: 0.74287879, Cur Avg Loss: 1.04432741, Log Avg loss: 1.03893068, Global Avg Loss: 3.78436928, Time: 0.0095 Steps: 16790, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001897, Sample Num: 30352, Cur Loss: 0.77021253, Cur Avg Loss: 1.04295229, Log Avg loss: 0.78346644, Global Avg Loss: 3.78258303, Time: 0.0132 Steps: 16800, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001907, Sample Num: 30512, Cur Loss: 1.53843403, Cur Avg Loss: 1.04275458, Log Avg loss: 1.00524835, Global Avg Loss: 3.78093084, Time: 0.0090 Steps: 16810, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001917, Sample Num: 30672, Cur Loss: 0.48977259, Cur Avg Loss: 1.04253241, Log Avg loss: 1.00016483, Global Avg Loss: 3.77927759, Time: 0.0121 Steps: 16820, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001927, Sample Num: 30832, Cur Loss: 1.28022087, Cur Avg Loss: 1.04315903, Log Avg loss: 1.16328275, Global Avg Loss: 3.77772322, Time: 0.0067 Steps: 16830, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001937, Sample Num: 30992, Cur Loss: 0.81635100, Cur Avg Loss: 1.04277640, Log Avg loss: 0.96904375, Global Avg Loss: 3.77605536, Time: 0.0121 Steps: 16840, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001947, Sample Num: 31152, Cur Loss: 2.64966941, Cur Avg Loss: 1.04432410, Log Avg loss: 1.34411361, Global Avg Loss: 3.77461207, Time: 0.0073 Steps: 16850, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001957, Sample Num: 31312, Cur Loss: 0.53110427, Cur Avg Loss: 1.04342905, Log Avg loss: 0.86916313, Global Avg Loss: 3.77288879, Time: 0.0110 Steps: 16860, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001967, Sample Num: 31472, Cur Loss: 1.42982101, Cur Avg Loss: 1.04368529, Log Avg loss: 1.09383138, Global Avg Loss: 3.77130073, Time: 0.0077 Steps: 16870, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001977, Sample Num: 31632, Cur Loss: 0.65906411, Cur Avg Loss: 1.04323130, Log Avg loss: 0.95393123, Global Avg Loss: 3.76963167, Time: 0.0114 Steps: 16880, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001987, Sample Num: 31792, Cur Loss: 1.25009811, Cur Avg Loss: 1.04465398, Log Avg loss: 1.32591708, Global Avg Loss: 3.76818483, Time: 0.0122 Steps: 16890, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 001997, Sample Num: 31952, Cur Loss: 2.22174621, Cur Avg Loss: 1.04562703, Log Avg loss: 1.23897339, Global Avg Loss: 3.76668826, Time: 0.0067 Steps: 16900, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002007, Sample Num: 32112, Cur Loss: 0.91085762, Cur Avg Loss: 1.04571820, Log Avg loss: 1.06392398, Global Avg Loss: 3.76508994, Time: 0.0072 Steps: 16910, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002017, Sample Num: 32272, Cur Loss: 0.92808414, Cur Avg Loss: 1.04517139, Log Avg loss: 0.93542756, Global Avg Loss: 3.76341756, Time: 0.0132 Steps: 16920, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002027, Sample Num: 32432, Cur Loss: 1.19201970, Cur Avg Loss: 1.04534375, Log Avg loss: 1.08010796, Global Avg Loss: 3.76183261, Time: 0.0067 Steps: 16930, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002037, Sample Num: 32592, Cur Loss: 0.69173616, Cur Avg Loss: 1.04345140, Log Avg loss: 0.65987277, Global Avg Loss: 3.76000147, Time: 0.0127 Steps: 16940, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002047, Sample Num: 32752, Cur Loss: 1.79326010, Cur Avg Loss: 1.04448537, Log Avg loss: 1.25510519, Global Avg Loss: 3.75852365, Time: 0.0135 Steps: 16950, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002057, Sample Num: 32912, Cur Loss: 0.47232655, Cur Avg Loss: 1.04364780, Log Avg loss: 0.87219701, Global Avg Loss: 3.75682181, Time: 0.0106 Steps: 16960, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002067, Sample Num: 33072, Cur Loss: 1.08861578, Cur Avg Loss: 1.04315892, Log Avg loss: 0.94259484, Global Avg Loss: 3.75516346, Time: 0.0227 Steps: 16970, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002077, Sample Num: 33232, Cur Loss: 0.97106743, Cur Avg Loss: 1.04297096, Log Avg loss: 1.00412004, Global Avg Loss: 3.75354329, Time: 0.0069 Steps: 16980, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002087, Sample Num: 33392, Cur Loss: 0.95438784, Cur Avg Loss: 1.04256066, Log Avg loss: 0.95734163, Global Avg Loss: 3.75189750, Time: 0.0083 Steps: 16990, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002097, Sample Num: 33552, Cur Loss: 1.12410438, Cur Avg Loss: 1.04220790, Log Avg loss: 0.96858732, Global Avg Loss: 3.75026026, Time: 0.0073 Steps: 17000, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002107, Sample Num: 33712, Cur Loss: 1.22941244, Cur Avg Loss: 1.04175570, Log Avg loss: 0.94692946, Global Avg Loss: 3.74861221, Time: 0.0159 Steps: 17010, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002117, Sample Num: 33872, Cur Loss: 0.97430670, Cur Avg Loss: 1.04068745, Log Avg loss: 0.81560733, Global Avg Loss: 3.74688894, Time: 0.0080 Steps: 17020, Updated lr: 0.000085 Training, Epoch: 0008, Batch: 002127, Sample Num: 34032, Cur Loss: 0.46799439, Cur Avg Loss: 1.04047965, Log Avg loss: 0.99648882, Global Avg Loss: 3.74527390, Time: 0.0065 Steps: 17030, Updated lr: 0.000085 ***** Running evaluation checkpoint-17032 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-17032 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.111011, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.205552, "eval_total_loss": 847.50312, "eval_mae": 0.948924, "eval_mse": 1.205554, "eval_r2": 0.233671, "eval_sp_statistic": 0.552405, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.588933, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 4.023517, "test_total_loss": 2019.805743, "test_mae": 1.851434, "test_mse": 4.022808, "test_r2": -1.596356, "test_sp_statistic": 0.254653, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.385706, "test_ps_pvalue": 0.0, "lr": 8.479658605974395e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.744913079199316, "train_cur_epoch_loss": 2214.445189446211, "train_cur_epoch_avg_loss": 1.0401339546482906, "train_cur_epoch_time": 22.111010789871216, "train_cur_epoch_avg_time": 0.01038563212300198, "epoch": 8, "step": 17032} ################################################## Training, Epoch: 0009, Batch: 000008, Sample Num: 128, Cur Loss: 0.83145851, Cur Avg Loss: 1.08846497, Log Avg loss: 1.00526830, Global Avg Loss: 3.74366592, Time: 0.0164 Steps: 17040, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000018, Sample Num: 288, Cur Loss: 0.62668777, Cur Avg Loss: 1.05939311, Log Avg loss: 1.03613562, Global Avg Loss: 3.74207793, Time: 0.0191 Steps: 17050, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000028, Sample Num: 448, Cur Loss: 0.64282441, Cur Avg Loss: 1.06418681, Log Avg loss: 1.07281547, Global Avg Loss: 3.74051329, Time: 0.0067 Steps: 17060, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000038, Sample Num: 608, Cur Loss: 0.97240484, Cur Avg Loss: 1.07132829, Log Avg loss: 1.09132443, Global Avg Loss: 3.73896134, Time: 0.0123 Steps: 17070, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000048, Sample Num: 768, Cur Loss: 2.03891873, Cur Avg Loss: 1.07847918, Log Avg loss: 1.10565259, Global Avg Loss: 3.73741959, Time: 0.0133 Steps: 17080, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000058, Sample Num: 928, Cur Loss: 0.72057575, Cur Avg Loss: 1.02675219, Log Avg loss: 0.77846259, Global Avg Loss: 3.73568819, Time: 0.0067 Steps: 17090, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000068, Sample Num: 1088, Cur Loss: 1.10368097, Cur Avg Loss: 1.01261396, Log Avg loss: 0.93061228, Global Avg Loss: 3.73404780, Time: 0.0159 Steps: 17100, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000078, Sample Num: 1248, Cur Loss: 0.61720675, Cur Avg Loss: 0.98669534, Log Avg loss: 0.81044871, Global Avg Loss: 3.73233909, Time: 0.0110 Steps: 17110, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000088, Sample Num: 1408, Cur Loss: 1.44952297, Cur Avg Loss: 1.02226450, Log Avg loss: 1.29970398, Global Avg Loss: 3.73091816, Time: 0.0077 Steps: 17120, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000098, Sample Num: 1568, Cur Loss: 0.72846508, Cur Avg Loss: 1.00800329, Log Avg loss: 0.88250464, Global Avg Loss: 3.72925533, Time: 0.0118 Steps: 17130, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000108, Sample Num: 1728, Cur Loss: 0.88792205, Cur Avg Loss: 1.00969613, Log Avg loss: 1.02628596, Global Avg Loss: 3.72767834, Time: 0.0122 Steps: 17140, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000118, Sample Num: 1888, Cur Loss: 1.26299059, Cur Avg Loss: 1.01403375, Log Avg loss: 1.06088000, Global Avg Loss: 3.72612336, Time: 0.0112 Steps: 17150, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000128, Sample Num: 2048, Cur Loss: 1.51764417, Cur Avg Loss: 1.01443562, Log Avg loss: 1.01917765, Global Avg Loss: 3.72454588, Time: 0.0068 Steps: 17160, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000138, Sample Num: 2208, Cur Loss: 1.18247640, Cur Avg Loss: 1.00106485, Log Avg loss: 0.82991908, Global Avg Loss: 3.72286002, Time: 0.0154 Steps: 17170, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000148, Sample Num: 2368, Cur Loss: 0.66083848, Cur Avg Loss: 0.98953118, Log Avg loss: 0.83036646, Global Avg Loss: 3.72117638, Time: 0.0129 Steps: 17180, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000158, Sample Num: 2528, Cur Loss: 0.89638275, Cur Avg Loss: 0.98494087, Log Avg loss: 0.91700432, Global Avg Loss: 3.71954510, Time: 0.0112 Steps: 17190, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000168, Sample Num: 2688, Cur Loss: 0.98897767, Cur Avg Loss: 0.99724857, Log Avg loss: 1.19171016, Global Avg Loss: 3.71807543, Time: 0.0226 Steps: 17200, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000178, Sample Num: 2848, Cur Loss: 1.14467812, Cur Avg Loss: 0.99772664, Log Avg loss: 1.00575821, Global Avg Loss: 3.71649941, Time: 0.0097 Steps: 17210, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000188, Sample Num: 3008, Cur Loss: 1.21746504, Cur Avg Loss: 0.99438032, Log Avg loss: 0.93481591, Global Avg Loss: 3.71488403, Time: 0.0069 Steps: 17220, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000198, Sample Num: 3168, Cur Loss: 2.08504796, Cur Avg Loss: 0.99300612, Log Avg loss: 0.96717110, Global Avg Loss: 3.71328931, Time: 0.0120 Steps: 17230, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000208, Sample Num: 3328, Cur Loss: 1.27977097, Cur Avg Loss: 1.00519870, Log Avg loss: 1.24661179, Global Avg Loss: 3.71185852, Time: 0.0138 Steps: 17240, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000218, Sample Num: 3488, Cur Loss: 0.96243608, Cur Avg Loss: 1.01184865, Log Avg loss: 1.15016763, Global Avg Loss: 3.71037348, Time: 0.0125 Steps: 17250, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000228, Sample Num: 3648, Cur Loss: 1.22228098, Cur Avg Loss: 1.00810098, Log Avg loss: 0.92640185, Global Avg Loss: 3.70876052, Time: 0.0146 Steps: 17260, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000238, Sample Num: 3808, Cur Loss: 1.00755715, Cur Avg Loss: 1.00675440, Log Avg loss: 0.97605233, Global Avg Loss: 3.70717818, Time: 0.0109 Steps: 17270, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000248, Sample Num: 3968, Cur Loss: 0.77138752, Cur Avg Loss: 1.00939061, Log Avg loss: 1.07213247, Global Avg Loss: 3.70565327, Time: 0.0125 Steps: 17280, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000258, Sample Num: 4128, Cur Loss: 0.86073923, Cur Avg Loss: 1.01488532, Log Avg loss: 1.15115417, Global Avg Loss: 3.70417582, Time: 0.0142 Steps: 17290, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000268, Sample Num: 4288, Cur Loss: 0.48734048, Cur Avg Loss: 1.01315069, Log Avg loss: 0.96839719, Global Avg Loss: 3.70259445, Time: 0.0114 Steps: 17300, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000278, Sample Num: 4448, Cur Loss: 0.94806588, Cur Avg Loss: 1.01314911, Log Avg loss: 1.01310675, Global Avg Loss: 3.70104073, Time: 0.0071 Steps: 17310, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000288, Sample Num: 4608, Cur Loss: 0.92819977, Cur Avg Loss: 1.01451800, Log Avg loss: 1.05257296, Global Avg Loss: 3.69951159, Time: 0.0075 Steps: 17320, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000298, Sample Num: 4768, Cur Loss: 1.22565889, Cur Avg Loss: 1.01270063, Log Avg loss: 0.96036037, Global Avg Loss: 3.69793101, Time: 0.0119 Steps: 17330, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000308, Sample Num: 4928, Cur Loss: 0.57345557, Cur Avg Loss: 1.00406117, Log Avg loss: 0.74660540, Global Avg Loss: 3.69622897, Time: 0.0065 Steps: 17340, Updated lr: 0.000085 Training, Epoch: 0009, Batch: 000318, Sample Num: 5088, Cur Loss: 1.09623170, Cur Avg Loss: 1.00805008, Log Avg loss: 1.13090851, Global Avg Loss: 3.69475040, Time: 0.0065 Steps: 17350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000328, Sample Num: 5248, Cur Loss: 1.47453153, Cur Avg Loss: 1.00852667, Log Avg loss: 1.02368234, Global Avg Loss: 3.69321177, Time: 0.0127 Steps: 17360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000338, Sample Num: 5408, Cur Loss: 1.03867745, Cur Avg Loss: 1.00912111, Log Avg loss: 1.02861851, Global Avg Loss: 3.69167775, Time: 0.0120 Steps: 17370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000348, Sample Num: 5568, Cur Loss: 1.07524121, Cur Avg Loss: 1.00352194, Log Avg loss: 0.81427014, Global Avg Loss: 3.69002216, Time: 0.0145 Steps: 17380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000358, Sample Num: 5728, Cur Loss: 0.82129544, Cur Avg Loss: 1.00275520, Log Avg loss: 0.97607274, Global Avg Loss: 3.68846153, Time: 0.0115 Steps: 17390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000368, Sample Num: 5888, Cur Loss: 0.82541192, Cur Avg Loss: 0.99580104, Log Avg loss: 0.74684206, Global Avg Loss: 3.68677094, Time: 0.0113 Steps: 17400, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000378, Sample Num: 6048, Cur Loss: 0.63008946, Cur Avg Loss: 0.99788026, Log Avg loss: 1.07439539, Global Avg Loss: 3.68527044, Time: 0.0141 Steps: 17410, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000388, Sample Num: 6208, Cur Loss: 0.81991929, Cur Avg Loss: 0.99986270, Log Avg loss: 1.07479890, Global Avg Loss: 3.68377189, Time: 0.0110 Steps: 17420, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000398, Sample Num: 6368, Cur Loss: 0.46987295, Cur Avg Loss: 1.00235090, Log Avg loss: 1.09889334, Global Avg Loss: 3.68228888, Time: 0.0115 Steps: 17430, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000408, Sample Num: 6528, Cur Loss: 0.68261409, Cur Avg Loss: 1.00256977, Log Avg loss: 1.01128053, Global Avg Loss: 3.68075734, Time: 0.0116 Steps: 17440, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000418, Sample Num: 6688, Cur Loss: 0.95460427, Cur Avg Loss: 1.00662490, Log Avg loss: 1.17207440, Global Avg Loss: 3.67931970, Time: 0.0114 Steps: 17450, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000428, Sample Num: 6848, Cur Loss: 0.57521868, Cur Avg Loss: 1.00894527, Log Avg loss: 1.10593676, Global Avg Loss: 3.67784583, Time: 0.0078 Steps: 17460, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000438, Sample Num: 7008, Cur Loss: 0.56910884, Cur Avg Loss: 1.00593055, Log Avg loss: 0.87690024, Global Avg Loss: 3.67624254, Time: 0.0067 Steps: 17470, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000448, Sample Num: 7168, Cur Loss: 0.61604571, Cur Avg Loss: 1.00343455, Log Avg loss: 0.89410988, Global Avg Loss: 3.67465093, Time: 0.0184 Steps: 17480, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000458, Sample Num: 7328, Cur Loss: 1.78175163, Cur Avg Loss: 1.00632875, Log Avg loss: 1.13598888, Global Avg Loss: 3.67319944, Time: 0.0065 Steps: 17490, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000468, Sample Num: 7488, Cur Loss: 1.01179123, Cur Avg Loss: 1.00355659, Log Avg loss: 0.87659159, Global Avg Loss: 3.67160137, Time: 0.0072 Steps: 17500, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000478, Sample Num: 7648, Cur Loss: 0.41751587, Cur Avg Loss: 1.00001206, Log Avg loss: 0.83412820, Global Avg Loss: 3.66998089, Time: 0.0115 Steps: 17510, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000488, Sample Num: 7808, Cur Loss: 0.58381927, Cur Avg Loss: 1.00137949, Log Avg loss: 1.06674242, Global Avg Loss: 3.66849502, Time: 0.0068 Steps: 17520, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000498, Sample Num: 7968, Cur Loss: 0.58300644, Cur Avg Loss: 0.99957381, Log Avg loss: 0.91145667, Global Avg Loss: 3.66692227, Time: 0.0067 Steps: 17530, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000508, Sample Num: 8128, Cur Loss: 1.27842259, Cur Avg Loss: 1.00457145, Log Avg loss: 1.25345418, Global Avg Loss: 3.66554629, Time: 0.0132 Steps: 17540, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000518, Sample Num: 8288, Cur Loss: 1.17414880, Cur Avg Loss: 1.00650171, Log Avg loss: 1.10455901, Global Avg Loss: 3.66408703, Time: 0.0073 Steps: 17550, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000528, Sample Num: 8448, Cur Loss: 0.34804392, Cur Avg Loss: 1.00092908, Log Avg loss: 0.71226677, Global Avg Loss: 3.66240604, Time: 0.0067 Steps: 17560, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000538, Sample Num: 8608, Cur Loss: 0.78280485, Cur Avg Loss: 0.99446052, Log Avg loss: 0.65292053, Global Avg Loss: 3.66069319, Time: 0.0131 Steps: 17570, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000548, Sample Num: 8768, Cur Loss: 1.20096314, Cur Avg Loss: 0.99249947, Log Avg loss: 0.88699510, Global Avg Loss: 3.65911543, Time: 0.0066 Steps: 17580, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000558, Sample Num: 8928, Cur Loss: 1.04496133, Cur Avg Loss: 0.99185628, Log Avg loss: 0.95660921, Global Avg Loss: 3.65757904, Time: 0.0090 Steps: 17590, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000568, Sample Num: 9088, Cur Loss: 1.08302116, Cur Avg Loss: 0.99051135, Log Avg loss: 0.91546442, Global Avg Loss: 3.65602102, Time: 0.0119 Steps: 17600, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000578, Sample Num: 9248, Cur Loss: 0.65172410, Cur Avg Loss: 0.98832324, Log Avg loss: 0.86403843, Global Avg Loss: 3.65443557, Time: 0.0202 Steps: 17610, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000588, Sample Num: 9408, Cur Loss: 0.96835905, Cur Avg Loss: 0.98884900, Log Avg loss: 1.01923795, Global Avg Loss: 3.65294000, Time: 0.0068 Steps: 17620, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000598, Sample Num: 9568, Cur Loss: 1.35813701, Cur Avg Loss: 0.98740076, Log Avg loss: 0.90224438, Global Avg Loss: 3.65137976, Time: 0.0110 Steps: 17630, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000608, Sample Num: 9728, Cur Loss: 0.73622912, Cur Avg Loss: 0.98626269, Log Avg loss: 0.91820597, Global Avg Loss: 3.64983034, Time: 0.0067 Steps: 17640, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000618, Sample Num: 9888, Cur Loss: 0.56593919, Cur Avg Loss: 0.98472231, Log Avg loss: 0.89106701, Global Avg Loss: 3.64826731, Time: 0.0117 Steps: 17650, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000628, Sample Num: 10048, Cur Loss: 1.17176759, Cur Avg Loss: 0.98622234, Log Avg loss: 1.07892450, Global Avg Loss: 3.64681241, Time: 0.0068 Steps: 17660, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000638, Sample Num: 10208, Cur Loss: 0.73310578, Cur Avg Loss: 0.98708746, Log Avg loss: 1.04141684, Global Avg Loss: 3.64533794, Time: 0.0122 Steps: 17670, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000648, Sample Num: 10368, Cur Loss: 1.13872027, Cur Avg Loss: 0.99032175, Log Avg loss: 1.19666980, Global Avg Loss: 3.64395294, Time: 0.0065 Steps: 17680, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000658, Sample Num: 10528, Cur Loss: 0.75302219, Cur Avg Loss: 0.98920636, Log Avg loss: 0.91692858, Global Avg Loss: 3.64241138, Time: 0.0069 Steps: 17690, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000668, Sample Num: 10688, Cur Loss: 0.75110716, Cur Avg Loss: 0.99040496, Log Avg loss: 1.06927324, Global Avg Loss: 3.64095763, Time: 0.0104 Steps: 17700, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000678, Sample Num: 10848, Cur Loss: 0.58309340, Cur Avg Loss: 0.99041472, Log Avg loss: 0.99106626, Global Avg Loss: 3.63946136, Time: 0.0124 Steps: 17710, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000688, Sample Num: 11008, Cur Loss: 1.20503795, Cur Avg Loss: 0.98924784, Log Avg loss: 0.91013390, Global Avg Loss: 3.63792111, Time: 0.0117 Steps: 17720, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000698, Sample Num: 11168, Cur Loss: 2.01416731, Cur Avg Loss: 0.99017318, Log Avg loss: 1.05383625, Global Avg Loss: 3.63646365, Time: 0.0103 Steps: 17730, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000708, Sample Num: 11328, Cur Loss: 0.63361239, Cur Avg Loss: 0.98751648, Log Avg loss: 0.80207915, Global Avg Loss: 3.63486591, Time: 0.0068 Steps: 17740, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000718, Sample Num: 11488, Cur Loss: 1.58818269, Cur Avg Loss: 0.98711575, Log Avg loss: 0.95874404, Global Avg Loss: 3.63335824, Time: 0.0136 Steps: 17750, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000728, Sample Num: 11648, Cur Loss: 1.56783450, Cur Avg Loss: 0.98737297, Log Avg loss: 1.00584151, Global Avg Loss: 3.63187878, Time: 0.0068 Steps: 17760, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000738, Sample Num: 11808, Cur Loss: 0.35122076, Cur Avg Loss: 0.98622220, Log Avg loss: 0.90244611, Global Avg Loss: 3.63034280, Time: 0.0068 Steps: 17770, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000748, Sample Num: 11968, Cur Loss: 0.58253878, Cur Avg Loss: 0.98751115, Log Avg loss: 1.08263506, Global Avg Loss: 3.62890989, Time: 0.0085 Steps: 17780, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000758, Sample Num: 12128, Cur Loss: 2.09756684, Cur Avg Loss: 0.99065147, Log Avg loss: 1.22554767, Global Avg Loss: 3.62755893, Time: 0.0115 Steps: 17790, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000768, Sample Num: 12288, Cur Loss: 0.42036930, Cur Avg Loss: 0.98988066, Log Avg loss: 0.93145353, Global Avg Loss: 3.62604426, Time: 0.0073 Steps: 17800, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000778, Sample Num: 12448, Cur Loss: 1.08630240, Cur Avg Loss: 0.98902107, Log Avg loss: 0.92300430, Global Avg Loss: 3.62452656, Time: 0.0157 Steps: 17810, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000788, Sample Num: 12608, Cur Loss: 1.10969329, Cur Avg Loss: 0.99158698, Log Avg loss: 1.19121467, Global Avg Loss: 3.62316106, Time: 0.0064 Steps: 17820, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000798, Sample Num: 12768, Cur Loss: 0.80191892, Cur Avg Loss: 0.99015233, Log Avg loss: 0.87710209, Global Avg Loss: 3.62162093, Time: 0.0065 Steps: 17830, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000808, Sample Num: 12928, Cur Loss: 0.47673881, Cur Avg Loss: 0.98903192, Log Avg loss: 0.89962297, Global Avg Loss: 3.62009514, Time: 0.0111 Steps: 17840, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000818, Sample Num: 13088, Cur Loss: 0.22521120, Cur Avg Loss: 0.98538954, Log Avg loss: 0.69108523, Global Avg Loss: 3.61845424, Time: 0.0067 Steps: 17850, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000828, Sample Num: 13248, Cur Loss: 1.60090494, Cur Avg Loss: 0.98334659, Log Avg loss: 0.81623327, Global Avg Loss: 3.61688525, Time: 0.0090 Steps: 17860, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000838, Sample Num: 13408, Cur Loss: 1.12588930, Cur Avg Loss: 0.99001507, Log Avg loss: 1.54216523, Global Avg Loss: 3.61572424, Time: 0.0066 Steps: 17870, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000848, Sample Num: 13568, Cur Loss: 0.92133743, Cur Avg Loss: 0.98905069, Log Avg loss: 0.90823577, Global Avg Loss: 3.61420999, Time: 0.0066 Steps: 17880, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000858, Sample Num: 13728, Cur Loss: 0.31719339, Cur Avg Loss: 0.98960969, Log Avg loss: 1.03701327, Global Avg Loss: 3.61276941, Time: 0.0068 Steps: 17890, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000868, Sample Num: 13888, Cur Loss: 1.44194913, Cur Avg Loss: 0.99187221, Log Avg loss: 1.18599604, Global Avg Loss: 3.61141367, Time: 0.0123 Steps: 17900, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000878, Sample Num: 14048, Cur Loss: 1.51771915, Cur Avg Loss: 0.99209721, Log Avg loss: 1.01162731, Global Avg Loss: 3.60996208, Time: 0.0116 Steps: 17910, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000888, Sample Num: 14208, Cur Loss: 1.44491208, Cur Avg Loss: 0.99240585, Log Avg loss: 1.01950479, Global Avg Loss: 3.60851652, Time: 0.0067 Steps: 17920, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000898, Sample Num: 14368, Cur Loss: 0.48370227, Cur Avg Loss: 0.99123825, Log Avg loss: 0.88755520, Global Avg Loss: 3.60699897, Time: 0.0137 Steps: 17930, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000908, Sample Num: 14528, Cur Loss: 1.22449517, Cur Avg Loss: 0.99271676, Log Avg loss: 1.12548658, Global Avg Loss: 3.60561574, Time: 0.0072 Steps: 17940, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000918, Sample Num: 14688, Cur Loss: 1.40491784, Cur Avg Loss: 0.99193113, Log Avg loss: 0.92059593, Global Avg Loss: 3.60411991, Time: 0.0065 Steps: 17950, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000928, Sample Num: 14848, Cur Loss: 0.90842175, Cur Avg Loss: 0.99317933, Log Avg loss: 1.10776462, Global Avg Loss: 3.60272995, Time: 0.0067 Steps: 17960, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000938, Sample Num: 15008, Cur Loss: 2.41565228, Cur Avg Loss: 0.99631754, Log Avg loss: 1.28754312, Global Avg Loss: 3.60144159, Time: 0.0120 Steps: 17970, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000948, Sample Num: 15168, Cur Loss: 0.83738410, Cur Avg Loss: 0.99573434, Log Avg loss: 0.94103032, Global Avg Loss: 3.59996194, Time: 0.0064 Steps: 17980, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000958, Sample Num: 15328, Cur Loss: 0.82186174, Cur Avg Loss: 0.99609278, Log Avg loss: 1.03007294, Global Avg Loss: 3.59853343, Time: 0.0224 Steps: 17990, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000968, Sample Num: 15488, Cur Loss: 1.12085903, Cur Avg Loss: 0.99836617, Log Avg loss: 1.21615621, Global Avg Loss: 3.59720989, Time: 0.0123 Steps: 18000, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000978, Sample Num: 15648, Cur Loss: 0.88550365, Cur Avg Loss: 0.99694320, Log Avg loss: 0.85920042, Global Avg Loss: 3.59568962, Time: 0.0112 Steps: 18010, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000988, Sample Num: 15808, Cur Loss: 1.32530129, Cur Avg Loss: 0.99805455, Log Avg loss: 1.10674441, Global Avg Loss: 3.59430841, Time: 0.0123 Steps: 18020, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 000998, Sample Num: 15968, Cur Loss: 1.46718347, Cur Avg Loss: 0.99962717, Log Avg loss: 1.15500144, Global Avg Loss: 3.59295549, Time: 0.0117 Steps: 18030, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001008, Sample Num: 16128, Cur Loss: 1.48555171, Cur Avg Loss: 1.00275064, Log Avg loss: 1.31447317, Global Avg Loss: 3.59169247, Time: 0.0116 Steps: 18040, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001018, Sample Num: 16288, Cur Loss: 0.35489392, Cur Avg Loss: 1.00115191, Log Avg loss: 0.84000037, Global Avg Loss: 3.59016799, Time: 0.0072 Steps: 18050, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001028, Sample Num: 16448, Cur Loss: 0.81991529, Cur Avg Loss: 1.00274112, Log Avg loss: 1.16452295, Global Avg Loss: 3.58882489, Time: 0.0073 Steps: 18060, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001038, Sample Num: 16608, Cur Loss: 1.53656769, Cur Avg Loss: 1.00290811, Log Avg loss: 1.02007421, Global Avg Loss: 3.58740333, Time: 0.0152 Steps: 18070, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001048, Sample Num: 16768, Cur Loss: 0.40154067, Cur Avg Loss: 1.00209973, Log Avg loss: 0.91819021, Global Avg Loss: 3.58592700, Time: 0.0125 Steps: 18080, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001058, Sample Num: 16928, Cur Loss: 2.06786156, Cur Avg Loss: 1.00372088, Log Avg loss: 1.17361759, Global Avg Loss: 3.58459349, Time: 0.0120 Steps: 18090, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001068, Sample Num: 17088, Cur Loss: 1.02892041, Cur Avg Loss: 1.00691231, Log Avg loss: 1.34456463, Global Avg Loss: 3.58335591, Time: 0.0151 Steps: 18100, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001078, Sample Num: 17248, Cur Loss: 0.53018165, Cur Avg Loss: 1.00606719, Log Avg loss: 0.91580918, Global Avg Loss: 3.58188294, Time: 0.0066 Steps: 18110, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001088, Sample Num: 17408, Cur Loss: 0.79872847, Cur Avg Loss: 1.00481810, Log Avg loss: 0.87016596, Global Avg Loss: 3.58038640, Time: 0.0122 Steps: 18120, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001098, Sample Num: 17568, Cur Loss: 0.71740592, Cur Avg Loss: 1.00420970, Log Avg loss: 0.93801540, Global Avg Loss: 3.57892895, Time: 0.0120 Steps: 18130, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001108, Sample Num: 17728, Cur Loss: 2.04038048, Cur Avg Loss: 1.00629220, Log Avg loss: 1.23495092, Global Avg Loss: 3.57763679, Time: 0.0236 Steps: 18140, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001118, Sample Num: 17888, Cur Loss: 0.68840069, Cur Avg Loss: 1.00451063, Log Avg loss: 0.80711244, Global Avg Loss: 3.57611033, Time: 0.0067 Steps: 18150, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001128, Sample Num: 18048, Cur Loss: 1.55488431, Cur Avg Loss: 1.00502533, Log Avg loss: 1.06256953, Global Avg Loss: 3.57472622, Time: 0.0071 Steps: 18160, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001138, Sample Num: 18208, Cur Loss: 0.62668848, Cur Avg Loss: 1.00384630, Log Avg loss: 0.87085153, Global Avg Loss: 3.57323812, Time: 0.0085 Steps: 18170, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001148, Sample Num: 18368, Cur Loss: 0.54057711, Cur Avg Loss: 1.00315148, Log Avg loss: 0.92408125, Global Avg Loss: 3.57178094, Time: 0.0115 Steps: 18180, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001158, Sample Num: 18528, Cur Loss: 0.90269989, Cur Avg Loss: 1.00197449, Log Avg loss: 0.86685530, Global Avg Loss: 3.57029390, Time: 0.0139 Steps: 18190, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001168, Sample Num: 18688, Cur Loss: 0.30652082, Cur Avg Loss: 1.00171929, Log Avg loss: 0.97216738, Global Avg Loss: 3.56886636, Time: 0.0115 Steps: 18200, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001178, Sample Num: 18848, Cur Loss: 1.31923175, Cur Avg Loss: 1.00128424, Log Avg loss: 0.95046988, Global Avg Loss: 3.56742847, Time: 0.0067 Steps: 18210, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001188, Sample Num: 19008, Cur Loss: 1.33859050, Cur Avg Loss: 1.00231084, Log Avg loss: 1.12324480, Global Avg Loss: 3.56608698, Time: 0.0158 Steps: 18220, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001198, Sample Num: 19168, Cur Loss: 0.12966529, Cur Avg Loss: 1.00235626, Log Avg loss: 1.00775277, Global Avg Loss: 3.56468362, Time: 0.0116 Steps: 18230, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001208, Sample Num: 19328, Cur Loss: 0.84512699, Cur Avg Loss: 1.00015246, Log Avg loss: 0.73613685, Global Avg Loss: 3.56313288, Time: 0.0071 Steps: 18240, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001218, Sample Num: 19488, Cur Loss: 0.79221117, Cur Avg Loss: 0.99952996, Log Avg loss: 0.92433140, Global Avg Loss: 3.56168696, Time: 0.0116 Steps: 18250, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001228, Sample Num: 19648, Cur Loss: 0.41404808, Cur Avg Loss: 0.99856006, Log Avg loss: 0.88042625, Global Avg Loss: 3.56021858, Time: 0.0231 Steps: 18260, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001238, Sample Num: 19808, Cur Loss: 1.38499141, Cur Avg Loss: 0.99855411, Log Avg loss: 0.99782422, Global Avg Loss: 3.55881607, Time: 0.0108 Steps: 18270, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001248, Sample Num: 19968, Cur Loss: 1.15767407, Cur Avg Loss: 1.00062956, Log Avg loss: 1.25757004, Global Avg Loss: 3.55755718, Time: 0.0133 Steps: 18280, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001258, Sample Num: 20128, Cur Loss: 1.33866429, Cur Avg Loss: 1.00150362, Log Avg loss: 1.11058624, Global Avg Loss: 3.55621931, Time: 0.0110 Steps: 18290, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001268, Sample Num: 20288, Cur Loss: 1.17258763, Cur Avg Loss: 1.00136647, Log Avg loss: 0.98411309, Global Avg Loss: 3.55481378, Time: 0.0121 Steps: 18300, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001278, Sample Num: 20448, Cur Loss: 0.72622311, Cur Avg Loss: 1.00048081, Log Avg loss: 0.88817865, Global Avg Loss: 3.55335740, Time: 0.0067 Steps: 18310, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001288, Sample Num: 20608, Cur Loss: 0.47585180, Cur Avg Loss: 1.00077190, Log Avg loss: 1.03797377, Global Avg Loss: 3.55198438, Time: 0.0120 Steps: 18320, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001298, Sample Num: 20768, Cur Loss: 1.71505189, Cur Avg Loss: 1.00068956, Log Avg loss: 0.99008368, Global Avg Loss: 3.55058672, Time: 0.0076 Steps: 18330, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001308, Sample Num: 20928, Cur Loss: 1.45171010, Cur Avg Loss: 1.00015741, Log Avg loss: 0.93108479, Global Avg Loss: 3.54915842, Time: 0.0115 Steps: 18340, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001318, Sample Num: 21088, Cur Loss: 0.71921355, Cur Avg Loss: 0.99953343, Log Avg loss: 0.91791678, Global Avg Loss: 3.54772450, Time: 0.0122 Steps: 18350, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001328, Sample Num: 21248, Cur Loss: 1.60019970, Cur Avg Loss: 1.00030496, Log Avg loss: 1.10199177, Global Avg Loss: 3.54639240, Time: 0.0109 Steps: 18360, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001338, Sample Num: 21408, Cur Loss: 0.92453271, Cur Avg Loss: 1.00118893, Log Avg loss: 1.11858038, Global Avg Loss: 3.54507079, Time: 0.0117 Steps: 18370, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001348, Sample Num: 21568, Cur Loss: 0.42737135, Cur Avg Loss: 1.00062616, Log Avg loss: 0.92532739, Global Avg Loss: 3.54364546, Time: 0.0140 Steps: 18380, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001358, Sample Num: 21728, Cur Loss: 0.79610366, Cur Avg Loss: 1.00005116, Log Avg loss: 0.92254119, Global Avg Loss: 3.54222018, Time: 0.0138 Steps: 18390, Updated lr: 0.000084 Training, Epoch: 0009, Batch: 001368, Sample Num: 21888, Cur Loss: 0.75864983, Cur Avg Loss: 0.99913468, Log Avg loss: 0.87467688, Global Avg Loss: 3.54077042, Time: 0.0115 Steps: 18400, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001378, Sample Num: 22048, Cur Loss: 1.03826618, Cur Avg Loss: 0.99946276, Log Avg loss: 1.04434395, Global Avg Loss: 3.53941441, Time: 0.0125 Steps: 18410, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001388, Sample Num: 22208, Cur Loss: 1.21288490, Cur Avg Loss: 1.00035286, Log Avg loss: 1.12300870, Global Avg Loss: 3.53810257, Time: 0.0076 Steps: 18420, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001398, Sample Num: 22368, Cur Loss: 0.85096765, Cur Avg Loss: 1.00036826, Log Avg loss: 1.00250684, Global Avg Loss: 3.53672677, Time: 0.0088 Steps: 18430, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001408, Sample Num: 22528, Cur Loss: 0.91575706, Cur Avg Loss: 1.00143089, Log Avg loss: 1.14998611, Global Avg Loss: 3.53543244, Time: 0.0117 Steps: 18440, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001418, Sample Num: 22688, Cur Loss: 0.93790579, Cur Avg Loss: 1.00114691, Log Avg loss: 0.96116209, Global Avg Loss: 3.53403718, Time: 0.0113 Steps: 18450, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001428, Sample Num: 22848, Cur Loss: 0.68455505, Cur Avg Loss: 1.00127449, Log Avg loss: 1.01936583, Global Avg Loss: 3.53267495, Time: 0.0148 Steps: 18460, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001438, Sample Num: 23008, Cur Loss: 1.41739321, Cur Avg Loss: 1.00151574, Log Avg loss: 1.03596569, Global Avg Loss: 3.53132318, Time: 0.0069 Steps: 18470, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001448, Sample Num: 23168, Cur Loss: 1.17846048, Cur Avg Loss: 1.00273917, Log Avg loss: 1.17866902, Global Avg Loss: 3.53005010, Time: 0.0113 Steps: 18480, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001458, Sample Num: 23328, Cur Loss: 0.61147243, Cur Avg Loss: 1.00120819, Log Avg loss: 0.77952143, Global Avg Loss: 3.52856253, Time: 0.0187 Steps: 18490, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001468, Sample Num: 23488, Cur Loss: 1.33057046, Cur Avg Loss: 1.00192680, Log Avg loss: 1.10670045, Global Avg Loss: 3.52725341, Time: 0.0084 Steps: 18500, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001478, Sample Num: 23648, Cur Loss: 0.93605202, Cur Avg Loss: 1.00038239, Log Avg loss: 0.77366319, Global Avg Loss: 3.52576579, Time: 0.0225 Steps: 18510, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001488, Sample Num: 23808, Cur Loss: 0.89000106, Cur Avg Loss: 1.00005469, Log Avg loss: 0.95162109, Global Avg Loss: 3.52437586, Time: 0.0073 Steps: 18520, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001498, Sample Num: 23968, Cur Loss: 0.67481697, Cur Avg Loss: 1.00082749, Log Avg loss: 1.11581932, Global Avg Loss: 3.52307605, Time: 0.0166 Steps: 18530, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001508, Sample Num: 24128, Cur Loss: 1.36499667, Cur Avg Loss: 0.99982195, Log Avg loss: 0.84919298, Global Avg Loss: 3.52163382, Time: 0.0094 Steps: 18540, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001518, Sample Num: 24288, Cur Loss: 0.81216276, Cur Avg Loss: 0.99941588, Log Avg loss: 0.93817969, Global Avg Loss: 3.52024112, Time: 0.0167 Steps: 18550, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001528, Sample Num: 24448, Cur Loss: 0.81435227, Cur Avg Loss: 0.99970095, Log Avg loss: 1.04297556, Global Avg Loss: 3.51890639, Time: 0.0183 Steps: 18560, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001538, Sample Num: 24608, Cur Loss: 1.31364441, Cur Avg Loss: 0.99915974, Log Avg loss: 0.91646155, Global Avg Loss: 3.51750497, Time: 0.0132 Steps: 18570, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001548, Sample Num: 24768, Cur Loss: 1.14947069, Cur Avg Loss: 0.99763895, Log Avg loss: 0.76374235, Global Avg Loss: 3.51602286, Time: 0.0217 Steps: 18580, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001558, Sample Num: 24928, Cur Loss: 2.05407500, Cur Avg Loss: 0.99833132, Log Avg loss: 1.10551011, Global Avg Loss: 3.51472618, Time: 0.0112 Steps: 18590, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001568, Sample Num: 25088, Cur Loss: 1.48040438, Cur Avg Loss: 0.99690328, Log Avg loss: 0.77441521, Global Avg Loss: 3.51325290, Time: 0.0077 Steps: 18600, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001578, Sample Num: 25248, Cur Loss: 1.98074591, Cur Avg Loss: 0.99821414, Log Avg loss: 1.20375559, Global Avg Loss: 3.51201190, Time: 0.0068 Steps: 18610, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001588, Sample Num: 25408, Cur Loss: 1.47392178, Cur Avg Loss: 0.99761684, Log Avg loss: 0.90336316, Global Avg Loss: 3.51061091, Time: 0.0119 Steps: 18620, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001598, Sample Num: 25568, Cur Loss: 1.34514546, Cur Avg Loss: 0.99789040, Log Avg loss: 1.04133234, Global Avg Loss: 3.50928548, Time: 0.0067 Steps: 18630, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001608, Sample Num: 25728, Cur Loss: 0.48763955, Cur Avg Loss: 0.99784435, Log Avg loss: 0.99048480, Global Avg Loss: 3.50793419, Time: 0.0129 Steps: 18640, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001618, Sample Num: 25888, Cur Loss: 2.17354345, Cur Avg Loss: 0.99811880, Log Avg loss: 1.04225094, Global Avg Loss: 3.50661211, Time: 0.0140 Steps: 18650, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001628, Sample Num: 26048, Cur Loss: 0.79384309, Cur Avg Loss: 0.99828610, Log Avg loss: 1.02535586, Global Avg Loss: 3.50528239, Time: 0.0136 Steps: 18660, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001638, Sample Num: 26208, Cur Loss: 0.67012739, Cur Avg Loss: 0.99760679, Log Avg loss: 0.88701475, Global Avg Loss: 3.50387999, Time: 0.0068 Steps: 18670, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001648, Sample Num: 26368, Cur Loss: 1.52712607, Cur Avg Loss: 0.99788531, Log Avg loss: 1.04350722, Global Avg Loss: 3.50256288, Time: 0.0070 Steps: 18680, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001658, Sample Num: 26528, Cur Loss: 0.68192798, Cur Avg Loss: 0.99702903, Log Avg loss: 0.85591388, Global Avg Loss: 3.50114680, Time: 0.0069 Steps: 18690, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001668, Sample Num: 26688, Cur Loss: 0.67211521, Cur Avg Loss: 0.99633026, Log Avg loss: 0.88047332, Global Avg Loss: 3.49974537, Time: 0.0069 Steps: 18700, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001678, Sample Num: 26848, Cur Loss: 0.63401276, Cur Avg Loss: 0.99583060, Log Avg loss: 0.91248792, Global Avg Loss: 3.49836255, Time: 0.0140 Steps: 18710, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001688, Sample Num: 27008, Cur Loss: 1.54514551, Cur Avg Loss: 0.99696081, Log Avg loss: 1.18660904, Global Avg Loss: 3.49712764, Time: 0.0125 Steps: 18720, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001698, Sample Num: 27168, Cur Loss: 0.40640557, Cur Avg Loss: 0.99604758, Log Avg loss: 0.84189517, Global Avg Loss: 3.49571000, Time: 0.0144 Steps: 18730, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001708, Sample Num: 27328, Cur Loss: 0.33105054, Cur Avg Loss: 0.99504055, Log Avg loss: 0.82404762, Global Avg Loss: 3.49428436, Time: 0.0117 Steps: 18740, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001718, Sample Num: 27488, Cur Loss: 1.56528020, Cur Avg Loss: 0.99532903, Log Avg loss: 1.04460083, Global Avg Loss: 3.49297786, Time: 0.0068 Steps: 18750, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001728, Sample Num: 27648, Cur Loss: 0.95836222, Cur Avg Loss: 0.99614575, Log Avg loss: 1.13645847, Global Avg Loss: 3.49172172, Time: 0.0134 Steps: 18760, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001738, Sample Num: 27808, Cur Loss: 1.16285634, Cur Avg Loss: 0.99653631, Log Avg loss: 1.06402494, Global Avg Loss: 3.49042833, Time: 0.0109 Steps: 18770, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001748, Sample Num: 27968, Cur Loss: 0.62225354, Cur Avg Loss: 0.99814068, Log Avg loss: 1.27698059, Global Avg Loss: 3.48924971, Time: 0.0068 Steps: 18780, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001758, Sample Num: 28128, Cur Loss: 1.32052755, Cur Avg Loss: 0.99658869, Log Avg loss: 0.72529966, Global Avg Loss: 3.48777874, Time: 0.0134 Steps: 18790, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001768, Sample Num: 28288, Cur Loss: 0.92790079, Cur Avg Loss: 0.99561066, Log Avg loss: 0.82367407, Global Avg Loss: 3.48636166, Time: 0.0104 Steps: 18800, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001778, Sample Num: 28448, Cur Loss: 0.24445421, Cur Avg Loss: 0.99469632, Log Avg loss: 0.83304042, Global Avg Loss: 3.48495107, Time: 0.0068 Steps: 18810, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001788, Sample Num: 28608, Cur Loss: 1.04121447, Cur Avg Loss: 0.99475504, Log Avg loss: 1.00519516, Global Avg Loss: 3.48363345, Time: 0.0117 Steps: 18820, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001798, Sample Num: 28768, Cur Loss: 1.09305692, Cur Avg Loss: 0.99422025, Log Avg loss: 0.89859931, Global Avg Loss: 3.48226062, Time: 0.0225 Steps: 18830, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001808, Sample Num: 28928, Cur Loss: 0.93198603, Cur Avg Loss: 0.99567776, Log Avg loss: 1.25773972, Global Avg Loss: 3.48107988, Time: 0.0074 Steps: 18840, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001818, Sample Num: 29088, Cur Loss: 1.12531114, Cur Avg Loss: 0.99724763, Log Avg loss: 1.28107906, Global Avg Loss: 3.47991277, Time: 0.0067 Steps: 18850, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001828, Sample Num: 29248, Cur Loss: 1.62913871, Cur Avg Loss: 0.99767490, Log Avg loss: 1.07535236, Global Avg Loss: 3.47863782, Time: 0.0065 Steps: 18860, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001838, Sample Num: 29408, Cur Loss: 0.77980030, Cur Avg Loss: 0.99670441, Log Avg loss: 0.81929881, Global Avg Loss: 3.47722852, Time: 0.0065 Steps: 18870, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001848, Sample Num: 29568, Cur Loss: 0.44591841, Cur Avg Loss: 0.99701189, Log Avg loss: 1.05352633, Global Avg Loss: 3.47594478, Time: 0.0164 Steps: 18880, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001858, Sample Num: 29728, Cur Loss: 0.82460833, Cur Avg Loss: 0.99738553, Log Avg loss: 1.06643417, Global Avg Loss: 3.47466924, Time: 0.0066 Steps: 18890, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001868, Sample Num: 29888, Cur Loss: 0.17881742, Cur Avg Loss: 0.99682422, Log Avg loss: 0.89253319, Global Avg Loss: 3.47330303, Time: 0.0115 Steps: 18900, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001878, Sample Num: 30048, Cur Loss: 0.23539422, Cur Avg Loss: 0.99577281, Log Avg loss: 0.79936914, Global Avg Loss: 3.47188899, Time: 0.0110 Steps: 18910, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001888, Sample Num: 30208, Cur Loss: 0.98119283, Cur Avg Loss: 0.99655327, Log Avg loss: 1.14312472, Global Avg Loss: 3.47065815, Time: 0.0077 Steps: 18920, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001898, Sample Num: 30368, Cur Loss: 0.60783076, Cur Avg Loss: 0.99709461, Log Avg loss: 1.09929976, Global Avg Loss: 3.46940545, Time: 0.0067 Steps: 18930, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001908, Sample Num: 30528, Cur Loss: 0.89117730, Cur Avg Loss: 0.99618813, Log Avg loss: 0.82413846, Global Avg Loss: 3.46800879, Time: 0.0089 Steps: 18940, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001918, Sample Num: 30688, Cur Loss: 0.44372693, Cur Avg Loss: 0.99549211, Log Avg loss: 0.86269108, Global Avg Loss: 3.46663395, Time: 0.0117 Steps: 18950, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001928, Sample Num: 30848, Cur Loss: 1.66514349, Cur Avg Loss: 0.99514106, Log Avg loss: 0.92780906, Global Avg Loss: 3.46529491, Time: 0.0207 Steps: 18960, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001938, Sample Num: 31008, Cur Loss: 1.25011849, Cur Avg Loss: 0.99556364, Log Avg loss: 1.07703747, Global Avg Loss: 3.46403595, Time: 0.0182 Steps: 18970, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001948, Sample Num: 31168, Cur Loss: 0.55459934, Cur Avg Loss: 0.99454118, Log Avg loss: 0.79638744, Global Avg Loss: 3.46263044, Time: 0.0067 Steps: 18980, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001958, Sample Num: 31328, Cur Loss: 0.80907482, Cur Avg Loss: 0.99364047, Log Avg loss: 0.81818311, Global Avg Loss: 3.46123789, Time: 0.0068 Steps: 18990, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001968, Sample Num: 31488, Cur Loss: 1.01090848, Cur Avg Loss: 0.99206308, Log Avg loss: 0.68320923, Global Avg Loss: 3.45977577, Time: 0.0120 Steps: 19000, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001978, Sample Num: 31648, Cur Loss: 0.33143136, Cur Avg Loss: 0.99252290, Log Avg loss: 1.08301581, Global Avg Loss: 3.45852551, Time: 0.0108 Steps: 19010, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001988, Sample Num: 31808, Cur Loss: 0.80201912, Cur Avg Loss: 0.99378815, Log Avg loss: 1.24405609, Global Avg Loss: 3.45736122, Time: 0.0108 Steps: 19020, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 001998, Sample Num: 31968, Cur Loss: 0.81031895, Cur Avg Loss: 0.99310398, Log Avg loss: 0.85708927, Global Avg Loss: 3.45599481, Time: 0.0068 Steps: 19030, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002008, Sample Num: 32128, Cur Loss: 0.73773181, Cur Avg Loss: 0.99455977, Log Avg loss: 1.28542702, Global Avg Loss: 3.45485481, Time: 0.0068 Steps: 19040, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002018, Sample Num: 32288, Cur Loss: 1.20610869, Cur Avg Loss: 0.99429406, Log Avg loss: 0.94094027, Global Avg Loss: 3.45353517, Time: 0.0123 Steps: 19050, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002028, Sample Num: 32448, Cur Loss: 0.99687380, Cur Avg Loss: 0.99400412, Log Avg loss: 0.93549464, Global Avg Loss: 3.45221406, Time: 0.0218 Steps: 19060, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002038, Sample Num: 32608, Cur Loss: 1.01562309, Cur Avg Loss: 0.99433465, Log Avg loss: 1.06136508, Global Avg Loss: 3.45096033, Time: 0.0112 Steps: 19070, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002048, Sample Num: 32768, Cur Loss: 0.59870505, Cur Avg Loss: 0.99370583, Log Avg loss: 0.86555189, Global Avg Loss: 3.44960530, Time: 0.0084 Steps: 19080, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002058, Sample Num: 32928, Cur Loss: 0.89751118, Cur Avg Loss: 0.99338041, Log Avg loss: 0.92673604, Global Avg Loss: 3.44828373, Time: 0.0068 Steps: 19090, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002068, Sample Num: 33088, Cur Loss: 1.34281957, Cur Avg Loss: 0.99293729, Log Avg loss: 0.90174236, Global Avg Loss: 3.44695047, Time: 0.0088 Steps: 19100, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002078, Sample Num: 33248, Cur Loss: 0.48742166, Cur Avg Loss: 0.99276332, Log Avg loss: 0.95678687, Global Avg Loss: 3.44564740, Time: 0.0091 Steps: 19110, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002088, Sample Num: 33408, Cur Loss: 0.75544643, Cur Avg Loss: 0.99193238, Log Avg loss: 0.81926299, Global Avg Loss: 3.44427376, Time: 0.0112 Steps: 19120, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002098, Sample Num: 33568, Cur Loss: 0.45389014, Cur Avg Loss: 0.99131036, Log Avg loss: 0.86143272, Global Avg Loss: 3.44292361, Time: 0.0114 Steps: 19130, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002108, Sample Num: 33728, Cur Loss: 0.41456568, Cur Avg Loss: 0.99091898, Log Avg loss: 0.90880563, Global Avg Loss: 3.44159962, Time: 0.0113 Steps: 19140, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002118, Sample Num: 33888, Cur Loss: 0.76946181, Cur Avg Loss: 0.99002377, Log Avg loss: 0.80131488, Global Avg Loss: 3.44022088, Time: 0.0068 Steps: 19150, Updated lr: 0.000083 Training, Epoch: 0009, Batch: 002128, Sample Num: 34048, Cur Loss: 0.58612919, Cur Avg Loss: 0.98920825, Log Avg loss: 0.81648103, Global Avg Loss: 3.43885150, Time: 0.0126 Steps: 19160, Updated lr: 0.000083 ***** Running evaluation checkpoint-19161 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-19161 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.448681, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.001192, "eval_total_loss": 703.837768, "eval_mae": 0.809697, "eval_mse": 1.001368, "eval_r2": 0.363465, "eval_sp_statistic": 0.587733, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.614717, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 2.200972, "test_total_loss": 1104.887912, "test_mae": 1.367623, "test_mse": 2.200649, "test_r2": -0.420319, "test_sp_statistic": 0.310338, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.45301, "test_ps_pvalue": 0.0, "lr": 8.277761972498815e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.4387135319230944, "train_cur_epoch_loss": 2105.830420255661, "train_cur_epoch_avg_loss": 0.9891171537133213, "train_cur_epoch_time": 22.448681116104126, "train_cur_epoch_avg_time": 0.010544237255098227, "epoch": 9, "step": 19161} ################################################## Training, Epoch: 0010, Batch: 000009, Sample Num: 144, Cur Loss: 0.47397774, Cur Avg Loss: 0.76009204, Log Avg loss: 0.76360900, Global Avg Loss: 3.43745596, Time: 0.0114 Steps: 19170, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000019, Sample Num: 304, Cur Loss: 0.78412509, Cur Avg Loss: 1.04160321, Log Avg loss: 1.29496326, Global Avg Loss: 3.43633892, Time: 0.0140 Steps: 19180, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000029, Sample Num: 464, Cur Loss: 1.04348016, Cur Avg Loss: 1.04790143, Log Avg loss: 1.05986805, Global Avg Loss: 3.43510053, Time: 0.0159 Steps: 19190, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000039, Sample Num: 624, Cur Loss: 0.62340081, Cur Avg Loss: 1.03914087, Log Avg loss: 1.01373523, Global Avg Loss: 3.43383940, Time: 0.0111 Steps: 19200, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000049, Sample Num: 784, Cur Loss: 0.64328051, Cur Avg Loss: 0.99250550, Log Avg loss: 0.81062757, Global Avg Loss: 3.43247386, Time: 0.0107 Steps: 19210, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000059, Sample Num: 944, Cur Loss: 0.67061424, Cur Avg Loss: 0.95525936, Log Avg loss: 0.77275329, Global Avg Loss: 3.43109003, Time: 0.0161 Steps: 19220, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000069, Sample Num: 1104, Cur Loss: 1.54880214, Cur Avg Loss: 0.98968028, Log Avg loss: 1.19276370, Global Avg Loss: 3.42992605, Time: 0.0071 Steps: 19230, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000079, Sample Num: 1264, Cur Loss: 0.79891741, Cur Avg Loss: 0.97869374, Log Avg loss: 0.90288660, Global Avg Loss: 3.42861262, Time: 0.0065 Steps: 19240, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000089, Sample Num: 1424, Cur Loss: 0.96661758, Cur Avg Loss: 0.98347005, Log Avg loss: 1.02120291, Global Avg Loss: 3.42736202, Time: 0.0141 Steps: 19250, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000099, Sample Num: 1584, Cur Loss: 1.35169148, Cur Avg Loss: 0.98581057, Log Avg loss: 1.00664124, Global Avg Loss: 3.42610515, Time: 0.0115 Steps: 19260, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000109, Sample Num: 1744, Cur Loss: 0.95087153, Cur Avg Loss: 0.98565454, Log Avg loss: 0.98410980, Global Avg Loss: 3.42483790, Time: 0.0111 Steps: 19270, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000119, Sample Num: 1904, Cur Loss: 0.47820580, Cur Avg Loss: 0.97347688, Log Avg loss: 0.84074044, Global Avg Loss: 3.42349760, Time: 0.0087 Steps: 19280, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000129, Sample Num: 2064, Cur Loss: 1.25955713, Cur Avg Loss: 0.98162529, Log Avg loss: 1.07859131, Global Avg Loss: 3.42228199, Time: 0.0066 Steps: 19290, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000139, Sample Num: 2224, Cur Loss: 0.49941128, Cur Avg Loss: 0.97506839, Log Avg loss: 0.89048441, Global Avg Loss: 3.42097018, Time: 0.0118 Steps: 19300, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000149, Sample Num: 2384, Cur Loss: 0.78161407, Cur Avg Loss: 0.96886873, Log Avg loss: 0.88269347, Global Avg Loss: 3.41965569, Time: 0.0116 Steps: 19310, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000159, Sample Num: 2544, Cur Loss: 1.16249156, Cur Avg Loss: 0.97542989, Log Avg loss: 1.07319115, Global Avg Loss: 3.41844117, Time: 0.0067 Steps: 19320, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000169, Sample Num: 2704, Cur Loss: 1.04914880, Cur Avg Loss: 0.97651780, Log Avg loss: 0.99381557, Global Avg Loss: 3.41718683, Time: 0.0109 Steps: 19330, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000179, Sample Num: 2864, Cur Loss: 0.61907727, Cur Avg Loss: 0.96844014, Log Avg loss: 0.83192773, Global Avg Loss: 3.41585009, Time: 0.0100 Steps: 19340, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000189, Sample Num: 3024, Cur Loss: 0.66082752, Cur Avg Loss: 0.97451485, Log Avg loss: 1.08325208, Global Avg Loss: 3.41464461, Time: 0.0121 Steps: 19350, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000199, Sample Num: 3184, Cur Loss: 1.06626892, Cur Avg Loss: 0.98751976, Log Avg loss: 1.23331249, Global Avg Loss: 3.41351789, Time: 0.0124 Steps: 19360, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000209, Sample Num: 3344, Cur Loss: 0.76411170, Cur Avg Loss: 0.98527328, Log Avg loss: 0.94056833, Global Avg Loss: 3.41224120, Time: 0.0095 Steps: 19370, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000219, Sample Num: 3504, Cur Loss: 0.63188332, Cur Avg Loss: 0.97705818, Log Avg loss: 0.80536265, Global Avg Loss: 3.41089606, Time: 0.0149 Steps: 19380, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000229, Sample Num: 3664, Cur Loss: 0.33683527, Cur Avg Loss: 0.96817319, Log Avg loss: 0.77359197, Global Avg Loss: 3.40953593, Time: 0.0065 Steps: 19390, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000239, Sample Num: 3824, Cur Loss: 0.45761576, Cur Avg Loss: 0.96245415, Log Avg loss: 0.83148814, Global Avg Loss: 3.40820704, Time: 0.0166 Steps: 19400, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000249, Sample Num: 3984, Cur Loss: 1.31270516, Cur Avg Loss: 0.98126242, Log Avg loss: 1.43077998, Global Avg Loss: 3.40718827, Time: 0.0119 Steps: 19410, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000259, Sample Num: 4144, Cur Loss: 0.58792603, Cur Avg Loss: 0.98173756, Log Avg loss: 0.99356850, Global Avg Loss: 3.40594542, Time: 0.0073 Steps: 19420, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000269, Sample Num: 4304, Cur Loss: 0.57269657, Cur Avg Loss: 0.97958795, Log Avg loss: 0.92391325, Global Avg Loss: 3.40466800, Time: 0.0064 Steps: 19430, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000279, Sample Num: 4464, Cur Loss: 1.34731758, Cur Avg Loss: 0.97670786, Log Avg loss: 0.89923334, Global Avg Loss: 3.40337919, Time: 0.0118 Steps: 19440, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000289, Sample Num: 4624, Cur Loss: 1.27908993, Cur Avg Loss: 0.98008602, Log Avg loss: 1.07433670, Global Avg Loss: 3.40218174, Time: 0.0118 Steps: 19450, Updated lr: 0.000083 Training, Epoch: 0010, Batch: 000299, Sample Num: 4784, Cur Loss: 0.36424187, Cur Avg Loss: 0.97513613, Log Avg loss: 0.83208417, Global Avg Loss: 3.40086103, Time: 0.0065 Steps: 19460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000309, Sample Num: 4944, Cur Loss: 1.26710463, Cur Avg Loss: 0.98265417, Log Avg loss: 1.20744376, Global Avg Loss: 3.39973447, Time: 0.0121 Steps: 19470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000319, Sample Num: 5104, Cur Loss: 0.48724014, Cur Avg Loss: 0.97808520, Log Avg loss: 0.83690392, Global Avg Loss: 3.39841885, Time: 0.0067 Steps: 19480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000329, Sample Num: 5264, Cur Loss: 0.41028681, Cur Avg Loss: 0.98103745, Log Avg loss: 1.07521435, Global Avg Loss: 3.39722685, Time: 0.0085 Steps: 19490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000339, Sample Num: 5424, Cur Loss: 0.73323327, Cur Avg Loss: 0.98094438, Log Avg loss: 0.97788222, Global Avg Loss: 3.39598616, Time: 0.0123 Steps: 19500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000349, Sample Num: 5584, Cur Loss: 0.82040489, Cur Avg Loss: 0.97907891, Log Avg loss: 0.91583964, Global Avg Loss: 3.39471494, Time: 0.0096 Steps: 19510, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000359, Sample Num: 5744, Cur Loss: 1.05530512, Cur Avg Loss: 0.97916172, Log Avg loss: 0.98205166, Global Avg Loss: 3.39347895, Time: 0.0066 Steps: 19520, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000369, Sample Num: 5904, Cur Loss: 1.11008155, Cur Avg Loss: 0.98487353, Log Avg loss: 1.18992766, Global Avg Loss: 3.39235066, Time: 0.0068 Steps: 19530, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000379, Sample Num: 6064, Cur Loss: 0.48592156, Cur Avg Loss: 0.98622734, Log Avg loss: 1.03618286, Global Avg Loss: 3.39114484, Time: 0.0071 Steps: 19540, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000389, Sample Num: 6224, Cur Loss: 0.38998318, Cur Avg Loss: 0.98572665, Log Avg loss: 0.96675048, Global Avg Loss: 3.38990474, Time: 0.0071 Steps: 19550, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000399, Sample Num: 6384, Cur Loss: 1.50687408, Cur Avg Loss: 0.98123215, Log Avg loss: 0.80639593, Global Avg Loss: 3.38858393, Time: 0.0127 Steps: 19560, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000409, Sample Num: 6544, Cur Loss: 1.36938095, Cur Avg Loss: 0.98026272, Log Avg loss: 0.94158256, Global Avg Loss: 3.38733354, Time: 0.0068 Steps: 19570, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000419, Sample Num: 6704, Cur Loss: 1.91369724, Cur Avg Loss: 0.98140694, Log Avg loss: 1.02820570, Global Avg Loss: 3.38612868, Time: 0.0121 Steps: 19580, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000429, Sample Num: 6864, Cur Loss: 0.70288408, Cur Avg Loss: 0.98961010, Log Avg loss: 1.33332248, Global Avg Loss: 3.38508079, Time: 0.0227 Steps: 19590, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000439, Sample Num: 7024, Cur Loss: 0.99670732, Cur Avg Loss: 0.98847095, Log Avg loss: 0.93960142, Global Avg Loss: 3.38383310, Time: 0.0106 Steps: 19600, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000449, Sample Num: 7184, Cur Loss: 0.36790609, Cur Avg Loss: 0.98373672, Log Avg loss: 0.77590407, Global Avg Loss: 3.38250320, Time: 0.0225 Steps: 19610, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000459, Sample Num: 7344, Cur Loss: 0.71120936, Cur Avg Loss: 0.98142248, Log Avg loss: 0.87751291, Global Avg Loss: 3.38122645, Time: 0.0138 Steps: 19620, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000469, Sample Num: 7504, Cur Loss: 1.36152911, Cur Avg Loss: 0.98590026, Log Avg loss: 1.19143057, Global Avg Loss: 3.38011091, Time: 0.0070 Steps: 19630, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000479, Sample Num: 7664, Cur Loss: 0.55801839, Cur Avg Loss: 0.98367149, Log Avg loss: 0.87914212, Global Avg Loss: 3.37883751, Time: 0.0068 Steps: 19640, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000489, Sample Num: 7824, Cur Loss: 1.72919834, Cur Avg Loss: 0.98357007, Log Avg loss: 0.97871211, Global Avg Loss: 3.37761607, Time: 0.0133 Steps: 19650, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000499, Sample Num: 7984, Cur Loss: 1.04930663, Cur Avg Loss: 0.98442050, Log Avg loss: 1.02600656, Global Avg Loss: 3.37641993, Time: 0.0108 Steps: 19660, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000509, Sample Num: 8144, Cur Loss: 1.38906956, Cur Avg Loss: 0.98205473, Log Avg loss: 0.86400263, Global Avg Loss: 3.37514265, Time: 0.0117 Steps: 19670, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000519, Sample Num: 8304, Cur Loss: 0.99403441, Cur Avg Loss: 0.97941386, Log Avg loss: 0.84499365, Global Avg Loss: 3.37385700, Time: 0.0115 Steps: 19680, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000529, Sample Num: 8464, Cur Loss: 2.93648529, Cur Avg Loss: 0.98518848, Log Avg loss: 1.28489139, Global Avg Loss: 3.37279607, Time: 0.0073 Steps: 19690, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000539, Sample Num: 8624, Cur Loss: 1.12464166, Cur Avg Loss: 0.98723882, Log Avg loss: 1.09570164, Global Avg Loss: 3.37164019, Time: 0.0066 Steps: 19700, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000549, Sample Num: 8784, Cur Loss: 0.94771254, Cur Avg Loss: 0.98675801, Log Avg loss: 0.96084210, Global Avg Loss: 3.37041705, Time: 0.0068 Steps: 19710, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000559, Sample Num: 8944, Cur Loss: 0.90314072, Cur Avg Loss: 0.98254955, Log Avg loss: 0.75150514, Global Avg Loss: 3.36908901, Time: 0.0156 Steps: 19720, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000569, Sample Num: 9104, Cur Loss: 0.89341545, Cur Avg Loss: 0.97897814, Log Avg loss: 0.77933624, Global Avg Loss: 3.36777641, Time: 0.0167 Steps: 19730, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000579, Sample Num: 9264, Cur Loss: 1.45061016, Cur Avg Loss: 0.97764635, Log Avg loss: 0.90186761, Global Avg Loss: 3.36652721, Time: 0.0134 Steps: 19740, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000589, Sample Num: 9424, Cur Loss: 0.68793142, Cur Avg Loss: 0.97601834, Log Avg loss: 0.88175657, Global Avg Loss: 3.36526910, Time: 0.0078 Steps: 19750, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000599, Sample Num: 9584, Cur Loss: 0.59865260, Cur Avg Loss: 0.97607135, Log Avg loss: 0.97919364, Global Avg Loss: 3.36406158, Time: 0.0067 Steps: 19760, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000609, Sample Num: 9744, Cur Loss: 0.65972018, Cur Avg Loss: 0.97365678, Log Avg loss: 0.82902420, Global Avg Loss: 3.36277931, Time: 0.0114 Steps: 19770, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000619, Sample Num: 9904, Cur Loss: 0.66637152, Cur Avg Loss: 0.97100789, Log Avg loss: 0.80969042, Global Avg Loss: 3.36148857, Time: 0.0169 Steps: 19780, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000629, Sample Num: 10064, Cur Loss: 1.42947650, Cur Avg Loss: 0.97056776, Log Avg loss: 0.94332346, Global Avg Loss: 3.36026666, Time: 0.0139 Steps: 19790, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000639, Sample Num: 10224, Cur Loss: 1.32707715, Cur Avg Loss: 0.97314120, Log Avg loss: 1.13501061, Global Avg Loss: 3.35914279, Time: 0.0139 Steps: 19800, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000649, Sample Num: 10384, Cur Loss: 1.39880133, Cur Avg Loss: 0.97127609, Log Avg loss: 0.85209568, Global Avg Loss: 3.35787724, Time: 0.0116 Steps: 19810, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000659, Sample Num: 10544, Cur Loss: 1.58115160, Cur Avg Loss: 0.97416359, Log Avg loss: 1.16156250, Global Avg Loss: 3.35676911, Time: 0.0142 Steps: 19820, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000669, Sample Num: 10704, Cur Loss: 1.43399382, Cur Avg Loss: 0.97136290, Log Avg loss: 0.78679770, Global Avg Loss: 3.35547311, Time: 0.0107 Steps: 19830, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000679, Sample Num: 10864, Cur Loss: 0.60015905, Cur Avg Loss: 0.97039125, Log Avg loss: 0.90538754, Global Avg Loss: 3.35423819, Time: 0.0068 Steps: 19840, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000689, Sample Num: 11024, Cur Loss: 1.13842726, Cur Avg Loss: 0.96912659, Log Avg loss: 0.88325652, Global Avg Loss: 3.35299336, Time: 0.0083 Steps: 19850, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000699, Sample Num: 11184, Cur Loss: 0.76338750, Cur Avg Loss: 0.96612595, Log Avg loss: 0.75938177, Global Avg Loss: 3.35168741, Time: 0.0116 Steps: 19860, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000709, Sample Num: 11344, Cur Loss: 0.71613097, Cur Avg Loss: 0.96613669, Log Avg loss: 0.96688700, Global Avg Loss: 3.35048721, Time: 0.0121 Steps: 19870, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000719, Sample Num: 11504, Cur Loss: 1.12047553, Cur Avg Loss: 0.96603496, Log Avg loss: 0.95882272, Global Avg Loss: 3.34928416, Time: 0.0126 Steps: 19880, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000729, Sample Num: 11664, Cur Loss: 0.51708758, Cur Avg Loss: 0.96443697, Log Avg loss: 0.84954158, Global Avg Loss: 3.34802738, Time: 0.0066 Steps: 19890, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000739, Sample Num: 11824, Cur Loss: 1.77695155, Cur Avg Loss: 0.96814759, Log Avg loss: 1.23865163, Global Avg Loss: 3.34696739, Time: 0.0066 Steps: 19900, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000749, Sample Num: 11984, Cur Loss: 1.02131093, Cur Avg Loss: 0.96416672, Log Avg loss: 0.66998028, Global Avg Loss: 3.34562285, Time: 0.0067 Steps: 19910, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000759, Sample Num: 12144, Cur Loss: 0.40641838, Cur Avg Loss: 0.96382101, Log Avg loss: 0.93792747, Global Avg Loss: 3.34441416, Time: 0.0118 Steps: 19920, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000769, Sample Num: 12304, Cur Loss: 1.12130606, Cur Avg Loss: 0.96295113, Log Avg loss: 0.89692717, Global Avg Loss: 3.34318612, Time: 0.0134 Steps: 19930, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000779, Sample Num: 12464, Cur Loss: 0.49035871, Cur Avg Loss: 0.96034222, Log Avg loss: 0.75971721, Global Avg Loss: 3.34189050, Time: 0.0096 Steps: 19940, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000789, Sample Num: 12624, Cur Loss: 0.92290747, Cur Avg Loss: 0.96018326, Log Avg loss: 0.94779996, Global Avg Loss: 3.34069045, Time: 0.0115 Steps: 19950, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000799, Sample Num: 12784, Cur Loss: 0.54344308, Cur Avg Loss: 0.95964789, Log Avg loss: 0.91740710, Global Avg Loss: 3.33947639, Time: 0.0161 Steps: 19960, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000809, Sample Num: 12944, Cur Loss: 0.34013987, Cur Avg Loss: 0.95624094, Log Avg loss: 0.68402550, Global Avg Loss: 3.33814667, Time: 0.0109 Steps: 19970, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000819, Sample Num: 13104, Cur Loss: 0.36794877, Cur Avg Loss: 0.95401589, Log Avg loss: 0.77400924, Global Avg Loss: 3.33686331, Time: 0.0109 Steps: 19980, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000829, Sample Num: 13264, Cur Loss: 0.86621875, Cur Avg Loss: 0.95299969, Log Avg loss: 0.86977322, Global Avg Loss: 3.33562915, Time: 0.0125 Steps: 19990, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000839, Sample Num: 13424, Cur Loss: 1.29899526, Cur Avg Loss: 0.95361063, Log Avg loss: 1.00425780, Global Avg Loss: 3.33446347, Time: 0.0064 Steps: 20000, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000849, Sample Num: 13584, Cur Loss: 0.63858169, Cur Avg Loss: 0.95364319, Log Avg loss: 0.95637487, Global Avg Loss: 3.33327502, Time: 0.0111 Steps: 20010, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000859, Sample Num: 13744, Cur Loss: 1.04803658, Cur Avg Loss: 0.95306461, Log Avg loss: 0.90394332, Global Avg Loss: 3.33206156, Time: 0.0117 Steps: 20020, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000869, Sample Num: 13904, Cur Loss: 0.87639242, Cur Avg Loss: 0.95140456, Log Avg loss: 0.80880584, Global Avg Loss: 3.33080182, Time: 0.0066 Steps: 20030, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000879, Sample Num: 14064, Cur Loss: 0.86710942, Cur Avg Loss: 0.95103888, Log Avg loss: 0.91926177, Global Avg Loss: 3.32959846, Time: 0.0109 Steps: 20040, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000889, Sample Num: 14224, Cur Loss: 1.15744162, Cur Avg Loss: 0.94991691, Log Avg loss: 0.85129536, Global Avg Loss: 3.32836240, Time: 0.0121 Steps: 20050, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000899, Sample Num: 14384, Cur Loss: 1.45089507, Cur Avg Loss: 0.95382948, Log Avg loss: 1.30165720, Global Avg Loss: 3.32735208, Time: 0.0066 Steps: 20060, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000909, Sample Num: 14544, Cur Loss: 0.77525878, Cur Avg Loss: 0.95555564, Log Avg loss: 1.11073687, Global Avg Loss: 3.32624764, Time: 0.0120 Steps: 20070, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000919, Sample Num: 14704, Cur Loss: 0.71033132, Cur Avg Loss: 0.95418748, Log Avg loss: 0.82982222, Global Avg Loss: 3.32500440, Time: 0.0108 Steps: 20080, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000929, Sample Num: 14864, Cur Loss: 1.27588260, Cur Avg Loss: 0.95345832, Log Avg loss: 0.88644864, Global Avg Loss: 3.32379058, Time: 0.0064 Steps: 20090, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000939, Sample Num: 15024, Cur Loss: 0.65559733, Cur Avg Loss: 0.95414119, Log Avg loss: 1.01757934, Global Avg Loss: 3.32264321, Time: 0.0114 Steps: 20100, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000949, Sample Num: 15184, Cur Loss: 0.68752193, Cur Avg Loss: 0.95346620, Log Avg loss: 0.89008530, Global Avg Loss: 3.32143359, Time: 0.0124 Steps: 20110, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000959, Sample Num: 15344, Cur Loss: 0.43618381, Cur Avg Loss: 0.95575229, Log Avg loss: 1.17270140, Global Avg Loss: 3.32036563, Time: 0.0079 Steps: 20120, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000969, Sample Num: 15504, Cur Loss: 0.54916722, Cur Avg Loss: 0.95571958, Log Avg loss: 0.95258285, Global Avg Loss: 3.31918938, Time: 0.0068 Steps: 20130, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000979, Sample Num: 15664, Cur Loss: 1.52856147, Cur Avg Loss: 0.95694988, Log Avg loss: 1.07616628, Global Avg Loss: 3.31807567, Time: 0.0120 Steps: 20140, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000989, Sample Num: 15824, Cur Loss: 2.77278209, Cur Avg Loss: 0.95853163, Log Avg loss: 1.11338479, Global Avg Loss: 3.31698153, Time: 0.0227 Steps: 20150, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 000999, Sample Num: 15984, Cur Loss: 0.94728667, Cur Avg Loss: 0.95814701, Log Avg loss: 0.92010803, Global Avg Loss: 3.31579260, Time: 0.0114 Steps: 20160, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001009, Sample Num: 16144, Cur Loss: 0.95821083, Cur Avg Loss: 0.95745678, Log Avg loss: 0.88850285, Global Avg Loss: 3.31458919, Time: 0.0132 Steps: 20170, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001019, Sample Num: 16304, Cur Loss: 1.11225665, Cur Avg Loss: 0.96125860, Log Avg loss: 1.34486230, Global Avg Loss: 3.31361311, Time: 0.0064 Steps: 20180, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001029, Sample Num: 16464, Cur Loss: 1.81147718, Cur Avg Loss: 0.96300325, Log Avg loss: 1.14078337, Global Avg Loss: 3.31253692, Time: 0.0074 Steps: 20190, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001039, Sample Num: 16624, Cur Loss: 1.72324693, Cur Avg Loss: 0.96432248, Log Avg loss: 1.10007077, Global Avg Loss: 3.31144164, Time: 0.0128 Steps: 20200, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001049, Sample Num: 16784, Cur Loss: 1.05369759, Cur Avg Loss: 0.96430603, Log Avg loss: 0.96259690, Global Avg Loss: 3.31027942, Time: 0.0116 Steps: 20210, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001059, Sample Num: 16944, Cur Loss: 1.16368484, Cur Avg Loss: 0.96451862, Log Avg loss: 0.98681932, Global Avg Loss: 3.30913033, Time: 0.0105 Steps: 20220, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001069, Sample Num: 17104, Cur Loss: 0.53798199, Cur Avg Loss: 0.96375379, Log Avg loss: 0.88275806, Global Avg Loss: 3.30793093, Time: 0.0071 Steps: 20230, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001079, Sample Num: 17264, Cur Loss: 1.25664032, Cur Avg Loss: 0.96130068, Log Avg loss: 0.69906397, Global Avg Loss: 3.30664197, Time: 0.0115 Steps: 20240, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001089, Sample Num: 17424, Cur Loss: 0.97787023, Cur Avg Loss: 0.96025425, Log Avg loss: 0.84734374, Global Avg Loss: 3.30542750, Time: 0.0128 Steps: 20250, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001099, Sample Num: 17584, Cur Loss: 0.40559137, Cur Avg Loss: 0.95943309, Log Avg loss: 0.87000940, Global Avg Loss: 3.30422542, Time: 0.0128 Steps: 20260, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001109, Sample Num: 17744, Cur Loss: 0.89980549, Cur Avg Loss: 0.95923288, Log Avg loss: 0.93722986, Global Avg Loss: 3.30305768, Time: 0.0066 Steps: 20270, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001119, Sample Num: 17904, Cur Loss: 1.27229524, Cur Avg Loss: 0.96006958, Log Avg loss: 1.05285908, Global Avg Loss: 3.30194812, Time: 0.0067 Steps: 20280, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001129, Sample Num: 18064, Cur Loss: 1.27985382, Cur Avg Loss: 0.96003408, Log Avg loss: 0.95606139, Global Avg Loss: 3.30079194, Time: 0.0065 Steps: 20290, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001139, Sample Num: 18224, Cur Loss: 0.96908498, Cur Avg Loss: 0.96149900, Log Avg loss: 1.12688879, Global Avg Loss: 3.29972105, Time: 0.0098 Steps: 20300, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001149, Sample Num: 18384, Cur Loss: 0.73689961, Cur Avg Loss: 0.96022695, Log Avg loss: 0.81534023, Global Avg Loss: 3.29849782, Time: 0.0066 Steps: 20310, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001159, Sample Num: 18544, Cur Loss: 0.80473244, Cur Avg Loss: 0.95880525, Log Avg loss: 0.79545195, Global Avg Loss: 3.29726601, Time: 0.0084 Steps: 20320, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001169, Sample Num: 18704, Cur Loss: 0.27463615, Cur Avg Loss: 0.95834422, Log Avg loss: 0.90491097, Global Avg Loss: 3.29608925, Time: 0.0066 Steps: 20330, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001179, Sample Num: 18864, Cur Loss: 1.24526525, Cur Avg Loss: 0.95892638, Log Avg loss: 1.02698098, Global Avg Loss: 3.29497366, Time: 0.0066 Steps: 20340, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001189, Sample Num: 19024, Cur Loss: 1.78219295, Cur Avg Loss: 0.96156658, Log Avg loss: 1.27284577, Global Avg Loss: 3.29397998, Time: 0.0127 Steps: 20350, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001199, Sample Num: 19184, Cur Loss: 0.75408787, Cur Avg Loss: 0.96019706, Log Avg loss: 0.79736182, Global Avg Loss: 3.29275375, Time: 0.0072 Steps: 20360, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001209, Sample Num: 19344, Cur Loss: 1.64900792, Cur Avg Loss: 0.96009258, Log Avg loss: 0.94756513, Global Avg Loss: 3.29160245, Time: 0.0133 Steps: 20370, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001219, Sample Num: 19504, Cur Loss: 1.27913475, Cur Avg Loss: 0.96136649, Log Avg loss: 1.11538211, Global Avg Loss: 3.29053463, Time: 0.0132 Steps: 20380, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001229, Sample Num: 19664, Cur Loss: 0.54000717, Cur Avg Loss: 0.96125288, Log Avg loss: 0.94740359, Global Avg Loss: 3.28938547, Time: 0.0066 Steps: 20390, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001239, Sample Num: 19824, Cur Loss: 1.00728130, Cur Avg Loss: 0.96230788, Log Avg loss: 1.09196802, Global Avg Loss: 3.28830831, Time: 0.0066 Steps: 20400, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001249, Sample Num: 19984, Cur Loss: 0.52093279, Cur Avg Loss: 0.96128471, Log Avg loss: 0.83451417, Global Avg Loss: 3.28710606, Time: 0.0107 Steps: 20410, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001259, Sample Num: 20144, Cur Loss: 0.63424826, Cur Avg Loss: 0.96239515, Log Avg loss: 1.10108834, Global Avg Loss: 3.28603553, Time: 0.0108 Steps: 20420, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001269, Sample Num: 20304, Cur Loss: 0.49116704, Cur Avg Loss: 0.96225121, Log Avg loss: 0.94412937, Global Avg Loss: 3.28488922, Time: 0.0107 Steps: 20430, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001279, Sample Num: 20464, Cur Loss: 1.03117776, Cur Avg Loss: 0.96318017, Log Avg loss: 1.08106551, Global Avg Loss: 3.28381103, Time: 0.0064 Steps: 20440, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001289, Sample Num: 20624, Cur Loss: 1.30034852, Cur Avg Loss: 0.96224821, Log Avg loss: 0.84305026, Global Avg Loss: 3.28261750, Time: 0.0091 Steps: 20450, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001299, Sample Num: 20784, Cur Loss: 0.30111626, Cur Avg Loss: 0.96255716, Log Avg loss: 1.00238138, Global Avg Loss: 3.28150302, Time: 0.0094 Steps: 20460, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001309, Sample Num: 20944, Cur Loss: 1.38675976, Cur Avg Loss: 0.96277594, Log Avg loss: 0.99119511, Global Avg Loss: 3.28038416, Time: 0.0114 Steps: 20470, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001319, Sample Num: 21104, Cur Loss: 0.21657535, Cur Avg Loss: 0.96115233, Log Avg loss: 0.74862178, Global Avg Loss: 3.27914794, Time: 0.0110 Steps: 20480, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001329, Sample Num: 21264, Cur Loss: 0.34050706, Cur Avg Loss: 0.96407111, Log Avg loss: 1.34905824, Global Avg Loss: 3.27820598, Time: 0.0088 Steps: 20490, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001339, Sample Num: 21424, Cur Loss: 1.11324596, Cur Avg Loss: 0.96376502, Log Avg loss: 0.92308586, Global Avg Loss: 3.27705714, Time: 0.0117 Steps: 20500, Updated lr: 0.000082 Training, Epoch: 0010, Batch: 001349, Sample Num: 21584, Cur Loss: 1.16238594, Cur Avg Loss: 0.96437067, Log Avg loss: 1.04546649, Global Avg Loss: 3.27596909, Time: 0.0094 Steps: 20510, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001359, Sample Num: 21744, Cur Loss: 1.83851027, Cur Avg Loss: 0.96356329, Log Avg loss: 0.85464768, Global Avg Loss: 3.27478911, Time: 0.0111 Steps: 20520, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001369, Sample Num: 21904, Cur Loss: 0.83722353, Cur Avg Loss: 0.96373487, Log Avg loss: 0.98705274, Global Avg Loss: 3.27367477, Time: 0.0116 Steps: 20530, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001379, Sample Num: 22064, Cur Loss: 0.91021574, Cur Avg Loss: 0.96393549, Log Avg loss: 0.99140113, Global Avg Loss: 3.27256363, Time: 0.0122 Steps: 20540, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001389, Sample Num: 22224, Cur Loss: 0.27408767, Cur Avg Loss: 0.96356156, Log Avg loss: 0.91199580, Global Avg Loss: 3.27141494, Time: 0.0119 Steps: 20550, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001399, Sample Num: 22384, Cur Loss: 2.86902142, Cur Avg Loss: 0.96427801, Log Avg loss: 1.06379302, Global Avg Loss: 3.27034119, Time: 0.0073 Steps: 20560, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001409, Sample Num: 22544, Cur Loss: 1.63526273, Cur Avg Loss: 0.96472138, Log Avg loss: 1.02674925, Global Avg Loss: 3.26925048, Time: 0.0111 Steps: 20570, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001419, Sample Num: 22704, Cur Loss: 0.72783208, Cur Avg Loss: 0.96327227, Log Avg loss: 0.75909225, Global Avg Loss: 3.26803077, Time: 0.0167 Steps: 20580, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001429, Sample Num: 22864, Cur Loss: 1.81100631, Cur Avg Loss: 0.96393652, Log Avg loss: 1.05819361, Global Avg Loss: 3.26695752, Time: 0.0113 Steps: 20590, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001439, Sample Num: 23024, Cur Loss: 0.90666330, Cur Avg Loss: 0.96298443, Log Avg loss: 0.82693063, Global Avg Loss: 3.26577304, Time: 0.0066 Steps: 20600, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001449, Sample Num: 23184, Cur Loss: 0.83779478, Cur Avg Loss: 0.96282036, Log Avg loss: 0.93921080, Global Avg Loss: 3.26464419, Time: 0.0109 Steps: 20610, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001459, Sample Num: 23344, Cur Loss: 0.47675890, Cur Avg Loss: 0.96202240, Log Avg loss: 0.84639895, Global Avg Loss: 3.26347142, Time: 0.0126 Steps: 20620, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001469, Sample Num: 23504, Cur Loss: 0.85770768, Cur Avg Loss: 0.96121598, Log Avg loss: 0.84355856, Global Avg Loss: 3.26229841, Time: 0.0065 Steps: 20630, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001479, Sample Num: 23664, Cur Loss: 1.32468092, Cur Avg Loss: 0.96037378, Log Avg loss: 0.83665429, Global Avg Loss: 3.26112320, Time: 0.0230 Steps: 20640, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001489, Sample Num: 23824, Cur Loss: 0.30862027, Cur Avg Loss: 0.95921809, Log Avg loss: 0.78829259, Global Avg Loss: 3.25992570, Time: 0.0119 Steps: 20650, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001499, Sample Num: 23984, Cur Loss: 2.10684323, Cur Avg Loss: 0.95935591, Log Avg loss: 0.97987688, Global Avg Loss: 3.25882210, Time: 0.0099 Steps: 20660, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001509, Sample Num: 24144, Cur Loss: 0.63525558, Cur Avg Loss: 0.96089504, Log Avg loss: 1.19160966, Global Avg Loss: 3.25782199, Time: 0.0068 Steps: 20670, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001519, Sample Num: 24304, Cur Loss: 0.68901509, Cur Avg Loss: 0.96142796, Log Avg loss: 1.04184598, Global Avg Loss: 3.25675044, Time: 0.0097 Steps: 20680, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001529, Sample Num: 24464, Cur Loss: 0.85407591, Cur Avg Loss: 0.96196808, Log Avg loss: 1.04401300, Global Avg Loss: 3.25568097, Time: 0.0238 Steps: 20690, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001539, Sample Num: 24624, Cur Loss: 0.94874060, Cur Avg Loss: 0.96341172, Log Avg loss: 1.18414420, Global Avg Loss: 3.25468022, Time: 0.0096 Steps: 20700, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001549, Sample Num: 24784, Cur Loss: 1.38257957, Cur Avg Loss: 0.96386198, Log Avg loss: 1.03315621, Global Avg Loss: 3.25360754, Time: 0.0067 Steps: 20710, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001559, Sample Num: 24944, Cur Loss: 0.58613306, Cur Avg Loss: 0.96340419, Log Avg loss: 0.89249274, Global Avg Loss: 3.25246801, Time: 0.0080 Steps: 20720, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001569, Sample Num: 25104, Cur Loss: 0.40613091, Cur Avg Loss: 0.96240592, Log Avg loss: 0.80677650, Global Avg Loss: 3.25128822, Time: 0.0118 Steps: 20730, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001579, Sample Num: 25264, Cur Loss: 0.82729149, Cur Avg Loss: 0.96292486, Log Avg loss: 1.04434549, Global Avg Loss: 3.25022412, Time: 0.0088 Steps: 20740, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001589, Sample Num: 25424, Cur Loss: 1.42084849, Cur Avg Loss: 0.96320278, Log Avg loss: 1.00708675, Global Avg Loss: 3.24914309, Time: 0.0106 Steps: 20750, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001599, Sample Num: 25584, Cur Loss: 0.80973971, Cur Avg Loss: 0.96298436, Log Avg loss: 0.92827761, Global Avg Loss: 3.24802514, Time: 0.0119 Steps: 20760, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001609, Sample Num: 25744, Cur Loss: 1.27093053, Cur Avg Loss: 0.96251604, Log Avg loss: 0.88763096, Global Avg Loss: 3.24688870, Time: 0.0117 Steps: 20770, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001619, Sample Num: 25904, Cur Loss: 1.43424511, Cur Avg Loss: 0.96331035, Log Avg loss: 1.09111518, Global Avg Loss: 3.24585127, Time: 0.0114 Steps: 20780, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001629, Sample Num: 26064, Cur Loss: 1.87544847, Cur Avg Loss: 0.96375552, Log Avg loss: 1.03582883, Global Avg Loss: 3.24478825, Time: 0.0130 Steps: 20790, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001639, Sample Num: 26224, Cur Loss: 1.39489961, Cur Avg Loss: 0.96492243, Log Avg loss: 1.15501153, Global Avg Loss: 3.24378355, Time: 0.0108 Steps: 20800, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001649, Sample Num: 26384, Cur Loss: 1.90059388, Cur Avg Loss: 0.96598426, Log Avg loss: 1.14001911, Global Avg Loss: 3.24277261, Time: 0.0118 Steps: 20810, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001659, Sample Num: 26544, Cur Loss: 0.75098461, Cur Avg Loss: 0.96611584, Log Avg loss: 0.98781307, Global Avg Loss: 3.24168954, Time: 0.0066 Steps: 20820, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001669, Sample Num: 26704, Cur Loss: 0.64695036, Cur Avg Loss: 0.96477314, Log Avg loss: 0.74201841, Global Avg Loss: 3.24048950, Time: 0.0117 Steps: 20830, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001679, Sample Num: 26864, Cur Loss: 0.53978777, Cur Avg Loss: 0.96357848, Log Avg loss: 0.76418959, Global Avg Loss: 3.23930126, Time: 0.0138 Steps: 20840, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001689, Sample Num: 27024, Cur Loss: 0.49864215, Cur Avg Loss: 0.96413489, Log Avg loss: 1.05755655, Global Avg Loss: 3.23825486, Time: 0.0130 Steps: 20850, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001699, Sample Num: 27184, Cur Loss: 0.39462039, Cur Avg Loss: 0.96384387, Log Avg loss: 0.91469143, Global Avg Loss: 3.23714097, Time: 0.0089 Steps: 20860, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001709, Sample Num: 27344, Cur Loss: 0.25263107, Cur Avg Loss: 0.96295623, Log Avg loss: 0.81214628, Global Avg Loss: 3.23597902, Time: 0.0122 Steps: 20870, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001719, Sample Num: 27504, Cur Loss: 0.68144482, Cur Avg Loss: 0.96242193, Log Avg loss: 0.87110897, Global Avg Loss: 3.23484642, Time: 0.0083 Steps: 20880, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001729, Sample Num: 27664, Cur Loss: 0.56256986, Cur Avg Loss: 0.96247175, Log Avg loss: 0.97103681, Global Avg Loss: 3.23376274, Time: 0.0116 Steps: 20890, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001739, Sample Num: 27824, Cur Loss: 0.47483629, Cur Avg Loss: 0.96203975, Log Avg loss: 0.88734637, Global Avg Loss: 3.23264005, Time: 0.0111 Steps: 20900, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001749, Sample Num: 27984, Cur Loss: 0.55212510, Cur Avg Loss: 0.96204302, Log Avg loss: 0.96261108, Global Avg Loss: 3.23155443, Time: 0.0104 Steps: 20910, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001759, Sample Num: 28144, Cur Loss: 0.30198085, Cur Avg Loss: 0.96093507, Log Avg loss: 0.76715541, Global Avg Loss: 3.23037642, Time: 0.0109 Steps: 20920, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001769, Sample Num: 28304, Cur Loss: 0.58128893, Cur Avg Loss: 0.96048011, Log Avg loss: 0.88045216, Global Avg Loss: 3.22925367, Time: 0.0112 Steps: 20930, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001779, Sample Num: 28464, Cur Loss: 1.06714272, Cur Avg Loss: 0.95948834, Log Avg loss: 0.78404441, Global Avg Loss: 3.22808595, Time: 0.0137 Steps: 20940, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001789, Sample Num: 28624, Cur Loss: 1.56494784, Cur Avg Loss: 0.95940295, Log Avg loss: 0.94421195, Global Avg Loss: 3.22699579, Time: 0.0121 Steps: 20950, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001799, Sample Num: 28784, Cur Loss: 0.39703289, Cur Avg Loss: 0.95912744, Log Avg loss: 0.90983983, Global Avg Loss: 3.22589028, Time: 0.0066 Steps: 20960, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001809, Sample Num: 28944, Cur Loss: 0.46324354, Cur Avg Loss: 0.95959030, Log Avg loss: 1.04285794, Global Avg Loss: 3.22484925, Time: 0.0097 Steps: 20970, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001819, Sample Num: 29104, Cur Loss: 1.22632945, Cur Avg Loss: 0.96063795, Log Avg loss: 1.15015743, Global Avg Loss: 3.22386036, Time: 0.0155 Steps: 20980, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001829, Sample Num: 29264, Cur Loss: 0.70080483, Cur Avg Loss: 0.96072317, Log Avg loss: 0.97622451, Global Avg Loss: 3.22278955, Time: 0.0091 Steps: 20990, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001839, Sample Num: 29424, Cur Loss: 1.08950603, Cur Avg Loss: 0.96091853, Log Avg loss: 0.99664995, Global Avg Loss: 3.22172948, Time: 0.0162 Steps: 21000, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001849, Sample Num: 29584, Cur Loss: 0.70198750, Cur Avg Loss: 0.96045991, Log Avg loss: 0.87612025, Global Avg Loss: 3.22061306, Time: 0.0133 Steps: 21010, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001859, Sample Num: 29744, Cur Loss: 1.12222171, Cur Avg Loss: 0.95971464, Log Avg loss: 0.82191390, Global Avg Loss: 3.21947191, Time: 0.0108 Steps: 21020, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001869, Sample Num: 29904, Cur Loss: 0.47700888, Cur Avg Loss: 0.96108767, Log Avg loss: 1.21633423, Global Avg Loss: 3.21851939, Time: 0.0066 Steps: 21030, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001879, Sample Num: 30064, Cur Loss: 0.43335003, Cur Avg Loss: 0.96118249, Log Avg loss: 0.97890418, Global Avg Loss: 3.21745494, Time: 0.0117 Steps: 21040, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001889, Sample Num: 30224, Cur Loss: 1.14004707, Cur Avg Loss: 0.96161124, Log Avg loss: 1.04217422, Global Avg Loss: 3.21642155, Time: 0.0230 Steps: 21050, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001899, Sample Num: 30384, Cur Loss: 0.86420566, Cur Avg Loss: 0.96135625, Log Avg loss: 0.91318737, Global Avg Loss: 3.21532790, Time: 0.0072 Steps: 21060, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001909, Sample Num: 30544, Cur Loss: 0.80975193, Cur Avg Loss: 0.96132103, Log Avg loss: 0.95463400, Global Avg Loss: 3.21425495, Time: 0.0114 Steps: 21070, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001919, Sample Num: 30704, Cur Loss: 0.82459795, Cur Avg Loss: 0.96127002, Log Avg loss: 0.95153128, Global Avg Loss: 3.21318155, Time: 0.0067 Steps: 21080, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001929, Sample Num: 30864, Cur Loss: 0.72449470, Cur Avg Loss: 0.96066471, Log Avg loss: 0.84450636, Global Avg Loss: 3.21205843, Time: 0.0155 Steps: 21090, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001939, Sample Num: 31024, Cur Loss: 0.59651321, Cur Avg Loss: 0.96066022, Log Avg loss: 0.95979456, Global Avg Loss: 3.21099100, Time: 0.0110 Steps: 21100, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001949, Sample Num: 31184, Cur Loss: 0.45315325, Cur Avg Loss: 0.96062229, Log Avg loss: 0.95326651, Global Avg Loss: 3.20992150, Time: 0.0067 Steps: 21110, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001959, Sample Num: 31344, Cur Loss: 0.91847134, Cur Avg Loss: 0.96095969, Log Avg loss: 1.02671963, Global Avg Loss: 3.20888779, Time: 0.0073 Steps: 21120, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001969, Sample Num: 31504, Cur Loss: 0.65745693, Cur Avg Loss: 0.96095428, Log Avg loss: 0.95989467, Global Avg Loss: 3.20782342, Time: 0.0110 Steps: 21130, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001979, Sample Num: 31664, Cur Loss: 0.97407055, Cur Avg Loss: 0.96019506, Log Avg loss: 0.81070410, Global Avg Loss: 3.20668950, Time: 0.0125 Steps: 21140, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001989, Sample Num: 31824, Cur Loss: 1.08902287, Cur Avg Loss: 0.96035748, Log Avg loss: 0.99249985, Global Avg Loss: 3.20564260, Time: 0.0067 Steps: 21150, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 001999, Sample Num: 31984, Cur Loss: 1.04268789, Cur Avg Loss: 0.96010730, Log Avg loss: 0.91034754, Global Avg Loss: 3.20455787, Time: 0.0067 Steps: 21160, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002009, Sample Num: 32144, Cur Loss: 0.67382628, Cur Avg Loss: 0.96109432, Log Avg loss: 1.15839845, Global Avg Loss: 3.20359133, Time: 0.0123 Steps: 21170, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002019, Sample Num: 32304, Cur Loss: 0.95374024, Cur Avg Loss: 0.96019377, Log Avg loss: 0.77927381, Global Avg Loss: 3.20244670, Time: 0.0113 Steps: 21180, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002029, Sample Num: 32464, Cur Loss: 0.80079526, Cur Avg Loss: 0.96111558, Log Avg loss: 1.14723012, Global Avg Loss: 3.20147681, Time: 0.0067 Steps: 21190, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002039, Sample Num: 32624, Cur Loss: 1.43346262, Cur Avg Loss: 0.96147044, Log Avg loss: 1.03347074, Global Avg Loss: 3.20045416, Time: 0.0123 Steps: 21200, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002049, Sample Num: 32784, Cur Loss: 0.64953226, Cur Avg Loss: 0.96102167, Log Avg loss: 0.86951714, Global Avg Loss: 3.19935518, Time: 0.0143 Steps: 21210, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002059, Sample Num: 32944, Cur Loss: 0.32778922, Cur Avg Loss: 0.96055570, Log Avg loss: 0.86507853, Global Avg Loss: 3.19825514, Time: 0.0127 Steps: 21220, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002069, Sample Num: 33104, Cur Loss: 1.04331625, Cur Avg Loss: 0.96044854, Log Avg loss: 0.93838517, Global Avg Loss: 3.19719067, Time: 0.0115 Steps: 21230, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002079, Sample Num: 33264, Cur Loss: 1.05543256, Cur Avg Loss: 0.96089514, Log Avg loss: 1.05329565, Global Avg Loss: 3.19618131, Time: 0.0067 Steps: 21240, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002089, Sample Num: 33424, Cur Loss: 0.38814622, Cur Avg Loss: 0.96023927, Log Avg loss: 0.82388335, Global Avg Loss: 3.19506493, Time: 0.0117 Steps: 21250, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002099, Sample Num: 33584, Cur Loss: 0.67995471, Cur Avg Loss: 0.95914551, Log Avg loss: 0.73065895, Global Avg Loss: 3.19390576, Time: 0.0069 Steps: 21260, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002109, Sample Num: 33744, Cur Loss: 0.79628778, Cur Avg Loss: 0.95923973, Log Avg loss: 0.97901708, Global Avg Loss: 3.19286444, Time: 0.0110 Steps: 21270, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002119, Sample Num: 33904, Cur Loss: 1.93937445, Cur Avg Loss: 0.95854266, Log Avg loss: 0.81153193, Global Avg Loss: 3.19174539, Time: 0.0111 Steps: 21280, Updated lr: 0.000081 Training, Epoch: 0010, Batch: 002129, Sample Num: 34055, Cur Loss: 0.51534903, Cur Avg Loss: 0.95839129, Log Avg loss: 0.92631420, Global Avg Loss: 3.19068131, Time: 0.0039 Steps: 21290, Updated lr: 0.000081 ***** Running evaluation checkpoint-21290 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-21290 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.356856, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.082695, "eval_total_loss": 761.134809, "eval_mae": 0.884113, "eval_mse": 1.082744, "eval_r2": 0.311737, "eval_sp_statistic": 0.58419, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.625434, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 2.536853, "test_total_loss": 1273.50026, "test_mae": 1.472086, "test_mse": 2.536451, "test_r2": -0.637048, "test_sp_statistic": 0.32678, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.470968, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 3.190681307327515, "train_cur_epoch_loss": 2040.4150478243828, "train_cur_epoch_avg_loss": 0.9583912859673005, "train_cur_epoch_time": 22.356855869293213, "train_cur_epoch_avg_time": 0.010501106561434107, "epoch": 10, "step": 21290} ################################################## Training, Epoch: 0011, Batch: 000010, Sample Num: 160, Cur Loss: 1.29180968, Cur Avg Loss: 1.06762019, Log Avg loss: 1.06762019, Global Avg Loss: 3.18968457, Time: 0.0121 Steps: 21300, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000020, Sample Num: 320, Cur Loss: 0.63068122, Cur Avg Loss: 0.97293240, Log Avg loss: 0.87824461, Global Avg Loss: 3.18859989, Time: 0.0113 Steps: 21310, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000030, Sample Num: 480, Cur Loss: 0.58608204, Cur Avg Loss: 0.98361618, Log Avg loss: 1.00498373, Global Avg Loss: 3.18757568, Time: 0.0114 Steps: 21320, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000040, Sample Num: 640, Cur Loss: 0.45368302, Cur Avg Loss: 0.96480192, Log Avg loss: 0.90835916, Global Avg Loss: 3.18650713, Time: 0.0078 Steps: 21330, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000050, Sample Num: 800, Cur Loss: 0.38984281, Cur Avg Loss: 0.93978443, Log Avg loss: 0.83971444, Global Avg Loss: 3.18540742, Time: 0.0072 Steps: 21340, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000060, Sample Num: 960, Cur Loss: 0.74960536, Cur Avg Loss: 0.94949001, Log Avg loss: 0.99801790, Global Avg Loss: 3.18438288, Time: 0.0131 Steps: 21350, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000070, Sample Num: 1120, Cur Loss: 1.25922751, Cur Avg Loss: 0.94598654, Log Avg loss: 0.92496576, Global Avg Loss: 3.18332510, Time: 0.0121 Steps: 21360, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000080, Sample Num: 1280, Cur Loss: 1.28264391, Cur Avg Loss: 0.94348804, Log Avg loss: 0.92599854, Global Avg Loss: 3.18226879, Time: 0.0224 Steps: 21370, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000090, Sample Num: 1440, Cur Loss: 0.72192699, Cur Avg Loss: 0.96292199, Log Avg loss: 1.11839362, Global Avg Loss: 3.18130346, Time: 0.0136 Steps: 21380, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000100, Sample Num: 1600, Cur Loss: 0.88094163, Cur Avg Loss: 0.96279021, Log Avg loss: 0.96160413, Global Avg Loss: 3.18026573, Time: 0.0071 Steps: 21390, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000110, Sample Num: 1760, Cur Loss: 1.07860529, Cur Avg Loss: 0.94076279, Log Avg loss: 0.72048858, Global Avg Loss: 3.17911631, Time: 0.0066 Steps: 21400, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000120, Sample Num: 1920, Cur Loss: 0.87213254, Cur Avg Loss: 0.95215674, Log Avg loss: 1.07749025, Global Avg Loss: 3.17813470, Time: 0.0114 Steps: 21410, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000130, Sample Num: 2080, Cur Loss: 1.57434523, Cur Avg Loss: 0.95532756, Log Avg loss: 0.99337743, Global Avg Loss: 3.17711473, Time: 0.0115 Steps: 21420, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000140, Sample Num: 2240, Cur Loss: 0.60002488, Cur Avg Loss: 0.95685345, Log Avg loss: 0.97668994, Global Avg Loss: 3.17608794, Time: 0.0113 Steps: 21430, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000150, Sample Num: 2400, Cur Loss: 0.81271261, Cur Avg Loss: 0.95308853, Log Avg loss: 0.90037975, Global Avg Loss: 3.17502651, Time: 0.0136 Steps: 21440, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000160, Sample Num: 2560, Cur Loss: 1.04418826, Cur Avg Loss: 0.95693810, Log Avg loss: 1.01468159, Global Avg Loss: 3.17401935, Time: 0.0071 Steps: 21450, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000170, Sample Num: 2720, Cur Loss: 0.81569016, Cur Avg Loss: 0.94324921, Log Avg loss: 0.72422701, Global Avg Loss: 3.17287779, Time: 0.0160 Steps: 21460, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000180, Sample Num: 2880, Cur Loss: 0.88609147, Cur Avg Loss: 0.94612780, Log Avg loss: 0.99506373, Global Avg Loss: 3.17186344, Time: 0.0131 Steps: 21470, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000190, Sample Num: 3040, Cur Loss: 1.55654883, Cur Avg Loss: 0.95481716, Log Avg loss: 1.11122569, Global Avg Loss: 3.17090411, Time: 0.0067 Steps: 21480, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000200, Sample Num: 3200, Cur Loss: 0.29335210, Cur Avg Loss: 0.95401046, Log Avg loss: 0.93868316, Global Avg Loss: 3.16986539, Time: 0.0065 Steps: 21490, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000210, Sample Num: 3360, Cur Loss: 0.88873255, Cur Avg Loss: 0.95214740, Log Avg loss: 0.91488624, Global Avg Loss: 3.16881656, Time: 0.0065 Steps: 21500, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000220, Sample Num: 3520, Cur Loss: 1.34251237, Cur Avg Loss: 0.95362465, Log Avg loss: 0.98464694, Global Avg Loss: 3.16780114, Time: 0.0180 Steps: 21510, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000230, Sample Num: 3680, Cur Loss: 0.88231826, Cur Avg Loss: 0.95167509, Log Avg loss: 0.90878459, Global Avg Loss: 3.16675141, Time: 0.0074 Steps: 21520, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000240, Sample Num: 3840, Cur Loss: 0.52936399, Cur Avg Loss: 0.94697271, Log Avg loss: 0.83881806, Global Avg Loss: 3.16567016, Time: 0.0064 Steps: 21530, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000250, Sample Num: 4000, Cur Loss: 0.46543437, Cur Avg Loss: 0.94769312, Log Avg loss: 0.96498291, Global Avg Loss: 3.16464848, Time: 0.0065 Steps: 21540, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000260, Sample Num: 4160, Cur Loss: 0.64183837, Cur Avg Loss: 0.94884776, Log Avg loss: 0.97771381, Global Avg Loss: 3.16363366, Time: 0.0064 Steps: 21550, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000270, Sample Num: 4320, Cur Loss: 0.36248651, Cur Avg Loss: 0.94379729, Log Avg loss: 0.81248518, Global Avg Loss: 3.16254315, Time: 0.0066 Steps: 21560, Updated lr: 0.000081 Training, Epoch: 0011, Batch: 000280, Sample Num: 4480, Cur Loss: 0.46160841, Cur Avg Loss: 0.94836931, Log Avg loss: 1.07181389, Global Avg Loss: 3.16157387, Time: 0.0067 Steps: 21570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000290, Sample Num: 4640, Cur Loss: 1.02729034, Cur Avg Loss: 0.94301202, Log Avg loss: 0.79300779, Global Avg Loss: 3.16047630, Time: 0.0070 Steps: 21580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000300, Sample Num: 4800, Cur Loss: 0.61018395, Cur Avg Loss: 0.94919576, Log Avg loss: 1.12852416, Global Avg Loss: 3.15953514, Time: 0.0154 Steps: 21590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000310, Sample Num: 4960, Cur Loss: 0.64552188, Cur Avg Loss: 0.95658814, Log Avg loss: 1.17835943, Global Avg Loss: 3.15861793, Time: 0.0114 Steps: 21600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000320, Sample Num: 5120, Cur Loss: 0.69251966, Cur Avg Loss: 0.96000250, Log Avg loss: 1.06584787, Global Avg Loss: 3.15764951, Time: 0.0117 Steps: 21610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000330, Sample Num: 5280, Cur Loss: 1.43339241, Cur Avg Loss: 0.95816485, Log Avg loss: 0.89936003, Global Avg Loss: 3.15660497, Time: 0.0068 Steps: 21620, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000340, Sample Num: 5440, Cur Loss: 0.57609868, Cur Avg Loss: 0.95676023, Log Avg loss: 0.91040766, Global Avg Loss: 3.15556651, Time: 0.0068 Steps: 21630, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000350, Sample Num: 5600, Cur Loss: 0.56185865, Cur Avg Loss: 0.95665408, Log Avg loss: 0.95304501, Global Avg Loss: 3.15454870, Time: 0.0068 Steps: 21640, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000360, Sample Num: 5760, Cur Loss: 0.64884079, Cur Avg Loss: 0.94818151, Log Avg loss: 0.65164173, Global Avg Loss: 3.15339263, Time: 0.0152 Steps: 21650, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000370, Sample Num: 5920, Cur Loss: 1.12122822, Cur Avg Loss: 0.94728396, Log Avg loss: 0.91497199, Global Avg Loss: 3.15235919, Time: 0.0111 Steps: 21660, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000380, Sample Num: 6080, Cur Loss: 0.97922969, Cur Avg Loss: 0.94230470, Log Avg loss: 0.75807212, Global Avg Loss: 3.15125431, Time: 0.0111 Steps: 21670, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000390, Sample Num: 6240, Cur Loss: 1.60896587, Cur Avg Loss: 0.93956609, Log Avg loss: 0.83549888, Global Avg Loss: 3.15018615, Time: 0.0072 Steps: 21680, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000400, Sample Num: 6400, Cur Loss: 1.02312732, Cur Avg Loss: 0.93856766, Log Avg loss: 0.89962886, Global Avg Loss: 3.14914855, Time: 0.0066 Steps: 21690, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000410, Sample Num: 6560, Cur Loss: 0.63915014, Cur Avg Loss: 0.93989549, Log Avg loss: 0.99300877, Global Avg Loss: 3.14815494, Time: 0.0132 Steps: 21700, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000420, Sample Num: 6720, Cur Loss: 1.07023799, Cur Avg Loss: 0.94403727, Log Avg loss: 1.11385003, Global Avg Loss: 3.14721790, Time: 0.0129 Steps: 21710, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000430, Sample Num: 6880, Cur Loss: 1.48405969, Cur Avg Loss: 0.94175533, Log Avg loss: 0.84591407, Global Avg Loss: 3.14615837, Time: 0.0134 Steps: 21720, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000440, Sample Num: 7040, Cur Loss: 0.65773034, Cur Avg Loss: 0.93920415, Log Avg loss: 0.82950348, Global Avg Loss: 3.14509226, Time: 0.0107 Steps: 21730, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000450, Sample Num: 7200, Cur Loss: 1.61324167, Cur Avg Loss: 0.93833512, Log Avg loss: 0.90009750, Global Avg Loss: 3.14405961, Time: 0.0117 Steps: 21740, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000460, Sample Num: 7360, Cur Loss: 0.98232931, Cur Avg Loss: 0.93668718, Log Avg loss: 0.86253023, Global Avg Loss: 3.14301063, Time: 0.0118 Steps: 21750, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000470, Sample Num: 7520, Cur Loss: 0.40043128, Cur Avg Loss: 0.93723647, Log Avg loss: 0.96250357, Global Avg Loss: 3.14200856, Time: 0.0116 Steps: 21760, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000480, Sample Num: 7680, Cur Loss: 0.74793357, Cur Avg Loss: 0.93744024, Log Avg loss: 0.94701743, Global Avg Loss: 3.14100029, Time: 0.0068 Steps: 21770, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000490, Sample Num: 7840, Cur Loss: 0.98029417, Cur Avg Loss: 0.93444278, Log Avg loss: 0.79056484, Global Avg Loss: 3.13992112, Time: 0.0129 Steps: 21780, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000500, Sample Num: 8000, Cur Loss: 1.02816916, Cur Avg Loss: 0.93580710, Log Avg loss: 1.00265892, Global Avg Loss: 3.13894027, Time: 0.0109 Steps: 21790, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000510, Sample Num: 8160, Cur Loss: 0.52054453, Cur Avg Loss: 0.93841750, Log Avg loss: 1.06893748, Global Avg Loss: 3.13799073, Time: 0.0117 Steps: 21800, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000520, Sample Num: 8320, Cur Loss: 0.48772591, Cur Avg Loss: 0.94208475, Log Avg loss: 1.12911452, Global Avg Loss: 3.13706965, Time: 0.0069 Steps: 21810, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000530, Sample Num: 8480, Cur Loss: 0.94447690, Cur Avg Loss: 0.93933214, Log Avg loss: 0.79619642, Global Avg Loss: 3.13599684, Time: 0.0068 Steps: 21820, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000540, Sample Num: 8640, Cur Loss: 1.34041977, Cur Avg Loss: 0.93808482, Log Avg loss: 0.87197651, Global Avg Loss: 3.13495973, Time: 0.0113 Steps: 21830, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000550, Sample Num: 8800, Cur Loss: 0.94475830, Cur Avg Loss: 0.93451237, Log Avg loss: 0.74160017, Global Avg Loss: 3.13386387, Time: 0.0117 Steps: 21840, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000560, Sample Num: 8960, Cur Loss: 1.53401721, Cur Avg Loss: 0.93294461, Log Avg loss: 0.84671776, Global Avg Loss: 3.13281712, Time: 0.0153 Steps: 21850, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000570, Sample Num: 9120, Cur Loss: 2.77806473, Cur Avg Loss: 0.93629107, Log Avg loss: 1.12369267, Global Avg Loss: 3.13189803, Time: 0.0068 Steps: 21860, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000580, Sample Num: 9280, Cur Loss: 1.30515635, Cur Avg Loss: 0.93418361, Log Avg loss: 0.81405873, Global Avg Loss: 3.13083820, Time: 0.0076 Steps: 21870, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000590, Sample Num: 9440, Cur Loss: 0.59888572, Cur Avg Loss: 0.93478491, Log Avg loss: 0.96966020, Global Avg Loss: 3.12985046, Time: 0.0068 Steps: 21880, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000600, Sample Num: 9600, Cur Loss: 0.61456251, Cur Avg Loss: 0.93456284, Log Avg loss: 0.92146077, Global Avg Loss: 3.12884161, Time: 0.0106 Steps: 21890, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000610, Sample Num: 9760, Cur Loss: 0.96771485, Cur Avg Loss: 0.93744679, Log Avg loss: 1.11048391, Global Avg Loss: 3.12791998, Time: 0.0068 Steps: 21900, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000620, Sample Num: 9920, Cur Loss: 1.01256716, Cur Avg Loss: 0.93735392, Log Avg loss: 0.93168840, Global Avg Loss: 3.12691759, Time: 0.0153 Steps: 21910, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000630, Sample Num: 10080, Cur Loss: 1.48072314, Cur Avg Loss: 0.93787637, Log Avg loss: 0.97026829, Global Avg Loss: 3.12593372, Time: 0.0128 Steps: 21920, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000640, Sample Num: 10240, Cur Loss: 0.78236216, Cur Avg Loss: 0.93175505, Log Avg loss: 0.54611230, Global Avg Loss: 3.12475733, Time: 0.0069 Steps: 21930, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000650, Sample Num: 10400, Cur Loss: 0.45665768, Cur Avg Loss: 0.92746748, Log Avg loss: 0.65306319, Global Avg Loss: 3.12363076, Time: 0.0073 Steps: 21940, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000660, Sample Num: 10560, Cur Loss: 0.58314162, Cur Avg Loss: 0.92665369, Log Avg loss: 0.87375681, Global Avg Loss: 3.12260576, Time: 0.0068 Steps: 21950, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000670, Sample Num: 10720, Cur Loss: 0.93707597, Cur Avg Loss: 0.92563764, Log Avg loss: 0.85857884, Global Avg Loss: 3.12157478, Time: 0.0132 Steps: 21960, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000680, Sample Num: 10880, Cur Loss: 0.63516098, Cur Avg Loss: 0.92278631, Log Avg loss: 0.73174658, Global Avg Loss: 3.12048702, Time: 0.0122 Steps: 21970, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000690, Sample Num: 11040, Cur Loss: 0.49538058, Cur Avg Loss: 0.92266489, Log Avg loss: 0.91440893, Global Avg Loss: 3.11948334, Time: 0.0082 Steps: 21980, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000700, Sample Num: 11200, Cur Loss: 0.66403866, Cur Avg Loss: 0.92787224, Log Avg loss: 1.28717940, Global Avg Loss: 3.11865010, Time: 0.0119 Steps: 21990, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000710, Sample Num: 11360, Cur Loss: 0.63816798, Cur Avg Loss: 0.92486030, Log Avg loss: 0.71402399, Global Avg Loss: 3.11755708, Time: 0.0082 Steps: 22000, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000720, Sample Num: 11520, Cur Loss: 0.56915450, Cur Avg Loss: 0.92093841, Log Avg loss: 0.64248426, Global Avg Loss: 3.11643256, Time: 0.0120 Steps: 22010, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000730, Sample Num: 11680, Cur Loss: 1.34611082, Cur Avg Loss: 0.91987666, Log Avg loss: 0.84343060, Global Avg Loss: 3.11540032, Time: 0.0068 Steps: 22020, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000740, Sample Num: 11840, Cur Loss: 0.88095725, Cur Avg Loss: 0.92195730, Log Avg loss: 1.07384429, Global Avg Loss: 3.11447360, Time: 0.0068 Steps: 22030, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000750, Sample Num: 12000, Cur Loss: 2.30039763, Cur Avg Loss: 0.92153173, Log Avg loss: 0.89003961, Global Avg Loss: 3.11346433, Time: 0.0068 Steps: 22040, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000760, Sample Num: 12160, Cur Loss: 0.71471328, Cur Avg Loss: 0.91967473, Log Avg loss: 0.78039970, Global Avg Loss: 3.11240625, Time: 0.0068 Steps: 22050, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000770, Sample Num: 12320, Cur Loss: 0.55094993, Cur Avg Loss: 0.92067071, Log Avg loss: 0.99636486, Global Avg Loss: 3.11144703, Time: 0.0105 Steps: 22060, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000780, Sample Num: 12480, Cur Loss: 1.48846674, Cur Avg Loss: 0.91935220, Log Avg loss: 0.81782758, Global Avg Loss: 3.11040778, Time: 0.0136 Steps: 22070, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000790, Sample Num: 12640, Cur Loss: 1.22228026, Cur Avg Loss: 0.92061380, Log Avg loss: 1.01901860, Global Avg Loss: 3.10946060, Time: 0.0068 Steps: 22080, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000800, Sample Num: 12800, Cur Loss: 0.99218947, Cur Avg Loss: 0.91754720, Log Avg loss: 0.67528536, Global Avg Loss: 3.10835866, Time: 0.0067 Steps: 22090, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000810, Sample Num: 12960, Cur Loss: 0.79754436, Cur Avg Loss: 0.91839636, Log Avg loss: 0.98632897, Global Avg Loss: 3.10739847, Time: 0.0068 Steps: 22100, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000820, Sample Num: 13120, Cur Loss: 0.84114486, Cur Avg Loss: 0.91927472, Log Avg loss: 0.99042229, Global Avg Loss: 3.10644099, Time: 0.0070 Steps: 22110, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000830, Sample Num: 13280, Cur Loss: 0.59675491, Cur Avg Loss: 0.91935959, Log Avg loss: 0.92631887, Global Avg Loss: 3.10545540, Time: 0.0067 Steps: 22120, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000840, Sample Num: 13440, Cur Loss: 0.91207057, Cur Avg Loss: 0.91896772, Log Avg loss: 0.88644239, Global Avg Loss: 3.10445268, Time: 0.0124 Steps: 22130, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000850, Sample Num: 13600, Cur Loss: 1.66920471, Cur Avg Loss: 0.92081654, Log Avg loss: 1.07611765, Global Avg Loss: 3.10353654, Time: 0.0114 Steps: 22140, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000860, Sample Num: 13760, Cur Loss: 0.68446934, Cur Avg Loss: 0.92156033, Log Avg loss: 0.98478249, Global Avg Loss: 3.10258000, Time: 0.0114 Steps: 22150, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000870, Sample Num: 13920, Cur Loss: 2.07702589, Cur Avg Loss: 0.92463390, Log Avg loss: 1.18896107, Global Avg Loss: 3.10171645, Time: 0.0219 Steps: 22160, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000880, Sample Num: 14080, Cur Loss: 1.60793722, Cur Avg Loss: 0.92396630, Log Avg loss: 0.86588478, Global Avg Loss: 3.10070796, Time: 0.0119 Steps: 22170, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000890, Sample Num: 14240, Cur Loss: 1.13917994, Cur Avg Loss: 0.92669759, Log Avg loss: 1.16705095, Global Avg Loss: 3.09983615, Time: 0.0112 Steps: 22180, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000900, Sample Num: 14400, Cur Loss: 0.73506927, Cur Avg Loss: 0.92835575, Log Avg loss: 1.07593216, Global Avg Loss: 3.09892407, Time: 0.0234 Steps: 22190, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000910, Sample Num: 14560, Cur Loss: 0.89916229, Cur Avg Loss: 0.92676929, Log Avg loss: 0.78398807, Global Avg Loss: 3.09788131, Time: 0.0116 Steps: 22200, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000920, Sample Num: 14720, Cur Loss: 1.53433609, Cur Avg Loss: 0.92410148, Log Avg loss: 0.68133015, Global Avg Loss: 3.09679326, Time: 0.0072 Steps: 22210, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000930, Sample Num: 14880, Cur Loss: 0.78190506, Cur Avg Loss: 0.92457795, Log Avg loss: 0.96841323, Global Avg Loss: 3.09583540, Time: 0.0114 Steps: 22220, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000940, Sample Num: 15040, Cur Loss: 0.83056647, Cur Avg Loss: 0.92578418, Log Avg loss: 1.03796370, Global Avg Loss: 3.09490968, Time: 0.0079 Steps: 22230, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000950, Sample Num: 15200, Cur Loss: 0.61288261, Cur Avg Loss: 0.92653805, Log Avg loss: 0.99740188, Global Avg Loss: 3.09396655, Time: 0.0136 Steps: 22240, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000960, Sample Num: 15360, Cur Loss: 1.20829570, Cur Avg Loss: 0.92678278, Log Avg loss: 0.95003254, Global Avg Loss: 3.09300299, Time: 0.0155 Steps: 22250, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000970, Sample Num: 15520, Cur Loss: 0.97545803, Cur Avg Loss: 0.92466159, Log Avg loss: 0.72102745, Global Avg Loss: 3.09193741, Time: 0.0072 Steps: 22260, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000980, Sample Num: 15680, Cur Loss: 0.89584225, Cur Avg Loss: 0.92636869, Log Avg loss: 1.09195725, Global Avg Loss: 3.09103935, Time: 0.0151 Steps: 22270, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 000990, Sample Num: 15840, Cur Loss: 1.50051427, Cur Avg Loss: 0.92953860, Log Avg loss: 1.24018936, Global Avg Loss: 3.09020863, Time: 0.0068 Steps: 22280, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001000, Sample Num: 16000, Cur Loss: 1.06092465, Cur Avg Loss: 0.92773354, Log Avg loss: 0.74903295, Global Avg Loss: 3.08915830, Time: 0.0068 Steps: 22290, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001010, Sample Num: 16160, Cur Loss: 1.14144897, Cur Avg Loss: 0.92870916, Log Avg loss: 1.02627065, Global Avg Loss: 3.08823324, Time: 0.0068 Steps: 22300, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001020, Sample Num: 16320, Cur Loss: 0.83447456, Cur Avg Loss: 0.92930065, Log Avg loss: 0.98904141, Global Avg Loss: 3.08729232, Time: 0.0158 Steps: 22310, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001030, Sample Num: 16480, Cur Loss: 1.51594055, Cur Avg Loss: 0.92949578, Log Avg loss: 0.94939938, Global Avg Loss: 3.08633448, Time: 0.0067 Steps: 22320, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001040, Sample Num: 16640, Cur Loss: 0.92742765, Cur Avg Loss: 0.92800186, Log Avg loss: 0.77412830, Global Avg Loss: 3.08529901, Time: 0.0069 Steps: 22330, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001050, Sample Num: 16800, Cur Loss: 0.57436740, Cur Avg Loss: 0.92629726, Log Avg loss: 0.74901801, Global Avg Loss: 3.08425323, Time: 0.0086 Steps: 22340, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001060, Sample Num: 16960, Cur Loss: 0.80560672, Cur Avg Loss: 0.92657640, Log Avg loss: 0.95588706, Global Avg Loss: 3.08330094, Time: 0.0156 Steps: 22350, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001070, Sample Num: 17120, Cur Loss: 0.43556106, Cur Avg Loss: 0.92698171, Log Avg loss: 0.96994389, Global Avg Loss: 3.08235579, Time: 0.0109 Steps: 22360, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001080, Sample Num: 17280, Cur Loss: 1.68733001, Cur Avg Loss: 0.92711198, Log Avg loss: 0.94105110, Global Avg Loss: 3.08139857, Time: 0.0086 Steps: 22370, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001090, Sample Num: 17440, Cur Loss: 1.27046275, Cur Avg Loss: 0.92765279, Log Avg loss: 0.98605992, Global Avg Loss: 3.08046231, Time: 0.0138 Steps: 22380, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001100, Sample Num: 17600, Cur Loss: 0.44995052, Cur Avg Loss: 0.92707680, Log Avg loss: 0.86429407, Global Avg Loss: 3.07947251, Time: 0.0084 Steps: 22390, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001110, Sample Num: 17760, Cur Loss: 0.63766015, Cur Avg Loss: 0.92582059, Log Avg loss: 0.78763747, Global Avg Loss: 3.07844937, Time: 0.0115 Steps: 22400, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001120, Sample Num: 17920, Cur Loss: 0.76110393, Cur Avg Loss: 0.92684824, Log Avg loss: 1.04091804, Global Avg Loss: 3.07754016, Time: 0.0117 Steps: 22410, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001130, Sample Num: 18080, Cur Loss: 0.52185440, Cur Avg Loss: 0.92704899, Log Avg loss: 0.94953229, Global Avg Loss: 3.07659101, Time: 0.0081 Steps: 22420, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001140, Sample Num: 18240, Cur Loss: 1.02326965, Cur Avg Loss: 0.92656481, Log Avg loss: 0.87185208, Global Avg Loss: 3.07560807, Time: 0.0127 Steps: 22430, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001150, Sample Num: 18400, Cur Loss: 0.97960484, Cur Avg Loss: 0.92685560, Log Avg loss: 0.96000655, Global Avg Loss: 3.07466528, Time: 0.0157 Steps: 22440, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001160, Sample Num: 18560, Cur Loss: 0.71349227, Cur Avg Loss: 0.92570354, Log Avg loss: 0.79321573, Global Avg Loss: 3.07364905, Time: 0.0093 Steps: 22450, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001170, Sample Num: 18720, Cur Loss: 0.66762352, Cur Avg Loss: 0.92517073, Log Avg loss: 0.86336542, Global Avg Loss: 3.07266495, Time: 0.0103 Steps: 22460, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001180, Sample Num: 18880, Cur Loss: 0.99172854, Cur Avg Loss: 0.92565399, Log Avg loss: 0.98219523, Global Avg Loss: 3.07173461, Time: 0.0090 Steps: 22470, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001190, Sample Num: 19040, Cur Loss: 0.85876650, Cur Avg Loss: 0.92837805, Log Avg loss: 1.24981762, Global Avg Loss: 3.07092415, Time: 0.0092 Steps: 22480, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001200, Sample Num: 19200, Cur Loss: 1.08871913, Cur Avg Loss: 0.93010359, Log Avg loss: 1.13544216, Global Avg Loss: 3.07006355, Time: 0.0113 Steps: 22490, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001210, Sample Num: 19360, Cur Loss: 1.14910960, Cur Avg Loss: 0.93111542, Log Avg loss: 1.05253512, Global Avg Loss: 3.06916688, Time: 0.0092 Steps: 22500, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001220, Sample Num: 19520, Cur Loss: 0.77336496, Cur Avg Loss: 0.93132416, Log Avg loss: 0.95658223, Global Avg Loss: 3.06822837, Time: 0.0099 Steps: 22510, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001230, Sample Num: 19680, Cur Loss: 0.45695889, Cur Avg Loss: 0.93129228, Log Avg loss: 0.92740211, Global Avg Loss: 3.06727773, Time: 0.0108 Steps: 22520, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001240, Sample Num: 19840, Cur Loss: 0.43685046, Cur Avg Loss: 0.93064373, Log Avg loss: 0.85087301, Global Avg Loss: 3.06629398, Time: 0.0115 Steps: 22530, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001250, Sample Num: 20000, Cur Loss: 0.26099086, Cur Avg Loss: 0.92987469, Log Avg loss: 0.83451275, Global Avg Loss: 3.06530383, Time: 0.0123 Steps: 22540, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001260, Sample Num: 20160, Cur Loss: 1.14100969, Cur Avg Loss: 0.92995784, Log Avg loss: 0.94035217, Global Avg Loss: 3.06436150, Time: 0.0087 Steps: 22550, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001270, Sample Num: 20320, Cur Loss: 0.81153798, Cur Avg Loss: 0.92913998, Log Avg loss: 0.82608901, Global Avg Loss: 3.06336936, Time: 0.0068 Steps: 22560, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001280, Sample Num: 20480, Cur Loss: 1.34138155, Cur Avg Loss: 0.92753446, Log Avg loss: 0.72363353, Global Avg Loss: 3.06233270, Time: 0.0121 Steps: 22570, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001290, Sample Num: 20640, Cur Loss: 1.26155496, Cur Avg Loss: 0.92727146, Log Avg loss: 0.89360802, Global Avg Loss: 3.06137224, Time: 0.0139 Steps: 22580, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001300, Sample Num: 20800, Cur Loss: 1.58045077, Cur Avg Loss: 0.92679755, Log Avg loss: 0.86566276, Global Avg Loss: 3.06040026, Time: 0.0150 Steps: 22590, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001310, Sample Num: 20960, Cur Loss: 0.66718322, Cur Avg Loss: 0.92567535, Log Avg loss: 0.77979030, Global Avg Loss: 3.05939114, Time: 0.0168 Steps: 22600, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001320, Sample Num: 21120, Cur Loss: 1.54903555, Cur Avg Loss: 0.92558592, Log Avg loss: 0.91386949, Global Avg Loss: 3.05844221, Time: 0.0068 Steps: 22610, Updated lr: 0.000080 Training, Epoch: 0011, Batch: 001330, Sample Num: 21280, Cur Loss: 1.09553027, Cur Avg Loss: 0.92664366, Log Avg loss: 1.06626620, Global Avg Loss: 3.05756150, Time: 0.0114 Steps: 22620, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001340, Sample Num: 21440, Cur Loss: 0.38966632, Cur Avg Loss: 0.92642987, Log Avg loss: 0.89799485, Global Avg Loss: 3.05660721, Time: 0.0085 Steps: 22630, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001350, Sample Num: 21600, Cur Loss: 0.78466648, Cur Avg Loss: 0.92689424, Log Avg loss: 0.98912003, Global Avg Loss: 3.05569400, Time: 0.0066 Steps: 22640, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001360, Sample Num: 21760, Cur Loss: 0.47525781, Cur Avg Loss: 0.92540464, Log Avg loss: 0.72430917, Global Avg Loss: 3.05466470, Time: 0.0065 Steps: 22650, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001370, Sample Num: 21920, Cur Loss: 0.80101264, Cur Avg Loss: 0.92545198, Log Avg loss: 0.93188966, Global Avg Loss: 3.05372790, Time: 0.0140 Steps: 22660, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001380, Sample Num: 22080, Cur Loss: 0.45892343, Cur Avg Loss: 0.92491134, Log Avg loss: 0.85084379, Global Avg Loss: 3.05275618, Time: 0.0067 Steps: 22670, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001390, Sample Num: 22240, Cur Loss: 0.11891061, Cur Avg Loss: 0.92403693, Log Avg loss: 0.80336898, Global Avg Loss: 3.05176439, Time: 0.0109 Steps: 22680, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001400, Sample Num: 22400, Cur Loss: 0.83009887, Cur Avg Loss: 0.92252146, Log Avg loss: 0.71187016, Global Avg Loss: 3.05073315, Time: 0.0122 Steps: 22690, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001410, Sample Num: 22560, Cur Loss: 0.96501708, Cur Avg Loss: 0.92257206, Log Avg loss: 0.92965613, Global Avg Loss: 3.04979875, Time: 0.0225 Steps: 22700, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001420, Sample Num: 22720, Cur Loss: 1.87254071, Cur Avg Loss: 0.92340476, Log Avg loss: 1.04081591, Global Avg Loss: 3.04891413, Time: 0.0168 Steps: 22710, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001430, Sample Num: 22880, Cur Loss: 0.93565130, Cur Avg Loss: 0.92424851, Log Avg loss: 1.04406073, Global Avg Loss: 3.04803171, Time: 0.0114 Steps: 22720, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001440, Sample Num: 23040, Cur Loss: 1.24363935, Cur Avg Loss: 0.92407571, Log Avg loss: 0.89936546, Global Avg Loss: 3.04708641, Time: 0.0108 Steps: 22730, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001450, Sample Num: 23200, Cur Loss: 0.56506753, Cur Avg Loss: 0.92466855, Log Avg loss: 1.01003717, Global Avg Loss: 3.04619061, Time: 0.0136 Steps: 22740, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001460, Sample Num: 23360, Cur Loss: 1.54784775, Cur Avg Loss: 0.92531437, Log Avg loss: 1.01895861, Global Avg Loss: 3.04529952, Time: 0.0071 Steps: 22750, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001470, Sample Num: 23520, Cur Loss: 1.03655171, Cur Avg Loss: 0.92634727, Log Avg loss: 1.07715110, Global Avg Loss: 3.04443478, Time: 0.0065 Steps: 22760, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001480, Sample Num: 23680, Cur Loss: 1.16406202, Cur Avg Loss: 0.92586908, Log Avg loss: 0.85557461, Global Avg Loss: 3.04347349, Time: 0.0118 Steps: 22770, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001490, Sample Num: 23840, Cur Loss: 1.15874612, Cur Avg Loss: 0.92617734, Log Avg loss: 0.97180028, Global Avg Loss: 3.04256406, Time: 0.0136 Steps: 22780, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001500, Sample Num: 24000, Cur Loss: 0.66799700, Cur Avg Loss: 0.92423213, Log Avg loss: 0.63439580, Global Avg Loss: 3.04150738, Time: 0.0068 Steps: 22790, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001510, Sample Num: 24160, Cur Loss: 1.43532705, Cur Avg Loss: 0.92503649, Log Avg loss: 1.04569103, Global Avg Loss: 3.04063202, Time: 0.0145 Steps: 22800, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001520, Sample Num: 24320, Cur Loss: 0.70649004, Cur Avg Loss: 0.92619394, Log Avg loss: 1.10096892, Global Avg Loss: 3.03978167, Time: 0.0065 Steps: 22810, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001530, Sample Num: 24480, Cur Loss: 1.03608704, Cur Avg Loss: 0.92677777, Log Avg loss: 1.01551857, Global Avg Loss: 3.03889461, Time: 0.0123 Steps: 22820, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001540, Sample Num: 24640, Cur Loss: 0.22535630, Cur Avg Loss: 0.92765658, Log Avg loss: 1.06211592, Global Avg Loss: 3.03802874, Time: 0.0225 Steps: 22830, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001550, Sample Num: 24800, Cur Loss: 0.73185360, Cur Avg Loss: 0.92794912, Log Avg loss: 0.97300024, Global Avg Loss: 3.03712461, Time: 0.0110 Steps: 22840, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001560, Sample Num: 24960, Cur Loss: 0.47097579, Cur Avg Loss: 0.92792151, Log Avg loss: 0.92364200, Global Avg Loss: 3.03619968, Time: 0.0068 Steps: 22850, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001570, Sample Num: 25120, Cur Loss: 0.64201486, Cur Avg Loss: 0.92692543, Log Avg loss: 0.77153623, Global Avg Loss: 3.03520901, Time: 0.0118 Steps: 22860, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001580, Sample Num: 25280, Cur Loss: 0.80903119, Cur Avg Loss: 0.92818296, Log Avg loss: 1.12561589, Global Avg Loss: 3.03437403, Time: 0.0114 Steps: 22870, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001590, Sample Num: 25440, Cur Loss: 1.13639355, Cur Avg Loss: 0.92901617, Log Avg loss: 1.06066186, Global Avg Loss: 3.03351140, Time: 0.0158 Steps: 22880, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001600, Sample Num: 25600, Cur Loss: 1.09949160, Cur Avg Loss: 0.92840798, Log Avg loss: 0.83170668, Global Avg Loss: 3.03254949, Time: 0.0138 Steps: 22890, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001610, Sample Num: 25760, Cur Loss: 0.50318742, Cur Avg Loss: 0.92900691, Log Avg loss: 1.02483618, Global Avg Loss: 3.03167276, Time: 0.0113 Steps: 22900, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001620, Sample Num: 25920, Cur Loss: 1.11748874, Cur Avg Loss: 0.92897251, Log Avg loss: 0.92343404, Global Avg Loss: 3.03075253, Time: 0.0137 Steps: 22910, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001630, Sample Num: 26080, Cur Loss: 1.14208794, Cur Avg Loss: 0.92977536, Log Avg loss: 1.05983729, Global Avg Loss: 3.02989262, Time: 0.0068 Steps: 22920, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001640, Sample Num: 26240, Cur Loss: 1.08470702, Cur Avg Loss: 0.92990728, Log Avg loss: 0.95140889, Global Avg Loss: 3.02898617, Time: 0.0148 Steps: 22930, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001650, Sample Num: 26400, Cur Loss: 0.61372113, Cur Avg Loss: 0.92885821, Log Avg loss: 0.75681156, Global Avg Loss: 3.02799569, Time: 0.0090 Steps: 22940, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001660, Sample Num: 26560, Cur Loss: 0.32624096, Cur Avg Loss: 0.92836143, Log Avg loss: 0.84639206, Global Avg Loss: 3.02704510, Time: 0.0117 Steps: 22950, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001670, Sample Num: 26720, Cur Loss: 0.94628549, Cur Avg Loss: 0.92919862, Log Avg loss: 1.06817327, Global Avg Loss: 3.02619193, Time: 0.0228 Steps: 22960, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001680, Sample Num: 26880, Cur Loss: 0.67137039, Cur Avg Loss: 0.92950976, Log Avg loss: 0.98146949, Global Avg Loss: 3.02530176, Time: 0.0073 Steps: 22970, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001690, Sample Num: 27040, Cur Loss: 0.36724484, Cur Avg Loss: 0.92882655, Log Avg loss: 0.81404780, Global Avg Loss: 3.02433951, Time: 0.0068 Steps: 22980, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001700, Sample Num: 27200, Cur Loss: 0.81382251, Cur Avg Loss: 0.92879313, Log Avg loss: 0.92314502, Global Avg Loss: 3.02342555, Time: 0.0119 Steps: 22990, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001710, Sample Num: 27360, Cur Loss: 0.93194771, Cur Avg Loss: 0.93027432, Log Avg loss: 1.18207583, Global Avg Loss: 3.02262496, Time: 0.0119 Steps: 23000, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001720, Sample Num: 27520, Cur Loss: 0.35733503, Cur Avg Loss: 0.92992179, Log Avg loss: 0.86963998, Global Avg Loss: 3.02168929, Time: 0.0068 Steps: 23010, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001730, Sample Num: 27680, Cur Loss: 1.34529591, Cur Avg Loss: 0.92875657, Log Avg loss: 0.72833860, Global Avg Loss: 3.02069305, Time: 0.0125 Steps: 23020, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001740, Sample Num: 27840, Cur Loss: 0.97440314, Cur Avg Loss: 0.92897252, Log Avg loss: 0.96633081, Global Avg Loss: 3.01980101, Time: 0.0109 Steps: 23030, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001750, Sample Num: 28000, Cur Loss: 0.64482683, Cur Avg Loss: 0.92879884, Log Avg loss: 0.89857912, Global Avg Loss: 3.01888034, Time: 0.0174 Steps: 23040, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001760, Sample Num: 28160, Cur Loss: 1.52600121, Cur Avg Loss: 0.92915218, Log Avg loss: 0.99098613, Global Avg Loss: 3.01800056, Time: 0.0123 Steps: 23050, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001770, Sample Num: 28320, Cur Loss: 0.86918950, Cur Avg Loss: 0.92832734, Log Avg loss: 0.78315584, Global Avg Loss: 3.01703141, Time: 0.0144 Steps: 23060, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001780, Sample Num: 28480, Cur Loss: 1.07870936, Cur Avg Loss: 0.92741735, Log Avg loss: 0.76635017, Global Avg Loss: 3.01605583, Time: 0.0118 Steps: 23070, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001790, Sample Num: 28640, Cur Loss: 1.31344473, Cur Avg Loss: 0.92751189, Log Avg loss: 0.94433866, Global Avg Loss: 3.01515820, Time: 0.0142 Steps: 23080, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001800, Sample Num: 28800, Cur Loss: 0.98816061, Cur Avg Loss: 0.92786806, Log Avg loss: 0.99162228, Global Avg Loss: 3.01428183, Time: 0.0120 Steps: 23090, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001810, Sample Num: 28960, Cur Loss: 1.07072425, Cur Avg Loss: 0.92746736, Log Avg loss: 0.85534213, Global Avg Loss: 3.01334723, Time: 0.0119 Steps: 23100, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001820, Sample Num: 29120, Cur Loss: 2.17187190, Cur Avg Loss: 0.92710233, Log Avg loss: 0.86103238, Global Avg Loss: 3.01241589, Time: 0.0145 Steps: 23110, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001830, Sample Num: 29280, Cur Loss: 1.16905200, Cur Avg Loss: 0.92653558, Log Avg loss: 0.82338700, Global Avg Loss: 3.01146908, Time: 0.0088 Steps: 23120, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001840, Sample Num: 29440, Cur Loss: 1.95847702, Cur Avg Loss: 0.92800203, Log Avg loss: 1.19636250, Global Avg Loss: 3.01068434, Time: 0.0153 Steps: 23130, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001850, Sample Num: 29600, Cur Loss: 0.74621582, Cur Avg Loss: 0.92889196, Log Avg loss: 1.09263917, Global Avg Loss: 3.00985545, Time: 0.0066 Steps: 23140, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001860, Sample Num: 29760, Cur Loss: 0.82608223, Cur Avg Loss: 0.92896235, Log Avg loss: 0.94198427, Global Avg Loss: 3.00896220, Time: 0.0108 Steps: 23150, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001870, Sample Num: 29920, Cur Loss: 0.64492869, Cur Avg Loss: 0.92907767, Log Avg loss: 0.95052678, Global Avg Loss: 3.00807341, Time: 0.0097 Steps: 23160, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001880, Sample Num: 30080, Cur Loss: 0.48391643, Cur Avg Loss: 0.92932905, Log Avg loss: 0.97633744, Global Avg Loss: 3.00719653, Time: 0.0109 Steps: 23170, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001890, Sample Num: 30240, Cur Loss: 0.39122349, Cur Avg Loss: 0.92895598, Log Avg loss: 0.85881796, Global Avg Loss: 3.00626971, Time: 0.0068 Steps: 23180, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001900, Sample Num: 30400, Cur Loss: 0.45683971, Cur Avg Loss: 0.92936957, Log Avg loss: 1.00753897, Global Avg Loss: 3.00540781, Time: 0.0118 Steps: 23190, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001910, Sample Num: 30560, Cur Loss: 1.78680146, Cur Avg Loss: 0.92790282, Log Avg loss: 0.64921985, Global Avg Loss: 3.00439222, Time: 0.0122 Steps: 23200, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001920, Sample Num: 30720, Cur Loss: 0.99137592, Cur Avg Loss: 0.92850754, Log Avg loss: 1.04400878, Global Avg Loss: 3.00354759, Time: 0.0107 Steps: 23210, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001930, Sample Num: 30880, Cur Loss: 1.21598005, Cur Avg Loss: 0.92812866, Log Avg loss: 0.85538508, Global Avg Loss: 3.00262245, Time: 0.0137 Steps: 23220, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001940, Sample Num: 31040, Cur Loss: 1.15614223, Cur Avg Loss: 0.92863099, Log Avg loss: 1.02558017, Global Avg Loss: 3.00177138, Time: 0.0114 Steps: 23230, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001950, Sample Num: 31200, Cur Loss: 1.92989814, Cur Avg Loss: 0.92939310, Log Avg loss: 1.07724251, Global Avg Loss: 3.00094327, Time: 0.0109 Steps: 23240, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001960, Sample Num: 31360, Cur Loss: 0.97716612, Cur Avg Loss: 0.92899721, Log Avg loss: 0.85179884, Global Avg Loss: 3.00001891, Time: 0.0116 Steps: 23250, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001970, Sample Num: 31520, Cur Loss: 0.70984042, Cur Avg Loss: 0.92796046, Log Avg loss: 0.72475657, Global Avg Loss: 2.99904072, Time: 0.0117 Steps: 23260, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001980, Sample Num: 31680, Cur Loss: 0.90982950, Cur Avg Loss: 0.92812972, Log Avg loss: 0.96147346, Global Avg Loss: 2.99816510, Time: 0.0072 Steps: 23270, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 001990, Sample Num: 31840, Cur Loss: 1.72782135, Cur Avg Loss: 0.92743040, Log Avg loss: 0.78896657, Global Avg Loss: 2.99721613, Time: 0.0115 Steps: 23280, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002000, Sample Num: 32000, Cur Loss: 0.53402174, Cur Avg Loss: 0.92583786, Log Avg loss: 0.60892086, Global Avg Loss: 2.99619067, Time: 0.0088 Steps: 23290, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002010, Sample Num: 32160, Cur Loss: 0.77325761, Cur Avg Loss: 0.92556725, Log Avg loss: 0.87144516, Global Avg Loss: 2.99527876, Time: 0.0123 Steps: 23300, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002020, Sample Num: 32320, Cur Loss: 1.27591300, Cur Avg Loss: 0.92675781, Log Avg loss: 1.16606198, Global Avg Loss: 2.99449403, Time: 0.0084 Steps: 23310, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002030, Sample Num: 32480, Cur Loss: 1.44141388, Cur Avg Loss: 0.92694709, Log Avg loss: 0.96518030, Global Avg Loss: 2.99362383, Time: 0.0073 Steps: 23320, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002040, Sample Num: 32640, Cur Loss: 1.04412627, Cur Avg Loss: 0.92697937, Log Avg loss: 0.93353288, Global Avg Loss: 2.99274080, Time: 0.0139 Steps: 23330, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002050, Sample Num: 32800, Cur Loss: 0.88038194, Cur Avg Loss: 0.92699184, Log Avg loss: 0.92953596, Global Avg Loss: 2.99185683, Time: 0.0123 Steps: 23340, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002060, Sample Num: 32960, Cur Loss: 0.14230540, Cur Avg Loss: 0.92624420, Log Avg loss: 0.77297777, Global Avg Loss: 2.99090656, Time: 0.0157 Steps: 23350, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002070, Sample Num: 33120, Cur Loss: 0.71947342, Cur Avg Loss: 0.92625375, Log Avg loss: 0.92822151, Global Avg Loss: 2.99002356, Time: 0.0116 Steps: 23360, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002080, Sample Num: 33280, Cur Loss: 0.94758677, Cur Avg Loss: 0.92590556, Log Avg loss: 0.85383009, Global Avg Loss: 2.98910948, Time: 0.0136 Steps: 23370, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002090, Sample Num: 33440, Cur Loss: 1.21411991, Cur Avg Loss: 0.92486801, Log Avg loss: 0.70905774, Global Avg Loss: 2.98813427, Time: 0.0141 Steps: 23380, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002100, Sample Num: 33600, Cur Loss: 0.49172026, Cur Avg Loss: 0.92468039, Log Avg loss: 0.88546629, Global Avg Loss: 2.98723531, Time: 0.0114 Steps: 23390, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002110, Sample Num: 33760, Cur Loss: 1.14437425, Cur Avg Loss: 0.92447844, Log Avg loss: 0.88207094, Global Avg Loss: 2.98633566, Time: 0.0089 Steps: 23400, Updated lr: 0.000079 Training, Epoch: 0011, Batch: 002120, Sample Num: 33920, Cur Loss: 0.44793794, Cur Avg Loss: 0.92438038, Log Avg loss: 0.90368849, Global Avg Loss: 2.98544602, Time: 0.0125 Steps: 23410, Updated lr: 0.000079 ***** Running evaluation checkpoint-23419 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-23419 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.076758, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.965565, "eval_total_loss": 678.791966, "eval_mae": 0.786442, "eval_mse": 0.965766, "eval_r2": 0.386096, "eval_sp_statistic": 0.604699, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.631922, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.462768, "test_total_loss": 734.309735, "test_mae": 1.064902, "test_mse": 1.462745, "test_r2": 0.055932, "test_sp_statistic": 0.37404, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.499883, "test_ps_pvalue": 0.0, "lr": 7.873968705547652e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.9845280289793763, "train_cur_epoch_loss": 1965.0568776652217, "train_cur_epoch_avg_loss": 0.9229952454979905, "train_cur_epoch_time": 22.076757669448853, "train_cur_epoch_avg_time": 0.01036954329236677, "epoch": 11, "step": 23419} ################################################## Training, Epoch: 0012, Batch: 000001, Sample Num: 16, Cur Loss: 0.28885186, Cur Avg Loss: 0.28885186, Log Avg loss: 0.56593266, Global Avg Loss: 2.98441293, Time: 0.0121 Steps: 23420, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000011, Sample Num: 176, Cur Loss: 1.05082560, Cur Avg Loss: 1.17179308, Log Avg loss: 1.26008720, Global Avg Loss: 2.98367698, Time: 0.0117 Steps: 23430, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000021, Sample Num: 336, Cur Loss: 0.49435437, Cur Avg Loss: 1.04214494, Log Avg loss: 0.89953199, Global Avg Loss: 2.98278784, Time: 0.0110 Steps: 23440, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000031, Sample Num: 496, Cur Loss: 1.02081442, Cur Avg Loss: 1.00424667, Log Avg loss: 0.92466031, Global Avg Loss: 2.98191017, Time: 0.0115 Steps: 23450, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000041, Sample Num: 656, Cur Loss: 0.85088241, Cur Avg Loss: 0.96007864, Log Avg loss: 0.82315772, Global Avg Loss: 2.98098999, Time: 0.0142 Steps: 23460, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000051, Sample Num: 816, Cur Loss: 1.26566994, Cur Avg Loss: 0.93461078, Log Avg loss: 0.83019256, Global Avg Loss: 2.98007359, Time: 0.0160 Steps: 23470, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000061, Sample Num: 976, Cur Loss: 0.51252449, Cur Avg Loss: 0.95514077, Log Avg loss: 1.05984372, Global Avg Loss: 2.97925577, Time: 0.0118 Steps: 23480, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000071, Sample Num: 1136, Cur Loss: 0.20099394, Cur Avg Loss: 0.92626148, Log Avg loss: 0.75009782, Global Avg Loss: 2.97830679, Time: 0.0065 Steps: 23490, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000081, Sample Num: 1296, Cur Loss: 0.50941539, Cur Avg Loss: 0.94032185, Log Avg loss: 1.04015046, Global Avg Loss: 2.97748204, Time: 0.0067 Steps: 23500, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000091, Sample Num: 1456, Cur Loss: 1.25065947, Cur Avg Loss: 0.94613334, Log Avg loss: 0.99320645, Global Avg Loss: 2.97663803, Time: 0.0098 Steps: 23510, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000101, Sample Num: 1616, Cur Loss: 0.31277734, Cur Avg Loss: 0.93641404, Log Avg loss: 0.84796839, Global Avg Loss: 2.97573298, Time: 0.0120 Steps: 23520, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000111, Sample Num: 1776, Cur Loss: 0.84228951, Cur Avg Loss: 0.93092263, Log Avg loss: 0.87545942, Global Avg Loss: 2.97484039, Time: 0.0117 Steps: 23530, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000121, Sample Num: 1936, Cur Loss: 0.47601563, Cur Avg Loss: 0.93219865, Log Avg loss: 0.94636243, Global Avg Loss: 2.97397867, Time: 0.0103 Steps: 23540, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000131, Sample Num: 2096, Cur Loss: 0.58851457, Cur Avg Loss: 0.92021969, Log Avg loss: 0.77527426, Global Avg Loss: 2.97304504, Time: 0.0127 Steps: 23550, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000141, Sample Num: 2256, Cur Loss: 0.86304098, Cur Avg Loss: 0.91224025, Log Avg loss: 0.80770961, Global Avg Loss: 2.97212597, Time: 0.0138 Steps: 23560, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000151, Sample Num: 2416, Cur Loss: 0.81145096, Cur Avg Loss: 0.91072421, Log Avg loss: 0.88934796, Global Avg Loss: 2.97124231, Time: 0.0068 Steps: 23570, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000161, Sample Num: 2576, Cur Loss: 0.62409186, Cur Avg Loss: 0.91013952, Log Avg loss: 0.90131069, Global Avg Loss: 2.97036448, Time: 0.0068 Steps: 23580, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000171, Sample Num: 2736, Cur Loss: 0.36981624, Cur Avg Loss: 0.91346722, Log Avg loss: 0.96704322, Global Avg Loss: 2.96951525, Time: 0.0116 Steps: 23590, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000181, Sample Num: 2896, Cur Loss: 0.76129317, Cur Avg Loss: 0.90605251, Log Avg loss: 0.77926104, Global Avg Loss: 2.96858718, Time: 0.0068 Steps: 23600, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000191, Sample Num: 3056, Cur Loss: 0.37248442, Cur Avg Loss: 0.89224181, Log Avg loss: 0.64226801, Global Avg Loss: 2.96760187, Time: 0.0070 Steps: 23610, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000201, Sample Num: 3216, Cur Loss: 0.80436909, Cur Avg Loss: 0.89671905, Log Avg loss: 0.98223453, Global Avg Loss: 2.96676132, Time: 0.0206 Steps: 23620, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000211, Sample Num: 3376, Cur Loss: 0.68840516, Cur Avg Loss: 0.89228695, Log Avg loss: 0.80320163, Global Avg Loss: 2.96584572, Time: 0.0065 Steps: 23630, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000221, Sample Num: 3536, Cur Loss: 0.78110343, Cur Avg Loss: 0.90536071, Log Avg loss: 1.18121698, Global Avg Loss: 2.96509080, Time: 0.0065 Steps: 23640, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000231, Sample Num: 3696, Cur Loss: 0.71124655, Cur Avg Loss: 0.91021709, Log Avg loss: 1.01754329, Global Avg Loss: 2.96426732, Time: 0.0110 Steps: 23650, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000241, Sample Num: 3856, Cur Loss: 1.24512815, Cur Avg Loss: 0.90755644, Log Avg loss: 0.84609520, Global Avg Loss: 2.96337206, Time: 0.0226 Steps: 23660, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000251, Sample Num: 4016, Cur Loss: 0.70471078, Cur Avg Loss: 0.90940121, Log Avg loss: 0.95386020, Global Avg Loss: 2.96252309, Time: 0.0067 Steps: 23670, Updated lr: 0.000079 Training, Epoch: 0012, Batch: 000261, Sample Num: 4176, Cur Loss: 1.05163109, Cur Avg Loss: 0.90729285, Log Avg loss: 0.85437302, Global Avg Loss: 2.96163283, Time: 0.0160 Steps: 23680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000271, Sample Num: 4336, Cur Loss: 0.34646976, Cur Avg Loss: 0.89620894, Log Avg loss: 0.60691907, Global Avg Loss: 2.96063886, Time: 0.0119 Steps: 23690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000281, Sample Num: 4496, Cur Loss: 0.53155535, Cur Avg Loss: 0.89153431, Log Avg loss: 0.76485178, Global Avg Loss: 2.95971237, Time: 0.0107 Steps: 23700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000291, Sample Num: 4656, Cur Loss: 0.92743266, Cur Avg Loss: 0.88405902, Log Avg loss: 0.67400328, Global Avg Loss: 2.95874834, Time: 0.0078 Steps: 23710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000301, Sample Num: 4816, Cur Loss: 1.03284407, Cur Avg Loss: 0.88039952, Log Avg loss: 0.77390806, Global Avg Loss: 2.95782724, Time: 0.0068 Steps: 23720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000311, Sample Num: 4976, Cur Loss: 0.63939750, Cur Avg Loss: 0.88316770, Log Avg loss: 0.96648992, Global Avg Loss: 2.95698808, Time: 0.0113 Steps: 23730, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000321, Sample Num: 5136, Cur Loss: 0.37954408, Cur Avg Loss: 0.88056030, Log Avg loss: 0.79947033, Global Avg Loss: 2.95607927, Time: 0.0066 Steps: 23740, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000331, Sample Num: 5296, Cur Loss: 1.39380956, Cur Avg Loss: 0.89209337, Log Avg loss: 1.26230473, Global Avg Loss: 2.95536610, Time: 0.0110 Steps: 23750, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000341, Sample Num: 5456, Cur Loss: 0.28819808, Cur Avg Loss: 0.88820383, Log Avg loss: 0.75946006, Global Avg Loss: 2.95444189, Time: 0.0063 Steps: 23760, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000351, Sample Num: 5616, Cur Loss: 0.76975507, Cur Avg Loss: 0.88883629, Log Avg loss: 0.91040312, Global Avg Loss: 2.95358197, Time: 0.0112 Steps: 23770, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000361, Sample Num: 5776, Cur Loss: 2.17039132, Cur Avg Loss: 0.89478515, Log Avg loss: 1.10359020, Global Avg Loss: 2.95280401, Time: 0.0129 Steps: 23780, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000371, Sample Num: 5936, Cur Loss: 1.44527757, Cur Avg Loss: 0.90147174, Log Avg loss: 1.14285780, Global Avg Loss: 2.95204321, Time: 0.0067 Steps: 23790, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000381, Sample Num: 6096, Cur Loss: 1.01118684, Cur Avg Loss: 0.89644515, Log Avg loss: 0.70995868, Global Avg Loss: 2.95110116, Time: 0.0068 Steps: 23800, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000391, Sample Num: 6256, Cur Loss: 0.38803685, Cur Avg Loss: 0.89049271, Log Avg loss: 0.66370449, Global Avg Loss: 2.95014047, Time: 0.0106 Steps: 23810, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000401, Sample Num: 6416, Cur Loss: 0.51392543, Cur Avg Loss: 0.88325116, Log Avg loss: 0.60010681, Global Avg Loss: 2.94915389, Time: 0.0132 Steps: 23820, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000411, Sample Num: 6576, Cur Loss: 0.76239836, Cur Avg Loss: 0.88378554, Log Avg loss: 0.90521405, Global Avg Loss: 2.94829617, Time: 0.0118 Steps: 23830, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000421, Sample Num: 6736, Cur Loss: 0.72406960, Cur Avg Loss: 0.88429798, Log Avg loss: 0.90535918, Global Avg Loss: 2.94743923, Time: 0.0106 Steps: 23840, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000431, Sample Num: 6896, Cur Loss: 0.84014547, Cur Avg Loss: 0.88898306, Log Avg loss: 1.08622501, Global Avg Loss: 2.94665885, Time: 0.0106 Steps: 23850, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000441, Sample Num: 7056, Cur Loss: 0.42937189, Cur Avg Loss: 0.88984740, Log Avg loss: 0.92710057, Global Avg Loss: 2.94581243, Time: 0.0078 Steps: 23860, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000451, Sample Num: 7216, Cur Loss: 0.32604027, Cur Avg Loss: 0.89476491, Log Avg loss: 1.11162681, Global Avg Loss: 2.94504403, Time: 0.0118 Steps: 23870, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000461, Sample Num: 7376, Cur Loss: 0.59636086, Cur Avg Loss: 0.88963368, Log Avg loss: 0.65821538, Global Avg Loss: 2.94408639, Time: 0.0220 Steps: 23880, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000471, Sample Num: 7536, Cur Loss: 0.50794041, Cur Avg Loss: 0.88904529, Log Avg loss: 0.86192070, Global Avg Loss: 2.94321483, Time: 0.0119 Steps: 23890, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000481, Sample Num: 7696, Cur Loss: 0.59258723, Cur Avg Loss: 0.89250581, Log Avg loss: 1.05549595, Global Avg Loss: 2.94242499, Time: 0.0066 Steps: 23900, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000491, Sample Num: 7856, Cur Loss: 0.89314932, Cur Avg Loss: 0.89334965, Log Avg loss: 0.93393830, Global Avg Loss: 2.94158497, Time: 0.0083 Steps: 23910, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000501, Sample Num: 8016, Cur Loss: 0.74557132, Cur Avg Loss: 0.89382856, Log Avg loss: 0.91734335, Global Avg Loss: 2.94073871, Time: 0.0110 Steps: 23920, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000511, Sample Num: 8176, Cur Loss: 0.32756156, Cur Avg Loss: 0.89732571, Log Avg loss: 1.07253288, Global Avg Loss: 2.93995802, Time: 0.0099 Steps: 23930, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000521, Sample Num: 8336, Cur Loss: 0.92170864, Cur Avg Loss: 0.89932870, Log Avg loss: 1.00168149, Global Avg Loss: 2.93914838, Time: 0.0066 Steps: 23940, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000531, Sample Num: 8496, Cur Loss: 0.63654351, Cur Avg Loss: 0.90490980, Log Avg loss: 1.19568497, Global Avg Loss: 2.93842042, Time: 0.0109 Steps: 23950, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000541, Sample Num: 8656, Cur Loss: 0.55315375, Cur Avg Loss: 0.90808813, Log Avg loss: 1.07685745, Global Avg Loss: 2.93764347, Time: 0.0064 Steps: 23960, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000551, Sample Num: 8816, Cur Loss: 0.75940645, Cur Avg Loss: 0.90619048, Log Avg loss: 0.80352791, Global Avg Loss: 2.93675314, Time: 0.0067 Steps: 23970, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000561, Sample Num: 8976, Cur Loss: 0.52616358, Cur Avg Loss: 0.90070734, Log Avg loss: 0.59858621, Global Avg Loss: 2.93577810, Time: 0.0114 Steps: 23980, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000571, Sample Num: 9136, Cur Loss: 0.52276462, Cur Avg Loss: 0.90135072, Log Avg loss: 0.93744422, Global Avg Loss: 2.93494511, Time: 0.0110 Steps: 23990, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000581, Sample Num: 9296, Cur Loss: 0.94769812, Cur Avg Loss: 0.90277661, Log Avg loss: 0.98419506, Global Avg Loss: 2.93413230, Time: 0.0111 Steps: 24000, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000591, Sample Num: 9456, Cur Loss: 0.63604754, Cur Avg Loss: 0.89997693, Log Avg loss: 0.73731534, Global Avg Loss: 2.93321734, Time: 0.0221 Steps: 24010, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000601, Sample Num: 9616, Cur Loss: 1.65842748, Cur Avg Loss: 0.89765005, Log Avg loss: 0.76013129, Global Avg Loss: 2.93231264, Time: 0.0139 Steps: 24020, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000611, Sample Num: 9776, Cur Loss: 0.73916227, Cur Avg Loss: 0.89765276, Log Avg loss: 0.89781595, Global Avg Loss: 2.93146599, Time: 0.0104 Steps: 24030, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000621, Sample Num: 9936, Cur Loss: 0.86097383, Cur Avg Loss: 0.89663031, Log Avg loss: 0.83415838, Global Avg Loss: 2.93059357, Time: 0.0118 Steps: 24040, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000631, Sample Num: 10096, Cur Loss: 0.88951129, Cur Avg Loss: 0.89512752, Log Avg loss: 0.80180431, Global Avg Loss: 2.92970841, Time: 0.0071 Steps: 24050, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000641, Sample Num: 10256, Cur Loss: 0.66355222, Cur Avg Loss: 0.89345568, Log Avg loss: 0.78796274, Global Avg Loss: 2.92881825, Time: 0.0066 Steps: 24060, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000651, Sample Num: 10416, Cur Loss: 1.30654633, Cur Avg Loss: 0.89325582, Log Avg loss: 0.88044443, Global Avg Loss: 2.92796724, Time: 0.0112 Steps: 24070, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000661, Sample Num: 10576, Cur Loss: 0.46230429, Cur Avg Loss: 0.89026361, Log Avg loss: 0.69547133, Global Avg Loss: 2.92704012, Time: 0.0153 Steps: 24080, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000671, Sample Num: 10736, Cur Loss: 0.38788781, Cur Avg Loss: 0.88756363, Log Avg loss: 0.70909434, Global Avg Loss: 2.92611943, Time: 0.0067 Steps: 24090, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000681, Sample Num: 10896, Cur Loss: 1.06307781, Cur Avg Loss: 0.88799864, Log Avg loss: 0.91718827, Global Avg Loss: 2.92528585, Time: 0.0156 Steps: 24100, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000691, Sample Num: 11056, Cur Loss: 1.45731783, Cur Avg Loss: 0.88996723, Log Avg loss: 1.02402800, Global Avg Loss: 2.92449727, Time: 0.0067 Steps: 24110, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000701, Sample Num: 11216, Cur Loss: 0.57666701, Cur Avg Loss: 0.88993001, Log Avg loss: 0.88735819, Global Avg Loss: 2.92365269, Time: 0.0115 Steps: 24120, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000711, Sample Num: 11376, Cur Loss: 0.51327920, Cur Avg Loss: 0.88803208, Log Avg loss: 0.75498718, Global Avg Loss: 2.92275395, Time: 0.0072 Steps: 24130, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000721, Sample Num: 11536, Cur Loss: 0.65112996, Cur Avg Loss: 0.88566799, Log Avg loss: 0.71758094, Global Avg Loss: 2.92184045, Time: 0.0067 Steps: 24140, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000731, Sample Num: 11696, Cur Loss: 1.20789218, Cur Avg Loss: 0.88568623, Log Avg loss: 0.88700165, Global Avg Loss: 2.92099787, Time: 0.0065 Steps: 24150, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000741, Sample Num: 11856, Cur Loss: 0.76391542, Cur Avg Loss: 0.88405388, Log Avg loss: 0.76472920, Global Avg Loss: 2.92010537, Time: 0.0070 Steps: 24160, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000751, Sample Num: 12016, Cur Loss: 0.49604622, Cur Avg Loss: 0.88457934, Log Avg loss: 0.92351599, Global Avg Loss: 2.91927931, Time: 0.0146 Steps: 24170, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000761, Sample Num: 12176, Cur Loss: 1.15471864, Cur Avg Loss: 0.88712842, Log Avg loss: 1.07856421, Global Avg Loss: 2.91851806, Time: 0.0118 Steps: 24180, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000771, Sample Num: 12336, Cur Loss: 0.69650131, Cur Avg Loss: 0.88708897, Log Avg loss: 0.88408642, Global Avg Loss: 2.91767704, Time: 0.0128 Steps: 24190, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000781, Sample Num: 12496, Cur Loss: 1.41927576, Cur Avg Loss: 0.88874421, Log Avg loss: 1.01636351, Global Avg Loss: 2.91689137, Time: 0.0115 Steps: 24200, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000791, Sample Num: 12656, Cur Loss: 0.96359336, Cur Avg Loss: 0.88970765, Log Avg loss: 0.96495242, Global Avg Loss: 2.91608512, Time: 0.0148 Steps: 24210, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000801, Sample Num: 12816, Cur Loss: 1.18587589, Cur Avg Loss: 0.89121823, Log Avg loss: 1.01070508, Global Avg Loss: 2.91529842, Time: 0.0106 Steps: 24220, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000811, Sample Num: 12976, Cur Loss: 1.18349087, Cur Avg Loss: 0.89076116, Log Avg loss: 0.85414969, Global Avg Loss: 2.91444776, Time: 0.0071 Steps: 24230, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000821, Sample Num: 13136, Cur Loss: 1.37617564, Cur Avg Loss: 0.89258693, Log Avg loss: 1.04065657, Global Avg Loss: 2.91367474, Time: 0.0153 Steps: 24240, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000831, Sample Num: 13296, Cur Loss: 0.70153582, Cur Avg Loss: 0.89300567, Log Avg loss: 0.92738433, Global Avg Loss: 2.91285565, Time: 0.0152 Steps: 24250, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000841, Sample Num: 13456, Cur Loss: 1.00375140, Cur Avg Loss: 0.89194958, Log Avg loss: 0.80418869, Global Avg Loss: 2.91198646, Time: 0.0221 Steps: 24260, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000851, Sample Num: 13616, Cur Loss: 1.25227892, Cur Avg Loss: 0.88904372, Log Avg loss: 0.64466099, Global Avg Loss: 2.91105225, Time: 0.0123 Steps: 24270, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000861, Sample Num: 13776, Cur Loss: 0.48515943, Cur Avg Loss: 0.88859206, Log Avg loss: 0.85015539, Global Avg Loss: 2.91020345, Time: 0.0110 Steps: 24280, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000871, Sample Num: 13936, Cur Loss: 1.24480736, Cur Avg Loss: 0.88859293, Log Avg loss: 0.88866782, Global Avg Loss: 2.90937120, Time: 0.0131 Steps: 24290, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000881, Sample Num: 14096, Cur Loss: 1.00280309, Cur Avg Loss: 0.88959292, Log Avg loss: 0.97669279, Global Avg Loss: 2.90857585, Time: 0.0064 Steps: 24300, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000891, Sample Num: 14256, Cur Loss: 1.37675786, Cur Avg Loss: 0.89111754, Log Avg loss: 1.02543634, Global Avg Loss: 2.90780122, Time: 0.0068 Steps: 24310, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000901, Sample Num: 14416, Cur Loss: 1.00812006, Cur Avg Loss: 0.89173105, Log Avg loss: 0.94639501, Global Avg Loss: 2.90699472, Time: 0.0083 Steps: 24320, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000911, Sample Num: 14576, Cur Loss: 0.80903912, Cur Avg Loss: 0.89094519, Log Avg loss: 0.82013900, Global Avg Loss: 2.90613699, Time: 0.0114 Steps: 24330, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000921, Sample Num: 14736, Cur Loss: 0.37703979, Cur Avg Loss: 0.89152894, Log Avg loss: 0.94470832, Global Avg Loss: 2.90533114, Time: 0.0114 Steps: 24340, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000931, Sample Num: 14896, Cur Loss: 0.75653696, Cur Avg Loss: 0.89256024, Log Avg loss: 0.98754308, Global Avg Loss: 2.90454355, Time: 0.0109 Steps: 24350, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000941, Sample Num: 15056, Cur Loss: 0.87961185, Cur Avg Loss: 0.89369413, Log Avg loss: 0.99925967, Global Avg Loss: 2.90376142, Time: 0.0117 Steps: 24360, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000951, Sample Num: 15216, Cur Loss: 1.14598751, Cur Avg Loss: 0.89455121, Log Avg loss: 0.97520244, Global Avg Loss: 2.90297005, Time: 0.0181 Steps: 24370, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000961, Sample Num: 15376, Cur Loss: 0.53204346, Cur Avg Loss: 0.89459349, Log Avg loss: 0.89861413, Global Avg Loss: 2.90214792, Time: 0.0070 Steps: 24380, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000971, Sample Num: 15536, Cur Loss: 0.65370893, Cur Avg Loss: 0.89313233, Log Avg loss: 0.75271496, Global Avg Loss: 2.90126664, Time: 0.0108 Steps: 24390, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000981, Sample Num: 15696, Cur Loss: 0.91622329, Cur Avg Loss: 0.89273203, Log Avg loss: 0.85386268, Global Avg Loss: 2.90042754, Time: 0.0126 Steps: 24400, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 000991, Sample Num: 15856, Cur Loss: 0.87820852, Cur Avg Loss: 0.89260145, Log Avg loss: 0.87979174, Global Avg Loss: 2.89959975, Time: 0.0120 Steps: 24410, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001001, Sample Num: 16016, Cur Loss: 2.16953230, Cur Avg Loss: 0.89320787, Log Avg loss: 0.95330357, Global Avg Loss: 2.89880274, Time: 0.0112 Steps: 24420, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001011, Sample Num: 16176, Cur Loss: 0.15839757, Cur Avg Loss: 0.89163846, Log Avg loss: 0.73454073, Global Avg Loss: 2.89791684, Time: 0.0139 Steps: 24430, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001021, Sample Num: 16336, Cur Loss: 0.62947136, Cur Avg Loss: 0.89355748, Log Avg loss: 1.08757074, Global Avg Loss: 2.89717611, Time: 0.0114 Steps: 24440, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001031, Sample Num: 16496, Cur Loss: 1.07836318, Cur Avg Loss: 0.89434903, Log Avg loss: 0.97516592, Global Avg Loss: 2.89639001, Time: 0.0116 Steps: 24450, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001041, Sample Num: 16656, Cur Loss: 0.84193587, Cur Avg Loss: 0.89419963, Log Avg loss: 0.87879683, Global Avg Loss: 2.89556516, Time: 0.0106 Steps: 24460, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001051, Sample Num: 16816, Cur Loss: 1.71989465, Cur Avg Loss: 0.89494268, Log Avg loss: 0.97229372, Global Avg Loss: 2.89477919, Time: 0.0122 Steps: 24470, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001061, Sample Num: 16976, Cur Loss: 0.44973496, Cur Avg Loss: 0.89599998, Log Avg loss: 1.00712229, Global Avg Loss: 2.89400808, Time: 0.0084 Steps: 24480, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001071, Sample Num: 17136, Cur Loss: 0.46273184, Cur Avg Loss: 0.89515589, Log Avg loss: 0.80559849, Global Avg Loss: 2.89315532, Time: 0.0066 Steps: 24490, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001081, Sample Num: 17296, Cur Loss: 0.54377186, Cur Avg Loss: 0.89449096, Log Avg loss: 0.82327650, Global Avg Loss: 2.89231048, Time: 0.0071 Steps: 24500, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001091, Sample Num: 17456, Cur Loss: 0.69588375, Cur Avg Loss: 0.89426169, Log Avg loss: 0.86947787, Global Avg Loss: 2.89148517, Time: 0.0159 Steps: 24510, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001101, Sample Num: 17616, Cur Loss: 0.45787215, Cur Avg Loss: 0.89492471, Log Avg loss: 0.96725976, Global Avg Loss: 2.89070041, Time: 0.0116 Steps: 24520, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001111, Sample Num: 17776, Cur Loss: 1.29626703, Cur Avg Loss: 0.89526151, Log Avg loss: 0.93234391, Global Avg Loss: 2.88990206, Time: 0.0071 Steps: 24530, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001121, Sample Num: 17936, Cur Loss: 0.72870660, Cur Avg Loss: 0.89652832, Log Avg loss: 1.03727012, Global Avg Loss: 2.88914711, Time: 0.0070 Steps: 24540, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001131, Sample Num: 18096, Cur Loss: 1.18072844, Cur Avg Loss: 0.89776980, Log Avg loss: 1.03694053, Global Avg Loss: 2.88839265, Time: 0.0111 Steps: 24550, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001141, Sample Num: 18256, Cur Loss: 1.33496833, Cur Avg Loss: 0.89673142, Log Avg loss: 0.77929013, Global Avg Loss: 2.88753390, Time: 0.0066 Steps: 24560, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001151, Sample Num: 18416, Cur Loss: 0.84411943, Cur Avg Loss: 0.89795549, Log Avg loss: 1.03762166, Global Avg Loss: 2.88678098, Time: 0.0111 Steps: 24570, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001161, Sample Num: 18576, Cur Loss: 0.63185859, Cur Avg Loss: 0.89849217, Log Avg loss: 0.96026365, Global Avg Loss: 2.88599721, Time: 0.0152 Steps: 24580, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001171, Sample Num: 18736, Cur Loss: 1.27621400, Cur Avg Loss: 0.89749707, Log Avg loss: 0.78196711, Global Avg Loss: 2.88514156, Time: 0.0119 Steps: 24590, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001181, Sample Num: 18896, Cur Loss: 0.37407434, Cur Avg Loss: 0.89566570, Log Avg loss: 0.68121130, Global Avg Loss: 2.88424565, Time: 0.0105 Steps: 24600, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001191, Sample Num: 19056, Cur Loss: 1.02482891, Cur Avg Loss: 0.89767575, Log Avg loss: 1.13506284, Global Avg Loss: 2.88353489, Time: 0.0118 Steps: 24610, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001201, Sample Num: 19216, Cur Loss: 1.74194515, Cur Avg Loss: 0.89647053, Log Avg loss: 0.75292865, Global Avg Loss: 2.88266950, Time: 0.0131 Steps: 24620, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001211, Sample Num: 19376, Cur Loss: 0.29577637, Cur Avg Loss: 0.89688835, Log Avg loss: 0.94706891, Global Avg Loss: 2.88188363, Time: 0.0117 Steps: 24630, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001221, Sample Num: 19536, Cur Loss: 0.58578300, Cur Avg Loss: 0.89733644, Log Avg loss: 0.95159965, Global Avg Loss: 2.88110023, Time: 0.0107 Steps: 24640, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001231, Sample Num: 19696, Cur Loss: 0.24450442, Cur Avg Loss: 0.89650455, Log Avg loss: 0.79493123, Global Avg Loss: 2.88025392, Time: 0.0066 Steps: 24650, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001241, Sample Num: 19856, Cur Loss: 0.56497681, Cur Avg Loss: 0.89391393, Log Avg loss: 0.57500900, Global Avg Loss: 2.87931910, Time: 0.0110 Steps: 24660, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001251, Sample Num: 20016, Cur Loss: 0.72543800, Cur Avg Loss: 0.89471324, Log Avg loss: 0.99390744, Global Avg Loss: 2.87855485, Time: 0.0065 Steps: 24670, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001261, Sample Num: 20176, Cur Loss: 1.04149246, Cur Avg Loss: 0.89452724, Log Avg loss: 0.87125897, Global Avg Loss: 2.87774152, Time: 0.0070 Steps: 24680, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001271, Sample Num: 20336, Cur Loss: 1.04188025, Cur Avg Loss: 0.89495199, Log Avg loss: 0.94851272, Global Avg Loss: 2.87696014, Time: 0.0068 Steps: 24690, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001281, Sample Num: 20496, Cur Loss: 0.22686192, Cur Avg Loss: 0.89382600, Log Avg loss: 0.75071203, Global Avg Loss: 2.87609931, Time: 0.0094 Steps: 24700, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001291, Sample Num: 20656, Cur Loss: 1.61289024, Cur Avg Loss: 0.89548328, Log Avg loss: 1.10778086, Global Avg Loss: 2.87538368, Time: 0.0121 Steps: 24710, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001301, Sample Num: 20816, Cur Loss: 0.36871386, Cur Avg Loss: 0.89496686, Log Avg loss: 0.82829703, Global Avg Loss: 2.87455557, Time: 0.0068 Steps: 24720, Updated lr: 0.000078 Training, Epoch: 0012, Batch: 001311, Sample Num: 20976, Cur Loss: 0.64271867, Cur Avg Loss: 0.89388671, Log Avg loss: 0.75335970, Global Avg Loss: 2.87369783, Time: 0.0141 Steps: 24730, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001321, Sample Num: 21136, Cur Loss: 1.08656597, Cur Avg Loss: 0.89370420, Log Avg loss: 0.86977659, Global Avg Loss: 2.87288784, Time: 0.0068 Steps: 24740, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001331, Sample Num: 21296, Cur Loss: 0.71447510, Cur Avg Loss: 0.89408905, Log Avg loss: 0.94492873, Global Avg Loss: 2.87210887, Time: 0.0099 Steps: 24750, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001341, Sample Num: 21456, Cur Loss: 1.18700624, Cur Avg Loss: 0.89423423, Log Avg loss: 0.91355705, Global Avg Loss: 2.87131785, Time: 0.0122 Steps: 24760, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001351, Sample Num: 21616, Cur Loss: 1.07279074, Cur Avg Loss: 0.89342490, Log Avg loss: 0.78489369, Global Avg Loss: 2.87047553, Time: 0.0071 Steps: 24770, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001361, Sample Num: 21776, Cur Loss: 0.97256076, Cur Avg Loss: 0.89431284, Log Avg loss: 1.01427390, Global Avg Loss: 2.86972646, Time: 0.0066 Steps: 24780, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001371, Sample Num: 21936, Cur Loss: 0.64871156, Cur Avg Loss: 0.89455045, Log Avg loss: 0.92688938, Global Avg Loss: 2.86894274, Time: 0.0198 Steps: 24790, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001381, Sample Num: 22096, Cur Loss: 1.34562159, Cur Avg Loss: 0.89506820, Log Avg loss: 0.96605123, Global Avg Loss: 2.86817545, Time: 0.0115 Steps: 24800, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001391, Sample Num: 22256, Cur Loss: 0.44796711, Cur Avg Loss: 0.89595924, Log Avg loss: 1.01901146, Global Avg Loss: 2.86743012, Time: 0.0065 Steps: 24810, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001401, Sample Num: 22416, Cur Loss: 0.67730290, Cur Avg Loss: 0.89511845, Log Avg loss: 0.77816540, Global Avg Loss: 2.86658835, Time: 0.0110 Steps: 24820, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001411, Sample Num: 22576, Cur Loss: 0.60810721, Cur Avg Loss: 0.89463155, Log Avg loss: 0.82641705, Global Avg Loss: 2.86576669, Time: 0.0116 Steps: 24830, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001421, Sample Num: 22736, Cur Loss: 0.57906127, Cur Avg Loss: 0.89497268, Log Avg loss: 0.94310563, Global Avg Loss: 2.86499268, Time: 0.0069 Steps: 24840, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001431, Sample Num: 22896, Cur Loss: 0.83388090, Cur Avg Loss: 0.89472690, Log Avg loss: 0.85980212, Global Avg Loss: 2.86418576, Time: 0.0123 Steps: 24850, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001441, Sample Num: 23056, Cur Loss: 0.93525469, Cur Avg Loss: 0.89373764, Log Avg loss: 0.75217326, Global Avg Loss: 2.86333620, Time: 0.0101 Steps: 24860, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001451, Sample Num: 23216, Cur Loss: 0.67583078, Cur Avg Loss: 0.89368201, Log Avg loss: 0.88566580, Global Avg Loss: 2.86254099, Time: 0.0068 Steps: 24870, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001461, Sample Num: 23376, Cur Loss: 0.85717499, Cur Avg Loss: 0.89470104, Log Avg loss: 1.04256275, Global Avg Loss: 2.86180949, Time: 0.0065 Steps: 24880, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001471, Sample Num: 23536, Cur Loss: 0.63166237, Cur Avg Loss: 0.89320533, Log Avg loss: 0.67468199, Global Avg Loss: 2.86093077, Time: 0.0135 Steps: 24890, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001481, Sample Num: 23696, Cur Loss: 1.31946278, Cur Avg Loss: 0.89317914, Log Avg loss: 0.88932735, Global Avg Loss: 2.86013896, Time: 0.0080 Steps: 24900, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001491, Sample Num: 23856, Cur Loss: 0.93584597, Cur Avg Loss: 0.89333044, Log Avg loss: 0.91573737, Global Avg Loss: 2.85935839, Time: 0.0110 Steps: 24910, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001501, Sample Num: 24016, Cur Loss: 1.79384971, Cur Avg Loss: 0.89443206, Log Avg loss: 1.05868331, Global Avg Loss: 2.85863581, Time: 0.0067 Steps: 24920, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001511, Sample Num: 24176, Cur Loss: 1.15270865, Cur Avg Loss: 0.89546251, Log Avg loss: 1.05013356, Global Avg Loss: 2.85791038, Time: 0.0067 Steps: 24930, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001521, Sample Num: 24336, Cur Loss: 0.52722740, Cur Avg Loss: 0.89555662, Log Avg loss: 0.90977664, Global Avg Loss: 2.85712925, Time: 0.0067 Steps: 24940, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001531, Sample Num: 24496, Cur Loss: 0.38433823, Cur Avg Loss: 0.89546095, Log Avg loss: 0.88090886, Global Avg Loss: 2.85633718, Time: 0.0133 Steps: 24950, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001541, Sample Num: 24656, Cur Loss: 2.23949528, Cur Avg Loss: 0.89546992, Log Avg loss: 0.89684362, Global Avg Loss: 2.85555213, Time: 0.0188 Steps: 24960, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001551, Sample Num: 24816, Cur Loss: 1.48049879, Cur Avg Loss: 0.89723490, Log Avg loss: 1.16921814, Global Avg Loss: 2.85487678, Time: 0.0121 Steps: 24970, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001561, Sample Num: 24976, Cur Loss: 0.73966539, Cur Avg Loss: 0.89741293, Log Avg loss: 0.92502581, Global Avg Loss: 2.85410422, Time: 0.0076 Steps: 24980, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001571, Sample Num: 25136, Cur Loss: 1.43999279, Cur Avg Loss: 0.89799024, Log Avg loss: 0.98810844, Global Avg Loss: 2.85335753, Time: 0.0068 Steps: 24990, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001581, Sample Num: 25296, Cur Loss: 0.24510157, Cur Avg Loss: 0.89798021, Log Avg loss: 0.89640413, Global Avg Loss: 2.85257474, Time: 0.0135 Steps: 25000, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001591, Sample Num: 25456, Cur Loss: 1.92694890, Cur Avg Loss: 0.89941326, Log Avg loss: 1.12597937, Global Avg Loss: 2.85188438, Time: 0.0119 Steps: 25010, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001601, Sample Num: 25616, Cur Loss: 0.78603226, Cur Avg Loss: 0.89896817, Log Avg loss: 0.82815343, Global Avg Loss: 2.85107554, Time: 0.0115 Steps: 25020, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001611, Sample Num: 25776, Cur Loss: 1.07314372, Cur Avg Loss: 0.89840567, Log Avg loss: 0.80834973, Global Avg Loss: 2.85025943, Time: 0.0095 Steps: 25030, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001621, Sample Num: 25936, Cur Loss: 1.36422777, Cur Avg Loss: 0.89931630, Log Avg loss: 1.04601855, Global Avg Loss: 2.84953888, Time: 0.0126 Steps: 25040, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001631, Sample Num: 26096, Cur Loss: 0.98848850, Cur Avg Loss: 0.89955339, Log Avg loss: 0.93798515, Global Avg Loss: 2.84877579, Time: 0.0085 Steps: 25050, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001641, Sample Num: 26256, Cur Loss: 0.74625009, Cur Avg Loss: 0.89776235, Log Avg loss: 0.60564409, Global Avg Loss: 2.84788068, Time: 0.0106 Steps: 25060, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001651, Sample Num: 26416, Cur Loss: 0.73810637, Cur Avg Loss: 0.89757024, Log Avg loss: 0.86604500, Global Avg Loss: 2.84709016, Time: 0.0065 Steps: 25070, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001661, Sample Num: 26576, Cur Loss: 0.40733835, Cur Avg Loss: 0.89658008, Log Avg loss: 0.73310487, Global Avg Loss: 2.84624727, Time: 0.0114 Steps: 25080, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001671, Sample Num: 26736, Cur Loss: 0.87930936, Cur Avg Loss: 0.89591801, Log Avg loss: 0.78594838, Global Avg Loss: 2.84542610, Time: 0.0082 Steps: 25090, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001681, Sample Num: 26896, Cur Loss: 0.87494171, Cur Avg Loss: 0.89564156, Log Avg loss: 0.84944696, Global Avg Loss: 2.84463089, Time: 0.0090 Steps: 25100, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001691, Sample Num: 27056, Cur Loss: 1.00704420, Cur Avg Loss: 0.89591313, Log Avg loss: 0.94156356, Global Avg Loss: 2.84387300, Time: 0.0117 Steps: 25110, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001701, Sample Num: 27216, Cur Loss: 0.95874506, Cur Avg Loss: 0.89531976, Log Avg loss: 0.79498056, Global Avg Loss: 2.84305736, Time: 0.0110 Steps: 25120, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001711, Sample Num: 27376, Cur Loss: 0.32609499, Cur Avg Loss: 0.89413424, Log Avg loss: 0.69247768, Global Avg Loss: 2.84220158, Time: 0.0112 Steps: 25130, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001721, Sample Num: 27536, Cur Loss: 1.62847877, Cur Avg Loss: 0.89433761, Log Avg loss: 0.92913388, Global Avg Loss: 2.84144061, Time: 0.0137 Steps: 25140, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001731, Sample Num: 27696, Cur Loss: 0.43663657, Cur Avg Loss: 0.89378275, Log Avg loss: 0.79829185, Global Avg Loss: 2.84062822, Time: 0.0154 Steps: 25150, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001741, Sample Num: 27856, Cur Loss: 0.30908167, Cur Avg Loss: 0.89381616, Log Avg loss: 0.89959944, Global Avg Loss: 2.83985675, Time: 0.0115 Steps: 25160, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001751, Sample Num: 28016, Cur Loss: 0.52130085, Cur Avg Loss: 0.89432814, Log Avg loss: 0.98346434, Global Avg Loss: 2.83911921, Time: 0.0134 Steps: 25170, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001761, Sample Num: 28176, Cur Loss: 1.05204797, Cur Avg Loss: 0.89402894, Log Avg loss: 0.84163837, Global Avg Loss: 2.83832593, Time: 0.0114 Steps: 25180, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001771, Sample Num: 28336, Cur Loss: 0.90865445, Cur Avg Loss: 0.89408649, Log Avg loss: 0.90422049, Global Avg Loss: 2.83755812, Time: 0.0111 Steps: 25190, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001781, Sample Num: 28496, Cur Loss: 0.57754397, Cur Avg Loss: 0.89377007, Log Avg loss: 0.83773274, Global Avg Loss: 2.83676454, Time: 0.0082 Steps: 25200, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001791, Sample Num: 28656, Cur Loss: 0.75115234, Cur Avg Loss: 0.89363638, Log Avg loss: 0.86982644, Global Avg Loss: 2.83598432, Time: 0.0067 Steps: 25210, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001801, Sample Num: 28816, Cur Loss: 0.41546905, Cur Avg Loss: 0.89364941, Log Avg loss: 0.89598351, Global Avg Loss: 2.83521509, Time: 0.0089 Steps: 25220, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001811, Sample Num: 28976, Cur Loss: 1.31317067, Cur Avg Loss: 0.89380727, Log Avg loss: 0.92223774, Global Avg Loss: 2.83445687, Time: 0.0067 Steps: 25230, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001821, Sample Num: 29136, Cur Loss: 0.87514579, Cur Avg Loss: 0.89420058, Log Avg loss: 0.96542887, Global Avg Loss: 2.83371637, Time: 0.0109 Steps: 25240, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001831, Sample Num: 29296, Cur Loss: 0.79417133, Cur Avg Loss: 0.89512348, Log Avg loss: 1.06318253, Global Avg Loss: 2.83301517, Time: 0.0069 Steps: 25250, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001841, Sample Num: 29456, Cur Loss: 1.22414970, Cur Avg Loss: 0.89679216, Log Avg loss: 1.20232847, Global Avg Loss: 2.83236961, Time: 0.0068 Steps: 25260, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001851, Sample Num: 29616, Cur Loss: 0.77234221, Cur Avg Loss: 0.89758696, Log Avg loss: 1.04390958, Global Avg Loss: 2.83166187, Time: 0.0073 Steps: 25270, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001861, Sample Num: 29776, Cur Loss: 1.50927806, Cur Avg Loss: 0.89816813, Log Avg loss: 1.00574310, Global Avg Loss: 2.83093959, Time: 0.0067 Steps: 25280, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001871, Sample Num: 29936, Cur Loss: 1.23449898, Cur Avg Loss: 0.89711365, Log Avg loss: 0.70087423, Global Avg Loss: 2.83009733, Time: 0.0114 Steps: 25290, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001881, Sample Num: 30096, Cur Loss: 1.39426255, Cur Avg Loss: 0.89722320, Log Avg loss: 0.91771977, Global Avg Loss: 2.82934145, Time: 0.0108 Steps: 25300, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001891, Sample Num: 30256, Cur Loss: 0.37477967, Cur Avg Loss: 0.89613873, Log Avg loss: 0.69214961, Global Avg Loss: 2.82849705, Time: 0.0068 Steps: 25310, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001901, Sample Num: 30416, Cur Loss: 1.05518520, Cur Avg Loss: 0.89680518, Log Avg loss: 1.02283158, Global Avg Loss: 2.82778391, Time: 0.0074 Steps: 25320, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001911, Sample Num: 30576, Cur Loss: 0.61409706, Cur Avg Loss: 0.89719753, Log Avg loss: 0.97178319, Global Avg Loss: 2.82705118, Time: 0.0119 Steps: 25330, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001921, Sample Num: 30736, Cur Loss: 0.56189954, Cur Avg Loss: 0.89566212, Log Avg loss: 0.60224537, Global Avg Loss: 2.82617320, Time: 0.0199 Steps: 25340, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001931, Sample Num: 30896, Cur Loss: 0.91186440, Cur Avg Loss: 0.89449242, Log Avg loss: 0.66979195, Global Avg Loss: 2.82532255, Time: 0.0065 Steps: 25350, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001941, Sample Num: 31056, Cur Loss: 1.74022114, Cur Avg Loss: 0.89479110, Log Avg loss: 0.95246741, Global Avg Loss: 2.82458405, Time: 0.0070 Steps: 25360, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001951, Sample Num: 31216, Cur Loss: 0.45534390, Cur Avg Loss: 0.89411897, Log Avg loss: 0.76365820, Global Avg Loss: 2.82377170, Time: 0.0112 Steps: 25370, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001961, Sample Num: 31376, Cur Loss: 0.76456678, Cur Avg Loss: 0.89390689, Log Avg loss: 0.85253031, Global Avg Loss: 2.82299501, Time: 0.0098 Steps: 25380, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001971, Sample Num: 31536, Cur Loss: 0.59004784, Cur Avg Loss: 0.89473102, Log Avg loss: 1.05634253, Global Avg Loss: 2.82229920, Time: 0.0068 Steps: 25390, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001981, Sample Num: 31696, Cur Loss: 0.71327394, Cur Avg Loss: 0.89556076, Log Avg loss: 1.05910224, Global Avg Loss: 2.82160503, Time: 0.0064 Steps: 25400, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 001991, Sample Num: 31856, Cur Loss: 0.92813993, Cur Avg Loss: 0.89534737, Log Avg loss: 0.85307578, Global Avg Loss: 2.82083032, Time: 0.0108 Steps: 25410, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002001, Sample Num: 32016, Cur Loss: 0.36307794, Cur Avg Loss: 0.89506362, Log Avg loss: 0.83856818, Global Avg Loss: 2.82005052, Time: 0.0097 Steps: 25420, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002011, Sample Num: 32176, Cur Loss: 0.91049415, Cur Avg Loss: 0.89470873, Log Avg loss: 0.82369590, Global Avg Loss: 2.81926548, Time: 0.0112 Steps: 25430, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002021, Sample Num: 32336, Cur Loss: 0.66422778, Cur Avg Loss: 0.89469494, Log Avg loss: 0.89192142, Global Avg Loss: 2.81850788, Time: 0.0067 Steps: 25440, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002031, Sample Num: 32496, Cur Loss: 0.32362843, Cur Avg Loss: 0.89367789, Log Avg loss: 0.68813216, Global Avg Loss: 2.81767079, Time: 0.0066 Steps: 25450, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002041, Sample Num: 32656, Cur Loss: 1.16199851, Cur Avg Loss: 0.89380732, Log Avg loss: 0.92009450, Global Avg Loss: 2.81692548, Time: 0.0067 Steps: 25460, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002051, Sample Num: 32816, Cur Loss: 1.12757516, Cur Avg Loss: 0.89301907, Log Avg loss: 0.73213624, Global Avg Loss: 2.81610695, Time: 0.0084 Steps: 25470, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002061, Sample Num: 32976, Cur Loss: 0.34715608, Cur Avg Loss: 0.89211523, Log Avg loss: 0.70673878, Global Avg Loss: 2.81527910, Time: 0.0131 Steps: 25480, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002071, Sample Num: 33136, Cur Loss: 1.23648167, Cur Avg Loss: 0.89105625, Log Avg loss: 0.67280037, Global Avg Loss: 2.81443858, Time: 0.0071 Steps: 25490, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002081, Sample Num: 33296, Cur Loss: 1.63233125, Cur Avg Loss: 0.89157208, Log Avg loss: 0.99840003, Global Avg Loss: 2.81372641, Time: 0.0119 Steps: 25500, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002091, Sample Num: 33456, Cur Loss: 0.55627227, Cur Avg Loss: 0.89038516, Log Avg loss: 0.64338662, Global Avg Loss: 2.81287563, Time: 0.0068 Steps: 25510, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002101, Sample Num: 33616, Cur Loss: 0.76789564, Cur Avg Loss: 0.88990626, Log Avg loss: 0.78976918, Global Avg Loss: 2.81208287, Time: 0.0069 Steps: 25520, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002111, Sample Num: 33776, Cur Loss: 0.79278588, Cur Avg Loss: 0.88977634, Log Avg loss: 0.86247923, Global Avg Loss: 2.81131922, Time: 0.0110 Steps: 25530, Updated lr: 0.000077 Training, Epoch: 0012, Batch: 002121, Sample Num: 33936, Cur Loss: 1.17673230, Cur Avg Loss: 0.88893272, Log Avg loss: 0.71084550, Global Avg Loss: 2.81049680, Time: 0.0067 Steps: 25540, Updated lr: 0.000077 ***** Running evaluation checkpoint-25548 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-25548 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.283795, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.93739, "eval_total_loss": 658.985128, "eval_mae": 0.783994, "eval_mse": 0.937546, "eval_r2": 0.404034, "eval_sp_statistic": 0.616373, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.654142, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.707442, "test_total_loss": 857.136126, "test_mae": 1.183435, "test_mse": 1.707271, "test_r2": -0.101887, "test_sp_statistic": 0.370093, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.506977, "test_ps_pvalue": 0.0, "lr": 7.672072072072073e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.8099224767046196, "train_cur_epoch_loss": 1893.2375241816044, "train_cur_epoch_avg_loss": 0.8892614016822942, "train_cur_epoch_time": 22.28379464149475, "train_cur_epoch_avg_time": 0.0104667894041779, "epoch": 12, "step": 25548} ################################################## Training, Epoch: 0013, Batch: 000002, Sample Num: 32, Cur Loss: 0.52264345, Cur Avg Loss: 0.55518571, Log Avg loss: 0.89215955, Global Avg Loss: 2.80974598, Time: 0.0132 Steps: 25550, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000012, Sample Num: 192, Cur Loss: 0.18601941, Cur Avg Loss: 0.85207082, Log Avg loss: 0.91144784, Global Avg Loss: 2.80900330, Time: 0.0075 Steps: 25560, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000022, Sample Num: 352, Cur Loss: 0.94205832, Cur Avg Loss: 0.95294631, Log Avg loss: 1.07399690, Global Avg Loss: 2.80832477, Time: 0.0123 Steps: 25570, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000032, Sample Num: 512, Cur Loss: 0.51951939, Cur Avg Loss: 0.90929737, Log Avg loss: 0.81326970, Global Avg Loss: 2.80754484, Time: 0.0134 Steps: 25580, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000042, Sample Num: 672, Cur Loss: 1.02509153, Cur Avg Loss: 0.95373979, Log Avg loss: 1.09595554, Global Avg Loss: 2.80687599, Time: 0.0184 Steps: 25590, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000052, Sample Num: 832, Cur Loss: 0.68859851, Cur Avg Loss: 0.95253335, Log Avg loss: 0.94746630, Global Avg Loss: 2.80614966, Time: 0.0067 Steps: 25600, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000062, Sample Num: 992, Cur Loss: 0.76607299, Cur Avg Loss: 0.92088256, Log Avg loss: 0.75629848, Global Avg Loss: 2.80534924, Time: 0.0086 Steps: 25610, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000072, Sample Num: 1152, Cur Loss: 0.33294845, Cur Avg Loss: 0.90771441, Log Avg loss: 0.82607187, Global Avg Loss: 2.80457669, Time: 0.0066 Steps: 25620, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000082, Sample Num: 1312, Cur Loss: 0.74464881, Cur Avg Loss: 0.87117778, Log Avg loss: 0.60811403, Global Avg Loss: 2.80371970, Time: 0.0133 Steps: 25630, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000092, Sample Num: 1472, Cur Loss: 0.93311918, Cur Avg Loss: 0.87296218, Log Avg loss: 0.88759429, Global Avg Loss: 2.80297239, Time: 0.0068 Steps: 25640, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000102, Sample Num: 1632, Cur Loss: 0.27480540, Cur Avg Loss: 0.87879414, Log Avg loss: 0.93244816, Global Avg Loss: 2.80224314, Time: 0.0067 Steps: 25650, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000112, Sample Num: 1792, Cur Loss: 0.27561724, Cur Avg Loss: 0.89434075, Log Avg loss: 1.05291619, Global Avg Loss: 2.80156140, Time: 0.0067 Steps: 25660, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000122, Sample Num: 1952, Cur Loss: 1.38272381, Cur Avg Loss: 0.87732229, Log Avg loss: 0.68671552, Global Avg Loss: 2.80073754, Time: 0.0071 Steps: 25670, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000132, Sample Num: 2112, Cur Loss: 1.16511083, Cur Avg Loss: 0.89684028, Log Avg loss: 1.13495980, Global Avg Loss: 2.80008888, Time: 0.0086 Steps: 25680, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000142, Sample Num: 2272, Cur Loss: 0.24590987, Cur Avg Loss: 0.87942867, Log Avg loss: 0.64959532, Global Avg Loss: 2.79925178, Time: 0.0140 Steps: 25690, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000152, Sample Num: 2432, Cur Loss: 1.72639251, Cur Avg Loss: 0.88409295, Log Avg loss: 0.95032579, Global Avg Loss: 2.79853236, Time: 0.0130 Steps: 25700, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000162, Sample Num: 2592, Cur Loss: 1.14902043, Cur Avg Loss: 0.87639153, Log Avg loss: 0.75932998, Global Avg Loss: 2.79773920, Time: 0.0065 Steps: 25710, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000172, Sample Num: 2752, Cur Loss: 1.34787703, Cur Avg Loss: 0.87650488, Log Avg loss: 0.87834106, Global Avg Loss: 2.79699293, Time: 0.0067 Steps: 25720, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000182, Sample Num: 2912, Cur Loss: 1.46164894, Cur Avg Loss: 0.87572488, Log Avg loss: 0.86230898, Global Avg Loss: 2.79624102, Time: 0.0118 Steps: 25730, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000192, Sample Num: 3072, Cur Loss: 1.34725356, Cur Avg Loss: 0.87438734, Log Avg loss: 0.85004403, Global Avg Loss: 2.79548492, Time: 0.0108 Steps: 25740, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000202, Sample Num: 3232, Cur Loss: 0.27309930, Cur Avg Loss: 0.86347449, Log Avg loss: 0.65394768, Global Avg Loss: 2.79465325, Time: 0.0068 Steps: 25750, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000212, Sample Num: 3392, Cur Loss: 0.86408961, Cur Avg Loss: 0.87979077, Log Avg loss: 1.20937979, Global Avg Loss: 2.79403785, Time: 0.0157 Steps: 25760, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000222, Sample Num: 3552, Cur Loss: 1.27285480, Cur Avg Loss: 0.87698311, Log Avg loss: 0.81746063, Global Avg Loss: 2.79327085, Time: 0.0068 Steps: 25770, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000232, Sample Num: 3712, Cur Loss: 0.80547357, Cur Avg Loss: 0.87531763, Log Avg loss: 0.83834390, Global Avg Loss: 2.79251253, Time: 0.0067 Steps: 25780, Updated lr: 0.000077 Training, Epoch: 0013, Batch: 000242, Sample Num: 3872, Cur Loss: 0.90540373, Cur Avg Loss: 0.87963160, Log Avg loss: 0.97971574, Global Avg Loss: 2.79180963, Time: 0.0120 Steps: 25790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000252, Sample Num: 4032, Cur Loss: 1.60670877, Cur Avg Loss: 0.88165116, Log Avg loss: 0.93052461, Global Avg Loss: 2.79108820, Time: 0.0078 Steps: 25800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000262, Sample Num: 4192, Cur Loss: 1.07920241, Cur Avg Loss: 0.87986968, Log Avg loss: 0.83497631, Global Avg Loss: 2.79033031, Time: 0.0127 Steps: 25810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000272, Sample Num: 4352, Cur Loss: 0.34715658, Cur Avg Loss: 0.87813712, Log Avg loss: 0.83274419, Global Avg Loss: 2.78957214, Time: 0.0108 Steps: 25820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000282, Sample Num: 4512, Cur Loss: 0.27249876, Cur Avg Loss: 0.87758364, Log Avg loss: 0.86252894, Global Avg Loss: 2.78882609, Time: 0.0127 Steps: 25830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000292, Sample Num: 4672, Cur Loss: 1.27439332, Cur Avg Loss: 0.87753100, Log Avg loss: 0.87604639, Global Avg Loss: 2.78808585, Time: 0.0073 Steps: 25840, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000302, Sample Num: 4832, Cur Loss: 1.21632648, Cur Avg Loss: 0.87432969, Log Avg loss: 0.78085168, Global Avg Loss: 2.78730936, Time: 0.0068 Steps: 25850, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000312, Sample Num: 4992, Cur Loss: 0.70261300, Cur Avg Loss: 0.88269419, Log Avg loss: 1.13530184, Global Avg Loss: 2.78667053, Time: 0.0069 Steps: 25860, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000322, Sample Num: 5152, Cur Loss: 0.32465404, Cur Avg Loss: 0.87954809, Log Avg loss: 0.78139003, Global Avg Loss: 2.78589540, Time: 0.0098 Steps: 25870, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000332, Sample Num: 5312, Cur Loss: 0.58814323, Cur Avg Loss: 0.88332752, Log Avg loss: 1.00502508, Global Avg Loss: 2.78520727, Time: 0.0141 Steps: 25880, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000342, Sample Num: 5472, Cur Loss: 0.64480555, Cur Avg Loss: 0.88131992, Log Avg loss: 0.81466759, Global Avg Loss: 2.78444615, Time: 0.0068 Steps: 25890, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000352, Sample Num: 5632, Cur Loss: 1.05865240, Cur Avg Loss: 0.87778452, Log Avg loss: 0.75687397, Global Avg Loss: 2.78366330, Time: 0.0112 Steps: 25900, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000362, Sample Num: 5792, Cur Loss: 0.48645228, Cur Avg Loss: 0.87561390, Log Avg loss: 0.79920781, Global Avg Loss: 2.78289740, Time: 0.0117 Steps: 25910, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000372, Sample Num: 5952, Cur Loss: 0.98006928, Cur Avg Loss: 0.87499395, Log Avg loss: 0.85255179, Global Avg Loss: 2.78215267, Time: 0.0144 Steps: 25920, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000382, Sample Num: 6112, Cur Loss: 0.73550886, Cur Avg Loss: 0.87614211, Log Avg loss: 0.91885363, Global Avg Loss: 2.78143408, Time: 0.0119 Steps: 25930, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000392, Sample Num: 6272, Cur Loss: 0.96487284, Cur Avg Loss: 0.87997009, Log Avg loss: 1.02619905, Global Avg Loss: 2.78075743, Time: 0.0073 Steps: 25940, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000402, Sample Num: 6432, Cur Loss: 0.30130494, Cur Avg Loss: 0.87703804, Log Avg loss: 0.76210147, Global Avg Loss: 2.77997953, Time: 0.0074 Steps: 25950, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000412, Sample Num: 6592, Cur Loss: 1.33746791, Cur Avg Loss: 0.87138887, Log Avg loss: 0.64429241, Global Avg Loss: 2.77915684, Time: 0.0111 Steps: 25960, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000422, Sample Num: 6752, Cur Loss: 0.79508120, Cur Avg Loss: 0.86771914, Log Avg loss: 0.71652611, Global Avg Loss: 2.77836261, Time: 0.0217 Steps: 25970, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000432, Sample Num: 6912, Cur Loss: 0.68480074, Cur Avg Loss: 0.86428792, Log Avg loss: 0.71949079, Global Avg Loss: 2.77757012, Time: 0.0076 Steps: 25980, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000442, Sample Num: 7072, Cur Loss: 1.41213584, Cur Avg Loss: 0.86682990, Log Avg loss: 0.97664308, Global Avg Loss: 2.77687719, Time: 0.0110 Steps: 25990, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000452, Sample Num: 7232, Cur Loss: 1.32940829, Cur Avg Loss: 0.86869800, Log Avg loss: 0.95126801, Global Avg Loss: 2.77617504, Time: 0.0134 Steps: 26000, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000462, Sample Num: 7392, Cur Loss: 0.90225768, Cur Avg Loss: 0.87102764, Log Avg loss: 0.97632741, Global Avg Loss: 2.77548305, Time: 0.0116 Steps: 26010, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000472, Sample Num: 7552, Cur Loss: 0.51741540, Cur Avg Loss: 0.87507047, Log Avg loss: 1.06184950, Global Avg Loss: 2.77482447, Time: 0.0076 Steps: 26020, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000482, Sample Num: 7712, Cur Loss: 1.14189065, Cur Avg Loss: 0.87792227, Log Avg loss: 1.01252727, Global Avg Loss: 2.77414744, Time: 0.0072 Steps: 26030, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000492, Sample Num: 7872, Cur Loss: 0.74669170, Cur Avg Loss: 0.87740026, Log Avg loss: 0.85223917, Global Avg Loss: 2.77340938, Time: 0.0123 Steps: 26040, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000502, Sample Num: 8032, Cur Loss: 1.01042795, Cur Avg Loss: 0.88095021, Log Avg loss: 1.05560797, Global Avg Loss: 2.77274996, Time: 0.0124 Steps: 26050, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000512, Sample Num: 8192, Cur Loss: 0.40925378, Cur Avg Loss: 0.87982395, Log Avg loss: 0.82328529, Global Avg Loss: 2.77200189, Time: 0.0154 Steps: 26060, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000522, Sample Num: 8352, Cur Loss: 0.20233813, Cur Avg Loss: 0.87374560, Log Avg loss: 0.56253442, Global Avg Loss: 2.77115438, Time: 0.0111 Steps: 26070, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000532, Sample Num: 8512, Cur Loss: 0.43449569, Cur Avg Loss: 0.86671637, Log Avg loss: 0.49979066, Global Avg Loss: 2.77028346, Time: 0.0111 Steps: 26080, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000542, Sample Num: 8672, Cur Loss: 1.19174278, Cur Avg Loss: 0.86951839, Log Avg loss: 1.01858569, Global Avg Loss: 2.76961205, Time: 0.0066 Steps: 26090, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000552, Sample Num: 8832, Cur Loss: 0.23977235, Cur Avg Loss: 0.86716661, Log Avg loss: 0.73969982, Global Avg Loss: 2.76883431, Time: 0.0117 Steps: 26100, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000562, Sample Num: 8992, Cur Loss: 1.20364857, Cur Avg Loss: 0.86858554, Log Avg loss: 0.94691077, Global Avg Loss: 2.76813652, Time: 0.0127 Steps: 26110, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000572, Sample Num: 9152, Cur Loss: 0.51489317, Cur Avg Loss: 0.86856429, Log Avg loss: 0.86737018, Global Avg Loss: 2.76740881, Time: 0.0085 Steps: 26120, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000582, Sample Num: 9312, Cur Loss: 0.88234603, Cur Avg Loss: 0.86544905, Log Avg loss: 0.68725691, Global Avg Loss: 2.76661274, Time: 0.0140 Steps: 26130, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000592, Sample Num: 9472, Cur Loss: 0.87197953, Cur Avg Loss: 0.86731172, Log Avg loss: 0.97571917, Global Avg Loss: 2.76592762, Time: 0.0106 Steps: 26140, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000602, Sample Num: 9632, Cur Loss: 1.59983730, Cur Avg Loss: 0.86749776, Log Avg loss: 0.87851150, Global Avg Loss: 2.76520585, Time: 0.0108 Steps: 26150, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000612, Sample Num: 9792, Cur Loss: 0.70009851, Cur Avg Loss: 0.86866755, Log Avg loss: 0.93908905, Global Avg Loss: 2.76450780, Time: 0.0132 Steps: 26160, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000622, Sample Num: 9952, Cur Loss: 0.46958780, Cur Avg Loss: 0.86625331, Log Avg loss: 0.71850165, Global Avg Loss: 2.76372598, Time: 0.0067 Steps: 26170, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000632, Sample Num: 10112, Cur Loss: 0.49194318, Cur Avg Loss: 0.86670505, Log Avg loss: 0.89480326, Global Avg Loss: 2.76301211, Time: 0.0070 Steps: 26180, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000642, Sample Num: 10272, Cur Loss: 0.63107765, Cur Avg Loss: 0.86720257, Log Avg loss: 0.89864556, Global Avg Loss: 2.76230025, Time: 0.0157 Steps: 26190, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000652, Sample Num: 10432, Cur Loss: 0.65281647, Cur Avg Loss: 0.87054978, Log Avg loss: 1.08544098, Global Avg Loss: 2.76166022, Time: 0.0110 Steps: 26200, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000662, Sample Num: 10592, Cur Loss: 0.71414143, Cur Avg Loss: 0.86975274, Log Avg loss: 0.81778576, Global Avg Loss: 2.76091857, Time: 0.0115 Steps: 26210, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000672, Sample Num: 10752, Cur Loss: 0.53426880, Cur Avg Loss: 0.86948881, Log Avg loss: 0.85201684, Global Avg Loss: 2.76019054, Time: 0.0226 Steps: 26220, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000682, Sample Num: 10912, Cur Loss: 1.49066055, Cur Avg Loss: 0.86924259, Log Avg loss: 0.85269606, Global Avg Loss: 2.75946332, Time: 0.0115 Steps: 26230, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000692, Sample Num: 11072, Cur Loss: 1.28555429, Cur Avg Loss: 0.86770327, Log Avg loss: 0.76272221, Global Avg Loss: 2.75870237, Time: 0.0110 Steps: 26240, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000702, Sample Num: 11232, Cur Loss: 0.83882344, Cur Avg Loss: 0.86451536, Log Avg loss: 0.64391145, Global Avg Loss: 2.75789673, Time: 0.0105 Steps: 26250, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000712, Sample Num: 11392, Cur Loss: 1.15261960, Cur Avg Loss: 0.86369725, Log Avg loss: 0.80626649, Global Avg Loss: 2.75715354, Time: 0.0112 Steps: 26260, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000722, Sample Num: 11552, Cur Loss: 1.04618144, Cur Avg Loss: 0.86394552, Log Avg loss: 0.88162187, Global Avg Loss: 2.75643959, Time: 0.0063 Steps: 26270, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000732, Sample Num: 11712, Cur Loss: 0.81394225, Cur Avg Loss: 0.86233379, Log Avg loss: 0.74596743, Global Avg Loss: 2.75567457, Time: 0.0065 Steps: 26280, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000742, Sample Num: 11872, Cur Loss: 0.62188667, Cur Avg Loss: 0.86302517, Log Avg loss: 0.91363350, Global Avg Loss: 2.75497391, Time: 0.0086 Steps: 26290, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000752, Sample Num: 12032, Cur Loss: 0.80828309, Cur Avg Loss: 0.86142409, Log Avg loss: 0.74262396, Global Avg Loss: 2.75420876, Time: 0.0106 Steps: 26300, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000762, Sample Num: 12192, Cur Loss: 0.72493935, Cur Avg Loss: 0.86038315, Log Avg loss: 0.78210445, Global Avg Loss: 2.75345919, Time: 0.0104 Steps: 26310, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000772, Sample Num: 12352, Cur Loss: 1.21469307, Cur Avg Loss: 0.86215821, Log Avg loss: 0.99741845, Global Avg Loss: 2.75279201, Time: 0.0171 Steps: 26320, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000782, Sample Num: 12512, Cur Loss: 1.01571548, Cur Avg Loss: 0.86282619, Log Avg loss: 0.91439355, Global Avg Loss: 2.75209379, Time: 0.0112 Steps: 26330, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000792, Sample Num: 12672, Cur Loss: 0.63850403, Cur Avg Loss: 0.86308949, Log Avg loss: 0.88367990, Global Avg Loss: 2.75138445, Time: 0.0096 Steps: 26340, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000802, Sample Num: 12832, Cur Loss: 1.52380872, Cur Avg Loss: 0.86569789, Log Avg loss: 1.07228351, Global Avg Loss: 2.75074722, Time: 0.0111 Steps: 26350, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000812, Sample Num: 12992, Cur Loss: 0.60513705, Cur Avg Loss: 0.86638564, Log Avg loss: 0.92154279, Global Avg Loss: 2.75005328, Time: 0.0135 Steps: 26360, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000822, Sample Num: 13152, Cur Loss: 0.41859555, Cur Avg Loss: 0.86608998, Log Avg loss: 0.84208238, Global Avg Loss: 2.74932975, Time: 0.0110 Steps: 26370, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000832, Sample Num: 13312, Cur Loss: 1.25410533, Cur Avg Loss: 0.86583791, Log Avg loss: 0.84511764, Global Avg Loss: 2.74860791, Time: 0.0070 Steps: 26380, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000842, Sample Num: 13472, Cur Loss: 0.96465790, Cur Avg Loss: 0.86655524, Log Avg loss: 0.92623751, Global Avg Loss: 2.74791735, Time: 0.0067 Steps: 26390, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000852, Sample Num: 13632, Cur Loss: 1.25197089, Cur Avg Loss: 0.86730487, Log Avg loss: 0.93042383, Global Avg Loss: 2.74722891, Time: 0.0115 Steps: 26400, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000862, Sample Num: 13792, Cur Loss: 1.03414214, Cur Avg Loss: 0.86865103, Log Avg loss: 0.98334367, Global Avg Loss: 2.74656102, Time: 0.0120 Steps: 26410, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000872, Sample Num: 13952, Cur Loss: 1.25481701, Cur Avg Loss: 0.86752772, Log Avg loss: 0.77069786, Global Avg Loss: 2.74581316, Time: 0.0068 Steps: 26420, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000882, Sample Num: 14112, Cur Loss: 1.00278616, Cur Avg Loss: 0.86901505, Log Avg loss: 0.99871073, Global Avg Loss: 2.74515213, Time: 0.0134 Steps: 26430, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000892, Sample Num: 14272, Cur Loss: 0.67941886, Cur Avg Loss: 0.86921917, Log Avg loss: 0.88722241, Global Avg Loss: 2.74444943, Time: 0.0111 Steps: 26440, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000902, Sample Num: 14432, Cur Loss: 0.71546364, Cur Avg Loss: 0.87001814, Log Avg loss: 0.94128593, Global Avg Loss: 2.74376770, Time: 0.0069 Steps: 26450, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000912, Sample Num: 14592, Cur Loss: 0.73724365, Cur Avg Loss: 0.87080123, Log Avg loss: 0.94143609, Global Avg Loss: 2.74308655, Time: 0.0067 Steps: 26460, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000922, Sample Num: 14752, Cur Loss: 1.34408045, Cur Avg Loss: 0.87398273, Log Avg loss: 1.16413558, Global Avg Loss: 2.74249005, Time: 0.0109 Steps: 26470, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000932, Sample Num: 14912, Cur Loss: 0.43843937, Cur Avg Loss: 0.87320037, Log Avg loss: 0.80106709, Global Avg Loss: 2.74175688, Time: 0.0123 Steps: 26480, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000942, Sample Num: 15072, Cur Loss: 0.48373151, Cur Avg Loss: 0.87195476, Log Avg loss: 0.75586357, Global Avg Loss: 2.74100720, Time: 0.0106 Steps: 26490, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000952, Sample Num: 15232, Cur Loss: 0.89803755, Cur Avg Loss: 0.87040926, Log Avg loss: 0.72482380, Global Avg Loss: 2.74024638, Time: 0.0111 Steps: 26500, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000962, Sample Num: 15392, Cur Loss: 0.63583362, Cur Avg Loss: 0.86915278, Log Avg loss: 0.74953503, Global Avg Loss: 2.73949545, Time: 0.0067 Steps: 26510, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000972, Sample Num: 15552, Cur Loss: 1.35205138, Cur Avg Loss: 0.86712294, Log Avg loss: 0.67185277, Global Avg Loss: 2.73871580, Time: 0.0064 Steps: 26520, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000982, Sample Num: 15712, Cur Loss: 0.50849217, Cur Avg Loss: 0.86659918, Log Avg loss: 0.81569020, Global Avg Loss: 2.73799095, Time: 0.0096 Steps: 26530, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 000992, Sample Num: 15872, Cur Loss: 1.14606297, Cur Avg Loss: 0.86721080, Log Avg loss: 0.92727168, Global Avg Loss: 2.73730869, Time: 0.0111 Steps: 26540, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001002, Sample Num: 16032, Cur Loss: 0.71295971, Cur Avg Loss: 0.86480114, Log Avg loss: 0.62576229, Global Avg Loss: 2.73651338, Time: 0.0067 Steps: 26550, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001012, Sample Num: 16192, Cur Loss: 1.43337667, Cur Avg Loss: 0.86768461, Log Avg loss: 1.15660847, Global Avg Loss: 2.73591853, Time: 0.0171 Steps: 26560, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001022, Sample Num: 16352, Cur Loss: 0.47423583, Cur Avg Loss: 0.86617333, Log Avg loss: 0.71323171, Global Avg Loss: 2.73515727, Time: 0.0067 Steps: 26570, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001032, Sample Num: 16512, Cur Loss: 2.11158657, Cur Avg Loss: 0.86787164, Log Avg loss: 1.04143932, Global Avg Loss: 2.73452005, Time: 0.0108 Steps: 26580, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001042, Sample Num: 16672, Cur Loss: 0.71816033, Cur Avg Loss: 0.86599200, Log Avg loss: 0.67201245, Global Avg Loss: 2.73374438, Time: 0.0103 Steps: 26590, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001052, Sample Num: 16832, Cur Loss: 0.87009072, Cur Avg Loss: 0.86757342, Log Avg loss: 1.03235764, Global Avg Loss: 2.73310476, Time: 0.0099 Steps: 26600, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001062, Sample Num: 16992, Cur Loss: 0.88002372, Cur Avg Loss: 0.86626167, Log Avg loss: 0.72826588, Global Avg Loss: 2.73235135, Time: 0.0104 Steps: 26610, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001072, Sample Num: 17152, Cur Loss: 1.80882823, Cur Avg Loss: 0.86609694, Log Avg loss: 0.84860290, Global Avg Loss: 2.73164370, Time: 0.0104 Steps: 26620, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001082, Sample Num: 17312, Cur Loss: 1.31373930, Cur Avg Loss: 0.86860405, Log Avg loss: 1.13736556, Global Avg Loss: 2.73104502, Time: 0.0118 Steps: 26630, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001092, Sample Num: 17472, Cur Loss: 0.43626213, Cur Avg Loss: 0.86985162, Log Avg loss: 1.00483903, Global Avg Loss: 2.73039705, Time: 0.0075 Steps: 26640, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001102, Sample Num: 17632, Cur Loss: 1.09866071, Cur Avg Loss: 0.86990001, Log Avg loss: 0.87518434, Global Avg Loss: 2.72970091, Time: 0.0119 Steps: 26650, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001112, Sample Num: 17792, Cur Loss: 0.33843791, Cur Avg Loss: 0.86965397, Log Avg loss: 0.84253967, Global Avg Loss: 2.72899305, Time: 0.0068 Steps: 26660, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001122, Sample Num: 17952, Cur Loss: 0.84908223, Cur Avg Loss: 0.86998513, Log Avg loss: 0.90681110, Global Avg Loss: 2.72830981, Time: 0.0088 Steps: 26670, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001132, Sample Num: 18112, Cur Loss: 0.36256698, Cur Avg Loss: 0.86741327, Log Avg loss: 0.57885043, Global Avg Loss: 2.72750417, Time: 0.0064 Steps: 26680, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001142, Sample Num: 18272, Cur Loss: 1.03984129, Cur Avg Loss: 0.86647303, Log Avg loss: 0.76003751, Global Avg Loss: 2.72676702, Time: 0.0126 Steps: 26690, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001152, Sample Num: 18432, Cur Loss: 0.65670007, Cur Avg Loss: 0.86699976, Log Avg loss: 0.92715259, Global Avg Loss: 2.72609300, Time: 0.0205 Steps: 26700, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001162, Sample Num: 18592, Cur Loss: 0.65041935, Cur Avg Loss: 0.86653064, Log Avg loss: 0.81248783, Global Avg Loss: 2.72537656, Time: 0.0067 Steps: 26710, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001172, Sample Num: 18752, Cur Loss: 0.81815827, Cur Avg Loss: 0.86753487, Log Avg loss: 0.98422690, Global Avg Loss: 2.72472494, Time: 0.0126 Steps: 26720, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001182, Sample Num: 18912, Cur Loss: 1.34679794, Cur Avg Loss: 0.86588316, Log Avg loss: 0.67230179, Global Avg Loss: 2.72395710, Time: 0.0228 Steps: 26730, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001192, Sample Num: 19072, Cur Loss: 0.70552003, Cur Avg Loss: 0.86524501, Log Avg loss: 0.78981625, Global Avg Loss: 2.72323379, Time: 0.0071 Steps: 26740, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001202, Sample Num: 19232, Cur Loss: 1.01672101, Cur Avg Loss: 0.86729060, Log Avg loss: 1.11112470, Global Avg Loss: 2.72263113, Time: 0.0067 Steps: 26750, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001212, Sample Num: 19392, Cur Loss: 0.63351387, Cur Avg Loss: 0.86816396, Log Avg loss: 0.97314230, Global Avg Loss: 2.72197736, Time: 0.0127 Steps: 26760, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001222, Sample Num: 19552, Cur Loss: 0.72397435, Cur Avg Loss: 0.86935800, Log Avg loss: 1.01407574, Global Avg Loss: 2.72133937, Time: 0.0124 Steps: 26770, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001232, Sample Num: 19712, Cur Loss: 0.93957859, Cur Avg Loss: 0.86743355, Log Avg loss: 0.63226511, Global Avg Loss: 2.72055928, Time: 0.0113 Steps: 26780, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001242, Sample Num: 19872, Cur Loss: 0.91867429, Cur Avg Loss: 0.86797656, Log Avg loss: 0.93487599, Global Avg Loss: 2.71989273, Time: 0.0112 Steps: 26790, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001252, Sample Num: 20032, Cur Loss: 0.88676286, Cur Avg Loss: 0.86781557, Log Avg loss: 0.84782031, Global Avg Loss: 2.71919420, Time: 0.0168 Steps: 26800, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001262, Sample Num: 20192, Cur Loss: 2.15528059, Cur Avg Loss: 0.86836456, Log Avg loss: 0.93709835, Global Avg Loss: 2.71852949, Time: 0.0067 Steps: 26810, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001272, Sample Num: 20352, Cur Loss: 0.71232289, Cur Avg Loss: 0.86774346, Log Avg loss: 0.78936029, Global Avg Loss: 2.71781018, Time: 0.0070 Steps: 26820, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001282, Sample Num: 20512, Cur Loss: 0.43962103, Cur Avg Loss: 0.86647050, Log Avg loss: 0.70455050, Global Avg Loss: 2.71705981, Time: 0.0068 Steps: 26830, Updated lr: 0.000076 Training, Epoch: 0013, Batch: 001292, Sample Num: 20672, Cur Loss: 1.15283239, Cur Avg Loss: 0.86793177, Log Avg loss: 1.05526668, Global Avg Loss: 2.71644066, Time: 0.0068 Steps: 26840, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001302, Sample Num: 20832, Cur Loss: 0.84735513, Cur Avg Loss: 0.86771540, Log Avg loss: 0.83975979, Global Avg Loss: 2.71574171, Time: 0.0124 Steps: 26850, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001312, Sample Num: 20992, Cur Loss: 0.68384713, Cur Avg Loss: 0.86794022, Log Avg loss: 0.89721194, Global Avg Loss: 2.71506467, Time: 0.0140 Steps: 26860, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001322, Sample Num: 21152, Cur Loss: 1.30255151, Cur Avg Loss: 0.86962562, Log Avg loss: 1.09074936, Global Avg Loss: 2.71446016, Time: 0.0070 Steps: 26870, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001332, Sample Num: 21312, Cur Loss: 0.68449664, Cur Avg Loss: 0.86970185, Log Avg loss: 0.87977961, Global Avg Loss: 2.71377762, Time: 0.0084 Steps: 26880, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001342, Sample Num: 21472, Cur Loss: 1.39485908, Cur Avg Loss: 0.86988709, Log Avg loss: 0.89456158, Global Avg Loss: 2.71310108, Time: 0.0107 Steps: 26890, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001352, Sample Num: 21632, Cur Loss: 0.86020303, Cur Avg Loss: 0.87283490, Log Avg loss: 1.26843071, Global Avg Loss: 2.71256402, Time: 0.0138 Steps: 26900, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001362, Sample Num: 21792, Cur Loss: 0.44275129, Cur Avg Loss: 0.87358951, Log Avg loss: 0.97561285, Global Avg Loss: 2.71191856, Time: 0.0108 Steps: 26910, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001372, Sample Num: 21952, Cur Loss: 0.75068688, Cur Avg Loss: 0.87338160, Log Avg loss: 0.84506418, Global Avg Loss: 2.71122507, Time: 0.0068 Steps: 26920, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001382, Sample Num: 22112, Cur Loss: 1.08301663, Cur Avg Loss: 0.87299645, Log Avg loss: 0.82015431, Global Avg Loss: 2.71052286, Time: 0.0142 Steps: 26930, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001392, Sample Num: 22272, Cur Loss: 1.00082028, Cur Avg Loss: 0.87235576, Log Avg loss: 0.78381214, Global Avg Loss: 2.70980767, Time: 0.0140 Steps: 26940, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001402, Sample Num: 22432, Cur Loss: 0.38162327, Cur Avg Loss: 0.87104795, Log Avg loss: 0.68900066, Global Avg Loss: 2.70905784, Time: 0.0122 Steps: 26950, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001412, Sample Num: 22592, Cur Loss: 0.55661070, Cur Avg Loss: 0.86983884, Log Avg loss: 0.70032158, Global Avg Loss: 2.70831276, Time: 0.0114 Steps: 26960, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001422, Sample Num: 22752, Cur Loss: 0.57255149, Cur Avg Loss: 0.86925909, Log Avg loss: 0.78739883, Global Avg Loss: 2.70760051, Time: 0.0116 Steps: 26970, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001432, Sample Num: 22912, Cur Loss: 0.48146319, Cur Avg Loss: 0.86766619, Log Avg loss: 0.64115591, Global Avg Loss: 2.70683460, Time: 0.0106 Steps: 26980, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001442, Sample Num: 23072, Cur Loss: 1.27880442, Cur Avg Loss: 0.86650151, Log Avg loss: 0.69971836, Global Avg Loss: 2.70609095, Time: 0.0117 Steps: 26990, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001452, Sample Num: 23232, Cur Loss: 1.17899764, Cur Avg Loss: 0.86540412, Log Avg loss: 0.70716144, Global Avg Loss: 2.70535060, Time: 0.0159 Steps: 27000, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001462, Sample Num: 23392, Cur Loss: 0.33087841, Cur Avg Loss: 0.86412266, Log Avg loss: 0.67805481, Global Avg Loss: 2.70460003, Time: 0.0117 Steps: 27010, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001472, Sample Num: 23552, Cur Loss: 1.76027358, Cur Avg Loss: 0.86433370, Log Avg loss: 0.89518746, Global Avg Loss: 2.70393037, Time: 0.0129 Steps: 27020, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001482, Sample Num: 23712, Cur Loss: 0.93377507, Cur Avg Loss: 0.86394659, Log Avg loss: 0.80696374, Global Avg Loss: 2.70322857, Time: 0.0153 Steps: 27030, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001492, Sample Num: 23872, Cur Loss: 1.31807649, Cur Avg Loss: 0.86312755, Log Avg loss: 0.74174520, Global Avg Loss: 2.70250317, Time: 0.0070 Steps: 27040, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001502, Sample Num: 24032, Cur Loss: 0.92482591, Cur Avg Loss: 0.86366263, Log Avg loss: 0.94349713, Global Avg Loss: 2.70185289, Time: 0.0181 Steps: 27050, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001512, Sample Num: 24192, Cur Loss: 0.90360940, Cur Avg Loss: 0.86393541, Log Avg loss: 0.90490674, Global Avg Loss: 2.70118883, Time: 0.0090 Steps: 27060, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001522, Sample Num: 24352, Cur Loss: 1.13317788, Cur Avg Loss: 0.86572364, Log Avg loss: 1.13610489, Global Avg Loss: 2.70061067, Time: 0.0110 Steps: 27070, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001532, Sample Num: 24512, Cur Loss: 1.03285539, Cur Avg Loss: 0.86612325, Log Avg loss: 0.92694342, Global Avg Loss: 2.69995570, Time: 0.0108 Steps: 27080, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001542, Sample Num: 24672, Cur Loss: 0.70013183, Cur Avg Loss: 0.86556054, Log Avg loss: 0.77935281, Global Avg Loss: 2.69924673, Time: 0.0094 Steps: 27090, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001552, Sample Num: 24832, Cur Loss: 0.42079890, Cur Avg Loss: 0.86662576, Log Avg loss: 1.03088333, Global Avg Loss: 2.69863109, Time: 0.0071 Steps: 27100, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001562, Sample Num: 24992, Cur Loss: 1.00855052, Cur Avg Loss: 0.86649513, Log Avg loss: 0.84622154, Global Avg Loss: 2.69794780, Time: 0.0122 Steps: 27110, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001572, Sample Num: 25152, Cur Loss: 0.53297257, Cur Avg Loss: 0.86586010, Log Avg loss: 0.76666875, Global Avg Loss: 2.69723568, Time: 0.0066 Steps: 27120, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001582, Sample Num: 25312, Cur Loss: 0.83064353, Cur Avg Loss: 0.86527055, Log Avg loss: 0.77259338, Global Avg Loss: 2.69652626, Time: 0.0237 Steps: 27130, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001592, Sample Num: 25472, Cur Loss: 0.29705340, Cur Avg Loss: 0.86696860, Log Avg loss: 1.13559910, Global Avg Loss: 2.69595112, Time: 0.0067 Steps: 27140, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001602, Sample Num: 25632, Cur Loss: 0.45155656, Cur Avg Loss: 0.86633815, Log Avg loss: 0.76597103, Global Avg Loss: 2.69524026, Time: 0.0096 Steps: 27150, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001612, Sample Num: 25792, Cur Loss: 0.61908591, Cur Avg Loss: 0.86481364, Log Avg loss: 0.62058765, Global Avg Loss: 2.69447640, Time: 0.0066 Steps: 27160, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001622, Sample Num: 25952, Cur Loss: 0.41633499, Cur Avg Loss: 0.86459155, Log Avg loss: 0.82878938, Global Avg Loss: 2.69378973, Time: 0.0113 Steps: 27170, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001632, Sample Num: 26112, Cur Loss: 0.94695926, Cur Avg Loss: 0.86460532, Log Avg loss: 0.86683988, Global Avg Loss: 2.69311756, Time: 0.0115 Steps: 27180, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001642, Sample Num: 26272, Cur Loss: 0.96976054, Cur Avg Loss: 0.86613418, Log Avg loss: 1.11564401, Global Avg Loss: 2.69253739, Time: 0.0128 Steps: 27190, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001652, Sample Num: 26432, Cur Loss: 1.66186428, Cur Avg Loss: 0.86798080, Log Avg loss: 1.17119585, Global Avg Loss: 2.69197808, Time: 0.0156 Steps: 27200, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001662, Sample Num: 26592, Cur Loss: 1.20262957, Cur Avg Loss: 0.86837694, Log Avg loss: 0.93381810, Global Avg Loss: 2.69133193, Time: 0.0080 Steps: 27210, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001672, Sample Num: 26752, Cur Loss: 0.82797277, Cur Avg Loss: 0.86905742, Log Avg loss: 0.98215309, Global Avg Loss: 2.69070402, Time: 0.0069 Steps: 27220, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001682, Sample Num: 26912, Cur Loss: 0.22697735, Cur Avg Loss: 0.86805080, Log Avg loss: 0.69974486, Global Avg Loss: 2.68997286, Time: 0.0068 Steps: 27230, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001692, Sample Num: 27072, Cur Loss: 0.71534473, Cur Avg Loss: 0.86839075, Log Avg loss: 0.92556973, Global Avg Loss: 2.68932513, Time: 0.0065 Steps: 27240, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001702, Sample Num: 27232, Cur Loss: 1.40478992, Cur Avg Loss: 0.86852908, Log Avg loss: 0.89193511, Global Avg Loss: 2.68866554, Time: 0.0121 Steps: 27250, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001712, Sample Num: 27392, Cur Loss: 0.53358972, Cur Avg Loss: 0.86734508, Log Avg loss: 0.66582786, Global Avg Loss: 2.68792349, Time: 0.0118 Steps: 27260, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001722, Sample Num: 27552, Cur Loss: 0.37975478, Cur Avg Loss: 0.86637142, Log Avg loss: 0.69968074, Global Avg Loss: 2.68719439, Time: 0.0215 Steps: 27270, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001732, Sample Num: 27712, Cur Loss: 1.10348415, Cur Avg Loss: 0.86521069, Log Avg loss: 0.66533393, Global Avg Loss: 2.68645324, Time: 0.0116 Steps: 27280, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001742, Sample Num: 27872, Cur Loss: 0.56955141, Cur Avg Loss: 0.86435949, Log Avg loss: 0.71693118, Global Avg Loss: 2.68573154, Time: 0.0066 Steps: 27290, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001752, Sample Num: 28032, Cur Loss: 1.70297062, Cur Avg Loss: 0.86514503, Log Avg loss: 1.00198570, Global Avg Loss: 2.68511478, Time: 0.0122 Steps: 27300, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001762, Sample Num: 28192, Cur Loss: 0.54916799, Cur Avg Loss: 0.86533434, Log Avg loss: 0.89850150, Global Avg Loss: 2.68446058, Time: 0.0104 Steps: 27310, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001772, Sample Num: 28352, Cur Loss: 0.48606303, Cur Avg Loss: 0.86449552, Log Avg loss: 0.71669599, Global Avg Loss: 2.68374032, Time: 0.0072 Steps: 27320, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001782, Sample Num: 28512, Cur Loss: 0.67536312, Cur Avg Loss: 0.86421436, Log Avg loss: 0.81439246, Global Avg Loss: 2.68305633, Time: 0.0122 Steps: 27330, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001792, Sample Num: 28672, Cur Loss: 0.72216332, Cur Avg Loss: 0.86344668, Log Avg loss: 0.72664555, Global Avg Loss: 2.68234074, Time: 0.0174 Steps: 27340, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001802, Sample Num: 28832, Cur Loss: 0.33136791, Cur Avg Loss: 0.86337078, Log Avg loss: 0.84977083, Global Avg Loss: 2.68167070, Time: 0.0094 Steps: 27350, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001812, Sample Num: 28992, Cur Loss: 0.14340132, Cur Avg Loss: 0.86301051, Log Avg loss: 0.79808882, Global Avg Loss: 2.68098225, Time: 0.0105 Steps: 27360, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001822, Sample Num: 29152, Cur Loss: 0.72606027, Cur Avg Loss: 0.86310599, Log Avg loss: 0.88040794, Global Avg Loss: 2.68032439, Time: 0.0113 Steps: 27370, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001832, Sample Num: 29312, Cur Loss: 0.13071546, Cur Avg Loss: 0.86336026, Log Avg loss: 0.90968777, Global Avg Loss: 2.67967770, Time: 0.0065 Steps: 27380, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001842, Sample Num: 29472, Cur Loss: 1.70498383, Cur Avg Loss: 0.86526093, Log Avg loss: 1.21346371, Global Avg Loss: 2.67914239, Time: 0.0103 Steps: 27390, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001852, Sample Num: 29632, Cur Loss: 0.46937293, Cur Avg Loss: 0.86480224, Log Avg loss: 0.78031102, Global Avg Loss: 2.67844939, Time: 0.0063 Steps: 27400, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001862, Sample Num: 29792, Cur Loss: 1.12316155, Cur Avg Loss: 0.86537055, Log Avg loss: 0.97062097, Global Avg Loss: 2.67782632, Time: 0.0157 Steps: 27410, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001872, Sample Num: 29952, Cur Loss: 0.70967865, Cur Avg Loss: 0.86477751, Log Avg loss: 0.75435523, Global Avg Loss: 2.67712483, Time: 0.0179 Steps: 27420, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001882, Sample Num: 30112, Cur Loss: 0.87226909, Cur Avg Loss: 0.86553213, Log Avg loss: 1.00679589, Global Avg Loss: 2.67651589, Time: 0.0106 Steps: 27430, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001892, Sample Num: 30272, Cur Loss: 0.44583976, Cur Avg Loss: 0.86530862, Log Avg loss: 0.82324349, Global Avg Loss: 2.67584050, Time: 0.0114 Steps: 27440, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001902, Sample Num: 30432, Cur Loss: 1.18068099, Cur Avg Loss: 0.86548419, Log Avg loss: 0.89870359, Global Avg Loss: 2.67519309, Time: 0.0118 Steps: 27450, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001912, Sample Num: 30592, Cur Loss: 1.04815912, Cur Avg Loss: 0.86505374, Log Avg loss: 0.78318167, Global Avg Loss: 2.67450409, Time: 0.0225 Steps: 27460, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001922, Sample Num: 30752, Cur Loss: 0.86022627, Cur Avg Loss: 0.86453663, Log Avg loss: 0.76566480, Global Avg Loss: 2.67380920, Time: 0.0229 Steps: 27470, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001932, Sample Num: 30912, Cur Loss: 0.23235501, Cur Avg Loss: 0.86385982, Log Avg loss: 0.73377665, Global Avg Loss: 2.67310322, Time: 0.0109 Steps: 27480, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001942, Sample Num: 31072, Cur Loss: 0.46845362, Cur Avg Loss: 0.86393209, Log Avg loss: 0.87789482, Global Avg Loss: 2.67245018, Time: 0.0098 Steps: 27490, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001952, Sample Num: 31232, Cur Loss: 0.39230412, Cur Avg Loss: 0.86428122, Log Avg loss: 0.93208266, Global Avg Loss: 2.67181732, Time: 0.0131 Steps: 27500, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001962, Sample Num: 31392, Cur Loss: 1.01457095, Cur Avg Loss: 0.86363396, Log Avg loss: 0.73728863, Global Avg Loss: 2.67111411, Time: 0.0067 Steps: 27510, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001972, Sample Num: 31552, Cur Loss: 0.54538643, Cur Avg Loss: 0.86257318, Log Avg loss: 0.65444762, Global Avg Loss: 2.67038131, Time: 0.0120 Steps: 27520, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001982, Sample Num: 31712, Cur Loss: 0.92876327, Cur Avg Loss: 0.86338371, Log Avg loss: 1.02322062, Global Avg Loss: 2.66978300, Time: 0.0133 Steps: 27530, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 001992, Sample Num: 31872, Cur Loss: 0.31477195, Cur Avg Loss: 0.86256438, Log Avg loss: 0.70017399, Global Avg Loss: 2.66906782, Time: 0.0115 Steps: 27540, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002002, Sample Num: 32032, Cur Loss: 1.31273174, Cur Avg Loss: 0.86162237, Log Avg loss: 0.67397356, Global Avg Loss: 2.66834365, Time: 0.0110 Steps: 27550, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002012, Sample Num: 32192, Cur Loss: 1.40993631, Cur Avg Loss: 0.86141333, Log Avg loss: 0.81956268, Global Avg Loss: 2.66767282, Time: 0.0114 Steps: 27560, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002022, Sample Num: 32352, Cur Loss: 1.10381770, Cur Avg Loss: 0.86174654, Log Avg loss: 0.92878850, Global Avg Loss: 2.66704211, Time: 0.0068 Steps: 27570, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002032, Sample Num: 32512, Cur Loss: 1.08954728, Cur Avg Loss: 0.86170924, Log Avg loss: 0.85416869, Global Avg Loss: 2.66638479, Time: 0.0097 Steps: 27580, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002042, Sample Num: 32672, Cur Loss: 0.46439332, Cur Avg Loss: 0.86223104, Log Avg loss: 0.96825978, Global Avg Loss: 2.66576931, Time: 0.0117 Steps: 27590, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002052, Sample Num: 32832, Cur Loss: 1.49920511, Cur Avg Loss: 0.86274961, Log Avg loss: 0.96864117, Global Avg Loss: 2.66515441, Time: 0.0113 Steps: 27600, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002062, Sample Num: 32992, Cur Loss: 0.47571266, Cur Avg Loss: 0.86185856, Log Avg loss: 0.67901576, Global Avg Loss: 2.66443505, Time: 0.0166 Steps: 27610, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002072, Sample Num: 33152, Cur Loss: 0.95681155, Cur Avg Loss: 0.86171390, Log Avg loss: 0.83188510, Global Avg Loss: 2.66377157, Time: 0.0066 Steps: 27620, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002082, Sample Num: 33312, Cur Loss: 0.42896214, Cur Avg Loss: 0.86102514, Log Avg loss: 0.71831397, Global Avg Loss: 2.66306745, Time: 0.0111 Steps: 27630, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002092, Sample Num: 33472, Cur Loss: 0.21911559, Cur Avg Loss: 0.85997853, Log Avg loss: 0.64207355, Global Avg Loss: 2.66233627, Time: 0.0107 Steps: 27640, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002102, Sample Num: 33632, Cur Loss: 0.56049359, Cur Avg Loss: 0.85985411, Log Avg loss: 0.83382501, Global Avg Loss: 2.66167496, Time: 0.0073 Steps: 27650, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002112, Sample Num: 33792, Cur Loss: 0.75617647, Cur Avg Loss: 0.85939702, Log Avg loss: 0.76331845, Global Avg Loss: 2.66098865, Time: 0.0146 Steps: 27660, Updated lr: 0.000075 Training, Epoch: 0013, Batch: 002122, Sample Num: 33952, Cur Loss: 0.24196462, Cur Avg Loss: 0.86013217, Log Avg loss: 1.01539528, Global Avg Loss: 2.66039392, Time: 0.0155 Steps: 27670, Updated lr: 0.000075 ***** Running evaluation checkpoint-27677 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-27677 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.396060, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.092248, "eval_total_loss": 767.850042, "eval_mae": 0.899582, "eval_mse": 1.092209, "eval_r2": 0.30572, "eval_sp_statistic": 0.617239, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.668426, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.887934, "test_total_loss": 947.74303, "test_mae": 1.256842, "test_mse": 1.887684, "test_r2": -0.218328, "test_sp_statistic": 0.389087, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.525226, "test_ps_pvalue": 0.0, "lr": 7.470175438596491e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.6599080141755023, "train_cur_epoch_loss": 1830.374673485756, "train_cur_epoch_avg_loss": 0.8597344638260949, "train_cur_epoch_time": 22.396060466766357, "train_cur_epoch_avg_time": 0.010519521121073912, "epoch": 13, "step": 27677} ################################################## Training, Epoch: 0014, Batch: 000003, Sample Num: 48, Cur Loss: 0.46551371, Cur Avg Loss: 0.68153820, Log Avg loss: 0.72188216, Global Avg Loss: 2.65969360, Time: 0.0198 Steps: 27680, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000013, Sample Num: 208, Cur Loss: 0.82666469, Cur Avg Loss: 0.96300930, Log Avg loss: 1.04745063, Global Avg Loss: 2.65911135, Time: 0.0065 Steps: 27690, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000023, Sample Num: 368, Cur Loss: 0.87284225, Cur Avg Loss: 0.88049827, Log Avg loss: 0.77323394, Global Avg Loss: 2.65843053, Time: 0.0070 Steps: 27700, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000033, Sample Num: 528, Cur Loss: 0.94739407, Cur Avg Loss: 0.85154753, Log Avg loss: 0.78496083, Global Avg Loss: 2.65775443, Time: 0.0068 Steps: 27710, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000043, Sample Num: 688, Cur Loss: 0.67138165, Cur Avg Loss: 0.88429878, Log Avg loss: 0.99237790, Global Avg Loss: 2.65715364, Time: 0.0204 Steps: 27720, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000053, Sample Num: 848, Cur Loss: 0.62473714, Cur Avg Loss: 0.88754267, Log Avg loss: 0.90149137, Global Avg Loss: 2.65652051, Time: 0.0072 Steps: 27730, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000063, Sample Num: 1008, Cur Loss: 0.22494376, Cur Avg Loss: 0.89357940, Log Avg loss: 0.92557411, Global Avg Loss: 2.65589653, Time: 0.0236 Steps: 27740, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000073, Sample Num: 1168, Cur Loss: 0.24382459, Cur Avg Loss: 0.88699110, Log Avg loss: 0.84548482, Global Avg Loss: 2.65524412, Time: 0.0067 Steps: 27750, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000083, Sample Num: 1328, Cur Loss: 0.52831519, Cur Avg Loss: 0.89161059, Log Avg loss: 0.92533286, Global Avg Loss: 2.65462096, Time: 0.0064 Steps: 27760, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000093, Sample Num: 1488, Cur Loss: 0.79103112, Cur Avg Loss: 0.89758445, Log Avg loss: 0.94716749, Global Avg Loss: 2.65400610, Time: 0.0173 Steps: 27770, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000103, Sample Num: 1648, Cur Loss: 0.41598156, Cur Avg Loss: 0.88605872, Log Avg loss: 0.77886946, Global Avg Loss: 2.65333111, Time: 0.0066 Steps: 27780, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000113, Sample Num: 1808, Cur Loss: 0.50713158, Cur Avg Loss: 0.87281644, Log Avg loss: 0.73642095, Global Avg Loss: 2.65264132, Time: 0.0114 Steps: 27790, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000123, Sample Num: 1968, Cur Loss: 0.72248888, Cur Avg Loss: 0.86972185, Log Avg loss: 0.83475292, Global Avg Loss: 2.65198741, Time: 0.0068 Steps: 27800, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000133, Sample Num: 2128, Cur Loss: 0.28435209, Cur Avg Loss: 0.87372091, Log Avg loss: 0.92290941, Global Avg Loss: 2.65136566, Time: 0.0186 Steps: 27810, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000143, Sample Num: 2288, Cur Loss: 1.23263144, Cur Avg Loss: 0.86631258, Log Avg loss: 0.76778174, Global Avg Loss: 2.65068860, Time: 0.0066 Steps: 27820, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000153, Sample Num: 2448, Cur Loss: 0.51873147, Cur Avg Loss: 0.86235263, Log Avg loss: 0.80572541, Global Avg Loss: 2.65002566, Time: 0.0064 Steps: 27830, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000163, Sample Num: 2608, Cur Loss: 0.80843413, Cur Avg Loss: 0.87122132, Log Avg loss: 1.00691215, Global Avg Loss: 2.64943546, Time: 0.0200 Steps: 27840, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000173, Sample Num: 2768, Cur Loss: 1.18588841, Cur Avg Loss: 0.87696532, Log Avg loss: 0.97059261, Global Avg Loss: 2.64883264, Time: 0.0064 Steps: 27850, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000183, Sample Num: 2928, Cur Loss: 0.67939883, Cur Avg Loss: 0.86984048, Log Avg loss: 0.74658073, Global Avg Loss: 2.64814985, Time: 0.0065 Steps: 27860, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000193, Sample Num: 3088, Cur Loss: 0.75414610, Cur Avg Loss: 0.85893102, Log Avg loss: 0.65928784, Global Avg Loss: 2.64743623, Time: 0.0111 Steps: 27870, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000203, Sample Num: 3248, Cur Loss: 0.63675416, Cur Avg Loss: 0.84981820, Log Avg loss: 0.67394076, Global Avg Loss: 2.64672838, Time: 0.0084 Steps: 27880, Updated lr: 0.000075 Training, Epoch: 0014, Batch: 000213, Sample Num: 3408, Cur Loss: 0.46952692, Cur Avg Loss: 0.85242664, Log Avg loss: 0.90537800, Global Avg Loss: 2.64610402, Time: 0.0072 Steps: 27890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000223, Sample Num: 3568, Cur Loss: 0.22610077, Cur Avg Loss: 0.84935226, Log Avg loss: 0.78386813, Global Avg Loss: 2.64543655, Time: 0.0131 Steps: 27900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000233, Sample Num: 3728, Cur Loss: 1.13461304, Cur Avg Loss: 0.84716768, Log Avg loss: 0.79845140, Global Avg Loss: 2.64477478, Time: 0.0084 Steps: 27910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000243, Sample Num: 3888, Cur Loss: 0.83087003, Cur Avg Loss: 0.84050265, Log Avg loss: 0.68520743, Global Avg Loss: 2.64407293, Time: 0.0107 Steps: 27920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000253, Sample Num: 4048, Cur Loss: 0.71935523, Cur Avg Loss: 0.84182472, Log Avg loss: 0.87395102, Global Avg Loss: 2.64343916, Time: 0.0131 Steps: 27930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000263, Sample Num: 4208, Cur Loss: 0.68090349, Cur Avg Loss: 0.84412152, Log Avg loss: 0.90223058, Global Avg Loss: 2.64281597, Time: 0.0121 Steps: 27940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000273, Sample Num: 4368, Cur Loss: 0.67589515, Cur Avg Loss: 0.84307934, Log Avg loss: 0.81566992, Global Avg Loss: 2.64216225, Time: 0.0111 Steps: 27950, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000283, Sample Num: 4528, Cur Loss: 0.31525388, Cur Avg Loss: 0.83982575, Log Avg loss: 0.75100293, Global Avg Loss: 2.64148587, Time: 0.0121 Steps: 27960, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000293, Sample Num: 4688, Cur Loss: 1.92575979, Cur Avg Loss: 0.83957615, Log Avg loss: 0.83251244, Global Avg Loss: 2.64083911, Time: 0.0066 Steps: 27970, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000303, Sample Num: 4848, Cur Loss: 0.35754728, Cur Avg Loss: 0.83927672, Log Avg loss: 0.83050326, Global Avg Loss: 2.64019210, Time: 0.0123 Steps: 27980, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000313, Sample Num: 5008, Cur Loss: 0.56877738, Cur Avg Loss: 0.83358168, Log Avg loss: 0.66102214, Global Avg Loss: 2.63948500, Time: 0.0217 Steps: 27990, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000323, Sample Num: 5168, Cur Loss: 1.36907828, Cur Avg Loss: 0.83512393, Log Avg loss: 0.88339635, Global Avg Loss: 2.63885783, Time: 0.0085 Steps: 28000, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000333, Sample Num: 5328, Cur Loss: 0.77118134, Cur Avg Loss: 0.83532994, Log Avg loss: 0.84198414, Global Avg Loss: 2.63821631, Time: 0.0089 Steps: 28010, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000343, Sample Num: 5488, Cur Loss: 1.52463388, Cur Avg Loss: 0.83929070, Log Avg loss: 0.97118391, Global Avg Loss: 2.63762137, Time: 0.0066 Steps: 28020, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000353, Sample Num: 5648, Cur Loss: 0.47926378, Cur Avg Loss: 0.83706619, Log Avg loss: 0.76076559, Global Avg Loss: 2.63695178, Time: 0.0068 Steps: 28030, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000363, Sample Num: 5808, Cur Loss: 0.35572094, Cur Avg Loss: 0.83658262, Log Avg loss: 0.81951241, Global Avg Loss: 2.63630362, Time: 0.0067 Steps: 28040, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000373, Sample Num: 5968, Cur Loss: 1.35564268, Cur Avg Loss: 0.83736202, Log Avg loss: 0.86565448, Global Avg Loss: 2.63567238, Time: 0.0073 Steps: 28050, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000383, Sample Num: 6128, Cur Loss: 1.36884379, Cur Avg Loss: 0.83687726, Log Avg loss: 0.81879563, Global Avg Loss: 2.63502488, Time: 0.0132 Steps: 28060, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000393, Sample Num: 6288, Cur Loss: 0.81129313, Cur Avg Loss: 0.83267893, Log Avg loss: 0.67188266, Global Avg Loss: 2.63432551, Time: 0.0135 Steps: 28070, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000403, Sample Num: 6448, Cur Loss: 0.66030139, Cur Avg Loss: 0.83261112, Log Avg loss: 0.82994640, Global Avg Loss: 2.63368292, Time: 0.0107 Steps: 28080, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000413, Sample Num: 6608, Cur Loss: 0.96581435, Cur Avg Loss: 0.82801047, Log Avg loss: 0.64260416, Global Avg Loss: 2.63297410, Time: 0.0120 Steps: 28090, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000423, Sample Num: 6768, Cur Loss: 0.67199743, Cur Avg Loss: 0.82933388, Log Avg loss: 0.88399083, Global Avg Loss: 2.63235168, Time: 0.0111 Steps: 28100, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000433, Sample Num: 6928, Cur Loss: 0.76253986, Cur Avg Loss: 0.82676685, Log Avg loss: 0.71818136, Global Avg Loss: 2.63167073, Time: 0.0122 Steps: 28110, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000443, Sample Num: 7088, Cur Loss: 1.55982232, Cur Avg Loss: 0.82760941, Log Avg loss: 0.86409244, Global Avg Loss: 2.63104214, Time: 0.0095 Steps: 28120, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000453, Sample Num: 7248, Cur Loss: 0.69898719, Cur Avg Loss: 0.82317463, Log Avg loss: 0.62671375, Global Avg Loss: 2.63032962, Time: 0.0159 Steps: 28130, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000463, Sample Num: 7408, Cur Loss: 0.94913280, Cur Avg Loss: 0.82445812, Log Avg loss: 0.88260010, Global Avg Loss: 2.62970854, Time: 0.0104 Steps: 28140, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000473, Sample Num: 7568, Cur Loss: 0.90120834, Cur Avg Loss: 0.82732513, Log Avg loss: 0.96006763, Global Avg Loss: 2.62911541, Time: 0.0114 Steps: 28150, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000483, Sample Num: 7728, Cur Loss: 0.78753942, Cur Avg Loss: 0.82912386, Log Avg loss: 0.91420405, Global Avg Loss: 2.62850643, Time: 0.0066 Steps: 28160, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000493, Sample Num: 7888, Cur Loss: 0.60361350, Cur Avg Loss: 0.82842020, Log Avg loss: 0.79443319, Global Avg Loss: 2.62785535, Time: 0.0068 Steps: 28170, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000503, Sample Num: 8048, Cur Loss: 0.78438377, Cur Avg Loss: 0.82838974, Log Avg loss: 0.82688797, Global Avg Loss: 2.62721626, Time: 0.0134 Steps: 28180, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000513, Sample Num: 8208, Cur Loss: 0.42358714, Cur Avg Loss: 0.82906470, Log Avg loss: 0.86301518, Global Avg Loss: 2.62659043, Time: 0.0106 Steps: 28190, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000523, Sample Num: 8368, Cur Loss: 0.65237653, Cur Avg Loss: 0.83038055, Log Avg loss: 0.89788397, Global Avg Loss: 2.62597742, Time: 0.0071 Steps: 28200, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000533, Sample Num: 8528, Cur Loss: 0.68320525, Cur Avg Loss: 0.83037002, Log Avg loss: 0.82981925, Global Avg Loss: 2.62534071, Time: 0.0072 Steps: 28210, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000543, Sample Num: 8688, Cur Loss: 0.39477730, Cur Avg Loss: 0.82572494, Log Avg loss: 0.57814220, Global Avg Loss: 2.62461526, Time: 0.0129 Steps: 28220, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000553, Sample Num: 8848, Cur Loss: 0.73337901, Cur Avg Loss: 0.82270330, Log Avg loss: 0.65862829, Global Avg Loss: 2.62391885, Time: 0.0132 Steps: 28230, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000563, Sample Num: 9008, Cur Loss: 0.90114093, Cur Avg Loss: 0.82760955, Log Avg loss: 1.09892503, Global Avg Loss: 2.62337883, Time: 0.0087 Steps: 28240, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000573, Sample Num: 9168, Cur Loss: 0.74361116, Cur Avg Loss: 0.82962687, Log Avg loss: 0.94320227, Global Avg Loss: 2.62278408, Time: 0.0092 Steps: 28250, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000583, Sample Num: 9328, Cur Loss: 0.93987799, Cur Avg Loss: 0.82921741, Log Avg loss: 0.80575536, Global Avg Loss: 2.62214111, Time: 0.0127 Steps: 28260, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000593, Sample Num: 9488, Cur Loss: 0.48383403, Cur Avg Loss: 0.82920819, Log Avg loss: 0.82867020, Global Avg Loss: 2.62150671, Time: 0.0073 Steps: 28270, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000603, Sample Num: 9648, Cur Loss: 0.37323409, Cur Avg Loss: 0.82535240, Log Avg loss: 0.59670408, Global Avg Loss: 2.62079072, Time: 0.0132 Steps: 28280, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000613, Sample Num: 9808, Cur Loss: 1.06894445, Cur Avg Loss: 0.82635402, Log Avg loss: 0.88675205, Global Avg Loss: 2.62017777, Time: 0.0071 Steps: 28290, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000623, Sample Num: 9968, Cur Loss: 0.36629629, Cur Avg Loss: 0.82605535, Log Avg loss: 0.80774701, Global Avg Loss: 2.61953734, Time: 0.0074 Steps: 28300, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000633, Sample Num: 10128, Cur Loss: 1.36585581, Cur Avg Loss: 0.82460104, Log Avg loss: 0.73399719, Global Avg Loss: 2.61887130, Time: 0.0070 Steps: 28310, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000643, Sample Num: 10288, Cur Loss: 0.37433940, Cur Avg Loss: 0.82383055, Log Avg loss: 0.77505866, Global Avg Loss: 2.61822024, Time: 0.0070 Steps: 28320, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000653, Sample Num: 10448, Cur Loss: 0.63835788, Cur Avg Loss: 0.82487501, Log Avg loss: 0.89203369, Global Avg Loss: 2.61761092, Time: 0.0077 Steps: 28330, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000663, Sample Num: 10608, Cur Loss: 0.94646573, Cur Avg Loss: 0.82714724, Log Avg loss: 0.97552413, Global Avg Loss: 2.61703150, Time: 0.0073 Steps: 28340, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000673, Sample Num: 10768, Cur Loss: 0.45080662, Cur Avg Loss: 0.82873439, Log Avg loss: 0.93396193, Global Avg Loss: 2.61643783, Time: 0.0216 Steps: 28350, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000683, Sample Num: 10928, Cur Loss: 0.64422011, Cur Avg Loss: 0.82574558, Log Avg loss: 0.62459909, Global Avg Loss: 2.61573548, Time: 0.0190 Steps: 28360, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000693, Sample Num: 11088, Cur Loss: 0.69383538, Cur Avg Loss: 0.82575361, Log Avg loss: 0.82630225, Global Avg Loss: 2.61510474, Time: 0.0071 Steps: 28370, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000703, Sample Num: 11248, Cur Loss: 0.67036164, Cur Avg Loss: 0.82257041, Log Avg loss: 0.60197450, Global Avg Loss: 2.61439539, Time: 0.0069 Steps: 28380, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000713, Sample Num: 11408, Cur Loss: 1.15371323, Cur Avg Loss: 0.82146162, Log Avg loss: 0.74351324, Global Avg Loss: 2.61373639, Time: 0.0076 Steps: 28390, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000723, Sample Num: 11568, Cur Loss: 0.65543270, Cur Avg Loss: 0.81821212, Log Avg loss: 0.58652301, Global Avg Loss: 2.61302259, Time: 0.0171 Steps: 28400, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000733, Sample Num: 11728, Cur Loss: 0.98676294, Cur Avg Loss: 0.81736200, Log Avg loss: 0.75589855, Global Avg Loss: 2.61236890, Time: 0.0077 Steps: 28410, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000743, Sample Num: 11888, Cur Loss: 0.56410611, Cur Avg Loss: 0.81732113, Log Avg loss: 0.81432547, Global Avg Loss: 2.61173623, Time: 0.0075 Steps: 28420, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000753, Sample Num: 12048, Cur Loss: 0.78657097, Cur Avg Loss: 0.81854831, Log Avg loss: 0.90972764, Global Avg Loss: 2.61113757, Time: 0.0071 Steps: 28430, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000763, Sample Num: 12208, Cur Loss: 0.63919675, Cur Avg Loss: 0.81722062, Log Avg loss: 0.71724522, Global Avg Loss: 2.61047164, Time: 0.0068 Steps: 28440, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000773, Sample Num: 12368, Cur Loss: 0.54614812, Cur Avg Loss: 0.81522563, Log Avg loss: 0.66300839, Global Avg Loss: 2.60978712, Time: 0.0135 Steps: 28450, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000783, Sample Num: 12528, Cur Loss: 0.77790564, Cur Avg Loss: 0.81607762, Log Avg loss: 0.88193641, Global Avg Loss: 2.60918000, Time: 0.0088 Steps: 28460, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000793, Sample Num: 12688, Cur Loss: 0.52219093, Cur Avg Loss: 0.81411474, Log Avg loss: 0.66042099, Global Avg Loss: 2.60849551, Time: 0.0073 Steps: 28470, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000803, Sample Num: 12848, Cur Loss: 1.09266877, Cur Avg Loss: 0.81470782, Log Avg loss: 0.86173862, Global Avg Loss: 2.60788218, Time: 0.0132 Steps: 28480, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000813, Sample Num: 13008, Cur Loss: 0.52526879, Cur Avg Loss: 0.81556875, Log Avg loss: 0.88470175, Global Avg Loss: 2.60727734, Time: 0.0128 Steps: 28490, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000823, Sample Num: 13168, Cur Loss: 0.87418187, Cur Avg Loss: 0.81588017, Log Avg loss: 0.84119881, Global Avg Loss: 2.60665767, Time: 0.0145 Steps: 28500, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000833, Sample Num: 13328, Cur Loss: 0.75843865, Cur Avg Loss: 0.81380117, Log Avg loss: 0.64269943, Global Avg Loss: 2.60596880, Time: 0.0076 Steps: 28510, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000843, Sample Num: 13488, Cur Loss: 1.01122010, Cur Avg Loss: 0.81572590, Log Avg loss: 0.97605559, Global Avg Loss: 2.60539730, Time: 0.0110 Steps: 28520, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000853, Sample Num: 13648, Cur Loss: 1.25175035, Cur Avg Loss: 0.81667935, Log Avg loss: 0.89705567, Global Avg Loss: 2.60479851, Time: 0.0071 Steps: 28530, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000863, Sample Num: 13808, Cur Loss: 0.64448148, Cur Avg Loss: 0.81651345, Log Avg loss: 0.80236155, Global Avg Loss: 2.60416697, Time: 0.0066 Steps: 28540, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000873, Sample Num: 13968, Cur Loss: 0.74099374, Cur Avg Loss: 0.81879277, Log Avg loss: 1.01549823, Global Avg Loss: 2.60361051, Time: 0.0085 Steps: 28550, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000883, Sample Num: 14128, Cur Loss: 0.59276760, Cur Avg Loss: 0.82020643, Log Avg loss: 0.94361876, Global Avg Loss: 2.60302929, Time: 0.0141 Steps: 28560, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000893, Sample Num: 14288, Cur Loss: 1.14587688, Cur Avg Loss: 0.81944588, Log Avg loss: 0.75228959, Global Avg Loss: 2.60238149, Time: 0.0112 Steps: 28570, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000903, Sample Num: 14448, Cur Loss: 0.22644761, Cur Avg Loss: 0.81962473, Log Avg loss: 0.83559663, Global Avg Loss: 2.60176330, Time: 0.0175 Steps: 28580, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000913, Sample Num: 14608, Cur Loss: 1.59405971, Cur Avg Loss: 0.82191476, Log Avg loss: 1.02870417, Global Avg Loss: 2.60121309, Time: 0.0065 Steps: 28590, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000923, Sample Num: 14768, Cur Loss: 1.39421880, Cur Avg Loss: 0.82066401, Log Avg loss: 0.70647014, Global Avg Loss: 2.60055059, Time: 0.0070 Steps: 28600, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000933, Sample Num: 14928, Cur Loss: 0.55286306, Cur Avg Loss: 0.82125173, Log Avg loss: 0.87549871, Global Avg Loss: 2.59994764, Time: 0.0136 Steps: 28610, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000943, Sample Num: 15088, Cur Loss: 0.48866928, Cur Avg Loss: 0.82342483, Log Avg loss: 1.02617446, Global Avg Loss: 2.59939775, Time: 0.0084 Steps: 28620, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000953, Sample Num: 15248, Cur Loss: 0.35988015, Cur Avg Loss: 0.82405203, Log Avg loss: 0.88319761, Global Avg Loss: 2.59879831, Time: 0.0076 Steps: 28630, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000963, Sample Num: 15408, Cur Loss: 0.94501686, Cur Avg Loss: 0.82372250, Log Avg loss: 0.79231788, Global Avg Loss: 2.59816756, Time: 0.0132 Steps: 28640, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000973, Sample Num: 15568, Cur Loss: 0.50685161, Cur Avg Loss: 0.82445086, Log Avg loss: 0.89459263, Global Avg Loss: 2.59757294, Time: 0.0069 Steps: 28650, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000983, Sample Num: 15728, Cur Loss: 0.79485518, Cur Avg Loss: 0.82371134, Log Avg loss: 0.75175603, Global Avg Loss: 2.59692890, Time: 0.0068 Steps: 28660, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 000993, Sample Num: 15888, Cur Loss: 0.57930911, Cur Avg Loss: 0.82385900, Log Avg loss: 0.83837363, Global Avg Loss: 2.59631552, Time: 0.0112 Steps: 28670, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001003, Sample Num: 16048, Cur Loss: 1.25196993, Cur Avg Loss: 0.82361475, Log Avg loss: 0.79936059, Global Avg Loss: 2.59568897, Time: 0.0154 Steps: 28680, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001013, Sample Num: 16208, Cur Loss: 0.68594921, Cur Avg Loss: 0.82228745, Log Avg loss: 0.68915929, Global Avg Loss: 2.59502444, Time: 0.0063 Steps: 28690, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001023, Sample Num: 16368, Cur Loss: 1.53787601, Cur Avg Loss: 0.82413942, Log Avg loss: 1.01174355, Global Avg Loss: 2.59447278, Time: 0.0085 Steps: 28700, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001033, Sample Num: 16528, Cur Loss: 1.72498202, Cur Avg Loss: 0.82419347, Log Avg loss: 0.82972322, Global Avg Loss: 2.59385810, Time: 0.0064 Steps: 28710, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001043, Sample Num: 16688, Cur Loss: 1.85844767, Cur Avg Loss: 0.82465001, Log Avg loss: 0.87181078, Global Avg Loss: 2.59325850, Time: 0.0121 Steps: 28720, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001053, Sample Num: 16848, Cur Loss: 0.79540431, Cur Avg Loss: 0.82420058, Log Avg loss: 0.77732449, Global Avg Loss: 2.59262643, Time: 0.0080 Steps: 28730, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001063, Sample Num: 17008, Cur Loss: 0.43022078, Cur Avg Loss: 0.82520869, Log Avg loss: 0.93136351, Global Avg Loss: 2.59204840, Time: 0.0107 Steps: 28740, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001073, Sample Num: 17168, Cur Loss: 1.56229198, Cur Avg Loss: 0.82553538, Log Avg loss: 0.86026224, Global Avg Loss: 2.59144604, Time: 0.0068 Steps: 28750, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001083, Sample Num: 17328, Cur Loss: 0.46980572, Cur Avg Loss: 0.82566481, Log Avg loss: 0.83955269, Global Avg Loss: 2.59083689, Time: 0.0071 Steps: 28760, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001093, Sample Num: 17488, Cur Loss: 0.87738782, Cur Avg Loss: 0.82622653, Log Avg loss: 0.88706037, Global Avg Loss: 2.59024469, Time: 0.0066 Steps: 28770, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001103, Sample Num: 17648, Cur Loss: 0.63459206, Cur Avg Loss: 0.82710357, Log Avg loss: 0.92296387, Global Avg Loss: 2.58966537, Time: 0.0134 Steps: 28780, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001113, Sample Num: 17808, Cur Loss: 0.40341288, Cur Avg Loss: 0.82765178, Log Avg loss: 0.88812021, Global Avg Loss: 2.58907435, Time: 0.0063 Steps: 28790, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001123, Sample Num: 17968, Cur Loss: 0.30414480, Cur Avg Loss: 0.82805993, Log Avg loss: 0.87348694, Global Avg Loss: 2.58847866, Time: 0.0109 Steps: 28800, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001133, Sample Num: 18128, Cur Loss: 0.52907372, Cur Avg Loss: 0.82838220, Log Avg loss: 0.86457250, Global Avg Loss: 2.58788029, Time: 0.0114 Steps: 28810, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001143, Sample Num: 18288, Cur Loss: 0.65239882, Cur Avg Loss: 0.82926092, Log Avg loss: 0.92882065, Global Avg Loss: 2.58730463, Time: 0.0093 Steps: 28820, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001153, Sample Num: 18448, Cur Loss: 0.63394308, Cur Avg Loss: 0.82752562, Log Avg loss: 0.62917988, Global Avg Loss: 2.58662543, Time: 0.0159 Steps: 28830, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001163, Sample Num: 18608, Cur Loss: 0.72292340, Cur Avg Loss: 0.82826464, Log Avg loss: 0.91347390, Global Avg Loss: 2.58604528, Time: 0.0066 Steps: 28840, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001173, Sample Num: 18768, Cur Loss: 0.42957360, Cur Avg Loss: 0.82598168, Log Avg loss: 0.56047367, Global Avg Loss: 2.58534318, Time: 0.0134 Steps: 28850, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001183, Sample Num: 18928, Cur Loss: 0.21865922, Cur Avg Loss: 0.82492157, Log Avg loss: 0.70057006, Global Avg Loss: 2.58469010, Time: 0.0068 Steps: 28860, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001193, Sample Num: 19088, Cur Loss: 0.56313455, Cur Avg Loss: 0.82470982, Log Avg loss: 0.79966078, Global Avg Loss: 2.58407180, Time: 0.0079 Steps: 28870, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001203, Sample Num: 19248, Cur Loss: 0.73204064, Cur Avg Loss: 0.82434162, Log Avg loss: 0.78041435, Global Avg Loss: 2.58344727, Time: 0.0117 Steps: 28880, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001213, Sample Num: 19408, Cur Loss: 0.63863552, Cur Avg Loss: 0.82357603, Log Avg loss: 0.73147561, Global Avg Loss: 2.58280622, Time: 0.0106 Steps: 28890, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001223, Sample Num: 19568, Cur Loss: 0.84015429, Cur Avg Loss: 0.82343018, Log Avg loss: 0.80573873, Global Avg Loss: 2.58219132, Time: 0.0158 Steps: 28900, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001233, Sample Num: 19728, Cur Loss: 0.41985750, Cur Avg Loss: 0.82179698, Log Avg loss: 0.62205686, Global Avg Loss: 2.58151331, Time: 0.0066 Steps: 28910, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001243, Sample Num: 19888, Cur Loss: 0.79416150, Cur Avg Loss: 0.82261459, Log Avg loss: 0.92342623, Global Avg Loss: 2.58093997, Time: 0.0066 Steps: 28920, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001253, Sample Num: 20048, Cur Loss: 0.38696292, Cur Avg Loss: 0.82317007, Log Avg loss: 0.89221629, Global Avg Loss: 2.58035625, Time: 0.0070 Steps: 28930, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001263, Sample Num: 20208, Cur Loss: 1.93221748, Cur Avg Loss: 0.82479424, Log Avg loss: 1.02830197, Global Avg Loss: 2.57981995, Time: 0.0115 Steps: 28940, Updated lr: 0.000074 Training, Epoch: 0014, Batch: 001273, Sample Num: 20368, Cur Loss: 1.57902551, Cur Avg Loss: 0.82502882, Log Avg loss: 0.85465725, Global Avg Loss: 2.57922403, Time: 0.0112 Steps: 28950, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001283, Sample Num: 20528, Cur Loss: 0.65062511, Cur Avg Loss: 0.82655472, Log Avg loss: 1.02080162, Global Avg Loss: 2.57868591, Time: 0.0194 Steps: 28960, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001293, Sample Num: 20688, Cur Loss: 0.79259574, Cur Avg Loss: 0.82740727, Log Avg loss: 0.93678876, Global Avg Loss: 2.57811915, Time: 0.0067 Steps: 28970, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001303, Sample Num: 20848, Cur Loss: 1.26318002, Cur Avg Loss: 0.82809964, Log Avg loss: 0.91762343, Global Avg Loss: 2.57754617, Time: 0.0069 Steps: 28980, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001313, Sample Num: 21008, Cur Loss: 0.74648088, Cur Avg Loss: 0.82790150, Log Avg loss: 0.80208331, Global Avg Loss: 2.57693373, Time: 0.0113 Steps: 28990, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001323, Sample Num: 21168, Cur Loss: 0.35948721, Cur Avg Loss: 0.82651522, Log Avg loss: 0.64449713, Global Avg Loss: 2.57626737, Time: 0.0135 Steps: 29000, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001333, Sample Num: 21328, Cur Loss: 0.80077207, Cur Avg Loss: 0.82614578, Log Avg loss: 0.77726893, Global Avg Loss: 2.57564724, Time: 0.0116 Steps: 29010, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001343, Sample Num: 21488, Cur Loss: 0.89260799, Cur Avg Loss: 0.82568721, Log Avg loss: 0.76455920, Global Avg Loss: 2.57502316, Time: 0.0088 Steps: 29020, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001353, Sample Num: 21648, Cur Loss: 1.25997579, Cur Avg Loss: 0.82664573, Log Avg loss: 0.95537498, Global Avg Loss: 2.57446524, Time: 0.0080 Steps: 29030, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001363, Sample Num: 21808, Cur Loss: 0.94534510, Cur Avg Loss: 0.82818169, Log Avg loss: 1.03599701, Global Avg Loss: 2.57393546, Time: 0.0115 Steps: 29040, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001373, Sample Num: 21968, Cur Loss: 1.68243515, Cur Avg Loss: 0.82791210, Log Avg loss: 0.79116739, Global Avg Loss: 2.57332177, Time: 0.0131 Steps: 29050, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001383, Sample Num: 22128, Cur Loss: 0.91902059, Cur Avg Loss: 0.82902838, Log Avg loss: 0.98229350, Global Avg Loss: 2.57277427, Time: 0.0120 Steps: 29060, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001393, Sample Num: 22288, Cur Loss: 0.92263699, Cur Avg Loss: 0.83050954, Log Avg loss: 1.03535476, Global Avg Loss: 2.57224540, Time: 0.0069 Steps: 29070, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001403, Sample Num: 22448, Cur Loss: 0.66847962, Cur Avg Loss: 0.83069030, Log Avg loss: 0.85586972, Global Avg Loss: 2.57165518, Time: 0.0120 Steps: 29080, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001413, Sample Num: 22608, Cur Loss: 0.27084535, Cur Avg Loss: 0.82967558, Log Avg loss: 0.68731101, Global Avg Loss: 2.57100742, Time: 0.0203 Steps: 29090, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001423, Sample Num: 22768, Cur Loss: 0.35990018, Cur Avg Loss: 0.82863073, Log Avg loss: 0.68099305, Global Avg Loss: 2.57035793, Time: 0.0086 Steps: 29100, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001433, Sample Num: 22928, Cur Loss: 0.09682006, Cur Avg Loss: 0.82736642, Log Avg loss: 0.64745471, Global Avg Loss: 2.56969736, Time: 0.0065 Steps: 29110, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001443, Sample Num: 23088, Cur Loss: 0.94336152, Cur Avg Loss: 0.82726946, Log Avg loss: 0.81337595, Global Avg Loss: 2.56909423, Time: 0.0129 Steps: 29120, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001453, Sample Num: 23248, Cur Loss: 0.85775483, Cur Avg Loss: 0.82787629, Log Avg loss: 0.91544121, Global Avg Loss: 2.56852655, Time: 0.0121 Steps: 29130, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001463, Sample Num: 23408, Cur Loss: 0.36544693, Cur Avg Loss: 0.82660596, Log Avg loss: 0.64202636, Global Avg Loss: 2.56786543, Time: 0.0112 Steps: 29140, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001473, Sample Num: 23568, Cur Loss: 1.39739430, Cur Avg Loss: 0.82733220, Log Avg loss: 0.93358167, Global Avg Loss: 2.56730478, Time: 0.0064 Steps: 29150, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001483, Sample Num: 23728, Cur Loss: 1.00853848, Cur Avg Loss: 0.82546547, Log Avg loss: 0.55049671, Global Avg Loss: 2.56661315, Time: 0.0111 Steps: 29160, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001493, Sample Num: 23888, Cur Loss: 1.12955761, Cur Avg Loss: 0.82697239, Log Avg loss: 1.05044869, Global Avg Loss: 2.56609338, Time: 0.0129 Steps: 29170, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001503, Sample Num: 24048, Cur Loss: 0.71241939, Cur Avg Loss: 0.82612197, Log Avg loss: 0.69915416, Global Avg Loss: 2.56545358, Time: 0.0108 Steps: 29180, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001513, Sample Num: 24208, Cur Loss: 0.32477307, Cur Avg Loss: 0.82670805, Log Avg loss: 0.91479483, Global Avg Loss: 2.56488809, Time: 0.0066 Steps: 29190, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001523, Sample Num: 24368, Cur Loss: 0.30719060, Cur Avg Loss: 0.82493091, Log Avg loss: 0.55604953, Global Avg Loss: 2.56420013, Time: 0.0065 Steps: 29200, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001533, Sample Num: 24528, Cur Loss: 0.44577646, Cur Avg Loss: 0.82549971, Log Avg loss: 0.91212863, Global Avg Loss: 2.56363455, Time: 0.0092 Steps: 29210, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001543, Sample Num: 24688, Cur Loss: 0.67955601, Cur Avg Loss: 0.82652066, Log Avg loss: 0.98303258, Global Avg Loss: 2.56309362, Time: 0.0129 Steps: 29220, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001553, Sample Num: 24848, Cur Loss: 0.24114937, Cur Avg Loss: 0.82528174, Log Avg loss: 0.63411684, Global Avg Loss: 2.56243369, Time: 0.0071 Steps: 29230, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001563, Sample Num: 25008, Cur Loss: 1.39277828, Cur Avg Loss: 0.82582877, Log Avg loss: 0.91078176, Global Avg Loss: 2.56186883, Time: 0.0068 Steps: 29240, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001573, Sample Num: 25168, Cur Loss: 1.10105658, Cur Avg Loss: 0.82680245, Log Avg loss: 0.97898839, Global Avg Loss: 2.56132767, Time: 0.0066 Steps: 29250, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001583, Sample Num: 25328, Cur Loss: 2.11763668, Cur Avg Loss: 0.82797900, Log Avg loss: 1.01305118, Global Avg Loss: 2.56079853, Time: 0.0072 Steps: 29260, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001593, Sample Num: 25488, Cur Loss: 0.62806773, Cur Avg Loss: 0.82819354, Log Avg loss: 0.86215522, Global Avg Loss: 2.56021819, Time: 0.0072 Steps: 29270, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001603, Sample Num: 25648, Cur Loss: 0.37316167, Cur Avg Loss: 0.82827787, Log Avg loss: 0.84171052, Global Avg Loss: 2.55963127, Time: 0.0065 Steps: 29280, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001613, Sample Num: 25808, Cur Loss: 1.17809844, Cur Avg Loss: 0.82806792, Log Avg loss: 0.79441305, Global Avg Loss: 2.55902860, Time: 0.0066 Steps: 29290, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001623, Sample Num: 25968, Cur Loss: 0.81408209, Cur Avg Loss: 0.82767305, Log Avg loss: 0.76398124, Global Avg Loss: 2.55841595, Time: 0.0067 Steps: 29300, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001633, Sample Num: 26128, Cur Loss: 0.92466795, Cur Avg Loss: 0.82691229, Log Avg loss: 0.70344137, Global Avg Loss: 2.55778307, Time: 0.0076 Steps: 29310, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001643, Sample Num: 26288, Cur Loss: 0.83024293, Cur Avg Loss: 0.82742860, Log Avg loss: 0.91174183, Global Avg Loss: 2.55722167, Time: 0.0172 Steps: 29320, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001653, Sample Num: 26448, Cur Loss: 0.71240580, Cur Avg Loss: 0.82751464, Log Avg loss: 0.84165023, Global Avg Loss: 2.55663675, Time: 0.0102 Steps: 29330, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001663, Sample Num: 26608, Cur Loss: 0.42977136, Cur Avg Loss: 0.82746488, Log Avg loss: 0.81924015, Global Avg Loss: 2.55604459, Time: 0.0068 Steps: 29340, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001673, Sample Num: 26768, Cur Loss: 1.01407123, Cur Avg Loss: 0.82778238, Log Avg loss: 0.88058149, Global Avg Loss: 2.55547373, Time: 0.0157 Steps: 29350, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001683, Sample Num: 26928, Cur Loss: 0.95867437, Cur Avg Loss: 0.82731575, Log Avg loss: 0.74924912, Global Avg Loss: 2.55485853, Time: 0.0137 Steps: 29360, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001693, Sample Num: 27088, Cur Loss: 0.76787198, Cur Avg Loss: 0.82727087, Log Avg loss: 0.81971708, Global Avg Loss: 2.55426775, Time: 0.0065 Steps: 29370, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001703, Sample Num: 27248, Cur Loss: 1.45410609, Cur Avg Loss: 0.82786113, Log Avg loss: 0.92779205, Global Avg Loss: 2.55371415, Time: 0.0066 Steps: 29380, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001713, Sample Num: 27408, Cur Loss: 1.00377679, Cur Avg Loss: 0.82694394, Log Avg loss: 0.67074661, Global Avg Loss: 2.55307346, Time: 0.0109 Steps: 29390, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001723, Sample Num: 27568, Cur Loss: 1.37495732, Cur Avg Loss: 0.82773769, Log Avg loss: 0.96370776, Global Avg Loss: 2.55253286, Time: 0.0109 Steps: 29400, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001733, Sample Num: 27728, Cur Loss: 0.29278922, Cur Avg Loss: 0.82726075, Log Avg loss: 0.74508341, Global Avg Loss: 2.55191829, Time: 0.0230 Steps: 29410, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001743, Sample Num: 27888, Cur Loss: 0.71036375, Cur Avg Loss: 0.82802592, Log Avg loss: 0.96063016, Global Avg Loss: 2.55137741, Time: 0.0125 Steps: 29420, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001753, Sample Num: 28048, Cur Loss: 0.33880371, Cur Avg Loss: 0.82679151, Log Avg loss: 0.61163481, Global Avg Loss: 2.55071830, Time: 0.0115 Steps: 29430, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001763, Sample Num: 28208, Cur Loss: 0.25446972, Cur Avg Loss: 0.82551504, Log Avg loss: 0.60174829, Global Avg Loss: 2.55005629, Time: 0.0073 Steps: 29440, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001773, Sample Num: 28368, Cur Loss: 0.49622333, Cur Avg Loss: 0.82521730, Log Avg loss: 0.77272721, Global Avg Loss: 2.54945278, Time: 0.0117 Steps: 29450, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001783, Sample Num: 28528, Cur Loss: 0.56280142, Cur Avg Loss: 0.82441860, Log Avg loss: 0.68280802, Global Avg Loss: 2.54881916, Time: 0.0105 Steps: 29460, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001793, Sample Num: 28688, Cur Loss: 0.70380068, Cur Avg Loss: 0.82450305, Log Avg loss: 0.83956021, Global Avg Loss: 2.54823916, Time: 0.0079 Steps: 29470, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001803, Sample Num: 28848, Cur Loss: 1.75711942, Cur Avg Loss: 0.82483075, Log Avg loss: 0.88358802, Global Avg Loss: 2.54767449, Time: 0.0067 Steps: 29480, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001813, Sample Num: 29008, Cur Loss: 0.49991298, Cur Avg Loss: 0.82363910, Log Avg loss: 0.60878406, Global Avg Loss: 2.54701702, Time: 0.0071 Steps: 29490, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001823, Sample Num: 29168, Cur Loss: 0.87368774, Cur Avg Loss: 0.82330987, Log Avg loss: 0.76362160, Global Avg Loss: 2.54641247, Time: 0.0106 Steps: 29500, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001833, Sample Num: 29328, Cur Loss: 0.49984598, Cur Avg Loss: 0.82322196, Log Avg loss: 0.80719460, Global Avg Loss: 2.54582311, Time: 0.0071 Steps: 29510, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001843, Sample Num: 29488, Cur Loss: 0.82937384, Cur Avg Loss: 0.82484940, Log Avg loss: 1.12315965, Global Avg Loss: 2.54534118, Time: 0.0068 Steps: 29520, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001853, Sample Num: 29648, Cur Loss: 0.51398104, Cur Avg Loss: 0.82556049, Log Avg loss: 0.95661532, Global Avg Loss: 2.54480317, Time: 0.0096 Steps: 29530, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001863, Sample Num: 29808, Cur Loss: 2.23491287, Cur Avg Loss: 0.82675795, Log Avg loss: 1.04864609, Global Avg Loss: 2.54429669, Time: 0.0113 Steps: 29540, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001873, Sample Num: 29968, Cur Loss: 1.31419742, Cur Avg Loss: 0.82711286, Log Avg loss: 0.89323276, Global Avg Loss: 2.54373795, Time: 0.0111 Steps: 29550, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001883, Sample Num: 30128, Cur Loss: 0.37617844, Cur Avg Loss: 0.82617020, Log Avg loss: 0.64961007, Global Avg Loss: 2.54309718, Time: 0.0110 Steps: 29560, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001893, Sample Num: 30288, Cur Loss: 0.76398277, Cur Avg Loss: 0.82581599, Log Avg loss: 0.75911786, Global Avg Loss: 2.54249387, Time: 0.0067 Steps: 29570, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001903, Sample Num: 30448, Cur Loss: 0.85836160, Cur Avg Loss: 0.82448203, Log Avg loss: 0.57196350, Global Avg Loss: 2.54182770, Time: 0.0131 Steps: 29580, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001913, Sample Num: 30608, Cur Loss: 0.87176645, Cur Avg Loss: 0.82340242, Log Avg loss: 0.61795423, Global Avg Loss: 2.54117752, Time: 0.0109 Steps: 29590, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001923, Sample Num: 30768, Cur Loss: 1.06485343, Cur Avg Loss: 0.82409095, Log Avg loss: 0.95580488, Global Avg Loss: 2.54064193, Time: 0.0114 Steps: 29600, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001933, Sample Num: 30928, Cur Loss: 0.99677467, Cur Avg Loss: 0.82414156, Log Avg loss: 0.83387564, Global Avg Loss: 2.54006551, Time: 0.0071 Steps: 29610, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001943, Sample Num: 31088, Cur Loss: 1.13654077, Cur Avg Loss: 0.82449912, Log Avg loss: 0.89361444, Global Avg Loss: 2.53950965, Time: 0.0115 Steps: 29620, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001953, Sample Num: 31248, Cur Loss: 0.81881070, Cur Avg Loss: 0.82408384, Log Avg loss: 0.74339517, Global Avg Loss: 2.53890347, Time: 0.0080 Steps: 29630, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001963, Sample Num: 31408, Cur Loss: 0.97490901, Cur Avg Loss: 0.82345133, Log Avg loss: 0.69992300, Global Avg Loss: 2.53828303, Time: 0.0110 Steps: 29640, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001973, Sample Num: 31568, Cur Loss: 0.61575699, Cur Avg Loss: 0.82369245, Log Avg loss: 0.87102242, Global Avg Loss: 2.53772072, Time: 0.0116 Steps: 29650, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001983, Sample Num: 31728, Cur Loss: 0.71105689, Cur Avg Loss: 0.82311685, Log Avg loss: 0.70955203, Global Avg Loss: 2.53710434, Time: 0.0066 Steps: 29660, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 001993, Sample Num: 31888, Cur Loss: 1.04189754, Cur Avg Loss: 0.82306291, Log Avg loss: 0.81236730, Global Avg Loss: 2.53652304, Time: 0.0112 Steps: 29670, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002003, Sample Num: 32048, Cur Loss: 0.52428293, Cur Avg Loss: 0.82398773, Log Avg loss: 1.00830310, Global Avg Loss: 2.53600814, Time: 0.0072 Steps: 29680, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002013, Sample Num: 32208, Cur Loss: 1.81006765, Cur Avg Loss: 0.82400777, Log Avg loss: 0.82802204, Global Avg Loss: 2.53543286, Time: 0.0148 Steps: 29690, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002023, Sample Num: 32368, Cur Loss: 0.78828603, Cur Avg Loss: 0.82281113, Log Avg loss: 0.58192809, Global Avg Loss: 2.53477512, Time: 0.0113 Steps: 29700, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002033, Sample Num: 32528, Cur Loss: 0.17329320, Cur Avg Loss: 0.82257654, Log Avg loss: 0.77511930, Global Avg Loss: 2.53418284, Time: 0.0115 Steps: 29710, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002043, Sample Num: 32688, Cur Loss: 0.96902269, Cur Avg Loss: 0.82362939, Log Avg loss: 1.03767385, Global Avg Loss: 2.53367931, Time: 0.0071 Steps: 29720, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002053, Sample Num: 32848, Cur Loss: 0.85026425, Cur Avg Loss: 0.82317637, Log Avg loss: 0.73062436, Global Avg Loss: 2.53307283, Time: 0.0068 Steps: 29730, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002063, Sample Num: 33008, Cur Loss: 0.58851326, Cur Avg Loss: 0.82245652, Log Avg loss: 0.67467147, Global Avg Loss: 2.53244795, Time: 0.0235 Steps: 29740, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002073, Sample Num: 33168, Cur Loss: 0.96471059, Cur Avg Loss: 0.82195699, Log Avg loss: 0.71890258, Global Avg Loss: 2.53183835, Time: 0.0111 Steps: 29750, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002083, Sample Num: 33328, Cur Loss: 0.89687717, Cur Avg Loss: 0.82084445, Log Avg loss: 0.59021530, Global Avg Loss: 2.53118592, Time: 0.0111 Steps: 29760, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002093, Sample Num: 33488, Cur Loss: 0.30269369, Cur Avg Loss: 0.82046800, Log Avg loss: 0.74205282, Global Avg Loss: 2.53058494, Time: 0.0070 Steps: 29770, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002103, Sample Num: 33648, Cur Loss: 1.10332024, Cur Avg Loss: 0.82115929, Log Avg loss: 0.96584802, Global Avg Loss: 2.53005951, Time: 0.0066 Steps: 29780, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002113, Sample Num: 33808, Cur Loss: 0.55729830, Cur Avg Loss: 0.82119609, Log Avg loss: 0.82893481, Global Avg Loss: 2.52948847, Time: 0.0112 Steps: 29790, Updated lr: 0.000073 Training, Epoch: 0014, Batch: 002123, Sample Num: 33968, Cur Loss: 0.34751672, Cur Avg Loss: 0.82051912, Log Avg loss: 0.67747396, Global Avg Loss: 2.52886699, Time: 0.0066 Steps: 29800, Updated lr: 0.000073 ***** Running evaluation checkpoint-29806 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-29806 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.574943, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.123618, "eval_total_loss": 789.903198, "eval_mae": 0.910215, "eval_mse": 1.123624, "eval_r2": 0.285751, "eval_sp_statistic": 0.610474, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.668635, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 2.219182, "test_total_loss": 1114.029566, "test_mae": 1.366958, "test_mse": 2.218809, "test_r2": -0.432039, "test_sp_statistic": 0.386936, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.523946, "test_ps_pvalue": 0.0, "lr": 7.26827880512091e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.528501858481691, "train_cur_epoch_loss": 1746.2522855699062, "train_cur_epoch_avg_loss": 0.8202218344621448, "train_cur_epoch_time": 22.57494306564331, "train_cur_epoch_avg_time": 0.010603543008756839, "epoch": 14, "step": 29806} ################################################## Training, Epoch: 0015, Batch: 000004, Sample Num: 64, Cur Loss: 0.86488491, Cur Avg Loss: 0.82023765, Log Avg loss: 0.75711515, Global Avg Loss: 2.52827264, Time: 0.0220 Steps: 29810, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000014, Sample Num: 224, Cur Loss: 0.51795489, Cur Avg Loss: 0.77545517, Log Avg loss: 0.75754218, Global Avg Loss: 2.52767883, Time: 0.0071 Steps: 29820, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000024, Sample Num: 384, Cur Loss: 0.82590318, Cur Avg Loss: 0.75687648, Log Avg loss: 0.73086632, Global Avg Loss: 2.52707648, Time: 0.0118 Steps: 29830, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000034, Sample Num: 544, Cur Loss: 0.43133691, Cur Avg Loss: 0.80861654, Log Avg loss: 0.93279266, Global Avg Loss: 2.52654220, Time: 0.0122 Steps: 29840, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000044, Sample Num: 704, Cur Loss: 0.53352457, Cur Avg Loss: 0.75940660, Log Avg loss: 0.59209283, Global Avg Loss: 2.52589415, Time: 0.0070 Steps: 29850, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000054, Sample Num: 864, Cur Loss: 0.90735638, Cur Avg Loss: 0.76618660, Log Avg loss: 0.79601858, Global Avg Loss: 2.52531482, Time: 0.0067 Steps: 29860, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000064, Sample Num: 1024, Cur Loss: 1.29404449, Cur Avg Loss: 0.77434044, Log Avg loss: 0.81837120, Global Avg Loss: 2.52474336, Time: 0.0068 Steps: 29870, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000074, Sample Num: 1184, Cur Loss: 0.67741430, Cur Avg Loss: 0.78315780, Log Avg loss: 0.83958887, Global Avg Loss: 2.52417939, Time: 0.0228 Steps: 29880, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000084, Sample Num: 1344, Cur Loss: 0.44497237, Cur Avg Loss: 0.76823297, Log Avg loss: 0.65778926, Global Avg Loss: 2.52355497, Time: 0.0066 Steps: 29890, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000094, Sample Num: 1504, Cur Loss: 0.65787184, Cur Avg Loss: 0.76907141, Log Avg loss: 0.77611425, Global Avg Loss: 2.52297054, Time: 0.0071 Steps: 29900, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000104, Sample Num: 1664, Cur Loss: 1.14265144, Cur Avg Loss: 0.77508988, Log Avg loss: 0.83166349, Global Avg Loss: 2.52240507, Time: 0.0195 Steps: 29910, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000114, Sample Num: 1824, Cur Loss: 1.60501933, Cur Avg Loss: 0.78204565, Log Avg loss: 0.85438573, Global Avg Loss: 2.52184758, Time: 0.0157 Steps: 29920, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000124, Sample Num: 1984, Cur Loss: 0.47832164, Cur Avg Loss: 0.78887001, Log Avg loss: 0.86666772, Global Avg Loss: 2.52129456, Time: 0.0063 Steps: 29930, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000134, Sample Num: 2144, Cur Loss: 0.44509998, Cur Avg Loss: 0.78889265, Log Avg loss: 0.78917330, Global Avg Loss: 2.52071603, Time: 0.0066 Steps: 29940, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000144, Sample Num: 2304, Cur Loss: 1.29394722, Cur Avg Loss: 0.80608338, Log Avg loss: 1.03643925, Global Avg Loss: 2.52022045, Time: 0.0114 Steps: 29950, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000154, Sample Num: 2464, Cur Loss: 0.78129745, Cur Avg Loss: 0.79809417, Log Avg loss: 0.68304946, Global Avg Loss: 2.51960724, Time: 0.0063 Steps: 29960, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000164, Sample Num: 2624, Cur Loss: 0.92936438, Cur Avg Loss: 0.81066278, Log Avg loss: 1.00421944, Global Avg Loss: 2.51910160, Time: 0.0063 Steps: 29970, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000174, Sample Num: 2784, Cur Loss: 0.32270342, Cur Avg Loss: 0.81086189, Log Avg loss: 0.81412731, Global Avg Loss: 2.51853290, Time: 0.0069 Steps: 29980, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000184, Sample Num: 2944, Cur Loss: 0.52936757, Cur Avg Loss: 0.81153992, Log Avg loss: 0.82333761, Global Avg Loss: 2.51796765, Time: 0.0063 Steps: 29990, Updated lr: 0.000073 Training, Epoch: 0015, Batch: 000194, Sample Num: 3104, Cur Loss: 1.08567631, Cur Avg Loss: 0.81637176, Log Avg loss: 0.90527765, Global Avg Loss: 2.51743008, Time: 0.0188 Steps: 30000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000204, Sample Num: 3264, Cur Loss: 1.42437327, Cur Avg Loss: 0.80741221, Log Avg loss: 0.63359694, Global Avg Loss: 2.51680235, Time: 0.0077 Steps: 30010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000214, Sample Num: 3424, Cur Loss: 1.37515283, Cur Avg Loss: 0.80963650, Log Avg loss: 0.85501193, Global Avg Loss: 2.51624879, Time: 0.0066 Steps: 30020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000224, Sample Num: 3584, Cur Loss: 1.25460446, Cur Avg Loss: 0.80177623, Log Avg loss: 0.63356645, Global Avg Loss: 2.51562185, Time: 0.0068 Steps: 30030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000234, Sample Num: 3744, Cur Loss: 0.83766657, Cur Avg Loss: 0.80771955, Log Avg loss: 0.94085003, Global Avg Loss: 2.51509763, Time: 0.0184 Steps: 30040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000244, Sample Num: 3904, Cur Loss: 0.57121909, Cur Avg Loss: 0.80933826, Log Avg loss: 0.84721608, Global Avg Loss: 2.51454259, Time: 0.0064 Steps: 30050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000254, Sample Num: 4064, Cur Loss: 0.81384605, Cur Avg Loss: 0.80347124, Log Avg loss: 0.66031581, Global Avg Loss: 2.51392575, Time: 0.0163 Steps: 30060, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000264, Sample Num: 4224, Cur Loss: 0.77849293, Cur Avg Loss: 0.80309120, Log Avg loss: 0.79343833, Global Avg Loss: 2.51335359, Time: 0.0092 Steps: 30070, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000274, Sample Num: 4384, Cur Loss: 0.88780373, Cur Avg Loss: 0.79713941, Log Avg loss: 0.64001209, Global Avg Loss: 2.51273080, Time: 0.0072 Steps: 30080, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000284, Sample Num: 4544, Cur Loss: 1.20559657, Cur Avg Loss: 0.80152321, Log Avg loss: 0.92163931, Global Avg Loss: 2.51220203, Time: 0.0071 Steps: 30090, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000294, Sample Num: 4704, Cur Loss: 1.12969542, Cur Avg Loss: 0.80167255, Log Avg loss: 0.80591389, Global Avg Loss: 2.51163515, Time: 0.0067 Steps: 30100, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000304, Sample Num: 4864, Cur Loss: 0.75942981, Cur Avg Loss: 0.80189516, Log Avg loss: 0.80843978, Global Avg Loss: 2.51106950, Time: 0.0109 Steps: 30110, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000314, Sample Num: 5024, Cur Loss: 0.45952505, Cur Avg Loss: 0.80363583, Log Avg loss: 0.85655234, Global Avg Loss: 2.51052019, Time: 0.0159 Steps: 30120, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000324, Sample Num: 5184, Cur Loss: 0.47280717, Cur Avg Loss: 0.80302436, Log Avg loss: 0.78382404, Global Avg Loss: 2.50994711, Time: 0.0126 Steps: 30130, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000334, Sample Num: 5344, Cur Loss: 0.47416544, Cur Avg Loss: 0.80546114, Log Avg loss: 0.88441297, Global Avg Loss: 2.50940778, Time: 0.0065 Steps: 30140, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000344, Sample Num: 5504, Cur Loss: 0.40525159, Cur Avg Loss: 0.80290561, Log Avg loss: 0.71755077, Global Avg Loss: 2.50881346, Time: 0.0068 Steps: 30150, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000354, Sample Num: 5664, Cur Loss: 0.99657398, Cur Avg Loss: 0.80449589, Log Avg loss: 0.85920163, Global Avg Loss: 2.50826651, Time: 0.0116 Steps: 30160, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000364, Sample Num: 5824, Cur Loss: 0.90693581, Cur Avg Loss: 0.80247434, Log Avg loss: 0.73091158, Global Avg Loss: 2.50767740, Time: 0.0066 Steps: 30170, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000374, Sample Num: 5984, Cur Loss: 0.37880784, Cur Avg Loss: 0.79695469, Log Avg loss: 0.59603911, Global Avg Loss: 2.50704398, Time: 0.0066 Steps: 30180, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000384, Sample Num: 6144, Cur Loss: 0.65675974, Cur Avg Loss: 0.80001657, Log Avg loss: 0.91453117, Global Avg Loss: 2.50651649, Time: 0.0072 Steps: 30190, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000394, Sample Num: 6304, Cur Loss: 0.53450894, Cur Avg Loss: 0.80443098, Log Avg loss: 0.97394428, Global Avg Loss: 2.50600901, Time: 0.0089 Steps: 30200, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000404, Sample Num: 6464, Cur Loss: 1.47192490, Cur Avg Loss: 0.81011872, Log Avg loss: 1.03421547, Global Avg Loss: 2.50552183, Time: 0.0068 Steps: 30210, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000414, Sample Num: 6624, Cur Loss: 0.42647141, Cur Avg Loss: 0.81120208, Log Avg loss: 0.85497013, Global Avg Loss: 2.50497565, Time: 0.0120 Steps: 30220, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000424, Sample Num: 6784, Cur Loss: 0.82030004, Cur Avg Loss: 0.80864577, Log Avg loss: 0.70281453, Global Avg Loss: 2.50437950, Time: 0.0230 Steps: 30230, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000434, Sample Num: 6944, Cur Loss: 0.28048262, Cur Avg Loss: 0.80909006, Log Avg loss: 0.82792784, Global Avg Loss: 2.50382512, Time: 0.0069 Steps: 30240, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000444, Sample Num: 7104, Cur Loss: 0.55591619, Cur Avg Loss: 0.80709297, Log Avg loss: 0.72041906, Global Avg Loss: 2.50323556, Time: 0.0066 Steps: 30250, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000454, Sample Num: 7264, Cur Loss: 0.92663723, Cur Avg Loss: 0.80902860, Log Avg loss: 0.89497085, Global Avg Loss: 2.50270408, Time: 0.0068 Steps: 30260, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000464, Sample Num: 7424, Cur Loss: 0.55113912, Cur Avg Loss: 0.80625166, Log Avg loss: 0.68017835, Global Avg Loss: 2.50210199, Time: 0.0116 Steps: 30270, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000474, Sample Num: 7584, Cur Loss: 0.82555908, Cur Avg Loss: 0.80196580, Log Avg loss: 0.60310194, Global Avg Loss: 2.50147484, Time: 0.0093 Steps: 30280, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000484, Sample Num: 7744, Cur Loss: 0.30468059, Cur Avg Loss: 0.79671027, Log Avg loss: 0.54759821, Global Avg Loss: 2.50082978, Time: 0.0110 Steps: 30290, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000494, Sample Num: 7904, Cur Loss: 0.54306316, Cur Avg Loss: 0.79448337, Log Avg loss: 0.68670152, Global Avg Loss: 2.50023106, Time: 0.0112 Steps: 30300, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000504, Sample Num: 8064, Cur Loss: 0.85028297, Cur Avg Loss: 0.79353028, Log Avg loss: 0.74644732, Global Avg Loss: 2.49965245, Time: 0.0108 Steps: 30310, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000514, Sample Num: 8224, Cur Loss: 0.42560777, Cur Avg Loss: 0.79914402, Log Avg loss: 1.08207666, Global Avg Loss: 2.49918491, Time: 0.0069 Steps: 30320, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000524, Sample Num: 8384, Cur Loss: 0.63952774, Cur Avg Loss: 0.80394524, Log Avg loss: 1.05072803, Global Avg Loss: 2.49870734, Time: 0.0066 Steps: 30330, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000534, Sample Num: 8544, Cur Loss: 1.17871439, Cur Avg Loss: 0.80397446, Log Avg loss: 0.80550561, Global Avg Loss: 2.49814927, Time: 0.0073 Steps: 30340, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000544, Sample Num: 8704, Cur Loss: 0.68229771, Cur Avg Loss: 0.80125555, Log Avg loss: 0.65606577, Global Avg Loss: 2.49754232, Time: 0.0120 Steps: 30350, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000554, Sample Num: 8864, Cur Loss: 0.91812932, Cur Avg Loss: 0.80070560, Log Avg loss: 0.77078846, Global Avg Loss: 2.49697356, Time: 0.0067 Steps: 30360, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000564, Sample Num: 9024, Cur Loss: 1.33954835, Cur Avg Loss: 0.80063868, Log Avg loss: 0.79693138, Global Avg Loss: 2.49641378, Time: 0.0127 Steps: 30370, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000574, Sample Num: 9184, Cur Loss: 1.00170279, Cur Avg Loss: 0.80003472, Log Avg loss: 0.76597098, Global Avg Loss: 2.49584418, Time: 0.0124 Steps: 30380, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000584, Sample Num: 9344, Cur Loss: 0.59497583, Cur Avg Loss: 0.79866089, Log Avg loss: 0.71980300, Global Avg Loss: 2.49525977, Time: 0.0067 Steps: 30390, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000594, Sample Num: 9504, Cur Loss: 0.34460706, Cur Avg Loss: 0.79632483, Log Avg loss: 0.65989917, Global Avg Loss: 2.49465603, Time: 0.0066 Steps: 30400, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000604, Sample Num: 9664, Cur Loss: 0.94285166, Cur Avg Loss: 0.79549591, Log Avg loss: 0.74625773, Global Avg Loss: 2.49408109, Time: 0.0079 Steps: 30410, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000614, Sample Num: 9824, Cur Loss: 0.69412929, Cur Avg Loss: 0.79424187, Log Avg loss: 0.71849806, Global Avg Loss: 2.49349740, Time: 0.0104 Steps: 30420, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000624, Sample Num: 9984, Cur Loss: 0.43886271, Cur Avg Loss: 0.79222995, Log Avg loss: 0.66869786, Global Avg Loss: 2.49289773, Time: 0.0097 Steps: 30430, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000634, Sample Num: 10144, Cur Loss: 1.03650427, Cur Avg Loss: 0.79368229, Log Avg loss: 0.88430890, Global Avg Loss: 2.49236928, Time: 0.0085 Steps: 30440, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000644, Sample Num: 10304, Cur Loss: 0.79803252, Cur Avg Loss: 0.79172355, Log Avg loss: 0.66753925, Global Avg Loss: 2.49177000, Time: 0.0136 Steps: 30450, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000654, Sample Num: 10464, Cur Loss: 0.48709947, Cur Avg Loss: 0.78991873, Log Avg loss: 0.67368844, Global Avg Loss: 2.49117312, Time: 0.0067 Steps: 30460, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000664, Sample Num: 10624, Cur Loss: 0.53747690, Cur Avg Loss: 0.79128366, Log Avg loss: 0.88055006, Global Avg Loss: 2.49064453, Time: 0.0118 Steps: 30470, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000674, Sample Num: 10784, Cur Loss: 0.48970932, Cur Avg Loss: 0.78994267, Log Avg loss: 0.70090093, Global Avg Loss: 2.49005734, Time: 0.0130 Steps: 30480, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000684, Sample Num: 10944, Cur Loss: 1.23051071, Cur Avg Loss: 0.79080019, Log Avg loss: 0.84859663, Global Avg Loss: 2.48951898, Time: 0.0065 Steps: 30490, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000694, Sample Num: 11104, Cur Loss: 1.03551579, Cur Avg Loss: 0.79140592, Log Avg loss: 0.83283832, Global Avg Loss: 2.48897581, Time: 0.0067 Steps: 30500, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000704, Sample Num: 11264, Cur Loss: 1.53963184, Cur Avg Loss: 0.79265629, Log Avg loss: 0.87943166, Global Avg Loss: 2.48844826, Time: 0.0227 Steps: 30510, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000714, Sample Num: 11424, Cur Loss: 0.55547273, Cur Avg Loss: 0.79061171, Log Avg loss: 0.64667353, Global Avg Loss: 2.48784480, Time: 0.0202 Steps: 30520, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000724, Sample Num: 11584, Cur Loss: 0.93728882, Cur Avg Loss: 0.79269638, Log Avg loss: 0.94154183, Global Avg Loss: 2.48733831, Time: 0.0066 Steps: 30530, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000734, Sample Num: 11744, Cur Loss: 1.24156201, Cur Avg Loss: 0.79415592, Log Avg loss: 0.89982620, Global Avg Loss: 2.48681850, Time: 0.0118 Steps: 30540, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000744, Sample Num: 11904, Cur Loss: 1.40006363, Cur Avg Loss: 0.79614253, Log Avg loss: 0.94195988, Global Avg Loss: 2.48631281, Time: 0.0143 Steps: 30550, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000754, Sample Num: 12064, Cur Loss: 0.19303422, Cur Avg Loss: 0.79265249, Log Avg loss: 0.53299352, Global Avg Loss: 2.48567364, Time: 0.0070 Steps: 30560, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000764, Sample Num: 12224, Cur Loss: 0.44331059, Cur Avg Loss: 0.79454896, Log Avg loss: 0.93754302, Global Avg Loss: 2.48516722, Time: 0.0068 Steps: 30570, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000774, Sample Num: 12384, Cur Loss: 0.54857910, Cur Avg Loss: 0.79499926, Log Avg loss: 0.82940166, Global Avg Loss: 2.48462576, Time: 0.0130 Steps: 30580, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000784, Sample Num: 12544, Cur Loss: 0.46896285, Cur Avg Loss: 0.79372352, Log Avg loss: 0.69498152, Global Avg Loss: 2.48404072, Time: 0.0115 Steps: 30590, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000794, Sample Num: 12704, Cur Loss: 1.47698545, Cur Avg Loss: 0.79637374, Log Avg loss: 1.00415093, Global Avg Loss: 2.48355710, Time: 0.0093 Steps: 30600, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000804, Sample Num: 12864, Cur Loss: 0.53447342, Cur Avg Loss: 0.79732860, Log Avg loss: 0.87314441, Global Avg Loss: 2.48303099, Time: 0.0127 Steps: 30610, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000814, Sample Num: 13024, Cur Loss: 0.54908776, Cur Avg Loss: 0.79662980, Log Avg loss: 0.74044676, Global Avg Loss: 2.48246189, Time: 0.0182 Steps: 30620, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000824, Sample Num: 13184, Cur Loss: 0.35085177, Cur Avg Loss: 0.79725685, Log Avg loss: 0.84829835, Global Avg Loss: 2.48192837, Time: 0.0068 Steps: 30630, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000834, Sample Num: 13344, Cur Loss: 0.41387871, Cur Avg Loss: 0.79929044, Log Avg loss: 0.96685856, Global Avg Loss: 2.48143390, Time: 0.0137 Steps: 30640, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000844, Sample Num: 13504, Cur Loss: 0.16816434, Cur Avg Loss: 0.80108656, Log Avg loss: 0.95088289, Global Avg Loss: 2.48093453, Time: 0.0066 Steps: 30650, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000854, Sample Num: 13664, Cur Loss: 1.14484942, Cur Avg Loss: 0.80284469, Log Avg loss: 0.95123050, Global Avg Loss: 2.48043561, Time: 0.0132 Steps: 30660, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000864, Sample Num: 13824, Cur Loss: 0.35737428, Cur Avg Loss: 0.80256427, Log Avg loss: 0.77861677, Global Avg Loss: 2.47988073, Time: 0.0101 Steps: 30670, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000874, Sample Num: 13984, Cur Loss: 0.82536054, Cur Avg Loss: 0.80247127, Log Avg loss: 0.79443558, Global Avg Loss: 2.47933137, Time: 0.0076 Steps: 30680, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000884, Sample Num: 14144, Cur Loss: 0.61784244, Cur Avg Loss: 0.80037185, Log Avg loss: 0.61688262, Global Avg Loss: 2.47872451, Time: 0.0136 Steps: 30690, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000894, Sample Num: 14304, Cur Loss: 0.88648832, Cur Avg Loss: 0.80102410, Log Avg loss: 0.85868302, Global Avg Loss: 2.47819681, Time: 0.0108 Steps: 30700, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000904, Sample Num: 14464, Cur Loss: 0.42813167, Cur Avg Loss: 0.80163199, Log Avg loss: 0.85597790, Global Avg Loss: 2.47766857, Time: 0.0137 Steps: 30710, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000914, Sample Num: 14624, Cur Loss: 1.08134341, Cur Avg Loss: 0.80160302, Log Avg loss: 0.79898368, Global Avg Loss: 2.47712212, Time: 0.0068 Steps: 30720, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000924, Sample Num: 14784, Cur Loss: 0.49103272, Cur Avg Loss: 0.80205144, Log Avg loss: 0.84303744, Global Avg Loss: 2.47659037, Time: 0.0068 Steps: 30730, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000934, Sample Num: 14944, Cur Loss: 1.21784151, Cur Avg Loss: 0.80333328, Log Avg loss: 0.92177517, Global Avg Loss: 2.47608457, Time: 0.0073 Steps: 30740, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000944, Sample Num: 15104, Cur Loss: 1.32348168, Cur Avg Loss: 0.80400251, Log Avg loss: 0.86650858, Global Avg Loss: 2.47556113, Time: 0.0068 Steps: 30750, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000954, Sample Num: 15264, Cur Loss: 1.10068750, Cur Avg Loss: 0.80352434, Log Avg loss: 0.75838481, Global Avg Loss: 2.47500288, Time: 0.0124 Steps: 30760, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000964, Sample Num: 15424, Cur Loss: 0.51351011, Cur Avg Loss: 0.80214443, Log Avg loss: 0.67050081, Global Avg Loss: 2.47441643, Time: 0.0117 Steps: 30770, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000974, Sample Num: 15584, Cur Loss: 0.35264140, Cur Avg Loss: 0.80125893, Log Avg loss: 0.71589695, Global Avg Loss: 2.47384511, Time: 0.0067 Steps: 30780, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000984, Sample Num: 15744, Cur Loss: 0.48062038, Cur Avg Loss: 0.80087466, Log Avg loss: 0.76344641, Global Avg Loss: 2.47328961, Time: 0.0069 Steps: 30790, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 000994, Sample Num: 15904, Cur Loss: 1.08599186, Cur Avg Loss: 0.80119483, Log Avg loss: 0.83270046, Global Avg Loss: 2.47275695, Time: 0.0117 Steps: 30800, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001004, Sample Num: 16064, Cur Loss: 0.56793606, Cur Avg Loss: 0.80086288, Log Avg loss: 0.76786613, Global Avg Loss: 2.47220359, Time: 0.0068 Steps: 30810, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001014, Sample Num: 16224, Cur Loss: 1.22498739, Cur Avg Loss: 0.80336396, Log Avg loss: 1.05447271, Global Avg Loss: 2.47174359, Time: 0.0121 Steps: 30820, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001024, Sample Num: 16384, Cur Loss: 0.87311697, Cur Avg Loss: 0.80340225, Log Avg loss: 0.80728486, Global Avg Loss: 2.47120371, Time: 0.0177 Steps: 30830, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001034, Sample Num: 16544, Cur Loss: 0.63081568, Cur Avg Loss: 0.80299199, Log Avg loss: 0.76098134, Global Avg Loss: 2.47064916, Time: 0.0112 Steps: 30840, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001044, Sample Num: 16704, Cur Loss: 1.47162580, Cur Avg Loss: 0.80446306, Log Avg loss: 0.95657211, Global Avg Loss: 2.47015837, Time: 0.0072 Steps: 30850, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001054, Sample Num: 16864, Cur Loss: 1.13232768, Cur Avg Loss: 0.80551300, Log Avg loss: 0.91512629, Global Avg Loss: 2.46965447, Time: 0.0065 Steps: 30860, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001064, Sample Num: 17024, Cur Loss: 0.99919546, Cur Avg Loss: 0.80621407, Log Avg loss: 0.88010724, Global Avg Loss: 2.46913956, Time: 0.0072 Steps: 30870, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001074, Sample Num: 17184, Cur Loss: 1.11875296, Cur Avg Loss: 0.80729798, Log Avg loss: 0.92262525, Global Avg Loss: 2.46863874, Time: 0.0067 Steps: 30880, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001084, Sample Num: 17344, Cur Loss: 0.91544157, Cur Avg Loss: 0.80868955, Log Avg loss: 0.95814496, Global Avg Loss: 2.46814975, Time: 0.0140 Steps: 30890, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001094, Sample Num: 17504, Cur Loss: 1.47048497, Cur Avg Loss: 0.80976621, Log Avg loss: 0.92647583, Global Avg Loss: 2.46765083, Time: 0.0071 Steps: 30900, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001104, Sample Num: 17664, Cur Loss: 0.76942366, Cur Avg Loss: 0.80987791, Log Avg loss: 0.82209753, Global Avg Loss: 2.46711846, Time: 0.0236 Steps: 30910, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001114, Sample Num: 17824, Cur Loss: 0.48800257, Cur Avg Loss: 0.80842821, Log Avg loss: 0.64838188, Global Avg Loss: 2.46653025, Time: 0.0068 Steps: 30920, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001124, Sample Num: 17984, Cur Loss: 0.47513944, Cur Avg Loss: 0.80736637, Log Avg loss: 0.68907777, Global Avg Loss: 2.46595558, Time: 0.0121 Steps: 30930, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001134, Sample Num: 18144, Cur Loss: 0.54346228, Cur Avg Loss: 0.80565719, Log Avg loss: 0.61354431, Global Avg Loss: 2.46535687, Time: 0.0086 Steps: 30940, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001144, Sample Num: 18304, Cur Loss: 0.53193462, Cur Avg Loss: 0.80448944, Log Avg loss: 0.67206756, Global Avg Loss: 2.46477746, Time: 0.0138 Steps: 30950, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001154, Sample Num: 18464, Cur Loss: 1.10546350, Cur Avg Loss: 0.80344019, Log Avg loss: 0.68340575, Global Avg Loss: 2.46420208, Time: 0.0073 Steps: 30960, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001164, Sample Num: 18624, Cur Loss: 1.06255424, Cur Avg Loss: 0.80510381, Log Avg loss: 0.99708522, Global Avg Loss: 2.46372836, Time: 0.0092 Steps: 30970, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001174, Sample Num: 18784, Cur Loss: 1.28238416, Cur Avg Loss: 0.80490508, Log Avg loss: 0.78177285, Global Avg Loss: 2.46318544, Time: 0.0112 Steps: 30980, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001184, Sample Num: 18944, Cur Loss: 0.81252897, Cur Avg Loss: 0.80480235, Log Avg loss: 0.79274222, Global Avg Loss: 2.46264641, Time: 0.0139 Steps: 30990, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001194, Sample Num: 19104, Cur Loss: 0.80871201, Cur Avg Loss: 0.80588346, Log Avg loss: 0.93388667, Global Avg Loss: 2.46215327, Time: 0.0081 Steps: 31000, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001204, Sample Num: 19264, Cur Loss: 0.88487911, Cur Avg Loss: 0.80717979, Log Avg loss: 0.96196159, Global Avg Loss: 2.46166949, Time: 0.0108 Steps: 31010, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001214, Sample Num: 19424, Cur Loss: 0.23028663, Cur Avg Loss: 0.80739099, Log Avg loss: 0.83281903, Global Avg Loss: 2.46114439, Time: 0.0137 Steps: 31020, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001224, Sample Num: 19584, Cur Loss: 0.75012428, Cur Avg Loss: 0.80860883, Log Avg loss: 0.95645465, Global Avg Loss: 2.46065948, Time: 0.0068 Steps: 31030, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001234, Sample Num: 19744, Cur Loss: 0.67029595, Cur Avg Loss: 0.80714842, Log Avg loss: 0.62839485, Global Avg Loss: 2.46006919, Time: 0.0090 Steps: 31040, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001244, Sample Num: 19904, Cur Loss: 0.69430792, Cur Avg Loss: 0.80619864, Log Avg loss: 0.68899516, Global Avg Loss: 2.45949879, Time: 0.0114 Steps: 31050, Updated lr: 0.000072 Training, Epoch: 0015, Batch: 001254, Sample Num: 20064, Cur Loss: 0.68014765, Cur Avg Loss: 0.80596235, Log Avg loss: 0.77656831, Global Avg Loss: 2.45895696, Time: 0.0067 Steps: 31060, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001264, Sample Num: 20224, Cur Loss: 0.45339066, Cur Avg Loss: 0.80602677, Log Avg loss: 0.81410447, Global Avg Loss: 2.45842756, Time: 0.0070 Steps: 31070, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001274, Sample Num: 20384, Cur Loss: 0.75611109, Cur Avg Loss: 0.80511876, Log Avg loss: 0.69034666, Global Avg Loss: 2.45785868, Time: 0.0118 Steps: 31080, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001284, Sample Num: 20544, Cur Loss: 0.89553356, Cur Avg Loss: 0.80562023, Log Avg loss: 0.86950775, Global Avg Loss: 2.45734779, Time: 0.0067 Steps: 31090, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001294, Sample Num: 20704, Cur Loss: 0.26590931, Cur Avg Loss: 0.80591722, Log Avg loss: 0.84405105, Global Avg Loss: 2.45682904, Time: 0.0064 Steps: 31100, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001304, Sample Num: 20864, Cur Loss: 0.28759119, Cur Avg Loss: 0.80511153, Log Avg loss: 0.70085428, Global Avg Loss: 2.45626460, Time: 0.0067 Steps: 31110, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001314, Sample Num: 21024, Cur Loss: 0.81127632, Cur Avg Loss: 0.80563545, Log Avg loss: 0.87395558, Global Avg Loss: 2.45575615, Time: 0.0067 Steps: 31120, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001324, Sample Num: 21184, Cur Loss: 0.49675548, Cur Avg Loss: 0.80470078, Log Avg loss: 0.68188472, Global Avg Loss: 2.45518632, Time: 0.0067 Steps: 31130, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001334, Sample Num: 21344, Cur Loss: 0.81427372, Cur Avg Loss: 0.80566140, Log Avg loss: 0.93284789, Global Avg Loss: 2.45469745, Time: 0.0065 Steps: 31140, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001344, Sample Num: 21504, Cur Loss: 0.57757252, Cur Avg Loss: 0.80677446, Log Avg loss: 0.95525647, Global Avg Loss: 2.45421609, Time: 0.0067 Steps: 31150, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001354, Sample Num: 21664, Cur Loss: 0.21896723, Cur Avg Loss: 0.80516013, Log Avg loss: 0.58819382, Global Avg Loss: 2.45361724, Time: 0.0121 Steps: 31160, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001364, Sample Num: 21824, Cur Loss: 0.88464797, Cur Avg Loss: 0.80583167, Log Avg loss: 0.89675800, Global Avg Loss: 2.45311777, Time: 0.0122 Steps: 31170, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001374, Sample Num: 21984, Cur Loss: 0.89165664, Cur Avg Loss: 0.80550484, Log Avg loss: 0.76092562, Global Avg Loss: 2.45257505, Time: 0.0128 Steps: 31180, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001384, Sample Num: 22144, Cur Loss: 0.37753195, Cur Avg Loss: 0.80495740, Log Avg loss: 0.72973946, Global Avg Loss: 2.45202268, Time: 0.0140 Steps: 31190, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001394, Sample Num: 22304, Cur Loss: 0.87414193, Cur Avg Loss: 0.80531467, Log Avg loss: 0.85475990, Global Avg Loss: 2.45151074, Time: 0.0112 Steps: 31200, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001404, Sample Num: 22464, Cur Loss: 0.57469922, Cur Avg Loss: 0.80599656, Log Avg loss: 0.90105301, Global Avg Loss: 2.45101396, Time: 0.0120 Steps: 31210, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001414, Sample Num: 22624, Cur Loss: 0.81915355, Cur Avg Loss: 0.80676972, Log Avg loss: 0.91532113, Global Avg Loss: 2.45052206, Time: 0.0116 Steps: 31220, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001424, Sample Num: 22784, Cur Loss: 1.18561029, Cur Avg Loss: 0.80926922, Log Avg loss: 1.16269831, Global Avg Loss: 2.45010969, Time: 0.0088 Steps: 31230, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001434, Sample Num: 22944, Cur Loss: 0.35145020, Cur Avg Loss: 0.80964171, Log Avg loss: 0.86268411, Global Avg Loss: 2.44960156, Time: 0.0068 Steps: 31240, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001444, Sample Num: 23104, Cur Loss: 1.30338454, Cur Avg Loss: 0.81067619, Log Avg loss: 0.95902128, Global Avg Loss: 2.44912457, Time: 0.0067 Steps: 31250, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001454, Sample Num: 23264, Cur Loss: 0.92786181, Cur Avg Loss: 0.80985608, Log Avg loss: 0.69143248, Global Avg Loss: 2.44856229, Time: 0.0067 Steps: 31260, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001464, Sample Num: 23424, Cur Loss: 0.85683668, Cur Avg Loss: 0.81210609, Log Avg loss: 1.13925762, Global Avg Loss: 2.44814358, Time: 0.0067 Steps: 31270, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001474, Sample Num: 23584, Cur Loss: 1.19411004, Cur Avg Loss: 0.81157031, Log Avg loss: 0.73313108, Global Avg Loss: 2.44759530, Time: 0.0107 Steps: 31280, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001484, Sample Num: 23744, Cur Loss: 0.47648534, Cur Avg Loss: 0.81100890, Log Avg loss: 0.72825777, Global Avg Loss: 2.44704582, Time: 0.0139 Steps: 31290, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001494, Sample Num: 23904, Cur Loss: 0.40981284, Cur Avg Loss: 0.80927539, Log Avg loss: 0.55202189, Global Avg Loss: 2.44644038, Time: 0.0124 Steps: 31300, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001504, Sample Num: 24064, Cur Loss: 0.48644143, Cur Avg Loss: 0.80850282, Log Avg loss: 0.69308061, Global Avg Loss: 2.44588038, Time: 0.0110 Steps: 31310, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001514, Sample Num: 24224, Cur Loss: 1.39127302, Cur Avg Loss: 0.80790736, Log Avg loss: 0.71835045, Global Avg Loss: 2.44532880, Time: 0.0115 Steps: 31320, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001524, Sample Num: 24384, Cur Loss: 0.40336198, Cur Avg Loss: 0.80592623, Log Avg loss: 0.50598359, Global Avg Loss: 2.44470980, Time: 0.0127 Steps: 31330, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001534, Sample Num: 24544, Cur Loss: 0.63061595, Cur Avg Loss: 0.80458672, Log Avg loss: 0.60044506, Global Avg Loss: 2.44412133, Time: 0.0117 Steps: 31340, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001544, Sample Num: 24704, Cur Loss: 1.27254105, Cur Avg Loss: 0.80486846, Log Avg loss: 0.84808711, Global Avg Loss: 2.44361223, Time: 0.0119 Steps: 31350, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001554, Sample Num: 24864, Cur Loss: 0.24676518, Cur Avg Loss: 0.80412378, Log Avg loss: 0.68914501, Global Avg Loss: 2.44305277, Time: 0.0110 Steps: 31360, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001564, Sample Num: 25024, Cur Loss: 0.70590883, Cur Avg Loss: 0.80307040, Log Avg loss: 0.63937596, Global Avg Loss: 2.44247780, Time: 0.0070 Steps: 31370, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001574, Sample Num: 25184, Cur Loss: 0.64394403, Cur Avg Loss: 0.80458500, Log Avg loss: 1.04146817, Global Avg Loss: 2.44203133, Time: 0.0116 Steps: 31380, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001584, Sample Num: 25344, Cur Loss: 0.35574901, Cur Avg Loss: 0.80318358, Log Avg loss: 0.58259978, Global Avg Loss: 2.44143897, Time: 0.0068 Steps: 31390, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001594, Sample Num: 25504, Cur Loss: 0.61137652, Cur Avg Loss: 0.80273835, Log Avg loss: 0.73221355, Global Avg Loss: 2.44089463, Time: 0.0142 Steps: 31400, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001604, Sample Num: 25664, Cur Loss: 0.36890620, Cur Avg Loss: 0.80249400, Log Avg loss: 0.76354546, Global Avg Loss: 2.44036061, Time: 0.0110 Steps: 31410, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001614, Sample Num: 25824, Cur Loss: 0.45439723, Cur Avg Loss: 0.80114663, Log Avg loss: 0.58502809, Global Avg Loss: 2.43977012, Time: 0.0090 Steps: 31420, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001624, Sample Num: 25984, Cur Loss: 1.46848655, Cur Avg Loss: 0.80313603, Log Avg loss: 1.12422545, Global Avg Loss: 2.43935155, Time: 0.0110 Steps: 31430, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001634, Sample Num: 26144, Cur Loss: 0.36927682, Cur Avg Loss: 0.80304767, Log Avg loss: 0.78869785, Global Avg Loss: 2.43882654, Time: 0.0070 Steps: 31440, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001644, Sample Num: 26304, Cur Loss: 0.90409517, Cur Avg Loss: 0.80335343, Log Avg loss: 0.85331466, Global Avg Loss: 2.43832240, Time: 0.0160 Steps: 31450, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001654, Sample Num: 26464, Cur Loss: 0.76732969, Cur Avg Loss: 0.80368217, Log Avg loss: 0.85772739, Global Avg Loss: 2.43781998, Time: 0.0118 Steps: 31460, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001664, Sample Num: 26624, Cur Loss: 0.71974432, Cur Avg Loss: 0.80334151, Log Avg loss: 0.74699590, Global Avg Loss: 2.43728270, Time: 0.0110 Steps: 31470, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001674, Sample Num: 26784, Cur Loss: 0.88747901, Cur Avg Loss: 0.80416897, Log Avg loss: 0.94185860, Global Avg Loss: 2.43680766, Time: 0.0109 Steps: 31480, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001684, Sample Num: 26944, Cur Loss: 0.36125132, Cur Avg Loss: 0.80351525, Log Avg loss: 0.69408183, Global Avg Loss: 2.43625424, Time: 0.0085 Steps: 31490, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001694, Sample Num: 27104, Cur Loss: 0.70325840, Cur Avg Loss: 0.80286326, Log Avg loss: 0.69306941, Global Avg Loss: 2.43570085, Time: 0.0066 Steps: 31500, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001704, Sample Num: 27264, Cur Loss: 0.70378947, Cur Avg Loss: 0.80338227, Log Avg loss: 0.89130114, Global Avg Loss: 2.43521072, Time: 0.0118 Steps: 31510, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001714, Sample Num: 27424, Cur Loss: 1.48538458, Cur Avg Loss: 0.80445633, Log Avg loss: 0.98747728, Global Avg Loss: 2.43475141, Time: 0.0117 Steps: 31520, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001724, Sample Num: 27584, Cur Loss: 0.52621341, Cur Avg Loss: 0.80414520, Log Avg loss: 0.75081727, Global Avg Loss: 2.43421734, Time: 0.0082 Steps: 31530, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001734, Sample Num: 27744, Cur Loss: 0.45912188, Cur Avg Loss: 0.80426049, Log Avg loss: 0.82413545, Global Avg Loss: 2.43370685, Time: 0.0154 Steps: 31540, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001744, Sample Num: 27904, Cur Loss: 0.76019621, Cur Avg Loss: 0.80341483, Log Avg loss: 0.65677747, Global Avg Loss: 2.43314364, Time: 0.0074 Steps: 31550, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001754, Sample Num: 28064, Cur Loss: 0.57082260, Cur Avg Loss: 0.80366674, Log Avg loss: 0.84760066, Global Avg Loss: 2.43264125, Time: 0.0142 Steps: 31560, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001764, Sample Num: 28224, Cur Loss: 0.26362130, Cur Avg Loss: 0.80336560, Log Avg loss: 0.75054525, Global Avg Loss: 2.43210844, Time: 0.0110 Steps: 31570, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001774, Sample Num: 28384, Cur Loss: 1.10036027, Cur Avg Loss: 0.80373982, Log Avg loss: 0.86975319, Global Avg Loss: 2.43161371, Time: 0.0117 Steps: 31580, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001784, Sample Num: 28544, Cur Loss: 0.71933454, Cur Avg Loss: 0.80392591, Log Avg loss: 0.83693703, Global Avg Loss: 2.43110890, Time: 0.0119 Steps: 31590, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001794, Sample Num: 28704, Cur Loss: 0.49132237, Cur Avg Loss: 0.80291721, Log Avg loss: 0.62296655, Global Avg Loss: 2.43053670, Time: 0.0139 Steps: 31600, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001804, Sample Num: 28864, Cur Loss: 0.67233431, Cur Avg Loss: 0.80340543, Log Avg loss: 0.89099212, Global Avg Loss: 2.43004966, Time: 0.0089 Steps: 31610, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001814, Sample Num: 29024, Cur Loss: 0.47911489, Cur Avg Loss: 0.80253577, Log Avg loss: 0.64564805, Global Avg Loss: 2.42948533, Time: 0.0129 Steps: 31620, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001824, Sample Num: 29184, Cur Loss: 2.21348739, Cur Avg Loss: 0.80344021, Log Avg loss: 0.96750649, Global Avg Loss: 2.42902312, Time: 0.0105 Steps: 31630, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001834, Sample Num: 29344, Cur Loss: 0.35107240, Cur Avg Loss: 0.80237155, Log Avg loss: 0.60744776, Global Avg Loss: 2.42844740, Time: 0.0071 Steps: 31640, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001844, Sample Num: 29504, Cur Loss: 0.65196407, Cur Avg Loss: 0.80235981, Log Avg loss: 0.80020628, Global Avg Loss: 2.42793295, Time: 0.0084 Steps: 31650, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001854, Sample Num: 29664, Cur Loss: 1.43774700, Cur Avg Loss: 0.80378833, Log Avg loss: 1.06720750, Global Avg Loss: 2.42750316, Time: 0.0112 Steps: 31660, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001864, Sample Num: 29824, Cur Loss: 0.22749418, Cur Avg Loss: 0.80310574, Log Avg loss: 0.67655383, Global Avg Loss: 2.42695028, Time: 0.0132 Steps: 31670, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001874, Sample Num: 29984, Cur Loss: 1.29960549, Cur Avg Loss: 0.80247150, Log Avg loss: 0.68424794, Global Avg Loss: 2.42640019, Time: 0.0105 Steps: 31680, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001884, Sample Num: 30144, Cur Loss: 1.64331937, Cur Avg Loss: 0.80234925, Log Avg loss: 0.77944030, Global Avg Loss: 2.42588048, Time: 0.0071 Steps: 31690, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001894, Sample Num: 30304, Cur Loss: 0.67264289, Cur Avg Loss: 0.80137207, Log Avg loss: 0.61727070, Global Avg Loss: 2.42530994, Time: 0.0119 Steps: 31700, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001904, Sample Num: 30464, Cur Loss: 0.61717546, Cur Avg Loss: 0.80014286, Log Avg loss: 0.56733170, Global Avg Loss: 2.42472401, Time: 0.0107 Steps: 31710, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001914, Sample Num: 30624, Cur Loss: 0.67037207, Cur Avg Loss: 0.79885345, Log Avg loss: 0.55334995, Global Avg Loss: 2.42413404, Time: 0.0066 Steps: 31720, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001924, Sample Num: 30784, Cur Loss: 0.98899460, Cur Avg Loss: 0.79857093, Log Avg loss: 0.74449660, Global Avg Loss: 2.42360469, Time: 0.0066 Steps: 31730, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001934, Sample Num: 30944, Cur Loss: 0.85758114, Cur Avg Loss: 0.79959538, Log Avg loss: 0.99669952, Global Avg Loss: 2.42315513, Time: 0.0135 Steps: 31740, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001944, Sample Num: 31104, Cur Loss: 1.96420372, Cur Avg Loss: 0.80068053, Log Avg loss: 1.01054861, Global Avg Loss: 2.42271022, Time: 0.0142 Steps: 31750, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001954, Sample Num: 31264, Cur Loss: 0.31797963, Cur Avg Loss: 0.79914268, Log Avg loss: 0.50018350, Global Avg Loss: 2.42210489, Time: 0.0085 Steps: 31760, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001964, Sample Num: 31424, Cur Loss: 0.34380233, Cur Avg Loss: 0.79915528, Log Avg loss: 0.80161765, Global Avg Loss: 2.42159482, Time: 0.0119 Steps: 31770, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001974, Sample Num: 31584, Cur Loss: 0.31755757, Cur Avg Loss: 0.80006637, Log Avg loss: 0.97900375, Global Avg Loss: 2.42114089, Time: 0.0084 Steps: 31780, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001984, Sample Num: 31744, Cur Loss: 0.46824372, Cur Avg Loss: 0.80078291, Log Avg loss: 0.94222846, Global Avg Loss: 2.42067567, Time: 0.0071 Steps: 31790, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 001994, Sample Num: 31904, Cur Loss: 0.32902810, Cur Avg Loss: 0.80071148, Log Avg loss: 0.78654049, Global Avg Loss: 2.42016180, Time: 0.0070 Steps: 31800, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002004, Sample Num: 32064, Cur Loss: 0.27459911, Cur Avg Loss: 0.80185400, Log Avg loss: 1.02967267, Global Avg Loss: 2.41972467, Time: 0.0132 Steps: 31810, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002014, Sample Num: 32224, Cur Loss: 0.36905548, Cur Avg Loss: 0.80204135, Log Avg loss: 0.83958674, Global Avg Loss: 2.41922809, Time: 0.0084 Steps: 31820, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002024, Sample Num: 32384, Cur Loss: 0.45988020, Cur Avg Loss: 0.80146084, Log Avg loss: 0.68454559, Global Avg Loss: 2.41868310, Time: 0.0117 Steps: 31830, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002034, Sample Num: 32544, Cur Loss: 0.46668744, Cur Avg Loss: 0.80187763, Log Avg loss: 0.88623467, Global Avg Loss: 2.41820181, Time: 0.0114 Steps: 31840, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002044, Sample Num: 32704, Cur Loss: 0.38096744, Cur Avg Loss: 0.80219027, Log Avg loss: 0.86578274, Global Avg Loss: 2.41771439, Time: 0.0118 Steps: 31850, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002054, Sample Num: 32864, Cur Loss: 0.87193167, Cur Avg Loss: 0.80246488, Log Avg loss: 0.85859411, Global Avg Loss: 2.41722502, Time: 0.0109 Steps: 31860, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002064, Sample Num: 33024, Cur Loss: 0.58324647, Cur Avg Loss: 0.80153105, Log Avg loss: 0.60972375, Global Avg Loss: 2.41665788, Time: 0.0118 Steps: 31870, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002074, Sample Num: 33184, Cur Loss: 0.49126238, Cur Avg Loss: 0.80051333, Log Avg loss: 0.59045460, Global Avg Loss: 2.41608504, Time: 0.0132 Steps: 31880, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002084, Sample Num: 33344, Cur Loss: 0.28899911, Cur Avg Loss: 0.79933993, Log Avg loss: 0.55597621, Global Avg Loss: 2.41550175, Time: 0.0068 Steps: 31890, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002094, Sample Num: 33504, Cur Loss: 0.54345238, Cur Avg Loss: 0.79948099, Log Avg loss: 0.82887843, Global Avg Loss: 2.41500438, Time: 0.0119 Steps: 31900, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002104, Sample Num: 33664, Cur Loss: 0.52434456, Cur Avg Loss: 0.79969963, Log Avg loss: 0.84548348, Global Avg Loss: 2.41451252, Time: 0.0108 Steps: 31910, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002114, Sample Num: 33824, Cur Loss: 0.23500052, Cur Avg Loss: 0.79940978, Log Avg loss: 0.73842454, Global Avg Loss: 2.41398743, Time: 0.0109 Steps: 31920, Updated lr: 0.000071 Training, Epoch: 0015, Batch: 002124, Sample Num: 33984, Cur Loss: 0.26623100, Cur Avg Loss: 0.79899918, Log Avg loss: 0.71219859, Global Avg Loss: 2.41345445, Time: 0.0151 Steps: 31930, Updated lr: 0.000071 ***** Running evaluation checkpoint-31935 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-31935 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.178018, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.003394, "eval_total_loss": 705.385938, "eval_mae": 0.846443, "eval_mse": 1.003481, "eval_r2": 0.362121, "eval_sp_statistic": 0.635847, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.682042, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.772113, "test_total_loss": 889.60052, "test_mae": 1.208369, "test_mse": 1.771915, "test_r2": -0.14361, "test_sp_statistic": 0.417384, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.540457, "test_ps_pvalue": 0.0, "lr": 7.066382171645329e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.4131593351702803, "train_cur_epoch_loss": 1699.7169747576118, "train_cur_epoch_avg_loss": 0.7983640088105269, "train_cur_epoch_time": 22.17801833152771, "train_cur_epoch_avg_time": 0.010417105839139365, "epoch": 15, "step": 31935} ################################################## Training, Epoch: 0016, Batch: 000005, Sample Num: 80, Cur Loss: 1.00799310, Cur Avg Loss: 0.87635758, Log Avg loss: 0.70245070, Global Avg Loss: 2.41291876, Time: 0.0235 Steps: 31940, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000015, Sample Num: 240, Cur Loss: 0.86821294, Cur Avg Loss: 0.84651020, Log Avg loss: 0.83158651, Global Avg Loss: 2.41242382, Time: 0.0065 Steps: 31950, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000025, Sample Num: 400, Cur Loss: 1.90237856, Cur Avg Loss: 0.78036920, Log Avg loss: 0.68115771, Global Avg Loss: 2.41188212, Time: 0.0196 Steps: 31960, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000035, Sample Num: 560, Cur Loss: 0.54335207, Cur Avg Loss: 0.81065257, Log Avg loss: 0.88636100, Global Avg Loss: 2.41140495, Time: 0.0067 Steps: 31970, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000045, Sample Num: 720, Cur Loss: 0.60773373, Cur Avg Loss: 0.80748298, Log Avg loss: 0.79638938, Global Avg Loss: 2.41089994, Time: 0.0064 Steps: 31980, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000055, Sample Num: 880, Cur Loss: 1.28659534, Cur Avg Loss: 0.80875121, Log Avg loss: 0.81445829, Global Avg Loss: 2.41040090, Time: 0.0064 Steps: 31990, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000065, Sample Num: 1040, Cur Loss: 0.27215391, Cur Avg Loss: 0.79019232, Log Avg loss: 0.68811842, Global Avg Loss: 2.40986268, Time: 0.0172 Steps: 32000, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000075, Sample Num: 1200, Cur Loss: 0.52094263, Cur Avg Loss: 0.75902686, Log Avg loss: 0.55645133, Global Avg Loss: 2.40928367, Time: 0.0066 Steps: 32010, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000085, Sample Num: 1360, Cur Loss: 0.52354103, Cur Avg Loss: 0.75272557, Log Avg loss: 0.70546591, Global Avg Loss: 2.40875156, Time: 0.0225 Steps: 32020, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000095, Sample Num: 1520, Cur Loss: 0.74551690, Cur Avg Loss: 0.75209204, Log Avg loss: 0.74670703, Global Avg Loss: 2.40823266, Time: 0.0192 Steps: 32030, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000105, Sample Num: 1680, Cur Loss: 1.10431862, Cur Avg Loss: 0.75588882, Log Avg loss: 0.79195824, Global Avg Loss: 2.40772821, Time: 0.0069 Steps: 32040, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000115, Sample Num: 1840, Cur Loss: 0.48497805, Cur Avg Loss: 0.74601996, Log Avg loss: 0.64239691, Global Avg Loss: 2.40717740, Time: 0.0073 Steps: 32050, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000125, Sample Num: 2000, Cur Loss: 0.56477189, Cur Avg Loss: 0.75089007, Log Avg loss: 0.80689629, Global Avg Loss: 2.40667825, Time: 0.0071 Steps: 32060, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000135, Sample Num: 2160, Cur Loss: 1.71579385, Cur Avg Loss: 0.77933311, Log Avg loss: 1.13487114, Global Avg Loss: 2.40628168, Time: 0.0151 Steps: 32070, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000145, Sample Num: 2320, Cur Loss: 1.40170658, Cur Avg Loss: 0.77771433, Log Avg loss: 0.75586081, Global Avg Loss: 2.40576721, Time: 0.0189 Steps: 32080, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000155, Sample Num: 2480, Cur Loss: 0.86389792, Cur Avg Loss: 0.78969508, Log Avg loss: 0.96341593, Global Avg Loss: 2.40531773, Time: 0.0067 Steps: 32090, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000165, Sample Num: 2640, Cur Loss: 0.97702736, Cur Avg Loss: 0.79360902, Log Avg loss: 0.85427516, Global Avg Loss: 2.40483454, Time: 0.0089 Steps: 32100, Updated lr: 0.000071 Training, Epoch: 0016, Batch: 000175, Sample Num: 2800, Cur Loss: 0.19329855, Cur Avg Loss: 0.79522669, Log Avg loss: 0.82191816, Global Avg Loss: 2.40434158, Time: 0.0123 Steps: 32110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000185, Sample Num: 2960, Cur Loss: 1.50449133, Cur Avg Loss: 0.79940809, Log Avg loss: 0.87258271, Global Avg Loss: 2.40386469, Time: 0.0065 Steps: 32120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000195, Sample Num: 3120, Cur Loss: 0.91281676, Cur Avg Loss: 0.79673480, Log Avg loss: 0.74727892, Global Avg Loss: 2.40334910, Time: 0.0135 Steps: 32130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000205, Sample Num: 3280, Cur Loss: 0.62868094, Cur Avg Loss: 0.78857768, Log Avg loss: 0.62951372, Global Avg Loss: 2.40279719, Time: 0.0113 Steps: 32140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000215, Sample Num: 3440, Cur Loss: 0.81744432, Cur Avg Loss: 0.78898288, Log Avg loss: 0.79728950, Global Avg Loss: 2.40229781, Time: 0.0121 Steps: 32150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000225, Sample Num: 3600, Cur Loss: 0.78309131, Cur Avg Loss: 0.79004495, Log Avg loss: 0.81287957, Global Avg Loss: 2.40180359, Time: 0.0153 Steps: 32160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000235, Sample Num: 3760, Cur Loss: 1.26172960, Cur Avg Loss: 0.79559287, Log Avg loss: 0.92042110, Global Avg Loss: 2.40134311, Time: 0.0122 Steps: 32170, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000245, Sample Num: 3920, Cur Loss: 0.32049400, Cur Avg Loss: 0.79060834, Log Avg loss: 0.67347174, Global Avg Loss: 2.40080617, Time: 0.0115 Steps: 32180, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000255, Sample Num: 4080, Cur Loss: 0.62121177, Cur Avg Loss: 0.78818559, Log Avg loss: 0.72882821, Global Avg Loss: 2.40028676, Time: 0.0090 Steps: 32190, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000265, Sample Num: 4240, Cur Loss: 1.12062097, Cur Avg Loss: 0.79020830, Log Avg loss: 0.84178740, Global Avg Loss: 2.39980275, Time: 0.0225 Steps: 32200, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000275, Sample Num: 4400, Cur Loss: 1.10251474, Cur Avg Loss: 0.78882719, Log Avg loss: 0.75222780, Global Avg Loss: 2.39929124, Time: 0.0116 Steps: 32210, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000285, Sample Num: 4560, Cur Loss: 0.64256251, Cur Avg Loss: 0.78702115, Log Avg loss: 0.73735519, Global Avg Loss: 2.39877543, Time: 0.0067 Steps: 32220, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000295, Sample Num: 4720, Cur Loss: 0.55552250, Cur Avg Loss: 0.78558727, Log Avg loss: 0.74472166, Global Avg Loss: 2.39826223, Time: 0.0131 Steps: 32230, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000305, Sample Num: 4880, Cur Loss: 0.38356459, Cur Avg Loss: 0.78466021, Log Avg loss: 0.75731184, Global Avg Loss: 2.39775325, Time: 0.0116 Steps: 32240, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000315, Sample Num: 5040, Cur Loss: 0.42737848, Cur Avg Loss: 0.79373705, Log Avg loss: 1.07058075, Global Avg Loss: 2.39734172, Time: 0.0121 Steps: 32250, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000325, Sample Num: 5200, Cur Loss: 0.28497201, Cur Avg Loss: 0.79560309, Log Avg loss: 0.85438330, Global Avg Loss: 2.39686343, Time: 0.0106 Steps: 32260, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000335, Sample Num: 5360, Cur Loss: 0.42594165, Cur Avg Loss: 0.79188174, Log Avg loss: 0.67093791, Global Avg Loss: 2.39632859, Time: 0.0112 Steps: 32270, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000345, Sample Num: 5520, Cur Loss: 1.33637285, Cur Avg Loss: 0.79067947, Log Avg loss: 0.75040351, Global Avg Loss: 2.39581870, Time: 0.0142 Steps: 32280, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000355, Sample Num: 5680, Cur Loss: 0.22879863, Cur Avg Loss: 0.78675912, Log Avg loss: 0.65150705, Global Avg Loss: 2.39527850, Time: 0.0118 Steps: 32290, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000365, Sample Num: 5840, Cur Loss: 0.74643302, Cur Avg Loss: 0.78473947, Log Avg loss: 0.71304178, Global Avg Loss: 2.39475769, Time: 0.0155 Steps: 32300, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000375, Sample Num: 6000, Cur Loss: 0.61374944, Cur Avg Loss: 0.78208725, Log Avg loss: 0.68528118, Global Avg Loss: 2.39422860, Time: 0.0066 Steps: 32310, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000385, Sample Num: 6160, Cur Loss: 0.62741858, Cur Avg Loss: 0.78128919, Log Avg loss: 0.75136183, Global Avg Loss: 2.39372029, Time: 0.0140 Steps: 32320, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000395, Sample Num: 6320, Cur Loss: 0.56470335, Cur Avg Loss: 0.78258422, Log Avg loss: 0.83244319, Global Avg Loss: 2.39323737, Time: 0.0084 Steps: 32330, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000405, Sample Num: 6480, Cur Loss: 0.43293351, Cur Avg Loss: 0.78081870, Log Avg loss: 0.71108034, Global Avg Loss: 2.39271722, Time: 0.0067 Steps: 32340, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000415, Sample Num: 6640, Cur Loss: 0.50572753, Cur Avg Loss: 0.77963058, Log Avg loss: 0.73151169, Global Avg Loss: 2.39220371, Time: 0.0155 Steps: 32350, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000425, Sample Num: 6800, Cur Loss: 1.00243938, Cur Avg Loss: 0.77316993, Log Avg loss: 0.50505309, Global Avg Loss: 2.39162054, Time: 0.0075 Steps: 32360, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000435, Sample Num: 6960, Cur Loss: 0.85592139, Cur Avg Loss: 0.77595325, Log Avg loss: 0.89424448, Global Avg Loss: 2.39115796, Time: 0.0067 Steps: 32370, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000445, Sample Num: 7120, Cur Loss: 0.62725586, Cur Avg Loss: 0.77834509, Log Avg loss: 0.88238983, Global Avg Loss: 2.39069200, Time: 0.0133 Steps: 32380, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000455, Sample Num: 7280, Cur Loss: 1.09307837, Cur Avg Loss: 0.78135993, Log Avg loss: 0.91552050, Global Avg Loss: 2.39023656, Time: 0.0098 Steps: 32390, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000465, Sample Num: 7440, Cur Loss: 0.53534579, Cur Avg Loss: 0.77858623, Log Avg loss: 0.65238305, Global Avg Loss: 2.38970018, Time: 0.0125 Steps: 32400, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000475, Sample Num: 7600, Cur Loss: 1.63589215, Cur Avg Loss: 0.77937129, Log Avg loss: 0.81587645, Global Avg Loss: 2.38921459, Time: 0.0119 Steps: 32410, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000485, Sample Num: 7760, Cur Loss: 0.54934126, Cur Avg Loss: 0.78006322, Log Avg loss: 0.81292979, Global Avg Loss: 2.38872838, Time: 0.0073 Steps: 32420, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000495, Sample Num: 7920, Cur Loss: 1.93502784, Cur Avg Loss: 0.78451771, Log Avg loss: 1.00056030, Global Avg Loss: 2.38830033, Time: 0.0096 Steps: 32430, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000505, Sample Num: 8080, Cur Loss: 0.80155206, Cur Avg Loss: 0.78341835, Log Avg loss: 0.72900005, Global Avg Loss: 2.38778883, Time: 0.0114 Steps: 32440, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000515, Sample Num: 8240, Cur Loss: 1.10430264, Cur Avg Loss: 0.78199428, Log Avg loss: 0.71007905, Global Avg Loss: 2.38727182, Time: 0.0109 Steps: 32450, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000525, Sample Num: 8400, Cur Loss: 0.26681054, Cur Avg Loss: 0.77786759, Log Avg loss: 0.56534286, Global Avg Loss: 2.38671053, Time: 0.0111 Steps: 32460, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000535, Sample Num: 8560, Cur Loss: 0.45410645, Cur Avg Loss: 0.77856283, Log Avg loss: 0.81506284, Global Avg Loss: 2.38622650, Time: 0.0069 Steps: 32470, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000545, Sample Num: 8720, Cur Loss: 0.45266846, Cur Avg Loss: 0.77860329, Log Avg loss: 0.78076805, Global Avg Loss: 2.38573221, Time: 0.0091 Steps: 32480, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000555, Sample Num: 8880, Cur Loss: 1.05937481, Cur Avg Loss: 0.78098568, Log Avg loss: 0.91082626, Global Avg Loss: 2.38527825, Time: 0.0066 Steps: 32490, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000565, Sample Num: 9040, Cur Loss: 0.35914025, Cur Avg Loss: 0.77785786, Log Avg loss: 0.60426339, Global Avg Loss: 2.38473025, Time: 0.0066 Steps: 32500, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000575, Sample Num: 9200, Cur Loss: 0.56377733, Cur Avg Loss: 0.77925357, Log Avg loss: 0.85811140, Global Avg Loss: 2.38426066, Time: 0.0066 Steps: 32510, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000585, Sample Num: 9360, Cur Loss: 0.24888083, Cur Avg Loss: 0.78141811, Log Avg loss: 0.90587922, Global Avg Loss: 2.38380606, Time: 0.0066 Steps: 32520, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000595, Sample Num: 9520, Cur Loss: 0.89306670, Cur Avg Loss: 0.78040485, Log Avg loss: 0.72112896, Global Avg Loss: 2.38329494, Time: 0.0222 Steps: 32530, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000605, Sample Num: 9680, Cur Loss: 0.64500880, Cur Avg Loss: 0.77892776, Log Avg loss: 0.69104095, Global Avg Loss: 2.38277488, Time: 0.0064 Steps: 32540, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000615, Sample Num: 9840, Cur Loss: 0.43647766, Cur Avg Loss: 0.77775587, Log Avg loss: 0.70685628, Global Avg Loss: 2.38226001, Time: 0.0066 Steps: 32550, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000625, Sample Num: 10000, Cur Loss: 1.27051187, Cur Avg Loss: 0.77620291, Log Avg loss: 0.68069614, Global Avg Loss: 2.38173741, Time: 0.0078 Steps: 32560, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000635, Sample Num: 10160, Cur Loss: 0.72058326, Cur Avg Loss: 0.77540186, Log Avg loss: 0.72533625, Global Avg Loss: 2.38122885, Time: 0.0116 Steps: 32570, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000645, Sample Num: 10320, Cur Loss: 0.24906084, Cur Avg Loss: 0.77536895, Log Avg loss: 0.77327908, Global Avg Loss: 2.38073531, Time: 0.0186 Steps: 32580, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000655, Sample Num: 10480, Cur Loss: 0.64519179, Cur Avg Loss: 0.77206201, Log Avg loss: 0.55876466, Global Avg Loss: 2.38017625, Time: 0.0064 Steps: 32590, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000665, Sample Num: 10640, Cur Loss: 0.94315618, Cur Avg Loss: 0.76918071, Log Avg loss: 0.58045555, Global Avg Loss: 2.37962419, Time: 0.0072 Steps: 32600, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000675, Sample Num: 10800, Cur Loss: 0.68482530, Cur Avg Loss: 0.77012722, Log Avg loss: 0.83306987, Global Avg Loss: 2.37914993, Time: 0.0067 Steps: 32610, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000685, Sample Num: 10960, Cur Loss: 0.68359590, Cur Avg Loss: 0.76881280, Log Avg loss: 0.68008929, Global Avg Loss: 2.37862907, Time: 0.0068 Steps: 32620, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000695, Sample Num: 11120, Cur Loss: 0.72249663, Cur Avg Loss: 0.76800968, Log Avg loss: 0.71299655, Global Avg Loss: 2.37811861, Time: 0.0228 Steps: 32630, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000705, Sample Num: 11280, Cur Loss: 0.29665187, Cur Avg Loss: 0.76598877, Log Avg loss: 0.62553499, Global Avg Loss: 2.37758166, Time: 0.0068 Steps: 32640, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000715, Sample Num: 11440, Cur Loss: 0.26568666, Cur Avg Loss: 0.76466772, Log Avg loss: 0.67153398, Global Avg Loss: 2.37705914, Time: 0.0066 Steps: 32650, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000725, Sample Num: 11600, Cur Loss: 0.61116433, Cur Avg Loss: 0.76453190, Log Avg loss: 0.75482047, Global Avg Loss: 2.37656243, Time: 0.0073 Steps: 32660, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000735, Sample Num: 11760, Cur Loss: 1.39526296, Cur Avg Loss: 0.76767256, Log Avg loss: 0.99537079, Global Avg Loss: 2.37613966, Time: 0.0066 Steps: 32670, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000745, Sample Num: 11920, Cur Loss: 1.20477533, Cur Avg Loss: 0.76898523, Log Avg loss: 0.86546609, Global Avg Loss: 2.37567740, Time: 0.0066 Steps: 32680, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000755, Sample Num: 12080, Cur Loss: 0.52244300, Cur Avg Loss: 0.76864208, Log Avg loss: 0.74307737, Global Avg Loss: 2.37517798, Time: 0.0070 Steps: 32690, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000765, Sample Num: 12240, Cur Loss: 0.75823939, Cur Avg Loss: 0.76916562, Log Avg loss: 0.80869326, Global Avg Loss: 2.37469893, Time: 0.0070 Steps: 32700, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000775, Sample Num: 12400, Cur Loss: 0.59449381, Cur Avg Loss: 0.77309392, Log Avg loss: 1.07360846, Global Avg Loss: 2.37430117, Time: 0.0100 Steps: 32710, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000785, Sample Num: 12560, Cur Loss: 1.02093148, Cur Avg Loss: 0.77553848, Log Avg loss: 0.96499253, Global Avg Loss: 2.37387045, Time: 0.0166 Steps: 32720, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000795, Sample Num: 12720, Cur Loss: 0.58344030, Cur Avg Loss: 0.77609863, Log Avg loss: 0.82006987, Global Avg Loss: 2.37339572, Time: 0.0071 Steps: 32730, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000805, Sample Num: 12880, Cur Loss: 0.80458426, Cur Avg Loss: 0.77729641, Log Avg loss: 0.87252022, Global Avg Loss: 2.37293729, Time: 0.0133 Steps: 32740, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000815, Sample Num: 13040, Cur Loss: 1.31512487, Cur Avg Loss: 0.77616502, Log Avg loss: 0.68508769, Global Avg Loss: 2.37242192, Time: 0.0071 Steps: 32750, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000825, Sample Num: 13200, Cur Loss: 0.74118924, Cur Avg Loss: 0.77882972, Log Avg loss: 0.99600332, Global Avg Loss: 2.37200177, Time: 0.0112 Steps: 32760, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000835, Sample Num: 13360, Cur Loss: 0.37481561, Cur Avg Loss: 0.77893216, Log Avg loss: 0.78738326, Global Avg Loss: 2.37151821, Time: 0.0112 Steps: 32770, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000845, Sample Num: 13520, Cur Loss: 0.59345168, Cur Avg Loss: 0.77655492, Log Avg loss: 0.57805527, Global Avg Loss: 2.37097109, Time: 0.0109 Steps: 32780, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000855, Sample Num: 13680, Cur Loss: 0.51299000, Cur Avg Loss: 0.77955868, Log Avg loss: 1.03337655, Global Avg Loss: 2.37056316, Time: 0.0155 Steps: 32790, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000865, Sample Num: 13840, Cur Loss: 1.55477667, Cur Avg Loss: 0.78109735, Log Avg loss: 0.91265321, Global Avg Loss: 2.37011868, Time: 0.0069 Steps: 32800, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000875, Sample Num: 14000, Cur Loss: 0.71016085, Cur Avg Loss: 0.78041342, Log Avg loss: 0.72125386, Global Avg Loss: 2.36961613, Time: 0.0073 Steps: 32810, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000885, Sample Num: 14160, Cur Loss: 1.02265143, Cur Avg Loss: 0.78057449, Log Avg loss: 0.79466821, Global Avg Loss: 2.36913625, Time: 0.0077 Steps: 32820, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000895, Sample Num: 14320, Cur Loss: 0.65962374, Cur Avg Loss: 0.78125818, Log Avg loss: 0.84176450, Global Avg Loss: 2.36867102, Time: 0.0226 Steps: 32830, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000905, Sample Num: 14480, Cur Loss: 0.51539361, Cur Avg Loss: 0.78161402, Log Avg loss: 0.81346222, Global Avg Loss: 2.36819744, Time: 0.0070 Steps: 32840, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000915, Sample Num: 14640, Cur Loss: 0.50018698, Cur Avg Loss: 0.78159771, Log Avg loss: 0.78012106, Global Avg Loss: 2.36771401, Time: 0.0089 Steps: 32850, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000925, Sample Num: 14800, Cur Loss: 1.33641601, Cur Avg Loss: 0.78553217, Log Avg loss: 1.14553590, Global Avg Loss: 2.36734208, Time: 0.0112 Steps: 32860, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000935, Sample Num: 14960, Cur Loss: 1.48549223, Cur Avg Loss: 0.78445213, Log Avg loss: 0.68454793, Global Avg Loss: 2.36683012, Time: 0.0105 Steps: 32870, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000945, Sample Num: 15120, Cur Loss: 0.70302916, Cur Avg Loss: 0.78326765, Log Avg loss: 0.67251906, Global Avg Loss: 2.36631482, Time: 0.0074 Steps: 32880, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000955, Sample Num: 15280, Cur Loss: 0.48280597, Cur Avg Loss: 0.78519785, Log Avg loss: 0.96760164, Global Avg Loss: 2.36588955, Time: 0.0072 Steps: 32890, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000965, Sample Num: 15440, Cur Loss: 1.49195457, Cur Avg Loss: 0.78643039, Log Avg loss: 0.90413792, Global Avg Loss: 2.36544525, Time: 0.0122 Steps: 32900, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000975, Sample Num: 15600, Cur Loss: 1.30398285, Cur Avg Loss: 0.78534979, Log Avg loss: 0.68107228, Global Avg Loss: 2.36493344, Time: 0.0232 Steps: 32910, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000985, Sample Num: 15760, Cur Loss: 1.74093533, Cur Avg Loss: 0.78689166, Log Avg loss: 0.93722344, Global Avg Loss: 2.36449975, Time: 0.0069 Steps: 32920, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 000995, Sample Num: 15920, Cur Loss: 0.27790642, Cur Avg Loss: 0.78522614, Log Avg loss: 0.62117253, Global Avg Loss: 2.36397034, Time: 0.0067 Steps: 32930, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001005, Sample Num: 16080, Cur Loss: 0.52813345, Cur Avg Loss: 0.78433966, Log Avg loss: 0.69613488, Global Avg Loss: 2.36346402, Time: 0.0110 Steps: 32940, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001015, Sample Num: 16240, Cur Loss: 0.54047275, Cur Avg Loss: 0.78405540, Log Avg loss: 0.75548710, Global Avg Loss: 2.36297601, Time: 0.0071 Steps: 32950, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001025, Sample Num: 16400, Cur Loss: 0.41013980, Cur Avg Loss: 0.78378581, Log Avg loss: 0.75642258, Global Avg Loss: 2.36248859, Time: 0.0174 Steps: 32960, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001035, Sample Num: 16560, Cur Loss: 0.79727167, Cur Avg Loss: 0.78286976, Log Avg loss: 0.68897429, Global Avg Loss: 2.36198100, Time: 0.0069 Steps: 32970, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001045, Sample Num: 16720, Cur Loss: 0.94319999, Cur Avg Loss: 0.78146537, Log Avg loss: 0.63611111, Global Avg Loss: 2.36145769, Time: 0.0185 Steps: 32980, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001055, Sample Num: 16880, Cur Loss: 0.83171082, Cur Avg Loss: 0.78284378, Log Avg loss: 0.92688795, Global Avg Loss: 2.36102284, Time: 0.0115 Steps: 32990, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001065, Sample Num: 17040, Cur Loss: 0.71699846, Cur Avg Loss: 0.78166631, Log Avg loss: 0.65744315, Global Avg Loss: 2.36050661, Time: 0.0064 Steps: 33000, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001075, Sample Num: 17200, Cur Loss: 0.45897785, Cur Avg Loss: 0.77869872, Log Avg loss: 0.46264993, Global Avg Loss: 2.35993167, Time: 0.0233 Steps: 33010, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001085, Sample Num: 17360, Cur Loss: 1.42720389, Cur Avg Loss: 0.77803487, Log Avg loss: 0.70667176, Global Avg Loss: 2.35943099, Time: 0.0160 Steps: 33020, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001095, Sample Num: 17520, Cur Loss: 0.53257549, Cur Avg Loss: 0.77778699, Log Avg loss: 0.75089216, Global Avg Loss: 2.35894399, Time: 0.0120 Steps: 33030, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001105, Sample Num: 17680, Cur Loss: 0.77597678, Cur Avg Loss: 0.77819755, Log Avg loss: 0.82315392, Global Avg Loss: 2.35847917, Time: 0.0066 Steps: 33040, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001115, Sample Num: 17840, Cur Loss: 0.42958042, Cur Avg Loss: 0.77867155, Log Avg loss: 0.83104756, Global Avg Loss: 2.35801701, Time: 0.0127 Steps: 33050, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001125, Sample Num: 18000, Cur Loss: 0.66448545, Cur Avg Loss: 0.77851140, Log Avg loss: 0.76065462, Global Avg Loss: 2.35753384, Time: 0.0128 Steps: 33060, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001135, Sample Num: 18160, Cur Loss: 1.15670860, Cur Avg Loss: 0.78112602, Log Avg loss: 1.07527079, Global Avg Loss: 2.35714610, Time: 0.0113 Steps: 33070, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001145, Sample Num: 18320, Cur Loss: 1.51945591, Cur Avg Loss: 0.78012380, Log Avg loss: 0.66637236, Global Avg Loss: 2.35663498, Time: 0.0066 Steps: 33080, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001155, Sample Num: 18480, Cur Loss: 0.52072930, Cur Avg Loss: 0.78014443, Log Avg loss: 0.78250678, Global Avg Loss: 2.35615927, Time: 0.0072 Steps: 33090, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001165, Sample Num: 18640, Cur Loss: 0.95419830, Cur Avg Loss: 0.77996966, Log Avg loss: 0.75978339, Global Avg Loss: 2.35567698, Time: 0.0072 Steps: 33100, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001175, Sample Num: 18800, Cur Loss: 0.43919939, Cur Avg Loss: 0.77897134, Log Avg loss: 0.66266681, Global Avg Loss: 2.35516565, Time: 0.0090 Steps: 33110, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001185, Sample Num: 18960, Cur Loss: 0.34480083, Cur Avg Loss: 0.77939686, Log Avg loss: 0.82939612, Global Avg Loss: 2.35470497, Time: 0.0119 Steps: 33120, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001195, Sample Num: 19120, Cur Loss: 0.77066576, Cur Avg Loss: 0.78146248, Log Avg loss: 1.02623809, Global Avg Loss: 2.35430399, Time: 0.0064 Steps: 33130, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001205, Sample Num: 19280, Cur Loss: 1.48201227, Cur Avg Loss: 0.78175719, Log Avg loss: 0.81697485, Global Avg Loss: 2.35384010, Time: 0.0064 Steps: 33140, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001215, Sample Num: 19440, Cur Loss: 0.61869729, Cur Avg Loss: 0.78277069, Log Avg loss: 0.90489759, Global Avg Loss: 2.35340301, Time: 0.0099 Steps: 33150, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001225, Sample Num: 19600, Cur Loss: 0.82453859, Cur Avg Loss: 0.78329210, Log Avg loss: 0.84664333, Global Avg Loss: 2.35294862, Time: 0.0072 Steps: 33160, Updated lr: 0.000070 Training, Epoch: 0016, Batch: 001235, Sample Num: 19760, Cur Loss: 0.82268405, Cur Avg Loss: 0.78059973, Log Avg loss: 0.45078463, Global Avg Loss: 2.35237516, Time: 0.0162 Steps: 33170, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001245, Sample Num: 19920, Cur Loss: 1.31967556, Cur Avg Loss: 0.78051561, Log Avg loss: 0.77012648, Global Avg Loss: 2.35189829, Time: 0.0076 Steps: 33180, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001255, Sample Num: 20080, Cur Loss: 0.69027948, Cur Avg Loss: 0.77969710, Log Avg loss: 0.67779284, Global Avg Loss: 2.35139389, Time: 0.0233 Steps: 33190, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001265, Sample Num: 20240, Cur Loss: 0.81257606, Cur Avg Loss: 0.78093160, Log Avg loss: 0.93586190, Global Avg Loss: 2.35096753, Time: 0.0068 Steps: 33200, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001275, Sample Num: 20400, Cur Loss: 1.21080852, Cur Avg Loss: 0.78159258, Log Avg loss: 0.86520658, Global Avg Loss: 2.35052014, Time: 0.0063 Steps: 33210, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001285, Sample Num: 20560, Cur Loss: 0.66651976, Cur Avg Loss: 0.78318300, Log Avg loss: 0.98596139, Global Avg Loss: 2.35010938, Time: 0.0117 Steps: 33220, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001295, Sample Num: 20720, Cur Loss: 1.23442650, Cur Avg Loss: 0.78445435, Log Avg loss: 0.94782255, Global Avg Loss: 2.34968738, Time: 0.0066 Steps: 33230, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001305, Sample Num: 20880, Cur Loss: 0.50724792, Cur Avg Loss: 0.78342487, Log Avg loss: 0.65010725, Global Avg Loss: 2.34917608, Time: 0.0122 Steps: 33240, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001315, Sample Num: 21040, Cur Loss: 1.10117400, Cur Avg Loss: 0.78356470, Log Avg loss: 0.80181236, Global Avg Loss: 2.34871071, Time: 0.0071 Steps: 33250, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001325, Sample Num: 21200, Cur Loss: 0.39588600, Cur Avg Loss: 0.78204875, Log Avg loss: 0.58270157, Global Avg Loss: 2.34817973, Time: 0.0067 Steps: 33260, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001335, Sample Num: 21360, Cur Loss: 0.99311364, Cur Avg Loss: 0.78155756, Log Avg loss: 0.71647452, Global Avg Loss: 2.34768929, Time: 0.0135 Steps: 33270, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001345, Sample Num: 21520, Cur Loss: 0.57546514, Cur Avg Loss: 0.78127255, Log Avg loss: 0.74322335, Global Avg Loss: 2.34720718, Time: 0.0137 Steps: 33280, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001355, Sample Num: 21680, Cur Loss: 0.20563552, Cur Avg Loss: 0.78315318, Log Avg loss: 1.03609879, Global Avg Loss: 2.34681334, Time: 0.0072 Steps: 33290, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001365, Sample Num: 21840, Cur Loss: 0.59136611, Cur Avg Loss: 0.78311009, Log Avg loss: 0.77727058, Global Avg Loss: 2.34634200, Time: 0.0135 Steps: 33300, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001375, Sample Num: 22000, Cur Loss: 0.89780998, Cur Avg Loss: 0.78292006, Log Avg loss: 0.75698073, Global Avg Loss: 2.34586486, Time: 0.0072 Steps: 33310, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001385, Sample Num: 22160, Cur Loss: 0.68215507, Cur Avg Loss: 0.78235539, Log Avg loss: 0.70471410, Global Avg Loss: 2.34537232, Time: 0.0086 Steps: 33320, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001395, Sample Num: 22320, Cur Loss: 1.07874608, Cur Avg Loss: 0.78364589, Log Avg loss: 0.96238019, Global Avg Loss: 2.34495738, Time: 0.0067 Steps: 33330, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001405, Sample Num: 22480, Cur Loss: 0.54963017, Cur Avg Loss: 0.78375840, Log Avg loss: 0.79945390, Global Avg Loss: 2.34449382, Time: 0.0121 Steps: 33340, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001415, Sample Num: 22640, Cur Loss: 1.16052318, Cur Avg Loss: 0.78250593, Log Avg loss: 0.60653346, Global Avg Loss: 2.34397269, Time: 0.0121 Steps: 33350, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001425, Sample Num: 22800, Cur Loss: 1.19696116, Cur Avg Loss: 0.78371650, Log Avg loss: 0.95501240, Global Avg Loss: 2.34355634, Time: 0.0077 Steps: 33360, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001435, Sample Num: 22960, Cur Loss: 0.27771267, Cur Avg Loss: 0.78419563, Log Avg loss: 0.85247113, Global Avg Loss: 2.34310950, Time: 0.0067 Steps: 33370, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001445, Sample Num: 23120, Cur Loss: 0.55922872, Cur Avg Loss: 0.78460095, Log Avg loss: 0.84276513, Global Avg Loss: 2.34266003, Time: 0.0094 Steps: 33380, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001455, Sample Num: 23280, Cur Loss: 0.41465664, Cur Avg Loss: 0.78415011, Log Avg loss: 0.71900287, Global Avg Loss: 2.34217376, Time: 0.0143 Steps: 33390, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001465, Sample Num: 23440, Cur Loss: 0.78719819, Cur Avg Loss: 0.78381756, Log Avg loss: 0.73543130, Global Avg Loss: 2.34169270, Time: 0.0117 Steps: 33400, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001475, Sample Num: 23600, Cur Loss: 0.45782483, Cur Avg Loss: 0.78370621, Log Avg loss: 0.76739381, Global Avg Loss: 2.34122149, Time: 0.0113 Steps: 33410, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001485, Sample Num: 23760, Cur Loss: 0.49861324, Cur Avg Loss: 0.78474675, Log Avg loss: 0.93822590, Global Avg Loss: 2.34080168, Time: 0.0119 Steps: 33420, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001495, Sample Num: 23920, Cur Loss: 0.41678542, Cur Avg Loss: 0.78456781, Log Avg loss: 0.75799509, Global Avg Loss: 2.34032822, Time: 0.0125 Steps: 33430, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001505, Sample Num: 24080, Cur Loss: 0.96652812, Cur Avg Loss: 0.78412987, Log Avg loss: 0.71865832, Global Avg Loss: 2.33984327, Time: 0.0111 Steps: 33440, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001515, Sample Num: 24240, Cur Loss: 0.81155324, Cur Avg Loss: 0.78359142, Log Avg loss: 0.70255556, Global Avg Loss: 2.33935379, Time: 0.0123 Steps: 33450, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001525, Sample Num: 24400, Cur Loss: 0.79850447, Cur Avg Loss: 0.78229641, Log Avg loss: 0.58610122, Global Avg Loss: 2.33882981, Time: 0.0068 Steps: 33460, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001535, Sample Num: 24560, Cur Loss: 0.87213194, Cur Avg Loss: 0.78212547, Log Avg loss: 0.75605760, Global Avg Loss: 2.33835692, Time: 0.0068 Steps: 33470, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001545, Sample Num: 24720, Cur Loss: 0.48379499, Cur Avg Loss: 0.78310998, Log Avg loss: 0.93423292, Global Avg Loss: 2.33793752, Time: 0.0115 Steps: 33480, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001555, Sample Num: 24880, Cur Loss: 0.60697877, Cur Avg Loss: 0.78300622, Log Avg loss: 0.76697530, Global Avg Loss: 2.33746844, Time: 0.0083 Steps: 33490, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001565, Sample Num: 25040, Cur Loss: 0.63138700, Cur Avg Loss: 0.78327481, Log Avg loss: 0.82503967, Global Avg Loss: 2.33701697, Time: 0.0072 Steps: 33500, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001575, Sample Num: 25200, Cur Loss: 0.30718684, Cur Avg Loss: 0.78327972, Log Avg loss: 0.78404875, Global Avg Loss: 2.33655353, Time: 0.0098 Steps: 33510, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001585, Sample Num: 25360, Cur Loss: 0.72677869, Cur Avg Loss: 0.78229293, Log Avg loss: 0.62687282, Global Avg Loss: 2.33604349, Time: 0.0133 Steps: 33520, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001595, Sample Num: 25520, Cur Loss: 0.50123870, Cur Avg Loss: 0.78139896, Log Avg loss: 0.63970536, Global Avg Loss: 2.33553757, Time: 0.0140 Steps: 33530, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001605, Sample Num: 25680, Cur Loss: 0.94414556, Cur Avg Loss: 0.78332917, Log Avg loss: 1.09119753, Global Avg Loss: 2.33516657, Time: 0.0068 Steps: 33540, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001615, Sample Num: 25840, Cur Loss: 0.91928947, Cur Avg Loss: 0.78363342, Log Avg loss: 0.83246584, Global Avg Loss: 2.33471867, Time: 0.0109 Steps: 33550, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001625, Sample Num: 26000, Cur Loss: 0.87228763, Cur Avg Loss: 0.78441654, Log Avg loss: 0.91088948, Global Avg Loss: 2.33429441, Time: 0.0157 Steps: 33560, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001635, Sample Num: 26160, Cur Loss: 0.74497640, Cur Avg Loss: 0.78297633, Log Avg loss: 0.54894328, Global Avg Loss: 2.33376258, Time: 0.0113 Steps: 33570, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001645, Sample Num: 26320, Cur Loss: 1.05867147, Cur Avg Loss: 0.78234204, Log Avg loss: 0.67863546, Global Avg Loss: 2.33326969, Time: 0.0119 Steps: 33580, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001655, Sample Num: 26480, Cur Loss: 0.48663259, Cur Avg Loss: 0.78247617, Log Avg loss: 0.80453951, Global Avg Loss: 2.33281457, Time: 0.0068 Steps: 33590, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001665, Sample Num: 26640, Cur Loss: 1.04255164, Cur Avg Loss: 0.78216403, Log Avg loss: 0.73050575, Global Avg Loss: 2.33233769, Time: 0.0068 Steps: 33600, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001675, Sample Num: 26800, Cur Loss: 0.67940176, Cur Avg Loss: 0.78218845, Log Avg loss: 0.78625436, Global Avg Loss: 2.33187769, Time: 0.0109 Steps: 33610, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001685, Sample Num: 26960, Cur Loss: 1.01672602, Cur Avg Loss: 0.78187083, Log Avg loss: 0.72866931, Global Avg Loss: 2.33140082, Time: 0.0123 Steps: 33620, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001695, Sample Num: 27120, Cur Loss: 1.51844287, Cur Avg Loss: 0.78092373, Log Avg loss: 0.62133789, Global Avg Loss: 2.33089233, Time: 0.0123 Steps: 33630, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001705, Sample Num: 27280, Cur Loss: 0.77979410, Cur Avg Loss: 0.78166610, Log Avg loss: 0.90749694, Global Avg Loss: 2.33046921, Time: 0.0072 Steps: 33640, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001715, Sample Num: 27440, Cur Loss: 0.49127874, Cur Avg Loss: 0.78258483, Log Avg loss: 0.93922840, Global Avg Loss: 2.33005576, Time: 0.0120 Steps: 33650, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001725, Sample Num: 27600, Cur Loss: 0.38953745, Cur Avg Loss: 0.78201523, Log Avg loss: 0.68432882, Global Avg Loss: 2.32956683, Time: 0.0067 Steps: 33660, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001735, Sample Num: 27760, Cur Loss: 0.35738832, Cur Avg Loss: 0.78119004, Log Avg loss: 0.63884434, Global Avg Loss: 2.32906469, Time: 0.0114 Steps: 33670, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001745, Sample Num: 27920, Cur Loss: 0.82183695, Cur Avg Loss: 0.78190736, Log Avg loss: 0.90636340, Global Avg Loss: 2.32864227, Time: 0.0113 Steps: 33680, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001755, Sample Num: 28080, Cur Loss: 0.58661389, Cur Avg Loss: 0.78098018, Log Avg loss: 0.61918739, Global Avg Loss: 2.32813486, Time: 0.0068 Steps: 33690, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001765, Sample Num: 28240, Cur Loss: 0.88849473, Cur Avg Loss: 0.78110595, Log Avg loss: 0.80317743, Global Avg Loss: 2.32768236, Time: 0.0072 Steps: 33700, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001775, Sample Num: 28400, Cur Loss: 0.68596119, Cur Avg Loss: 0.78274508, Log Avg loss: 1.07205215, Global Avg Loss: 2.32730987, Time: 0.0210 Steps: 33710, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001785, Sample Num: 28560, Cur Loss: 1.71800613, Cur Avg Loss: 0.78380927, Log Avg loss: 0.97270244, Global Avg Loss: 2.32690815, Time: 0.0218 Steps: 33720, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001795, Sample Num: 28720, Cur Loss: 1.10923040, Cur Avg Loss: 0.78379111, Log Avg loss: 0.78054954, Global Avg Loss: 2.32644970, Time: 0.0243 Steps: 33730, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001805, Sample Num: 28880, Cur Loss: 0.51566637, Cur Avg Loss: 0.78365308, Log Avg loss: 0.75887769, Global Avg Loss: 2.32598510, Time: 0.0073 Steps: 33740, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001815, Sample Num: 29040, Cur Loss: 1.17788208, Cur Avg Loss: 0.78410256, Log Avg loss: 0.86523284, Global Avg Loss: 2.32555228, Time: 0.0089 Steps: 33750, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001825, Sample Num: 29200, Cur Loss: 0.69187129, Cur Avg Loss: 0.78308300, Log Avg loss: 0.59803374, Global Avg Loss: 2.32504058, Time: 0.0121 Steps: 33760, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001835, Sample Num: 29360, Cur Loss: 0.66951501, Cur Avg Loss: 0.78301636, Log Avg loss: 0.77085432, Global Avg Loss: 2.32458035, Time: 0.0112 Steps: 33770, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001845, Sample Num: 29520, Cur Loss: 1.81919003, Cur Avg Loss: 0.78319102, Log Avg loss: 0.81524162, Global Avg Loss: 2.32413353, Time: 0.0113 Steps: 33780, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001855, Sample Num: 29680, Cur Loss: 0.68248165, Cur Avg Loss: 0.78199098, Log Avg loss: 0.56058296, Global Avg Loss: 2.32361162, Time: 0.0116 Steps: 33790, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001865, Sample Num: 29840, Cur Loss: 0.54151809, Cur Avg Loss: 0.78206179, Log Avg loss: 0.79519613, Global Avg Loss: 2.32315943, Time: 0.0072 Steps: 33800, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001875, Sample Num: 30000, Cur Loss: 0.54170889, Cur Avg Loss: 0.78145716, Log Avg loss: 0.66869546, Global Avg Loss: 2.32267008, Time: 0.0066 Steps: 33810, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001885, Sample Num: 30160, Cur Loss: 1.24245512, Cur Avg Loss: 0.78095952, Log Avg loss: 0.68765066, Global Avg Loss: 2.32218664, Time: 0.0090 Steps: 33820, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001895, Sample Num: 30320, Cur Loss: 0.11066784, Cur Avg Loss: 0.78015547, Log Avg loss: 0.62859238, Global Avg Loss: 2.32168602, Time: 0.0119 Steps: 33830, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001905, Sample Num: 30480, Cur Loss: 0.54917806, Cur Avg Loss: 0.78056876, Log Avg loss: 0.85888746, Global Avg Loss: 2.32125375, Time: 0.0066 Steps: 33840, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001915, Sample Num: 30640, Cur Loss: 1.50280261, Cur Avg Loss: 0.78076276, Log Avg loss: 0.81772053, Global Avg Loss: 2.32080957, Time: 0.0069 Steps: 33850, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001925, Sample Num: 30800, Cur Loss: 0.91767156, Cur Avg Loss: 0.78065010, Log Avg loss: 0.75907387, Global Avg Loss: 2.32034834, Time: 0.0066 Steps: 33860, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001935, Sample Num: 30960, Cur Loss: 0.99282825, Cur Avg Loss: 0.78093976, Log Avg loss: 0.83670061, Global Avg Loss: 2.31991030, Time: 0.0108 Steps: 33870, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001945, Sample Num: 31120, Cur Loss: 1.16263187, Cur Avg Loss: 0.78122257, Log Avg loss: 0.83594681, Global Avg Loss: 2.31947229, Time: 0.0067 Steps: 33880, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001955, Sample Num: 31280, Cur Loss: 1.04934764, Cur Avg Loss: 0.78039664, Log Avg loss: 0.61975264, Global Avg Loss: 2.31897075, Time: 0.0129 Steps: 33890, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001965, Sample Num: 31440, Cur Loss: 1.35705495, Cur Avg Loss: 0.78032312, Log Avg loss: 0.76594908, Global Avg Loss: 2.31851263, Time: 0.0087 Steps: 33900, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001975, Sample Num: 31600, Cur Loss: 0.27407086, Cur Avg Loss: 0.77987422, Log Avg loss: 0.69166515, Global Avg Loss: 2.31803288, Time: 0.0098 Steps: 33910, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001985, Sample Num: 31760, Cur Loss: 0.84566540, Cur Avg Loss: 0.78028692, Log Avg loss: 0.86179654, Global Avg Loss: 2.31760356, Time: 0.0067 Steps: 33920, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 001995, Sample Num: 31920, Cur Loss: 1.07152939, Cur Avg Loss: 0.78200760, Log Avg loss: 1.12356279, Global Avg Loss: 2.31725165, Time: 0.0068 Steps: 33930, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002005, Sample Num: 32080, Cur Loss: 0.39753377, Cur Avg Loss: 0.78111740, Log Avg loss: 0.60352217, Global Avg Loss: 2.31674672, Time: 0.0200 Steps: 33940, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002015, Sample Num: 32240, Cur Loss: 1.29208028, Cur Avg Loss: 0.78084125, Log Avg loss: 0.72547230, Global Avg Loss: 2.31627801, Time: 0.0119 Steps: 33950, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002025, Sample Num: 32400, Cur Loss: 0.94977486, Cur Avg Loss: 0.78143776, Log Avg loss: 0.90163504, Global Avg Loss: 2.31586145, Time: 0.0132 Steps: 33960, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002035, Sample Num: 32560, Cur Loss: 0.37281373, Cur Avg Loss: 0.78097849, Log Avg loss: 0.68797545, Global Avg Loss: 2.31538224, Time: 0.0081 Steps: 33970, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002045, Sample Num: 32720, Cur Loss: 0.89499867, Cur Avg Loss: 0.77988625, Log Avg loss: 0.55761556, Global Avg Loss: 2.31486494, Time: 0.0116 Steps: 33980, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002055, Sample Num: 32880, Cur Loss: 0.47342065, Cur Avg Loss: 0.77958403, Log Avg loss: 0.71778057, Global Avg Loss: 2.31439507, Time: 0.0071 Steps: 33990, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002065, Sample Num: 33040, Cur Loss: 0.38826776, Cur Avg Loss: 0.77842684, Log Avg loss: 0.54062511, Global Avg Loss: 2.31387338, Time: 0.0120 Steps: 34000, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002075, Sample Num: 33200, Cur Loss: 0.41348618, Cur Avg Loss: 0.77808091, Log Avg loss: 0.70664595, Global Avg Loss: 2.31340080, Time: 0.0066 Steps: 34010, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002085, Sample Num: 33360, Cur Loss: 0.87392253, Cur Avg Loss: 0.77776350, Log Avg loss: 0.71189997, Global Avg Loss: 2.31293005, Time: 0.0064 Steps: 34020, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002095, Sample Num: 33520, Cur Loss: 0.96051061, Cur Avg Loss: 0.77741969, Log Avg loss: 0.70573611, Global Avg Loss: 2.31245776, Time: 0.0111 Steps: 34030, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002105, Sample Num: 33680, Cur Loss: 1.41741431, Cur Avg Loss: 0.77825333, Log Avg loss: 0.95290059, Global Avg Loss: 2.31205836, Time: 0.0070 Steps: 34040, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002115, Sample Num: 33840, Cur Loss: 0.32766041, Cur Avg Loss: 0.77797549, Log Avg loss: 0.71949119, Global Avg Loss: 2.31159065, Time: 0.0071 Steps: 34050, Updated lr: 0.000069 Training, Epoch: 0016, Batch: 002125, Sample Num: 34000, Cur Loss: 1.06543314, Cur Avg Loss: 0.77679043, Log Avg loss: 0.52614871, Global Avg Loss: 2.31106644, Time: 0.0092 Steps: 34060, Updated lr: 0.000069 ***** Running evaluation checkpoint-34064 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-34064 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.465372, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.872028, "eval_total_loss": 613.03559, "eval_mae": 0.761349, "eval_mse": 0.872146, "eval_r2": 0.445607, "eval_sp_statistic": 0.644619, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.694695, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.341742, "test_total_loss": 673.554337, "test_mae": 1.020863, "test_mse": 1.341673, "test_r2": 0.134072, "test_sp_statistic": 0.443243, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.563602, "test_ps_pvalue": 0.0, "lr": 6.86448553816975e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.310892366526186, "train_cur_epoch_loss": 1653.994204685092, "train_cur_epoch_avg_loss": 0.7768878368647685, "train_cur_epoch_time": 22.465372323989868, "train_cur_epoch_avg_time": 0.010552077183649538, "epoch": 16, "step": 34064} ################################################## Training, Epoch: 0017, Batch: 000006, Sample Num: 96, Cur Loss: 0.85961258, Cur Avg Loss: 0.79464048, Log Avg loss: 0.80823907, Global Avg Loss: 2.31062534, Time: 0.0118 Steps: 34070, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000016, Sample Num: 256, Cur Loss: 0.97258419, Cur Avg Loss: 0.67033570, Log Avg loss: 0.59575282, Global Avg Loss: 2.31012215, Time: 0.0091 Steps: 34080, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000026, Sample Num: 416, Cur Loss: 0.29207921, Cur Avg Loss: 0.80724002, Log Avg loss: 1.02628694, Global Avg Loss: 2.30974555, Time: 0.0067 Steps: 34090, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000036, Sample Num: 576, Cur Loss: 0.83028448, Cur Avg Loss: 0.77943370, Log Avg loss: 0.70713725, Global Avg Loss: 2.30927558, Time: 0.0119 Steps: 34100, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000046, Sample Num: 736, Cur Loss: 0.39444500, Cur Avg Loss: 0.73446020, Log Avg loss: 0.57255560, Global Avg Loss: 2.30876642, Time: 0.0125 Steps: 34110, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000056, Sample Num: 896, Cur Loss: 2.01874256, Cur Avg Loss: 0.80239137, Log Avg loss: 1.11487478, Global Avg Loss: 2.30841651, Time: 0.0081 Steps: 34120, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000066, Sample Num: 1056, Cur Loss: 0.67001414, Cur Avg Loss: 0.80292779, Log Avg loss: 0.80593175, Global Avg Loss: 2.30797629, Time: 0.0069 Steps: 34130, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000076, Sample Num: 1216, Cur Loss: 1.05735457, Cur Avg Loss: 0.78007917, Log Avg loss: 0.62927824, Global Avg Loss: 2.30748458, Time: 0.0068 Steps: 34140, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000086, Sample Num: 1376, Cur Loss: 0.91385627, Cur Avg Loss: 0.78304188, Log Avg loss: 0.80555845, Global Avg Loss: 2.30704478, Time: 0.0092 Steps: 34150, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000096, Sample Num: 1536, Cur Loss: 0.85931814, Cur Avg Loss: 0.77373479, Log Avg loss: 0.69369388, Global Avg Loss: 2.30657249, Time: 0.0069 Steps: 34160, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000106, Sample Num: 1696, Cur Loss: 0.38300607, Cur Avg Loss: 0.75552673, Log Avg loss: 0.58072931, Global Avg Loss: 2.30606741, Time: 0.0114 Steps: 34170, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000116, Sample Num: 1856, Cur Loss: 0.39080548, Cur Avg Loss: 0.74521094, Log Avg loss: 0.63586359, Global Avg Loss: 2.30557876, Time: 0.0083 Steps: 34180, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000126, Sample Num: 2016, Cur Loss: 0.43443888, Cur Avg Loss: 0.75918457, Log Avg loss: 0.92127872, Global Avg Loss: 2.30517388, Time: 0.0121 Steps: 34190, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000136, Sample Num: 2176, Cur Loss: 0.48014984, Cur Avg Loss: 0.76114741, Log Avg loss: 0.78587919, Global Avg Loss: 2.30472964, Time: 0.0070 Steps: 34200, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000146, Sample Num: 2336, Cur Loss: 0.30688828, Cur Avg Loss: 0.73664203, Log Avg loss: 0.40336881, Global Avg Loss: 2.30417385, Time: 0.0150 Steps: 34210, Updated lr: 0.000069 Training, Epoch: 0017, Batch: 000156, Sample Num: 2496, Cur Loss: 0.18123448, Cur Avg Loss: 0.72130459, Log Avg loss: 0.49737793, Global Avg Loss: 2.30364585, Time: 0.0065 Steps: 34220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000166, Sample Num: 2656, Cur Loss: 0.44492045, Cur Avg Loss: 0.71770164, Log Avg loss: 0.66149566, Global Avg Loss: 2.30316611, Time: 0.0067 Steps: 34230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000176, Sample Num: 2816, Cur Loss: 0.63319331, Cur Avg Loss: 0.72140625, Log Avg loss: 0.78290286, Global Avg Loss: 2.30272211, Time: 0.0120 Steps: 34240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000186, Sample Num: 2976, Cur Loss: 1.99610829, Cur Avg Loss: 0.72427994, Log Avg loss: 0.77485671, Global Avg Loss: 2.30227602, Time: 0.0067 Steps: 34250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000196, Sample Num: 3136, Cur Loss: 0.42169636, Cur Avg Loss: 0.72516149, Log Avg loss: 0.74155835, Global Avg Loss: 2.30182047, Time: 0.0129 Steps: 34260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000206, Sample Num: 3296, Cur Loss: 0.28926516, Cur Avg Loss: 0.73271308, Log Avg loss: 0.88072430, Global Avg Loss: 2.30140579, Time: 0.0106 Steps: 34270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000216, Sample Num: 3456, Cur Loss: 0.97645330, Cur Avg Loss: 0.73594944, Log Avg loss: 0.80261838, Global Avg Loss: 2.30096857, Time: 0.0086 Steps: 34280, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000226, Sample Num: 3616, Cur Loss: 1.61272609, Cur Avg Loss: 0.74564132, Log Avg loss: 0.95498592, Global Avg Loss: 2.30057604, Time: 0.0132 Steps: 34290, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000236, Sample Num: 3776, Cur Loss: 0.93885660, Cur Avg Loss: 0.74616721, Log Avg loss: 0.75805236, Global Avg Loss: 2.30012633, Time: 0.0115 Steps: 34300, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000246, Sample Num: 3936, Cur Loss: 1.01003635, Cur Avg Loss: 0.75064333, Log Avg loss: 0.85627989, Global Avg Loss: 2.29970550, Time: 0.0111 Steps: 34310, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000256, Sample Num: 4096, Cur Loss: 1.60653174, Cur Avg Loss: 0.75117545, Log Avg loss: 0.76426562, Global Avg Loss: 2.29925811, Time: 0.0138 Steps: 34320, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000266, Sample Num: 4256, Cur Loss: 0.22237733, Cur Avg Loss: 0.75813352, Log Avg loss: 0.93625991, Global Avg Loss: 2.29886109, Time: 0.0066 Steps: 34330, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000276, Sample Num: 4416, Cur Loss: 0.17608865, Cur Avg Loss: 0.75399853, Log Avg loss: 0.64400796, Global Avg Loss: 2.29837918, Time: 0.0071 Steps: 34340, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000286, Sample Num: 4576, Cur Loss: 0.46518582, Cur Avg Loss: 0.75064676, Log Avg loss: 0.65813791, Global Avg Loss: 2.29790168, Time: 0.0067 Steps: 34350, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000296, Sample Num: 4736, Cur Loss: 0.60690010, Cur Avg Loss: 0.74928394, Log Avg loss: 0.71030733, Global Avg Loss: 2.29743963, Time: 0.0067 Steps: 34360, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000306, Sample Num: 4896, Cur Loss: 1.04901171, Cur Avg Loss: 0.74947809, Log Avg loss: 0.75522495, Global Avg Loss: 2.29699092, Time: 0.0068 Steps: 34370, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000316, Sample Num: 5056, Cur Loss: 0.73640788, Cur Avg Loss: 0.75071001, Log Avg loss: 0.78840669, Global Avg Loss: 2.29655212, Time: 0.0068 Steps: 34380, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000326, Sample Num: 5216, Cur Loss: 0.86551768, Cur Avg Loss: 0.75016187, Log Avg loss: 0.73284055, Global Avg Loss: 2.29609742, Time: 0.0227 Steps: 34390, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000336, Sample Num: 5376, Cur Loss: 0.59676987, Cur Avg Loss: 0.74971615, Log Avg loss: 0.73518562, Global Avg Loss: 2.29564367, Time: 0.0090 Steps: 34400, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000346, Sample Num: 5536, Cur Loss: 0.88724923, Cur Avg Loss: 0.75021263, Log Avg loss: 0.76689442, Global Avg Loss: 2.29519939, Time: 0.0107 Steps: 34410, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000356, Sample Num: 5696, Cur Loss: 0.72211689, Cur Avg Loss: 0.74743065, Log Avg loss: 0.65117428, Global Avg Loss: 2.29472176, Time: 0.0067 Steps: 34420, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000366, Sample Num: 5856, Cur Loss: 0.84067243, Cur Avg Loss: 0.75004515, Log Avg loss: 0.84312137, Global Avg Loss: 2.29430015, Time: 0.0067 Steps: 34430, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000376, Sample Num: 6016, Cur Loss: 0.55947620, Cur Avg Loss: 0.74759703, Log Avg loss: 0.65799581, Global Avg Loss: 2.29382503, Time: 0.0111 Steps: 34440, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000386, Sample Num: 6176, Cur Loss: 0.74004030, Cur Avg Loss: 0.74285430, Log Avg loss: 0.56452767, Global Avg Loss: 2.29332306, Time: 0.0068 Steps: 34450, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000396, Sample Num: 6336, Cur Loss: 0.65193790, Cur Avg Loss: 0.74185182, Log Avg loss: 0.70315601, Global Avg Loss: 2.29286160, Time: 0.0067 Steps: 34460, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000406, Sample Num: 6496, Cur Loss: 0.73659027, Cur Avg Loss: 0.74284334, Log Avg loss: 0.78210768, Global Avg Loss: 2.29242332, Time: 0.0067 Steps: 34470, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000416, Sample Num: 6656, Cur Loss: 0.92466825, Cur Avg Loss: 0.74838706, Log Avg loss: 0.97346196, Global Avg Loss: 2.29204079, Time: 0.0068 Steps: 34480, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000426, Sample Num: 6816, Cur Loss: 2.39935398, Cur Avg Loss: 0.75221831, Log Avg loss: 0.91159811, Global Avg Loss: 2.29164055, Time: 0.0074 Steps: 34490, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000436, Sample Num: 6976, Cur Loss: 1.28273678, Cur Avg Loss: 0.75411002, Log Avg loss: 0.83469720, Global Avg Loss: 2.29121825, Time: 0.0158 Steps: 34500, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000446, Sample Num: 7136, Cur Loss: 0.96074265, Cur Avg Loss: 0.75181553, Log Avg loss: 0.65177559, Global Avg Loss: 2.29074318, Time: 0.0100 Steps: 34510, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000456, Sample Num: 7296, Cur Loss: 1.69216990, Cur Avg Loss: 0.75632288, Log Avg loss: 0.95735056, Global Avg Loss: 2.29035692, Time: 0.0225 Steps: 34520, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000466, Sample Num: 7456, Cur Loss: 0.96546608, Cur Avg Loss: 0.75593330, Log Avg loss: 0.73816869, Global Avg Loss: 2.28990740, Time: 0.0142 Steps: 34530, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000476, Sample Num: 7616, Cur Loss: 0.65480715, Cur Avg Loss: 0.75584538, Log Avg loss: 0.75174836, Global Avg Loss: 2.28946207, Time: 0.0114 Steps: 34540, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000486, Sample Num: 7776, Cur Loss: 0.83559167, Cur Avg Loss: 0.75551658, Log Avg loss: 0.73986534, Global Avg Loss: 2.28901356, Time: 0.0106 Steps: 34550, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000496, Sample Num: 7936, Cur Loss: 1.11642337, Cur Avg Loss: 0.75577547, Log Avg loss: 0.76835791, Global Avg Loss: 2.28857356, Time: 0.0121 Steps: 34560, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000506, Sample Num: 8096, Cur Loss: 0.42093730, Cur Avg Loss: 0.75507473, Log Avg loss: 0.72031763, Global Avg Loss: 2.28811991, Time: 0.0067 Steps: 34570, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000516, Sample Num: 8256, Cur Loss: 0.61813915, Cur Avg Loss: 0.75592518, Log Avg loss: 0.79895818, Global Avg Loss: 2.28768927, Time: 0.0201 Steps: 34580, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000526, Sample Num: 8416, Cur Loss: 0.43350771, Cur Avg Loss: 0.75579270, Log Avg loss: 0.74895690, Global Avg Loss: 2.28724442, Time: 0.0064 Steps: 34590, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000536, Sample Num: 8576, Cur Loss: 0.60230815, Cur Avg Loss: 0.75658794, Log Avg loss: 0.79841755, Global Avg Loss: 2.28681412, Time: 0.0065 Steps: 34600, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000546, Sample Num: 8736, Cur Loss: 0.59549153, Cur Avg Loss: 0.75526360, Log Avg loss: 0.68427865, Global Avg Loss: 2.28635110, Time: 0.0067 Steps: 34610, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000556, Sample Num: 8896, Cur Loss: 0.56452399, Cur Avg Loss: 0.75298059, Log Avg loss: 0.62832865, Global Avg Loss: 2.28587218, Time: 0.0066 Steps: 34620, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000566, Sample Num: 9056, Cur Loss: 0.84404379, Cur Avg Loss: 0.75249225, Log Avg loss: 0.72534041, Global Avg Loss: 2.28542155, Time: 0.0133 Steps: 34630, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000576, Sample Num: 9216, Cur Loss: 1.03262126, Cur Avg Loss: 0.75318318, Log Avg loss: 0.79228958, Global Avg Loss: 2.28499050, Time: 0.0071 Steps: 34640, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000586, Sample Num: 9376, Cur Loss: 1.23629057, Cur Avg Loss: 0.75242838, Log Avg loss: 0.70895183, Global Avg Loss: 2.28453566, Time: 0.0068 Steps: 34650, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000596, Sample Num: 9536, Cur Loss: 0.96908081, Cur Avg Loss: 0.75072960, Log Avg loss: 0.65118157, Global Avg Loss: 2.28406441, Time: 0.0085 Steps: 34660, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000606, Sample Num: 9696, Cur Loss: 0.76941562, Cur Avg Loss: 0.75258239, Log Avg loss: 0.86300836, Global Avg Loss: 2.28365453, Time: 0.0118 Steps: 34670, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000616, Sample Num: 9856, Cur Loss: 0.72456104, Cur Avg Loss: 0.75479090, Log Avg loss: 0.88862633, Global Avg Loss: 2.28325227, Time: 0.0162 Steps: 34680, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000626, Sample Num: 10016, Cur Loss: 0.72291297, Cur Avg Loss: 0.75375321, Log Avg loss: 0.68983190, Global Avg Loss: 2.28279294, Time: 0.0096 Steps: 34690, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000636, Sample Num: 10176, Cur Loss: 0.27126634, Cur Avg Loss: 0.74847504, Log Avg loss: 0.41806126, Global Avg Loss: 2.28225555, Time: 0.0091 Steps: 34700, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000646, Sample Num: 10336, Cur Loss: 0.53164715, Cur Avg Loss: 0.74791146, Log Avg loss: 0.71206813, Global Avg Loss: 2.28180318, Time: 0.0111 Steps: 34710, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000656, Sample Num: 10496, Cur Loss: 0.53287202, Cur Avg Loss: 0.74786447, Log Avg loss: 0.74482885, Global Avg Loss: 2.28136050, Time: 0.0109 Steps: 34720, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000666, Sample Num: 10656, Cur Loss: 0.52384859, Cur Avg Loss: 0.74440690, Log Avg loss: 0.51758996, Global Avg Loss: 2.28085265, Time: 0.0122 Steps: 34730, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000676, Sample Num: 10816, Cur Loss: 0.57742143, Cur Avg Loss: 0.74086062, Log Avg loss: 0.50467861, Global Avg Loss: 2.28034137, Time: 0.0114 Steps: 34740, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000686, Sample Num: 10976, Cur Loss: 1.66400456, Cur Avg Loss: 0.74254625, Log Avg loss: 0.85649516, Global Avg Loss: 2.27993163, Time: 0.0088 Steps: 34750, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000696, Sample Num: 11136, Cur Loss: 0.53998351, Cur Avg Loss: 0.74237779, Log Avg loss: 0.73082090, Global Avg Loss: 2.27948598, Time: 0.0085 Steps: 34760, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000706, Sample Num: 11296, Cur Loss: 0.59966183, Cur Avg Loss: 0.74336797, Log Avg loss: 0.81228501, Global Avg Loss: 2.27906400, Time: 0.0108 Steps: 34770, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000716, Sample Num: 11456, Cur Loss: 0.59734952, Cur Avg Loss: 0.74396240, Log Avg loss: 0.78592895, Global Avg Loss: 2.27863469, Time: 0.0112 Steps: 34780, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000726, Sample Num: 11616, Cur Loss: 0.42215711, Cur Avg Loss: 0.74388105, Log Avg loss: 0.73805666, Global Avg Loss: 2.27819187, Time: 0.0086 Steps: 34790, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000736, Sample Num: 11776, Cur Loss: 0.41028258, Cur Avg Loss: 0.74493849, Log Avg loss: 0.82170821, Global Avg Loss: 2.27777334, Time: 0.0132 Steps: 34800, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000746, Sample Num: 11936, Cur Loss: 1.04334641, Cur Avg Loss: 0.74543999, Log Avg loss: 0.78235055, Global Avg Loss: 2.27734375, Time: 0.0106 Steps: 34810, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000756, Sample Num: 12096, Cur Loss: 1.30509913, Cur Avg Loss: 0.74851366, Log Avg loss: 0.97780921, Global Avg Loss: 2.27697053, Time: 0.0150 Steps: 34820, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000766, Sample Num: 12256, Cur Loss: 0.23249328, Cur Avg Loss: 0.74803233, Log Avg loss: 0.71164419, Global Avg Loss: 2.27652111, Time: 0.0111 Steps: 34830, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000776, Sample Num: 12416, Cur Loss: 1.33444619, Cur Avg Loss: 0.74905194, Log Avg loss: 0.82715402, Global Avg Loss: 2.27610511, Time: 0.0088 Steps: 34840, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000786, Sample Num: 12576, Cur Loss: 0.36435205, Cur Avg Loss: 0.74820933, Log Avg loss: 0.68282285, Global Avg Loss: 2.27564792, Time: 0.0116 Steps: 34850, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000796, Sample Num: 12736, Cur Loss: 0.66679382, Cur Avg Loss: 0.74749739, Log Avg loss: 0.69153853, Global Avg Loss: 2.27519350, Time: 0.0064 Steps: 34860, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000806, Sample Num: 12896, Cur Loss: 0.90342975, Cur Avg Loss: 0.74927910, Log Avg loss: 0.89110343, Global Avg Loss: 2.27479657, Time: 0.0145 Steps: 34870, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000816, Sample Num: 13056, Cur Loss: 0.35849974, Cur Avg Loss: 0.74959040, Log Avg loss: 0.77468086, Global Avg Loss: 2.27436649, Time: 0.0076 Steps: 34880, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000826, Sample Num: 13216, Cur Loss: 0.46008170, Cur Avg Loss: 0.74879236, Log Avg loss: 0.68367283, Global Avg Loss: 2.27391058, Time: 0.0104 Steps: 34890, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000836, Sample Num: 13376, Cur Loss: 0.68477672, Cur Avg Loss: 0.74791042, Log Avg loss: 0.67506177, Global Avg Loss: 2.27345246, Time: 0.0110 Steps: 34900, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000846, Sample Num: 13536, Cur Loss: 0.72395581, Cur Avg Loss: 0.75062668, Log Avg loss: 0.97770596, Global Avg Loss: 2.27308129, Time: 0.0129 Steps: 34910, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000856, Sample Num: 13696, Cur Loss: 0.41354179, Cur Avg Loss: 0.75187932, Log Avg loss: 0.85785278, Global Avg Loss: 2.27267601, Time: 0.0093 Steps: 34920, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000866, Sample Num: 13856, Cur Loss: 1.30673504, Cur Avg Loss: 0.75120308, Log Avg loss: 0.69331729, Global Avg Loss: 2.27222386, Time: 0.0069 Steps: 34930, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000876, Sample Num: 14016, Cur Loss: 2.13064909, Cur Avg Loss: 0.75210285, Log Avg loss: 0.83002256, Global Avg Loss: 2.27181110, Time: 0.0067 Steps: 34940, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000886, Sample Num: 14176, Cur Loss: 0.34106985, Cur Avg Loss: 0.75189876, Log Avg loss: 0.73402055, Global Avg Loss: 2.27137110, Time: 0.0113 Steps: 34950, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000896, Sample Num: 14336, Cur Loss: 0.85078108, Cur Avg Loss: 0.75355721, Log Avg loss: 0.90049572, Global Avg Loss: 2.27097897, Time: 0.0066 Steps: 34960, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000906, Sample Num: 14496, Cur Loss: 0.96314460, Cur Avg Loss: 0.75618129, Log Avg loss: 0.99129946, Global Avg Loss: 2.27061303, Time: 0.0064 Steps: 34970, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000916, Sample Num: 14656, Cur Loss: 1.97914147, Cur Avg Loss: 0.75669879, Log Avg loss: 0.80358371, Global Avg Loss: 2.27019364, Time: 0.0113 Steps: 34980, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000926, Sample Num: 14816, Cur Loss: 0.79809821, Cur Avg Loss: 0.75656265, Log Avg loss: 0.74409261, Global Avg Loss: 2.26975749, Time: 0.0067 Steps: 34990, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000936, Sample Num: 14976, Cur Loss: 0.83193767, Cur Avg Loss: 0.75694598, Log Avg loss: 0.79244226, Global Avg Loss: 2.26933540, Time: 0.0068 Steps: 35000, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000946, Sample Num: 15136, Cur Loss: 0.74644703, Cur Avg Loss: 0.75632305, Log Avg loss: 0.69801691, Global Avg Loss: 2.26888658, Time: 0.0095 Steps: 35010, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000956, Sample Num: 15296, Cur Loss: 0.45186412, Cur Avg Loss: 0.75623257, Log Avg loss: 0.74767302, Global Avg Loss: 2.26845220, Time: 0.0111 Steps: 35020, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000966, Sample Num: 15456, Cur Loss: 0.55584627, Cur Avg Loss: 0.75648693, Log Avg loss: 0.78080400, Global Avg Loss: 2.26802752, Time: 0.0109 Steps: 35030, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000976, Sample Num: 15616, Cur Loss: 0.49611515, Cur Avg Loss: 0.75590322, Log Avg loss: 0.69951608, Global Avg Loss: 2.26757988, Time: 0.0066 Steps: 35040, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000986, Sample Num: 15776, Cur Loss: 1.14527047, Cur Avg Loss: 0.75727941, Log Avg loss: 0.89159569, Global Avg Loss: 2.26718731, Time: 0.0117 Steps: 35050, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 000996, Sample Num: 15936, Cur Loss: 0.95803940, Cur Avg Loss: 0.75757365, Log Avg loss: 0.78658597, Global Avg Loss: 2.26676500, Time: 0.0067 Steps: 35060, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001006, Sample Num: 16096, Cur Loss: 0.39395767, Cur Avg Loss: 0.75747514, Log Avg loss: 0.74766346, Global Avg Loss: 2.26633184, Time: 0.0069 Steps: 35070, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001016, Sample Num: 16256, Cur Loss: 1.35515881, Cur Avg Loss: 0.75727441, Log Avg loss: 0.73708140, Global Avg Loss: 2.26589591, Time: 0.0225 Steps: 35080, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001026, Sample Num: 16416, Cur Loss: 0.48188865, Cur Avg Loss: 0.75762815, Log Avg loss: 0.79356819, Global Avg Loss: 2.26547632, Time: 0.0219 Steps: 35090, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001036, Sample Num: 16576, Cur Loss: 0.61710155, Cur Avg Loss: 0.75541459, Log Avg loss: 0.52830312, Global Avg Loss: 2.26498140, Time: 0.0220 Steps: 35100, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001046, Sample Num: 16736, Cur Loss: 0.32706529, Cur Avg Loss: 0.75509918, Log Avg loss: 0.72242287, Global Avg Loss: 2.26454205, Time: 0.0185 Steps: 35110, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001056, Sample Num: 16896, Cur Loss: 0.60326576, Cur Avg Loss: 0.75603490, Log Avg loss: 0.85391063, Global Avg Loss: 2.26414039, Time: 0.0067 Steps: 35120, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001066, Sample Num: 17056, Cur Loss: 0.58352071, Cur Avg Loss: 0.75559573, Log Avg loss: 0.70921913, Global Avg Loss: 2.26369777, Time: 0.0063 Steps: 35130, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001076, Sample Num: 17216, Cur Loss: 1.31847167, Cur Avg Loss: 0.75522088, Log Avg loss: 0.71526198, Global Avg Loss: 2.26325712, Time: 0.0073 Steps: 35140, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001086, Sample Num: 17376, Cur Loss: 0.72087914, Cur Avg Loss: 0.75584959, Log Avg loss: 0.82349901, Global Avg Loss: 2.26284752, Time: 0.0137 Steps: 35150, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001096, Sample Num: 17536, Cur Loss: 0.43065625, Cur Avg Loss: 0.75583288, Log Avg loss: 0.75401847, Global Avg Loss: 2.26241838, Time: 0.0136 Steps: 35160, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001106, Sample Num: 17696, Cur Loss: 0.88354403, Cur Avg Loss: 0.75901446, Log Avg loss: 1.10771520, Global Avg Loss: 2.26209006, Time: 0.0065 Steps: 35170, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001116, Sample Num: 17856, Cur Loss: 0.75459868, Cur Avg Loss: 0.75963259, Log Avg loss: 0.82799765, Global Avg Loss: 2.26168242, Time: 0.0064 Steps: 35180, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001126, Sample Num: 18016, Cur Loss: 0.51141727, Cur Avg Loss: 0.75972901, Log Avg loss: 0.77048963, Global Avg Loss: 2.26125867, Time: 0.0065 Steps: 35190, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001136, Sample Num: 18176, Cur Loss: 0.42579538, Cur Avg Loss: 0.75925007, Log Avg loss: 0.70532119, Global Avg Loss: 2.26081664, Time: 0.0063 Steps: 35200, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001146, Sample Num: 18336, Cur Loss: 0.85181886, Cur Avg Loss: 0.75851155, Log Avg loss: 0.67461641, Global Avg Loss: 2.26036614, Time: 0.0065 Steps: 35210, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001156, Sample Num: 18496, Cur Loss: 0.52407742, Cur Avg Loss: 0.75861803, Log Avg loss: 0.77082006, Global Avg Loss: 2.25994321, Time: 0.0074 Steps: 35220, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001166, Sample Num: 18656, Cur Loss: 0.56130517, Cur Avg Loss: 0.75838335, Log Avg loss: 0.73125497, Global Avg Loss: 2.25950930, Time: 0.0066 Steps: 35230, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001176, Sample Num: 18816, Cur Loss: 0.63718760, Cur Avg Loss: 0.75959996, Log Avg loss: 0.90145610, Global Avg Loss: 2.25912393, Time: 0.0064 Steps: 35240, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001186, Sample Num: 18976, Cur Loss: 0.54336572, Cur Avg Loss: 0.75848245, Log Avg loss: 0.62706311, Global Avg Loss: 2.25866093, Time: 0.0170 Steps: 35250, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001196, Sample Num: 19136, Cur Loss: 0.91200626, Cur Avg Loss: 0.75813548, Log Avg loss: 0.71698575, Global Avg Loss: 2.25822370, Time: 0.0066 Steps: 35260, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001206, Sample Num: 19296, Cur Loss: 1.13867521, Cur Avg Loss: 0.75904962, Log Avg loss: 0.86838078, Global Avg Loss: 2.25782964, Time: 0.0115 Steps: 35270, Updated lr: 0.000068 Training, Epoch: 0017, Batch: 001216, Sample Num: 19456, Cur Loss: 1.39626229, Cur Avg Loss: 0.76012341, Log Avg loss: 0.88962140, Global Avg Loss: 2.25744183, Time: 0.0066 Steps: 35280, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001226, Sample Num: 19616, Cur Loss: 0.50573146, Cur Avg Loss: 0.75947392, Log Avg loss: 0.68049674, Global Avg Loss: 2.25699497, Time: 0.0067 Steps: 35290, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001236, Sample Num: 19776, Cur Loss: 0.31038728, Cur Avg Loss: 0.75828805, Log Avg loss: 0.61289984, Global Avg Loss: 2.25652922, Time: 0.0065 Steps: 35300, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001246, Sample Num: 19936, Cur Loss: 0.83975959, Cur Avg Loss: 0.76015115, Log Avg loss: 0.99043019, Global Avg Loss: 2.25617066, Time: 0.0066 Steps: 35310, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001256, Sample Num: 20096, Cur Loss: 0.69200730, Cur Avg Loss: 0.76112295, Log Avg loss: 0.88220959, Global Avg Loss: 2.25578165, Time: 0.0152 Steps: 35320, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001266, Sample Num: 20256, Cur Loss: 0.14533466, Cur Avg Loss: 0.76134354, Log Avg loss: 0.78905027, Global Avg Loss: 2.25536650, Time: 0.0067 Steps: 35330, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001276, Sample Num: 20416, Cur Loss: 0.65050566, Cur Avg Loss: 0.76245990, Log Avg loss: 0.90379038, Global Avg Loss: 2.25498405, Time: 0.0160 Steps: 35340, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001286, Sample Num: 20576, Cur Loss: 0.92369640, Cur Avg Loss: 0.76236198, Log Avg loss: 0.74986723, Global Avg Loss: 2.25455828, Time: 0.0117 Steps: 35350, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001296, Sample Num: 20736, Cur Loss: 0.77697408, Cur Avg Loss: 0.76201649, Log Avg loss: 0.71758630, Global Avg Loss: 2.25412361, Time: 0.0086 Steps: 35360, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001306, Sample Num: 20896, Cur Loss: 1.19108689, Cur Avg Loss: 0.76223617, Log Avg loss: 0.79070736, Global Avg Loss: 2.25370987, Time: 0.0139 Steps: 35370, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001316, Sample Num: 21056, Cur Loss: 0.67167437, Cur Avg Loss: 0.76215695, Log Avg loss: 0.75181100, Global Avg Loss: 2.25328536, Time: 0.0066 Steps: 35380, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001326, Sample Num: 21216, Cur Loss: 1.43095922, Cur Avg Loss: 0.76339827, Log Avg loss: 0.92675526, Global Avg Loss: 2.25291053, Time: 0.0064 Steps: 35390, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001336, Sample Num: 21376, Cur Loss: 0.32028586, Cur Avg Loss: 0.76228593, Log Avg loss: 0.61479001, Global Avg Loss: 2.25244778, Time: 0.0121 Steps: 35400, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001346, Sample Num: 21536, Cur Loss: 1.07483387, Cur Avg Loss: 0.76201861, Log Avg loss: 0.72630475, Global Avg Loss: 2.25201679, Time: 0.0081 Steps: 35410, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001356, Sample Num: 21696, Cur Loss: 0.42973202, Cur Avg Loss: 0.76219027, Log Avg loss: 0.78529554, Global Avg Loss: 2.25160270, Time: 0.0110 Steps: 35420, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001366, Sample Num: 21856, Cur Loss: 0.39202428, Cur Avg Loss: 0.76167650, Log Avg loss: 0.69200947, Global Avg Loss: 2.25116251, Time: 0.0071 Steps: 35430, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001376, Sample Num: 22016, Cur Loss: 0.27100232, Cur Avg Loss: 0.76145362, Log Avg loss: 0.73100828, Global Avg Loss: 2.25073357, Time: 0.0120 Steps: 35440, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001386, Sample Num: 22176, Cur Loss: 0.73093075, Cur Avg Loss: 0.76222510, Log Avg loss: 0.86837993, Global Avg Loss: 2.25034363, Time: 0.0110 Steps: 35450, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001396, Sample Num: 22336, Cur Loss: 0.80650866, Cur Avg Loss: 0.76283079, Log Avg loss: 0.84677969, Global Avg Loss: 2.24994781, Time: 0.0066 Steps: 35460, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001406, Sample Num: 22496, Cur Loss: 0.78384066, Cur Avg Loss: 0.76241074, Log Avg loss: 0.70377279, Global Avg Loss: 2.24951190, Time: 0.0145 Steps: 35470, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001416, Sample Num: 22656, Cur Loss: 0.54305232, Cur Avg Loss: 0.76387571, Log Avg loss: 0.96984931, Global Avg Loss: 2.24915123, Time: 0.0073 Steps: 35480, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001426, Sample Num: 22816, Cur Loss: 0.50626791, Cur Avg Loss: 0.76337041, Log Avg loss: 0.69181990, Global Avg Loss: 2.24871242, Time: 0.0113 Steps: 35490, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001436, Sample Num: 22976, Cur Loss: 0.83849013, Cur Avg Loss: 0.76414075, Log Avg loss: 0.87399139, Global Avg Loss: 2.24832517, Time: 0.0086 Steps: 35500, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001446, Sample Num: 23136, Cur Loss: 2.35687613, Cur Avg Loss: 0.76520472, Log Avg loss: 0.91799138, Global Avg Loss: 2.24795054, Time: 0.0201 Steps: 35510, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001456, Sample Num: 23296, Cur Loss: 0.25415182, Cur Avg Loss: 0.76504475, Log Avg loss: 0.74191282, Global Avg Loss: 2.24752654, Time: 0.0066 Steps: 35520, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001466, Sample Num: 23456, Cur Loss: 0.11772336, Cur Avg Loss: 0.76392462, Log Avg loss: 0.60083310, Global Avg Loss: 2.24706308, Time: 0.0127 Steps: 35530, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001476, Sample Num: 23616, Cur Loss: 0.95715731, Cur Avg Loss: 0.76468628, Log Avg loss: 0.87634588, Global Avg Loss: 2.24667739, Time: 0.0066 Steps: 35540, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001486, Sample Num: 23776, Cur Loss: 0.47203183, Cur Avg Loss: 0.76458413, Log Avg loss: 0.74950684, Global Avg Loss: 2.24625625, Time: 0.0105 Steps: 35550, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001496, Sample Num: 23936, Cur Loss: 0.65909374, Cur Avg Loss: 0.76364826, Log Avg loss: 0.62457794, Global Avg Loss: 2.24580021, Time: 0.0127 Steps: 35560, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001506, Sample Num: 24096, Cur Loss: 0.90587562, Cur Avg Loss: 0.76359002, Log Avg loss: 0.75487739, Global Avg Loss: 2.24538106, Time: 0.0066 Steps: 35570, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001516, Sample Num: 24256, Cur Loss: 0.33198193, Cur Avg Loss: 0.76261108, Log Avg loss: 0.61518255, Global Avg Loss: 2.24492288, Time: 0.0094 Steps: 35580, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001526, Sample Num: 24416, Cur Loss: 0.29836777, Cur Avg Loss: 0.76322524, Log Avg loss: 0.85633262, Global Avg Loss: 2.24453271, Time: 0.0071 Steps: 35590, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001536, Sample Num: 24576, Cur Loss: 0.38383242, Cur Avg Loss: 0.76197675, Log Avg loss: 0.57145683, Global Avg Loss: 2.24406275, Time: 0.0087 Steps: 35600, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001546, Sample Num: 24736, Cur Loss: 0.97479486, Cur Avg Loss: 0.76271791, Log Avg loss: 0.87656028, Global Avg Loss: 2.24367873, Time: 0.0113 Steps: 35610, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001556, Sample Num: 24896, Cur Loss: 0.69329005, Cur Avg Loss: 0.76323640, Log Avg loss: 0.84339464, Global Avg Loss: 2.24328561, Time: 0.0167 Steps: 35620, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001566, Sample Num: 25056, Cur Loss: 0.42468286, Cur Avg Loss: 0.76321656, Log Avg loss: 0.76013038, Global Avg Loss: 2.24286934, Time: 0.0066 Steps: 35630, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001576, Sample Num: 25216, Cur Loss: 0.45720029, Cur Avg Loss: 0.76215379, Log Avg loss: 0.59572340, Global Avg Loss: 2.24240718, Time: 0.0107 Steps: 35640, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001586, Sample Num: 25376, Cur Loss: 1.21174061, Cur Avg Loss: 0.76344650, Log Avg loss: 0.96717740, Global Avg Loss: 2.24204947, Time: 0.0158 Steps: 35650, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001596, Sample Num: 25536, Cur Loss: 0.39122179, Cur Avg Loss: 0.76229360, Log Avg loss: 0.57944372, Global Avg Loss: 2.24158323, Time: 0.0072 Steps: 35660, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001606, Sample Num: 25696, Cur Loss: 0.58383894, Cur Avg Loss: 0.76226202, Log Avg loss: 0.75722246, Global Avg Loss: 2.24116710, Time: 0.0164 Steps: 35670, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001616, Sample Num: 25856, Cur Loss: 0.38166758, Cur Avg Loss: 0.76210123, Log Avg loss: 0.73627773, Global Avg Loss: 2.24074532, Time: 0.0064 Steps: 35680, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001626, Sample Num: 26016, Cur Loss: 0.49919230, Cur Avg Loss: 0.76114488, Log Avg loss: 0.60659825, Global Avg Loss: 2.24028745, Time: 0.0132 Steps: 35690, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001636, Sample Num: 26176, Cur Loss: 0.75486517, Cur Avg Loss: 0.76099777, Log Avg loss: 0.73707855, Global Avg Loss: 2.23986638, Time: 0.0067 Steps: 35700, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001646, Sample Num: 26336, Cur Loss: 1.15696323, Cur Avg Loss: 0.76090014, Log Avg loss: 0.74492754, Global Avg Loss: 2.23944775, Time: 0.0069 Steps: 35710, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001656, Sample Num: 26496, Cur Loss: 0.41785741, Cur Avg Loss: 0.76022930, Log Avg loss: 0.64980957, Global Avg Loss: 2.23900272, Time: 0.0065 Steps: 35720, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001666, Sample Num: 26656, Cur Loss: 0.58006841, Cur Avg Loss: 0.76091576, Log Avg loss: 0.87459258, Global Avg Loss: 2.23862086, Time: 0.0133 Steps: 35730, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001676, Sample Num: 26816, Cur Loss: 0.76839858, Cur Avg Loss: 0.76179651, Log Avg loss: 0.90852988, Global Avg Loss: 2.23824870, Time: 0.0074 Steps: 35740, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001686, Sample Num: 26976, Cur Loss: 0.45804200, Cur Avg Loss: 0.76049536, Log Avg loss: 0.54242231, Global Avg Loss: 2.23777434, Time: 0.0064 Steps: 35750, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001696, Sample Num: 27136, Cur Loss: 0.78146470, Cur Avg Loss: 0.75995240, Log Avg loss: 0.66840965, Global Avg Loss: 2.23733548, Time: 0.0068 Steps: 35760, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001706, Sample Num: 27296, Cur Loss: 1.45089972, Cur Avg Loss: 0.76026032, Log Avg loss: 0.81248391, Global Avg Loss: 2.23693715, Time: 0.0067 Steps: 35770, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001716, Sample Num: 27456, Cur Loss: 0.19737878, Cur Avg Loss: 0.76040004, Log Avg loss: 0.78423544, Global Avg Loss: 2.23653114, Time: 0.0065 Steps: 35780, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001726, Sample Num: 27616, Cur Loss: 0.27703023, Cur Avg Loss: 0.76045012, Log Avg loss: 0.76904385, Global Avg Loss: 2.23612111, Time: 0.0066 Steps: 35790, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001736, Sample Num: 27776, Cur Loss: 0.60799634, Cur Avg Loss: 0.76089965, Log Avg loss: 0.83848885, Global Avg Loss: 2.23573071, Time: 0.0074 Steps: 35800, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001746, Sample Num: 27936, Cur Loss: 0.46136624, Cur Avg Loss: 0.75993127, Log Avg loss: 0.59182058, Global Avg Loss: 2.23527164, Time: 0.0072 Steps: 35810, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001756, Sample Num: 28096, Cur Loss: 1.22769070, Cur Avg Loss: 0.76038399, Log Avg loss: 0.83942986, Global Avg Loss: 2.23488196, Time: 0.0067 Steps: 35820, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001766, Sample Num: 28256, Cur Loss: 0.84498411, Cur Avg Loss: 0.76013625, Log Avg loss: 0.71663143, Global Avg Loss: 2.23445822, Time: 0.0070 Steps: 35830, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001776, Sample Num: 28416, Cur Loss: 0.49817094, Cur Avg Loss: 0.75945774, Log Avg loss: 0.63963453, Global Avg Loss: 2.23401324, Time: 0.0199 Steps: 35840, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001786, Sample Num: 28576, Cur Loss: 0.76764214, Cur Avg Loss: 0.75995042, Log Avg loss: 0.84744988, Global Avg Loss: 2.23362647, Time: 0.0075 Steps: 35850, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001796, Sample Num: 28736, Cur Loss: 0.78144187, Cur Avg Loss: 0.76092464, Log Avg loss: 0.93491989, Global Avg Loss: 2.23326431, Time: 0.0074 Steps: 35860, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001806, Sample Num: 28896, Cur Loss: 0.22488691, Cur Avg Loss: 0.76053742, Log Avg loss: 0.69099294, Global Avg Loss: 2.23283435, Time: 0.0150 Steps: 35870, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001816, Sample Num: 29056, Cur Loss: 0.63430929, Cur Avg Loss: 0.76004548, Log Avg loss: 0.67120092, Global Avg Loss: 2.23239911, Time: 0.0065 Steps: 35880, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001826, Sample Num: 29216, Cur Loss: 0.85486704, Cur Avg Loss: 0.76038041, Log Avg loss: 0.82120332, Global Avg Loss: 2.23200591, Time: 0.0135 Steps: 35890, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001836, Sample Num: 29376, Cur Loss: 0.70870435, Cur Avg Loss: 0.76043815, Log Avg loss: 0.77098260, Global Avg Loss: 2.23159894, Time: 0.0068 Steps: 35900, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001846, Sample Num: 29536, Cur Loss: 0.71734774, Cur Avg Loss: 0.76016865, Log Avg loss: 0.71068828, Global Avg Loss: 2.23117541, Time: 0.0068 Steps: 35910, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001856, Sample Num: 29696, Cur Loss: 0.74065995, Cur Avg Loss: 0.75987039, Log Avg loss: 0.70481026, Global Avg Loss: 2.23075047, Time: 0.0067 Steps: 35920, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001866, Sample Num: 29856, Cur Loss: 0.61750567, Cur Avg Loss: 0.75872128, Log Avg loss: 0.54544778, Global Avg Loss: 2.23028142, Time: 0.0081 Steps: 35930, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001876, Sample Num: 30016, Cur Loss: 0.84284312, Cur Avg Loss: 0.75847482, Log Avg loss: 0.71248520, Global Avg Loss: 2.22985911, Time: 0.0086 Steps: 35940, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001886, Sample Num: 30176, Cur Loss: 0.37594950, Cur Avg Loss: 0.75934537, Log Avg loss: 0.92266067, Global Avg Loss: 2.22949549, Time: 0.0110 Steps: 35950, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001896, Sample Num: 30336, Cur Loss: 0.92075771, Cur Avg Loss: 0.75825808, Log Avg loss: 0.55319423, Global Avg Loss: 2.22902934, Time: 0.0115 Steps: 35960, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001906, Sample Num: 30496, Cur Loss: 0.62438136, Cur Avg Loss: 0.75829008, Log Avg loss: 0.76435703, Global Avg Loss: 2.22862214, Time: 0.0067 Steps: 35970, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001916, Sample Num: 30656, Cur Loss: 0.26150578, Cur Avg Loss: 0.75794015, Log Avg loss: 0.69124505, Global Avg Loss: 2.22819486, Time: 0.0098 Steps: 35980, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001926, Sample Num: 30816, Cur Loss: 0.41117710, Cur Avg Loss: 0.75756303, Log Avg loss: 0.68530588, Global Avg Loss: 2.22776616, Time: 0.0072 Steps: 35990, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001936, Sample Num: 30976, Cur Loss: 0.80396092, Cur Avg Loss: 0.75777200, Log Avg loss: 0.79801920, Global Avg Loss: 2.22736900, Time: 0.0114 Steps: 36000, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001946, Sample Num: 31136, Cur Loss: 0.57753468, Cur Avg Loss: 0.75840411, Log Avg loss: 0.88078094, Global Avg Loss: 2.22699506, Time: 0.0072 Steps: 36010, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001956, Sample Num: 31296, Cur Loss: 0.89541179, Cur Avg Loss: 0.75889964, Log Avg loss: 0.85533049, Global Avg Loss: 2.22661425, Time: 0.0108 Steps: 36020, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001966, Sample Num: 31456, Cur Loss: 0.83541465, Cur Avg Loss: 0.75809709, Log Avg loss: 0.60111734, Global Avg Loss: 2.22616310, Time: 0.0063 Steps: 36030, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001976, Sample Num: 31616, Cur Loss: 0.31444126, Cur Avg Loss: 0.75797237, Log Avg loss: 0.73345285, Global Avg Loss: 2.22574892, Time: 0.0068 Steps: 36040, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001986, Sample Num: 31776, Cur Loss: 0.34842798, Cur Avg Loss: 0.75763569, Log Avg loss: 0.69110684, Global Avg Loss: 2.22532322, Time: 0.0068 Steps: 36050, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 001996, Sample Num: 31936, Cur Loss: 0.62030709, Cur Avg Loss: 0.75657414, Log Avg loss: 0.54575226, Global Avg Loss: 2.22485745, Time: 0.0097 Steps: 36060, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002006, Sample Num: 32096, Cur Loss: 1.00817490, Cur Avg Loss: 0.75756436, Log Avg loss: 0.95521063, Global Avg Loss: 2.22450545, Time: 0.0115 Steps: 36070, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002016, Sample Num: 32256, Cur Loss: 1.83394599, Cur Avg Loss: 0.75890028, Log Avg loss: 1.02688751, Global Avg Loss: 2.22417352, Time: 0.0106 Steps: 36080, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002026, Sample Num: 32416, Cur Loss: 0.54250425, Cur Avg Loss: 0.75866969, Log Avg loss: 0.71218103, Global Avg Loss: 2.22375457, Time: 0.0113 Steps: 36090, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002036, Sample Num: 32576, Cur Loss: 0.28747633, Cur Avg Loss: 0.75816424, Log Avg loss: 0.65576173, Global Avg Loss: 2.22332022, Time: 0.0156 Steps: 36100, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002046, Sample Num: 32736, Cur Loss: 0.86562097, Cur Avg Loss: 0.75838167, Log Avg loss: 0.80264899, Global Avg Loss: 2.22292679, Time: 0.0121 Steps: 36110, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002056, Sample Num: 32896, Cur Loss: 0.39463627, Cur Avg Loss: 0.75756174, Log Avg loss: 0.58980379, Global Avg Loss: 2.22247465, Time: 0.0122 Steps: 36120, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002066, Sample Num: 33056, Cur Loss: 1.27328336, Cur Avg Loss: 0.75825220, Log Avg loss: 0.90021114, Global Avg Loss: 2.22210868, Time: 0.0114 Steps: 36130, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002076, Sample Num: 33216, Cur Loss: 0.35047093, Cur Avg Loss: 0.75756138, Log Avg loss: 0.61483929, Global Avg Loss: 2.22166395, Time: 0.0121 Steps: 36140, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002086, Sample Num: 33376, Cur Loss: 0.64611310, Cur Avg Loss: 0.75790714, Log Avg loss: 0.82968608, Global Avg Loss: 2.22127889, Time: 0.0117 Steps: 36150, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002096, Sample Num: 33536, Cur Loss: 0.49890631, Cur Avg Loss: 0.75732977, Log Avg loss: 0.63689056, Global Avg Loss: 2.22084073, Time: 0.0157 Steps: 36160, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002106, Sample Num: 33696, Cur Loss: 0.72675610, Cur Avg Loss: 0.75641921, Log Avg loss: 0.56556625, Global Avg Loss: 2.22038309, Time: 0.0152 Steps: 36170, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002116, Sample Num: 33856, Cur Loss: 0.22714907, Cur Avg Loss: 0.75565297, Log Avg loss: 0.59428313, Global Avg Loss: 2.21993364, Time: 0.0112 Steps: 36180, Updated lr: 0.000067 Training, Epoch: 0017, Batch: 002126, Sample Num: 34016, Cur Loss: 0.90123141, Cur Avg Loss: 0.75571042, Log Avg loss: 0.76786527, Global Avg Loss: 2.21953241, Time: 0.0070 Steps: 36190, Updated lr: 0.000067 ***** Running evaluation checkpoint-36193 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-36193 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.865476, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.930457, "eval_total_loss": 654.111465, "eval_mae": 0.807277, "eval_mse": 0.930525, "eval_r2": 0.408497, "eval_sp_statistic": 0.63513, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.697374, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.380038, "test_total_loss": 692.77932, "test_mae": 1.041717, "test_mse": 1.37997, "test_r2": 0.109355, "test_sp_statistic": 0.456111, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.573282, "test_ps_pvalue": 0.0, "lr": 6.662588904694168e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.2194208263530757, "train_cur_epoch_loss": 1609.2603948488832, "train_cur_epoch_avg_loss": 0.7558761835833175, "train_cur_epoch_time": 21.86547613143921, "train_cur_epoch_avg_time": 0.010270303490577365, "epoch": 17, "step": 36193} ################################################## Training, Epoch: 0018, Batch: 000007, Sample Num: 112, Cur Loss: 0.94415939, Cur Avg Loss: 0.67435921, Log Avg loss: 0.73405622, Global Avg Loss: 2.21912206, Time: 0.0068 Steps: 36200, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000017, Sample Num: 272, Cur Loss: 0.83528101, Cur Avg Loss: 0.61405330, Log Avg loss: 0.57183917, Global Avg Loss: 2.21866713, Time: 0.0122 Steps: 36210, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000027, Sample Num: 432, Cur Loss: 1.28814209, Cur Avg Loss: 0.72002775, Log Avg loss: 0.90018431, Global Avg Loss: 2.21830311, Time: 0.0118 Steps: 36220, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000037, Sample Num: 592, Cur Loss: 0.40452310, Cur Avg Loss: 0.71493908, Log Avg loss: 0.70119967, Global Avg Loss: 2.21788437, Time: 0.0129 Steps: 36230, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000047, Sample Num: 752, Cur Loss: 2.01032329, Cur Avg Loss: 0.72637111, Log Avg loss: 0.76866964, Global Avg Loss: 2.21748448, Time: 0.0109 Steps: 36240, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000057, Sample Num: 912, Cur Loss: 0.67101967, Cur Avg Loss: 0.74085420, Log Avg loss: 0.80892469, Global Avg Loss: 2.21709591, Time: 0.0115 Steps: 36250, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000067, Sample Num: 1072, Cur Loss: 0.48649579, Cur Avg Loss: 0.74195861, Log Avg loss: 0.74825377, Global Avg Loss: 2.21669082, Time: 0.0069 Steps: 36260, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000077, Sample Num: 1232, Cur Loss: 0.30966505, Cur Avg Loss: 0.70967111, Log Avg loss: 0.49334483, Global Avg Loss: 2.21621568, Time: 0.0120 Steps: 36270, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000087, Sample Num: 1392, Cur Loss: 1.26477206, Cur Avg Loss: 0.71789168, Log Avg loss: 0.78119005, Global Avg Loss: 2.21582014, Time: 0.0073 Steps: 36280, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000097, Sample Num: 1552, Cur Loss: 1.02297902, Cur Avg Loss: 0.71310356, Log Avg loss: 0.67144691, Global Avg Loss: 2.21539457, Time: 0.0114 Steps: 36290, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000107, Sample Num: 1712, Cur Loss: 0.82284784, Cur Avg Loss: 0.71444033, Log Avg loss: 0.72740707, Global Avg Loss: 2.21498466, Time: 0.0074 Steps: 36300, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000117, Sample Num: 1872, Cur Loss: 0.92628670, Cur Avg Loss: 0.72587342, Log Avg loss: 0.84820751, Global Avg Loss: 2.21460824, Time: 0.0073 Steps: 36310, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000127, Sample Num: 2032, Cur Loss: 0.56226420, Cur Avg Loss: 0.73253245, Log Avg loss: 0.81044303, Global Avg Loss: 2.21422163, Time: 0.0137 Steps: 36320, Updated lr: 0.000067 Training, Epoch: 0018, Batch: 000137, Sample Num: 2192, Cur Loss: 0.52867055, Cur Avg Loss: 0.73185960, Log Avg loss: 0.72331448, Global Avg Loss: 2.21381125, Time: 0.0122 Steps: 36330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000147, Sample Num: 2352, Cur Loss: 0.35387582, Cur Avg Loss: 0.72722412, Log Avg loss: 0.66371796, Global Avg Loss: 2.21338470, Time: 0.0067 Steps: 36340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000157, Sample Num: 2512, Cur Loss: 1.16370833, Cur Avg Loss: 0.73195258, Log Avg loss: 0.80146092, Global Avg Loss: 2.21299627, Time: 0.0073 Steps: 36350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000167, Sample Num: 2672, Cur Loss: 0.74276108, Cur Avg Loss: 0.72667264, Log Avg loss: 0.64377770, Global Avg Loss: 2.21256469, Time: 0.0119 Steps: 36360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000177, Sample Num: 2832, Cur Loss: 0.54279101, Cur Avg Loss: 0.72037977, Log Avg loss: 0.61528876, Global Avg Loss: 2.21212552, Time: 0.0112 Steps: 36370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000187, Sample Num: 2992, Cur Loss: 0.63680840, Cur Avg Loss: 0.73217340, Log Avg loss: 0.94092065, Global Avg Loss: 2.21177610, Time: 0.0110 Steps: 36380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000197, Sample Num: 3152, Cur Loss: 0.61827886, Cur Avg Loss: 0.72865788, Log Avg loss: 0.66291759, Global Avg Loss: 2.21135047, Time: 0.0100 Steps: 36390, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000207, Sample Num: 3312, Cur Loss: 0.36513710, Cur Avg Loss: 0.73499842, Log Avg loss: 0.85990714, Global Avg Loss: 2.21097919, Time: 0.0117 Steps: 36400, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000217, Sample Num: 3472, Cur Loss: 0.86623293, Cur Avg Loss: 0.73904340, Log Avg loss: 0.82277443, Global Avg Loss: 2.21059792, Time: 0.0140 Steps: 36410, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000227, Sample Num: 3632, Cur Loss: 1.35248065, Cur Avg Loss: 0.73458262, Log Avg loss: 0.63778379, Global Avg Loss: 2.21016607, Time: 0.0067 Steps: 36420, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000237, Sample Num: 3792, Cur Loss: 0.61997890, Cur Avg Loss: 0.73685693, Log Avg loss: 0.78848367, Global Avg Loss: 2.20977582, Time: 0.0069 Steps: 36430, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000247, Sample Num: 3952, Cur Loss: 1.02397835, Cur Avg Loss: 0.73978536, Log Avg loss: 0.80918922, Global Avg Loss: 2.20939146, Time: 0.0134 Steps: 36440, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000257, Sample Num: 4112, Cur Loss: 0.66791403, Cur Avg Loss: 0.74181472, Log Avg loss: 0.79193978, Global Avg Loss: 2.20900259, Time: 0.0146 Steps: 36450, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000267, Sample Num: 4272, Cur Loss: 0.60539371, Cur Avg Loss: 0.74933664, Log Avg loss: 0.94265000, Global Avg Loss: 2.20865526, Time: 0.0066 Steps: 36460, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000277, Sample Num: 4432, Cur Loss: 0.37505215, Cur Avg Loss: 0.74663885, Log Avg loss: 0.67460782, Global Avg Loss: 2.20823463, Time: 0.0067 Steps: 36470, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000287, Sample Num: 4592, Cur Loss: 1.72330070, Cur Avg Loss: 0.74970842, Log Avg loss: 0.83473575, Global Avg Loss: 2.20785812, Time: 0.0066 Steps: 36480, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000297, Sample Num: 4752, Cur Loss: 0.92750216, Cur Avg Loss: 0.74922682, Log Avg loss: 0.73540485, Global Avg Loss: 2.20745460, Time: 0.0110 Steps: 36490, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000307, Sample Num: 4912, Cur Loss: 1.03198171, Cur Avg Loss: 0.74579414, Log Avg loss: 0.64384351, Global Avg Loss: 2.20702621, Time: 0.0113 Steps: 36500, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000317, Sample Num: 5072, Cur Loss: 0.23670703, Cur Avg Loss: 0.74277532, Log Avg loss: 0.65009739, Global Avg Loss: 2.20659977, Time: 0.0108 Steps: 36510, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000327, Sample Num: 5232, Cur Loss: 1.02215850, Cur Avg Loss: 0.74352133, Log Avg loss: 0.76717003, Global Avg Loss: 2.20620563, Time: 0.0067 Steps: 36520, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000337, Sample Num: 5392, Cur Loss: 1.49940705, Cur Avg Loss: 0.74763333, Log Avg loss: 0.88209554, Global Avg Loss: 2.20584315, Time: 0.0067 Steps: 36530, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000347, Sample Num: 5552, Cur Loss: 0.48999569, Cur Avg Loss: 0.74841530, Log Avg loss: 0.77476772, Global Avg Loss: 2.20545151, Time: 0.0113 Steps: 36540, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000357, Sample Num: 5712, Cur Loss: 0.82289732, Cur Avg Loss: 0.74866241, Log Avg loss: 0.75723712, Global Avg Loss: 2.20505528, Time: 0.0115 Steps: 36550, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000367, Sample Num: 5872, Cur Loss: 0.93269551, Cur Avg Loss: 0.75129638, Log Avg loss: 0.84532930, Global Avg Loss: 2.20468336, Time: 0.0123 Steps: 36560, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000377, Sample Num: 6032, Cur Loss: 0.45294026, Cur Avg Loss: 0.74734256, Log Avg loss: 0.60223735, Global Avg Loss: 2.20424518, Time: 0.0111 Steps: 36570, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000387, Sample Num: 6192, Cur Loss: 1.42383742, Cur Avg Loss: 0.74856963, Log Avg loss: 0.79483021, Global Avg Loss: 2.20385988, Time: 0.0114 Steps: 36580, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000397, Sample Num: 6352, Cur Loss: 1.10403752, Cur Avg Loss: 0.74847383, Log Avg loss: 0.74476616, Global Avg Loss: 2.20346111, Time: 0.0069 Steps: 36590, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000407, Sample Num: 6512, Cur Loss: 1.62465668, Cur Avg Loss: 0.75114200, Log Avg loss: 0.85706841, Global Avg Loss: 2.20309324, Time: 0.0191 Steps: 36600, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000417, Sample Num: 6672, Cur Loss: 0.97993916, Cur Avg Loss: 0.75036210, Log Avg loss: 0.71862017, Global Avg Loss: 2.20268776, Time: 0.0115 Steps: 36610, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000427, Sample Num: 6832, Cur Loss: 1.00325942, Cur Avg Loss: 0.75221801, Log Avg loss: 0.82960935, Global Avg Loss: 2.20231281, Time: 0.0108 Steps: 36620, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000437, Sample Num: 6992, Cur Loss: 0.58571929, Cur Avg Loss: 0.74676888, Log Avg loss: 0.51409096, Global Avg Loss: 2.20185192, Time: 0.0068 Steps: 36630, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000447, Sample Num: 7152, Cur Loss: 0.80765128, Cur Avg Loss: 0.74653863, Log Avg loss: 0.73647701, Global Avg Loss: 2.20145199, Time: 0.0071 Steps: 36640, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000457, Sample Num: 7312, Cur Loss: 0.15696529, Cur Avg Loss: 0.74261205, Log Avg loss: 0.56709363, Global Avg Loss: 2.20100605, Time: 0.0107 Steps: 36650, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000467, Sample Num: 7472, Cur Loss: 0.44795486, Cur Avg Loss: 0.74204738, Log Avg loss: 0.71624238, Global Avg Loss: 2.20060104, Time: 0.0069 Steps: 36660, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000477, Sample Num: 7632, Cur Loss: 0.35848290, Cur Avg Loss: 0.74089247, Log Avg loss: 0.68695775, Global Avg Loss: 2.20018826, Time: 0.0137 Steps: 36670, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000487, Sample Num: 7792, Cur Loss: 0.49353889, Cur Avg Loss: 0.73770042, Log Avg loss: 0.58543980, Global Avg Loss: 2.19974804, Time: 0.0141 Steps: 36680, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000497, Sample Num: 7952, Cur Loss: 0.43176210, Cur Avg Loss: 0.73650275, Log Avg loss: 0.67817646, Global Avg Loss: 2.19933333, Time: 0.0137 Steps: 36690, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000507, Sample Num: 8112, Cur Loss: 0.27139670, Cur Avg Loss: 0.73544479, Log Avg loss: 0.68286421, Global Avg Loss: 2.19892012, Time: 0.0169 Steps: 36700, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000517, Sample Num: 8272, Cur Loss: 0.19982333, Cur Avg Loss: 0.73543266, Log Avg loss: 0.73481739, Global Avg Loss: 2.19852129, Time: 0.0067 Steps: 36710, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000527, Sample Num: 8432, Cur Loss: 0.39759749, Cur Avg Loss: 0.73025870, Log Avg loss: 0.46276479, Global Avg Loss: 2.19804859, Time: 0.0070 Steps: 36720, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000537, Sample Num: 8592, Cur Loss: 0.97483605, Cur Avg Loss: 0.72972466, Log Avg loss: 0.70158081, Global Avg Loss: 2.19764117, Time: 0.0118 Steps: 36730, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000547, Sample Num: 8752, Cur Loss: 0.38530469, Cur Avg Loss: 0.72571449, Log Avg loss: 0.51036827, Global Avg Loss: 2.19718192, Time: 0.0116 Steps: 36740, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000557, Sample Num: 8912, Cur Loss: 1.32431352, Cur Avg Loss: 0.72375354, Log Avg loss: 0.61648993, Global Avg Loss: 2.19675180, Time: 0.0159 Steps: 36750, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000567, Sample Num: 9072, Cur Loss: 0.34550250, Cur Avg Loss: 0.72108593, Log Avg loss: 0.57250011, Global Avg Loss: 2.19630995, Time: 0.0112 Steps: 36760, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000577, Sample Num: 9232, Cur Loss: 0.43816906, Cur Avg Loss: 0.71917502, Log Avg loss: 0.61082601, Global Avg Loss: 2.19587876, Time: 0.0067 Steps: 36770, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000587, Sample Num: 9392, Cur Loss: 0.50352627, Cur Avg Loss: 0.72216189, Log Avg loss: 0.89450472, Global Avg Loss: 2.19552493, Time: 0.0143 Steps: 36780, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000597, Sample Num: 9552, Cur Loss: 1.12382329, Cur Avg Loss: 0.71928410, Log Avg loss: 0.55035774, Global Avg Loss: 2.19507775, Time: 0.0068 Steps: 36790, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000607, Sample Num: 9712, Cur Loss: 0.83590859, Cur Avg Loss: 0.71884116, Log Avg loss: 0.69239752, Global Avg Loss: 2.19466942, Time: 0.0067 Steps: 36800, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000617, Sample Num: 9872, Cur Loss: 0.23860297, Cur Avg Loss: 0.72110339, Log Avg loss: 0.85842053, Global Avg Loss: 2.19430640, Time: 0.0133 Steps: 36810, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000627, Sample Num: 10032, Cur Loss: 0.77932703, Cur Avg Loss: 0.72107932, Log Avg loss: 0.71959470, Global Avg Loss: 2.19390589, Time: 0.0114 Steps: 36820, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000637, Sample Num: 10192, Cur Loss: 0.43931034, Cur Avg Loss: 0.72445221, Log Avg loss: 0.93593194, Global Avg Loss: 2.19356432, Time: 0.0120 Steps: 36830, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000647, Sample Num: 10352, Cur Loss: 1.20315194, Cur Avg Loss: 0.72471264, Log Avg loss: 0.74130200, Global Avg Loss: 2.19317012, Time: 0.0112 Steps: 36840, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000657, Sample Num: 10512, Cur Loss: 0.41620314, Cur Avg Loss: 0.72372006, Log Avg loss: 0.65950063, Global Avg Loss: 2.19275392, Time: 0.0123 Steps: 36850, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000667, Sample Num: 10672, Cur Loss: 1.19996858, Cur Avg Loss: 0.72673750, Log Avg loss: 0.92498289, Global Avg Loss: 2.19240998, Time: 0.0067 Steps: 36860, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000677, Sample Num: 10832, Cur Loss: 0.25853103, Cur Avg Loss: 0.72452361, Log Avg loss: 0.57685746, Global Avg Loss: 2.19197181, Time: 0.0110 Steps: 36870, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000687, Sample Num: 10992, Cur Loss: 0.83600223, Cur Avg Loss: 0.72301529, Log Avg loss: 0.62090205, Global Avg Loss: 2.19154581, Time: 0.0070 Steps: 36880, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000697, Sample Num: 11152, Cur Loss: 0.69482738, Cur Avg Loss: 0.72305411, Log Avg loss: 0.72572117, Global Avg Loss: 2.19114846, Time: 0.0089 Steps: 36890, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000707, Sample Num: 11312, Cur Loss: 1.16286623, Cur Avg Loss: 0.72138016, Log Avg loss: 0.60470532, Global Avg Loss: 2.19071853, Time: 0.0064 Steps: 36900, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000717, Sample Num: 11472, Cur Loss: 0.53926498, Cur Avg Loss: 0.72195113, Log Avg loss: 0.76231887, Global Avg Loss: 2.19033153, Time: 0.0118 Steps: 36910, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000727, Sample Num: 11632, Cur Loss: 0.79648232, Cur Avg Loss: 0.72207326, Log Avg loss: 0.73082975, Global Avg Loss: 2.18993622, Time: 0.0110 Steps: 36920, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000737, Sample Num: 11792, Cur Loss: 1.11766994, Cur Avg Loss: 0.72162182, Log Avg loss: 0.68880226, Global Avg Loss: 2.18952974, Time: 0.0080 Steps: 36930, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000747, Sample Num: 11952, Cur Loss: 0.47090903, Cur Avg Loss: 0.72027363, Log Avg loss: 0.62091226, Global Avg Loss: 2.18910510, Time: 0.0109 Steps: 36940, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000757, Sample Num: 12112, Cur Loss: 0.80100477, Cur Avg Loss: 0.71818927, Log Avg loss: 0.56248720, Global Avg Loss: 2.18866488, Time: 0.0110 Steps: 36950, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000767, Sample Num: 12272, Cur Loss: 1.17580009, Cur Avg Loss: 0.72166129, Log Avg loss: 0.98449337, Global Avg Loss: 2.18833907, Time: 0.0104 Steps: 36960, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000777, Sample Num: 12432, Cur Loss: 0.62811393, Cur Avg Loss: 0.72182236, Log Avg loss: 0.73417681, Global Avg Loss: 2.18794574, Time: 0.0069 Steps: 36970, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000787, Sample Num: 12592, Cur Loss: 0.85858536, Cur Avg Loss: 0.72104731, Log Avg loss: 0.66082550, Global Avg Loss: 2.18753278, Time: 0.0067 Steps: 36980, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000797, Sample Num: 12752, Cur Loss: 0.57262409, Cur Avg Loss: 0.72313885, Log Avg loss: 0.88774357, Global Avg Loss: 2.18718139, Time: 0.0067 Steps: 36990, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000807, Sample Num: 12912, Cur Loss: 1.23520184, Cur Avg Loss: 0.72364760, Log Avg loss: 0.76419448, Global Avg Loss: 2.18679680, Time: 0.0073 Steps: 37000, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000817, Sample Num: 13072, Cur Loss: 0.78619230, Cur Avg Loss: 0.72464529, Log Avg loss: 0.80515925, Global Avg Loss: 2.18642348, Time: 0.0126 Steps: 37010, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000827, Sample Num: 13232, Cur Loss: 0.65933943, Cur Avg Loss: 0.72345237, Log Avg loss: 0.62599032, Global Avg Loss: 2.18600197, Time: 0.0070 Steps: 37020, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000837, Sample Num: 13392, Cur Loss: 1.41547287, Cur Avg Loss: 0.72485650, Log Avg loss: 0.84097858, Global Avg Loss: 2.18563875, Time: 0.0123 Steps: 37030, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000847, Sample Num: 13552, Cur Loss: 0.90850824, Cur Avg Loss: 0.72356936, Log Avg loss: 0.61583540, Global Avg Loss: 2.18521494, Time: 0.0109 Steps: 37040, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000857, Sample Num: 13712, Cur Loss: 1.15371370, Cur Avg Loss: 0.72337578, Log Avg loss: 0.70697958, Global Avg Loss: 2.18481595, Time: 0.0123 Steps: 37050, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000867, Sample Num: 13872, Cur Loss: 0.56847805, Cur Avg Loss: 0.72545473, Log Avg loss: 0.90362094, Global Avg Loss: 2.18447024, Time: 0.0113 Steps: 37060, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000877, Sample Num: 14032, Cur Loss: 1.02605057, Cur Avg Loss: 0.72588911, Log Avg loss: 0.76355006, Global Avg Loss: 2.18408694, Time: 0.0118 Steps: 37070, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000887, Sample Num: 14192, Cur Loss: 0.91080856, Cur Avg Loss: 0.72785462, Log Avg loss: 0.90022970, Global Avg Loss: 2.18374070, Time: 0.0108 Steps: 37080, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000897, Sample Num: 14352, Cur Loss: 1.31420302, Cur Avg Loss: 0.72833279, Log Avg loss: 0.77074644, Global Avg Loss: 2.18335973, Time: 0.0108 Steps: 37090, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000907, Sample Num: 14512, Cur Loss: 0.69519466, Cur Avg Loss: 0.72912887, Log Avg loss: 0.80053664, Global Avg Loss: 2.18298700, Time: 0.0151 Steps: 37100, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000917, Sample Num: 14672, Cur Loss: 0.38494313, Cur Avg Loss: 0.72725776, Log Avg loss: 0.55754867, Global Avg Loss: 2.18254900, Time: 0.0068 Steps: 37110, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000927, Sample Num: 14832, Cur Loss: 0.78752041, Cur Avg Loss: 0.72740131, Log Avg loss: 0.74056434, Global Avg Loss: 2.18216053, Time: 0.0155 Steps: 37120, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000937, Sample Num: 14992, Cur Loss: 0.47117370, Cur Avg Loss: 0.72772118, Log Avg loss: 0.75737325, Global Avg Loss: 2.18177680, Time: 0.0118 Steps: 37130, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000947, Sample Num: 15152, Cur Loss: 0.73571384, Cur Avg Loss: 0.72760169, Log Avg loss: 0.71640595, Global Avg Loss: 2.18138225, Time: 0.0117 Steps: 37140, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000957, Sample Num: 15312, Cur Loss: 0.92911774, Cur Avg Loss: 0.73006895, Log Avg loss: 0.96371791, Global Avg Loss: 2.18105448, Time: 0.0104 Steps: 37150, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000967, Sample Num: 15472, Cur Loss: 0.73223376, Cur Avg Loss: 0.73346893, Log Avg loss: 1.05884691, Global Avg Loss: 2.18075249, Time: 0.0071 Steps: 37160, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000977, Sample Num: 15632, Cur Loss: 0.43361774, Cur Avg Loss: 0.73375963, Log Avg loss: 0.76187118, Global Avg Loss: 2.18037076, Time: 0.0107 Steps: 37170, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000987, Sample Num: 15792, Cur Loss: 0.78249127, Cur Avg Loss: 0.73317718, Log Avg loss: 0.67627113, Global Avg Loss: 2.17996621, Time: 0.0074 Steps: 37180, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 000997, Sample Num: 15952, Cur Loss: 0.32084781, Cur Avg Loss: 0.73205582, Log Avg loss: 0.62137767, Global Avg Loss: 2.17954713, Time: 0.0068 Steps: 37190, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001007, Sample Num: 16112, Cur Loss: 0.71635675, Cur Avg Loss: 0.73469633, Log Avg loss: 0.99795541, Global Avg Loss: 2.17922949, Time: 0.0105 Steps: 37200, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001017, Sample Num: 16272, Cur Loss: 0.34853947, Cur Avg Loss: 0.73628646, Log Avg loss: 0.89641279, Global Avg Loss: 2.17888474, Time: 0.0107 Steps: 37210, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001027, Sample Num: 16432, Cur Loss: 0.86702222, Cur Avg Loss: 0.73615411, Log Avg loss: 0.72269335, Global Avg Loss: 2.17849350, Time: 0.0222 Steps: 37220, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001037, Sample Num: 16592, Cur Loss: 1.32965946, Cur Avg Loss: 0.73526978, Log Avg loss: 0.64444918, Global Avg Loss: 2.17808146, Time: 0.0071 Steps: 37230, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001047, Sample Num: 16752, Cur Loss: 0.61366415, Cur Avg Loss: 0.73462803, Log Avg loss: 0.66807869, Global Avg Loss: 2.17767598, Time: 0.0072 Steps: 37240, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001057, Sample Num: 16912, Cur Loss: 0.86354351, Cur Avg Loss: 0.73387014, Log Avg loss: 0.65451945, Global Avg Loss: 2.17726708, Time: 0.0065 Steps: 37250, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001067, Sample Num: 17072, Cur Loss: 0.67106473, Cur Avg Loss: 0.73297995, Log Avg loss: 0.63888685, Global Avg Loss: 2.17685420, Time: 0.0074 Steps: 37260, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001077, Sample Num: 17232, Cur Loss: 0.52601451, Cur Avg Loss: 0.73251368, Log Avg loss: 0.68276299, Global Avg Loss: 2.17645332, Time: 0.0112 Steps: 37270, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001087, Sample Num: 17392, Cur Loss: 0.60735720, Cur Avg Loss: 0.73443569, Log Avg loss: 0.94143599, Global Avg Loss: 2.17612204, Time: 0.0159 Steps: 37280, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001097, Sample Num: 17552, Cur Loss: 1.04763865, Cur Avg Loss: 0.73444323, Log Avg loss: 0.73526288, Global Avg Loss: 2.17573564, Time: 0.0108 Steps: 37290, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001107, Sample Num: 17712, Cur Loss: 1.18384302, Cur Avg Loss: 0.73617023, Log Avg loss: 0.92562203, Global Avg Loss: 2.17540049, Time: 0.0067 Steps: 37300, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001117, Sample Num: 17872, Cur Loss: 0.57661116, Cur Avg Loss: 0.73623249, Log Avg loss: 0.74312438, Global Avg Loss: 2.17501661, Time: 0.0073 Steps: 37310, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001127, Sample Num: 18032, Cur Loss: 1.43183565, Cur Avg Loss: 0.73722238, Log Avg loss: 0.84779324, Global Avg Loss: 2.17466098, Time: 0.0120 Steps: 37320, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001137, Sample Num: 18192, Cur Loss: 0.48821759, Cur Avg Loss: 0.73731968, Log Avg loss: 0.74828559, Global Avg Loss: 2.17427888, Time: 0.0066 Steps: 37330, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001147, Sample Num: 18352, Cur Loss: 0.19828558, Cur Avg Loss: 0.73567757, Log Avg loss: 0.54896958, Global Avg Loss: 2.17384360, Time: 0.0166 Steps: 37340, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001157, Sample Num: 18512, Cur Loss: 2.26403356, Cur Avg Loss: 0.73650264, Log Avg loss: 0.83113768, Global Avg Loss: 2.17348411, Time: 0.0085 Steps: 37350, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001167, Sample Num: 18672, Cur Loss: 1.63983226, Cur Avg Loss: 0.73745305, Log Avg loss: 0.84741616, Global Avg Loss: 2.17312917, Time: 0.0067 Steps: 37360, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001177, Sample Num: 18832, Cur Loss: 0.88467574, Cur Avg Loss: 0.73734088, Log Avg loss: 0.72424991, Global Avg Loss: 2.17274146, Time: 0.0070 Steps: 37370, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001187, Sample Num: 18992, Cur Loss: 0.83247060, Cur Avg Loss: 0.73931801, Log Avg loss: 0.97202706, Global Avg Loss: 2.17242024, Time: 0.0131 Steps: 37380, Updated lr: 0.000066 Training, Epoch: 0018, Batch: 001197, Sample Num: 19152, Cur Loss: 0.74851203, Cur Avg Loss: 0.74066712, Log Avg loss: 0.90080612, Global Avg Loss: 2.17208014, Time: 0.0123 Steps: 37390, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001207, Sample Num: 19312, Cur Loss: 0.38150871, Cur Avg Loss: 0.73942517, Log Avg loss: 0.59076303, Global Avg Loss: 2.17165733, Time: 0.0112 Steps: 37400, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001217, Sample Num: 19472, Cur Loss: 0.18868148, Cur Avg Loss: 0.73842001, Log Avg loss: 0.61709762, Global Avg Loss: 2.17124178, Time: 0.0112 Steps: 37410, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001227, Sample Num: 19632, Cur Loss: 0.54031295, Cur Avg Loss: 0.73907610, Log Avg loss: 0.81892259, Global Avg Loss: 2.17088039, Time: 0.0114 Steps: 37420, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001237, Sample Num: 19792, Cur Loss: 0.63956964, Cur Avg Loss: 0.73851619, Log Avg loss: 0.66981559, Global Avg Loss: 2.17047936, Time: 0.0074 Steps: 37430, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001247, Sample Num: 19952, Cur Loss: 0.68167657, Cur Avg Loss: 0.73949912, Log Avg loss: 0.86108695, Global Avg Loss: 2.17012963, Time: 0.0064 Steps: 37440, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001257, Sample Num: 20112, Cur Loss: 0.41822481, Cur Avg Loss: 0.73857524, Log Avg loss: 0.62336687, Global Avg Loss: 2.16971661, Time: 0.0148 Steps: 37450, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001267, Sample Num: 20272, Cur Loss: 0.34358567, Cur Avg Loss: 0.73650359, Log Avg loss: 0.47609804, Global Avg Loss: 2.16926450, Time: 0.0120 Steps: 37460, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001277, Sample Num: 20432, Cur Loss: 1.09213698, Cur Avg Loss: 0.73588900, Log Avg loss: 0.65801961, Global Avg Loss: 2.16886117, Time: 0.0068 Steps: 37470, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001287, Sample Num: 20592, Cur Loss: 0.62743509, Cur Avg Loss: 0.73683539, Log Avg loss: 0.85768969, Global Avg Loss: 2.16851134, Time: 0.0147 Steps: 37480, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001297, Sample Num: 20752, Cur Loss: 0.53854269, Cur Avg Loss: 0.73623831, Log Avg loss: 0.65939438, Global Avg Loss: 2.16810880, Time: 0.0135 Steps: 37490, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001307, Sample Num: 20912, Cur Loss: 0.59184206, Cur Avg Loss: 0.73796844, Log Avg loss: 0.96236572, Global Avg Loss: 2.16778727, Time: 0.0067 Steps: 37500, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001317, Sample Num: 21072, Cur Loss: 0.55708581, Cur Avg Loss: 0.73824094, Log Avg loss: 0.77385726, Global Avg Loss: 2.16741566, Time: 0.0087 Steps: 37510, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001327, Sample Num: 21232, Cur Loss: 2.62506866, Cur Avg Loss: 0.73896672, Log Avg loss: 0.83455165, Global Avg Loss: 2.16706042, Time: 0.0133 Steps: 37520, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001337, Sample Num: 21392, Cur Loss: 0.60764408, Cur Avg Loss: 0.73841198, Log Avg loss: 0.66479865, Global Avg Loss: 2.16666013, Time: 0.0092 Steps: 37530, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001347, Sample Num: 21552, Cur Loss: 0.70268261, Cur Avg Loss: 0.73908218, Log Avg loss: 0.82868798, Global Avg Loss: 2.16630372, Time: 0.0117 Steps: 37540, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001357, Sample Num: 21712, Cur Loss: 0.80217195, Cur Avg Loss: 0.73771099, Log Avg loss: 0.55301122, Global Avg Loss: 2.16587408, Time: 0.0073 Steps: 37550, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001367, Sample Num: 21872, Cur Loss: 0.61571157, Cur Avg Loss: 0.73843535, Log Avg loss: 0.83673093, Global Avg Loss: 2.16552021, Time: 0.0096 Steps: 37560, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001377, Sample Num: 22032, Cur Loss: 0.30073547, Cur Avg Loss: 0.73845919, Log Avg loss: 0.74171794, Global Avg Loss: 2.16514124, Time: 0.0120 Steps: 37570, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001387, Sample Num: 22192, Cur Loss: 0.50445318, Cur Avg Loss: 0.73889564, Log Avg loss: 0.79899538, Global Avg Loss: 2.16477771, Time: 0.0067 Steps: 37580, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001397, Sample Num: 22352, Cur Loss: 0.20003749, Cur Avg Loss: 0.73989791, Log Avg loss: 0.87891202, Global Avg Loss: 2.16443563, Time: 0.0118 Steps: 37590, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001407, Sample Num: 22512, Cur Loss: 0.23861419, Cur Avg Loss: 0.73838833, Log Avg loss: 0.52750074, Global Avg Loss: 2.16400028, Time: 0.0067 Steps: 37600, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001417, Sample Num: 22672, Cur Loss: 1.32392609, Cur Avg Loss: 0.73858460, Log Avg loss: 0.76619926, Global Avg Loss: 2.16362862, Time: 0.0086 Steps: 37610, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001427, Sample Num: 22832, Cur Loss: 0.23557152, Cur Avg Loss: 0.73766621, Log Avg loss: 0.60753056, Global Avg Loss: 2.16321498, Time: 0.0145 Steps: 37620, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001437, Sample Num: 22992, Cur Loss: 0.57176512, Cur Avg Loss: 0.73767613, Log Avg loss: 0.73909242, Global Avg Loss: 2.16283653, Time: 0.0067 Steps: 37630, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001447, Sample Num: 23152, Cur Loss: 0.97995263, Cur Avg Loss: 0.73641758, Log Avg loss: 0.55556393, Global Avg Loss: 2.16240952, Time: 0.0067 Steps: 37640, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001457, Sample Num: 23312, Cur Loss: 0.54486525, Cur Avg Loss: 0.73557916, Log Avg loss: 0.61425878, Global Avg Loss: 2.16199832, Time: 0.0107 Steps: 37650, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001467, Sample Num: 23472, Cur Loss: 0.43928969, Cur Avg Loss: 0.73475478, Log Avg loss: 0.61464326, Global Avg Loss: 2.16158745, Time: 0.0105 Steps: 37660, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001477, Sample Num: 23632, Cur Loss: 0.95367134, Cur Avg Loss: 0.73622784, Log Avg loss: 0.95232592, Global Avg Loss: 2.16126643, Time: 0.0155 Steps: 37670, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001487, Sample Num: 23792, Cur Loss: 0.28662571, Cur Avg Loss: 0.73672436, Log Avg loss: 0.81005946, Global Avg Loss: 2.16090783, Time: 0.0065 Steps: 37680, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001497, Sample Num: 23952, Cur Loss: 0.50210106, Cur Avg Loss: 0.73592872, Log Avg loss: 0.61761805, Global Avg Loss: 2.16049836, Time: 0.0106 Steps: 37690, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001507, Sample Num: 24112, Cur Loss: 0.48779422, Cur Avg Loss: 0.73503757, Log Avg loss: 0.60163245, Global Avg Loss: 2.16008487, Time: 0.0114 Steps: 37700, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001517, Sample Num: 24272, Cur Loss: 0.69805300, Cur Avg Loss: 0.73419847, Log Avg loss: 0.60774557, Global Avg Loss: 2.15967322, Time: 0.0073 Steps: 37710, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001527, Sample Num: 24432, Cur Loss: 2.28331709, Cur Avg Loss: 0.73584386, Log Avg loss: 0.98544895, Global Avg Loss: 2.15936192, Time: 0.0132 Steps: 37720, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001537, Sample Num: 24592, Cur Loss: 0.34806174, Cur Avg Loss: 0.73589948, Log Avg loss: 0.74439380, Global Avg Loss: 2.15898689, Time: 0.0196 Steps: 37730, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001547, Sample Num: 24752, Cur Loss: 0.38536069, Cur Avg Loss: 0.73631423, Log Avg loss: 0.80006014, Global Avg Loss: 2.15862682, Time: 0.0064 Steps: 37740, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001557, Sample Num: 24912, Cur Loss: 0.42877674, Cur Avg Loss: 0.73590377, Log Avg loss: 0.67240628, Global Avg Loss: 2.15823312, Time: 0.0067 Steps: 37750, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001567, Sample Num: 25072, Cur Loss: 0.63193893, Cur Avg Loss: 0.73521498, Log Avg loss: 0.62796956, Global Avg Loss: 2.15782786, Time: 0.0236 Steps: 37760, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001577, Sample Num: 25232, Cur Loss: 0.94582337, Cur Avg Loss: 0.73539671, Log Avg loss: 0.76387403, Global Avg Loss: 2.15745879, Time: 0.0182 Steps: 37770, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001587, Sample Num: 25392, Cur Loss: 0.28061128, Cur Avg Loss: 0.73450505, Log Avg loss: 0.59389008, Global Avg Loss: 2.15704493, Time: 0.0178 Steps: 37780, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001597, Sample Num: 25552, Cur Loss: 1.99591088, Cur Avg Loss: 0.73470530, Log Avg loss: 0.76648575, Global Avg Loss: 2.15667696, Time: 0.0221 Steps: 37790, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001607, Sample Num: 25712, Cur Loss: 0.34438094, Cur Avg Loss: 0.73470584, Log Avg loss: 0.73479219, Global Avg Loss: 2.15630080, Time: 0.0189 Steps: 37800, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001617, Sample Num: 25872, Cur Loss: 0.42294854, Cur Avg Loss: 0.73424733, Log Avg loss: 0.66056465, Global Avg Loss: 2.15590521, Time: 0.0064 Steps: 37810, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001627, Sample Num: 26032, Cur Loss: 0.31354377, Cur Avg Loss: 0.73340087, Log Avg loss: 0.59652757, Global Avg Loss: 2.15549289, Time: 0.0063 Steps: 37820, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001637, Sample Num: 26192, Cur Loss: 0.91365755, Cur Avg Loss: 0.73299969, Log Avg loss: 0.66772898, Global Avg Loss: 2.15509962, Time: 0.0066 Steps: 37830, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001647, Sample Num: 26352, Cur Loss: 1.47831643, Cur Avg Loss: 0.73339122, Log Avg loss: 0.79748375, Global Avg Loss: 2.15474084, Time: 0.0072 Steps: 37840, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001657, Sample Num: 26512, Cur Loss: 0.49963060, Cur Avg Loss: 0.73523388, Log Avg loss: 1.03872005, Global Avg Loss: 2.15444598, Time: 0.0065 Steps: 37850, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001667, Sample Num: 26672, Cur Loss: 0.99129510, Cur Avg Loss: 0.73434294, Log Avg loss: 0.58671394, Global Avg Loss: 2.15403190, Time: 0.0067 Steps: 37860, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001677, Sample Num: 26832, Cur Loss: 0.19581127, Cur Avg Loss: 0.73302821, Log Avg loss: 0.51386369, Global Avg Loss: 2.15359879, Time: 0.0069 Steps: 37870, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001687, Sample Num: 26992, Cur Loss: 0.80523998, Cur Avg Loss: 0.73305744, Log Avg loss: 0.73795854, Global Avg Loss: 2.15322508, Time: 0.0063 Steps: 37880, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001697, Sample Num: 27152, Cur Loss: 0.73183906, Cur Avg Loss: 0.73300959, Log Avg loss: 0.72493766, Global Avg Loss: 2.15284812, Time: 0.0071 Steps: 37890, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001707, Sample Num: 27312, Cur Loss: 0.79566610, Cur Avg Loss: 0.73312691, Log Avg loss: 0.75303656, Global Avg Loss: 2.15247878, Time: 0.0064 Steps: 37900, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001717, Sample Num: 27472, Cur Loss: 0.55345839, Cur Avg Loss: 0.73238096, Log Avg loss: 0.60504730, Global Avg Loss: 2.15207059, Time: 0.0219 Steps: 37910, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001727, Sample Num: 27632, Cur Loss: 0.58984667, Cur Avg Loss: 0.73164020, Log Avg loss: 0.60445169, Global Avg Loss: 2.15166246, Time: 0.0114 Steps: 37920, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001737, Sample Num: 27792, Cur Loss: 0.88300157, Cur Avg Loss: 0.73177025, Log Avg loss: 0.75422983, Global Avg Loss: 2.15129404, Time: 0.0065 Steps: 37930, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001747, Sample Num: 27952, Cur Loss: 0.78048080, Cur Avg Loss: 0.73181723, Log Avg loss: 0.73997664, Global Avg Loss: 2.15092205, Time: 0.0119 Steps: 37940, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001757, Sample Num: 28112, Cur Loss: 0.81790060, Cur Avg Loss: 0.73154083, Log Avg loss: 0.68325447, Global Avg Loss: 2.15053532, Time: 0.0108 Steps: 37950, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001767, Sample Num: 28272, Cur Loss: 1.45980406, Cur Avg Loss: 0.73278134, Log Avg loss: 0.95073805, Global Avg Loss: 2.15021925, Time: 0.0098 Steps: 37960, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001777, Sample Num: 28432, Cur Loss: 0.49020156, Cur Avg Loss: 0.73399623, Log Avg loss: 0.94866787, Global Avg Loss: 2.14990280, Time: 0.0187 Steps: 37970, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001787, Sample Num: 28592, Cur Loss: 0.91823101, Cur Avg Loss: 0.73517552, Log Avg loss: 0.94473568, Global Avg Loss: 2.14958548, Time: 0.0074 Steps: 37980, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001797, Sample Num: 28752, Cur Loss: 0.58623272, Cur Avg Loss: 0.73559808, Log Avg loss: 0.81110905, Global Avg Loss: 2.14923316, Time: 0.0152 Steps: 37990, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001807, Sample Num: 28912, Cur Loss: 1.05976522, Cur Avg Loss: 0.73579976, Log Avg loss: 0.77204253, Global Avg Loss: 2.14887074, Time: 0.0116 Steps: 38000, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001817, Sample Num: 29072, Cur Loss: 0.89153057, Cur Avg Loss: 0.73508098, Log Avg loss: 0.60519676, Global Avg Loss: 2.14846462, Time: 0.0111 Steps: 38010, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001827, Sample Num: 29232, Cur Loss: 0.52048886, Cur Avg Loss: 0.73452233, Log Avg loss: 0.63301531, Global Avg Loss: 2.14806602, Time: 0.0113 Steps: 38020, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001837, Sample Num: 29392, Cur Loss: 0.81620514, Cur Avg Loss: 0.73571182, Log Avg loss: 0.95303128, Global Avg Loss: 2.14775179, Time: 0.0111 Steps: 38030, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001847, Sample Num: 29552, Cur Loss: 0.40236127, Cur Avg Loss: 0.73537466, Log Avg loss: 0.67343883, Global Avg Loss: 2.14736422, Time: 0.0072 Steps: 38040, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001857, Sample Num: 29712, Cur Loss: 0.94233841, Cur Avg Loss: 0.73495251, Log Avg loss: 0.65698158, Global Avg Loss: 2.14697253, Time: 0.0067 Steps: 38050, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001867, Sample Num: 29872, Cur Loss: 0.43626493, Cur Avg Loss: 0.73459299, Log Avg loss: 0.66783063, Global Avg Loss: 2.14658390, Time: 0.0087 Steps: 38060, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001877, Sample Num: 30032, Cur Loss: 0.57337630, Cur Avg Loss: 0.73545782, Log Avg loss: 0.89692110, Global Avg Loss: 2.14625564, Time: 0.0137 Steps: 38070, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001887, Sample Num: 30192, Cur Loss: 0.59300923, Cur Avg Loss: 0.73590843, Log Avg loss: 0.82048857, Global Avg Loss: 2.14590749, Time: 0.0068 Steps: 38080, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001897, Sample Num: 30352, Cur Loss: 0.90377384, Cur Avg Loss: 0.73613727, Log Avg loss: 0.77931958, Global Avg Loss: 2.14554871, Time: 0.0116 Steps: 38090, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001907, Sample Num: 30512, Cur Loss: 0.75812662, Cur Avg Loss: 0.73656552, Log Avg loss: 0.81780462, Global Avg Loss: 2.14520022, Time: 0.0236 Steps: 38100, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001917, Sample Num: 30672, Cur Loss: 0.60922599, Cur Avg Loss: 0.73666272, Log Avg loss: 0.75519803, Global Avg Loss: 2.14483549, Time: 0.0124 Steps: 38110, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001927, Sample Num: 30832, Cur Loss: 1.15512240, Cur Avg Loss: 0.73658511, Log Avg loss: 0.72170640, Global Avg Loss: 2.14446216, Time: 0.0153 Steps: 38120, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001937, Sample Num: 30992, Cur Loss: 0.22665018, Cur Avg Loss: 0.73677863, Log Avg loss: 0.77407018, Global Avg Loss: 2.14410276, Time: 0.0064 Steps: 38130, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001947, Sample Num: 31152, Cur Loss: 0.58436114, Cur Avg Loss: 0.73627512, Log Avg loss: 0.63874491, Global Avg Loss: 2.14370807, Time: 0.0129 Steps: 38140, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001957, Sample Num: 31312, Cur Loss: 0.44391590, Cur Avg Loss: 0.73647883, Log Avg loss: 0.77614227, Global Avg Loss: 2.14334959, Time: 0.0232 Steps: 38150, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001967, Sample Num: 31472, Cur Loss: 0.51261508, Cur Avg Loss: 0.73759449, Log Avg loss: 0.95592929, Global Avg Loss: 2.14303843, Time: 0.0125 Steps: 38160, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001977, Sample Num: 31632, Cur Loss: 0.44855508, Cur Avg Loss: 0.73715710, Log Avg loss: 0.65112167, Global Avg Loss: 2.14264756, Time: 0.0068 Steps: 38170, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001987, Sample Num: 31792, Cur Loss: 0.66733849, Cur Avg Loss: 0.73679458, Log Avg loss: 0.66512393, Global Avg Loss: 2.14226058, Time: 0.0066 Steps: 38180, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 001997, Sample Num: 31952, Cur Loss: 0.57920635, Cur Avg Loss: 0.73616501, Log Avg loss: 0.61107116, Global Avg Loss: 2.14185964, Time: 0.0067 Steps: 38190, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002007, Sample Num: 32112, Cur Loss: 0.78685182, Cur Avg Loss: 0.73567558, Log Avg loss: 0.63793638, Global Avg Loss: 2.14146594, Time: 0.0071 Steps: 38200, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002017, Sample Num: 32272, Cur Loss: 0.51420975, Cur Avg Loss: 0.73500262, Log Avg loss: 0.59993922, Global Avg Loss: 2.14106250, Time: 0.0108 Steps: 38210, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002027, Sample Num: 32432, Cur Loss: 0.60787535, Cur Avg Loss: 0.73504596, Log Avg loss: 0.74378754, Global Avg Loss: 2.14069692, Time: 0.0066 Steps: 38220, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002037, Sample Num: 32592, Cur Loss: 1.17241442, Cur Avg Loss: 0.73472551, Log Avg loss: 0.66976976, Global Avg Loss: 2.14031216, Time: 0.0080 Steps: 38230, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002047, Sample Num: 32752, Cur Loss: 1.02446401, Cur Avg Loss: 0.73521625, Log Avg loss: 0.83518060, Global Avg Loss: 2.13997086, Time: 0.0122 Steps: 38240, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002057, Sample Num: 32912, Cur Loss: 0.61010313, Cur Avg Loss: 0.73444818, Log Avg loss: 0.57722470, Global Avg Loss: 2.13956230, Time: 0.0108 Steps: 38250, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002067, Sample Num: 33072, Cur Loss: 0.75685823, Cur Avg Loss: 0.73422147, Log Avg loss: 0.68758673, Global Avg Loss: 2.13918280, Time: 0.0115 Steps: 38260, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002077, Sample Num: 33232, Cur Loss: 0.87247515, Cur Avg Loss: 0.73373045, Log Avg loss: 0.63223573, Global Avg Loss: 2.13878903, Time: 0.0070 Steps: 38270, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002087, Sample Num: 33392, Cur Loss: 1.80379486, Cur Avg Loss: 0.73420838, Log Avg loss: 0.83347394, Global Avg Loss: 2.13844804, Time: 0.0064 Steps: 38280, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002097, Sample Num: 33552, Cur Loss: 0.84023839, Cur Avg Loss: 0.73274844, Log Avg loss: 0.42805911, Global Avg Loss: 2.13800134, Time: 0.0124 Steps: 38290, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002107, Sample Num: 33712, Cur Loss: 1.92732120, Cur Avg Loss: 0.73252199, Log Avg loss: 0.68503538, Global Avg Loss: 2.13762198, Time: 0.0069 Steps: 38300, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002117, Sample Num: 33872, Cur Loss: 0.25333673, Cur Avg Loss: 0.73294179, Log Avg loss: 0.82139482, Global Avg Loss: 2.13727841, Time: 0.0220 Steps: 38310, Updated lr: 0.000065 Training, Epoch: 0018, Batch: 002127, Sample Num: 34032, Cur Loss: 0.54821891, Cur Avg Loss: 0.73250916, Log Avg loss: 0.64092030, Global Avg Loss: 2.13688792, Time: 0.0103 Steps: 38320, Updated lr: 0.000065 ***** Running evaluation checkpoint-38322 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-38322 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.158927, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.802639, "eval_total_loss": 564.255096, "eval_mae": 0.710131, "eval_mse": 0.802821, "eval_r2": 0.489674, "eval_sp_statistic": 0.666586, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.711007, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.127228, "test_total_loss": 565.868468, "test_mae": 0.895907, "test_mse": 1.127275, "test_r2": 0.272447, "test_sp_statistic": 0.472221, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.586706, "test_ps_pvalue": 0.0, "lr": 6.460692271218587e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.136795612913074, "train_cur_epoch_loss": 1558.7835098579526, "train_cur_epoch_avg_loss": 0.732166984433045, "train_cur_epoch_time": 22.158926963806152, "train_cur_epoch_avg_time": 0.010408138545705098, "epoch": 18, "step": 38322} ################################################## Training, Epoch: 0019, Batch: 000008, Sample Num: 128, Cur Loss: 0.75766259, Cur Avg Loss: 0.81963138, Log Avg loss: 0.72935851, Global Avg Loss: 2.13652070, Time: 0.0141 Steps: 38330, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000018, Sample Num: 288, Cur Loss: 0.78195947, Cur Avg Loss: 0.86299607, Log Avg loss: 0.89768781, Global Avg Loss: 2.13619758, Time: 0.0066 Steps: 38340, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000028, Sample Num: 448, Cur Loss: 1.02312469, Cur Avg Loss: 0.87077060, Log Avg loss: 0.88476477, Global Avg Loss: 2.13587127, Time: 0.0068 Steps: 38350, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000038, Sample Num: 608, Cur Loss: 0.45839655, Cur Avg Loss: 0.84389350, Log Avg loss: 0.76863760, Global Avg Loss: 2.13551484, Time: 0.0164 Steps: 38360, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000048, Sample Num: 768, Cur Loss: 0.62987643, Cur Avg Loss: 0.80971583, Log Avg loss: 0.67984071, Global Avg Loss: 2.13513547, Time: 0.0126 Steps: 38370, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000058, Sample Num: 928, Cur Loss: 0.88809085, Cur Avg Loss: 0.83597078, Log Avg loss: 0.96199454, Global Avg Loss: 2.13482980, Time: 0.0096 Steps: 38380, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000068, Sample Num: 1088, Cur Loss: 0.46776375, Cur Avg Loss: 0.79242843, Log Avg loss: 0.53988280, Global Avg Loss: 2.13441434, Time: 0.0068 Steps: 38390, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000078, Sample Num: 1248, Cur Loss: 0.13806140, Cur Avg Loss: 0.76174443, Log Avg loss: 0.55309324, Global Avg Loss: 2.13400254, Time: 0.0117 Steps: 38400, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000088, Sample Num: 1408, Cur Loss: 0.46534812, Cur Avg Loss: 0.77554435, Log Avg loss: 0.88318372, Global Avg Loss: 2.13367689, Time: 0.0155 Steps: 38410, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000098, Sample Num: 1568, Cur Loss: 0.46998185, Cur Avg Loss: 0.75602217, Log Avg loss: 0.58422691, Global Avg Loss: 2.13327360, Time: 0.0121 Steps: 38420, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000108, Sample Num: 1728, Cur Loss: 0.83438277, Cur Avg Loss: 0.76634442, Log Avg loss: 0.86750252, Global Avg Loss: 2.13294423, Time: 0.0141 Steps: 38430, Updated lr: 0.000065 Training, Epoch: 0019, Batch: 000118, Sample Num: 1888, Cur Loss: 1.59771967, Cur Avg Loss: 0.76859628, Log Avg loss: 0.79291636, Global Avg Loss: 2.13259563, Time: 0.0064 Steps: 38440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000128, Sample Num: 2048, Cur Loss: 0.52371740, Cur Avg Loss: 0.76660625, Log Avg loss: 0.74312390, Global Avg Loss: 2.13223425, Time: 0.0125 Steps: 38450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000138, Sample Num: 2208, Cur Loss: 0.35301158, Cur Avg Loss: 0.75345915, Log Avg loss: 0.58517625, Global Avg Loss: 2.13183200, Time: 0.0134 Steps: 38460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000148, Sample Num: 2368, Cur Loss: 0.78747702, Cur Avg Loss: 0.76780072, Log Avg loss: 0.96571438, Global Avg Loss: 2.13152888, Time: 0.0162 Steps: 38470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000158, Sample Num: 2528, Cur Loss: 0.96763277, Cur Avg Loss: 0.77096579, Log Avg loss: 0.81780879, Global Avg Loss: 2.13118748, Time: 0.0116 Steps: 38480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000168, Sample Num: 2688, Cur Loss: 0.58141083, Cur Avg Loss: 0.77607279, Log Avg loss: 0.85676339, Global Avg Loss: 2.13085637, Time: 0.0114 Steps: 38490, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000178, Sample Num: 2848, Cur Loss: 1.43316829, Cur Avg Loss: 0.77466696, Log Avg loss: 0.75104908, Global Avg Loss: 2.13049798, Time: 0.0072 Steps: 38500, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000188, Sample Num: 3008, Cur Loss: 0.31766397, Cur Avg Loss: 0.76872251, Log Avg loss: 0.66291138, Global Avg Loss: 2.13011689, Time: 0.0072 Steps: 38510, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000198, Sample Num: 3168, Cur Loss: 0.30631411, Cur Avg Loss: 0.76372201, Log Avg loss: 0.66971262, Global Avg Loss: 2.12973776, Time: 0.0115 Steps: 38520, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000208, Sample Num: 3328, Cur Loss: 0.71672714, Cur Avg Loss: 0.75333758, Log Avg loss: 0.54772586, Global Avg Loss: 2.12932717, Time: 0.0205 Steps: 38530, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000218, Sample Num: 3488, Cur Loss: 1.88712668, Cur Avg Loss: 0.76443946, Log Avg loss: 0.99535846, Global Avg Loss: 2.12903293, Time: 0.0139 Steps: 38540, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000228, Sample Num: 3648, Cur Loss: 0.28837717, Cur Avg Loss: 0.75837186, Log Avg loss: 0.62609810, Global Avg Loss: 2.12864307, Time: 0.0067 Steps: 38550, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000238, Sample Num: 3808, Cur Loss: 0.70095479, Cur Avg Loss: 0.75202293, Log Avg loss: 0.60726746, Global Avg Loss: 2.12824852, Time: 0.0120 Steps: 38560, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000248, Sample Num: 3968, Cur Loss: 0.67551929, Cur Avg Loss: 0.74666022, Log Avg loss: 0.61902760, Global Avg Loss: 2.12785723, Time: 0.0126 Steps: 38570, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000258, Sample Num: 4128, Cur Loss: 1.56276512, Cur Avg Loss: 0.74584983, Log Avg loss: 0.72575235, Global Avg Loss: 2.12749380, Time: 0.0133 Steps: 38580, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000268, Sample Num: 4288, Cur Loss: 0.91412866, Cur Avg Loss: 0.74725380, Log Avg loss: 0.78347625, Global Avg Loss: 2.12714552, Time: 0.0111 Steps: 38590, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000278, Sample Num: 4448, Cur Loss: 0.69399691, Cur Avg Loss: 0.74378962, Log Avg loss: 0.65094961, Global Avg Loss: 2.12676308, Time: 0.0143 Steps: 38600, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000288, Sample Num: 4608, Cur Loss: 0.25963312, Cur Avg Loss: 0.73947362, Log Avg loss: 0.61948866, Global Avg Loss: 2.12637270, Time: 0.0114 Steps: 38610, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000298, Sample Num: 4768, Cur Loss: 0.49271145, Cur Avg Loss: 0.74531919, Log Avg loss: 0.91367157, Global Avg Loss: 2.12605869, Time: 0.0068 Steps: 38620, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000308, Sample Num: 4928, Cur Loss: 0.72008669, Cur Avg Loss: 0.74414389, Log Avg loss: 0.70911990, Global Avg Loss: 2.12569189, Time: 0.0109 Steps: 38630, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000318, Sample Num: 5088, Cur Loss: 0.30343449, Cur Avg Loss: 0.74687716, Log Avg loss: 0.83106184, Global Avg Loss: 2.12535684, Time: 0.0104 Steps: 38640, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000328, Sample Num: 5248, Cur Loss: 0.62558115, Cur Avg Loss: 0.73607720, Log Avg loss: 0.39263876, Global Avg Loss: 2.12490853, Time: 0.0117 Steps: 38650, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000338, Sample Num: 5408, Cur Loss: 0.32755002, Cur Avg Loss: 0.73193195, Log Avg loss: 0.59596753, Global Avg Loss: 2.12451305, Time: 0.0068 Steps: 38660, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000348, Sample Num: 5568, Cur Loss: 0.53401589, Cur Avg Loss: 0.72758378, Log Avg loss: 0.58061587, Global Avg Loss: 2.12411380, Time: 0.0108 Steps: 38670, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000358, Sample Num: 5728, Cur Loss: 0.88186842, Cur Avg Loss: 0.72850520, Log Avg loss: 0.76057029, Global Avg Loss: 2.12376128, Time: 0.0115 Steps: 38680, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000368, Sample Num: 5888, Cur Loss: 0.22834721, Cur Avg Loss: 0.72534551, Log Avg loss: 0.61222863, Global Avg Loss: 2.12337060, Time: 0.0067 Steps: 38690, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000378, Sample Num: 6048, Cur Loss: 0.63796782, Cur Avg Loss: 0.72734547, Log Avg loss: 0.80094410, Global Avg Loss: 2.12302889, Time: 0.0228 Steps: 38700, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000388, Sample Num: 6208, Cur Loss: 0.83005643, Cur Avg Loss: 0.72659345, Log Avg loss: 0.69816725, Global Avg Loss: 2.12266080, Time: 0.0141 Steps: 38710, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000398, Sample Num: 6368, Cur Loss: 0.40432483, Cur Avg Loss: 0.72429187, Log Avg loss: 0.63499055, Global Avg Loss: 2.12227659, Time: 0.0139 Steps: 38720, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000408, Sample Num: 6528, Cur Loss: 0.95867527, Cur Avg Loss: 0.72622702, Log Avg loss: 0.80324603, Global Avg Loss: 2.12193602, Time: 0.0108 Steps: 38730, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000418, Sample Num: 6688, Cur Loss: 1.15973592, Cur Avg Loss: 0.72972818, Log Avg loss: 0.87257519, Global Avg Loss: 2.12161352, Time: 0.0067 Steps: 38740, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000428, Sample Num: 6848, Cur Loss: 0.83102590, Cur Avg Loss: 0.72648950, Log Avg loss: 0.59111302, Global Avg Loss: 2.12121855, Time: 0.0067 Steps: 38750, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000438, Sample Num: 7008, Cur Loss: 1.00076008, Cur Avg Loss: 0.72932889, Log Avg loss: 0.85085465, Global Avg Loss: 2.12089080, Time: 0.0123 Steps: 38760, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000448, Sample Num: 7168, Cur Loss: 0.94811416, Cur Avg Loss: 0.72885234, Log Avg loss: 0.70797949, Global Avg Loss: 2.12052637, Time: 0.0117 Steps: 38770, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000458, Sample Num: 7328, Cur Loss: 0.50715160, Cur Avg Loss: 0.72603786, Log Avg loss: 0.59994911, Global Avg Loss: 2.12013427, Time: 0.0107 Steps: 38780, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000468, Sample Num: 7488, Cur Loss: 1.02739704, Cur Avg Loss: 0.72842586, Log Avg loss: 0.83779628, Global Avg Loss: 2.11980368, Time: 0.0071 Steps: 38790, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000478, Sample Num: 7648, Cur Loss: 0.31363386, Cur Avg Loss: 0.72799237, Log Avg loss: 0.70770520, Global Avg Loss: 2.11943974, Time: 0.0125 Steps: 38800, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000488, Sample Num: 7808, Cur Loss: 0.61433423, Cur Avg Loss: 0.72868325, Log Avg loss: 0.76170720, Global Avg Loss: 2.11908990, Time: 0.0116 Steps: 38810, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000498, Sample Num: 7968, Cur Loss: 0.75338995, Cur Avg Loss: 0.72975511, Log Avg loss: 0.78206176, Global Avg Loss: 2.11874548, Time: 0.0109 Steps: 38820, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000508, Sample Num: 8128, Cur Loss: 0.29677224, Cur Avg Loss: 0.73081885, Log Avg loss: 0.78379321, Global Avg Loss: 2.11840169, Time: 0.0133 Steps: 38830, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000518, Sample Num: 8288, Cur Loss: 0.35248774, Cur Avg Loss: 0.72891579, Log Avg loss: 0.63224040, Global Avg Loss: 2.11801905, Time: 0.0077 Steps: 38840, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000528, Sample Num: 8448, Cur Loss: 0.34070283, Cur Avg Loss: 0.72637126, Log Avg loss: 0.59456431, Global Avg Loss: 2.11762691, Time: 0.0227 Steps: 38850, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000538, Sample Num: 8608, Cur Loss: 0.95771784, Cur Avg Loss: 0.72426573, Log Avg loss: 0.61309379, Global Avg Loss: 2.11723974, Time: 0.0124 Steps: 38860, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000548, Sample Num: 8768, Cur Loss: 0.52742994, Cur Avg Loss: 0.72321162, Log Avg loss: 0.66650077, Global Avg Loss: 2.11686652, Time: 0.0119 Steps: 38870, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000558, Sample Num: 8928, Cur Loss: 0.53163797, Cur Avg Loss: 0.72122700, Log Avg loss: 0.61246974, Global Avg Loss: 2.11647958, Time: 0.0148 Steps: 38880, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000568, Sample Num: 9088, Cur Loss: 0.61070281, Cur Avg Loss: 0.72248570, Log Avg loss: 0.79272115, Global Avg Loss: 2.11613920, Time: 0.0071 Steps: 38890, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000578, Sample Num: 9248, Cur Loss: 0.36482948, Cur Avg Loss: 0.71994743, Log Avg loss: 0.57577360, Global Avg Loss: 2.11574322, Time: 0.0067 Steps: 38900, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000588, Sample Num: 9408, Cur Loss: 0.79029298, Cur Avg Loss: 0.71720253, Log Avg loss: 0.55854748, Global Avg Loss: 2.11534301, Time: 0.0067 Steps: 38910, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000598, Sample Num: 9568, Cur Loss: 0.35257801, Cur Avg Loss: 0.71416947, Log Avg loss: 0.53582556, Global Avg Loss: 2.11493717, Time: 0.0107 Steps: 38920, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000608, Sample Num: 9728, Cur Loss: 0.45043725, Cur Avg Loss: 0.71380820, Log Avg loss: 0.69220419, Global Avg Loss: 2.11457171, Time: 0.0107 Steps: 38930, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000618, Sample Num: 9888, Cur Loss: 0.34295893, Cur Avg Loss: 0.71196312, Log Avg loss: 0.59978218, Global Avg Loss: 2.11418271, Time: 0.0067 Steps: 38940, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000628, Sample Num: 10048, Cur Loss: 0.42394489, Cur Avg Loss: 0.71218390, Log Avg loss: 0.72582782, Global Avg Loss: 2.11382626, Time: 0.0108 Steps: 38950, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000638, Sample Num: 10208, Cur Loss: 0.94692469, Cur Avg Loss: 0.71110501, Log Avg loss: 0.64335104, Global Avg Loss: 2.11344883, Time: 0.0132 Steps: 38960, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000648, Sample Num: 10368, Cur Loss: 0.56187779, Cur Avg Loss: 0.71009580, Log Avg loss: 0.64570833, Global Avg Loss: 2.11307220, Time: 0.0117 Steps: 38970, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000658, Sample Num: 10528, Cur Loss: 0.61841631, Cur Avg Loss: 0.70719567, Log Avg loss: 0.51926701, Global Avg Loss: 2.11266332, Time: 0.0120 Steps: 38980, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000668, Sample Num: 10688, Cur Loss: 0.86776495, Cur Avg Loss: 0.70686347, Log Avg loss: 0.68500466, Global Avg Loss: 2.11229716, Time: 0.0123 Steps: 38990, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000678, Sample Num: 10848, Cur Loss: 0.94143236, Cur Avg Loss: 0.70820885, Log Avg loss: 0.79808056, Global Avg Loss: 2.11196018, Time: 0.0138 Steps: 39000, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000688, Sample Num: 11008, Cur Loss: 0.72588313, Cur Avg Loss: 0.71029439, Log Avg loss: 0.85169407, Global Avg Loss: 2.11163712, Time: 0.0067 Steps: 39010, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000698, Sample Num: 11168, Cur Loss: 0.61556387, Cur Avg Loss: 0.71269450, Log Avg loss: 0.87782151, Global Avg Loss: 2.11132092, Time: 0.0068 Steps: 39020, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000708, Sample Num: 11328, Cur Loss: 0.70377058, Cur Avg Loss: 0.71380925, Log Avg loss: 0.79161885, Global Avg Loss: 2.11098279, Time: 0.0105 Steps: 39030, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000718, Sample Num: 11488, Cur Loss: 0.56799710, Cur Avg Loss: 0.71359811, Log Avg loss: 0.69864953, Global Avg Loss: 2.11062103, Time: 0.0066 Steps: 39040, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000728, Sample Num: 11648, Cur Loss: 0.45634648, Cur Avg Loss: 0.71473508, Log Avg loss: 0.79636964, Global Avg Loss: 2.11028447, Time: 0.0068 Steps: 39050, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000738, Sample Num: 11808, Cur Loss: 0.57481104, Cur Avg Loss: 0.71527309, Log Avg loss: 0.75443999, Global Avg Loss: 2.10993735, Time: 0.0067 Steps: 39060, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000748, Sample Num: 11968, Cur Loss: 0.37890396, Cur Avg Loss: 0.71277009, Log Avg loss: 0.52804898, Global Avg Loss: 2.10953247, Time: 0.0125 Steps: 39070, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000758, Sample Num: 12128, Cur Loss: 0.74940002, Cur Avg Loss: 0.71402000, Log Avg loss: 0.80751307, Global Avg Loss: 2.10919930, Time: 0.0114 Steps: 39080, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000768, Sample Num: 12288, Cur Loss: 0.85178858, Cur Avg Loss: 0.71283571, Log Avg loss: 0.62306665, Global Avg Loss: 2.10881912, Time: 0.0073 Steps: 39090, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000778, Sample Num: 12448, Cur Loss: 0.60404992, Cur Avg Loss: 0.71142205, Log Avg loss: 0.60285309, Global Avg Loss: 2.10843396, Time: 0.0108 Steps: 39100, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000788, Sample Num: 12608, Cur Loss: 1.08442390, Cur Avg Loss: 0.71154808, Log Avg loss: 0.72135270, Global Avg Loss: 2.10807930, Time: 0.0067 Steps: 39110, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000798, Sample Num: 12768, Cur Loss: 0.68165648, Cur Avg Loss: 0.71044021, Log Avg loss: 0.62314048, Global Avg Loss: 2.10769971, Time: 0.0067 Steps: 39120, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000808, Sample Num: 12928, Cur Loss: 1.79064989, Cur Avg Loss: 0.71229277, Log Avg loss: 0.86012705, Global Avg Loss: 2.10738089, Time: 0.0125 Steps: 39130, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000818, Sample Num: 13088, Cur Loss: 0.39986411, Cur Avg Loss: 0.71044273, Log Avg loss: 0.56095927, Global Avg Loss: 2.10698579, Time: 0.0116 Steps: 39140, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000828, Sample Num: 13248, Cur Loss: 0.94661963, Cur Avg Loss: 0.71056189, Log Avg loss: 0.72030925, Global Avg Loss: 2.10663159, Time: 0.0109 Steps: 39150, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000838, Sample Num: 13408, Cur Loss: 0.76616228, Cur Avg Loss: 0.71072313, Log Avg loss: 0.72407416, Global Avg Loss: 2.10627854, Time: 0.0073 Steps: 39160, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000848, Sample Num: 13568, Cur Loss: 0.48253575, Cur Avg Loss: 0.70997443, Log Avg loss: 0.64723277, Global Avg Loss: 2.10590605, Time: 0.0072 Steps: 39170, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000858, Sample Num: 13728, Cur Loss: 0.51201224, Cur Avg Loss: 0.70999049, Log Avg loss: 0.71135282, Global Avg Loss: 2.10555011, Time: 0.0123 Steps: 39180, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000868, Sample Num: 13888, Cur Loss: 1.50166893, Cur Avg Loss: 0.71246249, Log Avg loss: 0.92455989, Global Avg Loss: 2.10524876, Time: 0.0125 Steps: 39190, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000878, Sample Num: 14048, Cur Loss: 0.99041229, Cur Avg Loss: 0.71239499, Log Avg loss: 0.70653569, Global Avg Loss: 2.10489195, Time: 0.0111 Steps: 39200, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000888, Sample Num: 14208, Cur Loss: 1.48235941, Cur Avg Loss: 0.71295952, Log Avg loss: 0.76252541, Global Avg Loss: 2.10454959, Time: 0.0103 Steps: 39210, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000898, Sample Num: 14368, Cur Loss: 0.42405307, Cur Avg Loss: 0.71462005, Log Avg loss: 0.86207497, Global Avg Loss: 2.10423280, Time: 0.0127 Steps: 39220, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000908, Sample Num: 14528, Cur Loss: 1.20860839, Cur Avg Loss: 0.71750442, Log Avg loss: 0.97652150, Global Avg Loss: 2.10394534, Time: 0.0136 Steps: 39230, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000918, Sample Num: 14688, Cur Loss: 0.49305850, Cur Avg Loss: 0.71761043, Log Avg loss: 0.72723559, Global Avg Loss: 2.10359449, Time: 0.0121 Steps: 39240, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000928, Sample Num: 14848, Cur Loss: 0.79575825, Cur Avg Loss: 0.71777246, Log Avg loss: 0.73264702, Global Avg Loss: 2.10324521, Time: 0.0167 Steps: 39250, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000938, Sample Num: 15008, Cur Loss: 0.75328493, Cur Avg Loss: 0.71929876, Log Avg loss: 0.86093963, Global Avg Loss: 2.10292878, Time: 0.0144 Steps: 39260, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000948, Sample Num: 15168, Cur Loss: 0.64402574, Cur Avg Loss: 0.72073519, Log Avg loss: 0.85547190, Global Avg Loss: 2.10261111, Time: 0.0066 Steps: 39270, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000958, Sample Num: 15328, Cur Loss: 0.96135879, Cur Avg Loss: 0.72170975, Log Avg loss: 0.81409810, Global Avg Loss: 2.10228308, Time: 0.0108 Steps: 39280, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000968, Sample Num: 15488, Cur Loss: 0.89888537, Cur Avg Loss: 0.72204468, Log Avg loss: 0.75413163, Global Avg Loss: 2.10193995, Time: 0.0077 Steps: 39290, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000978, Sample Num: 15648, Cur Loss: 0.23778556, Cur Avg Loss: 0.72180031, Log Avg loss: 0.69814459, Global Avg Loss: 2.10158275, Time: 0.0107 Steps: 39300, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000988, Sample Num: 15808, Cur Loss: 1.35744345, Cur Avg Loss: 0.72199215, Log Avg loss: 0.74075408, Global Avg Loss: 2.10123657, Time: 0.0115 Steps: 39310, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 000998, Sample Num: 15968, Cur Loss: 0.67142099, Cur Avg Loss: 0.72229792, Log Avg loss: 0.75250845, Global Avg Loss: 2.10089356, Time: 0.0110 Steps: 39320, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001008, Sample Num: 16128, Cur Loss: 1.38359499, Cur Avg Loss: 0.72272907, Log Avg loss: 0.76575755, Global Avg Loss: 2.10055409, Time: 0.0110 Steps: 39330, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001018, Sample Num: 16288, Cur Loss: 0.43641549, Cur Avg Loss: 0.72147223, Log Avg loss: 0.59478305, Global Avg Loss: 2.10017133, Time: 0.0067 Steps: 39340, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001028, Sample Num: 16448, Cur Loss: 0.96220243, Cur Avg Loss: 0.72171032, Log Avg loss: 0.74594766, Global Avg Loss: 2.09982718, Time: 0.0109 Steps: 39350, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001038, Sample Num: 16608, Cur Loss: 0.95514417, Cur Avg Loss: 0.72265786, Log Avg loss: 0.82006493, Global Avg Loss: 2.09950204, Time: 0.0129 Steps: 39360, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001048, Sample Num: 16768, Cur Loss: 0.23986217, Cur Avg Loss: 0.72196580, Log Avg loss: 0.65013019, Global Avg Loss: 2.09913390, Time: 0.0066 Steps: 39370, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001058, Sample Num: 16928, Cur Loss: 1.54755569, Cur Avg Loss: 0.72225682, Log Avg loss: 0.75275567, Global Avg Loss: 2.09879201, Time: 0.0068 Steps: 39380, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001068, Sample Num: 17088, Cur Loss: 0.77595079, Cur Avg Loss: 0.72448276, Log Avg loss: 0.95998705, Global Avg Loss: 2.09850290, Time: 0.0086 Steps: 39390, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001078, Sample Num: 17248, Cur Loss: 0.36980698, Cur Avg Loss: 0.72445759, Log Avg loss: 0.72176939, Global Avg Loss: 2.09815347, Time: 0.0152 Steps: 39400, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001088, Sample Num: 17408, Cur Loss: 0.73596346, Cur Avg Loss: 0.72502494, Log Avg loss: 0.78618525, Global Avg Loss: 2.09782057, Time: 0.0237 Steps: 39410, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001098, Sample Num: 17568, Cur Loss: 0.82782519, Cur Avg Loss: 0.72620037, Log Avg loss: 0.85408750, Global Avg Loss: 2.09750506, Time: 0.0116 Steps: 39420, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001108, Sample Num: 17728, Cur Loss: 0.91474652, Cur Avg Loss: 0.72531014, Log Avg loss: 0.62756281, Global Avg Loss: 2.09713226, Time: 0.0112 Steps: 39430, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001118, Sample Num: 17888, Cur Loss: 0.41041699, Cur Avg Loss: 0.72411758, Log Avg loss: 0.59198134, Global Avg Loss: 2.09675063, Time: 0.0084 Steps: 39440, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001128, Sample Num: 18048, Cur Loss: 0.51760650, Cur Avg Loss: 0.72397081, Log Avg loss: 0.70756275, Global Avg Loss: 2.09639849, Time: 0.0113 Steps: 39450, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001138, Sample Num: 18208, Cur Loss: 0.94688064, Cur Avg Loss: 0.72349379, Log Avg loss: 0.66968534, Global Avg Loss: 2.09603693, Time: 0.0110 Steps: 39460, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001148, Sample Num: 18368, Cur Loss: 0.45312729, Cur Avg Loss: 0.72362108, Log Avg loss: 0.73810736, Global Avg Loss: 2.09569289, Time: 0.0136 Steps: 39470, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001158, Sample Num: 18528, Cur Loss: 0.72968990, Cur Avg Loss: 0.72345434, Log Avg loss: 0.70431200, Global Avg Loss: 2.09534047, Time: 0.0140 Steps: 39480, Updated lr: 0.000064 Training, Epoch: 0019, Batch: 001168, Sample Num: 18688, Cur Loss: 0.55801922, Cur Avg Loss: 0.72329306, Log Avg loss: 0.70461651, Global Avg Loss: 2.09498829, Time: 0.0066 Steps: 39490, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001178, Sample Num: 18848, Cur Loss: 0.97291565, Cur Avg Loss: 0.72298881, Log Avg loss: 0.68745291, Global Avg Loss: 2.09463196, Time: 0.0094 Steps: 39500, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001188, Sample Num: 19008, Cur Loss: 0.63553226, Cur Avg Loss: 0.72235380, Log Avg loss: 0.64754981, Global Avg Loss: 2.09426570, Time: 0.0119 Steps: 39510, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001198, Sample Num: 19168, Cur Loss: 0.37597451, Cur Avg Loss: 0.72253662, Log Avg loss: 0.74425585, Global Avg Loss: 2.09392410, Time: 0.0071 Steps: 39520, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001208, Sample Num: 19328, Cur Loss: 0.63639432, Cur Avg Loss: 0.72201957, Log Avg loss: 0.66007641, Global Avg Loss: 2.09356137, Time: 0.0089 Steps: 39530, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001218, Sample Num: 19488, Cur Loss: 1.28867483, Cur Avg Loss: 0.72286720, Log Avg loss: 0.82526083, Global Avg Loss: 2.09324061, Time: 0.0101 Steps: 39540, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001228, Sample Num: 19648, Cur Loss: 1.20644665, Cur Avg Loss: 0.72308744, Log Avg loss: 0.74991309, Global Avg Loss: 2.09290096, Time: 0.0067 Steps: 39550, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001238, Sample Num: 19808, Cur Loss: 0.23682322, Cur Avg Loss: 0.72025239, Log Avg loss: 0.37210803, Global Avg Loss: 2.09246597, Time: 0.0069 Steps: 39560, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001248, Sample Num: 19968, Cur Loss: 0.55335128, Cur Avg Loss: 0.72043655, Log Avg loss: 0.74323527, Global Avg Loss: 2.09212500, Time: 0.0102 Steps: 39570, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001258, Sample Num: 20128, Cur Loss: 0.78351820, Cur Avg Loss: 0.72148341, Log Avg loss: 0.85213245, Global Avg Loss: 2.09181171, Time: 0.0070 Steps: 39580, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001268, Sample Num: 20288, Cur Loss: 0.38891578, Cur Avg Loss: 0.72171345, Log Avg loss: 0.75065211, Global Avg Loss: 2.09147295, Time: 0.0092 Steps: 39590, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001278, Sample Num: 20448, Cur Loss: 0.87265706, Cur Avg Loss: 0.72143452, Log Avg loss: 0.68606558, Global Avg Loss: 2.09111805, Time: 0.0064 Steps: 39600, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001288, Sample Num: 20608, Cur Loss: 0.64188296, Cur Avg Loss: 0.72007233, Log Avg loss: 0.54598498, Global Avg Loss: 2.09072796, Time: 0.0121 Steps: 39610, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001298, Sample Num: 20768, Cur Loss: 0.35768014, Cur Avg Loss: 0.71859944, Log Avg loss: 0.52889081, Global Avg Loss: 2.09033376, Time: 0.0116 Steps: 39620, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001308, Sample Num: 20928, Cur Loss: 1.67748535, Cur Avg Loss: 0.71909759, Log Avg loss: 0.78375738, Global Avg Loss: 2.09000407, Time: 0.0067 Steps: 39630, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001318, Sample Num: 21088, Cur Loss: 0.24724291, Cur Avg Loss: 0.71934730, Log Avg loss: 0.75201000, Global Avg Loss: 2.08966653, Time: 0.0066 Steps: 39640, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001328, Sample Num: 21248, Cur Loss: 0.70418745, Cur Avg Loss: 0.71892020, Log Avg loss: 0.66262760, Global Avg Loss: 2.08930662, Time: 0.0120 Steps: 39650, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001338, Sample Num: 21408, Cur Loss: 1.00337315, Cur Avg Loss: 0.71959789, Log Avg loss: 0.80959580, Global Avg Loss: 2.08898395, Time: 0.0129 Steps: 39660, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001348, Sample Num: 21568, Cur Loss: 0.71648502, Cur Avg Loss: 0.71952790, Log Avg loss: 0.71016261, Global Avg Loss: 2.08863638, Time: 0.0131 Steps: 39670, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001358, Sample Num: 21728, Cur Loss: 0.18075441, Cur Avg Loss: 0.71825116, Log Avg loss: 0.54614656, Global Avg Loss: 2.08824764, Time: 0.0084 Steps: 39680, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001368, Sample Num: 21888, Cur Loss: 0.40983987, Cur Avg Loss: 0.71844200, Log Avg loss: 0.74435810, Global Avg Loss: 2.08790905, Time: 0.0090 Steps: 39690, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001378, Sample Num: 22048, Cur Loss: 0.63788855, Cur Avg Loss: 0.71855079, Log Avg loss: 0.73343309, Global Avg Loss: 2.08756787, Time: 0.0108 Steps: 39700, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001388, Sample Num: 22208, Cur Loss: 0.84916013, Cur Avg Loss: 0.71915845, Log Avg loss: 0.80289512, Global Avg Loss: 2.08724436, Time: 0.0110 Steps: 39710, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001398, Sample Num: 22368, Cur Loss: 0.70067477, Cur Avg Loss: 0.71776195, Log Avg loss: 0.52392786, Global Avg Loss: 2.08685077, Time: 0.0083 Steps: 39720, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001408, Sample Num: 22528, Cur Loss: 0.88803375, Cur Avg Loss: 0.72037504, Log Avg loss: 1.08568491, Global Avg Loss: 2.08659878, Time: 0.0066 Steps: 39730, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001418, Sample Num: 22688, Cur Loss: 0.70161736, Cur Avg Loss: 0.72101874, Log Avg loss: 0.81165135, Global Avg Loss: 2.08627796, Time: 0.0104 Steps: 39740, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001428, Sample Num: 22848, Cur Loss: 0.45050350, Cur Avg Loss: 0.72298391, Log Avg loss: 1.00164506, Global Avg Loss: 2.08600509, Time: 0.0067 Steps: 39750, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001438, Sample Num: 23008, Cur Loss: 0.67741966, Cur Avg Loss: 0.72305543, Log Avg loss: 0.73326890, Global Avg Loss: 2.08566487, Time: 0.0067 Steps: 39760, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001448, Sample Num: 23168, Cur Loss: 0.60362393, Cur Avg Loss: 0.72282971, Log Avg loss: 0.69037122, Global Avg Loss: 2.08531403, Time: 0.0133 Steps: 39770, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001458, Sample Num: 23328, Cur Loss: 0.52117079, Cur Avg Loss: 0.72228836, Log Avg loss: 0.64389972, Global Avg Loss: 2.08495168, Time: 0.0066 Steps: 39780, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001468, Sample Num: 23488, Cur Loss: 1.47169352, Cur Avg Loss: 0.72308864, Log Avg loss: 0.83977057, Global Avg Loss: 2.08463874, Time: 0.0112 Steps: 39790, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001478, Sample Num: 23648, Cur Loss: 0.43747553, Cur Avg Loss: 0.72298199, Log Avg loss: 0.70732456, Global Avg Loss: 2.08429268, Time: 0.0119 Steps: 39800, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001488, Sample Num: 23808, Cur Loss: 0.70030999, Cur Avg Loss: 0.72346535, Log Avg loss: 0.79490651, Global Avg Loss: 2.08396880, Time: 0.0067 Steps: 39810, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001498, Sample Num: 23968, Cur Loss: 0.17783368, Cur Avg Loss: 0.72290838, Log Avg loss: 0.64003181, Global Avg Loss: 2.08360618, Time: 0.0067 Steps: 39820, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001508, Sample Num: 24128, Cur Loss: 0.68053395, Cur Avg Loss: 0.72299643, Log Avg loss: 0.73618606, Global Avg Loss: 2.08326789, Time: 0.0135 Steps: 39830, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001518, Sample Num: 24288, Cur Loss: 0.69614398, Cur Avg Loss: 0.72301044, Log Avg loss: 0.72512237, Global Avg Loss: 2.08292699, Time: 0.0111 Steps: 39840, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001528, Sample Num: 24448, Cur Loss: 0.52184725, Cur Avg Loss: 0.72257380, Log Avg loss: 0.65629270, Global Avg Loss: 2.08256899, Time: 0.0121 Steps: 39850, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001538, Sample Num: 24608, Cur Loss: 0.16122325, Cur Avg Loss: 0.72176975, Log Avg loss: 0.59891055, Global Avg Loss: 2.08219677, Time: 0.0072 Steps: 39860, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001548, Sample Num: 24768, Cur Loss: 1.14186871, Cur Avg Loss: 0.72348840, Log Avg loss: 0.98781705, Global Avg Loss: 2.08192229, Time: 0.0067 Steps: 39870, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001558, Sample Num: 24928, Cur Loss: 0.93677461, Cur Avg Loss: 0.72270817, Log Avg loss: 0.60192883, Global Avg Loss: 2.08155117, Time: 0.0070 Steps: 39880, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001568, Sample Num: 25088, Cur Loss: 0.34350401, Cur Avg Loss: 0.72201204, Log Avg loss: 0.61355505, Global Avg Loss: 2.08118316, Time: 0.0236 Steps: 39890, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001578, Sample Num: 25248, Cur Loss: 1.18762159, Cur Avg Loss: 0.72417972, Log Avg loss: 1.06407146, Global Avg Loss: 2.08092825, Time: 0.0116 Steps: 39900, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001588, Sample Num: 25408, Cur Loss: 0.82116145, Cur Avg Loss: 0.72418826, Log Avg loss: 0.72553654, Global Avg Loss: 2.08058864, Time: 0.0073 Steps: 39910, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001598, Sample Num: 25568, Cur Loss: 0.15979652, Cur Avg Loss: 0.72387652, Log Avg loss: 0.67437137, Global Avg Loss: 2.08023638, Time: 0.0064 Steps: 39920, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001608, Sample Num: 25728, Cur Loss: 0.20808601, Cur Avg Loss: 0.72286644, Log Avg loss: 0.56145558, Global Avg Loss: 2.07985602, Time: 0.0173 Steps: 39930, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001618, Sample Num: 25888, Cur Loss: 0.35899019, Cur Avg Loss: 0.72237659, Log Avg loss: 0.64360905, Global Avg Loss: 2.07949641, Time: 0.0116 Steps: 39940, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001628, Sample Num: 26048, Cur Loss: 0.96703076, Cur Avg Loss: 0.72370351, Log Avg loss: 0.93839983, Global Avg Loss: 2.07921078, Time: 0.0075 Steps: 39950, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001638, Sample Num: 26208, Cur Loss: 0.63935792, Cur Avg Loss: 0.72345689, Log Avg loss: 0.68330626, Global Avg Loss: 2.07886146, Time: 0.0112 Steps: 39960, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001648, Sample Num: 26368, Cur Loss: 0.98406088, Cur Avg Loss: 0.72407720, Log Avg loss: 0.82568414, Global Avg Loss: 2.07854793, Time: 0.0116 Steps: 39970, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001658, Sample Num: 26528, Cur Loss: 1.01197314, Cur Avg Loss: 0.72416556, Log Avg loss: 0.73872791, Global Avg Loss: 2.07821281, Time: 0.0109 Steps: 39980, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001668, Sample Num: 26688, Cur Loss: 0.44324115, Cur Avg Loss: 0.72425671, Log Avg loss: 0.73936928, Global Avg Loss: 2.07787801, Time: 0.0067 Steps: 39990, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001678, Sample Num: 26848, Cur Loss: 0.26867190, Cur Avg Loss: 0.72446627, Log Avg loss: 0.75942078, Global Avg Loss: 2.07754840, Time: 0.0096 Steps: 40000, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001688, Sample Num: 27008, Cur Loss: 0.40711641, Cur Avg Loss: 0.72460935, Log Avg loss: 0.74861768, Global Avg Loss: 2.07721625, Time: 0.0099 Steps: 40010, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001698, Sample Num: 27168, Cur Loss: 0.66578507, Cur Avg Loss: 0.72401912, Log Avg loss: 0.62438831, Global Avg Loss: 2.07685322, Time: 0.0121 Steps: 40020, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001708, Sample Num: 27328, Cur Loss: 0.88627249, Cur Avg Loss: 0.72384730, Log Avg loss: 0.69467281, Global Avg Loss: 2.07650794, Time: 0.0122 Steps: 40030, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001718, Sample Num: 27488, Cur Loss: 0.93431854, Cur Avg Loss: 0.72271624, Log Avg loss: 0.52953123, Global Avg Loss: 2.07612158, Time: 0.0126 Steps: 40040, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001728, Sample Num: 27648, Cur Loss: 0.99792624, Cur Avg Loss: 0.72206640, Log Avg loss: 0.61042303, Global Avg Loss: 2.07575561, Time: 0.0106 Steps: 40050, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001738, Sample Num: 27808, Cur Loss: 0.45426673, Cur Avg Loss: 0.72220005, Log Avg loss: 0.74529494, Global Avg Loss: 2.07542349, Time: 0.0182 Steps: 40060, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001748, Sample Num: 27968, Cur Loss: 0.35712078, Cur Avg Loss: 0.72265489, Log Avg loss: 0.80170590, Global Avg Loss: 2.07510562, Time: 0.0111 Steps: 40070, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001758, Sample Num: 28128, Cur Loss: 0.27483702, Cur Avg Loss: 0.72210876, Log Avg loss: 0.62664474, Global Avg Loss: 2.07474423, Time: 0.0115 Steps: 40080, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001768, Sample Num: 28288, Cur Loss: 0.91922450, Cur Avg Loss: 0.72189164, Log Avg loss: 0.68372277, Global Avg Loss: 2.07439725, Time: 0.0068 Steps: 40090, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001778, Sample Num: 28448, Cur Loss: 0.29631066, Cur Avg Loss: 0.72133108, Log Avg loss: 0.62222413, Global Avg Loss: 2.07403512, Time: 0.0068 Steps: 40100, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001788, Sample Num: 28608, Cur Loss: 1.24681187, Cur Avg Loss: 0.72086253, Log Avg loss: 0.63755463, Global Avg Loss: 2.07367698, Time: 0.0109 Steps: 40110, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001798, Sample Num: 28768, Cur Loss: 1.12214649, Cur Avg Loss: 0.72160534, Log Avg loss: 0.85441940, Global Avg Loss: 2.07337308, Time: 0.0127 Steps: 40120, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001808, Sample Num: 28928, Cur Loss: 0.72638994, Cur Avg Loss: 0.72227796, Log Avg loss: 0.84321549, Global Avg Loss: 2.07306653, Time: 0.0122 Steps: 40130, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001818, Sample Num: 29088, Cur Loss: 0.66470182, Cur Avg Loss: 0.72301341, Log Avg loss: 0.85598231, Global Avg Loss: 2.07276332, Time: 0.0067 Steps: 40140, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001828, Sample Num: 29248, Cur Loss: 1.09068513, Cur Avg Loss: 0.72355137, Log Avg loss: 0.82135338, Global Avg Loss: 2.07245164, Time: 0.0114 Steps: 40150, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001838, Sample Num: 29408, Cur Loss: 0.45926896, Cur Avg Loss: 0.72235920, Log Avg loss: 0.50442956, Global Avg Loss: 2.07206120, Time: 0.0142 Steps: 40160, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001848, Sample Num: 29568, Cur Loss: 0.35456824, Cur Avg Loss: 0.72210693, Log Avg loss: 0.67573921, Global Avg Loss: 2.07171359, Time: 0.0116 Steps: 40170, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001858, Sample Num: 29728, Cur Loss: 0.90756238, Cur Avg Loss: 0.72193779, Log Avg loss: 0.69068067, Global Avg Loss: 2.07136988, Time: 0.0121 Steps: 40180, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001868, Sample Num: 29888, Cur Loss: 0.79508710, Cur Avg Loss: 0.72118989, Log Avg loss: 0.58223138, Global Avg Loss: 2.07099936, Time: 0.0118 Steps: 40190, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001878, Sample Num: 30048, Cur Loss: 0.18253100, Cur Avg Loss: 0.71988863, Log Avg loss: 0.47681187, Global Avg Loss: 2.07060279, Time: 0.0116 Steps: 40200, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001888, Sample Num: 30208, Cur Loss: 1.21775806, Cur Avg Loss: 0.71966089, Log Avg loss: 0.67689156, Global Avg Loss: 2.07025619, Time: 0.0067 Steps: 40210, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001898, Sample Num: 30368, Cur Loss: 0.57884216, Cur Avg Loss: 0.71919663, Log Avg loss: 0.63154479, Global Avg Loss: 2.06989848, Time: 0.0116 Steps: 40220, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001908, Sample Num: 30528, Cur Loss: 0.53143269, Cur Avg Loss: 0.71914350, Log Avg loss: 0.70905932, Global Avg Loss: 2.06956021, Time: 0.0072 Steps: 40230, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001918, Sample Num: 30688, Cur Loss: 0.81422341, Cur Avg Loss: 0.71850341, Log Avg loss: 0.59637394, Global Avg Loss: 2.06919411, Time: 0.0090 Steps: 40240, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001928, Sample Num: 30848, Cur Loss: 0.89380014, Cur Avg Loss: 0.71823226, Log Avg loss: 0.66622628, Global Avg Loss: 2.06884555, Time: 0.0067 Steps: 40250, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001938, Sample Num: 31008, Cur Loss: 0.99499995, Cur Avg Loss: 0.71824682, Log Avg loss: 0.72105369, Global Avg Loss: 2.06851078, Time: 0.0118 Steps: 40260, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001948, Sample Num: 31168, Cur Loss: 1.25717425, Cur Avg Loss: 0.71828983, Log Avg loss: 0.72662597, Global Avg Loss: 2.06817755, Time: 0.0226 Steps: 40270, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001958, Sample Num: 31328, Cur Loss: 0.46790987, Cur Avg Loss: 0.71790047, Log Avg loss: 0.64205334, Global Avg Loss: 2.06782350, Time: 0.0110 Steps: 40280, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001968, Sample Num: 31488, Cur Loss: 0.91419059, Cur Avg Loss: 0.71786464, Log Avg loss: 0.71084736, Global Avg Loss: 2.06748670, Time: 0.0085 Steps: 40290, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001978, Sample Num: 31648, Cur Loss: 0.85608864, Cur Avg Loss: 0.71820335, Log Avg loss: 0.78486223, Global Avg Loss: 2.06716843, Time: 0.0087 Steps: 40300, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001988, Sample Num: 31808, Cur Loss: 0.46228221, Cur Avg Loss: 0.71737827, Log Avg loss: 0.55417827, Global Avg Loss: 2.06679309, Time: 0.0128 Steps: 40310, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 001998, Sample Num: 31968, Cur Loss: 0.31284177, Cur Avg Loss: 0.71732246, Log Avg loss: 0.70622739, Global Avg Loss: 2.06645565, Time: 0.0067 Steps: 40320, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002008, Sample Num: 32128, Cur Loss: 0.66698635, Cur Avg Loss: 0.71729823, Log Avg loss: 0.71245562, Global Avg Loss: 2.06611992, Time: 0.0113 Steps: 40330, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002018, Sample Num: 32288, Cur Loss: 0.74832755, Cur Avg Loss: 0.71758690, Log Avg loss: 0.77555309, Global Avg Loss: 2.06580000, Time: 0.0113 Steps: 40340, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002028, Sample Num: 32448, Cur Loss: 0.53686750, Cur Avg Loss: 0.71768600, Log Avg loss: 0.73768465, Global Avg Loss: 2.06547085, Time: 0.0070 Steps: 40350, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002038, Sample Num: 32608, Cur Loss: 0.32540500, Cur Avg Loss: 0.71675217, Log Avg loss: 0.52737060, Global Avg Loss: 2.06508975, Time: 0.0066 Steps: 40360, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002048, Sample Num: 32768, Cur Loss: 0.72794884, Cur Avg Loss: 0.71704718, Log Avg loss: 0.77716933, Global Avg Loss: 2.06477072, Time: 0.0156 Steps: 40370, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002058, Sample Num: 32928, Cur Loss: 1.61773181, Cur Avg Loss: 0.71606682, Log Avg loss: 0.51529048, Global Avg Loss: 2.06438700, Time: 0.0067 Steps: 40380, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002068, Sample Num: 33088, Cur Loss: 0.66552132, Cur Avg Loss: 0.71584281, Log Avg loss: 0.66974182, Global Avg Loss: 2.06404170, Time: 0.0068 Steps: 40390, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002078, Sample Num: 33248, Cur Loss: 0.54453975, Cur Avg Loss: 0.71544727, Log Avg loss: 0.63364801, Global Avg Loss: 2.06368765, Time: 0.0067 Steps: 40400, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002088, Sample Num: 33408, Cur Loss: 0.24285519, Cur Avg Loss: 0.71513695, Log Avg loss: 0.65065257, Global Avg Loss: 2.06333797, Time: 0.0106 Steps: 40410, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002098, Sample Num: 33568, Cur Loss: 0.46633267, Cur Avg Loss: 0.71552770, Log Avg loss: 0.79711641, Global Avg Loss: 2.06302471, Time: 0.0109 Steps: 40420, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002108, Sample Num: 33728, Cur Loss: 0.90240216, Cur Avg Loss: 0.71517189, Log Avg loss: 0.64052451, Global Avg Loss: 2.06267286, Time: 0.0095 Steps: 40430, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002118, Sample Num: 33888, Cur Loss: 0.34607089, Cur Avg Loss: 0.71437169, Log Avg loss: 0.54568768, Global Avg Loss: 2.06229774, Time: 0.0068 Steps: 40440, Updated lr: 0.000063 Training, Epoch: 0019, Batch: 002128, Sample Num: 34048, Cur Loss: 0.32308766, Cur Avg Loss: 0.71399631, Log Avg loss: 0.63449117, Global Avg Loss: 2.06194476, Time: 0.0128 Steps: 40450, Updated lr: 0.000063 ***** Running evaluation checkpoint-40451 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-40451 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.043491, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.824886, "eval_total_loss": 579.89457, "eval_mae": 0.7315, "eval_mse": 0.825, "eval_r2": 0.475576, "eval_sp_statistic": 0.637189, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.70999, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.078896, "test_total_loss": 541.605637, "test_mae": 0.859508, "test_mse": 1.078983, "test_r2": 0.303615, "test_sp_statistic": 0.482466, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.595428, "test_ps_pvalue": 0.0, "lr": 6.258795637743007e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 2.061916976929527, "train_cur_epoch_loss": 1520.3221557214856, "train_cur_epoch_avg_loss": 0.714101529225686, "train_cur_epoch_time": 22.04349136352539, "train_cur_epoch_avg_time": 0.010353917972534238, "epoch": 19, "step": 40451} ################################################## Training, Epoch: 0020, Batch: 000009, Sample Num: 144, Cur Loss: 0.42034507, Cur Avg Loss: 0.63013706, Log Avg loss: 0.66092481, Global Avg Loss: 2.06159849, Time: 0.0135 Steps: 40460, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000019, Sample Num: 304, Cur Loss: 0.59729785, Cur Avg Loss: 0.84471819, Log Avg loss: 1.03784121, Global Avg Loss: 2.06134552, Time: 0.0159 Steps: 40470, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000029, Sample Num: 464, Cur Loss: 0.75877094, Cur Avg Loss: 0.81981437, Log Avg loss: 0.77249713, Global Avg Loss: 2.06102713, Time: 0.0090 Steps: 40480, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000039, Sample Num: 624, Cur Loss: 0.21866682, Cur Avg Loss: 0.81860454, Log Avg loss: 0.81509604, Global Avg Loss: 2.06071942, Time: 0.0068 Steps: 40490, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000049, Sample Num: 784, Cur Loss: 1.03451455, Cur Avg Loss: 0.80843433, Log Avg loss: 0.76877048, Global Avg Loss: 2.06040042, Time: 0.0071 Steps: 40500, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000059, Sample Num: 944, Cur Loss: 0.30481181, Cur Avg Loss: 0.79345060, Log Avg loss: 0.72003031, Global Avg Loss: 2.06006954, Time: 0.0068 Steps: 40510, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000069, Sample Num: 1104, Cur Loss: 0.51440209, Cur Avg Loss: 0.75321468, Log Avg loss: 0.51582275, Global Avg Loss: 2.05968844, Time: 0.0068 Steps: 40520, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000079, Sample Num: 1264, Cur Loss: 0.86209178, Cur Avg Loss: 0.72967966, Log Avg loss: 0.56728804, Global Avg Loss: 2.05932022, Time: 0.0111 Steps: 40530, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000089, Sample Num: 1424, Cur Loss: 0.63807148, Cur Avg Loss: 0.71719881, Log Avg loss: 0.61860007, Global Avg Loss: 2.05896483, Time: 0.0138 Steps: 40540, Updated lr: 0.000063 Training, Epoch: 0020, Batch: 000099, Sample Num: 1584, Cur Loss: 1.01199567, Cur Avg Loss: 0.70625959, Log Avg loss: 0.60890059, Global Avg Loss: 2.05860723, Time: 0.0135 Steps: 40550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000109, Sample Num: 1744, Cur Loss: 0.34639713, Cur Avg Loss: 0.72376268, Log Avg loss: 0.89704323, Global Avg Loss: 2.05832085, Time: 0.0068 Steps: 40560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000119, Sample Num: 1904, Cur Loss: 0.32156345, Cur Avg Loss: 0.71711318, Log Avg loss: 0.64463363, Global Avg Loss: 2.05797240, Time: 0.0109 Steps: 40570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000129, Sample Num: 2064, Cur Loss: 0.29286891, Cur Avg Loss: 0.72338158, Log Avg loss: 0.79797553, Global Avg Loss: 2.05766190, Time: 0.0117 Steps: 40580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000139, Sample Num: 2224, Cur Loss: 0.68172598, Cur Avg Loss: 0.72576117, Log Avg loss: 0.75645792, Global Avg Loss: 2.05734133, Time: 0.0067 Steps: 40590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000149, Sample Num: 2384, Cur Loss: 0.71922529, Cur Avg Loss: 0.72325114, Log Avg loss: 0.68836172, Global Avg Loss: 2.05700414, Time: 0.0116 Steps: 40600, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000159, Sample Num: 2544, Cur Loss: 0.76620609, Cur Avg Loss: 0.72034430, Log Avg loss: 0.67703242, Global Avg Loss: 2.05666433, Time: 0.0117 Steps: 40610, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000169, Sample Num: 2704, Cur Loss: 0.31245840, Cur Avg Loss: 0.71514674, Log Avg loss: 0.63250551, Global Avg Loss: 2.05631372, Time: 0.0069 Steps: 40620, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000179, Sample Num: 2864, Cur Loss: 0.35636219, Cur Avg Loss: 0.70241535, Log Avg loss: 0.48725479, Global Avg Loss: 2.05592754, Time: 0.0128 Steps: 40630, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000189, Sample Num: 3024, Cur Loss: 0.14455771, Cur Avg Loss: 0.70159089, Log Avg loss: 0.68683304, Global Avg Loss: 2.05559066, Time: 0.0113 Steps: 40640, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000199, Sample Num: 3184, Cur Loss: 0.63607287, Cur Avg Loss: 0.70787649, Log Avg loss: 0.82667434, Global Avg Loss: 2.05528834, Time: 0.0110 Steps: 40650, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000209, Sample Num: 3344, Cur Loss: 1.02417874, Cur Avg Loss: 0.69907897, Log Avg loss: 0.52400828, Global Avg Loss: 2.05491173, Time: 0.0067 Steps: 40660, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000219, Sample Num: 3504, Cur Loss: 0.39209837, Cur Avg Loss: 0.69913940, Log Avg loss: 0.70040242, Global Avg Loss: 2.05457869, Time: 0.0067 Steps: 40670, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000229, Sample Num: 3664, Cur Loss: 0.57730353, Cur Avg Loss: 0.69264114, Log Avg loss: 0.55032923, Global Avg Loss: 2.05420891, Time: 0.0110 Steps: 40680, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000239, Sample Num: 3824, Cur Loss: 0.60073483, Cur Avg Loss: 0.69391133, Log Avg loss: 0.72299881, Global Avg Loss: 2.05388175, Time: 0.0108 Steps: 40690, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000249, Sample Num: 3984, Cur Loss: 0.39266735, Cur Avg Loss: 0.69436093, Log Avg loss: 0.70510633, Global Avg Loss: 2.05355036, Time: 0.0068 Steps: 40700, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000259, Sample Num: 4144, Cur Loss: 0.36930358, Cur Avg Loss: 0.69943942, Log Avg loss: 0.82589383, Global Avg Loss: 2.05324879, Time: 0.0088 Steps: 40710, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000269, Sample Num: 4304, Cur Loss: 0.40224409, Cur Avg Loss: 0.69484180, Log Avg loss: 0.57576330, Global Avg Loss: 2.05288595, Time: 0.0068 Steps: 40720, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000279, Sample Num: 4464, Cur Loss: 1.04413867, Cur Avg Loss: 0.69928339, Log Avg loss: 0.81876225, Global Avg Loss: 2.05258295, Time: 0.0122 Steps: 40730, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000289, Sample Num: 4624, Cur Loss: 0.83100748, Cur Avg Loss: 0.69697211, Log Avg loss: 0.63248730, Global Avg Loss: 2.05223438, Time: 0.0120 Steps: 40740, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000299, Sample Num: 4784, Cur Loss: 0.60941172, Cur Avg Loss: 0.69768895, Log Avg loss: 0.71840586, Global Avg Loss: 2.05190706, Time: 0.0107 Steps: 40750, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000309, Sample Num: 4944, Cur Loss: 0.35024053, Cur Avg Loss: 0.69573640, Log Avg loss: 0.63735493, Global Avg Loss: 2.05156001, Time: 0.0106 Steps: 40760, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000319, Sample Num: 5104, Cur Loss: 0.17119583, Cur Avg Loss: 0.69629014, Log Avg loss: 0.71340084, Global Avg Loss: 2.05123179, Time: 0.0106 Steps: 40770, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000329, Sample Num: 5264, Cur Loss: 1.27298737, Cur Avg Loss: 0.69790022, Log Avg loss: 0.74926177, Global Avg Loss: 2.05091253, Time: 0.0080 Steps: 40780, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000339, Sample Num: 5424, Cur Loss: 0.71666765, Cur Avg Loss: 0.69734847, Log Avg loss: 0.67919577, Global Avg Loss: 2.05057624, Time: 0.0069 Steps: 40790, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000349, Sample Num: 5584, Cur Loss: 0.73075444, Cur Avg Loss: 0.69767546, Log Avg loss: 0.70876056, Global Avg Loss: 2.05024736, Time: 0.0065 Steps: 40800, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000359, Sample Num: 5744, Cur Loss: 0.20423919, Cur Avg Loss: 0.69418290, Log Avg loss: 0.57229242, Global Avg Loss: 2.04988521, Time: 0.0083 Steps: 40810, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000369, Sample Num: 5904, Cur Loss: 0.66897136, Cur Avg Loss: 0.69066244, Log Avg loss: 0.56427800, Global Avg Loss: 2.04952127, Time: 0.0069 Steps: 40820, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000379, Sample Num: 6064, Cur Loss: 0.93345404, Cur Avg Loss: 0.69211772, Log Avg loss: 0.74581771, Global Avg Loss: 2.04920197, Time: 0.0088 Steps: 40830, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000389, Sample Num: 6224, Cur Loss: 1.95519853, Cur Avg Loss: 0.69666431, Log Avg loss: 0.86897994, Global Avg Loss: 2.04891298, Time: 0.0067 Steps: 40840, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000399, Sample Num: 6384, Cur Loss: 0.65731287, Cur Avg Loss: 0.69932422, Log Avg loss: 0.80279466, Global Avg Loss: 2.04860793, Time: 0.0092 Steps: 40850, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000409, Sample Num: 6544, Cur Loss: 0.74955851, Cur Avg Loss: 0.70793977, Log Avg loss: 1.05170009, Global Avg Loss: 2.04836395, Time: 0.0111 Steps: 40860, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000419, Sample Num: 6704, Cur Loss: 0.47856709, Cur Avg Loss: 0.70836826, Log Avg loss: 0.72589363, Global Avg Loss: 2.04804037, Time: 0.0111 Steps: 40870, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000429, Sample Num: 6864, Cur Loss: 0.41690019, Cur Avg Loss: 0.70960062, Log Avg loss: 0.76123671, Global Avg Loss: 2.04772559, Time: 0.0107 Steps: 40880, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000439, Sample Num: 7024, Cur Loss: 0.67346537, Cur Avg Loss: 0.70617363, Log Avg loss: 0.55915542, Global Avg Loss: 2.04736155, Time: 0.0108 Steps: 40890, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000449, Sample Num: 7184, Cur Loss: 0.21433619, Cur Avg Loss: 0.70090291, Log Avg loss: 0.46951826, Global Avg Loss: 2.04697577, Time: 0.0105 Steps: 40900, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000459, Sample Num: 7344, Cur Loss: 0.94131333, Cur Avg Loss: 0.69626398, Log Avg loss: 0.48797613, Global Avg Loss: 2.04659469, Time: 0.0064 Steps: 40910, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000469, Sample Num: 7504, Cur Loss: 0.34972477, Cur Avg Loss: 0.69592512, Log Avg loss: 0.68037134, Global Avg Loss: 2.04626081, Time: 0.0113 Steps: 40920, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000479, Sample Num: 7664, Cur Loss: 0.73335224, Cur Avg Loss: 0.69443954, Log Avg loss: 0.62476585, Global Avg Loss: 2.04591352, Time: 0.0119 Steps: 40930, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000489, Sample Num: 7824, Cur Loss: 1.72000372, Cur Avg Loss: 0.69671082, Log Avg loss: 0.80550540, Global Avg Loss: 2.04561053, Time: 0.0153 Steps: 40940, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000499, Sample Num: 7984, Cur Loss: 0.54580361, Cur Avg Loss: 0.69916166, Log Avg loss: 0.81900784, Global Avg Loss: 2.04531100, Time: 0.0065 Steps: 40950, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000509, Sample Num: 8144, Cur Loss: 0.88925970, Cur Avg Loss: 0.70028034, Log Avg loss: 0.75610250, Global Avg Loss: 2.04499625, Time: 0.0104 Steps: 40960, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000519, Sample Num: 8304, Cur Loss: 1.07788587, Cur Avg Loss: 0.70536317, Log Avg loss: 0.96407880, Global Avg Loss: 2.04473242, Time: 0.0073 Steps: 40970, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000529, Sample Num: 8464, Cur Loss: 0.75167561, Cur Avg Loss: 0.70621697, Log Avg loss: 0.75052966, Global Avg Loss: 2.04441660, Time: 0.0134 Steps: 40980, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000539, Sample Num: 8624, Cur Loss: 0.96481264, Cur Avg Loss: 0.70277143, Log Avg loss: 0.52050214, Global Avg Loss: 2.04404483, Time: 0.0090 Steps: 40990, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000549, Sample Num: 8784, Cur Loss: 0.33869174, Cur Avg Loss: 0.69985858, Log Avg loss: 0.54285598, Global Avg Loss: 2.04367868, Time: 0.0136 Steps: 41000, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000559, Sample Num: 8944, Cur Loss: 0.82881969, Cur Avg Loss: 0.69955517, Log Avg loss: 0.68289789, Global Avg Loss: 2.04334687, Time: 0.0215 Steps: 41010, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000569, Sample Num: 9104, Cur Loss: 0.47100574, Cur Avg Loss: 0.70252155, Log Avg loss: 0.86834244, Global Avg Loss: 2.04306042, Time: 0.0120 Steps: 41020, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000579, Sample Num: 9264, Cur Loss: 1.24642539, Cur Avg Loss: 0.70114207, Log Avg loss: 0.62264965, Global Avg Loss: 2.04271423, Time: 0.0113 Steps: 41030, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000589, Sample Num: 9424, Cur Loss: 1.15745878, Cur Avg Loss: 0.70013889, Log Avg loss: 0.64205484, Global Avg Loss: 2.04237294, Time: 0.0068 Steps: 41040, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000599, Sample Num: 9584, Cur Loss: 0.90531862, Cur Avg Loss: 0.69878806, Log Avg loss: 0.61922372, Global Avg Loss: 2.04202625, Time: 0.0102 Steps: 41050, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000609, Sample Num: 9744, Cur Loss: 0.64435536, Cur Avg Loss: 0.69903454, Log Avg loss: 0.71379884, Global Avg Loss: 2.04170277, Time: 0.0115 Steps: 41060, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000619, Sample Num: 9904, Cur Loss: 1.19508564, Cur Avg Loss: 0.70240414, Log Avg loss: 0.90761272, Global Avg Loss: 2.04142663, Time: 0.0071 Steps: 41070, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000629, Sample Num: 10064, Cur Loss: 0.48821807, Cur Avg Loss: 0.70300347, Log Avg loss: 0.74010205, Global Avg Loss: 2.04110985, Time: 0.0109 Steps: 41080, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000639, Sample Num: 10224, Cur Loss: 0.98348862, Cur Avg Loss: 0.70132225, Log Avg loss: 0.59557360, Global Avg Loss: 2.04075806, Time: 0.0068 Steps: 41090, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000649, Sample Num: 10384, Cur Loss: 0.36988282, Cur Avg Loss: 0.69788203, Log Avg loss: 0.47805177, Global Avg Loss: 2.04037784, Time: 0.0068 Steps: 41100, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000659, Sample Num: 10544, Cur Loss: 0.35314384, Cur Avg Loss: 0.69595083, Log Avg loss: 0.57061608, Global Avg Loss: 2.04002032, Time: 0.0117 Steps: 41110, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000669, Sample Num: 10704, Cur Loss: 0.52984571, Cur Avg Loss: 0.69388226, Log Avg loss: 0.55756379, Global Avg Loss: 2.03965980, Time: 0.0068 Steps: 41120, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000679, Sample Num: 10864, Cur Loss: 0.49582130, Cur Avg Loss: 0.69460944, Log Avg loss: 0.74325775, Global Avg Loss: 2.03934460, Time: 0.0089 Steps: 41130, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000689, Sample Num: 11024, Cur Loss: 0.70077097, Cur Avg Loss: 0.69369742, Log Avg loss: 0.63177071, Global Avg Loss: 2.03900246, Time: 0.0113 Steps: 41140, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000699, Sample Num: 11184, Cur Loss: 0.53878653, Cur Avg Loss: 0.69214729, Log Avg loss: 0.58534370, Global Avg Loss: 2.03864920, Time: 0.0067 Steps: 41150, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000709, Sample Num: 11344, Cur Loss: 0.40678281, Cur Avg Loss: 0.69146658, Log Avg loss: 0.64388497, Global Avg Loss: 2.03831034, Time: 0.0111 Steps: 41160, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000719, Sample Num: 11504, Cur Loss: 0.36451206, Cur Avg Loss: 0.69022150, Log Avg loss: 0.60194539, Global Avg Loss: 2.03796145, Time: 0.0067 Steps: 41170, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000729, Sample Num: 11664, Cur Loss: 0.56667668, Cur Avg Loss: 0.68890865, Log Avg loss: 0.59451469, Global Avg Loss: 2.03761093, Time: 0.0128 Steps: 41180, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000739, Sample Num: 11824, Cur Loss: 0.31357449, Cur Avg Loss: 0.68829781, Log Avg loss: 0.64376783, Global Avg Loss: 2.03727254, Time: 0.0102 Steps: 41190, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000749, Sample Num: 11984, Cur Loss: 1.04523301, Cur Avg Loss: 0.68918885, Log Avg loss: 0.75503664, Global Avg Loss: 2.03696131, Time: 0.0227 Steps: 41200, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000759, Sample Num: 12144, Cur Loss: 0.21950302, Cur Avg Loss: 0.68746855, Log Avg loss: 0.55861763, Global Avg Loss: 2.03660258, Time: 0.0086 Steps: 41210, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000769, Sample Num: 12304, Cur Loss: 0.70542157, Cur Avg Loss: 0.68686469, Log Avg loss: 0.64103187, Global Avg Loss: 2.03626401, Time: 0.0150 Steps: 41220, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000779, Sample Num: 12464, Cur Loss: 1.30266166, Cur Avg Loss: 0.68587074, Log Avg loss: 0.60943592, Global Avg Loss: 2.03591795, Time: 0.0118 Steps: 41230, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000789, Sample Num: 12624, Cur Loss: 0.42991275, Cur Avg Loss: 0.68484587, Log Avg loss: 0.60500868, Global Avg Loss: 2.03557098, Time: 0.0120 Steps: 41240, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000799, Sample Num: 12784, Cur Loss: 0.73928165, Cur Avg Loss: 0.68572813, Log Avg loss: 0.75533848, Global Avg Loss: 2.03526062, Time: 0.0112 Steps: 41250, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000809, Sample Num: 12944, Cur Loss: 0.74345636, Cur Avg Loss: 0.68791112, Log Avg loss: 0.86233152, Global Avg Loss: 2.03497634, Time: 0.0068 Steps: 41260, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000819, Sample Num: 13104, Cur Loss: 0.48717424, Cur Avg Loss: 0.68699593, Log Avg loss: 0.61295778, Global Avg Loss: 2.03463177, Time: 0.0126 Steps: 41270, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000829, Sample Num: 13264, Cur Loss: 0.73148674, Cur Avg Loss: 0.68808052, Log Avg loss: 0.77690825, Global Avg Loss: 2.03432709, Time: 0.0121 Steps: 41280, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000839, Sample Num: 13424, Cur Loss: 0.55424351, Cur Avg Loss: 0.68721129, Log Avg loss: 0.61515230, Global Avg Loss: 2.03398338, Time: 0.0068 Steps: 41290, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000849, Sample Num: 13584, Cur Loss: 0.21393891, Cur Avg Loss: 0.68802600, Log Avg loss: 0.75638011, Global Avg Loss: 2.03367404, Time: 0.0073 Steps: 41300, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000859, Sample Num: 13744, Cur Loss: 0.57767916, Cur Avg Loss: 0.68687760, Log Avg loss: 0.58937855, Global Avg Loss: 2.03332441, Time: 0.0103 Steps: 41310, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000869, Sample Num: 13904, Cur Loss: 0.84734362, Cur Avg Loss: 0.68795359, Log Avg loss: 0.78038073, Global Avg Loss: 2.03302118, Time: 0.0069 Steps: 41320, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000879, Sample Num: 14064, Cur Loss: 0.96054578, Cur Avg Loss: 0.68862718, Log Avg loss: 0.74716221, Global Avg Loss: 2.03271006, Time: 0.0109 Steps: 41330, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000889, Sample Num: 14224, Cur Loss: 0.23424014, Cur Avg Loss: 0.68892748, Log Avg loss: 0.71532423, Global Avg Loss: 2.03239139, Time: 0.0067 Steps: 41340, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000899, Sample Num: 14384, Cur Loss: 0.22788468, Cur Avg Loss: 0.68901161, Log Avg loss: 0.69649014, Global Avg Loss: 2.03206832, Time: 0.0068 Steps: 41350, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000909, Sample Num: 14544, Cur Loss: 0.77252519, Cur Avg Loss: 0.69055301, Log Avg loss: 0.82912531, Global Avg Loss: 2.03177747, Time: 0.0068 Steps: 41360, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000919, Sample Num: 14704, Cur Loss: 0.42710954, Cur Avg Loss: 0.69063487, Log Avg loss: 0.69807594, Global Avg Loss: 2.03145509, Time: 0.0126 Steps: 41370, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000929, Sample Num: 14864, Cur Loss: 0.18176109, Cur Avg Loss: 0.69083775, Log Avg loss: 0.70948222, Global Avg Loss: 2.03113562, Time: 0.0073 Steps: 41380, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000939, Sample Num: 15024, Cur Loss: 0.67285848, Cur Avg Loss: 0.69153286, Log Avg loss: 0.75610842, Global Avg Loss: 2.03082757, Time: 0.0068 Steps: 41390, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000949, Sample Num: 15184, Cur Loss: 0.40994194, Cur Avg Loss: 0.69131467, Log Avg loss: 0.67082668, Global Avg Loss: 2.03049906, Time: 0.0070 Steps: 41400, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000959, Sample Num: 15344, Cur Loss: 0.49749681, Cur Avg Loss: 0.69134908, Log Avg loss: 0.69461487, Global Avg Loss: 2.03017646, Time: 0.0193 Steps: 41410, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000969, Sample Num: 15504, Cur Loss: 0.57314855, Cur Avg Loss: 0.69314468, Log Avg loss: 0.86534247, Global Avg Loss: 2.02989524, Time: 0.0120 Steps: 41420, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000979, Sample Num: 15664, Cur Loss: 0.91906172, Cur Avg Loss: 0.69417964, Log Avg loss: 0.79446777, Global Avg Loss: 2.02959704, Time: 0.0154 Steps: 41430, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000989, Sample Num: 15824, Cur Loss: 1.11026263, Cur Avg Loss: 0.69403408, Log Avg loss: 0.67978353, Global Avg Loss: 2.02927132, Time: 0.0117 Steps: 41440, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 000999, Sample Num: 15984, Cur Loss: 0.74849284, Cur Avg Loss: 0.69338274, Log Avg loss: 0.62896533, Global Avg Loss: 2.02893349, Time: 0.0082 Steps: 41450, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001009, Sample Num: 16144, Cur Loss: 0.60795021, Cur Avg Loss: 0.69482774, Log Avg loss: 0.83918265, Global Avg Loss: 2.02864652, Time: 0.0120 Steps: 41460, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001019, Sample Num: 16304, Cur Loss: 0.58579504, Cur Avg Loss: 0.69552025, Log Avg loss: 0.76539455, Global Avg Loss: 2.02834190, Time: 0.0101 Steps: 41470, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001029, Sample Num: 16464, Cur Loss: 0.98169625, Cur Avg Loss: 0.69499489, Log Avg loss: 0.64146128, Global Avg Loss: 2.02800755, Time: 0.0163 Steps: 41480, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001039, Sample Num: 16624, Cur Loss: 0.34864154, Cur Avg Loss: 0.69385753, Log Avg loss: 0.57682315, Global Avg Loss: 2.02765779, Time: 0.0121 Steps: 41490, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001049, Sample Num: 16784, Cur Loss: 0.32118678, Cur Avg Loss: 0.69124107, Log Avg loss: 0.41939105, Global Avg Loss: 2.02727025, Time: 0.0064 Steps: 41500, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001059, Sample Num: 16944, Cur Loss: 0.42425451, Cur Avg Loss: 0.68981869, Log Avg loss: 0.54061095, Global Avg Loss: 2.02691211, Time: 0.0075 Steps: 41510, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001069, Sample Num: 17104, Cur Loss: 1.00008559, Cur Avg Loss: 0.69015039, Log Avg loss: 0.72527677, Global Avg Loss: 2.02659861, Time: 0.0067 Steps: 41520, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001079, Sample Num: 17264, Cur Loss: 0.42570037, Cur Avg Loss: 0.69057851, Log Avg loss: 0.73634437, Global Avg Loss: 2.02628793, Time: 0.0067 Steps: 41530, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001089, Sample Num: 17424, Cur Loss: 0.36802548, Cur Avg Loss: 0.69243026, Log Avg loss: 0.89223423, Global Avg Loss: 2.02601493, Time: 0.0067 Steps: 41540, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001099, Sample Num: 17584, Cur Loss: 0.68017793, Cur Avg Loss: 0.69199747, Log Avg loss: 0.64486703, Global Avg Loss: 2.02568252, Time: 0.0100 Steps: 41550, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001109, Sample Num: 17744, Cur Loss: 1.31325746, Cur Avg Loss: 0.69219057, Log Avg loss: 0.71341237, Global Avg Loss: 2.02536677, Time: 0.0085 Steps: 41560, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001119, Sample Num: 17904, Cur Loss: 0.42056462, Cur Avg Loss: 0.69212330, Log Avg loss: 0.68466335, Global Avg Loss: 2.02504425, Time: 0.0115 Steps: 41570, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001129, Sample Num: 18064, Cur Loss: 0.94710851, Cur Avg Loss: 0.69368943, Log Avg loss: 0.86893928, Global Avg Loss: 2.02476621, Time: 0.0141 Steps: 41580, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001139, Sample Num: 18224, Cur Loss: 0.80647242, Cur Avg Loss: 0.69459646, Log Avg loss: 0.79699949, Global Avg Loss: 2.02447100, Time: 0.0069 Steps: 41590, Updated lr: 0.000062 Training, Epoch: 0020, Batch: 001149, Sample Num: 18384, Cur Loss: 0.60097301, Cur Avg Loss: 0.69468510, Log Avg loss: 0.70478199, Global Avg Loss: 2.02415377, Time: 0.0142 Steps: 41600, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001159, Sample Num: 18544, Cur Loss: 0.41952345, Cur Avg Loss: 0.69290738, Log Avg loss: 0.48864644, Global Avg Loss: 2.02378475, Time: 0.0074 Steps: 41610, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001169, Sample Num: 18704, Cur Loss: 1.03029919, Cur Avg Loss: 0.69222775, Log Avg loss: 0.61345892, Global Avg Loss: 2.02344589, Time: 0.0164 Steps: 41620, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001179, Sample Num: 18864, Cur Loss: 0.48020083, Cur Avg Loss: 0.69172416, Log Avg loss: 0.63285477, Global Avg Loss: 2.02311185, Time: 0.0073 Steps: 41630, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001189, Sample Num: 19024, Cur Loss: 1.08319271, Cur Avg Loss: 0.69190369, Log Avg loss: 0.71307040, Global Avg Loss: 2.02279724, Time: 0.0117 Steps: 41640, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001199, Sample Num: 19184, Cur Loss: 0.73089200, Cur Avg Loss: 0.69305186, Log Avg loss: 0.82956921, Global Avg Loss: 2.02251075, Time: 0.0067 Steps: 41650, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001209, Sample Num: 19344, Cur Loss: 0.65851378, Cur Avg Loss: 0.69384869, Log Avg loss: 0.78938826, Global Avg Loss: 2.02221476, Time: 0.0073 Steps: 41660, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001219, Sample Num: 19504, Cur Loss: 1.01010990, Cur Avg Loss: 0.69481061, Log Avg loss: 0.81110629, Global Avg Loss: 2.02192411, Time: 0.0113 Steps: 41670, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001229, Sample Num: 19664, Cur Loss: 0.15187731, Cur Avg Loss: 0.69572083, Log Avg loss: 0.80667748, Global Avg Loss: 2.02163255, Time: 0.0238 Steps: 41680, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001239, Sample Num: 19824, Cur Loss: 0.46196935, Cur Avg Loss: 0.69499730, Log Avg loss: 0.60607537, Global Avg Loss: 2.02129300, Time: 0.0122 Steps: 41690, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001249, Sample Num: 19984, Cur Loss: 0.40525264, Cur Avg Loss: 0.69394941, Log Avg loss: 0.56411581, Global Avg Loss: 2.02094356, Time: 0.0138 Steps: 41700, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001259, Sample Num: 20144, Cur Loss: 0.98746461, Cur Avg Loss: 0.69523024, Log Avg loss: 0.85520530, Global Avg Loss: 2.02066407, Time: 0.0067 Steps: 41710, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001269, Sample Num: 20304, Cur Loss: 0.93571740, Cur Avg Loss: 0.69578455, Log Avg loss: 0.76557274, Global Avg Loss: 2.02036324, Time: 0.0239 Steps: 41720, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001279, Sample Num: 20464, Cur Loss: 0.44393411, Cur Avg Loss: 0.69490861, Log Avg loss: 0.58375162, Global Avg Loss: 2.02001897, Time: 0.0067 Steps: 41730, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001289, Sample Num: 20624, Cur Loss: 0.68672097, Cur Avg Loss: 0.69417773, Log Avg loss: 0.60069844, Global Avg Loss: 2.01967893, Time: 0.0068 Steps: 41740, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001299, Sample Num: 20784, Cur Loss: 0.42962712, Cur Avg Loss: 0.69493636, Log Avg loss: 0.79272335, Global Avg Loss: 2.01938505, Time: 0.0108 Steps: 41750, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001309, Sample Num: 20944, Cur Loss: 0.63136333, Cur Avg Loss: 0.69432878, Log Avg loss: 0.61540378, Global Avg Loss: 2.01904885, Time: 0.0073 Steps: 41760, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001319, Sample Num: 21104, Cur Loss: 0.91671985, Cur Avg Loss: 0.69492400, Log Avg loss: 0.77283913, Global Avg Loss: 2.01875050, Time: 0.0072 Steps: 41770, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001329, Sample Num: 21264, Cur Loss: 0.50564522, Cur Avg Loss: 0.69437921, Log Avg loss: 0.62252178, Global Avg Loss: 2.01841631, Time: 0.0072 Steps: 41780, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001339, Sample Num: 21424, Cur Loss: 0.48823550, Cur Avg Loss: 0.69366914, Log Avg loss: 0.59930086, Global Avg Loss: 2.01807673, Time: 0.0227 Steps: 41790, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001349, Sample Num: 21584, Cur Loss: 0.58930063, Cur Avg Loss: 0.69403177, Log Avg loss: 0.74258690, Global Avg Loss: 2.01777159, Time: 0.0123 Steps: 41800, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001359, Sample Num: 21744, Cur Loss: 0.58004802, Cur Avg Loss: 0.69423801, Log Avg loss: 0.72205982, Global Avg Loss: 2.01746169, Time: 0.0123 Steps: 41810, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001369, Sample Num: 21904, Cur Loss: 1.53546178, Cur Avg Loss: 0.69519788, Log Avg loss: 0.82564425, Global Avg Loss: 2.01717670, Time: 0.0067 Steps: 41820, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001379, Sample Num: 22064, Cur Loss: 1.13242054, Cur Avg Loss: 0.69475439, Log Avg loss: 0.63404060, Global Avg Loss: 2.01684604, Time: 0.0099 Steps: 41830, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001389, Sample Num: 22224, Cur Loss: 0.44312447, Cur Avg Loss: 0.69502121, Log Avg loss: 0.73181598, Global Avg Loss: 2.01653891, Time: 0.0113 Steps: 41840, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001399, Sample Num: 22384, Cur Loss: 0.47602683, Cur Avg Loss: 0.69563596, Log Avg loss: 0.78102503, Global Avg Loss: 2.01624369, Time: 0.0123 Steps: 41850, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001409, Sample Num: 22544, Cur Loss: 0.69464064, Cur Avg Loss: 0.69459423, Log Avg loss: 0.54885649, Global Avg Loss: 2.01589314, Time: 0.0065 Steps: 41860, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001419, Sample Num: 22704, Cur Loss: 0.99355906, Cur Avg Loss: 0.69475530, Log Avg loss: 0.71744977, Global Avg Loss: 2.01558303, Time: 0.0106 Steps: 41870, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001429, Sample Num: 22864, Cur Loss: 0.42492595, Cur Avg Loss: 0.69475602, Log Avg loss: 0.69485725, Global Avg Loss: 2.01526767, Time: 0.0064 Steps: 41880, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001439, Sample Num: 23024, Cur Loss: 1.37599194, Cur Avg Loss: 0.69745594, Log Avg loss: 1.08327529, Global Avg Loss: 2.01504518, Time: 0.0066 Steps: 41890, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001449, Sample Num: 23184, Cur Loss: 0.47888860, Cur Avg Loss: 0.69720625, Log Avg loss: 0.66127538, Global Avg Loss: 2.01472209, Time: 0.0067 Steps: 41900, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001459, Sample Num: 23344, Cur Loss: 1.00056100, Cur Avg Loss: 0.69754196, Log Avg loss: 0.74618728, Global Avg Loss: 2.01441941, Time: 0.0067 Steps: 41910, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001469, Sample Num: 23504, Cur Loss: 0.53211576, Cur Avg Loss: 0.69829283, Log Avg loss: 0.80784467, Global Avg Loss: 2.01413158, Time: 0.0130 Steps: 41920, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001479, Sample Num: 23664, Cur Loss: 0.90096921, Cur Avg Loss: 0.69736547, Log Avg loss: 0.56113527, Global Avg Loss: 2.01378505, Time: 0.0128 Steps: 41930, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001489, Sample Num: 23824, Cur Loss: 0.40789410, Cur Avg Loss: 0.69697937, Log Avg loss: 0.63987602, Global Avg Loss: 2.01345746, Time: 0.0114 Steps: 41940, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001499, Sample Num: 23984, Cur Loss: 0.25171027, Cur Avg Loss: 0.69586651, Log Avg loss: 0.53016119, Global Avg Loss: 2.01310387, Time: 0.0065 Steps: 41950, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001509, Sample Num: 24144, Cur Loss: 0.23440221, Cur Avg Loss: 0.69626725, Log Avg loss: 0.75633843, Global Avg Loss: 2.01280436, Time: 0.0113 Steps: 41960, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001519, Sample Num: 24304, Cur Loss: 0.52710789, Cur Avg Loss: 0.69574393, Log Avg loss: 0.61677428, Global Avg Loss: 2.01247173, Time: 0.0068 Steps: 41970, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001529, Sample Num: 24464, Cur Loss: 0.25096112, Cur Avg Loss: 0.69418774, Log Avg loss: 0.45780360, Global Avg Loss: 2.01210140, Time: 0.0137 Steps: 41980, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001539, Sample Num: 24624, Cur Loss: 1.21545875, Cur Avg Loss: 0.69461333, Log Avg loss: 0.75968476, Global Avg Loss: 2.01180313, Time: 0.0064 Steps: 41990, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001549, Sample Num: 24784, Cur Loss: 0.81368840, Cur Avg Loss: 0.69416232, Log Avg loss: 0.62475336, Global Avg Loss: 2.01147288, Time: 0.0065 Steps: 42000, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001559, Sample Num: 24944, Cur Loss: 0.59595239, Cur Avg Loss: 0.69461452, Log Avg loss: 0.76465894, Global Avg Loss: 2.01117609, Time: 0.0065 Steps: 42010, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001569, Sample Num: 25104, Cur Loss: 0.55610418, Cur Avg Loss: 0.69367096, Log Avg loss: 0.54657031, Global Avg Loss: 2.01082754, Time: 0.0071 Steps: 42020, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001579, Sample Num: 25264, Cur Loss: 1.06630445, Cur Avg Loss: 0.69407659, Log Avg loss: 0.75771953, Global Avg Loss: 2.01052940, Time: 0.0068 Steps: 42030, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001589, Sample Num: 25424, Cur Loss: 0.68208241, Cur Avg Loss: 0.69397573, Log Avg loss: 0.67805125, Global Avg Loss: 2.01021244, Time: 0.0103 Steps: 42040, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001599, Sample Num: 25584, Cur Loss: 0.72305495, Cur Avg Loss: 0.69333943, Log Avg loss: 0.59223099, Global Avg Loss: 2.00987523, Time: 0.0162 Steps: 42050, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001609, Sample Num: 25744, Cur Loss: 0.26470616, Cur Avg Loss: 0.69279921, Log Avg loss: 0.60641852, Global Avg Loss: 2.00954155, Time: 0.0072 Steps: 42060, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001619, Sample Num: 25904, Cur Loss: 0.37930268, Cur Avg Loss: 0.69284983, Log Avg loss: 0.70099452, Global Avg Loss: 2.00923051, Time: 0.0071 Steps: 42070, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001629, Sample Num: 26064, Cur Loss: 0.45280099, Cur Avg Loss: 0.69231328, Log Avg loss: 0.60544550, Global Avg Loss: 2.00889691, Time: 0.0113 Steps: 42080, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001639, Sample Num: 26224, Cur Loss: 0.63020694, Cur Avg Loss: 0.69217311, Log Avg loss: 0.66933868, Global Avg Loss: 2.00857865, Time: 0.0086 Steps: 42090, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001649, Sample Num: 26384, Cur Loss: 0.70311028, Cur Avg Loss: 0.69183335, Log Avg loss: 0.63614788, Global Avg Loss: 2.00825266, Time: 0.0130 Steps: 42100, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001659, Sample Num: 26544, Cur Loss: 0.47979727, Cur Avg Loss: 0.69062173, Log Avg loss: 0.49082451, Global Avg Loss: 2.00789231, Time: 0.0108 Steps: 42110, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001669, Sample Num: 26704, Cur Loss: 0.57238483, Cur Avg Loss: 0.69086081, Log Avg loss: 0.73052512, Global Avg Loss: 2.00758904, Time: 0.0148 Steps: 42120, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001679, Sample Num: 26864, Cur Loss: 0.98670578, Cur Avg Loss: 0.69219089, Log Avg loss: 0.91418088, Global Avg Loss: 2.00732951, Time: 0.0130 Steps: 42130, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001689, Sample Num: 27024, Cur Loss: 1.20375121, Cur Avg Loss: 0.69140517, Log Avg loss: 0.55948173, Global Avg Loss: 2.00698593, Time: 0.0086 Steps: 42140, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001699, Sample Num: 27184, Cur Loss: 0.67442620, Cur Avg Loss: 0.69236609, Log Avg loss: 0.85466602, Global Avg Loss: 2.00671254, Time: 0.0130 Steps: 42150, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001709, Sample Num: 27344, Cur Loss: 0.88522714, Cur Avg Loss: 0.69301222, Log Avg loss: 0.80278995, Global Avg Loss: 2.00642698, Time: 0.0131 Steps: 42160, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001719, Sample Num: 27504, Cur Loss: 0.46332529, Cur Avg Loss: 0.69270791, Log Avg loss: 0.64070178, Global Avg Loss: 2.00610312, Time: 0.0221 Steps: 42170, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001729, Sample Num: 27664, Cur Loss: 0.86143386, Cur Avg Loss: 0.69332535, Log Avg loss: 0.79946220, Global Avg Loss: 2.00581705, Time: 0.0128 Steps: 42180, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001739, Sample Num: 27824, Cur Loss: 0.29503834, Cur Avg Loss: 0.69280769, Log Avg loss: 0.60330560, Global Avg Loss: 2.00548462, Time: 0.0064 Steps: 42190, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001749, Sample Num: 27984, Cur Loss: 1.13915694, Cur Avg Loss: 0.69368423, Log Avg loss: 0.84611430, Global Avg Loss: 2.00520989, Time: 0.0127 Steps: 42200, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001759, Sample Num: 28144, Cur Loss: 0.90630269, Cur Avg Loss: 0.69408104, Log Avg loss: 0.76348264, Global Avg Loss: 2.00491571, Time: 0.0112 Steps: 42210, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001769, Sample Num: 28304, Cur Loss: 0.42618620, Cur Avg Loss: 0.69347228, Log Avg loss: 0.58639138, Global Avg Loss: 2.00457973, Time: 0.0097 Steps: 42220, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001779, Sample Num: 28464, Cur Loss: 0.46483058, Cur Avg Loss: 0.69446692, Log Avg loss: 0.87041791, Global Avg Loss: 2.00431116, Time: 0.0102 Steps: 42230, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001789, Sample Num: 28624, Cur Loss: 0.96796823, Cur Avg Loss: 0.69459256, Log Avg loss: 0.71694397, Global Avg Loss: 2.00400639, Time: 0.0115 Steps: 42240, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001799, Sample Num: 28784, Cur Loss: 0.91862261, Cur Avg Loss: 0.69503700, Log Avg loss: 0.77454866, Global Avg Loss: 2.00371539, Time: 0.0086 Steps: 42250, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001809, Sample Num: 28944, Cur Loss: 0.45893598, Cur Avg Loss: 0.69534827, Log Avg loss: 0.75134472, Global Avg Loss: 2.00341904, Time: 0.0135 Steps: 42260, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001819, Sample Num: 29104, Cur Loss: 0.46761695, Cur Avg Loss: 0.69550184, Log Avg loss: 0.72328218, Global Avg Loss: 2.00311619, Time: 0.0066 Steps: 42270, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001829, Sample Num: 29264, Cur Loss: 0.86214107, Cur Avg Loss: 0.69548697, Log Avg loss: 0.69278282, Global Avg Loss: 2.00280627, Time: 0.0121 Steps: 42280, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001839, Sample Num: 29424, Cur Loss: 0.95064390, Cur Avg Loss: 0.69644533, Log Avg loss: 0.87172937, Global Avg Loss: 2.00253882, Time: 0.0120 Steps: 42290, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001849, Sample Num: 29584, Cur Loss: 0.63426203, Cur Avg Loss: 0.69658734, Log Avg loss: 0.72270361, Global Avg Loss: 2.00223626, Time: 0.0069 Steps: 42300, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001859, Sample Num: 29744, Cur Loss: 1.73215210, Cur Avg Loss: 0.69627420, Log Avg loss: 0.63837510, Global Avg Loss: 2.00191391, Time: 0.0104 Steps: 42310, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001869, Sample Num: 29904, Cur Loss: 0.44222420, Cur Avg Loss: 0.69598254, Log Avg loss: 0.64176180, Global Avg Loss: 2.00159251, Time: 0.0129 Steps: 42320, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001879, Sample Num: 30064, Cur Loss: 0.80955511, Cur Avg Loss: 0.69606624, Log Avg loss: 0.71171032, Global Avg Loss: 2.00128779, Time: 0.0118 Steps: 42330, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001889, Sample Num: 30224, Cur Loss: 0.24381626, Cur Avg Loss: 0.69611637, Log Avg loss: 0.70553545, Global Avg Loss: 2.00098175, Time: 0.0087 Steps: 42340, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001899, Sample Num: 30384, Cur Loss: 0.43805429, Cur Avg Loss: 0.69577926, Log Avg loss: 0.63209892, Global Avg Loss: 2.00065852, Time: 0.0102 Steps: 42350, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001909, Sample Num: 30544, Cur Loss: 1.31150663, Cur Avg Loss: 0.69640025, Log Avg loss: 0.81432692, Global Avg Loss: 2.00037846, Time: 0.0073 Steps: 42360, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001919, Sample Num: 30704, Cur Loss: 0.37409225, Cur Avg Loss: 0.69529273, Log Avg loss: 0.48386722, Global Avg Loss: 2.00002054, Time: 0.0069 Steps: 42370, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001929, Sample Num: 30864, Cur Loss: 0.21771282, Cur Avg Loss: 0.69470341, Log Avg loss: 0.58161216, Global Avg Loss: 1.99968585, Time: 0.0117 Steps: 42380, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001939, Sample Num: 31024, Cur Loss: 0.59052002, Cur Avg Loss: 0.69387230, Log Avg loss: 0.53355199, Global Avg Loss: 1.99933999, Time: 0.0110 Steps: 42390, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001949, Sample Num: 31184, Cur Loss: 0.66600800, Cur Avg Loss: 0.69361325, Log Avg loss: 0.64338367, Global Avg Loss: 1.99902019, Time: 0.0065 Steps: 42400, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001959, Sample Num: 31344, Cur Loss: 1.09675074, Cur Avg Loss: 0.69339243, Log Avg loss: 0.65035453, Global Avg Loss: 1.99870218, Time: 0.0119 Steps: 42410, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001969, Sample Num: 31504, Cur Loss: 0.38866523, Cur Avg Loss: 0.69339072, Log Avg loss: 0.69305543, Global Avg Loss: 1.99839439, Time: 0.0120 Steps: 42420, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001979, Sample Num: 31664, Cur Loss: 0.50096381, Cur Avg Loss: 0.69306964, Log Avg loss: 0.62984881, Global Avg Loss: 1.99807185, Time: 0.0068 Steps: 42430, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001989, Sample Num: 31824, Cur Loss: 0.83989644, Cur Avg Loss: 0.69301264, Log Avg loss: 0.68173267, Global Avg Loss: 1.99776168, Time: 0.0118 Steps: 42440, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 001999, Sample Num: 31984, Cur Loss: 0.51233929, Cur Avg Loss: 0.69284846, Log Avg loss: 0.66019274, Global Avg Loss: 1.99744659, Time: 0.0067 Steps: 42450, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002009, Sample Num: 32144, Cur Loss: 0.64531291, Cur Avg Loss: 0.69291464, Log Avg loss: 0.70614305, Global Avg Loss: 1.99714247, Time: 0.0121 Steps: 42460, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002019, Sample Num: 32304, Cur Loss: 0.22545946, Cur Avg Loss: 0.69275715, Log Avg loss: 0.66111727, Global Avg Loss: 1.99682789, Time: 0.0115 Steps: 42470, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002029, Sample Num: 32464, Cur Loss: 0.28024495, Cur Avg Loss: 0.69243908, Log Avg loss: 0.62822146, Global Avg Loss: 1.99650571, Time: 0.0134 Steps: 42480, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002039, Sample Num: 32624, Cur Loss: 0.28561431, Cur Avg Loss: 0.69299795, Log Avg loss: 0.80639313, Global Avg Loss: 1.99622562, Time: 0.0112 Steps: 42490, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002049, Sample Num: 32784, Cur Loss: 0.84238017, Cur Avg Loss: 0.69262964, Log Avg loss: 0.61753073, Global Avg Loss: 1.99590122, Time: 0.0178 Steps: 42500, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002059, Sample Num: 32944, Cur Loss: 1.28738594, Cur Avg Loss: 0.69287172, Log Avg loss: 0.74247326, Global Avg Loss: 1.99560636, Time: 0.0077 Steps: 42510, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002069, Sample Num: 33104, Cur Loss: 1.10713351, Cur Avg Loss: 0.69353586, Log Avg loss: 0.83028372, Global Avg Loss: 1.99533230, Time: 0.0110 Steps: 42520, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002079, Sample Num: 33264, Cur Loss: 0.35980177, Cur Avg Loss: 0.69248499, Log Avg loss: 0.47506015, Global Avg Loss: 1.99497484, Time: 0.0069 Steps: 42530, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002089, Sample Num: 33424, Cur Loss: 0.33492902, Cur Avg Loss: 0.69148043, Log Avg loss: 0.48263130, Global Avg Loss: 1.99461933, Time: 0.0225 Steps: 42540, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002099, Sample Num: 33584, Cur Loss: 0.44478372, Cur Avg Loss: 0.69018153, Log Avg loss: 0.41884174, Global Avg Loss: 1.99424899, Time: 0.0091 Steps: 42550, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002109, Sample Num: 33744, Cur Loss: 0.47069752, Cur Avg Loss: 0.69060149, Log Avg loss: 0.77875135, Global Avg Loss: 1.99396340, Time: 0.0109 Steps: 42560, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002119, Sample Num: 33904, Cur Loss: 0.83766758, Cur Avg Loss: 0.69034574, Log Avg loss: 0.63640756, Global Avg Loss: 1.99364450, Time: 0.0157 Steps: 42570, Updated lr: 0.000061 Training, Epoch: 0020, Batch: 002129, Sample Num: 34055, Cur Loss: 0.06807954, Cur Avg Loss: 0.68989777, Log Avg loss: 0.59497327, Global Avg Loss: 1.99331602, Time: 0.0036 Steps: 42580, Updated lr: 0.000061 ***** Running evaluation checkpoint-42580 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-42580 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.359771, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.886886, "eval_total_loss": 623.480509, "eval_mae": 0.784069, "eval_mse": 0.886934, "eval_r2": 0.436206, "eval_sp_statistic": 0.651375, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.72021, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.151954, "test_total_loss": 578.280928, "test_mae": 0.921944, "test_mse": 1.151956, "test_r2": 0.256517, "test_sp_statistic": 0.495151, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.605664, "test_ps_pvalue": 0.0, "lr": 6.056899004267425e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.9933160166002848, "train_cur_epoch_loss": 1468.7923530638218, "train_cur_epoch_avg_loss": 0.6898977703446791, "train_cur_epoch_time": 22.35977077484131, "train_cur_epoch_avg_time": 0.010502475704481592, "epoch": 20, "step": 42580} ################################################## Training, Epoch: 0021, Batch: 000010, Sample Num: 160, Cur Loss: 0.36085328, Cur Avg Loss: 0.76464287, Log Avg loss: 0.76464287, Global Avg Loss: 1.99302753, Time: 0.0116 Steps: 42590, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000020, Sample Num: 320, Cur Loss: 0.36052966, Cur Avg Loss: 0.64554172, Log Avg loss: 0.52644057, Global Avg Loss: 1.99268326, Time: 0.0065 Steps: 42600, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000030, Sample Num: 480, Cur Loss: 0.31136316, Cur Avg Loss: 0.58649952, Log Avg loss: 0.46841511, Global Avg Loss: 1.99232553, Time: 0.0116 Steps: 42610, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000040, Sample Num: 640, Cur Loss: 0.20721604, Cur Avg Loss: 0.64576131, Log Avg loss: 0.82354669, Global Avg Loss: 1.99205130, Time: 0.0121 Steps: 42620, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000050, Sample Num: 800, Cur Loss: 0.83691615, Cur Avg Loss: 0.65748950, Log Avg loss: 0.70440227, Global Avg Loss: 1.99174925, Time: 0.0118 Steps: 42630, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000060, Sample Num: 960, Cur Loss: 0.29446322, Cur Avg Loss: 0.65739546, Log Avg loss: 0.65692523, Global Avg Loss: 1.99143620, Time: 0.0098 Steps: 42640, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000070, Sample Num: 1120, Cur Loss: 0.32039863, Cur Avg Loss: 0.63732881, Log Avg loss: 0.51692893, Global Avg Loss: 1.99109048, Time: 0.0105 Steps: 42650, Updated lr: 0.000061 Training, Epoch: 0021, Batch: 000080, Sample Num: 1280, Cur Loss: 0.42884719, Cur Avg Loss: 0.62872789, Log Avg loss: 0.56852147, Global Avg Loss: 1.99075701, Time: 0.0116 Steps: 42660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000090, Sample Num: 1440, Cur Loss: 0.36538431, Cur Avg Loss: 0.62879059, Log Avg loss: 0.62929216, Global Avg Loss: 1.99043795, Time: 0.0106 Steps: 42670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000100, Sample Num: 1600, Cur Loss: 1.27523744, Cur Avg Loss: 0.64389055, Log Avg loss: 0.77979021, Global Avg Loss: 1.99015429, Time: 0.0118 Steps: 42680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000110, Sample Num: 1760, Cur Loss: 0.71183002, Cur Avg Loss: 0.65272546, Log Avg loss: 0.74107458, Global Avg Loss: 1.98986170, Time: 0.0070 Steps: 42690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000120, Sample Num: 1920, Cur Loss: 1.26862335, Cur Avg Loss: 0.65548975, Log Avg loss: 0.68589689, Global Avg Loss: 1.98955632, Time: 0.0067 Steps: 42700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000130, Sample Num: 2080, Cur Loss: 0.49626714, Cur Avg Loss: 0.64817455, Log Avg loss: 0.56039210, Global Avg Loss: 1.98922170, Time: 0.0067 Steps: 42710, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000140, Sample Num: 2240, Cur Loss: 0.35695961, Cur Avg Loss: 0.64420643, Log Avg loss: 0.59262091, Global Avg Loss: 1.98889478, Time: 0.0067 Steps: 42720, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000150, Sample Num: 2400, Cur Loss: 0.61813903, Cur Avg Loss: 0.64783701, Log Avg loss: 0.69866519, Global Avg Loss: 1.98859283, Time: 0.0231 Steps: 42730, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000160, Sample Num: 2560, Cur Loss: 0.84127557, Cur Avg Loss: 0.65086831, Log Avg loss: 0.69633784, Global Avg Loss: 1.98829048, Time: 0.0066 Steps: 42740, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000170, Sample Num: 2720, Cur Loss: 0.77475953, Cur Avg Loss: 0.65767224, Log Avg loss: 0.76653498, Global Avg Loss: 1.98800468, Time: 0.0066 Steps: 42750, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000180, Sample Num: 2880, Cur Loss: 0.70146656, Cur Avg Loss: 0.65233595, Log Avg loss: 0.56161903, Global Avg Loss: 1.98767111, Time: 0.0106 Steps: 42760, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000190, Sample Num: 3040, Cur Loss: 0.84363079, Cur Avg Loss: 0.64664491, Log Avg loss: 0.54420630, Global Avg Loss: 1.98733361, Time: 0.0115 Steps: 42770, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000200, Sample Num: 3200, Cur Loss: 0.68361175, Cur Avg Loss: 0.65147167, Log Avg loss: 0.74318008, Global Avg Loss: 1.98704278, Time: 0.0065 Steps: 42780, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000210, Sample Num: 3360, Cur Loss: 1.72581577, Cur Avg Loss: 0.65686022, Log Avg loss: 0.76463113, Global Avg Loss: 1.98675711, Time: 0.0066 Steps: 42790, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000220, Sample Num: 3520, Cur Loss: 0.59801614, Cur Avg Loss: 0.66090877, Log Avg loss: 0.74592835, Global Avg Loss: 1.98646719, Time: 0.0066 Steps: 42800, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000230, Sample Num: 3680, Cur Loss: 0.63265103, Cur Avg Loss: 0.65682932, Log Avg loss: 0.56708151, Global Avg Loss: 1.98613564, Time: 0.0112 Steps: 42810, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000240, Sample Num: 3840, Cur Loss: 0.50897646, Cur Avg Loss: 0.66153360, Log Avg loss: 0.76973191, Global Avg Loss: 1.98585157, Time: 0.0118 Steps: 42820, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000250, Sample Num: 4000, Cur Loss: 1.07822597, Cur Avg Loss: 0.65982448, Log Avg loss: 0.61880571, Global Avg Loss: 1.98553239, Time: 0.0076 Steps: 42830, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000260, Sample Num: 4160, Cur Loss: 0.69539207, Cur Avg Loss: 0.66415526, Log Avg loss: 0.77242473, Global Avg Loss: 1.98524921, Time: 0.0110 Steps: 42840, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000270, Sample Num: 4320, Cur Loss: 0.43219888, Cur Avg Loss: 0.65897965, Log Avg loss: 0.52441384, Global Avg Loss: 1.98490830, Time: 0.0191 Steps: 42850, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000280, Sample Num: 4480, Cur Loss: 0.52884227, Cur Avg Loss: 0.65900973, Log Avg loss: 0.65982176, Global Avg Loss: 1.98459913, Time: 0.0071 Steps: 42860, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000290, Sample Num: 4640, Cur Loss: 0.52961433, Cur Avg Loss: 0.65591350, Log Avg loss: 0.56921912, Global Avg Loss: 1.98426897, Time: 0.0114 Steps: 42870, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000300, Sample Num: 4800, Cur Loss: 0.83613336, Cur Avg Loss: 0.66196499, Log Avg loss: 0.83745835, Global Avg Loss: 1.98400153, Time: 0.0106 Steps: 42880, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000310, Sample Num: 4960, Cur Loss: 1.00362992, Cur Avg Loss: 0.66249637, Log Avg loss: 0.67843779, Global Avg Loss: 1.98369713, Time: 0.0072 Steps: 42890, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000320, Sample Num: 5120, Cur Loss: 0.34259903, Cur Avg Loss: 0.65931219, Log Avg loss: 0.56060238, Global Avg Loss: 1.98336541, Time: 0.0066 Steps: 42900, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000330, Sample Num: 5280, Cur Loss: 0.57171613, Cur Avg Loss: 0.66055987, Log Avg loss: 0.70048565, Global Avg Loss: 1.98306644, Time: 0.0066 Steps: 42910, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000340, Sample Num: 5440, Cur Loss: 0.69469619, Cur Avg Loss: 0.66289399, Log Avg loss: 0.73991984, Global Avg Loss: 1.98277679, Time: 0.0118 Steps: 42920, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000350, Sample Num: 5600, Cur Loss: 0.67749244, Cur Avg Loss: 0.66288359, Log Avg loss: 0.66253020, Global Avg Loss: 1.98246926, Time: 0.0120 Steps: 42930, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000360, Sample Num: 5760, Cur Loss: 0.96605408, Cur Avg Loss: 0.66802262, Log Avg loss: 0.84788861, Global Avg Loss: 1.98220503, Time: 0.0134 Steps: 42940, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000370, Sample Num: 5920, Cur Loss: 0.44049329, Cur Avg Loss: 0.66476854, Log Avg loss: 0.54762150, Global Avg Loss: 1.98187102, Time: 0.0067 Steps: 42950, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000380, Sample Num: 6080, Cur Loss: 0.96543491, Cur Avg Loss: 0.66683314, Log Avg loss: 0.74322365, Global Avg Loss: 1.98158270, Time: 0.0067 Steps: 42960, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000390, Sample Num: 6240, Cur Loss: 0.37175035, Cur Avg Loss: 0.66302944, Log Avg loss: 0.51848884, Global Avg Loss: 1.98124220, Time: 0.0118 Steps: 42970, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000400, Sample Num: 6400, Cur Loss: 0.46139252, Cur Avg Loss: 0.66918662, Log Avg loss: 0.90931642, Global Avg Loss: 1.98099280, Time: 0.0073 Steps: 42980, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000410, Sample Num: 6560, Cur Loss: 1.32532334, Cur Avg Loss: 0.67189893, Log Avg loss: 0.78039160, Global Avg Loss: 1.98071353, Time: 0.0113 Steps: 42990, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000420, Sample Num: 6720, Cur Loss: 0.53674823, Cur Avg Loss: 0.67129171, Log Avg loss: 0.64639563, Global Avg Loss: 1.98040322, Time: 0.0067 Steps: 43000, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000430, Sample Num: 6880, Cur Loss: 1.18200541, Cur Avg Loss: 0.67187027, Log Avg loss: 0.69616956, Global Avg Loss: 1.98010463, Time: 0.0067 Steps: 43010, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000440, Sample Num: 7040, Cur Loss: 1.14485693, Cur Avg Loss: 0.67130209, Log Avg loss: 0.64687046, Global Avg Loss: 1.97979472, Time: 0.0108 Steps: 43020, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000450, Sample Num: 7200, Cur Loss: 0.29698762, Cur Avg Loss: 0.66933486, Log Avg loss: 0.58277654, Global Avg Loss: 1.97947006, Time: 0.0066 Steps: 43030, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000460, Sample Num: 7360, Cur Loss: 0.85113859, Cur Avg Loss: 0.67055718, Log Avg loss: 0.72556166, Global Avg Loss: 1.97917872, Time: 0.0065 Steps: 43040, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000470, Sample Num: 7520, Cur Loss: 0.50862610, Cur Avg Loss: 0.66767246, Log Avg loss: 0.53497545, Global Avg Loss: 1.97884325, Time: 0.0071 Steps: 43050, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000480, Sample Num: 7680, Cur Loss: 0.39176130, Cur Avg Loss: 0.66612659, Log Avg loss: 0.59347073, Global Avg Loss: 1.97852152, Time: 0.0065 Steps: 43060, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000490, Sample Num: 7840, Cur Loss: 1.46167231, Cur Avg Loss: 0.66658300, Log Avg loss: 0.68849080, Global Avg Loss: 1.97822200, Time: 0.0123 Steps: 43070, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000500, Sample Num: 8000, Cur Loss: 0.24736254, Cur Avg Loss: 0.66602677, Log Avg loss: 0.63877137, Global Avg Loss: 1.97791108, Time: 0.0067 Steps: 43080, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000510, Sample Num: 8160, Cur Loss: 0.62093657, Cur Avg Loss: 0.67131173, Log Avg loss: 0.93555968, Global Avg Loss: 1.97766918, Time: 0.0067 Steps: 43090, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000520, Sample Num: 8320, Cur Loss: 0.30600590, Cur Avg Loss: 0.67171335, Log Avg loss: 0.69219616, Global Avg Loss: 1.97737093, Time: 0.0069 Steps: 43100, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000530, Sample Num: 8480, Cur Loss: 0.73760986, Cur Avg Loss: 0.67066648, Log Avg loss: 0.61622932, Global Avg Loss: 1.97705519, Time: 0.0120 Steps: 43110, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000540, Sample Num: 8640, Cur Loss: 0.89829141, Cur Avg Loss: 0.66784329, Log Avg loss: 0.51821421, Global Avg Loss: 1.97671687, Time: 0.0118 Steps: 43120, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000550, Sample Num: 8800, Cur Loss: 0.62462986, Cur Avg Loss: 0.66597136, Log Avg loss: 0.56488676, Global Avg Loss: 1.97638953, Time: 0.0068 Steps: 43130, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000560, Sample Num: 8960, Cur Loss: 0.60312635, Cur Avg Loss: 0.66269347, Log Avg loss: 0.48240988, Global Avg Loss: 1.97604322, Time: 0.0112 Steps: 43140, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000570, Sample Num: 9120, Cur Loss: 0.52230698, Cur Avg Loss: 0.66250892, Log Avg loss: 0.65217406, Global Avg Loss: 1.97573641, Time: 0.0067 Steps: 43150, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000580, Sample Num: 9280, Cur Loss: 0.28447104, Cur Avg Loss: 0.66054948, Log Avg loss: 0.54886123, Global Avg Loss: 1.97540581, Time: 0.0114 Steps: 43160, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000590, Sample Num: 9440, Cur Loss: 1.08259368, Cur Avg Loss: 0.66128759, Log Avg loss: 0.70409789, Global Avg Loss: 1.97511132, Time: 0.0106 Steps: 43170, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000600, Sample Num: 9600, Cur Loss: 0.13999093, Cur Avg Loss: 0.66157710, Log Avg loss: 0.67865848, Global Avg Loss: 1.97481108, Time: 0.0132 Steps: 43180, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000610, Sample Num: 9760, Cur Loss: 0.44188908, Cur Avg Loss: 0.66017421, Log Avg loss: 0.57600043, Global Avg Loss: 1.97448720, Time: 0.0078 Steps: 43190, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000620, Sample Num: 9920, Cur Loss: 0.68319017, Cur Avg Loss: 0.66183796, Log Avg loss: 0.76332691, Global Avg Loss: 1.97420684, Time: 0.0068 Steps: 43200, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000630, Sample Num: 10080, Cur Loss: 0.68222380, Cur Avg Loss: 0.66060005, Log Avg loss: 0.58384993, Global Avg Loss: 1.97388507, Time: 0.0116 Steps: 43210, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000640, Sample Num: 10240, Cur Loss: 0.35030127, Cur Avg Loss: 0.66052963, Log Avg loss: 0.65609310, Global Avg Loss: 1.97358017, Time: 0.0148 Steps: 43220, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000650, Sample Num: 10400, Cur Loss: 0.36782825, Cur Avg Loss: 0.65980156, Log Avg loss: 0.61320505, Global Avg Loss: 1.97326549, Time: 0.0118 Steps: 43230, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000660, Sample Num: 10560, Cur Loss: 1.37766480, Cur Avg Loss: 0.66168673, Log Avg loss: 0.78422245, Global Avg Loss: 1.97299050, Time: 0.0109 Steps: 43240, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000670, Sample Num: 10720, Cur Loss: 0.47546756, Cur Avg Loss: 0.66046429, Log Avg loss: 0.57978354, Global Avg Loss: 1.97266837, Time: 0.0118 Steps: 43250, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000680, Sample Num: 10880, Cur Loss: 0.81839073, Cur Avg Loss: 0.66006154, Log Avg loss: 0.63307730, Global Avg Loss: 1.97235871, Time: 0.0074 Steps: 43260, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000690, Sample Num: 11040, Cur Loss: 0.46038815, Cur Avg Loss: 0.65936686, Log Avg loss: 0.61212874, Global Avg Loss: 1.97204435, Time: 0.0073 Steps: 43270, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000700, Sample Num: 11200, Cur Loss: 0.39714465, Cur Avg Loss: 0.65892728, Log Avg loss: 0.62859599, Global Avg Loss: 1.97173394, Time: 0.0118 Steps: 43280, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000710, Sample Num: 11360, Cur Loss: 0.30915794, Cur Avg Loss: 0.65891887, Log Avg loss: 0.65832994, Global Avg Loss: 1.97143055, Time: 0.0109 Steps: 43290, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000720, Sample Num: 11520, Cur Loss: 0.37920737, Cur Avg Loss: 0.65869818, Log Avg loss: 0.64302931, Global Avg Loss: 1.97112376, Time: 0.0118 Steps: 43300, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000730, Sample Num: 11680, Cur Loss: 0.39429247, Cur Avg Loss: 0.65629773, Log Avg loss: 0.48346517, Global Avg Loss: 1.97078027, Time: 0.0073 Steps: 43310, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000740, Sample Num: 11840, Cur Loss: 0.35196671, Cur Avg Loss: 0.65731266, Log Avg loss: 0.73140260, Global Avg Loss: 1.97049417, Time: 0.0119 Steps: 43320, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000750, Sample Num: 12000, Cur Loss: 0.69909990, Cur Avg Loss: 0.65701529, Log Avg loss: 0.63501045, Global Avg Loss: 1.97018596, Time: 0.0068 Steps: 43330, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000760, Sample Num: 12160, Cur Loss: 1.40895975, Cur Avg Loss: 0.65821835, Log Avg loss: 0.74844774, Global Avg Loss: 1.96990406, Time: 0.0068 Steps: 43340, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000770, Sample Num: 12320, Cur Loss: 0.69649971, Cur Avg Loss: 0.65949932, Log Avg loss: 0.75685287, Global Avg Loss: 1.96962423, Time: 0.0078 Steps: 43350, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000780, Sample Num: 12480, Cur Loss: 1.23395491, Cur Avg Loss: 0.66075237, Log Avg loss: 0.75723730, Global Avg Loss: 1.96934462, Time: 0.0144 Steps: 43360, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000790, Sample Num: 12640, Cur Loss: 1.23120010, Cur Avg Loss: 0.66201295, Log Avg loss: 0.76033764, Global Avg Loss: 1.96906586, Time: 0.0116 Steps: 43370, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000800, Sample Num: 12800, Cur Loss: 0.30439246, Cur Avg Loss: 0.66263135, Log Avg loss: 0.71148551, Global Avg Loss: 1.96877596, Time: 0.0072 Steps: 43380, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000810, Sample Num: 12960, Cur Loss: 0.77179831, Cur Avg Loss: 0.66321725, Log Avg loss: 0.71008924, Global Avg Loss: 1.96848587, Time: 0.0109 Steps: 43390, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000820, Sample Num: 13120, Cur Loss: 0.72845972, Cur Avg Loss: 0.66454985, Log Avg loss: 0.77249040, Global Avg Loss: 1.96821030, Time: 0.0117 Steps: 43400, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000830, Sample Num: 13280, Cur Loss: 0.69708848, Cur Avg Loss: 0.66399196, Log Avg loss: 0.61824459, Global Avg Loss: 1.96789932, Time: 0.0098 Steps: 43410, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000840, Sample Num: 13440, Cur Loss: 0.73340619, Cur Avg Loss: 0.66251915, Log Avg loss: 0.54027625, Global Avg Loss: 1.96757052, Time: 0.0066 Steps: 43420, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000850, Sample Num: 13600, Cur Loss: 1.37044597, Cur Avg Loss: 0.66333049, Log Avg loss: 0.73148260, Global Avg Loss: 1.96728591, Time: 0.0068 Steps: 43430, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000860, Sample Num: 13760, Cur Loss: 0.76861465, Cur Avg Loss: 0.66292467, Log Avg loss: 0.62843003, Global Avg Loss: 1.96697770, Time: 0.0068 Steps: 43440, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000870, Sample Num: 13920, Cur Loss: 1.05555880, Cur Avg Loss: 0.66360539, Log Avg loss: 0.72214725, Global Avg Loss: 1.96669120, Time: 0.0110 Steps: 43450, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000880, Sample Num: 14080, Cur Loss: 0.48095471, Cur Avg Loss: 0.66273233, Log Avg loss: 0.58677682, Global Avg Loss: 1.96637369, Time: 0.0077 Steps: 43460, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000890, Sample Num: 14240, Cur Loss: 1.14178216, Cur Avg Loss: 0.66263249, Log Avg loss: 0.65384595, Global Avg Loss: 1.96607175, Time: 0.0066 Steps: 43470, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000900, Sample Num: 14400, Cur Loss: 1.57349277, Cur Avg Loss: 0.66594387, Log Avg loss: 0.96065703, Global Avg Loss: 1.96584051, Time: 0.0114 Steps: 43480, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000910, Sample Num: 14560, Cur Loss: 0.78696334, Cur Avg Loss: 0.66707595, Log Avg loss: 0.76896319, Global Avg Loss: 1.96556530, Time: 0.0117 Steps: 43490, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000920, Sample Num: 14720, Cur Loss: 1.63337374, Cur Avg Loss: 0.67100150, Log Avg loss: 1.02822605, Global Avg Loss: 1.96534982, Time: 0.0087 Steps: 43500, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000930, Sample Num: 14880, Cur Loss: 0.92827046, Cur Avg Loss: 0.67024743, Log Avg loss: 0.60087366, Global Avg Loss: 1.96503622, Time: 0.0068 Steps: 43510, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000940, Sample Num: 15040, Cur Loss: 0.43742144, Cur Avg Loss: 0.67060975, Log Avg loss: 0.70430495, Global Avg Loss: 1.96474653, Time: 0.0157 Steps: 43520, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000950, Sample Num: 15200, Cur Loss: 0.55040234, Cur Avg Loss: 0.67183646, Log Avg loss: 0.78714763, Global Avg Loss: 1.96447601, Time: 0.0067 Steps: 43530, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000960, Sample Num: 15360, Cur Loss: 0.26076952, Cur Avg Loss: 0.66956776, Log Avg loss: 0.45404068, Global Avg Loss: 1.96412910, Time: 0.0090 Steps: 43540, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000970, Sample Num: 15520, Cur Loss: 0.88738072, Cur Avg Loss: 0.67104487, Log Avg loss: 0.81284745, Global Avg Loss: 1.96386474, Time: 0.0070 Steps: 43550, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000980, Sample Num: 15680, Cur Loss: 0.94056523, Cur Avg Loss: 0.67091574, Log Avg loss: 0.65839028, Global Avg Loss: 1.96356505, Time: 0.0073 Steps: 43560, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 000990, Sample Num: 15840, Cur Loss: 0.41772810, Cur Avg Loss: 0.67038676, Log Avg loss: 0.61854703, Global Avg Loss: 1.96325634, Time: 0.0119 Steps: 43570, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001000, Sample Num: 16000, Cur Loss: 0.32572514, Cur Avg Loss: 0.66823496, Log Avg loss: 0.45520669, Global Avg Loss: 1.96291030, Time: 0.0125 Steps: 43580, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001010, Sample Num: 16160, Cur Loss: 0.30674100, Cur Avg Loss: 0.66872565, Log Avg loss: 0.71779481, Global Avg Loss: 1.96262466, Time: 0.0068 Steps: 43590, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001020, Sample Num: 16320, Cur Loss: 0.98400730, Cur Avg Loss: 0.66963087, Log Avg loss: 0.76105803, Global Avg Loss: 1.96234907, Time: 0.0115 Steps: 43600, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001030, Sample Num: 16480, Cur Loss: 1.16367221, Cur Avg Loss: 0.67071333, Log Avg loss: 0.78112421, Global Avg Loss: 1.96207821, Time: 0.0115 Steps: 43610, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001040, Sample Num: 16640, Cur Loss: 0.49384993, Cur Avg Loss: 0.67105133, Log Avg loss: 0.70586534, Global Avg Loss: 1.96179022, Time: 0.0120 Steps: 43620, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001050, Sample Num: 16800, Cur Loss: 0.46733758, Cur Avg Loss: 0.66927148, Log Avg loss: 0.48416717, Global Avg Loss: 1.96145155, Time: 0.0084 Steps: 43630, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001060, Sample Num: 16960, Cur Loss: 0.85060823, Cur Avg Loss: 0.66977688, Log Avg loss: 0.72284423, Global Avg Loss: 1.96116772, Time: 0.0112 Steps: 43640, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001070, Sample Num: 17120, Cur Loss: 1.08637691, Cur Avg Loss: 0.67078989, Log Avg loss: 0.77816826, Global Avg Loss: 1.96089670, Time: 0.0067 Steps: 43650, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001080, Sample Num: 17280, Cur Loss: 0.44721884, Cur Avg Loss: 0.67070232, Log Avg loss: 0.66133297, Global Avg Loss: 1.96059905, Time: 0.0105 Steps: 43660, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001090, Sample Num: 17440, Cur Loss: 0.72306871, Cur Avg Loss: 0.67149063, Log Avg loss: 0.75662785, Global Avg Loss: 1.96032335, Time: 0.0119 Steps: 43670, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001100, Sample Num: 17600, Cur Loss: 0.26382488, Cur Avg Loss: 0.67057716, Log Avg loss: 0.57100887, Global Avg Loss: 1.96000529, Time: 0.0109 Steps: 43680, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001110, Sample Num: 17760, Cur Loss: 0.27416354, Cur Avg Loss: 0.66969967, Log Avg loss: 0.57317593, Global Avg Loss: 1.95968786, Time: 0.0105 Steps: 43690, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001120, Sample Num: 17920, Cur Loss: 0.19551796, Cur Avg Loss: 0.67032681, Log Avg loss: 0.73993867, Global Avg Loss: 1.95940874, Time: 0.0066 Steps: 43700, Updated lr: 0.000060 Training, Epoch: 0021, Batch: 001130, Sample Num: 18080, Cur Loss: 0.53544164, Cur Avg Loss: 0.66970029, Log Avg loss: 0.59953045, Global Avg Loss: 1.95909763, Time: 0.0090 Steps: 43710, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001140, Sample Num: 18240, Cur Loss: 0.86754161, Cur Avg Loss: 0.67165273, Log Avg loss: 0.89227848, Global Avg Loss: 1.95885362, Time: 0.0071 Steps: 43720, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001150, Sample Num: 18400, Cur Loss: 0.81102961, Cur Avg Loss: 0.67115587, Log Avg loss: 0.61451327, Global Avg Loss: 1.95854620, Time: 0.0083 Steps: 43730, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001160, Sample Num: 18560, Cur Loss: 0.61590755, Cur Avg Loss: 0.66942277, Log Avg loss: 0.47011647, Global Avg Loss: 1.95820591, Time: 0.0154 Steps: 43740, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001170, Sample Num: 18720, Cur Loss: 0.52314734, Cur Avg Loss: 0.66921189, Log Avg loss: 0.64474981, Global Avg Loss: 1.95790569, Time: 0.0125 Steps: 43750, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001180, Sample Num: 18880, Cur Loss: 1.22283804, Cur Avg Loss: 0.67025051, Log Avg loss: 0.79176889, Global Avg Loss: 1.95763920, Time: 0.0166 Steps: 43760, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001190, Sample Num: 19040, Cur Loss: 0.41823703, Cur Avg Loss: 0.67161243, Log Avg loss: 0.83231908, Global Avg Loss: 1.95738211, Time: 0.0125 Steps: 43770, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001200, Sample Num: 19200, Cur Loss: 0.62389582, Cur Avg Loss: 0.67216425, Log Avg loss: 0.73783081, Global Avg Loss: 1.95710354, Time: 0.0100 Steps: 43780, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001210, Sample Num: 19360, Cur Loss: 1.60867620, Cur Avg Loss: 0.67248113, Log Avg loss: 0.71050666, Global Avg Loss: 1.95681887, Time: 0.0116 Steps: 43790, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001220, Sample Num: 19520, Cur Loss: 0.55587512, Cur Avg Loss: 0.67334934, Log Avg loss: 0.77840315, Global Avg Loss: 1.95654982, Time: 0.0120 Steps: 43800, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001230, Sample Num: 19680, Cur Loss: 0.39574319, Cur Avg Loss: 0.67303362, Log Avg loss: 0.63451605, Global Avg Loss: 1.95624806, Time: 0.0121 Steps: 43810, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001240, Sample Num: 19840, Cur Loss: 0.98596478, Cur Avg Loss: 0.67368402, Log Avg loss: 0.75368279, Global Avg Loss: 1.95597362, Time: 0.0070 Steps: 43820, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001250, Sample Num: 20000, Cur Loss: 0.17661628, Cur Avg Loss: 0.67233833, Log Avg loss: 0.50547343, Global Avg Loss: 1.95564269, Time: 0.0121 Steps: 43830, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001260, Sample Num: 20160, Cur Loss: 1.08278334, Cur Avg Loss: 0.67109832, Log Avg loss: 0.51609715, Global Avg Loss: 1.95531432, Time: 0.0116 Steps: 43840, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001270, Sample Num: 20320, Cur Loss: 0.38290757, Cur Avg Loss: 0.67242698, Log Avg loss: 0.83983826, Global Avg Loss: 1.95505994, Time: 0.0067 Steps: 43850, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001280, Sample Num: 20480, Cur Loss: 0.79814386, Cur Avg Loss: 0.67278505, Log Avg loss: 0.71825886, Global Avg Loss: 1.95477795, Time: 0.0130 Steps: 43860, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001290, Sample Num: 20640, Cur Loss: 0.31085825, Cur Avg Loss: 0.67173031, Log Avg loss: 0.53672400, Global Avg Loss: 1.95445471, Time: 0.0131 Steps: 43870, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001300, Sample Num: 20800, Cur Loss: 0.50692046, Cur Avg Loss: 0.67147377, Log Avg loss: 0.63838007, Global Avg Loss: 1.95415478, Time: 0.0109 Steps: 43880, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001310, Sample Num: 20960, Cur Loss: 0.49669129, Cur Avg Loss: 0.67124687, Log Avg loss: 0.64174982, Global Avg Loss: 1.95385576, Time: 0.0067 Steps: 43890, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001320, Sample Num: 21120, Cur Loss: 0.80681038, Cur Avg Loss: 0.67129162, Log Avg loss: 0.67715362, Global Avg Loss: 1.95356494, Time: 0.0091 Steps: 43900, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001330, Sample Num: 21280, Cur Loss: 0.67196345, Cur Avg Loss: 0.67154971, Log Avg loss: 0.70561794, Global Avg Loss: 1.95328074, Time: 0.0107 Steps: 43910, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001340, Sample Num: 21440, Cur Loss: 0.29931760, Cur Avg Loss: 0.67127314, Log Avg loss: 0.63448920, Global Avg Loss: 1.95298046, Time: 0.0118 Steps: 43920, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001350, Sample Num: 21600, Cur Loss: 0.32361400, Cur Avg Loss: 0.67134877, Log Avg loss: 0.68148341, Global Avg Loss: 1.95269103, Time: 0.0091 Steps: 43930, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001360, Sample Num: 21760, Cur Loss: 1.50675452, Cur Avg Loss: 0.67238075, Log Avg loss: 0.81169817, Global Avg Loss: 1.95243136, Time: 0.0129 Steps: 43940, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001370, Sample Num: 21920, Cur Loss: 0.38706523, Cur Avg Loss: 0.67231673, Log Avg loss: 0.66360937, Global Avg Loss: 1.95213811, Time: 0.0159 Steps: 43950, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001380, Sample Num: 22080, Cur Loss: 0.82556272, Cur Avg Loss: 0.67242483, Log Avg loss: 0.68723540, Global Avg Loss: 1.95185037, Time: 0.0068 Steps: 43960, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001390, Sample Num: 22240, Cur Loss: 0.45486334, Cur Avg Loss: 0.67226078, Log Avg loss: 0.64962191, Global Avg Loss: 1.95155421, Time: 0.0067 Steps: 43970, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001400, Sample Num: 22400, Cur Loss: 0.20089650, Cur Avg Loss: 0.67264481, Log Avg loss: 0.72602406, Global Avg Loss: 1.95127555, Time: 0.0118 Steps: 43980, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001410, Sample Num: 22560, Cur Loss: 0.91880137, Cur Avg Loss: 0.67397590, Log Avg loss: 0.86032969, Global Avg Loss: 1.95102755, Time: 0.0068 Steps: 43990, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001420, Sample Num: 22720, Cur Loss: 1.23504114, Cur Avg Loss: 0.67393061, Log Avg loss: 0.66754422, Global Avg Loss: 1.95073585, Time: 0.0109 Steps: 44000, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001430, Sample Num: 22880, Cur Loss: 0.83422393, Cur Avg Loss: 0.67379643, Log Avg loss: 0.65474252, Global Avg Loss: 1.95044137, Time: 0.0113 Steps: 44010, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001440, Sample Num: 23040, Cur Loss: 0.70665896, Cur Avg Loss: 0.67391640, Log Avg loss: 0.69107283, Global Avg Loss: 1.95015528, Time: 0.0161 Steps: 44020, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001450, Sample Num: 23200, Cur Loss: 0.52530348, Cur Avg Loss: 0.67402543, Log Avg loss: 0.68972547, Global Avg Loss: 1.94986902, Time: 0.0139 Steps: 44030, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001460, Sample Num: 23360, Cur Loss: 0.33979341, Cur Avg Loss: 0.67490628, Log Avg loss: 0.80262984, Global Avg Loss: 1.94960852, Time: 0.0069 Steps: 44040, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001470, Sample Num: 23520, Cur Loss: 0.41508943, Cur Avg Loss: 0.67347484, Log Avg loss: 0.46448463, Global Avg Loss: 1.94927137, Time: 0.0115 Steps: 44050, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001480, Sample Num: 23680, Cur Loss: 0.51979107, Cur Avg Loss: 0.67333102, Log Avg loss: 0.65218848, Global Avg Loss: 1.94897698, Time: 0.0126 Steps: 44060, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001490, Sample Num: 23840, Cur Loss: 0.62437600, Cur Avg Loss: 0.67334232, Log Avg loss: 0.67501548, Global Avg Loss: 1.94868791, Time: 0.0114 Steps: 44070, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001500, Sample Num: 24000, Cur Loss: 0.26572257, Cur Avg Loss: 0.67213749, Log Avg loss: 0.49261809, Global Avg Loss: 1.94835758, Time: 0.0224 Steps: 44080, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001510, Sample Num: 24160, Cur Loss: 0.55713856, Cur Avg Loss: 0.67190416, Log Avg loss: 0.63690380, Global Avg Loss: 1.94806013, Time: 0.0068 Steps: 44090, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001520, Sample Num: 24320, Cur Loss: 0.91228628, Cur Avg Loss: 0.67116663, Log Avg loss: 0.55980026, Global Avg Loss: 1.94774533, Time: 0.0087 Steps: 44100, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001530, Sample Num: 24480, Cur Loss: 0.55290192, Cur Avg Loss: 0.67131577, Log Avg loss: 0.69398468, Global Avg Loss: 1.94746110, Time: 0.0094 Steps: 44110, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001540, Sample Num: 24640, Cur Loss: 1.21478689, Cur Avg Loss: 0.67134243, Log Avg loss: 0.67542062, Global Avg Loss: 1.94717279, Time: 0.0069 Steps: 44120, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001550, Sample Num: 24800, Cur Loss: 0.15707032, Cur Avg Loss: 0.67007974, Log Avg loss: 0.47562676, Global Avg Loss: 1.94683933, Time: 0.0068 Steps: 44130, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001560, Sample Num: 24960, Cur Loss: 0.56529379, Cur Avg Loss: 0.67032152, Log Avg loss: 0.70779756, Global Avg Loss: 1.94655862, Time: 0.0073 Steps: 44140, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001570, Sample Num: 25120, Cur Loss: 0.79067320, Cur Avg Loss: 0.66986086, Log Avg loss: 0.59799665, Global Avg Loss: 1.94625317, Time: 0.0105 Steps: 44150, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001580, Sample Num: 25280, Cur Loss: 0.64220691, Cur Avg Loss: 0.67006608, Log Avg loss: 0.70228543, Global Avg Loss: 1.94597148, Time: 0.0065 Steps: 44160, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001590, Sample Num: 25440, Cur Loss: 0.53398228, Cur Avg Loss: 0.67012734, Log Avg loss: 0.67980736, Global Avg Loss: 1.94568482, Time: 0.0090 Steps: 44170, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001600, Sample Num: 25600, Cur Loss: 0.82878560, Cur Avg Loss: 0.67130532, Log Avg loss: 0.85860430, Global Avg Loss: 1.94543876, Time: 0.0156 Steps: 44180, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001610, Sample Num: 25760, Cur Loss: 0.61166000, Cur Avg Loss: 0.67084916, Log Avg loss: 0.59786327, Global Avg Loss: 1.94513381, Time: 0.0099 Steps: 44190, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001620, Sample Num: 25920, Cur Loss: 0.84622860, Cur Avg Loss: 0.67077417, Log Avg loss: 0.65870124, Global Avg Loss: 1.94484276, Time: 0.0071 Steps: 44200, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001630, Sample Num: 26080, Cur Loss: 0.34308171, Cur Avg Loss: 0.67152301, Log Avg loss: 0.79283413, Global Avg Loss: 1.94458219, Time: 0.0112 Steps: 44210, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001640, Sample Num: 26240, Cur Loss: 0.42457116, Cur Avg Loss: 0.67144783, Log Avg loss: 0.65919325, Global Avg Loss: 1.94429151, Time: 0.0067 Steps: 44220, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001650, Sample Num: 26400, Cur Loss: 0.99826181, Cur Avg Loss: 0.67173333, Log Avg loss: 0.71855643, Global Avg Loss: 1.94401438, Time: 0.0120 Steps: 44230, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001660, Sample Num: 26560, Cur Loss: 1.60483360, Cur Avg Loss: 0.67213688, Log Avg loss: 0.73872302, Global Avg Loss: 1.94374194, Time: 0.0067 Steps: 44240, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001670, Sample Num: 26720, Cur Loss: 1.26811790, Cur Avg Loss: 0.67190823, Log Avg loss: 0.63395077, Global Avg Loss: 1.94344594, Time: 0.0129 Steps: 44250, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001680, Sample Num: 26880, Cur Loss: 0.47538275, Cur Avg Loss: 0.67164151, Log Avg loss: 0.62710011, Global Avg Loss: 1.94314853, Time: 0.0105 Steps: 44260, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001690, Sample Num: 27040, Cur Loss: 0.22139353, Cur Avg Loss: 0.67098338, Log Avg loss: 0.56041692, Global Avg Loss: 1.94283618, Time: 0.0068 Steps: 44270, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001700, Sample Num: 27200, Cur Loss: 1.91579151, Cur Avg Loss: 0.67165297, Log Avg loss: 0.78481412, Global Avg Loss: 1.94257466, Time: 0.0109 Steps: 44280, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001710, Sample Num: 27360, Cur Loss: 0.50132608, Cur Avg Loss: 0.67139856, Log Avg loss: 0.62814955, Global Avg Loss: 1.94227789, Time: 0.0132 Steps: 44290, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001720, Sample Num: 27520, Cur Loss: 0.87856090, Cur Avg Loss: 0.67140287, Log Avg loss: 0.67213902, Global Avg Loss: 1.94199117, Time: 0.0114 Steps: 44300, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001730, Sample Num: 27680, Cur Loss: 0.82104254, Cur Avg Loss: 0.67170722, Log Avg loss: 0.72405542, Global Avg Loss: 1.94171631, Time: 0.0115 Steps: 44310, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001740, Sample Num: 27840, Cur Loss: 0.88785458, Cur Avg Loss: 0.67244627, Log Avg loss: 0.80030145, Global Avg Loss: 1.94145877, Time: 0.0162 Steps: 44320, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001750, Sample Num: 28000, Cur Loss: 0.83968514, Cur Avg Loss: 0.67252294, Log Avg loss: 0.68586414, Global Avg Loss: 1.94117553, Time: 0.0072 Steps: 44330, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001760, Sample Num: 28160, Cur Loss: 1.00684690, Cur Avg Loss: 0.67229739, Log Avg loss: 0.63282576, Global Avg Loss: 1.94088046, Time: 0.0113 Steps: 44340, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001770, Sample Num: 28320, Cur Loss: 1.50914931, Cur Avg Loss: 0.67302141, Log Avg loss: 0.80044926, Global Avg Loss: 1.94062331, Time: 0.0072 Steps: 44350, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001780, Sample Num: 28480, Cur Loss: 1.08569372, Cur Avg Loss: 0.67231976, Log Avg loss: 0.54812800, Global Avg Loss: 1.94030940, Time: 0.0159 Steps: 44360, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001790, Sample Num: 28640, Cur Loss: 1.47698617, Cur Avg Loss: 0.67347495, Log Avg loss: 0.87909819, Global Avg Loss: 1.94007023, Time: 0.0114 Steps: 44370, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001800, Sample Num: 28800, Cur Loss: 0.42621601, Cur Avg Loss: 0.67267127, Log Avg loss: 0.52881293, Global Avg Loss: 1.93975224, Time: 0.0111 Steps: 44380, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001810, Sample Num: 28960, Cur Loss: 0.41396326, Cur Avg Loss: 0.67236238, Log Avg loss: 0.61676259, Global Avg Loss: 1.93945420, Time: 0.0128 Steps: 44390, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001820, Sample Num: 29120, Cur Loss: 1.69908881, Cur Avg Loss: 0.67260031, Log Avg loss: 0.71566603, Global Avg Loss: 1.93917857, Time: 0.0109 Steps: 44400, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001830, Sample Num: 29280, Cur Loss: 0.51927179, Cur Avg Loss: 0.67250127, Log Avg loss: 0.65447528, Global Avg Loss: 1.93888929, Time: 0.0115 Steps: 44410, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001840, Sample Num: 29440, Cur Loss: 0.22564352, Cur Avg Loss: 0.67222633, Log Avg loss: 0.62191177, Global Avg Loss: 1.93859281, Time: 0.0100 Steps: 44420, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001850, Sample Num: 29600, Cur Loss: 0.33848947, Cur Avg Loss: 0.67266187, Log Avg loss: 0.75280143, Global Avg Loss: 1.93832592, Time: 0.0118 Steps: 44430, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001860, Sample Num: 29760, Cur Loss: 0.47428036, Cur Avg Loss: 0.67265490, Log Avg loss: 0.67136565, Global Avg Loss: 1.93804082, Time: 0.0067 Steps: 44440, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001870, Sample Num: 29920, Cur Loss: 0.37736890, Cur Avg Loss: 0.67164721, Log Avg loss: 0.48421743, Global Avg Loss: 1.93771375, Time: 0.0117 Steps: 44450, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001880, Sample Num: 30080, Cur Loss: 0.89453113, Cur Avg Loss: 0.67183504, Log Avg loss: 0.70695887, Global Avg Loss: 1.93743693, Time: 0.0118 Steps: 44460, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001890, Sample Num: 30240, Cur Loss: 0.21003374, Cur Avg Loss: 0.67137467, Log Avg loss: 0.58482465, Global Avg Loss: 1.93713277, Time: 0.0068 Steps: 44470, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001900, Sample Num: 30400, Cur Loss: 0.77674687, Cur Avg Loss: 0.67145695, Log Avg loss: 0.68700734, Global Avg Loss: 1.93685171, Time: 0.0068 Steps: 44480, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001910, Sample Num: 30560, Cur Loss: 1.27209127, Cur Avg Loss: 0.67228884, Log Avg loss: 0.83034901, Global Avg Loss: 1.93660300, Time: 0.0120 Steps: 44490, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001920, Sample Num: 30720, Cur Loss: 0.39441767, Cur Avg Loss: 0.67351594, Log Avg loss: 0.90789244, Global Avg Loss: 1.93637183, Time: 0.0122 Steps: 44500, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001930, Sample Num: 30880, Cur Loss: 0.90761578, Cur Avg Loss: 0.67326964, Log Avg loss: 0.62597981, Global Avg Loss: 1.93607743, Time: 0.0068 Steps: 44510, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001940, Sample Num: 31040, Cur Loss: 0.53693676, Cur Avg Loss: 0.67262483, Log Avg loss: 0.54817610, Global Avg Loss: 1.93576568, Time: 0.0068 Steps: 44520, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001950, Sample Num: 31200, Cur Loss: 0.34773639, Cur Avg Loss: 0.67230933, Log Avg loss: 0.61110309, Global Avg Loss: 1.93546821, Time: 0.0114 Steps: 44530, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001960, Sample Num: 31360, Cur Loss: 0.45263508, Cur Avg Loss: 0.67193640, Log Avg loss: 0.59921374, Global Avg Loss: 1.93516819, Time: 0.0151 Steps: 44540, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001970, Sample Num: 31520, Cur Loss: 0.84104055, Cur Avg Loss: 0.67161315, Log Avg loss: 0.60825768, Global Avg Loss: 1.93487035, Time: 0.0219 Steps: 44550, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001980, Sample Num: 31680, Cur Loss: 0.84170222, Cur Avg Loss: 0.67168781, Log Avg loss: 0.68639539, Global Avg Loss: 1.93459017, Time: 0.0067 Steps: 44560, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 001990, Sample Num: 31840, Cur Loss: 0.48356691, Cur Avg Loss: 0.67164953, Log Avg loss: 0.66406925, Global Avg Loss: 1.93430511, Time: 0.0071 Steps: 44570, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002000, Sample Num: 32000, Cur Loss: 0.96174753, Cur Avg Loss: 0.67149657, Log Avg loss: 0.64105859, Global Avg Loss: 1.93401501, Time: 0.0095 Steps: 44580, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002010, Sample Num: 32160, Cur Loss: 0.32688406, Cur Avg Loss: 0.67169797, Log Avg loss: 0.71197713, Global Avg Loss: 1.93374095, Time: 0.0123 Steps: 44590, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002020, Sample Num: 32320, Cur Loss: 0.28582537, Cur Avg Loss: 0.67189087, Log Avg loss: 0.71066345, Global Avg Loss: 1.93346672, Time: 0.0138 Steps: 44600, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002030, Sample Num: 32480, Cur Loss: 0.54947180, Cur Avg Loss: 0.67153968, Log Avg loss: 0.60060072, Global Avg Loss: 1.93316793, Time: 0.0117 Steps: 44610, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002040, Sample Num: 32640, Cur Loss: 0.45711881, Cur Avg Loss: 0.67106412, Log Avg loss: 0.57452391, Global Avg Loss: 1.93286344, Time: 0.0152 Steps: 44620, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002050, Sample Num: 32800, Cur Loss: 0.52445459, Cur Avg Loss: 0.67043862, Log Avg loss: 0.54283648, Global Avg Loss: 1.93255199, Time: 0.0107 Steps: 44630, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002060, Sample Num: 32960, Cur Loss: 0.53057766, Cur Avg Loss: 0.67010493, Log Avg loss: 0.60169882, Global Avg Loss: 1.93225386, Time: 0.0113 Steps: 44640, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002070, Sample Num: 33120, Cur Loss: 0.36227274, Cur Avg Loss: 0.66989669, Log Avg loss: 0.62699928, Global Avg Loss: 1.93196153, Time: 0.0108 Steps: 44650, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002080, Sample Num: 33280, Cur Loss: 1.07123995, Cur Avg Loss: 0.66999103, Log Avg loss: 0.68952006, Global Avg Loss: 1.93168333, Time: 0.0226 Steps: 44660, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002090, Sample Num: 33440, Cur Loss: 0.48284340, Cur Avg Loss: 0.66944179, Log Avg loss: 0.55519930, Global Avg Loss: 1.93137518, Time: 0.0116 Steps: 44670, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002100, Sample Num: 33600, Cur Loss: 0.62105072, Cur Avg Loss: 0.66938622, Log Avg loss: 0.65777234, Global Avg Loss: 1.93109013, Time: 0.0071 Steps: 44680, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002110, Sample Num: 33760, Cur Loss: 1.26731920, Cur Avg Loss: 0.66928243, Log Avg loss: 0.64748593, Global Avg Loss: 1.93080291, Time: 0.0116 Steps: 44690, Updated lr: 0.000059 Training, Epoch: 0021, Batch: 002120, Sample Num: 33920, Cur Loss: 0.38068908, Cur Avg Loss: 0.66923250, Log Avg loss: 0.65869736, Global Avg Loss: 1.93051832, Time: 0.0232 Steps: 44700, Updated lr: 0.000059 ***** Running evaluation checkpoint-44709 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-44709 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.290458, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.746648, "eval_total_loss": 524.893276, "eval_mae": 0.675248, "eval_mse": 0.746819, "eval_r2": 0.525273, "eval_sp_statistic": 0.688271, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.732482, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.963572, "test_total_loss": 483.713099, "test_mae": 0.735333, "test_mse": 0.963768, "test_r2": 0.377975, "test_sp_statistic": 0.513473, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.61573, "test_ps_pvalue": 0.0, "lr": 5.855002370791844e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.9302400274865021, "train_cur_epoch_loss": 1423.7054020538926, "train_cur_epoch_avg_loss": 0.6687202452108467, "train_cur_epoch_time": 22.29045820236206, "train_cur_epoch_avg_time": 0.010469919305947421, "epoch": 21, "step": 44709} ################################################## Training, Epoch: 0022, Batch: 000001, Sample Num: 16, Cur Loss: 0.38704062, Cur Avg Loss: 0.38704062, Log Avg loss: 0.53195484, Global Avg Loss: 1.93020551, Time: 0.0075 Steps: 44710, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000011, Sample Num: 176, Cur Loss: 0.53345770, Cur Avg Loss: 0.54671585, Log Avg loss: 0.56268337, Global Avg Loss: 1.92989972, Time: 0.0066 Steps: 44720, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000021, Sample Num: 336, Cur Loss: 0.68093979, Cur Avg Loss: 0.60509609, Log Avg loss: 0.66931435, Global Avg Loss: 1.92961789, Time: 0.0134 Steps: 44730, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000031, Sample Num: 496, Cur Loss: 0.49339110, Cur Avg Loss: 0.62903904, Log Avg loss: 0.67931924, Global Avg Loss: 1.92933844, Time: 0.0066 Steps: 44740, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000041, Sample Num: 656, Cur Loss: 0.30912489, Cur Avg Loss: 0.61263175, Log Avg loss: 0.56176917, Global Avg Loss: 1.92903283, Time: 0.0088 Steps: 44750, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000051, Sample Num: 816, Cur Loss: 0.40295318, Cur Avg Loss: 0.59559745, Log Avg loss: 0.52575681, Global Avg Loss: 1.92871932, Time: 0.0139 Steps: 44760, Updated lr: 0.000059 Training, Epoch: 0022, Batch: 000061, Sample Num: 976, Cur Loss: 0.69253367, Cur Avg Loss: 0.59701438, Log Avg loss: 0.60424072, Global Avg Loss: 1.92842348, Time: 0.0115 Steps: 44770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000071, Sample Num: 1136, Cur Loss: 1.07429934, Cur Avg Loss: 0.60716118, Log Avg loss: 0.66905662, Global Avg Loss: 1.92814225, Time: 0.0078 Steps: 44780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000081, Sample Num: 1296, Cur Loss: 0.43431848, Cur Avg Loss: 0.62603045, Log Avg loss: 0.76000227, Global Avg Loss: 1.92788144, Time: 0.0133 Steps: 44790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000091, Sample Num: 1456, Cur Loss: 0.32198930, Cur Avg Loss: 0.62773490, Log Avg loss: 0.64154101, Global Avg Loss: 1.92759431, Time: 0.0118 Steps: 44800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000101, Sample Num: 1616, Cur Loss: 0.61196303, Cur Avg Loss: 0.62980658, Log Avg loss: 0.64865878, Global Avg Loss: 1.92730890, Time: 0.0123 Steps: 44810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000111, Sample Num: 1776, Cur Loss: 0.42721963, Cur Avg Loss: 0.62708640, Log Avg loss: 0.59961266, Global Avg Loss: 1.92701267, Time: 0.0069 Steps: 44820, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000121, Sample Num: 1936, Cur Loss: 0.38716173, Cur Avg Loss: 0.61429423, Log Avg loss: 0.47230112, Global Avg Loss: 1.92668818, Time: 0.0117 Steps: 44830, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000131, Sample Num: 2096, Cur Loss: 0.37705088, Cur Avg Loss: 0.60936536, Log Avg loss: 0.54972601, Global Avg Loss: 1.92638109, Time: 0.0064 Steps: 44840, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000141, Sample Num: 2256, Cur Loss: 0.20789804, Cur Avg Loss: 0.60738411, Log Avg loss: 0.58142980, Global Avg Loss: 1.92608122, Time: 0.0070 Steps: 44850, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000151, Sample Num: 2416, Cur Loss: 0.21318379, Cur Avg Loss: 0.62373614, Log Avg loss: 0.85429970, Global Avg Loss: 1.92584230, Time: 0.0221 Steps: 44860, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000161, Sample Num: 2576, Cur Loss: 0.34848639, Cur Avg Loss: 0.63058161, Log Avg loss: 0.73394814, Global Avg Loss: 1.92557667, Time: 0.0065 Steps: 44870, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000171, Sample Num: 2736, Cur Loss: 0.93216830, Cur Avg Loss: 0.63812930, Log Avg loss: 0.75964717, Global Avg Loss: 1.92531688, Time: 0.0109 Steps: 44880, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000181, Sample Num: 2896, Cur Loss: 1.18789196, Cur Avg Loss: 0.64160102, Log Avg loss: 0.70096742, Global Avg Loss: 1.92504413, Time: 0.0065 Steps: 44890, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000191, Sample Num: 3056, Cur Loss: 0.43011528, Cur Avg Loss: 0.64338938, Log Avg loss: 0.67575883, Global Avg Loss: 1.92476590, Time: 0.0065 Steps: 44900, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000201, Sample Num: 3216, Cur Loss: 0.92130214, Cur Avg Loss: 0.63907022, Log Avg loss: 0.55657427, Global Avg Loss: 1.92446124, Time: 0.0118 Steps: 44910, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000211, Sample Num: 3376, Cur Loss: 0.82220685, Cur Avg Loss: 0.63930523, Log Avg loss: 0.64402874, Global Avg Loss: 1.92417620, Time: 0.0107 Steps: 44920, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000221, Sample Num: 3536, Cur Loss: 1.02029121, Cur Avg Loss: 0.65589479, Log Avg loss: 1.00593457, Global Avg Loss: 1.92397183, Time: 0.0110 Steps: 44930, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000231, Sample Num: 3696, Cur Loss: 0.89998150, Cur Avg Loss: 0.65016718, Log Avg loss: 0.52358713, Global Avg Loss: 1.92366021, Time: 0.0116 Steps: 44940, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000241, Sample Num: 3856, Cur Loss: 1.12105870, Cur Avg Loss: 0.65622619, Log Avg loss: 0.79618916, Global Avg Loss: 1.92340939, Time: 0.0073 Steps: 44950, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000251, Sample Num: 4016, Cur Loss: 0.49500322, Cur Avg Loss: 0.65481153, Log Avg loss: 0.62071829, Global Avg Loss: 1.92311964, Time: 0.0065 Steps: 44960, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000261, Sample Num: 4176, Cur Loss: 0.75227159, Cur Avg Loss: 0.65474375, Log Avg loss: 0.65304258, Global Avg Loss: 1.92283721, Time: 0.0119 Steps: 44970, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000271, Sample Num: 4336, Cur Loss: 0.59315240, Cur Avg Loss: 0.65326684, Log Avg loss: 0.61471928, Global Avg Loss: 1.92254639, Time: 0.0063 Steps: 44980, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000281, Sample Num: 4496, Cur Loss: 0.26524508, Cur Avg Loss: 0.65488823, Log Avg loss: 0.69882799, Global Avg Loss: 1.92227439, Time: 0.0125 Steps: 44990, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000291, Sample Num: 4656, Cur Loss: 0.25597453, Cur Avg Loss: 0.64761774, Log Avg loss: 0.44331699, Global Avg Loss: 1.92194574, Time: 0.0121 Steps: 45000, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000301, Sample Num: 4816, Cur Loss: 0.76314706, Cur Avg Loss: 0.65000660, Log Avg loss: 0.71952232, Global Avg Loss: 1.92167859, Time: 0.0067 Steps: 45010, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000311, Sample Num: 4976, Cur Loss: 0.14252609, Cur Avg Loss: 0.64876308, Log Avg loss: 0.61133326, Global Avg Loss: 1.92138753, Time: 0.0114 Steps: 45020, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000321, Sample Num: 5136, Cur Loss: 0.59978610, Cur Avg Loss: 0.65084700, Log Avg loss: 0.71565699, Global Avg Loss: 1.92111977, Time: 0.0119 Steps: 45030, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000331, Sample Num: 5296, Cur Loss: 0.22414550, Cur Avg Loss: 0.64768812, Log Avg loss: 0.54628797, Global Avg Loss: 1.92081452, Time: 0.0066 Steps: 45040, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000341, Sample Num: 5456, Cur Loss: 0.50242150, Cur Avg Loss: 0.64187434, Log Avg loss: 0.44943820, Global Avg Loss: 1.92048791, Time: 0.0104 Steps: 45050, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000351, Sample Num: 5616, Cur Loss: 0.52707624, Cur Avg Loss: 0.63954399, Log Avg loss: 0.56007918, Global Avg Loss: 1.92018600, Time: 0.0066 Steps: 45060, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000361, Sample Num: 5776, Cur Loss: 0.54304987, Cur Avg Loss: 0.63772569, Log Avg loss: 0.57390307, Global Avg Loss: 1.91988729, Time: 0.0112 Steps: 45070, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000371, Sample Num: 5936, Cur Loss: 0.37515008, Cur Avg Loss: 0.63874960, Log Avg loss: 0.67571304, Global Avg Loss: 1.91961130, Time: 0.0107 Steps: 45080, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000381, Sample Num: 6096, Cur Loss: 0.27496022, Cur Avg Loss: 0.64429295, Log Avg loss: 0.84995111, Global Avg Loss: 1.91937407, Time: 0.0116 Steps: 45090, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000391, Sample Num: 6256, Cur Loss: 1.42821169, Cur Avg Loss: 0.64852918, Log Avg loss: 0.80992944, Global Avg Loss: 1.91912808, Time: 0.0082 Steps: 45100, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000401, Sample Num: 6416, Cur Loss: 1.11664379, Cur Avg Loss: 0.65370127, Log Avg loss: 0.85593029, Global Avg Loss: 1.91889239, Time: 0.0102 Steps: 45110, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000411, Sample Num: 6576, Cur Loss: 0.61152393, Cur Avg Loss: 0.65870381, Log Avg loss: 0.85930541, Global Avg Loss: 1.91865755, Time: 0.0134 Steps: 45120, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000421, Sample Num: 6736, Cur Loss: 0.62701905, Cur Avg Loss: 0.65599231, Log Avg loss: 0.54454989, Global Avg Loss: 1.91835307, Time: 0.0066 Steps: 45130, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000431, Sample Num: 6896, Cur Loss: 1.20680392, Cur Avg Loss: 0.65451440, Log Avg loss: 0.59229409, Global Avg Loss: 1.91805931, Time: 0.0109 Steps: 45140, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000441, Sample Num: 7056, Cur Loss: 0.39433298, Cur Avg Loss: 0.65214010, Log Avg loss: 0.54980779, Global Avg Loss: 1.91775626, Time: 0.0106 Steps: 45150, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000451, Sample Num: 7216, Cur Loss: 0.21333086, Cur Avg Loss: 0.64961330, Log Avg loss: 0.53818152, Global Avg Loss: 1.91745077, Time: 0.0184 Steps: 45160, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000461, Sample Num: 7376, Cur Loss: 0.98214471, Cur Avg Loss: 0.64919933, Log Avg loss: 0.63052940, Global Avg Loss: 1.91716587, Time: 0.0067 Steps: 45170, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000471, Sample Num: 7536, Cur Loss: 0.63619840, Cur Avg Loss: 0.64782093, Log Avg loss: 0.58427667, Global Avg Loss: 1.91687085, Time: 0.0071 Steps: 45180, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000481, Sample Num: 7696, Cur Loss: 0.19650689, Cur Avg Loss: 0.64785961, Log Avg loss: 0.64968134, Global Avg Loss: 1.91659044, Time: 0.0066 Steps: 45190, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000491, Sample Num: 7856, Cur Loss: 0.41185915, Cur Avg Loss: 0.64858212, Log Avg loss: 0.68333480, Global Avg Loss: 1.91631759, Time: 0.0066 Steps: 45200, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000501, Sample Num: 8016, Cur Loss: 1.11663854, Cur Avg Loss: 0.65475886, Log Avg loss: 0.95803698, Global Avg Loss: 1.91610563, Time: 0.0071 Steps: 45210, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000511, Sample Num: 8176, Cur Loss: 0.16142538, Cur Avg Loss: 0.65205842, Log Avg loss: 0.51676618, Global Avg Loss: 1.91579618, Time: 0.0064 Steps: 45220, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000521, Sample Num: 8336, Cur Loss: 0.56204802, Cur Avg Loss: 0.65440488, Log Avg loss: 0.77430887, Global Avg Loss: 1.91554381, Time: 0.0072 Steps: 45230, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000531, Sample Num: 8496, Cur Loss: 0.50944090, Cur Avg Loss: 0.65484151, Log Avg loss: 0.67759016, Global Avg Loss: 1.91527016, Time: 0.0087 Steps: 45240, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000541, Sample Num: 8656, Cur Loss: 0.25446534, Cur Avg Loss: 0.65315866, Log Avg loss: 0.56379921, Global Avg Loss: 1.91497150, Time: 0.0110 Steps: 45250, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000551, Sample Num: 8816, Cur Loss: 0.48341733, Cur Avg Loss: 0.65419888, Log Avg loss: 0.71047469, Global Avg Loss: 1.91470537, Time: 0.0066 Steps: 45260, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000561, Sample Num: 8976, Cur Loss: 0.32921511, Cur Avg Loss: 0.65552599, Log Avg loss: 0.72864991, Global Avg Loss: 1.91444337, Time: 0.0074 Steps: 45270, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000571, Sample Num: 9136, Cur Loss: 0.36388314, Cur Avg Loss: 0.65469933, Log Avg loss: 0.60832344, Global Avg Loss: 1.91415492, Time: 0.0109 Steps: 45280, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000581, Sample Num: 9296, Cur Loss: 0.58303589, Cur Avg Loss: 0.65605782, Log Avg loss: 0.73362773, Global Avg Loss: 1.91389426, Time: 0.0114 Steps: 45290, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000591, Sample Num: 9456, Cur Loss: 0.66445220, Cur Avg Loss: 0.65419491, Log Avg loss: 0.54595997, Global Avg Loss: 1.91359229, Time: 0.0118 Steps: 45300, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000601, Sample Num: 9616, Cur Loss: 0.18192056, Cur Avg Loss: 0.65248024, Log Avg loss: 0.55114320, Global Avg Loss: 1.91329159, Time: 0.0114 Steps: 45310, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000611, Sample Num: 9776, Cur Loss: 0.99261081, Cur Avg Loss: 0.65240743, Log Avg loss: 0.64803144, Global Avg Loss: 1.91301241, Time: 0.0168 Steps: 45320, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000621, Sample Num: 9936, Cur Loss: 0.62896901, Cur Avg Loss: 0.65193098, Log Avg loss: 0.62281982, Global Avg Loss: 1.91272779, Time: 0.0075 Steps: 45330, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000631, Sample Num: 10096, Cur Loss: 0.62170798, Cur Avg Loss: 0.65360733, Log Avg loss: 0.75770857, Global Avg Loss: 1.91247304, Time: 0.0069 Steps: 45340, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000641, Sample Num: 10256, Cur Loss: 0.87693655, Cur Avg Loss: 0.65358817, Log Avg loss: 0.65237958, Global Avg Loss: 1.91219518, Time: 0.0067 Steps: 45350, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000651, Sample Num: 10416, Cur Loss: 0.27885807, Cur Avg Loss: 0.64957006, Log Avg loss: 0.39200911, Global Avg Loss: 1.91186004, Time: 0.0123 Steps: 45360, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000661, Sample Num: 10576, Cur Loss: 1.03671479, Cur Avg Loss: 0.64959205, Log Avg loss: 0.65102330, Global Avg Loss: 1.91158214, Time: 0.0113 Steps: 45370, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000671, Sample Num: 10736, Cur Loss: 0.21588878, Cur Avg Loss: 0.64734773, Log Avg loss: 0.49899862, Global Avg Loss: 1.91127086, Time: 0.0086 Steps: 45380, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000681, Sample Num: 10896, Cur Loss: 0.47808996, Cur Avg Loss: 0.64472365, Log Avg loss: 0.46864778, Global Avg Loss: 1.91095303, Time: 0.0067 Steps: 45390, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000691, Sample Num: 11056, Cur Loss: 0.39116776, Cur Avg Loss: 0.64380060, Log Avg loss: 0.58094066, Global Avg Loss: 1.91066008, Time: 0.0114 Steps: 45400, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000701, Sample Num: 11216, Cur Loss: 0.53140688, Cur Avg Loss: 0.64344653, Log Avg loss: 0.61898052, Global Avg Loss: 1.91037563, Time: 0.0073 Steps: 45410, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000711, Sample Num: 11376, Cur Loss: 0.48863345, Cur Avg Loss: 0.64613218, Log Avg loss: 0.83439589, Global Avg Loss: 1.91013874, Time: 0.0071 Steps: 45420, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000721, Sample Num: 11536, Cur Loss: 0.51843917, Cur Avg Loss: 0.64509864, Log Avg loss: 0.57161430, Global Avg Loss: 1.90984410, Time: 0.0089 Steps: 45430, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000731, Sample Num: 11696, Cur Loss: 0.36248744, Cur Avg Loss: 0.64590152, Log Avg loss: 0.70378920, Global Avg Loss: 1.90957868, Time: 0.0076 Steps: 45440, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000741, Sample Num: 11856, Cur Loss: 0.65261632, Cur Avg Loss: 0.64466573, Log Avg loss: 0.55432971, Global Avg Loss: 1.90928050, Time: 0.0111 Steps: 45450, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000751, Sample Num: 12016, Cur Loss: 0.59857267, Cur Avg Loss: 0.64439027, Log Avg loss: 0.62397824, Global Avg Loss: 1.90899777, Time: 0.0119 Steps: 45460, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000761, Sample Num: 12176, Cur Loss: 0.68262875, Cur Avg Loss: 0.64526179, Log Avg loss: 0.71071280, Global Avg Loss: 1.90873423, Time: 0.0129 Steps: 45470, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000771, Sample Num: 12336, Cur Loss: 0.43071201, Cur Avg Loss: 0.64609960, Log Avg loss: 0.70985729, Global Avg Loss: 1.90847063, Time: 0.0105 Steps: 45480, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000781, Sample Num: 12496, Cur Loss: 0.63077766, Cur Avg Loss: 0.64565244, Log Avg loss: 0.61117643, Global Avg Loss: 1.90818545, Time: 0.0114 Steps: 45490, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000791, Sample Num: 12656, Cur Loss: 0.47003004, Cur Avg Loss: 0.64525594, Log Avg loss: 0.61428902, Global Avg Loss: 1.90790107, Time: 0.0070 Steps: 45500, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000801, Sample Num: 12816, Cur Loss: 0.65943253, Cur Avg Loss: 0.64539334, Log Avg loss: 0.65626193, Global Avg Loss: 1.90762605, Time: 0.0136 Steps: 45510, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000811, Sample Num: 12976, Cur Loss: 1.11475599, Cur Avg Loss: 0.64761510, Log Avg loss: 0.82557841, Global Avg Loss: 1.90738834, Time: 0.0117 Steps: 45520, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000821, Sample Num: 13136, Cur Loss: 0.40127298, Cur Avg Loss: 0.64681153, Log Avg loss: 0.58164130, Global Avg Loss: 1.90709716, Time: 0.0123 Steps: 45530, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000831, Sample Num: 13296, Cur Loss: 0.75650609, Cur Avg Loss: 0.64819580, Log Avg loss: 0.76184429, Global Avg Loss: 1.90684568, Time: 0.0066 Steps: 45540, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000841, Sample Num: 13456, Cur Loss: 0.45597678, Cur Avg Loss: 0.64692014, Log Avg loss: 0.54091307, Global Avg Loss: 1.90654580, Time: 0.0123 Steps: 45550, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000851, Sample Num: 13616, Cur Loss: 0.15237887, Cur Avg Loss: 0.64438759, Log Avg loss: 0.43140040, Global Avg Loss: 1.90622202, Time: 0.0064 Steps: 45560, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000861, Sample Num: 13776, Cur Loss: 0.19174993, Cur Avg Loss: 0.64500962, Log Avg loss: 0.69794406, Global Avg Loss: 1.90595687, Time: 0.0069 Steps: 45570, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000871, Sample Num: 13936, Cur Loss: 0.52585250, Cur Avg Loss: 0.64410009, Log Avg loss: 0.56578980, Global Avg Loss: 1.90566285, Time: 0.0067 Steps: 45580, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000881, Sample Num: 14096, Cur Loss: 0.36739984, Cur Avg Loss: 0.64651860, Log Avg loss: 0.85717058, Global Avg Loss: 1.90543286, Time: 0.0230 Steps: 45590, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000891, Sample Num: 14256, Cur Loss: 0.56295580, Cur Avg Loss: 0.64726178, Log Avg loss: 0.71273619, Global Avg Loss: 1.90517131, Time: 0.0112 Steps: 45600, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000901, Sample Num: 14416, Cur Loss: 0.85467607, Cur Avg Loss: 0.64757393, Log Avg loss: 0.67538595, Global Avg Loss: 1.90490168, Time: 0.0130 Steps: 45610, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000911, Sample Num: 14576, Cur Loss: 0.67670548, Cur Avg Loss: 0.64689103, Log Avg loss: 0.58536210, Global Avg Loss: 1.90461243, Time: 0.0066 Steps: 45620, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000921, Sample Num: 14736, Cur Loss: 0.38710716, Cur Avg Loss: 0.64738931, Log Avg loss: 0.69278306, Global Avg Loss: 1.90434685, Time: 0.0121 Steps: 45630, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000931, Sample Num: 14896, Cur Loss: 0.51283765, Cur Avg Loss: 0.64704524, Log Avg loss: 0.61535574, Global Avg Loss: 1.90406443, Time: 0.0064 Steps: 45640, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000941, Sample Num: 15056, Cur Loss: 1.08388901, Cur Avg Loss: 0.64750067, Log Avg loss: 0.68990112, Global Avg Loss: 1.90379846, Time: 0.0069 Steps: 45650, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000951, Sample Num: 15216, Cur Loss: 0.43004847, Cur Avg Loss: 0.64725269, Log Avg loss: 0.62391852, Global Avg Loss: 1.90351815, Time: 0.0124 Steps: 45660, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000961, Sample Num: 15376, Cur Loss: 0.66613829, Cur Avg Loss: 0.64665503, Log Avg loss: 0.58981676, Global Avg Loss: 1.90323050, Time: 0.0119 Steps: 45670, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000971, Sample Num: 15536, Cur Loss: 0.40956751, Cur Avg Loss: 0.64551013, Log Avg loss: 0.53548609, Global Avg Loss: 1.90293108, Time: 0.0172 Steps: 45680, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000981, Sample Num: 15696, Cur Loss: 1.12889469, Cur Avg Loss: 0.64597121, Log Avg loss: 0.69074157, Global Avg Loss: 1.90266577, Time: 0.0107 Steps: 45690, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 000991, Sample Num: 15856, Cur Loss: 0.58698511, Cur Avg Loss: 0.64814918, Log Avg loss: 0.86180840, Global Avg Loss: 1.90243801, Time: 0.0066 Steps: 45700, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001001, Sample Num: 16016, Cur Loss: 2.21977830, Cur Avg Loss: 0.65173394, Log Avg loss: 1.00698351, Global Avg Loss: 1.90224211, Time: 0.0070 Steps: 45710, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001011, Sample Num: 16176, Cur Loss: 0.77299035, Cur Avg Loss: 0.65439192, Log Avg loss: 0.92045515, Global Avg Loss: 1.90202738, Time: 0.0085 Steps: 45720, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001021, Sample Num: 16336, Cur Loss: 0.82559699, Cur Avg Loss: 0.65371569, Log Avg loss: 0.58534881, Global Avg Loss: 1.90173945, Time: 0.0123 Steps: 45730, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001031, Sample Num: 16496, Cur Loss: 1.15620255, Cur Avg Loss: 0.65502232, Log Avg loss: 0.78842986, Global Avg Loss: 1.90149605, Time: 0.0150 Steps: 45740, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001041, Sample Num: 16656, Cur Loss: 1.14163971, Cur Avg Loss: 0.65629242, Log Avg loss: 0.78723933, Global Avg Loss: 1.90125250, Time: 0.0077 Steps: 45750, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001051, Sample Num: 16816, Cur Loss: 0.84680820, Cur Avg Loss: 0.65666865, Log Avg loss: 0.69583420, Global Avg Loss: 1.90098908, Time: 0.0068 Steps: 45760, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001061, Sample Num: 16976, Cur Loss: 0.50218219, Cur Avg Loss: 0.65754279, Log Avg loss: 0.74941517, Global Avg Loss: 1.90073748, Time: 0.0088 Steps: 45770, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001071, Sample Num: 17136, Cur Loss: 0.41821450, Cur Avg Loss: 0.65787963, Log Avg loss: 0.69361797, Global Avg Loss: 1.90047380, Time: 0.0088 Steps: 45780, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001081, Sample Num: 17296, Cur Loss: 0.93604970, Cur Avg Loss: 0.65803502, Log Avg loss: 0.67467736, Global Avg Loss: 1.90020610, Time: 0.0118 Steps: 45790, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001091, Sample Num: 17456, Cur Loss: 0.23277834, Cur Avg Loss: 0.65648170, Log Avg loss: 0.48856863, Global Avg Loss: 1.89989788, Time: 0.0116 Steps: 45800, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001101, Sample Num: 17616, Cur Loss: 0.56523818, Cur Avg Loss: 0.65651907, Log Avg loss: 0.66059505, Global Avg Loss: 1.89962735, Time: 0.0115 Steps: 45810, Updated lr: 0.000058 Training, Epoch: 0022, Batch: 001111, Sample Num: 17776, Cur Loss: 1.52442193, Cur Avg Loss: 0.65683731, Log Avg loss: 0.69187555, Global Avg Loss: 1.89936376, Time: 0.0066 Steps: 45820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001121, Sample Num: 17936, Cur Loss: 0.60831755, Cur Avg Loss: 0.65658683, Log Avg loss: 0.62875935, Global Avg Loss: 1.89908652, Time: 0.0071 Steps: 45830, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001131, Sample Num: 18096, Cur Loss: 0.24923769, Cur Avg Loss: 0.65572870, Log Avg loss: 0.55953216, Global Avg Loss: 1.89879430, Time: 0.0136 Steps: 45840, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001141, Sample Num: 18256, Cur Loss: 0.69237542, Cur Avg Loss: 0.65590553, Log Avg loss: 0.67590467, Global Avg Loss: 1.89852758, Time: 0.0120 Steps: 45850, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001151, Sample Num: 18416, Cur Loss: 0.22588938, Cur Avg Loss: 0.65500653, Log Avg loss: 0.55243102, Global Avg Loss: 1.89823406, Time: 0.0112 Steps: 45860, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001161, Sample Num: 18576, Cur Loss: 0.89831138, Cur Avg Loss: 0.65463826, Log Avg loss: 0.61225027, Global Avg Loss: 1.89795370, Time: 0.0146 Steps: 45870, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001171, Sample Num: 18736, Cur Loss: 0.65558773, Cur Avg Loss: 0.65392093, Log Avg loss: 0.57063887, Global Avg Loss: 1.89766440, Time: 0.0066 Steps: 45880, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001181, Sample Num: 18896, Cur Loss: 0.55179882, Cur Avg Loss: 0.65351599, Log Avg loss: 0.60609691, Global Avg Loss: 1.89738295, Time: 0.0066 Steps: 45890, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001191, Sample Num: 19056, Cur Loss: 0.68158269, Cur Avg Loss: 0.65380075, Log Avg loss: 0.68743123, Global Avg Loss: 1.89711935, Time: 0.0133 Steps: 45900, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001201, Sample Num: 19216, Cur Loss: 0.33326605, Cur Avg Loss: 0.65287879, Log Avg loss: 0.54307358, Global Avg Loss: 1.89682441, Time: 0.0066 Steps: 45910, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001211, Sample Num: 19376, Cur Loss: 0.58733827, Cur Avg Loss: 0.65374677, Log Avg loss: 0.75799078, Global Avg Loss: 1.89657641, Time: 0.0119 Steps: 45920, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001221, Sample Num: 19536, Cur Loss: 0.42886156, Cur Avg Loss: 0.65361076, Log Avg loss: 0.63714049, Global Avg Loss: 1.89630220, Time: 0.0110 Steps: 45930, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001231, Sample Num: 19696, Cur Loss: 0.51147979, Cur Avg Loss: 0.65551472, Log Avg loss: 0.88798794, Global Avg Loss: 1.89608272, Time: 0.0227 Steps: 45940, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001241, Sample Num: 19856, Cur Loss: 0.63713986, Cur Avg Loss: 0.65630703, Log Avg loss: 0.75384019, Global Avg Loss: 1.89583413, Time: 0.0095 Steps: 45950, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001251, Sample Num: 20016, Cur Loss: 1.45104480, Cur Avg Loss: 0.65713180, Log Avg loss: 0.75948642, Global Avg Loss: 1.89558689, Time: 0.0107 Steps: 45960, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001261, Sample Num: 20176, Cur Loss: 0.68567145, Cur Avg Loss: 0.65801366, Log Avg loss: 0.76833446, Global Avg Loss: 1.89534167, Time: 0.0111 Steps: 45970, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001271, Sample Num: 20336, Cur Loss: 0.87908977, Cur Avg Loss: 0.65836346, Log Avg loss: 0.70247297, Global Avg Loss: 1.89508224, Time: 0.0123 Steps: 45980, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001281, Sample Num: 20496, Cur Loss: 0.40972826, Cur Avg Loss: 0.65851547, Log Avg loss: 0.67783592, Global Avg Loss: 1.89481756, Time: 0.0156 Steps: 45990, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001291, Sample Num: 20656, Cur Loss: 0.40734214, Cur Avg Loss: 0.65783049, Log Avg loss: 0.57008396, Global Avg Loss: 1.89452958, Time: 0.0112 Steps: 46000, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001301, Sample Num: 20816, Cur Loss: 1.05645895, Cur Avg Loss: 0.65772813, Log Avg loss: 0.64451383, Global Avg Loss: 1.89425789, Time: 0.0107 Steps: 46010, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001311, Sample Num: 20976, Cur Loss: 0.85417712, Cur Avg Loss: 0.65686769, Log Avg loss: 0.54492416, Global Avg Loss: 1.89396469, Time: 0.0107 Steps: 46020, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001321, Sample Num: 21136, Cur Loss: 0.79863214, Cur Avg Loss: 0.65636310, Log Avg loss: 0.59021158, Global Avg Loss: 1.89368145, Time: 0.0076 Steps: 46030, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001331, Sample Num: 21296, Cur Loss: 0.91640055, Cur Avg Loss: 0.65661616, Log Avg loss: 0.69004601, Global Avg Loss: 1.89342002, Time: 0.0077 Steps: 46040, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001341, Sample Num: 21456, Cur Loss: 0.48048240, Cur Avg Loss: 0.65648542, Log Avg loss: 0.63908384, Global Avg Loss: 1.89314763, Time: 0.0078 Steps: 46050, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001351, Sample Num: 21616, Cur Loss: 0.93317574, Cur Avg Loss: 0.65675941, Log Avg loss: 0.69350094, Global Avg Loss: 1.89288718, Time: 0.0064 Steps: 46060, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001361, Sample Num: 21776, Cur Loss: 1.12572813, Cur Avg Loss: 0.65789570, Log Avg loss: 0.81140912, Global Avg Loss: 1.89265243, Time: 0.0087 Steps: 46070, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001371, Sample Num: 21936, Cur Loss: 0.54332483, Cur Avg Loss: 0.65881067, Log Avg loss: 0.78333785, Global Avg Loss: 1.89241169, Time: 0.0089 Steps: 46080, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001381, Sample Num: 22096, Cur Loss: 0.43133879, Cur Avg Loss: 0.65797053, Log Avg loss: 0.54278766, Global Avg Loss: 1.89211887, Time: 0.0067 Steps: 46090, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001391, Sample Num: 22256, Cur Loss: 1.68638098, Cur Avg Loss: 0.65768522, Log Avg loss: 0.61828383, Global Avg Loss: 1.89184255, Time: 0.0112 Steps: 46100, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001401, Sample Num: 22416, Cur Loss: 0.66054833, Cur Avg Loss: 0.65703041, Log Avg loss: 0.56594590, Global Avg Loss: 1.89155500, Time: 0.0138 Steps: 46110, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001411, Sample Num: 22576, Cur Loss: 0.76645023, Cur Avg Loss: 0.65736134, Log Avg loss: 0.70372498, Global Avg Loss: 1.89129745, Time: 0.0069 Steps: 46120, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001421, Sample Num: 22736, Cur Loss: 0.89053226, Cur Avg Loss: 0.65719901, Log Avg loss: 0.63429343, Global Avg Loss: 1.89102496, Time: 0.0067 Steps: 46130, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001431, Sample Num: 22896, Cur Loss: 0.91706693, Cur Avg Loss: 0.65754795, Log Avg loss: 0.70713307, Global Avg Loss: 1.89076837, Time: 0.0112 Steps: 46140, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001441, Sample Num: 23056, Cur Loss: 0.56639326, Cur Avg Loss: 0.65719289, Log Avg loss: 0.60638379, Global Avg Loss: 1.89049006, Time: 0.0067 Steps: 46150, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001451, Sample Num: 23216, Cur Loss: 0.97052521, Cur Avg Loss: 0.65791963, Log Avg loss: 0.76264284, Global Avg Loss: 1.89024573, Time: 0.0114 Steps: 46160, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001461, Sample Num: 23376, Cur Loss: 0.31236854, Cur Avg Loss: 0.65822996, Log Avg loss: 0.70325805, Global Avg Loss: 1.88998864, Time: 0.0108 Steps: 46170, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001471, Sample Num: 23536, Cur Loss: 0.78815478, Cur Avg Loss: 0.65832968, Log Avg loss: 0.67289880, Global Avg Loss: 1.88972508, Time: 0.0088 Steps: 46180, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001481, Sample Num: 23696, Cur Loss: 1.15287387, Cur Avg Loss: 0.65827475, Log Avg loss: 0.65019469, Global Avg Loss: 1.88945673, Time: 0.0067 Steps: 46190, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001491, Sample Num: 23856, Cur Loss: 0.62174225, Cur Avg Loss: 0.65748978, Log Avg loss: 0.54123571, Global Avg Loss: 1.88916491, Time: 0.0067 Steps: 46200, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001501, Sample Num: 24016, Cur Loss: 0.55264270, Cur Avg Loss: 0.65986303, Log Avg loss: 1.01371534, Global Avg Loss: 1.88897546, Time: 0.0067 Steps: 46210, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001511, Sample Num: 24176, Cur Loss: 0.67139125, Cur Avg Loss: 0.65938617, Log Avg loss: 0.58780886, Global Avg Loss: 1.88869394, Time: 0.0111 Steps: 46220, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001521, Sample Num: 24336, Cur Loss: 0.64135712, Cur Avg Loss: 0.65972700, Log Avg loss: 0.71122668, Global Avg Loss: 1.88843924, Time: 0.0067 Steps: 46230, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001531, Sample Num: 24496, Cur Loss: 0.64262509, Cur Avg Loss: 0.66081958, Log Avg loss: 0.82700076, Global Avg Loss: 1.88820969, Time: 0.0148 Steps: 46240, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001541, Sample Num: 24656, Cur Loss: 0.58484054, Cur Avg Loss: 0.66177434, Log Avg loss: 0.80794814, Global Avg Loss: 1.88797612, Time: 0.0069 Steps: 46250, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001551, Sample Num: 24816, Cur Loss: 0.93640280, Cur Avg Loss: 0.66123002, Log Avg loss: 0.57735031, Global Avg Loss: 1.88769280, Time: 0.0073 Steps: 46260, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001561, Sample Num: 24976, Cur Loss: 0.71307099, Cur Avg Loss: 0.66117678, Log Avg loss: 0.65291960, Global Avg Loss: 1.88742594, Time: 0.0068 Steps: 46270, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001571, Sample Num: 25136, Cur Loss: 0.83719218, Cur Avg Loss: 0.66144107, Log Avg loss: 0.70269657, Global Avg Loss: 1.88716995, Time: 0.0121 Steps: 46280, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001581, Sample Num: 25296, Cur Loss: 0.96780860, Cur Avg Loss: 0.66270186, Log Avg loss: 0.86077157, Global Avg Loss: 1.88694822, Time: 0.0067 Steps: 46290, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001591, Sample Num: 25456, Cur Loss: 0.33750278, Cur Avg Loss: 0.66197174, Log Avg loss: 0.54653958, Global Avg Loss: 1.88665871, Time: 0.0118 Steps: 46300, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001601, Sample Num: 25616, Cur Loss: 0.53585619, Cur Avg Loss: 0.66305516, Log Avg loss: 0.83542797, Global Avg Loss: 1.88643171, Time: 0.0114 Steps: 46310, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001611, Sample Num: 25776, Cur Loss: 0.48369795, Cur Avg Loss: 0.66266108, Log Avg loss: 0.59956931, Global Avg Loss: 1.88615389, Time: 0.0127 Steps: 46320, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001621, Sample Num: 25936, Cur Loss: 0.45758861, Cur Avg Loss: 0.66258953, Log Avg loss: 0.65106255, Global Avg Loss: 1.88588731, Time: 0.0069 Steps: 46330, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001631, Sample Num: 26096, Cur Loss: 0.79069996, Cur Avg Loss: 0.66287812, Log Avg loss: 0.70965841, Global Avg Loss: 1.88563348, Time: 0.0090 Steps: 46340, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001641, Sample Num: 26256, Cur Loss: 0.69483793, Cur Avg Loss: 0.66291070, Log Avg loss: 0.66822392, Global Avg Loss: 1.88537083, Time: 0.0070 Steps: 46350, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001651, Sample Num: 26416, Cur Loss: 0.50422478, Cur Avg Loss: 0.66263303, Log Avg loss: 0.61706822, Global Avg Loss: 1.88509725, Time: 0.0224 Steps: 46360, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001661, Sample Num: 26576, Cur Loss: 0.36002055, Cur Avg Loss: 0.66124234, Log Avg loss: 0.43163935, Global Avg Loss: 1.88478380, Time: 0.0107 Steps: 46370, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001671, Sample Num: 26736, Cur Loss: 0.51621926, Cur Avg Loss: 0.66056006, Log Avg loss: 0.54723380, Global Avg Loss: 1.88449541, Time: 0.0066 Steps: 46380, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001681, Sample Num: 26896, Cur Loss: 0.44444513, Cur Avg Loss: 0.66094438, Log Avg loss: 0.72516367, Global Avg Loss: 1.88424550, Time: 0.0112 Steps: 46390, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001691, Sample Num: 27056, Cur Loss: 0.49588531, Cur Avg Loss: 0.66105841, Log Avg loss: 0.68022650, Global Avg Loss: 1.88398602, Time: 0.0126 Steps: 46400, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001701, Sample Num: 27216, Cur Loss: 0.47887948, Cur Avg Loss: 0.66098102, Log Avg loss: 0.64789497, Global Avg Loss: 1.88371967, Time: 0.0139 Steps: 46410, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001711, Sample Num: 27376, Cur Loss: 0.13822636, Cur Avg Loss: 0.65995526, Log Avg loss: 0.48547258, Global Avg Loss: 1.88341846, Time: 0.0115 Steps: 46420, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001721, Sample Num: 27536, Cur Loss: 0.63719201, Cur Avg Loss: 0.66008085, Log Avg loss: 0.68156974, Global Avg Loss: 1.88315961, Time: 0.0068 Steps: 46430, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001731, Sample Num: 27696, Cur Loss: 0.46884185, Cur Avg Loss: 0.66068506, Log Avg loss: 0.76466915, Global Avg Loss: 1.88291876, Time: 0.0067 Steps: 46440, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001741, Sample Num: 27856, Cur Loss: 0.30284539, Cur Avg Loss: 0.66159154, Log Avg loss: 0.81850438, Global Avg Loss: 1.88268961, Time: 0.0068 Steps: 46450, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001751, Sample Num: 28016, Cur Loss: 0.36772612, Cur Avg Loss: 0.66111859, Log Avg loss: 0.57877659, Global Avg Loss: 1.88240895, Time: 0.0073 Steps: 46460, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001761, Sample Num: 28176, Cur Loss: 0.62041616, Cur Avg Loss: 0.66142608, Log Avg loss: 0.71526807, Global Avg Loss: 1.88215779, Time: 0.0068 Steps: 46470, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001771, Sample Num: 28336, Cur Loss: 0.37987643, Cur Avg Loss: 0.66058999, Log Avg loss: 0.51335489, Global Avg Loss: 1.88186330, Time: 0.0068 Steps: 46480, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001781, Sample Num: 28496, Cur Loss: 1.00470591, Cur Avg Loss: 0.66124918, Log Avg loss: 0.77799087, Global Avg Loss: 1.88162586, Time: 0.0089 Steps: 46490, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001791, Sample Num: 28656, Cur Loss: 0.18344449, Cur Avg Loss: 0.66087577, Log Avg loss: 0.59437227, Global Avg Loss: 1.88134903, Time: 0.0111 Steps: 46500, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001801, Sample Num: 28816, Cur Loss: 0.78621858, Cur Avg Loss: 0.66042172, Log Avg loss: 0.57910071, Global Avg Loss: 1.88106904, Time: 0.0112 Steps: 46510, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001811, Sample Num: 28976, Cur Loss: 0.43689936, Cur Avg Loss: 0.66043865, Log Avg loss: 0.66348916, Global Avg Loss: 1.88080730, Time: 0.0234 Steps: 46520, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001821, Sample Num: 29136, Cur Loss: 0.71710098, Cur Avg Loss: 0.66039777, Log Avg loss: 0.65299350, Global Avg Loss: 1.88054343, Time: 0.0111 Steps: 46530, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001831, Sample Num: 29296, Cur Loss: 1.76957226, Cur Avg Loss: 0.66080332, Log Avg loss: 0.73465313, Global Avg Loss: 1.88029721, Time: 0.0111 Steps: 46540, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001841, Sample Num: 29456, Cur Loss: 0.92374617, Cur Avg Loss: 0.66138374, Log Avg loss: 0.76766027, Global Avg Loss: 1.88005819, Time: 0.0120 Steps: 46550, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001851, Sample Num: 29616, Cur Loss: 1.01897204, Cur Avg Loss: 0.66171447, Log Avg loss: 0.72260157, Global Avg Loss: 1.87980960, Time: 0.0068 Steps: 46560, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001861, Sample Num: 29776, Cur Loss: 0.65132892, Cur Avg Loss: 0.66172259, Log Avg loss: 0.66322461, Global Avg Loss: 1.87954836, Time: 0.0068 Steps: 46570, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001871, Sample Num: 29936, Cur Loss: 0.60400689, Cur Avg Loss: 0.66237720, Log Avg loss: 0.78420132, Global Avg Loss: 1.87931321, Time: 0.0120 Steps: 46580, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001881, Sample Num: 30096, Cur Loss: 0.34347075, Cur Avg Loss: 0.66168273, Log Avg loss: 0.53174685, Global Avg Loss: 1.87902397, Time: 0.0082 Steps: 46590, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001891, Sample Num: 30256, Cur Loss: 0.57612050, Cur Avg Loss: 0.66089323, Log Avg loss: 0.51238832, Global Avg Loss: 1.87873070, Time: 0.0136 Steps: 46600, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001901, Sample Num: 30416, Cur Loss: 0.56404054, Cur Avg Loss: 0.66117118, Log Avg loss: 0.71373231, Global Avg Loss: 1.87848075, Time: 0.0116 Steps: 46610, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001911, Sample Num: 30576, Cur Loss: 0.64135557, Cur Avg Loss: 0.66142909, Log Avg loss: 0.71045597, Global Avg Loss: 1.87823021, Time: 0.0122 Steps: 46620, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001921, Sample Num: 30736, Cur Loss: 0.99065793, Cur Avg Loss: 0.66142352, Log Avg loss: 0.66036041, Global Avg Loss: 1.87796903, Time: 0.0109 Steps: 46630, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001931, Sample Num: 30896, Cur Loss: 0.27089211, Cur Avg Loss: 0.66099266, Log Avg loss: 0.57822475, Global Avg Loss: 1.87769036, Time: 0.0066 Steps: 46640, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001941, Sample Num: 31056, Cur Loss: 1.06446350, Cur Avg Loss: 0.66068283, Log Avg loss: 0.60085457, Global Avg Loss: 1.87741665, Time: 0.0066 Steps: 46650, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001951, Sample Num: 31216, Cur Loss: 0.25593311, Cur Avg Loss: 0.66049533, Log Avg loss: 0.62410043, Global Avg Loss: 1.87714804, Time: 0.0145 Steps: 46660, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001961, Sample Num: 31376, Cur Loss: 0.28888088, Cur Avg Loss: 0.66092920, Log Avg loss: 0.74557684, Global Avg Loss: 1.87690558, Time: 0.0122 Steps: 46670, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001971, Sample Num: 31536, Cur Loss: 0.63056535, Cur Avg Loss: 0.66081995, Log Avg loss: 0.63939788, Global Avg Loss: 1.87664048, Time: 0.0067 Steps: 46680, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001981, Sample Num: 31696, Cur Loss: 0.43819919, Cur Avg Loss: 0.66099764, Log Avg loss: 0.69602005, Global Avg Loss: 1.87638761, Time: 0.0118 Steps: 46690, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 001991, Sample Num: 31856, Cur Loss: 0.70460403, Cur Avg Loss: 0.66108439, Log Avg loss: 0.67826909, Global Avg Loss: 1.87613106, Time: 0.0120 Steps: 46700, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002001, Sample Num: 32016, Cur Loss: 0.59637523, Cur Avg Loss: 0.66115349, Log Avg loss: 0.67491037, Global Avg Loss: 1.87587389, Time: 0.0067 Steps: 46710, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002011, Sample Num: 32176, Cur Loss: 0.62348616, Cur Avg Loss: 0.66103360, Log Avg loss: 0.63704349, Global Avg Loss: 1.87560873, Time: 0.0066 Steps: 46720, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002021, Sample Num: 32336, Cur Loss: 0.17837197, Cur Avg Loss: 0.66039421, Log Avg loss: 0.53181273, Global Avg Loss: 1.87532117, Time: 0.0067 Steps: 46730, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002031, Sample Num: 32496, Cur Loss: 0.54611284, Cur Avg Loss: 0.66048487, Log Avg loss: 0.67880893, Global Avg Loss: 1.87506517, Time: 0.0112 Steps: 46740, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002041, Sample Num: 32656, Cur Loss: 0.58577907, Cur Avg Loss: 0.65938344, Log Avg loss: 0.43568227, Global Avg Loss: 1.87475728, Time: 0.0067 Steps: 46750, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002051, Sample Num: 32816, Cur Loss: 0.62643540, Cur Avg Loss: 0.65936571, Log Avg loss: 0.65574631, Global Avg Loss: 1.87449659, Time: 0.0202 Steps: 46760, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002061, Sample Num: 32976, Cur Loss: 0.63674784, Cur Avg Loss: 0.65840880, Log Avg loss: 0.46214721, Global Avg Loss: 1.87419461, Time: 0.0160 Steps: 46770, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002071, Sample Num: 33136, Cur Loss: 1.02856326, Cur Avg Loss: 0.65841959, Log Avg loss: 0.66064341, Global Avg Loss: 1.87393519, Time: 0.0066 Steps: 46780, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002081, Sample Num: 33296, Cur Loss: 0.17297010, Cur Avg Loss: 0.65771230, Log Avg loss: 0.51123180, Global Avg Loss: 1.87364396, Time: 0.0066 Steps: 46790, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002091, Sample Num: 33456, Cur Loss: 0.40096679, Cur Avg Loss: 0.65823387, Log Avg loss: 0.76677427, Global Avg Loss: 1.87340744, Time: 0.0126 Steps: 46800, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002101, Sample Num: 33616, Cur Loss: 0.42346108, Cur Avg Loss: 0.65865954, Log Avg loss: 0.74766641, Global Avg Loss: 1.87316695, Time: 0.0144 Steps: 46810, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002111, Sample Num: 33776, Cur Loss: 0.23770413, Cur Avg Loss: 0.65833382, Log Avg loss: 0.58989920, Global Avg Loss: 1.87289287, Time: 0.0160 Steps: 46820, Updated lr: 0.000057 Training, Epoch: 0022, Batch: 002121, Sample Num: 33936, Cur Loss: 0.86353862, Cur Avg Loss: 0.65891430, Log Avg loss: 0.78145392, Global Avg Loss: 1.87265980, Time: 0.0124 Steps: 46830, Updated lr: 0.000057 ***** Running evaluation checkpoint-46838 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-46838 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.167289, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.712114, "eval_total_loss": 500.615983, "eval_mae": 0.655034, "eval_mse": 0.712307, "eval_r2": 0.547211, "eval_sp_statistic": 0.696155, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.744349, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 1.082789, "test_total_loss": 543.559992, "test_mae": 0.878446, "test_mse": 1.082827, "test_r2": 0.301134, "test_sp_statistic": 0.504287, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.611874, "test_ps_pvalue": 0.0, "lr": 5.653105737316264e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.872422116041989, "train_cur_epoch_loss": 1401.4056822806597, "train_cur_epoch_avg_loss": 0.6582459757072145, "train_cur_epoch_time": 22.16728925704956, "train_cur_epoch_avg_time": 0.010412066349013415, "epoch": 22, "step": 46838} ################################################## Training, Epoch: 0023, Batch: 000002, Sample Num: 32, Cur Loss: 1.60296702, Cur Avg Loss: 0.92802978, Log Avg loss: 0.57045160, Global Avg Loss: 1.87238179, Time: 0.0102 Steps: 46840, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000012, Sample Num: 192, Cur Loss: 0.37035871, Cur Avg Loss: 0.58437873, Log Avg loss: 0.51564852, Global Avg Loss: 1.87209220, Time: 0.0131 Steps: 46850, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000022, Sample Num: 352, Cur Loss: 0.62715888, Cur Avg Loss: 0.52914352, Log Avg loss: 0.46286126, Global Avg Loss: 1.87179147, Time: 0.0069 Steps: 46860, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000032, Sample Num: 512, Cur Loss: 0.76917160, Cur Avg Loss: 0.54028285, Log Avg loss: 0.56478938, Global Avg Loss: 1.87151261, Time: 0.0064 Steps: 46870, Updated lr: 0.000057 Training, Epoch: 0023, Batch: 000042, Sample Num: 672, Cur Loss: 0.28016371, Cur Avg Loss: 0.52148483, Log Avg loss: 0.46133117, Global Avg Loss: 1.87121181, Time: 0.0155 Steps: 46880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000052, Sample Num: 832, Cur Loss: 0.58398813, Cur Avg Loss: 0.53980751, Log Avg loss: 0.61676275, Global Avg Loss: 1.87094428, Time: 0.0158 Steps: 46890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000062, Sample Num: 992, Cur Loss: 1.00832820, Cur Avg Loss: 0.57690322, Log Avg loss: 0.76980095, Global Avg Loss: 1.87070949, Time: 0.0088 Steps: 46900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000072, Sample Num: 1152, Cur Loss: 0.63007998, Cur Avg Loss: 0.59791679, Log Avg loss: 0.72820092, Global Avg Loss: 1.87046594, Time: 0.0068 Steps: 46910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000082, Sample Num: 1312, Cur Loss: 0.55115801, Cur Avg Loss: 0.59451962, Log Avg loss: 0.57005995, Global Avg Loss: 1.87018878, Time: 0.0114 Steps: 46920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000092, Sample Num: 1472, Cur Loss: 0.35098201, Cur Avg Loss: 0.60348541, Log Avg loss: 0.67700491, Global Avg Loss: 1.86993454, Time: 0.0239 Steps: 46930, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000102, Sample Num: 1632, Cur Loss: 0.60473037, Cur Avg Loss: 0.59949513, Log Avg loss: 0.56278453, Global Avg Loss: 1.86965606, Time: 0.0072 Steps: 46940, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000112, Sample Num: 1792, Cur Loss: 0.59689915, Cur Avg Loss: 0.58911098, Log Avg loss: 0.48319272, Global Avg Loss: 1.86936076, Time: 0.0072 Steps: 46950, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000122, Sample Num: 1952, Cur Loss: 0.36335313, Cur Avg Loss: 0.58877932, Log Avg loss: 0.58506475, Global Avg Loss: 1.86908727, Time: 0.0117 Steps: 46960, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000132, Sample Num: 2112, Cur Loss: 1.28851163, Cur Avg Loss: 0.60137269, Log Avg loss: 0.75501176, Global Avg Loss: 1.86885008, Time: 0.0145 Steps: 46970, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000142, Sample Num: 2272, Cur Loss: 0.56230479, Cur Avg Loss: 0.60925493, Log Avg loss: 0.71330042, Global Avg Loss: 1.86860411, Time: 0.0103 Steps: 46980, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000152, Sample Num: 2432, Cur Loss: 1.25425553, Cur Avg Loss: 0.62077346, Log Avg loss: 0.78433667, Global Avg Loss: 1.86837337, Time: 0.0104 Steps: 46990, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000162, Sample Num: 2592, Cur Loss: 0.91120988, Cur Avg Loss: 0.62745195, Log Avg loss: 0.72896504, Global Avg Loss: 1.86813094, Time: 0.0134 Steps: 47000, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000172, Sample Num: 2752, Cur Loss: 0.48966315, Cur Avg Loss: 0.62833040, Log Avg loss: 0.64256115, Global Avg Loss: 1.86787024, Time: 0.0067 Steps: 47010, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000182, Sample Num: 2912, Cur Loss: 0.30825269, Cur Avg Loss: 0.62604640, Log Avg loss: 0.58676160, Global Avg Loss: 1.86759778, Time: 0.0152 Steps: 47020, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000192, Sample Num: 3072, Cur Loss: 0.70120466, Cur Avg Loss: 0.62629202, Log Avg loss: 0.63076234, Global Avg Loss: 1.86733479, Time: 0.0065 Steps: 47030, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000202, Sample Num: 3232, Cur Loss: 0.25974384, Cur Avg Loss: 0.62649951, Log Avg loss: 0.63048331, Global Avg Loss: 1.86707185, Time: 0.0111 Steps: 47040, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000212, Sample Num: 3392, Cur Loss: 0.48630819, Cur Avg Loss: 0.63909465, Log Avg loss: 0.89351651, Global Avg Loss: 1.86686493, Time: 0.0110 Steps: 47050, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000222, Sample Num: 3552, Cur Loss: 0.27960411, Cur Avg Loss: 0.63319270, Log Avg loss: 0.50807128, Global Avg Loss: 1.86657620, Time: 0.0110 Steps: 47060, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000232, Sample Num: 3712, Cur Loss: 1.75333762, Cur Avg Loss: 0.63804298, Log Avg loss: 0.74571931, Global Avg Loss: 1.86633807, Time: 0.0117 Steps: 47070, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000242, Sample Num: 3872, Cur Loss: 0.30732408, Cur Avg Loss: 0.63377352, Log Avg loss: 0.53472195, Global Avg Loss: 1.86605523, Time: 0.0069 Steps: 47080, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000252, Sample Num: 4032, Cur Loss: 0.75508738, Cur Avg Loss: 0.63323736, Log Avg loss: 0.62026231, Global Avg Loss: 1.86579067, Time: 0.0071 Steps: 47090, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000262, Sample Num: 4192, Cur Loss: 0.33320189, Cur Avg Loss: 0.62561446, Log Avg loss: 0.43351749, Global Avg Loss: 1.86548658, Time: 0.0079 Steps: 47100, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000272, Sample Num: 4352, Cur Loss: 0.59164286, Cur Avg Loss: 0.62603128, Log Avg loss: 0.63695201, Global Avg Loss: 1.86522580, Time: 0.0090 Steps: 47110, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000282, Sample Num: 4512, Cur Loss: 0.32632822, Cur Avg Loss: 0.63595109, Log Avg loss: 0.90576979, Global Avg Loss: 1.86502218, Time: 0.0073 Steps: 47120, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000292, Sample Num: 4672, Cur Loss: 0.30886102, Cur Avg Loss: 0.63605247, Log Avg loss: 0.63891143, Global Avg Loss: 1.86476203, Time: 0.0068 Steps: 47130, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000302, Sample Num: 4832, Cur Loss: 0.61856097, Cur Avg Loss: 0.63395971, Log Avg loss: 0.57285119, Global Avg Loss: 1.86448797, Time: 0.0117 Steps: 47140, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000312, Sample Num: 4992, Cur Loss: 0.30079409, Cur Avg Loss: 0.63725619, Log Avg loss: 0.73680973, Global Avg Loss: 1.86424880, Time: 0.0071 Steps: 47150, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000322, Sample Num: 5152, Cur Loss: 0.45423824, Cur Avg Loss: 0.64143536, Log Avg loss: 0.77182562, Global Avg Loss: 1.86401716, Time: 0.0176 Steps: 47160, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000332, Sample Num: 5312, Cur Loss: 0.67862570, Cur Avg Loss: 0.64301179, Log Avg loss: 0.69377270, Global Avg Loss: 1.86376907, Time: 0.0152 Steps: 47170, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000342, Sample Num: 5472, Cur Loss: 0.38766664, Cur Avg Loss: 0.64216244, Log Avg loss: 0.61396393, Global Avg Loss: 1.86350417, Time: 0.0105 Steps: 47180, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000352, Sample Num: 5632, Cur Loss: 0.56492954, Cur Avg Loss: 0.64158401, Log Avg loss: 0.62180197, Global Avg Loss: 1.86324104, Time: 0.0066 Steps: 47190, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000362, Sample Num: 5792, Cur Loss: 0.67853415, Cur Avg Loss: 0.64220416, Log Avg loss: 0.66403336, Global Avg Loss: 1.86298697, Time: 0.0122 Steps: 47200, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000372, Sample Num: 5952, Cur Loss: 0.14014255, Cur Avg Loss: 0.63727462, Log Avg loss: 0.45882531, Global Avg Loss: 1.86268954, Time: 0.0119 Steps: 47210, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000382, Sample Num: 6112, Cur Loss: 0.37314075, Cur Avg Loss: 0.63944965, Log Avg loss: 0.72036072, Global Avg Loss: 1.86244762, Time: 0.0108 Steps: 47220, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000392, Sample Num: 6272, Cur Loss: 0.21567112, Cur Avg Loss: 0.63896636, Log Avg loss: 0.62050479, Global Avg Loss: 1.86218467, Time: 0.0067 Steps: 47230, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000402, Sample Num: 6432, Cur Loss: 0.64816403, Cur Avg Loss: 0.64357379, Log Avg loss: 0.82418505, Global Avg Loss: 1.86196494, Time: 0.0133 Steps: 47240, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000412, Sample Num: 6592, Cur Loss: 0.26288685, Cur Avg Loss: 0.64249958, Log Avg loss: 0.59931599, Global Avg Loss: 1.86169771, Time: 0.0103 Steps: 47250, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000422, Sample Num: 6752, Cur Loss: 0.62313342, Cur Avg Loss: 0.64065877, Log Avg loss: 0.56481752, Global Avg Loss: 1.86142330, Time: 0.0067 Steps: 47260, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000432, Sample Num: 6912, Cur Loss: 0.99431562, Cur Avg Loss: 0.64034461, Log Avg loss: 0.62708706, Global Avg Loss: 1.86116217, Time: 0.0069 Steps: 47270, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000442, Sample Num: 7072, Cur Loss: 1.14591599, Cur Avg Loss: 0.63939739, Log Avg loss: 0.59847763, Global Avg Loss: 1.86089511, Time: 0.0121 Steps: 47280, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000452, Sample Num: 7232, Cur Loss: 0.22090036, Cur Avg Loss: 0.63916890, Log Avg loss: 0.62906943, Global Avg Loss: 1.86063462, Time: 0.0116 Steps: 47290, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000462, Sample Num: 7392, Cur Loss: 1.23787737, Cur Avg Loss: 0.64272703, Log Avg loss: 0.80355453, Global Avg Loss: 1.86041114, Time: 0.0116 Steps: 47300, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000472, Sample Num: 7552, Cur Loss: 0.10556123, Cur Avg Loss: 0.63979787, Log Avg loss: 0.50447098, Global Avg Loss: 1.86012453, Time: 0.0110 Steps: 47310, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000482, Sample Num: 7712, Cur Loss: 0.73443955, Cur Avg Loss: 0.63754081, Log Avg loss: 0.53100761, Global Avg Loss: 1.85984365, Time: 0.0109 Steps: 47320, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000492, Sample Num: 7872, Cur Loss: 1.16046619, Cur Avg Loss: 0.63731644, Log Avg loss: 0.62650151, Global Avg Loss: 1.85958307, Time: 0.0112 Steps: 47330, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000502, Sample Num: 8032, Cur Loss: 0.57338077, Cur Avg Loss: 0.63424061, Log Avg loss: 0.48291003, Global Avg Loss: 1.85929227, Time: 0.0065 Steps: 47340, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000512, Sample Num: 8192, Cur Loss: 0.78657234, Cur Avg Loss: 0.63715048, Log Avg loss: 0.78322585, Global Avg Loss: 1.85906501, Time: 0.0154 Steps: 47350, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000522, Sample Num: 8352, Cur Loss: 0.61131251, Cur Avg Loss: 0.63352240, Log Avg loss: 0.44776482, Global Avg Loss: 1.85876701, Time: 0.0069 Steps: 47360, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000532, Sample Num: 8512, Cur Loss: 0.37163535, Cur Avg Loss: 0.63217185, Log Avg loss: 0.56167303, Global Avg Loss: 1.85849319, Time: 0.0113 Steps: 47370, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000542, Sample Num: 8672, Cur Loss: 0.58431017, Cur Avg Loss: 0.63409798, Log Avg loss: 0.73656784, Global Avg Loss: 1.85825640, Time: 0.0071 Steps: 47380, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000552, Sample Num: 8832, Cur Loss: 0.35893452, Cur Avg Loss: 0.63374688, Log Avg loss: 0.61471730, Global Avg Loss: 1.85799399, Time: 0.0070 Steps: 47390, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000562, Sample Num: 8992, Cur Loss: 0.47541499, Cur Avg Loss: 0.63337819, Log Avg loss: 0.61302678, Global Avg Loss: 1.85773134, Time: 0.0064 Steps: 47400, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000572, Sample Num: 9152, Cur Loss: 0.28645635, Cur Avg Loss: 0.63191019, Log Avg loss: 0.54940869, Global Avg Loss: 1.85745538, Time: 0.0111 Steps: 47410, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000582, Sample Num: 9312, Cur Loss: 0.64664137, Cur Avg Loss: 0.63340211, Log Avg loss: 0.71873955, Global Avg Loss: 1.85721525, Time: 0.0109 Steps: 47420, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000592, Sample Num: 9472, Cur Loss: 0.28634489, Cur Avg Loss: 0.63467737, Log Avg loss: 0.70889789, Global Avg Loss: 1.85697314, Time: 0.0117 Steps: 47430, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000602, Sample Num: 9632, Cur Loss: 0.37909222, Cur Avg Loss: 0.63371532, Log Avg loss: 0.57676184, Global Avg Loss: 1.85670328, Time: 0.0122 Steps: 47440, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000612, Sample Num: 9792, Cur Loss: 0.55400467, Cur Avg Loss: 0.63471083, Log Avg loss: 0.69464028, Global Avg Loss: 1.85645838, Time: 0.0086 Steps: 47450, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000622, Sample Num: 9952, Cur Loss: 0.86685443, Cur Avg Loss: 0.63448366, Log Avg loss: 0.62058107, Global Avg Loss: 1.85619798, Time: 0.0067 Steps: 47460, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000632, Sample Num: 10112, Cur Loss: 0.47001350, Cur Avg Loss: 0.63346932, Log Avg loss: 0.57037732, Global Avg Loss: 1.85592711, Time: 0.0073 Steps: 47470, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000642, Sample Num: 10272, Cur Loss: 0.94013679, Cur Avg Loss: 0.63435819, Log Avg loss: 0.69053474, Global Avg Loss: 1.85568166, Time: 0.0115 Steps: 47480, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000652, Sample Num: 10432, Cur Loss: 0.55849755, Cur Avg Loss: 0.63413841, Log Avg loss: 0.62002870, Global Avg Loss: 1.85542146, Time: 0.0065 Steps: 47490, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000662, Sample Num: 10592, Cur Loss: 0.53986979, Cur Avg Loss: 0.63309759, Log Avg loss: 0.56523606, Global Avg Loss: 1.85514985, Time: 0.0142 Steps: 47500, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000672, Sample Num: 10752, Cur Loss: 0.94691014, Cur Avg Loss: 0.63384020, Log Avg loss: 0.68300093, Global Avg Loss: 1.85490313, Time: 0.0066 Steps: 47510, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000682, Sample Num: 10912, Cur Loss: 0.42052126, Cur Avg Loss: 0.63293556, Log Avg loss: 0.57214359, Global Avg Loss: 1.85463319, Time: 0.0113 Steps: 47520, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000692, Sample Num: 11072, Cur Loss: 1.31820357, Cur Avg Loss: 0.63194922, Log Avg loss: 0.56468119, Global Avg Loss: 1.85436179, Time: 0.0066 Steps: 47530, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000702, Sample Num: 11232, Cur Loss: 0.26003739, Cur Avg Loss: 0.63129295, Log Avg loss: 0.58587892, Global Avg Loss: 1.85409497, Time: 0.0069 Steps: 47540, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000712, Sample Num: 11392, Cur Loss: 0.47349715, Cur Avg Loss: 0.63034647, Log Avg loss: 0.56390373, Global Avg Loss: 1.85382363, Time: 0.0119 Steps: 47550, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000722, Sample Num: 11552, Cur Loss: 0.68013924, Cur Avg Loss: 0.63026206, Log Avg loss: 0.62425189, Global Avg Loss: 1.85356510, Time: 0.0115 Steps: 47560, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000732, Sample Num: 11712, Cur Loss: 0.79585314, Cur Avg Loss: 0.63384679, Log Avg loss: 0.89266399, Global Avg Loss: 1.85336311, Time: 0.0127 Steps: 47570, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000742, Sample Num: 11872, Cur Loss: 0.70820028, Cur Avg Loss: 0.63269949, Log Avg loss: 0.54871747, Global Avg Loss: 1.85308890, Time: 0.0117 Steps: 47580, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000752, Sample Num: 12032, Cur Loss: 0.49175018, Cur Avg Loss: 0.63106787, Log Avg loss: 0.51000152, Global Avg Loss: 1.85280668, Time: 0.0066 Steps: 47590, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000762, Sample Num: 12192, Cur Loss: 0.24315694, Cur Avg Loss: 0.62848680, Log Avg loss: 0.43439014, Global Avg Loss: 1.85250870, Time: 0.0066 Steps: 47600, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000772, Sample Num: 12352, Cur Loss: 0.70139110, Cur Avg Loss: 0.62635781, Log Avg loss: 0.46412892, Global Avg Loss: 1.85221708, Time: 0.0072 Steps: 47610, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000782, Sample Num: 12512, Cur Loss: 0.88440084, Cur Avg Loss: 0.62823003, Log Avg loss: 0.77276516, Global Avg Loss: 1.85199040, Time: 0.0221 Steps: 47620, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000792, Sample Num: 12672, Cur Loss: 0.46640679, Cur Avg Loss: 0.62871303, Log Avg loss: 0.66648408, Global Avg Loss: 1.85174150, Time: 0.0064 Steps: 47630, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000802, Sample Num: 12832, Cur Loss: 0.32132217, Cur Avg Loss: 0.62975714, Log Avg loss: 0.71245055, Global Avg Loss: 1.85150236, Time: 0.0066 Steps: 47640, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000812, Sample Num: 12992, Cur Loss: 1.05625463, Cur Avg Loss: 0.62809958, Log Avg loss: 0.49516283, Global Avg Loss: 1.85121771, Time: 0.0112 Steps: 47650, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000822, Sample Num: 13152, Cur Loss: 0.53088492, Cur Avg Loss: 0.62734445, Log Avg loss: 0.56602822, Global Avg Loss: 1.85094805, Time: 0.0066 Steps: 47660, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000832, Sample Num: 13312, Cur Loss: 0.50904989, Cur Avg Loss: 0.62943493, Log Avg loss: 0.80127253, Global Avg Loss: 1.85072786, Time: 0.0132 Steps: 47670, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000842, Sample Num: 13472, Cur Loss: 1.02942967, Cur Avg Loss: 0.62981257, Log Avg loss: 0.66123241, Global Avg Loss: 1.85047838, Time: 0.0073 Steps: 47680, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000852, Sample Num: 13632, Cur Loss: 0.27344838, Cur Avg Loss: 0.62892793, Log Avg loss: 0.55444100, Global Avg Loss: 1.85020662, Time: 0.0115 Steps: 47690, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000862, Sample Num: 13792, Cur Loss: 1.20684695, Cur Avg Loss: 0.62860493, Log Avg loss: 0.60108527, Global Avg Loss: 1.84994475, Time: 0.0129 Steps: 47700, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000872, Sample Num: 13952, Cur Loss: 0.54503310, Cur Avg Loss: 0.62958850, Log Avg loss: 0.71437205, Global Avg Loss: 1.84970673, Time: 0.0126 Steps: 47710, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000882, Sample Num: 14112, Cur Loss: 0.60132635, Cur Avg Loss: 0.63065520, Log Avg loss: 0.72367158, Global Avg Loss: 1.84947077, Time: 0.0123 Steps: 47720, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000892, Sample Num: 14272, Cur Loss: 1.35378373, Cur Avg Loss: 0.63274726, Log Avg loss: 0.81726700, Global Avg Loss: 1.84925451, Time: 0.0067 Steps: 47730, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000902, Sample Num: 14432, Cur Loss: 0.37430245, Cur Avg Loss: 0.63339796, Log Avg loss: 0.69144021, Global Avg Loss: 1.84901198, Time: 0.0067 Steps: 47740, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000912, Sample Num: 14592, Cur Loss: 0.40129712, Cur Avg Loss: 0.63304854, Log Avg loss: 0.60153106, Global Avg Loss: 1.84875073, Time: 0.0065 Steps: 47750, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000922, Sample Num: 14752, Cur Loss: 0.33106372, Cur Avg Loss: 0.63425310, Log Avg loss: 0.74410922, Global Avg Loss: 1.84851944, Time: 0.0117 Steps: 47760, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000932, Sample Num: 14912, Cur Loss: 1.31312072, Cur Avg Loss: 0.63644266, Log Avg loss: 0.83832031, Global Avg Loss: 1.84830797, Time: 0.0086 Steps: 47770, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000942, Sample Num: 15072, Cur Loss: 0.83240026, Cur Avg Loss: 0.63627258, Log Avg loss: 0.62042070, Global Avg Loss: 1.84805098, Time: 0.0123 Steps: 47780, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000952, Sample Num: 15232, Cur Loss: 0.25666052, Cur Avg Loss: 0.63691741, Log Avg loss: 0.69766042, Global Avg Loss: 1.84781026, Time: 0.0112 Steps: 47790, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000962, Sample Num: 15392, Cur Loss: 0.68189937, Cur Avg Loss: 0.63701708, Log Avg loss: 0.64650553, Global Avg Loss: 1.84755894, Time: 0.0067 Steps: 47800, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000972, Sample Num: 15552, Cur Loss: 1.09194720, Cur Avg Loss: 0.63787910, Log Avg loss: 0.72080589, Global Avg Loss: 1.84732327, Time: 0.0118 Steps: 47810, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000982, Sample Num: 15712, Cur Loss: 0.44524682, Cur Avg Loss: 0.63902598, Log Avg loss: 0.75050227, Global Avg Loss: 1.84709391, Time: 0.0066 Steps: 47820, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 000992, Sample Num: 15872, Cur Loss: 0.77577341, Cur Avg Loss: 0.63816918, Log Avg loss: 0.55403140, Global Avg Loss: 1.84682356, Time: 0.0085 Steps: 47830, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001002, Sample Num: 16032, Cur Loss: 0.69352603, Cur Avg Loss: 0.63856964, Log Avg loss: 0.67829568, Global Avg Loss: 1.84657930, Time: 0.0121 Steps: 47840, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001012, Sample Num: 16192, Cur Loss: 0.60951656, Cur Avg Loss: 0.63883514, Log Avg loss: 0.66543798, Global Avg Loss: 1.84633246, Time: 0.0067 Steps: 47850, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001022, Sample Num: 16352, Cur Loss: 0.31605968, Cur Avg Loss: 0.63787367, Log Avg loss: 0.54057263, Global Avg Loss: 1.84605963, Time: 0.0122 Steps: 47860, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001032, Sample Num: 16512, Cur Loss: 0.80694163, Cur Avg Loss: 0.63834244, Log Avg loss: 0.68625105, Global Avg Loss: 1.84581735, Time: 0.0229 Steps: 47870, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001042, Sample Num: 16672, Cur Loss: 0.25627261, Cur Avg Loss: 0.63754169, Log Avg loss: 0.55490390, Global Avg Loss: 1.84554773, Time: 0.0067 Steps: 47880, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001052, Sample Num: 16832, Cur Loss: 0.51431847, Cur Avg Loss: 0.63679602, Log Avg loss: 0.55909799, Global Avg Loss: 1.84527911, Time: 0.0108 Steps: 47890, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001062, Sample Num: 16992, Cur Loss: 0.26189995, Cur Avg Loss: 0.63700495, Log Avg loss: 0.65898417, Global Avg Loss: 1.84503145, Time: 0.0122 Steps: 47900, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001072, Sample Num: 17152, Cur Loss: 0.78566974, Cur Avg Loss: 0.63869386, Log Avg loss: 0.81805622, Global Avg Loss: 1.84481709, Time: 0.0070 Steps: 47910, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001082, Sample Num: 17312, Cur Loss: 1.30882633, Cur Avg Loss: 0.63823338, Log Avg loss: 0.58886918, Global Avg Loss: 1.84455500, Time: 0.0112 Steps: 47920, Updated lr: 0.000056 Training, Epoch: 0023, Batch: 001092, Sample Num: 17472, Cur Loss: 0.19114837, Cur Avg Loss: 0.63854982, Log Avg loss: 0.67278945, Global Avg Loss: 1.84431053, Time: 0.0067 Steps: 47930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001102, Sample Num: 17632, Cur Loss: 0.42235374, Cur Avg Loss: 0.63704164, Log Avg loss: 0.47234842, Global Avg Loss: 1.84402434, Time: 0.0108 Steps: 47940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001112, Sample Num: 17792, Cur Loss: 0.64216304, Cur Avg Loss: 0.63796362, Log Avg loss: 0.73956486, Global Avg Loss: 1.84379401, Time: 0.0126 Steps: 47950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001122, Sample Num: 17952, Cur Loss: 0.65302569, Cur Avg Loss: 0.63776714, Log Avg loss: 0.61591905, Global Avg Loss: 1.84353799, Time: 0.0126 Steps: 47960, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001132, Sample Num: 18112, Cur Loss: 0.34233919, Cur Avg Loss: 0.63720939, Log Avg loss: 0.57462990, Global Avg Loss: 1.84327346, Time: 0.0132 Steps: 47970, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001142, Sample Num: 18272, Cur Loss: 1.25394797, Cur Avg Loss: 0.63678171, Log Avg loss: 0.58836806, Global Avg Loss: 1.84301192, Time: 0.0108 Steps: 47980, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001152, Sample Num: 18432, Cur Loss: 0.43222785, Cur Avg Loss: 0.63684602, Log Avg loss: 0.64419056, Global Avg Loss: 1.84276211, Time: 0.0111 Steps: 47990, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001162, Sample Num: 18592, Cur Loss: 1.03092468, Cur Avg Loss: 0.63732605, Log Avg loss: 0.69262561, Global Avg Loss: 1.84252250, Time: 0.0096 Steps: 48000, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001172, Sample Num: 18752, Cur Loss: 0.33911395, Cur Avg Loss: 0.63569501, Log Avg loss: 0.44616826, Global Avg Loss: 1.84223165, Time: 0.0106 Steps: 48010, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001182, Sample Num: 18912, Cur Loss: 0.50557750, Cur Avg Loss: 0.63681379, Log Avg loss: 0.76793452, Global Avg Loss: 1.84200793, Time: 0.0085 Steps: 48020, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001192, Sample Num: 19072, Cur Loss: 1.02775049, Cur Avg Loss: 0.63621586, Log Avg loss: 0.56554085, Global Avg Loss: 1.84174217, Time: 0.0116 Steps: 48030, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001202, Sample Num: 19232, Cur Loss: 0.60173643, Cur Avg Loss: 0.63724690, Log Avg loss: 0.76014633, Global Avg Loss: 1.84151702, Time: 0.0072 Steps: 48040, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001212, Sample Num: 19392, Cur Loss: 1.14463735, Cur Avg Loss: 0.63728579, Log Avg loss: 0.64195997, Global Avg Loss: 1.84126738, Time: 0.0067 Steps: 48050, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001222, Sample Num: 19552, Cur Loss: 0.95699608, Cur Avg Loss: 0.63840035, Log Avg loss: 0.77348508, Global Avg Loss: 1.84104520, Time: 0.0071 Steps: 48060, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001232, Sample Num: 19712, Cur Loss: 0.92659378, Cur Avg Loss: 0.64036805, Log Avg loss: 0.88082196, Global Avg Loss: 1.84084544, Time: 0.0067 Steps: 48070, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001242, Sample Num: 19872, Cur Loss: 0.41039228, Cur Avg Loss: 0.64011753, Log Avg loss: 0.60925317, Global Avg Loss: 1.84058929, Time: 0.0072 Steps: 48080, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001252, Sample Num: 20032, Cur Loss: 0.84681094, Cur Avg Loss: 0.64200879, Log Avg loss: 0.87690340, Global Avg Loss: 1.84038890, Time: 0.0078 Steps: 48090, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001262, Sample Num: 20192, Cur Loss: 0.65160209, Cur Avg Loss: 0.64186039, Log Avg loss: 0.62328101, Global Avg Loss: 1.84013586, Time: 0.0108 Steps: 48100, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001272, Sample Num: 20352, Cur Loss: 1.25342298, Cur Avg Loss: 0.64108930, Log Avg loss: 0.54377734, Global Avg Loss: 1.83986640, Time: 0.0067 Steps: 48110, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001282, Sample Num: 20512, Cur Loss: 0.44068205, Cur Avg Loss: 0.64045832, Log Avg loss: 0.56019760, Global Avg Loss: 1.83960047, Time: 0.0230 Steps: 48120, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001292, Sample Num: 20672, Cur Loss: 0.77968597, Cur Avg Loss: 0.63891270, Log Avg loss: 0.44076374, Global Avg Loss: 1.83930983, Time: 0.0143 Steps: 48130, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001302, Sample Num: 20832, Cur Loss: 0.35376868, Cur Avg Loss: 0.63774333, Log Avg loss: 0.48666166, Global Avg Loss: 1.83902885, Time: 0.0067 Steps: 48140, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001312, Sample Num: 20992, Cur Loss: 0.76922494, Cur Avg Loss: 0.63961468, Log Avg loss: 0.88326425, Global Avg Loss: 1.83883035, Time: 0.0067 Steps: 48150, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001322, Sample Num: 21152, Cur Loss: 0.67716289, Cur Avg Loss: 0.64053797, Log Avg loss: 0.76167304, Global Avg Loss: 1.83860669, Time: 0.0067 Steps: 48160, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001332, Sample Num: 21312, Cur Loss: 0.99072766, Cur Avg Loss: 0.63999518, Log Avg loss: 0.56823919, Global Avg Loss: 1.83834297, Time: 0.0140 Steps: 48170, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001342, Sample Num: 21472, Cur Loss: 0.26080269, Cur Avg Loss: 0.64030616, Log Avg loss: 0.68172840, Global Avg Loss: 1.83810290, Time: 0.0067 Steps: 48180, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001352, Sample Num: 21632, Cur Loss: 0.39673847, Cur Avg Loss: 0.63889879, Log Avg loss: 0.45002926, Global Avg Loss: 1.83781486, Time: 0.0106 Steps: 48190, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001362, Sample Num: 21792, Cur Loss: 0.46705911, Cur Avg Loss: 0.63896606, Log Avg loss: 0.64806085, Global Avg Loss: 1.83756803, Time: 0.0072 Steps: 48200, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001372, Sample Num: 21952, Cur Loss: 0.67559683, Cur Avg Loss: 0.63941950, Log Avg loss: 0.70117883, Global Avg Loss: 1.83733231, Time: 0.0117 Steps: 48210, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001382, Sample Num: 22112, Cur Loss: 0.60387754, Cur Avg Loss: 0.63944351, Log Avg loss: 0.64273673, Global Avg Loss: 1.83708457, Time: 0.0116 Steps: 48220, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001392, Sample Num: 22272, Cur Loss: 0.36487532, Cur Avg Loss: 0.63893353, Log Avg loss: 0.56845426, Global Avg Loss: 1.83682153, Time: 0.0118 Steps: 48230, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001402, Sample Num: 22432, Cur Loss: 0.68165046, Cur Avg Loss: 0.64031980, Log Avg loss: 0.83328831, Global Avg Loss: 1.83661350, Time: 0.0142 Steps: 48240, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001412, Sample Num: 22592, Cur Loss: 0.24729428, Cur Avg Loss: 0.63955301, Log Avg loss: 0.53204954, Global Avg Loss: 1.83634313, Time: 0.0109 Steps: 48250, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001422, Sample Num: 22752, Cur Loss: 0.42957342, Cur Avg Loss: 0.63989775, Log Avg loss: 0.68857456, Global Avg Loss: 1.83610530, Time: 0.0126 Steps: 48260, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001432, Sample Num: 22912, Cur Loss: 0.16683838, Cur Avg Loss: 0.64002502, Log Avg loss: 0.65812288, Global Avg Loss: 1.83586126, Time: 0.0069 Steps: 48270, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001442, Sample Num: 23072, Cur Loss: 0.61079931, Cur Avg Loss: 0.63951173, Log Avg loss: 0.56600959, Global Avg Loss: 1.83559824, Time: 0.0140 Steps: 48280, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001452, Sample Num: 23232, Cur Loss: 0.43558457, Cur Avg Loss: 0.63890118, Log Avg loss: 0.55085906, Global Avg Loss: 1.83533219, Time: 0.0067 Steps: 48290, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001462, Sample Num: 23392, Cur Loss: 0.72619206, Cur Avg Loss: 0.64016925, Log Avg loss: 0.82429324, Global Avg Loss: 1.83512287, Time: 0.0103 Steps: 48300, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001472, Sample Num: 23552, Cur Loss: 0.75716722, Cur Avg Loss: 0.64106365, Log Avg loss: 0.77182540, Global Avg Loss: 1.83490277, Time: 0.0116 Steps: 48310, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001482, Sample Num: 23712, Cur Loss: 1.37989759, Cur Avg Loss: 0.64300596, Log Avg loss: 0.92891443, Global Avg Loss: 1.83471527, Time: 0.0087 Steps: 48320, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001492, Sample Num: 23872, Cur Loss: 0.81379396, Cur Avg Loss: 0.64398737, Log Avg loss: 0.78943124, Global Avg Loss: 1.83449899, Time: 0.0138 Steps: 48330, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001502, Sample Num: 24032, Cur Loss: 0.52692926, Cur Avg Loss: 0.64335849, Log Avg loss: 0.54952975, Global Avg Loss: 1.83423317, Time: 0.0153 Steps: 48340, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001512, Sample Num: 24192, Cur Loss: 0.59293187, Cur Avg Loss: 0.64378264, Log Avg loss: 0.70749011, Global Avg Loss: 1.83400013, Time: 0.0069 Steps: 48350, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001522, Sample Num: 24352, Cur Loss: 0.26488012, Cur Avg Loss: 0.64376745, Log Avg loss: 0.64147025, Global Avg Loss: 1.83375354, Time: 0.0120 Steps: 48360, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001532, Sample Num: 24512, Cur Loss: 0.57705563, Cur Avg Loss: 0.64292359, Log Avg loss: 0.51448893, Global Avg Loss: 1.83348079, Time: 0.0067 Steps: 48370, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001542, Sample Num: 24672, Cur Loss: 0.95928156, Cur Avg Loss: 0.64318124, Log Avg loss: 0.68265282, Global Avg Loss: 1.83324292, Time: 0.0073 Steps: 48380, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001552, Sample Num: 24832, Cur Loss: 0.93822312, Cur Avg Loss: 0.64335395, Log Avg loss: 0.66998570, Global Avg Loss: 1.83300253, Time: 0.0093 Steps: 48390, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001562, Sample Num: 24992, Cur Loss: 0.35466200, Cur Avg Loss: 0.64369982, Log Avg loss: 0.69737979, Global Avg Loss: 1.83276790, Time: 0.0123 Steps: 48400, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001572, Sample Num: 25152, Cur Loss: 0.42098913, Cur Avg Loss: 0.64309517, Log Avg loss: 0.54864789, Global Avg Loss: 1.83250264, Time: 0.0111 Steps: 48410, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001582, Sample Num: 25312, Cur Loss: 0.38349068, Cur Avg Loss: 0.64286417, Log Avg loss: 0.60655130, Global Avg Loss: 1.83224945, Time: 0.0117 Steps: 48420, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001592, Sample Num: 25472, Cur Loss: 0.58605826, Cur Avg Loss: 0.64320752, Log Avg loss: 0.69752500, Global Avg Loss: 1.83201514, Time: 0.0117 Steps: 48430, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001602, Sample Num: 25632, Cur Loss: 1.32881069, Cur Avg Loss: 0.64461708, Log Avg loss: 0.86901886, Global Avg Loss: 1.83181634, Time: 0.0072 Steps: 48440, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001612, Sample Num: 25792, Cur Loss: 1.07909381, Cur Avg Loss: 0.64530107, Log Avg loss: 0.75487754, Global Avg Loss: 1.83159406, Time: 0.0119 Steps: 48450, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001622, Sample Num: 25952, Cur Loss: 0.48021373, Cur Avg Loss: 0.64450417, Log Avg loss: 0.51604274, Global Avg Loss: 1.83132259, Time: 0.0118 Steps: 48460, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001632, Sample Num: 26112, Cur Loss: 0.34259415, Cur Avg Loss: 0.64402013, Log Avg loss: 0.56550898, Global Avg Loss: 1.83106144, Time: 0.0087 Steps: 48470, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001642, Sample Num: 26272, Cur Loss: 1.05795527, Cur Avg Loss: 0.64524913, Log Avg loss: 0.84582273, Global Avg Loss: 1.83085821, Time: 0.0069 Steps: 48480, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001652, Sample Num: 26432, Cur Loss: 0.73129576, Cur Avg Loss: 0.64453359, Log Avg loss: 0.52704149, Global Avg Loss: 1.83058933, Time: 0.0107 Steps: 48490, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001662, Sample Num: 26592, Cur Loss: 1.04674292, Cur Avg Loss: 0.64454381, Log Avg loss: 0.64623260, Global Avg Loss: 1.83034513, Time: 0.0086 Steps: 48500, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001672, Sample Num: 26752, Cur Loss: 0.77170247, Cur Avg Loss: 0.64390028, Log Avg loss: 0.53694548, Global Avg Loss: 1.83007851, Time: 0.0111 Steps: 48510, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001682, Sample Num: 26912, Cur Loss: 0.23985630, Cur Avg Loss: 0.64315398, Log Avg loss: 0.51837171, Global Avg Loss: 1.82980816, Time: 0.0163 Steps: 48520, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001692, Sample Num: 27072, Cur Loss: 0.30189151, Cur Avg Loss: 0.64358113, Log Avg loss: 0.71542840, Global Avg Loss: 1.82957854, Time: 0.0123 Steps: 48530, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001702, Sample Num: 27232, Cur Loss: 0.65959328, Cur Avg Loss: 0.64272588, Log Avg loss: 0.49801704, Global Avg Loss: 1.82930421, Time: 0.0125 Steps: 48540, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001712, Sample Num: 27392, Cur Loss: 0.39231241, Cur Avg Loss: 0.64356934, Log Avg loss: 0.78712705, Global Avg Loss: 1.82908955, Time: 0.0114 Steps: 48550, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001722, Sample Num: 27552, Cur Loss: 0.24897546, Cur Avg Loss: 0.64351736, Log Avg loss: 0.63461827, Global Avg Loss: 1.82884357, Time: 0.0071 Steps: 48560, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001732, Sample Num: 27712, Cur Loss: 0.59178853, Cur Avg Loss: 0.64210129, Log Avg loss: 0.39825322, Global Avg Loss: 1.82854903, Time: 0.0090 Steps: 48570, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001742, Sample Num: 27872, Cur Loss: 0.53619587, Cur Avg Loss: 0.64265411, Log Avg loss: 0.73840246, Global Avg Loss: 1.82832463, Time: 0.0081 Steps: 48580, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001752, Sample Num: 28032, Cur Loss: 0.25780901, Cur Avg Loss: 0.64277873, Log Avg loss: 0.66448868, Global Avg Loss: 1.82808511, Time: 0.0094 Steps: 48590, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001762, Sample Num: 28192, Cur Loss: 0.44604957, Cur Avg Loss: 0.64179696, Log Avg loss: 0.46979093, Global Avg Loss: 1.82780562, Time: 0.0087 Steps: 48600, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001772, Sample Num: 28352, Cur Loss: 1.30490839, Cur Avg Loss: 0.64284069, Log Avg loss: 0.82674613, Global Avg Loss: 1.82759969, Time: 0.0070 Steps: 48610, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001782, Sample Num: 28512, Cur Loss: 0.70344645, Cur Avg Loss: 0.64345907, Log Avg loss: 0.75303441, Global Avg Loss: 1.82737867, Time: 0.0126 Steps: 48620, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001792, Sample Num: 28672, Cur Loss: 0.50962341, Cur Avg Loss: 0.64308078, Log Avg loss: 0.57567078, Global Avg Loss: 1.82712128, Time: 0.0146 Steps: 48630, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001802, Sample Num: 28832, Cur Loss: 0.37474966, Cur Avg Loss: 0.64262328, Log Avg loss: 0.56063802, Global Avg Loss: 1.82686090, Time: 0.0116 Steps: 48640, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001812, Sample Num: 28992, Cur Loss: 0.46041426, Cur Avg Loss: 0.64348995, Log Avg loss: 0.79966428, Global Avg Loss: 1.82664976, Time: 0.0066 Steps: 48650, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001822, Sample Num: 29152, Cur Loss: 0.08688930, Cur Avg Loss: 0.64293519, Log Avg loss: 0.54241386, Global Avg Loss: 1.82638584, Time: 0.0066 Steps: 48660, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001832, Sample Num: 29312, Cur Loss: 0.31314525, Cur Avg Loss: 0.64257862, Log Avg loss: 0.57761165, Global Avg Loss: 1.82612926, Time: 0.0127 Steps: 48670, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001842, Sample Num: 29472, Cur Loss: 0.86711711, Cur Avg Loss: 0.64259394, Log Avg loss: 0.64539987, Global Avg Loss: 1.82588671, Time: 0.0066 Steps: 48680, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001852, Sample Num: 29632, Cur Loss: 1.14196873, Cur Avg Loss: 0.64264082, Log Avg loss: 0.65127575, Global Avg Loss: 1.82564547, Time: 0.0227 Steps: 48690, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001862, Sample Num: 29792, Cur Loss: 1.27306962, Cur Avg Loss: 0.64223706, Log Avg loss: 0.56746174, Global Avg Loss: 1.82538711, Time: 0.0067 Steps: 48700, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001872, Sample Num: 29952, Cur Loss: 0.51762682, Cur Avg Loss: 0.64158528, Log Avg loss: 0.52022264, Global Avg Loss: 1.82511917, Time: 0.0071 Steps: 48710, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001882, Sample Num: 30112, Cur Loss: 0.63117039, Cur Avg Loss: 0.64044509, Log Avg loss: 0.42700270, Global Avg Loss: 1.82483220, Time: 0.0068 Steps: 48720, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001892, Sample Num: 30272, Cur Loss: 0.49449870, Cur Avg Loss: 0.64036240, Log Avg loss: 0.62479873, Global Avg Loss: 1.82458594, Time: 0.0116 Steps: 48730, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001902, Sample Num: 30432, Cur Loss: 1.06878042, Cur Avg Loss: 0.64047705, Log Avg loss: 0.66216954, Global Avg Loss: 1.82434744, Time: 0.0121 Steps: 48740, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001912, Sample Num: 30592, Cur Loss: 1.09978938, Cur Avg Loss: 0.64048498, Log Avg loss: 0.64199330, Global Avg Loss: 1.82410491, Time: 0.0065 Steps: 48750, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001922, Sample Num: 30752, Cur Loss: 0.33940232, Cur Avg Loss: 0.64027572, Log Avg loss: 0.60026487, Global Avg Loss: 1.82385392, Time: 0.0110 Steps: 48760, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001932, Sample Num: 30912, Cur Loss: 0.58288932, Cur Avg Loss: 0.63963803, Log Avg loss: 0.51707370, Global Avg Loss: 1.82358597, Time: 0.0113 Steps: 48770, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001942, Sample Num: 31072, Cur Loss: 0.76621813, Cur Avg Loss: 0.63985378, Log Avg loss: 0.68153809, Global Avg Loss: 1.82335185, Time: 0.0085 Steps: 48780, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001952, Sample Num: 31232, Cur Loss: 0.40676507, Cur Avg Loss: 0.63907559, Log Avg loss: 0.48795027, Global Avg Loss: 1.82307814, Time: 0.0085 Steps: 48790, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001962, Sample Num: 31392, Cur Loss: 0.54575950, Cur Avg Loss: 0.63893122, Log Avg loss: 0.61075021, Global Avg Loss: 1.82282972, Time: 0.0199 Steps: 48800, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001972, Sample Num: 31552, Cur Loss: 0.26437861, Cur Avg Loss: 0.63929971, Log Avg loss: 0.71159779, Global Avg Loss: 1.82260205, Time: 0.0103 Steps: 48810, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001982, Sample Num: 31712, Cur Loss: 1.11635029, Cur Avg Loss: 0.63989249, Log Avg loss: 0.75678923, Global Avg Loss: 1.82238374, Time: 0.0067 Steps: 48820, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 001992, Sample Num: 31872, Cur Loss: 1.04505944, Cur Avg Loss: 0.64056416, Log Avg loss: 0.77368836, Global Avg Loss: 1.82216897, Time: 0.0120 Steps: 48830, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002002, Sample Num: 32032, Cur Loss: 0.38452029, Cur Avg Loss: 0.64032898, Log Avg loss: 0.59348111, Global Avg Loss: 1.82191740, Time: 0.0125 Steps: 48840, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002012, Sample Num: 32192, Cur Loss: 1.10921693, Cur Avg Loss: 0.63996540, Log Avg loss: 0.56717736, Global Avg Loss: 1.82166054, Time: 0.0073 Steps: 48850, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002022, Sample Num: 32352, Cur Loss: 0.87440062, Cur Avg Loss: 0.64061662, Log Avg loss: 0.77164185, Global Avg Loss: 1.82144564, Time: 0.0085 Steps: 48860, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002032, Sample Num: 32512, Cur Loss: 0.13303411, Cur Avg Loss: 0.64120217, Log Avg loss: 0.75959889, Global Avg Loss: 1.82122836, Time: 0.0123 Steps: 48870, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002042, Sample Num: 32672, Cur Loss: 0.98835039, Cur Avg Loss: 0.64108200, Log Avg loss: 0.61666367, Global Avg Loss: 1.82098193, Time: 0.0122 Steps: 48880, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002052, Sample Num: 32832, Cur Loss: 0.36879864, Cur Avg Loss: 0.64082010, Log Avg loss: 0.58734001, Global Avg Loss: 1.82072960, Time: 0.0068 Steps: 48890, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002062, Sample Num: 32992, Cur Loss: 0.21979696, Cur Avg Loss: 0.64137802, Log Avg loss: 0.75586484, Global Avg Loss: 1.82051183, Time: 0.0071 Steps: 48900, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002072, Sample Num: 33152, Cur Loss: 0.35106111, Cur Avg Loss: 0.64099766, Log Avg loss: 0.56256711, Global Avg Loss: 1.82025464, Time: 0.0067 Steps: 48910, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002082, Sample Num: 33312, Cur Loss: 0.68962663, Cur Avg Loss: 0.64036932, Log Avg loss: 0.51017664, Global Avg Loss: 1.81998684, Time: 0.0066 Steps: 48920, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002092, Sample Num: 33472, Cur Loss: 0.34418857, Cur Avg Loss: 0.64040931, Log Avg loss: 0.64873513, Global Avg Loss: 1.81974746, Time: 0.0114 Steps: 48930, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002102, Sample Num: 33632, Cur Loss: 0.94895989, Cur Avg Loss: 0.64062759, Log Avg loss: 0.68629275, Global Avg Loss: 1.81951586, Time: 0.0114 Steps: 48940, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002112, Sample Num: 33792, Cur Loss: 1.65091276, Cur Avg Loss: 0.64033400, Log Avg loss: 0.57861960, Global Avg Loss: 1.81926236, Time: 0.0066 Steps: 48950, Updated lr: 0.000055 Training, Epoch: 0023, Batch: 002122, Sample Num: 33952, Cur Loss: 0.41648763, Cur Avg Loss: 0.64025895, Log Avg loss: 0.62440844, Global Avg Loss: 1.81901831, Time: 0.0065 Steps: 48960, Updated lr: 0.000055 ***** Running evaluation checkpoint-48967 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-48967 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.205679, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.717606, "eval_total_loss": 504.476816, "eval_mae": 0.605543, "eval_mse": 0.717883, "eval_r2": 0.543666, "eval_sp_statistic": 0.706843, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.745892, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.961037, "test_total_loss": 482.440658, "test_mae": 0.725056, "test_mse": 0.961256, "test_r2": 0.379597, "test_sp_statistic": 0.51812, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.618515, "test_ps_pvalue": 0.0, "lr": 5.451209103840683e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.8188161036324468, "train_cur_epoch_loss": 1361.4610753953457, "train_cur_epoch_avg_loss": 0.6394838306225202, "train_cur_epoch_time": 22.205679416656494, "train_cur_epoch_avg_time": 0.010430098363859322, "epoch": 23, "step": 48967} ################################################## Training, Epoch: 0024, Batch: 000003, Sample Num: 48, Cur Loss: 0.81182164, Cur Avg Loss: 0.60338054, Log Avg loss: 0.46417345, Global Avg Loss: 1.81874164, Time: 0.0067 Steps: 48970, Updated lr: 0.000055 Training, Epoch: 0024, Batch: 000013, Sample Num: 208, Cur Loss: 0.46985048, Cur Avg Loss: 0.67678134, Log Avg loss: 0.69880157, Global Avg Loss: 1.81851299, Time: 0.0230 Steps: 48980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000023, Sample Num: 368, Cur Loss: 0.68623292, Cur Avg Loss: 0.65773354, Log Avg loss: 0.63297140, Global Avg Loss: 1.81827099, Time: 0.0106 Steps: 48990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000033, Sample Num: 528, Cur Loss: 0.80791777, Cur Avg Loss: 0.68240009, Log Avg loss: 0.73913318, Global Avg Loss: 1.81805076, Time: 0.0131 Steps: 49000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000043, Sample Num: 688, Cur Loss: 0.83186555, Cur Avg Loss: 0.68632516, Log Avg loss: 0.69927787, Global Avg Loss: 1.81782249, Time: 0.0118 Steps: 49010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000053, Sample Num: 848, Cur Loss: 0.23302516, Cur Avg Loss: 0.64409608, Log Avg loss: 0.46251106, Global Avg Loss: 1.81754601, Time: 0.0069 Steps: 49020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000063, Sample Num: 1008, Cur Loss: 0.92859650, Cur Avg Loss: 0.62842452, Log Avg loss: 0.54536523, Global Avg Loss: 1.81728654, Time: 0.0066 Steps: 49030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000073, Sample Num: 1168, Cur Loss: 0.48108730, Cur Avg Loss: 0.64273707, Log Avg loss: 0.73290615, Global Avg Loss: 1.81706542, Time: 0.0107 Steps: 49040, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000083, Sample Num: 1328, Cur Loss: 0.36011717, Cur Avg Loss: 0.64728296, Log Avg loss: 0.68046791, Global Avg Loss: 1.81683369, Time: 0.0085 Steps: 49050, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000093, Sample Num: 1488, Cur Loss: 0.33435339, Cur Avg Loss: 0.63864199, Log Avg loss: 0.56692195, Global Avg Loss: 1.81657892, Time: 0.0111 Steps: 49060, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000103, Sample Num: 1648, Cur Loss: 0.71370864, Cur Avg Loss: 0.63557854, Log Avg loss: 0.60708851, Global Avg Loss: 1.81633244, Time: 0.0094 Steps: 49070, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000113, Sample Num: 1808, Cur Loss: 0.47221410, Cur Avg Loss: 0.63336610, Log Avg loss: 0.61057799, Global Avg Loss: 1.81608677, Time: 0.0118 Steps: 49080, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000123, Sample Num: 1968, Cur Loss: 0.22002777, Cur Avg Loss: 0.62785811, Log Avg loss: 0.56561782, Global Avg Loss: 1.81583204, Time: 0.0117 Steps: 49090, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000133, Sample Num: 2128, Cur Loss: 0.88870764, Cur Avg Loss: 0.62111578, Log Avg loss: 0.53818506, Global Avg Loss: 1.81557182, Time: 0.0066 Steps: 49100, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000143, Sample Num: 2288, Cur Loss: 0.53029931, Cur Avg Loss: 0.61837481, Log Avg loss: 0.58191994, Global Avg Loss: 1.81532062, Time: 0.0096 Steps: 49110, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000153, Sample Num: 2448, Cur Loss: 0.75289726, Cur Avg Loss: 0.62048049, Log Avg loss: 0.65059171, Global Avg Loss: 1.81508350, Time: 0.0117 Steps: 49120, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000163, Sample Num: 2608, Cur Loss: 0.16782077, Cur Avg Loss: 0.62595213, Log Avg loss: 0.70966823, Global Avg Loss: 1.81485850, Time: 0.0063 Steps: 49130, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000173, Sample Num: 2768, Cur Loss: 0.45225412, Cur Avg Loss: 0.63207032, Log Avg loss: 0.73179677, Global Avg Loss: 1.81463810, Time: 0.0085 Steps: 49140, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000183, Sample Num: 2928, Cur Loss: 0.85506320, Cur Avg Loss: 0.63991232, Log Avg loss: 0.77557885, Global Avg Loss: 1.81442670, Time: 0.0151 Steps: 49150, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000193, Sample Num: 3088, Cur Loss: 0.31833914, Cur Avg Loss: 0.63189716, Log Avg loss: 0.48521976, Global Avg Loss: 1.81415631, Time: 0.0147 Steps: 49160, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000203, Sample Num: 3248, Cur Loss: 0.62890565, Cur Avg Loss: 0.62688224, Log Avg loss: 0.53009425, Global Avg Loss: 1.81389516, Time: 0.0066 Steps: 49170, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000213, Sample Num: 3408, Cur Loss: 1.04783988, Cur Avg Loss: 0.62532041, Log Avg loss: 0.59361535, Global Avg Loss: 1.81364704, Time: 0.0113 Steps: 49180, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000223, Sample Num: 3568, Cur Loss: 0.23294708, Cur Avg Loss: 0.61889841, Log Avg loss: 0.48210989, Global Avg Loss: 1.81337635, Time: 0.0113 Steps: 49190, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000233, Sample Num: 3728, Cur Loss: 0.36460507, Cur Avg Loss: 0.61547670, Log Avg loss: 0.53917247, Global Avg Loss: 1.81311736, Time: 0.0065 Steps: 49200, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000243, Sample Num: 3888, Cur Loss: 0.79390657, Cur Avg Loss: 0.61977142, Log Avg loss: 0.71983833, Global Avg Loss: 1.81289520, Time: 0.0070 Steps: 49210, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000253, Sample Num: 4048, Cur Loss: 0.55379295, Cur Avg Loss: 0.61954929, Log Avg loss: 0.61415173, Global Avg Loss: 1.81265165, Time: 0.0115 Steps: 49220, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000263, Sample Num: 4208, Cur Loss: 1.55235791, Cur Avg Loss: 0.62122547, Log Avg loss: 0.66363277, Global Avg Loss: 1.81241825, Time: 0.0110 Steps: 49230, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000273, Sample Num: 4368, Cur Loss: 0.38654274, Cur Avg Loss: 0.62061104, Log Avg loss: 0.60445144, Global Avg Loss: 1.81217293, Time: 0.0123 Steps: 49240, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000283, Sample Num: 4528, Cur Loss: 0.51944053, Cur Avg Loss: 0.61651310, Log Avg loss: 0.50463952, Global Avg Loss: 1.81190744, Time: 0.0063 Steps: 49250, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000293, Sample Num: 4688, Cur Loss: 0.28331700, Cur Avg Loss: 0.61416776, Log Avg loss: 0.54779463, Global Avg Loss: 1.81165082, Time: 0.0209 Steps: 49260, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000303, Sample Num: 4848, Cur Loss: 0.37954476, Cur Avg Loss: 0.61360207, Log Avg loss: 0.59702718, Global Avg Loss: 1.81140429, Time: 0.0067 Steps: 49270, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000313, Sample Num: 5008, Cur Loss: 0.33426133, Cur Avg Loss: 0.61487198, Log Avg loss: 0.65335016, Global Avg Loss: 1.81116930, Time: 0.0222 Steps: 49280, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000323, Sample Num: 5168, Cur Loss: 0.48421776, Cur Avg Loss: 0.61235579, Log Avg loss: 0.53359918, Global Avg Loss: 1.81091010, Time: 0.0186 Steps: 49290, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000333, Sample Num: 5328, Cur Loss: 0.53875309, Cur Avg Loss: 0.60895848, Log Avg loss: 0.49922535, Global Avg Loss: 1.81064404, Time: 0.0065 Steps: 49300, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000343, Sample Num: 5488, Cur Loss: 0.50662005, Cur Avg Loss: 0.60820419, Log Avg loss: 0.58308632, Global Avg Loss: 1.81039510, Time: 0.0237 Steps: 49310, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000353, Sample Num: 5648, Cur Loss: 0.62596607, Cur Avg Loss: 0.61387006, Log Avg loss: 0.80820957, Global Avg Loss: 1.81019190, Time: 0.0069 Steps: 49320, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000363, Sample Num: 5808, Cur Loss: 1.09149528, Cur Avg Loss: 0.61405059, Log Avg loss: 0.62042319, Global Avg Loss: 1.80995071, Time: 0.0219 Steps: 49330, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000373, Sample Num: 5968, Cur Loss: 0.09104966, Cur Avg Loss: 0.61290782, Log Avg loss: 0.57142514, Global Avg Loss: 1.80969969, Time: 0.0174 Steps: 49340, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000383, Sample Num: 6128, Cur Loss: 0.40482968, Cur Avg Loss: 0.61242595, Log Avg loss: 0.59445229, Global Avg Loss: 1.80945344, Time: 0.0073 Steps: 49350, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000393, Sample Num: 6288, Cur Loss: 0.77219409, Cur Avg Loss: 0.61630364, Log Avg loss: 0.76481907, Global Avg Loss: 1.80924180, Time: 0.0069 Steps: 49360, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000403, Sample Num: 6448, Cur Loss: 0.47665852, Cur Avg Loss: 0.61755073, Log Avg loss: 0.66656132, Global Avg Loss: 1.80901035, Time: 0.0064 Steps: 49370, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000413, Sample Num: 6608, Cur Loss: 0.75476289, Cur Avg Loss: 0.61840213, Log Avg loss: 0.65271385, Global Avg Loss: 1.80877619, Time: 0.0064 Steps: 49380, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000423, Sample Num: 6768, Cur Loss: 0.56020218, Cur Avg Loss: 0.62278519, Log Avg loss: 0.80380555, Global Avg Loss: 1.80857271, Time: 0.0069 Steps: 49390, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000433, Sample Num: 6928, Cur Loss: 0.27495277, Cur Avg Loss: 0.61950213, Log Avg loss: 0.48062861, Global Avg Loss: 1.80830390, Time: 0.0068 Steps: 49400, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000443, Sample Num: 7088, Cur Loss: 0.59125090, Cur Avg Loss: 0.62049141, Log Avg loss: 0.66332732, Global Avg Loss: 1.80807217, Time: 0.0071 Steps: 49410, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000453, Sample Num: 7248, Cur Loss: 0.29180634, Cur Avg Loss: 0.61724954, Log Avg loss: 0.47363450, Global Avg Loss: 1.80780215, Time: 0.0069 Steps: 49420, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000463, Sample Num: 7408, Cur Loss: 0.75669408, Cur Avg Loss: 0.61806870, Log Avg loss: 0.65517684, Global Avg Loss: 1.80756897, Time: 0.0069 Steps: 49430, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000473, Sample Num: 7568, Cur Loss: 0.36114082, Cur Avg Loss: 0.62012218, Log Avg loss: 0.71519825, Global Avg Loss: 1.80734802, Time: 0.0109 Steps: 49440, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000483, Sample Num: 7728, Cur Loss: 0.27061051, Cur Avg Loss: 0.62172091, Log Avg loss: 0.69734091, Global Avg Loss: 1.80712355, Time: 0.0105 Steps: 49450, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000493, Sample Num: 7888, Cur Loss: 0.39670575, Cur Avg Loss: 0.62012096, Log Avg loss: 0.54284323, Global Avg Loss: 1.80686793, Time: 0.0115 Steps: 49460, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000503, Sample Num: 8048, Cur Loss: 0.46743149, Cur Avg Loss: 0.62108854, Log Avg loss: 0.66879004, Global Avg Loss: 1.80663788, Time: 0.0122 Steps: 49470, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000513, Sample Num: 8208, Cur Loss: 0.22411887, Cur Avg Loss: 0.62156513, Log Avg loss: 0.64553786, Global Avg Loss: 1.80640321, Time: 0.0141 Steps: 49480, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000523, Sample Num: 8368, Cur Loss: 0.68935448, Cur Avg Loss: 0.61794477, Log Avg loss: 0.43222029, Global Avg Loss: 1.80612555, Time: 0.0140 Steps: 49490, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000533, Sample Num: 8528, Cur Loss: 0.46484852, Cur Avg Loss: 0.62055867, Log Avg loss: 0.75726552, Global Avg Loss: 1.80591365, Time: 0.0067 Steps: 49500, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000543, Sample Num: 8688, Cur Loss: 0.33382487, Cur Avg Loss: 0.61906288, Log Avg loss: 0.53933753, Global Avg Loss: 1.80565783, Time: 0.0158 Steps: 49510, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000553, Sample Num: 8848, Cur Loss: 0.56558549, Cur Avg Loss: 0.62001712, Log Avg loss: 0.67183227, Global Avg Loss: 1.80542887, Time: 0.0228 Steps: 49520, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000563, Sample Num: 9008, Cur Loss: 0.63573670, Cur Avg Loss: 0.61901638, Log Avg loss: 0.56367557, Global Avg Loss: 1.80517816, Time: 0.0066 Steps: 49530, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000573, Sample Num: 9168, Cur Loss: 0.68056571, Cur Avg Loss: 0.61848503, Log Avg loss: 0.58856958, Global Avg Loss: 1.80493258, Time: 0.0095 Steps: 49540, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000583, Sample Num: 9328, Cur Loss: 1.41827989, Cur Avg Loss: 0.61880442, Log Avg loss: 0.63710553, Global Avg Loss: 1.80469689, Time: 0.0116 Steps: 49550, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000593, Sample Num: 9488, Cur Loss: 0.51310116, Cur Avg Loss: 0.61812574, Log Avg loss: 0.57855856, Global Avg Loss: 1.80444949, Time: 0.0067 Steps: 49560, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000603, Sample Num: 9648, Cur Loss: 0.96706426, Cur Avg Loss: 0.62063101, Log Avg loss: 0.76919406, Global Avg Loss: 1.80424064, Time: 0.0120 Steps: 49570, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000613, Sample Num: 9808, Cur Loss: 0.43835518, Cur Avg Loss: 0.62065211, Log Avg loss: 0.62192429, Global Avg Loss: 1.80400218, Time: 0.0117 Steps: 49580, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000623, Sample Num: 9968, Cur Loss: 0.76801360, Cur Avg Loss: 0.61944098, Log Avg loss: 0.54519877, Global Avg Loss: 1.80374833, Time: 0.0067 Steps: 49590, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000633, Sample Num: 10128, Cur Loss: 0.72090811, Cur Avg Loss: 0.61968498, Log Avg loss: 0.63488626, Global Avg Loss: 1.80351268, Time: 0.0109 Steps: 49600, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000643, Sample Num: 10288, Cur Loss: 0.55500346, Cur Avg Loss: 0.61645033, Log Avg loss: 0.41169705, Global Avg Loss: 1.80323212, Time: 0.0225 Steps: 49610, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000653, Sample Num: 10448, Cur Loss: 0.64011014, Cur Avg Loss: 0.61639506, Log Avg loss: 0.61284119, Global Avg Loss: 1.80299222, Time: 0.0117 Steps: 49620, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000663, Sample Num: 10608, Cur Loss: 0.43680960, Cur Avg Loss: 0.61585132, Log Avg loss: 0.58034505, Global Avg Loss: 1.80274587, Time: 0.0126 Steps: 49630, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000673, Sample Num: 10768, Cur Loss: 0.60981548, Cur Avg Loss: 0.61600529, Log Avg loss: 0.62621294, Global Avg Loss: 1.80250886, Time: 0.0067 Steps: 49640, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000683, Sample Num: 10928, Cur Loss: 0.84672135, Cur Avg Loss: 0.61664025, Log Avg loss: 0.65937316, Global Avg Loss: 1.80227862, Time: 0.0126 Steps: 49650, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000693, Sample Num: 11088, Cur Loss: 0.51719773, Cur Avg Loss: 0.61502418, Log Avg loss: 0.50464679, Global Avg Loss: 1.80201732, Time: 0.0133 Steps: 49660, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000703, Sample Num: 11248, Cur Loss: 0.31304604, Cur Avg Loss: 0.61424785, Log Avg loss: 0.56044812, Global Avg Loss: 1.80176735, Time: 0.0067 Steps: 49670, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000713, Sample Num: 11408, Cur Loss: 0.18307281, Cur Avg Loss: 0.61292126, Log Avg loss: 0.51966200, Global Avg Loss: 1.80150928, Time: 0.0211 Steps: 49680, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000723, Sample Num: 11568, Cur Loss: 0.57574987, Cur Avg Loss: 0.61331138, Log Avg loss: 0.64112712, Global Avg Loss: 1.80127576, Time: 0.0110 Steps: 49690, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000733, Sample Num: 11728, Cur Loss: 0.83651233, Cur Avg Loss: 0.61370666, Log Avg loss: 0.64228518, Global Avg Loss: 1.80104256, Time: 0.0136 Steps: 49700, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000743, Sample Num: 11888, Cur Loss: 0.26278120, Cur Avg Loss: 0.61560860, Log Avg loss: 0.75502055, Global Avg Loss: 1.80083213, Time: 0.0110 Steps: 49710, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000753, Sample Num: 12048, Cur Loss: 1.22041655, Cur Avg Loss: 0.61675668, Log Avg loss: 0.70205951, Global Avg Loss: 1.80061114, Time: 0.0067 Steps: 49720, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000763, Sample Num: 12208, Cur Loss: 0.49782264, Cur Avg Loss: 0.61634435, Log Avg loss: 0.58529575, Global Avg Loss: 1.80036676, Time: 0.0139 Steps: 49730, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000773, Sample Num: 12368, Cur Loss: 0.57714498, Cur Avg Loss: 0.61676487, Log Avg loss: 0.64885041, Global Avg Loss: 1.80013525, Time: 0.0131 Steps: 49740, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000783, Sample Num: 12528, Cur Loss: 0.22993501, Cur Avg Loss: 0.61651373, Log Avg loss: 0.59710110, Global Avg Loss: 1.79989344, Time: 0.0067 Steps: 49750, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000793, Sample Num: 12688, Cur Loss: 0.84293407, Cur Avg Loss: 0.61620838, Log Avg loss: 0.59229936, Global Avg Loss: 1.79965075, Time: 0.0132 Steps: 49760, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000803, Sample Num: 12848, Cur Loss: 0.44896537, Cur Avg Loss: 0.61945698, Log Avg loss: 0.87707106, Global Avg Loss: 1.79946538, Time: 0.0154 Steps: 49770, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000813, Sample Num: 13008, Cur Loss: 0.49609876, Cur Avg Loss: 0.61825569, Log Avg loss: 0.52179170, Global Avg Loss: 1.79920872, Time: 0.0093 Steps: 49780, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000823, Sample Num: 13168, Cur Loss: 0.48189533, Cur Avg Loss: 0.62013506, Log Avg loss: 0.77292831, Global Avg Loss: 1.79900260, Time: 0.0107 Steps: 49790, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000833, Sample Num: 13328, Cur Loss: 0.97786313, Cur Avg Loss: 0.62020364, Log Avg loss: 0.62584775, Global Avg Loss: 1.79876702, Time: 0.0067 Steps: 49800, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000843, Sample Num: 13488, Cur Loss: 0.58257079, Cur Avg Loss: 0.62038866, Log Avg loss: 0.63580031, Global Avg Loss: 1.79853354, Time: 0.0122 Steps: 49810, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000853, Sample Num: 13648, Cur Loss: 0.42230445, Cur Avg Loss: 0.61954773, Log Avg loss: 0.54865733, Global Avg Loss: 1.79828266, Time: 0.0071 Steps: 49820, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000863, Sample Num: 13808, Cur Loss: 0.48273498, Cur Avg Loss: 0.62022898, Log Avg loss: 0.67834005, Global Avg Loss: 1.79805791, Time: 0.0167 Steps: 49830, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000873, Sample Num: 13968, Cur Loss: 0.54092121, Cur Avg Loss: 0.62056406, Log Avg loss: 0.64948151, Global Avg Loss: 1.79782746, Time: 0.0067 Steps: 49840, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000883, Sample Num: 14128, Cur Loss: 0.88738126, Cur Avg Loss: 0.62097474, Log Avg loss: 0.65682695, Global Avg Loss: 1.79759857, Time: 0.0078 Steps: 49850, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000893, Sample Num: 14288, Cur Loss: 0.53896821, Cur Avg Loss: 0.62325550, Log Avg loss: 0.82464657, Global Avg Loss: 1.79740344, Time: 0.0186 Steps: 49860, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000903, Sample Num: 14448, Cur Loss: 0.60735571, Cur Avg Loss: 0.62404589, Log Avg loss: 0.69462723, Global Avg Loss: 1.79718231, Time: 0.0073 Steps: 49870, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000913, Sample Num: 14608, Cur Loss: 0.43483967, Cur Avg Loss: 0.62495021, Log Avg loss: 0.70661089, Global Avg Loss: 1.79696367, Time: 0.0064 Steps: 49880, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000923, Sample Num: 14768, Cur Loss: 0.65294415, Cur Avg Loss: 0.62604740, Log Avg loss: 0.72622080, Global Avg Loss: 1.79674905, Time: 0.0069 Steps: 49890, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000933, Sample Num: 14928, Cur Loss: 0.49995700, Cur Avg Loss: 0.62478045, Log Avg loss: 0.50784066, Global Avg Loss: 1.79649075, Time: 0.0236 Steps: 49900, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000943, Sample Num: 15088, Cur Loss: 0.86682594, Cur Avg Loss: 0.62570730, Log Avg loss: 0.71218280, Global Avg Loss: 1.79627349, Time: 0.0114 Steps: 49910, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000953, Sample Num: 15248, Cur Loss: 0.86248612, Cur Avg Loss: 0.62667345, Log Avg loss: 0.71778147, Global Avg Loss: 1.79605745, Time: 0.0123 Steps: 49920, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000963, Sample Num: 15408, Cur Loss: 0.36331272, Cur Avg Loss: 0.62550290, Log Avg loss: 0.51394914, Global Avg Loss: 1.79580067, Time: 0.0071 Steps: 49930, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000973, Sample Num: 15568, Cur Loss: 0.52076650, Cur Avg Loss: 0.62547674, Log Avg loss: 0.62295767, Global Avg Loss: 1.79556582, Time: 0.0067 Steps: 49940, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000983, Sample Num: 15728, Cur Loss: 0.71617031, Cur Avg Loss: 0.62650323, Log Avg loss: 0.72638058, Global Avg Loss: 1.79535177, Time: 0.0071 Steps: 49950, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 000993, Sample Num: 15888, Cur Loss: 0.87550724, Cur Avg Loss: 0.62799130, Log Avg loss: 0.77426878, Global Avg Loss: 1.79514739, Time: 0.0064 Steps: 49960, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001003, Sample Num: 16048, Cur Loss: 0.71971440, Cur Avg Loss: 0.62783070, Log Avg loss: 0.61188241, Global Avg Loss: 1.79491059, Time: 0.0073 Steps: 49970, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001013, Sample Num: 16208, Cur Loss: 0.23745093, Cur Avg Loss: 0.62645567, Log Avg loss: 0.48854020, Global Avg Loss: 1.79464921, Time: 0.0065 Steps: 49980, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001023, Sample Num: 16368, Cur Loss: 0.57030427, Cur Avg Loss: 0.62589096, Log Avg loss: 0.56868645, Global Avg Loss: 1.79440397, Time: 0.0066 Steps: 49990, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001033, Sample Num: 16528, Cur Loss: 1.59170973, Cur Avg Loss: 0.62742740, Log Avg loss: 0.78460502, Global Avg Loss: 1.79420201, Time: 0.0155 Steps: 50000, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001043, Sample Num: 16688, Cur Loss: 1.12599242, Cur Avg Loss: 0.62917478, Log Avg loss: 0.80967923, Global Avg Loss: 1.79400515, Time: 0.0225 Steps: 50010, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001053, Sample Num: 16848, Cur Loss: 0.65292448, Cur Avg Loss: 0.62903380, Log Avg loss: 0.61432998, Global Avg Loss: 1.79376931, Time: 0.0111 Steps: 50020, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001063, Sample Num: 17008, Cur Loss: 0.44225135, Cur Avg Loss: 0.62980241, Log Avg loss: 0.71073643, Global Avg Loss: 1.79355283, Time: 0.0163 Steps: 50030, Updated lr: 0.000054 Training, Epoch: 0024, Batch: 001073, Sample Num: 17168, Cur Loss: 0.59967285, Cur Avg Loss: 0.63016226, Log Avg loss: 0.66841424, Global Avg Loss: 1.79332798, Time: 0.0140 Steps: 50040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001083, Sample Num: 17328, Cur Loss: 0.76099151, Cur Avg Loss: 0.62995122, Log Avg loss: 0.60730685, Global Avg Loss: 1.79309102, Time: 0.0094 Steps: 50050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001093, Sample Num: 17488, Cur Loss: 1.05892003, Cur Avg Loss: 0.63097314, Log Avg loss: 0.74164700, Global Avg Loss: 1.79288098, Time: 0.0090 Steps: 50060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001103, Sample Num: 17648, Cur Loss: 0.31383431, Cur Avg Loss: 0.62970581, Log Avg loss: 0.49118617, Global Avg Loss: 1.79262100, Time: 0.0125 Steps: 50070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001113, Sample Num: 17808, Cur Loss: 1.00649226, Cur Avg Loss: 0.62929837, Log Avg loss: 0.58435834, Global Avg Loss: 1.79237974, Time: 0.0114 Steps: 50080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001123, Sample Num: 17968, Cur Loss: 0.85951197, Cur Avg Loss: 0.62980627, Log Avg loss: 0.68633515, Global Avg Loss: 1.79215893, Time: 0.0115 Steps: 50090, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001133, Sample Num: 18128, Cur Loss: 0.45317990, Cur Avg Loss: 0.62903155, Log Avg loss: 0.54203100, Global Avg Loss: 1.79190940, Time: 0.0069 Steps: 50100, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001143, Sample Num: 18288, Cur Loss: 1.06049120, Cur Avg Loss: 0.63165116, Log Avg loss: 0.92845332, Global Avg Loss: 1.79173709, Time: 0.0115 Steps: 50110, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001153, Sample Num: 18448, Cur Loss: 0.44205588, Cur Avg Loss: 0.63039990, Log Avg loss: 0.48737979, Global Avg Loss: 1.79147684, Time: 0.0085 Steps: 50120, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001163, Sample Num: 18608, Cur Loss: 0.39555806, Cur Avg Loss: 0.63102454, Log Avg loss: 0.70304650, Global Avg Loss: 1.79125972, Time: 0.0120 Steps: 50130, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001173, Sample Num: 18768, Cur Loss: 0.34004390, Cur Avg Loss: 0.63063255, Log Avg loss: 0.58504411, Global Avg Loss: 1.79101915, Time: 0.0066 Steps: 50140, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001183, Sample Num: 18928, Cur Loss: 0.26492754, Cur Avg Loss: 0.62985130, Log Avg loss: 0.53820989, Global Avg Loss: 1.79076934, Time: 0.0068 Steps: 50150, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001193, Sample Num: 19088, Cur Loss: 0.53950757, Cur Avg Loss: 0.63081423, Log Avg loss: 0.74472924, Global Avg Loss: 1.79056080, Time: 0.0153 Steps: 50160, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001203, Sample Num: 19248, Cur Loss: 0.45361072, Cur Avg Loss: 0.63066691, Log Avg loss: 0.61309145, Global Avg Loss: 1.79032610, Time: 0.0113 Steps: 50170, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001213, Sample Num: 19408, Cur Loss: 0.63250196, Cur Avg Loss: 0.62920803, Log Avg loss: 0.45370454, Global Avg Loss: 1.79005973, Time: 0.0066 Steps: 50180, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001223, Sample Num: 19568, Cur Loss: 0.36363718, Cur Avg Loss: 0.62949090, Log Avg loss: 0.66380382, Global Avg Loss: 1.78983534, Time: 0.0065 Steps: 50190, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001233, Sample Num: 19728, Cur Loss: 1.50408518, Cur Avg Loss: 0.62935628, Log Avg loss: 0.61289237, Global Avg Loss: 1.78960089, Time: 0.0132 Steps: 50200, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001243, Sample Num: 19888, Cur Loss: 0.65736985, Cur Avg Loss: 0.62859776, Log Avg loss: 0.53507236, Global Avg Loss: 1.78935103, Time: 0.0214 Steps: 50210, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001253, Sample Num: 20048, Cur Loss: 0.45218697, Cur Avg Loss: 0.62792018, Log Avg loss: 0.54369614, Global Avg Loss: 1.78910299, Time: 0.0129 Steps: 50220, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001263, Sample Num: 20208, Cur Loss: 0.50634867, Cur Avg Loss: 0.62666949, Log Avg loss: 0.46995775, Global Avg Loss: 1.78884037, Time: 0.0136 Steps: 50230, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001273, Sample Num: 20368, Cur Loss: 0.64137447, Cur Avg Loss: 0.62670496, Log Avg loss: 0.63118476, Global Avg Loss: 1.78860994, Time: 0.0066 Steps: 50240, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001283, Sample Num: 20528, Cur Loss: 0.19198546, Cur Avg Loss: 0.62577708, Log Avg loss: 0.50765890, Global Avg Loss: 1.78835503, Time: 0.0141 Steps: 50250, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001293, Sample Num: 20688, Cur Loss: 0.44239891, Cur Avg Loss: 0.62482591, Log Avg loss: 0.50279100, Global Avg Loss: 1.78809925, Time: 0.0113 Steps: 50260, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001303, Sample Num: 20848, Cur Loss: 0.72072840, Cur Avg Loss: 0.62457434, Log Avg loss: 0.59204607, Global Avg Loss: 1.78786132, Time: 0.0066 Steps: 50270, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001313, Sample Num: 21008, Cur Loss: 0.80595440, Cur Avg Loss: 0.62596811, Log Avg loss: 0.80757553, Global Avg Loss: 1.78766635, Time: 0.0143 Steps: 50280, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001323, Sample Num: 21168, Cur Loss: 0.24236745, Cur Avg Loss: 0.62590437, Log Avg loss: 0.61753610, Global Avg Loss: 1.78743368, Time: 0.0119 Steps: 50290, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001333, Sample Num: 21328, Cur Loss: 0.28939024, Cur Avg Loss: 0.62628143, Log Avg loss: 0.67616561, Global Avg Loss: 1.78721275, Time: 0.0111 Steps: 50300, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001343, Sample Num: 21488, Cur Loss: 0.76539642, Cur Avg Loss: 0.62671833, Log Avg loss: 0.68495708, Global Avg Loss: 1.78699366, Time: 0.0066 Steps: 50310, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001353, Sample Num: 21648, Cur Loss: 0.30422378, Cur Avg Loss: 0.62619386, Log Avg loss: 0.55575799, Global Avg Loss: 1.78674898, Time: 0.0155 Steps: 50320, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001363, Sample Num: 21808, Cur Loss: 0.33071375, Cur Avg Loss: 0.62558395, Log Avg loss: 0.54306347, Global Avg Loss: 1.78650187, Time: 0.0069 Steps: 50330, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001373, Sample Num: 21968, Cur Loss: 0.22014484, Cur Avg Loss: 0.62603178, Log Avg loss: 0.68707041, Global Avg Loss: 1.78628347, Time: 0.0091 Steps: 50340, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001383, Sample Num: 22128, Cur Loss: 0.15242213, Cur Avg Loss: 0.62673033, Log Avg loss: 0.72264190, Global Avg Loss: 1.78607222, Time: 0.0068 Steps: 50350, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001393, Sample Num: 22288, Cur Loss: 0.79136455, Cur Avg Loss: 0.62583177, Log Avg loss: 0.50156041, Global Avg Loss: 1.78581715, Time: 0.0119 Steps: 50360, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001403, Sample Num: 22448, Cur Loss: 0.68296123, Cur Avg Loss: 0.62586363, Log Avg loss: 0.63030244, Global Avg Loss: 1.78558775, Time: 0.0120 Steps: 50370, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001413, Sample Num: 22608, Cur Loss: 0.48343116, Cur Avg Loss: 0.62602282, Log Avg loss: 0.64835736, Global Avg Loss: 1.78536202, Time: 0.0232 Steps: 50380, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001423, Sample Num: 22768, Cur Loss: 0.22621629, Cur Avg Loss: 0.62541108, Log Avg loss: 0.53897155, Global Avg Loss: 1.78511467, Time: 0.0220 Steps: 50390, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001433, Sample Num: 22928, Cur Loss: 0.90039873, Cur Avg Loss: 0.62593765, Log Avg loss: 0.70086810, Global Avg Loss: 1.78489954, Time: 0.0115 Steps: 50400, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001443, Sample Num: 23088, Cur Loss: 0.63348234, Cur Avg Loss: 0.62735312, Log Avg loss: 0.83019062, Global Avg Loss: 1.78471015, Time: 0.0111 Steps: 50410, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001453, Sample Num: 23248, Cur Loss: 0.34763184, Cur Avg Loss: 0.62834033, Log Avg loss: 0.77079392, Global Avg Loss: 1.78450906, Time: 0.0068 Steps: 50420, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001463, Sample Num: 23408, Cur Loss: 0.54599518, Cur Avg Loss: 0.62746357, Log Avg loss: 0.50007051, Global Avg Loss: 1.78425436, Time: 0.0234 Steps: 50430, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001473, Sample Num: 23568, Cur Loss: 0.81500983, Cur Avg Loss: 0.62681002, Log Avg loss: 0.53119646, Global Avg Loss: 1.78400593, Time: 0.0066 Steps: 50440, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001483, Sample Num: 23728, Cur Loss: 0.41965687, Cur Avg Loss: 0.62683240, Log Avg loss: 0.63012859, Global Avg Loss: 1.78377722, Time: 0.0116 Steps: 50450, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001493, Sample Num: 23888, Cur Loss: 1.20303988, Cur Avg Loss: 0.62650769, Log Avg loss: 0.57835291, Global Avg Loss: 1.78353833, Time: 0.0143 Steps: 50460, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001503, Sample Num: 24048, Cur Loss: 0.49402127, Cur Avg Loss: 0.62661748, Log Avg loss: 0.64300976, Global Avg Loss: 1.78331235, Time: 0.0124 Steps: 50470, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001513, Sample Num: 24208, Cur Loss: 0.84777272, Cur Avg Loss: 0.62660562, Log Avg loss: 0.62482325, Global Avg Loss: 1.78308285, Time: 0.0198 Steps: 50480, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001523, Sample Num: 24368, Cur Loss: 1.23218799, Cur Avg Loss: 0.62629599, Log Avg loss: 0.57944930, Global Avg Loss: 1.78284446, Time: 0.0064 Steps: 50490, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001533, Sample Num: 24528, Cur Loss: 0.49480134, Cur Avg Loss: 0.62588055, Log Avg loss: 0.56260781, Global Avg Loss: 1.78260283, Time: 0.0072 Steps: 50500, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001543, Sample Num: 24688, Cur Loss: 0.34552580, Cur Avg Loss: 0.62450640, Log Avg loss: 0.41384891, Global Avg Loss: 1.78233185, Time: 0.0069 Steps: 50510, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001553, Sample Num: 24848, Cur Loss: 0.25707033, Cur Avg Loss: 0.62510135, Log Avg loss: 0.71690231, Global Avg Loss: 1.78212095, Time: 0.0114 Steps: 50520, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001563, Sample Num: 25008, Cur Loss: 0.69237393, Cur Avg Loss: 0.62503043, Log Avg loss: 0.61401735, Global Avg Loss: 1.78188978, Time: 0.0117 Steps: 50530, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001573, Sample Num: 25168, Cur Loss: 0.60319209, Cur Avg Loss: 0.62509963, Log Avg loss: 0.63591548, Global Avg Loss: 1.78166304, Time: 0.0076 Steps: 50540, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001583, Sample Num: 25328, Cur Loss: 0.87346530, Cur Avg Loss: 0.62622939, Log Avg loss: 0.80394087, Global Avg Loss: 1.78146962, Time: 0.0109 Steps: 50550, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001593, Sample Num: 25488, Cur Loss: 1.24189472, Cur Avg Loss: 0.62709187, Log Avg loss: 0.76362211, Global Avg Loss: 1.78126830, Time: 0.0068 Steps: 50560, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001603, Sample Num: 25648, Cur Loss: 0.77926993, Cur Avg Loss: 0.62689934, Log Avg loss: 0.59622950, Global Avg Loss: 1.78103397, Time: 0.0070 Steps: 50570, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001613, Sample Num: 25808, Cur Loss: 0.97914636, Cur Avg Loss: 0.62708139, Log Avg loss: 0.65626357, Global Avg Loss: 1.78081159, Time: 0.0122 Steps: 50580, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001623, Sample Num: 25968, Cur Loss: 0.32430112, Cur Avg Loss: 0.62602741, Log Avg loss: 0.45602027, Global Avg Loss: 1.78054973, Time: 0.0069 Steps: 50590, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001633, Sample Num: 26128, Cur Loss: 0.37515891, Cur Avg Loss: 0.62651861, Log Avg loss: 0.70624009, Global Avg Loss: 1.78033741, Time: 0.0114 Steps: 50600, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001643, Sample Num: 26288, Cur Loss: 0.37442556, Cur Avg Loss: 0.62689633, Log Avg loss: 0.68857893, Global Avg Loss: 1.78012169, Time: 0.0112 Steps: 50610, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001653, Sample Num: 26448, Cur Loss: 0.71831369, Cur Avg Loss: 0.62666391, Log Avg loss: 0.58847748, Global Avg Loss: 1.77988628, Time: 0.0071 Steps: 50620, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001663, Sample Num: 26608, Cur Loss: 0.32222903, Cur Avg Loss: 0.62587930, Log Avg loss: 0.49618298, Global Avg Loss: 1.77963274, Time: 0.0137 Steps: 50630, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001673, Sample Num: 26768, Cur Loss: 0.43634456, Cur Avg Loss: 0.62524227, Log Avg loss: 0.51930358, Global Avg Loss: 1.77938386, Time: 0.0068 Steps: 50640, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001683, Sample Num: 26928, Cur Loss: 0.57608008, Cur Avg Loss: 0.62468531, Log Avg loss: 0.53150633, Global Avg Loss: 1.77913748, Time: 0.0069 Steps: 50650, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001693, Sample Num: 27088, Cur Loss: 0.52168095, Cur Avg Loss: 0.62509086, Log Avg loss: 0.69334447, Global Avg Loss: 1.77892315, Time: 0.0152 Steps: 50660, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001703, Sample Num: 27248, Cur Loss: 0.85648644, Cur Avg Loss: 0.62537707, Log Avg loss: 0.67383355, Global Avg Loss: 1.77870506, Time: 0.0066 Steps: 50670, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001713, Sample Num: 27408, Cur Loss: 0.36175644, Cur Avg Loss: 0.62465980, Log Avg loss: 0.50250855, Global Avg Loss: 1.77845324, Time: 0.0115 Steps: 50680, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001723, Sample Num: 27568, Cur Loss: 0.70222270, Cur Avg Loss: 0.62562391, Log Avg loss: 0.79077565, Global Avg Loss: 1.77825840, Time: 0.0142 Steps: 50690, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001733, Sample Num: 27728, Cur Loss: 0.57724679, Cur Avg Loss: 0.62716762, Log Avg loss: 0.89314922, Global Avg Loss: 1.77808382, Time: 0.0109 Steps: 50700, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001743, Sample Num: 27888, Cur Loss: 0.52616382, Cur Avg Loss: 0.62776947, Log Avg loss: 0.73206895, Global Avg Loss: 1.77787755, Time: 0.0117 Steps: 50710, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001753, Sample Num: 28048, Cur Loss: 0.99475765, Cur Avg Loss: 0.62721523, Log Avg loss: 0.53061139, Global Avg Loss: 1.77763163, Time: 0.0108 Steps: 50720, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001763, Sample Num: 28208, Cur Loss: 0.45449972, Cur Avg Loss: 0.62696808, Log Avg loss: 0.58364292, Global Avg Loss: 1.77739627, Time: 0.0227 Steps: 50730, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001773, Sample Num: 28368, Cur Loss: 0.52478349, Cur Avg Loss: 0.62749343, Log Avg loss: 0.72011293, Global Avg Loss: 1.77718790, Time: 0.0123 Steps: 50740, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001783, Sample Num: 28528, Cur Loss: 0.62629187, Cur Avg Loss: 0.62791442, Log Avg loss: 0.70255644, Global Avg Loss: 1.77697615, Time: 0.0111 Steps: 50750, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001793, Sample Num: 28688, Cur Loss: 0.40580177, Cur Avg Loss: 0.62743738, Log Avg loss: 0.54237986, Global Avg Loss: 1.77673293, Time: 0.0149 Steps: 50760, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001803, Sample Num: 28848, Cur Loss: 0.40461913, Cur Avg Loss: 0.62837507, Log Avg loss: 0.79650383, Global Avg Loss: 1.77653985, Time: 0.0113 Steps: 50770, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001813, Sample Num: 29008, Cur Loss: 0.37099484, Cur Avg Loss: 0.62840566, Log Avg loss: 0.63392169, Global Avg Loss: 1.77631484, Time: 0.0086 Steps: 50780, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001823, Sample Num: 29168, Cur Loss: 0.53277177, Cur Avg Loss: 0.62833972, Log Avg loss: 0.61638481, Global Avg Loss: 1.77608646, Time: 0.0072 Steps: 50790, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001833, Sample Num: 29328, Cur Loss: 0.26009095, Cur Avg Loss: 0.62776400, Log Avg loss: 0.52280973, Global Avg Loss: 1.77583976, Time: 0.0125 Steps: 50800, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001843, Sample Num: 29488, Cur Loss: 2.18310118, Cur Avg Loss: 0.62835019, Log Avg loss: 0.73579909, Global Avg Loss: 1.77563506, Time: 0.0128 Steps: 50810, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001853, Sample Num: 29648, Cur Loss: 0.37390086, Cur Avg Loss: 0.62760774, Log Avg loss: 0.49077440, Global Avg Loss: 1.77538224, Time: 0.0113 Steps: 50820, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001863, Sample Num: 29808, Cur Loss: 0.54098481, Cur Avg Loss: 0.62749546, Log Avg loss: 0.60668947, Global Avg Loss: 1.77515232, Time: 0.0127 Steps: 50830, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001873, Sample Num: 29968, Cur Loss: 0.30103207, Cur Avg Loss: 0.62829585, Log Avg loss: 0.77740810, Global Avg Loss: 1.77495606, Time: 0.0067 Steps: 50840, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001883, Sample Num: 30128, Cur Loss: 0.68785334, Cur Avg Loss: 0.62800909, Log Avg loss: 0.57429939, Global Avg Loss: 1.77471995, Time: 0.0131 Steps: 50850, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001893, Sample Num: 30288, Cur Loss: 0.37530962, Cur Avg Loss: 0.62825249, Log Avg loss: 0.67408355, Global Avg Loss: 1.77450354, Time: 0.0076 Steps: 50860, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001903, Sample Num: 30448, Cur Loss: 0.37366292, Cur Avg Loss: 0.62848720, Log Avg loss: 0.67291817, Global Avg Loss: 1.77428699, Time: 0.0153 Steps: 50870, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001913, Sample Num: 30608, Cur Loss: 0.31907833, Cur Avg Loss: 0.62819157, Log Avg loss: 0.57193417, Global Avg Loss: 1.77405068, Time: 0.0067 Steps: 50880, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001923, Sample Num: 30768, Cur Loss: 0.29182178, Cur Avg Loss: 0.62788942, Log Avg loss: 0.57008705, Global Avg Loss: 1.77381410, Time: 0.0072 Steps: 50890, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001933, Sample Num: 30928, Cur Loss: 0.32323146, Cur Avg Loss: 0.62728731, Log Avg loss: 0.51150180, Global Avg Loss: 1.77356610, Time: 0.0168 Steps: 50900, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001943, Sample Num: 31088, Cur Loss: 0.43474132, Cur Avg Loss: 0.62776845, Log Avg loss: 0.72077356, Global Avg Loss: 1.77335931, Time: 0.0120 Steps: 50910, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001953, Sample Num: 31248, Cur Loss: 0.46017587, Cur Avg Loss: 0.62794612, Log Avg loss: 0.66246737, Global Avg Loss: 1.77314114, Time: 0.0092 Steps: 50920, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001963, Sample Num: 31408, Cur Loss: 1.15237010, Cur Avg Loss: 0.62829549, Log Avg loss: 0.69652693, Global Avg Loss: 1.77292975, Time: 0.0089 Steps: 50930, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001973, Sample Num: 31568, Cur Loss: 0.15144461, Cur Avg Loss: 0.62734793, Log Avg loss: 0.44134296, Global Avg Loss: 1.77266835, Time: 0.0144 Steps: 50940, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001983, Sample Num: 31728, Cur Loss: 0.72195637, Cur Avg Loss: 0.62687178, Log Avg loss: 0.53292556, Global Avg Loss: 1.77242502, Time: 0.0112 Steps: 50950, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 001993, Sample Num: 31888, Cur Loss: 0.40132576, Cur Avg Loss: 0.62611612, Log Avg loss: 0.47626950, Global Avg Loss: 1.77217067, Time: 0.0127 Steps: 50960, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002003, Sample Num: 32048, Cur Loss: 0.22372150, Cur Avg Loss: 0.62604847, Log Avg loss: 0.61256496, Global Avg Loss: 1.77194317, Time: 0.0160 Steps: 50970, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002013, Sample Num: 32208, Cur Loss: 0.51851237, Cur Avg Loss: 0.62582829, Log Avg loss: 0.58172768, Global Avg Loss: 1.77170970, Time: 0.0234 Steps: 50980, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002023, Sample Num: 32368, Cur Loss: 0.34566337, Cur Avg Loss: 0.62503268, Log Avg loss: 0.46487602, Global Avg Loss: 1.77145341, Time: 0.0138 Steps: 50990, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002033, Sample Num: 32528, Cur Loss: 0.64714223, Cur Avg Loss: 0.62563565, Log Avg loss: 0.74761618, Global Avg Loss: 1.77125266, Time: 0.0084 Steps: 51000, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002043, Sample Num: 32688, Cur Loss: 0.27614170, Cur Avg Loss: 0.62553861, Log Avg loss: 0.60580999, Global Avg Loss: 1.77102418, Time: 0.0091 Steps: 51010, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002053, Sample Num: 32848, Cur Loss: 0.71680403, Cur Avg Loss: 0.62582015, Log Avg loss: 0.68333909, Global Avg Loss: 1.77081099, Time: 0.0200 Steps: 51020, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002063, Sample Num: 33008, Cur Loss: 0.39203754, Cur Avg Loss: 0.62569609, Log Avg loss: 0.60022768, Global Avg Loss: 1.77058160, Time: 0.0193 Steps: 51030, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002073, Sample Num: 33168, Cur Loss: 0.47304237, Cur Avg Loss: 0.62562007, Log Avg loss: 0.60993621, Global Avg Loss: 1.77035420, Time: 0.0067 Steps: 51040, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002083, Sample Num: 33328, Cur Loss: 0.27438423, Cur Avg Loss: 0.62530922, Log Avg loss: 0.56087092, Global Avg Loss: 1.77011728, Time: 0.0201 Steps: 51050, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002093, Sample Num: 33488, Cur Loss: 0.19625872, Cur Avg Loss: 0.62441739, Log Avg loss: 0.43864857, Global Avg Loss: 1.76985652, Time: 0.0066 Steps: 51060, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002103, Sample Num: 33648, Cur Loss: 0.72308576, Cur Avg Loss: 0.62497984, Log Avg loss: 0.74269993, Global Avg Loss: 1.76965539, Time: 0.0084 Steps: 51070, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002113, Sample Num: 33808, Cur Loss: 0.76406121, Cur Avg Loss: 0.62443075, Log Avg loss: 0.50895717, Global Avg Loss: 1.76940858, Time: 0.0115 Steps: 51080, Updated lr: 0.000053 Training, Epoch: 0024, Batch: 002123, Sample Num: 33968, Cur Loss: 0.26680395, Cur Avg Loss: 0.62416740, Log Avg loss: 0.56852208, Global Avg Loss: 1.76917353, Time: 0.0107 Steps: 51090, Updated lr: 0.000052 ***** Running evaluation checkpoint-51096 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-51096 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.266252, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.78954, "eval_total_loss": 555.046616, "eval_mae": 0.731147, "eval_mse": 0.789625, "eval_r2": 0.498063, "eval_sp_statistic": 0.70089, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.751711, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.090118, "test_total_loss": 547.239187, "test_mae": 0.886414, "test_mse": 1.090157, "test_r2": 0.296403, "test_sp_statistic": 0.52532, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.626686, "test_ps_pvalue": 0.0, "lr": 5.249312470365102e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.769033649244862, "train_cur_epoch_loss": 1328.5751952454448, "train_cur_epoch_avg_loss": 0.6240371983304109, "train_cur_epoch_time": 22.266252279281616, "train_cur_epoch_avg_time": 0.010458549684960834, "epoch": 24, "step": 51096} ################################################## Training, Epoch: 0025, Batch: 000004, Sample Num: 64, Cur Loss: 0.77792597, Cur Avg Loss: 0.51771293, Log Avg loss: 0.55386558, Global Avg Loss: 1.76893570, Time: 0.0069 Steps: 51100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000014, Sample Num: 224, Cur Loss: 0.76593572, Cur Avg Loss: 0.54730691, Log Avg loss: 0.55914450, Global Avg Loss: 1.76869900, Time: 0.0070 Steps: 51110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000024, Sample Num: 384, Cur Loss: 0.41897130, Cur Avg Loss: 0.55741314, Log Avg loss: 0.57156188, Global Avg Loss: 1.76846481, Time: 0.0108 Steps: 51120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000034, Sample Num: 544, Cur Loss: 0.76801032, Cur Avg Loss: 0.59401075, Log Avg loss: 0.68184501, Global Avg Loss: 1.76825229, Time: 0.0191 Steps: 51130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000044, Sample Num: 704, Cur Loss: 2.03730893, Cur Avg Loss: 0.61611675, Log Avg loss: 0.69127715, Global Avg Loss: 1.76804170, Time: 0.0153 Steps: 51140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000054, Sample Num: 864, Cur Loss: 0.60196102, Cur Avg Loss: 0.63962589, Log Avg loss: 0.74306608, Global Avg Loss: 1.76784131, Time: 0.0112 Steps: 51150, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000064, Sample Num: 1024, Cur Loss: 0.56515741, Cur Avg Loss: 0.64339300, Log Avg loss: 0.66373545, Global Avg Loss: 1.76762550, Time: 0.0126 Steps: 51160, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000074, Sample Num: 1184, Cur Loss: 1.11717677, Cur Avg Loss: 0.64431511, Log Avg loss: 0.65021661, Global Avg Loss: 1.76740713, Time: 0.0067 Steps: 51170, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000084, Sample Num: 1344, Cur Loss: 0.52342826, Cur Avg Loss: 0.63323288, Log Avg loss: 0.55122436, Global Avg Loss: 1.76716950, Time: 0.0067 Steps: 51180, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000094, Sample Num: 1504, Cur Loss: 1.04673088, Cur Avg Loss: 0.63152551, Log Avg loss: 0.61718363, Global Avg Loss: 1.76694485, Time: 0.0126 Steps: 51190, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000104, Sample Num: 1664, Cur Loss: 0.83659816, Cur Avg Loss: 0.61840418, Log Avg loss: 0.49506366, Global Avg Loss: 1.76669643, Time: 0.0068 Steps: 51200, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000114, Sample Num: 1824, Cur Loss: 0.71610224, Cur Avg Loss: 0.61828959, Log Avg loss: 0.61709788, Global Avg Loss: 1.76647195, Time: 0.0123 Steps: 51210, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000124, Sample Num: 1984, Cur Loss: 0.20974869, Cur Avg Loss: 0.61003020, Log Avg loss: 0.51587315, Global Avg Loss: 1.76622778, Time: 0.0117 Steps: 51220, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000134, Sample Num: 2144, Cur Loss: 0.65093607, Cur Avg Loss: 0.60468828, Log Avg loss: 0.53844842, Global Avg Loss: 1.76598812, Time: 0.0113 Steps: 51230, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000144, Sample Num: 2304, Cur Loss: 1.08179843, Cur Avg Loss: 0.60764127, Log Avg loss: 0.64721133, Global Avg Loss: 1.76576978, Time: 0.0068 Steps: 51240, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000154, Sample Num: 2464, Cur Loss: 0.76903117, Cur Avg Loss: 0.59890473, Log Avg loss: 0.47309854, Global Avg Loss: 1.76551755, Time: 0.0069 Steps: 51250, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000164, Sample Num: 2624, Cur Loss: 0.76622504, Cur Avg Loss: 0.60431579, Log Avg loss: 0.68764619, Global Avg Loss: 1.76530728, Time: 0.0068 Steps: 51260, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000174, Sample Num: 2784, Cur Loss: 0.78496528, Cur Avg Loss: 0.60126445, Log Avg loss: 0.55122246, Global Avg Loss: 1.76507048, Time: 0.0067 Steps: 51270, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000184, Sample Num: 2944, Cur Loss: 0.55494547, Cur Avg Loss: 0.60689131, Log Avg loss: 0.70479859, Global Avg Loss: 1.76486372, Time: 0.0077 Steps: 51280, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000194, Sample Num: 3104, Cur Loss: 0.26500526, Cur Avg Loss: 0.59322770, Log Avg loss: 0.34181723, Global Avg Loss: 1.76458626, Time: 0.0110 Steps: 51290, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000204, Sample Num: 3264, Cur Loss: 0.66356164, Cur Avg Loss: 0.59935439, Log Avg loss: 0.71821221, Global Avg Loss: 1.76438229, Time: 0.0067 Steps: 51300, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000214, Sample Num: 3424, Cur Loss: 0.36318666, Cur Avg Loss: 0.59229746, Log Avg loss: 0.44833608, Global Avg Loss: 1.76412580, Time: 0.0132 Steps: 51310, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000224, Sample Num: 3584, Cur Loss: 0.35664162, Cur Avg Loss: 0.59446617, Log Avg loss: 0.64087664, Global Avg Loss: 1.76390693, Time: 0.0076 Steps: 51320, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000234, Sample Num: 3744, Cur Loss: 0.52603984, Cur Avg Loss: 0.59823532, Log Avg loss: 0.68266419, Global Avg Loss: 1.76369629, Time: 0.0067 Steps: 51330, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000244, Sample Num: 3904, Cur Loss: 0.36620027, Cur Avg Loss: 0.59086580, Log Avg loss: 0.41841919, Global Avg Loss: 1.76343425, Time: 0.0064 Steps: 51340, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000254, Sample Num: 4064, Cur Loss: 0.47886491, Cur Avg Loss: 0.59306689, Log Avg loss: 0.64677329, Global Avg Loss: 1.76321679, Time: 0.0164 Steps: 51350, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000264, Sample Num: 4224, Cur Loss: 0.45476803, Cur Avg Loss: 0.59646815, Log Avg loss: 0.68286016, Global Avg Loss: 1.76300644, Time: 0.0069 Steps: 51360, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000274, Sample Num: 4384, Cur Loss: 0.21354476, Cur Avg Loss: 0.59902427, Log Avg loss: 0.66650604, Global Avg Loss: 1.76279299, Time: 0.0063 Steps: 51370, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000284, Sample Num: 4544, Cur Loss: 0.38694537, Cur Avg Loss: 0.59839825, Log Avg loss: 0.58124528, Global Avg Loss: 1.76256303, Time: 0.0176 Steps: 51380, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000294, Sample Num: 4704, Cur Loss: 0.36757997, Cur Avg Loss: 0.59908793, Log Avg loss: 0.61867476, Global Avg Loss: 1.76234044, Time: 0.0069 Steps: 51390, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000304, Sample Num: 4864, Cur Loss: 0.65869880, Cur Avg Loss: 0.59635627, Log Avg loss: 0.51604538, Global Avg Loss: 1.76209797, Time: 0.0066 Steps: 51400, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000314, Sample Num: 5024, Cur Loss: 0.99445444, Cur Avg Loss: 0.59691235, Log Avg loss: 0.61381719, Global Avg Loss: 1.76187461, Time: 0.0064 Steps: 51410, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000324, Sample Num: 5184, Cur Loss: 0.36372834, Cur Avg Loss: 0.59849255, Log Avg loss: 0.64811077, Global Avg Loss: 1.76165801, Time: 0.0069 Steps: 51420, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000334, Sample Num: 5344, Cur Loss: 0.78633463, Cur Avg Loss: 0.59797079, Log Avg loss: 0.58106591, Global Avg Loss: 1.76142846, Time: 0.0069 Steps: 51430, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000344, Sample Num: 5504, Cur Loss: 0.49265581, Cur Avg Loss: 0.59502340, Log Avg loss: 0.49658051, Global Avg Loss: 1.76118257, Time: 0.0065 Steps: 51440, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000354, Sample Num: 5664, Cur Loss: 0.51877487, Cur Avg Loss: 0.59856729, Log Avg loss: 0.72047734, Global Avg Loss: 1.76098029, Time: 0.0088 Steps: 51450, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000364, Sample Num: 5824, Cur Loss: 0.47859457, Cur Avg Loss: 0.59990030, Log Avg loss: 0.64708860, Global Avg Loss: 1.76076384, Time: 0.0123 Steps: 51460, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000374, Sample Num: 5984, Cur Loss: 2.08036971, Cur Avg Loss: 0.61183230, Log Avg loss: 1.04615708, Global Avg Loss: 1.76062500, Time: 0.0114 Steps: 51470, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000384, Sample Num: 6144, Cur Loss: 0.80551505, Cur Avg Loss: 0.61768554, Log Avg loss: 0.83659700, Global Avg Loss: 1.76044550, Time: 0.0118 Steps: 51480, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000394, Sample Num: 6304, Cur Loss: 0.32852837, Cur Avg Loss: 0.62077287, Log Avg loss: 0.73932605, Global Avg Loss: 1.76024719, Time: 0.0067 Steps: 51490, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000404, Sample Num: 6464, Cur Loss: 0.62926698, Cur Avg Loss: 0.61986985, Log Avg loss: 0.58429108, Global Avg Loss: 1.76001885, Time: 0.0227 Steps: 51500, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000414, Sample Num: 6624, Cur Loss: 0.30132622, Cur Avg Loss: 0.62162795, Log Avg loss: 0.69265527, Global Avg Loss: 1.75981163, Time: 0.0113 Steps: 51510, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000424, Sample Num: 6784, Cur Loss: 0.54525542, Cur Avg Loss: 0.61878273, Log Avg loss: 0.50099048, Global Avg Loss: 1.75956730, Time: 0.0070 Steps: 51520, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000434, Sample Num: 6944, Cur Loss: 0.78909594, Cur Avg Loss: 0.61867298, Log Avg loss: 0.61401976, Global Avg Loss: 1.75934499, Time: 0.0067 Steps: 51530, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000444, Sample Num: 7104, Cur Loss: 0.54624587, Cur Avg Loss: 0.61772051, Log Avg loss: 0.57638314, Global Avg Loss: 1.75911547, Time: 0.0105 Steps: 51540, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000454, Sample Num: 7264, Cur Loss: 0.25392109, Cur Avg Loss: 0.61353850, Log Avg loss: 0.42785735, Global Avg Loss: 1.75885722, Time: 0.0116 Steps: 51550, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000464, Sample Num: 7424, Cur Loss: 0.48899880, Cur Avg Loss: 0.61577655, Log Avg loss: 0.71738367, Global Avg Loss: 1.75865523, Time: 0.0117 Steps: 51560, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000474, Sample Num: 7584, Cur Loss: 0.36252642, Cur Avg Loss: 0.61465314, Log Avg loss: 0.56252698, Global Avg Loss: 1.75842329, Time: 0.0115 Steps: 51570, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000484, Sample Num: 7744, Cur Loss: 0.35350728, Cur Avg Loss: 0.61017265, Log Avg loss: 0.39779751, Global Avg Loss: 1.75815950, Time: 0.0121 Steps: 51580, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000494, Sample Num: 7904, Cur Loss: 0.23982400, Cur Avg Loss: 0.60837983, Log Avg loss: 0.52160736, Global Avg Loss: 1.75791981, Time: 0.0119 Steps: 51590, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000504, Sample Num: 8064, Cur Loss: 0.27956635, Cur Avg Loss: 0.60975358, Log Avg loss: 0.67761713, Global Avg Loss: 1.75771045, Time: 0.0069 Steps: 51600, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000514, Sample Num: 8224, Cur Loss: 0.33800972, Cur Avg Loss: 0.60833790, Log Avg loss: 0.53698718, Global Avg Loss: 1.75747392, Time: 0.0076 Steps: 51610, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000524, Sample Num: 8384, Cur Loss: 0.18447384, Cur Avg Loss: 0.60502417, Log Avg loss: 0.43469884, Global Avg Loss: 1.75721767, Time: 0.0108 Steps: 51620, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000534, Sample Num: 8544, Cur Loss: 0.95179236, Cur Avg Loss: 0.60585172, Log Avg loss: 0.64921509, Global Avg Loss: 1.75700306, Time: 0.0074 Steps: 51630, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000544, Sample Num: 8704, Cur Loss: 0.26566768, Cur Avg Loss: 0.60541962, Log Avg loss: 0.58234572, Global Avg Loss: 1.75677559, Time: 0.0225 Steps: 51640, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000554, Sample Num: 8864, Cur Loss: 0.42639509, Cur Avg Loss: 0.60435398, Log Avg loss: 0.54638315, Global Avg Loss: 1.75654125, Time: 0.0123 Steps: 51650, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000564, Sample Num: 9024, Cur Loss: 0.50253117, Cur Avg Loss: 0.60454077, Log Avg loss: 0.61488854, Global Avg Loss: 1.75632025, Time: 0.0097 Steps: 51660, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000574, Sample Num: 9184, Cur Loss: 0.42702007, Cur Avg Loss: 0.60379550, Log Avg loss: 0.56176263, Global Avg Loss: 1.75608906, Time: 0.0110 Steps: 51670, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000584, Sample Num: 9344, Cur Loss: 0.72531724, Cur Avg Loss: 0.60670575, Log Avg loss: 0.77375384, Global Avg Loss: 1.75589898, Time: 0.0121 Steps: 51680, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000594, Sample Num: 9504, Cur Loss: 0.52755278, Cur Avg Loss: 0.60600483, Log Avg loss: 0.56507096, Global Avg Loss: 1.75566861, Time: 0.0121 Steps: 51690, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000604, Sample Num: 9664, Cur Loss: 0.32478613, Cur Avg Loss: 0.60516133, Log Avg loss: 0.55505751, Global Avg Loss: 1.75543638, Time: 0.0087 Steps: 51700, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000614, Sample Num: 9824, Cur Loss: 0.41920584, Cur Avg Loss: 0.60492672, Log Avg loss: 0.59075661, Global Avg Loss: 1.75521115, Time: 0.0066 Steps: 51710, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000624, Sample Num: 9984, Cur Loss: 0.77545655, Cur Avg Loss: 0.60555530, Log Avg loss: 0.64415016, Global Avg Loss: 1.75499632, Time: 0.0116 Steps: 51720, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000634, Sample Num: 10144, Cur Loss: 0.58160019, Cur Avg Loss: 0.60470062, Log Avg loss: 0.55136815, Global Avg Loss: 1.75476365, Time: 0.0113 Steps: 51730, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000644, Sample Num: 10304, Cur Loss: 0.37690675, Cur Avg Loss: 0.60278862, Log Avg loss: 0.48156810, Global Avg Loss: 1.75451757, Time: 0.0144 Steps: 51740, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000654, Sample Num: 10464, Cur Loss: 0.66036332, Cur Avg Loss: 0.60505314, Log Avg loss: 0.75088809, Global Avg Loss: 1.75432363, Time: 0.0067 Steps: 51750, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000664, Sample Num: 10624, Cur Loss: 0.72217768, Cur Avg Loss: 0.60489718, Log Avg loss: 0.59469777, Global Avg Loss: 1.75409960, Time: 0.0117 Steps: 51760, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000674, Sample Num: 10784, Cur Loss: 0.96594572, Cur Avg Loss: 0.60527913, Log Avg loss: 0.63064018, Global Avg Loss: 1.75388259, Time: 0.0119 Steps: 51770, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000684, Sample Num: 10944, Cur Loss: 0.36693200, Cur Avg Loss: 0.60456349, Log Avg loss: 0.55632981, Global Avg Loss: 1.75365131, Time: 0.0108 Steps: 51780, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000694, Sample Num: 11104, Cur Loss: 0.14643332, Cur Avg Loss: 0.60585308, Log Avg loss: 0.69406058, Global Avg Loss: 1.75344672, Time: 0.0107 Steps: 51790, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000704, Sample Num: 11264, Cur Loss: 1.07200885, Cur Avg Loss: 0.60561394, Log Avg loss: 0.58901811, Global Avg Loss: 1.75322192, Time: 0.0116 Steps: 51800, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000714, Sample Num: 11424, Cur Loss: 0.28457069, Cur Avg Loss: 0.60594773, Log Avg loss: 0.62944637, Global Avg Loss: 1.75300502, Time: 0.0138 Steps: 51810, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000724, Sample Num: 11584, Cur Loss: 0.68111962, Cur Avg Loss: 0.60565029, Log Avg loss: 0.58441305, Global Avg Loss: 1.75277951, Time: 0.0067 Steps: 51820, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000734, Sample Num: 11744, Cur Loss: 1.07364190, Cur Avg Loss: 0.60946124, Log Avg loss: 0.88537408, Global Avg Loss: 1.75261215, Time: 0.0071 Steps: 51830, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000744, Sample Num: 11904, Cur Loss: 0.72001374, Cur Avg Loss: 0.60921738, Log Avg loss: 0.59131790, Global Avg Loss: 1.75238814, Time: 0.0096 Steps: 51840, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000754, Sample Num: 12064, Cur Loss: 0.23843271, Cur Avg Loss: 0.60743828, Log Avg loss: 0.47507308, Global Avg Loss: 1.75214179, Time: 0.0121 Steps: 51850, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000764, Sample Num: 12224, Cur Loss: 0.36046728, Cur Avg Loss: 0.60908751, Log Avg loss: 0.73343927, Global Avg Loss: 1.75194536, Time: 0.0071 Steps: 51860, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000774, Sample Num: 12384, Cur Loss: 1.15585232, Cur Avg Loss: 0.61021715, Log Avg loss: 0.69652209, Global Avg Loss: 1.75174188, Time: 0.0234 Steps: 51870, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000784, Sample Num: 12544, Cur Loss: 0.34934223, Cur Avg Loss: 0.60923539, Log Avg loss: 0.53324685, Global Avg Loss: 1.75150701, Time: 0.0065 Steps: 51880, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000794, Sample Num: 12704, Cur Loss: 0.78540742, Cur Avg Loss: 0.61069822, Log Avg loss: 0.72538422, Global Avg Loss: 1.75130926, Time: 0.0070 Steps: 51890, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000804, Sample Num: 12864, Cur Loss: 0.48720959, Cur Avg Loss: 0.61021464, Log Avg loss: 0.57181851, Global Avg Loss: 1.75108200, Time: 0.0072 Steps: 51900, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000814, Sample Num: 13024, Cur Loss: 0.74002099, Cur Avg Loss: 0.60920728, Log Avg loss: 0.52821558, Global Avg Loss: 1.75084643, Time: 0.0069 Steps: 51910, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000824, Sample Num: 13184, Cur Loss: 0.21559355, Cur Avg Loss: 0.60823362, Log Avg loss: 0.52897747, Global Avg Loss: 1.75061109, Time: 0.0068 Steps: 51920, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000834, Sample Num: 13344, Cur Loss: 0.36746848, Cur Avg Loss: 0.60849876, Log Avg loss: 0.63034618, Global Avg Loss: 1.75039537, Time: 0.0067 Steps: 51930, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000844, Sample Num: 13504, Cur Loss: 0.19149503, Cur Avg Loss: 0.60807603, Log Avg loss: 0.57282086, Global Avg Loss: 1.75016865, Time: 0.0068 Steps: 51940, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000854, Sample Num: 13664, Cur Loss: 0.43828386, Cur Avg Loss: 0.60664660, Log Avg loss: 0.48600248, Global Avg Loss: 1.74992530, Time: 0.0072 Steps: 51950, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000864, Sample Num: 13824, Cur Loss: 1.21857405, Cur Avg Loss: 0.61011239, Log Avg loss: 0.90609101, Global Avg Loss: 1.74976290, Time: 0.0201 Steps: 51960, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000874, Sample Num: 13984, Cur Loss: 0.91201830, Cur Avg Loss: 0.61073158, Log Avg loss: 0.66422918, Global Avg Loss: 1.74955403, Time: 0.0065 Steps: 51970, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000884, Sample Num: 14144, Cur Loss: 0.28516418, Cur Avg Loss: 0.61018165, Log Avg loss: 0.56211806, Global Avg Loss: 1.74932559, Time: 0.0073 Steps: 51980, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000894, Sample Num: 14304, Cur Loss: 0.57398534, Cur Avg Loss: 0.60925515, Log Avg loss: 0.52735280, Global Avg Loss: 1.74909055, Time: 0.0117 Steps: 51990, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000904, Sample Num: 14464, Cur Loss: 0.80410814, Cur Avg Loss: 0.60938026, Log Avg loss: 0.62056512, Global Avg Loss: 1.74887352, Time: 0.0069 Steps: 52000, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000914, Sample Num: 14624, Cur Loss: 0.54496634, Cur Avg Loss: 0.61107569, Log Avg loss: 0.76434195, Global Avg Loss: 1.74868422, Time: 0.0218 Steps: 52010, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000924, Sample Num: 14784, Cur Loss: 0.91857749, Cur Avg Loss: 0.60974262, Log Avg loss: 0.48790065, Global Avg Loss: 1.74844186, Time: 0.0065 Steps: 52020, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000934, Sample Num: 14944, Cur Loss: 0.25982291, Cur Avg Loss: 0.60945750, Log Avg loss: 0.58311167, Global Avg Loss: 1.74821789, Time: 0.0071 Steps: 52030, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000944, Sample Num: 15104, Cur Loss: 0.96527100, Cur Avg Loss: 0.61259444, Log Avg loss: 0.90558463, Global Avg Loss: 1.74805597, Time: 0.0153 Steps: 52040, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000954, Sample Num: 15264, Cur Loss: 0.24487926, Cur Avg Loss: 0.61281888, Log Avg loss: 0.63400614, Global Avg Loss: 1.74784193, Time: 0.0067 Steps: 52050, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000964, Sample Num: 15424, Cur Loss: 0.52072191, Cur Avg Loss: 0.61455117, Log Avg loss: 0.77981178, Global Avg Loss: 1.74765599, Time: 0.0224 Steps: 52060, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000974, Sample Num: 15584, Cur Loss: 0.65075111, Cur Avg Loss: 0.61532983, Log Avg loss: 0.69039283, Global Avg Loss: 1.74745294, Time: 0.0065 Steps: 52070, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000984, Sample Num: 15744, Cur Loss: 0.55298305, Cur Avg Loss: 0.61710914, Log Avg loss: 0.79041344, Global Avg Loss: 1.74726918, Time: 0.0075 Steps: 52080, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 000994, Sample Num: 15904, Cur Loss: 0.54727298, Cur Avg Loss: 0.61607487, Log Avg loss: 0.51430262, Global Avg Loss: 1.74703248, Time: 0.0067 Steps: 52090, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001004, Sample Num: 16064, Cur Loss: 0.52170503, Cur Avg Loss: 0.61521214, Log Avg loss: 0.52945756, Global Avg Loss: 1.74679878, Time: 0.0116 Steps: 52100, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001014, Sample Num: 16224, Cur Loss: 0.70553160, Cur Avg Loss: 0.61554871, Log Avg loss: 0.64934015, Global Avg Loss: 1.74658817, Time: 0.0066 Steps: 52110, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001024, Sample Num: 16384, Cur Loss: 0.33804762, Cur Avg Loss: 0.61532849, Log Avg loss: 0.59299825, Global Avg Loss: 1.74636684, Time: 0.0125 Steps: 52120, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001034, Sample Num: 16544, Cur Loss: 0.94808733, Cur Avg Loss: 0.61447288, Log Avg loss: 0.52685821, Global Avg Loss: 1.74613290, Time: 0.0089 Steps: 52130, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001044, Sample Num: 16704, Cur Loss: 0.27817023, Cur Avg Loss: 0.61344318, Log Avg loss: 0.50697267, Global Avg Loss: 1.74589524, Time: 0.0134 Steps: 52140, Updated lr: 0.000052 Training, Epoch: 0025, Batch: 001054, Sample Num: 16864, Cur Loss: 0.44769537, Cur Avg Loss: 0.61386789, Log Avg loss: 0.65820728, Global Avg Loss: 1.74568667, Time: 0.0113 Steps: 52150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001064, Sample Num: 17024, Cur Loss: 0.45705166, Cur Avg Loss: 0.61454142, Log Avg loss: 0.68553120, Global Avg Loss: 1.74548342, Time: 0.0115 Steps: 52160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001074, Sample Num: 17184, Cur Loss: 0.21043891, Cur Avg Loss: 0.61423534, Log Avg loss: 0.58166863, Global Avg Loss: 1.74526034, Time: 0.0071 Steps: 52170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001084, Sample Num: 17344, Cur Loss: 0.60139734, Cur Avg Loss: 0.61391484, Log Avg loss: 0.57949352, Global Avg Loss: 1.74503693, Time: 0.0125 Steps: 52180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001094, Sample Num: 17504, Cur Loss: 0.83601606, Cur Avg Loss: 0.61380390, Log Avg loss: 0.60177716, Global Avg Loss: 1.74481787, Time: 0.0118 Steps: 52190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001104, Sample Num: 17664, Cur Loss: 0.50609523, Cur Avg Loss: 0.61309441, Log Avg loss: 0.53547648, Global Avg Loss: 1.74458620, Time: 0.0108 Steps: 52200, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001114, Sample Num: 17824, Cur Loss: 1.43575799, Cur Avg Loss: 0.61444069, Log Avg loss: 0.76306967, Global Avg Loss: 1.74439820, Time: 0.0107 Steps: 52210, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001124, Sample Num: 17984, Cur Loss: 0.15202232, Cur Avg Loss: 0.61319532, Log Avg loss: 0.47446155, Global Avg Loss: 1.74415501, Time: 0.0118 Steps: 52220, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001134, Sample Num: 18144, Cur Loss: 0.47181445, Cur Avg Loss: 0.61329675, Log Avg loss: 0.62469757, Global Avg Loss: 1.74394068, Time: 0.0208 Steps: 52230, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001144, Sample Num: 18304, Cur Loss: 0.44322777, Cur Avg Loss: 0.61248514, Log Avg loss: 0.52044883, Global Avg Loss: 1.74370648, Time: 0.0090 Steps: 52240, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001154, Sample Num: 18464, Cur Loss: 1.29311574, Cur Avg Loss: 0.61235940, Log Avg loss: 0.59797416, Global Avg Loss: 1.74348720, Time: 0.0066 Steps: 52250, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001164, Sample Num: 18624, Cur Loss: 0.31913847, Cur Avg Loss: 0.61224609, Log Avg loss: 0.59917019, Global Avg Loss: 1.74326823, Time: 0.0186 Steps: 52260, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001174, Sample Num: 18784, Cur Loss: 1.06444812, Cur Avg Loss: 0.61186302, Log Avg loss: 0.56727399, Global Avg Loss: 1.74304325, Time: 0.0067 Steps: 52270, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001184, Sample Num: 18944, Cur Loss: 0.35775474, Cur Avg Loss: 0.61230263, Log Avg loss: 0.66391236, Global Avg Loss: 1.74283683, Time: 0.0150 Steps: 52280, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001194, Sample Num: 19104, Cur Loss: 1.19452870, Cur Avg Loss: 0.61343124, Log Avg loss: 0.74705933, Global Avg Loss: 1.74264640, Time: 0.0109 Steps: 52290, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001204, Sample Num: 19264, Cur Loss: 0.51589286, Cur Avg Loss: 0.61205091, Log Avg loss: 0.44723976, Global Avg Loss: 1.74239871, Time: 0.0118 Steps: 52300, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001214, Sample Num: 19424, Cur Loss: 0.52666175, Cur Avg Loss: 0.61243174, Log Avg loss: 0.65828317, Global Avg Loss: 1.74219146, Time: 0.0115 Steps: 52310, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001224, Sample Num: 19584, Cur Loss: 0.39771277, Cur Avg Loss: 0.61349165, Log Avg loss: 0.74216461, Global Avg Loss: 1.74200033, Time: 0.0068 Steps: 52320, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001234, Sample Num: 19744, Cur Loss: 0.19122830, Cur Avg Loss: 0.61285683, Log Avg loss: 0.53515444, Global Avg Loss: 1.74176971, Time: 0.0172 Steps: 52330, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001244, Sample Num: 19904, Cur Loss: 0.45505002, Cur Avg Loss: 0.61376218, Log Avg loss: 0.72548258, Global Avg Loss: 1.74157553, Time: 0.0066 Steps: 52340, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001254, Sample Num: 20064, Cur Loss: 0.25843742, Cur Avg Loss: 0.61415101, Log Avg loss: 0.66252130, Global Avg Loss: 1.74136941, Time: 0.0066 Steps: 52350, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001264, Sample Num: 20224, Cur Loss: 0.64308918, Cur Avg Loss: 0.61403051, Log Avg loss: 0.59892071, Global Avg Loss: 1.74115122, Time: 0.0069 Steps: 52360, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001274, Sample Num: 20384, Cur Loss: 0.61861706, Cur Avg Loss: 0.61397402, Log Avg loss: 0.60683322, Global Avg Loss: 1.74093462, Time: 0.0067 Steps: 52370, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001284, Sample Num: 20544, Cur Loss: 0.42630908, Cur Avg Loss: 0.61386333, Log Avg loss: 0.59976154, Global Avg Loss: 1.74071676, Time: 0.0118 Steps: 52380, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001294, Sample Num: 20704, Cur Loss: 1.02688789, Cur Avg Loss: 0.61373918, Log Avg loss: 0.59779806, Global Avg Loss: 1.74049860, Time: 0.0109 Steps: 52390, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001304, Sample Num: 20864, Cur Loss: 0.46365404, Cur Avg Loss: 0.61377727, Log Avg loss: 0.61870584, Global Avg Loss: 1.74028452, Time: 0.0067 Steps: 52400, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001314, Sample Num: 21024, Cur Loss: 1.10288298, Cur Avg Loss: 0.61451391, Log Avg loss: 0.71057160, Global Avg Loss: 1.74008805, Time: 0.0109 Steps: 52410, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001324, Sample Num: 21184, Cur Loss: 0.25613496, Cur Avg Loss: 0.61407810, Log Avg loss: 0.55681299, Global Avg Loss: 1.73986232, Time: 0.0113 Steps: 52420, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001334, Sample Num: 21344, Cur Loss: 0.18563314, Cur Avg Loss: 0.61408158, Log Avg loss: 0.61454327, Global Avg Loss: 1.73964769, Time: 0.0118 Steps: 52430, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001344, Sample Num: 21504, Cur Loss: 0.72795141, Cur Avg Loss: 0.61307418, Log Avg loss: 0.47868656, Global Avg Loss: 1.73940723, Time: 0.0097 Steps: 52440, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001354, Sample Num: 21664, Cur Loss: 0.80203009, Cur Avg Loss: 0.61422111, Log Avg loss: 0.76836826, Global Avg Loss: 1.73922209, Time: 0.0114 Steps: 52450, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001364, Sample Num: 21824, Cur Loss: 0.29341772, Cur Avg Loss: 0.61459642, Log Avg loss: 0.66541340, Global Avg Loss: 1.73901740, Time: 0.0155 Steps: 52460, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001374, Sample Num: 21984, Cur Loss: 0.32783943, Cur Avg Loss: 0.61443983, Log Avg loss: 0.59308043, Global Avg Loss: 1.73879900, Time: 0.0114 Steps: 52470, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001384, Sample Num: 22144, Cur Loss: 0.68039155, Cur Avg Loss: 0.61418076, Log Avg loss: 0.57858540, Global Avg Loss: 1.73857793, Time: 0.0065 Steps: 52480, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001394, Sample Num: 22304, Cur Loss: 0.76162899, Cur Avg Loss: 0.61339243, Log Avg loss: 0.50428745, Global Avg Loss: 1.73834278, Time: 0.0069 Steps: 52490, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001404, Sample Num: 22464, Cur Loss: 0.55903536, Cur Avg Loss: 0.61281590, Log Avg loss: 0.53244693, Global Avg Loss: 1.73811308, Time: 0.0215 Steps: 52500, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001414, Sample Num: 22624, Cur Loss: 0.88179791, Cur Avg Loss: 0.61439358, Log Avg loss: 0.83589989, Global Avg Loss: 1.73794127, Time: 0.0172 Steps: 52510, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001424, Sample Num: 22784, Cur Loss: 1.01286662, Cur Avg Loss: 0.61451646, Log Avg loss: 0.63189282, Global Avg Loss: 1.73773067, Time: 0.0206 Steps: 52520, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001434, Sample Num: 22944, Cur Loss: 0.24537432, Cur Avg Loss: 0.61408820, Log Avg loss: 0.55310308, Global Avg Loss: 1.73750516, Time: 0.0114 Steps: 52530, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001444, Sample Num: 23104, Cur Loss: 0.24009441, Cur Avg Loss: 0.61412607, Log Avg loss: 0.61955616, Global Avg Loss: 1.73729237, Time: 0.0065 Steps: 52540, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001454, Sample Num: 23264, Cur Loss: 0.81106377, Cur Avg Loss: 0.61431447, Log Avg loss: 0.64152028, Global Avg Loss: 1.73708386, Time: 0.0227 Steps: 52550, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001464, Sample Num: 23424, Cur Loss: 1.68033540, Cur Avg Loss: 0.61417929, Log Avg loss: 0.59452392, Global Avg Loss: 1.73686647, Time: 0.0075 Steps: 52560, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001474, Sample Num: 23584, Cur Loss: 0.25821036, Cur Avg Loss: 0.61376692, Log Avg loss: 0.55339596, Global Avg Loss: 1.73664135, Time: 0.0067 Steps: 52570, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001484, Sample Num: 23744, Cur Loss: 0.42888618, Cur Avg Loss: 0.61411316, Log Avg loss: 0.66514932, Global Avg Loss: 1.73643757, Time: 0.0204 Steps: 52580, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001494, Sample Num: 23904, Cur Loss: 1.15301442, Cur Avg Loss: 0.61398099, Log Avg loss: 0.59436675, Global Avg Loss: 1.73622040, Time: 0.0218 Steps: 52590, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001504, Sample Num: 24064, Cur Loss: 0.81084037, Cur Avg Loss: 0.61415196, Log Avg loss: 0.63969497, Global Avg Loss: 1.73601194, Time: 0.0068 Steps: 52600, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001514, Sample Num: 24224, Cur Loss: 0.43344912, Cur Avg Loss: 0.61422599, Log Avg loss: 0.62536038, Global Avg Loss: 1.73580083, Time: 0.0066 Steps: 52610, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001524, Sample Num: 24384, Cur Loss: 0.90579098, Cur Avg Loss: 0.61408732, Log Avg loss: 0.59309222, Global Avg Loss: 1.73558366, Time: 0.0164 Steps: 52620, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001534, Sample Num: 24544, Cur Loss: 0.73748058, Cur Avg Loss: 0.61303632, Log Avg loss: 0.45286386, Global Avg Loss: 1.73533994, Time: 0.0199 Steps: 52630, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001544, Sample Num: 24704, Cur Loss: 0.28090549, Cur Avg Loss: 0.61286003, Log Avg loss: 0.58581795, Global Avg Loss: 1.73512157, Time: 0.0124 Steps: 52640, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001554, Sample Num: 24864, Cur Loss: 0.24097472, Cur Avg Loss: 0.61234073, Log Avg loss: 0.53215956, Global Avg Loss: 1.73489308, Time: 0.0126 Steps: 52650, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001564, Sample Num: 25024, Cur Loss: 0.97547084, Cur Avg Loss: 0.61231497, Log Avg loss: 0.60831254, Global Avg Loss: 1.73467915, Time: 0.0115 Steps: 52660, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001574, Sample Num: 25184, Cur Loss: 0.17901443, Cur Avg Loss: 0.61259673, Log Avg loss: 0.65666455, Global Avg Loss: 1.73447448, Time: 0.0068 Steps: 52670, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001584, Sample Num: 25344, Cur Loss: 1.60048270, Cur Avg Loss: 0.61423555, Log Avg loss: 0.87218478, Global Avg Loss: 1.73431079, Time: 0.0216 Steps: 52680, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001594, Sample Num: 25504, Cur Loss: 0.31269076, Cur Avg Loss: 0.61383200, Log Avg loss: 0.54991074, Global Avg Loss: 1.73408600, Time: 0.0157 Steps: 52690, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001604, Sample Num: 25664, Cur Loss: 1.34794927, Cur Avg Loss: 0.61464554, Log Avg loss: 0.74432260, Global Avg Loss: 1.73389819, Time: 0.0139 Steps: 52700, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001614, Sample Num: 25824, Cur Loss: 0.89954340, Cur Avg Loss: 0.61434056, Log Avg loss: 0.56542238, Global Avg Loss: 1.73367651, Time: 0.0088 Steps: 52710, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001624, Sample Num: 25984, Cur Loss: 0.64217317, Cur Avg Loss: 0.61466404, Log Avg loss: 0.66687373, Global Avg Loss: 1.73347416, Time: 0.0072 Steps: 52720, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001634, Sample Num: 26144, Cur Loss: 0.23274085, Cur Avg Loss: 0.61371333, Log Avg loss: 0.45931706, Global Avg Loss: 1.73323252, Time: 0.0115 Steps: 52730, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001644, Sample Num: 26304, Cur Loss: 0.51027346, Cur Avg Loss: 0.61412926, Log Avg loss: 0.68209296, Global Avg Loss: 1.73303322, Time: 0.0066 Steps: 52740, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001654, Sample Num: 26464, Cur Loss: 0.35125303, Cur Avg Loss: 0.61373143, Log Avg loss: 0.54832851, Global Avg Loss: 1.73280863, Time: 0.0066 Steps: 52750, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001664, Sample Num: 26624, Cur Loss: 0.60146093, Cur Avg Loss: 0.61460737, Log Avg loss: 0.75948830, Global Avg Loss: 1.73262415, Time: 0.0066 Steps: 52760, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001674, Sample Num: 26784, Cur Loss: 2.42142558, Cur Avg Loss: 0.61567687, Log Avg loss: 0.79364132, Global Avg Loss: 1.73244621, Time: 0.0073 Steps: 52770, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001684, Sample Num: 26944, Cur Loss: 1.29674888, Cur Avg Loss: 0.61565870, Log Avg loss: 0.61261736, Global Avg Loss: 1.73223404, Time: 0.0122 Steps: 52780, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001694, Sample Num: 27104, Cur Loss: 0.40012157, Cur Avg Loss: 0.61497647, Log Avg loss: 0.50008863, Global Avg Loss: 1.73200063, Time: 0.0114 Steps: 52790, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001704, Sample Num: 27264, Cur Loss: 0.32031491, Cur Avg Loss: 0.61529106, Log Avg loss: 0.66858298, Global Avg Loss: 1.73179923, Time: 0.0145 Steps: 52800, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001714, Sample Num: 27424, Cur Loss: 1.01424766, Cur Avg Loss: 0.61493044, Log Avg loss: 0.55348032, Global Avg Loss: 1.73157611, Time: 0.0072 Steps: 52810, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001724, Sample Num: 27584, Cur Loss: 0.47407523, Cur Avg Loss: 0.61488381, Log Avg loss: 0.60689137, Global Avg Loss: 1.73136318, Time: 0.0141 Steps: 52820, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001734, Sample Num: 27744, Cur Loss: 0.29404008, Cur Avg Loss: 0.61564142, Log Avg loss: 0.74625382, Global Avg Loss: 1.73117671, Time: 0.0111 Steps: 52830, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001744, Sample Num: 27904, Cur Loss: 0.39371690, Cur Avg Loss: 0.61536392, Log Avg loss: 0.56724515, Global Avg Loss: 1.73095643, Time: 0.0107 Steps: 52840, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001754, Sample Num: 28064, Cur Loss: 0.61588311, Cur Avg Loss: 0.61547122, Log Avg loss: 0.63418466, Global Avg Loss: 1.73074891, Time: 0.0113 Steps: 52850, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001764, Sample Num: 28224, Cur Loss: 0.34166270, Cur Avg Loss: 0.61473305, Log Avg loss: 0.48525726, Global Avg Loss: 1.73051329, Time: 0.0107 Steps: 52860, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001774, Sample Num: 28384, Cur Loss: 0.29769573, Cur Avg Loss: 0.61575807, Log Avg loss: 0.79657116, Global Avg Loss: 1.73033664, Time: 0.0067 Steps: 52870, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001784, Sample Num: 28544, Cur Loss: 0.35943067, Cur Avg Loss: 0.61445209, Log Avg loss: 0.38277123, Global Avg Loss: 1.73008181, Time: 0.0067 Steps: 52880, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001794, Sample Num: 28704, Cur Loss: 1.25140023, Cur Avg Loss: 0.61443240, Log Avg loss: 0.61091991, Global Avg Loss: 1.72987020, Time: 0.0115 Steps: 52890, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001804, Sample Num: 28864, Cur Loss: 0.56764925, Cur Avg Loss: 0.61384540, Log Avg loss: 0.50853726, Global Avg Loss: 1.72963933, Time: 0.0068 Steps: 52900, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001814, Sample Num: 29024, Cur Loss: 0.77072686, Cur Avg Loss: 0.61381023, Log Avg loss: 0.60746603, Global Avg Loss: 1.72942724, Time: 0.0067 Steps: 52910, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001824, Sample Num: 29184, Cur Loss: 0.48391855, Cur Avg Loss: 0.61384455, Log Avg loss: 0.62007113, Global Avg Loss: 1.72921761, Time: 0.0111 Steps: 52920, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001834, Sample Num: 29344, Cur Loss: 0.97199512, Cur Avg Loss: 0.61367372, Log Avg loss: 0.58251301, Global Avg Loss: 1.72900096, Time: 0.0110 Steps: 52930, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001844, Sample Num: 29504, Cur Loss: 0.98438913, Cur Avg Loss: 0.61317270, Log Avg loss: 0.52128592, Global Avg Loss: 1.72877283, Time: 0.0137 Steps: 52940, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001854, Sample Num: 29664, Cur Loss: 1.24053276, Cur Avg Loss: 0.61264667, Log Avg loss: 0.51564747, Global Avg Loss: 1.72854373, Time: 0.0123 Steps: 52950, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001864, Sample Num: 29824, Cur Loss: 1.42437720, Cur Avg Loss: 0.61327228, Log Avg loss: 0.72925959, Global Avg Loss: 1.72835504, Time: 0.0126 Steps: 52960, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001874, Sample Num: 29984, Cur Loss: 0.29767686, Cur Avg Loss: 0.61358973, Log Avg loss: 0.67276199, Global Avg Loss: 1.72815576, Time: 0.0115 Steps: 52970, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001884, Sample Num: 30144, Cur Loss: 0.89132684, Cur Avg Loss: 0.61385463, Log Avg loss: 0.66349824, Global Avg Loss: 1.72795480, Time: 0.0137 Steps: 52980, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001894, Sample Num: 30304, Cur Loss: 0.26716524, Cur Avg Loss: 0.61398856, Log Avg loss: 0.63922063, Global Avg Loss: 1.72774934, Time: 0.0114 Steps: 52990, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001904, Sample Num: 30464, Cur Loss: 0.32875460, Cur Avg Loss: 0.61323866, Log Avg loss: 0.47120801, Global Avg Loss: 1.72751226, Time: 0.0229 Steps: 53000, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001914, Sample Num: 30624, Cur Loss: 0.37745118, Cur Avg Loss: 0.61321426, Log Avg loss: 0.60856855, Global Avg Loss: 1.72730118, Time: 0.0134 Steps: 53010, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001924, Sample Num: 30784, Cur Loss: 0.57318008, Cur Avg Loss: 0.61325245, Log Avg loss: 0.62056170, Global Avg Loss: 1.72709244, Time: 0.0111 Steps: 53020, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001934, Sample Num: 30944, Cur Loss: 0.87108338, Cur Avg Loss: 0.61403066, Log Avg loss: 0.76375737, Global Avg Loss: 1.72691078, Time: 0.0105 Steps: 53030, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001944, Sample Num: 31104, Cur Loss: 1.03221738, Cur Avg Loss: 0.61561672, Log Avg loss: 0.92236103, Global Avg Loss: 1.72675909, Time: 0.0109 Steps: 53040, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001954, Sample Num: 31264, Cur Loss: 0.75148344, Cur Avg Loss: 0.61586916, Log Avg loss: 0.66494301, Global Avg Loss: 1.72655894, Time: 0.0072 Steps: 53050, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001964, Sample Num: 31424, Cur Loss: 0.31158423, Cur Avg Loss: 0.61580649, Log Avg loss: 0.60356064, Global Avg Loss: 1.72634729, Time: 0.0087 Steps: 53060, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001974, Sample Num: 31584, Cur Loss: 0.57978415, Cur Avg Loss: 0.61520457, Log Avg loss: 0.49698913, Global Avg Loss: 1.72611564, Time: 0.0066 Steps: 53070, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001984, Sample Num: 31744, Cur Loss: 0.55952716, Cur Avg Loss: 0.61570445, Log Avg loss: 0.71438066, Global Avg Loss: 1.72592504, Time: 0.0105 Steps: 53080, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 001994, Sample Num: 31904, Cur Loss: 0.24419403, Cur Avg Loss: 0.61425714, Log Avg loss: 0.32710968, Global Avg Loss: 1.72566156, Time: 0.0104 Steps: 53090, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002004, Sample Num: 32064, Cur Loss: 0.42309561, Cur Avg Loss: 0.61412288, Log Avg loss: 0.58735151, Global Avg Loss: 1.72544719, Time: 0.0112 Steps: 53100, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002014, Sample Num: 32224, Cur Loss: 0.66742289, Cur Avg Loss: 0.61366282, Log Avg loss: 0.52146621, Global Avg Loss: 1.72522049, Time: 0.0128 Steps: 53110, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002024, Sample Num: 32384, Cur Loss: 0.26921067, Cur Avg Loss: 0.61300338, Log Avg loss: 0.48019377, Global Avg Loss: 1.72498611, Time: 0.0075 Steps: 53120, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002034, Sample Num: 32544, Cur Loss: 0.69763404, Cur Avg Loss: 0.61267974, Log Avg loss: 0.54717369, Global Avg Loss: 1.72476443, Time: 0.0065 Steps: 53130, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002044, Sample Num: 32704, Cur Loss: 0.37918174, Cur Avg Loss: 0.61214358, Log Avg loss: 0.50309003, Global Avg Loss: 1.72453453, Time: 0.0068 Steps: 53140, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002054, Sample Num: 32864, Cur Loss: 0.28828108, Cur Avg Loss: 0.61149891, Log Avg loss: 0.47972802, Global Avg Loss: 1.72430032, Time: 0.0104 Steps: 53150, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002064, Sample Num: 33024, Cur Loss: 0.29913467, Cur Avg Loss: 0.61147359, Log Avg loss: 0.60627315, Global Avg Loss: 1.72409001, Time: 0.0112 Steps: 53160, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002074, Sample Num: 33184, Cur Loss: 0.39806345, Cur Avg Loss: 0.61089234, Log Avg loss: 0.49092115, Global Avg Loss: 1.72385808, Time: 0.0114 Steps: 53170, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002084, Sample Num: 33344, Cur Loss: 0.39417183, Cur Avg Loss: 0.61070736, Log Avg loss: 0.57234342, Global Avg Loss: 1.72364155, Time: 0.0065 Steps: 53180, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002094, Sample Num: 33504, Cur Loss: 0.14375934, Cur Avg Loss: 0.61126097, Log Avg loss: 0.72663201, Global Avg Loss: 1.72345410, Time: 0.0067 Steps: 53190, Updated lr: 0.000051 Training, Epoch: 0025, Batch: 002104, Sample Num: 33664, Cur Loss: 0.19993269, Cur Avg Loss: 0.61056817, Log Avg loss: 0.46549678, Global Avg Loss: 1.72321765, Time: 0.0135 Steps: 53200, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002114, Sample Num: 33824, Cur Loss: 0.85783601, Cur Avg Loss: 0.61081894, Log Avg loss: 0.66357986, Global Avg Loss: 1.72301850, Time: 0.0117 Steps: 53210, Updated lr: 0.000050 Training, Epoch: 0025, Batch: 002124, Sample Num: 33984, Cur Loss: 0.17937678, Cur Avg Loss: 0.61021478, Log Avg loss: 0.48249577, Global Avg Loss: 1.72278541, Time: 0.0118 Steps: 53220, Updated lr: 0.000050 ***** Running evaluation checkpoint-53225 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-53225 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.394940, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.676102, "eval_total_loss": 475.300033, "eval_mae": 0.619126, "eval_mse": 0.676304, "eval_r2": 0.570096, "eval_sp_statistic": 0.705303, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.755596, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.91554, "test_total_loss": 459.600871, "test_mae": 0.720219, "test_mse": 0.915719, "test_r2": 0.408987, "test_sp_statistic": 0.539385, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.641254, "test_ps_pvalue": 0.0, "lr": 5.0474158368895205e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.7226573270133763, "train_cur_epoch_loss": 1297.8928884714842, "train_cur_epoch_avg_loss": 0.6096255934577192, "train_cur_epoch_time": 22.39494037628174, "train_cur_epoch_avg_time": 0.01051899500999612, "epoch": 25, "step": 53225} ################################################## Training, Epoch: 0026, Batch: 000005, Sample Num: 80, Cur Loss: 0.97912359, Cur Avg Loss: 0.65117365, Log Avg loss: 0.50525697, Global Avg Loss: 1.72255668, Time: 0.0119 Steps: 53230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000015, Sample Num: 240, Cur Loss: 0.37853512, Cur Avg Loss: 0.59427846, Log Avg loss: 0.56583087, Global Avg Loss: 1.72233941, Time: 0.0068 Steps: 53240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000025, Sample Num: 400, Cur Loss: 0.21330167, Cur Avg Loss: 0.57907169, Log Avg loss: 0.55626153, Global Avg Loss: 1.72212043, Time: 0.0122 Steps: 53250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000035, Sample Num: 560, Cur Loss: 0.43866625, Cur Avg Loss: 0.57989357, Log Avg loss: 0.58194827, Global Avg Loss: 1.72190636, Time: 0.0113 Steps: 53260, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000045, Sample Num: 720, Cur Loss: 1.01406169, Cur Avg Loss: 0.55682355, Log Avg loss: 0.47607849, Global Avg Loss: 1.72167249, Time: 0.0227 Steps: 53270, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000055, Sample Num: 880, Cur Loss: 0.53365606, Cur Avg Loss: 0.58358561, Log Avg loss: 0.70401489, Global Avg Loss: 1.72148148, Time: 0.0072 Steps: 53280, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000065, Sample Num: 1040, Cur Loss: 0.40951151, Cur Avg Loss: 0.57635585, Log Avg loss: 0.53659213, Global Avg Loss: 1.72125914, Time: 0.0109 Steps: 53290, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000075, Sample Num: 1200, Cur Loss: 0.27031335, Cur Avg Loss: 0.58548044, Log Avg loss: 0.64479028, Global Avg Loss: 1.72105717, Time: 0.0115 Steps: 53300, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000085, Sample Num: 1360, Cur Loss: 0.88876694, Cur Avg Loss: 0.61762641, Log Avg loss: 0.85872123, Global Avg Loss: 1.72089541, Time: 0.0117 Steps: 53310, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000095, Sample Num: 1520, Cur Loss: 0.29472786, Cur Avg Loss: 0.62031610, Log Avg loss: 0.64317840, Global Avg Loss: 1.72069329, Time: 0.0090 Steps: 53320, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000105, Sample Num: 1680, Cur Loss: 0.34932941, Cur Avg Loss: 0.60816968, Log Avg loss: 0.49277873, Global Avg Loss: 1.72046304, Time: 0.0069 Steps: 53330, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000115, Sample Num: 1840, Cur Loss: 1.01166141, Cur Avg Loss: 0.61559629, Log Avg loss: 0.69357572, Global Avg Loss: 1.72027053, Time: 0.0111 Steps: 53340, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000125, Sample Num: 2000, Cur Loss: 0.30412871, Cur Avg Loss: 0.60182747, Log Avg loss: 0.44348599, Global Avg Loss: 1.72003120, Time: 0.0069 Steps: 53350, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000135, Sample Num: 2160, Cur Loss: 0.22281183, Cur Avg Loss: 0.60111442, Log Avg loss: 0.59220133, Global Avg Loss: 1.71981984, Time: 0.0115 Steps: 53360, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000145, Sample Num: 2320, Cur Loss: 0.43323943, Cur Avg Loss: 0.59611238, Log Avg loss: 0.52858488, Global Avg Loss: 1.71959664, Time: 0.0068 Steps: 53370, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000155, Sample Num: 2480, Cur Loss: 0.43790489, Cur Avg Loss: 0.60004905, Log Avg loss: 0.65713073, Global Avg Loss: 1.71939760, Time: 0.0068 Steps: 53380, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000165, Sample Num: 2640, Cur Loss: 0.53125000, Cur Avg Loss: 0.60528637, Log Avg loss: 0.68646476, Global Avg Loss: 1.71920413, Time: 0.0111 Steps: 53390, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000175, Sample Num: 2800, Cur Loss: 0.68057263, Cur Avg Loss: 0.59832498, Log Avg loss: 0.48346214, Global Avg Loss: 1.71897272, Time: 0.0137 Steps: 53400, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000185, Sample Num: 2960, Cur Loss: 1.08681965, Cur Avg Loss: 0.60553640, Log Avg loss: 0.73173626, Global Avg Loss: 1.71878788, Time: 0.0118 Steps: 53410, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000195, Sample Num: 3120, Cur Loss: 0.35427481, Cur Avg Loss: 0.60716420, Log Avg loss: 0.63727842, Global Avg Loss: 1.71858542, Time: 0.0088 Steps: 53420, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000205, Sample Num: 3280, Cur Loss: 0.23074704, Cur Avg Loss: 0.61061787, Log Avg loss: 0.67796440, Global Avg Loss: 1.71839066, Time: 0.0077 Steps: 53430, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000215, Sample Num: 3440, Cur Loss: 0.57145977, Cur Avg Loss: 0.60211763, Log Avg loss: 0.42786267, Global Avg Loss: 1.71814917, Time: 0.0069 Steps: 53440, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000225, Sample Num: 3600, Cur Loss: 0.71868634, Cur Avg Loss: 0.60085280, Log Avg loss: 0.57365906, Global Avg Loss: 1.71793504, Time: 0.0064 Steps: 53450, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000235, Sample Num: 3760, Cur Loss: 0.22033560, Cur Avg Loss: 0.59577807, Log Avg loss: 0.48159662, Global Avg Loss: 1.71770378, Time: 0.0067 Steps: 53460, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000245, Sample Num: 3920, Cur Loss: 0.15921178, Cur Avg Loss: 0.60172439, Log Avg loss: 0.74146305, Global Avg Loss: 1.71752120, Time: 0.0068 Steps: 53470, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000255, Sample Num: 4080, Cur Loss: 0.34780049, Cur Avg Loss: 0.59854114, Log Avg loss: 0.52055131, Global Avg Loss: 1.71729739, Time: 0.0232 Steps: 53480, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000265, Sample Num: 4240, Cur Loss: 0.59394133, Cur Avg Loss: 0.59796357, Log Avg loss: 0.58323557, Global Avg Loss: 1.71708537, Time: 0.0221 Steps: 53490, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000275, Sample Num: 4400, Cur Loss: 0.22467311, Cur Avg Loss: 0.59089277, Log Avg loss: 0.40351658, Global Avg Loss: 1.71683985, Time: 0.0067 Steps: 53500, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000285, Sample Num: 4560, Cur Loss: 0.37933511, Cur Avg Loss: 0.58956971, Log Avg loss: 0.55318573, Global Avg Loss: 1.71662238, Time: 0.0097 Steps: 53510, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000295, Sample Num: 4720, Cur Loss: 1.30513418, Cur Avg Loss: 0.59352390, Log Avg loss: 0.70621829, Global Avg Loss: 1.71643359, Time: 0.0117 Steps: 53520, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000305, Sample Num: 4880, Cur Loss: 0.55896461, Cur Avg Loss: 0.60270023, Log Avg loss: 0.87340173, Global Avg Loss: 1.71627610, Time: 0.0118 Steps: 53530, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000315, Sample Num: 5040, Cur Loss: 0.75772834, Cur Avg Loss: 0.60162372, Log Avg loss: 0.56879040, Global Avg Loss: 1.71606178, Time: 0.0119 Steps: 53540, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000325, Sample Num: 5200, Cur Loss: 0.62166893, Cur Avg Loss: 0.59904716, Log Avg loss: 0.51788552, Global Avg Loss: 1.71583803, Time: 0.0116 Steps: 53550, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000335, Sample Num: 5360, Cur Loss: 0.58745635, Cur Avg Loss: 0.59872273, Log Avg loss: 0.58817858, Global Avg Loss: 1.71562749, Time: 0.0228 Steps: 53560, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000345, Sample Num: 5520, Cur Loss: 0.92556977, Cur Avg Loss: 0.60091132, Log Avg loss: 0.67422926, Global Avg Loss: 1.71543309, Time: 0.0121 Steps: 53570, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000355, Sample Num: 5680, Cur Loss: 0.66834456, Cur Avg Loss: 0.59993682, Log Avg loss: 0.56631659, Global Avg Loss: 1.71521862, Time: 0.0155 Steps: 53580, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000365, Sample Num: 5840, Cur Loss: 0.34409407, Cur Avg Loss: 0.59635083, Log Avg loss: 0.46904824, Global Avg Loss: 1.71498608, Time: 0.0118 Steps: 53590, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000375, Sample Num: 6000, Cur Loss: 0.95535523, Cur Avg Loss: 0.59688090, Log Avg loss: 0.61622820, Global Avg Loss: 1.71478109, Time: 0.0068 Steps: 53600, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000385, Sample Num: 6160, Cur Loss: 0.83440769, Cur Avg Loss: 0.59690300, Log Avg loss: 0.59773199, Global Avg Loss: 1.71457273, Time: 0.0115 Steps: 53610, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000395, Sample Num: 6320, Cur Loss: 0.54355729, Cur Avg Loss: 0.59747592, Log Avg loss: 0.61953304, Global Avg Loss: 1.71436850, Time: 0.0067 Steps: 53620, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000405, Sample Num: 6480, Cur Loss: 0.35851160, Cur Avg Loss: 0.59410806, Log Avg loss: 0.46107784, Global Avg Loss: 1.71413481, Time: 0.0073 Steps: 53630, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000415, Sample Num: 6640, Cur Loss: 0.74446273, Cur Avg Loss: 0.59197022, Log Avg loss: 0.50538761, Global Avg Loss: 1.71390947, Time: 0.0115 Steps: 53640, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000425, Sample Num: 6800, Cur Loss: 0.22468144, Cur Avg Loss: 0.59602906, Log Avg loss: 0.76447096, Global Avg Loss: 1.71373250, Time: 0.0069 Steps: 53650, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000435, Sample Num: 6960, Cur Loss: 0.65724605, Cur Avg Loss: 0.59764754, Log Avg loss: 0.66643281, Global Avg Loss: 1.71353733, Time: 0.0067 Steps: 53660, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000445, Sample Num: 7120, Cur Loss: 0.50132692, Cur Avg Loss: 0.60073859, Log Avg loss: 0.73519938, Global Avg Loss: 1.71335504, Time: 0.0133 Steps: 53670, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000455, Sample Num: 7280, Cur Loss: 0.58213341, Cur Avg Loss: 0.60051006, Log Avg loss: 0.59034034, Global Avg Loss: 1.71314583, Time: 0.0154 Steps: 53680, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000465, Sample Num: 7440, Cur Loss: 0.48435616, Cur Avg Loss: 0.59962594, Log Avg loss: 0.55939871, Global Avg Loss: 1.71293094, Time: 0.0117 Steps: 53690, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000475, Sample Num: 7600, Cur Loss: 0.97759187, Cur Avg Loss: 0.60061657, Log Avg loss: 0.64668093, Global Avg Loss: 1.71273239, Time: 0.0164 Steps: 53700, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000485, Sample Num: 7760, Cur Loss: 1.02160239, Cur Avg Loss: 0.60197116, Log Avg loss: 0.66631418, Global Avg Loss: 1.71253756, Time: 0.0073 Steps: 53710, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000495, Sample Num: 7920, Cur Loss: 0.47850907, Cur Avg Loss: 0.60116659, Log Avg loss: 0.56214486, Global Avg Loss: 1.71232341, Time: 0.0080 Steps: 53720, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000505, Sample Num: 8080, Cur Loss: 0.56024188, Cur Avg Loss: 0.60051792, Log Avg loss: 0.56840860, Global Avg Loss: 1.71211051, Time: 0.0120 Steps: 53730, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000515, Sample Num: 8240, Cur Loss: 0.80388331, Cur Avg Loss: 0.60245436, Log Avg loss: 0.70024474, Global Avg Loss: 1.71192222, Time: 0.0166 Steps: 53740, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000525, Sample Num: 8400, Cur Loss: 0.50362134, Cur Avg Loss: 0.59955297, Log Avg loss: 0.45013140, Global Avg Loss: 1.71168747, Time: 0.0066 Steps: 53750, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000535, Sample Num: 8560, Cur Loss: 1.07496524, Cur Avg Loss: 0.60034036, Log Avg loss: 0.64167825, Global Avg Loss: 1.71148844, Time: 0.0179 Steps: 53760, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000545, Sample Num: 8720, Cur Loss: 0.46462548, Cur Avg Loss: 0.59939101, Log Avg loss: 0.54860086, Global Avg Loss: 1.71127217, Time: 0.0123 Steps: 53770, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000555, Sample Num: 8880, Cur Loss: 1.03747618, Cur Avg Loss: 0.59813247, Log Avg loss: 0.52954194, Global Avg Loss: 1.71105243, Time: 0.0128 Steps: 53780, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000565, Sample Num: 9040, Cur Loss: 0.71984565, Cur Avg Loss: 0.59859136, Log Avg loss: 0.62405944, Global Avg Loss: 1.71085035, Time: 0.0116 Steps: 53790, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000575, Sample Num: 9200, Cur Loss: 1.33161736, Cur Avg Loss: 0.59996426, Log Avg loss: 0.67753325, Global Avg Loss: 1.71065828, Time: 0.0119 Steps: 53800, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000585, Sample Num: 9360, Cur Loss: 0.38854706, Cur Avg Loss: 0.59811830, Log Avg loss: 0.49197581, Global Avg Loss: 1.71043181, Time: 0.0116 Steps: 53810, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000595, Sample Num: 9520, Cur Loss: 1.21720338, Cur Avg Loss: 0.59860544, Log Avg loss: 0.62710299, Global Avg Loss: 1.71023052, Time: 0.0121 Steps: 53820, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000605, Sample Num: 9680, Cur Loss: 0.33479029, Cur Avg Loss: 0.59738991, Log Avg loss: 0.52506623, Global Avg Loss: 1.71001035, Time: 0.0065 Steps: 53830, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000615, Sample Num: 9840, Cur Loss: 0.23075199, Cur Avg Loss: 0.59541073, Log Avg loss: 0.47566995, Global Avg Loss: 1.70978109, Time: 0.0137 Steps: 53840, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000625, Sample Num: 10000, Cur Loss: 0.51444620, Cur Avg Loss: 0.59608034, Log Avg loss: 0.63726140, Global Avg Loss: 1.70958192, Time: 0.0119 Steps: 53850, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000635, Sample Num: 10160, Cur Loss: 0.70015955, Cur Avg Loss: 0.59612283, Log Avg loss: 0.59877824, Global Avg Loss: 1.70937568, Time: 0.0110 Steps: 53860, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000645, Sample Num: 10320, Cur Loss: 0.25901616, Cur Avg Loss: 0.59623194, Log Avg loss: 0.60316054, Global Avg Loss: 1.70917033, Time: 0.0232 Steps: 53870, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000655, Sample Num: 10480, Cur Loss: 0.46096760, Cur Avg Loss: 0.59679505, Log Avg loss: 0.63311556, Global Avg Loss: 1.70897062, Time: 0.0079 Steps: 53880, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000665, Sample Num: 10640, Cur Loss: 0.95293540, Cur Avg Loss: 0.60078200, Log Avg loss: 0.86192759, Global Avg Loss: 1.70881344, Time: 0.0077 Steps: 53890, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000675, Sample Num: 10800, Cur Loss: 0.61752295, Cur Avg Loss: 0.59885705, Log Avg loss: 0.47084786, Global Avg Loss: 1.70858376, Time: 0.0088 Steps: 53900, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000685, Sample Num: 10960, Cur Loss: 0.26315933, Cur Avg Loss: 0.59646097, Log Avg loss: 0.43472538, Global Avg Loss: 1.70834747, Time: 0.0110 Steps: 53910, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000695, Sample Num: 11120, Cur Loss: 1.02282631, Cur Avg Loss: 0.59731716, Log Avg loss: 0.65596612, Global Avg Loss: 1.70815229, Time: 0.0224 Steps: 53920, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000705, Sample Num: 11280, Cur Loss: 0.41883743, Cur Avg Loss: 0.59674704, Log Avg loss: 0.55712385, Global Avg Loss: 1.70793886, Time: 0.0066 Steps: 53930, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000715, Sample Num: 11440, Cur Loss: 0.15916508, Cur Avg Loss: 0.59392739, Log Avg loss: 0.39514208, Global Avg Loss: 1.70769548, Time: 0.0111 Steps: 53940, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000725, Sample Num: 11600, Cur Loss: 0.22025283, Cur Avg Loss: 0.59161224, Log Avg loss: 0.42607902, Global Avg Loss: 1.70745793, Time: 0.0109 Steps: 53950, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000735, Sample Num: 11760, Cur Loss: 1.14690590, Cur Avg Loss: 0.59343463, Log Avg loss: 0.72555759, Global Avg Loss: 1.70727596, Time: 0.0143 Steps: 53960, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000745, Sample Num: 11920, Cur Loss: 0.51888204, Cur Avg Loss: 0.59303577, Log Avg loss: 0.56371993, Global Avg Loss: 1.70706407, Time: 0.0069 Steps: 53970, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000755, Sample Num: 12080, Cur Loss: 0.23102978, Cur Avg Loss: 0.59484059, Log Avg loss: 0.72929947, Global Avg Loss: 1.70688294, Time: 0.0110 Steps: 53980, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000765, Sample Num: 12240, Cur Loss: 0.59358495, Cur Avg Loss: 0.59493687, Log Avg loss: 0.60220608, Global Avg Loss: 1.70667833, Time: 0.0219 Steps: 53990, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000775, Sample Num: 12400, Cur Loss: 0.47998673, Cur Avg Loss: 0.59330485, Log Avg loss: 0.46845501, Global Avg Loss: 1.70644903, Time: 0.0070 Steps: 54000, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000785, Sample Num: 12560, Cur Loss: 0.51832259, Cur Avg Loss: 0.59259826, Log Avg loss: 0.53783777, Global Avg Loss: 1.70623266, Time: 0.0066 Steps: 54010, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000795, Sample Num: 12720, Cur Loss: 0.35931775, Cur Avg Loss: 0.59186165, Log Avg loss: 0.53403803, Global Avg Loss: 1.70601567, Time: 0.0072 Steps: 54020, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000805, Sample Num: 12880, Cur Loss: 0.37069029, Cur Avg Loss: 0.59466425, Log Avg loss: 0.81747052, Global Avg Loss: 1.70585121, Time: 0.0067 Steps: 54030, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000815, Sample Num: 13040, Cur Loss: 0.58153403, Cur Avg Loss: 0.59647886, Log Avg loss: 0.74255535, Global Avg Loss: 1.70567296, Time: 0.0185 Steps: 54040, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000825, Sample Num: 13200, Cur Loss: 0.16385266, Cur Avg Loss: 0.59622404, Log Avg loss: 0.57545590, Global Avg Loss: 1.70546385, Time: 0.0068 Steps: 54050, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000835, Sample Num: 13360, Cur Loss: 1.47619653, Cur Avg Loss: 0.59809436, Log Avg loss: 0.75239624, Global Avg Loss: 1.70528755, Time: 0.0065 Steps: 54060, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000845, Sample Num: 13520, Cur Loss: 0.31872711, Cur Avg Loss: 0.59701842, Log Avg loss: 0.50717716, Global Avg Loss: 1.70506597, Time: 0.0199 Steps: 54070, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000855, Sample Num: 13680, Cur Loss: 0.55682766, Cur Avg Loss: 0.59791849, Log Avg loss: 0.67397416, Global Avg Loss: 1.70487531, Time: 0.0065 Steps: 54080, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000865, Sample Num: 13840, Cur Loss: 0.31399941, Cur Avg Loss: 0.59803987, Log Avg loss: 0.60841851, Global Avg Loss: 1.70467260, Time: 0.0073 Steps: 54090, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000875, Sample Num: 14000, Cur Loss: 0.27989477, Cur Avg Loss: 0.60071422, Log Avg loss: 0.83204476, Global Avg Loss: 1.70451130, Time: 0.0078 Steps: 54100, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000885, Sample Num: 14160, Cur Loss: 0.55910891, Cur Avg Loss: 0.60152380, Log Avg loss: 0.67236243, Global Avg Loss: 1.70432055, Time: 0.0066 Steps: 54110, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000895, Sample Num: 14320, Cur Loss: 0.75212270, Cur Avg Loss: 0.60033247, Log Avg loss: 0.49489947, Global Avg Loss: 1.70409708, Time: 0.0212 Steps: 54120, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000905, Sample Num: 14480, Cur Loss: 0.48344916, Cur Avg Loss: 0.60022931, Log Avg loss: 0.59099680, Global Avg Loss: 1.70389144, Time: 0.0168 Steps: 54130, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000915, Sample Num: 14640, Cur Loss: 1.00446022, Cur Avg Loss: 0.60109782, Log Avg loss: 0.67969820, Global Avg Loss: 1.70370227, Time: 0.0066 Steps: 54140, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000925, Sample Num: 14800, Cur Loss: 0.74940658, Cur Avg Loss: 0.60229435, Log Avg loss: 0.71177636, Global Avg Loss: 1.70351909, Time: 0.0064 Steps: 54150, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000935, Sample Num: 14960, Cur Loss: 0.63046551, Cur Avg Loss: 0.60202118, Log Avg loss: 0.57675351, Global Avg Loss: 1.70331104, Time: 0.0072 Steps: 54160, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000945, Sample Num: 15120, Cur Loss: 0.54967761, Cur Avg Loss: 0.60363147, Log Avg loss: 0.75419336, Global Avg Loss: 1.70313583, Time: 0.0076 Steps: 54170, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000955, Sample Num: 15280, Cur Loss: 0.40123069, Cur Avg Loss: 0.60203631, Log Avg loss: 0.45129353, Global Avg Loss: 1.70290478, Time: 0.0069 Steps: 54180, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000965, Sample Num: 15440, Cur Loss: 0.68191159, Cur Avg Loss: 0.60065723, Log Avg loss: 0.46895489, Global Avg Loss: 1.70267707, Time: 0.0179 Steps: 54190, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000975, Sample Num: 15600, Cur Loss: 0.41759771, Cur Avg Loss: 0.60015791, Log Avg loss: 0.55197366, Global Avg Loss: 1.70246476, Time: 0.0097 Steps: 54200, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000985, Sample Num: 15760, Cur Loss: 0.41829252, Cur Avg Loss: 0.60063200, Log Avg loss: 0.64685571, Global Avg Loss: 1.70227004, Time: 0.0120 Steps: 54210, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 000995, Sample Num: 15920, Cur Loss: 0.75536406, Cur Avg Loss: 0.60133837, Log Avg loss: 0.67091588, Global Avg Loss: 1.70207982, Time: 0.0205 Steps: 54220, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001005, Sample Num: 16080, Cur Loss: 0.50810838, Cur Avg Loss: 0.60191253, Log Avg loss: 0.65904134, Global Avg Loss: 1.70188749, Time: 0.0068 Steps: 54230, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001015, Sample Num: 16240, Cur Loss: 0.83858454, Cur Avg Loss: 0.60249359, Log Avg loss: 0.66089070, Global Avg Loss: 1.70169556, Time: 0.0068 Steps: 54240, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001025, Sample Num: 16400, Cur Loss: 1.03139329, Cur Avg Loss: 0.60269286, Log Avg loss: 0.62291850, Global Avg Loss: 1.70149671, Time: 0.0131 Steps: 54250, Updated lr: 0.000050 Training, Epoch: 0026, Batch: 001035, Sample Num: 16560, Cur Loss: 0.81962299, Cur Avg Loss: 0.60178452, Log Avg loss: 0.50867983, Global Avg Loss: 1.70127687, Time: 0.0073 Steps: 54260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001045, Sample Num: 16720, Cur Loss: 1.28191328, Cur Avg Loss: 0.60194626, Log Avg loss: 0.61868561, Global Avg Loss: 1.70107739, Time: 0.0116 Steps: 54270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001055, Sample Num: 16880, Cur Loss: 1.19793069, Cur Avg Loss: 0.60391434, Log Avg loss: 0.80957937, Global Avg Loss: 1.70091315, Time: 0.0111 Steps: 54280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001065, Sample Num: 17040, Cur Loss: 0.81488663, Cur Avg Loss: 0.60373801, Log Avg loss: 0.58513501, Global Avg Loss: 1.70070763, Time: 0.0068 Steps: 54290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001075, Sample Num: 17200, Cur Loss: 0.12600774, Cur Avg Loss: 0.60252190, Log Avg loss: 0.47300636, Global Avg Loss: 1.70048153, Time: 0.0068 Steps: 54300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001085, Sample Num: 17360, Cur Loss: 0.67899954, Cur Avg Loss: 0.60327718, Log Avg loss: 0.68447007, Global Avg Loss: 1.70029446, Time: 0.0140 Steps: 54310, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001095, Sample Num: 17520, Cur Loss: 0.29686147, Cur Avg Loss: 0.60235954, Log Avg loss: 0.50279553, Global Avg Loss: 1.70007400, Time: 0.0118 Steps: 54320, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001105, Sample Num: 17680, Cur Loss: 1.12243748, Cur Avg Loss: 0.60364916, Log Avg loss: 0.74486227, Global Avg Loss: 1.69989819, Time: 0.0118 Steps: 54330, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001115, Sample Num: 17840, Cur Loss: 0.55148041, Cur Avg Loss: 0.60301011, Log Avg loss: 0.53239450, Global Avg Loss: 1.69968334, Time: 0.0110 Steps: 54340, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001125, Sample Num: 18000, Cur Loss: 0.42425555, Cur Avg Loss: 0.60391657, Log Avg loss: 0.70498784, Global Avg Loss: 1.69950032, Time: 0.0117 Steps: 54350, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001135, Sample Num: 18160, Cur Loss: 0.58980435, Cur Avg Loss: 0.60433399, Log Avg loss: 0.65129266, Global Avg Loss: 1.69930749, Time: 0.0105 Steps: 54360, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001145, Sample Num: 18320, Cur Loss: 0.44143909, Cur Avg Loss: 0.60500845, Log Avg loss: 0.68156075, Global Avg Loss: 1.69912030, Time: 0.0090 Steps: 54370, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001155, Sample Num: 18480, Cur Loss: 0.55330992, Cur Avg Loss: 0.60653063, Log Avg loss: 0.78082027, Global Avg Loss: 1.69895144, Time: 0.0071 Steps: 54380, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001165, Sample Num: 18640, Cur Loss: 0.69159967, Cur Avg Loss: 0.60590322, Log Avg loss: 0.53343702, Global Avg Loss: 1.69873715, Time: 0.0110 Steps: 54390, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001175, Sample Num: 18800, Cur Loss: 1.77565062, Cur Avg Loss: 0.60740147, Log Avg loss: 0.78194729, Global Avg Loss: 1.69856862, Time: 0.0119 Steps: 54400, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001185, Sample Num: 18960, Cur Loss: 0.34537792, Cur Avg Loss: 0.60686752, Log Avg loss: 0.54412879, Global Avg Loss: 1.69835645, Time: 0.0134 Steps: 54410, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001195, Sample Num: 19120, Cur Loss: 0.85192025, Cur Avg Loss: 0.60575446, Log Avg loss: 0.47385686, Global Avg Loss: 1.69813144, Time: 0.0068 Steps: 54420, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001205, Sample Num: 19280, Cur Loss: 0.95193851, Cur Avg Loss: 0.60501645, Log Avg loss: 0.51682384, Global Avg Loss: 1.69791440, Time: 0.0155 Steps: 54430, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001215, Sample Num: 19440, Cur Loss: 0.42645001, Cur Avg Loss: 0.60463445, Log Avg loss: 0.55860348, Global Avg Loss: 1.69770513, Time: 0.0073 Steps: 54440, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001225, Sample Num: 19600, Cur Loss: 0.30393896, Cur Avg Loss: 0.60466893, Log Avg loss: 0.60885838, Global Avg Loss: 1.69750515, Time: 0.0111 Steps: 54450, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001235, Sample Num: 19760, Cur Loss: 0.27180138, Cur Avg Loss: 0.60394049, Log Avg loss: 0.51470709, Global Avg Loss: 1.69728797, Time: 0.0109 Steps: 54460, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001245, Sample Num: 19920, Cur Loss: 0.97145891, Cur Avg Loss: 0.60353763, Log Avg loss: 0.55378408, Global Avg Loss: 1.69707804, Time: 0.0080 Steps: 54470, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001255, Sample Num: 20080, Cur Loss: 0.28259972, Cur Avg Loss: 0.60362815, Log Avg loss: 0.61489763, Global Avg Loss: 1.69687940, Time: 0.0069 Steps: 54480, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001265, Sample Num: 20240, Cur Loss: 0.49928552, Cur Avg Loss: 0.60274643, Log Avg loss: 0.49209102, Global Avg Loss: 1.69665829, Time: 0.0113 Steps: 54490, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001275, Sample Num: 20400, Cur Loss: 0.64259040, Cur Avg Loss: 0.60193519, Log Avg loss: 0.49931289, Global Avg Loss: 1.69643860, Time: 0.0074 Steps: 54500, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001285, Sample Num: 20560, Cur Loss: 0.33654842, Cur Avg Loss: 0.60177250, Log Avg loss: 0.58102963, Global Avg Loss: 1.69623397, Time: 0.0113 Steps: 54510, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001295, Sample Num: 20720, Cur Loss: 0.31152967, Cur Avg Loss: 0.60115274, Log Avg loss: 0.52151304, Global Avg Loss: 1.69601851, Time: 0.0110 Steps: 54520, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001305, Sample Num: 20880, Cur Loss: 0.27409312, Cur Avg Loss: 0.60100372, Log Avg loss: 0.58170547, Global Avg Loss: 1.69581416, Time: 0.0089 Steps: 54530, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001315, Sample Num: 21040, Cur Loss: 0.43296444, Cur Avg Loss: 0.60005999, Log Avg loss: 0.47690374, Global Avg Loss: 1.69559067, Time: 0.0106 Steps: 54540, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001325, Sample Num: 21200, Cur Loss: 0.28428379, Cur Avg Loss: 0.60164695, Log Avg loss: 0.81033186, Global Avg Loss: 1.69542839, Time: 0.0066 Steps: 54550, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001335, Sample Num: 21360, Cur Loss: 0.74256313, Cur Avg Loss: 0.60198989, Log Avg loss: 0.64743003, Global Avg Loss: 1.69523630, Time: 0.0110 Steps: 54560, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001345, Sample Num: 21520, Cur Loss: 0.81692082, Cur Avg Loss: 0.60350550, Log Avg loss: 0.80583990, Global Avg Loss: 1.69507332, Time: 0.0087 Steps: 54570, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001355, Sample Num: 21680, Cur Loss: 0.49564028, Cur Avg Loss: 0.60305469, Log Avg loss: 0.54242027, Global Avg Loss: 1.69486214, Time: 0.0166 Steps: 54580, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001365, Sample Num: 21840, Cur Loss: 0.77379996, Cur Avg Loss: 0.60398564, Log Avg loss: 0.73012864, Global Avg Loss: 1.69468541, Time: 0.0227 Steps: 54590, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001375, Sample Num: 22000, Cur Loss: 0.26973420, Cur Avg Loss: 0.60374268, Log Avg loss: 0.57057885, Global Avg Loss: 1.69447953, Time: 0.0110 Steps: 54600, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001385, Sample Num: 22160, Cur Loss: 0.52631700, Cur Avg Loss: 0.60332621, Log Avg loss: 0.54606199, Global Avg Loss: 1.69426924, Time: 0.0067 Steps: 54610, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001395, Sample Num: 22320, Cur Loss: 0.29415956, Cur Avg Loss: 0.60330514, Log Avg loss: 0.60038755, Global Avg Loss: 1.69406897, Time: 0.0075 Steps: 54620, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001405, Sample Num: 22480, Cur Loss: 0.76136553, Cur Avg Loss: 0.60357751, Log Avg loss: 0.64157208, Global Avg Loss: 1.69387631, Time: 0.0065 Steps: 54630, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001415, Sample Num: 22640, Cur Loss: 0.88017303, Cur Avg Loss: 0.60335277, Log Avg loss: 0.57177756, Global Avg Loss: 1.69367094, Time: 0.0217 Steps: 54640, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001425, Sample Num: 22800, Cur Loss: 0.48711789, Cur Avg Loss: 0.60294275, Log Avg loss: 0.54492408, Global Avg Loss: 1.69346074, Time: 0.0067 Steps: 54650, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001435, Sample Num: 22960, Cur Loss: 0.61237311, Cur Avg Loss: 0.60348195, Log Avg loss: 0.68031875, Global Avg Loss: 1.69327539, Time: 0.0065 Steps: 54660, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001445, Sample Num: 23120, Cur Loss: 0.40067571, Cur Avg Loss: 0.60264372, Log Avg loss: 0.48235773, Global Avg Loss: 1.69305389, Time: 0.0066 Steps: 54670, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001455, Sample Num: 23280, Cur Loss: 0.55792683, Cur Avg Loss: 0.60180700, Log Avg loss: 0.48090126, Global Avg Loss: 1.69283221, Time: 0.0073 Steps: 54680, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001465, Sample Num: 23440, Cur Loss: 0.39818120, Cur Avg Loss: 0.60110050, Log Avg loss: 0.49830359, Global Avg Loss: 1.69261380, Time: 0.0124 Steps: 54690, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001475, Sample Num: 23600, Cur Loss: 0.31767109, Cur Avg Loss: 0.60007923, Log Avg loss: 0.45046395, Global Avg Loss: 1.69238671, Time: 0.0202 Steps: 54700, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001485, Sample Num: 23760, Cur Loss: 1.43898010, Cur Avg Loss: 0.60019639, Log Avg loss: 0.61747719, Global Avg Loss: 1.69219024, Time: 0.0068 Steps: 54710, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001495, Sample Num: 23920, Cur Loss: 0.23267448, Cur Avg Loss: 0.60102802, Log Avg loss: 0.72452450, Global Avg Loss: 1.69201340, Time: 0.0067 Steps: 54720, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001505, Sample Num: 24080, Cur Loss: 0.49672696, Cur Avg Loss: 0.60113145, Log Avg loss: 0.61659525, Global Avg Loss: 1.69181690, Time: 0.0078 Steps: 54730, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001515, Sample Num: 24240, Cur Loss: 0.65990591, Cur Avg Loss: 0.60159691, Log Avg loss: 0.67164894, Global Avg Loss: 1.69163054, Time: 0.0064 Steps: 54740, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001525, Sample Num: 24400, Cur Loss: 0.43448281, Cur Avg Loss: 0.60290851, Log Avg loss: 0.80161585, Global Avg Loss: 1.69146798, Time: 0.0063 Steps: 54750, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001535, Sample Num: 24560, Cur Loss: 0.47656727, Cur Avg Loss: 0.60264156, Log Avg loss: 0.56193108, Global Avg Loss: 1.69126171, Time: 0.0064 Steps: 54760, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001545, Sample Num: 24720, Cur Loss: 1.28373551, Cur Avg Loss: 0.60267592, Log Avg loss: 0.60795087, Global Avg Loss: 1.69106391, Time: 0.0065 Steps: 54770, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001555, Sample Num: 24880, Cur Loss: 0.49051216, Cur Avg Loss: 0.60215489, Log Avg loss: 0.52165482, Global Avg Loss: 1.69085044, Time: 0.0064 Steps: 54780, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001565, Sample Num: 25040, Cur Loss: 0.84338200, Cur Avg Loss: 0.60202647, Log Avg loss: 0.58205775, Global Avg Loss: 1.69064807, Time: 0.0113 Steps: 54790, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001575, Sample Num: 25200, Cur Loss: 0.28147119, Cur Avg Loss: 0.60131499, Log Avg loss: 0.48996808, Global Avg Loss: 1.69042897, Time: 0.0110 Steps: 54800, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001585, Sample Num: 25360, Cur Loss: 0.55739307, Cur Avg Loss: 0.60109374, Log Avg loss: 0.56624720, Global Avg Loss: 1.69022386, Time: 0.0109 Steps: 54810, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001595, Sample Num: 25520, Cur Loss: 0.35711464, Cur Avg Loss: 0.60069958, Log Avg loss: 0.53822440, Global Avg Loss: 1.69001372, Time: 0.0064 Steps: 54820, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001605, Sample Num: 25680, Cur Loss: 0.90010691, Cur Avg Loss: 0.60020502, Log Avg loss: 0.52132347, Global Avg Loss: 1.68980057, Time: 0.0066 Steps: 54830, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001615, Sample Num: 25840, Cur Loss: 0.94229311, Cur Avg Loss: 0.60078287, Log Avg loss: 0.69352824, Global Avg Loss: 1.68961890, Time: 0.0112 Steps: 54840, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001625, Sample Num: 26000, Cur Loss: 0.60088253, Cur Avg Loss: 0.60080990, Log Avg loss: 0.60517367, Global Avg Loss: 1.68942119, Time: 0.0071 Steps: 54850, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001635, Sample Num: 26160, Cur Loss: 0.88622302, Cur Avg Loss: 0.60053675, Log Avg loss: 0.55615142, Global Avg Loss: 1.68921462, Time: 0.0063 Steps: 54860, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001645, Sample Num: 26320, Cur Loss: 0.33523017, Cur Avg Loss: 0.60094550, Log Avg loss: 0.66777594, Global Avg Loss: 1.68902846, Time: 0.0110 Steps: 54870, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001655, Sample Num: 26480, Cur Loss: 0.66369569, Cur Avg Loss: 0.60138965, Log Avg loss: 0.67445103, Global Avg Loss: 1.68884359, Time: 0.0131 Steps: 54880, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001665, Sample Num: 26640, Cur Loss: 0.41952401, Cur Avg Loss: 0.60052462, Log Avg loss: 0.45736248, Global Avg Loss: 1.68861923, Time: 0.0070 Steps: 54890, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001675, Sample Num: 26800, Cur Loss: 0.43638140, Cur Avg Loss: 0.60043294, Log Avg loss: 0.58516794, Global Avg Loss: 1.68841824, Time: 0.0065 Steps: 54900, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001685, Sample Num: 26960, Cur Loss: 0.42914614, Cur Avg Loss: 0.59964755, Log Avg loss: 0.46809592, Global Avg Loss: 1.68819600, Time: 0.0132 Steps: 54910, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001695, Sample Num: 27120, Cur Loss: 0.91283512, Cur Avg Loss: 0.59947392, Log Avg loss: 0.57021691, Global Avg Loss: 1.68799243, Time: 0.0107 Steps: 54920, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001705, Sample Num: 27280, Cur Loss: 0.63922662, Cur Avg Loss: 0.59906932, Log Avg loss: 0.53048931, Global Avg Loss: 1.68778171, Time: 0.0116 Steps: 54930, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001715, Sample Num: 27440, Cur Loss: 0.96980703, Cur Avg Loss: 0.59987421, Log Avg loss: 0.73710799, Global Avg Loss: 1.68760867, Time: 0.0201 Steps: 54940, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001725, Sample Num: 27600, Cur Loss: 0.34665480, Cur Avg Loss: 0.59929113, Log Avg loss: 0.49929385, Global Avg Loss: 1.68739242, Time: 0.0115 Steps: 54950, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001735, Sample Num: 27760, Cur Loss: 1.35630584, Cur Avg Loss: 0.60076021, Log Avg loss: 0.85417556, Global Avg Loss: 1.68724082, Time: 0.0092 Steps: 54960, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001745, Sample Num: 27920, Cur Loss: 0.41682819, Cur Avg Loss: 0.60126721, Log Avg loss: 0.68923270, Global Avg Loss: 1.68705926, Time: 0.0118 Steps: 54970, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001755, Sample Num: 28080, Cur Loss: 0.59166873, Cur Avg Loss: 0.60095705, Log Avg loss: 0.54683293, Global Avg Loss: 1.68685187, Time: 0.0119 Steps: 54980, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001765, Sample Num: 28240, Cur Loss: 0.79716325, Cur Avg Loss: 0.60207492, Log Avg loss: 0.79826214, Global Avg Loss: 1.68669028, Time: 0.0117 Steps: 54990, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001775, Sample Num: 28400, Cur Loss: 0.83220923, Cur Avg Loss: 0.60259771, Log Avg loss: 0.69486972, Global Avg Loss: 1.68650995, Time: 0.0107 Steps: 55000, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001785, Sample Num: 28560, Cur Loss: 0.93901926, Cur Avg Loss: 0.60379697, Log Avg loss: 0.81666451, Global Avg Loss: 1.68635182, Time: 0.0086 Steps: 55010, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001795, Sample Num: 28720, Cur Loss: 0.24354264, Cur Avg Loss: 0.60295934, Log Avg loss: 0.45344266, Global Avg Loss: 1.68612774, Time: 0.0136 Steps: 55020, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001805, Sample Num: 28880, Cur Loss: 0.30688789, Cur Avg Loss: 0.60166573, Log Avg loss: 0.36946379, Global Avg Loss: 1.68588848, Time: 0.0085 Steps: 55030, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001815, Sample Num: 29040, Cur Loss: 1.06002903, Cur Avg Loss: 0.60218829, Log Avg loss: 0.69650927, Global Avg Loss: 1.68570872, Time: 0.0220 Steps: 55040, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001825, Sample Num: 29200, Cur Loss: 0.38163468, Cur Avg Loss: 0.60208652, Log Avg loss: 0.58361588, Global Avg Loss: 1.68550852, Time: 0.0107 Steps: 55050, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001835, Sample Num: 29360, Cur Loss: 0.53570855, Cur Avg Loss: 0.60160706, Log Avg loss: 0.51410590, Global Avg Loss: 1.68529577, Time: 0.0119 Steps: 55060, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001845, Sample Num: 29520, Cur Loss: 0.30274111, Cur Avg Loss: 0.60089393, Log Avg loss: 0.47003355, Global Avg Loss: 1.68507510, Time: 0.0107 Steps: 55070, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001855, Sample Num: 29680, Cur Loss: 0.26855969, Cur Avg Loss: 0.60050043, Log Avg loss: 0.52790124, Global Avg Loss: 1.68486501, Time: 0.0067 Steps: 55080, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001865, Sample Num: 29840, Cur Loss: 0.44868374, Cur Avg Loss: 0.59980289, Log Avg loss: 0.47040897, Global Avg Loss: 1.68464456, Time: 0.0118 Steps: 55090, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001875, Sample Num: 30000, Cur Loss: 0.47654486, Cur Avg Loss: 0.59962372, Log Avg loss: 0.56620790, Global Avg Loss: 1.68444157, Time: 0.0123 Steps: 55100, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001885, Sample Num: 30160, Cur Loss: 0.63806200, Cur Avg Loss: 0.59974953, Log Avg loss: 0.62333944, Global Avg Loss: 1.68424903, Time: 0.0064 Steps: 55110, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001895, Sample Num: 30320, Cur Loss: 0.61567611, Cur Avg Loss: 0.60002483, Log Avg loss: 0.65191845, Global Avg Loss: 1.68406174, Time: 0.0105 Steps: 55120, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001905, Sample Num: 30480, Cur Loss: 0.63612896, Cur Avg Loss: 0.60008871, Log Avg loss: 0.61219306, Global Avg Loss: 1.68386732, Time: 0.0066 Steps: 55130, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001915, Sample Num: 30640, Cur Loss: 0.42026764, Cur Avg Loss: 0.60040475, Log Avg loss: 0.66061194, Global Avg Loss: 1.68368174, Time: 0.0129 Steps: 55140, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001925, Sample Num: 30800, Cur Loss: 0.51727128, Cur Avg Loss: 0.59978392, Log Avg loss: 0.48089417, Global Avg Loss: 1.68346365, Time: 0.0072 Steps: 55150, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001935, Sample Num: 30960, Cur Loss: 0.65718591, Cur Avg Loss: 0.59941199, Log Avg loss: 0.52781473, Global Avg Loss: 1.68325414, Time: 0.0108 Steps: 55160, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001945, Sample Num: 31120, Cur Loss: 0.30230314, Cur Avg Loss: 0.59931921, Log Avg loss: 0.58136770, Global Avg Loss: 1.68305442, Time: 0.0133 Steps: 55170, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001955, Sample Num: 31280, Cur Loss: 0.45707488, Cur Avg Loss: 0.60023457, Log Avg loss: 0.77827080, Global Avg Loss: 1.68289045, Time: 0.0065 Steps: 55180, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001965, Sample Num: 31440, Cur Loss: 0.21754318, Cur Avg Loss: 0.60015601, Log Avg loss: 0.58479808, Global Avg Loss: 1.68269148, Time: 0.0152 Steps: 55190, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001975, Sample Num: 31600, Cur Loss: 1.23844123, Cur Avg Loss: 0.60005115, Log Avg loss: 0.57944640, Global Avg Loss: 1.68249162, Time: 0.0067 Steps: 55200, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001985, Sample Num: 31760, Cur Loss: 0.67061156, Cur Avg Loss: 0.60092449, Log Avg loss: 0.77340994, Global Avg Loss: 1.68232696, Time: 0.0107 Steps: 55210, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 001995, Sample Num: 31920, Cur Loss: 0.90929717, Cur Avg Loss: 0.60074972, Log Avg loss: 0.56605612, Global Avg Loss: 1.68212481, Time: 0.0128 Steps: 55220, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002005, Sample Num: 32080, Cur Loss: 0.34304518, Cur Avg Loss: 0.60094523, Log Avg loss: 0.63994967, Global Avg Loss: 1.68193611, Time: 0.0120 Steps: 55230, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002015, Sample Num: 32240, Cur Loss: 0.98684680, Cur Avg Loss: 0.60060664, Log Avg loss: 0.53272020, Global Avg Loss: 1.68172807, Time: 0.0069 Steps: 55240, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002025, Sample Num: 32400, Cur Loss: 0.34110618, Cur Avg Loss: 0.60060758, Log Avg loss: 0.60079628, Global Avg Loss: 1.68153243, Time: 0.0156 Steps: 55250, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002035, Sample Num: 32560, Cur Loss: 0.86650378, Cur Avg Loss: 0.60099940, Log Avg loss: 0.68034261, Global Avg Loss: 1.68135125, Time: 0.0078 Steps: 55260, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002045, Sample Num: 32720, Cur Loss: 0.48195320, Cur Avg Loss: 0.60106600, Log Avg loss: 0.61462048, Global Avg Loss: 1.68115825, Time: 0.0149 Steps: 55270, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002055, Sample Num: 32880, Cur Loss: 0.34479639, Cur Avg Loss: 0.60103648, Log Avg loss: 0.59499941, Global Avg Loss: 1.68096176, Time: 0.0194 Steps: 55280, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002065, Sample Num: 33040, Cur Loss: 0.32624984, Cur Avg Loss: 0.60059167, Log Avg loss: 0.50918270, Global Avg Loss: 1.68074983, Time: 0.0104 Steps: 55290, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002075, Sample Num: 33200, Cur Loss: 0.68418533, Cur Avg Loss: 0.60002154, Log Avg loss: 0.48229013, Global Avg Loss: 1.68053311, Time: 0.0107 Steps: 55300, Updated lr: 0.000049 Training, Epoch: 0026, Batch: 002085, Sample Num: 33360, Cur Loss: 0.32670432, Cur Avg Loss: 0.59946107, Log Avg loss: 0.48316407, Global Avg Loss: 1.68031663, Time: 0.0107 Steps: 55310, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002095, Sample Num: 33520, Cur Loss: 0.40560114, Cur Avg Loss: 0.59913906, Log Avg loss: 0.53199847, Global Avg Loss: 1.68010905, Time: 0.0066 Steps: 55320, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002105, Sample Num: 33680, Cur Loss: 0.11344017, Cur Avg Loss: 0.59917629, Log Avg loss: 0.60697670, Global Avg Loss: 1.67991510, Time: 0.0087 Steps: 55330, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002115, Sample Num: 33840, Cur Loss: 0.59398901, Cur Avg Loss: 0.59925144, Log Avg loss: 0.61506952, Global Avg Loss: 1.67972268, Time: 0.0066 Steps: 55340, Updated lr: 0.000048 Training, Epoch: 0026, Batch: 002125, Sample Num: 34000, Cur Loss: 1.04158628, Cur Avg Loss: 0.59886301, Log Avg loss: 0.51671053, Global Avg Loss: 1.67951256, Time: 0.0118 Steps: 55350, Updated lr: 0.000048 ***** Running evaluation checkpoint-55354 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-55354 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.316186, Avg time per batch (s): 0.010000 {"eval_avg_loss": 1.029872, "eval_total_loss": 724.000264, "eval_mae": 0.875253, "eval_mse": 1.029833, "eval_r2": 0.34537, "eval_sp_statistic": 0.689741, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.753825, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 1.188142, "test_total_loss": 596.447369, "test_mae": 0.948661, "test_mse": 1.188133, "test_r2": 0.233169, "test_sp_statistic": 0.53785, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.634512, "test_ps_pvalue": 0.0, "lr": 4.845519203413941e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6794347912892842, "train_cur_epoch_loss": 1274.9972067400813, "train_cur_epoch_avg_loss": 0.5988713981869804, "train_cur_epoch_time": 22.31618595123291, "train_cur_epoch_avg_time": 0.01048200373472659, "epoch": 26, "step": 55354} ################################################## Training, Epoch: 0027, Batch: 000006, Sample Num: 96, Cur Loss: 0.51790440, Cur Avg Loss: 0.48429984, Log Avg loss: 0.53191148, Global Avg Loss: 1.67930526, Time: 0.0112 Steps: 55360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000016, Sample Num: 256, Cur Loss: 0.54454166, Cur Avg Loss: 0.52040607, Log Avg loss: 0.54206981, Global Avg Loss: 1.67909987, Time: 0.0071 Steps: 55370, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000026, Sample Num: 416, Cur Loss: 1.25388789, Cur Avg Loss: 0.55922051, Log Avg loss: 0.62132361, Global Avg Loss: 1.67890887, Time: 0.0111 Steps: 55380, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000036, Sample Num: 576, Cur Loss: 0.43333828, Cur Avg Loss: 0.53881782, Log Avg loss: 0.48577082, Global Avg Loss: 1.67869346, Time: 0.0117 Steps: 55390, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000046, Sample Num: 736, Cur Loss: 0.56954539, Cur Avg Loss: 0.55385670, Log Avg loss: 0.60799668, Global Avg Loss: 1.67850020, Time: 0.0101 Steps: 55400, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000056, Sample Num: 896, Cur Loss: 1.07883263, Cur Avg Loss: 0.57246484, Log Avg loss: 0.65806230, Global Avg Loss: 1.67831603, Time: 0.0072 Steps: 55410, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000066, Sample Num: 1056, Cur Loss: 0.67660540, Cur Avg Loss: 0.56696066, Log Avg loss: 0.53613725, Global Avg Loss: 1.67810994, Time: 0.0089 Steps: 55420, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000076, Sample Num: 1216, Cur Loss: 0.86054230, Cur Avg Loss: 0.57243165, Log Avg loss: 0.60854017, Global Avg Loss: 1.67791698, Time: 0.0066 Steps: 55430, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000086, Sample Num: 1376, Cur Loss: 0.37632990, Cur Avg Loss: 0.58680730, Log Avg loss: 0.69606223, Global Avg Loss: 1.67773988, Time: 0.0067 Steps: 55440, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000096, Sample Num: 1536, Cur Loss: 0.49943870, Cur Avg Loss: 0.59236821, Log Avg loss: 0.64019205, Global Avg Loss: 1.67755276, Time: 0.0072 Steps: 55450, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000106, Sample Num: 1696, Cur Loss: 0.41771820, Cur Avg Loss: 0.59573860, Log Avg loss: 0.62809433, Global Avg Loss: 1.67736354, Time: 0.0097 Steps: 55460, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000116, Sample Num: 1856, Cur Loss: 0.46468163, Cur Avg Loss: 0.59433746, Log Avg loss: 0.57948541, Global Avg Loss: 1.67716561, Time: 0.0068 Steps: 55470, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000126, Sample Num: 2016, Cur Loss: 0.36159006, Cur Avg Loss: 0.58271379, Log Avg loss: 0.44787920, Global Avg Loss: 1.67694404, Time: 0.0135 Steps: 55480, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000136, Sample Num: 2176, Cur Loss: 0.54453981, Cur Avg Loss: 0.58371533, Log Avg loss: 0.59633470, Global Avg Loss: 1.67674930, Time: 0.0108 Steps: 55490, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000146, Sample Num: 2336, Cur Loss: 0.62073535, Cur Avg Loss: 0.58060381, Log Avg loss: 0.53828711, Global Avg Loss: 1.67654417, Time: 0.0065 Steps: 55500, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000156, Sample Num: 2496, Cur Loss: 0.30817065, Cur Avg Loss: 0.57568720, Log Avg loss: 0.50390481, Global Avg Loss: 1.67633292, Time: 0.0117 Steps: 55510, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000166, Sample Num: 2656, Cur Loss: 0.29559153, Cur Avg Loss: 0.57997181, Log Avg loss: 0.64681167, Global Avg Loss: 1.67614749, Time: 0.0116 Steps: 55520, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000176, Sample Num: 2816, Cur Loss: 0.37547556, Cur Avg Loss: 0.57191414, Log Avg loss: 0.43815687, Global Avg Loss: 1.67592455, Time: 0.0135 Steps: 55530, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000186, Sample Num: 2976, Cur Loss: 0.19665480, Cur Avg Loss: 0.57628960, Log Avg loss: 0.65329758, Global Avg Loss: 1.67574043, Time: 0.0132 Steps: 55540, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000196, Sample Num: 3136, Cur Loss: 0.40568465, Cur Avg Loss: 0.57702568, Log Avg loss: 0.59071679, Global Avg Loss: 1.67554510, Time: 0.0153 Steps: 55550, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000206, Sample Num: 3296, Cur Loss: 0.09526587, Cur Avg Loss: 0.57253874, Log Avg loss: 0.48459481, Global Avg Loss: 1.67533075, Time: 0.0064 Steps: 55560, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000216, Sample Num: 3456, Cur Loss: 1.79893887, Cur Avg Loss: 0.58049933, Log Avg loss: 0.74448738, Global Avg Loss: 1.67516324, Time: 0.0064 Steps: 55570, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000226, Sample Num: 3616, Cur Loss: 0.13987505, Cur Avg Loss: 0.57531797, Log Avg loss: 0.46340063, Global Avg Loss: 1.67494522, Time: 0.0065 Steps: 55580, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000236, Sample Num: 3776, Cur Loss: 0.28240728, Cur Avg Loss: 0.57615844, Log Avg loss: 0.59515300, Global Avg Loss: 1.67475098, Time: 0.0107 Steps: 55590, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000246, Sample Num: 3936, Cur Loss: 0.61467212, Cur Avg Loss: 0.58126364, Log Avg loss: 0.70174640, Global Avg Loss: 1.67457598, Time: 0.0086 Steps: 55600, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000256, Sample Num: 4096, Cur Loss: 0.28560024, Cur Avg Loss: 0.58569238, Log Avg loss: 0.69463945, Global Avg Loss: 1.67439976, Time: 0.0085 Steps: 55610, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000266, Sample Num: 4256, Cur Loss: 0.66260862, Cur Avg Loss: 0.58843930, Log Avg loss: 0.65876039, Global Avg Loss: 1.67421716, Time: 0.0098 Steps: 55620, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000276, Sample Num: 4416, Cur Loss: 0.41505456, Cur Avg Loss: 0.58341550, Log Avg loss: 0.44978238, Global Avg Loss: 1.67399705, Time: 0.0063 Steps: 55630, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000286, Sample Num: 4576, Cur Loss: 0.62937421, Cur Avg Loss: 0.58372156, Log Avg loss: 0.59216877, Global Avg Loss: 1.67380262, Time: 0.0095 Steps: 55640, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000296, Sample Num: 4736, Cur Loss: 0.92394680, Cur Avg Loss: 0.59164387, Log Avg loss: 0.81822209, Global Avg Loss: 1.67364888, Time: 0.0064 Steps: 55650, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000306, Sample Num: 4896, Cur Loss: 0.94182295, Cur Avg Loss: 0.59287001, Log Avg loss: 0.62916377, Global Avg Loss: 1.67346122, Time: 0.0076 Steps: 55660, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000316, Sample Num: 5056, Cur Loss: 0.74293602, Cur Avg Loss: 0.59230659, Log Avg loss: 0.57506601, Global Avg Loss: 1.67326392, Time: 0.0224 Steps: 55670, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000326, Sample Num: 5216, Cur Loss: 0.36358666, Cur Avg Loss: 0.58924672, Log Avg loss: 0.49255464, Global Avg Loss: 1.67305187, Time: 0.0075 Steps: 55680, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000336, Sample Num: 5376, Cur Loss: 0.41716337, Cur Avg Loss: 0.58750174, Log Avg loss: 0.53061532, Global Avg Loss: 1.67284672, Time: 0.0210 Steps: 55690, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000346, Sample Num: 5536, Cur Loss: 1.45000982, Cur Avg Loss: 0.58945066, Log Avg loss: 0.65493448, Global Avg Loss: 1.67266397, Time: 0.0078 Steps: 55700, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000356, Sample Num: 5696, Cur Loss: 0.67400110, Cur Avg Loss: 0.58759403, Log Avg loss: 0.52335449, Global Avg Loss: 1.67245767, Time: 0.0066 Steps: 55710, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000366, Sample Num: 5856, Cur Loss: 0.37012365, Cur Avg Loss: 0.58620740, Log Avg loss: 0.53684365, Global Avg Loss: 1.67225386, Time: 0.0065 Steps: 55720, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000376, Sample Num: 6016, Cur Loss: 1.95153630, Cur Avg Loss: 0.59016011, Log Avg loss: 0.73482931, Global Avg Loss: 1.67208566, Time: 0.0067 Steps: 55730, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000386, Sample Num: 6176, Cur Loss: 0.67767656, Cur Avg Loss: 0.58710065, Log Avg loss: 0.47206472, Global Avg Loss: 1.67187037, Time: 0.0106 Steps: 55740, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000396, Sample Num: 6336, Cur Loss: 0.42859048, Cur Avg Loss: 0.58732634, Log Avg loss: 0.59603812, Global Avg Loss: 1.67167739, Time: 0.0064 Steps: 55750, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000406, Sample Num: 6496, Cur Loss: 0.78472745, Cur Avg Loss: 0.58467035, Log Avg loss: 0.47949314, Global Avg Loss: 1.67146359, Time: 0.0152 Steps: 55760, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000416, Sample Num: 6656, Cur Loss: 1.51993859, Cur Avg Loss: 0.58374057, Log Avg loss: 0.54599133, Global Avg Loss: 1.67126178, Time: 0.0066 Steps: 55770, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000426, Sample Num: 6816, Cur Loss: 0.24190293, Cur Avg Loss: 0.58239538, Log Avg loss: 0.52643545, Global Avg Loss: 1.67105654, Time: 0.0067 Steps: 55780, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000436, Sample Num: 6976, Cur Loss: 0.24825647, Cur Avg Loss: 0.58729812, Log Avg loss: 0.79615517, Global Avg Loss: 1.67089972, Time: 0.0114 Steps: 55790, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000446, Sample Num: 7136, Cur Loss: 1.07907903, Cur Avg Loss: 0.58704243, Log Avg loss: 0.57589413, Global Avg Loss: 1.67070348, Time: 0.0066 Steps: 55800, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000456, Sample Num: 7296, Cur Loss: 0.36259773, Cur Avg Loss: 0.58810391, Log Avg loss: 0.63544592, Global Avg Loss: 1.67051799, Time: 0.0114 Steps: 55810, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000466, Sample Num: 7456, Cur Loss: 0.47922248, Cur Avg Loss: 0.59061078, Log Avg loss: 0.70492407, Global Avg Loss: 1.67034500, Time: 0.0089 Steps: 55820, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000476, Sample Num: 7616, Cur Loss: 0.64157939, Cur Avg Loss: 0.59249307, Log Avg loss: 0.68020768, Global Avg Loss: 1.67016765, Time: 0.0082 Steps: 55830, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000486, Sample Num: 7776, Cur Loss: 0.19470058, Cur Avg Loss: 0.59039187, Log Avg loss: 0.49037506, Global Avg Loss: 1.66995637, Time: 0.0066 Steps: 55840, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000496, Sample Num: 7936, Cur Loss: 0.57379627, Cur Avg Loss: 0.59297166, Log Avg loss: 0.71834936, Global Avg Loss: 1.66978599, Time: 0.0097 Steps: 55850, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000506, Sample Num: 8096, Cur Loss: 0.28447986, Cur Avg Loss: 0.59594340, Log Avg loss: 0.74334165, Global Avg Loss: 1.66962014, Time: 0.0120 Steps: 55860, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000516, Sample Num: 8256, Cur Loss: 0.35717046, Cur Avg Loss: 0.59444190, Log Avg loss: 0.51846585, Global Avg Loss: 1.66941409, Time: 0.0109 Steps: 55870, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000526, Sample Num: 8416, Cur Loss: 1.17771626, Cur Avg Loss: 0.59584757, Log Avg loss: 0.66838036, Global Avg Loss: 1.66923495, Time: 0.0077 Steps: 55880, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000536, Sample Num: 8576, Cur Loss: 0.34895858, Cur Avg Loss: 0.59519619, Log Avg loss: 0.56093344, Global Avg Loss: 1.66903665, Time: 0.0077 Steps: 55890, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000546, Sample Num: 8736, Cur Loss: 0.24408995, Cur Avg Loss: 0.59572542, Log Avg loss: 0.62409204, Global Avg Loss: 1.66884972, Time: 0.0072 Steps: 55900, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000556, Sample Num: 8896, Cur Loss: 0.13998327, Cur Avg Loss: 0.59234707, Log Avg loss: 0.40788925, Global Avg Loss: 1.66862419, Time: 0.0219 Steps: 55910, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000566, Sample Num: 9056, Cur Loss: 0.69827068, Cur Avg Loss: 0.59333637, Log Avg loss: 0.64834168, Global Avg Loss: 1.66844174, Time: 0.0066 Steps: 55920, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000576, Sample Num: 9216, Cur Loss: 0.52094895, Cur Avg Loss: 0.59349903, Log Avg loss: 0.60270555, Global Avg Loss: 1.66825119, Time: 0.0116 Steps: 55930, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000586, Sample Num: 9376, Cur Loss: 0.24985024, Cur Avg Loss: 0.59118853, Log Avg loss: 0.45810349, Global Avg Loss: 1.66803486, Time: 0.0064 Steps: 55940, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000596, Sample Num: 9536, Cur Loss: 1.00929296, Cur Avg Loss: 0.59138346, Log Avg loss: 0.60280644, Global Avg Loss: 1.66784447, Time: 0.0110 Steps: 55950, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000606, Sample Num: 9696, Cur Loss: 0.21922322, Cur Avg Loss: 0.59052734, Log Avg loss: 0.53950271, Global Avg Loss: 1.66764283, Time: 0.0067 Steps: 55960, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000616, Sample Num: 9856, Cur Loss: 0.32023519, Cur Avg Loss: 0.58912672, Log Avg loss: 0.50424915, Global Avg Loss: 1.66743497, Time: 0.0080 Steps: 55970, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000626, Sample Num: 10016, Cur Loss: 0.56709498, Cur Avg Loss: 0.58765672, Log Avg loss: 0.49710440, Global Avg Loss: 1.66722591, Time: 0.0086 Steps: 55980, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000636, Sample Num: 10176, Cur Loss: 0.77800417, Cur Avg Loss: 0.58945360, Log Avg loss: 0.70193840, Global Avg Loss: 1.66705351, Time: 0.0110 Steps: 55990, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000646, Sample Num: 10336, Cur Loss: 0.18441454, Cur Avg Loss: 0.58800625, Log Avg loss: 0.49595490, Global Avg Loss: 1.66684438, Time: 0.0067 Steps: 56000, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000656, Sample Num: 10496, Cur Loss: 0.38604021, Cur Avg Loss: 0.58671646, Log Avg loss: 0.50339606, Global Avg Loss: 1.66663666, Time: 0.0114 Steps: 56010, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000666, Sample Num: 10656, Cur Loss: 0.77607405, Cur Avg Loss: 0.58499049, Log Avg loss: 0.47176685, Global Avg Loss: 1.66642337, Time: 0.0132 Steps: 56020, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000676, Sample Num: 10816, Cur Loss: 0.57971716, Cur Avg Loss: 0.58531419, Log Avg loss: 0.60687237, Global Avg Loss: 1.66623426, Time: 0.0068 Steps: 56030, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000686, Sample Num: 10976, Cur Loss: 0.23538059, Cur Avg Loss: 0.58347903, Log Avg loss: 0.45942248, Global Avg Loss: 1.66601892, Time: 0.0199 Steps: 56040, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000696, Sample Num: 11136, Cur Loss: 1.06280589, Cur Avg Loss: 0.58242716, Log Avg loss: 0.51026918, Global Avg Loss: 1.66581272, Time: 0.0082 Steps: 56050, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000706, Sample Num: 11296, Cur Loss: 1.88901222, Cur Avg Loss: 0.58398985, Log Avg loss: 0.69275298, Global Avg Loss: 1.66563914, Time: 0.0080 Steps: 56060, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000716, Sample Num: 11456, Cur Loss: 0.92525685, Cur Avg Loss: 0.58409988, Log Avg loss: 0.59186797, Global Avg Loss: 1.66544764, Time: 0.0067 Steps: 56070, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000726, Sample Num: 11616, Cur Loss: 0.22191933, Cur Avg Loss: 0.58305978, Log Avg loss: 0.50858863, Global Avg Loss: 1.66524135, Time: 0.0066 Steps: 56080, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000736, Sample Num: 11776, Cur Loss: 0.26289570, Cur Avg Loss: 0.58372829, Log Avg loss: 0.63226182, Global Avg Loss: 1.66505718, Time: 0.0128 Steps: 56090, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000746, Sample Num: 11936, Cur Loss: 0.38849503, Cur Avg Loss: 0.58250966, Log Avg loss: 0.49281843, Global Avg Loss: 1.66484823, Time: 0.0069 Steps: 56100, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000756, Sample Num: 12096, Cur Loss: 0.45144439, Cur Avg Loss: 0.58017377, Log Avg loss: 0.40591660, Global Avg Loss: 1.66462386, Time: 0.0069 Steps: 56110, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000766, Sample Num: 12256, Cur Loss: 0.66270727, Cur Avg Loss: 0.58002409, Log Avg loss: 0.56870798, Global Avg Loss: 1.66442858, Time: 0.0067 Steps: 56120, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000776, Sample Num: 12416, Cur Loss: 0.36625415, Cur Avg Loss: 0.58134646, Log Avg loss: 0.68264060, Global Avg Loss: 1.66425367, Time: 0.0124 Steps: 56130, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000786, Sample Num: 12576, Cur Loss: 0.69285309, Cur Avg Loss: 0.58150092, Log Avg loss: 0.59348673, Global Avg Loss: 1.66406293, Time: 0.0134 Steps: 56140, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000796, Sample Num: 12736, Cur Loss: 0.30474705, Cur Avg Loss: 0.58287475, Log Avg loss: 0.69085793, Global Avg Loss: 1.66388961, Time: 0.0158 Steps: 56150, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000806, Sample Num: 12896, Cur Loss: 0.53495747, Cur Avg Loss: 0.58289032, Log Avg loss: 0.58412917, Global Avg Loss: 1.66369735, Time: 0.0064 Steps: 56160, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000816, Sample Num: 13056, Cur Loss: 0.49041104, Cur Avg Loss: 0.58352446, Log Avg loss: 0.63463665, Global Avg Loss: 1.66351414, Time: 0.0064 Steps: 56170, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000826, Sample Num: 13216, Cur Loss: 0.44039220, Cur Avg Loss: 0.58532486, Log Avg loss: 0.73223720, Global Avg Loss: 1.66334838, Time: 0.0065 Steps: 56180, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000836, Sample Num: 13376, Cur Loss: 0.29503918, Cur Avg Loss: 0.58589060, Log Avg loss: 0.63262112, Global Avg Loss: 1.66316494, Time: 0.0184 Steps: 56190, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000846, Sample Num: 13536, Cur Loss: 0.64823246, Cur Avg Loss: 0.58773302, Log Avg loss: 0.74175892, Global Avg Loss: 1.66300099, Time: 0.0069 Steps: 56200, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000856, Sample Num: 13696, Cur Loss: 0.27661204, Cur Avg Loss: 0.58741468, Log Avg loss: 0.56048356, Global Avg Loss: 1.66280485, Time: 0.0064 Steps: 56210, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000866, Sample Num: 13856, Cur Loss: 1.27662718, Cur Avg Loss: 0.58845263, Log Avg loss: 0.67730106, Global Avg Loss: 1.66262955, Time: 0.0066 Steps: 56220, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000876, Sample Num: 14016, Cur Loss: 0.81101090, Cur Avg Loss: 0.58706700, Log Avg loss: 0.46707152, Global Avg Loss: 1.66241693, Time: 0.0071 Steps: 56230, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000886, Sample Num: 14176, Cur Loss: 0.62964225, Cur Avg Loss: 0.58794803, Log Avg loss: 0.66512577, Global Avg Loss: 1.66223961, Time: 0.0064 Steps: 56240, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000896, Sample Num: 14336, Cur Loss: 0.30320466, Cur Avg Loss: 0.58831910, Log Avg loss: 0.62119649, Global Avg Loss: 1.66205453, Time: 0.0066 Steps: 56250, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000906, Sample Num: 14496, Cur Loss: 0.53293610, Cur Avg Loss: 0.58732773, Log Avg loss: 0.49850028, Global Avg Loss: 1.66184771, Time: 0.0064 Steps: 56260, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000916, Sample Num: 14656, Cur Loss: 0.50235528, Cur Avg Loss: 0.58819279, Log Avg loss: 0.66656731, Global Avg Loss: 1.66167084, Time: 0.0071 Steps: 56270, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000926, Sample Num: 14816, Cur Loss: 0.49843001, Cur Avg Loss: 0.58853739, Log Avg loss: 0.62010254, Global Avg Loss: 1.66148577, Time: 0.0064 Steps: 56280, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000936, Sample Num: 14976, Cur Loss: 0.12464910, Cur Avg Loss: 0.58772091, Log Avg loss: 0.51211550, Global Avg Loss: 1.66128158, Time: 0.0066 Steps: 56290, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000946, Sample Num: 15136, Cur Loss: 0.72259349, Cur Avg Loss: 0.59010353, Log Avg loss: 0.81311688, Global Avg Loss: 1.66113093, Time: 0.0065 Steps: 56300, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000956, Sample Num: 15296, Cur Loss: 1.02185607, Cur Avg Loss: 0.59082946, Log Avg loss: 0.65950248, Global Avg Loss: 1.66095305, Time: 0.0066 Steps: 56310, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000966, Sample Num: 15456, Cur Loss: 0.35890228, Cur Avg Loss: 0.59099084, Log Avg loss: 0.60641866, Global Avg Loss: 1.66076581, Time: 0.0068 Steps: 56320, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000976, Sample Num: 15616, Cur Loss: 0.67369968, Cur Avg Loss: 0.59061858, Log Avg loss: 0.55465783, Global Avg Loss: 1.66056945, Time: 0.0097 Steps: 56330, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000986, Sample Num: 15776, Cur Loss: 0.71541470, Cur Avg Loss: 0.59333516, Log Avg loss: 0.85847382, Global Avg Loss: 1.66042708, Time: 0.0225 Steps: 56340, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 000996, Sample Num: 15936, Cur Loss: 0.44259056, Cur Avg Loss: 0.59374668, Log Avg loss: 0.63432181, Global Avg Loss: 1.66024499, Time: 0.0064 Steps: 56350, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001006, Sample Num: 16096, Cur Loss: 0.84285235, Cur Avg Loss: 0.59394907, Log Avg loss: 0.61410756, Global Avg Loss: 1.66005937, Time: 0.0064 Steps: 56360, Updated lr: 0.000048 Training, Epoch: 0027, Batch: 001016, Sample Num: 16256, Cur Loss: 0.39307261, Cur Avg Loss: 0.59236112, Log Avg loss: 0.43261328, Global Avg Loss: 1.65984162, Time: 0.0090 Steps: 56370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001026, Sample Num: 16416, Cur Loss: 0.40818080, Cur Avg Loss: 0.59048705, Log Avg loss: 0.40008156, Global Avg Loss: 1.65961818, Time: 0.0110 Steps: 56380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001036, Sample Num: 16576, Cur Loss: 0.83262813, Cur Avg Loss: 0.59155166, Log Avg loss: 0.70078020, Global Avg Loss: 1.65944815, Time: 0.0066 Steps: 56390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001046, Sample Num: 16736, Cur Loss: 0.12972848, Cur Avg Loss: 0.59168901, Log Avg loss: 0.60591927, Global Avg Loss: 1.65926135, Time: 0.0112 Steps: 56400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001056, Sample Num: 16896, Cur Loss: 0.24718028, Cur Avg Loss: 0.59113653, Log Avg loss: 0.53334699, Global Avg Loss: 1.65906176, Time: 0.0112 Steps: 56410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001066, Sample Num: 17056, Cur Loss: 0.27824444, Cur Avg Loss: 0.58925723, Log Avg loss: 0.39080305, Global Avg Loss: 1.65883697, Time: 0.0068 Steps: 56420, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001076, Sample Num: 17216, Cur Loss: 0.62122536, Cur Avg Loss: 0.58983388, Log Avg loss: 0.65130481, Global Avg Loss: 1.65865842, Time: 0.0114 Steps: 56430, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001086, Sample Num: 17376, Cur Loss: 0.54242647, Cur Avg Loss: 0.58911907, Log Avg loss: 0.51220568, Global Avg Loss: 1.65845529, Time: 0.0070 Steps: 56440, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001096, Sample Num: 17536, Cur Loss: 0.32434022, Cur Avg Loss: 0.58803936, Log Avg loss: 0.47078217, Global Avg Loss: 1.65824490, Time: 0.0135 Steps: 56450, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001106, Sample Num: 17696, Cur Loss: 0.43247104, Cur Avg Loss: 0.58780872, Log Avg loss: 0.56253143, Global Avg Loss: 1.65805083, Time: 0.0118 Steps: 56460, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001116, Sample Num: 17856, Cur Loss: 0.34572956, Cur Avg Loss: 0.58779242, Log Avg loss: 0.58598887, Global Avg Loss: 1.65786098, Time: 0.0073 Steps: 56470, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001126, Sample Num: 18016, Cur Loss: 0.69317281, Cur Avg Loss: 0.58838170, Log Avg loss: 0.65414529, Global Avg Loss: 1.65768327, Time: 0.0069 Steps: 56480, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001136, Sample Num: 18176, Cur Loss: 0.51721925, Cur Avg Loss: 0.58866660, Log Avg loss: 0.62074667, Global Avg Loss: 1.65749971, Time: 0.0117 Steps: 56490, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001146, Sample Num: 18336, Cur Loss: 0.34455901, Cur Avg Loss: 0.58908496, Log Avg loss: 0.63661101, Global Avg Loss: 1.65731902, Time: 0.0073 Steps: 56500, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001156, Sample Num: 18496, Cur Loss: 0.55631882, Cur Avg Loss: 0.58860361, Log Avg loss: 0.53344033, Global Avg Loss: 1.65712014, Time: 0.0067 Steps: 56510, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001166, Sample Num: 18656, Cur Loss: 0.37022638, Cur Avg Loss: 0.58858146, Log Avg loss: 0.58602144, Global Avg Loss: 1.65693063, Time: 0.0110 Steps: 56520, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001176, Sample Num: 18816, Cur Loss: 0.33284864, Cur Avg Loss: 0.58953351, Log Avg loss: 0.70054196, Global Avg Loss: 1.65676145, Time: 0.0111 Steps: 56530, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001186, Sample Num: 18976, Cur Loss: 0.37032452, Cur Avg Loss: 0.59025618, Log Avg loss: 0.67524295, Global Avg Loss: 1.65658785, Time: 0.0136 Steps: 56540, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001196, Sample Num: 19136, Cur Loss: 0.23368995, Cur Avg Loss: 0.58996189, Log Avg loss: 0.55505825, Global Avg Loss: 1.65639307, Time: 0.0065 Steps: 56550, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001206, Sample Num: 19296, Cur Loss: 0.26580054, Cur Avg Loss: 0.58892409, Log Avg loss: 0.46480359, Global Avg Loss: 1.65618239, Time: 0.0065 Steps: 56560, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001216, Sample Num: 19456, Cur Loss: 0.40195537, Cur Avg Loss: 0.58925795, Log Avg loss: 0.62952141, Global Avg Loss: 1.65600090, Time: 0.0069 Steps: 56570, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001226, Sample Num: 19616, Cur Loss: 0.78546071, Cur Avg Loss: 0.59050136, Log Avg loss: 0.74169970, Global Avg Loss: 1.65583931, Time: 0.0224 Steps: 56580, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001236, Sample Num: 19776, Cur Loss: 0.41356575, Cur Avg Loss: 0.59047566, Log Avg loss: 0.58732572, Global Avg Loss: 1.65565049, Time: 0.0081 Steps: 56590, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001246, Sample Num: 19936, Cur Loss: 0.25885549, Cur Avg Loss: 0.59103293, Log Avg loss: 0.65991124, Global Avg Loss: 1.65547457, Time: 0.0112 Steps: 56600, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001256, Sample Num: 20096, Cur Loss: 0.85274404, Cur Avg Loss: 0.59139438, Log Avg loss: 0.63643132, Global Avg Loss: 1.65529456, Time: 0.0134 Steps: 56610, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001266, Sample Num: 20256, Cur Loss: 0.35978734, Cur Avg Loss: 0.59234127, Log Avg loss: 0.71127006, Global Avg Loss: 1.65512783, Time: 0.0112 Steps: 56620, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001276, Sample Num: 20416, Cur Loss: 0.88858318, Cur Avg Loss: 0.59183954, Log Avg loss: 0.52831999, Global Avg Loss: 1.65492885, Time: 0.0119 Steps: 56630, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001286, Sample Num: 20576, Cur Loss: 0.61271721, Cur Avg Loss: 0.59195443, Log Avg loss: 0.60661532, Global Avg Loss: 1.65474376, Time: 0.0095 Steps: 56640, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001296, Sample Num: 20736, Cur Loss: 0.27355629, Cur Avg Loss: 0.59184979, Log Avg loss: 0.57839269, Global Avg Loss: 1.65455376, Time: 0.0067 Steps: 56650, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001306, Sample Num: 20896, Cur Loss: 0.19475128, Cur Avg Loss: 0.59091813, Log Avg loss: 0.47017526, Global Avg Loss: 1.65434473, Time: 0.0068 Steps: 56660, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001316, Sample Num: 21056, Cur Loss: 1.25410783, Cur Avg Loss: 0.59146734, Log Avg loss: 0.66319384, Global Avg Loss: 1.65416983, Time: 0.0073 Steps: 56670, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001326, Sample Num: 21216, Cur Loss: 0.52808738, Cur Avg Loss: 0.59226028, Log Avg loss: 0.69661199, Global Avg Loss: 1.65400089, Time: 0.0152 Steps: 56680, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001336, Sample Num: 21376, Cur Loss: 1.65033984, Cur Avg Loss: 0.59171334, Log Avg loss: 0.51918822, Global Avg Loss: 1.65380071, Time: 0.0107 Steps: 56690, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001346, Sample Num: 21536, Cur Loss: 0.30209360, Cur Avg Loss: 0.59146138, Log Avg loss: 0.55779994, Global Avg Loss: 1.65360742, Time: 0.0126 Steps: 56700, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001356, Sample Num: 21696, Cur Loss: 0.23470762, Cur Avg Loss: 0.59100366, Log Avg loss: 0.52939484, Global Avg Loss: 1.65340918, Time: 0.0108 Steps: 56710, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001366, Sample Num: 21856, Cur Loss: 0.45630044, Cur Avg Loss: 0.58996841, Log Avg loss: 0.44958794, Global Avg Loss: 1.65319694, Time: 0.0106 Steps: 56720, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001376, Sample Num: 22016, Cur Loss: 0.49116462, Cur Avg Loss: 0.59062959, Log Avg loss: 0.68094704, Global Avg Loss: 1.65302556, Time: 0.0066 Steps: 56730, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001386, Sample Num: 22176, Cur Loss: 0.52052152, Cur Avg Loss: 0.59043810, Log Avg loss: 0.56408931, Global Avg Loss: 1.65283364, Time: 0.0066 Steps: 56740, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001396, Sample Num: 22336, Cur Loss: 0.47675359, Cur Avg Loss: 0.59146240, Log Avg loss: 0.73342931, Global Avg Loss: 1.65267163, Time: 0.0127 Steps: 56750, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001406, Sample Num: 22496, Cur Loss: 0.68875992, Cur Avg Loss: 0.59115829, Log Avg loss: 0.54870490, Global Avg Loss: 1.65247713, Time: 0.0067 Steps: 56760, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001416, Sample Num: 22656, Cur Loss: 0.65196204, Cur Avg Loss: 0.59167203, Log Avg loss: 0.66390482, Global Avg Loss: 1.65230300, Time: 0.0070 Steps: 56770, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001426, Sample Num: 22816, Cur Loss: 0.20406766, Cur Avg Loss: 0.59141199, Log Avg loss: 0.55458962, Global Avg Loss: 1.65210967, Time: 0.0180 Steps: 56780, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001436, Sample Num: 22976, Cur Loss: 0.45766824, Cur Avg Loss: 0.59053990, Log Avg loss: 0.46617965, Global Avg Loss: 1.65190084, Time: 0.0066 Steps: 56790, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001446, Sample Num: 23136, Cur Loss: 0.94858617, Cur Avg Loss: 0.59163803, Log Avg loss: 0.74932936, Global Avg Loss: 1.65174194, Time: 0.0072 Steps: 56800, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001456, Sample Num: 23296, Cur Loss: 0.31038660, Cur Avg Loss: 0.59152247, Log Avg loss: 0.57481322, Global Avg Loss: 1.65155237, Time: 0.0065 Steps: 56810, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001466, Sample Num: 23456, Cur Loss: 0.47977927, Cur Avg Loss: 0.59001862, Log Avg loss: 0.37105784, Global Avg Loss: 1.65132701, Time: 0.0066 Steps: 56820, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001476, Sample Num: 23616, Cur Loss: 0.22469184, Cur Avg Loss: 0.58935060, Log Avg loss: 0.49141811, Global Avg Loss: 1.65112291, Time: 0.0166 Steps: 56830, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001486, Sample Num: 23776, Cur Loss: 0.71596742, Cur Avg Loss: 0.59020396, Log Avg loss: 0.71616067, Global Avg Loss: 1.65095842, Time: 0.0064 Steps: 56840, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001496, Sample Num: 23936, Cur Loss: 1.10766149, Cur Avg Loss: 0.59111916, Log Avg loss: 0.72711746, Global Avg Loss: 1.65079591, Time: 0.0148 Steps: 56850, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001506, Sample Num: 24096, Cur Loss: 1.14513004, Cur Avg Loss: 0.59101225, Log Avg loss: 0.57501806, Global Avg Loss: 1.65060672, Time: 0.0133 Steps: 56860, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001516, Sample Num: 24256, Cur Loss: 0.51391149, Cur Avg Loss: 0.58998445, Log Avg loss: 0.43519857, Global Avg Loss: 1.65039300, Time: 0.0066 Steps: 56870, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001526, Sample Num: 24416, Cur Loss: 0.49024320, Cur Avg Loss: 0.59041371, Log Avg loss: 0.65548887, Global Avg Loss: 1.65021809, Time: 0.0070 Steps: 56880, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001536, Sample Num: 24576, Cur Loss: 1.09714675, Cur Avg Loss: 0.59061496, Log Avg loss: 0.62132601, Global Avg Loss: 1.65003723, Time: 0.0074 Steps: 56890, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001546, Sample Num: 24736, Cur Loss: 0.16320524, Cur Avg Loss: 0.59052658, Log Avg loss: 0.57695112, Global Avg Loss: 1.64984864, Time: 0.0068 Steps: 56900, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001556, Sample Num: 24896, Cur Loss: 0.15525924, Cur Avg Loss: 0.58953605, Log Avg loss: 0.43640028, Global Avg Loss: 1.64963542, Time: 0.0122 Steps: 56910, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001566, Sample Num: 25056, Cur Loss: 0.45749575, Cur Avg Loss: 0.58918489, Log Avg loss: 0.53454498, Global Avg Loss: 1.64943951, Time: 0.0120 Steps: 56920, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001576, Sample Num: 25216, Cur Loss: 0.51537538, Cur Avg Loss: 0.58872101, Log Avg loss: 0.51607761, Global Avg Loss: 1.64924043, Time: 0.0095 Steps: 56930, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001586, Sample Num: 25376, Cur Loss: 1.04649293, Cur Avg Loss: 0.58785398, Log Avg loss: 0.45121015, Global Avg Loss: 1.64903003, Time: 0.0234 Steps: 56940, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001596, Sample Num: 25536, Cur Loss: 0.98099697, Cur Avg Loss: 0.58849172, Log Avg loss: 0.68963628, Global Avg Loss: 1.64886157, Time: 0.0112 Steps: 56950, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001606, Sample Num: 25696, Cur Loss: 1.00725627, Cur Avg Loss: 0.59029837, Log Avg loss: 0.87864004, Global Avg Loss: 1.64872635, Time: 0.0071 Steps: 56960, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001616, Sample Num: 25856, Cur Loss: 0.48987436, Cur Avg Loss: 0.59085522, Log Avg loss: 0.68028621, Global Avg Loss: 1.64855635, Time: 0.0116 Steps: 56970, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001626, Sample Num: 26016, Cur Loss: 0.39298001, Cur Avg Loss: 0.59048184, Log Avg loss: 0.53014274, Global Avg Loss: 1.64836007, Time: 0.0067 Steps: 56980, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001636, Sample Num: 26176, Cur Loss: 1.65550876, Cur Avg Loss: 0.59061102, Log Avg loss: 0.61161607, Global Avg Loss: 1.64817816, Time: 0.0080 Steps: 56990, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001646, Sample Num: 26336, Cur Loss: 0.25644550, Cur Avg Loss: 0.58960741, Log Avg loss: 0.42541745, Global Avg Loss: 1.64796364, Time: 0.0067 Steps: 57000, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001656, Sample Num: 26496, Cur Loss: 0.38440940, Cur Avg Loss: 0.59011800, Log Avg loss: 0.67415963, Global Avg Loss: 1.64779282, Time: 0.0115 Steps: 57010, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001666, Sample Num: 26656, Cur Loss: 0.43746522, Cur Avg Loss: 0.59046063, Log Avg loss: 0.64720123, Global Avg Loss: 1.64761734, Time: 0.0069 Steps: 57020, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001676, Sample Num: 26816, Cur Loss: 0.68139428, Cur Avg Loss: 0.59041042, Log Avg loss: 0.58204458, Global Avg Loss: 1.64743050, Time: 0.0069 Steps: 57030, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001686, Sample Num: 26976, Cur Loss: 0.17779800, Cur Avg Loss: 0.59015302, Log Avg loss: 0.54701395, Global Avg Loss: 1.64723758, Time: 0.0068 Steps: 57040, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001696, Sample Num: 27136, Cur Loss: 0.60617769, Cur Avg Loss: 0.59057498, Log Avg loss: 0.66171717, Global Avg Loss: 1.64706483, Time: 0.0088 Steps: 57050, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001706, Sample Num: 27296, Cur Loss: 0.50257623, Cur Avg Loss: 0.59017339, Log Avg loss: 0.52206402, Global Avg Loss: 1.64686767, Time: 0.0085 Steps: 57060, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001716, Sample Num: 27456, Cur Loss: 0.75723845, Cur Avg Loss: 0.59094422, Log Avg loss: 0.72244775, Global Avg Loss: 1.64670569, Time: 0.0120 Steps: 57070, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001726, Sample Num: 27616, Cur Loss: 0.49575540, Cur Avg Loss: 0.58981999, Log Avg loss: 0.39690117, Global Avg Loss: 1.64648673, Time: 0.0110 Steps: 57080, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001736, Sample Num: 27776, Cur Loss: 0.79463756, Cur Avg Loss: 0.59014948, Log Avg loss: 0.64701971, Global Avg Loss: 1.64631166, Time: 0.0069 Steps: 57090, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001746, Sample Num: 27936, Cur Loss: 0.18283392, Cur Avg Loss: 0.59063390, Log Avg loss: 0.67472928, Global Avg Loss: 1.64614151, Time: 0.0119 Steps: 57100, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001756, Sample Num: 28096, Cur Loss: 0.52380407, Cur Avg Loss: 0.59108089, Log Avg loss: 0.66912613, Global Avg Loss: 1.64597043, Time: 0.0115 Steps: 57110, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001766, Sample Num: 28256, Cur Loss: 0.87538129, Cur Avg Loss: 0.59209258, Log Avg loss: 0.76974503, Global Avg Loss: 1.64581703, Time: 0.0093 Steps: 57120, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001776, Sample Num: 28416, Cur Loss: 0.40962619, Cur Avg Loss: 0.59157916, Log Avg loss: 0.50090874, Global Avg Loss: 1.64561663, Time: 0.0113 Steps: 57130, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001786, Sample Num: 28576, Cur Loss: 0.85602528, Cur Avg Loss: 0.59162010, Log Avg loss: 0.59889049, Global Avg Loss: 1.64543344, Time: 0.0074 Steps: 57140, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001796, Sample Num: 28736, Cur Loss: 0.68651575, Cur Avg Loss: 0.59215994, Log Avg loss: 0.68857606, Global Avg Loss: 1.64526601, Time: 0.0114 Steps: 57150, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001806, Sample Num: 28896, Cur Loss: 1.59605813, Cur Avg Loss: 0.59185906, Log Avg loss: 0.53782113, Global Avg Loss: 1.64507227, Time: 0.0065 Steps: 57160, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001816, Sample Num: 29056, Cur Loss: 0.49046946, Cur Avg Loss: 0.59154699, Log Avg loss: 0.53518718, Global Avg Loss: 1.64487813, Time: 0.0066 Steps: 57170, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001826, Sample Num: 29216, Cur Loss: 0.23549031, Cur Avg Loss: 0.59152753, Log Avg loss: 0.58799251, Global Avg Loss: 1.64469330, Time: 0.0103 Steps: 57180, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001836, Sample Num: 29376, Cur Loss: 0.68383831, Cur Avg Loss: 0.59324237, Log Avg loss: 0.90637294, Global Avg Loss: 1.64456420, Time: 0.0114 Steps: 57190, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001846, Sample Num: 29536, Cur Loss: 0.43373859, Cur Avg Loss: 0.59265873, Log Avg loss: 0.48550180, Global Avg Loss: 1.64436156, Time: 0.0128 Steps: 57200, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001856, Sample Num: 29696, Cur Loss: 0.97398937, Cur Avg Loss: 0.59277754, Log Avg loss: 0.61470994, Global Avg Loss: 1.64418159, Time: 0.0225 Steps: 57210, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001866, Sample Num: 29856, Cur Loss: 0.36231482, Cur Avg Loss: 0.59276539, Log Avg loss: 0.59051078, Global Avg Loss: 1.64399744, Time: 0.0073 Steps: 57220, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001876, Sample Num: 30016, Cur Loss: 0.41743320, Cur Avg Loss: 0.59268761, Log Avg loss: 0.57817384, Global Avg Loss: 1.64381121, Time: 0.0188 Steps: 57230, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001886, Sample Num: 30176, Cur Loss: 0.37735507, Cur Avg Loss: 0.59156618, Log Avg loss: 0.38118679, Global Avg Loss: 1.64359062, Time: 0.0118 Steps: 57240, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001896, Sample Num: 30336, Cur Loss: 0.27707660, Cur Avg Loss: 0.59177258, Log Avg loss: 0.63069904, Global Avg Loss: 1.64341370, Time: 0.0146 Steps: 57250, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001906, Sample Num: 30496, Cur Loss: 1.18794847, Cur Avg Loss: 0.59247272, Log Avg loss: 0.72521947, Global Avg Loss: 1.64325334, Time: 0.0067 Steps: 57260, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001916, Sample Num: 30656, Cur Loss: 0.32501531, Cur Avg Loss: 0.59230667, Log Avg loss: 0.56065671, Global Avg Loss: 1.64306431, Time: 0.0114 Steps: 57270, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001926, Sample Num: 30816, Cur Loss: 0.34754124, Cur Avg Loss: 0.59275276, Log Avg loss: 0.67822457, Global Avg Loss: 1.64289587, Time: 0.0068 Steps: 57280, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001936, Sample Num: 30976, Cur Loss: 0.64112449, Cur Avg Loss: 0.59184530, Log Avg loss: 0.41706817, Global Avg Loss: 1.64268190, Time: 0.0081 Steps: 57290, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001946, Sample Num: 31136, Cur Loss: 0.58548003, Cur Avg Loss: 0.59263185, Log Avg loss: 0.74490864, Global Avg Loss: 1.64252522, Time: 0.0071 Steps: 57300, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001956, Sample Num: 31296, Cur Loss: 0.62837100, Cur Avg Loss: 0.59230432, Log Avg loss: 0.52856606, Global Avg Loss: 1.64233084, Time: 0.0110 Steps: 57310, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001966, Sample Num: 31456, Cur Loss: 0.71585774, Cur Avg Loss: 0.59206355, Log Avg loss: 0.54496801, Global Avg Loss: 1.64213940, Time: 0.0082 Steps: 57320, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001976, Sample Num: 31616, Cur Loss: 0.61534154, Cur Avg Loss: 0.59189201, Log Avg loss: 0.55816915, Global Avg Loss: 1.64195032, Time: 0.0114 Steps: 57330, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001986, Sample Num: 31776, Cur Loss: 0.47009355, Cur Avg Loss: 0.59159699, Log Avg loss: 0.53329976, Global Avg Loss: 1.64175698, Time: 0.0119 Steps: 57340, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 001996, Sample Num: 31936, Cur Loss: 0.62711340, Cur Avg Loss: 0.59114405, Log Avg loss: 0.50119005, Global Avg Loss: 1.64155810, Time: 0.0187 Steps: 57350, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002006, Sample Num: 32096, Cur Loss: 0.47047761, Cur Avg Loss: 0.59091795, Log Avg loss: 0.54578927, Global Avg Loss: 1.64136707, Time: 0.0069 Steps: 57360, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002016, Sample Num: 32256, Cur Loss: 0.19577923, Cur Avg Loss: 0.59023257, Log Avg loss: 0.45274495, Global Avg Loss: 1.64115988, Time: 0.0068 Steps: 57370, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002026, Sample Num: 32416, Cur Loss: 0.93541443, Cur Avg Loss: 0.59016169, Log Avg loss: 0.57587150, Global Avg Loss: 1.64097422, Time: 0.0067 Steps: 57380, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002036, Sample Num: 32576, Cur Loss: 0.48167944, Cur Avg Loss: 0.59039526, Log Avg loss: 0.63771696, Global Avg Loss: 1.64079941, Time: 0.0125 Steps: 57390, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002046, Sample Num: 32736, Cur Loss: 0.08091593, Cur Avg Loss: 0.59002508, Log Avg loss: 0.51465747, Global Avg Loss: 1.64060322, Time: 0.0108 Steps: 57400, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002056, Sample Num: 32896, Cur Loss: 0.71746862, Cur Avg Loss: 0.58965332, Log Avg loss: 0.51359006, Global Avg Loss: 1.64040691, Time: 0.0114 Steps: 57410, Updated lr: 0.000047 Training, Epoch: 0027, Batch: 002066, Sample Num: 33056, Cur Loss: 0.12915343, Cur Avg Loss: 0.58885686, Log Avg loss: 0.42510429, Global Avg Loss: 1.64019526, Time: 0.0107 Steps: 57420, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002076, Sample Num: 33216, Cur Loss: 0.67298311, Cur Avg Loss: 0.58901034, Log Avg loss: 0.62072068, Global Avg Loss: 1.64001774, Time: 0.0065 Steps: 57430, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002086, Sample Num: 33376, Cur Loss: 0.26213706, Cur Avg Loss: 0.58875996, Log Avg loss: 0.53678143, Global Avg Loss: 1.63982567, Time: 0.0068 Steps: 57440, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002096, Sample Num: 33536, Cur Loss: 0.19453841, Cur Avg Loss: 0.58890501, Log Avg loss: 0.61916072, Global Avg Loss: 1.63964801, Time: 0.0155 Steps: 57450, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002106, Sample Num: 33696, Cur Loss: 0.25454491, Cur Avg Loss: 0.58900031, Log Avg loss: 0.60897640, Global Avg Loss: 1.63946864, Time: 0.0067 Steps: 57460, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002116, Sample Num: 33856, Cur Loss: 0.33229324, Cur Avg Loss: 0.58885085, Log Avg loss: 0.55737336, Global Avg Loss: 1.63928035, Time: 0.0072 Steps: 57470, Updated lr: 0.000046 Training, Epoch: 0027, Batch: 002126, Sample Num: 34016, Cur Loss: 0.49580044, Cur Avg Loss: 0.58826534, Log Avg loss: 0.46437300, Global Avg Loss: 1.63907595, Time: 0.0100 Steps: 57480, Updated lr: 0.000046 ***** Running evaluation checkpoint-57483 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-57483 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.919765, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.653182, "eval_total_loss": 459.186967, "eval_mae": 0.609466, "eval_mse": 0.653378, "eval_r2": 0.58467, "eval_sp_statistic": 0.721107, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.765652, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.914259, "test_total_loss": 458.957797, "test_mae": 0.744533, "test_mse": 0.914406, "test_r2": 0.409834, "test_sp_statistic": 0.549675, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.647477, "test_ps_pvalue": 0.0, "lr": 4.6436225699383594e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6390031491213024, "train_cur_epoch_loss": 1251.3845839127898, "train_cur_epoch_avg_loss": 0.5877804527537763, "train_cur_epoch_time": 21.91976499557495, "train_cur_epoch_avg_time": 0.010295803191909324, "epoch": 27, "step": 57483} ################################################## Training, Epoch: 0028, Batch: 000007, Sample Num: 112, Cur Loss: 0.37166190, Cur Avg Loss: 0.54886252, Log Avg loss: 0.45745011, Global Avg Loss: 1.63887041, Time: 0.0092 Steps: 57490, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000017, Sample Num: 272, Cur Loss: 0.22334287, Cur Avg Loss: 0.54375665, Log Avg loss: 0.54018253, Global Avg Loss: 1.63867934, Time: 0.0090 Steps: 57500, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000027, Sample Num: 432, Cur Loss: 0.29045671, Cur Avg Loss: 0.54880392, Log Avg loss: 0.55738429, Global Avg Loss: 1.63849132, Time: 0.0067 Steps: 57510, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000037, Sample Num: 592, Cur Loss: 0.27504408, Cur Avg Loss: 0.51723814, Log Avg loss: 0.43201053, Global Avg Loss: 1.63828157, Time: 0.0092 Steps: 57520, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000047, Sample Num: 752, Cur Loss: 0.86271858, Cur Avg Loss: 0.51773511, Log Avg loss: 0.51957392, Global Avg Loss: 1.63808711, Time: 0.0066 Steps: 57530, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000057, Sample Num: 912, Cur Loss: 0.44459134, Cur Avg Loss: 0.52943036, Log Avg loss: 0.58439801, Global Avg Loss: 1.63790399, Time: 0.0065 Steps: 57540, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000067, Sample Num: 1072, Cur Loss: 1.12374914, Cur Avg Loss: 0.56507285, Log Avg loss: 0.76823502, Global Avg Loss: 1.63775287, Time: 0.0070 Steps: 57550, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000077, Sample Num: 1232, Cur Loss: 0.37801528, Cur Avg Loss: 0.56553828, Log Avg loss: 0.56865672, Global Avg Loss: 1.63756714, Time: 0.0067 Steps: 57560, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000087, Sample Num: 1392, Cur Loss: 0.53415453, Cur Avg Loss: 0.56272925, Log Avg loss: 0.54109966, Global Avg Loss: 1.63737668, Time: 0.0066 Steps: 57570, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000097, Sample Num: 1552, Cur Loss: 0.60511833, Cur Avg Loss: 0.54472861, Log Avg loss: 0.38812310, Global Avg Loss: 1.63715972, Time: 0.0129 Steps: 57580, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000107, Sample Num: 1712, Cur Loss: 0.65500128, Cur Avg Loss: 0.54661998, Log Avg loss: 0.56496624, Global Avg Loss: 1.63697354, Time: 0.0131 Steps: 57590, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000117, Sample Num: 1872, Cur Loss: 0.58934271, Cur Avg Loss: 0.53805517, Log Avg loss: 0.44641176, Global Avg Loss: 1.63676685, Time: 0.0120 Steps: 57600, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000127, Sample Num: 2032, Cur Loss: 0.77489936, Cur Avg Loss: 0.55552619, Log Avg loss: 0.75993706, Global Avg Loss: 1.63661465, Time: 0.0067 Steps: 57610, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000137, Sample Num: 2192, Cur Loss: 0.50189471, Cur Avg Loss: 0.55593842, Log Avg loss: 0.56117376, Global Avg Loss: 1.63642800, Time: 0.0105 Steps: 57620, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000147, Sample Num: 2352, Cur Loss: 0.71408218, Cur Avg Loss: 0.56002519, Log Avg loss: 0.61601391, Global Avg Loss: 1.63625094, Time: 0.0119 Steps: 57630, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000157, Sample Num: 2512, Cur Loss: 1.28588951, Cur Avg Loss: 0.56101107, Log Avg loss: 0.57550360, Global Avg Loss: 1.63606691, Time: 0.0110 Steps: 57640, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000167, Sample Num: 2672, Cur Loss: 0.60196710, Cur Avg Loss: 0.55690060, Log Avg loss: 0.49236614, Global Avg Loss: 1.63586852, Time: 0.0071 Steps: 57650, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000177, Sample Num: 2832, Cur Loss: 0.78521514, Cur Avg Loss: 0.56216666, Log Avg loss: 0.65010992, Global Avg Loss: 1.63569756, Time: 0.0114 Steps: 57660, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000187, Sample Num: 2992, Cur Loss: 0.55443537, Cur Avg Loss: 0.56175870, Log Avg loss: 0.55453771, Global Avg Loss: 1.63551009, Time: 0.0112 Steps: 57670, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000197, Sample Num: 3152, Cur Loss: 0.44086555, Cur Avg Loss: 0.56708814, Log Avg loss: 0.66674871, Global Avg Loss: 1.63534214, Time: 0.0068 Steps: 57680, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000207, Sample Num: 3312, Cur Loss: 1.24467468, Cur Avg Loss: 0.57505660, Log Avg loss: 0.73203536, Global Avg Loss: 1.63518556, Time: 0.0146 Steps: 57690, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000217, Sample Num: 3472, Cur Loss: 0.34409371, Cur Avg Loss: 0.57428184, Log Avg loss: 0.55824419, Global Avg Loss: 1.63499891, Time: 0.0153 Steps: 57700, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000227, Sample Num: 3632, Cur Loss: 0.40342167, Cur Avg Loss: 0.57203184, Log Avg loss: 0.52320678, Global Avg Loss: 1.63480626, Time: 0.0072 Steps: 57710, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000237, Sample Num: 3792, Cur Loss: 0.87495536, Cur Avg Loss: 0.57100262, Log Avg loss: 0.54763950, Global Avg Loss: 1.63461791, Time: 0.0071 Steps: 57720, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000247, Sample Num: 3952, Cur Loss: 0.32137641, Cur Avg Loss: 0.56441183, Log Avg loss: 0.40820995, Global Avg Loss: 1.63440547, Time: 0.0068 Steps: 57730, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000257, Sample Num: 4112, Cur Loss: 0.59275651, Cur Avg Loss: 0.56619398, Log Avg loss: 0.61021304, Global Avg Loss: 1.63422809, Time: 0.0144 Steps: 57740, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000267, Sample Num: 4272, Cur Loss: 0.29736000, Cur Avg Loss: 0.56907829, Log Avg loss: 0.64320509, Global Avg Loss: 1.63405648, Time: 0.0134 Steps: 57750, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000277, Sample Num: 4432, Cur Loss: 0.63371837, Cur Avg Loss: 0.57401972, Log Avg loss: 0.70595611, Global Avg Loss: 1.63389580, Time: 0.0137 Steps: 57760, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000287, Sample Num: 4592, Cur Loss: 1.05270290, Cur Avg Loss: 0.57310303, Log Avg loss: 0.54771056, Global Avg Loss: 1.63370778, Time: 0.0127 Steps: 57770, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000297, Sample Num: 4752, Cur Loss: 0.68795383, Cur Avg Loss: 0.57262101, Log Avg loss: 0.55878721, Global Avg Loss: 1.63352175, Time: 0.0071 Steps: 57780, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000307, Sample Num: 4912, Cur Loss: 0.71877080, Cur Avg Loss: 0.57995911, Log Avg loss: 0.79790067, Global Avg Loss: 1.63337715, Time: 0.0158 Steps: 57790, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000317, Sample Num: 5072, Cur Loss: 0.20032953, Cur Avg Loss: 0.58053841, Log Avg loss: 0.59832270, Global Avg Loss: 1.63319807, Time: 0.0119 Steps: 57800, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000327, Sample Num: 5232, Cur Loss: 0.61973852, Cur Avg Loss: 0.58261864, Log Avg loss: 0.64856210, Global Avg Loss: 1.63302775, Time: 0.0118 Steps: 57810, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000337, Sample Num: 5392, Cur Loss: 1.33480000, Cur Avg Loss: 0.58118674, Log Avg loss: 0.53436340, Global Avg Loss: 1.63283774, Time: 0.0073 Steps: 57820, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000347, Sample Num: 5552, Cur Loss: 0.41626436, Cur Avg Loss: 0.58029098, Log Avg loss: 0.55010402, Global Avg Loss: 1.63265051, Time: 0.0113 Steps: 57830, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000357, Sample Num: 5712, Cur Loss: 0.36455303, Cur Avg Loss: 0.58152444, Log Avg loss: 0.62432543, Global Avg Loss: 1.63247618, Time: 0.0107 Steps: 57840, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000367, Sample Num: 5872, Cur Loss: 1.14098895, Cur Avg Loss: 0.58410384, Log Avg loss: 0.67618836, Global Avg Loss: 1.63231088, Time: 0.0158 Steps: 57850, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000377, Sample Num: 6032, Cur Loss: 0.97923750, Cur Avg Loss: 0.58599349, Log Avg loss: 0.65534372, Global Avg Loss: 1.63214202, Time: 0.0099 Steps: 57860, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000387, Sample Num: 6192, Cur Loss: 0.42173451, Cur Avg Loss: 0.58681771, Log Avg loss: 0.61789089, Global Avg Loss: 1.63196676, Time: 0.0067 Steps: 57870, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000397, Sample Num: 6352, Cur Loss: 0.15130647, Cur Avg Loss: 0.58247225, Log Avg loss: 0.41430302, Global Avg Loss: 1.63175638, Time: 0.0114 Steps: 57880, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000407, Sample Num: 6512, Cur Loss: 0.14368218, Cur Avg Loss: 0.58107212, Log Avg loss: 0.52548672, Global Avg Loss: 1.63156529, Time: 0.0063 Steps: 57890, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000417, Sample Num: 6672, Cur Loss: 0.32153198, Cur Avg Loss: 0.57926858, Log Avg loss: 0.50586469, Global Avg Loss: 1.63137086, Time: 0.0110 Steps: 57900, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000427, Sample Num: 6832, Cur Loss: 0.52251494, Cur Avg Loss: 0.57696655, Log Avg loss: 0.48097182, Global Avg Loss: 1.63117221, Time: 0.0141 Steps: 57910, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000437, Sample Num: 6992, Cur Loss: 0.56474704, Cur Avg Loss: 0.57484061, Log Avg loss: 0.48406283, Global Avg Loss: 1.63097416, Time: 0.0067 Steps: 57920, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000447, Sample Num: 7152, Cur Loss: 0.30099571, Cur Avg Loss: 0.57344942, Log Avg loss: 0.51265447, Global Avg Loss: 1.63078111, Time: 0.0074 Steps: 57930, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000457, Sample Num: 7312, Cur Loss: 0.15593754, Cur Avg Loss: 0.57656460, Log Avg loss: 0.71581318, Global Avg Loss: 1.63062320, Time: 0.0102 Steps: 57940, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000467, Sample Num: 7472, Cur Loss: 0.28138506, Cur Avg Loss: 0.57438054, Log Avg loss: 0.47456906, Global Avg Loss: 1.63042371, Time: 0.0157 Steps: 57950, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000477, Sample Num: 7632, Cur Loss: 0.07335536, Cur Avg Loss: 0.57376898, Log Avg loss: 0.54520909, Global Avg Loss: 1.63023647, Time: 0.0085 Steps: 57960, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000487, Sample Num: 7792, Cur Loss: 0.31386060, Cur Avg Loss: 0.57081786, Log Avg loss: 0.43004965, Global Avg Loss: 1.63002943, Time: 0.0110 Steps: 57970, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000497, Sample Num: 7952, Cur Loss: 0.31366247, Cur Avg Loss: 0.57061890, Log Avg loss: 0.56092950, Global Avg Loss: 1.62984504, Time: 0.0115 Steps: 57980, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000507, Sample Num: 8112, Cur Loss: 0.27133408, Cur Avg Loss: 0.57162474, Log Avg loss: 0.62161470, Global Avg Loss: 1.62967118, Time: 0.0136 Steps: 57990, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000517, Sample Num: 8272, Cur Loss: 0.59597254, Cur Avg Loss: 0.56980495, Log Avg loss: 0.47754154, Global Avg Loss: 1.62947254, Time: 0.0068 Steps: 58000, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000527, Sample Num: 8432, Cur Loss: 0.31999499, Cur Avg Loss: 0.56829857, Log Avg loss: 0.49041916, Global Avg Loss: 1.62927618, Time: 0.0114 Steps: 58010, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000537, Sample Num: 8592, Cur Loss: 0.69105238, Cur Avg Loss: 0.56747430, Log Avg loss: 0.52403517, Global Avg Loss: 1.62908569, Time: 0.0116 Steps: 58020, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000547, Sample Num: 8752, Cur Loss: 0.54601455, Cur Avg Loss: 0.56682898, Log Avg loss: 0.53217519, Global Avg Loss: 1.62889667, Time: 0.0068 Steps: 58030, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000557, Sample Num: 8912, Cur Loss: 0.40488625, Cur Avg Loss: 0.56552168, Log Avg loss: 0.49401248, Global Avg Loss: 1.62870113, Time: 0.0107 Steps: 58040, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000567, Sample Num: 9072, Cur Loss: 0.16835320, Cur Avg Loss: 0.56650291, Log Avg loss: 0.62115717, Global Avg Loss: 1.62852757, Time: 0.0113 Steps: 58050, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000577, Sample Num: 9232, Cur Loss: 0.36957711, Cur Avg Loss: 0.56894890, Log Avg loss: 0.70763665, Global Avg Loss: 1.62836896, Time: 0.0114 Steps: 58060, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000587, Sample Num: 9392, Cur Loss: 0.76902592, Cur Avg Loss: 0.56895333, Log Avg loss: 0.56920895, Global Avg Loss: 1.62818656, Time: 0.0110 Steps: 58070, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000597, Sample Num: 9552, Cur Loss: 0.35945848, Cur Avg Loss: 0.56728331, Log Avg loss: 0.46925327, Global Avg Loss: 1.62798702, Time: 0.0125 Steps: 58080, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000607, Sample Num: 9712, Cur Loss: 0.60894024, Cur Avg Loss: 0.56664523, Log Avg loss: 0.52855200, Global Avg Loss: 1.62779776, Time: 0.0097 Steps: 58090, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000617, Sample Num: 9872, Cur Loss: 0.47437981, Cur Avg Loss: 0.56709032, Log Avg loss: 0.59410686, Global Avg Loss: 1.62761984, Time: 0.0067 Steps: 58100, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000627, Sample Num: 10032, Cur Loss: 0.44196793, Cur Avg Loss: 0.57015834, Log Avg loss: 0.75945506, Global Avg Loss: 1.62747044, Time: 0.0067 Steps: 58110, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000637, Sample Num: 10192, Cur Loss: 0.21434198, Cur Avg Loss: 0.56891183, Log Avg loss: 0.49075568, Global Avg Loss: 1.62727486, Time: 0.0115 Steps: 58120, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000647, Sample Num: 10352, Cur Loss: 0.69032174, Cur Avg Loss: 0.57020183, Log Avg loss: 0.65237487, Global Avg Loss: 1.62710715, Time: 0.0116 Steps: 58130, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000657, Sample Num: 10512, Cur Loss: 0.60439026, Cur Avg Loss: 0.57038351, Log Avg loss: 0.58213829, Global Avg Loss: 1.62692742, Time: 0.0067 Steps: 58140, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000667, Sample Num: 10672, Cur Loss: 0.39930442, Cur Avg Loss: 0.57152022, Log Avg loss: 0.64620245, Global Avg Loss: 1.62675876, Time: 0.0115 Steps: 58150, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000677, Sample Num: 10832, Cur Loss: 0.89964575, Cur Avg Loss: 0.57289504, Log Avg loss: 0.66459526, Global Avg Loss: 1.62659333, Time: 0.0140 Steps: 58160, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000687, Sample Num: 10992, Cur Loss: 0.31956267, Cur Avg Loss: 0.57136787, Log Avg loss: 0.46797856, Global Avg Loss: 1.62639415, Time: 0.0110 Steps: 58170, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000697, Sample Num: 11152, Cur Loss: 0.99419397, Cur Avg Loss: 0.57145676, Log Avg loss: 0.57756353, Global Avg Loss: 1.62621388, Time: 0.0130 Steps: 58180, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000707, Sample Num: 11312, Cur Loss: 0.52808028, Cur Avg Loss: 0.57087577, Log Avg loss: 0.53038057, Global Avg Loss: 1.62602556, Time: 0.0087 Steps: 58190, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000717, Sample Num: 11472, Cur Loss: 0.61664999, Cur Avg Loss: 0.56990024, Log Avg loss: 0.50093047, Global Avg Loss: 1.62583224, Time: 0.0076 Steps: 58200, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000727, Sample Num: 11632, Cur Loss: 0.59705234, Cur Avg Loss: 0.56971520, Log Avg loss: 0.55644780, Global Avg Loss: 1.62564853, Time: 0.0082 Steps: 58210, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000737, Sample Num: 11792, Cur Loss: 0.47544366, Cur Avg Loss: 0.56873212, Log Avg loss: 0.49726232, Global Avg Loss: 1.62545472, Time: 0.0066 Steps: 58220, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000747, Sample Num: 11952, Cur Loss: 0.51226205, Cur Avg Loss: 0.56863142, Log Avg loss: 0.56120931, Global Avg Loss: 1.62527195, Time: 0.0115 Steps: 58230, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000757, Sample Num: 12112, Cur Loss: 1.06083584, Cur Avg Loss: 0.56790270, Log Avg loss: 0.51346734, Global Avg Loss: 1.62508105, Time: 0.0115 Steps: 58240, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000767, Sample Num: 12272, Cur Loss: 0.22921343, Cur Avg Loss: 0.56639964, Log Avg loss: 0.45261833, Global Avg Loss: 1.62487977, Time: 0.0231 Steps: 58250, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000777, Sample Num: 12432, Cur Loss: 0.63102806, Cur Avg Loss: 0.56871441, Log Avg loss: 0.74625720, Global Avg Loss: 1.62472896, Time: 0.0113 Steps: 58260, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000787, Sample Num: 12592, Cur Loss: 0.86346376, Cur Avg Loss: 0.56912354, Log Avg loss: 0.60091329, Global Avg Loss: 1.62455326, Time: 0.0063 Steps: 58270, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000797, Sample Num: 12752, Cur Loss: 0.74371517, Cur Avg Loss: 0.56867433, Log Avg loss: 0.53332082, Global Avg Loss: 1.62436602, Time: 0.0138 Steps: 58280, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000807, Sample Num: 12912, Cur Loss: 0.35588211, Cur Avg Loss: 0.56761705, Log Avg loss: 0.48335257, Global Avg Loss: 1.62417027, Time: 0.0073 Steps: 58290, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000817, Sample Num: 13072, Cur Loss: 0.40447551, Cur Avg Loss: 0.56810807, Log Avg loss: 0.60773307, Global Avg Loss: 1.62399592, Time: 0.0194 Steps: 58300, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000827, Sample Num: 13232, Cur Loss: 0.65389252, Cur Avg Loss: 0.56862530, Log Avg loss: 0.61088305, Global Avg Loss: 1.62382218, Time: 0.0085 Steps: 58310, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000837, Sample Num: 13392, Cur Loss: 0.48510680, Cur Avg Loss: 0.56842189, Log Avg loss: 0.55159937, Global Avg Loss: 1.62363833, Time: 0.0068 Steps: 58320, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000847, Sample Num: 13552, Cur Loss: 0.75671971, Cur Avg Loss: 0.56914887, Log Avg loss: 0.62999786, Global Avg Loss: 1.62346798, Time: 0.0065 Steps: 58330, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000857, Sample Num: 13712, Cur Loss: 1.12145901, Cur Avg Loss: 0.57118191, Log Avg loss: 0.74337997, Global Avg Loss: 1.62331712, Time: 0.0067 Steps: 58340, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000867, Sample Num: 13872, Cur Loss: 0.49504718, Cur Avg Loss: 0.57183176, Log Avg loss: 0.62752381, Global Avg Loss: 1.62314646, Time: 0.0073 Steps: 58350, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000877, Sample Num: 14032, Cur Loss: 0.58266705, Cur Avg Loss: 0.57079048, Log Avg loss: 0.48051159, Global Avg Loss: 1.62295067, Time: 0.0070 Steps: 58360, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000887, Sample Num: 14192, Cur Loss: 0.60470068, Cur Avg Loss: 0.57100974, Log Avg loss: 0.59023894, Global Avg Loss: 1.62277375, Time: 0.0073 Steps: 58370, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000897, Sample Num: 14352, Cur Loss: 0.56296974, Cur Avg Loss: 0.57238078, Log Avg loss: 0.69399174, Global Avg Loss: 1.62261466, Time: 0.0065 Steps: 58380, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000907, Sample Num: 14512, Cur Loss: 0.69137156, Cur Avg Loss: 0.57239990, Log Avg loss: 0.57411505, Global Avg Loss: 1.62243509, Time: 0.0103 Steps: 58390, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000917, Sample Num: 14672, Cur Loss: 0.20475864, Cur Avg Loss: 0.57173390, Log Avg loss: 0.51132744, Global Avg Loss: 1.62224483, Time: 0.0070 Steps: 58400, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000927, Sample Num: 14832, Cur Loss: 0.40843529, Cur Avg Loss: 0.57230538, Log Avg loss: 0.62471013, Global Avg Loss: 1.62207405, Time: 0.0069 Steps: 58410, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000937, Sample Num: 14992, Cur Loss: 0.56830370, Cur Avg Loss: 0.57142818, Log Avg loss: 0.49011227, Global Avg Loss: 1.62188028, Time: 0.0067 Steps: 58420, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000947, Sample Num: 15152, Cur Loss: 0.82898593, Cur Avg Loss: 0.57345902, Log Avg loss: 0.76374890, Global Avg Loss: 1.62173342, Time: 0.0095 Steps: 58430, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000957, Sample Num: 15312, Cur Loss: 0.95247829, Cur Avg Loss: 0.57537174, Log Avg loss: 0.75650566, Global Avg Loss: 1.62158537, Time: 0.0068 Steps: 58440, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000967, Sample Num: 15472, Cur Loss: 0.57107538, Cur Avg Loss: 0.57492846, Log Avg loss: 0.53250662, Global Avg Loss: 1.62139904, Time: 0.0111 Steps: 58450, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000977, Sample Num: 15632, Cur Loss: 0.21016081, Cur Avg Loss: 0.57538158, Log Avg loss: 0.61919882, Global Avg Loss: 1.62122761, Time: 0.0070 Steps: 58460, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000987, Sample Num: 15792, Cur Loss: 0.50225526, Cur Avg Loss: 0.57588576, Log Avg loss: 0.62514415, Global Avg Loss: 1.62105725, Time: 0.0118 Steps: 58470, Updated lr: 0.000046 Training, Epoch: 0028, Batch: 000997, Sample Num: 15952, Cur Loss: 0.45018670, Cur Avg Loss: 0.57461508, Log Avg loss: 0.44919889, Global Avg Loss: 1.62085686, Time: 0.0107 Steps: 58480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001007, Sample Num: 16112, Cur Loss: 0.60326171, Cur Avg Loss: 0.57608458, Log Avg loss: 0.72259403, Global Avg Loss: 1.62070329, Time: 0.0099 Steps: 58490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001017, Sample Num: 16272, Cur Loss: 0.46480495, Cur Avg Loss: 0.57593480, Log Avg loss: 0.56085130, Global Avg Loss: 1.62052211, Time: 0.0168 Steps: 58500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001027, Sample Num: 16432, Cur Loss: 0.23040734, Cur Avg Loss: 0.57477515, Log Avg loss: 0.45683935, Global Avg Loss: 1.62032323, Time: 0.0130 Steps: 58510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001037, Sample Num: 16592, Cur Loss: 0.27740741, Cur Avg Loss: 0.57295816, Log Avg loss: 0.38635266, Global Avg Loss: 1.62011237, Time: 0.0156 Steps: 58520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001047, Sample Num: 16752, Cur Loss: 0.19839808, Cur Avg Loss: 0.57424588, Log Avg loss: 0.70778299, Global Avg Loss: 1.61995649, Time: 0.0120 Steps: 58530, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001057, Sample Num: 16912, Cur Loss: 0.50118345, Cur Avg Loss: 0.57578013, Log Avg loss: 0.73641535, Global Avg Loss: 1.61980556, Time: 0.0077 Steps: 58540, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001067, Sample Num: 17072, Cur Loss: 0.78670001, Cur Avg Loss: 0.57522785, Log Avg loss: 0.51685270, Global Avg Loss: 1.61961718, Time: 0.0219 Steps: 58550, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001077, Sample Num: 17232, Cur Loss: 0.70566756, Cur Avg Loss: 0.57411179, Log Avg loss: 0.45502806, Global Avg Loss: 1.61941831, Time: 0.0082 Steps: 58560, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001087, Sample Num: 17392, Cur Loss: 0.33746541, Cur Avg Loss: 0.57363641, Log Avg loss: 0.52243786, Global Avg Loss: 1.61923102, Time: 0.0073 Steps: 58570, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001097, Sample Num: 17552, Cur Loss: 0.29272592, Cur Avg Loss: 0.57583013, Log Avg loss: 0.81428708, Global Avg Loss: 1.61909361, Time: 0.0110 Steps: 58580, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001107, Sample Num: 17712, Cur Loss: 0.56346214, Cur Avg Loss: 0.57551924, Log Avg loss: 0.54141501, Global Avg Loss: 1.61890967, Time: 0.0108 Steps: 58590, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001117, Sample Num: 17872, Cur Loss: 0.26734543, Cur Avg Loss: 0.57532499, Log Avg loss: 0.55382154, Global Avg Loss: 1.61872792, Time: 0.0066 Steps: 58600, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001127, Sample Num: 18032, Cur Loss: 0.54883277, Cur Avg Loss: 0.57527841, Log Avg loss: 0.57007471, Global Avg Loss: 1.61854900, Time: 0.0075 Steps: 58610, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001137, Sample Num: 18192, Cur Loss: 0.59727085, Cur Avg Loss: 0.57434202, Log Avg loss: 0.46881189, Global Avg Loss: 1.61835286, Time: 0.0070 Steps: 58620, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001147, Sample Num: 18352, Cur Loss: 0.66422260, Cur Avg Loss: 0.57409895, Log Avg loss: 0.54646117, Global Avg Loss: 1.61817004, Time: 0.0118 Steps: 58630, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001157, Sample Num: 18512, Cur Loss: 0.89010310, Cur Avg Loss: 0.57345017, Log Avg loss: 0.49903501, Global Avg Loss: 1.61797919, Time: 0.0115 Steps: 58640, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001167, Sample Num: 18672, Cur Loss: 0.85817623, Cur Avg Loss: 0.57401856, Log Avg loss: 0.63978194, Global Avg Loss: 1.61781241, Time: 0.0133 Steps: 58650, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001177, Sample Num: 18832, Cur Loss: 0.48426715, Cur Avg Loss: 0.57337363, Log Avg loss: 0.49810961, Global Avg Loss: 1.61762153, Time: 0.0115 Steps: 58660, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001187, Sample Num: 18992, Cur Loss: 0.52862942, Cur Avg Loss: 0.57418226, Log Avg loss: 0.66935818, Global Avg Loss: 1.61745990, Time: 0.0077 Steps: 58670, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001197, Sample Num: 19152, Cur Loss: 1.18442023, Cur Avg Loss: 0.57338359, Log Avg loss: 0.47858127, Global Avg Loss: 1.61726582, Time: 0.0148 Steps: 58680, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001207, Sample Num: 19312, Cur Loss: 0.54680985, Cur Avg Loss: 0.57266595, Log Avg loss: 0.48676488, Global Avg Loss: 1.61707320, Time: 0.0112 Steps: 58690, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001217, Sample Num: 19472, Cur Loss: 0.70924574, Cur Avg Loss: 0.57317940, Log Avg loss: 0.63515228, Global Avg Loss: 1.61690592, Time: 0.0064 Steps: 58700, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001227, Sample Num: 19632, Cur Loss: 1.93682265, Cur Avg Loss: 0.57521732, Log Avg loss: 0.82323247, Global Avg Loss: 1.61677073, Time: 0.0064 Steps: 58710, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001237, Sample Num: 19792, Cur Loss: 0.40486881, Cur Avg Loss: 0.57512445, Log Avg loss: 0.56372963, Global Avg Loss: 1.61659140, Time: 0.0111 Steps: 58720, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001247, Sample Num: 19952, Cur Loss: 0.31950709, Cur Avg Loss: 0.57649506, Log Avg loss: 0.74603946, Global Avg Loss: 1.61644317, Time: 0.0118 Steps: 58730, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001257, Sample Num: 20112, Cur Loss: 0.65045798, Cur Avg Loss: 0.57698457, Log Avg loss: 0.63802612, Global Avg Loss: 1.61627660, Time: 0.0112 Steps: 58740, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001267, Sample Num: 20272, Cur Loss: 0.54637116, Cur Avg Loss: 0.57646040, Log Avg loss: 0.51057223, Global Avg Loss: 1.61608840, Time: 0.0132 Steps: 58750, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001277, Sample Num: 20432, Cur Loss: 0.89008313, Cur Avg Loss: 0.57720376, Log Avg loss: 0.67138811, Global Avg Loss: 1.61592762, Time: 0.0064 Steps: 58760, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001287, Sample Num: 20592, Cur Loss: 0.29260939, Cur Avg Loss: 0.57606080, Log Avg loss: 0.43010455, Global Avg Loss: 1.61572585, Time: 0.0115 Steps: 58770, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001297, Sample Num: 20752, Cur Loss: 0.34585845, Cur Avg Loss: 0.57622206, Log Avg loss: 0.59697552, Global Avg Loss: 1.61555254, Time: 0.0128 Steps: 58780, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001307, Sample Num: 20912, Cur Loss: 0.28416112, Cur Avg Loss: 0.57558357, Log Avg loss: 0.49277264, Global Avg Loss: 1.61536155, Time: 0.0164 Steps: 58790, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001317, Sample Num: 21072, Cur Loss: 0.50519460, Cur Avg Loss: 0.57714394, Log Avg loss: 0.78108383, Global Avg Loss: 1.61521967, Time: 0.0066 Steps: 58800, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001327, Sample Num: 21232, Cur Loss: 0.69451648, Cur Avg Loss: 0.57684475, Log Avg loss: 0.53744118, Global Avg Loss: 1.61503641, Time: 0.0220 Steps: 58810, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001337, Sample Num: 21392, Cur Loss: 0.74905014, Cur Avg Loss: 0.57710250, Log Avg loss: 0.61130669, Global Avg Loss: 1.61486576, Time: 0.0063 Steps: 58820, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001347, Sample Num: 21552, Cur Loss: 0.57020837, Cur Avg Loss: 0.57631751, Log Avg loss: 0.47136336, Global Avg Loss: 1.61467139, Time: 0.0064 Steps: 58830, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001357, Sample Num: 21712, Cur Loss: 0.53391486, Cur Avg Loss: 0.57602516, Log Avg loss: 0.53664538, Global Avg Loss: 1.61448817, Time: 0.0070 Steps: 58840, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001367, Sample Num: 21872, Cur Loss: 0.32964611, Cur Avg Loss: 0.57541069, Log Avg loss: 0.49202834, Global Avg Loss: 1.61429744, Time: 0.0124 Steps: 58850, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001377, Sample Num: 22032, Cur Loss: 0.24059853, Cur Avg Loss: 0.57487764, Log Avg loss: 0.50200861, Global Avg Loss: 1.61410847, Time: 0.0070 Steps: 58860, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001387, Sample Num: 22192, Cur Loss: 1.12920427, Cur Avg Loss: 0.57501379, Log Avg loss: 0.59376270, Global Avg Loss: 1.61393515, Time: 0.0190 Steps: 58870, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001397, Sample Num: 22352, Cur Loss: 0.60021710, Cur Avg Loss: 0.57529052, Log Avg loss: 0.61367214, Global Avg Loss: 1.61376527, Time: 0.0066 Steps: 58880, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001407, Sample Num: 22512, Cur Loss: 0.55611533, Cur Avg Loss: 0.57584749, Log Avg loss: 0.65365638, Global Avg Loss: 1.61360223, Time: 0.0068 Steps: 58890, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001417, Sample Num: 22672, Cur Loss: 0.26233667, Cur Avg Loss: 0.57648034, Log Avg loss: 0.66552306, Global Avg Loss: 1.61344127, Time: 0.0065 Steps: 58900, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001427, Sample Num: 22832, Cur Loss: 1.66728950, Cur Avg Loss: 0.57859440, Log Avg loss: 0.87815574, Global Avg Loss: 1.61331645, Time: 0.0115 Steps: 58910, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001437, Sample Num: 22992, Cur Loss: 0.65421921, Cur Avg Loss: 0.57735178, Log Avg loss: 0.40003049, Global Avg Loss: 1.61311053, Time: 0.0067 Steps: 58920, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001447, Sample Num: 23152, Cur Loss: 1.10993731, Cur Avg Loss: 0.57763702, Log Avg loss: 0.61862640, Global Avg Loss: 1.61294177, Time: 0.0064 Steps: 58930, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001457, Sample Num: 23312, Cur Loss: 0.81563795, Cur Avg Loss: 0.57768332, Log Avg loss: 0.58438206, Global Avg Loss: 1.61276727, Time: 0.0153 Steps: 58940, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001467, Sample Num: 23472, Cur Loss: 0.77042693, Cur Avg Loss: 0.57830672, Log Avg loss: 0.66913677, Global Avg Loss: 1.61260719, Time: 0.0138 Steps: 58950, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001477, Sample Num: 23632, Cur Loss: 1.31523871, Cur Avg Loss: 0.57833691, Log Avg loss: 0.58276523, Global Avg Loss: 1.61243252, Time: 0.0065 Steps: 58960, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001487, Sample Num: 23792, Cur Loss: 0.21901423, Cur Avg Loss: 0.57731721, Log Avg loss: 0.42670827, Global Avg Loss: 1.61223145, Time: 0.0072 Steps: 58970, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001497, Sample Num: 23952, Cur Loss: 0.39910385, Cur Avg Loss: 0.57654458, Log Avg loss: 0.46165346, Global Avg Loss: 1.61203637, Time: 0.0069 Steps: 58980, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001507, Sample Num: 24112, Cur Loss: 0.69940537, Cur Avg Loss: 0.57624373, Log Avg loss: 0.53120730, Global Avg Loss: 1.61185315, Time: 0.0072 Steps: 58990, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001517, Sample Num: 24272, Cur Loss: 0.44896764, Cur Avg Loss: 0.57692905, Log Avg loss: 0.68020625, Global Avg Loss: 1.61169524, Time: 0.0070 Steps: 59000, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001527, Sample Num: 24432, Cur Loss: 0.57473356, Cur Avg Loss: 0.57670872, Log Avg loss: 0.54328451, Global Avg Loss: 1.61151419, Time: 0.0067 Steps: 59010, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001537, Sample Num: 24592, Cur Loss: 0.28101656, Cur Avg Loss: 0.57616119, Log Avg loss: 0.49255325, Global Avg Loss: 1.61132460, Time: 0.0176 Steps: 59020, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001547, Sample Num: 24752, Cur Loss: 0.46546441, Cur Avg Loss: 0.57523305, Log Avg loss: 0.43257806, Global Avg Loss: 1.61112491, Time: 0.0121 Steps: 59030, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001557, Sample Num: 24912, Cur Loss: 0.50597346, Cur Avg Loss: 0.57669615, Log Avg loss: 0.80303715, Global Avg Loss: 1.61098804, Time: 0.0072 Steps: 59040, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001567, Sample Num: 25072, Cur Loss: 0.36122394, Cur Avg Loss: 0.57749119, Log Avg loss: 0.70128012, Global Avg Loss: 1.61083398, Time: 0.0074 Steps: 59050, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001577, Sample Num: 25232, Cur Loss: 0.70850962, Cur Avg Loss: 0.57796794, Log Avg loss: 0.65267397, Global Avg Loss: 1.61067175, Time: 0.0123 Steps: 59060, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001587, Sample Num: 25392, Cur Loss: 0.26605871, Cur Avg Loss: 0.57682778, Log Avg loss: 0.39702485, Global Avg Loss: 1.61046629, Time: 0.0076 Steps: 59070, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001597, Sample Num: 25552, Cur Loss: 1.08786309, Cur Avg Loss: 0.57665584, Log Avg loss: 0.54936832, Global Avg Loss: 1.61028669, Time: 0.0098 Steps: 59080, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001607, Sample Num: 25712, Cur Loss: 0.37395290, Cur Avg Loss: 0.57717639, Log Avg loss: 0.66030929, Global Avg Loss: 1.61012592, Time: 0.0080 Steps: 59090, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001617, Sample Num: 25872, Cur Loss: 1.99508214, Cur Avg Loss: 0.57736777, Log Avg loss: 0.60812160, Global Avg Loss: 1.60995637, Time: 0.0121 Steps: 59100, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001627, Sample Num: 26032, Cur Loss: 1.01185417, Cur Avg Loss: 0.57732509, Log Avg loss: 0.57042353, Global Avg Loss: 1.60978051, Time: 0.0066 Steps: 59110, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001637, Sample Num: 26192, Cur Loss: 0.57685524, Cur Avg Loss: 0.57673942, Log Avg loss: 0.48145156, Global Avg Loss: 1.60958966, Time: 0.0065 Steps: 59120, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001647, Sample Num: 26352, Cur Loss: 0.29824111, Cur Avg Loss: 0.57733494, Log Avg loss: 0.67482161, Global Avg Loss: 1.60943157, Time: 0.0109 Steps: 59130, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001657, Sample Num: 26512, Cur Loss: 0.43236077, Cur Avg Loss: 0.57804067, Log Avg loss: 0.69427439, Global Avg Loss: 1.60927682, Time: 0.0122 Steps: 59140, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001667, Sample Num: 26672, Cur Loss: 0.67723173, Cur Avg Loss: 0.57769968, Log Avg loss: 0.52119833, Global Avg Loss: 1.60909287, Time: 0.0118 Steps: 59150, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001677, Sample Num: 26832, Cur Loss: 0.46398836, Cur Avg Loss: 0.57909098, Log Avg loss: 0.81101965, Global Avg Loss: 1.60895797, Time: 0.0232 Steps: 59160, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001687, Sample Num: 26992, Cur Loss: 0.37604794, Cur Avg Loss: 0.57974590, Log Avg loss: 0.68957664, Global Avg Loss: 1.60880259, Time: 0.0120 Steps: 59170, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001697, Sample Num: 27152, Cur Loss: 0.60458308, Cur Avg Loss: 0.58054378, Log Avg loss: 0.71514548, Global Avg Loss: 1.60865159, Time: 0.0072 Steps: 59180, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001707, Sample Num: 27312, Cur Loss: 0.28132832, Cur Avg Loss: 0.57945175, Log Avg loss: 0.39413376, Global Avg Loss: 1.60844640, Time: 0.0071 Steps: 59190, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001717, Sample Num: 27472, Cur Loss: 0.35338241, Cur Avg Loss: 0.57892491, Log Avg loss: 0.48899383, Global Avg Loss: 1.60825730, Time: 0.0089 Steps: 59200, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001727, Sample Num: 27632, Cur Loss: 1.13837898, Cur Avg Loss: 0.57988214, Log Avg loss: 0.74423898, Global Avg Loss: 1.60811137, Time: 0.0111 Steps: 59210, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001737, Sample Num: 27792, Cur Loss: 0.70879555, Cur Avg Loss: 0.58130782, Log Avg loss: 0.82752279, Global Avg Loss: 1.60797956, Time: 0.0106 Steps: 59220, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001747, Sample Num: 27952, Cur Loss: 0.38754728, Cur Avg Loss: 0.58133223, Log Avg loss: 0.58557290, Global Avg Loss: 1.60780695, Time: 0.0067 Steps: 59230, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001757, Sample Num: 28112, Cur Loss: 0.67275101, Cur Avg Loss: 0.58055284, Log Avg loss: 0.44439225, Global Avg Loss: 1.60761056, Time: 0.0070 Steps: 59240, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001767, Sample Num: 28272, Cur Loss: 0.49252170, Cur Avg Loss: 0.58011918, Log Avg loss: 0.50392547, Global Avg Loss: 1.60742428, Time: 0.0112 Steps: 59250, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001777, Sample Num: 28432, Cur Loss: 0.54045403, Cur Avg Loss: 0.57966052, Log Avg loss: 0.49861573, Global Avg Loss: 1.60723717, Time: 0.0096 Steps: 59260, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001787, Sample Num: 28592, Cur Loss: 0.88886958, Cur Avg Loss: 0.57909876, Log Avg loss: 0.47927290, Global Avg Loss: 1.60704686, Time: 0.0138 Steps: 59270, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001797, Sample Num: 28752, Cur Loss: 0.42953017, Cur Avg Loss: 0.57941881, Log Avg loss: 0.63661221, Global Avg Loss: 1.60688316, Time: 0.0120 Steps: 59280, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001807, Sample Num: 28912, Cur Loss: 0.73067856, Cur Avg Loss: 0.57901518, Log Avg loss: 0.50648292, Global Avg Loss: 1.60669756, Time: 0.0159 Steps: 59290, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001817, Sample Num: 29072, Cur Loss: 0.85812485, Cur Avg Loss: 0.57963637, Log Avg loss: 0.69188519, Global Avg Loss: 1.60654329, Time: 0.0115 Steps: 59300, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001827, Sample Num: 29232, Cur Loss: 0.57602608, Cur Avg Loss: 0.58000911, Log Avg loss: 0.64773561, Global Avg Loss: 1.60638163, Time: 0.0073 Steps: 59310, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001837, Sample Num: 29392, Cur Loss: 0.20967323, Cur Avg Loss: 0.57996176, Log Avg loss: 0.57131216, Global Avg Loss: 1.60620714, Time: 0.0108 Steps: 59320, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001847, Sample Num: 29552, Cur Loss: 0.20107514, Cur Avg Loss: 0.57924133, Log Avg loss: 0.44689710, Global Avg Loss: 1.60601174, Time: 0.0067 Steps: 59330, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001857, Sample Num: 29712, Cur Loss: 0.26449680, Cur Avg Loss: 0.57873151, Log Avg loss: 0.48456837, Global Avg Loss: 1.60582276, Time: 0.0071 Steps: 59340, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001867, Sample Num: 29872, Cur Loss: 0.26752698, Cur Avg Loss: 0.57797094, Log Avg loss: 0.43673398, Global Avg Loss: 1.60562578, Time: 0.0096 Steps: 59350, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001877, Sample Num: 30032, Cur Loss: 0.60968065, Cur Avg Loss: 0.57774230, Log Avg loss: 0.53505385, Global Avg Loss: 1.60544542, Time: 0.0076 Steps: 59360, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001887, Sample Num: 30192, Cur Loss: 0.48539090, Cur Avg Loss: 0.57748674, Log Avg loss: 0.52951808, Global Avg Loss: 1.60526420, Time: 0.0070 Steps: 59370, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001897, Sample Num: 30352, Cur Loss: 1.19500911, Cur Avg Loss: 0.57725496, Log Avg loss: 0.53351899, Global Avg Loss: 1.60508371, Time: 0.0129 Steps: 59380, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001907, Sample Num: 30512, Cur Loss: 0.30806255, Cur Avg Loss: 0.57676963, Log Avg loss: 0.48470179, Global Avg Loss: 1.60489506, Time: 0.0112 Steps: 59390, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001917, Sample Num: 30672, Cur Loss: 0.67718005, Cur Avg Loss: 0.57765916, Log Avg loss: 0.74729351, Global Avg Loss: 1.60475068, Time: 0.0122 Steps: 59400, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001927, Sample Num: 30832, Cur Loss: 0.56349993, Cur Avg Loss: 0.57820025, Log Avg loss: 0.68192635, Global Avg Loss: 1.60459535, Time: 0.0065 Steps: 59410, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001937, Sample Num: 30992, Cur Loss: 0.17944229, Cur Avg Loss: 0.57826076, Log Avg loss: 0.58992040, Global Avg Loss: 1.60442459, Time: 0.0106 Steps: 59420, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001947, Sample Num: 31152, Cur Loss: 0.41723546, Cur Avg Loss: 0.57812940, Log Avg loss: 0.55268494, Global Avg Loss: 1.60424762, Time: 0.0111 Steps: 59430, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001957, Sample Num: 31312, Cur Loss: 0.36143649, Cur Avg Loss: 0.57804715, Log Avg loss: 0.56203418, Global Avg Loss: 1.60407228, Time: 0.0100 Steps: 59440, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001967, Sample Num: 31472, Cur Loss: 0.20908862, Cur Avg Loss: 0.57706841, Log Avg loss: 0.38552965, Global Avg Loss: 1.60386731, Time: 0.0067 Steps: 59450, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001977, Sample Num: 31632, Cur Loss: 0.61369026, Cur Avg Loss: 0.57760281, Log Avg loss: 0.68271900, Global Avg Loss: 1.60371239, Time: 0.0114 Steps: 59460, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001987, Sample Num: 31792, Cur Loss: 0.97578567, Cur Avg Loss: 0.57753150, Log Avg loss: 0.56343246, Global Avg Loss: 1.60353747, Time: 0.0066 Steps: 59470, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 001997, Sample Num: 31952, Cur Loss: 0.85898769, Cur Avg Loss: 0.57794986, Log Avg loss: 0.66107949, Global Avg Loss: 1.60337902, Time: 0.0064 Steps: 59480, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002007, Sample Num: 32112, Cur Loss: 0.75967640, Cur Avg Loss: 0.57781566, Log Avg loss: 0.55101400, Global Avg Loss: 1.60320212, Time: 0.0168 Steps: 59490, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002017, Sample Num: 32272, Cur Loss: 0.61232501, Cur Avg Loss: 0.57788906, Log Avg loss: 0.59262043, Global Avg Loss: 1.60303227, Time: 0.0069 Steps: 59500, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002027, Sample Num: 32432, Cur Loss: 0.39260471, Cur Avg Loss: 0.57771517, Log Avg loss: 0.54264171, Global Avg Loss: 1.60285409, Time: 0.0185 Steps: 59510, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002037, Sample Num: 32592, Cur Loss: 0.57754225, Cur Avg Loss: 0.57755897, Log Avg loss: 0.54589876, Global Avg Loss: 1.60267651, Time: 0.0064 Steps: 59520, Updated lr: 0.000045 Training, Epoch: 0028, Batch: 002047, Sample Num: 32752, Cur Loss: 0.81850219, Cur Avg Loss: 0.57778699, Log Avg loss: 0.62423330, Global Avg Loss: 1.60251214, Time: 0.0186 Steps: 59530, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002057, Sample Num: 32912, Cur Loss: 0.31054944, Cur Avg Loss: 0.57753056, Log Avg loss: 0.52503968, Global Avg Loss: 1.60233118, Time: 0.0092 Steps: 59540, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002067, Sample Num: 33072, Cur Loss: 1.03810251, Cur Avg Loss: 0.57817374, Log Avg loss: 0.71047628, Global Avg Loss: 1.60218141, Time: 0.0069 Steps: 59550, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002077, Sample Num: 33232, Cur Loss: 0.20387739, Cur Avg Loss: 0.57739723, Log Avg loss: 0.41689177, Global Avg Loss: 1.60198241, Time: 0.0072 Steps: 59560, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002087, Sample Num: 33392, Cur Loss: 0.78486139, Cur Avg Loss: 0.57701856, Log Avg loss: 0.49836895, Global Avg Loss: 1.60179714, Time: 0.0116 Steps: 59570, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002097, Sample Num: 33552, Cur Loss: 0.96953750, Cur Avg Loss: 0.57685911, Log Avg loss: 0.54358292, Global Avg Loss: 1.60161953, Time: 0.0067 Steps: 59580, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002107, Sample Num: 33712, Cur Loss: 0.31286442, Cur Avg Loss: 0.57701122, Log Avg loss: 0.60890910, Global Avg Loss: 1.60145294, Time: 0.0067 Steps: 59590, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002117, Sample Num: 33872, Cur Loss: 0.78380144, Cur Avg Loss: 0.57704139, Log Avg loss: 0.58339830, Global Avg Loss: 1.60128213, Time: 0.0109 Steps: 59600, Updated lr: 0.000044 Training, Epoch: 0028, Batch: 002127, Sample Num: 34032, Cur Loss: 0.75514627, Cur Avg Loss: 0.57686060, Log Avg loss: 0.53858557, Global Avg Loss: 1.60110385, Time: 0.0156 Steps: 59610, Updated lr: 0.000044 ***** Running evaluation checkpoint-59612 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-59612 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.115857, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.735164, "eval_total_loss": 516.820025, "eval_mae": 0.698386, "eval_mse": 0.735239, "eval_r2": 0.532634, "eval_sp_statistic": 0.704315, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.766248, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.927133, "test_total_loss": 465.420732, "test_mae": 0.766494, "test_mse": 0.927258, "test_r2": 0.40154, "test_sp_statistic": 0.554807, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.650574, "test_ps_pvalue": 0.0, "lr": 4.441725936462779e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.6010687871662586, "train_cur_epoch_loss": 1228.0945196151733, "train_cur_epoch_avg_loss": 0.5768410143800721, "train_cur_epoch_time": 22.115856885910034, "train_cur_epoch_avg_time": 0.010387908354114624, "epoch": 28, "step": 59612} ################################################## Training, Epoch: 0029, Batch: 000008, Sample Num: 128, Cur Loss: 0.28066319, Cur Avg Loss: 0.54644240, Log Avg loss: 0.54835711, Global Avg Loss: 1.60092727, Time: 0.0115 Steps: 59620, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000018, Sample Num: 288, Cur Loss: 0.43562049, Cur Avg Loss: 0.60720423, Log Avg loss: 0.65581370, Global Avg Loss: 1.60076878, Time: 0.0117 Steps: 59630, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000028, Sample Num: 448, Cur Loss: 0.47835380, Cur Avg Loss: 0.60331778, Log Avg loss: 0.59632218, Global Avg Loss: 1.60060036, Time: 0.0066 Steps: 59640, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000038, Sample Num: 608, Cur Loss: 0.26167494, Cur Avg Loss: 0.58531651, Log Avg loss: 0.53491295, Global Avg Loss: 1.60042170, Time: 0.0124 Steps: 59650, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000048, Sample Num: 768, Cur Loss: 0.39608172, Cur Avg Loss: 0.55373413, Log Avg loss: 0.43372107, Global Avg Loss: 1.60022614, Time: 0.0103 Steps: 59660, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000058, Sample Num: 928, Cur Loss: 0.54611295, Cur Avg Loss: 0.54309462, Log Avg loss: 0.49202497, Global Avg Loss: 1.60004042, Time: 0.0068 Steps: 59670, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000068, Sample Num: 1088, Cur Loss: 0.38747495, Cur Avg Loss: 0.52703135, Log Avg loss: 0.43386440, Global Avg Loss: 1.59984502, Time: 0.0111 Steps: 59680, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000078, Sample Num: 1248, Cur Loss: 1.14411294, Cur Avg Loss: 0.54132000, Log Avg loss: 0.63848281, Global Avg Loss: 1.59968396, Time: 0.0067 Steps: 59690, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000088, Sample Num: 1408, Cur Loss: 0.21166925, Cur Avg Loss: 0.54051437, Log Avg loss: 0.53423043, Global Avg Loss: 1.59950549, Time: 0.0067 Steps: 59700, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000098, Sample Num: 1568, Cur Loss: 0.30025786, Cur Avg Loss: 0.55129248, Log Avg loss: 0.64613992, Global Avg Loss: 1.59934582, Time: 0.0068 Steps: 59710, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000108, Sample Num: 1728, Cur Loss: 0.40265840, Cur Avg Loss: 0.54629806, Log Avg loss: 0.49735266, Global Avg Loss: 1.59916130, Time: 0.0159 Steps: 59720, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000118, Sample Num: 1888, Cur Loss: 0.52217722, Cur Avg Loss: 0.54155671, Log Avg loss: 0.49035018, Global Avg Loss: 1.59897566, Time: 0.0120 Steps: 59730, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000128, Sample Num: 2048, Cur Loss: 0.38468105, Cur Avg Loss: 0.54897675, Log Avg loss: 0.63653316, Global Avg Loss: 1.59881456, Time: 0.0108 Steps: 59740, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000138, Sample Num: 2208, Cur Loss: 0.47726476, Cur Avg Loss: 0.54320534, Log Avg loss: 0.46933130, Global Avg Loss: 1.59862552, Time: 0.0102 Steps: 59750, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000148, Sample Num: 2368, Cur Loss: 0.42489743, Cur Avg Loss: 0.54939301, Log Avg loss: 0.63478290, Global Avg Loss: 1.59846424, Time: 0.0116 Steps: 59760, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000158, Sample Num: 2528, Cur Loss: 0.40839210, Cur Avg Loss: 0.55334867, Log Avg loss: 0.61189248, Global Avg Loss: 1.59829917, Time: 0.0115 Steps: 59770, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000168, Sample Num: 2688, Cur Loss: 0.17740500, Cur Avg Loss: 0.55865863, Log Avg loss: 0.64255593, Global Avg Loss: 1.59813930, Time: 0.0155 Steps: 59780, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000178, Sample Num: 2848, Cur Loss: 0.75567091, Cur Avg Loss: 0.55139161, Log Avg loss: 0.42930567, Global Avg Loss: 1.59794381, Time: 0.0073 Steps: 59790, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000188, Sample Num: 3008, Cur Loss: 0.54926652, Cur Avg Loss: 0.54546926, Log Avg loss: 0.44005141, Global Avg Loss: 1.59775018, Time: 0.0155 Steps: 59800, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000198, Sample Num: 3168, Cur Loss: 0.96154583, Cur Avg Loss: 0.54803740, Log Avg loss: 0.59631847, Global Avg Loss: 1.59758274, Time: 0.0066 Steps: 59810, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000208, Sample Num: 3328, Cur Loss: 1.30056155, Cur Avg Loss: 0.55156293, Log Avg loss: 0.62136850, Global Avg Loss: 1.59741955, Time: 0.0072 Steps: 59820, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000218, Sample Num: 3488, Cur Loss: 0.47349247, Cur Avg Loss: 0.56212506, Log Avg loss: 0.78181740, Global Avg Loss: 1.59728323, Time: 0.0065 Steps: 59830, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000228, Sample Num: 3648, Cur Loss: 0.52144670, Cur Avg Loss: 0.55494821, Log Avg loss: 0.39849281, Global Avg Loss: 1.59708290, Time: 0.0066 Steps: 59840, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000238, Sample Num: 3808, Cur Loss: 0.74617267, Cur Avg Loss: 0.55953036, Log Avg loss: 0.66400328, Global Avg Loss: 1.59692700, Time: 0.0112 Steps: 59850, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000248, Sample Num: 3968, Cur Loss: 0.70165539, Cur Avg Loss: 0.56211522, Log Avg loss: 0.62363499, Global Avg Loss: 1.59676440, Time: 0.0117 Steps: 59860, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000258, Sample Num: 4128, Cur Loss: 0.22040682, Cur Avg Loss: 0.55701223, Log Avg loss: 0.43045803, Global Avg Loss: 1.59656960, Time: 0.0106 Steps: 59870, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000268, Sample Num: 4288, Cur Loss: 0.46023339, Cur Avg Loss: 0.55695486, Log Avg loss: 0.55547468, Global Avg Loss: 1.59639573, Time: 0.0112 Steps: 59880, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000278, Sample Num: 4448, Cur Loss: 1.22108150, Cur Avg Loss: 0.55800907, Log Avg loss: 0.58626190, Global Avg Loss: 1.59622707, Time: 0.0123 Steps: 59890, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000288, Sample Num: 4608, Cur Loss: 0.42436555, Cur Avg Loss: 0.55890979, Log Avg loss: 0.58394985, Global Avg Loss: 1.59605807, Time: 0.0114 Steps: 59900, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000298, Sample Num: 4768, Cur Loss: 0.43169284, Cur Avg Loss: 0.55269448, Log Avg loss: 0.37369350, Global Avg Loss: 1.59585404, Time: 0.0226 Steps: 59910, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000308, Sample Num: 4928, Cur Loss: 0.33233991, Cur Avg Loss: 0.55164453, Log Avg loss: 0.52035618, Global Avg Loss: 1.59567455, Time: 0.0123 Steps: 59920, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000318, Sample Num: 5088, Cur Loss: 0.57248443, Cur Avg Loss: 0.55157911, Log Avg loss: 0.54956423, Global Avg Loss: 1.59549999, Time: 0.0067 Steps: 59930, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000328, Sample Num: 5248, Cur Loss: 0.32791871, Cur Avg Loss: 0.55124993, Log Avg loss: 0.54078203, Global Avg Loss: 1.59532403, Time: 0.0115 Steps: 59940, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000338, Sample Num: 5408, Cur Loss: 0.55348140, Cur Avg Loss: 0.55124809, Log Avg loss: 0.55118762, Global Avg Loss: 1.59514986, Time: 0.0093 Steps: 59950, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000348, Sample Num: 5568, Cur Loss: 0.60082012, Cur Avg Loss: 0.55056239, Log Avg loss: 0.52738569, Global Avg Loss: 1.59497179, Time: 0.0156 Steps: 59960, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000358, Sample Num: 5728, Cur Loss: 0.56045365, Cur Avg Loss: 0.55352520, Log Avg loss: 0.65663104, Global Avg Loss: 1.59481532, Time: 0.0118 Steps: 59970, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000368, Sample Num: 5888, Cur Loss: 0.96218175, Cur Avg Loss: 0.55149088, Log Avg loss: 0.47866206, Global Avg Loss: 1.59462923, Time: 0.0086 Steps: 59980, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000378, Sample Num: 6048, Cur Loss: 1.10382712, Cur Avg Loss: 0.55125384, Log Avg loss: 0.54253071, Global Avg Loss: 1.59445385, Time: 0.0073 Steps: 59990, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000388, Sample Num: 6208, Cur Loss: 0.53910768, Cur Avg Loss: 0.55318826, Log Avg loss: 0.62630935, Global Avg Loss: 1.59429249, Time: 0.0105 Steps: 60000, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000398, Sample Num: 6368, Cur Loss: 0.44515377, Cur Avg Loss: 0.55177150, Log Avg loss: 0.49680134, Global Avg Loss: 1.59410961, Time: 0.0108 Steps: 60010, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000408, Sample Num: 6528, Cur Loss: 1.29735911, Cur Avg Loss: 0.55232598, Log Avg loss: 0.57439442, Global Avg Loss: 1.59393971, Time: 0.0068 Steps: 60020, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000418, Sample Num: 6688, Cur Loss: 1.16980672, Cur Avg Loss: 0.55249390, Log Avg loss: 0.55934499, Global Avg Loss: 1.59376737, Time: 0.0138 Steps: 60030, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000428, Sample Num: 6848, Cur Loss: 0.45122209, Cur Avg Loss: 0.55036634, Log Avg loss: 0.46143422, Global Avg Loss: 1.59357877, Time: 0.0066 Steps: 60040, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000438, Sample Num: 7008, Cur Loss: 0.37700790, Cur Avg Loss: 0.55195613, Log Avg loss: 0.61999906, Global Avg Loss: 1.59341664, Time: 0.0066 Steps: 60050, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000448, Sample Num: 7168, Cur Loss: 0.27346501, Cur Avg Loss: 0.55390298, Log Avg loss: 0.63917535, Global Avg Loss: 1.59325776, Time: 0.0068 Steps: 60060, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000458, Sample Num: 7328, Cur Loss: 0.46083409, Cur Avg Loss: 0.55276454, Log Avg loss: 0.50176242, Global Avg Loss: 1.59307606, Time: 0.0111 Steps: 60070, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000468, Sample Num: 7488, Cur Loss: 1.04547644, Cur Avg Loss: 0.55718356, Log Avg loss: 0.75957456, Global Avg Loss: 1.59293732, Time: 0.0065 Steps: 60080, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000478, Sample Num: 7648, Cur Loss: 0.68663603, Cur Avg Loss: 0.55878597, Log Avg loss: 0.63377856, Global Avg Loss: 1.59277770, Time: 0.0126 Steps: 60090, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000488, Sample Num: 7808, Cur Loss: 0.94365501, Cur Avg Loss: 0.55865203, Log Avg loss: 0.55224981, Global Avg Loss: 1.59260457, Time: 0.0105 Steps: 60100, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000498, Sample Num: 7968, Cur Loss: 0.48738378, Cur Avg Loss: 0.55918798, Log Avg loss: 0.58534231, Global Avg Loss: 1.59243700, Time: 0.0066 Steps: 60110, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000508, Sample Num: 8128, Cur Loss: 0.17711207, Cur Avg Loss: 0.56039387, Log Avg loss: 0.62044704, Global Avg Loss: 1.59227533, Time: 0.0134 Steps: 60120, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000518, Sample Num: 8288, Cur Loss: 0.88755548, Cur Avg Loss: 0.56264899, Log Avg loss: 0.67720912, Global Avg Loss: 1.59212315, Time: 0.0075 Steps: 60130, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000528, Sample Num: 8448, Cur Loss: 0.76811492, Cur Avg Loss: 0.55992192, Log Avg loss: 0.41865985, Global Avg Loss: 1.59192802, Time: 0.0139 Steps: 60140, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000538, Sample Num: 8608, Cur Loss: 0.71770591, Cur Avg Loss: 0.56104755, Log Avg loss: 0.62048058, Global Avg Loss: 1.59176652, Time: 0.0067 Steps: 60150, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000548, Sample Num: 8768, Cur Loss: 0.63729656, Cur Avg Loss: 0.56358187, Log Avg loss: 0.69992862, Global Avg Loss: 1.59161827, Time: 0.0112 Steps: 60160, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000558, Sample Num: 8928, Cur Loss: 0.74698704, Cur Avg Loss: 0.56317889, Log Avg loss: 0.54109565, Global Avg Loss: 1.59144368, Time: 0.0124 Steps: 60170, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000568, Sample Num: 9088, Cur Loss: 0.97652698, Cur Avg Loss: 0.56334096, Log Avg loss: 0.57238417, Global Avg Loss: 1.59127435, Time: 0.0106 Steps: 60180, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000578, Sample Num: 9248, Cur Loss: 0.31275076, Cur Avg Loss: 0.56549636, Log Avg loss: 0.68792299, Global Avg Loss: 1.59112426, Time: 0.0232 Steps: 60190, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000588, Sample Num: 9408, Cur Loss: 0.36819828, Cur Avg Loss: 0.56354493, Log Avg loss: 0.45075278, Global Avg Loss: 1.59093483, Time: 0.0065 Steps: 60200, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000598, Sample Num: 9568, Cur Loss: 0.54437208, Cur Avg Loss: 0.56342789, Log Avg loss: 0.55654582, Global Avg Loss: 1.59076304, Time: 0.0070 Steps: 60210, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000608, Sample Num: 9728, Cur Loss: 0.19716112, Cur Avg Loss: 0.56134010, Log Avg loss: 0.43648992, Global Avg Loss: 1.59057136, Time: 0.0112 Steps: 60220, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000618, Sample Num: 9888, Cur Loss: 0.43009296, Cur Avg Loss: 0.56340664, Log Avg loss: 0.68905216, Global Avg Loss: 1.59042168, Time: 0.0108 Steps: 60230, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000628, Sample Num: 10048, Cur Loss: 0.34881070, Cur Avg Loss: 0.56355528, Log Avg loss: 0.57274149, Global Avg Loss: 1.59025274, Time: 0.0139 Steps: 60240, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000638, Sample Num: 10208, Cur Loss: 0.57809240, Cur Avg Loss: 0.56289287, Log Avg loss: 0.52129332, Global Avg Loss: 1.59007532, Time: 0.0113 Steps: 60250, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000648, Sample Num: 10368, Cur Loss: 0.84501672, Cur Avg Loss: 0.56440614, Log Avg loss: 0.66095288, Global Avg Loss: 1.58992114, Time: 0.0107 Steps: 60260, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000658, Sample Num: 10528, Cur Loss: 1.05509460, Cur Avg Loss: 0.56496828, Log Avg loss: 0.60139512, Global Avg Loss: 1.58975712, Time: 0.0129 Steps: 60270, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000668, Sample Num: 10688, Cur Loss: 0.78730202, Cur Avg Loss: 0.56431349, Log Avg loss: 0.52122853, Global Avg Loss: 1.58957986, Time: 0.0067 Steps: 60280, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000678, Sample Num: 10848, Cur Loss: 0.34084475, Cur Avg Loss: 0.56425324, Log Avg loss: 0.56022839, Global Avg Loss: 1.58940913, Time: 0.0066 Steps: 60290, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000688, Sample Num: 11008, Cur Loss: 0.56974518, Cur Avg Loss: 0.56455339, Log Avg loss: 0.58490372, Global Avg Loss: 1.58924254, Time: 0.0227 Steps: 60300, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000698, Sample Num: 11168, Cur Loss: 0.09869857, Cur Avg Loss: 0.56385974, Log Avg loss: 0.51613658, Global Avg Loss: 1.58906461, Time: 0.0067 Steps: 60310, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000708, Sample Num: 11328, Cur Loss: 0.12184812, Cur Avg Loss: 0.56422432, Log Avg loss: 0.58967144, Global Avg Loss: 1.58889893, Time: 0.0067 Steps: 60320, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000718, Sample Num: 11488, Cur Loss: 0.46105939, Cur Avg Loss: 0.56320676, Log Avg loss: 0.49116373, Global Avg Loss: 1.58871697, Time: 0.0121 Steps: 60330, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000728, Sample Num: 11648, Cur Loss: 0.43297333, Cur Avg Loss: 0.56014138, Log Avg loss: 0.34004723, Global Avg Loss: 1.58851003, Time: 0.0070 Steps: 60340, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000738, Sample Num: 11808, Cur Loss: 0.24059770, Cur Avg Loss: 0.55923175, Log Avg loss: 0.49301038, Global Avg Loss: 1.58832851, Time: 0.0110 Steps: 60350, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000748, Sample Num: 11968, Cur Loss: 0.41594300, Cur Avg Loss: 0.55952677, Log Avg loss: 0.58129970, Global Avg Loss: 1.58816167, Time: 0.0067 Steps: 60360, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000758, Sample Num: 12128, Cur Loss: 0.15707168, Cur Avg Loss: 0.55941379, Log Avg loss: 0.55096264, Global Avg Loss: 1.58798987, Time: 0.0067 Steps: 60370, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000768, Sample Num: 12288, Cur Loss: 0.68999773, Cur Avg Loss: 0.55939681, Log Avg loss: 0.55810964, Global Avg Loss: 1.58781930, Time: 0.0115 Steps: 60380, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000778, Sample Num: 12448, Cur Loss: 0.38266703, Cur Avg Loss: 0.55967724, Log Avg loss: 0.58121459, Global Avg Loss: 1.58765262, Time: 0.0066 Steps: 60390, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000788, Sample Num: 12608, Cur Loss: 0.91513759, Cur Avg Loss: 0.55955586, Log Avg loss: 0.55011235, Global Avg Loss: 1.58748084, Time: 0.0208 Steps: 60400, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000798, Sample Num: 12768, Cur Loss: 0.43787500, Cur Avg Loss: 0.55899318, Log Avg loss: 0.51465429, Global Avg Loss: 1.58730325, Time: 0.0066 Steps: 60410, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000808, Sample Num: 12928, Cur Loss: 1.38431060, Cur Avg Loss: 0.56054320, Log Avg loss: 0.68423406, Global Avg Loss: 1.58715378, Time: 0.0066 Steps: 60420, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000818, Sample Num: 13088, Cur Loss: 0.84840077, Cur Avg Loss: 0.56341064, Log Avg loss: 0.79510000, Global Avg Loss: 1.58702271, Time: 0.0066 Steps: 60430, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000828, Sample Num: 13248, Cur Loss: 0.48010439, Cur Avg Loss: 0.56342996, Log Avg loss: 0.56501042, Global Avg Loss: 1.58685362, Time: 0.0064 Steps: 60440, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000838, Sample Num: 13408, Cur Loss: 0.40018034, Cur Avg Loss: 0.56431669, Log Avg loss: 0.63773763, Global Avg Loss: 1.58669661, Time: 0.0222 Steps: 60450, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000848, Sample Num: 13568, Cur Loss: 0.62329096, Cur Avg Loss: 0.56478945, Log Avg loss: 0.60440724, Global Avg Loss: 1.58653414, Time: 0.0160 Steps: 60460, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000858, Sample Num: 13728, Cur Loss: 0.40802056, Cur Avg Loss: 0.56508885, Log Avg loss: 0.59047761, Global Avg Loss: 1.58636942, Time: 0.0111 Steps: 60470, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000868, Sample Num: 13888, Cur Loss: 0.25686008, Cur Avg Loss: 0.56416145, Log Avg loss: 0.48459043, Global Avg Loss: 1.58618725, Time: 0.0117 Steps: 60480, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000878, Sample Num: 14048, Cur Loss: 0.49441049, Cur Avg Loss: 0.56328161, Log Avg loss: 0.48691158, Global Avg Loss: 1.58600552, Time: 0.0179 Steps: 60490, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000888, Sample Num: 14208, Cur Loss: 1.17187059, Cur Avg Loss: 0.56467499, Log Avg loss: 0.68701389, Global Avg Loss: 1.58585692, Time: 0.0064 Steps: 60500, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000898, Sample Num: 14368, Cur Loss: 0.72708386, Cur Avg Loss: 0.56625058, Log Avg loss: 0.70616282, Global Avg Loss: 1.58571154, Time: 0.0117 Steps: 60510, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000908, Sample Num: 14528, Cur Loss: 0.52377510, Cur Avg Loss: 0.56600122, Log Avg loss: 0.54360940, Global Avg Loss: 1.58553935, Time: 0.0073 Steps: 60520, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000918, Sample Num: 14688, Cur Loss: 0.86346567, Cur Avg Loss: 0.56642366, Log Avg loss: 0.60478101, Global Avg Loss: 1.58537732, Time: 0.0163 Steps: 60530, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000928, Sample Num: 14848, Cur Loss: 1.00289094, Cur Avg Loss: 0.56669279, Log Avg loss: 0.59139877, Global Avg Loss: 1.58521314, Time: 0.0113 Steps: 60540, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000938, Sample Num: 15008, Cur Loss: 1.60513234, Cur Avg Loss: 0.56900754, Log Avg loss: 0.78381632, Global Avg Loss: 1.58508079, Time: 0.0067 Steps: 60550, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000948, Sample Num: 15168, Cur Loss: 0.52164102, Cur Avg Loss: 0.56887131, Log Avg loss: 0.55609290, Global Avg Loss: 1.58491087, Time: 0.0097 Steps: 60560, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000958, Sample Num: 15328, Cur Loss: 0.81359297, Cur Avg Loss: 0.56860305, Log Avg loss: 0.54317241, Global Avg Loss: 1.58473889, Time: 0.0068 Steps: 60570, Updated lr: 0.000044 Training, Epoch: 0029, Batch: 000968, Sample Num: 15488, Cur Loss: 1.23671973, Cur Avg Loss: 0.56854431, Log Avg loss: 0.56291650, Global Avg Loss: 1.58457021, Time: 0.0116 Steps: 60580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000978, Sample Num: 15648, Cur Loss: 0.43412572, Cur Avg Loss: 0.56745964, Log Avg loss: 0.46246392, Global Avg Loss: 1.58438502, Time: 0.0119 Steps: 60590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000988, Sample Num: 15808, Cur Loss: 0.60167718, Cur Avg Loss: 0.56766304, Log Avg loss: 0.58755554, Global Avg Loss: 1.58422052, Time: 0.0064 Steps: 60600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 000998, Sample Num: 15968, Cur Loss: 0.26539800, Cur Avg Loss: 0.56755181, Log Avg loss: 0.55656217, Global Avg Loss: 1.58405097, Time: 0.0111 Steps: 60610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001008, Sample Num: 16128, Cur Loss: 0.83855826, Cur Avg Loss: 0.56799939, Log Avg loss: 0.61266739, Global Avg Loss: 1.58389073, Time: 0.0114 Steps: 60620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001018, Sample Num: 16288, Cur Loss: 1.08702183, Cur Avg Loss: 0.56825711, Log Avg loss: 0.59423577, Global Avg Loss: 1.58372750, Time: 0.0072 Steps: 60630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001028, Sample Num: 16448, Cur Loss: 0.75876868, Cur Avg Loss: 0.56776482, Log Avg loss: 0.51764971, Global Avg Loss: 1.58355169, Time: 0.0111 Steps: 60640, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001038, Sample Num: 16608, Cur Loss: 0.56791818, Cur Avg Loss: 0.56802619, Log Avg loss: 0.59489519, Global Avg Loss: 1.58338868, Time: 0.0067 Steps: 60650, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001048, Sample Num: 16768, Cur Loss: 0.10406938, Cur Avg Loss: 0.56936764, Log Avg loss: 0.70861023, Global Avg Loss: 1.58324447, Time: 0.0113 Steps: 60660, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001058, Sample Num: 16928, Cur Loss: 0.20713627, Cur Avg Loss: 0.56908584, Log Avg loss: 0.53955285, Global Avg Loss: 1.58307245, Time: 0.0065 Steps: 60670, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001068, Sample Num: 17088, Cur Loss: 1.01977849, Cur Avg Loss: 0.56958469, Log Avg loss: 0.62236327, Global Avg Loss: 1.58291412, Time: 0.0187 Steps: 60680, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001078, Sample Num: 17248, Cur Loss: 0.29001307, Cur Avg Loss: 0.56929831, Log Avg loss: 0.53871291, Global Avg Loss: 1.58274207, Time: 0.0123 Steps: 60690, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001088, Sample Num: 17408, Cur Loss: 0.46566695, Cur Avg Loss: 0.56942919, Log Avg loss: 0.58353817, Global Avg Loss: 1.58257745, Time: 0.0112 Steps: 60700, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001098, Sample Num: 17568, Cur Loss: 0.51332837, Cur Avg Loss: 0.56903429, Log Avg loss: 0.52606874, Global Avg Loss: 1.58240343, Time: 0.0068 Steps: 60710, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001108, Sample Num: 17728, Cur Loss: 0.09493464, Cur Avg Loss: 0.56775005, Log Avg loss: 0.42674105, Global Avg Loss: 1.58221310, Time: 0.0122 Steps: 60720, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001118, Sample Num: 17888, Cur Loss: 0.36259255, Cur Avg Loss: 0.56635201, Log Avg loss: 0.41144867, Global Avg Loss: 1.58202032, Time: 0.0067 Steps: 60730, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001128, Sample Num: 18048, Cur Loss: 0.05786110, Cur Avg Loss: 0.56502827, Log Avg loss: 0.41703390, Global Avg Loss: 1.58182852, Time: 0.0072 Steps: 60740, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001138, Sample Num: 18208, Cur Loss: 0.57714111, Cur Avg Loss: 0.56615006, Log Avg loss: 0.69268853, Global Avg Loss: 1.58168216, Time: 0.0119 Steps: 60750, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001148, Sample Num: 18368, Cur Loss: 0.34083575, Cur Avg Loss: 0.56588862, Log Avg loss: 0.53613688, Global Avg Loss: 1.58151008, Time: 0.0068 Steps: 60760, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001158, Sample Num: 18528, Cur Loss: 0.61182529, Cur Avg Loss: 0.56484826, Log Avg loss: 0.44541410, Global Avg Loss: 1.58132313, Time: 0.0065 Steps: 60770, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001168, Sample Num: 18688, Cur Loss: 0.56104660, Cur Avg Loss: 0.56657626, Log Avg loss: 0.76667874, Global Avg Loss: 1.58118910, Time: 0.0101 Steps: 60780, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001178, Sample Num: 18848, Cur Loss: 0.40370587, Cur Avg Loss: 0.56555857, Log Avg loss: 0.44669244, Global Avg Loss: 1.58100248, Time: 0.0111 Steps: 60790, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001188, Sample Num: 19008, Cur Loss: 1.01577210, Cur Avg Loss: 0.56616592, Log Avg loss: 0.63771223, Global Avg Loss: 1.58084733, Time: 0.0137 Steps: 60800, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001198, Sample Num: 19168, Cur Loss: 0.58015180, Cur Avg Loss: 0.56467061, Log Avg loss: 0.38702724, Global Avg Loss: 1.58065101, Time: 0.0090 Steps: 60810, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001208, Sample Num: 19328, Cur Loss: 0.75497276, Cur Avg Loss: 0.56383196, Log Avg loss: 0.46336236, Global Avg Loss: 1.58046731, Time: 0.0066 Steps: 60820, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001218, Sample Num: 19488, Cur Loss: 0.40962049, Cur Avg Loss: 0.56467726, Log Avg loss: 0.66678870, Global Avg Loss: 1.58031710, Time: 0.0109 Steps: 60830, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001228, Sample Num: 19648, Cur Loss: 0.66546297, Cur Avg Loss: 0.56461328, Log Avg loss: 0.55682152, Global Avg Loss: 1.58014888, Time: 0.0127 Steps: 60840, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001238, Sample Num: 19808, Cur Loss: 0.67069280, Cur Avg Loss: 0.56560560, Log Avg loss: 0.68746183, Global Avg Loss: 1.58000217, Time: 0.0125 Steps: 60850, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001248, Sample Num: 19968, Cur Loss: 0.95691741, Cur Avg Loss: 0.56576397, Log Avg loss: 0.58537064, Global Avg Loss: 1.57983874, Time: 0.0066 Steps: 60860, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001258, Sample Num: 20128, Cur Loss: 0.62203300, Cur Avg Loss: 0.56644181, Log Avg loss: 0.65103622, Global Avg Loss: 1.57968616, Time: 0.0068 Steps: 60870, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001268, Sample Num: 20288, Cur Loss: 0.29828441, Cur Avg Loss: 0.56725880, Log Avg loss: 0.67003638, Global Avg Loss: 1.57953674, Time: 0.0134 Steps: 60880, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001278, Sample Num: 20448, Cur Loss: 0.47447443, Cur Avg Loss: 0.56709122, Log Avg loss: 0.54584091, Global Avg Loss: 1.57936698, Time: 0.0071 Steps: 60890, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001288, Sample Num: 20608, Cur Loss: 0.57451069, Cur Avg Loss: 0.56691580, Log Avg loss: 0.54449832, Global Avg Loss: 1.57919705, Time: 0.0069 Steps: 60900, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001298, Sample Num: 20768, Cur Loss: 0.54387808, Cur Avg Loss: 0.56694775, Log Avg loss: 0.57106258, Global Avg Loss: 1.57903153, Time: 0.0067 Steps: 60910, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001308, Sample Num: 20928, Cur Loss: 0.45194435, Cur Avg Loss: 0.56751657, Log Avg loss: 0.64134895, Global Avg Loss: 1.57887761, Time: 0.0069 Steps: 60920, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001318, Sample Num: 21088, Cur Loss: 0.12871610, Cur Avg Loss: 0.56701118, Log Avg loss: 0.50090660, Global Avg Loss: 1.57870069, Time: 0.0182 Steps: 60930, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001328, Sample Num: 21248, Cur Loss: 0.39296412, Cur Avg Loss: 0.56684321, Log Avg loss: 0.54470391, Global Avg Loss: 1.57853102, Time: 0.0070 Steps: 60940, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001338, Sample Num: 21408, Cur Loss: 0.67881584, Cur Avg Loss: 0.56680093, Log Avg loss: 0.56118634, Global Avg Loss: 1.57836410, Time: 0.0064 Steps: 60950, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001348, Sample Num: 21568, Cur Loss: 0.41695184, Cur Avg Loss: 0.56820668, Log Avg loss: 0.75629653, Global Avg Loss: 1.57822925, Time: 0.0063 Steps: 60960, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001358, Sample Num: 21728, Cur Loss: 1.30631709, Cur Avg Loss: 0.56787754, Log Avg loss: 0.52350904, Global Avg Loss: 1.57805626, Time: 0.0074 Steps: 60970, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001368, Sample Num: 21888, Cur Loss: 0.30211335, Cur Avg Loss: 0.56810714, Log Avg loss: 0.59928691, Global Avg Loss: 1.57789575, Time: 0.0064 Steps: 60980, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001378, Sample Num: 22048, Cur Loss: 0.61420453, Cur Avg Loss: 0.56836349, Log Avg loss: 0.60343290, Global Avg Loss: 1.57773598, Time: 0.0072 Steps: 60990, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001388, Sample Num: 22208, Cur Loss: 0.49316275, Cur Avg Loss: 0.56820791, Log Avg loss: 0.54676901, Global Avg Loss: 1.57756697, Time: 0.0067 Steps: 61000, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001398, Sample Num: 22368, Cur Loss: 1.02311778, Cur Avg Loss: 0.56964674, Log Avg loss: 0.76935629, Global Avg Loss: 1.57743450, Time: 0.0068 Steps: 61010, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001408, Sample Num: 22528, Cur Loss: 0.39115137, Cur Avg Loss: 0.56994905, Log Avg loss: 0.61221191, Global Avg Loss: 1.57727632, Time: 0.0071 Steps: 61020, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001418, Sample Num: 22688, Cur Loss: 0.30955616, Cur Avg Loss: 0.56978703, Log Avg loss: 0.54697359, Global Avg Loss: 1.57710750, Time: 0.0069 Steps: 61030, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001428, Sample Num: 22848, Cur Loss: 0.35941023, Cur Avg Loss: 0.57016172, Log Avg loss: 0.62329358, Global Avg Loss: 1.57695124, Time: 0.0118 Steps: 61040, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001438, Sample Num: 23008, Cur Loss: 0.36341900, Cur Avg Loss: 0.57037522, Log Avg loss: 0.60086331, Global Avg Loss: 1.57679135, Time: 0.0215 Steps: 61050, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001448, Sample Num: 23168, Cur Loss: 0.30431098, Cur Avg Loss: 0.56997982, Log Avg loss: 0.51312026, Global Avg Loss: 1.57661715, Time: 0.0070 Steps: 61060, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001458, Sample Num: 23328, Cur Loss: 1.11016774, Cur Avg Loss: 0.57083087, Log Avg loss: 0.69406348, Global Avg Loss: 1.57647264, Time: 0.0071 Steps: 61070, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001468, Sample Num: 23488, Cur Loss: 0.61792946, Cur Avg Loss: 0.57067327, Log Avg loss: 0.54769551, Global Avg Loss: 1.57630421, Time: 0.0069 Steps: 61080, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001478, Sample Num: 23648, Cur Loss: 0.41316444, Cur Avg Loss: 0.56919267, Log Avg loss: 0.35184052, Global Avg Loss: 1.57610377, Time: 0.0068 Steps: 61090, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001488, Sample Num: 23808, Cur Loss: 0.75756031, Cur Avg Loss: 0.56863464, Log Avg loss: 0.48615827, Global Avg Loss: 1.57592538, Time: 0.0071 Steps: 61100, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001498, Sample Num: 23968, Cur Loss: 0.34525669, Cur Avg Loss: 0.56824947, Log Avg loss: 0.51093576, Global Avg Loss: 1.57575111, Time: 0.0069 Steps: 61110, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001508, Sample Num: 24128, Cur Loss: 0.31031993, Cur Avg Loss: 0.56741771, Log Avg loss: 0.44282007, Global Avg Loss: 1.57556575, Time: 0.0073 Steps: 61120, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001518, Sample Num: 24288, Cur Loss: 1.33148265, Cur Avg Loss: 0.56812947, Log Avg loss: 0.67546313, Global Avg Loss: 1.57541850, Time: 0.0117 Steps: 61130, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001528, Sample Num: 24448, Cur Loss: 0.47018608, Cur Avg Loss: 0.56756150, Log Avg loss: 0.48134378, Global Avg Loss: 1.57523956, Time: 0.0065 Steps: 61140, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001538, Sample Num: 24608, Cur Loss: 0.43677986, Cur Avg Loss: 0.56619266, Log Avg loss: 0.35703387, Global Avg Loss: 1.57504034, Time: 0.0074 Steps: 61150, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001548, Sample Num: 24768, Cur Loss: 0.78424752, Cur Avg Loss: 0.56682220, Log Avg loss: 0.66364451, Global Avg Loss: 1.57489132, Time: 0.0226 Steps: 61160, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001558, Sample Num: 24928, Cur Loss: 1.14990711, Cur Avg Loss: 0.56794762, Log Avg loss: 0.74216314, Global Avg Loss: 1.57475519, Time: 0.0129 Steps: 61170, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001568, Sample Num: 25088, Cur Loss: 0.31600595, Cur Avg Loss: 0.56730033, Log Avg loss: 0.46645218, Global Avg Loss: 1.57457403, Time: 0.0068 Steps: 61180, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001578, Sample Num: 25248, Cur Loss: 1.12500811, Cur Avg Loss: 0.56968288, Log Avg loss: 0.94326691, Global Avg Loss: 1.57447086, Time: 0.0154 Steps: 61190, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001588, Sample Num: 25408, Cur Loss: 1.11791885, Cur Avg Loss: 0.57053822, Log Avg loss: 0.70551137, Global Avg Loss: 1.57432888, Time: 0.0123 Steps: 61200, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001598, Sample Num: 25568, Cur Loss: 0.44196278, Cur Avg Loss: 0.57088338, Log Avg loss: 0.62569412, Global Avg Loss: 1.57417390, Time: 0.0068 Steps: 61210, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001608, Sample Num: 25728, Cur Loss: 0.15304115, Cur Avg Loss: 0.57073753, Log Avg loss: 0.54743033, Global Avg Loss: 1.57400618, Time: 0.0079 Steps: 61220, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001618, Sample Num: 25888, Cur Loss: 0.29918337, Cur Avg Loss: 0.57109118, Log Avg loss: 0.62795926, Global Avg Loss: 1.57385168, Time: 0.0144 Steps: 61230, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001628, Sample Num: 26048, Cur Loss: 0.23981899, Cur Avg Loss: 0.57017322, Log Avg loss: 0.42164600, Global Avg Loss: 1.57366353, Time: 0.0083 Steps: 61240, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001638, Sample Num: 26208, Cur Loss: 1.17951703, Cur Avg Loss: 0.57012358, Log Avg loss: 0.56204317, Global Avg Loss: 1.57349837, Time: 0.0117 Steps: 61250, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001648, Sample Num: 26368, Cur Loss: 0.33306259, Cur Avg Loss: 0.56930820, Log Avg loss: 0.43574887, Global Avg Loss: 1.57331264, Time: 0.0069 Steps: 61260, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001658, Sample Num: 26528, Cur Loss: 0.29796416, Cur Avg Loss: 0.56936397, Log Avg loss: 0.57855519, Global Avg Loss: 1.57315029, Time: 0.0078 Steps: 61270, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001668, Sample Num: 26688, Cur Loss: 1.19916737, Cur Avg Loss: 0.56947817, Log Avg loss: 0.58841239, Global Avg Loss: 1.57298959, Time: 0.0074 Steps: 61280, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001678, Sample Num: 26848, Cur Loss: 0.68266463, Cur Avg Loss: 0.56957528, Log Avg loss: 0.58577342, Global Avg Loss: 1.57282852, Time: 0.0094 Steps: 61290, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001688, Sample Num: 27008, Cur Loss: 0.47172284, Cur Avg Loss: 0.56969734, Log Avg loss: 0.59017816, Global Avg Loss: 1.57266822, Time: 0.0124 Steps: 61300, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001698, Sample Num: 27168, Cur Loss: 0.39215997, Cur Avg Loss: 0.57027988, Log Avg loss: 0.66861216, Global Avg Loss: 1.57252076, Time: 0.0068 Steps: 61310, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001708, Sample Num: 27328, Cur Loss: 0.27515921, Cur Avg Loss: 0.56959424, Log Avg loss: 0.45317354, Global Avg Loss: 1.57233822, Time: 0.0068 Steps: 61320, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001718, Sample Num: 27488, Cur Loss: 0.28283906, Cur Avg Loss: 0.56918797, Log Avg loss: 0.49979693, Global Avg Loss: 1.57216334, Time: 0.0071 Steps: 61330, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001728, Sample Num: 27648, Cur Loss: 0.52692747, Cur Avg Loss: 0.56953686, Log Avg loss: 0.62947582, Global Avg Loss: 1.57200965, Time: 0.0190 Steps: 61340, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001738, Sample Num: 27808, Cur Loss: 0.21769565, Cur Avg Loss: 0.56937226, Log Avg loss: 0.54093011, Global Avg Loss: 1.57184159, Time: 0.0107 Steps: 61350, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001748, Sample Num: 27968, Cur Loss: 0.60430396, Cur Avg Loss: 0.56931926, Log Avg loss: 0.56010811, Global Avg Loss: 1.57167670, Time: 0.0111 Steps: 61360, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001758, Sample Num: 28128, Cur Loss: 0.17515954, Cur Avg Loss: 0.56973186, Log Avg loss: 0.64185422, Global Avg Loss: 1.57152519, Time: 0.0135 Steps: 61370, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001768, Sample Num: 28288, Cur Loss: 0.64352691, Cur Avg Loss: 0.57021879, Log Avg loss: 0.65582056, Global Avg Loss: 1.57137601, Time: 0.0068 Steps: 61380, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001778, Sample Num: 28448, Cur Loss: 0.28955334, Cur Avg Loss: 0.56967793, Log Avg loss: 0.47405329, Global Avg Loss: 1.57119726, Time: 0.0118 Steps: 61390, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001788, Sample Num: 28608, Cur Loss: 0.26937738, Cur Avg Loss: 0.56938621, Log Avg loss: 0.51751959, Global Avg Loss: 1.57102565, Time: 0.0073 Steps: 61400, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001798, Sample Num: 28768, Cur Loss: 0.51971388, Cur Avg Loss: 0.56897281, Log Avg loss: 0.49505589, Global Avg Loss: 1.57085044, Time: 0.0101 Steps: 61410, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001808, Sample Num: 28928, Cur Loss: 0.49475729, Cur Avg Loss: 0.56926717, Log Avg loss: 0.62219354, Global Avg Loss: 1.57069599, Time: 0.0068 Steps: 61420, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001818, Sample Num: 29088, Cur Loss: 0.61251533, Cur Avg Loss: 0.57021816, Log Avg loss: 0.74215738, Global Avg Loss: 1.57056111, Time: 0.0066 Steps: 61430, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001828, Sample Num: 29248, Cur Loss: 0.44028056, Cur Avg Loss: 0.57033327, Log Avg loss: 0.59125925, Global Avg Loss: 1.57040172, Time: 0.0110 Steps: 61440, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001838, Sample Num: 29408, Cur Loss: 0.25222597, Cur Avg Loss: 0.57004311, Log Avg loss: 0.51700253, Global Avg Loss: 1.57023030, Time: 0.0071 Steps: 61450, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001848, Sample Num: 29568, Cur Loss: 0.40958917, Cur Avg Loss: 0.56974823, Log Avg loss: 0.51554892, Global Avg Loss: 1.57005869, Time: 0.0067 Steps: 61460, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001858, Sample Num: 29728, Cur Loss: 0.26046759, Cur Avg Loss: 0.56920859, Log Avg loss: 0.46948421, Global Avg Loss: 1.56987965, Time: 0.0118 Steps: 61470, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001868, Sample Num: 29888, Cur Loss: 0.96570659, Cur Avg Loss: 0.56955067, Log Avg loss: 0.63310888, Global Avg Loss: 1.56972728, Time: 0.0108 Steps: 61480, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001878, Sample Num: 30048, Cur Loss: 0.60623747, Cur Avg Loss: 0.56923481, Log Avg loss: 0.51023248, Global Avg Loss: 1.56955498, Time: 0.0063 Steps: 61490, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001888, Sample Num: 30208, Cur Loss: 0.50347972, Cur Avg Loss: 0.56875365, Log Avg loss: 0.47839130, Global Avg Loss: 1.56937755, Time: 0.0067 Steps: 61500, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001898, Sample Num: 30368, Cur Loss: 1.32605124, Cur Avg Loss: 0.56863431, Log Avg loss: 0.54610248, Global Avg Loss: 1.56921119, Time: 0.0066 Steps: 61510, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001908, Sample Num: 30528, Cur Loss: 0.39110065, Cur Avg Loss: 0.56818759, Log Avg loss: 0.48339976, Global Avg Loss: 1.56903470, Time: 0.0071 Steps: 61520, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001918, Sample Num: 30688, Cur Loss: 0.45304242, Cur Avg Loss: 0.56861789, Log Avg loss: 0.65071974, Global Avg Loss: 1.56888545, Time: 0.0067 Steps: 61530, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001928, Sample Num: 30848, Cur Loss: 0.61128473, Cur Avg Loss: 0.56922577, Log Avg loss: 0.68581717, Global Avg Loss: 1.56874195, Time: 0.0111 Steps: 61540, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001938, Sample Num: 31008, Cur Loss: 0.33498704, Cur Avg Loss: 0.56912085, Log Avg loss: 0.54889283, Global Avg Loss: 1.56857626, Time: 0.0117 Steps: 61550, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001948, Sample Num: 31168, Cur Loss: 0.63488758, Cur Avg Loss: 0.56925250, Log Avg loss: 0.59476640, Global Avg Loss: 1.56841807, Time: 0.0067 Steps: 61560, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001958, Sample Num: 31328, Cur Loss: 0.59904087, Cur Avg Loss: 0.56932427, Log Avg loss: 0.58330429, Global Avg Loss: 1.56825807, Time: 0.0069 Steps: 61570, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001968, Sample Num: 31488, Cur Loss: 0.24212456, Cur Avg Loss: 0.56839681, Log Avg loss: 0.38680092, Global Avg Loss: 1.56806621, Time: 0.0223 Steps: 61580, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001978, Sample Num: 31648, Cur Loss: 0.36954290, Cur Avg Loss: 0.56878482, Log Avg loss: 0.64514395, Global Avg Loss: 1.56791636, Time: 0.0136 Steps: 61590, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001988, Sample Num: 31808, Cur Loss: 0.63957709, Cur Avg Loss: 0.56886515, Log Avg loss: 0.58475528, Global Avg Loss: 1.56775676, Time: 0.0071 Steps: 61600, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 001998, Sample Num: 31968, Cur Loss: 0.31383768, Cur Avg Loss: 0.56772836, Log Avg loss: 0.34173356, Global Avg Loss: 1.56755776, Time: 0.0202 Steps: 61610, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002008, Sample Num: 32128, Cur Loss: 0.39895251, Cur Avg Loss: 0.56758353, Log Avg loss: 0.53864773, Global Avg Loss: 1.56739079, Time: 0.0073 Steps: 61620, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002018, Sample Num: 32288, Cur Loss: 1.05303872, Cur Avg Loss: 0.56788191, Log Avg loss: 0.62779675, Global Avg Loss: 1.56723833, Time: 0.0146 Steps: 61630, Updated lr: 0.000043 Training, Epoch: 0029, Batch: 002028, Sample Num: 32448, Cur Loss: 1.09004390, Cur Avg Loss: 0.56858974, Log Avg loss: 0.71142919, Global Avg Loss: 1.56709949, Time: 0.0158 Steps: 61640, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002038, Sample Num: 32608, Cur Loss: 0.66027701, Cur Avg Loss: 0.56927682, Log Avg loss: 0.70861655, Global Avg Loss: 1.56696024, Time: 0.0063 Steps: 61650, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002048, Sample Num: 32768, Cur Loss: 0.25065011, Cur Avg Loss: 0.56889876, Log Avg loss: 0.49185091, Global Avg Loss: 1.56678588, Time: 0.0086 Steps: 61660, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002058, Sample Num: 32928, Cur Loss: 0.15375522, Cur Avg Loss: 0.56834212, Log Avg loss: 0.45434194, Global Avg Loss: 1.56660549, Time: 0.0070 Steps: 61670, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002068, Sample Num: 33088, Cur Loss: 1.65552199, Cur Avg Loss: 0.56850134, Log Avg loss: 0.60126885, Global Avg Loss: 1.56644898, Time: 0.0111 Steps: 61680, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002078, Sample Num: 33248, Cur Loss: 0.75327510, Cur Avg Loss: 0.56813234, Log Avg loss: 0.49182174, Global Avg Loss: 1.56627479, Time: 0.0119 Steps: 61690, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002088, Sample Num: 33408, Cur Loss: 0.44777751, Cur Avg Loss: 0.56774616, Log Avg loss: 0.48749889, Global Avg Loss: 1.56609994, Time: 0.0160 Steps: 61700, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002098, Sample Num: 33568, Cur Loss: 0.38784611, Cur Avg Loss: 0.56741953, Log Avg loss: 0.49921815, Global Avg Loss: 1.56592706, Time: 0.0127 Steps: 61710, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002108, Sample Num: 33728, Cur Loss: 0.26370239, Cur Avg Loss: 0.56766534, Log Avg loss: 0.61923689, Global Avg Loss: 1.56577367, Time: 0.0071 Steps: 61720, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002118, Sample Num: 33888, Cur Loss: 0.47294500, Cur Avg Loss: 0.56716088, Log Avg loss: 0.46082079, Global Avg Loss: 1.56559467, Time: 0.0068 Steps: 61730, Updated lr: 0.000042 Training, Epoch: 0029, Batch: 002128, Sample Num: 34048, Cur Loss: 0.14413038, Cur Avg Loss: 0.56682631, Log Avg loss: 0.49596461, Global Avg Loss: 1.56542143, Time: 0.0068 Steps: 61740, Updated lr: 0.000042 ***** Running evaluation checkpoint-61741 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-61741 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.617841, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.653783, "eval_total_loss": 459.609376, "eval_mae": 0.632403, "eval_mse": 0.653917, "eval_r2": 0.584327, "eval_sp_statistic": 0.721333, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.774242, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.917951, "test_total_loss": 460.811624, "test_mae": 0.75275, "test_mse": 0.918095, "test_r2": 0.407453, "test_sp_statistic": 0.558599, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.64833, "test_ps_pvalue": 0.0, "lr": 4.239829302987198e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.565412681531264, "train_cur_epoch_loss": 1207.231829866767, "train_cur_epoch_avg_loss": 0.567041723751417, "train_cur_epoch_time": 22.61784052848816, "train_cur_epoch_avg_time": 0.010623692122352353, "epoch": 29, "step": 61741} ################################################## Training, Epoch: 0030, Batch: 000009, Sample Num: 144, Cur Loss: 0.44505650, Cur Avg Loss: 0.42990557, Log Avg loss: 0.48945916, Global Avg Loss: 1.56524718, Time: 0.0124 Steps: 61750, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000019, Sample Num: 304, Cur Loss: 0.17519006, Cur Avg Loss: 0.48040147, Log Avg loss: 0.52584778, Global Avg Loss: 1.56507889, Time: 0.0110 Steps: 61760, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000029, Sample Num: 464, Cur Loss: 0.61167181, Cur Avg Loss: 0.46208997, Log Avg loss: 0.42729813, Global Avg Loss: 1.56489469, Time: 0.0158 Steps: 61770, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000039, Sample Num: 624, Cur Loss: 0.91042769, Cur Avg Loss: 0.50934115, Log Avg loss: 0.64636958, Global Avg Loss: 1.56474601, Time: 0.0069 Steps: 61780, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000049, Sample Num: 784, Cur Loss: 0.51838732, Cur Avg Loss: 0.52471602, Log Avg loss: 0.58467802, Global Avg Loss: 1.56458740, Time: 0.0079 Steps: 61790, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000059, Sample Num: 944, Cur Loss: 0.92795348, Cur Avg Loss: 0.52401049, Log Avg loss: 0.52055338, Global Avg Loss: 1.56441846, Time: 0.0078 Steps: 61800, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000069, Sample Num: 1104, Cur Loss: 0.49082962, Cur Avg Loss: 0.52780255, Log Avg loss: 0.55017571, Global Avg Loss: 1.56425437, Time: 0.0108 Steps: 61810, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000079, Sample Num: 1264, Cur Loss: 0.28195465, Cur Avg Loss: 0.53171788, Log Avg loss: 0.55873368, Global Avg Loss: 1.56409172, Time: 0.0118 Steps: 61820, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000089, Sample Num: 1424, Cur Loss: 0.61568135, Cur Avg Loss: 0.52750785, Log Avg loss: 0.49424862, Global Avg Loss: 1.56391869, Time: 0.0070 Steps: 61830, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000099, Sample Num: 1584, Cur Loss: 1.39230204, Cur Avg Loss: 0.53434269, Log Avg loss: 0.59517274, Global Avg Loss: 1.56376204, Time: 0.0113 Steps: 61840, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000109, Sample Num: 1744, Cur Loss: 0.59707165, Cur Avg Loss: 0.52706727, Log Avg loss: 0.45504057, Global Avg Loss: 1.56358278, Time: 0.0118 Steps: 61850, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000119, Sample Num: 1904, Cur Loss: 0.39997020, Cur Avg Loss: 0.52953454, Log Avg loss: 0.55642776, Global Avg Loss: 1.56341996, Time: 0.0066 Steps: 61860, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000129, Sample Num: 2064, Cur Loss: 0.37773731, Cur Avg Loss: 0.53484016, Log Avg loss: 0.59797708, Global Avg Loss: 1.56326392, Time: 0.0068 Steps: 61870, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000139, Sample Num: 2224, Cur Loss: 0.54873466, Cur Avg Loss: 0.53862853, Log Avg loss: 0.58749846, Global Avg Loss: 1.56310623, Time: 0.0067 Steps: 61880, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000149, Sample Num: 2384, Cur Loss: 0.41188937, Cur Avg Loss: 0.53477819, Log Avg loss: 0.48125848, Global Avg Loss: 1.56293143, Time: 0.0115 Steps: 61890, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000159, Sample Num: 2544, Cur Loss: 0.41060737, Cur Avg Loss: 0.53534696, Log Avg loss: 0.54382169, Global Avg Loss: 1.56276679, Time: 0.0114 Steps: 61900, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000169, Sample Num: 2704, Cur Loss: 0.46653175, Cur Avg Loss: 0.54601834, Log Avg loss: 0.71569319, Global Avg Loss: 1.56262997, Time: 0.0161 Steps: 61910, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000179, Sample Num: 2864, Cur Loss: 0.68785906, Cur Avg Loss: 0.54956407, Log Avg loss: 0.60948694, Global Avg Loss: 1.56247604, Time: 0.0118 Steps: 61920, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000189, Sample Num: 3024, Cur Loss: 1.05089259, Cur Avg Loss: 0.54749727, Log Avg loss: 0.51050158, Global Avg Loss: 1.56230617, Time: 0.0120 Steps: 61930, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000199, Sample Num: 3184, Cur Loss: 0.31832162, Cur Avg Loss: 0.54180002, Log Avg loss: 0.43412196, Global Avg Loss: 1.56212403, Time: 0.0071 Steps: 61940, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000209, Sample Num: 3344, Cur Loss: 0.38288322, Cur Avg Loss: 0.53943960, Log Avg loss: 0.49246724, Global Avg Loss: 1.56195137, Time: 0.0122 Steps: 61950, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000219, Sample Num: 3504, Cur Loss: 0.81425428, Cur Avg Loss: 0.54061332, Log Avg loss: 0.56514403, Global Avg Loss: 1.56179049, Time: 0.0097 Steps: 61960, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000229, Sample Num: 3664, Cur Loss: 0.40968907, Cur Avg Loss: 0.53992415, Log Avg loss: 0.52483143, Global Avg Loss: 1.56162316, Time: 0.0066 Steps: 61970, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000239, Sample Num: 3824, Cur Loss: 1.78617322, Cur Avg Loss: 0.54865299, Log Avg loss: 0.74854338, Global Avg Loss: 1.56149197, Time: 0.0118 Steps: 61980, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000249, Sample Num: 3984, Cur Loss: 0.41448838, Cur Avg Loss: 0.54610952, Log Avg loss: 0.48532065, Global Avg Loss: 1.56131837, Time: 0.0122 Steps: 61990, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000259, Sample Num: 4144, Cur Loss: 0.50705510, Cur Avg Loss: 0.54630110, Log Avg loss: 0.55107142, Global Avg Loss: 1.56115543, Time: 0.0073 Steps: 62000, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000269, Sample Num: 4304, Cur Loss: 0.59933770, Cur Avg Loss: 0.55443524, Log Avg loss: 0.76510940, Global Avg Loss: 1.56102705, Time: 0.0106 Steps: 62010, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000279, Sample Num: 4464, Cur Loss: 0.71074605, Cur Avg Loss: 0.55152600, Log Avg loss: 0.47326743, Global Avg Loss: 1.56085166, Time: 0.0069 Steps: 62020, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000289, Sample Num: 4624, Cur Loss: 0.41118413, Cur Avg Loss: 0.55077294, Log Avg loss: 0.52976274, Global Avg Loss: 1.56068544, Time: 0.0115 Steps: 62030, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000299, Sample Num: 4784, Cur Loss: 0.30824155, Cur Avg Loss: 0.55021027, Log Avg loss: 0.53394910, Global Avg Loss: 1.56051994, Time: 0.0140 Steps: 62040, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000309, Sample Num: 4944, Cur Loss: 0.93845379, Cur Avg Loss: 0.54826379, Log Avg loss: 0.49006381, Global Avg Loss: 1.56034743, Time: 0.0125 Steps: 62050, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000319, Sample Num: 5104, Cur Loss: 0.49147820, Cur Avg Loss: 0.54998563, Log Avg loss: 0.60319063, Global Avg Loss: 1.56019320, Time: 0.0140 Steps: 62060, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000329, Sample Num: 5264, Cur Loss: 0.62086457, Cur Avg Loss: 0.55335870, Log Avg loss: 0.66095960, Global Avg Loss: 1.56004832, Time: 0.0121 Steps: 62070, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000339, Sample Num: 5424, Cur Loss: 0.47245061, Cur Avg Loss: 0.55310321, Log Avg loss: 0.54469761, Global Avg Loss: 1.55988477, Time: 0.0101 Steps: 62080, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000349, Sample Num: 5584, Cur Loss: 0.35176095, Cur Avg Loss: 0.55050085, Log Avg loss: 0.46228074, Global Avg Loss: 1.55970799, Time: 0.0123 Steps: 62090, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000359, Sample Num: 5744, Cur Loss: 0.53440297, Cur Avg Loss: 0.55074564, Log Avg loss: 0.55928901, Global Avg Loss: 1.55954689, Time: 0.0136 Steps: 62100, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000369, Sample Num: 5904, Cur Loss: 0.87730682, Cur Avg Loss: 0.55165083, Log Avg loss: 0.58414708, Global Avg Loss: 1.55938985, Time: 0.0069 Steps: 62110, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000379, Sample Num: 6064, Cur Loss: 0.43130606, Cur Avg Loss: 0.55018012, Log Avg loss: 0.49591104, Global Avg Loss: 1.55921865, Time: 0.0112 Steps: 62120, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000389, Sample Num: 6224, Cur Loss: 0.39699313, Cur Avg Loss: 0.55020398, Log Avg loss: 0.55110821, Global Avg Loss: 1.55905639, Time: 0.0107 Steps: 62130, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000399, Sample Num: 6384, Cur Loss: 0.30565199, Cur Avg Loss: 0.54830675, Log Avg loss: 0.47450447, Global Avg Loss: 1.55888186, Time: 0.0112 Steps: 62140, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000409, Sample Num: 6544, Cur Loss: 0.60770726, Cur Avg Loss: 0.55054601, Log Avg loss: 0.63989239, Global Avg Loss: 1.55873399, Time: 0.0117 Steps: 62150, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000419, Sample Num: 6704, Cur Loss: 0.27567995, Cur Avg Loss: 0.55090114, Log Avg loss: 0.56542581, Global Avg Loss: 1.55857419, Time: 0.0089 Steps: 62160, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000429, Sample Num: 6864, Cur Loss: 0.40805626, Cur Avg Loss: 0.55170426, Log Avg loss: 0.58535529, Global Avg Loss: 1.55841765, Time: 0.0111 Steps: 62170, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000439, Sample Num: 7024, Cur Loss: 0.48035991, Cur Avg Loss: 0.54949701, Log Avg loss: 0.45480591, Global Avg Loss: 1.55824017, Time: 0.0087 Steps: 62180, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000449, Sample Num: 7184, Cur Loss: 0.88398337, Cur Avg Loss: 0.55030052, Log Avg loss: 0.58557453, Global Avg Loss: 1.55808376, Time: 0.0125 Steps: 62190, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000459, Sample Num: 7344, Cur Loss: 0.29088798, Cur Avg Loss: 0.54674168, Log Avg loss: 0.38694979, Global Avg Loss: 1.55789548, Time: 0.0071 Steps: 62200, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000469, Sample Num: 7504, Cur Loss: 0.74212575, Cur Avg Loss: 0.54937523, Log Avg loss: 0.67025511, Global Avg Loss: 1.55775279, Time: 0.0083 Steps: 62210, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000479, Sample Num: 7664, Cur Loss: 1.19258237, Cur Avg Loss: 0.54971600, Log Avg loss: 0.56569797, Global Avg Loss: 1.55759335, Time: 0.0159 Steps: 62220, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000489, Sample Num: 7824, Cur Loss: 0.30693027, Cur Avg Loss: 0.54959614, Log Avg loss: 0.54385520, Global Avg Loss: 1.55743045, Time: 0.0068 Steps: 62230, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000499, Sample Num: 7984, Cur Loss: 1.62126577, Cur Avg Loss: 0.54935455, Log Avg loss: 0.53754086, Global Avg Loss: 1.55726659, Time: 0.0067 Steps: 62240, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000509, Sample Num: 8144, Cur Loss: 0.24958260, Cur Avg Loss: 0.54779842, Log Avg loss: 0.47014747, Global Avg Loss: 1.55709195, Time: 0.0117 Steps: 62250, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000519, Sample Num: 8304, Cur Loss: 0.19305287, Cur Avg Loss: 0.54667468, Log Avg loss: 0.48947596, Global Avg Loss: 1.55692047, Time: 0.0072 Steps: 62260, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000529, Sample Num: 8464, Cur Loss: 0.79643190, Cur Avg Loss: 0.54643321, Log Avg loss: 0.53390118, Global Avg Loss: 1.55675618, Time: 0.0071 Steps: 62270, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000539, Sample Num: 8624, Cur Loss: 0.53275973, Cur Avg Loss: 0.54660053, Log Avg loss: 0.55545161, Global Avg Loss: 1.55659541, Time: 0.0102 Steps: 62280, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000549, Sample Num: 8784, Cur Loss: 0.37065786, Cur Avg Loss: 0.54462260, Log Avg loss: 0.43801227, Global Avg Loss: 1.55641583, Time: 0.0075 Steps: 62290, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000559, Sample Num: 8944, Cur Loss: 0.43927163, Cur Avg Loss: 0.54250118, Log Avg loss: 0.42603522, Global Avg Loss: 1.55623439, Time: 0.0130 Steps: 62300, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000569, Sample Num: 9104, Cur Loss: 0.65803945, Cur Avg Loss: 0.54520502, Log Avg loss: 0.69634973, Global Avg Loss: 1.55609639, Time: 0.0071 Steps: 62310, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000579, Sample Num: 9264, Cur Loss: 0.91815472, Cur Avg Loss: 0.54524809, Log Avg loss: 0.54769870, Global Avg Loss: 1.55593458, Time: 0.0072 Steps: 62320, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000589, Sample Num: 9424, Cur Loss: 0.24343398, Cur Avg Loss: 0.54513971, Log Avg loss: 0.53886447, Global Avg Loss: 1.55577140, Time: 0.0069 Steps: 62330, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000599, Sample Num: 9584, Cur Loss: 0.73357046, Cur Avg Loss: 0.54705581, Log Avg loss: 0.65991413, Global Avg Loss: 1.55562770, Time: 0.0093 Steps: 62340, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000609, Sample Num: 9744, Cur Loss: 0.18457867, Cur Avg Loss: 0.54720596, Log Avg loss: 0.55620001, Global Avg Loss: 1.55546741, Time: 0.0076 Steps: 62350, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000619, Sample Num: 9904, Cur Loss: 0.72449780, Cur Avg Loss: 0.54689712, Log Avg loss: 0.52808858, Global Avg Loss: 1.55530266, Time: 0.0115 Steps: 62360, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000629, Sample Num: 10064, Cur Loss: 0.95022708, Cur Avg Loss: 0.54627244, Log Avg loss: 0.50760516, Global Avg Loss: 1.55513468, Time: 0.0075 Steps: 62370, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000639, Sample Num: 10224, Cur Loss: 0.75302398, Cur Avg Loss: 0.54549969, Log Avg loss: 0.49689364, Global Avg Loss: 1.55496503, Time: 0.0074 Steps: 62380, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000649, Sample Num: 10384, Cur Loss: 0.54038656, Cur Avg Loss: 0.54617256, Log Avg loss: 0.58916875, Global Avg Loss: 1.55481023, Time: 0.0071 Steps: 62390, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000659, Sample Num: 10544, Cur Loss: 1.09040272, Cur Avg Loss: 0.54517984, Log Avg loss: 0.48075208, Global Avg Loss: 1.55463811, Time: 0.0183 Steps: 62400, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000669, Sample Num: 10704, Cur Loss: 0.68000650, Cur Avg Loss: 0.54380063, Log Avg loss: 0.45291082, Global Avg Loss: 1.55446158, Time: 0.0085 Steps: 62410, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000679, Sample Num: 10864, Cur Loss: 0.34148866, Cur Avg Loss: 0.54341980, Log Avg loss: 0.51794218, Global Avg Loss: 1.55429552, Time: 0.0074 Steps: 62420, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000689, Sample Num: 11024, Cur Loss: 0.35690022, Cur Avg Loss: 0.54423091, Log Avg loss: 0.59930541, Global Avg Loss: 1.55414255, Time: 0.0129 Steps: 62430, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000699, Sample Num: 11184, Cur Loss: 0.42477944, Cur Avg Loss: 0.54516303, Log Avg loss: 0.60938644, Global Avg Loss: 1.55399124, Time: 0.0073 Steps: 62440, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000709, Sample Num: 11344, Cur Loss: 1.16972947, Cur Avg Loss: 0.54481140, Log Avg loss: 0.52023249, Global Avg Loss: 1.55382571, Time: 0.0071 Steps: 62450, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000719, Sample Num: 11504, Cur Loss: 0.69830000, Cur Avg Loss: 0.54482594, Log Avg loss: 0.54585624, Global Avg Loss: 1.55366433, Time: 0.0073 Steps: 62460, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000729, Sample Num: 11664, Cur Loss: 0.70975614, Cur Avg Loss: 0.54375061, Log Avg loss: 0.46643452, Global Avg Loss: 1.55349029, Time: 0.0071 Steps: 62470, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000739, Sample Num: 11824, Cur Loss: 0.74038368, Cur Avg Loss: 0.54294361, Log Avg loss: 0.48411330, Global Avg Loss: 1.55331914, Time: 0.0128 Steps: 62480, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000749, Sample Num: 11984, Cur Loss: 0.22043030, Cur Avg Loss: 0.54342939, Log Avg loss: 0.57932894, Global Avg Loss: 1.55316327, Time: 0.0105 Steps: 62490, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000759, Sample Num: 12144, Cur Loss: 0.41703832, Cur Avg Loss: 0.54357989, Log Avg loss: 0.55485211, Global Avg Loss: 1.55300354, Time: 0.0072 Steps: 62500, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000769, Sample Num: 12304, Cur Loss: 0.32642299, Cur Avg Loss: 0.54232305, Log Avg loss: 0.44692870, Global Avg Loss: 1.55282660, Time: 0.0192 Steps: 62510, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000779, Sample Num: 12464, Cur Loss: 1.21978927, Cur Avg Loss: 0.54474619, Log Avg loss: 0.73108577, Global Avg Loss: 1.55269516, Time: 0.0133 Steps: 62520, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000789, Sample Num: 12624, Cur Loss: 0.35666773, Cur Avg Loss: 0.54572792, Log Avg loss: 0.62220468, Global Avg Loss: 1.55254636, Time: 0.0137 Steps: 62530, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000799, Sample Num: 12784, Cur Loss: 0.24860093, Cur Avg Loss: 0.54446903, Log Avg loss: 0.44514235, Global Avg Loss: 1.55236929, Time: 0.0130 Steps: 62540, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000809, Sample Num: 12944, Cur Loss: 0.45541471, Cur Avg Loss: 0.54372223, Log Avg loss: 0.48405333, Global Avg Loss: 1.55219849, Time: 0.0088 Steps: 62550, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000819, Sample Num: 13104, Cur Loss: 0.28434223, Cur Avg Loss: 0.54300956, Log Avg loss: 0.48535481, Global Avg Loss: 1.55202796, Time: 0.0068 Steps: 62560, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000829, Sample Num: 13264, Cur Loss: 0.47178707, Cur Avg Loss: 0.54379708, Log Avg loss: 0.60829461, Global Avg Loss: 1.55187713, Time: 0.0191 Steps: 62570, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000839, Sample Num: 13424, Cur Loss: 0.78498399, Cur Avg Loss: 0.54571594, Log Avg loss: 0.70478906, Global Avg Loss: 1.55174177, Time: 0.0072 Steps: 62580, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000849, Sample Num: 13584, Cur Loss: 0.69845176, Cur Avg Loss: 0.54609371, Log Avg loss: 0.57778891, Global Avg Loss: 1.55158616, Time: 0.0114 Steps: 62590, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000859, Sample Num: 13744, Cur Loss: 1.16355729, Cur Avg Loss: 0.54629645, Log Avg loss: 0.56350944, Global Avg Loss: 1.55142832, Time: 0.0150 Steps: 62600, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000869, Sample Num: 13904, Cur Loss: 0.82430691, Cur Avg Loss: 0.54665643, Log Avg loss: 0.57757812, Global Avg Loss: 1.55127278, Time: 0.0094 Steps: 62610, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000879, Sample Num: 14064, Cur Loss: 0.67306793, Cur Avg Loss: 0.54881695, Log Avg loss: 0.73656610, Global Avg Loss: 1.55114268, Time: 0.0072 Steps: 62620, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000889, Sample Num: 14224, Cur Loss: 0.41712326, Cur Avg Loss: 0.54958889, Log Avg loss: 0.61744318, Global Avg Loss: 1.55099360, Time: 0.0067 Steps: 62630, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000899, Sample Num: 14384, Cur Loss: 0.50455910, Cur Avg Loss: 0.54863752, Log Avg loss: 0.46406030, Global Avg Loss: 1.55082008, Time: 0.0067 Steps: 62640, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000909, Sample Num: 14544, Cur Loss: 0.86834681, Cur Avg Loss: 0.54942639, Log Avg loss: 0.62034629, Global Avg Loss: 1.55067156, Time: 0.0117 Steps: 62650, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000919, Sample Num: 14704, Cur Loss: 0.59840506, Cur Avg Loss: 0.55109691, Log Avg loss: 0.70294702, Global Avg Loss: 1.55053627, Time: 0.0119 Steps: 62660, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000929, Sample Num: 14864, Cur Loss: 0.99939203, Cur Avg Loss: 0.55142145, Log Avg loss: 0.58124620, Global Avg Loss: 1.55038160, Time: 0.0067 Steps: 62670, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000939, Sample Num: 15024, Cur Loss: 0.23980553, Cur Avg Loss: 0.55039778, Log Avg loss: 0.45529880, Global Avg Loss: 1.55020689, Time: 0.0110 Steps: 62680, Updated lr: 0.000042 Training, Epoch: 0030, Batch: 000949, Sample Num: 15184, Cur Loss: 0.71968061, Cur Avg Loss: 0.55012274, Log Avg loss: 0.52429642, Global Avg Loss: 1.55004324, Time: 0.0065 Steps: 62690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000959, Sample Num: 15344, Cur Loss: 0.59538937, Cur Avg Loss: 0.54851617, Log Avg loss: 0.39605264, Global Avg Loss: 1.54985919, Time: 0.0106 Steps: 62700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000969, Sample Num: 15504, Cur Loss: 0.51591456, Cur Avg Loss: 0.54923722, Log Avg loss: 0.61838616, Global Avg Loss: 1.54971066, Time: 0.0116 Steps: 62710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000979, Sample Num: 15664, Cur Loss: 0.18815622, Cur Avg Loss: 0.54806233, Log Avg loss: 0.43421524, Global Avg Loss: 1.54953280, Time: 0.0114 Steps: 62720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000989, Sample Num: 15824, Cur Loss: 0.81136394, Cur Avg Loss: 0.54940060, Log Avg loss: 0.68041752, Global Avg Loss: 1.54939425, Time: 0.0115 Steps: 62730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 000999, Sample Num: 15984, Cur Loss: 0.61541200, Cur Avg Loss: 0.54897494, Log Avg loss: 0.50687672, Global Avg Loss: 1.54922809, Time: 0.0118 Steps: 62740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001009, Sample Num: 16144, Cur Loss: 0.56938314, Cur Avg Loss: 0.54981868, Log Avg loss: 0.63410896, Global Avg Loss: 1.54908225, Time: 0.0068 Steps: 62750, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001019, Sample Num: 16304, Cur Loss: 1.70935023, Cur Avg Loss: 0.55201872, Log Avg loss: 0.77400222, Global Avg Loss: 1.54895875, Time: 0.0067 Steps: 62760, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001029, Sample Num: 16464, Cur Loss: 0.64379585, Cur Avg Loss: 0.55189601, Log Avg loss: 0.53939181, Global Avg Loss: 1.54879792, Time: 0.0071 Steps: 62770, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001039, Sample Num: 16624, Cur Loss: 0.21437167, Cur Avg Loss: 0.55082596, Log Avg loss: 0.44071805, Global Avg Loss: 1.54862142, Time: 0.0124 Steps: 62780, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001049, Sample Num: 16784, Cur Loss: 0.44723925, Cur Avg Loss: 0.55121476, Log Avg loss: 0.59161105, Global Avg Loss: 1.54846900, Time: 0.0117 Steps: 62790, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001059, Sample Num: 16944, Cur Loss: 0.36056602, Cur Avg Loss: 0.55190292, Log Avg loss: 0.62409161, Global Avg Loss: 1.54832181, Time: 0.0121 Steps: 62800, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001069, Sample Num: 17104, Cur Loss: 0.62476987, Cur Avg Loss: 0.55102610, Log Avg loss: 0.45817024, Global Avg Loss: 1.54814825, Time: 0.0123 Steps: 62810, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001079, Sample Num: 17264, Cur Loss: 0.33254421, Cur Avg Loss: 0.55074718, Log Avg loss: 0.52093039, Global Avg Loss: 1.54798473, Time: 0.0114 Steps: 62820, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001089, Sample Num: 17424, Cur Loss: 0.72586930, Cur Avg Loss: 0.55001598, Log Avg loss: 0.47112027, Global Avg Loss: 1.54781333, Time: 0.0073 Steps: 62830, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001099, Sample Num: 17584, Cur Loss: 1.60206401, Cur Avg Loss: 0.55049227, Log Avg loss: 0.60235959, Global Avg Loss: 1.54766288, Time: 0.0107 Steps: 62840, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001109, Sample Num: 17744, Cur Loss: 1.23177528, Cur Avg Loss: 0.55106656, Log Avg loss: 0.61418102, Global Avg Loss: 1.54751435, Time: 0.0117 Steps: 62850, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001119, Sample Num: 17904, Cur Loss: 0.73713732, Cur Avg Loss: 0.55196615, Log Avg loss: 0.65173066, Global Avg Loss: 1.54737185, Time: 0.0137 Steps: 62860, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001129, Sample Num: 18064, Cur Loss: 0.32686529, Cur Avg Loss: 0.55127362, Log Avg loss: 0.47377980, Global Avg Loss: 1.54720109, Time: 0.0118 Steps: 62870, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001139, Sample Num: 18224, Cur Loss: 0.39919341, Cur Avg Loss: 0.55104931, Log Avg loss: 0.52572519, Global Avg Loss: 1.54703864, Time: 0.0120 Steps: 62880, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001149, Sample Num: 18384, Cur Loss: 0.53849638, Cur Avg Loss: 0.55115599, Log Avg loss: 0.56330680, Global Avg Loss: 1.54688222, Time: 0.0231 Steps: 62890, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001159, Sample Num: 18544, Cur Loss: 0.73592317, Cur Avg Loss: 0.55181925, Log Avg loss: 0.62802754, Global Avg Loss: 1.54673613, Time: 0.0072 Steps: 62900, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001169, Sample Num: 18704, Cur Loss: 0.67441821, Cur Avg Loss: 0.55215199, Log Avg loss: 0.59071631, Global Avg Loss: 1.54658417, Time: 0.0077 Steps: 62910, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001179, Sample Num: 18864, Cur Loss: 1.30400085, Cur Avg Loss: 0.55196702, Log Avg loss: 0.53034380, Global Avg Loss: 1.54642266, Time: 0.0112 Steps: 62920, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001189, Sample Num: 19024, Cur Loss: 0.35514998, Cur Avg Loss: 0.55296907, Log Avg loss: 0.67111136, Global Avg Loss: 1.54628356, Time: 0.0080 Steps: 62930, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001199, Sample Num: 19184, Cur Loss: 0.81346452, Cur Avg Loss: 0.55379593, Log Avg loss: 0.65210990, Global Avg Loss: 1.54614150, Time: 0.0082 Steps: 62940, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001209, Sample Num: 19344, Cur Loss: 1.12810802, Cur Avg Loss: 0.55313689, Log Avg loss: 0.47411777, Global Avg Loss: 1.54597120, Time: 0.0118 Steps: 62950, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001219, Sample Num: 19504, Cur Loss: 0.32351163, Cur Avg Loss: 0.55280969, Log Avg loss: 0.51325148, Global Avg Loss: 1.54580717, Time: 0.0217 Steps: 62960, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001229, Sample Num: 19664, Cur Loss: 0.20812814, Cur Avg Loss: 0.55262217, Log Avg loss: 0.52976234, Global Avg Loss: 1.54564582, Time: 0.0066 Steps: 62970, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001239, Sample Num: 19824, Cur Loss: 0.40227818, Cur Avg Loss: 0.55328657, Log Avg loss: 0.63494151, Global Avg Loss: 1.54550121, Time: 0.0109 Steps: 62980, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001249, Sample Num: 19984, Cur Loss: 0.58305502, Cur Avg Loss: 0.55239864, Log Avg loss: 0.44238411, Global Avg Loss: 1.54532609, Time: 0.0161 Steps: 62990, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001259, Sample Num: 20144, Cur Loss: 0.62478524, Cur Avg Loss: 0.55451182, Log Avg loss: 0.81844826, Global Avg Loss: 1.54521071, Time: 0.0086 Steps: 63000, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001269, Sample Num: 20304, Cur Loss: 0.31048739, Cur Avg Loss: 0.55404934, Log Avg loss: 0.49582300, Global Avg Loss: 1.54504417, Time: 0.0116 Steps: 63010, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001279, Sample Num: 20464, Cur Loss: 0.26315320, Cur Avg Loss: 0.55341518, Log Avg loss: 0.47294036, Global Avg Loss: 1.54487405, Time: 0.0068 Steps: 63020, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001289, Sample Num: 20624, Cur Loss: 0.25670946, Cur Avg Loss: 0.55352115, Log Avg loss: 0.56707496, Global Avg Loss: 1.54471891, Time: 0.0066 Steps: 63030, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001299, Sample Num: 20784, Cur Loss: 0.35677600, Cur Avg Loss: 0.55269957, Log Avg loss: 0.44679728, Global Avg Loss: 1.54454475, Time: 0.0141 Steps: 63040, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001309, Sample Num: 20944, Cur Loss: 1.45603871, Cur Avg Loss: 0.55414912, Log Avg loss: 0.74244632, Global Avg Loss: 1.54441753, Time: 0.0064 Steps: 63050, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001319, Sample Num: 21104, Cur Loss: 0.72542608, Cur Avg Loss: 0.55373627, Log Avg loss: 0.49969364, Global Avg Loss: 1.54425186, Time: 0.0064 Steps: 63060, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001329, Sample Num: 21264, Cur Loss: 0.37534240, Cur Avg Loss: 0.55365752, Log Avg loss: 0.54327082, Global Avg Loss: 1.54409315, Time: 0.0099 Steps: 63070, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001339, Sample Num: 21424, Cur Loss: 0.48378631, Cur Avg Loss: 0.55345565, Log Avg loss: 0.52662648, Global Avg Loss: 1.54393186, Time: 0.0115 Steps: 63080, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001349, Sample Num: 21584, Cur Loss: 0.38484499, Cur Avg Loss: 0.55384603, Log Avg loss: 0.60611874, Global Avg Loss: 1.54378321, Time: 0.0128 Steps: 63090, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001359, Sample Num: 21744, Cur Loss: 0.13231701, Cur Avg Loss: 0.55397481, Log Avg loss: 0.57134651, Global Avg Loss: 1.54362910, Time: 0.0111 Steps: 63100, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001369, Sample Num: 21904, Cur Loss: 0.69133759, Cur Avg Loss: 0.55377757, Log Avg loss: 0.52697372, Global Avg Loss: 1.54346801, Time: 0.0109 Steps: 63110, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001379, Sample Num: 22064, Cur Loss: 0.34674537, Cur Avg Loss: 0.55425380, Log Avg loss: 0.61944881, Global Avg Loss: 1.54332162, Time: 0.0105 Steps: 63120, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001389, Sample Num: 22224, Cur Loss: 0.41800320, Cur Avg Loss: 0.55376168, Log Avg loss: 0.48589909, Global Avg Loss: 1.54315412, Time: 0.0081 Steps: 63130, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001399, Sample Num: 22384, Cur Loss: 0.59390438, Cur Avg Loss: 0.55358297, Log Avg loss: 0.52875932, Global Avg Loss: 1.54299346, Time: 0.0144 Steps: 63140, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001409, Sample Num: 22544, Cur Loss: 1.28880858, Cur Avg Loss: 0.55531212, Log Avg loss: 0.79722082, Global Avg Loss: 1.54287536, Time: 0.0135 Steps: 63150, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001419, Sample Num: 22704, Cur Loss: 1.28106308, Cur Avg Loss: 0.55529860, Log Avg loss: 0.55339332, Global Avg Loss: 1.54271870, Time: 0.0068 Steps: 63160, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001429, Sample Num: 22864, Cur Loss: 0.32076830, Cur Avg Loss: 0.55592916, Log Avg loss: 0.64540583, Global Avg Loss: 1.54257665, Time: 0.0092 Steps: 63170, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001439, Sample Num: 23024, Cur Loss: 0.71771562, Cur Avg Loss: 0.55733347, Log Avg loss: 0.75800927, Global Avg Loss: 1.54245247, Time: 0.0101 Steps: 63180, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001449, Sample Num: 23184, Cur Loss: 1.43705869, Cur Avg Loss: 0.55705229, Log Avg loss: 0.51658989, Global Avg Loss: 1.54229013, Time: 0.0109 Steps: 63190, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001459, Sample Num: 23344, Cur Loss: 0.45588917, Cur Avg Loss: 0.55705524, Log Avg loss: 0.55748401, Global Avg Loss: 1.54213430, Time: 0.0080 Steps: 63200, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001469, Sample Num: 23504, Cur Loss: 0.38620895, Cur Avg Loss: 0.55658757, Log Avg loss: 0.48835442, Global Avg Loss: 1.54196759, Time: 0.0134 Steps: 63210, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001479, Sample Num: 23664, Cur Loss: 0.29023850, Cur Avg Loss: 0.55726153, Log Avg loss: 0.65626534, Global Avg Loss: 1.54182749, Time: 0.0072 Steps: 63220, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001489, Sample Num: 23824, Cur Loss: 0.44210076, Cur Avg Loss: 0.55693216, Log Avg loss: 0.50821820, Global Avg Loss: 1.54166403, Time: 0.0233 Steps: 63230, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001499, Sample Num: 23984, Cur Loss: 0.59661293, Cur Avg Loss: 0.55690675, Log Avg loss: 0.55312391, Global Avg Loss: 1.54150771, Time: 0.0096 Steps: 63240, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001509, Sample Num: 24144, Cur Loss: 0.40971577, Cur Avg Loss: 0.55569306, Log Avg loss: 0.37376105, Global Avg Loss: 1.54132309, Time: 0.0132 Steps: 63250, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001519, Sample Num: 24304, Cur Loss: 0.23785126, Cur Avg Loss: 0.55618050, Log Avg loss: 0.62973447, Global Avg Loss: 1.54117898, Time: 0.0116 Steps: 63260, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001529, Sample Num: 24464, Cur Loss: 0.64697003, Cur Avg Loss: 0.55532534, Log Avg loss: 0.42542664, Global Avg Loss: 1.54100264, Time: 0.0110 Steps: 63270, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001539, Sample Num: 24624, Cur Loss: 1.07160187, Cur Avg Loss: 0.55582796, Log Avg loss: 0.63267862, Global Avg Loss: 1.54085910, Time: 0.0114 Steps: 63280, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001549, Sample Num: 24784, Cur Loss: 0.28017840, Cur Avg Loss: 0.55599941, Log Avg loss: 0.58238646, Global Avg Loss: 1.54070765, Time: 0.0066 Steps: 63290, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001559, Sample Num: 24944, Cur Loss: 0.65342134, Cur Avg Loss: 0.55573838, Log Avg loss: 0.51530381, Global Avg Loss: 1.54054566, Time: 0.0066 Steps: 63300, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001569, Sample Num: 25104, Cur Loss: 0.30075514, Cur Avg Loss: 0.55597685, Log Avg loss: 0.59315521, Global Avg Loss: 1.54039602, Time: 0.0083 Steps: 63310, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001579, Sample Num: 25264, Cur Loss: 0.39240193, Cur Avg Loss: 0.55606510, Log Avg loss: 0.56991044, Global Avg Loss: 1.54024275, Time: 0.0108 Steps: 63320, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001589, Sample Num: 25424, Cur Loss: 0.30629051, Cur Avg Loss: 0.55561703, Log Avg loss: 0.48486789, Global Avg Loss: 1.54007611, Time: 0.0131 Steps: 63330, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001599, Sample Num: 25584, Cur Loss: 0.74703765, Cur Avg Loss: 0.55654921, Log Avg loss: 0.70467224, Global Avg Loss: 1.53994421, Time: 0.0117 Steps: 63340, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001609, Sample Num: 25744, Cur Loss: 0.73129106, Cur Avg Loss: 0.55654395, Log Avg loss: 0.55570255, Global Avg Loss: 1.53978885, Time: 0.0112 Steps: 63350, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001619, Sample Num: 25904, Cur Loss: 0.99054074, Cur Avg Loss: 0.55676950, Log Avg loss: 0.59306088, Global Avg Loss: 1.53963943, Time: 0.0099 Steps: 63360, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001629, Sample Num: 26064, Cur Loss: 0.53751242, Cur Avg Loss: 0.55767637, Log Avg loss: 0.70449768, Global Avg Loss: 1.53950764, Time: 0.0068 Steps: 63370, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001639, Sample Num: 26224, Cur Loss: 0.61111736, Cur Avg Loss: 0.55779047, Log Avg loss: 0.57637798, Global Avg Loss: 1.53935568, Time: 0.0068 Steps: 63380, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001649, Sample Num: 26384, Cur Loss: 0.59179246, Cur Avg Loss: 0.55780381, Log Avg loss: 0.55999045, Global Avg Loss: 1.53920118, Time: 0.0074 Steps: 63390, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001659, Sample Num: 26544, Cur Loss: 0.10961641, Cur Avg Loss: 0.55811176, Log Avg loss: 0.60889266, Global Avg Loss: 1.53905444, Time: 0.0158 Steps: 63400, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001669, Sample Num: 26704, Cur Loss: 1.24127972, Cur Avg Loss: 0.55961150, Log Avg loss: 0.80841817, Global Avg Loss: 1.53893922, Time: 0.0110 Steps: 63410, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001679, Sample Num: 26864, Cur Loss: 0.92486525, Cur Avg Loss: 0.56032846, Log Avg loss: 0.67998912, Global Avg Loss: 1.53880378, Time: 0.0150 Steps: 63420, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001689, Sample Num: 27024, Cur Loss: 0.51941836, Cur Avg Loss: 0.56095808, Log Avg loss: 0.66667166, Global Avg Loss: 1.53866629, Time: 0.0120 Steps: 63430, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001699, Sample Num: 27184, Cur Loss: 0.84814119, Cur Avg Loss: 0.56082491, Log Avg loss: 0.53833165, Global Avg Loss: 1.53850860, Time: 0.0106 Steps: 63440, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001709, Sample Num: 27344, Cur Loss: 0.43079656, Cur Avg Loss: 0.56001379, Log Avg loss: 0.42220480, Global Avg Loss: 1.53833267, Time: 0.0073 Steps: 63450, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001719, Sample Num: 27504, Cur Loss: 0.25143260, Cur Avg Loss: 0.55950061, Log Avg loss: 0.47179773, Global Avg Loss: 1.53816461, Time: 0.0113 Steps: 63460, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001729, Sample Num: 27664, Cur Loss: 0.73569548, Cur Avg Loss: 0.55932156, Log Avg loss: 0.52854413, Global Avg Loss: 1.53800554, Time: 0.0067 Steps: 63470, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001739, Sample Num: 27824, Cur Loss: 0.34196550, Cur Avg Loss: 0.55926209, Log Avg loss: 0.54897903, Global Avg Loss: 1.53784973, Time: 0.0078 Steps: 63480, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001749, Sample Num: 27984, Cur Loss: 0.22487873, Cur Avg Loss: 0.55891172, Log Avg loss: 0.49798194, Global Avg Loss: 1.53768595, Time: 0.0068 Steps: 63490, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001759, Sample Num: 28144, Cur Loss: 0.82210308, Cur Avg Loss: 0.55873518, Log Avg loss: 0.52785925, Global Avg Loss: 1.53752692, Time: 0.0092 Steps: 63500, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001769, Sample Num: 28304, Cur Loss: 0.70087612, Cur Avg Loss: 0.55905032, Log Avg loss: 0.61448259, Global Avg Loss: 1.53738158, Time: 0.0225 Steps: 63510, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001779, Sample Num: 28464, Cur Loss: 0.79096794, Cur Avg Loss: 0.55937619, Log Avg loss: 0.61702337, Global Avg Loss: 1.53723669, Time: 0.0068 Steps: 63520, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001789, Sample Num: 28624, Cur Loss: 0.39901739, Cur Avg Loss: 0.55917390, Log Avg loss: 0.52318627, Global Avg Loss: 1.53707707, Time: 0.0109 Steps: 63530, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001799, Sample Num: 28784, Cur Loss: 1.02893555, Cur Avg Loss: 0.55937724, Log Avg loss: 0.59575439, Global Avg Loss: 1.53692893, Time: 0.0201 Steps: 63540, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001809, Sample Num: 28944, Cur Loss: 0.49463081, Cur Avg Loss: 0.55952696, Log Avg loss: 0.58646210, Global Avg Loss: 1.53677936, Time: 0.0072 Steps: 63550, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001819, Sample Num: 29104, Cur Loss: 0.13213837, Cur Avg Loss: 0.55900654, Log Avg loss: 0.46486309, Global Avg Loss: 1.53661072, Time: 0.0064 Steps: 63560, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001829, Sample Num: 29264, Cur Loss: 0.60165489, Cur Avg Loss: 0.55911379, Log Avg loss: 0.57862101, Global Avg Loss: 1.53646002, Time: 0.0065 Steps: 63570, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001839, Sample Num: 29424, Cur Loss: 0.43000728, Cur Avg Loss: 0.55842262, Log Avg loss: 0.43200796, Global Avg Loss: 1.53628631, Time: 0.0066 Steps: 63580, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001849, Sample Num: 29584, Cur Loss: 0.85511196, Cur Avg Loss: 0.55859099, Log Avg loss: 0.58955415, Global Avg Loss: 1.53613743, Time: 0.0073 Steps: 63590, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001859, Sample Num: 29744, Cur Loss: 0.23209657, Cur Avg Loss: 0.55844064, Log Avg loss: 0.53064162, Global Avg Loss: 1.53597933, Time: 0.0068 Steps: 63600, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001869, Sample Num: 29904, Cur Loss: 1.27939761, Cur Avg Loss: 0.55850099, Log Avg loss: 0.56971896, Global Avg Loss: 1.53582743, Time: 0.0226 Steps: 63610, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001879, Sample Num: 30064, Cur Loss: 0.53552043, Cur Avg Loss: 0.55788667, Log Avg loss: 0.44307122, Global Avg Loss: 1.53565567, Time: 0.0065 Steps: 63620, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001889, Sample Num: 30224, Cur Loss: 0.53668261, Cur Avg Loss: 0.55830110, Log Avg loss: 0.63617232, Global Avg Loss: 1.53551430, Time: 0.0070 Steps: 63630, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001899, Sample Num: 30384, Cur Loss: 0.22596717, Cur Avg Loss: 0.55763226, Log Avg loss: 0.43128825, Global Avg Loss: 1.53534079, Time: 0.0191 Steps: 63640, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001909, Sample Num: 30544, Cur Loss: 0.49778664, Cur Avg Loss: 0.55752111, Log Avg loss: 0.53641311, Global Avg Loss: 1.53518385, Time: 0.0065 Steps: 63650, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001919, Sample Num: 30704, Cur Loss: 0.46373484, Cur Avg Loss: 0.55650223, Log Avg loss: 0.36199946, Global Avg Loss: 1.53499956, Time: 0.0068 Steps: 63660, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001929, Sample Num: 30864, Cur Loss: 0.49583256, Cur Avg Loss: 0.55672698, Log Avg loss: 0.59985691, Global Avg Loss: 1.53485269, Time: 0.0225 Steps: 63670, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001939, Sample Num: 31024, Cur Loss: 0.18311152, Cur Avg Loss: 0.55743491, Log Avg loss: 0.69399340, Global Avg Loss: 1.53472064, Time: 0.0065 Steps: 63680, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001949, Sample Num: 31184, Cur Loss: 0.63833690, Cur Avg Loss: 0.55792409, Log Avg loss: 0.65277579, Global Avg Loss: 1.53458217, Time: 0.0102 Steps: 63690, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001959, Sample Num: 31344, Cur Loss: 0.44125527, Cur Avg Loss: 0.55806111, Log Avg loss: 0.58476645, Global Avg Loss: 1.53443306, Time: 0.0066 Steps: 63700, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001969, Sample Num: 31504, Cur Loss: 0.54677027, Cur Avg Loss: 0.55819073, Log Avg loss: 0.58358381, Global Avg Loss: 1.53428382, Time: 0.0069 Steps: 63710, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001979, Sample Num: 31664, Cur Loss: 0.90030420, Cur Avg Loss: 0.55797345, Log Avg loss: 0.51519046, Global Avg Loss: 1.53412388, Time: 0.0066 Steps: 63720, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001989, Sample Num: 31824, Cur Loss: 0.30564094, Cur Avg Loss: 0.55697913, Log Avg loss: 0.36020323, Global Avg Loss: 1.53393968, Time: 0.0153 Steps: 63730, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 001999, Sample Num: 31984, Cur Loss: 0.35305926, Cur Avg Loss: 0.55695122, Log Avg loss: 0.55140037, Global Avg Loss: 1.53378553, Time: 0.0066 Steps: 63740, Updated lr: 0.000041 Training, Epoch: 0030, Batch: 002009, Sample Num: 32144, Cur Loss: 0.69916469, Cur Avg Loss: 0.55743262, Log Avg loss: 0.65366509, Global Avg Loss: 1.53364747, Time: 0.0116 Steps: 63750, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002019, Sample Num: 32304, Cur Loss: 0.37690991, Cur Avg Loss: 0.55741018, Log Avg loss: 0.55290183, Global Avg Loss: 1.53349366, Time: 0.0112 Steps: 63760, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002029, Sample Num: 32464, Cur Loss: 0.56040955, Cur Avg Loss: 0.55783701, Log Avg loss: 0.64401305, Global Avg Loss: 1.53335417, Time: 0.0069 Steps: 63770, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002039, Sample Num: 32624, Cur Loss: 0.60065985, Cur Avg Loss: 0.55737173, Log Avg loss: 0.46296726, Global Avg Loss: 1.53318635, Time: 0.0066 Steps: 63780, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002049, Sample Num: 32784, Cur Loss: 0.68543053, Cur Avg Loss: 0.55791153, Log Avg loss: 0.66797711, Global Avg Loss: 1.53305071, Time: 0.0074 Steps: 63790, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002059, Sample Num: 32944, Cur Loss: 0.33508015, Cur Avg Loss: 0.55772085, Log Avg loss: 0.51865073, Global Avg Loss: 1.53289172, Time: 0.0108 Steps: 63800, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002069, Sample Num: 33104, Cur Loss: 0.35890675, Cur Avg Loss: 0.55740497, Log Avg loss: 0.49236457, Global Avg Loss: 1.53272865, Time: 0.0124 Steps: 63810, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002079, Sample Num: 33264, Cur Loss: 0.57188880, Cur Avg Loss: 0.55663295, Log Avg loss: 0.39690115, Global Avg Loss: 1.53255068, Time: 0.0160 Steps: 63820, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002089, Sample Num: 33424, Cur Loss: 0.55035472, Cur Avg Loss: 0.55678645, Log Avg loss: 0.58870028, Global Avg Loss: 1.53240281, Time: 0.0117 Steps: 63830, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002099, Sample Num: 33584, Cur Loss: 0.64086187, Cur Avg Loss: 0.55757439, Log Avg loss: 0.72217430, Global Avg Loss: 1.53227589, Time: 0.0112 Steps: 63840, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002109, Sample Num: 33744, Cur Loss: 0.19049305, Cur Avg Loss: 0.55700369, Log Avg loss: 0.43721346, Global Avg Loss: 1.53210439, Time: 0.0112 Steps: 63850, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002119, Sample Num: 33904, Cur Loss: 0.22249891, Cur Avg Loss: 0.55588717, Log Avg loss: 0.32041421, Global Avg Loss: 1.53191465, Time: 0.0107 Steps: 63860, Updated lr: 0.000040 Training, Epoch: 0030, Batch: 002129, Sample Num: 34055, Cur Loss: 0.33034626, Cur Avg Loss: 0.55597899, Log Avg loss: 0.57543586, Global Avg Loss: 1.53176489, Time: 0.0036 Steps: 63870, Updated lr: 0.000040 ***** Running evaluation checkpoint-63870 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-63870 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.426205, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.696203, "eval_total_loss": 489.430463, "eval_mae": 0.676845, "eval_mse": 0.696308, "eval_r2": 0.557381, "eval_sp_statistic": 0.727784, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.776488, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.963418, "test_total_loss": 483.635927, "test_mae": 0.797358, "test_mse": 0.963523, "test_r2": 0.378134, "test_sp_statistic": 0.554773, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.6437, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.5317648919425522, "train_cur_epoch_loss": 1183.6792779490352, "train_cur_epoch_avg_loss": 0.5559789938699085, "train_cur_epoch_time": 22.426205158233643, "train_cur_epoch_avg_time": 0.010533680205840133, "epoch": 30, "step": 63870} ################################################## Training, Epoch: 0031, Batch: 000010, Sample Num: 160, Cur Loss: 0.33875251, Cur Avg Loss: 0.79838999, Log Avg loss: 0.79838999, Global Avg Loss: 1.53165009, Time: 0.0072 Steps: 63880, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000020, Sample Num: 320, Cur Loss: 0.45631203, Cur Avg Loss: 0.75122978, Log Avg loss: 0.70406957, Global Avg Loss: 1.53152055, Time: 0.0130 Steps: 63890, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000030, Sample Num: 480, Cur Loss: 0.38649711, Cur Avg Loss: 0.63796933, Log Avg loss: 0.41144842, Global Avg Loss: 1.53134527, Time: 0.0108 Steps: 63900, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000040, Sample Num: 640, Cur Loss: 0.34270519, Cur Avg Loss: 0.57657944, Log Avg loss: 0.39240979, Global Avg Loss: 1.53116706, Time: 0.0086 Steps: 63910, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000050, Sample Num: 800, Cur Loss: 0.64564574, Cur Avg Loss: 0.56010301, Log Avg loss: 0.49419727, Global Avg Loss: 1.53100483, Time: 0.0116 Steps: 63920, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000060, Sample Num: 960, Cur Loss: 0.09482619, Cur Avg Loss: 0.53066900, Log Avg loss: 0.38349898, Global Avg Loss: 1.53082534, Time: 0.0067 Steps: 63930, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000070, Sample Num: 1120, Cur Loss: 0.62037456, Cur Avg Loss: 0.54210721, Log Avg loss: 0.61073647, Global Avg Loss: 1.53068144, Time: 0.0121 Steps: 63940, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000080, Sample Num: 1280, Cur Loss: 0.36892945, Cur Avg Loss: 0.55047898, Log Avg loss: 0.60908132, Global Avg Loss: 1.53053733, Time: 0.0113 Steps: 63950, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000090, Sample Num: 1440, Cur Loss: 0.31232339, Cur Avg Loss: 0.55241500, Log Avg loss: 0.56790318, Global Avg Loss: 1.53038682, Time: 0.0069 Steps: 63960, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000100, Sample Num: 1600, Cur Loss: 0.43559593, Cur Avg Loss: 0.55236162, Log Avg loss: 0.55188120, Global Avg Loss: 1.53023386, Time: 0.0075 Steps: 63970, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000110, Sample Num: 1760, Cur Loss: 0.49281999, Cur Avg Loss: 0.53959001, Log Avg loss: 0.41187394, Global Avg Loss: 1.53005906, Time: 0.0066 Steps: 63980, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000120, Sample Num: 1920, Cur Loss: 0.39015386, Cur Avg Loss: 0.54409370, Log Avg loss: 0.59363433, Global Avg Loss: 1.52991272, Time: 0.0088 Steps: 63990, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000130, Sample Num: 2080, Cur Loss: 0.92513198, Cur Avg Loss: 0.53741827, Log Avg loss: 0.45731306, Global Avg Loss: 1.52974513, Time: 0.0068 Steps: 64000, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000140, Sample Num: 2240, Cur Loss: 0.31855077, Cur Avg Loss: 0.53476124, Log Avg loss: 0.50021984, Global Avg Loss: 1.52958429, Time: 0.0115 Steps: 64010, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000150, Sample Num: 2400, Cur Loss: 0.61153567, Cur Avg Loss: 0.53391314, Log Avg loss: 0.52203975, Global Avg Loss: 1.52942691, Time: 0.0128 Steps: 64020, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000160, Sample Num: 2560, Cur Loss: 0.42453668, Cur Avg Loss: 0.52876216, Log Avg loss: 0.45149752, Global Avg Loss: 1.52925856, Time: 0.0070 Steps: 64030, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000170, Sample Num: 2720, Cur Loss: 0.92435086, Cur Avg Loss: 0.52569189, Log Avg loss: 0.47656747, Global Avg Loss: 1.52909418, Time: 0.0126 Steps: 64040, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000180, Sample Num: 2880, Cur Loss: 0.20047039, Cur Avg Loss: 0.52726799, Log Avg loss: 0.55406165, Global Avg Loss: 1.52894195, Time: 0.0066 Steps: 64050, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000190, Sample Num: 3040, Cur Loss: 0.34186190, Cur Avg Loss: 0.52104699, Log Avg loss: 0.40906901, Global Avg Loss: 1.52876713, Time: 0.0139 Steps: 64060, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000200, Sample Num: 3200, Cur Loss: 0.34929794, Cur Avg Loss: 0.52172717, Log Avg loss: 0.53465062, Global Avg Loss: 1.52861197, Time: 0.0191 Steps: 64070, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000210, Sample Num: 3360, Cur Loss: 0.61845070, Cur Avg Loss: 0.52922105, Log Avg loss: 0.67909873, Global Avg Loss: 1.52847940, Time: 0.0114 Steps: 64080, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000220, Sample Num: 3520, Cur Loss: 1.56322861, Cur Avg Loss: 0.53244827, Log Avg loss: 0.60021983, Global Avg Loss: 1.52833456, Time: 0.0066 Steps: 64090, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000230, Sample Num: 3680, Cur Loss: 0.21674854, Cur Avg Loss: 0.52820113, Log Avg loss: 0.43476394, Global Avg Loss: 1.52816396, Time: 0.0121 Steps: 64100, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000240, Sample Num: 3840, Cur Loss: 0.48559630, Cur Avg Loss: 0.53381343, Log Avg loss: 0.66289643, Global Avg Loss: 1.52802900, Time: 0.0118 Steps: 64110, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000250, Sample Num: 4000, Cur Loss: 0.43860796, Cur Avg Loss: 0.53148902, Log Avg loss: 0.47570322, Global Avg Loss: 1.52786488, Time: 0.0092 Steps: 64120, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000260, Sample Num: 4160, Cur Loss: 0.27785212, Cur Avg Loss: 0.52693509, Log Avg loss: 0.41308680, Global Avg Loss: 1.52769105, Time: 0.0076 Steps: 64130, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000270, Sample Num: 4320, Cur Loss: 0.37344527, Cur Avg Loss: 0.52905105, Log Avg loss: 0.58406599, Global Avg Loss: 1.52754393, Time: 0.0119 Steps: 64140, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000280, Sample Num: 4480, Cur Loss: 0.60280043, Cur Avg Loss: 0.52995343, Log Avg loss: 0.55431768, Global Avg Loss: 1.52739222, Time: 0.0124 Steps: 64150, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000290, Sample Num: 4640, Cur Loss: 0.30859417, Cur Avg Loss: 0.53329087, Log Avg loss: 0.62673921, Global Avg Loss: 1.52725184, Time: 0.0122 Steps: 64160, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000300, Sample Num: 4800, Cur Loss: 0.44093844, Cur Avg Loss: 0.52769719, Log Avg loss: 0.36548039, Global Avg Loss: 1.52707079, Time: 0.0067 Steps: 64170, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000310, Sample Num: 4960, Cur Loss: 0.27097490, Cur Avg Loss: 0.52361126, Log Avg loss: 0.40103335, Global Avg Loss: 1.52689534, Time: 0.0117 Steps: 64180, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000320, Sample Num: 5120, Cur Loss: 0.72205895, Cur Avg Loss: 0.52434597, Log Avg loss: 0.54712213, Global Avg Loss: 1.52674271, Time: 0.0100 Steps: 64190, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000330, Sample Num: 5280, Cur Loss: 1.56219447, Cur Avg Loss: 0.53283107, Log Avg loss: 0.80435427, Global Avg Loss: 1.52663019, Time: 0.0092 Steps: 64200, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000340, Sample Num: 5440, Cur Loss: 0.67861384, Cur Avg Loss: 0.53456940, Log Avg loss: 0.59193427, Global Avg Loss: 1.52648462, Time: 0.0072 Steps: 64210, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000350, Sample Num: 5600, Cur Loss: 0.91637290, Cur Avg Loss: 0.53279265, Log Avg loss: 0.47238308, Global Avg Loss: 1.52632048, Time: 0.0069 Steps: 64220, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000360, Sample Num: 5760, Cur Loss: 0.28616440, Cur Avg Loss: 0.53030573, Log Avg loss: 0.44326373, Global Avg Loss: 1.52615186, Time: 0.0084 Steps: 64230, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000370, Sample Num: 5920, Cur Loss: 0.32066777, Cur Avg Loss: 0.52715333, Log Avg loss: 0.41366694, Global Avg Loss: 1.52597868, Time: 0.0112 Steps: 64240, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000380, Sample Num: 6080, Cur Loss: 0.48100069, Cur Avg Loss: 0.52413366, Log Avg loss: 0.41240579, Global Avg Loss: 1.52580536, Time: 0.0067 Steps: 64250, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000390, Sample Num: 6240, Cur Loss: 0.30950695, Cur Avg Loss: 0.52466409, Log Avg loss: 0.54482024, Global Avg Loss: 1.52565270, Time: 0.0067 Steps: 64260, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000400, Sample Num: 6400, Cur Loss: 0.49255070, Cur Avg Loss: 0.53130769, Log Avg loss: 0.79040821, Global Avg Loss: 1.52553830, Time: 0.0067 Steps: 64270, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000410, Sample Num: 6560, Cur Loss: 0.84729624, Cur Avg Loss: 0.53472265, Log Avg loss: 0.67132099, Global Avg Loss: 1.52540541, Time: 0.0073 Steps: 64280, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000420, Sample Num: 6720, Cur Loss: 0.50873101, Cur Avg Loss: 0.53735920, Log Avg loss: 0.64545783, Global Avg Loss: 1.52526854, Time: 0.0119 Steps: 64290, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000430, Sample Num: 6880, Cur Loss: 0.52292228, Cur Avg Loss: 0.53693003, Log Avg loss: 0.51890489, Global Avg Loss: 1.52511203, Time: 0.0129 Steps: 64300, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000440, Sample Num: 7040, Cur Loss: 0.99830866, Cur Avg Loss: 0.53904098, Log Avg loss: 0.62981170, Global Avg Loss: 1.52497281, Time: 0.0068 Steps: 64310, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000450, Sample Num: 7200, Cur Loss: 0.54747730, Cur Avg Loss: 0.53939389, Log Avg loss: 0.55492218, Global Avg Loss: 1.52482200, Time: 0.0068 Steps: 64320, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000460, Sample Num: 7360, Cur Loss: 0.37509131, Cur Avg Loss: 0.54128590, Log Avg loss: 0.62642610, Global Avg Loss: 1.52468234, Time: 0.0117 Steps: 64330, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000470, Sample Num: 7520, Cur Loss: 0.69986951, Cur Avg Loss: 0.54088339, Log Avg loss: 0.52236798, Global Avg Loss: 1.52452656, Time: 0.0097 Steps: 64340, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000480, Sample Num: 7680, Cur Loss: 0.35733449, Cur Avg Loss: 0.53999934, Log Avg loss: 0.49844918, Global Avg Loss: 1.52436711, Time: 0.0145 Steps: 64350, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000490, Sample Num: 7840, Cur Loss: 0.51777554, Cur Avg Loss: 0.54251679, Log Avg loss: 0.66335423, Global Avg Loss: 1.52423333, Time: 0.0168 Steps: 64360, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000500, Sample Num: 8000, Cur Loss: 0.32248944, Cur Avg Loss: 0.54588769, Log Avg loss: 0.71106196, Global Avg Loss: 1.52410700, Time: 0.0226 Steps: 64370, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000510, Sample Num: 8160, Cur Loss: 0.22601001, Cur Avg Loss: 0.54426558, Log Avg loss: 0.46315999, Global Avg Loss: 1.52394220, Time: 0.0151 Steps: 64380, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000520, Sample Num: 8320, Cur Loss: 0.95990109, Cur Avg Loss: 0.54338892, Log Avg loss: 0.49867907, Global Avg Loss: 1.52378298, Time: 0.0067 Steps: 64390, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000530, Sample Num: 8480, Cur Loss: 0.18928346, Cur Avg Loss: 0.54032034, Log Avg loss: 0.38075408, Global Avg Loss: 1.52360549, Time: 0.0106 Steps: 64400, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000540, Sample Num: 8640, Cur Loss: 0.90774238, Cur Avg Loss: 0.53905226, Log Avg loss: 0.47184410, Global Avg Loss: 1.52344220, Time: 0.0112 Steps: 64410, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000550, Sample Num: 8800, Cur Loss: 0.60124081, Cur Avg Loss: 0.53831520, Log Avg loss: 0.49851391, Global Avg Loss: 1.52328310, Time: 0.0067 Steps: 64420, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000560, Sample Num: 8960, Cur Loss: 0.44582784, Cur Avg Loss: 0.53703086, Log Avg loss: 0.46639210, Global Avg Loss: 1.52311906, Time: 0.0134 Steps: 64430, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000570, Sample Num: 9120, Cur Loss: 0.15277079, Cur Avg Loss: 0.53598450, Log Avg loss: 0.47738879, Global Avg Loss: 1.52295678, Time: 0.0067 Steps: 64440, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000580, Sample Num: 9280, Cur Loss: 0.56682742, Cur Avg Loss: 0.53544382, Log Avg loss: 0.50462493, Global Avg Loss: 1.52279878, Time: 0.0096 Steps: 64450, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000590, Sample Num: 9440, Cur Loss: 0.82231647, Cur Avg Loss: 0.53530277, Log Avg loss: 0.52712198, Global Avg Loss: 1.52264431, Time: 0.0072 Steps: 64460, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000600, Sample Num: 9600, Cur Loss: 0.79065949, Cur Avg Loss: 0.53550466, Log Avg loss: 0.54741601, Global Avg Loss: 1.52249304, Time: 0.0121 Steps: 64470, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000610, Sample Num: 9760, Cur Loss: 0.93892109, Cur Avg Loss: 0.53457108, Log Avg loss: 0.47855637, Global Avg Loss: 1.52233114, Time: 0.0067 Steps: 64480, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000620, Sample Num: 9920, Cur Loss: 0.49330345, Cur Avg Loss: 0.53316172, Log Avg loss: 0.44719059, Global Avg Loss: 1.52216443, Time: 0.0067 Steps: 64490, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000630, Sample Num: 10080, Cur Loss: 0.53995514, Cur Avg Loss: 0.53229068, Log Avg loss: 0.47828635, Global Avg Loss: 1.52200259, Time: 0.0067 Steps: 64500, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000640, Sample Num: 10240, Cur Loss: 0.48199496, Cur Avg Loss: 0.53217229, Log Avg loss: 0.52471357, Global Avg Loss: 1.52184799, Time: 0.0160 Steps: 64510, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000650, Sample Num: 10400, Cur Loss: 0.46511993, Cur Avg Loss: 0.53138686, Log Avg loss: 0.48111911, Global Avg Loss: 1.52168669, Time: 0.0124 Steps: 64520, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000660, Sample Num: 10560, Cur Loss: 0.34514573, Cur Avg Loss: 0.53305677, Log Avg loss: 0.64160116, Global Avg Loss: 1.52155030, Time: 0.0112 Steps: 64530, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000670, Sample Num: 10720, Cur Loss: 0.32026237, Cur Avg Loss: 0.53211583, Log Avg loss: 0.47001393, Global Avg Loss: 1.52138738, Time: 0.0068 Steps: 64540, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000680, Sample Num: 10880, Cur Loss: 0.31264818, Cur Avg Loss: 0.53380597, Log Avg loss: 0.64704528, Global Avg Loss: 1.52125192, Time: 0.0119 Steps: 64550, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000690, Sample Num: 11040, Cur Loss: 0.48923552, Cur Avg Loss: 0.53227321, Log Avg loss: 0.42804573, Global Avg Loss: 1.52108259, Time: 0.0068 Steps: 64560, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000700, Sample Num: 11200, Cur Loss: 0.83171284, Cur Avg Loss: 0.53242997, Log Avg loss: 0.54324645, Global Avg Loss: 1.52093115, Time: 0.0116 Steps: 64570, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000710, Sample Num: 11360, Cur Loss: 0.40515012, Cur Avg Loss: 0.53209947, Log Avg loss: 0.50896444, Global Avg Loss: 1.52077445, Time: 0.0067 Steps: 64580, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000720, Sample Num: 11520, Cur Loss: 0.26648849, Cur Avg Loss: 0.53142779, Log Avg loss: 0.48373826, Global Avg Loss: 1.52061390, Time: 0.0067 Steps: 64590, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000730, Sample Num: 11680, Cur Loss: 0.39877313, Cur Avg Loss: 0.53178873, Log Avg loss: 0.55777667, Global Avg Loss: 1.52046485, Time: 0.0067 Steps: 64600, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000740, Sample Num: 11840, Cur Loss: 0.46858561, Cur Avg Loss: 0.53308428, Log Avg loss: 0.62765933, Global Avg Loss: 1.52032667, Time: 0.0067 Steps: 64610, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000750, Sample Num: 12000, Cur Loss: 0.33747154, Cur Avg Loss: 0.53294462, Log Avg loss: 0.52260956, Global Avg Loss: 1.52017227, Time: 0.0067 Steps: 64620, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000760, Sample Num: 12160, Cur Loss: 0.60292566, Cur Avg Loss: 0.53537451, Log Avg loss: 0.71761642, Global Avg Loss: 1.52004809, Time: 0.0067 Steps: 64630, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000770, Sample Num: 12320, Cur Loss: 0.81356287, Cur Avg Loss: 0.53605222, Log Avg loss: 0.58755787, Global Avg Loss: 1.51990383, Time: 0.0211 Steps: 64640, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000780, Sample Num: 12480, Cur Loss: 0.83491474, Cur Avg Loss: 0.53654419, Log Avg loss: 0.57442574, Global Avg Loss: 1.51975759, Time: 0.0065 Steps: 64650, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000790, Sample Num: 12640, Cur Loss: 0.31193525, Cur Avg Loss: 0.53505946, Log Avg loss: 0.41925071, Global Avg Loss: 1.51958739, Time: 0.0066 Steps: 64660, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000800, Sample Num: 12800, Cur Loss: 1.22451746, Cur Avg Loss: 0.53724464, Log Avg loss: 0.70987395, Global Avg Loss: 1.51946218, Time: 0.0067 Steps: 64670, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000810, Sample Num: 12960, Cur Loss: 0.77241427, Cur Avg Loss: 0.53793077, Log Avg loss: 0.59282157, Global Avg Loss: 1.51931892, Time: 0.0067 Steps: 64680, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000820, Sample Num: 13120, Cur Loss: 0.62490201, Cur Avg Loss: 0.53722772, Log Avg loss: 0.48028005, Global Avg Loss: 1.51915830, Time: 0.0064 Steps: 64690, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000830, Sample Num: 13280, Cur Loss: 0.68569088, Cur Avg Loss: 0.53822903, Log Avg loss: 0.62033683, Global Avg Loss: 1.51901938, Time: 0.0114 Steps: 64700, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000840, Sample Num: 13440, Cur Loss: 0.53297204, Cur Avg Loss: 0.53867967, Log Avg loss: 0.57608306, Global Avg Loss: 1.51887366, Time: 0.0085 Steps: 64710, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000850, Sample Num: 13600, Cur Loss: 0.50746900, Cur Avg Loss: 0.53877385, Log Avg loss: 0.54668486, Global Avg Loss: 1.51872345, Time: 0.0074 Steps: 64720, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000860, Sample Num: 13760, Cur Loss: 0.77953964, Cur Avg Loss: 0.53855884, Log Avg loss: 0.52028299, Global Avg Loss: 1.51856920, Time: 0.0067 Steps: 64730, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000870, Sample Num: 13920, Cur Loss: 0.43331552, Cur Avg Loss: 0.54020491, Log Avg loss: 0.68176653, Global Avg Loss: 1.51843994, Time: 0.0068 Steps: 64740, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000880, Sample Num: 14080, Cur Loss: 0.33375907, Cur Avg Loss: 0.54084973, Log Avg loss: 0.59694950, Global Avg Loss: 1.51829763, Time: 0.0089 Steps: 64750, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000890, Sample Num: 14240, Cur Loss: 0.52895713, Cur Avg Loss: 0.54083114, Log Avg loss: 0.53919523, Global Avg Loss: 1.51814644, Time: 0.0120 Steps: 64760, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000900, Sample Num: 14400, Cur Loss: 0.23091784, Cur Avg Loss: 0.54113601, Log Avg loss: 0.56826896, Global Avg Loss: 1.51799978, Time: 0.0112 Steps: 64770, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000910, Sample Num: 14560, Cur Loss: 0.34501550, Cur Avg Loss: 0.53896895, Log Avg loss: 0.34393410, Global Avg Loss: 1.51781855, Time: 0.0136 Steps: 64780, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000920, Sample Num: 14720, Cur Loss: 0.37403181, Cur Avg Loss: 0.53786073, Log Avg loss: 0.43701262, Global Avg Loss: 1.51765173, Time: 0.0239 Steps: 64790, Updated lr: 0.000040 Training, Epoch: 0031, Batch: 000930, Sample Num: 14880, Cur Loss: 1.65355229, Cur Avg Loss: 0.53899157, Log Avg loss: 0.64302834, Global Avg Loss: 1.51751676, Time: 0.0106 Steps: 64800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000940, Sample Num: 15040, Cur Loss: 0.23181683, Cur Avg Loss: 0.53787584, Log Avg loss: 0.43411347, Global Avg Loss: 1.51734959, Time: 0.0138 Steps: 64810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000950, Sample Num: 15200, Cur Loss: 1.13234293, Cur Avg Loss: 0.54061829, Log Avg loss: 0.79840871, Global Avg Loss: 1.51723868, Time: 0.0115 Steps: 64820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000960, Sample Num: 15360, Cur Loss: 0.38354462, Cur Avg Loss: 0.54080527, Log Avg loss: 0.55856771, Global Avg Loss: 1.51709080, Time: 0.0066 Steps: 64830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000970, Sample Num: 15520, Cur Loss: 0.42689368, Cur Avg Loss: 0.54104616, Log Avg loss: 0.56417209, Global Avg Loss: 1.51694384, Time: 0.0068 Steps: 64840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000980, Sample Num: 15680, Cur Loss: 0.23548545, Cur Avg Loss: 0.54187747, Log Avg loss: 0.62251445, Global Avg Loss: 1.51680591, Time: 0.0116 Steps: 64850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 000990, Sample Num: 15840, Cur Loss: 0.67286235, Cur Avg Loss: 0.54331723, Log Avg loss: 0.68441336, Global Avg Loss: 1.51667758, Time: 0.0081 Steps: 64860, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001000, Sample Num: 16000, Cur Loss: 0.53681654, Cur Avg Loss: 0.54278068, Log Avg loss: 0.48966268, Global Avg Loss: 1.51651926, Time: 0.0066 Steps: 64870, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001010, Sample Num: 16160, Cur Loss: 0.69913065, Cur Avg Loss: 0.54278232, Log Avg loss: 0.54294581, Global Avg Loss: 1.51636920, Time: 0.0065 Steps: 64880, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001020, Sample Num: 16320, Cur Loss: 0.95202219, Cur Avg Loss: 0.54254685, Log Avg loss: 0.51876474, Global Avg Loss: 1.51621546, Time: 0.0116 Steps: 64890, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001030, Sample Num: 16480, Cur Loss: 0.56400621, Cur Avg Loss: 0.54248137, Log Avg loss: 0.53580203, Global Avg Loss: 1.51606440, Time: 0.0073 Steps: 64900, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001040, Sample Num: 16640, Cur Loss: 0.57575274, Cur Avg Loss: 0.54189759, Log Avg loss: 0.48176825, Global Avg Loss: 1.51590506, Time: 0.0118 Steps: 64910, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001050, Sample Num: 16800, Cur Loss: 1.39859557, Cur Avg Loss: 0.54292664, Log Avg loss: 0.64994861, Global Avg Loss: 1.51577167, Time: 0.0117 Steps: 64920, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001060, Sample Num: 16960, Cur Loss: 0.44517639, Cur Avg Loss: 0.54332502, Log Avg loss: 0.58515440, Global Avg Loss: 1.51562834, Time: 0.0120 Steps: 64930, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001070, Sample Num: 17120, Cur Loss: 0.48918861, Cur Avg Loss: 0.54312791, Log Avg loss: 0.52223404, Global Avg Loss: 1.51547537, Time: 0.0097 Steps: 64940, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001080, Sample Num: 17280, Cur Loss: 0.47221285, Cur Avg Loss: 0.54342426, Log Avg loss: 0.57513371, Global Avg Loss: 1.51533059, Time: 0.0068 Steps: 64950, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001090, Sample Num: 17440, Cur Loss: 0.40125081, Cur Avg Loss: 0.54180750, Log Avg loss: 0.36719749, Global Avg Loss: 1.51515385, Time: 0.0070 Steps: 64960, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001100, Sample Num: 17600, Cur Loss: 1.59189117, Cur Avg Loss: 0.54314362, Log Avg loss: 0.68878081, Global Avg Loss: 1.51502665, Time: 0.0113 Steps: 64970, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001110, Sample Num: 17760, Cur Loss: 1.04714143, Cur Avg Loss: 0.54373034, Log Avg loss: 0.60826967, Global Avg Loss: 1.51488711, Time: 0.0137 Steps: 64980, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001120, Sample Num: 17920, Cur Loss: 0.27779520, Cur Avg Loss: 0.54371083, Log Avg loss: 0.54154573, Global Avg Loss: 1.51473734, Time: 0.0068 Steps: 64990, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001130, Sample Num: 18080, Cur Loss: 0.18579330, Cur Avg Loss: 0.54481722, Log Avg loss: 0.66873204, Global Avg Loss: 1.51460719, Time: 0.0068 Steps: 65000, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001140, Sample Num: 18240, Cur Loss: 0.20687869, Cur Avg Loss: 0.54586448, Log Avg loss: 0.66420544, Global Avg Loss: 1.51447638, Time: 0.0068 Steps: 65010, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001150, Sample Num: 18400, Cur Loss: 0.19880061, Cur Avg Loss: 0.54599355, Log Avg loss: 0.56070718, Global Avg Loss: 1.51432969, Time: 0.0067 Steps: 65020, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001160, Sample Num: 18560, Cur Loss: 1.40983272, Cur Avg Loss: 0.54686434, Log Avg loss: 0.64700532, Global Avg Loss: 1.51419631, Time: 0.0128 Steps: 65030, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001170, Sample Num: 18720, Cur Loss: 0.43183410, Cur Avg Loss: 0.54526310, Log Avg loss: 0.35951960, Global Avg Loss: 1.51401878, Time: 0.0140 Steps: 65040, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001180, Sample Num: 18880, Cur Loss: 0.56163764, Cur Avg Loss: 0.54432312, Log Avg loss: 0.43434464, Global Avg Loss: 1.51385280, Time: 0.0066 Steps: 65050, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001190, Sample Num: 19040, Cur Loss: 0.26391476, Cur Avg Loss: 0.54310662, Log Avg loss: 0.39955957, Global Avg Loss: 1.51368153, Time: 0.0124 Steps: 65060, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001200, Sample Num: 19200, Cur Loss: 0.94350779, Cur Avg Loss: 0.54370894, Log Avg loss: 0.61538550, Global Avg Loss: 1.51354348, Time: 0.0110 Steps: 65070, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001210, Sample Num: 19360, Cur Loss: 0.22282684, Cur Avg Loss: 0.54347030, Log Avg loss: 0.51483354, Global Avg Loss: 1.51339002, Time: 0.0218 Steps: 65080, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001220, Sample Num: 19520, Cur Loss: 0.28547072, Cur Avg Loss: 0.54332022, Log Avg loss: 0.52516100, Global Avg Loss: 1.51323820, Time: 0.0132 Steps: 65090, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001230, Sample Num: 19680, Cur Loss: 0.48190662, Cur Avg Loss: 0.54229395, Log Avg loss: 0.41708872, Global Avg Loss: 1.51306982, Time: 0.0119 Steps: 65100, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001240, Sample Num: 19840, Cur Loss: 0.68581492, Cur Avg Loss: 0.54128066, Log Avg loss: 0.41664531, Global Avg Loss: 1.51290142, Time: 0.0123 Steps: 65110, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001250, Sample Num: 20000, Cur Loss: 0.76526982, Cur Avg Loss: 0.54022640, Log Avg loss: 0.40949901, Global Avg Loss: 1.51273198, Time: 0.0109 Steps: 65120, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001260, Sample Num: 20160, Cur Loss: 0.83683527, Cur Avg Loss: 0.54237314, Log Avg loss: 0.81071483, Global Avg Loss: 1.51262419, Time: 0.0124 Steps: 65130, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001270, Sample Num: 20320, Cur Loss: 1.32760382, Cur Avg Loss: 0.54315420, Log Avg loss: 0.64156864, Global Avg Loss: 1.51249047, Time: 0.0145 Steps: 65140, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001280, Sample Num: 20480, Cur Loss: 1.36427474, Cur Avg Loss: 0.54348553, Log Avg loss: 0.58556369, Global Avg Loss: 1.51234820, Time: 0.0089 Steps: 65150, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001290, Sample Num: 20640, Cur Loss: 0.61777467, Cur Avg Loss: 0.54510994, Log Avg loss: 0.75303423, Global Avg Loss: 1.51223167, Time: 0.0115 Steps: 65160, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001300, Sample Num: 20800, Cur Loss: 0.90161061, Cur Avg Loss: 0.54657223, Log Avg loss: 0.73520851, Global Avg Loss: 1.51211244, Time: 0.0114 Steps: 65170, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001310, Sample Num: 20960, Cur Loss: 0.11895983, Cur Avg Loss: 0.54614430, Log Avg loss: 0.49051246, Global Avg Loss: 1.51195570, Time: 0.0142 Steps: 65180, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001320, Sample Num: 21120, Cur Loss: 0.55225074, Cur Avg Loss: 0.54554584, Log Avg loss: 0.46714847, Global Avg Loss: 1.51179543, Time: 0.0162 Steps: 65190, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001330, Sample Num: 21280, Cur Loss: 0.24282947, Cur Avg Loss: 0.54417115, Log Avg loss: 0.36271167, Global Avg Loss: 1.51161919, Time: 0.0117 Steps: 65200, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001340, Sample Num: 21440, Cur Loss: 1.32950342, Cur Avg Loss: 0.54438620, Log Avg loss: 0.57298820, Global Avg Loss: 1.51147525, Time: 0.0113 Steps: 65210, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001350, Sample Num: 21600, Cur Loss: 1.02015162, Cur Avg Loss: 0.54473593, Log Avg loss: 0.59159996, Global Avg Loss: 1.51133421, Time: 0.0139 Steps: 65220, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001360, Sample Num: 21760, Cur Loss: 0.50401300, Cur Avg Loss: 0.54411794, Log Avg loss: 0.46068826, Global Avg Loss: 1.51117314, Time: 0.0121 Steps: 65230, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001370, Sample Num: 21920, Cur Loss: 1.01521337, Cur Avg Loss: 0.54419041, Log Avg loss: 0.55404737, Global Avg Loss: 1.51102643, Time: 0.0107 Steps: 65240, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001380, Sample Num: 22080, Cur Loss: 0.44450426, Cur Avg Loss: 0.54494869, Log Avg loss: 0.64883218, Global Avg Loss: 1.51089430, Time: 0.0067 Steps: 65250, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001390, Sample Num: 22240, Cur Loss: 0.21116176, Cur Avg Loss: 0.54532427, Log Avg loss: 0.59715510, Global Avg Loss: 1.51075428, Time: 0.0072 Steps: 65260, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001400, Sample Num: 22400, Cur Loss: 0.36583894, Cur Avg Loss: 0.54569693, Log Avg loss: 0.59749622, Global Avg Loss: 1.51061436, Time: 0.0118 Steps: 65270, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001410, Sample Num: 22560, Cur Loss: 0.75752461, Cur Avg Loss: 0.54478957, Log Avg loss: 0.41775923, Global Avg Loss: 1.51044695, Time: 0.0071 Steps: 65280, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001420, Sample Num: 22720, Cur Loss: 0.46206179, Cur Avg Loss: 0.54529908, Log Avg loss: 0.61714005, Global Avg Loss: 1.51031013, Time: 0.0120 Steps: 65290, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001430, Sample Num: 22880, Cur Loss: 0.67000282, Cur Avg Loss: 0.54557682, Log Avg loss: 0.58501547, Global Avg Loss: 1.51016843, Time: 0.0117 Steps: 65300, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001440, Sample Num: 23040, Cur Loss: 0.53067791, Cur Avg Loss: 0.54515800, Log Avg loss: 0.48526637, Global Avg Loss: 1.51001150, Time: 0.0068 Steps: 65310, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001450, Sample Num: 23200, Cur Loss: 0.11368204, Cur Avg Loss: 0.54585366, Log Avg loss: 0.64602879, Global Avg Loss: 1.50987923, Time: 0.0074 Steps: 65320, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001460, Sample Num: 23360, Cur Loss: 0.15442464, Cur Avg Loss: 0.54563436, Log Avg loss: 0.51383589, Global Avg Loss: 1.50972677, Time: 0.0073 Steps: 65330, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001470, Sample Num: 23520, Cur Loss: 0.27908033, Cur Avg Loss: 0.54537049, Log Avg loss: 0.50684669, Global Avg Loss: 1.50957328, Time: 0.0072 Steps: 65340, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001480, Sample Num: 23680, Cur Loss: 0.69154710, Cur Avg Loss: 0.54585831, Log Avg loss: 0.61756758, Global Avg Loss: 1.50943679, Time: 0.0070 Steps: 65350, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001490, Sample Num: 23840, Cur Loss: 0.19540542, Cur Avg Loss: 0.54556429, Log Avg loss: 0.50204958, Global Avg Loss: 1.50928266, Time: 0.0110 Steps: 65360, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001500, Sample Num: 24000, Cur Loss: 0.60991365, Cur Avg Loss: 0.54511622, Log Avg loss: 0.47835389, Global Avg Loss: 1.50912495, Time: 0.0066 Steps: 65370, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001510, Sample Num: 24160, Cur Loss: 0.22713384, Cur Avg Loss: 0.54464098, Log Avg loss: 0.47335369, Global Avg Loss: 1.50896653, Time: 0.0067 Steps: 65380, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001520, Sample Num: 24320, Cur Loss: 1.86176050, Cur Avg Loss: 0.54442576, Log Avg loss: 0.51192829, Global Avg Loss: 1.50881405, Time: 0.0120 Steps: 65390, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001530, Sample Num: 24480, Cur Loss: 0.30611598, Cur Avg Loss: 0.54442866, Log Avg loss: 0.54486871, Global Avg Loss: 1.50866666, Time: 0.0142 Steps: 65400, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001540, Sample Num: 24640, Cur Loss: 0.24983811, Cur Avg Loss: 0.54457246, Log Avg loss: 0.56657386, Global Avg Loss: 1.50852263, Time: 0.0134 Steps: 65410, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001550, Sample Num: 24800, Cur Loss: 0.55810505, Cur Avg Loss: 0.54474137, Log Avg loss: 0.57075479, Global Avg Loss: 1.50837928, Time: 0.0112 Steps: 65420, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001560, Sample Num: 24960, Cur Loss: 0.60416466, Cur Avg Loss: 0.54472098, Log Avg loss: 0.54155913, Global Avg Loss: 1.50823152, Time: 0.0111 Steps: 65430, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001570, Sample Num: 25120, Cur Loss: 0.22066902, Cur Avg Loss: 0.54489053, Log Avg loss: 0.57134088, Global Avg Loss: 1.50808835, Time: 0.0112 Steps: 65440, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001580, Sample Num: 25280, Cur Loss: 0.60592049, Cur Avg Loss: 0.54475728, Log Avg loss: 0.52383753, Global Avg Loss: 1.50793797, Time: 0.0070 Steps: 65450, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001590, Sample Num: 25440, Cur Loss: 0.39114100, Cur Avg Loss: 0.54386560, Log Avg loss: 0.40297976, Global Avg Loss: 1.50776917, Time: 0.0231 Steps: 65460, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001600, Sample Num: 25600, Cur Loss: 0.44372758, Cur Avg Loss: 0.54375004, Log Avg loss: 0.52537612, Global Avg Loss: 1.50761912, Time: 0.0109 Steps: 65470, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001610, Sample Num: 25760, Cur Loss: 0.41860163, Cur Avg Loss: 0.54377751, Log Avg loss: 0.54817328, Global Avg Loss: 1.50747259, Time: 0.0068 Steps: 65480, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001620, Sample Num: 25920, Cur Loss: 0.53719091, Cur Avg Loss: 0.54330793, Log Avg loss: 0.46770426, Global Avg Loss: 1.50731383, Time: 0.0071 Steps: 65490, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001630, Sample Num: 26080, Cur Loss: 0.28966159, Cur Avg Loss: 0.54350091, Log Avg loss: 0.57476462, Global Avg Loss: 1.50717145, Time: 0.0118 Steps: 65500, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001640, Sample Num: 26240, Cur Loss: 0.54297924, Cur Avg Loss: 0.54304600, Log Avg loss: 0.46889490, Global Avg Loss: 1.50701296, Time: 0.0130 Steps: 65510, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001650, Sample Num: 26400, Cur Loss: 0.51947939, Cur Avg Loss: 0.54264822, Log Avg loss: 0.47741271, Global Avg Loss: 1.50685582, Time: 0.0106 Steps: 65520, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001660, Sample Num: 26560, Cur Loss: 1.48282802, Cur Avg Loss: 0.54383223, Log Avg loss: 0.73919441, Global Avg Loss: 1.50673867, Time: 0.0227 Steps: 65530, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001670, Sample Num: 26720, Cur Loss: 0.47056088, Cur Avg Loss: 0.54491976, Log Avg loss: 0.72544850, Global Avg Loss: 1.50661946, Time: 0.0133 Steps: 65540, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001680, Sample Num: 26880, Cur Loss: 0.34520653, Cur Avg Loss: 0.54474073, Log Avg loss: 0.51484377, Global Avg Loss: 1.50646816, Time: 0.0096 Steps: 65550, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001690, Sample Num: 27040, Cur Loss: 0.22809693, Cur Avg Loss: 0.54403478, Log Avg loss: 0.42543425, Global Avg Loss: 1.50630327, Time: 0.0068 Steps: 65560, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001700, Sample Num: 27200, Cur Loss: 0.30683920, Cur Avg Loss: 0.54383694, Log Avg loss: 0.51040329, Global Avg Loss: 1.50615139, Time: 0.0220 Steps: 65570, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001710, Sample Num: 27360, Cur Loss: 0.80809200, Cur Avg Loss: 0.54421728, Log Avg loss: 0.60887438, Global Avg Loss: 1.50601457, Time: 0.0068 Steps: 65580, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001720, Sample Num: 27520, Cur Loss: 0.40138930, Cur Avg Loss: 0.54453849, Log Avg loss: 0.59946628, Global Avg Loss: 1.50587635, Time: 0.0067 Steps: 65590, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001730, Sample Num: 27680, Cur Loss: 0.30034590, Cur Avg Loss: 0.54410053, Log Avg loss: 0.46877011, Global Avg Loss: 1.50571826, Time: 0.0155 Steps: 65600, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001740, Sample Num: 27840, Cur Loss: 0.45234704, Cur Avg Loss: 0.54416911, Log Avg loss: 0.55603316, Global Avg Loss: 1.50557351, Time: 0.0072 Steps: 65610, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001750, Sample Num: 28000, Cur Loss: 0.32896447, Cur Avg Loss: 0.54439775, Log Avg loss: 0.58418224, Global Avg Loss: 1.50543310, Time: 0.0094 Steps: 65620, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001760, Sample Num: 28160, Cur Loss: 0.53133738, Cur Avg Loss: 0.54624808, Log Avg loss: 0.87005540, Global Avg Loss: 1.50533628, Time: 0.0116 Steps: 65630, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001770, Sample Num: 28320, Cur Loss: 0.27214649, Cur Avg Loss: 0.54694039, Log Avg loss: 0.66878730, Global Avg Loss: 1.50520884, Time: 0.0120 Steps: 65640, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001780, Sample Num: 28480, Cur Loss: 0.30723822, Cur Avg Loss: 0.54780264, Log Avg loss: 0.70042130, Global Avg Loss: 1.50508625, Time: 0.0111 Steps: 65650, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001790, Sample Num: 28640, Cur Loss: 0.35108641, Cur Avg Loss: 0.54773077, Log Avg loss: 0.53493647, Global Avg Loss: 1.50493850, Time: 0.0108 Steps: 65660, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001800, Sample Num: 28800, Cur Loss: 0.33098310, Cur Avg Loss: 0.54830474, Log Avg loss: 0.65104616, Global Avg Loss: 1.50480847, Time: 0.0208 Steps: 65670, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001810, Sample Num: 28960, Cur Loss: 0.52428710, Cur Avg Loss: 0.54781703, Log Avg loss: 0.46002943, Global Avg Loss: 1.50464940, Time: 0.0066 Steps: 65680, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001820, Sample Num: 29120, Cur Loss: 0.61048031, Cur Avg Loss: 0.54739582, Log Avg loss: 0.47115696, Global Avg Loss: 1.50449207, Time: 0.0066 Steps: 65690, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001830, Sample Num: 29280, Cur Loss: 0.55122411, Cur Avg Loss: 0.54668271, Log Avg loss: 0.41689628, Global Avg Loss: 1.50432653, Time: 0.0068 Steps: 65700, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001840, Sample Num: 29440, Cur Loss: 0.53490967, Cur Avg Loss: 0.54668464, Log Avg loss: 0.54703752, Global Avg Loss: 1.50418085, Time: 0.0112 Steps: 65710, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001850, Sample Num: 29600, Cur Loss: 0.29899913, Cur Avg Loss: 0.54614990, Log Avg loss: 0.44775857, Global Avg Loss: 1.50402010, Time: 0.0150 Steps: 65720, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001860, Sample Num: 29760, Cur Loss: 0.25558004, Cur Avg Loss: 0.54597471, Log Avg loss: 0.51356371, Global Avg Loss: 1.50386941, Time: 0.0068 Steps: 65730, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001870, Sample Num: 29920, Cur Loss: 0.91056252, Cur Avg Loss: 0.54676741, Log Avg loss: 0.69421016, Global Avg Loss: 1.50374625, Time: 0.0070 Steps: 65740, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001880, Sample Num: 30080, Cur Loss: 0.47200167, Cur Avg Loss: 0.54662487, Log Avg loss: 0.51996934, Global Avg Loss: 1.50359663, Time: 0.0118 Steps: 65750, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001890, Sample Num: 30240, Cur Loss: 0.23000017, Cur Avg Loss: 0.54598075, Log Avg loss: 0.42488739, Global Avg Loss: 1.50343259, Time: 0.0141 Steps: 65760, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001900, Sample Num: 30400, Cur Loss: 0.39387169, Cur Avg Loss: 0.54614151, Log Avg loss: 0.57652441, Global Avg Loss: 1.50329166, Time: 0.0120 Steps: 65770, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001910, Sample Num: 30560, Cur Loss: 0.22326174, Cur Avg Loss: 0.54658469, Log Avg loss: 0.63078934, Global Avg Loss: 1.50315902, Time: 0.0113 Steps: 65780, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001920, Sample Num: 30720, Cur Loss: 0.14235035, Cur Avg Loss: 0.54679483, Log Avg loss: 0.58693107, Global Avg Loss: 1.50301976, Time: 0.0065 Steps: 65790, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001930, Sample Num: 30880, Cur Loss: 0.81393814, Cur Avg Loss: 0.54647730, Log Avg loss: 0.48551169, Global Avg Loss: 1.50286512, Time: 0.0135 Steps: 65800, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001940, Sample Num: 31040, Cur Loss: 0.60472792, Cur Avg Loss: 0.54735842, Log Avg loss: 0.71741522, Global Avg Loss: 1.50274577, Time: 0.0115 Steps: 65810, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001950, Sample Num: 31200, Cur Loss: 0.49301323, Cur Avg Loss: 0.54744549, Log Avg loss: 0.56433558, Global Avg Loss: 1.50260320, Time: 0.0071 Steps: 65820, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001960, Sample Num: 31360, Cur Loss: 0.18715625, Cur Avg Loss: 0.54791238, Log Avg loss: 0.63895608, Global Avg Loss: 1.50247200, Time: 0.0072 Steps: 65830, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001970, Sample Num: 31520, Cur Loss: 0.75210804, Cur Avg Loss: 0.54781531, Log Avg loss: 0.52879077, Global Avg Loss: 1.50232412, Time: 0.0067 Steps: 65840, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001980, Sample Num: 31680, Cur Loss: 0.34090549, Cur Avg Loss: 0.54760467, Log Avg loss: 0.50610792, Global Avg Loss: 1.50217283, Time: 0.0107 Steps: 65850, Updated lr: 0.000039 Training, Epoch: 0031, Batch: 001990, Sample Num: 31840, Cur Loss: 0.46304679, Cur Avg Loss: 0.54789423, Log Avg loss: 0.60522674, Global Avg Loss: 1.50203664, Time: 0.0067 Steps: 65860, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002000, Sample Num: 32000, Cur Loss: 0.58800918, Cur Avg Loss: 0.54783484, Log Avg loss: 0.53601720, Global Avg Loss: 1.50188999, Time: 0.0121 Steps: 65870, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002010, Sample Num: 32160, Cur Loss: 0.73934811, Cur Avg Loss: 0.54810303, Log Avg loss: 0.60174073, Global Avg Loss: 1.50175335, Time: 0.0067 Steps: 65880, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002020, Sample Num: 32320, Cur Loss: 1.12111485, Cur Avg Loss: 0.54777437, Log Avg loss: 0.48171275, Global Avg Loss: 1.50159854, Time: 0.0066 Steps: 65890, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002030, Sample Num: 32480, Cur Loss: 0.79446536, Cur Avg Loss: 0.54813710, Log Avg loss: 0.62140925, Global Avg Loss: 1.50146498, Time: 0.0073 Steps: 65900, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002040, Sample Num: 32640, Cur Loss: 0.07280207, Cur Avg Loss: 0.54790407, Log Avg loss: 0.50059834, Global Avg Loss: 1.50131312, Time: 0.0094 Steps: 65910, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002050, Sample Num: 32800, Cur Loss: 0.34363598, Cur Avg Loss: 0.54695000, Log Avg loss: 0.35232080, Global Avg Loss: 1.50113882, Time: 0.0117 Steps: 65920, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002060, Sample Num: 32960, Cur Loss: 0.38978207, Cur Avg Loss: 0.54652716, Log Avg loss: 0.45984358, Global Avg Loss: 1.50098088, Time: 0.0109 Steps: 65930, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002070, Sample Num: 33120, Cur Loss: 0.30158031, Cur Avg Loss: 0.54682018, Log Avg loss: 0.60718340, Global Avg Loss: 1.50084534, Time: 0.0132 Steps: 65940, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002080, Sample Num: 33280, Cur Loss: 0.15710114, Cur Avg Loss: 0.54630923, Log Avg loss: 0.44054146, Global Avg Loss: 1.50068456, Time: 0.0185 Steps: 65950, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002090, Sample Num: 33440, Cur Loss: 0.64346385, Cur Avg Loss: 0.54656823, Log Avg loss: 0.60044086, Global Avg Loss: 1.50054808, Time: 0.0073 Steps: 65960, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002100, Sample Num: 33600, Cur Loss: 0.79516602, Cur Avg Loss: 0.54740014, Log Avg loss: 0.72126929, Global Avg Loss: 1.50042995, Time: 0.0068 Steps: 65970, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002110, Sample Num: 33760, Cur Loss: 0.51200819, Cur Avg Loss: 0.54684781, Log Avg loss: 0.43085788, Global Avg Loss: 1.50026785, Time: 0.0114 Steps: 65980, Updated lr: 0.000038 Training, Epoch: 0031, Batch: 002120, Sample Num: 33920, Cur Loss: 0.38624439, Cur Avg Loss: 0.54589488, Log Avg loss: 0.34482665, Global Avg Loss: 1.50009275, Time: 0.0067 Steps: 65990, Updated lr: 0.000038 ***** Running evaluation checkpoint-65999 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-65999 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.195659, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.625909, "eval_total_loss": 440.014016, "eval_mae": 0.600381, "eval_mse": 0.626086, "eval_r2": 0.602019, "eval_sp_statistic": 0.733801, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.778048, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.891227, "test_total_loss": 447.395807, "test_mae": 0.696301, "test_mse": 0.891434, "test_r2": 0.424661, "test_sp_statistic": 0.566038, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.652333, "test_ps_pvalue": 0.0, "lr": 3.8360360360360366e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4999380639165052, "train_cur_epoch_loss": 1160.588632054627, "train_cur_epoch_avg_loss": 0.5451332231350996, "train_cur_epoch_time": 22.195658683776855, "train_cur_epoch_avg_time": 0.010425391584676775, "epoch": 31, "step": 65999} ################################################## Training, Epoch: 0032, Batch: 000001, Sample Num: 16, Cur Loss: 2.06104326, Cur Avg Loss: 2.06104326, Log Avg loss: 0.53525389, Global Avg Loss: 1.49994657, Time: 0.0222 Steps: 66000, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000011, Sample Num: 176, Cur Loss: 0.52550250, Cur Avg Loss: 0.55301284, Log Avg loss: 0.40220980, Global Avg Loss: 1.49978027, Time: 0.0119 Steps: 66010, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000021, Sample Num: 336, Cur Loss: 0.28259957, Cur Avg Loss: 0.45844695, Log Avg loss: 0.35442448, Global Avg Loss: 1.49960678, Time: 0.0063 Steps: 66020, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000031, Sample Num: 496, Cur Loss: 0.44998452, Cur Avg Loss: 0.45365882, Log Avg loss: 0.44360374, Global Avg Loss: 1.49944685, Time: 0.0064 Steps: 66030, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000041, Sample Num: 656, Cur Loss: 0.29623103, Cur Avg Loss: 0.46836208, Log Avg loss: 0.51394217, Global Avg Loss: 1.49929762, Time: 0.0063 Steps: 66040, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000051, Sample Num: 816, Cur Loss: 0.65622681, Cur Avg Loss: 0.44972232, Log Avg loss: 0.37329930, Global Avg Loss: 1.49912715, Time: 0.0064 Steps: 66050, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000061, Sample Num: 976, Cur Loss: 0.74228168, Cur Avg Loss: 0.47971486, Log Avg loss: 0.63267683, Global Avg Loss: 1.49899599, Time: 0.0237 Steps: 66060, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000071, Sample Num: 1136, Cur Loss: 0.45255095, Cur Avg Loss: 0.48419480, Log Avg loss: 0.51152245, Global Avg Loss: 1.49884653, Time: 0.0063 Steps: 66070, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000081, Sample Num: 1296, Cur Loss: 0.48492879, Cur Avg Loss: 0.48658825, Log Avg loss: 0.50358174, Global Avg Loss: 1.49869591, Time: 0.0203 Steps: 66080, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000091, Sample Num: 1456, Cur Loss: 0.73323357, Cur Avg Loss: 0.49460860, Log Avg loss: 0.55957338, Global Avg Loss: 1.49855382, Time: 0.0128 Steps: 66090, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000101, Sample Num: 1616, Cur Loss: 0.56379855, Cur Avg Loss: 0.50543775, Log Avg loss: 0.60398310, Global Avg Loss: 1.49841848, Time: 0.0094 Steps: 66100, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000111, Sample Num: 1776, Cur Loss: 0.50468552, Cur Avg Loss: 0.51569863, Log Avg loss: 0.61933346, Global Avg Loss: 1.49828551, Time: 0.0063 Steps: 66110, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000121, Sample Num: 1936, Cur Loss: 0.66721964, Cur Avg Loss: 0.51930807, Log Avg loss: 0.55937292, Global Avg Loss: 1.49814351, Time: 0.0239 Steps: 66120, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000131, Sample Num: 2096, Cur Loss: 0.28859955, Cur Avg Loss: 0.52308058, Log Avg loss: 0.56872788, Global Avg Loss: 1.49800296, Time: 0.0167 Steps: 66130, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000141, Sample Num: 2256, Cur Loss: 0.24793658, Cur Avg Loss: 0.51060396, Log Avg loss: 0.34716020, Global Avg Loss: 1.49782896, Time: 0.0227 Steps: 66140, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000151, Sample Num: 2416, Cur Loss: 0.18204409, Cur Avg Loss: 0.50220915, Log Avg loss: 0.38384239, Global Avg Loss: 1.49766056, Time: 0.0064 Steps: 66150, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000161, Sample Num: 2576, Cur Loss: 0.33223429, Cur Avg Loss: 0.49651184, Log Avg loss: 0.41048247, Global Avg Loss: 1.49749623, Time: 0.0064 Steps: 66160, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000171, Sample Num: 2736, Cur Loss: 0.24674466, Cur Avg Loss: 0.50372617, Log Avg loss: 0.61987690, Global Avg Loss: 1.49736360, Time: 0.0147 Steps: 66170, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000181, Sample Num: 2896, Cur Loss: 0.53075874, Cur Avg Loss: 0.49806591, Log Avg loss: 0.40127540, Global Avg Loss: 1.49719798, Time: 0.0066 Steps: 66180, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000191, Sample Num: 3056, Cur Loss: 0.30650851, Cur Avg Loss: 0.50269796, Log Avg loss: 0.58653805, Global Avg Loss: 1.49706040, Time: 0.0235 Steps: 66190, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000201, Sample Num: 3216, Cur Loss: 0.52712071, Cur Avg Loss: 0.50200104, Log Avg loss: 0.48868982, Global Avg Loss: 1.49690807, Time: 0.0232 Steps: 66200, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000211, Sample Num: 3376, Cur Loss: 1.00581217, Cur Avg Loss: 0.50763377, Log Avg loss: 0.62085167, Global Avg Loss: 1.49677576, Time: 0.0063 Steps: 66210, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000221, Sample Num: 3536, Cur Loss: 0.79584169, Cur Avg Loss: 0.51533582, Log Avg loss: 0.67784922, Global Avg Loss: 1.49665209, Time: 0.0064 Steps: 66220, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000231, Sample Num: 3696, Cur Loss: 0.48397255, Cur Avg Loss: 0.52016024, Log Avg loss: 0.62677991, Global Avg Loss: 1.49652075, Time: 0.0065 Steps: 66230, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000241, Sample Num: 3856, Cur Loss: 0.34466887, Cur Avg Loss: 0.51756611, Log Avg loss: 0.45764173, Global Avg Loss: 1.49636391, Time: 0.0065 Steps: 66240, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000251, Sample Num: 4016, Cur Loss: 0.73739308, Cur Avg Loss: 0.51910738, Log Avg loss: 0.55625196, Global Avg Loss: 1.49622201, Time: 0.0063 Steps: 66250, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000261, Sample Num: 4176, Cur Loss: 0.38265106, Cur Avg Loss: 0.52034356, Log Avg loss: 0.55137164, Global Avg Loss: 1.49607941, Time: 0.0064 Steps: 66260, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000271, Sample Num: 4336, Cur Loss: 0.31458336, Cur Avg Loss: 0.51757554, Log Avg loss: 0.44533025, Global Avg Loss: 1.49592086, Time: 0.0064 Steps: 66270, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000281, Sample Num: 4496, Cur Loss: 0.42900974, Cur Avg Loss: 0.51798430, Log Avg loss: 0.52906168, Global Avg Loss: 1.49577498, Time: 0.0233 Steps: 66280, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000291, Sample Num: 4656, Cur Loss: 0.44449842, Cur Avg Loss: 0.51255863, Log Avg loss: 0.36009735, Global Avg Loss: 1.49560366, Time: 0.0065 Steps: 66290, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000301, Sample Num: 4816, Cur Loss: 0.15967228, Cur Avg Loss: 0.51794299, Log Avg loss: 0.67462770, Global Avg Loss: 1.49547984, Time: 0.0066 Steps: 66300, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000311, Sample Num: 4976, Cur Loss: 0.25658751, Cur Avg Loss: 0.51450577, Log Avg loss: 0.41104565, Global Avg Loss: 1.49531630, Time: 0.0066 Steps: 66310, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000321, Sample Num: 5136, Cur Loss: 0.16974375, Cur Avg Loss: 0.52222030, Log Avg loss: 0.76214220, Global Avg Loss: 1.49520574, Time: 0.0066 Steps: 66320, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000331, Sample Num: 5296, Cur Loss: 0.26764220, Cur Avg Loss: 0.51659195, Log Avg loss: 0.33592188, Global Avg Loss: 1.49503097, Time: 0.0231 Steps: 66330, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000341, Sample Num: 5456, Cur Loss: 0.18171626, Cur Avg Loss: 0.51702931, Log Avg loss: 0.53150578, Global Avg Loss: 1.49488573, Time: 0.0106 Steps: 66340, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000351, Sample Num: 5616, Cur Loss: 0.24789387, Cur Avg Loss: 0.51817936, Log Avg loss: 0.55739626, Global Avg Loss: 1.49474443, Time: 0.0143 Steps: 66350, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000361, Sample Num: 5776, Cur Loss: 0.28343788, Cur Avg Loss: 0.52146122, Log Avg loss: 0.63665445, Global Avg Loss: 1.49461513, Time: 0.0064 Steps: 66360, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000371, Sample Num: 5936, Cur Loss: 0.36115611, Cur Avg Loss: 0.52098963, Log Avg loss: 0.50396511, Global Avg Loss: 1.49446586, Time: 0.0064 Steps: 66370, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000381, Sample Num: 6096, Cur Loss: 0.62836277, Cur Avg Loss: 0.52478134, Log Avg loss: 0.66545380, Global Avg Loss: 1.49434098, Time: 0.0228 Steps: 66380, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000391, Sample Num: 6256, Cur Loss: 0.58872914, Cur Avg Loss: 0.52837943, Log Avg loss: 0.66546652, Global Avg Loss: 1.49421613, Time: 0.0154 Steps: 66390, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000401, Sample Num: 6416, Cur Loss: 0.38393182, Cur Avg Loss: 0.52909755, Log Avg loss: 0.55717627, Global Avg Loss: 1.49407501, Time: 0.0065 Steps: 66400, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000411, Sample Num: 6576, Cur Loss: 0.73362178, Cur Avg Loss: 0.52560312, Log Avg loss: 0.38547656, Global Avg Loss: 1.49390807, Time: 0.0064 Steps: 66410, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000421, Sample Num: 6736, Cur Loss: 0.48637211, Cur Avg Loss: 0.52542924, Log Avg loss: 0.51828243, Global Avg Loss: 1.49376119, Time: 0.0152 Steps: 66420, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000431, Sample Num: 6896, Cur Loss: 0.18148573, Cur Avg Loss: 0.52739884, Log Avg loss: 0.61031923, Global Avg Loss: 1.49362820, Time: 0.0064 Steps: 66430, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000441, Sample Num: 7056, Cur Loss: 0.09176510, Cur Avg Loss: 0.52693309, Log Avg loss: 0.50685930, Global Avg Loss: 1.49347968, Time: 0.0066 Steps: 66440, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000451, Sample Num: 7216, Cur Loss: 0.64392066, Cur Avg Loss: 0.52955093, Log Avg loss: 0.64499753, Global Avg Loss: 1.49335199, Time: 0.0232 Steps: 66450, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000461, Sample Num: 7376, Cur Loss: 0.49466920, Cur Avg Loss: 0.53031639, Log Avg loss: 0.56483871, Global Avg Loss: 1.49321228, Time: 0.0221 Steps: 66460, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000471, Sample Num: 7536, Cur Loss: 0.21532241, Cur Avg Loss: 0.52953974, Log Avg loss: 0.49373621, Global Avg Loss: 1.49306192, Time: 0.0063 Steps: 66470, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000481, Sample Num: 7696, Cur Loss: 1.02696693, Cur Avg Loss: 0.53302288, Log Avg loss: 0.69707884, Global Avg Loss: 1.49294218, Time: 0.0064 Steps: 66480, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000491, Sample Num: 7856, Cur Loss: 0.14264950, Cur Avg Loss: 0.53140657, Log Avg loss: 0.45366211, Global Avg Loss: 1.49278588, Time: 0.0065 Steps: 66490, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000501, Sample Num: 8016, Cur Loss: 0.44695169, Cur Avg Loss: 0.52856210, Log Avg loss: 0.38889834, Global Avg Loss: 1.49261988, Time: 0.0066 Steps: 66500, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000511, Sample Num: 8176, Cur Loss: 0.39905357, Cur Avg Loss: 0.52954826, Log Avg loss: 0.57895495, Global Avg Loss: 1.49248251, Time: 0.0232 Steps: 66510, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000521, Sample Num: 8336, Cur Loss: 0.71590275, Cur Avg Loss: 0.52893555, Log Avg loss: 0.49762636, Global Avg Loss: 1.49233295, Time: 0.0083 Steps: 66520, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000531, Sample Num: 8496, Cur Loss: 0.38616526, Cur Avg Loss: 0.52904556, Log Avg loss: 0.53477697, Global Avg Loss: 1.49218902, Time: 0.0066 Steps: 66530, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000541, Sample Num: 8656, Cur Loss: 0.57645828, Cur Avg Loss: 0.52871835, Log Avg loss: 0.51134361, Global Avg Loss: 1.49204161, Time: 0.0122 Steps: 66540, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000551, Sample Num: 8816, Cur Loss: 0.51332688, Cur Avg Loss: 0.53015530, Log Avg loss: 0.60789393, Global Avg Loss: 1.49190876, Time: 0.0094 Steps: 66550, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000561, Sample Num: 8976, Cur Loss: 0.66039574, Cur Avg Loss: 0.53134333, Log Avg loss: 0.59680367, Global Avg Loss: 1.49177428, Time: 0.0112 Steps: 66560, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000571, Sample Num: 9136, Cur Loss: 0.21613616, Cur Avg Loss: 0.53021353, Log Avg loss: 0.46683174, Global Avg Loss: 1.49162031, Time: 0.0069 Steps: 66570, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000581, Sample Num: 9296, Cur Loss: 0.77452898, Cur Avg Loss: 0.52995411, Log Avg loss: 0.51514170, Global Avg Loss: 1.49147365, Time: 0.0114 Steps: 66580, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000591, Sample Num: 9456, Cur Loss: 0.30850127, Cur Avg Loss: 0.52753679, Log Avg loss: 0.38709051, Global Avg Loss: 1.49130780, Time: 0.0099 Steps: 66590, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000601, Sample Num: 9616, Cur Loss: 0.23080555, Cur Avg Loss: 0.52714740, Log Avg loss: 0.50413422, Global Avg Loss: 1.49115958, Time: 0.0098 Steps: 66600, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000611, Sample Num: 9776, Cur Loss: 1.10870790, Cur Avg Loss: 0.52901544, Log Avg loss: 0.64128439, Global Avg Loss: 1.49103199, Time: 0.0134 Steps: 66610, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000621, Sample Num: 9936, Cur Loss: 0.21470034, Cur Avg Loss: 0.52720484, Log Avg loss: 0.41657750, Global Avg Loss: 1.49087071, Time: 0.0079 Steps: 66620, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000631, Sample Num: 10096, Cur Loss: 0.52456641, Cur Avg Loss: 0.52439163, Log Avg loss: 0.34969144, Global Avg Loss: 1.49069944, Time: 0.0071 Steps: 66630, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000641, Sample Num: 10256, Cur Loss: 0.30383199, Cur Avg Loss: 0.52633930, Log Avg loss: 0.64923689, Global Avg Loss: 1.49057317, Time: 0.0102 Steps: 66640, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000651, Sample Num: 10416, Cur Loss: 0.79209936, Cur Avg Loss: 0.52505459, Log Avg loss: 0.44270451, Global Avg Loss: 1.49041595, Time: 0.0133 Steps: 66650, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000661, Sample Num: 10576, Cur Loss: 0.19164565, Cur Avg Loss: 0.52380639, Log Avg loss: 0.44254898, Global Avg Loss: 1.49025875, Time: 0.0070 Steps: 66660, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000671, Sample Num: 10736, Cur Loss: 0.46911541, Cur Avg Loss: 0.52359006, Log Avg loss: 0.50929059, Global Avg Loss: 1.49011161, Time: 0.0115 Steps: 66670, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000681, Sample Num: 10896, Cur Loss: 0.29031062, Cur Avg Loss: 0.52105711, Log Avg loss: 0.35109611, Global Avg Loss: 1.48994079, Time: 0.0127 Steps: 66680, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000691, Sample Num: 11056, Cur Loss: 1.03990948, Cur Avg Loss: 0.52083149, Log Avg loss: 0.50546664, Global Avg Loss: 1.48979317, Time: 0.0109 Steps: 66690, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000701, Sample Num: 11216, Cur Loss: 0.91375744, Cur Avg Loss: 0.52319678, Log Avg loss: 0.68663860, Global Avg Loss: 1.48967276, Time: 0.0108 Steps: 66700, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000711, Sample Num: 11376, Cur Loss: 0.14119977, Cur Avg Loss: 0.52210742, Log Avg loss: 0.44574309, Global Avg Loss: 1.48951627, Time: 0.0087 Steps: 66710, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000721, Sample Num: 11536, Cur Loss: 0.22478122, Cur Avg Loss: 0.52296907, Log Avg loss: 0.58423260, Global Avg Loss: 1.48938059, Time: 0.0071 Steps: 66720, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000731, Sample Num: 11696, Cur Loss: 0.46699834, Cur Avg Loss: 0.52266078, Log Avg loss: 0.50043305, Global Avg Loss: 1.48923239, Time: 0.0104 Steps: 66730, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000741, Sample Num: 11856, Cur Loss: 0.80822253, Cur Avg Loss: 0.52435169, Log Avg loss: 0.64795684, Global Avg Loss: 1.48910634, Time: 0.0103 Steps: 66740, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000751, Sample Num: 12016, Cur Loss: 0.97238564, Cur Avg Loss: 0.52529678, Log Avg loss: 0.59532848, Global Avg Loss: 1.48897244, Time: 0.0113 Steps: 66750, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000761, Sample Num: 12176, Cur Loss: 1.19594955, Cur Avg Loss: 0.52648684, Log Avg loss: 0.61585995, Global Avg Loss: 1.48884165, Time: 0.0127 Steps: 66760, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000771, Sample Num: 12336, Cur Loss: 0.47335815, Cur Avg Loss: 0.52589752, Log Avg loss: 0.48105045, Global Avg Loss: 1.48869072, Time: 0.0071 Steps: 66770, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000781, Sample Num: 12496, Cur Loss: 0.32493055, Cur Avg Loss: 0.52706812, Log Avg loss: 0.61732113, Global Avg Loss: 1.48856023, Time: 0.0072 Steps: 66780, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000791, Sample Num: 12656, Cur Loss: 0.72105211, Cur Avg Loss: 0.52834503, Log Avg loss: 0.62807184, Global Avg Loss: 1.48843140, Time: 0.0075 Steps: 66790, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000801, Sample Num: 12816, Cur Loss: 0.30410659, Cur Avg Loss: 0.52815752, Log Avg loss: 0.51332530, Global Avg Loss: 1.48828543, Time: 0.0071 Steps: 66800, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000811, Sample Num: 12976, Cur Loss: 0.26983038, Cur Avg Loss: 0.52989739, Log Avg loss: 0.66926128, Global Avg Loss: 1.48816284, Time: 0.0114 Steps: 66810, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000821, Sample Num: 13136, Cur Loss: 0.55971122, Cur Avg Loss: 0.52881260, Log Avg loss: 0.44083628, Global Avg Loss: 1.48800610, Time: 0.0095 Steps: 66820, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000831, Sample Num: 13296, Cur Loss: 0.28577995, Cur Avg Loss: 0.52978009, Log Avg loss: 0.60921102, Global Avg Loss: 1.48787460, Time: 0.0066 Steps: 66830, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000841, Sample Num: 13456, Cur Loss: 0.50646293, Cur Avg Loss: 0.53141833, Log Avg loss: 0.66755580, Global Avg Loss: 1.48775187, Time: 0.0064 Steps: 66840, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000851, Sample Num: 13616, Cur Loss: 0.59390920, Cur Avg Loss: 0.53155526, Log Avg loss: 0.54307091, Global Avg Loss: 1.48761056, Time: 0.0065 Steps: 66850, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000861, Sample Num: 13776, Cur Loss: 0.26856712, Cur Avg Loss: 0.53137411, Log Avg loss: 0.51595842, Global Avg Loss: 1.48746523, Time: 0.0065 Steps: 66860, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000871, Sample Num: 13936, Cur Loss: 0.55089760, Cur Avg Loss: 0.53061177, Log Avg loss: 0.46497431, Global Avg Loss: 1.48731232, Time: 0.0067 Steps: 66870, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000881, Sample Num: 14096, Cur Loss: 0.12317351, Cur Avg Loss: 0.52921393, Log Avg loss: 0.40746236, Global Avg Loss: 1.48715086, Time: 0.0072 Steps: 66880, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000891, Sample Num: 14256, Cur Loss: 0.37201801, Cur Avg Loss: 0.52921004, Log Avg loss: 0.52886746, Global Avg Loss: 1.48700760, Time: 0.0219 Steps: 66890, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000901, Sample Num: 14416, Cur Loss: 0.57833630, Cur Avg Loss: 0.52892005, Log Avg loss: 0.50308138, Global Avg Loss: 1.48686053, Time: 0.0066 Steps: 66900, Updated lr: 0.000038 Training, Epoch: 0032, Batch: 000911, Sample Num: 14576, Cur Loss: 0.25780752, Cur Avg Loss: 0.52881413, Log Avg loss: 0.51927047, Global Avg Loss: 1.48671592, Time: 0.0066 Steps: 66910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000921, Sample Num: 14736, Cur Loss: 0.42467681, Cur Avg Loss: 0.52832179, Log Avg loss: 0.48346989, Global Avg Loss: 1.48656600, Time: 0.0068 Steps: 66920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000931, Sample Num: 14896, Cur Loss: 0.67143476, Cur Avg Loss: 0.52837107, Log Avg loss: 0.53290969, Global Avg Loss: 1.48642351, Time: 0.0097 Steps: 66930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000941, Sample Num: 15056, Cur Loss: 0.51475406, Cur Avg Loss: 0.52870477, Log Avg loss: 0.55977207, Global Avg Loss: 1.48628508, Time: 0.0123 Steps: 66940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000951, Sample Num: 15216, Cur Loss: 0.22172371, Cur Avg Loss: 0.52970242, Log Avg loss: 0.62358202, Global Avg Loss: 1.48615623, Time: 0.0065 Steps: 66950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000961, Sample Num: 15376, Cur Loss: 0.32288823, Cur Avg Loss: 0.53056784, Log Avg loss: 0.61286896, Global Avg Loss: 1.48602581, Time: 0.0068 Steps: 66960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000971, Sample Num: 15536, Cur Loss: 0.55800140, Cur Avg Loss: 0.53321883, Log Avg loss: 0.78797915, Global Avg Loss: 1.48592157, Time: 0.0116 Steps: 66970, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000981, Sample Num: 15696, Cur Loss: 0.32816970, Cur Avg Loss: 0.53214849, Log Avg loss: 0.42821779, Global Avg Loss: 1.48576366, Time: 0.0116 Steps: 66980, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 000991, Sample Num: 15856, Cur Loss: 0.48563415, Cur Avg Loss: 0.53204437, Log Avg loss: 0.52183058, Global Avg Loss: 1.48561977, Time: 0.0072 Steps: 66990, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001001, Sample Num: 16016, Cur Loss: 0.35874552, Cur Avg Loss: 0.53278970, Log Avg loss: 0.60665223, Global Avg Loss: 1.48548858, Time: 0.0120 Steps: 67000, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001011, Sample Num: 16176, Cur Loss: 0.48157907, Cur Avg Loss: 0.53396331, Log Avg loss: 0.65144192, Global Avg Loss: 1.48536411, Time: 0.0067 Steps: 67010, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001021, Sample Num: 16336, Cur Loss: 0.92311180, Cur Avg Loss: 0.53599295, Log Avg loss: 0.74118944, Global Avg Loss: 1.48525307, Time: 0.0124 Steps: 67020, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001031, Sample Num: 16496, Cur Loss: 0.28129512, Cur Avg Loss: 0.53656311, Log Avg loss: 0.59477574, Global Avg Loss: 1.48512023, Time: 0.0077 Steps: 67030, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001041, Sample Num: 16656, Cur Loss: 0.26968640, Cur Avg Loss: 0.53668165, Log Avg loss: 0.54890324, Global Avg Loss: 1.48498058, Time: 0.0074 Steps: 67040, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001051, Sample Num: 16816, Cur Loss: 0.63606787, Cur Avg Loss: 0.53749042, Log Avg loss: 0.62168384, Global Avg Loss: 1.48485182, Time: 0.0119 Steps: 67050, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001061, Sample Num: 16976, Cur Loss: 0.83659697, Cur Avg Loss: 0.53784034, Log Avg loss: 0.57461616, Global Avg Loss: 1.48471609, Time: 0.0118 Steps: 67060, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001071, Sample Num: 17136, Cur Loss: 0.48260695, Cur Avg Loss: 0.53767495, Log Avg loss: 0.52012802, Global Avg Loss: 1.48457227, Time: 0.0106 Steps: 67070, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001081, Sample Num: 17296, Cur Loss: 0.36814946, Cur Avg Loss: 0.53733490, Log Avg loss: 0.50091542, Global Avg Loss: 1.48442563, Time: 0.0067 Steps: 67080, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001091, Sample Num: 17456, Cur Loss: 1.04885018, Cur Avg Loss: 0.53877739, Log Avg loss: 0.69471061, Global Avg Loss: 1.48430792, Time: 0.0067 Steps: 67090, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001101, Sample Num: 17616, Cur Loss: 0.62566149, Cur Avg Loss: 0.53872534, Log Avg loss: 0.53304636, Global Avg Loss: 1.48416615, Time: 0.0110 Steps: 67100, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001111, Sample Num: 17776, Cur Loss: 0.61410189, Cur Avg Loss: 0.53843732, Log Avg loss: 0.50672651, Global Avg Loss: 1.48402051, Time: 0.0104 Steps: 67110, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001121, Sample Num: 17936, Cur Loss: 0.38814667, Cur Avg Loss: 0.53888259, Log Avg loss: 0.58835157, Global Avg Loss: 1.48388706, Time: 0.0123 Steps: 67120, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001131, Sample Num: 18096, Cur Loss: 0.46776047, Cur Avg Loss: 0.53893528, Log Avg loss: 0.54484185, Global Avg Loss: 1.48374718, Time: 0.0117 Steps: 67130, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001141, Sample Num: 18256, Cur Loss: 0.27727643, Cur Avg Loss: 0.53968284, Log Avg loss: 0.62423273, Global Avg Loss: 1.48361916, Time: 0.0073 Steps: 67140, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001151, Sample Num: 18416, Cur Loss: 0.47273439, Cur Avg Loss: 0.53891464, Log Avg loss: 0.45126216, Global Avg Loss: 1.48346542, Time: 0.0116 Steps: 67150, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001161, Sample Num: 18576, Cur Loss: 0.27867675, Cur Avg Loss: 0.53812577, Log Avg loss: 0.44732701, Global Avg Loss: 1.48331114, Time: 0.0067 Steps: 67160, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001171, Sample Num: 18736, Cur Loss: 0.94466680, Cur Avg Loss: 0.53889235, Log Avg loss: 0.62789200, Global Avg Loss: 1.48318379, Time: 0.0069 Steps: 67170, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001181, Sample Num: 18896, Cur Loss: 0.28442776, Cur Avg Loss: 0.53786719, Log Avg loss: 0.41782112, Global Avg Loss: 1.48302521, Time: 0.0104 Steps: 67180, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001191, Sample Num: 19056, Cur Loss: 0.41197702, Cur Avg Loss: 0.53735040, Log Avg loss: 0.47631750, Global Avg Loss: 1.48287538, Time: 0.0137 Steps: 67190, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001201, Sample Num: 19216, Cur Loss: 0.21969390, Cur Avg Loss: 0.53757007, Log Avg loss: 0.56373329, Global Avg Loss: 1.48273860, Time: 0.0183 Steps: 67200, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001211, Sample Num: 19376, Cur Loss: 0.44080415, Cur Avg Loss: 0.53797274, Log Avg loss: 0.58633282, Global Avg Loss: 1.48260523, Time: 0.0107 Steps: 67210, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001221, Sample Num: 19536, Cur Loss: 0.53191626, Cur Avg Loss: 0.53700588, Log Avg loss: 0.41991947, Global Avg Loss: 1.48244714, Time: 0.0071 Steps: 67220, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001231, Sample Num: 19696, Cur Loss: 0.34388390, Cur Avg Loss: 0.53793573, Log Avg loss: 0.65147064, Global Avg Loss: 1.48232353, Time: 0.0107 Steps: 67230, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001241, Sample Num: 19856, Cur Loss: 0.33690077, Cur Avg Loss: 0.53636553, Log Avg loss: 0.34307305, Global Avg Loss: 1.48215410, Time: 0.0067 Steps: 67240, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001251, Sample Num: 20016, Cur Loss: 0.19185486, Cur Avg Loss: 0.53761915, Log Avg loss: 0.69319372, Global Avg Loss: 1.48203679, Time: 0.0087 Steps: 67250, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001261, Sample Num: 20176, Cur Loss: 0.54366505, Cur Avg Loss: 0.53824618, Log Avg loss: 0.61668790, Global Avg Loss: 1.48190813, Time: 0.0116 Steps: 67260, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001271, Sample Num: 20336, Cur Loss: 0.29689616, Cur Avg Loss: 0.53811224, Log Avg loss: 0.52122287, Global Avg Loss: 1.48176532, Time: 0.0118 Steps: 67270, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001281, Sample Num: 20496, Cur Loss: 0.24975276, Cur Avg Loss: 0.53834012, Log Avg loss: 0.56730288, Global Avg Loss: 1.48162940, Time: 0.0155 Steps: 67280, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001291, Sample Num: 20656, Cur Loss: 0.71804208, Cur Avg Loss: 0.53824257, Log Avg loss: 0.52574632, Global Avg Loss: 1.48148734, Time: 0.0108 Steps: 67290, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001301, Sample Num: 20816, Cur Loss: 0.32362816, Cur Avg Loss: 0.53859928, Log Avg loss: 0.58465144, Global Avg Loss: 1.48135409, Time: 0.0086 Steps: 67300, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001311, Sample Num: 20976, Cur Loss: 0.42938089, Cur Avg Loss: 0.53775584, Log Avg loss: 0.42802376, Global Avg Loss: 1.48119760, Time: 0.0105 Steps: 67310, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001321, Sample Num: 21136, Cur Loss: 0.31522962, Cur Avg Loss: 0.53926461, Log Avg loss: 0.73706407, Global Avg Loss: 1.48108706, Time: 0.0082 Steps: 67320, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001331, Sample Num: 21296, Cur Loss: 0.26796561, Cur Avg Loss: 0.53954152, Log Avg loss: 0.57612187, Global Avg Loss: 1.48095265, Time: 0.0064 Steps: 67330, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001341, Sample Num: 21456, Cur Loss: 0.47410375, Cur Avg Loss: 0.53940048, Log Avg loss: 0.52062754, Global Avg Loss: 1.48081004, Time: 0.0065 Steps: 67340, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001351, Sample Num: 21616, Cur Loss: 0.74763656, Cur Avg Loss: 0.54064270, Log Avg loss: 0.70722419, Global Avg Loss: 1.48069518, Time: 0.0066 Steps: 67350, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001361, Sample Num: 21776, Cur Loss: 0.97739935, Cur Avg Loss: 0.54005041, Log Avg loss: 0.46003285, Global Avg Loss: 1.48054366, Time: 0.0066 Steps: 67360, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001371, Sample Num: 21936, Cur Loss: 1.24230814, Cur Avg Loss: 0.54031547, Log Avg loss: 0.57639050, Global Avg Loss: 1.48040945, Time: 0.0074 Steps: 67370, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001381, Sample Num: 22096, Cur Loss: 0.71983933, Cur Avg Loss: 0.54126579, Log Avg loss: 0.67155363, Global Avg Loss: 1.48028941, Time: 0.0067 Steps: 67380, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001391, Sample Num: 22256, Cur Loss: 0.65202332, Cur Avg Loss: 0.54164102, Log Avg loss: 0.59346034, Global Avg Loss: 1.48015781, Time: 0.0069 Steps: 67390, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001401, Sample Num: 22416, Cur Loss: 0.81959003, Cur Avg Loss: 0.54284246, Log Avg loss: 0.70996353, Global Avg Loss: 1.48004354, Time: 0.0075 Steps: 67400, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001411, Sample Num: 22576, Cur Loss: 0.75564736, Cur Avg Loss: 0.54292843, Log Avg loss: 0.55497209, Global Avg Loss: 1.47990631, Time: 0.0228 Steps: 67410, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001421, Sample Num: 22736, Cur Loss: 0.27105004, Cur Avg Loss: 0.54201320, Log Avg loss: 0.41287511, Global Avg Loss: 1.47974804, Time: 0.0067 Steps: 67420, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001431, Sample Num: 22896, Cur Loss: 0.97336292, Cur Avg Loss: 0.54187800, Log Avg loss: 0.52266529, Global Avg Loss: 1.47960611, Time: 0.0073 Steps: 67430, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001441, Sample Num: 23056, Cur Loss: 0.55008876, Cur Avg Loss: 0.54217935, Log Avg loss: 0.58530242, Global Avg Loss: 1.47947350, Time: 0.0073 Steps: 67440, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001451, Sample Num: 23216, Cur Loss: 0.25263906, Cur Avg Loss: 0.54158241, Log Avg loss: 0.45556409, Global Avg Loss: 1.47932170, Time: 0.0068 Steps: 67450, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001461, Sample Num: 23376, Cur Loss: 0.22819084, Cur Avg Loss: 0.54053963, Log Avg loss: 0.38923138, Global Avg Loss: 1.47916010, Time: 0.0065 Steps: 67460, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001471, Sample Num: 23536, Cur Loss: 0.39080811, Cur Avg Loss: 0.54067977, Log Avg loss: 0.56115521, Global Avg Loss: 1.47902404, Time: 0.0069 Steps: 67470, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001481, Sample Num: 23696, Cur Loss: 0.57968581, Cur Avg Loss: 0.54049919, Log Avg loss: 0.51393600, Global Avg Loss: 1.47888103, Time: 0.0064 Steps: 67480, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001491, Sample Num: 23856, Cur Loss: 0.91022742, Cur Avg Loss: 0.54183774, Log Avg loss: 0.74007688, Global Avg Loss: 1.47877156, Time: 0.0104 Steps: 67490, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001501, Sample Num: 24016, Cur Loss: 0.29104745, Cur Avg Loss: 0.54043726, Log Avg loss: 0.33162480, Global Avg Loss: 1.47860161, Time: 0.0108 Steps: 67500, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001511, Sample Num: 24176, Cur Loss: 0.28221565, Cur Avg Loss: 0.54033281, Log Avg loss: 0.52465516, Global Avg Loss: 1.47846030, Time: 0.0152 Steps: 67510, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001521, Sample Num: 24336, Cur Loss: 1.08615494, Cur Avg Loss: 0.54009318, Log Avg loss: 0.50388462, Global Avg Loss: 1.47831597, Time: 0.0071 Steps: 67520, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001531, Sample Num: 24496, Cur Loss: 0.82041579, Cur Avg Loss: 0.53964173, Log Avg loss: 0.47097664, Global Avg Loss: 1.47816680, Time: 0.0067 Steps: 67530, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001541, Sample Num: 24656, Cur Loss: 0.34471160, Cur Avg Loss: 0.53935967, Log Avg loss: 0.49617710, Global Avg Loss: 1.47802140, Time: 0.0147 Steps: 67540, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001551, Sample Num: 24816, Cur Loss: 0.41884074, Cur Avg Loss: 0.53891821, Log Avg loss: 0.47088782, Global Avg Loss: 1.47787231, Time: 0.0133 Steps: 67550, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001561, Sample Num: 24976, Cur Loss: 0.37480623, Cur Avg Loss: 0.53954146, Log Avg loss: 0.63620849, Global Avg Loss: 1.47774773, Time: 0.0115 Steps: 67560, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001571, Sample Num: 25136, Cur Loss: 0.67669916, Cur Avg Loss: 0.53877741, Log Avg loss: 0.41950918, Global Avg Loss: 1.47759111, Time: 0.0088 Steps: 67570, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001581, Sample Num: 25296, Cur Loss: 0.59064317, Cur Avg Loss: 0.53859940, Log Avg loss: 0.51063360, Global Avg Loss: 1.47744803, Time: 0.0116 Steps: 67580, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001591, Sample Num: 25456, Cur Loss: 0.42267603, Cur Avg Loss: 0.53904219, Log Avg loss: 0.60904737, Global Avg Loss: 1.47731955, Time: 0.0108 Steps: 67590, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001601, Sample Num: 25616, Cur Loss: 0.27664882, Cur Avg Loss: 0.53979272, Log Avg loss: 0.65920235, Global Avg Loss: 1.47719853, Time: 0.0067 Steps: 67600, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001611, Sample Num: 25776, Cur Loss: 0.15122229, Cur Avg Loss: 0.53983523, Log Avg loss: 0.54664115, Global Avg Loss: 1.47706089, Time: 0.0122 Steps: 67610, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001621, Sample Num: 25936, Cur Loss: 0.41715598, Cur Avg Loss: 0.54016325, Log Avg loss: 0.59300706, Global Avg Loss: 1.47693015, Time: 0.0067 Steps: 67620, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001631, Sample Num: 26096, Cur Loss: 0.19674119, Cur Avg Loss: 0.54009863, Log Avg loss: 0.52962321, Global Avg Loss: 1.47679008, Time: 0.0107 Steps: 67630, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001641, Sample Num: 26256, Cur Loss: 0.32225090, Cur Avg Loss: 0.54018638, Log Avg loss: 0.55449814, Global Avg Loss: 1.47665373, Time: 0.0106 Steps: 67640, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001651, Sample Num: 26416, Cur Loss: 0.74542916, Cur Avg Loss: 0.54050351, Log Avg loss: 0.59254572, Global Avg Loss: 1.47652304, Time: 0.0066 Steps: 67650, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001661, Sample Num: 26576, Cur Loss: 0.29953209, Cur Avg Loss: 0.54080145, Log Avg loss: 0.58999126, Global Avg Loss: 1.47639201, Time: 0.0118 Steps: 67660, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001671, Sample Num: 26736, Cur Loss: 0.43897170, Cur Avg Loss: 0.54170589, Log Avg loss: 0.69193325, Global Avg Loss: 1.47627609, Time: 0.0112 Steps: 67670, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001681, Sample Num: 26896, Cur Loss: 0.79395461, Cur Avg Loss: 0.54275831, Log Avg loss: 0.71861798, Global Avg Loss: 1.47616414, Time: 0.0134 Steps: 67680, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001691, Sample Num: 27056, Cur Loss: 0.28964412, Cur Avg Loss: 0.54188190, Log Avg loss: 0.39455699, Global Avg Loss: 1.47600435, Time: 0.0083 Steps: 67690, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001701, Sample Num: 27216, Cur Loss: 0.47560781, Cur Avg Loss: 0.54268887, Log Avg loss: 0.67914786, Global Avg Loss: 1.47588665, Time: 0.0067 Steps: 67700, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001711, Sample Num: 27376, Cur Loss: 0.44547752, Cur Avg Loss: 0.54217503, Log Avg loss: 0.45477064, Global Avg Loss: 1.47573584, Time: 0.0113 Steps: 67710, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001721, Sample Num: 27536, Cur Loss: 0.67589033, Cur Avg Loss: 0.54183317, Log Avg loss: 0.48334087, Global Avg Loss: 1.47558930, Time: 0.0084 Steps: 67720, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001731, Sample Num: 27696, Cur Loss: 0.99690306, Cur Avg Loss: 0.54222483, Log Avg loss: 0.60962923, Global Avg Loss: 1.47546144, Time: 0.0121 Steps: 67730, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001741, Sample Num: 27856, Cur Loss: 0.74648929, Cur Avg Loss: 0.54197857, Log Avg loss: 0.49935085, Global Avg Loss: 1.47531735, Time: 0.0158 Steps: 67740, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001751, Sample Num: 28016, Cur Loss: 1.34143555, Cur Avg Loss: 0.54321546, Log Avg loss: 0.75855856, Global Avg Loss: 1.47521155, Time: 0.0111 Steps: 67750, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001761, Sample Num: 28176, Cur Loss: 0.33660656, Cur Avg Loss: 0.54221776, Log Avg loss: 0.36751921, Global Avg Loss: 1.47504808, Time: 0.0073 Steps: 67760, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001771, Sample Num: 28336, Cur Loss: 0.34270388, Cur Avg Loss: 0.54300727, Log Avg loss: 0.68204043, Global Avg Loss: 1.47493106, Time: 0.0066 Steps: 67770, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001781, Sample Num: 28496, Cur Loss: 0.46542996, Cur Avg Loss: 0.54249412, Log Avg loss: 0.45161645, Global Avg Loss: 1.47478009, Time: 0.0122 Steps: 67780, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001791, Sample Num: 28656, Cur Loss: 0.22008526, Cur Avg Loss: 0.54220495, Log Avg loss: 0.49070293, Global Avg Loss: 1.47463492, Time: 0.0071 Steps: 67790, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001801, Sample Num: 28816, Cur Loss: 0.67799556, Cur Avg Loss: 0.54268397, Log Avg loss: 0.62847684, Global Avg Loss: 1.47451012, Time: 0.0095 Steps: 67800, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001811, Sample Num: 28976, Cur Loss: 0.16705613, Cur Avg Loss: 0.54168065, Log Avg loss: 0.36098281, Global Avg Loss: 1.47434591, Time: 0.0122 Steps: 67810, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001821, Sample Num: 29136, Cur Loss: 0.95104444, Cur Avg Loss: 0.54167124, Log Avg loss: 0.53996729, Global Avg Loss: 1.47420813, Time: 0.0065 Steps: 67820, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001831, Sample Num: 29296, Cur Loss: 0.68945068, Cur Avg Loss: 0.54183717, Log Avg loss: 0.57205317, Global Avg Loss: 1.47407513, Time: 0.0066 Steps: 67830, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001841, Sample Num: 29456, Cur Loss: 0.18475156, Cur Avg Loss: 0.54191043, Log Avg loss: 0.55532394, Global Avg Loss: 1.47393970, Time: 0.0074 Steps: 67840, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001851, Sample Num: 29616, Cur Loss: 0.39676228, Cur Avg Loss: 0.54110424, Log Avg loss: 0.39268474, Global Avg Loss: 1.47378034, Time: 0.0065 Steps: 67850, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001861, Sample Num: 29776, Cur Loss: 0.47650877, Cur Avg Loss: 0.54120330, Log Avg loss: 0.55953890, Global Avg Loss: 1.47364562, Time: 0.0067 Steps: 67860, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001871, Sample Num: 29936, Cur Loss: 1.38773298, Cur Avg Loss: 0.54143330, Log Avg loss: 0.58423634, Global Avg Loss: 1.47351457, Time: 0.0066 Steps: 67870, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001881, Sample Num: 30096, Cur Loss: 0.48419461, Cur Avg Loss: 0.54175382, Log Avg loss: 0.60172208, Global Avg Loss: 1.47338614, Time: 0.0126 Steps: 67880, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001891, Sample Num: 30256, Cur Loss: 0.42882684, Cur Avg Loss: 0.54144773, Log Avg loss: 0.48387399, Global Avg Loss: 1.47324039, Time: 0.0073 Steps: 67890, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001901, Sample Num: 30416, Cur Loss: 0.84098792, Cur Avg Loss: 0.54122665, Log Avg loss: 0.49942020, Global Avg Loss: 1.47309697, Time: 0.0067 Steps: 67900, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001911, Sample Num: 30576, Cur Loss: 0.11904003, Cur Avg Loss: 0.54140265, Log Avg loss: 0.57486013, Global Avg Loss: 1.47296470, Time: 0.0084 Steps: 67910, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001921, Sample Num: 30736, Cur Loss: 0.52771920, Cur Avg Loss: 0.54151227, Log Avg loss: 0.56245974, Global Avg Loss: 1.47283064, Time: 0.0064 Steps: 67920, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001931, Sample Num: 30896, Cur Loss: 0.23910286, Cur Avg Loss: 0.54109458, Log Avg loss: 0.46085696, Global Avg Loss: 1.47268167, Time: 0.0225 Steps: 67930, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001941, Sample Num: 31056, Cur Loss: 0.39853480, Cur Avg Loss: 0.54043160, Log Avg loss: 0.41240907, Global Avg Loss: 1.47252561, Time: 0.0114 Steps: 67940, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001951, Sample Num: 31216, Cur Loss: 0.45419741, Cur Avg Loss: 0.54070815, Log Avg loss: 0.59438820, Global Avg Loss: 1.47239638, Time: 0.0074 Steps: 67950, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001961, Sample Num: 31376, Cur Loss: 0.25867116, Cur Avg Loss: 0.54055364, Log Avg loss: 0.51040785, Global Avg Loss: 1.47225483, Time: 0.0165 Steps: 67960, Updated lr: 0.000037 Training, Epoch: 0032, Batch: 001971, Sample Num: 31536, Cur Loss: 0.74203420, Cur Avg Loss: 0.54087120, Log Avg loss: 0.60314384, Global Avg Loss: 1.47212696, Time: 0.0066 Steps: 67970, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001981, Sample Num: 31696, Cur Loss: 0.21146119, Cur Avg Loss: 0.54042125, Log Avg loss: 0.45173672, Global Avg Loss: 1.47197686, Time: 0.0071 Steps: 67980, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 001991, Sample Num: 31856, Cur Loss: 0.86550808, Cur Avg Loss: 0.54042000, Log Avg loss: 0.54017177, Global Avg Loss: 1.47183981, Time: 0.0107 Steps: 67990, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002001, Sample Num: 32016, Cur Loss: 0.32436827, Cur Avg Loss: 0.54051657, Log Avg loss: 0.55974511, Global Avg Loss: 1.47170568, Time: 0.0156 Steps: 68000, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002011, Sample Num: 32176, Cur Loss: 0.79191756, Cur Avg Loss: 0.54008629, Log Avg loss: 0.45398616, Global Avg Loss: 1.47155603, Time: 0.0064 Steps: 68010, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002021, Sample Num: 32336, Cur Loss: 0.39957735, Cur Avg Loss: 0.53994835, Log Avg loss: 0.51220924, Global Avg Loss: 1.47141499, Time: 0.0085 Steps: 68020, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002031, Sample Num: 32496, Cur Loss: 0.54685271, Cur Avg Loss: 0.54020703, Log Avg loss: 0.59248708, Global Avg Loss: 1.47128580, Time: 0.0064 Steps: 68030, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002041, Sample Num: 32656, Cur Loss: 0.56162077, Cur Avg Loss: 0.54019942, Log Avg loss: 0.53865245, Global Avg Loss: 1.47114873, Time: 0.0156 Steps: 68040, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002051, Sample Num: 32816, Cur Loss: 0.53232986, Cur Avg Loss: 0.54058647, Log Avg loss: 0.61958399, Global Avg Loss: 1.47102359, Time: 0.0072 Steps: 68050, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002061, Sample Num: 32976, Cur Loss: 0.55233425, Cur Avg Loss: 0.54058578, Log Avg loss: 0.54044362, Global Avg Loss: 1.47088686, Time: 0.0072 Steps: 68060, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002071, Sample Num: 33136, Cur Loss: 0.60761404, Cur Avg Loss: 0.54062185, Log Avg loss: 0.54805715, Global Avg Loss: 1.47075129, Time: 0.0115 Steps: 68070, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002081, Sample Num: 33296, Cur Loss: 0.20637648, Cur Avg Loss: 0.54006057, Log Avg loss: 0.42381853, Global Avg Loss: 1.47059751, Time: 0.0134 Steps: 68080, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002091, Sample Num: 33456, Cur Loss: 0.41743577, Cur Avg Loss: 0.53993321, Log Avg loss: 0.51343075, Global Avg Loss: 1.47045693, Time: 0.0113 Steps: 68090, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002101, Sample Num: 33616, Cur Loss: 0.79463673, Cur Avg Loss: 0.53953395, Log Avg loss: 0.45604717, Global Avg Loss: 1.47030798, Time: 0.0066 Steps: 68100, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002111, Sample Num: 33776, Cur Loss: 1.16738260, Cur Avg Loss: 0.53964792, Log Avg loss: 0.56359349, Global Avg Loss: 1.47017485, Time: 0.0116 Steps: 68110, Updated lr: 0.000036 Training, Epoch: 0032, Batch: 002121, Sample Num: 33936, Cur Loss: 0.76245439, Cur Avg Loss: 0.53928260, Log Avg loss: 0.46216305, Global Avg Loss: 1.47002687, Time: 0.0118 Steps: 68120, Updated lr: 0.000036 ***** Running evaluation checkpoint-68128 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-68128 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.604951, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.652647, "eval_total_loss": 458.810706, "eval_mae": 0.629673, "eval_mse": 0.652782, "eval_r2": 0.585049, "eval_sp_statistic": 0.711813, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.773148, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.871255, "test_total_loss": 437.369888, "test_mae": 0.696554, "test_mse": 0.871436, "test_r2": 0.437568, "test_sp_statistic": 0.574929, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.662235, "test_ps_pvalue": 0.0, "lr": 3.634139402560455e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4698963432827377, "train_cur_epoch_loss": 1146.6857947409153, "train_cur_epoch_avg_loss": 0.5386030036359396, "train_cur_epoch_time": 21.60495138168335, "train_cur_epoch_avg_time": 0.010147933951002042, "epoch": 32, "step": 68128} ################################################## Training, Epoch: 0033, Batch: 000002, Sample Num: 32, Cur Loss: 0.43931401, Cur Avg Loss: 0.35936905, Log Avg loss: 0.35861441, Global Avg Loss: 1.46986374, Time: 0.0105 Steps: 68130, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000012, Sample Num: 192, Cur Loss: 0.35418892, Cur Avg Loss: 0.57623300, Log Avg loss: 0.61960579, Global Avg Loss: 1.46973896, Time: 0.0148 Steps: 68140, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000022, Sample Num: 352, Cur Loss: 0.55921781, Cur Avg Loss: 0.54924798, Log Avg loss: 0.51686594, Global Avg Loss: 1.46959914, Time: 0.0066 Steps: 68150, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000032, Sample Num: 512, Cur Loss: 0.58609360, Cur Avg Loss: 0.61662223, Log Avg loss: 0.76484559, Global Avg Loss: 1.46949575, Time: 0.0155 Steps: 68160, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000042, Sample Num: 672, Cur Loss: 0.48111188, Cur Avg Loss: 0.61006387, Log Avg loss: 0.58907713, Global Avg Loss: 1.46936659, Time: 0.0072 Steps: 68170, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000052, Sample Num: 832, Cur Loss: 0.60414988, Cur Avg Loss: 0.62219669, Log Avg loss: 0.67315450, Global Avg Loss: 1.46924981, Time: 0.0068 Steps: 68180, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000062, Sample Num: 992, Cur Loss: 0.24133566, Cur Avg Loss: 0.61347469, Log Avg loss: 0.56812031, Global Avg Loss: 1.46911766, Time: 0.0234 Steps: 68190, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000072, Sample Num: 1152, Cur Loss: 0.28916547, Cur Avg Loss: 0.61725384, Log Avg loss: 0.64068457, Global Avg Loss: 1.46899619, Time: 0.0068 Steps: 68200, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000082, Sample Num: 1312, Cur Loss: 0.11815697, Cur Avg Loss: 0.60218093, Log Avg loss: 0.49365594, Global Avg Loss: 1.46885320, Time: 0.0076 Steps: 68210, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000092, Sample Num: 1472, Cur Loss: 1.55165100, Cur Avg Loss: 0.59261748, Log Avg loss: 0.51419719, Global Avg Loss: 1.46871326, Time: 0.0067 Steps: 68220, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000102, Sample Num: 1632, Cur Loss: 0.25876468, Cur Avg Loss: 0.56558930, Log Avg loss: 0.31693010, Global Avg Loss: 1.46854446, Time: 0.0069 Steps: 68230, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000112, Sample Num: 1792, Cur Loss: 0.42273265, Cur Avg Loss: 0.55270746, Log Avg loss: 0.42131265, Global Avg Loss: 1.46839099, Time: 0.0074 Steps: 68240, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000122, Sample Num: 1952, Cur Loss: 0.63878912, Cur Avg Loss: 0.55191036, Log Avg loss: 0.54298286, Global Avg Loss: 1.46825540, Time: 0.0067 Steps: 68250, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000132, Sample Num: 2112, Cur Loss: 0.31674302, Cur Avg Loss: 0.55024394, Log Avg loss: 0.52991358, Global Avg Loss: 1.46811794, Time: 0.0067 Steps: 68260, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000142, Sample Num: 2272, Cur Loss: 0.45509881, Cur Avg Loss: 0.54308471, Log Avg loss: 0.44858296, Global Avg Loss: 1.46796860, Time: 0.0065 Steps: 68270, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000152, Sample Num: 2432, Cur Loss: 0.30941278, Cur Avg Loss: 0.54288771, Log Avg loss: 0.54009021, Global Avg Loss: 1.46783270, Time: 0.0065 Steps: 68280, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000162, Sample Num: 2592, Cur Loss: 0.42911929, Cur Avg Loss: 0.54141368, Log Avg loss: 0.51900842, Global Avg Loss: 1.46769376, Time: 0.0066 Steps: 68290, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000172, Sample Num: 2752, Cur Loss: 0.75509638, Cur Avg Loss: 0.53973065, Log Avg loss: 0.51246563, Global Avg Loss: 1.46755391, Time: 0.0068 Steps: 68300, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000182, Sample Num: 2912, Cur Loss: 0.73785639, Cur Avg Loss: 0.53258144, Log Avg loss: 0.40961508, Global Avg Loss: 1.46739903, Time: 0.0068 Steps: 68310, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000192, Sample Num: 3072, Cur Loss: 0.93722814, Cur Avg Loss: 0.53160455, Log Avg loss: 0.51382506, Global Avg Loss: 1.46725946, Time: 0.0151 Steps: 68320, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000202, Sample Num: 3232, Cur Loss: 0.79660791, Cur Avg Loss: 0.53216513, Log Avg loss: 0.54292828, Global Avg Loss: 1.46712418, Time: 0.0067 Steps: 68330, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000212, Sample Num: 3392, Cur Loss: 0.24435964, Cur Avg Loss: 0.52948613, Log Avg loss: 0.47537029, Global Avg Loss: 1.46697906, Time: 0.0068 Steps: 68340, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000222, Sample Num: 3552, Cur Loss: 0.77630377, Cur Avg Loss: 0.52663192, Log Avg loss: 0.46612262, Global Avg Loss: 1.46683263, Time: 0.0074 Steps: 68350, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000232, Sample Num: 3712, Cur Loss: 0.37822187, Cur Avg Loss: 0.52560400, Log Avg loss: 0.50278428, Global Avg Loss: 1.46669161, Time: 0.0072 Steps: 68360, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000242, Sample Num: 3872, Cur Loss: 0.19334969, Cur Avg Loss: 0.52177234, Log Avg loss: 0.43287785, Global Avg Loss: 1.46654040, Time: 0.0129 Steps: 68370, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000252, Sample Num: 4032, Cur Loss: 0.19812825, Cur Avg Loss: 0.51806683, Log Avg loss: 0.42839345, Global Avg Loss: 1.46638858, Time: 0.0113 Steps: 68380, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000262, Sample Num: 4192, Cur Loss: 0.34511018, Cur Avg Loss: 0.51508171, Log Avg loss: 0.43985683, Global Avg Loss: 1.46623848, Time: 0.0071 Steps: 68390, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000272, Sample Num: 4352, Cur Loss: 0.46808001, Cur Avg Loss: 0.51673104, Log Avg loss: 0.55994343, Global Avg Loss: 1.46610598, Time: 0.0132 Steps: 68400, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000282, Sample Num: 4512, Cur Loss: 0.50832677, Cur Avg Loss: 0.51730217, Log Avg loss: 0.53283670, Global Avg Loss: 1.46596956, Time: 0.0113 Steps: 68410, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000292, Sample Num: 4672, Cur Loss: 1.65202117, Cur Avg Loss: 0.51784318, Log Avg loss: 0.53309965, Global Avg Loss: 1.46583321, Time: 0.0066 Steps: 68420, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000302, Sample Num: 4832, Cur Loss: 0.20760465, Cur Avg Loss: 0.51987463, Log Avg loss: 0.57919325, Global Avg Loss: 1.46570364, Time: 0.0138 Steps: 68430, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000312, Sample Num: 4992, Cur Loss: 1.23691714, Cur Avg Loss: 0.52104141, Log Avg loss: 0.55627809, Global Avg Loss: 1.46557076, Time: 0.0124 Steps: 68440, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000322, Sample Num: 5152, Cur Loss: 0.53315896, Cur Avg Loss: 0.51510049, Log Avg loss: 0.32974382, Global Avg Loss: 1.46540483, Time: 0.0112 Steps: 68450, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000332, Sample Num: 5312, Cur Loss: 1.32611775, Cur Avg Loss: 0.51746734, Log Avg loss: 0.59367975, Global Avg Loss: 1.46527749, Time: 0.0065 Steps: 68460, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000342, Sample Num: 5472, Cur Loss: 0.08915921, Cur Avg Loss: 0.51779611, Log Avg loss: 0.52871122, Global Avg Loss: 1.46514071, Time: 0.0066 Steps: 68470, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000352, Sample Num: 5632, Cur Loss: 0.46968979, Cur Avg Loss: 0.52005340, Log Avg loss: 0.59725275, Global Avg Loss: 1.46501397, Time: 0.0067 Steps: 68480, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000362, Sample Num: 5792, Cur Loss: 0.39486694, Cur Avg Loss: 0.52198995, Log Avg loss: 0.59015676, Global Avg Loss: 1.46488624, Time: 0.0067 Steps: 68490, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000372, Sample Num: 5952, Cur Loss: 0.32034087, Cur Avg Loss: 0.52279279, Log Avg loss: 0.55185561, Global Avg Loss: 1.46475295, Time: 0.0111 Steps: 68500, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000382, Sample Num: 6112, Cur Loss: 0.79879671, Cur Avg Loss: 0.52855177, Log Avg loss: 0.74278552, Global Avg Loss: 1.46464757, Time: 0.0084 Steps: 68510, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000392, Sample Num: 6272, Cur Loss: 0.36650860, Cur Avg Loss: 0.52783140, Log Avg loss: 0.50031332, Global Avg Loss: 1.46450683, Time: 0.0116 Steps: 68520, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000402, Sample Num: 6432, Cur Loss: 1.02730989, Cur Avg Loss: 0.52828998, Log Avg loss: 0.54626654, Global Avg Loss: 1.46437284, Time: 0.0068 Steps: 68530, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000412, Sample Num: 6592, Cur Loss: 0.19169550, Cur Avg Loss: 0.52552913, Log Avg loss: 0.41454299, Global Avg Loss: 1.46421967, Time: 0.0118 Steps: 68540, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000422, Sample Num: 6752, Cur Loss: 0.59878480, Cur Avg Loss: 0.52085985, Log Avg loss: 0.32848534, Global Avg Loss: 1.46405399, Time: 0.0121 Steps: 68550, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000432, Sample Num: 6912, Cur Loss: 1.31212068, Cur Avg Loss: 0.52551323, Log Avg loss: 0.72188594, Global Avg Loss: 1.46394574, Time: 0.0068 Steps: 68560, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000442, Sample Num: 7072, Cur Loss: 0.39059705, Cur Avg Loss: 0.52383344, Log Avg loss: 0.45126650, Global Avg Loss: 1.46379805, Time: 0.0119 Steps: 68570, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000452, Sample Num: 7232, Cur Loss: 0.49693796, Cur Avg Loss: 0.52201040, Log Avg loss: 0.44143216, Global Avg Loss: 1.46364898, Time: 0.0066 Steps: 68580, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000462, Sample Num: 7392, Cur Loss: 0.40970451, Cur Avg Loss: 0.52020717, Log Avg loss: 0.43870091, Global Avg Loss: 1.46349954, Time: 0.0068 Steps: 68590, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000472, Sample Num: 7552, Cur Loss: 0.62005252, Cur Avg Loss: 0.51757295, Log Avg loss: 0.39587209, Global Avg Loss: 1.46334391, Time: 0.0238 Steps: 68600, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000482, Sample Num: 7712, Cur Loss: 1.04602444, Cur Avg Loss: 0.52158049, Log Avg loss: 0.71073645, Global Avg Loss: 1.46323422, Time: 0.0123 Steps: 68610, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000492, Sample Num: 7872, Cur Loss: 0.35475075, Cur Avg Loss: 0.52171769, Log Avg loss: 0.52833066, Global Avg Loss: 1.46309798, Time: 0.0068 Steps: 68620, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000502, Sample Num: 8032, Cur Loss: 0.69497418, Cur Avg Loss: 0.52165443, Log Avg loss: 0.51854215, Global Avg Loss: 1.46296035, Time: 0.0111 Steps: 68630, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000512, Sample Num: 8192, Cur Loss: 0.37712735, Cur Avg Loss: 0.52093870, Log Avg loss: 0.48500910, Global Avg Loss: 1.46281787, Time: 0.0082 Steps: 68640, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000522, Sample Num: 8352, Cur Loss: 0.49988425, Cur Avg Loss: 0.52201308, Log Avg loss: 0.57702106, Global Avg Loss: 1.46268884, Time: 0.0069 Steps: 68650, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000532, Sample Num: 8512, Cur Loss: 0.65705001, Cur Avg Loss: 0.52345670, Log Avg loss: 0.59881348, Global Avg Loss: 1.46256302, Time: 0.0199 Steps: 68660, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000542, Sample Num: 8672, Cur Loss: 0.51578116, Cur Avg Loss: 0.52225327, Log Avg loss: 0.45823087, Global Avg Loss: 1.46241677, Time: 0.0067 Steps: 68670, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000552, Sample Num: 8832, Cur Loss: 0.43401361, Cur Avg Loss: 0.52382813, Log Avg loss: 0.60918582, Global Avg Loss: 1.46229253, Time: 0.0067 Steps: 68680, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000562, Sample Num: 8992, Cur Loss: 1.05388272, Cur Avg Loss: 0.52276285, Log Avg loss: 0.46395957, Global Avg Loss: 1.46214719, Time: 0.0169 Steps: 68690, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000572, Sample Num: 9152, Cur Loss: 0.23637718, Cur Avg Loss: 0.52375892, Log Avg loss: 0.57973802, Global Avg Loss: 1.46201875, Time: 0.0065 Steps: 68700, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000582, Sample Num: 9312, Cur Loss: 0.62767029, Cur Avg Loss: 0.52430607, Log Avg loss: 0.55560264, Global Avg Loss: 1.46188683, Time: 0.0075 Steps: 68710, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000592, Sample Num: 9472, Cur Loss: 0.49566829, Cur Avg Loss: 0.52330432, Log Avg loss: 0.46500267, Global Avg Loss: 1.46174177, Time: 0.0067 Steps: 68720, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000602, Sample Num: 9632, Cur Loss: 0.13606238, Cur Avg Loss: 0.52212448, Log Avg loss: 0.45227778, Global Avg Loss: 1.46159489, Time: 0.0066 Steps: 68730, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000612, Sample Num: 9792, Cur Loss: 0.30248427, Cur Avg Loss: 0.52191742, Log Avg loss: 0.50945284, Global Avg Loss: 1.46145638, Time: 0.0096 Steps: 68740, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000622, Sample Num: 9952, Cur Loss: 0.49300775, Cur Avg Loss: 0.52116209, Log Avg loss: 0.47493555, Global Avg Loss: 1.46131289, Time: 0.0067 Steps: 68750, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000632, Sample Num: 10112, Cur Loss: 0.46799541, Cur Avg Loss: 0.52045427, Log Avg loss: 0.47642790, Global Avg Loss: 1.46116965, Time: 0.0065 Steps: 68760, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000642, Sample Num: 10272, Cur Loss: 0.27349150, Cur Avg Loss: 0.51834629, Log Avg loss: 0.38512174, Global Avg Loss: 1.46101318, Time: 0.0078 Steps: 68770, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000652, Sample Num: 10432, Cur Loss: 0.44901934, Cur Avg Loss: 0.52043850, Log Avg loss: 0.65475843, Global Avg Loss: 1.46089596, Time: 0.0068 Steps: 68780, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000662, Sample Num: 10592, Cur Loss: 0.23530403, Cur Avg Loss: 0.52049400, Log Avg loss: 0.52411286, Global Avg Loss: 1.46075978, Time: 0.0119 Steps: 68790, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000672, Sample Num: 10752, Cur Loss: 0.59902304, Cur Avg Loss: 0.52018835, Log Avg loss: 0.49995455, Global Avg Loss: 1.46062013, Time: 0.0112 Steps: 68800, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000682, Sample Num: 10912, Cur Loss: 0.43437284, Cur Avg Loss: 0.52127341, Log Avg loss: 0.59418894, Global Avg Loss: 1.46049421, Time: 0.0097 Steps: 68810, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000692, Sample Num: 11072, Cur Loss: 0.49825424, Cur Avg Loss: 0.52169872, Log Avg loss: 0.55070516, Global Avg Loss: 1.46036201, Time: 0.0107 Steps: 68820, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000702, Sample Num: 11232, Cur Loss: 0.47730505, Cur Avg Loss: 0.52236224, Log Avg loss: 0.56827810, Global Avg Loss: 1.46023240, Time: 0.0117 Steps: 68830, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000712, Sample Num: 11392, Cur Loss: 0.56208992, Cur Avg Loss: 0.52245843, Log Avg loss: 0.52921051, Global Avg Loss: 1.46009716, Time: 0.0109 Steps: 68840, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000722, Sample Num: 11552, Cur Loss: 0.30555862, Cur Avg Loss: 0.52063171, Log Avg loss: 0.39056928, Global Avg Loss: 1.45994182, Time: 0.0120 Steps: 68850, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000732, Sample Num: 11712, Cur Loss: 0.38830194, Cur Avg Loss: 0.52080874, Log Avg loss: 0.53359007, Global Avg Loss: 1.45980729, Time: 0.0111 Steps: 68860, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000742, Sample Num: 11872, Cur Loss: 0.39568126, Cur Avg Loss: 0.51870798, Log Avg loss: 0.36493275, Global Avg Loss: 1.45964831, Time: 0.0066 Steps: 68870, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000752, Sample Num: 12032, Cur Loss: 0.74239349, Cur Avg Loss: 0.51833689, Log Avg loss: 0.49080181, Global Avg Loss: 1.45950766, Time: 0.0113 Steps: 68880, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000762, Sample Num: 12192, Cur Loss: 0.53748727, Cur Avg Loss: 0.51856921, Log Avg loss: 0.53603943, Global Avg Loss: 1.45937361, Time: 0.0150 Steps: 68890, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000772, Sample Num: 12352, Cur Loss: 0.50772643, Cur Avg Loss: 0.51849891, Log Avg loss: 0.51314237, Global Avg Loss: 1.45923627, Time: 0.0072 Steps: 68900, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000782, Sample Num: 12512, Cur Loss: 0.13098359, Cur Avg Loss: 0.51865326, Log Avg loss: 0.53056885, Global Avg Loss: 1.45910151, Time: 0.0139 Steps: 68910, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000792, Sample Num: 12672, Cur Loss: 0.91096866, Cur Avg Loss: 0.52111401, Log Avg loss: 0.71354477, Global Avg Loss: 1.45899333, Time: 0.0118 Steps: 68920, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000802, Sample Num: 12832, Cur Loss: 0.69163024, Cur Avg Loss: 0.52300194, Log Avg loss: 0.67252580, Global Avg Loss: 1.45887923, Time: 0.0108 Steps: 68930, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000812, Sample Num: 12992, Cur Loss: 0.65690881, Cur Avg Loss: 0.52276792, Log Avg loss: 0.50399997, Global Avg Loss: 1.45874073, Time: 0.0068 Steps: 68940, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000822, Sample Num: 13152, Cur Loss: 0.62039471, Cur Avg Loss: 0.52187669, Log Avg loss: 0.44950873, Global Avg Loss: 1.45859435, Time: 0.0072 Steps: 68950, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000832, Sample Num: 13312, Cur Loss: 0.43463540, Cur Avg Loss: 0.52276269, Log Avg loss: 0.59559203, Global Avg Loss: 1.45846921, Time: 0.0119 Steps: 68960, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000842, Sample Num: 13472, Cur Loss: 0.59607446, Cur Avg Loss: 0.52321985, Log Avg loss: 0.56125572, Global Avg Loss: 1.45833912, Time: 0.0113 Steps: 68970, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000852, Sample Num: 13632, Cur Loss: 0.32067448, Cur Avg Loss: 0.52279195, Log Avg loss: 0.48676217, Global Avg Loss: 1.45819827, Time: 0.0095 Steps: 68980, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000862, Sample Num: 13792, Cur Loss: 0.17025894, Cur Avg Loss: 0.52227396, Log Avg loss: 0.47814115, Global Avg Loss: 1.45805621, Time: 0.0106 Steps: 68990, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000872, Sample Num: 13952, Cur Loss: 0.34296876, Cur Avg Loss: 0.52175314, Log Avg loss: 0.47685863, Global Avg Loss: 1.45791401, Time: 0.0085 Steps: 69000, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000882, Sample Num: 14112, Cur Loss: 0.36436790, Cur Avg Loss: 0.52489385, Log Avg loss: 0.79876363, Global Avg Loss: 1.45781850, Time: 0.0063 Steps: 69010, Updated lr: 0.000036 Training, Epoch: 0033, Batch: 000892, Sample Num: 14272, Cur Loss: 0.62187701, Cur Avg Loss: 0.52570748, Log Avg loss: 0.59746978, Global Avg Loss: 1.45769384, Time: 0.0068 Steps: 69020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000902, Sample Num: 14432, Cur Loss: 1.07283688, Cur Avg Loss: 0.52550347, Log Avg loss: 0.50730590, Global Avg Loss: 1.45755617, Time: 0.0066 Steps: 69030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000912, Sample Num: 14592, Cur Loss: 0.09297961, Cur Avg Loss: 0.52578295, Log Avg loss: 0.55099203, Global Avg Loss: 1.45742486, Time: 0.0065 Steps: 69040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000922, Sample Num: 14752, Cur Loss: 0.84016740, Cur Avg Loss: 0.52661420, Log Avg loss: 0.60242459, Global Avg Loss: 1.45730103, Time: 0.0231 Steps: 69050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000932, Sample Num: 14912, Cur Loss: 0.37893981, Cur Avg Loss: 0.52523752, Log Avg loss: 0.39830733, Global Avg Loss: 1.45714769, Time: 0.0068 Steps: 69060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000942, Sample Num: 15072, Cur Loss: 0.70555282, Cur Avg Loss: 0.52563387, Log Avg loss: 0.56257342, Global Avg Loss: 1.45701817, Time: 0.0070 Steps: 69070, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000952, Sample Num: 15232, Cur Loss: 0.50547218, Cur Avg Loss: 0.52601272, Log Avg loss: 0.56170051, Global Avg Loss: 1.45688857, Time: 0.0107 Steps: 69080, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000962, Sample Num: 15392, Cur Loss: 0.47570309, Cur Avg Loss: 0.52681276, Log Avg loss: 0.60297650, Global Avg Loss: 1.45676497, Time: 0.0110 Steps: 69090, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000972, Sample Num: 15552, Cur Loss: 0.23661336, Cur Avg Loss: 0.52669909, Log Avg loss: 0.51576440, Global Avg Loss: 1.45662879, Time: 0.0120 Steps: 69100, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000982, Sample Num: 15712, Cur Loss: 0.94162333, Cur Avg Loss: 0.52817028, Log Avg loss: 0.67116984, Global Avg Loss: 1.45651514, Time: 0.0066 Steps: 69110, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 000992, Sample Num: 15872, Cur Loss: 0.78708571, Cur Avg Loss: 0.52893379, Log Avg loss: 0.60391009, Global Avg Loss: 1.45639179, Time: 0.0110 Steps: 69120, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001002, Sample Num: 16032, Cur Loss: 0.40412241, Cur Avg Loss: 0.52931601, Log Avg loss: 0.56723282, Global Avg Loss: 1.45626317, Time: 0.0072 Steps: 69130, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001012, Sample Num: 16192, Cur Loss: 0.75238633, Cur Avg Loss: 0.52918137, Log Avg loss: 0.51568988, Global Avg Loss: 1.45612713, Time: 0.0108 Steps: 69140, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001022, Sample Num: 16352, Cur Loss: 0.18184817, Cur Avg Loss: 0.52699815, Log Avg loss: 0.30605669, Global Avg Loss: 1.45596081, Time: 0.0119 Steps: 69150, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001032, Sample Num: 16512, Cur Loss: 0.62846541, Cur Avg Loss: 0.52735790, Log Avg loss: 0.56412421, Global Avg Loss: 1.45583186, Time: 0.0074 Steps: 69160, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001042, Sample Num: 16672, Cur Loss: 0.41681337, Cur Avg Loss: 0.52767515, Log Avg loss: 0.56041487, Global Avg Loss: 1.45570241, Time: 0.0070 Steps: 69170, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001052, Sample Num: 16832, Cur Loss: 0.95123464, Cur Avg Loss: 0.52820608, Log Avg loss: 0.58352908, Global Avg Loss: 1.45557634, Time: 0.0067 Steps: 69180, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001062, Sample Num: 16992, Cur Loss: 0.34214938, Cur Avg Loss: 0.52726307, Log Avg loss: 0.42805929, Global Avg Loss: 1.45542783, Time: 0.0066 Steps: 69190, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001072, Sample Num: 17152, Cur Loss: 0.14001240, Cur Avg Loss: 0.52618066, Log Avg loss: 0.41122782, Global Avg Loss: 1.45527693, Time: 0.0066 Steps: 69200, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001082, Sample Num: 17312, Cur Loss: 0.81448758, Cur Avg Loss: 0.52788180, Log Avg loss: 0.71024443, Global Avg Loss: 1.45516928, Time: 0.0072 Steps: 69210, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001092, Sample Num: 17472, Cur Loss: 0.34275782, Cur Avg Loss: 0.52639976, Log Avg loss: 0.36604352, Global Avg Loss: 1.45501194, Time: 0.0066 Steps: 69220, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001102, Sample Num: 17632, Cur Loss: 0.28691089, Cur Avg Loss: 0.52565029, Log Avg loss: 0.44380777, Global Avg Loss: 1.45486588, Time: 0.0154 Steps: 69230, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001112, Sample Num: 17792, Cur Loss: 2.43143296, Cur Avg Loss: 0.52627063, Log Avg loss: 0.59463163, Global Avg Loss: 1.45474164, Time: 0.0067 Steps: 69240, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001122, Sample Num: 17952, Cur Loss: 0.39433861, Cur Avg Loss: 0.52577465, Log Avg loss: 0.47062212, Global Avg Loss: 1.45459953, Time: 0.0122 Steps: 69250, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001132, Sample Num: 18112, Cur Loss: 0.51101208, Cur Avg Loss: 0.52677672, Log Avg loss: 0.63920885, Global Avg Loss: 1.45448180, Time: 0.0157 Steps: 69260, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001142, Sample Num: 18272, Cur Loss: 0.39645776, Cur Avg Loss: 0.52581909, Log Avg loss: 0.41741493, Global Avg Loss: 1.45433208, Time: 0.0068 Steps: 69270, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001152, Sample Num: 18432, Cur Loss: 0.56012475, Cur Avg Loss: 0.52644738, Log Avg loss: 0.59819846, Global Avg Loss: 1.45420851, Time: 0.0074 Steps: 69280, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001162, Sample Num: 18592, Cur Loss: 0.23965976, Cur Avg Loss: 0.52581125, Log Avg loss: 0.45252933, Global Avg Loss: 1.45406394, Time: 0.0119 Steps: 69290, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001172, Sample Num: 18752, Cur Loss: 0.44708872, Cur Avg Loss: 0.52603694, Log Avg loss: 0.55226224, Global Avg Loss: 1.45393381, Time: 0.0114 Steps: 69300, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001182, Sample Num: 18912, Cur Loss: 0.47778577, Cur Avg Loss: 0.52571456, Log Avg loss: 0.48793192, Global Avg Loss: 1.45379444, Time: 0.0111 Steps: 69310, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001192, Sample Num: 19072, Cur Loss: 0.59550601, Cur Avg Loss: 0.52677601, Log Avg loss: 0.65223910, Global Avg Loss: 1.45367881, Time: 0.0116 Steps: 69320, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001202, Sample Num: 19232, Cur Loss: 0.56293190, Cur Avg Loss: 0.52797583, Log Avg loss: 0.67099356, Global Avg Loss: 1.45356592, Time: 0.0114 Steps: 69330, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001212, Sample Num: 19392, Cur Loss: 1.27462852, Cur Avg Loss: 0.52923253, Log Avg loss: 0.68028859, Global Avg Loss: 1.45345440, Time: 0.0073 Steps: 69340, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001222, Sample Num: 19552, Cur Loss: 0.71358168, Cur Avg Loss: 0.52932393, Log Avg loss: 0.54040145, Global Avg Loss: 1.45332274, Time: 0.0084 Steps: 69350, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001232, Sample Num: 19712, Cur Loss: 0.41885293, Cur Avg Loss: 0.52919937, Log Avg loss: 0.51397808, Global Avg Loss: 1.45318731, Time: 0.0067 Steps: 69360, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001242, Sample Num: 19872, Cur Loss: 0.40627235, Cur Avg Loss: 0.52916212, Log Avg loss: 0.52457264, Global Avg Loss: 1.45305344, Time: 0.0112 Steps: 69370, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001252, Sample Num: 20032, Cur Loss: 0.43723643, Cur Avg Loss: 0.52905471, Log Avg loss: 0.51571511, Global Avg Loss: 1.45291834, Time: 0.0122 Steps: 69380, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001262, Sample Num: 20192, Cur Loss: 0.90772283, Cur Avg Loss: 0.52966350, Log Avg loss: 0.60588375, Global Avg Loss: 1.45279627, Time: 0.0114 Steps: 69390, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001272, Sample Num: 20352, Cur Loss: 0.79544628, Cur Avg Loss: 0.52957320, Log Avg loss: 0.51817688, Global Avg Loss: 1.45266160, Time: 0.0067 Steps: 69400, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001282, Sample Num: 20512, Cur Loss: 0.51281750, Cur Avg Loss: 0.52990464, Log Avg loss: 0.57206427, Global Avg Loss: 1.45253473, Time: 0.0113 Steps: 69410, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001292, Sample Num: 20672, Cur Loss: 0.13150176, Cur Avg Loss: 0.53000541, Log Avg loss: 0.54292403, Global Avg Loss: 1.45240370, Time: 0.0067 Steps: 69420, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001302, Sample Num: 20832, Cur Loss: 0.29506290, Cur Avg Loss: 0.53041608, Log Avg loss: 0.58347524, Global Avg Loss: 1.45227855, Time: 0.0066 Steps: 69430, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001312, Sample Num: 20992, Cur Loss: 0.35005689, Cur Avg Loss: 0.53014713, Log Avg loss: 0.49512924, Global Avg Loss: 1.45214071, Time: 0.0089 Steps: 69440, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001322, Sample Num: 21152, Cur Loss: 0.62591422, Cur Avg Loss: 0.53003471, Log Avg loss: 0.51528499, Global Avg Loss: 1.45200582, Time: 0.0088 Steps: 69450, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001332, Sample Num: 21312, Cur Loss: 1.04233956, Cur Avg Loss: 0.53196780, Log Avg loss: 0.78752295, Global Avg Loss: 1.45191015, Time: 0.0110 Steps: 69460, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001342, Sample Num: 21472, Cur Loss: 0.39914113, Cur Avg Loss: 0.53151803, Log Avg loss: 0.47160758, Global Avg Loss: 1.45176904, Time: 0.0119 Steps: 69470, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001352, Sample Num: 21632, Cur Loss: 0.69972038, Cur Avg Loss: 0.53231297, Log Avg loss: 0.63899412, Global Avg Loss: 1.45165206, Time: 0.0066 Steps: 69480, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001362, Sample Num: 21792, Cur Loss: 0.69688815, Cur Avg Loss: 0.53387629, Log Avg loss: 0.74523742, Global Avg Loss: 1.45155040, Time: 0.0111 Steps: 69490, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001372, Sample Num: 21952, Cur Loss: 0.73671877, Cur Avg Loss: 0.53488808, Log Avg loss: 0.67269442, Global Avg Loss: 1.45143834, Time: 0.0073 Steps: 69500, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001382, Sample Num: 22112, Cur Loss: 0.30480427, Cur Avg Loss: 0.53360430, Log Avg loss: 0.35746970, Global Avg Loss: 1.45128096, Time: 0.0072 Steps: 69510, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001392, Sample Num: 22272, Cur Loss: 0.26831329, Cur Avg Loss: 0.53426846, Log Avg loss: 0.62605527, Global Avg Loss: 1.45116225, Time: 0.0066 Steps: 69520, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001402, Sample Num: 22432, Cur Loss: 0.53384161, Cur Avg Loss: 0.53455829, Log Avg loss: 0.57490284, Global Avg Loss: 1.45103623, Time: 0.0111 Steps: 69530, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001412, Sample Num: 22592, Cur Loss: 0.40390599, Cur Avg Loss: 0.53568548, Log Avg loss: 0.69371722, Global Avg Loss: 1.45092732, Time: 0.0066 Steps: 69540, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001422, Sample Num: 22752, Cur Loss: 0.19093233, Cur Avg Loss: 0.53522118, Log Avg loss: 0.46966252, Global Avg Loss: 1.45078623, Time: 0.0083 Steps: 69550, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001432, Sample Num: 22912, Cur Loss: 0.73562253, Cur Avg Loss: 0.53504059, Log Avg loss: 0.50936023, Global Avg Loss: 1.45065089, Time: 0.0088 Steps: 69560, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001442, Sample Num: 23072, Cur Loss: 0.51494789, Cur Avg Loss: 0.53562326, Log Avg loss: 0.61906204, Global Avg Loss: 1.45053136, Time: 0.0115 Steps: 69570, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001452, Sample Num: 23232, Cur Loss: 0.18091679, Cur Avg Loss: 0.53526926, Log Avg loss: 0.48422158, Global Avg Loss: 1.45039248, Time: 0.0110 Steps: 69580, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001462, Sample Num: 23392, Cur Loss: 0.74551427, Cur Avg Loss: 0.53429247, Log Avg loss: 0.39246249, Global Avg Loss: 1.45024046, Time: 0.0064 Steps: 69590, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001472, Sample Num: 23552, Cur Loss: 0.82389998, Cur Avg Loss: 0.53406745, Log Avg loss: 0.50116958, Global Avg Loss: 1.45010410, Time: 0.0066 Steps: 69600, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001482, Sample Num: 23712, Cur Loss: 0.29426238, Cur Avg Loss: 0.53338256, Log Avg loss: 0.43256724, Global Avg Loss: 1.44995792, Time: 0.0111 Steps: 69610, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001492, Sample Num: 23872, Cur Loss: 0.37774599, Cur Avg Loss: 0.53322964, Log Avg loss: 0.51056720, Global Avg Loss: 1.44982299, Time: 0.0117 Steps: 69620, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001502, Sample Num: 24032, Cur Loss: 0.28274262, Cur Avg Loss: 0.53276973, Log Avg loss: 0.46415125, Global Avg Loss: 1.44968143, Time: 0.0122 Steps: 69630, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001512, Sample Num: 24192, Cur Loss: 0.39879864, Cur Avg Loss: 0.53293494, Log Avg loss: 0.55774877, Global Avg Loss: 1.44955336, Time: 0.0110 Steps: 69640, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001522, Sample Num: 24352, Cur Loss: 0.12285736, Cur Avg Loss: 0.53220086, Log Avg loss: 0.42120830, Global Avg Loss: 1.44940571, Time: 0.0113 Steps: 69650, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001532, Sample Num: 24512, Cur Loss: 1.53909385, Cur Avg Loss: 0.53238532, Log Avg loss: 0.56045991, Global Avg Loss: 1.44927810, Time: 0.0117 Steps: 69660, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001542, Sample Num: 24672, Cur Loss: 0.58899975, Cur Avg Loss: 0.53305109, Log Avg loss: 0.63504753, Global Avg Loss: 1.44916123, Time: 0.0141 Steps: 69670, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001552, Sample Num: 24832, Cur Loss: 1.08197379, Cur Avg Loss: 0.53400510, Log Avg loss: 0.68111294, Global Avg Loss: 1.44905100, Time: 0.0104 Steps: 69680, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001562, Sample Num: 24992, Cur Loss: 0.89420748, Cur Avg Loss: 0.53385907, Log Avg loss: 0.51119585, Global Avg Loss: 1.44891643, Time: 0.0070 Steps: 69690, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001572, Sample Num: 25152, Cur Loss: 0.89883983, Cur Avg Loss: 0.53392959, Log Avg loss: 0.54494445, Global Avg Loss: 1.44878673, Time: 0.0108 Steps: 69700, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001582, Sample Num: 25312, Cur Loss: 0.27218235, Cur Avg Loss: 0.53396219, Log Avg loss: 0.53908732, Global Avg Loss: 1.44865624, Time: 0.0065 Steps: 69710, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001592, Sample Num: 25472, Cur Loss: 0.80037081, Cur Avg Loss: 0.53382840, Log Avg loss: 0.51266232, Global Avg Loss: 1.44852199, Time: 0.0106 Steps: 69720, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001602, Sample Num: 25632, Cur Loss: 0.28941113, Cur Avg Loss: 0.53354056, Log Avg loss: 0.48771667, Global Avg Loss: 1.44838420, Time: 0.0066 Steps: 69730, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001612, Sample Num: 25792, Cur Loss: 0.39589205, Cur Avg Loss: 0.53325846, Log Avg loss: 0.48806612, Global Avg Loss: 1.44824650, Time: 0.0111 Steps: 69740, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001622, Sample Num: 25952, Cur Loss: 0.71217591, Cur Avg Loss: 0.53318515, Log Avg loss: 0.52136742, Global Avg Loss: 1.44811361, Time: 0.0224 Steps: 69750, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001632, Sample Num: 26112, Cur Loss: 0.81733143, Cur Avg Loss: 0.53295816, Log Avg loss: 0.49614004, Global Avg Loss: 1.44797715, Time: 0.0135 Steps: 69760, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001642, Sample Num: 26272, Cur Loss: 0.23657796, Cur Avg Loss: 0.53235410, Log Avg loss: 0.43377201, Global Avg Loss: 1.44783178, Time: 0.0066 Steps: 69770, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001652, Sample Num: 26432, Cur Loss: 0.50987554, Cur Avg Loss: 0.53166720, Log Avg loss: 0.41887708, Global Avg Loss: 1.44768433, Time: 0.0119 Steps: 69780, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001662, Sample Num: 26592, Cur Loss: 0.75205964, Cur Avg Loss: 0.53193989, Log Avg loss: 0.57698965, Global Avg Loss: 1.44755957, Time: 0.0093 Steps: 69790, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001672, Sample Num: 26752, Cur Loss: 0.65054768, Cur Avg Loss: 0.53254580, Log Avg loss: 0.63324745, Global Avg Loss: 1.44744290, Time: 0.0070 Steps: 69800, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001682, Sample Num: 26912, Cur Loss: 0.52613747, Cur Avg Loss: 0.53319205, Log Avg loss: 0.64124540, Global Avg Loss: 1.44732742, Time: 0.0112 Steps: 69810, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001692, Sample Num: 27072, Cur Loss: 0.61109853, Cur Avg Loss: 0.53353094, Log Avg loss: 0.59053135, Global Avg Loss: 1.44720470, Time: 0.0159 Steps: 69820, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001702, Sample Num: 27232, Cur Loss: 0.99979264, Cur Avg Loss: 0.53427022, Log Avg loss: 0.65935710, Global Avg Loss: 1.44709188, Time: 0.0117 Steps: 69830, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001712, Sample Num: 27392, Cur Loss: 0.83348149, Cur Avg Loss: 0.53458722, Log Avg loss: 0.58854104, Global Avg Loss: 1.44696895, Time: 0.0218 Steps: 69840, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001722, Sample Num: 27552, Cur Loss: 0.71835017, Cur Avg Loss: 0.53567736, Log Avg loss: 0.72230847, Global Avg Loss: 1.44686520, Time: 0.0171 Steps: 69850, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001732, Sample Num: 27712, Cur Loss: 0.61254001, Cur Avg Loss: 0.53534291, Log Avg loss: 0.47775121, Global Avg Loss: 1.44672648, Time: 0.0114 Steps: 69860, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001742, Sample Num: 27872, Cur Loss: 0.22000894, Cur Avg Loss: 0.53560701, Log Avg loss: 0.58134939, Global Avg Loss: 1.44660263, Time: 0.0116 Steps: 69870, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001752, Sample Num: 28032, Cur Loss: 1.10034978, Cur Avg Loss: 0.53634612, Log Avg loss: 0.66509861, Global Avg Loss: 1.44649079, Time: 0.0110 Steps: 69880, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001762, Sample Num: 28192, Cur Loss: 0.18183145, Cur Avg Loss: 0.53637316, Log Avg loss: 0.54111013, Global Avg Loss: 1.44636125, Time: 0.0112 Steps: 69890, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001772, Sample Num: 28352, Cur Loss: 0.40720558, Cur Avg Loss: 0.53600960, Log Avg loss: 0.47195094, Global Avg Loss: 1.44622185, Time: 0.0108 Steps: 69900, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001782, Sample Num: 28512, Cur Loss: 0.47486737, Cur Avg Loss: 0.53557584, Log Avg loss: 0.45871334, Global Avg Loss: 1.44608059, Time: 0.0068 Steps: 69910, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001792, Sample Num: 28672, Cur Loss: 0.34884849, Cur Avg Loss: 0.53544428, Log Avg loss: 0.51199932, Global Avg Loss: 1.44594700, Time: 0.0076 Steps: 69920, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001802, Sample Num: 28832, Cur Loss: 0.92594838, Cur Avg Loss: 0.53480007, Log Avg loss: 0.41935907, Global Avg Loss: 1.44580020, Time: 0.0069 Steps: 69930, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001812, Sample Num: 28992, Cur Loss: 0.36385047, Cur Avg Loss: 0.53434117, Log Avg loss: 0.45164596, Global Avg Loss: 1.44565805, Time: 0.0066 Steps: 69940, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001822, Sample Num: 29152, Cur Loss: 0.31400108, Cur Avg Loss: 0.53502186, Log Avg loss: 0.65836313, Global Avg Loss: 1.44554550, Time: 0.0066 Steps: 69950, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001832, Sample Num: 29312, Cur Loss: 0.51234537, Cur Avg Loss: 0.53447882, Log Avg loss: 0.43553692, Global Avg Loss: 1.44540113, Time: 0.0111 Steps: 69960, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001842, Sample Num: 29472, Cur Loss: 0.77048576, Cur Avg Loss: 0.53454883, Log Avg loss: 0.54737525, Global Avg Loss: 1.44527279, Time: 0.0071 Steps: 69970, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001852, Sample Num: 29632, Cur Loss: 0.70396596, Cur Avg Loss: 0.53480684, Log Avg loss: 0.58233244, Global Avg Loss: 1.44514948, Time: 0.0119 Steps: 69980, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001862, Sample Num: 29792, Cur Loss: 0.29433349, Cur Avg Loss: 0.53423909, Log Avg loss: 0.42909108, Global Avg Loss: 1.44500430, Time: 0.0136 Steps: 69990, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001872, Sample Num: 29952, Cur Loss: 0.45144045, Cur Avg Loss: 0.53383651, Log Avg loss: 0.45887651, Global Avg Loss: 1.44486343, Time: 0.0068 Steps: 70000, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001882, Sample Num: 30112, Cur Loss: 0.79862976, Cur Avg Loss: 0.53397822, Log Avg loss: 0.56050656, Global Avg Loss: 1.44473711, Time: 0.0073 Steps: 70010, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001892, Sample Num: 30272, Cur Loss: 0.84553111, Cur Avg Loss: 0.53365318, Log Avg loss: 0.47248076, Global Avg Loss: 1.44459826, Time: 0.0084 Steps: 70020, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001902, Sample Num: 30432, Cur Loss: 0.33456439, Cur Avg Loss: 0.53297828, Log Avg loss: 0.40528618, Global Avg Loss: 1.44444985, Time: 0.0092 Steps: 70030, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001912, Sample Num: 30592, Cur Loss: 0.32506308, Cur Avg Loss: 0.53334676, Log Avg loss: 0.60343315, Global Avg Loss: 1.44432977, Time: 0.0112 Steps: 70040, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001922, Sample Num: 30752, Cur Loss: 0.85093415, Cur Avg Loss: 0.53437124, Log Avg loss: 0.73025123, Global Avg Loss: 1.44422783, Time: 0.0145 Steps: 70050, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001932, Sample Num: 30912, Cur Loss: 0.40384376, Cur Avg Loss: 0.53395644, Log Avg loss: 0.45423144, Global Avg Loss: 1.44408652, Time: 0.0126 Steps: 70060, Updated lr: 0.000035 Training, Epoch: 0033, Batch: 001942, Sample Num: 31072, Cur Loss: 0.59068072, Cur Avg Loss: 0.53432909, Log Avg loss: 0.60632561, Global Avg Loss: 1.44396696, Time: 0.0128 Steps: 70070, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001952, Sample Num: 31232, Cur Loss: 0.40411633, Cur Avg Loss: 0.53436455, Log Avg loss: 0.54124982, Global Avg Loss: 1.44383815, Time: 0.0126 Steps: 70080, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001962, Sample Num: 31392, Cur Loss: 0.28255665, Cur Avg Loss: 0.53442376, Log Avg loss: 0.54598282, Global Avg Loss: 1.44371005, Time: 0.0078 Steps: 70090, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001972, Sample Num: 31552, Cur Loss: 0.26315111, Cur Avg Loss: 0.53479408, Log Avg loss: 0.60744990, Global Avg Loss: 1.44359076, Time: 0.0066 Steps: 70100, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001982, Sample Num: 31712, Cur Loss: 0.37705958, Cur Avg Loss: 0.53495618, Log Avg loss: 0.56692345, Global Avg Loss: 1.44346571, Time: 0.0083 Steps: 70110, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 001992, Sample Num: 31872, Cur Loss: 0.45553476, Cur Avg Loss: 0.53513037, Log Avg loss: 0.56965405, Global Avg Loss: 1.44334110, Time: 0.0068 Steps: 70120, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002002, Sample Num: 32032, Cur Loss: 0.72145915, Cur Avg Loss: 0.53540013, Log Avg loss: 0.58913670, Global Avg Loss: 1.44321929, Time: 0.0096 Steps: 70130, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002012, Sample Num: 32192, Cur Loss: 0.66085953, Cur Avg Loss: 0.53512813, Log Avg loss: 0.48067341, Global Avg Loss: 1.44308206, Time: 0.0123 Steps: 70140, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002022, Sample Num: 32352, Cur Loss: 0.17032072, Cur Avg Loss: 0.53426344, Log Avg loss: 0.36028790, Global Avg Loss: 1.44292771, Time: 0.0065 Steps: 70150, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002032, Sample Num: 32512, Cur Loss: 0.21684822, Cur Avg Loss: 0.53356856, Log Avg loss: 0.39306370, Global Avg Loss: 1.44277807, Time: 0.0090 Steps: 70160, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002042, Sample Num: 32672, Cur Loss: 0.70296121, Cur Avg Loss: 0.53301252, Log Avg loss: 0.42002562, Global Avg Loss: 1.44263232, Time: 0.0072 Steps: 70170, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002052, Sample Num: 32832, Cur Loss: 0.82561123, Cur Avg Loss: 0.53384423, Log Avg loss: 0.70367966, Global Avg Loss: 1.44252702, Time: 0.0269 Steps: 70180, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002062, Sample Num: 32992, Cur Loss: 0.57305086, Cur Avg Loss: 0.53343260, Log Avg loss: 0.44896463, Global Avg Loss: 1.44238547, Time: 0.0115 Steps: 70190, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002072, Sample Num: 33152, Cur Loss: 0.36278921, Cur Avg Loss: 0.53337885, Log Avg loss: 0.52229621, Global Avg Loss: 1.44225440, Time: 0.0105 Steps: 70200, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002082, Sample Num: 33312, Cur Loss: 0.22339776, Cur Avg Loss: 0.53358260, Log Avg loss: 0.57579912, Global Avg Loss: 1.44213099, Time: 0.0086 Steps: 70210, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002092, Sample Num: 33472, Cur Loss: 0.33105642, Cur Avg Loss: 0.53304048, Log Avg loss: 0.42017123, Global Avg Loss: 1.44198546, Time: 0.0118 Steps: 70220, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002102, Sample Num: 33632, Cur Loss: 0.36551014, Cur Avg Loss: 0.53243075, Log Avg loss: 0.40487652, Global Avg Loss: 1.44183778, Time: 0.0066 Steps: 70230, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002112, Sample Num: 33792, Cur Loss: 1.02118516, Cur Avg Loss: 0.53208267, Log Avg loss: 0.45891581, Global Avg Loss: 1.44169785, Time: 0.0069 Steps: 70240, Updated lr: 0.000034 Training, Epoch: 0033, Batch: 002122, Sample Num: 33952, Cur Loss: 1.14419603, Cur Avg Loss: 0.53192793, Log Avg loss: 0.49924694, Global Avg Loss: 1.44156369, Time: 0.0118 Steps: 70250, Updated lr: 0.000034 ***** Running evaluation checkpoint-70257 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-70257 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.346065, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.628172, "eval_total_loss": 441.604575, "eval_mae": 0.612947, "eval_mse": 0.628324, "eval_r2": 0.600596, "eval_sp_statistic": 0.733017, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.781597, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.903039, "test_total_loss": 453.325481, "test_mae": 0.738666, "test_mse": 0.903195, "test_r2": 0.41707, "test_sp_statistic": 0.567799, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.652623, "test_ps_pvalue": 0.0, "lr": 3.432242769084875e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4414727456260255, "train_cur_epoch_loss": 1132.4526142813265, "train_cur_epoch_avg_loss": 0.5319176206112384, "train_cur_epoch_time": 21.346065044403076, "train_cur_epoch_avg_time": 0.010026333980461755, "epoch": 33, "step": 70257} ################################################## Training, Epoch: 0034, Batch: 000003, Sample Num: 48, Cur Loss: 0.20339258, Cur Avg Loss: 0.28858423, Log Avg loss: 0.45672967, Global Avg Loss: 1.44142352, Time: 0.0066 Steps: 70260, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000013, Sample Num: 208, Cur Loss: 0.70753503, Cur Avg Loss: 0.36168914, Log Avg loss: 0.38362061, Global Avg Loss: 1.44127298, Time: 0.0068 Steps: 70270, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000023, Sample Num: 368, Cur Loss: 0.52037162, Cur Avg Loss: 0.43776762, Log Avg loss: 0.53666965, Global Avg Loss: 1.44114427, Time: 0.0089 Steps: 70280, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000033, Sample Num: 528, Cur Loss: 0.14293706, Cur Avg Loss: 0.49715356, Log Avg loss: 0.63374121, Global Avg Loss: 1.44102940, Time: 0.0087 Steps: 70290, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000043, Sample Num: 688, Cur Loss: 0.44980943, Cur Avg Loss: 0.47920508, Log Avg loss: 0.41997511, Global Avg Loss: 1.44088416, Time: 0.0071 Steps: 70300, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000053, Sample Num: 848, Cur Loss: 0.30133289, Cur Avg Loss: 0.50351405, Log Avg loss: 0.60804259, Global Avg Loss: 1.44076571, Time: 0.0067 Steps: 70310, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000063, Sample Num: 1008, Cur Loss: 0.97520941, Cur Avg Loss: 0.53354180, Log Avg loss: 0.69268887, Global Avg Loss: 1.44065933, Time: 0.0105 Steps: 70320, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000073, Sample Num: 1168, Cur Loss: 0.58803785, Cur Avg Loss: 0.52664124, Log Avg loss: 0.48316771, Global Avg Loss: 1.44052318, Time: 0.0108 Steps: 70330, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000083, Sample Num: 1328, Cur Loss: 0.38695887, Cur Avg Loss: 0.50225978, Log Avg loss: 0.32427512, Global Avg Loss: 1.44036449, Time: 0.0067 Steps: 70340, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000093, Sample Num: 1488, Cur Loss: 0.67499959, Cur Avg Loss: 0.49219389, Log Avg loss: 0.40864700, Global Avg Loss: 1.44021784, Time: 0.0109 Steps: 70350, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000103, Sample Num: 1648, Cur Loss: 0.67694676, Cur Avg Loss: 0.48931568, Log Avg loss: 0.46254835, Global Avg Loss: 1.44007888, Time: 0.0108 Steps: 70360, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000113, Sample Num: 1808, Cur Loss: 0.44621840, Cur Avg Loss: 0.50019297, Log Avg loss: 0.61222903, Global Avg Loss: 1.43996124, Time: 0.0116 Steps: 70370, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000123, Sample Num: 1968, Cur Loss: 0.88599926, Cur Avg Loss: 0.50461903, Log Avg loss: 0.55463355, Global Avg Loss: 1.43983545, Time: 0.0119 Steps: 70380, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000133, Sample Num: 2128, Cur Loss: 0.43638009, Cur Avg Loss: 0.49988132, Log Avg loss: 0.44160753, Global Avg Loss: 1.43969363, Time: 0.0113 Steps: 70390, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000143, Sample Num: 2288, Cur Loss: 0.60564417, Cur Avg Loss: 0.50030120, Log Avg loss: 0.50588550, Global Avg Loss: 1.43956099, Time: 0.0068 Steps: 70400, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000153, Sample Num: 2448, Cur Loss: 0.27557284, Cur Avg Loss: 0.50429857, Log Avg loss: 0.56146097, Global Avg Loss: 1.43943628, Time: 0.0107 Steps: 70410, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000163, Sample Num: 2608, Cur Loss: 0.71739978, Cur Avg Loss: 0.50323482, Log Avg loss: 0.48695945, Global Avg Loss: 1.43930102, Time: 0.0066 Steps: 70420, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000173, Sample Num: 2768, Cur Loss: 0.28036171, Cur Avg Loss: 0.49616266, Log Avg loss: 0.38088656, Global Avg Loss: 1.43915074, Time: 0.0112 Steps: 70430, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000183, Sample Num: 2928, Cur Loss: 0.24363083, Cur Avg Loss: 0.48995005, Log Avg loss: 0.38247180, Global Avg Loss: 1.43900073, Time: 0.0118 Steps: 70440, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000193, Sample Num: 3088, Cur Loss: 0.79194462, Cur Avg Loss: 0.49107198, Log Avg loss: 0.51160334, Global Avg Loss: 1.43886909, Time: 0.0068 Steps: 70450, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000203, Sample Num: 3248, Cur Loss: 0.35952014, Cur Avg Loss: 0.49203044, Log Avg loss: 0.51052865, Global Avg Loss: 1.43873734, Time: 0.0111 Steps: 70460, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000213, Sample Num: 3408, Cur Loss: 0.62823623, Cur Avg Loss: 0.50175130, Log Avg loss: 0.69908477, Global Avg Loss: 1.43863238, Time: 0.0097 Steps: 70470, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000223, Sample Num: 3568, Cur Loss: 0.54068345, Cur Avg Loss: 0.50279646, Log Avg loss: 0.52505845, Global Avg Loss: 1.43850276, Time: 0.0066 Steps: 70480, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000233, Sample Num: 3728, Cur Loss: 0.22157159, Cur Avg Loss: 0.50380792, Log Avg loss: 0.52636330, Global Avg Loss: 1.43837336, Time: 0.0072 Steps: 70490, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000243, Sample Num: 3888, Cur Loss: 0.32096040, Cur Avg Loss: 0.49874900, Log Avg loss: 0.38087623, Global Avg Loss: 1.43822336, Time: 0.0070 Steps: 70500, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000253, Sample Num: 4048, Cur Loss: 0.57974768, Cur Avg Loss: 0.49843228, Log Avg loss: 0.49073600, Global Avg Loss: 1.43808898, Time: 0.0157 Steps: 70510, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000263, Sample Num: 4208, Cur Loss: 0.58242106, Cur Avg Loss: 0.50783539, Log Avg loss: 0.74573407, Global Avg Loss: 1.43799080, Time: 0.0074 Steps: 70520, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000273, Sample Num: 4368, Cur Loss: 0.22503322, Cur Avg Loss: 0.50966760, Log Avg loss: 0.55785490, Global Avg Loss: 1.43786601, Time: 0.0105 Steps: 70530, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000283, Sample Num: 4528, Cur Loss: 0.39437044, Cur Avg Loss: 0.50911472, Log Avg loss: 0.49402107, Global Avg Loss: 1.43773221, Time: 0.0064 Steps: 70540, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000293, Sample Num: 4688, Cur Loss: 0.43229556, Cur Avg Loss: 0.50861896, Log Avg loss: 0.49458871, Global Avg Loss: 1.43759853, Time: 0.0120 Steps: 70550, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000303, Sample Num: 4848, Cur Loss: 0.77230006, Cur Avg Loss: 0.51430659, Log Avg loss: 0.68095422, Global Avg Loss: 1.43749129, Time: 0.0076 Steps: 70560, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000313, Sample Num: 5008, Cur Loss: 0.81941772, Cur Avg Loss: 0.51443100, Log Avg loss: 0.51820070, Global Avg Loss: 1.43736103, Time: 0.0218 Steps: 70570, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000323, Sample Num: 5168, Cur Loss: 0.70804590, Cur Avg Loss: 0.51689408, Log Avg loss: 0.59398835, Global Avg Loss: 1.43724153, Time: 0.0109 Steps: 70580, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000333, Sample Num: 5328, Cur Loss: 0.35662004, Cur Avg Loss: 0.51148473, Log Avg loss: 0.33676296, Global Avg Loss: 1.43708564, Time: 0.0063 Steps: 70590, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000343, Sample Num: 5488, Cur Loss: 0.47413853, Cur Avg Loss: 0.51235220, Log Avg loss: 0.54123898, Global Avg Loss: 1.43695875, Time: 0.0105 Steps: 70600, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000353, Sample Num: 5648, Cur Loss: 0.46984538, Cur Avg Loss: 0.51131232, Log Avg loss: 0.47564436, Global Avg Loss: 1.43682260, Time: 0.0132 Steps: 70610, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000363, Sample Num: 5808, Cur Loss: 0.58737075, Cur Avg Loss: 0.51424814, Log Avg loss: 0.61788258, Global Avg Loss: 1.43670664, Time: 0.0124 Steps: 70620, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000373, Sample Num: 5968, Cur Loss: 0.41721988, Cur Avg Loss: 0.51117607, Log Avg loss: 0.39965977, Global Avg Loss: 1.43655981, Time: 0.0120 Steps: 70630, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000383, Sample Num: 6128, Cur Loss: 0.66923201, Cur Avg Loss: 0.51000406, Log Avg loss: 0.46628831, Global Avg Loss: 1.43642246, Time: 0.0103 Steps: 70640, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000393, Sample Num: 6288, Cur Loss: 0.26027754, Cur Avg Loss: 0.50846120, Log Avg loss: 0.44936969, Global Avg Loss: 1.43628275, Time: 0.0135 Steps: 70650, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000403, Sample Num: 6448, Cur Loss: 0.49212146, Cur Avg Loss: 0.50503434, Log Avg loss: 0.37035849, Global Avg Loss: 1.43613189, Time: 0.0068 Steps: 70660, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000413, Sample Num: 6608, Cur Loss: 1.27852201, Cur Avg Loss: 0.50866757, Log Avg loss: 0.65508696, Global Avg Loss: 1.43602137, Time: 0.0112 Steps: 70670, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000423, Sample Num: 6768, Cur Loss: 0.43411297, Cur Avg Loss: 0.50945398, Log Avg loss: 0.54193278, Global Avg Loss: 1.43589487, Time: 0.0133 Steps: 70680, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000433, Sample Num: 6928, Cur Loss: 0.65876389, Cur Avg Loss: 0.51059234, Log Avg loss: 0.55874473, Global Avg Loss: 1.43577079, Time: 0.0106 Steps: 70690, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000443, Sample Num: 7088, Cur Loss: 0.27020782, Cur Avg Loss: 0.51034580, Log Avg loss: 0.49967060, Global Avg Loss: 1.43563839, Time: 0.0126 Steps: 70700, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000453, Sample Num: 7248, Cur Loss: 0.30107954, Cur Avg Loss: 0.51057633, Log Avg loss: 0.52078870, Global Avg Loss: 1.43550901, Time: 0.0145 Steps: 70710, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000463, Sample Num: 7408, Cur Loss: 0.63035882, Cur Avg Loss: 0.50815177, Log Avg loss: 0.39831918, Global Avg Loss: 1.43536234, Time: 0.0077 Steps: 70720, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000473, Sample Num: 7568, Cur Loss: 0.48556399, Cur Avg Loss: 0.50980553, Log Avg loss: 0.58637484, Global Avg Loss: 1.43524231, Time: 0.0080 Steps: 70730, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000483, Sample Num: 7728, Cur Loss: 0.64340013, Cur Avg Loss: 0.50882246, Log Avg loss: 0.46232334, Global Avg Loss: 1.43510478, Time: 0.0069 Steps: 70740, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000493, Sample Num: 7888, Cur Loss: 0.45104641, Cur Avg Loss: 0.50748379, Log Avg loss: 0.44282606, Global Avg Loss: 1.43496453, Time: 0.0080 Steps: 70750, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000503, Sample Num: 8048, Cur Loss: 0.87104690, Cur Avg Loss: 0.50872966, Log Avg loss: 0.57015087, Global Avg Loss: 1.43484231, Time: 0.0127 Steps: 70760, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000513, Sample Num: 8208, Cur Loss: 0.34512007, Cur Avg Loss: 0.50908951, Log Avg loss: 0.52719004, Global Avg Loss: 1.43471405, Time: 0.0152 Steps: 70770, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000523, Sample Num: 8368, Cur Loss: 0.56796056, Cur Avg Loss: 0.50827604, Log Avg loss: 0.46654519, Global Avg Loss: 1.43457727, Time: 0.0107 Steps: 70780, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000533, Sample Num: 8528, Cur Loss: 0.33836395, Cur Avg Loss: 0.50594196, Log Avg loss: 0.38386944, Global Avg Loss: 1.43442884, Time: 0.0067 Steps: 70790, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000543, Sample Num: 8688, Cur Loss: 0.41029564, Cur Avg Loss: 0.50324050, Log Avg loss: 0.35925281, Global Avg Loss: 1.43427698, Time: 0.0065 Steps: 70800, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000553, Sample Num: 8848, Cur Loss: 0.48492247, Cur Avg Loss: 0.50285646, Log Avg loss: 0.48200305, Global Avg Loss: 1.43414250, Time: 0.0070 Steps: 70810, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000563, Sample Num: 9008, Cur Loss: 1.32478154, Cur Avg Loss: 0.50278594, Log Avg loss: 0.49888609, Global Avg Loss: 1.43401044, Time: 0.0070 Steps: 70820, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000573, Sample Num: 9168, Cur Loss: 0.38673750, Cur Avg Loss: 0.50165165, Log Avg loss: 0.43779095, Global Avg Loss: 1.43386979, Time: 0.0120 Steps: 70830, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000583, Sample Num: 9328, Cur Loss: 0.94723362, Cur Avg Loss: 0.50555378, Log Avg loss: 0.72914568, Global Avg Loss: 1.43377031, Time: 0.0105 Steps: 70840, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000593, Sample Num: 9488, Cur Loss: 1.20101011, Cur Avg Loss: 0.50580864, Log Avg loss: 0.52066748, Global Avg Loss: 1.43364143, Time: 0.0137 Steps: 70850, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000603, Sample Num: 9648, Cur Loss: 0.36493227, Cur Avg Loss: 0.50415312, Log Avg loss: 0.40598076, Global Avg Loss: 1.43349640, Time: 0.0117 Steps: 70860, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000613, Sample Num: 9808, Cur Loss: 0.40366387, Cur Avg Loss: 0.50296748, Log Avg loss: 0.43147322, Global Avg Loss: 1.43335501, Time: 0.0116 Steps: 70870, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000623, Sample Num: 9968, Cur Loss: 0.56960583, Cur Avg Loss: 0.50233519, Log Avg loss: 0.46357584, Global Avg Loss: 1.43321819, Time: 0.0066 Steps: 70880, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000633, Sample Num: 10128, Cur Loss: 0.26643607, Cur Avg Loss: 0.50359425, Log Avg loss: 0.58203376, Global Avg Loss: 1.43309812, Time: 0.0088 Steps: 70890, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000643, Sample Num: 10288, Cur Loss: 0.29647872, Cur Avg Loss: 0.50203586, Log Avg loss: 0.40338961, Global Avg Loss: 1.43295289, Time: 0.0065 Steps: 70900, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000653, Sample Num: 10448, Cur Loss: 0.30627561, Cur Avg Loss: 0.50071473, Log Avg loss: 0.41576573, Global Avg Loss: 1.43280944, Time: 0.0116 Steps: 70910, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000663, Sample Num: 10608, Cur Loss: 0.29731226, Cur Avg Loss: 0.50074765, Log Avg loss: 0.50289754, Global Avg Loss: 1.43267832, Time: 0.0067 Steps: 70920, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000673, Sample Num: 10768, Cur Loss: 0.37432823, Cur Avg Loss: 0.50337677, Log Avg loss: 0.67768749, Global Avg Loss: 1.43257188, Time: 0.0066 Steps: 70930, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000683, Sample Num: 10928, Cur Loss: 0.62908125, Cur Avg Loss: 0.50201144, Log Avg loss: 0.41012478, Global Avg Loss: 1.43242775, Time: 0.0066 Steps: 70940, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000693, Sample Num: 11088, Cur Loss: 0.44460157, Cur Avg Loss: 0.50066501, Log Avg loss: 0.40870356, Global Avg Loss: 1.43228346, Time: 0.0130 Steps: 70950, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000703, Sample Num: 11248, Cur Loss: 0.35125911, Cur Avg Loss: 0.50296005, Log Avg loss: 0.66200644, Global Avg Loss: 1.43217491, Time: 0.0066 Steps: 70960, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000713, Sample Num: 11408, Cur Loss: 0.64126086, Cur Avg Loss: 0.50425749, Log Avg loss: 0.59546731, Global Avg Loss: 1.43205701, Time: 0.0066 Steps: 70970, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000723, Sample Num: 11568, Cur Loss: 0.58013225, Cur Avg Loss: 0.50847267, Log Avg loss: 0.80901538, Global Avg Loss: 1.43196924, Time: 0.0064 Steps: 70980, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000733, Sample Num: 11728, Cur Loss: 0.55549979, Cur Avg Loss: 0.50919636, Log Avg loss: 0.56151906, Global Avg Loss: 1.43184662, Time: 0.0063 Steps: 70990, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000743, Sample Num: 11888, Cur Loss: 0.93059272, Cur Avg Loss: 0.51065944, Log Avg loss: 0.61790315, Global Avg Loss: 1.43173198, Time: 0.0131 Steps: 71000, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000753, Sample Num: 12048, Cur Loss: 0.46246728, Cur Avg Loss: 0.50923825, Log Avg loss: 0.40364365, Global Avg Loss: 1.43158720, Time: 0.0065 Steps: 71010, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000763, Sample Num: 12208, Cur Loss: 0.60426718, Cur Avg Loss: 0.50825844, Log Avg loss: 0.43447941, Global Avg Loss: 1.43144680, Time: 0.0131 Steps: 71020, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000773, Sample Num: 12368, Cur Loss: 0.49465984, Cur Avg Loss: 0.50732084, Log Avg loss: 0.43578173, Global Avg Loss: 1.43130663, Time: 0.0075 Steps: 71030, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000783, Sample Num: 12528, Cur Loss: 0.87043649, Cur Avg Loss: 0.50961276, Log Avg loss: 0.68677800, Global Avg Loss: 1.43120182, Time: 0.0077 Steps: 71040, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000793, Sample Num: 12688, Cur Loss: 0.90684116, Cur Avg Loss: 0.51007886, Log Avg loss: 0.54657445, Global Avg Loss: 1.43107731, Time: 0.0115 Steps: 71050, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000803, Sample Num: 12848, Cur Loss: 0.19250417, Cur Avg Loss: 0.51073273, Log Avg loss: 0.56258469, Global Avg Loss: 1.43095510, Time: 0.0068 Steps: 71060, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000813, Sample Num: 13008, Cur Loss: 0.36046243, Cur Avg Loss: 0.51150575, Log Avg loss: 0.57357935, Global Avg Loss: 1.43083446, Time: 0.0068 Steps: 71070, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000823, Sample Num: 13168, Cur Loss: 0.32194275, Cur Avg Loss: 0.51403979, Log Avg loss: 0.72005756, Global Avg Loss: 1.43073446, Time: 0.0067 Steps: 71080, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000833, Sample Num: 13328, Cur Loss: 0.17610958, Cur Avg Loss: 0.51310580, Log Avg loss: 0.43623775, Global Avg Loss: 1.43059457, Time: 0.0106 Steps: 71090, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000843, Sample Num: 13488, Cur Loss: 0.31597713, Cur Avg Loss: 0.51151301, Log Avg loss: 0.37883349, Global Avg Loss: 1.43044664, Time: 0.0130 Steps: 71100, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000853, Sample Num: 13648, Cur Loss: 0.46403277, Cur Avg Loss: 0.51294790, Log Avg loss: 0.63390914, Global Avg Loss: 1.43033463, Time: 0.0068 Steps: 71110, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000863, Sample Num: 13808, Cur Loss: 0.33624321, Cur Avg Loss: 0.51194370, Log Avg loss: 0.42628603, Global Avg Loss: 1.43019345, Time: 0.0078 Steps: 71120, Updated lr: 0.000034 Training, Epoch: 0034, Batch: 000873, Sample Num: 13968, Cur Loss: 0.57165974, Cur Avg Loss: 0.51250865, Log Avg loss: 0.56126366, Global Avg Loss: 1.43007129, Time: 0.0126 Steps: 71130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000883, Sample Num: 14128, Cur Loss: 0.61631238, Cur Avg Loss: 0.51318071, Log Avg loss: 0.57185179, Global Avg Loss: 1.42995065, Time: 0.0086 Steps: 71140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000893, Sample Num: 14288, Cur Loss: 0.44424915, Cur Avg Loss: 0.51287001, Log Avg loss: 0.48543527, Global Avg Loss: 1.42981790, Time: 0.0066 Steps: 71150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000903, Sample Num: 14448, Cur Loss: 0.70339644, Cur Avg Loss: 0.51376179, Log Avg loss: 0.59339733, Global Avg Loss: 1.42970036, Time: 0.0066 Steps: 71160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000913, Sample Num: 14608, Cur Loss: 0.66171092, Cur Avg Loss: 0.51383395, Log Avg loss: 0.52034987, Global Avg Loss: 1.42957259, Time: 0.0116 Steps: 71170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000923, Sample Num: 14768, Cur Loss: 0.74595904, Cur Avg Loss: 0.51301012, Log Avg loss: 0.43779436, Global Avg Loss: 1.42943325, Time: 0.0118 Steps: 71180, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000933, Sample Num: 14928, Cur Loss: 0.64332360, Cur Avg Loss: 0.51573557, Log Avg loss: 0.76729512, Global Avg Loss: 1.42934024, Time: 0.0067 Steps: 71190, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000943, Sample Num: 15088, Cur Loss: 0.36962065, Cur Avg Loss: 0.51577830, Log Avg loss: 0.51976451, Global Avg Loss: 1.42921249, Time: 0.0071 Steps: 71200, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000953, Sample Num: 15248, Cur Loss: 0.39098576, Cur Avg Loss: 0.51555712, Log Avg loss: 0.49470009, Global Avg Loss: 1.42908126, Time: 0.0065 Steps: 71210, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000963, Sample Num: 15408, Cur Loss: 0.48523641, Cur Avg Loss: 0.51502305, Log Avg loss: 0.46412572, Global Avg Loss: 1.42894577, Time: 0.0068 Steps: 71220, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000973, Sample Num: 15568, Cur Loss: 0.52118313, Cur Avg Loss: 0.51517186, Log Avg loss: 0.52950234, Global Avg Loss: 1.42881950, Time: 0.0115 Steps: 71230, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000983, Sample Num: 15728, Cur Loss: 0.26606768, Cur Avg Loss: 0.51533287, Log Avg loss: 0.53099915, Global Avg Loss: 1.42869347, Time: 0.0067 Steps: 71240, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 000993, Sample Num: 15888, Cur Loss: 2.26199675, Cur Avg Loss: 0.51685845, Log Avg loss: 0.66682284, Global Avg Loss: 1.42858654, Time: 0.0067 Steps: 71250, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001003, Sample Num: 16048, Cur Loss: 0.44311652, Cur Avg Loss: 0.51661721, Log Avg loss: 0.49266212, Global Avg Loss: 1.42845520, Time: 0.0067 Steps: 71260, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001013, Sample Num: 16208, Cur Loss: 0.37531066, Cur Avg Loss: 0.51647921, Log Avg loss: 0.50263850, Global Avg Loss: 1.42832530, Time: 0.0067 Steps: 71270, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001023, Sample Num: 16368, Cur Loss: 0.27305862, Cur Avg Loss: 0.51613014, Log Avg loss: 0.48076939, Global Avg Loss: 1.42819237, Time: 0.0127 Steps: 71280, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001033, Sample Num: 16528, Cur Loss: 0.28778097, Cur Avg Loss: 0.51738388, Log Avg loss: 0.64564100, Global Avg Loss: 1.42808260, Time: 0.0113 Steps: 71290, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001043, Sample Num: 16688, Cur Loss: 0.31613067, Cur Avg Loss: 0.51770734, Log Avg loss: 0.55112058, Global Avg Loss: 1.42795960, Time: 0.0111 Steps: 71300, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001053, Sample Num: 16848, Cur Loss: 0.23623517, Cur Avg Loss: 0.51708216, Log Avg loss: 0.45187628, Global Avg Loss: 1.42782272, Time: 0.0108 Steps: 71310, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001063, Sample Num: 17008, Cur Loss: 0.34108171, Cur Avg Loss: 0.51697635, Log Avg loss: 0.50583500, Global Avg Loss: 1.42769345, Time: 0.0067 Steps: 71320, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001073, Sample Num: 17168, Cur Loss: 0.60872680, Cur Avg Loss: 0.51665952, Log Avg loss: 0.48297951, Global Avg Loss: 1.42756100, Time: 0.0115 Steps: 71330, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001083, Sample Num: 17328, Cur Loss: 0.67360401, Cur Avg Loss: 0.51721131, Log Avg loss: 0.57641841, Global Avg Loss: 1.42744170, Time: 0.0110 Steps: 71340, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001093, Sample Num: 17488, Cur Loss: 0.53388351, Cur Avg Loss: 0.51825986, Log Avg loss: 0.63181792, Global Avg Loss: 1.42733019, Time: 0.0195 Steps: 71350, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001103, Sample Num: 17648, Cur Loss: 0.54629481, Cur Avg Loss: 0.51749318, Log Avg loss: 0.43369562, Global Avg Loss: 1.42719094, Time: 0.0112 Steps: 71360, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001113, Sample Num: 17808, Cur Loss: 0.89427876, Cur Avg Loss: 0.51722267, Log Avg loss: 0.48738495, Global Avg Loss: 1.42705926, Time: 0.0123 Steps: 71370, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001123, Sample Num: 17968, Cur Loss: 0.28454095, Cur Avg Loss: 0.51745377, Log Avg loss: 0.54317588, Global Avg Loss: 1.42693543, Time: 0.0108 Steps: 71380, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001133, Sample Num: 18128, Cur Loss: 0.22672972, Cur Avg Loss: 0.51717597, Log Avg loss: 0.48597816, Global Avg Loss: 1.42680363, Time: 0.0068 Steps: 71390, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001143, Sample Num: 18288, Cur Loss: 0.46066877, Cur Avg Loss: 0.51873294, Log Avg loss: 0.69513783, Global Avg Loss: 1.42670115, Time: 0.0072 Steps: 71400, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001153, Sample Num: 18448, Cur Loss: 0.27156937, Cur Avg Loss: 0.51977025, Log Avg loss: 0.63833545, Global Avg Loss: 1.42659075, Time: 0.0072 Steps: 71410, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001163, Sample Num: 18608, Cur Loss: 0.18223876, Cur Avg Loss: 0.51957691, Log Avg loss: 0.49728470, Global Avg Loss: 1.42646064, Time: 0.0105 Steps: 71420, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001173, Sample Num: 18768, Cur Loss: 0.46991837, Cur Avg Loss: 0.51961003, Log Avg loss: 0.52346142, Global Avg Loss: 1.42633422, Time: 0.0087 Steps: 71430, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001183, Sample Num: 18928, Cur Loss: 0.39675930, Cur Avg Loss: 0.51927754, Log Avg loss: 0.48027700, Global Avg Loss: 1.42620179, Time: 0.0066 Steps: 71440, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001193, Sample Num: 19088, Cur Loss: 0.48131427, Cur Avg Loss: 0.51902906, Log Avg loss: 0.48963369, Global Avg Loss: 1.42607071, Time: 0.0069 Steps: 71450, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001203, Sample Num: 19248, Cur Loss: 0.75354999, Cur Avg Loss: 0.51857580, Log Avg loss: 0.46450183, Global Avg Loss: 1.42593615, Time: 0.0115 Steps: 71460, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001213, Sample Num: 19408, Cur Loss: 0.21530941, Cur Avg Loss: 0.51782475, Log Avg loss: 0.42747307, Global Avg Loss: 1.42579645, Time: 0.0066 Steps: 71470, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001223, Sample Num: 19568, Cur Loss: 0.29717097, Cur Avg Loss: 0.51680867, Log Avg loss: 0.39355809, Global Avg Loss: 1.42565204, Time: 0.0153 Steps: 71480, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001233, Sample Num: 19728, Cur Loss: 0.91216826, Cur Avg Loss: 0.51693177, Log Avg loss: 0.53198757, Global Avg Loss: 1.42552703, Time: 0.0111 Steps: 71490, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001243, Sample Num: 19888, Cur Loss: 0.39317420, Cur Avg Loss: 0.51747572, Log Avg loss: 0.58454499, Global Avg Loss: 1.42540941, Time: 0.0138 Steps: 71500, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001253, Sample Num: 20048, Cur Loss: 0.17225228, Cur Avg Loss: 0.51684971, Log Avg loss: 0.43903622, Global Avg Loss: 1.42527148, Time: 0.0106 Steps: 71510, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001263, Sample Num: 20208, Cur Loss: 0.49116886, Cur Avg Loss: 0.51841065, Log Avg loss: 0.71399623, Global Avg Loss: 1.42517203, Time: 0.0122 Steps: 71520, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001273, Sample Num: 20368, Cur Loss: 0.19531302, Cur Avg Loss: 0.51838897, Log Avg loss: 0.51565048, Global Avg Loss: 1.42504487, Time: 0.0153 Steps: 71530, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001283, Sample Num: 20528, Cur Loss: 0.45718974, Cur Avg Loss: 0.51880704, Log Avg loss: 0.57202799, Global Avg Loss: 1.42492564, Time: 0.0068 Steps: 71540, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001293, Sample Num: 20688, Cur Loss: 0.27340281, Cur Avg Loss: 0.51994948, Log Avg loss: 0.66652462, Global Avg Loss: 1.42481964, Time: 0.0099 Steps: 71550, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001303, Sample Num: 20848, Cur Loss: 0.15634015, Cur Avg Loss: 0.51907608, Log Avg loss: 0.40614512, Global Avg Loss: 1.42467729, Time: 0.0114 Steps: 71560, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001313, Sample Num: 21008, Cur Loss: 0.21174097, Cur Avg Loss: 0.51924800, Log Avg loss: 0.54164885, Global Avg Loss: 1.42455391, Time: 0.0067 Steps: 71570, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001323, Sample Num: 21168, Cur Loss: 0.62058878, Cur Avg Loss: 0.51977586, Log Avg loss: 0.58908413, Global Avg Loss: 1.42443719, Time: 0.0113 Steps: 71580, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001333, Sample Num: 21328, Cur Loss: 0.49223292, Cur Avg Loss: 0.51842386, Log Avg loss: 0.33955474, Global Avg Loss: 1.42428565, Time: 0.0108 Steps: 71590, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001343, Sample Num: 21488, Cur Loss: 0.38530737, Cur Avg Loss: 0.51891172, Log Avg loss: 0.58394366, Global Avg Loss: 1.42416828, Time: 0.0105 Steps: 71600, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001353, Sample Num: 21648, Cur Loss: 0.71073031, Cur Avg Loss: 0.51924151, Log Avg loss: 0.56353174, Global Avg Loss: 1.42404810, Time: 0.0068 Steps: 71610, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001363, Sample Num: 21808, Cur Loss: 0.30174807, Cur Avg Loss: 0.51923747, Log Avg loss: 0.51869145, Global Avg Loss: 1.42392169, Time: 0.0121 Steps: 71620, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001373, Sample Num: 21968, Cur Loss: 1.32715380, Cur Avg Loss: 0.52033892, Log Avg loss: 0.67046604, Global Avg Loss: 1.42381650, Time: 0.0110 Steps: 71630, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001383, Sample Num: 22128, Cur Loss: 0.56699568, Cur Avg Loss: 0.51973472, Log Avg loss: 0.43677836, Global Avg Loss: 1.42367872, Time: 0.0106 Steps: 71640, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001393, Sample Num: 22288, Cur Loss: 0.37206912, Cur Avg Loss: 0.51943877, Log Avg loss: 0.47850808, Global Avg Loss: 1.42354681, Time: 0.0081 Steps: 71650, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001403, Sample Num: 22448, Cur Loss: 0.76238239, Cur Avg Loss: 0.51890756, Log Avg loss: 0.44491021, Global Avg Loss: 1.42341024, Time: 0.0118 Steps: 71660, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001413, Sample Num: 22608, Cur Loss: 0.74500763, Cur Avg Loss: 0.51902907, Log Avg loss: 0.53607651, Global Avg Loss: 1.42328643, Time: 0.0068 Steps: 71670, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001423, Sample Num: 22768, Cur Loss: 0.36461937, Cur Avg Loss: 0.51867005, Log Avg loss: 0.46794158, Global Avg Loss: 1.42315316, Time: 0.0066 Steps: 71680, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001433, Sample Num: 22928, Cur Loss: 0.74979240, Cur Avg Loss: 0.51915381, Log Avg loss: 0.58799239, Global Avg Loss: 1.42303666, Time: 0.0114 Steps: 71690, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001443, Sample Num: 23088, Cur Loss: 0.32285765, Cur Avg Loss: 0.52008574, Log Avg loss: 0.65363148, Global Avg Loss: 1.42292935, Time: 0.0110 Steps: 71700, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001453, Sample Num: 23248, Cur Loss: 0.21752819, Cur Avg Loss: 0.51922192, Log Avg loss: 0.39457285, Global Avg Loss: 1.42278595, Time: 0.0123 Steps: 71710, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001463, Sample Num: 23408, Cur Loss: 0.58583248, Cur Avg Loss: 0.51841296, Log Avg loss: 0.40087020, Global Avg Loss: 1.42264346, Time: 0.0155 Steps: 71720, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001473, Sample Num: 23568, Cur Loss: 0.15796652, Cur Avg Loss: 0.51788220, Log Avg loss: 0.44023319, Global Avg Loss: 1.42250650, Time: 0.0132 Steps: 71730, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001483, Sample Num: 23728, Cur Loss: 0.44922489, Cur Avg Loss: 0.51772397, Log Avg loss: 0.49441593, Global Avg Loss: 1.42237713, Time: 0.0070 Steps: 71740, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001493, Sample Num: 23888, Cur Loss: 0.73838723, Cur Avg Loss: 0.51831674, Log Avg loss: 0.60622490, Global Avg Loss: 1.42226338, Time: 0.0076 Steps: 71750, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001503, Sample Num: 24048, Cur Loss: 1.23626983, Cur Avg Loss: 0.51921163, Log Avg loss: 0.65281866, Global Avg Loss: 1.42215616, Time: 0.0065 Steps: 71760, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001513, Sample Num: 24208, Cur Loss: 1.39275587, Cur Avg Loss: 0.52008392, Log Avg loss: 0.65118911, Global Avg Loss: 1.42204873, Time: 0.0111 Steps: 71770, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001523, Sample Num: 24368, Cur Loss: 0.63242120, Cur Avg Loss: 0.52049777, Log Avg loss: 0.58311306, Global Avg Loss: 1.42193186, Time: 0.0115 Steps: 71780, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001533, Sample Num: 24528, Cur Loss: 0.74221939, Cur Avg Loss: 0.52044665, Log Avg loss: 0.51266068, Global Avg Loss: 1.42180520, Time: 0.0106 Steps: 71790, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001543, Sample Num: 24688, Cur Loss: 0.57315040, Cur Avg Loss: 0.52035463, Log Avg loss: 0.50624901, Global Avg Loss: 1.42167769, Time: 0.0083 Steps: 71800, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001553, Sample Num: 24848, Cur Loss: 0.55227846, Cur Avg Loss: 0.52027122, Log Avg loss: 0.50740045, Global Avg Loss: 1.42155037, Time: 0.0085 Steps: 71810, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001563, Sample Num: 25008, Cur Loss: 0.57365328, Cur Avg Loss: 0.52020933, Log Avg loss: 0.51059814, Global Avg Loss: 1.42142353, Time: 0.0069 Steps: 71820, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001573, Sample Num: 25168, Cur Loss: 0.97698551, Cur Avg Loss: 0.52130174, Log Avg loss: 0.69204451, Global Avg Loss: 1.42132199, Time: 0.0069 Steps: 71830, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001583, Sample Num: 25328, Cur Loss: 0.42687204, Cur Avg Loss: 0.52159031, Log Avg loss: 0.56698309, Global Avg Loss: 1.42120306, Time: 0.0065 Steps: 71840, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001593, Sample Num: 25488, Cur Loss: 0.38635349, Cur Avg Loss: 0.52159369, Log Avg loss: 0.52212829, Global Avg Loss: 1.42107793, Time: 0.0110 Steps: 71850, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001603, Sample Num: 25648, Cur Loss: 0.22151080, Cur Avg Loss: 0.52197113, Log Avg loss: 0.58209813, Global Avg Loss: 1.42096118, Time: 0.0084 Steps: 71860, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001613, Sample Num: 25808, Cur Loss: 0.24870452, Cur Avg Loss: 0.52149527, Log Avg loss: 0.44521447, Global Avg Loss: 1.42082541, Time: 0.0073 Steps: 71870, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001623, Sample Num: 25968, Cur Loss: 0.21515633, Cur Avg Loss: 0.52072673, Log Avg loss: 0.39676058, Global Avg Loss: 1.42068295, Time: 0.0108 Steps: 71880, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001633, Sample Num: 26128, Cur Loss: 0.32777381, Cur Avg Loss: 0.52090745, Log Avg loss: 0.55023969, Global Avg Loss: 1.42056187, Time: 0.0095 Steps: 71890, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001643, Sample Num: 26288, Cur Loss: 0.34625083, Cur Avg Loss: 0.52077715, Log Avg loss: 0.49949886, Global Avg Loss: 1.42043376, Time: 0.0123 Steps: 71900, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001653, Sample Num: 26448, Cur Loss: 0.67561859, Cur Avg Loss: 0.52169129, Log Avg loss: 0.67188455, Global Avg Loss: 1.42032967, Time: 0.0133 Steps: 71910, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001663, Sample Num: 26608, Cur Loss: 0.26763594, Cur Avg Loss: 0.52130662, Log Avg loss: 0.45772064, Global Avg Loss: 1.42019582, Time: 0.0070 Steps: 71920, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001673, Sample Num: 26768, Cur Loss: 0.14771351, Cur Avg Loss: 0.52067347, Log Avg loss: 0.41538045, Global Avg Loss: 1.42005613, Time: 0.0088 Steps: 71930, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001683, Sample Num: 26928, Cur Loss: 0.43819469, Cur Avg Loss: 0.52020659, Log Avg loss: 0.44209783, Global Avg Loss: 1.41992019, Time: 0.0109 Steps: 71940, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001693, Sample Num: 27088, Cur Loss: 0.48564214, Cur Avg Loss: 0.52059736, Log Avg loss: 0.58636349, Global Avg Loss: 1.41980434, Time: 0.0067 Steps: 71950, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001703, Sample Num: 27248, Cur Loss: 1.16562533, Cur Avg Loss: 0.52137834, Log Avg loss: 0.65359769, Global Avg Loss: 1.41969786, Time: 0.0068 Steps: 71960, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001713, Sample Num: 27408, Cur Loss: 0.36815256, Cur Avg Loss: 0.52201963, Log Avg loss: 0.63123204, Global Avg Loss: 1.41958831, Time: 0.0115 Steps: 71970, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001723, Sample Num: 27568, Cur Loss: 0.37558013, Cur Avg Loss: 0.52176700, Log Avg loss: 0.47849151, Global Avg Loss: 1.41945756, Time: 0.0066 Steps: 71980, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001733, Sample Num: 27728, Cur Loss: 1.10366809, Cur Avg Loss: 0.52313920, Log Avg loss: 0.75956970, Global Avg Loss: 1.41936590, Time: 0.0088 Steps: 71990, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001743, Sample Num: 27888, Cur Loss: 0.81393659, Cur Avg Loss: 0.52336133, Log Avg loss: 0.56185594, Global Avg Loss: 1.41924680, Time: 0.0107 Steps: 72000, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001753, Sample Num: 28048, Cur Loss: 0.20154117, Cur Avg Loss: 0.52294893, Log Avg loss: 0.45106737, Global Avg Loss: 1.41911235, Time: 0.0114 Steps: 72010, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001763, Sample Num: 28208, Cur Loss: 0.21566725, Cur Avg Loss: 0.52277059, Log Avg loss: 0.49150829, Global Avg Loss: 1.41898355, Time: 0.0121 Steps: 72020, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001773, Sample Num: 28368, Cur Loss: 0.28980744, Cur Avg Loss: 0.52217673, Log Avg loss: 0.41747823, Global Avg Loss: 1.41884451, Time: 0.0109 Steps: 72030, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001783, Sample Num: 28528, Cur Loss: 0.50998449, Cur Avg Loss: 0.52253689, Log Avg loss: 0.58639320, Global Avg Loss: 1.41872896, Time: 0.0115 Steps: 72040, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001793, Sample Num: 28688, Cur Loss: 0.79074883, Cur Avg Loss: 0.52320446, Log Avg loss: 0.64223312, Global Avg Loss: 1.41862118, Time: 0.0194 Steps: 72050, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001803, Sample Num: 28848, Cur Loss: 0.97427964, Cur Avg Loss: 0.52347788, Log Avg loss: 0.57250093, Global Avg Loss: 1.41850376, Time: 0.0138 Steps: 72060, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001813, Sample Num: 29008, Cur Loss: 0.53901744, Cur Avg Loss: 0.52337655, Log Avg loss: 0.50510812, Global Avg Loss: 1.41837703, Time: 0.0067 Steps: 72070, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001823, Sample Num: 29168, Cur Loss: 0.13561779, Cur Avg Loss: 0.52358739, Log Avg loss: 0.56181097, Global Avg Loss: 1.41825819, Time: 0.0120 Steps: 72080, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001833, Sample Num: 29328, Cur Loss: 1.01702404, Cur Avg Loss: 0.52351770, Log Avg loss: 0.51081470, Global Avg Loss: 1.41813232, Time: 0.0157 Steps: 72090, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001843, Sample Num: 29488, Cur Loss: 0.64798993, Cur Avg Loss: 0.52249948, Log Avg loss: 0.33585835, Global Avg Loss: 1.41798221, Time: 0.0114 Steps: 72100, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001853, Sample Num: 29648, Cur Loss: 0.51030946, Cur Avg Loss: 0.52318625, Log Avg loss: 0.64975840, Global Avg Loss: 1.41787567, Time: 0.0194 Steps: 72110, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001863, Sample Num: 29808, Cur Loss: 0.37076661, Cur Avg Loss: 0.52352710, Log Avg loss: 0.58668741, Global Avg Loss: 1.41776042, Time: 0.0107 Steps: 72120, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001873, Sample Num: 29968, Cur Loss: 0.46062371, Cur Avg Loss: 0.52374970, Log Avg loss: 0.56521947, Global Avg Loss: 1.41764223, Time: 0.0117 Steps: 72130, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001883, Sample Num: 30128, Cur Loss: 0.35400486, Cur Avg Loss: 0.52332113, Log Avg loss: 0.44305077, Global Avg Loss: 1.41750713, Time: 0.0068 Steps: 72140, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001893, Sample Num: 30288, Cur Loss: 0.48058447, Cur Avg Loss: 0.52337248, Log Avg loss: 0.53304158, Global Avg Loss: 1.41738454, Time: 0.0068 Steps: 72150, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001903, Sample Num: 30448, Cur Loss: 0.14200020, Cur Avg Loss: 0.52230373, Log Avg loss: 0.31998963, Global Avg Loss: 1.41723247, Time: 0.0121 Steps: 72160, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001913, Sample Num: 30608, Cur Loss: 0.22122565, Cur Avg Loss: 0.52343734, Log Avg loss: 0.73916248, Global Avg Loss: 1.41713851, Time: 0.0068 Steps: 72170, Updated lr: 0.000033 Training, Epoch: 0034, Batch: 001923, Sample Num: 30768, Cur Loss: 0.68110245, Cur Avg Loss: 0.52314815, Log Avg loss: 0.46782570, Global Avg Loss: 1.41700699, Time: 0.0066 Steps: 72180, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001933, Sample Num: 30928, Cur Loss: 0.44616365, Cur Avg Loss: 0.52307285, Log Avg loss: 0.50859269, Global Avg Loss: 1.41688115, Time: 0.0117 Steps: 72190, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001943, Sample Num: 31088, Cur Loss: 0.18645522, Cur Avg Loss: 0.52270752, Log Avg loss: 0.45208979, Global Avg Loss: 1.41674753, Time: 0.0111 Steps: 72200, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001953, Sample Num: 31248, Cur Loss: 0.90862930, Cur Avg Loss: 0.52222532, Log Avg loss: 0.42853344, Global Avg Loss: 1.41661067, Time: 0.0114 Steps: 72210, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001963, Sample Num: 31408, Cur Loss: 0.62475193, Cur Avg Loss: 0.52297322, Log Avg loss: 0.66903893, Global Avg Loss: 1.41650716, Time: 0.0068 Steps: 72220, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001973, Sample Num: 31568, Cur Loss: 0.53277373, Cur Avg Loss: 0.52275858, Log Avg loss: 0.48062440, Global Avg Loss: 1.41637759, Time: 0.0119 Steps: 72230, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001983, Sample Num: 31728, Cur Loss: 0.26519370, Cur Avg Loss: 0.52303092, Log Avg loss: 0.57676289, Global Avg Loss: 1.41626136, Time: 0.0144 Steps: 72240, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 001993, Sample Num: 31888, Cur Loss: 0.25905377, Cur Avg Loss: 0.52268693, Log Avg loss: 0.45447455, Global Avg Loss: 1.41612825, Time: 0.0069 Steps: 72250, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002003, Sample Num: 32048, Cur Loss: 1.82072675, Cur Avg Loss: 0.52273276, Log Avg loss: 0.53186697, Global Avg Loss: 1.41600587, Time: 0.0067 Steps: 72260, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002013, Sample Num: 32208, Cur Loss: 0.29993021, Cur Avg Loss: 0.52312161, Log Avg loss: 0.60100797, Global Avg Loss: 1.41589310, Time: 0.0070 Steps: 72270, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002023, Sample Num: 32368, Cur Loss: 0.19215596, Cur Avg Loss: 0.52248269, Log Avg loss: 0.39386846, Global Avg Loss: 1.41575170, Time: 0.0123 Steps: 72280, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002033, Sample Num: 32528, Cur Loss: 0.63067347, Cur Avg Loss: 0.52302196, Log Avg loss: 0.63211544, Global Avg Loss: 1.41564330, Time: 0.0161 Steps: 72290, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002043, Sample Num: 32688, Cur Loss: 0.80107439, Cur Avg Loss: 0.52321210, Log Avg loss: 0.56186798, Global Avg Loss: 1.41552521, Time: 0.0116 Steps: 72300, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002053, Sample Num: 32848, Cur Loss: 0.25150722, Cur Avg Loss: 0.52291656, Log Avg loss: 0.46253825, Global Avg Loss: 1.41539342, Time: 0.0135 Steps: 72310, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002063, Sample Num: 33008, Cur Loss: 0.43979165, Cur Avg Loss: 0.52290000, Log Avg loss: 0.51950026, Global Avg Loss: 1.41526954, Time: 0.0233 Steps: 72320, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002073, Sample Num: 33168, Cur Loss: 0.25021243, Cur Avg Loss: 0.52233635, Log Avg loss: 0.40605482, Global Avg Loss: 1.41513001, Time: 0.0116 Steps: 72330, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002083, Sample Num: 33328, Cur Loss: 0.46029729, Cur Avg Loss: 0.52220801, Log Avg loss: 0.49560285, Global Avg Loss: 1.41500290, Time: 0.0153 Steps: 72340, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002093, Sample Num: 33488, Cur Loss: 0.47273511, Cur Avg Loss: 0.52287678, Log Avg loss: 0.66218086, Global Avg Loss: 1.41489885, Time: 0.0113 Steps: 72350, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002103, Sample Num: 33648, Cur Loss: 0.68983227, Cur Avg Loss: 0.52257983, Log Avg loss: 0.46042849, Global Avg Loss: 1.41476694, Time: 0.0119 Steps: 72360, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002113, Sample Num: 33808, Cur Loss: 0.32571563, Cur Avg Loss: 0.52203799, Log Avg loss: 0.40808879, Global Avg Loss: 1.41462784, Time: 0.0121 Steps: 72370, Updated lr: 0.000032 Training, Epoch: 0034, Batch: 002123, Sample Num: 33968, Cur Loss: 0.81046212, Cur Avg Loss: 0.52180210, Log Avg loss: 0.47195915, Global Avg Loss: 1.41449760, Time: 0.0134 Steps: 72380, Updated lr: 0.000032 ***** Running evaluation checkpoint-72386 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-72386 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.868383, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.611176, "eval_total_loss": 429.656818, "eval_mae": 0.565105, "eval_mse": 0.611387, "eval_r2": 0.611362, "eval_sp_statistic": 0.737744, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.784698, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.883369, "test_total_loss": 443.451446, "test_mae": 0.662676, "test_mse": 0.883608, "test_r2": 0.429712, "test_sp_statistic": 0.581204, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.666474, "test_ps_pvalue": 0.0, "lr": 3.2303461356092936e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.4144160524335398, "train_cur_epoch_loss": 1110.369682006538, "train_cur_epoch_avg_loss": 0.5215451770815115, "train_cur_epoch_time": 21.868383407592773, "train_cur_epoch_avg_time": 0.01027166905006706, "epoch": 34, "step": 72386} ################################################## Training, Epoch: 0035, Batch: 000004, Sample Num: 64, Cur Loss: 1.55886054, Cur Avg Loss: 1.18220552, Log Avg loss: 0.73126463, Global Avg Loss: 1.41440322, Time: 0.0113 Steps: 72390, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000014, Sample Num: 224, Cur Loss: 0.22480884, Cur Avg Loss: 0.59168357, Log Avg loss: 0.35547479, Global Avg Loss: 1.41425696, Time: 0.0117 Steps: 72400, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000024, Sample Num: 384, Cur Loss: 0.33873212, Cur Avg Loss: 0.53874866, Log Avg loss: 0.46463979, Global Avg Loss: 1.41412582, Time: 0.0072 Steps: 72410, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000034, Sample Num: 544, Cur Loss: 0.26660690, Cur Avg Loss: 0.51366636, Log Avg loss: 0.45346885, Global Avg Loss: 1.41399317, Time: 0.0088 Steps: 72420, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000044, Sample Num: 704, Cur Loss: 0.53587490, Cur Avg Loss: 0.52101859, Log Avg loss: 0.54601614, Global Avg Loss: 1.41387333, Time: 0.0112 Steps: 72430, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000054, Sample Num: 864, Cur Loss: 0.29602090, Cur Avg Loss: 0.51212505, Log Avg loss: 0.47299352, Global Avg Loss: 1.41374344, Time: 0.0124 Steps: 72440, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000064, Sample Num: 1024, Cur Loss: 0.32907838, Cur Avg Loss: 0.51938769, Log Avg loss: 0.55860591, Global Avg Loss: 1.41362541, Time: 0.0120 Steps: 72450, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000074, Sample Num: 1184, Cur Loss: 0.98918194, Cur Avg Loss: 0.52135649, Log Avg loss: 0.53395685, Global Avg Loss: 1.41350401, Time: 0.0075 Steps: 72460, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000084, Sample Num: 1344, Cur Loss: 0.52413720, Cur Avg Loss: 0.54289880, Log Avg loss: 0.70231187, Global Avg Loss: 1.41340588, Time: 0.0194 Steps: 72470, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000094, Sample Num: 1504, Cur Loss: 0.16807327, Cur Avg Loss: 0.52126377, Log Avg loss: 0.33952954, Global Avg Loss: 1.41325771, Time: 0.0121 Steps: 72480, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000104, Sample Num: 1664, Cur Loss: 0.09775050, Cur Avg Loss: 0.51438307, Log Avg loss: 0.44970451, Global Avg Loss: 1.41312479, Time: 0.0105 Steps: 72490, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000114, Sample Num: 1824, Cur Loss: 0.29216355, Cur Avg Loss: 0.50048991, Log Avg loss: 0.35600095, Global Avg Loss: 1.41297898, Time: 0.0105 Steps: 72500, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000124, Sample Num: 1984, Cur Loss: 0.30755740, Cur Avg Loss: 0.48990315, Log Avg loss: 0.36921414, Global Avg Loss: 1.41283503, Time: 0.0087 Steps: 72510, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000134, Sample Num: 2144, Cur Loss: 0.83821344, Cur Avg Loss: 0.49224317, Log Avg loss: 0.52125939, Global Avg Loss: 1.41271209, Time: 0.0068 Steps: 72520, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000144, Sample Num: 2304, Cur Loss: 0.61007893, Cur Avg Loss: 0.49647662, Log Avg loss: 0.55320480, Global Avg Loss: 1.41259359, Time: 0.0166 Steps: 72530, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000154, Sample Num: 2464, Cur Loss: 0.11909534, Cur Avg Loss: 0.49710775, Log Avg loss: 0.50619615, Global Avg Loss: 1.41246864, Time: 0.0108 Steps: 72540, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000164, Sample Num: 2624, Cur Loss: 0.14958513, Cur Avg Loss: 0.48603017, Log Avg loss: 0.31543545, Global Avg Loss: 1.41231743, Time: 0.0108 Steps: 72550, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000174, Sample Num: 2784, Cur Loss: 0.46042874, Cur Avg Loss: 0.49305354, Log Avg loss: 0.60823680, Global Avg Loss: 1.41220661, Time: 0.0083 Steps: 72560, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000184, Sample Num: 2944, Cur Loss: 1.25022435, Cur Avg Loss: 0.50630639, Log Avg loss: 0.73690595, Global Avg Loss: 1.41211356, Time: 0.0112 Steps: 72570, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000194, Sample Num: 3104, Cur Loss: 0.29075775, Cur Avg Loss: 0.51321861, Log Avg loss: 0.64040350, Global Avg Loss: 1.41200723, Time: 0.0067 Steps: 72580, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000204, Sample Num: 3264, Cur Loss: 0.20611301, Cur Avg Loss: 0.50941583, Log Avg loss: 0.43564176, Global Avg Loss: 1.41187273, Time: 0.0068 Steps: 72590, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000214, Sample Num: 3424, Cur Loss: 0.15972583, Cur Avg Loss: 0.50838421, Log Avg loss: 0.48733916, Global Avg Loss: 1.41174538, Time: 0.0119 Steps: 72600, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000224, Sample Num: 3584, Cur Loss: 1.10693538, Cur Avg Loss: 0.51238101, Log Avg loss: 0.59791258, Global Avg Loss: 1.41163330, Time: 0.0069 Steps: 72610, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000234, Sample Num: 3744, Cur Loss: 0.29293311, Cur Avg Loss: 0.50638771, Log Avg loss: 0.37213769, Global Avg Loss: 1.41149016, Time: 0.0110 Steps: 72620, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000244, Sample Num: 3904, Cur Loss: 0.61494392, Cur Avg Loss: 0.50276506, Log Avg loss: 0.41799509, Global Avg Loss: 1.41135337, Time: 0.0068 Steps: 72630, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000254, Sample Num: 4064, Cur Loss: 0.49228579, Cur Avg Loss: 0.50031127, Log Avg loss: 0.44043876, Global Avg Loss: 1.41121971, Time: 0.0090 Steps: 72640, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000264, Sample Num: 4224, Cur Loss: 0.39429992, Cur Avg Loss: 0.50148084, Log Avg loss: 0.53118793, Global Avg Loss: 1.41109857, Time: 0.0116 Steps: 72650, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000274, Sample Num: 4384, Cur Loss: 0.66146982, Cur Avg Loss: 0.49778555, Log Avg loss: 0.40023011, Global Avg Loss: 1.41095945, Time: 0.0082 Steps: 72660, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000284, Sample Num: 4544, Cur Loss: 0.60138857, Cur Avg Loss: 0.49824153, Log Avg loss: 0.51073514, Global Avg Loss: 1.41083557, Time: 0.0068 Steps: 72670, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000294, Sample Num: 4704, Cur Loss: 0.44446468, Cur Avg Loss: 0.49669315, Log Avg loss: 0.45271915, Global Avg Loss: 1.41070374, Time: 0.0071 Steps: 72680, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000304, Sample Num: 4864, Cur Loss: 0.42977580, Cur Avg Loss: 0.50322859, Log Avg loss: 0.69537062, Global Avg Loss: 1.41060534, Time: 0.0104 Steps: 72690, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000314, Sample Num: 5024, Cur Loss: 0.41936770, Cur Avg Loss: 0.50109996, Log Avg loss: 0.43638965, Global Avg Loss: 1.41047133, Time: 0.0085 Steps: 72700, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000324, Sample Num: 5184, Cur Loss: 0.54905689, Cur Avg Loss: 0.50204710, Log Avg loss: 0.53178712, Global Avg Loss: 1.41035048, Time: 0.0066 Steps: 72710, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000334, Sample Num: 5344, Cur Loss: 0.70104057, Cur Avg Loss: 0.50881831, Log Avg loss: 0.72820564, Global Avg Loss: 1.41025668, Time: 0.0072 Steps: 72720, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000344, Sample Num: 5504, Cur Loss: 0.59847063, Cur Avg Loss: 0.50872114, Log Avg loss: 0.50547568, Global Avg Loss: 1.41013228, Time: 0.0072 Steps: 72730, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000354, Sample Num: 5664, Cur Loss: 0.30354929, Cur Avg Loss: 0.51143085, Log Avg loss: 0.60464476, Global Avg Loss: 1.41002154, Time: 0.0067 Steps: 72740, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000364, Sample Num: 5824, Cur Loss: 0.47320983, Cur Avg Loss: 0.51149465, Log Avg loss: 0.51375340, Global Avg Loss: 1.40989834, Time: 0.0119 Steps: 72750, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000374, Sample Num: 5984, Cur Loss: 0.10918841, Cur Avg Loss: 0.51190819, Log Avg loss: 0.52696103, Global Avg Loss: 1.40977699, Time: 0.0124 Steps: 72760, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000384, Sample Num: 6144, Cur Loss: 0.23592159, Cur Avg Loss: 0.51112923, Log Avg loss: 0.48199587, Global Avg Loss: 1.40964950, Time: 0.0107 Steps: 72770, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000394, Sample Num: 6304, Cur Loss: 0.25586390, Cur Avg Loss: 0.51175760, Log Avg loss: 0.53588704, Global Avg Loss: 1.40952944, Time: 0.0106 Steps: 72780, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000404, Sample Num: 6464, Cur Loss: 0.38142213, Cur Avg Loss: 0.51107331, Log Avg loss: 0.48411229, Global Avg Loss: 1.40940231, Time: 0.0067 Steps: 72790, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000414, Sample Num: 6624, Cur Loss: 0.36397111, Cur Avg Loss: 0.51083916, Log Avg loss: 0.50137960, Global Avg Loss: 1.40927758, Time: 0.0071 Steps: 72800, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000424, Sample Num: 6784, Cur Loss: 0.62696135, Cur Avg Loss: 0.51376986, Log Avg loss: 0.63510088, Global Avg Loss: 1.40917125, Time: 0.0108 Steps: 72810, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000434, Sample Num: 6944, Cur Loss: 0.36727110, Cur Avg Loss: 0.51166101, Log Avg loss: 0.42224559, Global Avg Loss: 1.40903572, Time: 0.0110 Steps: 72820, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000444, Sample Num: 7104, Cur Loss: 0.72928071, Cur Avg Loss: 0.51252508, Log Avg loss: 0.55002592, Global Avg Loss: 1.40891777, Time: 0.0109 Steps: 72830, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000454, Sample Num: 7264, Cur Loss: 1.01410937, Cur Avg Loss: 0.51336536, Log Avg loss: 0.55067394, Global Avg Loss: 1.40879995, Time: 0.0067 Steps: 72840, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000464, Sample Num: 7424, Cur Loss: 0.63385212, Cur Avg Loss: 0.51286672, Log Avg loss: 0.49022825, Global Avg Loss: 1.40867386, Time: 0.0067 Steps: 72850, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000474, Sample Num: 7584, Cur Loss: 0.20523691, Cur Avg Loss: 0.51167255, Log Avg loss: 0.45626297, Global Avg Loss: 1.40854314, Time: 0.0082 Steps: 72860, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000484, Sample Num: 7744, Cur Loss: 0.19379929, Cur Avg Loss: 0.51017058, Log Avg loss: 0.43897750, Global Avg Loss: 1.40841009, Time: 0.0067 Steps: 72870, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000494, Sample Num: 7904, Cur Loss: 0.17957744, Cur Avg Loss: 0.50842947, Log Avg loss: 0.42415948, Global Avg Loss: 1.40827503, Time: 0.0067 Steps: 72880, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000504, Sample Num: 8064, Cur Loss: 0.11340915, Cur Avg Loss: 0.50923708, Log Avg loss: 0.54913317, Global Avg Loss: 1.40815717, Time: 0.0081 Steps: 72890, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000514, Sample Num: 8224, Cur Loss: 0.28389028, Cur Avg Loss: 0.50854580, Log Avg loss: 0.47370547, Global Avg Loss: 1.40802898, Time: 0.0073 Steps: 72900, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000524, Sample Num: 8384, Cur Loss: 0.56880301, Cur Avg Loss: 0.50639571, Log Avg loss: 0.39588101, Global Avg Loss: 1.40789016, Time: 0.0071 Steps: 72910, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000534, Sample Num: 8544, Cur Loss: 0.78229356, Cur Avg Loss: 0.50915038, Log Avg loss: 0.65349489, Global Avg Loss: 1.40778671, Time: 0.0067 Steps: 72920, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000544, Sample Num: 8704, Cur Loss: 0.38310596, Cur Avg Loss: 0.50887054, Log Avg loss: 0.49392710, Global Avg Loss: 1.40766140, Time: 0.0129 Steps: 72930, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000554, Sample Num: 8864, Cur Loss: 0.68547648, Cur Avg Loss: 0.50773117, Log Avg loss: 0.44574956, Global Avg Loss: 1.40752952, Time: 0.0223 Steps: 72940, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000564, Sample Num: 9024, Cur Loss: 0.75707734, Cur Avg Loss: 0.50963332, Log Avg loss: 0.61501247, Global Avg Loss: 1.40742089, Time: 0.0120 Steps: 72950, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000574, Sample Num: 9184, Cur Loss: 0.12699360, Cur Avg Loss: 0.50802086, Log Avg loss: 0.41707802, Global Avg Loss: 1.40728515, Time: 0.0120 Steps: 72960, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000584, Sample Num: 9344, Cur Loss: 0.64016110, Cur Avg Loss: 0.50862152, Log Avg loss: 0.54309926, Global Avg Loss: 1.40716672, Time: 0.0166 Steps: 72970, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000594, Sample Num: 9504, Cur Loss: 0.14829814, Cur Avg Loss: 0.50839879, Log Avg loss: 0.49539139, Global Avg Loss: 1.40704178, Time: 0.0067 Steps: 72980, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000604, Sample Num: 9664, Cur Loss: 0.63757586, Cur Avg Loss: 0.51041255, Log Avg loss: 0.63003015, Global Avg Loss: 1.40693533, Time: 0.0114 Steps: 72990, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000614, Sample Num: 9824, Cur Loss: 0.30334872, Cur Avg Loss: 0.50958584, Log Avg loss: 0.45965205, Global Avg Loss: 1.40680556, Time: 0.0111 Steps: 73000, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000624, Sample Num: 9984, Cur Loss: 0.73978978, Cur Avg Loss: 0.50880500, Log Avg loss: 0.46086154, Global Avg Loss: 1.40667600, Time: 0.0134 Steps: 73010, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000634, Sample Num: 10144, Cur Loss: 1.13205957, Cur Avg Loss: 0.50826288, Log Avg loss: 0.47443488, Global Avg Loss: 1.40654833, Time: 0.0118 Steps: 73020, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000644, Sample Num: 10304, Cur Loss: 1.00546086, Cur Avg Loss: 0.51292851, Log Avg loss: 0.80872955, Global Avg Loss: 1.40646647, Time: 0.0067 Steps: 73030, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000654, Sample Num: 10464, Cur Loss: 0.71384525, Cur Avg Loss: 0.51202790, Log Avg loss: 0.45402844, Global Avg Loss: 1.40633607, Time: 0.0070 Steps: 73040, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000664, Sample Num: 10624, Cur Loss: 0.39457625, Cur Avg Loss: 0.51164488, Log Avg loss: 0.48659548, Global Avg Loss: 1.40621017, Time: 0.0070 Steps: 73050, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000674, Sample Num: 10784, Cur Loss: 0.45926231, Cur Avg Loss: 0.51078896, Log Avg loss: 0.45395588, Global Avg Loss: 1.40607983, Time: 0.0076 Steps: 73060, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000684, Sample Num: 10944, Cur Loss: 1.32058620, Cur Avg Loss: 0.51083042, Log Avg loss: 0.51362451, Global Avg Loss: 1.40595769, Time: 0.0133 Steps: 73070, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000694, Sample Num: 11104, Cur Loss: 0.53438580, Cur Avg Loss: 0.50809876, Log Avg loss: 0.32125369, Global Avg Loss: 1.40580926, Time: 0.0115 Steps: 73080, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000704, Sample Num: 11264, Cur Loss: 0.72368163, Cur Avg Loss: 0.50745315, Log Avg loss: 0.46264719, Global Avg Loss: 1.40568022, Time: 0.0116 Steps: 73090, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000714, Sample Num: 11424, Cur Loss: 0.89015603, Cur Avg Loss: 0.50932733, Log Avg loss: 0.64127012, Global Avg Loss: 1.40557565, Time: 0.0067 Steps: 73100, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000724, Sample Num: 11584, Cur Loss: 0.74542201, Cur Avg Loss: 0.50869224, Log Avg loss: 0.46334687, Global Avg Loss: 1.40544677, Time: 0.0066 Steps: 73110, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000734, Sample Num: 11744, Cur Loss: 0.42151707, Cur Avg Loss: 0.50760196, Log Avg loss: 0.42866525, Global Avg Loss: 1.40531319, Time: 0.0064 Steps: 73120, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000744, Sample Num: 11904, Cur Loss: 0.39525649, Cur Avg Loss: 0.50593206, Log Avg loss: 0.38336146, Global Avg Loss: 1.40517344, Time: 0.0083 Steps: 73130, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000754, Sample Num: 12064, Cur Loss: 0.43994725, Cur Avg Loss: 0.50533789, Log Avg loss: 0.46113184, Global Avg Loss: 1.40504437, Time: 0.0070 Steps: 73140, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000764, Sample Num: 12224, Cur Loss: 0.30230826, Cur Avg Loss: 0.50606645, Log Avg loss: 0.56099971, Global Avg Loss: 1.40492898, Time: 0.0070 Steps: 73150, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000774, Sample Num: 12384, Cur Loss: 0.46554542, Cur Avg Loss: 0.50510693, Log Avg loss: 0.43179983, Global Avg Loss: 1.40479597, Time: 0.0087 Steps: 73160, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000784, Sample Num: 12544, Cur Loss: 0.79083472, Cur Avg Loss: 0.50633193, Log Avg loss: 0.60114700, Global Avg Loss: 1.40468614, Time: 0.0120 Steps: 73170, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000794, Sample Num: 12704, Cur Loss: 0.36480331, Cur Avg Loss: 0.50677455, Log Avg loss: 0.54147612, Global Avg Loss: 1.40456818, Time: 0.0072 Steps: 73180, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000804, Sample Num: 12864, Cur Loss: 0.60469508, Cur Avg Loss: 0.50700455, Log Avg loss: 0.52526629, Global Avg Loss: 1.40444804, Time: 0.0234 Steps: 73190, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000814, Sample Num: 13024, Cur Loss: 0.05698623, Cur Avg Loss: 0.50629483, Log Avg loss: 0.44923318, Global Avg Loss: 1.40431755, Time: 0.0077 Steps: 73200, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000824, Sample Num: 13184, Cur Loss: 0.43021029, Cur Avg Loss: 0.50427704, Log Avg loss: 0.34002935, Global Avg Loss: 1.40417217, Time: 0.0111 Steps: 73210, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000834, Sample Num: 13344, Cur Loss: 0.67965662, Cur Avg Loss: 0.50453163, Log Avg loss: 0.52550928, Global Avg Loss: 1.40405217, Time: 0.0113 Steps: 73220, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000844, Sample Num: 13504, Cur Loss: 0.37147576, Cur Avg Loss: 0.50470196, Log Avg loss: 0.51890796, Global Avg Loss: 1.40393130, Time: 0.0103 Steps: 73230, Updated lr: 0.000032 Training, Epoch: 0035, Batch: 000854, Sample Num: 13664, Cur Loss: 1.22279990, Cur Avg Loss: 0.50707812, Log Avg loss: 0.70762576, Global Avg Loss: 1.40383622, Time: 0.0086 Steps: 73240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000864, Sample Num: 13824, Cur Loss: 0.75713575, Cur Avg Loss: 0.50710080, Log Avg loss: 0.50903780, Global Avg Loss: 1.40371407, Time: 0.0072 Steps: 73250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000874, Sample Num: 13984, Cur Loss: 0.74087691, Cur Avg Loss: 0.50752869, Log Avg loss: 0.54449834, Global Avg Loss: 1.40359678, Time: 0.0067 Steps: 73260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000884, Sample Num: 14144, Cur Loss: 0.73403692, Cur Avg Loss: 0.50823653, Log Avg loss: 0.57010190, Global Avg Loss: 1.40348303, Time: 0.0109 Steps: 73270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000894, Sample Num: 14304, Cur Loss: 0.24141799, Cur Avg Loss: 0.50788394, Log Avg loss: 0.47671429, Global Avg Loss: 1.40335656, Time: 0.0109 Steps: 73280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000904, Sample Num: 14464, Cur Loss: 0.66404945, Cur Avg Loss: 0.50900060, Log Avg loss: 0.60883041, Global Avg Loss: 1.40324815, Time: 0.0110 Steps: 73290, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000914, Sample Num: 14624, Cur Loss: 0.28352207, Cur Avg Loss: 0.51088786, Log Avg loss: 0.68149601, Global Avg Loss: 1.40314968, Time: 0.0067 Steps: 73300, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000924, Sample Num: 14784, Cur Loss: 0.36740232, Cur Avg Loss: 0.50995509, Log Avg loss: 0.42470014, Global Avg Loss: 1.40301622, Time: 0.0066 Steps: 73310, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000934, Sample Num: 14944, Cur Loss: 0.70127237, Cur Avg Loss: 0.51058772, Log Avg loss: 0.56904270, Global Avg Loss: 1.40290247, Time: 0.0065 Steps: 73320, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000944, Sample Num: 15104, Cur Loss: 0.30722553, Cur Avg Loss: 0.51069452, Log Avg loss: 0.52066940, Global Avg Loss: 1.40278216, Time: 0.0064 Steps: 73330, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000954, Sample Num: 15264, Cur Loss: 0.39086103, Cur Avg Loss: 0.51091377, Log Avg loss: 0.53161154, Global Avg Loss: 1.40266338, Time: 0.0140 Steps: 73340, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000964, Sample Num: 15424, Cur Loss: 1.12139738, Cur Avg Loss: 0.51170332, Log Avg loss: 0.58702565, Global Avg Loss: 1.40255218, Time: 0.0117 Steps: 73350, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000974, Sample Num: 15584, Cur Loss: 0.48706287, Cur Avg Loss: 0.51273200, Log Avg loss: 0.61189687, Global Avg Loss: 1.40244440, Time: 0.0115 Steps: 73360, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000984, Sample Num: 15744, Cur Loss: 1.25662100, Cur Avg Loss: 0.51416467, Log Avg loss: 0.65370740, Global Avg Loss: 1.40234235, Time: 0.0207 Steps: 73370, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 000994, Sample Num: 15904, Cur Loss: 0.58055067, Cur Avg Loss: 0.51394206, Log Avg loss: 0.49203664, Global Avg Loss: 1.40221830, Time: 0.0117 Steps: 73380, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001004, Sample Num: 16064, Cur Loss: 0.55290759, Cur Avg Loss: 0.51427043, Log Avg loss: 0.54691020, Global Avg Loss: 1.40210176, Time: 0.0096 Steps: 73390, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001014, Sample Num: 16224, Cur Loss: 0.88064098, Cur Avg Loss: 0.51531359, Log Avg loss: 0.62004769, Global Avg Loss: 1.40199521, Time: 0.0121 Steps: 73400, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001024, Sample Num: 16384, Cur Loss: 0.53120589, Cur Avg Loss: 0.51432565, Log Avg loss: 0.41414762, Global Avg Loss: 1.40186064, Time: 0.0124 Steps: 73410, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001034, Sample Num: 16544, Cur Loss: 0.16233632, Cur Avg Loss: 0.51416536, Log Avg loss: 0.49775189, Global Avg Loss: 1.40173750, Time: 0.0066 Steps: 73420, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001044, Sample Num: 16704, Cur Loss: 0.47344953, Cur Avg Loss: 0.51417273, Log Avg loss: 0.51493469, Global Avg Loss: 1.40161673, Time: 0.0114 Steps: 73430, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001054, Sample Num: 16864, Cur Loss: 0.11850238, Cur Avg Loss: 0.51444560, Log Avg loss: 0.54293377, Global Avg Loss: 1.40149981, Time: 0.0068 Steps: 73440, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001064, Sample Num: 17024, Cur Loss: 0.28875038, Cur Avg Loss: 0.51605054, Log Avg loss: 0.68521100, Global Avg Loss: 1.40140229, Time: 0.0091 Steps: 73450, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001074, Sample Num: 17184, Cur Loss: 0.34320551, Cur Avg Loss: 0.51501740, Log Avg loss: 0.40509183, Global Avg Loss: 1.40126666, Time: 0.0119 Steps: 73460, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001084, Sample Num: 17344, Cur Loss: 0.30277780, Cur Avg Loss: 0.51425444, Log Avg loss: 0.43231235, Global Avg Loss: 1.40113478, Time: 0.0114 Steps: 73470, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001094, Sample Num: 17504, Cur Loss: 0.48502916, Cur Avg Loss: 0.51352885, Log Avg loss: 0.43487445, Global Avg Loss: 1.40100328, Time: 0.0111 Steps: 73480, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001104, Sample Num: 17664, Cur Loss: 1.29709220, Cur Avg Loss: 0.51403170, Log Avg loss: 0.56904410, Global Avg Loss: 1.40089007, Time: 0.0117 Steps: 73490, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001114, Sample Num: 17824, Cur Loss: 0.28217745, Cur Avg Loss: 0.51444881, Log Avg loss: 0.56049674, Global Avg Loss: 1.40077573, Time: 0.0120 Steps: 73500, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001124, Sample Num: 17984, Cur Loss: 0.78938019, Cur Avg Loss: 0.51541813, Log Avg loss: 0.62340078, Global Avg Loss: 1.40066998, Time: 0.0116 Steps: 73510, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001134, Sample Num: 18144, Cur Loss: 0.52101827, Cur Avg Loss: 0.51652583, Log Avg loss: 0.64103156, Global Avg Loss: 1.40056666, Time: 0.0089 Steps: 73520, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001144, Sample Num: 18304, Cur Loss: 1.00849652, Cur Avg Loss: 0.51693150, Log Avg loss: 0.56293393, Global Avg Loss: 1.40045274, Time: 0.0075 Steps: 73530, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001154, Sample Num: 18464, Cur Loss: 0.37683868, Cur Avg Loss: 0.51608233, Log Avg loss: 0.41893755, Global Avg Loss: 1.40031927, Time: 0.0158 Steps: 73540, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001164, Sample Num: 18624, Cur Loss: 0.44995469, Cur Avg Loss: 0.51569061, Log Avg loss: 0.47048657, Global Avg Loss: 1.40019285, Time: 0.0067 Steps: 73550, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001174, Sample Num: 18784, Cur Loss: 0.75322992, Cur Avg Loss: 0.51684847, Log Avg loss: 0.65162335, Global Avg Loss: 1.40009109, Time: 0.0083 Steps: 73560, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001184, Sample Num: 18944, Cur Loss: 0.21292345, Cur Avg Loss: 0.51623622, Log Avg loss: 0.44435795, Global Avg Loss: 1.39996118, Time: 0.0113 Steps: 73570, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001194, Sample Num: 19104, Cur Loss: 0.29962039, Cur Avg Loss: 0.51629526, Log Avg loss: 0.52328504, Global Avg Loss: 1.39984203, Time: 0.0111 Steps: 73580, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001204, Sample Num: 19264, Cur Loss: 0.34654999, Cur Avg Loss: 0.51695726, Log Avg loss: 0.59600086, Global Avg Loss: 1.39973280, Time: 0.0240 Steps: 73590, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001214, Sample Num: 19424, Cur Loss: 0.19703984, Cur Avg Loss: 0.51570464, Log Avg loss: 0.36488878, Global Avg Loss: 1.39959220, Time: 0.0123 Steps: 73600, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001224, Sample Num: 19584, Cur Loss: 1.57069385, Cur Avg Loss: 0.51598402, Log Avg loss: 0.54990093, Global Avg Loss: 1.39947677, Time: 0.0107 Steps: 73610, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001234, Sample Num: 19744, Cur Loss: 0.21802405, Cur Avg Loss: 0.51525249, Log Avg loss: 0.42571266, Global Avg Loss: 1.39934450, Time: 0.0106 Steps: 73620, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001244, Sample Num: 19904, Cur Loss: 0.35353583, Cur Avg Loss: 0.51588715, Log Avg loss: 0.59420452, Global Avg Loss: 1.39923515, Time: 0.0230 Steps: 73630, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001254, Sample Num: 20064, Cur Loss: 0.32788742, Cur Avg Loss: 0.51495840, Log Avg loss: 0.39942231, Global Avg Loss: 1.39909938, Time: 0.0115 Steps: 73640, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001264, Sample Num: 20224, Cur Loss: 0.42328623, Cur Avg Loss: 0.51496443, Log Avg loss: 0.51572072, Global Avg Loss: 1.39897944, Time: 0.0067 Steps: 73650, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001274, Sample Num: 20384, Cur Loss: 0.26142260, Cur Avg Loss: 0.51580242, Log Avg loss: 0.62172382, Global Avg Loss: 1.39887392, Time: 0.0123 Steps: 73660, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001284, Sample Num: 20544, Cur Loss: 0.40362877, Cur Avg Loss: 0.51575704, Log Avg loss: 0.50997520, Global Avg Loss: 1.39875326, Time: 0.0111 Steps: 73670, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001294, Sample Num: 20704, Cur Loss: 1.39594865, Cur Avg Loss: 0.51625322, Log Avg loss: 0.57996261, Global Avg Loss: 1.39864213, Time: 0.0135 Steps: 73680, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001304, Sample Num: 20864, Cur Loss: 0.87408429, Cur Avg Loss: 0.51546928, Log Avg loss: 0.41402768, Global Avg Loss: 1.39850851, Time: 0.0068 Steps: 73690, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001314, Sample Num: 21024, Cur Loss: 1.56302524, Cur Avg Loss: 0.51515301, Log Avg loss: 0.47391237, Global Avg Loss: 1.39838306, Time: 0.0068 Steps: 73700, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001324, Sample Num: 21184, Cur Loss: 0.41729000, Cur Avg Loss: 0.51533355, Log Avg loss: 0.53905600, Global Avg Loss: 1.39826648, Time: 0.0089 Steps: 73710, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001334, Sample Num: 21344, Cur Loss: 0.78534931, Cur Avg Loss: 0.51493881, Log Avg loss: 0.46267508, Global Avg Loss: 1.39813957, Time: 0.0067 Steps: 73720, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001344, Sample Num: 21504, Cur Loss: 0.27100790, Cur Avg Loss: 0.51534883, Log Avg loss: 0.57004544, Global Avg Loss: 1.39802725, Time: 0.0072 Steps: 73730, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001354, Sample Num: 21664, Cur Loss: 0.68259788, Cur Avg Loss: 0.51650335, Log Avg loss: 0.67167108, Global Avg Loss: 1.39792875, Time: 0.0066 Steps: 73740, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001364, Sample Num: 21824, Cur Loss: 0.43406427, Cur Avg Loss: 0.51723475, Log Avg loss: 0.61626618, Global Avg Loss: 1.39782276, Time: 0.0115 Steps: 73750, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001374, Sample Num: 21984, Cur Loss: 0.46436146, Cur Avg Loss: 0.51645179, Log Avg loss: 0.40965614, Global Avg Loss: 1.39768879, Time: 0.0068 Steps: 73760, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001384, Sample Num: 22144, Cur Loss: 0.64711171, Cur Avg Loss: 0.51636342, Log Avg loss: 0.50422110, Global Avg Loss: 1.39756767, Time: 0.0068 Steps: 73770, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001394, Sample Num: 22304, Cur Loss: 0.84795201, Cur Avg Loss: 0.51698248, Log Avg loss: 0.60266023, Global Avg Loss: 1.39745993, Time: 0.0124 Steps: 73780, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001404, Sample Num: 22464, Cur Loss: 0.57458705, Cur Avg Loss: 0.51628897, Log Avg loss: 0.41961411, Global Avg Loss: 1.39732742, Time: 0.0072 Steps: 73790, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001414, Sample Num: 22624, Cur Loss: 1.25840425, Cur Avg Loss: 0.51661892, Log Avg loss: 0.56294370, Global Avg Loss: 1.39721436, Time: 0.0112 Steps: 73800, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001424, Sample Num: 22784, Cur Loss: 0.28539094, Cur Avg Loss: 0.51617471, Log Avg loss: 0.45336384, Global Avg Loss: 1.39708648, Time: 0.0117 Steps: 73810, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001434, Sample Num: 22944, Cur Loss: 0.32879904, Cur Avg Loss: 0.51585544, Log Avg loss: 0.47039157, Global Avg Loss: 1.39696095, Time: 0.0114 Steps: 73820, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001444, Sample Num: 23104, Cur Loss: 1.20753312, Cur Avg Loss: 0.51645727, Log Avg loss: 0.60275973, Global Avg Loss: 1.39685337, Time: 0.0070 Steps: 73830, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001454, Sample Num: 23264, Cur Loss: 0.63909090, Cur Avg Loss: 0.51626163, Log Avg loss: 0.48801135, Global Avg Loss: 1.39673029, Time: 0.0113 Steps: 73840, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001464, Sample Num: 23424, Cur Loss: 0.37288043, Cur Avg Loss: 0.51595453, Log Avg loss: 0.47130152, Global Avg Loss: 1.39660498, Time: 0.0113 Steps: 73850, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001474, Sample Num: 23584, Cur Loss: 0.40670004, Cur Avg Loss: 0.51554257, Log Avg loss: 0.45523118, Global Avg Loss: 1.39647753, Time: 0.0064 Steps: 73860, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001484, Sample Num: 23744, Cur Loss: 0.61266422, Cur Avg Loss: 0.51476338, Log Avg loss: 0.39991126, Global Avg Loss: 1.39634262, Time: 0.0065 Steps: 73870, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001494, Sample Num: 23904, Cur Loss: 0.25638068, Cur Avg Loss: 0.51469384, Log Avg loss: 0.50437336, Global Avg Loss: 1.39622189, Time: 0.0067 Steps: 73880, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001504, Sample Num: 24064, Cur Loss: 0.76711011, Cur Avg Loss: 0.51496176, Log Avg loss: 0.55498965, Global Avg Loss: 1.39610804, Time: 0.0065 Steps: 73890, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001514, Sample Num: 24224, Cur Loss: 0.51851863, Cur Avg Loss: 0.51480328, Log Avg loss: 0.49096750, Global Avg Loss: 1.39598556, Time: 0.0074 Steps: 73900, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001524, Sample Num: 24384, Cur Loss: 0.61047876, Cur Avg Loss: 0.51472295, Log Avg loss: 0.50256213, Global Avg Loss: 1.39586468, Time: 0.0070 Steps: 73910, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001534, Sample Num: 24544, Cur Loss: 0.55681413, Cur Avg Loss: 0.51560574, Log Avg loss: 0.65014247, Global Avg Loss: 1.39576379, Time: 0.0077 Steps: 73920, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001544, Sample Num: 24704, Cur Loss: 0.29319996, Cur Avg Loss: 0.51611510, Log Avg loss: 0.59425107, Global Avg Loss: 1.39565538, Time: 0.0140 Steps: 73930, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001554, Sample Num: 24864, Cur Loss: 0.23615885, Cur Avg Loss: 0.51602697, Log Avg loss: 0.50242004, Global Avg Loss: 1.39553457, Time: 0.0123 Steps: 73940, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001564, Sample Num: 25024, Cur Loss: 0.17433536, Cur Avg Loss: 0.51518941, Log Avg loss: 0.38503264, Global Avg Loss: 1.39539793, Time: 0.0134 Steps: 73950, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001574, Sample Num: 25184, Cur Loss: 0.38011253, Cur Avg Loss: 0.51502646, Log Avg loss: 0.48953965, Global Avg Loss: 1.39527545, Time: 0.0222 Steps: 73960, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001584, Sample Num: 25344, Cur Loss: 1.87484145, Cur Avg Loss: 0.51565514, Log Avg loss: 0.61460956, Global Avg Loss: 1.39516991, Time: 0.0064 Steps: 73970, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001594, Sample Num: 25504, Cur Loss: 0.09665342, Cur Avg Loss: 0.51583562, Log Avg loss: 0.54442347, Global Avg Loss: 1.39505491, Time: 0.0066 Steps: 73980, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001604, Sample Num: 25664, Cur Loss: 0.31833857, Cur Avg Loss: 0.51572844, Log Avg loss: 0.49864423, Global Avg Loss: 1.39493376, Time: 0.0128 Steps: 73990, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001614, Sample Num: 25824, Cur Loss: 0.53225434, Cur Avg Loss: 0.51528376, Log Avg loss: 0.44395766, Global Avg Loss: 1.39480525, Time: 0.0065 Steps: 74000, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001624, Sample Num: 25984, Cur Loss: 0.71685958, Cur Avg Loss: 0.51549528, Log Avg loss: 0.54963389, Global Avg Loss: 1.39469105, Time: 0.0064 Steps: 74010, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001634, Sample Num: 26144, Cur Loss: 1.19591165, Cur Avg Loss: 0.51583790, Log Avg loss: 0.57148040, Global Avg Loss: 1.39457984, Time: 0.0133 Steps: 74020, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001644, Sample Num: 26304, Cur Loss: 0.42145106, Cur Avg Loss: 0.51473244, Log Avg loss: 0.33409951, Global Avg Loss: 1.39443659, Time: 0.0065 Steps: 74030, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001654, Sample Num: 26464, Cur Loss: 0.33675772, Cur Avg Loss: 0.51547473, Log Avg loss: 0.63750750, Global Avg Loss: 1.39433435, Time: 0.0067 Steps: 74040, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001664, Sample Num: 26624, Cur Loss: 0.72566998, Cur Avg Loss: 0.51650235, Log Avg loss: 0.68647058, Global Avg Loss: 1.39423876, Time: 0.0147 Steps: 74050, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001674, Sample Num: 26784, Cur Loss: 0.36874813, Cur Avg Loss: 0.51727400, Log Avg loss: 0.64567652, Global Avg Loss: 1.39413769, Time: 0.0066 Steps: 74060, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001684, Sample Num: 26944, Cur Loss: 0.36019808, Cur Avg Loss: 0.51668364, Log Avg loss: 0.41785762, Global Avg Loss: 1.39400588, Time: 0.0065 Steps: 74070, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001694, Sample Num: 27104, Cur Loss: 0.43199724, Cur Avg Loss: 0.51627442, Log Avg loss: 0.44736158, Global Avg Loss: 1.39387809, Time: 0.0068 Steps: 74080, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001704, Sample Num: 27264, Cur Loss: 0.25813350, Cur Avg Loss: 0.51567803, Log Avg loss: 0.41464915, Global Avg Loss: 1.39374593, Time: 0.0115 Steps: 74090, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001714, Sample Num: 27424, Cur Loss: 0.61332136, Cur Avg Loss: 0.51619172, Log Avg loss: 0.60372447, Global Avg Loss: 1.39363931, Time: 0.0065 Steps: 74100, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001724, Sample Num: 27584, Cur Loss: 0.88827884, Cur Avg Loss: 0.51704348, Log Avg loss: 0.66303557, Global Avg Loss: 1.39354073, Time: 0.0104 Steps: 74110, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001734, Sample Num: 27744, Cur Loss: 1.02985084, Cur Avg Loss: 0.51758157, Log Avg loss: 0.61034864, Global Avg Loss: 1.39343506, Time: 0.0066 Steps: 74120, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001744, Sample Num: 27904, Cur Loss: 0.37621945, Cur Avg Loss: 0.51872957, Log Avg loss: 0.71779182, Global Avg Loss: 1.39334392, Time: 0.0233 Steps: 74130, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001754, Sample Num: 28064, Cur Loss: 0.38610739, Cur Avg Loss: 0.51893200, Log Avg loss: 0.55423647, Global Avg Loss: 1.39323074, Time: 0.0066 Steps: 74140, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001764, Sample Num: 28224, Cur Loss: 0.27723482, Cur Avg Loss: 0.51891497, Log Avg loss: 0.51592852, Global Avg Loss: 1.39311243, Time: 0.0240 Steps: 74150, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001774, Sample Num: 28384, Cur Loss: 0.31441140, Cur Avg Loss: 0.51858560, Log Avg loss: 0.46048457, Global Avg Loss: 1.39298667, Time: 0.0066 Steps: 74160, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001784, Sample Num: 28544, Cur Loss: 0.82024735, Cur Avg Loss: 0.51810812, Log Avg loss: 0.43340363, Global Avg Loss: 1.39285729, Time: 0.0240 Steps: 74170, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001794, Sample Num: 28704, Cur Loss: 0.72127885, Cur Avg Loss: 0.51842747, Log Avg loss: 0.57539895, Global Avg Loss: 1.39274709, Time: 0.0069 Steps: 74180, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001804, Sample Num: 28864, Cur Loss: 0.43390539, Cur Avg Loss: 0.51836239, Log Avg loss: 0.50668744, Global Avg Loss: 1.39262766, Time: 0.0066 Steps: 74190, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001814, Sample Num: 29024, Cur Loss: 0.52314758, Cur Avg Loss: 0.51838998, Log Avg loss: 0.52336591, Global Avg Loss: 1.39251051, Time: 0.0075 Steps: 74200, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001824, Sample Num: 29184, Cur Loss: 0.71550399, Cur Avg Loss: 0.51919372, Log Avg loss: 0.66499324, Global Avg Loss: 1.39241247, Time: 0.0080 Steps: 74210, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001834, Sample Num: 29344, Cur Loss: 0.69347018, Cur Avg Loss: 0.51897558, Log Avg loss: 0.47918675, Global Avg Loss: 1.39228943, Time: 0.0230 Steps: 74220, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001844, Sample Num: 29504, Cur Loss: 0.30642831, Cur Avg Loss: 0.51878109, Log Avg loss: 0.48311051, Global Avg Loss: 1.39216695, Time: 0.0070 Steps: 74230, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001854, Sample Num: 29664, Cur Loss: 2.00539398, Cur Avg Loss: 0.51944831, Log Avg loss: 0.64248491, Global Avg Loss: 1.39206597, Time: 0.0066 Steps: 74240, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001864, Sample Num: 29824, Cur Loss: 0.58318388, Cur Avg Loss: 0.51899149, Log Avg loss: 0.43429647, Global Avg Loss: 1.39193698, Time: 0.0184 Steps: 74250, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001874, Sample Num: 29984, Cur Loss: 0.24972388, Cur Avg Loss: 0.51887353, Log Avg loss: 0.49688512, Global Avg Loss: 1.39181645, Time: 0.0228 Steps: 74260, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001884, Sample Num: 30144, Cur Loss: 0.28915095, Cur Avg Loss: 0.51950790, Log Avg loss: 0.63838935, Global Avg Loss: 1.39171500, Time: 0.0068 Steps: 74270, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001894, Sample Num: 30304, Cur Loss: 0.21793573, Cur Avg Loss: 0.51907425, Log Avg loss: 0.43737432, Global Avg Loss: 1.39158652, Time: 0.0066 Steps: 74280, Updated lr: 0.000031 Training, Epoch: 0035, Batch: 001904, Sample Num: 30464, Cur Loss: 0.64319897, Cur Avg Loss: 0.51942597, Log Avg loss: 0.58604174, Global Avg Loss: 1.39147809, Time: 0.0068 Steps: 74290, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001914, Sample Num: 30624, Cur Loss: 0.09675904, Cur Avg Loss: 0.51939400, Log Avg loss: 0.51330670, Global Avg Loss: 1.39135990, Time: 0.0065 Steps: 74300, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001924, Sample Num: 30784, Cur Loss: 0.43550265, Cur Avg Loss: 0.51892322, Log Avg loss: 0.42881717, Global Avg Loss: 1.39123037, Time: 0.0069 Steps: 74310, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001934, Sample Num: 30944, Cur Loss: 0.81866539, Cur Avg Loss: 0.51907755, Log Avg loss: 0.54877019, Global Avg Loss: 1.39111701, Time: 0.0143 Steps: 74320, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001944, Sample Num: 31104, Cur Loss: 0.22221643, Cur Avg Loss: 0.51897555, Log Avg loss: 0.49924800, Global Avg Loss: 1.39099702, Time: 0.0226 Steps: 74330, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001954, Sample Num: 31264, Cur Loss: 0.49124664, Cur Avg Loss: 0.51901219, Log Avg loss: 0.52613471, Global Avg Loss: 1.39088069, Time: 0.0066 Steps: 74340, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001964, Sample Num: 31424, Cur Loss: 0.49619830, Cur Avg Loss: 0.51865551, Log Avg loss: 0.44896111, Global Avg Loss: 1.39075400, Time: 0.0137 Steps: 74350, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001974, Sample Num: 31584, Cur Loss: 0.21430150, Cur Avg Loss: 0.51878912, Log Avg loss: 0.54503052, Global Avg Loss: 1.39064026, Time: 0.0156 Steps: 74360, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001984, Sample Num: 31744, Cur Loss: 0.59676820, Cur Avg Loss: 0.51923327, Log Avg loss: 0.60690790, Global Avg Loss: 1.39053488, Time: 0.0235 Steps: 74370, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 001994, Sample Num: 31904, Cur Loss: 0.55693495, Cur Avg Loss: 0.51961439, Log Avg loss: 0.59522920, Global Avg Loss: 1.39042796, Time: 0.0067 Steps: 74380, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002004, Sample Num: 32064, Cur Loss: 0.33042163, Cur Avg Loss: 0.51971190, Log Avg loss: 0.53915451, Global Avg Loss: 1.39031352, Time: 0.0160 Steps: 74390, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002014, Sample Num: 32224, Cur Loss: 0.33358487, Cur Avg Loss: 0.51921559, Log Avg loss: 0.41975623, Global Avg Loss: 1.39018307, Time: 0.0071 Steps: 74400, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002024, Sample Num: 32384, Cur Loss: 0.65117627, Cur Avg Loss: 0.51855712, Log Avg loss: 0.38594051, Global Avg Loss: 1.39004811, Time: 0.0120 Steps: 74410, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002034, Sample Num: 32544, Cur Loss: 0.23637365, Cur Avg Loss: 0.51810514, Log Avg loss: 0.42662531, Global Avg Loss: 1.38991865, Time: 0.0070 Steps: 74420, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002044, Sample Num: 32704, Cur Loss: 0.65617090, Cur Avg Loss: 0.51787611, Log Avg loss: 0.47129111, Global Avg Loss: 1.38979523, Time: 0.0106 Steps: 74430, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002054, Sample Num: 32864, Cur Loss: 0.24297185, Cur Avg Loss: 0.51760330, Log Avg loss: 0.46184005, Global Avg Loss: 1.38967057, Time: 0.0114 Steps: 74440, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002064, Sample Num: 33024, Cur Loss: 0.56007290, Cur Avg Loss: 0.51687319, Log Avg loss: 0.36690807, Global Avg Loss: 1.38953320, Time: 0.0067 Steps: 74450, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002074, Sample Num: 33184, Cur Loss: 0.15196118, Cur Avg Loss: 0.51659562, Log Avg loss: 0.45930667, Global Avg Loss: 1.38940827, Time: 0.0067 Steps: 74460, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002084, Sample Num: 33344, Cur Loss: 0.60946935, Cur Avg Loss: 0.51659850, Log Avg loss: 0.51719555, Global Avg Loss: 1.38929115, Time: 0.0066 Steps: 74470, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002094, Sample Num: 33504, Cur Loss: 0.37150866, Cur Avg Loss: 0.51664242, Log Avg loss: 0.52579572, Global Avg Loss: 1.38917521, Time: 0.0131 Steps: 74480, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002104, Sample Num: 33664, Cur Loss: 1.81876421, Cur Avg Loss: 0.51679403, Log Avg loss: 0.54854100, Global Avg Loss: 1.38906236, Time: 0.0067 Steps: 74490, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002114, Sample Num: 33824, Cur Loss: 0.54759145, Cur Avg Loss: 0.51679968, Log Avg loss: 0.51798753, Global Avg Loss: 1.38894543, Time: 0.0119 Steps: 74500, Updated lr: 0.000030 Training, Epoch: 0035, Batch: 002124, Sample Num: 33984, Cur Loss: 0.29184300, Cur Avg Loss: 0.51654543, Log Avg loss: 0.46279737, Global Avg Loss: 1.38882114, Time: 0.0110 Steps: 74510, Updated lr: 0.000030 ***** Running evaluation checkpoint-74515 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-74515 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 22.615600, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.600652, "eval_total_loss": 422.258179, "eval_mae": 0.579075, "eval_mse": 0.600817, "eval_r2": 0.618081, "eval_sp_statistic": 0.740477, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.786962, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.870704, "test_total_loss": 437.093469, "test_mae": 0.666843, "test_mse": 0.870925, "test_r2": 0.437897, "test_sp_statistic": 0.588876, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.667893, "test_ps_pvalue": 0.0, "lr": 3.0284495021337127e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.3887569180729684, "train_cur_epoch_loss": 1099.3013787530363, "train_cur_epoch_avg_loss": 0.5163463498135445, "train_cur_epoch_time": 22.615599632263184, "train_cur_epoch_avg_time": 0.010622639564238227, "epoch": 35, "step": 74515} ################################################## Training, Epoch: 0036, Batch: 000005, Sample Num: 80, Cur Loss: 0.32000306, Cur Avg Loss: 0.34256812, Log Avg loss: 0.38717260, Global Avg Loss: 1.38868672, Time: 0.0113 Steps: 74520, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000015, Sample Num: 240, Cur Loss: 0.66983104, Cur Avg Loss: 0.41831884, Log Avg loss: 0.45619419, Global Avg Loss: 1.38856161, Time: 0.0114 Steps: 74530, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000025, Sample Num: 400, Cur Loss: 0.39468020, Cur Avg Loss: 0.43455454, Log Avg loss: 0.45890809, Global Avg Loss: 1.38843689, Time: 0.0073 Steps: 74540, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000035, Sample Num: 560, Cur Loss: 0.53835237, Cur Avg Loss: 0.51245054, Log Avg loss: 0.70719056, Global Avg Loss: 1.38834551, Time: 0.0115 Steps: 74550, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000045, Sample Num: 720, Cur Loss: 0.54209185, Cur Avg Loss: 0.51552154, Log Avg loss: 0.52627004, Global Avg Loss: 1.38822988, Time: 0.0117 Steps: 74560, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000055, Sample Num: 880, Cur Loss: 0.41939017, Cur Avg Loss: 0.53267539, Log Avg loss: 0.60986770, Global Avg Loss: 1.38812550, Time: 0.0108 Steps: 74570, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000065, Sample Num: 1040, Cur Loss: 0.87263596, Cur Avg Loss: 0.51673242, Log Avg loss: 0.42904607, Global Avg Loss: 1.38799691, Time: 0.0068 Steps: 74580, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000075, Sample Num: 1200, Cur Loss: 0.55751425, Cur Avg Loss: 0.51378536, Log Avg loss: 0.49462952, Global Avg Loss: 1.38787714, Time: 0.0114 Steps: 74590, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000085, Sample Num: 1360, Cur Loss: 0.62360471, Cur Avg Loss: 0.53776133, Log Avg loss: 0.71758110, Global Avg Loss: 1.38778729, Time: 0.0089 Steps: 74600, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000095, Sample Num: 1520, Cur Loss: 0.59599113, Cur Avg Loss: 0.53686130, Log Avg loss: 0.52921098, Global Avg Loss: 1.38767221, Time: 0.0099 Steps: 74610, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000105, Sample Num: 1680, Cur Loss: 0.39104435, Cur Avg Loss: 0.53380512, Log Avg loss: 0.50477142, Global Avg Loss: 1.38755389, Time: 0.0078 Steps: 74620, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000115, Sample Num: 1840, Cur Loss: 0.16951549, Cur Avg Loss: 0.52046580, Log Avg loss: 0.38040292, Global Avg Loss: 1.38741894, Time: 0.0092 Steps: 74630, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000125, Sample Num: 2000, Cur Loss: 0.60586429, Cur Avg Loss: 0.51749450, Log Avg loss: 0.48332457, Global Avg Loss: 1.38729781, Time: 0.0113 Steps: 74640, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000135, Sample Num: 2160, Cur Loss: 0.56674069, Cur Avg Loss: 0.52775026, Log Avg loss: 0.65594731, Global Avg Loss: 1.38719984, Time: 0.0146 Steps: 74650, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000145, Sample Num: 2320, Cur Loss: 0.26710063, Cur Avg Loss: 0.52485493, Log Avg loss: 0.48576788, Global Avg Loss: 1.38707910, Time: 0.0115 Steps: 74660, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000155, Sample Num: 2480, Cur Loss: 0.66178280, Cur Avg Loss: 0.52235304, Log Avg loss: 0.48607563, Global Avg Loss: 1.38695844, Time: 0.0094 Steps: 74670, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000165, Sample Num: 2640, Cur Loss: 0.52362806, Cur Avg Loss: 0.52140807, Log Avg loss: 0.50676104, Global Avg Loss: 1.38684057, Time: 0.0124 Steps: 74680, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000175, Sample Num: 2800, Cur Loss: 0.42936763, Cur Avg Loss: 0.51165769, Log Avg loss: 0.35077648, Global Avg Loss: 1.38670186, Time: 0.0076 Steps: 74690, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000185, Sample Num: 2960, Cur Loss: 0.14757407, Cur Avg Loss: 0.50778624, Log Avg loss: 0.44003580, Global Avg Loss: 1.38657513, Time: 0.0113 Steps: 74700, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000195, Sample Num: 3120, Cur Loss: 0.61660665, Cur Avg Loss: 0.50476986, Log Avg loss: 0.44896687, Global Avg Loss: 1.38644963, Time: 0.0111 Steps: 74710, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000205, Sample Num: 3280, Cur Loss: 0.67716366, Cur Avg Loss: 0.50477293, Log Avg loss: 0.50483291, Global Avg Loss: 1.38633164, Time: 0.0123 Steps: 74720, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000215, Sample Num: 3440, Cur Loss: 0.46692312, Cur Avg Loss: 0.50418560, Log Avg loss: 0.49214528, Global Avg Loss: 1.38621199, Time: 0.0120 Steps: 74730, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000225, Sample Num: 3600, Cur Loss: 0.41258356, Cur Avg Loss: 0.50879636, Log Avg loss: 0.60792777, Global Avg Loss: 1.38610785, Time: 0.0144 Steps: 74740, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000235, Sample Num: 3760, Cur Loss: 0.24057652, Cur Avg Loss: 0.50814617, Log Avg loss: 0.49351673, Global Avg Loss: 1.38598844, Time: 0.0066 Steps: 74750, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000245, Sample Num: 3920, Cur Loss: 0.50899494, Cur Avg Loss: 0.50123790, Log Avg loss: 0.33889372, Global Avg Loss: 1.38584838, Time: 0.0109 Steps: 74760, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000255, Sample Num: 4080, Cur Loss: 0.80221438, Cur Avg Loss: 0.50085665, Log Avg loss: 0.49151590, Global Avg Loss: 1.38572877, Time: 0.0090 Steps: 74770, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000265, Sample Num: 4240, Cur Loss: 0.69780320, Cur Avg Loss: 0.50128042, Log Avg loss: 0.51208653, Global Avg Loss: 1.38561194, Time: 0.0068 Steps: 74780, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000275, Sample Num: 4400, Cur Loss: 0.48546511, Cur Avg Loss: 0.49983233, Log Avg loss: 0.46145793, Global Avg Loss: 1.38548838, Time: 0.0105 Steps: 74790, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000285, Sample Num: 4560, Cur Loss: 0.27553219, Cur Avg Loss: 0.49769114, Log Avg loss: 0.43880852, Global Avg Loss: 1.38536181, Time: 0.0111 Steps: 74800, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000295, Sample Num: 4720, Cur Loss: 1.16386604, Cur Avg Loss: 0.49849533, Log Avg loss: 0.52141458, Global Avg Loss: 1.38524633, Time: 0.0108 Steps: 74810, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000305, Sample Num: 4880, Cur Loss: 0.21408683, Cur Avg Loss: 0.49856616, Log Avg loss: 0.50065571, Global Avg Loss: 1.38512810, Time: 0.0134 Steps: 74820, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000315, Sample Num: 5040, Cur Loss: 0.95821518, Cur Avg Loss: 0.49768448, Log Avg loss: 0.47079327, Global Avg Loss: 1.38500591, Time: 0.0071 Steps: 74830, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000325, Sample Num: 5200, Cur Loss: 0.35921073, Cur Avg Loss: 0.49977690, Log Avg loss: 0.56568825, Global Avg Loss: 1.38489644, Time: 0.0123 Steps: 74840, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000335, Sample Num: 5360, Cur Loss: 0.38572025, Cur Avg Loss: 0.49852201, Log Avg loss: 0.45773810, Global Avg Loss: 1.38477257, Time: 0.0118 Steps: 74850, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000345, Sample Num: 5520, Cur Loss: 0.18592922, Cur Avg Loss: 0.49973315, Log Avg loss: 0.54030611, Global Avg Loss: 1.38465976, Time: 0.0071 Steps: 74860, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000355, Sample Num: 5680, Cur Loss: 0.55709982, Cur Avg Loss: 0.50069990, Log Avg loss: 0.53405296, Global Avg Loss: 1.38454615, Time: 0.0067 Steps: 74870, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000365, Sample Num: 5840, Cur Loss: 0.36375770, Cur Avg Loss: 0.49884825, Log Avg loss: 0.43311470, Global Avg Loss: 1.38441909, Time: 0.0105 Steps: 74880, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000375, Sample Num: 6000, Cur Loss: 0.36323249, Cur Avg Loss: 0.50108155, Log Avg loss: 0.58259685, Global Avg Loss: 1.38431202, Time: 0.0105 Steps: 74890, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000385, Sample Num: 6160, Cur Loss: 0.72743708, Cur Avg Loss: 0.49799111, Log Avg loss: 0.38209978, Global Avg Loss: 1.38417822, Time: 0.0107 Steps: 74900, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000395, Sample Num: 6320, Cur Loss: 0.50629783, Cur Avg Loss: 0.50037550, Log Avg loss: 0.59217443, Global Avg Loss: 1.38407249, Time: 0.0065 Steps: 74910, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000405, Sample Num: 6480, Cur Loss: 0.37361133, Cur Avg Loss: 0.49945318, Log Avg loss: 0.46302165, Global Avg Loss: 1.38394955, Time: 0.0115 Steps: 74920, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000415, Sample Num: 6640, Cur Loss: 0.28011441, Cur Avg Loss: 0.49761930, Log Avg loss: 0.42334710, Global Avg Loss: 1.38382135, Time: 0.0067 Steps: 74930, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000425, Sample Num: 6800, Cur Loss: 0.38689655, Cur Avg Loss: 0.49809384, Log Avg loss: 0.51778731, Global Avg Loss: 1.38370579, Time: 0.0066 Steps: 74940, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000435, Sample Num: 6960, Cur Loss: 1.08770692, Cur Avg Loss: 0.49947369, Log Avg loss: 0.55811701, Global Avg Loss: 1.38359563, Time: 0.0082 Steps: 74950, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000445, Sample Num: 7120, Cur Loss: 1.02123547, Cur Avg Loss: 0.49827922, Log Avg loss: 0.44631972, Global Avg Loss: 1.38347060, Time: 0.0071 Steps: 74960, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000455, Sample Num: 7280, Cur Loss: 0.45876735, Cur Avg Loss: 0.50240121, Log Avg loss: 0.68583001, Global Avg Loss: 1.38337754, Time: 0.0067 Steps: 74970, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000465, Sample Num: 7440, Cur Loss: 0.31656384, Cur Avg Loss: 0.50054432, Log Avg loss: 0.41605580, Global Avg Loss: 1.38324853, Time: 0.0176 Steps: 74980, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000475, Sample Num: 7600, Cur Loss: 0.53615677, Cur Avg Loss: 0.50198979, Log Avg loss: 0.56920406, Global Avg Loss: 1.38313998, Time: 0.0084 Steps: 74990, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000485, Sample Num: 7760, Cur Loss: 0.41632473, Cur Avg Loss: 0.50352724, Log Avg loss: 0.57655624, Global Avg Loss: 1.38303243, Time: 0.0067 Steps: 75000, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000495, Sample Num: 7920, Cur Loss: 0.44375643, Cur Avg Loss: 0.50384041, Log Avg loss: 0.51902913, Global Avg Loss: 1.38291725, Time: 0.0120 Steps: 75010, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000505, Sample Num: 8080, Cur Loss: 0.21766061, Cur Avg Loss: 0.50407191, Log Avg loss: 0.51553134, Global Avg Loss: 1.38280163, Time: 0.0066 Steps: 75020, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000515, Sample Num: 8240, Cur Loss: 0.27096021, Cur Avg Loss: 0.50482448, Log Avg loss: 0.54282889, Global Avg Loss: 1.38268968, Time: 0.0098 Steps: 75030, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000525, Sample Num: 8400, Cur Loss: 0.28851497, Cur Avg Loss: 0.50150930, Log Avg loss: 0.33077752, Global Avg Loss: 1.38254950, Time: 0.0115 Steps: 75040, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000535, Sample Num: 8560, Cur Loss: 0.62414968, Cur Avg Loss: 0.50145733, Log Avg loss: 0.49872914, Global Avg Loss: 1.38243173, Time: 0.0105 Steps: 75050, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000545, Sample Num: 8720, Cur Loss: 0.37063342, Cur Avg Loss: 0.50125828, Log Avg loss: 0.49060897, Global Avg Loss: 1.38231292, Time: 0.0109 Steps: 75060, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000555, Sample Num: 8880, Cur Loss: 0.47459191, Cur Avg Loss: 0.50001563, Log Avg loss: 0.43229158, Global Avg Loss: 1.38218637, Time: 0.0079 Steps: 75070, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000565, Sample Num: 9040, Cur Loss: 0.37115037, Cur Avg Loss: 0.49983402, Log Avg loss: 0.48975424, Global Avg Loss: 1.38206750, Time: 0.0114 Steps: 75080, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000575, Sample Num: 9200, Cur Loss: 0.47628450, Cur Avg Loss: 0.50035938, Log Avg loss: 0.53004226, Global Avg Loss: 1.38195403, Time: 0.0067 Steps: 75090, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000585, Sample Num: 9360, Cur Loss: 0.22376493, Cur Avg Loss: 0.50020348, Log Avg loss: 0.49123954, Global Avg Loss: 1.38183543, Time: 0.0134 Steps: 75100, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000595, Sample Num: 9520, Cur Loss: 0.34845415, Cur Avg Loss: 0.50060016, Log Avg loss: 0.52380549, Global Avg Loss: 1.38172119, Time: 0.0115 Steps: 75110, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000605, Sample Num: 9680, Cur Loss: 0.27124459, Cur Avg Loss: 0.50348133, Log Avg loss: 0.67491133, Global Avg Loss: 1.38162710, Time: 0.0073 Steps: 75120, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000615, Sample Num: 9840, Cur Loss: 0.33641100, Cur Avg Loss: 0.50241340, Log Avg loss: 0.43780365, Global Avg Loss: 1.38150148, Time: 0.0065 Steps: 75130, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000625, Sample Num: 10000, Cur Loss: 0.69544405, Cur Avg Loss: 0.50398566, Log Avg loss: 0.60067977, Global Avg Loss: 1.38139756, Time: 0.0227 Steps: 75140, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000635, Sample Num: 10160, Cur Loss: 0.27373734, Cur Avg Loss: 0.50266104, Log Avg loss: 0.41987177, Global Avg Loss: 1.38126961, Time: 0.0108 Steps: 75150, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000645, Sample Num: 10320, Cur Loss: 0.87604541, Cur Avg Loss: 0.50368235, Log Avg loss: 0.56853594, Global Avg Loss: 1.38116148, Time: 0.0070 Steps: 75160, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000655, Sample Num: 10480, Cur Loss: 0.68322635, Cur Avg Loss: 0.50438319, Log Avg loss: 0.54958688, Global Avg Loss: 1.38105085, Time: 0.0110 Steps: 75170, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000665, Sample Num: 10640, Cur Loss: 0.36377758, Cur Avg Loss: 0.50288396, Log Avg loss: 0.40468461, Global Avg Loss: 1.38092098, Time: 0.0108 Steps: 75180, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000675, Sample Num: 10800, Cur Loss: 0.59016597, Cur Avg Loss: 0.50383316, Log Avg loss: 0.56695514, Global Avg Loss: 1.38081273, Time: 0.0070 Steps: 75190, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000685, Sample Num: 10960, Cur Loss: 0.83323669, Cur Avg Loss: 0.50477873, Log Avg loss: 0.56860472, Global Avg Loss: 1.38070472, Time: 0.0066 Steps: 75200, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000695, Sample Num: 11120, Cur Loss: 0.32522920, Cur Avg Loss: 0.50356048, Log Avg loss: 0.42011020, Global Avg Loss: 1.38057700, Time: 0.0121 Steps: 75210, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000705, Sample Num: 11280, Cur Loss: 0.35493997, Cur Avg Loss: 0.50517225, Log Avg loss: 0.61719045, Global Avg Loss: 1.38047551, Time: 0.0110 Steps: 75220, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000715, Sample Num: 11440, Cur Loss: 0.20952646, Cur Avg Loss: 0.50393926, Log Avg loss: 0.41701332, Global Avg Loss: 1.38034745, Time: 0.0111 Steps: 75230, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000725, Sample Num: 11600, Cur Loss: 0.69999993, Cur Avg Loss: 0.50364099, Log Avg loss: 0.48231488, Global Avg Loss: 1.38022809, Time: 0.0067 Steps: 75240, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000735, Sample Num: 11760, Cur Loss: 0.10827174, Cur Avg Loss: 0.49985428, Log Avg loss: 0.22531731, Global Avg Loss: 1.38007461, Time: 0.0087 Steps: 75250, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000745, Sample Num: 11920, Cur Loss: 0.24819553, Cur Avg Loss: 0.50035239, Log Avg loss: 0.53696380, Global Avg Loss: 1.37996259, Time: 0.0113 Steps: 75260, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000755, Sample Num: 12080, Cur Loss: 0.24822970, Cur Avg Loss: 0.49981575, Log Avg loss: 0.45983580, Global Avg Loss: 1.37984034, Time: 0.0115 Steps: 75270, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000765, Sample Num: 12240, Cur Loss: 0.28141648, Cur Avg Loss: 0.49865798, Log Avg loss: 0.41124627, Global Avg Loss: 1.37971168, Time: 0.0066 Steps: 75280, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000775, Sample Num: 12400, Cur Loss: 0.20521654, Cur Avg Loss: 0.49874938, Log Avg loss: 0.50574186, Global Avg Loss: 1.37959560, Time: 0.0109 Steps: 75290, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000785, Sample Num: 12560, Cur Loss: 0.82682925, Cur Avg Loss: 0.49912543, Log Avg loss: 0.52826922, Global Avg Loss: 1.37948254, Time: 0.0065 Steps: 75300, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000795, Sample Num: 12720, Cur Loss: 0.20099244, Cur Avg Loss: 0.50060090, Log Avg loss: 0.61642528, Global Avg Loss: 1.37938122, Time: 0.0137 Steps: 75310, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000805, Sample Num: 12880, Cur Loss: 1.06477880, Cur Avg Loss: 0.50052635, Log Avg loss: 0.49459974, Global Avg Loss: 1.37926375, Time: 0.0089 Steps: 75320, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000815, Sample Num: 13040, Cur Loss: 0.32732219, Cur Avg Loss: 0.50171564, Log Avg loss: 0.59745347, Global Avg Loss: 1.37915996, Time: 0.0131 Steps: 75330, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000825, Sample Num: 13200, Cur Loss: 0.86510718, Cur Avg Loss: 0.50118144, Log Avg loss: 0.45764419, Global Avg Loss: 1.37903765, Time: 0.0114 Steps: 75340, Updated lr: 0.000030 Training, Epoch: 0036, Batch: 000835, Sample Num: 13360, Cur Loss: 0.43186867, Cur Avg Loss: 0.50298075, Log Avg loss: 0.65142353, Global Avg Loss: 1.37894108, Time: 0.0113 Steps: 75350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000845, Sample Num: 13520, Cur Loss: 0.43629318, Cur Avg Loss: 0.50495056, Log Avg loss: 0.66942979, Global Avg Loss: 1.37884693, Time: 0.0112 Steps: 75360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000855, Sample Num: 13680, Cur Loss: 0.30012953, Cur Avg Loss: 0.50487381, Log Avg loss: 0.49838859, Global Avg Loss: 1.37873012, Time: 0.0070 Steps: 75370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000865, Sample Num: 13840, Cur Loss: 0.27378964, Cur Avg Loss: 0.50437746, Log Avg loss: 0.46193968, Global Avg Loss: 1.37860849, Time: 0.0071 Steps: 75380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000875, Sample Num: 14000, Cur Loss: 0.39047313, Cur Avg Loss: 0.50370820, Log Avg loss: 0.44581715, Global Avg Loss: 1.37848476, Time: 0.0127 Steps: 75390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000885, Sample Num: 14160, Cur Loss: 1.04737663, Cur Avg Loss: 0.50540720, Log Avg loss: 0.65406948, Global Avg Loss: 1.37838869, Time: 0.0128 Steps: 75400, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000895, Sample Num: 14320, Cur Loss: 0.40548077, Cur Avg Loss: 0.50544428, Log Avg loss: 0.50872580, Global Avg Loss: 1.37827336, Time: 0.0067 Steps: 75410, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000905, Sample Num: 14480, Cur Loss: 0.54080570, Cur Avg Loss: 0.50615612, Log Avg loss: 0.56986610, Global Avg Loss: 1.37816618, Time: 0.0066 Steps: 75420, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000915, Sample Num: 14640, Cur Loss: 0.55853742, Cur Avg Loss: 0.50670084, Log Avg loss: 0.55599813, Global Avg Loss: 1.37805718, Time: 0.0150 Steps: 75430, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000925, Sample Num: 14800, Cur Loss: 1.08506656, Cur Avg Loss: 0.50730847, Log Avg loss: 0.56290636, Global Avg Loss: 1.37794913, Time: 0.0067 Steps: 75440, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000935, Sample Num: 14960, Cur Loss: 0.64494163, Cur Avg Loss: 0.50756507, Log Avg loss: 0.53130025, Global Avg Loss: 1.37783691, Time: 0.0123 Steps: 75450, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000945, Sample Num: 15120, Cur Loss: 0.56602228, Cur Avg Loss: 0.50653902, Log Avg loss: 0.41060404, Global Avg Loss: 1.37770873, Time: 0.0120 Steps: 75460, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000955, Sample Num: 15280, Cur Loss: 0.28057232, Cur Avg Loss: 0.50583830, Log Avg loss: 0.43961948, Global Avg Loss: 1.37758444, Time: 0.0119 Steps: 75470, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000965, Sample Num: 15440, Cur Loss: 0.41012609, Cur Avg Loss: 0.50584597, Log Avg loss: 0.50657845, Global Avg Loss: 1.37746904, Time: 0.0096 Steps: 75480, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000975, Sample Num: 15600, Cur Loss: 0.69349319, Cur Avg Loss: 0.50585778, Log Avg loss: 0.50699787, Global Avg Loss: 1.37735373, Time: 0.0112 Steps: 75490, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000985, Sample Num: 15760, Cur Loss: 0.43627679, Cur Avg Loss: 0.50678477, Log Avg loss: 0.59716671, Global Avg Loss: 1.37725039, Time: 0.0138 Steps: 75500, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 000995, Sample Num: 15920, Cur Loss: 0.30101883, Cur Avg Loss: 0.50630937, Log Avg loss: 0.45948224, Global Avg Loss: 1.37712885, Time: 0.0145 Steps: 75510, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001005, Sample Num: 16080, Cur Loss: 0.91794789, Cur Avg Loss: 0.50839992, Log Avg loss: 0.71640949, Global Avg Loss: 1.37704136, Time: 0.0063 Steps: 75520, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001015, Sample Num: 16240, Cur Loss: 0.23648596, Cur Avg Loss: 0.50826465, Log Avg loss: 0.49467003, Global Avg Loss: 1.37692454, Time: 0.0118 Steps: 75530, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001025, Sample Num: 16400, Cur Loss: 0.50640047, Cur Avg Loss: 0.50738898, Log Avg loss: 0.41850798, Global Avg Loss: 1.37679766, Time: 0.0262 Steps: 75540, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001035, Sample Num: 16560, Cur Loss: 1.49608779, Cur Avg Loss: 0.50850053, Log Avg loss: 0.62243520, Global Avg Loss: 1.37669781, Time: 0.0087 Steps: 75550, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001045, Sample Num: 16720, Cur Loss: 0.37215406, Cur Avg Loss: 0.50917909, Log Avg loss: 0.57941006, Global Avg Loss: 1.37659230, Time: 0.0121 Steps: 75560, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001055, Sample Num: 16880, Cur Loss: 0.24957737, Cur Avg Loss: 0.50767192, Log Avg loss: 0.35017258, Global Avg Loss: 1.37645647, Time: 0.0233 Steps: 75570, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001065, Sample Num: 17040, Cur Loss: 0.45146316, Cur Avg Loss: 0.50650378, Log Avg loss: 0.38326522, Global Avg Loss: 1.37632506, Time: 0.0065 Steps: 75580, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001075, Sample Num: 17200, Cur Loss: 0.35052603, Cur Avg Loss: 0.50668104, Log Avg loss: 0.52555858, Global Avg Loss: 1.37621251, Time: 0.0065 Steps: 75590, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001085, Sample Num: 17360, Cur Loss: 0.56198812, Cur Avg Loss: 0.50802841, Log Avg loss: 0.65287119, Global Avg Loss: 1.37611683, Time: 0.0070 Steps: 75600, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001095, Sample Num: 17520, Cur Loss: 0.27994892, Cur Avg Loss: 0.50861521, Log Avg loss: 0.57228278, Global Avg Loss: 1.37601052, Time: 0.0080 Steps: 75610, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001105, Sample Num: 17680, Cur Loss: 0.42494953, Cur Avg Loss: 0.50817368, Log Avg loss: 0.45982590, Global Avg Loss: 1.37588936, Time: 0.0067 Steps: 75620, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001115, Sample Num: 17840, Cur Loss: 0.31763601, Cur Avg Loss: 0.50862049, Log Avg loss: 0.55799301, Global Avg Loss: 1.37578122, Time: 0.0065 Steps: 75630, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001125, Sample Num: 18000, Cur Loss: 0.49032027, Cur Avg Loss: 0.50926605, Log Avg loss: 0.58124578, Global Avg Loss: 1.37567618, Time: 0.0090 Steps: 75640, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001135, Sample Num: 18160, Cur Loss: 0.15491080, Cur Avg Loss: 0.50808414, Log Avg loss: 0.37511907, Global Avg Loss: 1.37554392, Time: 0.0072 Steps: 75650, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001145, Sample Num: 18320, Cur Loss: 0.34176022, Cur Avg Loss: 0.51029787, Log Avg loss: 0.76155645, Global Avg Loss: 1.37546277, Time: 0.0112 Steps: 75660, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001155, Sample Num: 18480, Cur Loss: 0.30736381, Cur Avg Loss: 0.50862509, Log Avg loss: 0.31709232, Global Avg Loss: 1.37532290, Time: 0.0070 Steps: 75670, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001165, Sample Num: 18640, Cur Loss: 0.18596366, Cur Avg Loss: 0.50898717, Log Avg loss: 0.55080725, Global Avg Loss: 1.37521395, Time: 0.0064 Steps: 75680, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001175, Sample Num: 18800, Cur Loss: 0.34004691, Cur Avg Loss: 0.50876212, Log Avg loss: 0.48254405, Global Avg Loss: 1.37509601, Time: 0.0116 Steps: 75690, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001185, Sample Num: 18960, Cur Loss: 0.34059399, Cur Avg Loss: 0.50822535, Log Avg loss: 0.44515472, Global Avg Loss: 1.37497317, Time: 0.0064 Steps: 75700, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001195, Sample Num: 19120, Cur Loss: 0.21752606, Cur Avg Loss: 0.50703630, Log Avg loss: 0.36613411, Global Avg Loss: 1.37483992, Time: 0.0126 Steps: 75710, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001205, Sample Num: 19280, Cur Loss: 0.24005190, Cur Avg Loss: 0.50641451, Log Avg loss: 0.43211065, Global Avg Loss: 1.37471542, Time: 0.0065 Steps: 75720, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001215, Sample Num: 19440, Cur Loss: 0.35407740, Cur Avg Loss: 0.50544030, Log Avg loss: 0.38804733, Global Avg Loss: 1.37458513, Time: 0.0090 Steps: 75730, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001225, Sample Num: 19600, Cur Loss: 0.50377166, Cur Avg Loss: 0.50576047, Log Avg loss: 0.54466089, Global Avg Loss: 1.37447555, Time: 0.0100 Steps: 75740, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001235, Sample Num: 19760, Cur Loss: 0.57699615, Cur Avg Loss: 0.50677491, Log Avg loss: 0.63104456, Global Avg Loss: 1.37437741, Time: 0.0113 Steps: 75750, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001245, Sample Num: 19920, Cur Loss: 0.25341800, Cur Avg Loss: 0.50689520, Log Avg loss: 0.52175074, Global Avg Loss: 1.37426487, Time: 0.0118 Steps: 75760, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001255, Sample Num: 20080, Cur Loss: 0.18258771, Cur Avg Loss: 0.50731727, Log Avg loss: 0.55986470, Global Avg Loss: 1.37415738, Time: 0.0073 Steps: 75770, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001265, Sample Num: 20240, Cur Loss: 0.20329773, Cur Avg Loss: 0.50875809, Log Avg loss: 0.68958184, Global Avg Loss: 1.37406705, Time: 0.0069 Steps: 75780, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001275, Sample Num: 20400, Cur Loss: 0.92481124, Cur Avg Loss: 0.50914002, Log Avg loss: 0.55745359, Global Avg Loss: 1.37395930, Time: 0.0065 Steps: 75790, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001285, Sample Num: 20560, Cur Loss: 0.64187944, Cur Avg Loss: 0.50987035, Log Avg loss: 0.60298746, Global Avg Loss: 1.37385759, Time: 0.0066 Steps: 75800, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001295, Sample Num: 20720, Cur Loss: 0.39409363, Cur Avg Loss: 0.51030615, Log Avg loss: 0.56630682, Global Avg Loss: 1.37375106, Time: 0.0065 Steps: 75810, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001305, Sample Num: 20880, Cur Loss: 0.44766527, Cur Avg Loss: 0.51012052, Log Avg loss: 0.48608155, Global Avg Loss: 1.37363399, Time: 0.0066 Steps: 75820, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001315, Sample Num: 21040, Cur Loss: 0.69165516, Cur Avg Loss: 0.51035655, Log Avg loss: 0.54115777, Global Avg Loss: 1.37352421, Time: 0.0118 Steps: 75830, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001325, Sample Num: 21200, Cur Loss: 0.59890604, Cur Avg Loss: 0.50984882, Log Avg loss: 0.44308279, Global Avg Loss: 1.37340152, Time: 0.0237 Steps: 75840, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001335, Sample Num: 21360, Cur Loss: 1.47786069, Cur Avg Loss: 0.51011702, Log Avg loss: 0.54565357, Global Avg Loss: 1.37329239, Time: 0.0109 Steps: 75850, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001345, Sample Num: 21520, Cur Loss: 0.19684535, Cur Avg Loss: 0.50989965, Log Avg loss: 0.48088057, Global Avg Loss: 1.37317475, Time: 0.0066 Steps: 75860, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001355, Sample Num: 21680, Cur Loss: 0.53456533, Cur Avg Loss: 0.51024607, Log Avg loss: 0.55683897, Global Avg Loss: 1.37306716, Time: 0.0107 Steps: 75870, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001365, Sample Num: 21840, Cur Loss: 0.36664632, Cur Avg Loss: 0.51036050, Log Avg loss: 0.52586550, Global Avg Loss: 1.37295551, Time: 0.0113 Steps: 75880, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001375, Sample Num: 22000, Cur Loss: 0.39335775, Cur Avg Loss: 0.51051247, Log Avg loss: 0.53125713, Global Avg Loss: 1.37284460, Time: 0.0082 Steps: 75890, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001385, Sample Num: 22160, Cur Loss: 0.64789790, Cur Avg Loss: 0.51120477, Log Avg loss: 0.60639553, Global Avg Loss: 1.37274361, Time: 0.0078 Steps: 75900, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001395, Sample Num: 22320, Cur Loss: 0.39637700, Cur Avg Loss: 0.51053572, Log Avg loss: 0.41787274, Global Avg Loss: 1.37261782, Time: 0.0067 Steps: 75910, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001405, Sample Num: 22480, Cur Loss: 0.63353366, Cur Avg Loss: 0.51159945, Log Avg loss: 0.65998977, Global Avg Loss: 1.37252396, Time: 0.0121 Steps: 75920, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001415, Sample Num: 22640, Cur Loss: 0.68833363, Cur Avg Loss: 0.51161291, Log Avg loss: 0.51350338, Global Avg Loss: 1.37241083, Time: 0.0090 Steps: 75930, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001425, Sample Num: 22800, Cur Loss: 0.17948875, Cur Avg Loss: 0.51171065, Log Avg loss: 0.52554107, Global Avg Loss: 1.37229931, Time: 0.0087 Steps: 75940, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001435, Sample Num: 22960, Cur Loss: 0.62351489, Cur Avg Loss: 0.51230006, Log Avg loss: 0.59629134, Global Avg Loss: 1.37219713, Time: 0.0114 Steps: 75950, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001445, Sample Num: 23120, Cur Loss: 0.84120518, Cur Avg Loss: 0.51212256, Log Avg loss: 0.48665189, Global Avg Loss: 1.37208055, Time: 0.0117 Steps: 75960, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001455, Sample Num: 23280, Cur Loss: 0.46045858, Cur Avg Loss: 0.51296833, Log Avg loss: 0.63518132, Global Avg Loss: 1.37198355, Time: 0.0118 Steps: 75970, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001465, Sample Num: 23440, Cur Loss: 0.79496491, Cur Avg Loss: 0.51227957, Log Avg loss: 0.41206455, Global Avg Loss: 1.37185722, Time: 0.0118 Steps: 75980, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001475, Sample Num: 23600, Cur Loss: 0.26323518, Cur Avg Loss: 0.51279564, Log Avg loss: 0.58840024, Global Avg Loss: 1.37175412, Time: 0.0074 Steps: 75990, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001485, Sample Num: 23760, Cur Loss: 0.41109651, Cur Avg Loss: 0.51311252, Log Avg loss: 0.55985209, Global Avg Loss: 1.37164729, Time: 0.0134 Steps: 76000, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001495, Sample Num: 23920, Cur Loss: 0.71715724, Cur Avg Loss: 0.51249679, Log Avg loss: 0.42106107, Global Avg Loss: 1.37152223, Time: 0.0072 Steps: 76010, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001505, Sample Num: 24080, Cur Loss: 0.43922710, Cur Avg Loss: 0.51176645, Log Avg loss: 0.40258115, Global Avg Loss: 1.37139477, Time: 0.0110 Steps: 76020, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001515, Sample Num: 24240, Cur Loss: 0.50917256, Cur Avg Loss: 0.51138612, Log Avg loss: 0.45414567, Global Avg Loss: 1.37127412, Time: 0.0071 Steps: 76030, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001525, Sample Num: 24400, Cur Loss: 0.30415159, Cur Avg Loss: 0.51063339, Log Avg loss: 0.39659607, Global Avg Loss: 1.37114595, Time: 0.0118 Steps: 76040, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001535, Sample Num: 24560, Cur Loss: 0.31807423, Cur Avg Loss: 0.50926587, Log Avg loss: 0.30071779, Global Avg Loss: 1.37100519, Time: 0.0144 Steps: 76050, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001545, Sample Num: 24720, Cur Loss: 0.31761247, Cur Avg Loss: 0.50988183, Log Avg loss: 0.60443200, Global Avg Loss: 1.37090441, Time: 0.0068 Steps: 76060, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001555, Sample Num: 24880, Cur Loss: 0.23955846, Cur Avg Loss: 0.50976850, Log Avg loss: 0.49225883, Global Avg Loss: 1.37078890, Time: 0.0118 Steps: 76070, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001565, Sample Num: 25040, Cur Loss: 0.24531506, Cur Avg Loss: 0.50881754, Log Avg loss: 0.36094317, Global Avg Loss: 1.37065617, Time: 0.0107 Steps: 76080, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001575, Sample Num: 25200, Cur Loss: 0.22335023, Cur Avg Loss: 0.50808786, Log Avg loss: 0.39389381, Global Avg Loss: 1.37052780, Time: 0.0107 Steps: 76090, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001585, Sample Num: 25360, Cur Loss: 0.23951527, Cur Avg Loss: 0.50750558, Log Avg loss: 0.41579704, Global Avg Loss: 1.37040234, Time: 0.0116 Steps: 76100, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001595, Sample Num: 25520, Cur Loss: 0.54596782, Cur Avg Loss: 0.50854663, Log Avg loss: 0.67355303, Global Avg Loss: 1.37031078, Time: 0.0087 Steps: 76110, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001605, Sample Num: 25680, Cur Loss: 0.43641976, Cur Avg Loss: 0.50942367, Log Avg loss: 0.64931144, Global Avg Loss: 1.37021606, Time: 0.0066 Steps: 76120, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001615, Sample Num: 25840, Cur Loss: 0.15813154, Cur Avg Loss: 0.50994369, Log Avg loss: 0.59340566, Global Avg Loss: 1.37011403, Time: 0.0113 Steps: 76130, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001625, Sample Num: 26000, Cur Loss: 0.35695589, Cur Avg Loss: 0.50934074, Log Avg loss: 0.41196481, Global Avg Loss: 1.36998819, Time: 0.0109 Steps: 76140, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001635, Sample Num: 26160, Cur Loss: 0.24285968, Cur Avg Loss: 0.50895955, Log Avg loss: 0.44701571, Global Avg Loss: 1.36986698, Time: 0.0071 Steps: 76150, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001645, Sample Num: 26320, Cur Loss: 0.62165582, Cur Avg Loss: 0.50894167, Log Avg loss: 0.50601873, Global Avg Loss: 1.36975356, Time: 0.0067 Steps: 76160, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001655, Sample Num: 26480, Cur Loss: 0.28154072, Cur Avg Loss: 0.50910343, Log Avg loss: 0.53571318, Global Avg Loss: 1.36964406, Time: 0.0088 Steps: 76170, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001665, Sample Num: 26640, Cur Loss: 0.51380438, Cur Avg Loss: 0.50975769, Log Avg loss: 0.61803746, Global Avg Loss: 1.36954540, Time: 0.0127 Steps: 76180, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001675, Sample Num: 26800, Cur Loss: 0.37592769, Cur Avg Loss: 0.51010264, Log Avg loss: 0.56753773, Global Avg Loss: 1.36944013, Time: 0.0066 Steps: 76190, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001685, Sample Num: 26960, Cur Loss: 0.59097350, Cur Avg Loss: 0.51052477, Log Avg loss: 0.58123122, Global Avg Loss: 1.36933669, Time: 0.0067 Steps: 76200, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001695, Sample Num: 27120, Cur Loss: 0.99742460, Cur Avg Loss: 0.51091004, Log Avg loss: 0.57582761, Global Avg Loss: 1.36923257, Time: 0.0113 Steps: 76210, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001705, Sample Num: 27280, Cur Loss: 0.75491142, Cur Avg Loss: 0.51159328, Log Avg loss: 0.62740257, Global Avg Loss: 1.36913524, Time: 0.0068 Steps: 76220, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001715, Sample Num: 27440, Cur Loss: 0.52966750, Cur Avg Loss: 0.51328565, Log Avg loss: 0.80183530, Global Avg Loss: 1.36906082, Time: 0.0105 Steps: 76230, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001725, Sample Num: 27600, Cur Loss: 0.42694914, Cur Avg Loss: 0.51306355, Log Avg loss: 0.47497342, Global Avg Loss: 1.36894355, Time: 0.0066 Steps: 76240, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001735, Sample Num: 27760, Cur Loss: 0.36667824, Cur Avg Loss: 0.51300919, Log Avg loss: 0.50363204, Global Avg Loss: 1.36883007, Time: 0.0113 Steps: 76250, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001745, Sample Num: 27920, Cur Loss: 0.27256817, Cur Avg Loss: 0.51223737, Log Avg loss: 0.37832605, Global Avg Loss: 1.36870018, Time: 0.0158 Steps: 76260, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001755, Sample Num: 28080, Cur Loss: 0.30759120, Cur Avg Loss: 0.51276947, Log Avg loss: 0.60562150, Global Avg Loss: 1.36860013, Time: 0.0065 Steps: 76270, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001765, Sample Num: 28240, Cur Loss: 0.51917303, Cur Avg Loss: 0.51208841, Log Avg loss: 0.39256171, Global Avg Loss: 1.36847218, Time: 0.0066 Steps: 76280, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001775, Sample Num: 28400, Cur Loss: 0.34468371, Cur Avg Loss: 0.51195593, Log Avg loss: 0.48857380, Global Avg Loss: 1.36835684, Time: 0.0104 Steps: 76290, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001785, Sample Num: 28560, Cur Loss: 0.15798733, Cur Avg Loss: 0.51168574, Log Avg loss: 0.46372669, Global Avg Loss: 1.36823828, Time: 0.0106 Steps: 76300, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001795, Sample Num: 28720, Cur Loss: 0.74046957, Cur Avg Loss: 0.51092635, Log Avg loss: 0.37537470, Global Avg Loss: 1.36810817, Time: 0.0147 Steps: 76310, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001805, Sample Num: 28880, Cur Loss: 0.22302210, Cur Avg Loss: 0.51131744, Log Avg loss: 0.58151889, Global Avg Loss: 1.36800511, Time: 0.0074 Steps: 76320, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001815, Sample Num: 29040, Cur Loss: 0.59063900, Cur Avg Loss: 0.51085914, Log Avg loss: 0.42813505, Global Avg Loss: 1.36788197, Time: 0.0150 Steps: 76330, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001825, Sample Num: 29200, Cur Loss: 0.51305956, Cur Avg Loss: 0.51103296, Log Avg loss: 0.54258149, Global Avg Loss: 1.36777387, Time: 0.0070 Steps: 76340, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001835, Sample Num: 29360, Cur Loss: 0.35681811, Cur Avg Loss: 0.51145356, Log Avg loss: 0.58821346, Global Avg Loss: 1.36767176, Time: 0.0231 Steps: 76350, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001845, Sample Num: 29520, Cur Loss: 0.79909992, Cur Avg Loss: 0.51134207, Log Avg loss: 0.49088338, Global Avg Loss: 1.36755694, Time: 0.0167 Steps: 76360, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001855, Sample Num: 29680, Cur Loss: 0.61837196, Cur Avg Loss: 0.51108505, Log Avg loss: 0.46366494, Global Avg Loss: 1.36743858, Time: 0.0066 Steps: 76370, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001865, Sample Num: 29840, Cur Loss: 0.23450622, Cur Avg Loss: 0.51117025, Log Avg loss: 0.52697548, Global Avg Loss: 1.36732855, Time: 0.0094 Steps: 76380, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001875, Sample Num: 30000, Cur Loss: 0.35022485, Cur Avg Loss: 0.51108258, Log Avg loss: 0.49473179, Global Avg Loss: 1.36721432, Time: 0.0130 Steps: 76390, Updated lr: 0.000029 Training, Epoch: 0036, Batch: 001885, Sample Num: 30160, Cur Loss: 0.59153521, Cur Avg Loss: 0.51137106, Log Avg loss: 0.56545991, Global Avg Loss: 1.36710937, Time: 0.0117 Steps: 76400, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001895, Sample Num: 30320, Cur Loss: 0.84788787, Cur Avg Loss: 0.51103525, Log Avg loss: 0.44773637, Global Avg Loss: 1.36698905, Time: 0.0116 Steps: 76410, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001905, Sample Num: 30480, Cur Loss: 0.24823463, Cur Avg Loss: 0.51054735, Log Avg loss: 0.41808905, Global Avg Loss: 1.36686488, Time: 0.0139 Steps: 76420, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001915, Sample Num: 30640, Cur Loss: 0.24677286, Cur Avg Loss: 0.51040818, Log Avg loss: 0.48389724, Global Avg Loss: 1.36674936, Time: 0.0070 Steps: 76430, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001925, Sample Num: 30800, Cur Loss: 0.85340130, Cur Avg Loss: 0.51021491, Log Avg loss: 0.47320387, Global Avg Loss: 1.36663246, Time: 0.0066 Steps: 76440, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001935, Sample Num: 30960, Cur Loss: 0.40846515, Cur Avg Loss: 0.50994550, Log Avg loss: 0.45808432, Global Avg Loss: 1.36651362, Time: 0.0067 Steps: 76450, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001945, Sample Num: 31120, Cur Loss: 0.21123892, Cur Avg Loss: 0.50951390, Log Avg loss: 0.42599812, Global Avg Loss: 1.36639061, Time: 0.0066 Steps: 76460, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001955, Sample Num: 31280, Cur Loss: 0.25182080, Cur Avg Loss: 0.50976978, Log Avg loss: 0.55953849, Global Avg Loss: 1.36628510, Time: 0.0132 Steps: 76470, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001965, Sample Num: 31440, Cur Loss: 0.55114311, Cur Avg Loss: 0.50971937, Log Avg loss: 0.49986442, Global Avg Loss: 1.36617181, Time: 0.0066 Steps: 76480, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001975, Sample Num: 31600, Cur Loss: 0.55499715, Cur Avg Loss: 0.51018222, Log Avg loss: 0.60113221, Global Avg Loss: 1.36607180, Time: 0.0081 Steps: 76490, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001985, Sample Num: 31760, Cur Loss: 0.85295653, Cur Avg Loss: 0.51027832, Log Avg loss: 0.52925875, Global Avg Loss: 1.36596241, Time: 0.0165 Steps: 76500, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 001995, Sample Num: 31920, Cur Loss: 0.31507403, Cur Avg Loss: 0.50957772, Log Avg loss: 0.37050784, Global Avg Loss: 1.36583230, Time: 0.0071 Steps: 76510, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002005, Sample Num: 32080, Cur Loss: 0.38406056, Cur Avg Loss: 0.50997090, Log Avg loss: 0.58841179, Global Avg Loss: 1.36573070, Time: 0.0067 Steps: 76520, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002015, Sample Num: 32240, Cur Loss: 0.19803037, Cur Avg Loss: 0.50957318, Log Avg loss: 0.42982934, Global Avg Loss: 1.36560841, Time: 0.0095 Steps: 76530, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002025, Sample Num: 32400, Cur Loss: 0.34387982, Cur Avg Loss: 0.51015322, Log Avg loss: 0.62703057, Global Avg Loss: 1.36551192, Time: 0.0122 Steps: 76540, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002035, Sample Num: 32560, Cur Loss: 0.26554769, Cur Avg Loss: 0.50973892, Log Avg loss: 0.42584297, Global Avg Loss: 1.36538916, Time: 0.0131 Steps: 76550, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002045, Sample Num: 32720, Cur Loss: 0.48957953, Cur Avg Loss: 0.50998579, Log Avg loss: 0.56022450, Global Avg Loss: 1.36528400, Time: 0.0105 Steps: 76560, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002055, Sample Num: 32880, Cur Loss: 0.41689855, Cur Avg Loss: 0.50932525, Log Avg loss: 0.37424435, Global Avg Loss: 1.36515457, Time: 0.0114 Steps: 76570, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002065, Sample Num: 33040, Cur Loss: 0.62737429, Cur Avg Loss: 0.50881181, Log Avg loss: 0.40330047, Global Avg Loss: 1.36502896, Time: 0.0116 Steps: 76580, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002075, Sample Num: 33200, Cur Loss: 0.49554199, Cur Avg Loss: 0.50943107, Log Avg loss: 0.63730883, Global Avg Loss: 1.36493395, Time: 0.0086 Steps: 76590, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002085, Sample Num: 33360, Cur Loss: 0.26337481, Cur Avg Loss: 0.50951865, Log Avg loss: 0.52769158, Global Avg Loss: 1.36482465, Time: 0.0111 Steps: 76600, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002095, Sample Num: 33520, Cur Loss: 0.38129580, Cur Avg Loss: 0.50926840, Log Avg loss: 0.45708985, Global Avg Loss: 1.36470616, Time: 0.0071 Steps: 76610, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002105, Sample Num: 33680, Cur Loss: 0.40849629, Cur Avg Loss: 0.50953650, Log Avg loss: 0.56570379, Global Avg Loss: 1.36460188, Time: 0.0071 Steps: 76620, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002115, Sample Num: 33840, Cur Loss: 0.21446766, Cur Avg Loss: 0.50873831, Log Avg loss: 0.34071918, Global Avg Loss: 1.36446827, Time: 0.0067 Steps: 76630, Updated lr: 0.000028 Training, Epoch: 0036, Batch: 002125, Sample Num: 34000, Cur Loss: 0.17250034, Cur Avg Loss: 0.50842958, Log Avg loss: 0.44313393, Global Avg Loss: 1.36434805, Time: 0.0067 Steps: 76640, Updated lr: 0.000028 ***** Running evaluation checkpoint-76644 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-76644 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.612329, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.601036, "eval_total_loss": 422.528599, "eval_mae": 0.59225, "eval_mse": 0.601189, "eval_r2": 0.617845, "eval_sp_statistic": 0.7451, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.790035, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.869901, "test_total_loss": 436.690374, "test_mae": 0.700289, "test_mse": 0.870084, "test_r2": 0.43844, "test_sp_statistic": 0.582765, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.66267, "test_ps_pvalue": 0.0, "lr": 2.826552868658132e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.3642979038094887, "train_cur_epoch_loss": 1082.026789367199, "train_cur_epoch_avg_loss": 0.5082324045876933, "train_cur_epoch_time": 21.61232876777649, "train_cur_epoch_avg_time": 0.01015139913939713, "epoch": 36, "step": 76644} ################################################## Training, Epoch: 0037, Batch: 000006, Sample Num: 96, Cur Loss: 0.42511636, Cur Avg Loss: 0.50078261, Log Avg loss: 0.46186274, Global Avg Loss: 1.36423031, Time: 0.0090 Steps: 76650, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000016, Sample Num: 256, Cur Loss: 0.62813878, Cur Avg Loss: 0.44301085, Log Avg loss: 0.40834779, Global Avg Loss: 1.36410562, Time: 0.0114 Steps: 76660, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000026, Sample Num: 416, Cur Loss: 0.30412322, Cur Avg Loss: 0.41446222, Log Avg loss: 0.36878442, Global Avg Loss: 1.36397580, Time: 0.0112 Steps: 76670, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000036, Sample Num: 576, Cur Loss: 0.40306103, Cur Avg Loss: 0.47189309, Log Avg loss: 0.62121335, Global Avg Loss: 1.36387893, Time: 0.0111 Steps: 76680, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000046, Sample Num: 736, Cur Loss: 0.45081425, Cur Avg Loss: 0.50297283, Log Avg loss: 0.61485988, Global Avg Loss: 1.36378127, Time: 0.0065 Steps: 76690, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000056, Sample Num: 896, Cur Loss: 0.31019658, Cur Avg Loss: 0.51276373, Log Avg loss: 0.55780189, Global Avg Loss: 1.36367618, Time: 0.0072 Steps: 76700, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000066, Sample Num: 1056, Cur Loss: 0.21675833, Cur Avg Loss: 0.48468151, Log Avg loss: 0.32742107, Global Avg Loss: 1.36354110, Time: 0.0104 Steps: 76710, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000076, Sample Num: 1216, Cur Loss: 0.51023746, Cur Avg Loss: 0.47916833, Log Avg loss: 0.44278133, Global Avg Loss: 1.36342108, Time: 0.0078 Steps: 76720, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000086, Sample Num: 1376, Cur Loss: 0.21921545, Cur Avg Loss: 0.47936746, Log Avg loss: 0.48088083, Global Avg Loss: 1.36330606, Time: 0.0066 Steps: 76730, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000096, Sample Num: 1536, Cur Loss: 0.30704761, Cur Avg Loss: 0.46857149, Log Avg loss: 0.37572616, Global Avg Loss: 1.36317737, Time: 0.0067 Steps: 76740, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000106, Sample Num: 1696, Cur Loss: 0.69970447, Cur Avg Loss: 0.48835928, Log Avg loss: 0.67832212, Global Avg Loss: 1.36308814, Time: 0.0145 Steps: 76750, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000116, Sample Num: 1856, Cur Loss: 0.39037782, Cur Avg Loss: 0.48956050, Log Avg loss: 0.50229340, Global Avg Loss: 1.36297600, Time: 0.0105 Steps: 76760, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000126, Sample Num: 2016, Cur Loss: 0.87240064, Cur Avg Loss: 0.51114120, Log Avg loss: 0.76147731, Global Avg Loss: 1.36289765, Time: 0.0083 Steps: 76770, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000136, Sample Num: 2176, Cur Loss: 0.29334638, Cur Avg Loss: 0.50386327, Log Avg loss: 0.41216129, Global Avg Loss: 1.36277382, Time: 0.0139 Steps: 76780, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000146, Sample Num: 2336, Cur Loss: 1.36017525, Cur Avg Loss: 0.50384709, Log Avg loss: 0.50362706, Global Avg Loss: 1.36266194, Time: 0.0066 Steps: 76790, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000156, Sample Num: 2496, Cur Loss: 0.31751728, Cur Avg Loss: 0.49922154, Log Avg loss: 0.43168857, Global Avg Loss: 1.36254072, Time: 0.0070 Steps: 76800, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000166, Sample Num: 2656, Cur Loss: 0.97217047, Cur Avg Loss: 0.49835098, Log Avg loss: 0.48477020, Global Avg Loss: 1.36242644, Time: 0.0114 Steps: 76810, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000176, Sample Num: 2816, Cur Loss: 0.47613955, Cur Avg Loss: 0.49184566, Log Avg loss: 0.38385736, Global Avg Loss: 1.36229905, Time: 0.0159 Steps: 76820, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000186, Sample Num: 2976, Cur Loss: 0.53210348, Cur Avg Loss: 0.49653587, Log Avg loss: 0.57908361, Global Avg Loss: 1.36219711, Time: 0.0067 Steps: 76830, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000196, Sample Num: 3136, Cur Loss: 0.63998604, Cur Avg Loss: 0.49525712, Log Avg loss: 0.47147239, Global Avg Loss: 1.36208119, Time: 0.0125 Steps: 76840, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000206, Sample Num: 3296, Cur Loss: 0.88536793, Cur Avg Loss: 0.49654295, Log Avg loss: 0.52174509, Global Avg Loss: 1.36197185, Time: 0.0110 Steps: 76850, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000216, Sample Num: 3456, Cur Loss: 0.22784230, Cur Avg Loss: 0.49797507, Log Avg loss: 0.52747678, Global Avg Loss: 1.36186327, Time: 0.0172 Steps: 76860, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000226, Sample Num: 3616, Cur Loss: 0.94170344, Cur Avg Loss: 0.49536932, Log Avg loss: 0.43908520, Global Avg Loss: 1.36174323, Time: 0.0112 Steps: 76870, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000236, Sample Num: 3776, Cur Loss: 1.38720465, Cur Avg Loss: 0.49191290, Log Avg loss: 0.41379787, Global Avg Loss: 1.36161993, Time: 0.0087 Steps: 76880, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000246, Sample Num: 3936, Cur Loss: 0.19729927, Cur Avg Loss: 0.48742441, Log Avg loss: 0.38149591, Global Avg Loss: 1.36149246, Time: 0.0111 Steps: 76890, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000256, Sample Num: 4096, Cur Loss: 0.45386329, Cur Avg Loss: 0.48506723, Log Avg loss: 0.42708067, Global Avg Loss: 1.36137095, Time: 0.0140 Steps: 76900, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000266, Sample Num: 4256, Cur Loss: 0.53383571, Cur Avg Loss: 0.49133079, Log Avg loss: 0.65167790, Global Avg Loss: 1.36127867, Time: 0.0065 Steps: 76910, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000276, Sample Num: 4416, Cur Loss: 0.70146400, Cur Avg Loss: 0.49076554, Log Avg loss: 0.47572987, Global Avg Loss: 1.36116354, Time: 0.0115 Steps: 76920, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000286, Sample Num: 4576, Cur Loss: 0.84776008, Cur Avg Loss: 0.49541773, Log Avg loss: 0.62381825, Global Avg Loss: 1.36106770, Time: 0.0137 Steps: 76930, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000296, Sample Num: 4736, Cur Loss: 0.08349511, Cur Avg Loss: 0.49294017, Log Avg loss: 0.42208174, Global Avg Loss: 1.36094566, Time: 0.0106 Steps: 76940, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000306, Sample Num: 4896, Cur Loss: 0.70842612, Cur Avg Loss: 0.49171095, Log Avg loss: 0.45532615, Global Avg Loss: 1.36082797, Time: 0.0184 Steps: 76950, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000316, Sample Num: 5056, Cur Loss: 0.21024892, Cur Avg Loss: 0.48778858, Log Avg loss: 0.36776424, Global Avg Loss: 1.36069893, Time: 0.0110 Steps: 76960, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000326, Sample Num: 5216, Cur Loss: 0.91536438, Cur Avg Loss: 0.49212595, Log Avg loss: 0.62918671, Global Avg Loss: 1.36060389, Time: 0.0116 Steps: 76970, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000336, Sample Num: 5376, Cur Loss: 0.56618780, Cur Avg Loss: 0.49644665, Log Avg loss: 0.63730153, Global Avg Loss: 1.36050993, Time: 0.0074 Steps: 76980, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000346, Sample Num: 5536, Cur Loss: 0.34085226, Cur Avg Loss: 0.49649176, Log Avg loss: 0.49800748, Global Avg Loss: 1.36039790, Time: 0.0066 Steps: 76990, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000356, Sample Num: 5696, Cur Loss: 0.41534382, Cur Avg Loss: 0.49632519, Log Avg loss: 0.49056193, Global Avg Loss: 1.36028494, Time: 0.0071 Steps: 77000, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000366, Sample Num: 5856, Cur Loss: 0.38573182, Cur Avg Loss: 0.49492344, Log Avg loss: 0.44502110, Global Avg Loss: 1.36016609, Time: 0.0102 Steps: 77010, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000376, Sample Num: 6016, Cur Loss: 0.47296935, Cur Avg Loss: 0.49508777, Log Avg loss: 0.50110202, Global Avg Loss: 1.36005455, Time: 0.0068 Steps: 77020, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000386, Sample Num: 6176, Cur Loss: 0.59110886, Cur Avg Loss: 0.49389068, Log Avg loss: 0.44888034, Global Avg Loss: 1.35993626, Time: 0.0114 Steps: 77030, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000396, Sample Num: 6336, Cur Loss: 0.62929177, Cur Avg Loss: 0.49295600, Log Avg loss: 0.45687716, Global Avg Loss: 1.35981904, Time: 0.0113 Steps: 77040, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000406, Sample Num: 6496, Cur Loss: 0.21072747, Cur Avg Loss: 0.49377245, Log Avg loss: 0.52610406, Global Avg Loss: 1.35971084, Time: 0.0121 Steps: 77050, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000416, Sample Num: 6656, Cur Loss: 0.60899603, Cur Avg Loss: 0.49485921, Log Avg loss: 0.53898135, Global Avg Loss: 1.35960433, Time: 0.0069 Steps: 77060, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000426, Sample Num: 6816, Cur Loss: 0.68156362, Cur Avg Loss: 0.49390134, Log Avg loss: 0.45405423, Global Avg Loss: 1.35948684, Time: 0.0073 Steps: 77070, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000436, Sample Num: 6976, Cur Loss: 0.50406533, Cur Avg Loss: 0.49343579, Log Avg loss: 0.47360343, Global Avg Loss: 1.35937191, Time: 0.0116 Steps: 77080, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000446, Sample Num: 7136, Cur Loss: 0.87221634, Cur Avg Loss: 0.49546798, Log Avg loss: 0.58407139, Global Avg Loss: 1.35927134, Time: 0.0109 Steps: 77090, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000456, Sample Num: 7296, Cur Loss: 0.11616141, Cur Avg Loss: 0.49147429, Log Avg loss: 0.31335542, Global Avg Loss: 1.35913568, Time: 0.0114 Steps: 77100, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000466, Sample Num: 7456, Cur Loss: 0.38820300, Cur Avg Loss: 0.49045651, Log Avg loss: 0.44404615, Global Avg Loss: 1.35901701, Time: 0.0122 Steps: 77110, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000476, Sample Num: 7616, Cur Loss: 0.22557181, Cur Avg Loss: 0.48760225, Log Avg loss: 0.35459335, Global Avg Loss: 1.35888676, Time: 0.0068 Steps: 77120, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000486, Sample Num: 7776, Cur Loss: 0.29105914, Cur Avg Loss: 0.48951061, Log Avg loss: 0.58034851, Global Avg Loss: 1.35878583, Time: 0.0109 Steps: 77130, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000496, Sample Num: 7936, Cur Loss: 0.43155438, Cur Avg Loss: 0.49034512, Log Avg loss: 0.53090230, Global Avg Loss: 1.35867850, Time: 0.0072 Steps: 77140, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000506, Sample Num: 8096, Cur Loss: 0.29241565, Cur Avg Loss: 0.49428373, Log Avg loss: 0.68963919, Global Avg Loss: 1.35859178, Time: 0.0067 Steps: 77150, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000516, Sample Num: 8256, Cur Loss: 0.29562205, Cur Avg Loss: 0.49296223, Log Avg loss: 0.42609423, Global Avg Loss: 1.35847093, Time: 0.0109 Steps: 77160, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000526, Sample Num: 8416, Cur Loss: 0.58137417, Cur Avg Loss: 0.49318177, Log Avg loss: 0.50450978, Global Avg Loss: 1.35836027, Time: 0.0067 Steps: 77170, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000536, Sample Num: 8576, Cur Loss: 0.23031622, Cur Avg Loss: 0.49220696, Log Avg loss: 0.44093208, Global Avg Loss: 1.35824140, Time: 0.0072 Steps: 77180, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000546, Sample Num: 8736, Cur Loss: 0.24343011, Cur Avg Loss: 0.48904233, Log Avg loss: 0.31941804, Global Avg Loss: 1.35810682, Time: 0.0106 Steps: 77190, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000556, Sample Num: 8896, Cur Loss: 0.60210311, Cur Avg Loss: 0.48684102, Log Avg loss: 0.36664950, Global Avg Loss: 1.35797840, Time: 0.0117 Steps: 77200, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000566, Sample Num: 9056, Cur Loss: 0.34248036, Cur Avg Loss: 0.48664645, Log Avg loss: 0.47582867, Global Avg Loss: 1.35786414, Time: 0.0072 Steps: 77210, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000576, Sample Num: 9216, Cur Loss: 0.17825794, Cur Avg Loss: 0.48653990, Log Avg loss: 0.48050908, Global Avg Loss: 1.35775052, Time: 0.0156 Steps: 77220, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000586, Sample Num: 9376, Cur Loss: 0.90070283, Cur Avg Loss: 0.48695721, Log Avg loss: 0.51099420, Global Avg Loss: 1.35764088, Time: 0.0157 Steps: 77230, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000596, Sample Num: 9536, Cur Loss: 0.71590215, Cur Avg Loss: 0.48627249, Log Avg loss: 0.44614813, Global Avg Loss: 1.35752288, Time: 0.0165 Steps: 77240, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000606, Sample Num: 9696, Cur Loss: 0.43868852, Cur Avg Loss: 0.48655481, Log Avg loss: 0.50338058, Global Avg Loss: 1.35741231, Time: 0.0074 Steps: 77250, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000616, Sample Num: 9856, Cur Loss: 0.67174762, Cur Avg Loss: 0.48611424, Log Avg loss: 0.45941576, Global Avg Loss: 1.35729608, Time: 0.0120 Steps: 77260, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000626, Sample Num: 10016, Cur Loss: 0.87659168, Cur Avg Loss: 0.48623116, Log Avg loss: 0.49343356, Global Avg Loss: 1.35718428, Time: 0.0085 Steps: 77270, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000636, Sample Num: 10176, Cur Loss: 0.90589088, Cur Avg Loss: 0.48670809, Log Avg loss: 0.51656376, Global Avg Loss: 1.35707550, Time: 0.0122 Steps: 77280, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000646, Sample Num: 10336, Cur Loss: 0.60416454, Cur Avg Loss: 0.48694665, Log Avg loss: 0.50211902, Global Avg Loss: 1.35696489, Time: 0.0068 Steps: 77290, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000656, Sample Num: 10496, Cur Loss: 0.33351994, Cur Avg Loss: 0.48783865, Log Avg loss: 0.54546197, Global Avg Loss: 1.35685991, Time: 0.0112 Steps: 77300, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000666, Sample Num: 10656, Cur Loss: 0.64586639, Cur Avg Loss: 0.48933833, Log Avg loss: 0.58771726, Global Avg Loss: 1.35676042, Time: 0.0066 Steps: 77310, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000676, Sample Num: 10816, Cur Loss: 0.59145695, Cur Avg Loss: 0.48803156, Log Avg loss: 0.40100062, Global Avg Loss: 1.35663681, Time: 0.0111 Steps: 77320, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000686, Sample Num: 10976, Cur Loss: 0.12784481, Cur Avg Loss: 0.48633471, Log Avg loss: 0.37162777, Global Avg Loss: 1.35650943, Time: 0.0117 Steps: 77330, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000696, Sample Num: 11136, Cur Loss: 0.72488213, Cur Avg Loss: 0.48517229, Log Avg loss: 0.40543057, Global Avg Loss: 1.35638646, Time: 0.0085 Steps: 77340, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000706, Sample Num: 11296, Cur Loss: 0.34674099, Cur Avg Loss: 0.48619542, Log Avg loss: 0.55740543, Global Avg Loss: 1.35628316, Time: 0.0123 Steps: 77350, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000716, Sample Num: 11456, Cur Loss: 0.46152788, Cur Avg Loss: 0.48510777, Log Avg loss: 0.40831963, Global Avg Loss: 1.35616062, Time: 0.0073 Steps: 77360, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000726, Sample Num: 11616, Cur Loss: 0.56545770, Cur Avg Loss: 0.48519742, Log Avg loss: 0.49161616, Global Avg Loss: 1.35604888, Time: 0.0065 Steps: 77370, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000736, Sample Num: 11776, Cur Loss: 0.69081497, Cur Avg Loss: 0.48707880, Log Avg loss: 0.62366730, Global Avg Loss: 1.35595423, Time: 0.0069 Steps: 77380, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000746, Sample Num: 11936, Cur Loss: 0.14981277, Cur Avg Loss: 0.48727343, Log Avg loss: 0.50159819, Global Avg Loss: 1.35584384, Time: 0.0065 Steps: 77390, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000756, Sample Num: 12096, Cur Loss: 0.12961760, Cur Avg Loss: 0.48739227, Log Avg loss: 0.49625767, Global Avg Loss: 1.35573278, Time: 0.0065 Steps: 77400, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000766, Sample Num: 12256, Cur Loss: 0.34543175, Cur Avg Loss: 0.48882098, Log Avg loss: 0.59683092, Global Avg Loss: 1.35563474, Time: 0.0067 Steps: 77410, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000776, Sample Num: 12416, Cur Loss: 0.60263753, Cur Avg Loss: 0.48860184, Log Avg loss: 0.47181583, Global Avg Loss: 1.35552058, Time: 0.0122 Steps: 77420, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000786, Sample Num: 12576, Cur Loss: 0.92384511, Cur Avg Loss: 0.48980647, Log Avg loss: 0.58328563, Global Avg Loss: 1.35542085, Time: 0.0114 Steps: 77430, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000796, Sample Num: 12736, Cur Loss: 0.72495860, Cur Avg Loss: 0.49050620, Log Avg loss: 0.54550563, Global Avg Loss: 1.35531626, Time: 0.0118 Steps: 77440, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000806, Sample Num: 12896, Cur Loss: 0.37188941, Cur Avg Loss: 0.49129948, Log Avg loss: 0.55444392, Global Avg Loss: 1.35521286, Time: 0.0137 Steps: 77450, Updated lr: 0.000028 Training, Epoch: 0037, Batch: 000816, Sample Num: 13056, Cur Loss: 0.56399900, Cur Avg Loss: 0.49116871, Log Avg loss: 0.48062865, Global Avg Loss: 1.35509995, Time: 0.0124 Steps: 77460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000826, Sample Num: 13216, Cur Loss: 0.26186672, Cur Avg Loss: 0.49099168, Log Avg loss: 0.47654655, Global Avg Loss: 1.35498655, Time: 0.0142 Steps: 77470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000836, Sample Num: 13376, Cur Loss: 0.38623458, Cur Avg Loss: 0.49218834, Log Avg loss: 0.59103235, Global Avg Loss: 1.35488795, Time: 0.0122 Steps: 77480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000846, Sample Num: 13536, Cur Loss: 0.69855452, Cur Avg Loss: 0.49397120, Log Avg loss: 0.64301808, Global Avg Loss: 1.35479608, Time: 0.0119 Steps: 77490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000856, Sample Num: 13696, Cur Loss: 0.19923805, Cur Avg Loss: 0.49444425, Log Avg loss: 0.53446457, Global Avg Loss: 1.35469023, Time: 0.0131 Steps: 77500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000866, Sample Num: 13856, Cur Loss: 0.81147164, Cur Avg Loss: 0.49629592, Log Avg loss: 0.65479890, Global Avg Loss: 1.35459993, Time: 0.0108 Steps: 77510, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000876, Sample Num: 14016, Cur Loss: 0.52426296, Cur Avg Loss: 0.49636516, Log Avg loss: 0.50236139, Global Avg Loss: 1.35449000, Time: 0.0115 Steps: 77520, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000886, Sample Num: 14176, Cur Loss: 0.36463410, Cur Avg Loss: 0.49466995, Log Avg loss: 0.34616895, Global Avg Loss: 1.35435994, Time: 0.0095 Steps: 77530, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000896, Sample Num: 14336, Cur Loss: 1.20369053, Cur Avg Loss: 0.49597405, Log Avg loss: 0.61151790, Global Avg Loss: 1.35426414, Time: 0.0130 Steps: 77540, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000906, Sample Num: 14496, Cur Loss: 0.67298681, Cur Avg Loss: 0.49826935, Log Avg loss: 0.70392767, Global Avg Loss: 1.35418028, Time: 0.0155 Steps: 77550, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000916, Sample Num: 14656, Cur Loss: 0.10637669, Cur Avg Loss: 0.49819351, Log Avg loss: 0.49132282, Global Avg Loss: 1.35406903, Time: 0.0066 Steps: 77560, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000926, Sample Num: 14816, Cur Loss: 0.32698441, Cur Avg Loss: 0.49864867, Log Avg loss: 0.54034092, Global Avg Loss: 1.35396413, Time: 0.0112 Steps: 77570, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000936, Sample Num: 14976, Cur Loss: 0.37109238, Cur Avg Loss: 0.49821863, Log Avg loss: 0.45839698, Global Avg Loss: 1.35384869, Time: 0.0155 Steps: 77580, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000946, Sample Num: 15136, Cur Loss: 0.49986908, Cur Avg Loss: 0.49758013, Log Avg loss: 0.43781673, Global Avg Loss: 1.35373063, Time: 0.0159 Steps: 77590, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000956, Sample Num: 15296, Cur Loss: 0.42540500, Cur Avg Loss: 0.49797535, Log Avg loss: 0.53536278, Global Avg Loss: 1.35362517, Time: 0.0177 Steps: 77600, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000966, Sample Num: 15456, Cur Loss: 0.50292337, Cur Avg Loss: 0.49874416, Log Avg loss: 0.57224317, Global Avg Loss: 1.35352449, Time: 0.0133 Steps: 77610, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000976, Sample Num: 15616, Cur Loss: 0.49240530, Cur Avg Loss: 0.49889072, Log Avg loss: 0.51304756, Global Avg Loss: 1.35341621, Time: 0.0131 Steps: 77620, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000986, Sample Num: 15776, Cur Loss: 0.48593327, Cur Avg Loss: 0.49893097, Log Avg loss: 0.50285993, Global Avg Loss: 1.35330664, Time: 0.0122 Steps: 77630, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 000996, Sample Num: 15936, Cur Loss: 0.48290846, Cur Avg Loss: 0.49811667, Log Avg loss: 0.41782696, Global Avg Loss: 1.35318615, Time: 0.0116 Steps: 77640, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001006, Sample Num: 16096, Cur Loss: 0.50051188, Cur Avg Loss: 0.49775313, Log Avg loss: 0.46154370, Global Avg Loss: 1.35307132, Time: 0.0067 Steps: 77650, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001016, Sample Num: 16256, Cur Loss: 0.82261437, Cur Avg Loss: 0.49793367, Log Avg loss: 0.51609662, Global Avg Loss: 1.35296355, Time: 0.0068 Steps: 77660, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001026, Sample Num: 16416, Cur Loss: 0.08481391, Cur Avg Loss: 0.49798348, Log Avg loss: 0.50304420, Global Avg Loss: 1.35285412, Time: 0.0143 Steps: 77670, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001036, Sample Num: 16576, Cur Loss: 0.25265354, Cur Avg Loss: 0.49937355, Log Avg loss: 0.64199486, Global Avg Loss: 1.35276261, Time: 0.0118 Steps: 77680, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001046, Sample Num: 16736, Cur Loss: 0.92665398, Cur Avg Loss: 0.49958278, Log Avg loss: 0.52125885, Global Avg Loss: 1.35265558, Time: 0.0089 Steps: 77690, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001056, Sample Num: 16896, Cur Loss: 0.72195596, Cur Avg Loss: 0.49972985, Log Avg loss: 0.51511368, Global Avg Loss: 1.35254779, Time: 0.0066 Steps: 77700, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001066, Sample Num: 17056, Cur Loss: 0.31708622, Cur Avg Loss: 0.49900770, Log Avg loss: 0.42274864, Global Avg Loss: 1.35242814, Time: 0.0115 Steps: 77710, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001076, Sample Num: 17216, Cur Loss: 1.60265326, Cur Avg Loss: 0.50123741, Log Avg loss: 0.73892442, Global Avg Loss: 1.35234920, Time: 0.0133 Steps: 77720, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001086, Sample Num: 17376, Cur Loss: 0.31262636, Cur Avg Loss: 0.50109342, Log Avg loss: 0.48559992, Global Avg Loss: 1.35223769, Time: 0.0119 Steps: 77730, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001096, Sample Num: 17536, Cur Loss: 0.66484749, Cur Avg Loss: 0.50167960, Log Avg loss: 0.56533876, Global Avg Loss: 1.35213647, Time: 0.0066 Steps: 77740, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001106, Sample Num: 17696, Cur Loss: 0.39312303, Cur Avg Loss: 0.50213251, Log Avg loss: 0.55177096, Global Avg Loss: 1.35203353, Time: 0.0172 Steps: 77750, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001116, Sample Num: 17856, Cur Loss: 0.45609349, Cur Avg Loss: 0.50116788, Log Avg loss: 0.39448061, Global Avg Loss: 1.35191039, Time: 0.0108 Steps: 77760, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001126, Sample Num: 18016, Cur Loss: 0.37262475, Cur Avg Loss: 0.50202854, Log Avg loss: 0.59807747, Global Avg Loss: 1.35181346, Time: 0.0240 Steps: 77770, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001136, Sample Num: 18176, Cur Loss: 0.29713547, Cur Avg Loss: 0.50128622, Log Avg loss: 0.41770134, Global Avg Loss: 1.35169336, Time: 0.0113 Steps: 77780, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001146, Sample Num: 18336, Cur Loss: 0.50154841, Cur Avg Loss: 0.50145377, Log Avg loss: 0.52048706, Global Avg Loss: 1.35158651, Time: 0.0064 Steps: 77790, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001156, Sample Num: 18496, Cur Loss: 0.61909342, Cur Avg Loss: 0.50131207, Log Avg loss: 0.48507342, Global Avg Loss: 1.35147513, Time: 0.0178 Steps: 77800, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001166, Sample Num: 18656, Cur Loss: 0.29484275, Cur Avg Loss: 0.50227574, Log Avg loss: 0.61367580, Global Avg Loss: 1.35138031, Time: 0.0113 Steps: 77810, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001176, Sample Num: 18816, Cur Loss: 0.28791153, Cur Avg Loss: 0.50163558, Log Avg loss: 0.42699375, Global Avg Loss: 1.35126153, Time: 0.0066 Steps: 77820, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001186, Sample Num: 18976, Cur Loss: 0.40026385, Cur Avg Loss: 0.50116370, Log Avg loss: 0.44567010, Global Avg Loss: 1.35114517, Time: 0.0069 Steps: 77830, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001196, Sample Num: 19136, Cur Loss: 0.97633696, Cur Avg Loss: 0.50031093, Log Avg loss: 0.39917234, Global Avg Loss: 1.35102287, Time: 0.0066 Steps: 77840, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001206, Sample Num: 19296, Cur Loss: 0.91142160, Cur Avg Loss: 0.50027449, Log Avg loss: 0.49591597, Global Avg Loss: 1.35091303, Time: 0.0074 Steps: 77850, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001216, Sample Num: 19456, Cur Loss: 0.23715451, Cur Avg Loss: 0.50096136, Log Avg loss: 0.58379877, Global Avg Loss: 1.35081451, Time: 0.0073 Steps: 77860, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001226, Sample Num: 19616, Cur Loss: 0.32561949, Cur Avg Loss: 0.50183171, Log Avg loss: 0.60766527, Global Avg Loss: 1.35071907, Time: 0.0072 Steps: 77870, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001236, Sample Num: 19776, Cur Loss: 0.26903105, Cur Avg Loss: 0.50139979, Log Avg loss: 0.44844687, Global Avg Loss: 1.35060322, Time: 0.0074 Steps: 77880, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001246, Sample Num: 19936, Cur Loss: 0.86112076, Cur Avg Loss: 0.50196923, Log Avg loss: 0.57235163, Global Avg Loss: 1.35050330, Time: 0.0068 Steps: 77890, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001256, Sample Num: 20096, Cur Loss: 0.34980828, Cur Avg Loss: 0.50108042, Log Avg loss: 0.39033484, Global Avg Loss: 1.35038005, Time: 0.0073 Steps: 77900, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001266, Sample Num: 20256, Cur Loss: 0.49633360, Cur Avg Loss: 0.50104788, Log Avg loss: 0.49696157, Global Avg Loss: 1.35027051, Time: 0.0189 Steps: 77910, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001276, Sample Num: 20416, Cur Loss: 1.09958982, Cur Avg Loss: 0.50144752, Log Avg loss: 0.55204093, Global Avg Loss: 1.35016806, Time: 0.0159 Steps: 77920, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001286, Sample Num: 20576, Cur Loss: 0.26707819, Cur Avg Loss: 0.50135029, Log Avg loss: 0.48894374, Global Avg Loss: 1.35005755, Time: 0.0158 Steps: 77930, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001296, Sample Num: 20736, Cur Loss: 0.36178142, Cur Avg Loss: 0.50100293, Log Avg loss: 0.45633347, Global Avg Loss: 1.34994288, Time: 0.0115 Steps: 77940, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001306, Sample Num: 20896, Cur Loss: 0.26836014, Cur Avg Loss: 0.50032648, Log Avg loss: 0.41265806, Global Avg Loss: 1.34982264, Time: 0.0068 Steps: 77950, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001316, Sample Num: 21056, Cur Loss: 0.25099623, Cur Avg Loss: 0.50125753, Log Avg loss: 0.62285256, Global Avg Loss: 1.34972939, Time: 0.0156 Steps: 77960, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001326, Sample Num: 21216, Cur Loss: 0.42005366, Cur Avg Loss: 0.50160396, Log Avg loss: 0.54719379, Global Avg Loss: 1.34962646, Time: 0.0110 Steps: 77970, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001336, Sample Num: 21376, Cur Loss: 0.45146498, Cur Avg Loss: 0.50069556, Log Avg loss: 0.38024250, Global Avg Loss: 1.34950215, Time: 0.0076 Steps: 77980, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001346, Sample Num: 21536, Cur Loss: 0.63435936, Cur Avg Loss: 0.50115474, Log Avg loss: 0.56250071, Global Avg Loss: 1.34940124, Time: 0.0068 Steps: 77990, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001356, Sample Num: 21696, Cur Loss: 0.38908577, Cur Avg Loss: 0.50070889, Log Avg loss: 0.44069790, Global Avg Loss: 1.34928474, Time: 0.0128 Steps: 78000, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001366, Sample Num: 21856, Cur Loss: 0.67284441, Cur Avg Loss: 0.50048787, Log Avg loss: 0.47051794, Global Avg Loss: 1.34917209, Time: 0.0068 Steps: 78010, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001376, Sample Num: 22016, Cur Loss: 0.46777242, Cur Avg Loss: 0.50044622, Log Avg loss: 0.49475635, Global Avg Loss: 1.34906258, Time: 0.0156 Steps: 78020, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001386, Sample Num: 22176, Cur Loss: 1.13854051, Cur Avg Loss: 0.50128980, Log Avg loss: 0.61736597, Global Avg Loss: 1.34896881, Time: 0.0118 Steps: 78030, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001396, Sample Num: 22336, Cur Loss: 0.69665372, Cur Avg Loss: 0.50144290, Log Avg loss: 0.52266231, Global Avg Loss: 1.34886293, Time: 0.0069 Steps: 78040, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001406, Sample Num: 22496, Cur Loss: 0.30530763, Cur Avg Loss: 0.50136168, Log Avg loss: 0.49002373, Global Avg Loss: 1.34875289, Time: 0.0067 Steps: 78050, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001416, Sample Num: 22656, Cur Loss: 0.41456881, Cur Avg Loss: 0.50111664, Log Avg loss: 0.46666469, Global Avg Loss: 1.34863989, Time: 0.0070 Steps: 78060, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001426, Sample Num: 22816, Cur Loss: 0.47998291, Cur Avg Loss: 0.50076706, Log Avg loss: 0.45126666, Global Avg Loss: 1.34852494, Time: 0.0108 Steps: 78070, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001436, Sample Num: 22976, Cur Loss: 0.27693021, Cur Avg Loss: 0.50006512, Log Avg loss: 0.39996740, Global Avg Loss: 1.34840346, Time: 0.0066 Steps: 78080, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001446, Sample Num: 23136, Cur Loss: 0.90210307, Cur Avg Loss: 0.50076544, Log Avg loss: 0.60133115, Global Avg Loss: 1.34830779, Time: 0.0073 Steps: 78090, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001456, Sample Num: 23296, Cur Loss: 0.14891864, Cur Avg Loss: 0.49998605, Log Avg loss: 0.38728723, Global Avg Loss: 1.34818474, Time: 0.0067 Steps: 78100, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001466, Sample Num: 23456, Cur Loss: 0.64600581, Cur Avg Loss: 0.49932753, Log Avg loss: 0.40344683, Global Avg Loss: 1.34806379, Time: 0.0083 Steps: 78110, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001476, Sample Num: 23616, Cur Loss: 0.36355487, Cur Avg Loss: 0.49942992, Log Avg loss: 0.51444083, Global Avg Loss: 1.34795708, Time: 0.0124 Steps: 78120, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001486, Sample Num: 23776, Cur Loss: 0.46062949, Cur Avg Loss: 0.50046152, Log Avg loss: 0.65272549, Global Avg Loss: 1.34786810, Time: 0.0114 Steps: 78130, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001496, Sample Num: 23936, Cur Loss: 0.46017903, Cur Avg Loss: 0.50007044, Log Avg loss: 0.44195577, Global Avg Loss: 1.34775216, Time: 0.0117 Steps: 78140, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001506, Sample Num: 24096, Cur Loss: 0.25176862, Cur Avg Loss: 0.50086211, Log Avg loss: 0.61929574, Global Avg Loss: 1.34765895, Time: 0.0067 Steps: 78150, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001516, Sample Num: 24256, Cur Loss: 0.30220670, Cur Avg Loss: 0.50121922, Log Avg loss: 0.55500028, Global Avg Loss: 1.34755753, Time: 0.0113 Steps: 78160, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001526, Sample Num: 24416, Cur Loss: 0.35143876, Cur Avg Loss: 0.50169274, Log Avg loss: 0.57347819, Global Avg Loss: 1.34745851, Time: 0.0124 Steps: 78170, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001536, Sample Num: 24576, Cur Loss: 0.44129351, Cur Avg Loss: 0.50205209, Log Avg loss: 0.55688937, Global Avg Loss: 1.34735739, Time: 0.0075 Steps: 78180, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001546, Sample Num: 24736, Cur Loss: 0.42855349, Cur Avg Loss: 0.50225216, Log Avg loss: 0.53298259, Global Avg Loss: 1.34725323, Time: 0.0083 Steps: 78190, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001556, Sample Num: 24896, Cur Loss: 0.56498927, Cur Avg Loss: 0.50211949, Log Avg loss: 0.48160840, Global Avg Loss: 1.34714254, Time: 0.0111 Steps: 78200, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001566, Sample Num: 25056, Cur Loss: 0.48904157, Cur Avg Loss: 0.50207834, Log Avg loss: 0.49567613, Global Avg Loss: 1.34703367, Time: 0.0119 Steps: 78210, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001576, Sample Num: 25216, Cur Loss: 0.27106926, Cur Avg Loss: 0.50136066, Log Avg loss: 0.38897046, Global Avg Loss: 1.34691119, Time: 0.0108 Steps: 78220, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001586, Sample Num: 25376, Cur Loss: 0.52165067, Cur Avg Loss: 0.50156280, Log Avg loss: 0.53342104, Global Avg Loss: 1.34680720, Time: 0.0078 Steps: 78230, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001596, Sample Num: 25536, Cur Loss: 0.71180278, Cur Avg Loss: 0.50065816, Log Avg loss: 0.35718140, Global Avg Loss: 1.34668071, Time: 0.0066 Steps: 78240, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001606, Sample Num: 25696, Cur Loss: 0.31863660, Cur Avg Loss: 0.50095748, Log Avg loss: 0.54872916, Global Avg Loss: 1.34657874, Time: 0.0065 Steps: 78250, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001616, Sample Num: 25856, Cur Loss: 0.51939678, Cur Avg Loss: 0.50141171, Log Avg loss: 0.57436072, Global Avg Loss: 1.34648006, Time: 0.0065 Steps: 78260, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001626, Sample Num: 26016, Cur Loss: 0.42560506, Cur Avg Loss: 0.50219798, Log Avg loss: 0.62925956, Global Avg Loss: 1.34638843, Time: 0.0067 Steps: 78270, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001636, Sample Num: 26176, Cur Loss: 0.38102058, Cur Avg Loss: 0.50218690, Log Avg loss: 0.50038489, Global Avg Loss: 1.34628036, Time: 0.0067 Steps: 78280, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001646, Sample Num: 26336, Cur Loss: 0.24116431, Cur Avg Loss: 0.50242528, Log Avg loss: 0.54142470, Global Avg Loss: 1.34617755, Time: 0.0068 Steps: 78290, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001656, Sample Num: 26496, Cur Loss: 0.78661513, Cur Avg Loss: 0.50229111, Log Avg loss: 0.48020646, Global Avg Loss: 1.34606696, Time: 0.0086 Steps: 78300, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001666, Sample Num: 26656, Cur Loss: 0.60413736, Cur Avg Loss: 0.50311081, Log Avg loss: 0.63885430, Global Avg Loss: 1.34597665, Time: 0.0069 Steps: 78310, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001676, Sample Num: 26816, Cur Loss: 0.27129665, Cur Avg Loss: 0.50336265, Log Avg loss: 0.54531912, Global Avg Loss: 1.34587442, Time: 0.0068 Steps: 78320, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001686, Sample Num: 26976, Cur Loss: 0.35452849, Cur Avg Loss: 0.50251475, Log Avg loss: 0.36040678, Global Avg Loss: 1.34574861, Time: 0.0074 Steps: 78330, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001696, Sample Num: 27136, Cur Loss: 0.54171771, Cur Avg Loss: 0.50271204, Log Avg loss: 0.53597480, Global Avg Loss: 1.34564524, Time: 0.0073 Steps: 78340, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001706, Sample Num: 27296, Cur Loss: 0.33964467, Cur Avg Loss: 0.50360051, Log Avg loss: 0.65428425, Global Avg Loss: 1.34555700, Time: 0.0072 Steps: 78350, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001716, Sample Num: 27456, Cur Loss: 0.35039192, Cur Avg Loss: 0.50376095, Log Avg loss: 0.53113233, Global Avg Loss: 1.34545307, Time: 0.0125 Steps: 78360, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001726, Sample Num: 27616, Cur Loss: 0.41167447, Cur Avg Loss: 0.50344816, Log Avg loss: 0.44977298, Global Avg Loss: 1.34533878, Time: 0.0065 Steps: 78370, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001736, Sample Num: 27776, Cur Loss: 0.84120655, Cur Avg Loss: 0.50409577, Log Avg loss: 0.61587351, Global Avg Loss: 1.34524571, Time: 0.0074 Steps: 78380, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001746, Sample Num: 27936, Cur Loss: 0.35053921, Cur Avg Loss: 0.50433181, Log Avg loss: 0.54530934, Global Avg Loss: 1.34514366, Time: 0.0071 Steps: 78390, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001756, Sample Num: 28096, Cur Loss: 0.35303578, Cur Avg Loss: 0.50474314, Log Avg loss: 0.57656038, Global Avg Loss: 1.34504563, Time: 0.0116 Steps: 78400, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001766, Sample Num: 28256, Cur Loss: 0.29868197, Cur Avg Loss: 0.50461298, Log Avg loss: 0.48175691, Global Avg Loss: 1.34493553, Time: 0.0071 Steps: 78410, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001776, Sample Num: 28416, Cur Loss: 0.23966561, Cur Avg Loss: 0.50443054, Log Avg loss: 0.47221146, Global Avg Loss: 1.34482424, Time: 0.0107 Steps: 78420, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001786, Sample Num: 28576, Cur Loss: 0.63938457, Cur Avg Loss: 0.50405532, Log Avg loss: 0.43741654, Global Avg Loss: 1.34470855, Time: 0.0115 Steps: 78430, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001796, Sample Num: 28736, Cur Loss: 0.14856786, Cur Avg Loss: 0.50423534, Log Avg loss: 0.53638781, Global Avg Loss: 1.34460550, Time: 0.0068 Steps: 78440, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001806, Sample Num: 28896, Cur Loss: 2.15887451, Cur Avg Loss: 0.50580687, Log Avg loss: 0.78805245, Global Avg Loss: 1.34453455, Time: 0.0118 Steps: 78450, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001816, Sample Num: 29056, Cur Loss: 0.55903816, Cur Avg Loss: 0.50604592, Log Avg loss: 0.54921962, Global Avg Loss: 1.34443319, Time: 0.0109 Steps: 78460, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001826, Sample Num: 29216, Cur Loss: 0.50513315, Cur Avg Loss: 0.50605447, Log Avg loss: 0.50760707, Global Avg Loss: 1.34432655, Time: 0.0068 Steps: 78470, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001836, Sample Num: 29376, Cur Loss: 0.53726482, Cur Avg Loss: 0.50583653, Log Avg loss: 0.46603919, Global Avg Loss: 1.34421463, Time: 0.0113 Steps: 78480, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001846, Sample Num: 29536, Cur Loss: 0.41655579, Cur Avg Loss: 0.50544277, Log Avg loss: 0.43314997, Global Avg Loss: 1.34409856, Time: 0.0156 Steps: 78490, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001856, Sample Num: 29696, Cur Loss: 0.77471256, Cur Avg Loss: 0.50534590, Log Avg loss: 0.48746264, Global Avg Loss: 1.34398943, Time: 0.0068 Steps: 78500, Updated lr: 0.000027 Training, Epoch: 0037, Batch: 001866, Sample Num: 29856, Cur Loss: 0.53172779, Cur Avg Loss: 0.50507281, Log Avg loss: 0.45438856, Global Avg Loss: 1.34387612, Time: 0.0067 Steps: 78510, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001876, Sample Num: 30016, Cur Loss: 0.64135027, Cur Avg Loss: 0.50558287, Log Avg loss: 0.60075849, Global Avg Loss: 1.34378148, Time: 0.0068 Steps: 78520, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001886, Sample Num: 30176, Cur Loss: 0.15433258, Cur Avg Loss: 0.50532620, Log Avg loss: 0.45717535, Global Avg Loss: 1.34366858, Time: 0.0141 Steps: 78530, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001896, Sample Num: 30336, Cur Loss: 1.66235328, Cur Avg Loss: 0.50697393, Log Avg loss: 0.81773639, Global Avg Loss: 1.34360162, Time: 0.0071 Steps: 78540, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001906, Sample Num: 30496, Cur Loss: 0.34620133, Cur Avg Loss: 0.50667600, Log Avg loss: 0.45018846, Global Avg Loss: 1.34348788, Time: 0.0067 Steps: 78550, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001916, Sample Num: 30656, Cur Loss: 0.93197274, Cur Avg Loss: 0.50640975, Log Avg loss: 0.45566159, Global Avg Loss: 1.34337487, Time: 0.0067 Steps: 78560, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001926, Sample Num: 30816, Cur Loss: 0.35951272, Cur Avg Loss: 0.50584156, Log Avg loss: 0.39697626, Global Avg Loss: 1.34325441, Time: 0.0067 Steps: 78570, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001936, Sample Num: 30976, Cur Loss: 1.31450760, Cur Avg Loss: 0.50639658, Log Avg loss: 0.61329460, Global Avg Loss: 1.34316152, Time: 0.0113 Steps: 78580, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001946, Sample Num: 31136, Cur Loss: 0.35033071, Cur Avg Loss: 0.50578179, Log Avg loss: 0.38675805, Global Avg Loss: 1.34303983, Time: 0.0067 Steps: 78590, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001956, Sample Num: 31296, Cur Loss: 0.33307555, Cur Avg Loss: 0.50529012, Log Avg loss: 0.40961056, Global Avg Loss: 1.34292107, Time: 0.0071 Steps: 78600, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001966, Sample Num: 31456, Cur Loss: 0.41402000, Cur Avg Loss: 0.50537291, Log Avg loss: 0.52156699, Global Avg Loss: 1.34281658, Time: 0.0067 Steps: 78610, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001976, Sample Num: 31616, Cur Loss: 0.83936477, Cur Avg Loss: 0.50536442, Log Avg loss: 0.50369523, Global Avg Loss: 1.34270985, Time: 0.0130 Steps: 78620, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001986, Sample Num: 31776, Cur Loss: 0.24708262, Cur Avg Loss: 0.50448861, Log Avg loss: 0.33142880, Global Avg Loss: 1.34258124, Time: 0.0109 Steps: 78630, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 001996, Sample Num: 31936, Cur Loss: 0.21320185, Cur Avg Loss: 0.50396427, Log Avg loss: 0.39983093, Global Avg Loss: 1.34246136, Time: 0.0120 Steps: 78640, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002006, Sample Num: 32096, Cur Loss: 0.57806027, Cur Avg Loss: 0.50351771, Log Avg loss: 0.41438420, Global Avg Loss: 1.34234336, Time: 0.0105 Steps: 78650, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002016, Sample Num: 32256, Cur Loss: 0.18711460, Cur Avg Loss: 0.50266973, Log Avg loss: 0.33256465, Global Avg Loss: 1.34221498, Time: 0.0115 Steps: 78660, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002026, Sample Num: 32416, Cur Loss: 0.81341136, Cur Avg Loss: 0.50305439, Log Avg loss: 0.58060202, Global Avg Loss: 1.34211817, Time: 0.0067 Steps: 78670, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002036, Sample Num: 32576, Cur Loss: 0.55859768, Cur Avg Loss: 0.50322198, Log Avg loss: 0.53717599, Global Avg Loss: 1.34201587, Time: 0.0067 Steps: 78680, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002046, Sample Num: 32736, Cur Loss: 0.66869932, Cur Avg Loss: 0.50323085, Log Avg loss: 0.50503652, Global Avg Loss: 1.34190950, Time: 0.0105 Steps: 78690, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002056, Sample Num: 32896, Cur Loss: 0.44871557, Cur Avg Loss: 0.50349005, Log Avg loss: 0.55652156, Global Avg Loss: 1.34180971, Time: 0.0107 Steps: 78700, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002066, Sample Num: 33056, Cur Loss: 0.38049501, Cur Avg Loss: 0.50303502, Log Avg loss: 0.40948114, Global Avg Loss: 1.34169126, Time: 0.0116 Steps: 78710, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002076, Sample Num: 33216, Cur Loss: 0.36288407, Cur Avg Loss: 0.50243828, Log Avg loss: 0.37915207, Global Avg Loss: 1.34156898, Time: 0.0116 Steps: 78720, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002086, Sample Num: 33376, Cur Loss: 0.25987077, Cur Avg Loss: 0.50243600, Log Avg loss: 0.50196273, Global Avg Loss: 1.34146234, Time: 0.0065 Steps: 78730, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002096, Sample Num: 33536, Cur Loss: 0.87904978, Cur Avg Loss: 0.50215878, Log Avg loss: 0.44433068, Global Avg Loss: 1.34134840, Time: 0.0115 Steps: 78740, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002106, Sample Num: 33696, Cur Loss: 0.70156044, Cur Avg Loss: 0.50219727, Log Avg loss: 0.51026405, Global Avg Loss: 1.34124287, Time: 0.0157 Steps: 78750, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002116, Sample Num: 33856, Cur Loss: 0.26823458, Cur Avg Loss: 0.50235817, Log Avg loss: 0.53624485, Global Avg Loss: 1.34114066, Time: 0.0106 Steps: 78760, Updated lr: 0.000026 Training, Epoch: 0037, Batch: 002126, Sample Num: 34016, Cur Loss: 0.25578076, Cur Avg Loss: 0.50205909, Log Avg loss: 0.43877244, Global Avg Loss: 1.34102610, Time: 0.0067 Steps: 78770, Updated lr: 0.000026 ***** Running evaluation checkpoint-78773 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-78773 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.356624, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.594101, "eval_total_loss": 417.653287, "eval_mae": 0.575027, "eval_mse": 0.594274, "eval_r2": 0.622241, "eval_sp_statistic": 0.740429, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.789227, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.881013, "test_total_loss": 442.268323, "test_mae": 0.659174, "test_mse": 0.881247, "test_r2": 0.431235, "test_sp_statistic": 0.593572, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.668856, "test_ps_pvalue": 0.0, "lr": 2.624656235182551e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.3409894703934375, "train_cur_epoch_loss": 1068.51501172781, "train_cur_epoch_avg_loss": 0.5018858674155988, "train_cur_epoch_time": 21.356624364852905, "train_cur_epoch_avg_time": 0.010031293736426916, "epoch": 37, "step": 78773} ################################################## Training, Epoch: 0038, Batch: 000007, Sample Num: 112, Cur Loss: 0.18005492, Cur Avg Loss: 0.47340478, Log Avg loss: 0.44512270, Global Avg Loss: 1.34091238, Time: 0.0138 Steps: 78780, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000017, Sample Num: 272, Cur Loss: 0.70865571, Cur Avg Loss: 0.47450895, Log Avg loss: 0.47528187, Global Avg Loss: 1.34080252, Time: 0.0069 Steps: 78790, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000027, Sample Num: 432, Cur Loss: 0.25827974, Cur Avg Loss: 0.49585617, Log Avg loss: 0.53214643, Global Avg Loss: 1.34069989, Time: 0.0143 Steps: 78800, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000037, Sample Num: 592, Cur Loss: 0.23794484, Cur Avg Loss: 0.45831913, Log Avg loss: 0.35696914, Global Avg Loss: 1.34057507, Time: 0.0073 Steps: 78810, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000047, Sample Num: 752, Cur Loss: 0.41675842, Cur Avg Loss: 0.44662955, Log Avg loss: 0.40337810, Global Avg Loss: 1.34045617, Time: 0.0126 Steps: 78820, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000057, Sample Num: 912, Cur Loss: 0.52869713, Cur Avg Loss: 0.43876450, Log Avg loss: 0.40179875, Global Avg Loss: 1.34033709, Time: 0.0118 Steps: 78830, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000067, Sample Num: 1072, Cur Loss: 0.57461905, Cur Avg Loss: 0.45004819, Log Avg loss: 0.51436522, Global Avg Loss: 1.34023233, Time: 0.0119 Steps: 78840, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000077, Sample Num: 1232, Cur Loss: 0.65858710, Cur Avg Loss: 0.45014251, Log Avg loss: 0.45077450, Global Avg Loss: 1.34011952, Time: 0.0068 Steps: 78850, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000087, Sample Num: 1392, Cur Loss: 0.37564951, Cur Avg Loss: 0.43955844, Log Avg loss: 0.35806106, Global Avg Loss: 1.33999499, Time: 0.0082 Steps: 78860, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000097, Sample Num: 1552, Cur Loss: 0.53728735, Cur Avg Loss: 0.44733052, Log Avg loss: 0.51494766, Global Avg Loss: 1.33989038, Time: 0.0122 Steps: 78870, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000107, Sample Num: 1712, Cur Loss: 0.23870942, Cur Avg Loss: 0.44204288, Log Avg loss: 0.39075271, Global Avg Loss: 1.33977006, Time: 0.0155 Steps: 78880, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000117, Sample Num: 1872, Cur Loss: 0.67209369, Cur Avg Loss: 0.44812036, Log Avg loss: 0.51314939, Global Avg Loss: 1.33966528, Time: 0.0124 Steps: 78890, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000127, Sample Num: 2032, Cur Loss: 0.07227890, Cur Avg Loss: 0.44903989, Log Avg loss: 0.45979848, Global Avg Loss: 1.33955376, Time: 0.0115 Steps: 78900, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000137, Sample Num: 2192, Cur Loss: 0.44137555, Cur Avg Loss: 0.44537295, Log Avg loss: 0.39880278, Global Avg Loss: 1.33943454, Time: 0.0124 Steps: 78910, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000147, Sample Num: 2352, Cur Loss: 0.26092169, Cur Avg Loss: 0.45057136, Log Avg loss: 0.52178960, Global Avg Loss: 1.33933094, Time: 0.0129 Steps: 78920, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000157, Sample Num: 2512, Cur Loss: 0.71119022, Cur Avg Loss: 0.45616248, Log Avg loss: 0.53835197, Global Avg Loss: 1.33922946, Time: 0.0113 Steps: 78930, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000167, Sample Num: 2672, Cur Loss: 0.54645246, Cur Avg Loss: 0.45724564, Log Avg loss: 0.47425119, Global Avg Loss: 1.33911988, Time: 0.0117 Steps: 78940, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000177, Sample Num: 2832, Cur Loss: 0.63101864, Cur Avg Loss: 0.45932709, Log Avg loss: 0.49408720, Global Avg Loss: 1.33901285, Time: 0.0138 Steps: 78950, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000187, Sample Num: 2992, Cur Loss: 0.37067029, Cur Avg Loss: 0.46097274, Log Avg loss: 0.49010080, Global Avg Loss: 1.33890534, Time: 0.0080 Steps: 78960, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000197, Sample Num: 3152, Cur Loss: 0.44988409, Cur Avg Loss: 0.45932820, Log Avg loss: 0.42857536, Global Avg Loss: 1.33879006, Time: 0.0106 Steps: 78970, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000207, Sample Num: 3312, Cur Loss: 0.35647240, Cur Avg Loss: 0.46096544, Log Avg loss: 0.49321904, Global Avg Loss: 1.33868300, Time: 0.0125 Steps: 78980, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000217, Sample Num: 3472, Cur Loss: 0.98367083, Cur Avg Loss: 0.46147569, Log Avg loss: 0.47203784, Global Avg Loss: 1.33857328, Time: 0.0069 Steps: 78990, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000227, Sample Num: 3632, Cur Loss: 0.54114705, Cur Avg Loss: 0.46702921, Log Avg loss: 0.58754066, Global Avg Loss: 1.33847822, Time: 0.0066 Steps: 79000, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000237, Sample Num: 3792, Cur Loss: 0.66704935, Cur Avg Loss: 0.46822121, Log Avg loss: 0.49527956, Global Avg Loss: 1.33837150, Time: 0.0116 Steps: 79010, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000247, Sample Num: 3952, Cur Loss: 0.06880588, Cur Avg Loss: 0.47722456, Log Avg loss: 0.69060386, Global Avg Loss: 1.33828952, Time: 0.0115 Steps: 79020, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000257, Sample Num: 4112, Cur Loss: 0.21004431, Cur Avg Loss: 0.48215780, Log Avg loss: 0.60400903, Global Avg Loss: 1.33819661, Time: 0.0263 Steps: 79030, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000267, Sample Num: 4272, Cur Loss: 0.59033233, Cur Avg Loss: 0.48188767, Log Avg loss: 0.47494531, Global Avg Loss: 1.33808739, Time: 0.0068 Steps: 79040, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000277, Sample Num: 4432, Cur Loss: 0.34217286, Cur Avg Loss: 0.48374355, Log Avg loss: 0.53329545, Global Avg Loss: 1.33798559, Time: 0.0065 Steps: 79050, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000287, Sample Num: 4592, Cur Loss: 0.77122205, Cur Avg Loss: 0.48223381, Log Avg loss: 0.44041406, Global Avg Loss: 1.33787205, Time: 0.0079 Steps: 79060, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000297, Sample Num: 4752, Cur Loss: 0.12727070, Cur Avg Loss: 0.48079153, Log Avg loss: 0.43939813, Global Avg Loss: 1.33775842, Time: 0.0067 Steps: 79070, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000307, Sample Num: 4912, Cur Loss: 0.35645026, Cur Avg Loss: 0.47601480, Log Avg loss: 0.33414586, Global Avg Loss: 1.33763151, Time: 0.0066 Steps: 79080, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000317, Sample Num: 5072, Cur Loss: 0.57262230, Cur Avg Loss: 0.48027734, Log Avg loss: 0.61113736, Global Avg Loss: 1.33753966, Time: 0.0066 Steps: 79090, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000327, Sample Num: 5232, Cur Loss: 0.50437570, Cur Avg Loss: 0.47941380, Log Avg loss: 0.45203958, Global Avg Loss: 1.33742771, Time: 0.0064 Steps: 79100, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000337, Sample Num: 5392, Cur Loss: 0.27511647, Cur Avg Loss: 0.47418540, Log Avg loss: 0.30321660, Global Avg Loss: 1.33729698, Time: 0.0064 Steps: 79110, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000347, Sample Num: 5552, Cur Loss: 0.40976241, Cur Avg Loss: 0.47687386, Log Avg loss: 0.56747488, Global Avg Loss: 1.33719968, Time: 0.0079 Steps: 79120, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000357, Sample Num: 5712, Cur Loss: 0.42377716, Cur Avg Loss: 0.47885668, Log Avg loss: 0.54766053, Global Avg Loss: 1.33709990, Time: 0.0073 Steps: 79130, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000367, Sample Num: 5872, Cur Loss: 0.71154523, Cur Avg Loss: 0.48476814, Log Avg loss: 0.69580733, Global Avg Loss: 1.33701887, Time: 0.0171 Steps: 79140, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000377, Sample Num: 6032, Cur Loss: 0.49628884, Cur Avg Loss: 0.48286125, Log Avg loss: 0.41287867, Global Avg Loss: 1.33690211, Time: 0.0114 Steps: 79150, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000387, Sample Num: 6192, Cur Loss: 0.57057047, Cur Avg Loss: 0.48077259, Log Avg loss: 0.40202985, Global Avg Loss: 1.33678401, Time: 0.0071 Steps: 79160, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000397, Sample Num: 6352, Cur Loss: 0.51589942, Cur Avg Loss: 0.48478970, Log Avg loss: 0.64025180, Global Avg Loss: 1.33669603, Time: 0.0070 Steps: 79170, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000407, Sample Num: 6512, Cur Loss: 0.30007261, Cur Avg Loss: 0.48322788, Log Avg loss: 0.42122370, Global Avg Loss: 1.33658042, Time: 0.0086 Steps: 79180, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000417, Sample Num: 6672, Cur Loss: 0.33990639, Cur Avg Loss: 0.48359922, Log Avg loss: 0.49871295, Global Avg Loss: 1.33647461, Time: 0.0069 Steps: 79190, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000427, Sample Num: 6832, Cur Loss: 1.13473856, Cur Avg Loss: 0.48744910, Log Avg loss: 0.64798889, Global Avg Loss: 1.33638768, Time: 0.0091 Steps: 79200, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000437, Sample Num: 6992, Cur Loss: 0.32733658, Cur Avg Loss: 0.48706564, Log Avg loss: 0.47069219, Global Avg Loss: 1.33627839, Time: 0.0089 Steps: 79210, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000447, Sample Num: 7152, Cur Loss: 0.37824306, Cur Avg Loss: 0.48718627, Log Avg loss: 0.49245759, Global Avg Loss: 1.33617187, Time: 0.0121 Steps: 79220, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000457, Sample Num: 7312, Cur Loss: 0.75122499, Cur Avg Loss: 0.48804081, Log Avg loss: 0.52623895, Global Avg Loss: 1.33606965, Time: 0.0120 Steps: 79230, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000467, Sample Num: 7472, Cur Loss: 0.51110744, Cur Avg Loss: 0.48654079, Log Avg loss: 0.41798988, Global Avg Loss: 1.33595379, Time: 0.0113 Steps: 79240, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000477, Sample Num: 7632, Cur Loss: 0.33387285, Cur Avg Loss: 0.48889532, Log Avg loss: 0.59885177, Global Avg Loss: 1.33586078, Time: 0.0110 Steps: 79250, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000487, Sample Num: 7792, Cur Loss: 0.69174433, Cur Avg Loss: 0.49009552, Log Avg loss: 0.54734490, Global Avg Loss: 1.33576129, Time: 0.0073 Steps: 79260, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000497, Sample Num: 7952, Cur Loss: 0.21485327, Cur Avg Loss: 0.48812600, Log Avg loss: 0.39221041, Global Avg Loss: 1.33564226, Time: 0.0123 Steps: 79270, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000507, Sample Num: 8112, Cur Loss: 0.80788487, Cur Avg Loss: 0.49114190, Log Avg loss: 0.64103228, Global Avg Loss: 1.33555465, Time: 0.0119 Steps: 79280, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000517, Sample Num: 8272, Cur Loss: 0.29928145, Cur Avg Loss: 0.49187581, Log Avg loss: 0.52908475, Global Avg Loss: 1.33545294, Time: 0.0116 Steps: 79290, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000527, Sample Num: 8432, Cur Loss: 0.29922399, Cur Avg Loss: 0.49345516, Log Avg loss: 0.57510799, Global Avg Loss: 1.33535705, Time: 0.0115 Steps: 79300, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000537, Sample Num: 8592, Cur Loss: 0.38537845, Cur Avg Loss: 0.49307235, Log Avg loss: 0.47289811, Global Avg Loss: 1.33524831, Time: 0.0066 Steps: 79310, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000547, Sample Num: 8752, Cur Loss: 0.85170144, Cur Avg Loss: 0.49275008, Log Avg loss: 0.47544424, Global Avg Loss: 1.33513991, Time: 0.0127 Steps: 79320, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000557, Sample Num: 8912, Cur Loss: 0.27862817, Cur Avg Loss: 0.49024917, Log Avg loss: 0.35344919, Global Avg Loss: 1.33501616, Time: 0.0072 Steps: 79330, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000567, Sample Num: 9072, Cur Loss: 0.33990723, Cur Avg Loss: 0.48935530, Log Avg loss: 0.43956682, Global Avg Loss: 1.33490330, Time: 0.0110 Steps: 79340, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000577, Sample Num: 9232, Cur Loss: 0.17061514, Cur Avg Loss: 0.48665786, Log Avg loss: 0.33371283, Global Avg Loss: 1.33477713, Time: 0.0143 Steps: 79350, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000587, Sample Num: 9392, Cur Loss: 0.45561981, Cur Avg Loss: 0.48651018, Log Avg loss: 0.47798911, Global Avg Loss: 1.33466917, Time: 0.0112 Steps: 79360, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000597, Sample Num: 9552, Cur Loss: 0.90287215, Cur Avg Loss: 0.48688638, Log Avg loss: 0.50896923, Global Avg Loss: 1.33456513, Time: 0.0083 Steps: 79370, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000607, Sample Num: 9712, Cur Loss: 1.42613506, Cur Avg Loss: 0.48739129, Log Avg loss: 0.51753480, Global Avg Loss: 1.33446221, Time: 0.0157 Steps: 79380, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000617, Sample Num: 9872, Cur Loss: 0.58683258, Cur Avg Loss: 0.48551479, Log Avg loss: 0.37161109, Global Avg Loss: 1.33434093, Time: 0.0096 Steps: 79390, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000627, Sample Num: 10032, Cur Loss: 0.79923266, Cur Avg Loss: 0.48741199, Log Avg loss: 0.60446924, Global Avg Loss: 1.33424900, Time: 0.0072 Steps: 79400, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000637, Sample Num: 10192, Cur Loss: 0.19652832, Cur Avg Loss: 0.48883342, Log Avg loss: 0.57795674, Global Avg Loss: 1.33415376, Time: 0.0111 Steps: 79410, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000647, Sample Num: 10352, Cur Loss: 0.38067666, Cur Avg Loss: 0.48768342, Log Avg loss: 0.41442847, Global Avg Loss: 1.33403796, Time: 0.0066 Steps: 79420, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000657, Sample Num: 10512, Cur Loss: 0.66443700, Cur Avg Loss: 0.48574744, Log Avg loss: 0.36048991, Global Avg Loss: 1.33391539, Time: 0.0065 Steps: 79430, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000667, Sample Num: 10672, Cur Loss: 0.26826060, Cur Avg Loss: 0.48790687, Log Avg loss: 0.62978109, Global Avg Loss: 1.33382676, Time: 0.0063 Steps: 79440, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000677, Sample Num: 10832, Cur Loss: 0.20794713, Cur Avg Loss: 0.48885849, Log Avg loss: 0.55233191, Global Avg Loss: 1.33372839, Time: 0.0111 Steps: 79450, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000687, Sample Num: 10992, Cur Loss: 0.51798588, Cur Avg Loss: 0.48792625, Log Avg loss: 0.42481376, Global Avg Loss: 1.33361401, Time: 0.0067 Steps: 79460, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000697, Sample Num: 11152, Cur Loss: 0.95351106, Cur Avg Loss: 0.48872031, Log Avg loss: 0.54327181, Global Avg Loss: 1.33351455, Time: 0.0132 Steps: 79470, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000707, Sample Num: 11312, Cur Loss: 0.54429030, Cur Avg Loss: 0.48806110, Log Avg loss: 0.44211429, Global Avg Loss: 1.33340240, Time: 0.0065 Steps: 79480, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000717, Sample Num: 11472, Cur Loss: 0.29339525, Cur Avg Loss: 0.48832892, Log Avg loss: 0.50726410, Global Avg Loss: 1.33329847, Time: 0.0115 Steps: 79490, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000727, Sample Num: 11632, Cur Loss: 1.08332872, Cur Avg Loss: 0.48805789, Log Avg loss: 0.46862451, Global Avg Loss: 1.33318971, Time: 0.0071 Steps: 79500, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000737, Sample Num: 11792, Cur Loss: 0.77083284, Cur Avg Loss: 0.48805422, Log Avg loss: 0.48778760, Global Avg Loss: 1.33308338, Time: 0.0120 Steps: 79510, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000747, Sample Num: 11952, Cur Loss: 0.69333649, Cur Avg Loss: 0.48831921, Log Avg loss: 0.50784892, Global Avg Loss: 1.33297960, Time: 0.0119 Steps: 79520, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000757, Sample Num: 12112, Cur Loss: 0.29277134, Cur Avg Loss: 0.48651421, Log Avg loss: 0.35168107, Global Avg Loss: 1.33285622, Time: 0.0128 Steps: 79530, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000767, Sample Num: 12272, Cur Loss: 0.25443536, Cur Avg Loss: 0.48593408, Log Avg loss: 0.44201817, Global Avg Loss: 1.33274422, Time: 0.0104 Steps: 79540, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000777, Sample Num: 12432, Cur Loss: 0.61644816, Cur Avg Loss: 0.48764914, Log Avg loss: 0.61919412, Global Avg Loss: 1.33265452, Time: 0.0074 Steps: 79550, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000787, Sample Num: 12592, Cur Loss: 0.54338336, Cur Avg Loss: 0.48754183, Log Avg loss: 0.47920358, Global Avg Loss: 1.33254725, Time: 0.0067 Steps: 79560, Updated lr: 0.000026 Training, Epoch: 0038, Batch: 000797, Sample Num: 12752, Cur Loss: 0.49581221, Cur Avg Loss: 0.48723326, Log Avg loss: 0.46294932, Global Avg Loss: 1.33243796, Time: 0.0064 Steps: 79570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000807, Sample Num: 12912, Cur Loss: 0.50919616, Cur Avg Loss: 0.48656957, Log Avg loss: 0.43367301, Global Avg Loss: 1.33232502, Time: 0.0068 Steps: 79580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000817, Sample Num: 13072, Cur Loss: 0.38686520, Cur Avg Loss: 0.48751999, Log Avg loss: 0.56421892, Global Avg Loss: 1.33222851, Time: 0.0069 Steps: 79590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000827, Sample Num: 13232, Cur Loss: 0.15478361, Cur Avg Loss: 0.48821412, Log Avg loss: 0.54492418, Global Avg Loss: 1.33212961, Time: 0.0075 Steps: 79600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000837, Sample Num: 13392, Cur Loss: 0.46010250, Cur Avg Loss: 0.48987662, Log Avg loss: 0.62736568, Global Avg Loss: 1.33204108, Time: 0.0081 Steps: 79610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000847, Sample Num: 13552, Cur Loss: 0.47504851, Cur Avg Loss: 0.48879850, Log Avg loss: 0.39856024, Global Avg Loss: 1.33192384, Time: 0.0115 Steps: 79620, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000857, Sample Num: 13712, Cur Loss: 0.09110209, Cur Avg Loss: 0.48821742, Log Avg loss: 0.43899980, Global Avg Loss: 1.33181170, Time: 0.0107 Steps: 79630, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000867, Sample Num: 13872, Cur Loss: 0.33179411, Cur Avg Loss: 0.48809657, Log Avg loss: 0.47773927, Global Avg Loss: 1.33170446, Time: 0.0067 Steps: 79640, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000877, Sample Num: 14032, Cur Loss: 0.56465274, Cur Avg Loss: 0.48932904, Log Avg loss: 0.59618477, Global Avg Loss: 1.33161212, Time: 0.0110 Steps: 79650, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000887, Sample Num: 14192, Cur Loss: 0.63689482, Cur Avg Loss: 0.48855939, Log Avg loss: 0.42106098, Global Avg Loss: 1.33149781, Time: 0.0105 Steps: 79660, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000897, Sample Num: 14352, Cur Loss: 0.82222670, Cur Avg Loss: 0.48852846, Log Avg loss: 0.48578444, Global Avg Loss: 1.33139166, Time: 0.0120 Steps: 79670, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000907, Sample Num: 14512, Cur Loss: 0.34310001, Cur Avg Loss: 0.48868082, Log Avg loss: 0.50234777, Global Avg Loss: 1.33128761, Time: 0.0104 Steps: 79680, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000917, Sample Num: 14672, Cur Loss: 0.80553919, Cur Avg Loss: 0.48909724, Log Avg loss: 0.52686647, Global Avg Loss: 1.33118667, Time: 0.0067 Steps: 79690, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000927, Sample Num: 14832, Cur Loss: 0.24368018, Cur Avg Loss: 0.48971906, Log Avg loss: 0.54674007, Global Avg Loss: 1.33108824, Time: 0.0121 Steps: 79700, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000937, Sample Num: 14992, Cur Loss: 0.67598897, Cur Avg Loss: 0.49123587, Log Avg loss: 0.63184445, Global Avg Loss: 1.33100052, Time: 0.0117 Steps: 79710, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000947, Sample Num: 15152, Cur Loss: 0.28365713, Cur Avg Loss: 0.49074326, Log Avg loss: 0.44458552, Global Avg Loss: 1.33088933, Time: 0.0089 Steps: 79720, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000957, Sample Num: 15312, Cur Loss: 0.68547279, Cur Avg Loss: 0.49229044, Log Avg loss: 0.63880827, Global Avg Loss: 1.33080253, Time: 0.0094 Steps: 79730, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000967, Sample Num: 15472, Cur Loss: 0.21372348, Cur Avg Loss: 0.49233635, Log Avg loss: 0.49672960, Global Avg Loss: 1.33069793, Time: 0.0071 Steps: 79740, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000977, Sample Num: 15632, Cur Loss: 0.19849055, Cur Avg Loss: 0.49191266, Log Avg loss: 0.45094248, Global Avg Loss: 1.33058761, Time: 0.0073 Steps: 79750, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000987, Sample Num: 15792, Cur Loss: 0.21650985, Cur Avg Loss: 0.49175715, Log Avg loss: 0.47656384, Global Avg Loss: 1.33048054, Time: 0.0113 Steps: 79760, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 000997, Sample Num: 15952, Cur Loss: 1.06376696, Cur Avg Loss: 0.49197266, Log Avg loss: 0.51324340, Global Avg Loss: 1.33037809, Time: 0.0091 Steps: 79770, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001007, Sample Num: 16112, Cur Loss: 1.29728341, Cur Avg Loss: 0.49144454, Log Avg loss: 0.43879132, Global Avg Loss: 1.33026634, Time: 0.0162 Steps: 79780, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001017, Sample Num: 16272, Cur Loss: 1.03144193, Cur Avg Loss: 0.49339000, Log Avg loss: 0.68929683, Global Avg Loss: 1.33018600, Time: 0.0124 Steps: 79790, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001027, Sample Num: 16432, Cur Loss: 0.29568791, Cur Avg Loss: 0.49289579, Log Avg loss: 0.44263499, Global Avg Loss: 1.33007478, Time: 0.0067 Steps: 79800, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001037, Sample Num: 16592, Cur Loss: 0.30362543, Cur Avg Loss: 0.49187308, Log Avg loss: 0.38684040, Global Avg Loss: 1.32995660, Time: 0.0066 Steps: 79810, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001047, Sample Num: 16752, Cur Loss: 0.42056781, Cur Avg Loss: 0.49149606, Log Avg loss: 0.45239912, Global Avg Loss: 1.32984665, Time: 0.0103 Steps: 79820, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001057, Sample Num: 16912, Cur Loss: 0.36657283, Cur Avg Loss: 0.49087861, Log Avg loss: 0.42623231, Global Avg Loss: 1.32973346, Time: 0.0156 Steps: 79830, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001067, Sample Num: 17072, Cur Loss: 0.65481114, Cur Avg Loss: 0.49287239, Log Avg loss: 0.70361427, Global Avg Loss: 1.32965504, Time: 0.0123 Steps: 79840, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001077, Sample Num: 17232, Cur Loss: 0.16068985, Cur Avg Loss: 0.49301066, Log Avg loss: 0.50776401, Global Avg Loss: 1.32955211, Time: 0.0114 Steps: 79850, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001087, Sample Num: 17392, Cur Loss: 0.73194146, Cur Avg Loss: 0.49315034, Log Avg loss: 0.50819414, Global Avg Loss: 1.32944926, Time: 0.0114 Steps: 79860, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001097, Sample Num: 17552, Cur Loss: 0.51183820, Cur Avg Loss: 0.49348967, Log Avg loss: 0.53037547, Global Avg Loss: 1.32934921, Time: 0.0123 Steps: 79870, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001107, Sample Num: 17712, Cur Loss: 0.19110657, Cur Avg Loss: 0.49419612, Log Avg loss: 0.57169307, Global Avg Loss: 1.32925436, Time: 0.0117 Steps: 79880, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001117, Sample Num: 17872, Cur Loss: 0.34561861, Cur Avg Loss: 0.49373890, Log Avg loss: 0.44312502, Global Avg Loss: 1.32914345, Time: 0.0095 Steps: 79890, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001127, Sample Num: 18032, Cur Loss: 0.49493515, Cur Avg Loss: 0.49691034, Log Avg loss: 0.85115948, Global Avg Loss: 1.32908362, Time: 0.0070 Steps: 79900, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001137, Sample Num: 18192, Cur Loss: 0.47541589, Cur Avg Loss: 0.49747152, Log Avg loss: 0.56071720, Global Avg Loss: 1.32898747, Time: 0.0222 Steps: 79910, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001147, Sample Num: 18352, Cur Loss: 0.17944998, Cur Avg Loss: 0.49764545, Log Avg loss: 0.51742124, Global Avg Loss: 1.32888592, Time: 0.0109 Steps: 79920, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001157, Sample Num: 18512, Cur Loss: 0.21399990, Cur Avg Loss: 0.49622781, Log Avg loss: 0.33362435, Global Avg Loss: 1.32876141, Time: 0.0117 Steps: 79930, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001167, Sample Num: 18672, Cur Loss: 0.37088108, Cur Avg Loss: 0.49620718, Log Avg loss: 0.49382034, Global Avg Loss: 1.32865696, Time: 0.0118 Steps: 79940, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001177, Sample Num: 18832, Cur Loss: 0.60078967, Cur Avg Loss: 0.49579147, Log Avg loss: 0.44727785, Global Avg Loss: 1.32854672, Time: 0.0097 Steps: 79950, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001187, Sample Num: 18992, Cur Loss: 0.48879677, Cur Avg Loss: 0.49521684, Log Avg loss: 0.42758354, Global Avg Loss: 1.32843404, Time: 0.0239 Steps: 79960, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001197, Sample Num: 19152, Cur Loss: 0.39827850, Cur Avg Loss: 0.49514791, Log Avg loss: 0.48696520, Global Avg Loss: 1.32832882, Time: 0.0125 Steps: 79970, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001207, Sample Num: 19312, Cur Loss: 0.99655664, Cur Avg Loss: 0.49628872, Log Avg loss: 0.63284394, Global Avg Loss: 1.32824186, Time: 0.0071 Steps: 79980, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001217, Sample Num: 19472, Cur Loss: 0.25654441, Cur Avg Loss: 0.49531006, Log Avg loss: 0.37718597, Global Avg Loss: 1.32812296, Time: 0.0066 Steps: 79990, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001227, Sample Num: 19632, Cur Loss: 0.25771984, Cur Avg Loss: 0.49504099, Log Avg loss: 0.46229474, Global Avg Loss: 1.32801474, Time: 0.0068 Steps: 80000, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001237, Sample Num: 19792, Cur Loss: 1.64630151, Cur Avg Loss: 0.49530938, Log Avg loss: 0.52824073, Global Avg Loss: 1.32791478, Time: 0.0220 Steps: 80010, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001247, Sample Num: 19952, Cur Loss: 0.67335951, Cur Avg Loss: 0.49517846, Log Avg loss: 0.47898377, Global Avg Loss: 1.32780869, Time: 0.0111 Steps: 80020, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001257, Sample Num: 20112, Cur Loss: 1.00250614, Cur Avg Loss: 0.49575306, Log Avg loss: 0.56740623, Global Avg Loss: 1.32771367, Time: 0.0065 Steps: 80030, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001267, Sample Num: 20272, Cur Loss: 0.85264099, Cur Avg Loss: 0.49545521, Log Avg loss: 0.45801508, Global Avg Loss: 1.32760501, Time: 0.0070 Steps: 80040, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001277, Sample Num: 20432, Cur Loss: 0.14278269, Cur Avg Loss: 0.49508070, Log Avg loss: 0.44763085, Global Avg Loss: 1.32749509, Time: 0.0102 Steps: 80050, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001287, Sample Num: 20592, Cur Loss: 0.26432914, Cur Avg Loss: 0.49524199, Log Avg loss: 0.51583838, Global Avg Loss: 1.32739370, Time: 0.0128 Steps: 80060, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001297, Sample Num: 20752, Cur Loss: 0.60497814, Cur Avg Loss: 0.49501003, Log Avg loss: 0.46515730, Global Avg Loss: 1.32728602, Time: 0.0114 Steps: 80070, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001307, Sample Num: 20912, Cur Loss: 0.58737963, Cur Avg Loss: 0.49674708, Log Avg loss: 0.72204223, Global Avg Loss: 1.32721044, Time: 0.0073 Steps: 80080, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001317, Sample Num: 21072, Cur Loss: 0.45965821, Cur Avg Loss: 0.49653306, Log Avg loss: 0.46856049, Global Avg Loss: 1.32710323, Time: 0.0116 Steps: 80090, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001327, Sample Num: 21232, Cur Loss: 0.19631109, Cur Avg Loss: 0.49636528, Log Avg loss: 0.47426800, Global Avg Loss: 1.32699676, Time: 0.0230 Steps: 80100, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001337, Sample Num: 21392, Cur Loss: 0.83200163, Cur Avg Loss: 0.49736815, Log Avg loss: 0.63044897, Global Avg Loss: 1.32690981, Time: 0.0107 Steps: 80110, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001347, Sample Num: 21552, Cur Loss: 0.39244297, Cur Avg Loss: 0.49744934, Log Avg loss: 0.50830532, Global Avg Loss: 1.32680764, Time: 0.0096 Steps: 80120, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001357, Sample Num: 21712, Cur Loss: 0.81090987, Cur Avg Loss: 0.49781669, Log Avg loss: 0.54729884, Global Avg Loss: 1.32671036, Time: 0.0070 Steps: 80130, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001367, Sample Num: 21872, Cur Loss: 0.14506243, Cur Avg Loss: 0.49784305, Log Avg loss: 0.50141902, Global Avg Loss: 1.32660737, Time: 0.0107 Steps: 80140, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001377, Sample Num: 22032, Cur Loss: 0.23374015, Cur Avg Loss: 0.49745297, Log Avg loss: 0.44413022, Global Avg Loss: 1.32649727, Time: 0.0089 Steps: 80150, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001387, Sample Num: 22192, Cur Loss: 0.29612345, Cur Avg Loss: 0.49699892, Log Avg loss: 0.43447588, Global Avg Loss: 1.32638599, Time: 0.0109 Steps: 80160, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001397, Sample Num: 22352, Cur Loss: 0.21848848, Cur Avg Loss: 0.49662542, Log Avg loss: 0.44482121, Global Avg Loss: 1.32627603, Time: 0.0110 Steps: 80170, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001407, Sample Num: 22512, Cur Loss: 0.47644114, Cur Avg Loss: 0.49601176, Log Avg loss: 0.41028351, Global Avg Loss: 1.32616179, Time: 0.0085 Steps: 80180, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001417, Sample Num: 22672, Cur Loss: 0.17017063, Cur Avg Loss: 0.49567304, Log Avg loss: 0.44801472, Global Avg Loss: 1.32605228, Time: 0.0065 Steps: 80190, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001427, Sample Num: 22832, Cur Loss: 0.67036390, Cur Avg Loss: 0.49643824, Log Avg loss: 0.60486658, Global Avg Loss: 1.32596236, Time: 0.0128 Steps: 80200, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001437, Sample Num: 22992, Cur Loss: 0.69942927, Cur Avg Loss: 0.49616195, Log Avg loss: 0.45673634, Global Avg Loss: 1.32585399, Time: 0.0076 Steps: 80210, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001447, Sample Num: 23152, Cur Loss: 0.31322035, Cur Avg Loss: 0.49644385, Log Avg loss: 0.53695176, Global Avg Loss: 1.32575564, Time: 0.0148 Steps: 80220, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001457, Sample Num: 23312, Cur Loss: 0.44732895, Cur Avg Loss: 0.49599316, Log Avg loss: 0.43077931, Global Avg Loss: 1.32564409, Time: 0.0092 Steps: 80230, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001467, Sample Num: 23472, Cur Loss: 0.51191890, Cur Avg Loss: 0.49490081, Log Avg loss: 0.33574495, Global Avg Loss: 1.32552073, Time: 0.0111 Steps: 80240, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001477, Sample Num: 23632, Cur Loss: 0.28533420, Cur Avg Loss: 0.49473989, Log Avg loss: 0.47113304, Global Avg Loss: 1.32541426, Time: 0.0110 Steps: 80250, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001487, Sample Num: 23792, Cur Loss: 0.20603904, Cur Avg Loss: 0.49569339, Log Avg loss: 0.63652517, Global Avg Loss: 1.32532843, Time: 0.0073 Steps: 80260, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001497, Sample Num: 23952, Cur Loss: 0.23922892, Cur Avg Loss: 0.49568066, Log Avg loss: 0.49378780, Global Avg Loss: 1.32522483, Time: 0.0125 Steps: 80270, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001507, Sample Num: 24112, Cur Loss: 0.28586888, Cur Avg Loss: 0.49571132, Log Avg loss: 0.50030130, Global Avg Loss: 1.32512208, Time: 0.0096 Steps: 80280, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001517, Sample Num: 24272, Cur Loss: 0.21547943, Cur Avg Loss: 0.49667881, Log Avg loss: 0.64247894, Global Avg Loss: 1.32503706, Time: 0.0084 Steps: 80290, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001527, Sample Num: 24432, Cur Loss: 0.50784087, Cur Avg Loss: 0.49626709, Log Avg loss: 0.43380910, Global Avg Loss: 1.32492607, Time: 0.0111 Steps: 80300, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001537, Sample Num: 24592, Cur Loss: 0.46193415, Cur Avg Loss: 0.49547176, Log Avg loss: 0.37402479, Global Avg Loss: 1.32480767, Time: 0.0145 Steps: 80310, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001547, Sample Num: 24752, Cur Loss: 0.41197741, Cur Avg Loss: 0.49600854, Log Avg loss: 0.57851199, Global Avg Loss: 1.32471475, Time: 0.0072 Steps: 80320, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001557, Sample Num: 24912, Cur Loss: 0.77836341, Cur Avg Loss: 0.49679408, Log Avg loss: 0.61831716, Global Avg Loss: 1.32462681, Time: 0.0124 Steps: 80330, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001567, Sample Num: 25072, Cur Loss: 0.37961626, Cur Avg Loss: 0.49575152, Log Avg loss: 0.33342551, Global Avg Loss: 1.32450344, Time: 0.0067 Steps: 80340, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001577, Sample Num: 25232, Cur Loss: 0.35803047, Cur Avg Loss: 0.49475215, Log Avg loss: 0.33815019, Global Avg Loss: 1.32438068, Time: 0.0126 Steps: 80350, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001587, Sample Num: 25392, Cur Loss: 0.31604859, Cur Avg Loss: 0.49481642, Log Avg loss: 0.50495159, Global Avg Loss: 1.32427871, Time: 0.0113 Steps: 80360, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001597, Sample Num: 25552, Cur Loss: 0.67890561, Cur Avg Loss: 0.49538695, Log Avg loss: 0.58592996, Global Avg Loss: 1.32418684, Time: 0.0111 Steps: 80370, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001607, Sample Num: 25712, Cur Loss: 0.22530425, Cur Avg Loss: 0.49544122, Log Avg loss: 0.50410860, Global Avg Loss: 1.32408482, Time: 0.0112 Steps: 80380, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001617, Sample Num: 25872, Cur Loss: 0.98736113, Cur Avg Loss: 0.49504772, Log Avg loss: 0.43181249, Global Avg Loss: 1.32397382, Time: 0.0114 Steps: 80390, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001627, Sample Num: 26032, Cur Loss: 0.23123170, Cur Avg Loss: 0.49541423, Log Avg loss: 0.55467890, Global Avg Loss: 1.32387814, Time: 0.0113 Steps: 80400, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001637, Sample Num: 26192, Cur Loss: 0.90476441, Cur Avg Loss: 0.49566907, Log Avg loss: 0.53713096, Global Avg Loss: 1.32378030, Time: 0.0066 Steps: 80410, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001647, Sample Num: 26352, Cur Loss: 0.61883795, Cur Avg Loss: 0.49547800, Log Avg loss: 0.46420002, Global Avg Loss: 1.32367341, Time: 0.0117 Steps: 80420, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001657, Sample Num: 26512, Cur Loss: 1.03705859, Cur Avg Loss: 0.49612447, Log Avg loss: 0.60259792, Global Avg Loss: 1.32358376, Time: 0.0123 Steps: 80430, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001667, Sample Num: 26672, Cur Loss: 0.52742851, Cur Avg Loss: 0.49624633, Log Avg loss: 0.51643937, Global Avg Loss: 1.32348342, Time: 0.0112 Steps: 80440, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001677, Sample Num: 26832, Cur Loss: 0.43216360, Cur Avg Loss: 0.49703545, Log Avg loss: 0.62858054, Global Avg Loss: 1.32339704, Time: 0.0189 Steps: 80450, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001687, Sample Num: 26992, Cur Loss: 0.62965119, Cur Avg Loss: 0.49755347, Log Avg loss: 0.58442565, Global Avg Loss: 1.32330520, Time: 0.0069 Steps: 80460, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001697, Sample Num: 27152, Cur Loss: 0.40389058, Cur Avg Loss: 0.49699394, Log Avg loss: 0.40260131, Global Avg Loss: 1.32319078, Time: 0.0068 Steps: 80470, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001707, Sample Num: 27312, Cur Loss: 0.47912765, Cur Avg Loss: 0.49719667, Log Avg loss: 0.53160017, Global Avg Loss: 1.32309242, Time: 0.0107 Steps: 80480, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001717, Sample Num: 27472, Cur Loss: 0.44540662, Cur Avg Loss: 0.49651197, Log Avg loss: 0.37963421, Global Avg Loss: 1.32297521, Time: 0.0113 Steps: 80490, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001727, Sample Num: 27632, Cur Loss: 0.70701176, Cur Avg Loss: 0.49701497, Log Avg loss: 0.58337945, Global Avg Loss: 1.32288333, Time: 0.0076 Steps: 80500, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001737, Sample Num: 27792, Cur Loss: 0.76141965, Cur Avg Loss: 0.49661118, Log Avg loss: 0.42687674, Global Avg Loss: 1.32277204, Time: 0.0140 Steps: 80510, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001747, Sample Num: 27952, Cur Loss: 0.22805326, Cur Avg Loss: 0.49652821, Log Avg loss: 0.48211657, Global Avg Loss: 1.32266764, Time: 0.0107 Steps: 80520, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001757, Sample Num: 28112, Cur Loss: 0.22151467, Cur Avg Loss: 0.49619350, Log Avg loss: 0.43771964, Global Avg Loss: 1.32255775, Time: 0.0136 Steps: 80530, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001767, Sample Num: 28272, Cur Loss: 0.17096050, Cur Avg Loss: 0.49729454, Log Avg loss: 0.69074791, Global Avg Loss: 1.32247930, Time: 0.0067 Steps: 80540, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001777, Sample Num: 28432, Cur Loss: 0.51888627, Cur Avg Loss: 0.49727929, Log Avg loss: 0.49458448, Global Avg Loss: 1.32237652, Time: 0.0125 Steps: 80550, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001787, Sample Num: 28592, Cur Loss: 0.70058465, Cur Avg Loss: 0.49795538, Log Avg loss: 0.61809558, Global Avg Loss: 1.32228910, Time: 0.0090 Steps: 80560, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001797, Sample Num: 28752, Cur Loss: 0.55450463, Cur Avg Loss: 0.49767699, Log Avg loss: 0.44792892, Global Avg Loss: 1.32218058, Time: 0.0072 Steps: 80570, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001807, Sample Num: 28912, Cur Loss: 0.45212460, Cur Avg Loss: 0.49744481, Log Avg loss: 0.45572249, Global Avg Loss: 1.32207305, Time: 0.0107 Steps: 80580, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001817, Sample Num: 29072, Cur Loss: 0.54978442, Cur Avg Loss: 0.49794403, Log Avg loss: 0.58815327, Global Avg Loss: 1.32198198, Time: 0.0121 Steps: 80590, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001827, Sample Num: 29232, Cur Loss: 0.27386630, Cur Avg Loss: 0.49773155, Log Avg loss: 0.45912314, Global Avg Loss: 1.32187493, Time: 0.0123 Steps: 80600, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001837, Sample Num: 29392, Cur Loss: 0.36906129, Cur Avg Loss: 0.49704983, Log Avg loss: 0.37250035, Global Avg Loss: 1.32175715, Time: 0.0158 Steps: 80610, Updated lr: 0.000025 Training, Epoch: 0038, Batch: 001847, Sample Num: 29552, Cur Loss: 0.58394760, Cur Avg Loss: 0.49699125, Log Avg loss: 0.48623015, Global Avg Loss: 1.32165352, Time: 0.0066 Steps: 80620, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001857, Sample Num: 29712, Cur Loss: 0.27078715, Cur Avg Loss: 0.49719549, Log Avg loss: 0.53491837, Global Avg Loss: 1.32155594, Time: 0.0080 Steps: 80630, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001867, Sample Num: 29872, Cur Loss: 0.82618284, Cur Avg Loss: 0.49781387, Log Avg loss: 0.61264630, Global Avg Loss: 1.32146803, Time: 0.0116 Steps: 80640, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001877, Sample Num: 30032, Cur Loss: 0.30917203, Cur Avg Loss: 0.49743902, Log Avg loss: 0.42745428, Global Avg Loss: 1.32135718, Time: 0.0132 Steps: 80650, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001887, Sample Num: 30192, Cur Loss: 1.11864924, Cur Avg Loss: 0.49792925, Log Avg loss: 0.58994692, Global Avg Loss: 1.32126650, Time: 0.0116 Steps: 80660, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001897, Sample Num: 30352, Cur Loss: 0.35997537, Cur Avg Loss: 0.49830315, Log Avg loss: 0.56885721, Global Avg Loss: 1.32117323, Time: 0.0078 Steps: 80670, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001907, Sample Num: 30512, Cur Loss: 0.37742501, Cur Avg Loss: 0.49790386, Log Avg loss: 0.42215850, Global Avg Loss: 1.32106180, Time: 0.0065 Steps: 80680, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001917, Sample Num: 30672, Cur Loss: 0.17587160, Cur Avg Loss: 0.49826822, Log Avg loss: 0.56775244, Global Avg Loss: 1.32096844, Time: 0.0108 Steps: 80690, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001927, Sample Num: 30832, Cur Loss: 0.21421036, Cur Avg Loss: 0.49827855, Log Avg loss: 0.50025751, Global Avg Loss: 1.32086674, Time: 0.0136 Steps: 80700, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001937, Sample Num: 30992, Cur Loss: 0.29834083, Cur Avg Loss: 0.49757403, Log Avg loss: 0.36181446, Global Avg Loss: 1.32074792, Time: 0.0067 Steps: 80710, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001947, Sample Num: 31152, Cur Loss: 0.66519666, Cur Avg Loss: 0.49765664, Log Avg loss: 0.51365822, Global Avg Loss: 1.32064793, Time: 0.0067 Steps: 80720, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001957, Sample Num: 31312, Cur Loss: 0.16174006, Cur Avg Loss: 0.49757060, Log Avg loss: 0.48081757, Global Avg Loss: 1.32054390, Time: 0.0134 Steps: 80730, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001967, Sample Num: 31472, Cur Loss: 0.40010419, Cur Avg Loss: 0.49735698, Log Avg loss: 0.45555135, Global Avg Loss: 1.32043677, Time: 0.0144 Steps: 80740, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001977, Sample Num: 31632, Cur Loss: 0.29202396, Cur Avg Loss: 0.49747682, Log Avg loss: 0.52105008, Global Avg Loss: 1.32033777, Time: 0.0068 Steps: 80750, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001987, Sample Num: 31792, Cur Loss: 0.11796831, Cur Avg Loss: 0.49737956, Log Avg loss: 0.47815066, Global Avg Loss: 1.32023349, Time: 0.0131 Steps: 80760, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 001997, Sample Num: 31952, Cur Loss: 0.31621447, Cur Avg Loss: 0.49784528, Log Avg loss: 0.59038405, Global Avg Loss: 1.32014313, Time: 0.0115 Steps: 80770, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002007, Sample Num: 32112, Cur Loss: 0.48695204, Cur Avg Loss: 0.49722809, Log Avg loss: 0.37397458, Global Avg Loss: 1.32002600, Time: 0.0069 Steps: 80780, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002017, Sample Num: 32272, Cur Loss: 0.64717138, Cur Avg Loss: 0.49717889, Log Avg loss: 0.48730536, Global Avg Loss: 1.31992293, Time: 0.0120 Steps: 80790, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002027, Sample Num: 32432, Cur Loss: 0.48864388, Cur Avg Loss: 0.49685619, Log Avg loss: 0.43176737, Global Avg Loss: 1.31981301, Time: 0.0067 Steps: 80800, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002037, Sample Num: 32592, Cur Loss: 0.43532604, Cur Avg Loss: 0.49721933, Log Avg loss: 0.57082774, Global Avg Loss: 1.31972032, Time: 0.0122 Steps: 80810, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002047, Sample Num: 32752, Cur Loss: 0.33403850, Cur Avg Loss: 0.49685995, Log Avg loss: 0.42365410, Global Avg Loss: 1.31960945, Time: 0.0154 Steps: 80820, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002057, Sample Num: 32912, Cur Loss: 0.58619338, Cur Avg Loss: 0.49694737, Log Avg loss: 0.51484218, Global Avg Loss: 1.31950989, Time: 0.0114 Steps: 80830, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002067, Sample Num: 33072, Cur Loss: 1.01830232, Cur Avg Loss: 0.49696414, Log Avg loss: 0.50041435, Global Avg Loss: 1.31940857, Time: 0.0066 Steps: 80840, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002077, Sample Num: 33232, Cur Loss: 0.13708225, Cur Avg Loss: 0.49636318, Log Avg loss: 0.37214456, Global Avg Loss: 1.31929140, Time: 0.0066 Steps: 80850, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002087, Sample Num: 33392, Cur Loss: 0.27337033, Cur Avg Loss: 0.49558226, Log Avg loss: 0.33338511, Global Avg Loss: 1.31916947, Time: 0.0064 Steps: 80860, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002097, Sample Num: 33552, Cur Loss: 0.91331398, Cur Avg Loss: 0.49615014, Log Avg loss: 0.61466586, Global Avg Loss: 1.31908236, Time: 0.0067 Steps: 80870, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002107, Sample Num: 33712, Cur Loss: 0.64497328, Cur Avg Loss: 0.49574875, Log Avg loss: 0.41157875, Global Avg Loss: 1.31897016, Time: 0.0231 Steps: 80880, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002117, Sample Num: 33872, Cur Loss: 0.32337397, Cur Avg Loss: 0.49560127, Log Avg loss: 0.46452757, Global Avg Loss: 1.31886453, Time: 0.0110 Steps: 80890, Updated lr: 0.000024 Training, Epoch: 0038, Batch: 002127, Sample Num: 34032, Cur Loss: 0.74441159, Cur Avg Loss: 0.49577070, Log Avg loss: 0.53163719, Global Avg Loss: 1.31876722, Time: 0.0108 Steps: 80900, Updated lr: 0.000024 ***** Running evaluation checkpoint-80902 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-80902 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 21.877330, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.624582, "eval_total_loss": 439.081019, "eval_mae": 0.624525, "eval_mse": 0.624685, "eval_r2": 0.602909, "eval_sp_statistic": 0.736523, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.790926, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.871299, "test_total_loss": 437.391987, "test_mae": 0.717006, "test_mse": 0.87145, "test_r2": 0.437558, "test_sp_statistic": 0.588085, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.665576, "test_ps_pvalue": 0.0, "lr": 2.4227596017069703e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.3187500628100097, "train_cur_epoch_loss": 1055.7540301531553, "train_cur_epoch_avg_loss": 0.4958919822231824, "train_cur_epoch_time": 21.87733006477356, "train_cur_epoch_avg_time": 0.010275871331504725, "epoch": 38, "step": 80902} ################################################## Training, Epoch: 0039, Batch: 000008, Sample Num: 128, Cur Loss: 0.39759141, Cur Avg Loss: 0.44425592, Log Avg loss: 0.48038085, Global Avg Loss: 1.31866360, Time: 0.0076 Steps: 80910, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000018, Sample Num: 288, Cur Loss: 0.47370607, Cur Avg Loss: 0.37976667, Log Avg loss: 0.32817527, Global Avg Loss: 1.31854119, Time: 0.0068 Steps: 80920, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000028, Sample Num: 448, Cur Loss: 0.33376259, Cur Avg Loss: 0.39008760, Log Avg loss: 0.40866526, Global Avg Loss: 1.31842877, Time: 0.0066 Steps: 80930, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000038, Sample Num: 608, Cur Loss: 0.44717145, Cur Avg Loss: 0.40743732, Log Avg loss: 0.45601656, Global Avg Loss: 1.31832222, Time: 0.0065 Steps: 80940, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000048, Sample Num: 768, Cur Loss: 0.43439150, Cur Avg Loss: 0.42594114, Log Avg loss: 0.49625565, Global Avg Loss: 1.31822066, Time: 0.0073 Steps: 80950, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000058, Sample Num: 928, Cur Loss: 1.03442645, Cur Avg Loss: 0.46025429, Log Avg loss: 0.62495742, Global Avg Loss: 1.31813503, Time: 0.0067 Steps: 80960, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000068, Sample Num: 1088, Cur Loss: 0.26092014, Cur Avg Loss: 0.43782976, Log Avg loss: 0.30776748, Global Avg Loss: 1.31801025, Time: 0.0067 Steps: 80970, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000078, Sample Num: 1248, Cur Loss: 0.48855641, Cur Avg Loss: 0.42419890, Log Avg loss: 0.33150901, Global Avg Loss: 1.31788843, Time: 0.0240 Steps: 80980, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000088, Sample Num: 1408, Cur Loss: 0.74430871, Cur Avg Loss: 0.43225129, Log Avg loss: 0.49505999, Global Avg Loss: 1.31778683, Time: 0.0064 Steps: 80990, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000098, Sample Num: 1568, Cur Loss: 0.56912231, Cur Avg Loss: 0.44530893, Log Avg loss: 0.56021618, Global Avg Loss: 1.31769331, Time: 0.0198 Steps: 81000, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000108, Sample Num: 1728, Cur Loss: 0.52970707, Cur Avg Loss: 0.44594890, Log Avg loss: 0.45222060, Global Avg Loss: 1.31758647, Time: 0.0064 Steps: 81010, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000118, Sample Num: 1888, Cur Loss: 0.15518996, Cur Avg Loss: 0.45006823, Log Avg loss: 0.49455697, Global Avg Loss: 1.31748489, Time: 0.0206 Steps: 81020, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000128, Sample Num: 2048, Cur Loss: 0.29960859, Cur Avg Loss: 0.44352898, Log Avg loss: 0.36636584, Global Avg Loss: 1.31736751, Time: 0.0065 Steps: 81030, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000138, Sample Num: 2208, Cur Loss: 0.22930911, Cur Avg Loss: 0.44196249, Log Avg loss: 0.42191139, Global Avg Loss: 1.31725701, Time: 0.0065 Steps: 81040, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000148, Sample Num: 2368, Cur Loss: 0.23531435, Cur Avg Loss: 0.45122167, Log Avg loss: 0.57899829, Global Avg Loss: 1.31716593, Time: 0.0065 Steps: 81050, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000158, Sample Num: 2528, Cur Loss: 0.80339128, Cur Avg Loss: 0.45252924, Log Avg loss: 0.47188130, Global Avg Loss: 1.31706165, Time: 0.0067 Steps: 81060, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000168, Sample Num: 2688, Cur Loss: 0.31495050, Cur Avg Loss: 0.46285217, Log Avg loss: 0.62595452, Global Avg Loss: 1.31697640, Time: 0.0067 Steps: 81070, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000178, Sample Num: 2848, Cur Loss: 0.32397211, Cur Avg Loss: 0.46421513, Log Avg loss: 0.48711278, Global Avg Loss: 1.31687405, Time: 0.0068 Steps: 81080, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000188, Sample Num: 3008, Cur Loss: 0.25281259, Cur Avg Loss: 0.46183648, Log Avg loss: 0.41949653, Global Avg Loss: 1.31676338, Time: 0.0067 Steps: 81090, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000198, Sample Num: 3168, Cur Loss: 0.57659280, Cur Avg Loss: 0.46180054, Log Avg loss: 0.46112489, Global Avg Loss: 1.31665788, Time: 0.0077 Steps: 81100, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000208, Sample Num: 3328, Cur Loss: 0.67064226, Cur Avg Loss: 0.46446845, Log Avg loss: 0.51729299, Global Avg Loss: 1.31655933, Time: 0.0075 Steps: 81110, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000218, Sample Num: 3488, Cur Loss: 0.28545395, Cur Avg Loss: 0.46691470, Log Avg loss: 0.51779669, Global Avg Loss: 1.31646086, Time: 0.0076 Steps: 81120, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000228, Sample Num: 3648, Cur Loss: 0.76853663, Cur Avg Loss: 0.46619844, Log Avg loss: 0.45058405, Global Avg Loss: 1.31635413, Time: 0.0077 Steps: 81130, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000238, Sample Num: 3808, Cur Loss: 0.40075666, Cur Avg Loss: 0.47832705, Log Avg loss: 0.75485940, Global Avg Loss: 1.31628493, Time: 0.0076 Steps: 81140, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000248, Sample Num: 3968, Cur Loss: 0.14601418, Cur Avg Loss: 0.47878947, Log Avg loss: 0.48979507, Global Avg Loss: 1.31618309, Time: 0.0076 Steps: 81150, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000258, Sample Num: 4128, Cur Loss: 0.38838568, Cur Avg Loss: 0.47666236, Log Avg loss: 0.42391004, Global Avg Loss: 1.31607315, Time: 0.0078 Steps: 81160, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000268, Sample Num: 4288, Cur Loss: 0.17993495, Cur Avg Loss: 0.47501073, Log Avg loss: 0.43239869, Global Avg Loss: 1.31596428, Time: 0.0077 Steps: 81170, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000278, Sample Num: 4448, Cur Loss: 0.34536579, Cur Avg Loss: 0.47578541, Log Avg loss: 0.49654679, Global Avg Loss: 1.31586334, Time: 0.0075 Steps: 81180, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000288, Sample Num: 4608, Cur Loss: 0.42952660, Cur Avg Loss: 0.47793530, Log Avg loss: 0.53770229, Global Avg Loss: 1.31576750, Time: 0.0075 Steps: 81190, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000298, Sample Num: 4768, Cur Loss: 0.33207050, Cur Avg Loss: 0.47450053, Log Avg loss: 0.37557896, Global Avg Loss: 1.31565171, Time: 0.0075 Steps: 81200, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000308, Sample Num: 4928, Cur Loss: 0.30242255, Cur Avg Loss: 0.47379544, Log Avg loss: 0.45278400, Global Avg Loss: 1.31554546, Time: 0.0075 Steps: 81210, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000318, Sample Num: 5088, Cur Loss: 0.15949483, Cur Avg Loss: 0.47310628, Log Avg loss: 0.45187990, Global Avg Loss: 1.31543912, Time: 0.0076 Steps: 81220, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000328, Sample Num: 5248, Cur Loss: 0.38118911, Cur Avg Loss: 0.47516935, Log Avg loss: 0.54077516, Global Avg Loss: 1.31534375, Time: 0.0076 Steps: 81230, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000338, Sample Num: 5408, Cur Loss: 0.31517476, Cur Avg Loss: 0.47621347, Log Avg loss: 0.51046074, Global Avg Loss: 1.31524468, Time: 0.0076 Steps: 81240, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000348, Sample Num: 5568, Cur Loss: 0.57582206, Cur Avg Loss: 0.47872683, Log Avg loss: 0.56367829, Global Avg Loss: 1.31515218, Time: 0.0076 Steps: 81250, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000358, Sample Num: 5728, Cur Loss: 0.31353414, Cur Avg Loss: 0.48220039, Log Avg loss: 0.60308006, Global Avg Loss: 1.31506455, Time: 0.0075 Steps: 81260, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000368, Sample Num: 5888, Cur Loss: 0.51491469, Cur Avg Loss: 0.48674019, Log Avg loss: 0.64926524, Global Avg Loss: 1.31498263, Time: 0.0075 Steps: 81270, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000378, Sample Num: 6048, Cur Loss: 0.30864444, Cur Avg Loss: 0.48914173, Log Avg loss: 0.57751835, Global Avg Loss: 1.31489189, Time: 0.0075 Steps: 81280, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000388, Sample Num: 6208, Cur Loss: 0.22660704, Cur Avg Loss: 0.49037211, Log Avg loss: 0.53688051, Global Avg Loss: 1.31479619, Time: 0.0076 Steps: 81290, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000398, Sample Num: 6368, Cur Loss: 0.27196869, Cur Avg Loss: 0.48741001, Log Avg loss: 0.37248065, Global Avg Loss: 1.31468028, Time: 0.0076 Steps: 81300, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000408, Sample Num: 6528, Cur Loss: 0.23307295, Cur Avg Loss: 0.48523916, Log Avg loss: 0.39883923, Global Avg Loss: 1.31456764, Time: 0.0076 Steps: 81310, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000418, Sample Num: 6688, Cur Loss: 0.27581611, Cur Avg Loss: 0.48439836, Log Avg loss: 0.45009360, Global Avg Loss: 1.31446134, Time: 0.0075 Steps: 81320, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000428, Sample Num: 6848, Cur Loss: 0.90818501, Cur Avg Loss: 0.48268375, Log Avg loss: 0.41101327, Global Avg Loss: 1.31435025, Time: 0.0075 Steps: 81330, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000438, Sample Num: 7008, Cur Loss: 0.39915419, Cur Avg Loss: 0.48299577, Log Avg loss: 0.49634989, Global Avg Loss: 1.31424969, Time: 0.0075 Steps: 81340, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000448, Sample Num: 7168, Cur Loss: 0.79009283, Cur Avg Loss: 0.48021803, Log Avg loss: 0.35855315, Global Avg Loss: 1.31413221, Time: 0.0076 Steps: 81350, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000458, Sample Num: 7328, Cur Loss: 0.23001108, Cur Avg Loss: 0.47781888, Log Avg loss: 0.37033706, Global Avg Loss: 1.31401621, Time: 0.0076 Steps: 81360, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000468, Sample Num: 7488, Cur Loss: 0.61528522, Cur Avg Loss: 0.47806653, Log Avg loss: 0.48940893, Global Avg Loss: 1.31391487, Time: 0.0075 Steps: 81370, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000478, Sample Num: 7648, Cur Loss: 0.74137211, Cur Avg Loss: 0.47940755, Log Avg loss: 0.54216719, Global Avg Loss: 1.31382003, Time: 0.0075 Steps: 81380, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000488, Sample Num: 7808, Cur Loss: 1.18182015, Cur Avg Loss: 0.47858940, Log Avg loss: 0.43948187, Global Avg Loss: 1.31371261, Time: 0.0076 Steps: 81390, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000498, Sample Num: 7968, Cur Loss: 0.25874841, Cur Avg Loss: 0.47993902, Log Avg loss: 0.54580025, Global Avg Loss: 1.31361827, Time: 0.0076 Steps: 81400, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000508, Sample Num: 8128, Cur Loss: 1.09342790, Cur Avg Loss: 0.48309826, Log Avg loss: 0.64042883, Global Avg Loss: 1.31353558, Time: 0.0075 Steps: 81410, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000518, Sample Num: 8288, Cur Loss: 0.20184855, Cur Avg Loss: 0.48295919, Log Avg loss: 0.47589442, Global Avg Loss: 1.31343270, Time: 0.0076 Steps: 81420, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000528, Sample Num: 8448, Cur Loss: 0.46357042, Cur Avg Loss: 0.48341190, Log Avg loss: 0.50686192, Global Avg Loss: 1.31333365, Time: 0.0076 Steps: 81430, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000538, Sample Num: 8608, Cur Loss: 0.35891250, Cur Avg Loss: 0.48041230, Log Avg loss: 0.32203334, Global Avg Loss: 1.31321193, Time: 0.0076 Steps: 81440, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000548, Sample Num: 8768, Cur Loss: 0.28511578, Cur Avg Loss: 0.48053048, Log Avg loss: 0.48688870, Global Avg Loss: 1.31311048, Time: 0.0075 Steps: 81450, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000558, Sample Num: 8928, Cur Loss: 0.41681325, Cur Avg Loss: 0.48200606, Log Avg loss: 0.56286796, Global Avg Loss: 1.31301838, Time: 0.0076 Steps: 81460, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000568, Sample Num: 9088, Cur Loss: 0.11769192, Cur Avg Loss: 0.47961662, Log Avg loss: 0.34628580, Global Avg Loss: 1.31289972, Time: 0.0076 Steps: 81470, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000578, Sample Num: 9248, Cur Loss: 0.16577064, Cur Avg Loss: 0.47990884, Log Avg loss: 0.49650698, Global Avg Loss: 1.31279952, Time: 0.0076 Steps: 81480, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000588, Sample Num: 9408, Cur Loss: 0.42869836, Cur Avg Loss: 0.47901132, Log Avg loss: 0.42713472, Global Avg Loss: 1.31269084, Time: 0.0076 Steps: 81490, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000598, Sample Num: 9568, Cur Loss: 0.40952888, Cur Avg Loss: 0.48028833, Log Avg loss: 0.55537652, Global Avg Loss: 1.31259791, Time: 0.0076 Steps: 81500, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000608, Sample Num: 9728, Cur Loss: 0.54669762, Cur Avg Loss: 0.48066140, Log Avg loss: 0.50297097, Global Avg Loss: 1.31249859, Time: 0.0075 Steps: 81510, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000618, Sample Num: 9888, Cur Loss: 0.46602505, Cur Avg Loss: 0.47994251, Log Avg loss: 0.43623427, Global Avg Loss: 1.31239109, Time: 0.0075 Steps: 81520, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000628, Sample Num: 10048, Cur Loss: 0.34722763, Cur Avg Loss: 0.47983667, Log Avg loss: 0.47329556, Global Avg Loss: 1.31228818, Time: 0.0076 Steps: 81530, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000638, Sample Num: 10208, Cur Loss: 0.13325566, Cur Avg Loss: 0.47974469, Log Avg loss: 0.47396833, Global Avg Loss: 1.31218537, Time: 0.0076 Steps: 81540, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000648, Sample Num: 10368, Cur Loss: 0.71867996, Cur Avg Loss: 0.48090444, Log Avg loss: 0.55489613, Global Avg Loss: 1.31209250, Time: 0.0076 Steps: 81550, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000658, Sample Num: 10528, Cur Loss: 0.48045257, Cur Avg Loss: 0.48060487, Log Avg loss: 0.46119280, Global Avg Loss: 1.31198818, Time: 0.0077 Steps: 81560, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000668, Sample Num: 10688, Cur Loss: 0.24622443, Cur Avg Loss: 0.48019618, Log Avg loss: 0.45330446, Global Avg Loss: 1.31188291, Time: 0.0075 Steps: 81570, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000678, Sample Num: 10848, Cur Loss: 0.58271444, Cur Avg Loss: 0.48046127, Log Avg loss: 0.49816912, Global Avg Loss: 1.31178316, Time: 0.0076 Steps: 81580, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000688, Sample Num: 11008, Cur Loss: 0.83038014, Cur Avg Loss: 0.48099450, Log Avg loss: 0.51714782, Global Avg Loss: 1.31168577, Time: 0.0076 Steps: 81590, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000698, Sample Num: 11168, Cur Loss: 0.71611619, Cur Avg Loss: 0.48053603, Log Avg loss: 0.44899298, Global Avg Loss: 1.31158005, Time: 0.0076 Steps: 81600, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000708, Sample Num: 11328, Cur Loss: 0.59245706, Cur Avg Loss: 0.48108736, Log Avg loss: 0.51957074, Global Avg Loss: 1.31148300, Time: 0.0075 Steps: 81610, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000718, Sample Num: 11488, Cur Loss: 0.18521589, Cur Avg Loss: 0.47934722, Log Avg loss: 0.35614536, Global Avg Loss: 1.31136595, Time: 0.0076 Steps: 81620, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000728, Sample Num: 11648, Cur Loss: 0.57931751, Cur Avg Loss: 0.48248534, Log Avg loss: 0.70780181, Global Avg Loss: 1.31129201, Time: 0.0077 Steps: 81630, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000738, Sample Num: 11808, Cur Loss: 0.70128226, Cur Avg Loss: 0.48284616, Log Avg loss: 0.50911418, Global Avg Loss: 1.31119375, Time: 0.0076 Steps: 81640, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000748, Sample Num: 11968, Cur Loss: 0.84403956, Cur Avg Loss: 0.48296207, Log Avg loss: 0.49151623, Global Avg Loss: 1.31109336, Time: 0.0075 Steps: 81650, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000758, Sample Num: 12128, Cur Loss: 0.27247566, Cur Avg Loss: 0.48184251, Log Avg loss: 0.39809968, Global Avg Loss: 1.31098156, Time: 0.0075 Steps: 81660, Updated lr: 0.000024 Training, Epoch: 0039, Batch: 000768, Sample Num: 12288, Cur Loss: 0.27647364, Cur Avg Loss: 0.47995745, Log Avg loss: 0.33706976, Global Avg Loss: 1.31086231, Time: 0.0084 Steps: 81670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000778, Sample Num: 12448, Cur Loss: 0.30256695, Cur Avg Loss: 0.48057313, Log Avg loss: 0.52785711, Global Avg Loss: 1.31076645, Time: 0.0076 Steps: 81680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000788, Sample Num: 12608, Cur Loss: 0.38003275, Cur Avg Loss: 0.48197774, Log Avg loss: 0.59125635, Global Avg Loss: 1.31067837, Time: 0.0075 Steps: 81690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000798, Sample Num: 12768, Cur Loss: 0.13523965, Cur Avg Loss: 0.48221947, Log Avg loss: 0.50126753, Global Avg Loss: 1.31057930, Time: 0.0077 Steps: 81700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000808, Sample Num: 12928, Cur Loss: 0.23131913, Cur Avg Loss: 0.48387283, Log Avg loss: 0.61581098, Global Avg Loss: 1.31049427, Time: 0.0076 Steps: 81710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000818, Sample Num: 13088, Cur Loss: 0.18058297, Cur Avg Loss: 0.48405620, Log Avg loss: 0.49887275, Global Avg Loss: 1.31039495, Time: 0.0076 Steps: 81720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000828, Sample Num: 13248, Cur Loss: 0.20470427, Cur Avg Loss: 0.48607482, Log Avg loss: 0.65119759, Global Avg Loss: 1.31031430, Time: 0.0076 Steps: 81730, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000838, Sample Num: 13408, Cur Loss: 0.34271866, Cur Avg Loss: 0.48478930, Log Avg loss: 0.37834882, Global Avg Loss: 1.31020028, Time: 0.0077 Steps: 81740, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000848, Sample Num: 13568, Cur Loss: 0.19579123, Cur Avg Loss: 0.48397866, Log Avg loss: 0.41604676, Global Avg Loss: 1.31009091, Time: 0.0076 Steps: 81750, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000858, Sample Num: 13728, Cur Loss: 0.42746633, Cur Avg Loss: 0.48497703, Log Avg loss: 0.56963864, Global Avg Loss: 1.31000034, Time: 0.0075 Steps: 81760, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000868, Sample Num: 13888, Cur Loss: 0.85350382, Cur Avg Loss: 0.48721944, Log Avg loss: 0.67961874, Global Avg Loss: 1.30992325, Time: 0.0076 Steps: 81770, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000878, Sample Num: 14048, Cur Loss: 0.65609682, Cur Avg Loss: 0.48721749, Log Avg loss: 0.48704818, Global Avg Loss: 1.30982263, Time: 0.0068 Steps: 81780, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000888, Sample Num: 14208, Cur Loss: 0.28014779, Cur Avg Loss: 0.48694106, Log Avg loss: 0.46267026, Global Avg Loss: 1.30971905, Time: 0.0065 Steps: 81790, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000898, Sample Num: 14368, Cur Loss: 0.21286741, Cur Avg Loss: 0.48613213, Log Avg loss: 0.41429910, Global Avg Loss: 1.30960959, Time: 0.0066 Steps: 81800, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000908, Sample Num: 14528, Cur Loss: 0.35391641, Cur Avg Loss: 0.48623197, Log Avg loss: 0.49519725, Global Avg Loss: 1.30951004, Time: 0.0206 Steps: 81810, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000918, Sample Num: 14688, Cur Loss: 0.93146795, Cur Avg Loss: 0.48686295, Log Avg loss: 0.54415627, Global Avg Loss: 1.30941650, Time: 0.0064 Steps: 81820, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000928, Sample Num: 14848, Cur Loss: 0.45883468, Cur Avg Loss: 0.48705001, Log Avg loss: 0.50422248, Global Avg Loss: 1.30931810, Time: 0.0202 Steps: 81830, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000938, Sample Num: 15008, Cur Loss: 0.41178420, Cur Avg Loss: 0.48835246, Log Avg loss: 0.60921964, Global Avg Loss: 1.30923255, Time: 0.0066 Steps: 81840, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000948, Sample Num: 15168, Cur Loss: 0.17604432, Cur Avg Loss: 0.48799317, Log Avg loss: 0.45429142, Global Avg Loss: 1.30912810, Time: 0.0063 Steps: 81850, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000958, Sample Num: 15328, Cur Loss: 0.56101876, Cur Avg Loss: 0.48797957, Log Avg loss: 0.48669102, Global Avg Loss: 1.30902763, Time: 0.0064 Steps: 81860, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000968, Sample Num: 15488, Cur Loss: 0.33447587, Cur Avg Loss: 0.48699838, Log Avg loss: 0.39300027, Global Avg Loss: 1.30891574, Time: 0.0066 Steps: 81870, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000978, Sample Num: 15648, Cur Loss: 0.20092136, Cur Avg Loss: 0.48628503, Log Avg loss: 0.41723204, Global Avg Loss: 1.30880684, Time: 0.0066 Steps: 81880, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000988, Sample Num: 15808, Cur Loss: 0.94962305, Cur Avg Loss: 0.48650828, Log Avg loss: 0.50834263, Global Avg Loss: 1.30870909, Time: 0.0064 Steps: 81890, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 000998, Sample Num: 15968, Cur Loss: 0.13733783, Cur Avg Loss: 0.48694247, Log Avg loss: 0.52984015, Global Avg Loss: 1.30861399, Time: 0.0066 Steps: 81900, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001008, Sample Num: 16128, Cur Loss: 0.97166097, Cur Avg Loss: 0.48774230, Log Avg loss: 0.56756580, Global Avg Loss: 1.30852352, Time: 0.0066 Steps: 81910, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001018, Sample Num: 16288, Cur Loss: 1.11012018, Cur Avg Loss: 0.48818889, Log Avg loss: 0.53320440, Global Avg Loss: 1.30842888, Time: 0.0066 Steps: 81920, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001028, Sample Num: 16448, Cur Loss: 0.90290189, Cur Avg Loss: 0.48887042, Log Avg loss: 0.55825031, Global Avg Loss: 1.30833732, Time: 0.0217 Steps: 81930, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001038, Sample Num: 16608, Cur Loss: 0.39654818, Cur Avg Loss: 0.48859658, Log Avg loss: 0.46044564, Global Avg Loss: 1.30823384, Time: 0.0066 Steps: 81940, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001048, Sample Num: 16768, Cur Loss: 0.44484013, Cur Avg Loss: 0.48843781, Log Avg loss: 0.47195800, Global Avg Loss: 1.30813179, Time: 0.0176 Steps: 81950, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001058, Sample Num: 16928, Cur Loss: 0.81154847, Cur Avg Loss: 0.48733961, Log Avg loss: 0.37224790, Global Avg Loss: 1.30801760, Time: 0.0185 Steps: 81960, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001068, Sample Num: 17088, Cur Loss: 0.53077734, Cur Avg Loss: 0.48753051, Log Avg loss: 0.50772839, Global Avg Loss: 1.30791997, Time: 0.0073 Steps: 81970, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001078, Sample Num: 17248, Cur Loss: 0.59659612, Cur Avg Loss: 0.48768521, Log Avg loss: 0.50420718, Global Avg Loss: 1.30782194, Time: 0.0186 Steps: 81980, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001088, Sample Num: 17408, Cur Loss: 0.26703930, Cur Avg Loss: 0.48737624, Log Avg loss: 0.45406928, Global Avg Loss: 1.30771781, Time: 0.0064 Steps: 81990, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001098, Sample Num: 17568, Cur Loss: 0.40878731, Cur Avg Loss: 0.48777577, Log Avg loss: 0.53124390, Global Avg Loss: 1.30762311, Time: 0.0194 Steps: 82000, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001108, Sample Num: 17728, Cur Loss: 0.31170207, Cur Avg Loss: 0.48762039, Log Avg loss: 0.47056060, Global Avg Loss: 1.30752105, Time: 0.0197 Steps: 82010, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001118, Sample Num: 17888, Cur Loss: 0.22863862, Cur Avg Loss: 0.48748262, Log Avg loss: 0.47221771, Global Avg Loss: 1.30741920, Time: 0.0072 Steps: 82020, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001128, Sample Num: 18048, Cur Loss: 0.20560348, Cur Avg Loss: 0.48783848, Log Avg loss: 0.52762291, Global Avg Loss: 1.30732414, Time: 0.0066 Steps: 82030, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001138, Sample Num: 18208, Cur Loss: 0.59374601, Cur Avg Loss: 0.48723818, Log Avg loss: 0.41952481, Global Avg Loss: 1.30721593, Time: 0.0199 Steps: 82040, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001148, Sample Num: 18368, Cur Loss: 0.11480290, Cur Avg Loss: 0.48801446, Log Avg loss: 0.57635511, Global Avg Loss: 1.30712685, Time: 0.0225 Steps: 82050, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001158, Sample Num: 18528, Cur Loss: 0.82852459, Cur Avg Loss: 0.48759868, Log Avg loss: 0.43986635, Global Avg Loss: 1.30702117, Time: 0.0066 Steps: 82060, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001168, Sample Num: 18688, Cur Loss: 0.49990112, Cur Avg Loss: 0.48763994, Log Avg loss: 0.49241810, Global Avg Loss: 1.30692191, Time: 0.0065 Steps: 82070, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001178, Sample Num: 18848, Cur Loss: 0.11931794, Cur Avg Loss: 0.48712267, Log Avg loss: 0.42670592, Global Avg Loss: 1.30681467, Time: 0.0065 Steps: 82080, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001188, Sample Num: 19008, Cur Loss: 1.18836737, Cur Avg Loss: 0.48962566, Log Avg loss: 0.78447728, Global Avg Loss: 1.30675104, Time: 0.0065 Steps: 82090, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001198, Sample Num: 19168, Cur Loss: 0.24165359, Cur Avg Loss: 0.48989867, Log Avg loss: 0.52233235, Global Avg Loss: 1.30665550, Time: 0.0065 Steps: 82100, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001208, Sample Num: 19328, Cur Loss: 0.41916883, Cur Avg Loss: 0.48835558, Log Avg loss: 0.30349344, Global Avg Loss: 1.30653332, Time: 0.0064 Steps: 82110, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001218, Sample Num: 19488, Cur Loss: 0.68048418, Cur Avg Loss: 0.48848174, Log Avg loss: 0.50372191, Global Avg Loss: 1.30643556, Time: 0.0074 Steps: 82120, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001228, Sample Num: 19648, Cur Loss: 0.42571533, Cur Avg Loss: 0.48856278, Log Avg loss: 0.49843378, Global Avg Loss: 1.30633718, Time: 0.0074 Steps: 82130, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001238, Sample Num: 19808, Cur Loss: 0.43045288, Cur Avg Loss: 0.48779524, Log Avg loss: 0.39354158, Global Avg Loss: 1.30622605, Time: 0.0075 Steps: 82140, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001248, Sample Num: 19968, Cur Loss: 0.39162773, Cur Avg Loss: 0.48819588, Log Avg loss: 0.53779501, Global Avg Loss: 1.30613251, Time: 0.0076 Steps: 82150, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001258, Sample Num: 20128, Cur Loss: 0.45007569, Cur Avg Loss: 0.48834197, Log Avg loss: 0.50657339, Global Avg Loss: 1.30603520, Time: 0.0076 Steps: 82160, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001268, Sample Num: 20288, Cur Loss: 0.20603177, Cur Avg Loss: 0.48808977, Log Avg loss: 0.45636284, Global Avg Loss: 1.30593179, Time: 0.0076 Steps: 82170, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001278, Sample Num: 20448, Cur Loss: 0.35325506, Cur Avg Loss: 0.48828002, Log Avg loss: 0.51240392, Global Avg Loss: 1.30583523, Time: 0.0078 Steps: 82180, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001288, Sample Num: 20608, Cur Loss: 0.20927583, Cur Avg Loss: 0.49022939, Log Avg loss: 0.73935915, Global Avg Loss: 1.30576631, Time: 0.0075 Steps: 82190, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001298, Sample Num: 20768, Cur Loss: 0.45856446, Cur Avg Loss: 0.48965602, Log Avg loss: 0.41580570, Global Avg Loss: 1.30565804, Time: 0.0077 Steps: 82200, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001308, Sample Num: 20928, Cur Loss: 0.62752950, Cur Avg Loss: 0.48978771, Log Avg loss: 0.50688191, Global Avg Loss: 1.30556088, Time: 0.0076 Steps: 82210, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001318, Sample Num: 21088, Cur Loss: 0.79489112, Cur Avg Loss: 0.48963383, Log Avg loss: 0.46950602, Global Avg Loss: 1.30545919, Time: 0.0075 Steps: 82220, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001328, Sample Num: 21248, Cur Loss: 0.60900497, Cur Avg Loss: 0.49008288, Log Avg loss: 0.54926694, Global Avg Loss: 1.30536723, Time: 0.0076 Steps: 82230, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001338, Sample Num: 21408, Cur Loss: 0.52206075, Cur Avg Loss: 0.48994928, Log Avg loss: 0.47220715, Global Avg Loss: 1.30526593, Time: 0.0075 Steps: 82240, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001348, Sample Num: 21568, Cur Loss: 0.70521665, Cur Avg Loss: 0.49018427, Log Avg loss: 0.52162636, Global Avg Loss: 1.30517065, Time: 0.0077 Steps: 82250, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001358, Sample Num: 21728, Cur Loss: 0.67864287, Cur Avg Loss: 0.49075248, Log Avg loss: 0.56734674, Global Avg Loss: 1.30508096, Time: 0.0075 Steps: 82260, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001368, Sample Num: 21888, Cur Loss: 0.26679549, Cur Avg Loss: 0.49008123, Log Avg loss: 0.39892618, Global Avg Loss: 1.30497081, Time: 0.0076 Steps: 82270, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001378, Sample Num: 22048, Cur Loss: 0.95579058, Cur Avg Loss: 0.48978731, Log Avg loss: 0.44957910, Global Avg Loss: 1.30486685, Time: 0.0077 Steps: 82280, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001388, Sample Num: 22208, Cur Loss: 0.34363905, Cur Avg Loss: 0.49009043, Log Avg loss: 0.53186095, Global Avg Loss: 1.30477291, Time: 0.0076 Steps: 82290, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001398, Sample Num: 22368, Cur Loss: 0.38043183, Cur Avg Loss: 0.48981151, Log Avg loss: 0.45109709, Global Avg Loss: 1.30466919, Time: 0.0076 Steps: 82300, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001408, Sample Num: 22528, Cur Loss: 0.29474533, Cur Avg Loss: 0.48985044, Log Avg loss: 0.49529231, Global Avg Loss: 1.30457085, Time: 0.0077 Steps: 82310, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001418, Sample Num: 22688, Cur Loss: 0.65085959, Cur Avg Loss: 0.49091791, Log Avg loss: 0.64121718, Global Avg Loss: 1.30449027, Time: 0.0076 Steps: 82320, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001428, Sample Num: 22848, Cur Loss: 0.52350998, Cur Avg Loss: 0.49146672, Log Avg loss: 0.56928805, Global Avg Loss: 1.30440097, Time: 0.0076 Steps: 82330, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001438, Sample Num: 23008, Cur Loss: 0.22636129, Cur Avg Loss: 0.49133531, Log Avg loss: 0.47257114, Global Avg Loss: 1.30429995, Time: 0.0076 Steps: 82340, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001448, Sample Num: 23168, Cur Loss: 1.05451727, Cur Avg Loss: 0.49080147, Log Avg loss: 0.41403532, Global Avg Loss: 1.30419184, Time: 0.0075 Steps: 82350, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001458, Sample Num: 23328, Cur Loss: 0.08177796, Cur Avg Loss: 0.49095480, Log Avg loss: 0.51315612, Global Avg Loss: 1.30409580, Time: 0.0075 Steps: 82360, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001468, Sample Num: 23488, Cur Loss: 0.65780675, Cur Avg Loss: 0.49276344, Log Avg loss: 0.75646350, Global Avg Loss: 1.30402931, Time: 0.0075 Steps: 82370, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001478, Sample Num: 23648, Cur Loss: 0.48170355, Cur Avg Loss: 0.49288493, Log Avg loss: 0.51071886, Global Avg Loss: 1.30393301, Time: 0.0076 Steps: 82380, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001488, Sample Num: 23808, Cur Loss: 0.28214633, Cur Avg Loss: 0.49210831, Log Avg loss: 0.37732457, Global Avg Loss: 1.30382055, Time: 0.0076 Steps: 82390, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001498, Sample Num: 23968, Cur Loss: 0.23300606, Cur Avg Loss: 0.49161176, Log Avg loss: 0.41772491, Global Avg Loss: 1.30371301, Time: 0.0076 Steps: 82400, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001508, Sample Num: 24128, Cur Loss: 0.42801189, Cur Avg Loss: 0.49120762, Log Avg loss: 0.43066775, Global Avg Loss: 1.30360707, Time: 0.0076 Steps: 82410, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001518, Sample Num: 24288, Cur Loss: 0.22226933, Cur Avg Loss: 0.49194453, Log Avg loss: 0.60307077, Global Avg Loss: 1.30352207, Time: 0.0076 Steps: 82420, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001528, Sample Num: 24448, Cur Loss: 0.70554918, Cur Avg Loss: 0.49082246, Log Avg loss: 0.32049187, Global Avg Loss: 1.30340282, Time: 0.0076 Steps: 82430, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001538, Sample Num: 24608, Cur Loss: 0.36601114, Cur Avg Loss: 0.49126035, Log Avg loss: 0.55817081, Global Avg Loss: 1.30331242, Time: 0.0076 Steps: 82440, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001548, Sample Num: 24768, Cur Loss: 0.54905015, Cur Avg Loss: 0.49060912, Log Avg loss: 0.39044945, Global Avg Loss: 1.30320170, Time: 0.0076 Steps: 82450, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001558, Sample Num: 24928, Cur Loss: 0.29500613, Cur Avg Loss: 0.49054767, Log Avg loss: 0.48103425, Global Avg Loss: 1.30310200, Time: 0.0076 Steps: 82460, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001568, Sample Num: 25088, Cur Loss: 0.21288195, Cur Avg Loss: 0.49125298, Log Avg loss: 0.60114146, Global Avg Loss: 1.30301688, Time: 0.0076 Steps: 82470, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001578, Sample Num: 25248, Cur Loss: 0.28311428, Cur Avg Loss: 0.49096332, Log Avg loss: 0.44554381, Global Avg Loss: 1.30291292, Time: 0.0075 Steps: 82480, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001588, Sample Num: 25408, Cur Loss: 0.20498957, Cur Avg Loss: 0.49104347, Log Avg loss: 0.50369126, Global Avg Loss: 1.30281603, Time: 0.0076 Steps: 82490, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001598, Sample Num: 25568, Cur Loss: 1.05951524, Cur Avg Loss: 0.49305622, Log Avg loss: 0.81268128, Global Avg Loss: 1.30275662, Time: 0.0075 Steps: 82500, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001608, Sample Num: 25728, Cur Loss: 0.66798872, Cur Avg Loss: 0.49269892, Log Avg loss: 0.43560251, Global Avg Loss: 1.30265153, Time: 0.0076 Steps: 82510, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001618, Sample Num: 25888, Cur Loss: 0.26500535, Cur Avg Loss: 0.49286907, Log Avg loss: 0.52022858, Global Avg Loss: 1.30255671, Time: 0.0076 Steps: 82520, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001628, Sample Num: 26048, Cur Loss: 0.40651950, Cur Avg Loss: 0.49206837, Log Avg loss: 0.36251506, Global Avg Loss: 1.30244281, Time: 0.0075 Steps: 82530, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001638, Sample Num: 26208, Cur Loss: 0.14173537, Cur Avg Loss: 0.49203304, Log Avg loss: 0.48628184, Global Avg Loss: 1.30234393, Time: 0.0078 Steps: 82540, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001648, Sample Num: 26368, Cur Loss: 0.87888795, Cur Avg Loss: 0.49196543, Log Avg loss: 0.48089003, Global Avg Loss: 1.30224442, Time: 0.0075 Steps: 82550, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001658, Sample Num: 26528, Cur Loss: 0.71551228, Cur Avg Loss: 0.49154268, Log Avg loss: 0.42187445, Global Avg Loss: 1.30213778, Time: 0.0078 Steps: 82560, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001668, Sample Num: 26688, Cur Loss: 0.55774629, Cur Avg Loss: 0.49219512, Log Avg loss: 0.60036864, Global Avg Loss: 1.30205279, Time: 0.0076 Steps: 82570, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001678, Sample Num: 26848, Cur Loss: 0.32997525, Cur Avg Loss: 0.49155040, Log Avg loss: 0.38401241, Global Avg Loss: 1.30194162, Time: 0.0077 Steps: 82580, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001688, Sample Num: 27008, Cur Loss: 0.41064262, Cur Avg Loss: 0.49153411, Log Avg loss: 0.48879958, Global Avg Loss: 1.30184317, Time: 0.0077 Steps: 82590, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001698, Sample Num: 27168, Cur Loss: 0.43945831, Cur Avg Loss: 0.49185697, Log Avg loss: 0.54635701, Global Avg Loss: 1.30175170, Time: 0.0074 Steps: 82600, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001708, Sample Num: 27328, Cur Loss: 0.88458002, Cur Avg Loss: 0.49224880, Log Avg loss: 0.55878035, Global Avg Loss: 1.30166177, Time: 0.0075 Steps: 82610, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001718, Sample Num: 27488, Cur Loss: 0.49299848, Cur Avg Loss: 0.49257863, Log Avg loss: 0.54891372, Global Avg Loss: 1.30157066, Time: 0.0074 Steps: 82620, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001728, Sample Num: 27648, Cur Loss: 0.50577259, Cur Avg Loss: 0.49304059, Log Avg loss: 0.57240580, Global Avg Loss: 1.30148241, Time: 0.0074 Steps: 82630, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001738, Sample Num: 27808, Cur Loss: 0.45666471, Cur Avg Loss: 0.49338285, Log Avg loss: 0.55252567, Global Avg Loss: 1.30139178, Time: 0.0074 Steps: 82640, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001748, Sample Num: 27968, Cur Loss: 0.33907750, Cur Avg Loss: 0.49317525, Log Avg loss: 0.45709345, Global Avg Loss: 1.30128963, Time: 0.0075 Steps: 82650, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001758, Sample Num: 28128, Cur Loss: 0.49493381, Cur Avg Loss: 0.49337189, Log Avg loss: 0.52774580, Global Avg Loss: 1.30119605, Time: 0.0074 Steps: 82660, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001768, Sample Num: 28288, Cur Loss: 0.25625011, Cur Avg Loss: 0.49424313, Log Avg loss: 0.64740674, Global Avg Loss: 1.30111696, Time: 0.0075 Steps: 82670, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001778, Sample Num: 28448, Cur Loss: 0.23960619, Cur Avg Loss: 0.49429994, Log Avg loss: 0.50434299, Global Avg Loss: 1.30102060, Time: 0.0074 Steps: 82680, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001788, Sample Num: 28608, Cur Loss: 0.19878200, Cur Avg Loss: 0.49368155, Log Avg loss: 0.38373270, Global Avg Loss: 1.30090966, Time: 0.0075 Steps: 82690, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001798, Sample Num: 28768, Cur Loss: 0.46760893, Cur Avg Loss: 0.49350953, Log Avg loss: 0.46275250, Global Avg Loss: 1.30080832, Time: 0.0074 Steps: 82700, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001808, Sample Num: 28928, Cur Loss: 0.38721067, Cur Avg Loss: 0.49303204, Log Avg loss: 0.40717942, Global Avg Loss: 1.30070027, Time: 0.0074 Steps: 82710, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001818, Sample Num: 29088, Cur Loss: 0.39351997, Cur Avg Loss: 0.49296384, Log Avg loss: 0.48063320, Global Avg Loss: 1.30060113, Time: 0.0074 Steps: 82720, Updated lr: 0.000023 Training, Epoch: 0039, Batch: 001828, Sample Num: 29248, Cur Loss: 0.86208844, Cur Avg Loss: 0.49279891, Log Avg loss: 0.46281435, Global Avg Loss: 1.30049987, Time: 0.0074 Steps: 82730, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001838, Sample Num: 29408, Cur Loss: 0.57748687, Cur Avg Loss: 0.49306290, Log Avg loss: 0.54131994, Global Avg Loss: 1.30040811, Time: 0.0075 Steps: 82740, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001848, Sample Num: 29568, Cur Loss: 0.86407691, Cur Avg Loss: 0.49334677, Log Avg loss: 0.54552283, Global Avg Loss: 1.30031689, Time: 0.0083 Steps: 82750, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001858, Sample Num: 29728, Cur Loss: 0.23361312, Cur Avg Loss: 0.49326176, Log Avg loss: 0.47755220, Global Avg Loss: 1.30021747, Time: 0.0193 Steps: 82760, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001868, Sample Num: 29888, Cur Loss: 0.38171247, Cur Avg Loss: 0.49259788, Log Avg loss: 0.36924903, Global Avg Loss: 1.30010499, Time: 0.0237 Steps: 82770, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001878, Sample Num: 30048, Cur Loss: 0.24226141, Cur Avg Loss: 0.49216616, Log Avg loss: 0.41151941, Global Avg Loss: 1.29999765, Time: 0.0206 Steps: 82780, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001888, Sample Num: 30208, Cur Loss: 0.32380214, Cur Avg Loss: 0.49164050, Log Avg loss: 0.39292169, Global Avg Loss: 1.29988809, Time: 0.0066 Steps: 82790, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001898, Sample Num: 30368, Cur Loss: 0.39384052, Cur Avg Loss: 0.49184743, Log Avg loss: 0.53091708, Global Avg Loss: 1.29979522, Time: 0.0204 Steps: 82800, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001908, Sample Num: 30528, Cur Loss: 0.21525660, Cur Avg Loss: 0.49230574, Log Avg loss: 0.57929154, Global Avg Loss: 1.29970821, Time: 0.0199 Steps: 82810, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001918, Sample Num: 30688, Cur Loss: 0.32671666, Cur Avg Loss: 0.49192269, Log Avg loss: 0.41883717, Global Avg Loss: 1.29960185, Time: 0.0063 Steps: 82820, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001928, Sample Num: 30848, Cur Loss: 0.28909105, Cur Avg Loss: 0.49122913, Log Avg loss: 0.35820442, Global Avg Loss: 1.29948820, Time: 0.0066 Steps: 82830, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001938, Sample Num: 31008, Cur Loss: 0.53127307, Cur Avg Loss: 0.49174875, Log Avg loss: 0.59193108, Global Avg Loss: 1.29940278, Time: 0.0225 Steps: 82840, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001948, Sample Num: 31168, Cur Loss: 0.63827646, Cur Avg Loss: 0.49150759, Log Avg loss: 0.44477075, Global Avg Loss: 1.29929963, Time: 0.0064 Steps: 82850, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001958, Sample Num: 31328, Cur Loss: 0.20306045, Cur Avg Loss: 0.49247999, Log Avg loss: 0.68190358, Global Avg Loss: 1.29922512, Time: 0.0077 Steps: 82860, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001968, Sample Num: 31488, Cur Loss: 0.29655942, Cur Avg Loss: 0.49278864, Log Avg loss: 0.55322329, Global Avg Loss: 1.29913510, Time: 0.0163 Steps: 82870, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001978, Sample Num: 31648, Cur Loss: 0.25700462, Cur Avg Loss: 0.49338962, Log Avg loss: 0.61166121, Global Avg Loss: 1.29905215, Time: 0.0064 Steps: 82880, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001988, Sample Num: 31808, Cur Loss: 0.72744793, Cur Avg Loss: 0.49310169, Log Avg loss: 0.43615034, Global Avg Loss: 1.29894805, Time: 0.0064 Steps: 82890, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 001998, Sample Num: 31968, Cur Loss: 0.32587236, Cur Avg Loss: 0.49357926, Log Avg loss: 0.58851957, Global Avg Loss: 1.29886235, Time: 0.0064 Steps: 82900, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002008, Sample Num: 32128, Cur Loss: 0.35948730, Cur Avg Loss: 0.49421836, Log Avg loss: 0.62191145, Global Avg Loss: 1.29878070, Time: 0.0063 Steps: 82910, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002018, Sample Num: 32288, Cur Loss: 0.41127643, Cur Avg Loss: 0.49360527, Log Avg loss: 0.37049571, Global Avg Loss: 1.29866875, Time: 0.0240 Steps: 82920, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002028, Sample Num: 32448, Cur Loss: 0.22246151, Cur Avg Loss: 0.49308586, Log Avg loss: 0.38826953, Global Avg Loss: 1.29855897, Time: 0.0113 Steps: 82930, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002038, Sample Num: 32608, Cur Loss: 0.55627584, Cur Avg Loss: 0.49337827, Log Avg loss: 0.55267778, Global Avg Loss: 1.29846904, Time: 0.0064 Steps: 82940, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002048, Sample Num: 32768, Cur Loss: 0.29421788, Cur Avg Loss: 0.49358737, Log Avg loss: 0.53620380, Global Avg Loss: 1.29837715, Time: 0.0082 Steps: 82950, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002058, Sample Num: 32928, Cur Loss: 0.13124499, Cur Avg Loss: 0.49347089, Log Avg loss: 0.46961482, Global Avg Loss: 1.29827725, Time: 0.0074 Steps: 82960, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002068, Sample Num: 33088, Cur Loss: 0.26553261, Cur Avg Loss: 0.49252726, Log Avg loss: 0.29832888, Global Avg Loss: 1.29815673, Time: 0.0074 Steps: 82970, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002078, Sample Num: 33248, Cur Loss: 0.21408647, Cur Avg Loss: 0.49268886, Log Avg loss: 0.52610796, Global Avg Loss: 1.29806369, Time: 0.0075 Steps: 82980, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002088, Sample Num: 33408, Cur Loss: 0.64387524, Cur Avg Loss: 0.49238747, Log Avg loss: 0.42975840, Global Avg Loss: 1.29795906, Time: 0.0075 Steps: 82990, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002098, Sample Num: 33568, Cur Loss: 0.58117294, Cur Avg Loss: 0.49255619, Log Avg loss: 0.52778353, Global Avg Loss: 1.29786627, Time: 0.0074 Steps: 83000, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002108, Sample Num: 33728, Cur Loss: 0.19703671, Cur Avg Loss: 0.49185148, Log Avg loss: 0.34400495, Global Avg Loss: 1.29775136, Time: 0.0074 Steps: 83010, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002118, Sample Num: 33888, Cur Loss: 0.19550399, Cur Avg Loss: 0.49242637, Log Avg loss: 0.61361257, Global Avg Loss: 1.29766895, Time: 0.0074 Steps: 83020, Updated lr: 0.000022 Training, Epoch: 0039, Batch: 002128, Sample Num: 34048, Cur Loss: 0.23291746, Cur Avg Loss: 0.49234045, Log Avg loss: 0.47414165, Global Avg Loss: 1.29756977, Time: 0.0074 Steps: 83030, Updated lr: 0.000022 ***** Running evaluation checkpoint-83031 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-83031 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 18.020426, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.591425, "eval_total_loss": 415.771613, "eval_mae": 0.549516, "eval_mse": 0.591633, "eval_r2": 0.623919, "eval_sp_statistic": 0.755333, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.79469, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.970113, "test_total_loss": 486.996646, "test_mae": 0.668473, "test_mse": 0.970413, "test_r2": 0.373687, "test_sp_statistic": 0.595964, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.670101, "test_ps_pvalue": 0.0, "lr": 2.2208629682313894e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.297558065110583, "train_cur_epoch_loss": 1048.0261227414012, "train_cur_epoch_avg_loss": 0.49226215253236316, "train_cur_epoch_time": 18.02042579650879, "train_cur_epoch_avg_time": 0.00846426763574861, "epoch": 39, "step": 83031} ################################################## Training, Epoch: 0040, Batch: 000009, Sample Num: 144, Cur Loss: 0.82905811, Cur Avg Loss: 0.47695108, Log Avg loss: 0.46182116, Global Avg Loss: 1.29746913, Time: 0.0074 Steps: 83040, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000019, Sample Num: 304, Cur Loss: 0.34765562, Cur Avg Loss: 0.48437317, Log Avg loss: 0.49105305, Global Avg Loss: 1.29737203, Time: 0.0074 Steps: 83050, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000029, Sample Num: 464, Cur Loss: 0.09473397, Cur Avg Loss: 0.46649406, Log Avg loss: 0.43252376, Global Avg Loss: 1.29726790, Time: 0.0075 Steps: 83060, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000039, Sample Num: 624, Cur Loss: 0.33432215, Cur Avg Loss: 0.44706086, Log Avg loss: 0.39070459, Global Avg Loss: 1.29715877, Time: 0.0076 Steps: 83070, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000049, Sample Num: 784, Cur Loss: 0.34308925, Cur Avg Loss: 0.43855147, Log Avg loss: 0.40536483, Global Avg Loss: 1.29705143, Time: 0.0075 Steps: 83080, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000059, Sample Num: 944, Cur Loss: 0.43823898, Cur Avg Loss: 0.45414773, Log Avg loss: 0.53056940, Global Avg Loss: 1.29695918, Time: 0.0077 Steps: 83090, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000069, Sample Num: 1104, Cur Loss: 0.87861466, Cur Avg Loss: 0.44784215, Log Avg loss: 0.41063925, Global Avg Loss: 1.29685252, Time: 0.0077 Steps: 83100, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000079, Sample Num: 1264, Cur Loss: 0.23467311, Cur Avg Loss: 0.44500912, Log Avg loss: 0.42546123, Global Avg Loss: 1.29674768, Time: 0.0077 Steps: 83110, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000089, Sample Num: 1424, Cur Loss: 0.61725581, Cur Avg Loss: 0.46140436, Log Avg loss: 0.59092672, Global Avg Loss: 1.29666276, Time: 0.0077 Steps: 83120, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000099, Sample Num: 1584, Cur Loss: 0.31447804, Cur Avg Loss: 0.46355241, Log Avg loss: 0.48267010, Global Avg Loss: 1.29656484, Time: 0.0081 Steps: 83130, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000109, Sample Num: 1744, Cur Loss: 0.43559408, Cur Avg Loss: 0.47077670, Log Avg loss: 0.54229715, Global Avg Loss: 1.29647412, Time: 0.0078 Steps: 83140, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000119, Sample Num: 1904, Cur Loss: 0.48901483, Cur Avg Loss: 0.47782868, Log Avg loss: 0.55469524, Global Avg Loss: 1.29638491, Time: 0.0077 Steps: 83150, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000129, Sample Num: 2064, Cur Loss: 0.26702148, Cur Avg Loss: 0.47978822, Log Avg loss: 0.50310670, Global Avg Loss: 1.29628952, Time: 0.0076 Steps: 83160, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000139, Sample Num: 2224, Cur Loss: 0.38078737, Cur Avg Loss: 0.48209118, Log Avg loss: 0.51179942, Global Avg Loss: 1.29619520, Time: 0.0083 Steps: 83170, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000149, Sample Num: 2384, Cur Loss: 0.48995066, Cur Avg Loss: 0.48154690, Log Avg loss: 0.47398143, Global Avg Loss: 1.29609635, Time: 0.0079 Steps: 83180, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000159, Sample Num: 2544, Cur Loss: 0.40900534, Cur Avg Loss: 0.47923703, Log Avg loss: 0.44481995, Global Avg Loss: 1.29599402, Time: 0.0078 Steps: 83190, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000169, Sample Num: 2704, Cur Loss: 0.23492196, Cur Avg Loss: 0.47198468, Log Avg loss: 0.35667228, Global Avg Loss: 1.29588112, Time: 0.0082 Steps: 83200, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000179, Sample Num: 2864, Cur Loss: 0.45301703, Cur Avg Loss: 0.47179114, Log Avg loss: 0.46852033, Global Avg Loss: 1.29578169, Time: 0.0083 Steps: 83210, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000189, Sample Num: 3024, Cur Loss: 0.48371193, Cur Avg Loss: 0.46506471, Log Avg loss: 0.34466155, Global Avg Loss: 1.29566740, Time: 0.0079 Steps: 83220, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000199, Sample Num: 3184, Cur Loss: 0.35747144, Cur Avg Loss: 0.46607097, Log Avg loss: 0.48508926, Global Avg Loss: 1.29557001, Time: 0.0078 Steps: 83230, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000209, Sample Num: 3344, Cur Loss: 0.15801525, Cur Avg Loss: 0.46492677, Log Avg loss: 0.44215723, Global Avg Loss: 1.29546748, Time: 0.0083 Steps: 83240, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000219, Sample Num: 3504, Cur Loss: 0.51148307, Cur Avg Loss: 0.46539981, Log Avg loss: 0.47528643, Global Avg Loss: 1.29536896, Time: 0.0083 Steps: 83250, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000229, Sample Num: 3664, Cur Loss: 0.29052678, Cur Avg Loss: 0.46783572, Log Avg loss: 0.52118215, Global Avg Loss: 1.29527598, Time: 0.0078 Steps: 83260, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000239, Sample Num: 3824, Cur Loss: 0.25043729, Cur Avg Loss: 0.46576913, Log Avg loss: 0.41844412, Global Avg Loss: 1.29517068, Time: 0.0079 Steps: 83270, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000249, Sample Num: 3984, Cur Loss: 0.92288828, Cur Avg Loss: 0.46843054, Log Avg loss: 0.53203837, Global Avg Loss: 1.29507905, Time: 0.0083 Steps: 83280, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000259, Sample Num: 4144, Cur Loss: 0.30433404, Cur Avg Loss: 0.46583882, Log Avg loss: 0.40130490, Global Avg Loss: 1.29497174, Time: 0.0077 Steps: 83290, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000269, Sample Num: 4304, Cur Loss: 0.18111855, Cur Avg Loss: 0.46685234, Log Avg loss: 0.49310249, Global Avg Loss: 1.29487547, Time: 0.0077 Steps: 83300, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000279, Sample Num: 4464, Cur Loss: 0.75509012, Cur Avg Loss: 0.46821747, Log Avg loss: 0.50493963, Global Avg Loss: 1.29478066, Time: 0.0077 Steps: 83310, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000289, Sample Num: 4624, Cur Loss: 0.34258389, Cur Avg Loss: 0.46937775, Log Avg loss: 0.50174934, Global Avg Loss: 1.29468548, Time: 0.0077 Steps: 83320, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000299, Sample Num: 4784, Cur Loss: 0.28493214, Cur Avg Loss: 0.46920525, Log Avg loss: 0.46422010, Global Avg Loss: 1.29458582, Time: 0.0077 Steps: 83330, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000309, Sample Num: 4944, Cur Loss: 0.74306417, Cur Avg Loss: 0.46712709, Log Avg loss: 0.40499022, Global Avg Loss: 1.29447907, Time: 0.0077 Steps: 83340, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000319, Sample Num: 5104, Cur Loss: 0.12660161, Cur Avg Loss: 0.46864384, Log Avg loss: 0.51551125, Global Avg Loss: 1.29438562, Time: 0.0076 Steps: 83350, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000329, Sample Num: 5264, Cur Loss: 0.22727820, Cur Avg Loss: 0.46722439, Log Avg loss: 0.42194403, Global Avg Loss: 1.29428096, Time: 0.0077 Steps: 83360, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000339, Sample Num: 5424, Cur Loss: 0.82817709, Cur Avg Loss: 0.46687834, Log Avg loss: 0.45549310, Global Avg Loss: 1.29418035, Time: 0.0076 Steps: 83370, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000349, Sample Num: 5584, Cur Loss: 0.43774116, Cur Avg Loss: 0.46820438, Log Avg loss: 0.51315719, Global Avg Loss: 1.29408668, Time: 0.0076 Steps: 83380, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000359, Sample Num: 5744, Cur Loss: 0.33137470, Cur Avg Loss: 0.46679253, Log Avg loss: 0.41751899, Global Avg Loss: 1.29398156, Time: 0.0077 Steps: 83390, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000369, Sample Num: 5904, Cur Loss: 0.57109219, Cur Avg Loss: 0.46799409, Log Avg loss: 0.51113023, Global Avg Loss: 1.29388769, Time: 0.0078 Steps: 83400, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000379, Sample Num: 6064, Cur Loss: 0.50428957, Cur Avg Loss: 0.47201998, Log Avg loss: 0.62057507, Global Avg Loss: 1.29380697, Time: 0.0077 Steps: 83410, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000389, Sample Num: 6224, Cur Loss: 0.44777083, Cur Avg Loss: 0.47491366, Log Avg loss: 0.58458420, Global Avg Loss: 1.29372195, Time: 0.0076 Steps: 83420, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000399, Sample Num: 6384, Cur Loss: 0.94342732, Cur Avg Loss: 0.47526789, Log Avg loss: 0.48904755, Global Avg Loss: 1.29362550, Time: 0.0077 Steps: 83430, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000409, Sample Num: 6544, Cur Loss: 0.13472880, Cur Avg Loss: 0.47329502, Log Avg loss: 0.39457767, Global Avg Loss: 1.29351775, Time: 0.0077 Steps: 83440, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000419, Sample Num: 6704, Cur Loss: 0.31150120, Cur Avg Loss: 0.47359483, Log Avg loss: 0.48585687, Global Avg Loss: 1.29342097, Time: 0.0078 Steps: 83450, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000429, Sample Num: 6864, Cur Loss: 0.49959853, Cur Avg Loss: 0.47302682, Log Avg loss: 0.44922703, Global Avg Loss: 1.29331982, Time: 0.0077 Steps: 83460, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000439, Sample Num: 7024, Cur Loss: 0.41777718, Cur Avg Loss: 0.47621238, Log Avg loss: 0.61287322, Global Avg Loss: 1.29323830, Time: 0.0076 Steps: 83470, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000449, Sample Num: 7184, Cur Loss: 1.35419548, Cur Avg Loss: 0.47449233, Log Avg loss: 0.39898224, Global Avg Loss: 1.29313118, Time: 0.0076 Steps: 83480, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000459, Sample Num: 7344, Cur Loss: 0.46020079, Cur Avg Loss: 0.47198763, Log Avg loss: 0.35952661, Global Avg Loss: 1.29301936, Time: 0.0079 Steps: 83490, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000469, Sample Num: 7504, Cur Loss: 0.33574909, Cur Avg Loss: 0.47403808, Log Avg loss: 0.56815352, Global Avg Loss: 1.29293255, Time: 0.0076 Steps: 83500, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000479, Sample Num: 7664, Cur Loss: 1.59612036, Cur Avg Loss: 0.47687588, Log Avg loss: 0.60996874, Global Avg Loss: 1.29285076, Time: 0.0068 Steps: 83510, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000489, Sample Num: 7824, Cur Loss: 0.28624743, Cur Avg Loss: 0.47708620, Log Avg loss: 0.48716045, Global Avg Loss: 1.29275430, Time: 0.0067 Steps: 83520, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000499, Sample Num: 7984, Cur Loss: 0.55815184, Cur Avg Loss: 0.47939424, Log Avg loss: 0.59225763, Global Avg Loss: 1.29267043, Time: 0.0064 Steps: 83530, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000509, Sample Num: 8144, Cur Loss: 0.30651712, Cur Avg Loss: 0.47618450, Log Avg loss: 0.31601819, Global Avg Loss: 1.29255353, Time: 0.0115 Steps: 83540, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000519, Sample Num: 8304, Cur Loss: 0.21475983, Cur Avg Loss: 0.47428528, Log Avg loss: 0.37761492, Global Avg Loss: 1.29244402, Time: 0.0065 Steps: 83550, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000529, Sample Num: 8464, Cur Loss: 0.60632670, Cur Avg Loss: 0.47541809, Log Avg loss: 0.53421097, Global Avg Loss: 1.29235328, Time: 0.0066 Steps: 83560, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000539, Sample Num: 8624, Cur Loss: 0.33918220, Cur Avg Loss: 0.47581055, Log Avg loss: 0.49657166, Global Avg Loss: 1.29225805, Time: 0.0064 Steps: 83570, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000549, Sample Num: 8784, Cur Loss: 0.21810271, Cur Avg Loss: 0.47681947, Log Avg loss: 0.53120049, Global Avg Loss: 1.29216700, Time: 0.0188 Steps: 83580, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000559, Sample Num: 8944, Cur Loss: 0.24114013, Cur Avg Loss: 0.47652483, Log Avg loss: 0.46034915, Global Avg Loss: 1.29206749, Time: 0.0222 Steps: 83590, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000569, Sample Num: 9104, Cur Loss: 0.52253604, Cur Avg Loss: 0.47632624, Log Avg loss: 0.46522514, Global Avg Loss: 1.29196858, Time: 0.0067 Steps: 83600, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000579, Sample Num: 9264, Cur Loss: 0.85020083, Cur Avg Loss: 0.47652060, Log Avg loss: 0.48757964, Global Avg Loss: 1.29187237, Time: 0.0068 Steps: 83610, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000589, Sample Num: 9424, Cur Loss: 0.32870448, Cur Avg Loss: 0.47539977, Log Avg loss: 0.41050371, Global Avg Loss: 1.29176697, Time: 0.0067 Steps: 83620, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000599, Sample Num: 9584, Cur Loss: 0.85675162, Cur Avg Loss: 0.47528802, Log Avg loss: 0.46870583, Global Avg Loss: 1.29166855, Time: 0.0218 Steps: 83630, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000609, Sample Num: 9744, Cur Loss: 0.69476497, Cur Avg Loss: 0.47366170, Log Avg loss: 0.37624512, Global Avg Loss: 1.29155911, Time: 0.0076 Steps: 83640, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000619, Sample Num: 9904, Cur Loss: 0.39359730, Cur Avg Loss: 0.47259684, Log Avg loss: 0.40774682, Global Avg Loss: 1.29145345, Time: 0.0186 Steps: 83650, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000629, Sample Num: 10064, Cur Loss: 0.45189759, Cur Avg Loss: 0.47169776, Log Avg loss: 0.41604483, Global Avg Loss: 1.29134881, Time: 0.0064 Steps: 83660, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000639, Sample Num: 10224, Cur Loss: 0.36558771, Cur Avg Loss: 0.47170678, Log Avg loss: 0.47227425, Global Avg Loss: 1.29125092, Time: 0.0182 Steps: 83670, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000649, Sample Num: 10384, Cur Loss: 0.48454532, Cur Avg Loss: 0.47356582, Log Avg loss: 0.59235828, Global Avg Loss: 1.29116740, Time: 0.0065 Steps: 83680, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000659, Sample Num: 10544, Cur Loss: 0.26730248, Cur Avg Loss: 0.47501358, Log Avg loss: 0.56897316, Global Avg Loss: 1.29108110, Time: 0.0064 Steps: 83690, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000669, Sample Num: 10704, Cur Loss: 0.74413371, Cur Avg Loss: 0.47783777, Log Avg loss: 0.66395179, Global Avg Loss: 1.29100618, Time: 0.0067 Steps: 83700, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000679, Sample Num: 10864, Cur Loss: 1.02521539, Cur Avg Loss: 0.47798734, Log Avg loss: 0.48799339, Global Avg Loss: 1.29091025, Time: 0.0070 Steps: 83710, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000689, Sample Num: 11024, Cur Loss: 0.25921321, Cur Avg Loss: 0.47811737, Log Avg loss: 0.48694689, Global Avg Loss: 1.29081422, Time: 0.0230 Steps: 83720, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000699, Sample Num: 11184, Cur Loss: 0.19609645, Cur Avg Loss: 0.47593624, Log Avg loss: 0.32565603, Global Avg Loss: 1.29069895, Time: 0.0067 Steps: 83730, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000709, Sample Num: 11344, Cur Loss: 0.62328553, Cur Avg Loss: 0.47569079, Log Avg loss: 0.45853380, Global Avg Loss: 1.29059958, Time: 0.0067 Steps: 83740, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000719, Sample Num: 11504, Cur Loss: 0.87715966, Cur Avg Loss: 0.47611473, Log Avg loss: 0.50617200, Global Avg Loss: 1.29050591, Time: 0.0066 Steps: 83750, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000729, Sample Num: 11664, Cur Loss: 0.33001363, Cur Avg Loss: 0.47431013, Log Avg loss: 0.34455982, Global Avg Loss: 1.29039298, Time: 0.0066 Steps: 83760, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000739, Sample Num: 11824, Cur Loss: 0.48123050, Cur Avg Loss: 0.47216864, Log Avg loss: 0.31605414, Global Avg Loss: 1.29027667, Time: 0.0064 Steps: 83770, Updated lr: 0.000022 Training, Epoch: 0040, Batch: 000749, Sample Num: 11984, Cur Loss: 0.26693535, Cur Avg Loss: 0.47295506, Log Avg loss: 0.53107156, Global Avg Loss: 1.29018605, Time: 0.0066 Steps: 83780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000759, Sample Num: 12144, Cur Loss: 0.34589571, Cur Avg Loss: 0.47283098, Log Avg loss: 0.46353681, Global Avg Loss: 1.29008739, Time: 0.0065 Steps: 83790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000769, Sample Num: 12304, Cur Loss: 0.41092843, Cur Avg Loss: 0.47244718, Log Avg loss: 0.44331729, Global Avg Loss: 1.28998634, Time: 0.0066 Steps: 83800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000779, Sample Num: 12464, Cur Loss: 0.46042389, Cur Avg Loss: 0.47277705, Log Avg loss: 0.49814401, Global Avg Loss: 1.28989186, Time: 0.0065 Steps: 83810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000789, Sample Num: 12624, Cur Loss: 0.45708746, Cur Avg Loss: 0.47462049, Log Avg loss: 0.61822413, Global Avg Loss: 1.28981173, Time: 0.0065 Steps: 83820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000799, Sample Num: 12784, Cur Loss: 0.68776858, Cur Avg Loss: 0.47487300, Log Avg loss: 0.49479639, Global Avg Loss: 1.28971689, Time: 0.0065 Steps: 83830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000809, Sample Num: 12944, Cur Loss: 0.17656143, Cur Avg Loss: 0.47639221, Log Avg loss: 0.59777702, Global Avg Loss: 1.28963436, Time: 0.0078 Steps: 83840, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000819, Sample Num: 13104, Cur Loss: 0.57340384, Cur Avg Loss: 0.47603705, Log Avg loss: 0.44730461, Global Avg Loss: 1.28953391, Time: 0.0076 Steps: 83850, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000829, Sample Num: 13264, Cur Loss: 0.22409689, Cur Avg Loss: 0.47749191, Log Avg loss: 0.59664512, Global Avg Loss: 1.28945128, Time: 0.0076 Steps: 83860, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000839, Sample Num: 13424, Cur Loss: 0.73433846, Cur Avg Loss: 0.47896716, Log Avg loss: 0.60126525, Global Avg Loss: 1.28936923, Time: 0.0076 Steps: 83870, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000849, Sample Num: 13584, Cur Loss: 0.31767729, Cur Avg Loss: 0.47941827, Log Avg loss: 0.51726602, Global Avg Loss: 1.28927718, Time: 0.0077 Steps: 83880, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000859, Sample Num: 13744, Cur Loss: 0.39064139, Cur Avg Loss: 0.48101512, Log Avg loss: 0.61658799, Global Avg Loss: 1.28919699, Time: 0.0075 Steps: 83890, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000869, Sample Num: 13904, Cur Loss: 0.28728011, Cur Avg Loss: 0.48025485, Log Avg loss: 0.41494741, Global Avg Loss: 1.28909279, Time: 0.0077 Steps: 83900, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000879, Sample Num: 14064, Cur Loss: 0.42073551, Cur Avg Loss: 0.48190524, Log Avg loss: 0.62532461, Global Avg Loss: 1.28901369, Time: 0.0075 Steps: 83910, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000889, Sample Num: 14224, Cur Loss: 0.46057695, Cur Avg Loss: 0.48121291, Log Avg loss: 0.42035697, Global Avg Loss: 1.28891018, Time: 0.0075 Steps: 83920, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000899, Sample Num: 14384, Cur Loss: 0.81448889, Cur Avg Loss: 0.48133330, Log Avg loss: 0.49203615, Global Avg Loss: 1.28881523, Time: 0.0079 Steps: 83930, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000909, Sample Num: 14544, Cur Loss: 0.56225681, Cur Avg Loss: 0.48119020, Log Avg loss: 0.46832497, Global Avg Loss: 1.28871748, Time: 0.0075 Steps: 83940, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000919, Sample Num: 14704, Cur Loss: 0.76491559, Cur Avg Loss: 0.48155585, Log Avg loss: 0.51479312, Global Avg Loss: 1.28862530, Time: 0.0077 Steps: 83950, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000929, Sample Num: 14864, Cur Loss: 0.83041584, Cur Avg Loss: 0.48123511, Log Avg loss: 0.45175981, Global Avg Loss: 1.28852562, Time: 0.0075 Steps: 83960, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000939, Sample Num: 15024, Cur Loss: 0.15363212, Cur Avg Loss: 0.48103789, Log Avg loss: 0.46271572, Global Avg Loss: 1.28842727, Time: 0.0076 Steps: 83970, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000949, Sample Num: 15184, Cur Loss: 0.47102416, Cur Avg Loss: 0.48130555, Log Avg loss: 0.50643888, Global Avg Loss: 1.28833416, Time: 0.0074 Steps: 83980, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000959, Sample Num: 15344, Cur Loss: 0.31516510, Cur Avg Loss: 0.48147186, Log Avg loss: 0.49725462, Global Avg Loss: 1.28823997, Time: 0.0074 Steps: 83990, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000969, Sample Num: 15504, Cur Loss: 0.38543311, Cur Avg Loss: 0.48106051, Log Avg loss: 0.44161202, Global Avg Loss: 1.28813918, Time: 0.0074 Steps: 84000, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000979, Sample Num: 15664, Cur Loss: 0.79993773, Cur Avg Loss: 0.48140837, Log Avg loss: 0.51511661, Global Avg Loss: 1.28804717, Time: 0.0075 Steps: 84010, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000989, Sample Num: 15824, Cur Loss: 0.67511773, Cur Avg Loss: 0.48078191, Log Avg loss: 0.41945112, Global Avg Loss: 1.28794379, Time: 0.0074 Steps: 84020, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 000999, Sample Num: 15984, Cur Loss: 0.15206994, Cur Avg Loss: 0.48001452, Log Avg loss: 0.40411933, Global Avg Loss: 1.28783861, Time: 0.0074 Steps: 84030, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001009, Sample Num: 16144, Cur Loss: 0.31411740, Cur Avg Loss: 0.47993275, Log Avg loss: 0.47176433, Global Avg Loss: 1.28774150, Time: 0.0074 Steps: 84040, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001019, Sample Num: 16304, Cur Loss: 1.19673443, Cur Avg Loss: 0.48085481, Log Avg loss: 0.57389040, Global Avg Loss: 1.28765657, Time: 0.0074 Steps: 84050, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001029, Sample Num: 16464, Cur Loss: 0.51304489, Cur Avg Loss: 0.48098397, Log Avg loss: 0.49414563, Global Avg Loss: 1.28756217, Time: 0.0074 Steps: 84060, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001039, Sample Num: 16624, Cur Loss: 0.95525235, Cur Avg Loss: 0.48094644, Log Avg loss: 0.47708457, Global Avg Loss: 1.28746577, Time: 0.0075 Steps: 84070, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001049, Sample Num: 16784, Cur Loss: 0.27511877, Cur Avg Loss: 0.48079192, Log Avg loss: 0.46473718, Global Avg Loss: 1.28736792, Time: 0.0075 Steps: 84080, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001059, Sample Num: 16944, Cur Loss: 0.98492622, Cur Avg Loss: 0.48148851, Log Avg loss: 0.55456071, Global Avg Loss: 1.28728077, Time: 0.0074 Steps: 84090, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001069, Sample Num: 17104, Cur Loss: 0.28599399, Cur Avg Loss: 0.48135891, Log Avg loss: 0.46763414, Global Avg Loss: 1.28718331, Time: 0.0075 Steps: 84100, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001079, Sample Num: 17264, Cur Loss: 0.21116331, Cur Avg Loss: 0.48249850, Log Avg loss: 0.60432093, Global Avg Loss: 1.28710212, Time: 0.0075 Steps: 84110, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001089, Sample Num: 17424, Cur Loss: 0.55358887, Cur Avg Loss: 0.48225398, Log Avg loss: 0.45587043, Global Avg Loss: 1.28700331, Time: 0.0074 Steps: 84120, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001099, Sample Num: 17584, Cur Loss: 0.13004148, Cur Avg Loss: 0.48220221, Log Avg loss: 0.47656380, Global Avg Loss: 1.28690698, Time: 0.0074 Steps: 84130, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001109, Sample Num: 17744, Cur Loss: 1.34442329, Cur Avg Loss: 0.48267127, Log Avg loss: 0.53422111, Global Avg Loss: 1.28681752, Time: 0.0074 Steps: 84140, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001119, Sample Num: 17904, Cur Loss: 0.47702384, Cur Avg Loss: 0.48302854, Log Avg loss: 0.52265022, Global Avg Loss: 1.28672671, Time: 0.0074 Steps: 84150, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001129, Sample Num: 18064, Cur Loss: 0.42647794, Cur Avg Loss: 0.48380009, Log Avg loss: 0.57013651, Global Avg Loss: 1.28664156, Time: 0.0074 Steps: 84160, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001139, Sample Num: 18224, Cur Loss: 0.51073503, Cur Avg Loss: 0.48312472, Log Avg loss: 0.40687549, Global Avg Loss: 1.28653704, Time: 0.0075 Steps: 84170, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001149, Sample Num: 18384, Cur Loss: 0.31664777, Cur Avg Loss: 0.48463001, Log Avg loss: 0.65608210, Global Avg Loss: 1.28646215, Time: 0.0073 Steps: 84180, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001159, Sample Num: 18544, Cur Loss: 0.88014221, Cur Avg Loss: 0.48514562, Log Avg loss: 0.54438897, Global Avg Loss: 1.28637400, Time: 0.0074 Steps: 84190, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001169, Sample Num: 18704, Cur Loss: 0.52957213, Cur Avg Loss: 0.48460650, Log Avg loss: 0.42212290, Global Avg Loss: 1.28627136, Time: 0.0075 Steps: 84200, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001179, Sample Num: 18864, Cur Loss: 0.72654015, Cur Avg Loss: 0.48407320, Log Avg loss: 0.42173097, Global Avg Loss: 1.28616870, Time: 0.0075 Steps: 84210, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001189, Sample Num: 19024, Cur Loss: 0.31777167, Cur Avg Loss: 0.48547544, Log Avg loss: 0.65079925, Global Avg Loss: 1.28609326, Time: 0.0075 Steps: 84220, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001199, Sample Num: 19184, Cur Loss: 0.26505989, Cur Avg Loss: 0.48572567, Log Avg loss: 0.51547730, Global Avg Loss: 1.28600177, Time: 0.0074 Steps: 84230, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001209, Sample Num: 19344, Cur Loss: 0.07808879, Cur Avg Loss: 0.48386768, Log Avg loss: 0.26109579, Global Avg Loss: 1.28588010, Time: 0.0074 Steps: 84240, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001219, Sample Num: 19504, Cur Loss: 0.98792952, Cur Avg Loss: 0.48409266, Log Avg loss: 0.51129259, Global Avg Loss: 1.28578816, Time: 0.0075 Steps: 84250, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001229, Sample Num: 19664, Cur Loss: 0.39922071, Cur Avg Loss: 0.48418911, Log Avg loss: 0.49594611, Global Avg Loss: 1.28569442, Time: 0.0074 Steps: 84260, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001239, Sample Num: 19824, Cur Loss: 0.69489777, Cur Avg Loss: 0.48419946, Log Avg loss: 0.48547194, Global Avg Loss: 1.28559946, Time: 0.0075 Steps: 84270, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001249, Sample Num: 19984, Cur Loss: 0.12120952, Cur Avg Loss: 0.48374919, Log Avg loss: 0.42796068, Global Avg Loss: 1.28549770, Time: 0.0075 Steps: 84280, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001259, Sample Num: 20144, Cur Loss: 0.25792745, Cur Avg Loss: 0.48244860, Log Avg loss: 0.32000449, Global Avg Loss: 1.28538316, Time: 0.0075 Steps: 84290, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001269, Sample Num: 20304, Cur Loss: 0.25717938, Cur Avg Loss: 0.48275192, Log Avg loss: 0.52093978, Global Avg Loss: 1.28529248, Time: 0.0075 Steps: 84300, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001279, Sample Num: 20464, Cur Loss: 0.69163215, Cur Avg Loss: 0.48301985, Log Avg loss: 0.51701973, Global Avg Loss: 1.28520135, Time: 0.0075 Steps: 84310, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001289, Sample Num: 20624, Cur Loss: 0.11635402, Cur Avg Loss: 0.48237700, Log Avg loss: 0.40015747, Global Avg Loss: 1.28509639, Time: 0.0076 Steps: 84320, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001299, Sample Num: 20784, Cur Loss: 0.25599670, Cur Avg Loss: 0.48202610, Log Avg loss: 0.43679471, Global Avg Loss: 1.28499580, Time: 0.0075 Steps: 84330, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001309, Sample Num: 20944, Cur Loss: 0.67310774, Cur Avg Loss: 0.48182868, Log Avg loss: 0.45618333, Global Avg Loss: 1.28489753, Time: 0.0075 Steps: 84340, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001319, Sample Num: 21104, Cur Loss: 0.24974394, Cur Avg Loss: 0.48116677, Log Avg loss: 0.39452325, Global Avg Loss: 1.28479197, Time: 0.0075 Steps: 84350, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001329, Sample Num: 21264, Cur Loss: 1.10655880, Cur Avg Loss: 0.48209403, Log Avg loss: 0.60439893, Global Avg Loss: 1.28471132, Time: 0.0075 Steps: 84360, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001339, Sample Num: 21424, Cur Loss: 0.33072644, Cur Avg Loss: 0.48096715, Log Avg loss: 0.33120513, Global Avg Loss: 1.28459830, Time: 0.0075 Steps: 84370, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001349, Sample Num: 21584, Cur Loss: 0.82493246, Cur Avg Loss: 0.48059812, Log Avg loss: 0.43118584, Global Avg Loss: 1.28449716, Time: 0.0075 Steps: 84380, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001359, Sample Num: 21744, Cur Loss: 0.37731248, Cur Avg Loss: 0.48108921, Log Avg loss: 0.54733593, Global Avg Loss: 1.28440981, Time: 0.0075 Steps: 84390, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001369, Sample Num: 21904, Cur Loss: 0.47316319, Cur Avg Loss: 0.48110346, Log Avg loss: 0.48304011, Global Avg Loss: 1.28431486, Time: 0.0075 Steps: 84400, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001379, Sample Num: 22064, Cur Loss: 0.47133854, Cur Avg Loss: 0.48226313, Log Avg loss: 0.64102196, Global Avg Loss: 1.28423865, Time: 0.0075 Steps: 84410, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001389, Sample Num: 22224, Cur Loss: 0.64276230, Cur Avg Loss: 0.48219822, Log Avg loss: 0.47324760, Global Avg Loss: 1.28414259, Time: 0.0075 Steps: 84420, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001399, Sample Num: 22384, Cur Loss: 1.19311285, Cur Avg Loss: 0.48366201, Log Avg loss: 0.68698276, Global Avg Loss: 1.28407186, Time: 0.0075 Steps: 84430, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001409, Sample Num: 22544, Cur Loss: 0.43040293, Cur Avg Loss: 0.48390170, Log Avg loss: 0.51743369, Global Avg Loss: 1.28398107, Time: 0.0076 Steps: 84440, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001419, Sample Num: 22704, Cur Loss: 0.89636582, Cur Avg Loss: 0.48507218, Log Avg loss: 0.64999343, Global Avg Loss: 1.28390599, Time: 0.0076 Steps: 84450, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001429, Sample Num: 22864, Cur Loss: 0.61450845, Cur Avg Loss: 0.48548036, Log Avg loss: 0.54340061, Global Avg Loss: 1.28381832, Time: 0.0075 Steps: 84460, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001439, Sample Num: 23024, Cur Loss: 1.05185962, Cur Avg Loss: 0.48574723, Log Avg loss: 0.52388258, Global Avg Loss: 1.28372835, Time: 0.0075 Steps: 84470, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001449, Sample Num: 23184, Cur Loss: 0.30002540, Cur Avg Loss: 0.48596112, Log Avg loss: 0.51674100, Global Avg Loss: 1.28363756, Time: 0.0073 Steps: 84480, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001459, Sample Num: 23344, Cur Loss: 0.32457983, Cur Avg Loss: 0.48523586, Log Avg loss: 0.38014518, Global Avg Loss: 1.28353063, Time: 0.0067 Steps: 84490, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001469, Sample Num: 23504, Cur Loss: 0.64495605, Cur Avg Loss: 0.48441940, Log Avg loss: 0.36529806, Global Avg Loss: 1.28342196, Time: 0.0075 Steps: 84500, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001479, Sample Num: 23664, Cur Loss: 0.36079562, Cur Avg Loss: 0.48407195, Log Avg loss: 0.43303197, Global Avg Loss: 1.28332134, Time: 0.0180 Steps: 84510, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001489, Sample Num: 23824, Cur Loss: 0.60298693, Cur Avg Loss: 0.48327454, Log Avg loss: 0.36533683, Global Avg Loss: 1.28321272, Time: 0.0168 Steps: 84520, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001499, Sample Num: 23984, Cur Loss: 1.17265356, Cur Avg Loss: 0.48420226, Log Avg loss: 0.62234059, Global Avg Loss: 1.28313454, Time: 0.0071 Steps: 84530, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001509, Sample Num: 24144, Cur Loss: 0.78421056, Cur Avg Loss: 0.48443585, Log Avg loss: 0.51945052, Global Avg Loss: 1.28304421, Time: 0.0228 Steps: 84540, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001519, Sample Num: 24304, Cur Loss: 0.83399981, Cur Avg Loss: 0.48463695, Log Avg loss: 0.51498247, Global Avg Loss: 1.28295337, Time: 0.0065 Steps: 84550, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001529, Sample Num: 24464, Cur Loss: 0.64823604, Cur Avg Loss: 0.48459591, Log Avg loss: 0.47836179, Global Avg Loss: 1.28285822, Time: 0.0114 Steps: 84560, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001539, Sample Num: 24624, Cur Loss: 0.45127499, Cur Avg Loss: 0.48439758, Log Avg loss: 0.45407326, Global Avg Loss: 1.28276022, Time: 0.0134 Steps: 84570, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001549, Sample Num: 24784, Cur Loss: 0.27903765, Cur Avg Loss: 0.48336862, Log Avg loss: 0.32501243, Global Avg Loss: 1.28264698, Time: 0.0073 Steps: 84580, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001559, Sample Num: 24944, Cur Loss: 0.14964075, Cur Avg Loss: 0.48229317, Log Avg loss: 0.31570482, Global Avg Loss: 1.28253267, Time: 0.0064 Steps: 84590, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001569, Sample Num: 25104, Cur Loss: 0.52258193, Cur Avg Loss: 0.48230406, Log Avg loss: 0.48400251, Global Avg Loss: 1.28243828, Time: 0.0065 Steps: 84600, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001579, Sample Num: 25264, Cur Loss: 0.38668773, Cur Avg Loss: 0.48238264, Log Avg loss: 0.49471220, Global Avg Loss: 1.28234518, Time: 0.0065 Steps: 84610, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001589, Sample Num: 25424, Cur Loss: 0.26970947, Cur Avg Loss: 0.48284187, Log Avg loss: 0.55535418, Global Avg Loss: 1.28225927, Time: 0.0089 Steps: 84620, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001599, Sample Num: 25584, Cur Loss: 0.12380673, Cur Avg Loss: 0.48182149, Log Avg loss: 0.31968287, Global Avg Loss: 1.28214553, Time: 0.0224 Steps: 84630, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001609, Sample Num: 25744, Cur Loss: 0.76904714, Cur Avg Loss: 0.48230564, Log Avg loss: 0.55972156, Global Avg Loss: 1.28206018, Time: 0.0073 Steps: 84640, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001619, Sample Num: 25904, Cur Loss: 0.63461578, Cur Avg Loss: 0.48282286, Log Avg loss: 0.56604271, Global Avg Loss: 1.28197559, Time: 0.0066 Steps: 84650, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001629, Sample Num: 26064, Cur Loss: 0.63176757, Cur Avg Loss: 0.48278088, Log Avg loss: 0.47598519, Global Avg Loss: 1.28188039, Time: 0.0065 Steps: 84660, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001639, Sample Num: 26224, Cur Loss: 1.70608687, Cur Avg Loss: 0.48316218, Log Avg loss: 0.54527476, Global Avg Loss: 1.28179339, Time: 0.0200 Steps: 84670, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001649, Sample Num: 26384, Cur Loss: 0.42791903, Cur Avg Loss: 0.48277668, Log Avg loss: 0.41959391, Global Avg Loss: 1.28169157, Time: 0.0096 Steps: 84680, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001659, Sample Num: 26544, Cur Loss: 0.15846640, Cur Avg Loss: 0.48319425, Log Avg loss: 0.55205081, Global Avg Loss: 1.28160542, Time: 0.0172 Steps: 84690, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001669, Sample Num: 26704, Cur Loss: 0.79377156, Cur Avg Loss: 0.48326406, Log Avg loss: 0.49484658, Global Avg Loss: 1.28151253, Time: 0.0066 Steps: 84700, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001679, Sample Num: 26864, Cur Loss: 0.59394372, Cur Avg Loss: 0.48363062, Log Avg loss: 0.54480857, Global Avg Loss: 1.28142556, Time: 0.0065 Steps: 84710, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001689, Sample Num: 27024, Cur Loss: 0.66700709, Cur Avg Loss: 0.48419342, Log Avg loss: 0.57868771, Global Avg Loss: 1.28134262, Time: 0.0065 Steps: 84720, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001699, Sample Num: 27184, Cur Loss: 0.55671328, Cur Avg Loss: 0.48485906, Log Avg loss: 0.59728645, Global Avg Loss: 1.28126188, Time: 0.0065 Steps: 84730, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001709, Sample Num: 27344, Cur Loss: 1.25873542, Cur Avg Loss: 0.48462211, Log Avg loss: 0.44436362, Global Avg Loss: 1.28116312, Time: 0.0065 Steps: 84740, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001719, Sample Num: 27504, Cur Loss: 0.45572817, Cur Avg Loss: 0.48485342, Log Avg loss: 0.52438550, Global Avg Loss: 1.28107383, Time: 0.0068 Steps: 84750, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001729, Sample Num: 27664, Cur Loss: 0.56106949, Cur Avg Loss: 0.48554365, Log Avg loss: 0.60419275, Global Avg Loss: 1.28099397, Time: 0.0074 Steps: 84760, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001739, Sample Num: 27824, Cur Loss: 0.83016294, Cur Avg Loss: 0.48567179, Log Avg loss: 0.50782720, Global Avg Loss: 1.28090276, Time: 0.0074 Steps: 84770, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001749, Sample Num: 27984, Cur Loss: 0.63865733, Cur Avg Loss: 0.48593030, Log Avg loss: 0.53088541, Global Avg Loss: 1.28081429, Time: 0.0074 Steps: 84780, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001759, Sample Num: 28144, Cur Loss: 0.28878093, Cur Avg Loss: 0.48603914, Log Avg loss: 0.50507553, Global Avg Loss: 1.28072280, Time: 0.0074 Steps: 84790, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001769, Sample Num: 28304, Cur Loss: 1.24266887, Cur Avg Loss: 0.48660878, Log Avg loss: 0.58680782, Global Avg Loss: 1.28064097, Time: 0.0075 Steps: 84800, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001779, Sample Num: 28464, Cur Loss: 0.62609094, Cur Avg Loss: 0.48769630, Log Avg loss: 0.68007863, Global Avg Loss: 1.28057016, Time: 0.0074 Steps: 84810, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001789, Sample Num: 28624, Cur Loss: 0.62588853, Cur Avg Loss: 0.48708972, Log Avg loss: 0.37918032, Global Avg Loss: 1.28046389, Time: 0.0074 Steps: 84820, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001799, Sample Num: 28784, Cur Loss: 1.18043864, Cur Avg Loss: 0.48704658, Log Avg loss: 0.47932840, Global Avg Loss: 1.28036945, Time: 0.0075 Steps: 84830, Updated lr: 0.000021 Training, Epoch: 0040, Batch: 001809, Sample Num: 28944, Cur Loss: 0.52596027, Cur Avg Loss: 0.48725964, Log Avg loss: 0.52558851, Global Avg Loss: 1.28028049, Time: 0.0074 Steps: 84840, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001819, Sample Num: 29104, Cur Loss: 0.29680711, Cur Avg Loss: 0.48721401, Log Avg loss: 0.47896053, Global Avg Loss: 1.28018605, Time: 0.0074 Steps: 84850, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001829, Sample Num: 29264, Cur Loss: 0.60421628, Cur Avg Loss: 0.48759037, Log Avg loss: 0.55604923, Global Avg Loss: 1.28010071, Time: 0.0075 Steps: 84860, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001839, Sample Num: 29424, Cur Loss: 0.22931510, Cur Avg Loss: 0.48721623, Log Avg loss: 0.41878594, Global Avg Loss: 1.27999923, Time: 0.0074 Steps: 84870, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001849, Sample Num: 29584, Cur Loss: 0.60673356, Cur Avg Loss: 0.48689726, Log Avg loss: 0.42823978, Global Avg Loss: 1.27989888, Time: 0.0074 Steps: 84880, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001859, Sample Num: 29744, Cur Loss: 0.32668406, Cur Avg Loss: 0.48648399, Log Avg loss: 0.41007067, Global Avg Loss: 1.27979641, Time: 0.0074 Steps: 84890, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001869, Sample Num: 29904, Cur Loss: 0.23242202, Cur Avg Loss: 0.48675498, Log Avg loss: 0.53713186, Global Avg Loss: 1.27970894, Time: 0.0074 Steps: 84900, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001879, Sample Num: 30064, Cur Loss: 0.61671245, Cur Avg Loss: 0.48740198, Log Avg loss: 0.60832638, Global Avg Loss: 1.27962987, Time: 0.0075 Steps: 84910, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001889, Sample Num: 30224, Cur Loss: 0.16771644, Cur Avg Loss: 0.48733353, Log Avg loss: 0.47447062, Global Avg Loss: 1.27953505, Time: 0.0074 Steps: 84920, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001899, Sample Num: 30384, Cur Loss: 0.43892300, Cur Avg Loss: 0.48707577, Log Avg loss: 0.43838465, Global Avg Loss: 1.27943601, Time: 0.0074 Steps: 84930, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001909, Sample Num: 30544, Cur Loss: 0.60271633, Cur Avg Loss: 0.48734135, Log Avg loss: 0.53777650, Global Avg Loss: 1.27934870, Time: 0.0074 Steps: 84940, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001919, Sample Num: 30704, Cur Loss: 0.17039332, Cur Avg Loss: 0.48667461, Log Avg loss: 0.35939289, Global Avg Loss: 1.27924040, Time: 0.0074 Steps: 84950, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001929, Sample Num: 30864, Cur Loss: 0.75191224, Cur Avg Loss: 0.48702738, Log Avg loss: 0.55472360, Global Avg Loss: 1.27915513, Time: 0.0074 Steps: 84960, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001939, Sample Num: 31024, Cur Loss: 0.58756316, Cur Avg Loss: 0.48660804, Log Avg loss: 0.40571893, Global Avg Loss: 1.27905233, Time: 0.0075 Steps: 84970, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001949, Sample Num: 31184, Cur Loss: 0.91492993, Cur Avg Loss: 0.48664056, Log Avg loss: 0.49294463, Global Avg Loss: 1.27895983, Time: 0.0074 Steps: 84980, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001959, Sample Num: 31344, Cur Loss: 0.61172879, Cur Avg Loss: 0.48731748, Log Avg loss: 0.61924902, Global Avg Loss: 1.27888221, Time: 0.0075 Steps: 84990, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001969, Sample Num: 31504, Cur Loss: 0.22865732, Cur Avg Loss: 0.48688457, Log Avg loss: 0.40207852, Global Avg Loss: 1.27877905, Time: 0.0075 Steps: 85000, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001979, Sample Num: 31664, Cur Loss: 0.25825340, Cur Avg Loss: 0.48675289, Log Avg loss: 0.46082432, Global Avg Loss: 1.27868283, Time: 0.0074 Steps: 85010, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001989, Sample Num: 31824, Cur Loss: 0.16603307, Cur Avg Loss: 0.48612737, Log Avg loss: 0.36233715, Global Avg Loss: 1.27857505, Time: 0.0074 Steps: 85020, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 001999, Sample Num: 31984, Cur Loss: 1.16294432, Cur Avg Loss: 0.48597255, Log Avg loss: 0.45517875, Global Avg Loss: 1.27847822, Time: 0.0075 Steps: 85030, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002009, Sample Num: 32144, Cur Loss: 0.33531952, Cur Avg Loss: 0.48545632, Log Avg loss: 0.38226194, Global Avg Loss: 1.27837283, Time: 0.0076 Steps: 85040, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002019, Sample Num: 32304, Cur Loss: 0.49035951, Cur Avg Loss: 0.48554893, Log Avg loss: 0.50415387, Global Avg Loss: 1.27828180, Time: 0.0077 Steps: 85050, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002029, Sample Num: 32464, Cur Loss: 0.61503297, Cur Avg Loss: 0.48572321, Log Avg loss: 0.52091177, Global Avg Loss: 1.27819276, Time: 0.0076 Steps: 85060, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002039, Sample Num: 32624, Cur Loss: 0.15755874, Cur Avg Loss: 0.48575656, Log Avg loss: 0.49252301, Global Avg Loss: 1.27810040, Time: 0.0075 Steps: 85070, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002049, Sample Num: 32784, Cur Loss: 0.35914439, Cur Avg Loss: 0.48649599, Log Avg loss: 0.63726451, Global Avg Loss: 1.27802508, Time: 0.0151 Steps: 85080, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002059, Sample Num: 32944, Cur Loss: 0.32356155, Cur Avg Loss: 0.48673459, Log Avg loss: 0.53562422, Global Avg Loss: 1.27793783, Time: 0.0085 Steps: 85090, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002069, Sample Num: 33104, Cur Loss: 0.86050487, Cur Avg Loss: 0.48641869, Log Avg loss: 0.42137569, Global Avg Loss: 1.27783718, Time: 0.0067 Steps: 85100, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002079, Sample Num: 33264, Cur Loss: 0.30074894, Cur Avg Loss: 0.48634022, Log Avg loss: 0.47010535, Global Avg Loss: 1.27774228, Time: 0.0113 Steps: 85110, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002089, Sample Num: 33424, Cur Loss: 0.28349811, Cur Avg Loss: 0.48589600, Log Avg loss: 0.39354186, Global Avg Loss: 1.27763840, Time: 0.0076 Steps: 85120, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002099, Sample Num: 33584, Cur Loss: 0.34203562, Cur Avg Loss: 0.48549648, Log Avg loss: 0.40203768, Global Avg Loss: 1.27753554, Time: 0.0155 Steps: 85130, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002109, Sample Num: 33744, Cur Loss: 0.45371535, Cur Avg Loss: 0.48541094, Log Avg loss: 0.46745595, Global Avg Loss: 1.27744040, Time: 0.0199 Steps: 85140, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002119, Sample Num: 33904, Cur Loss: 0.34779736, Cur Avg Loss: 0.48499224, Log Avg loss: 0.39668715, Global Avg Loss: 1.27733696, Time: 0.0066 Steps: 85150, Updated lr: 0.000020 Training, Epoch: 0040, Batch: 002129, Sample Num: 34055, Cur Loss: 0.74651760, Cur Avg Loss: 0.48535037, Log Avg loss: 0.56123768, Global Avg Loss: 1.27725287, Time: 0.0092 Steps: 85160, Updated lr: 0.000020 ***** Running evaluation checkpoint-85160 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-85160 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.836050, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.599887, "eval_total_loss": 421.720242, "eval_mae": 0.604899, "eval_mse": 0.600002, "eval_r2": 0.618599, "eval_sp_statistic": 0.755406, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.797018, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.858371, "test_total_loss": 430.90206, "test_mae": 0.676599, "test_mse": 0.858573, "test_r2": 0.44587, "test_sp_statistic": 0.596902, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.668449, "test_ps_pvalue": 0.0, "lr": 2.0189663347558086e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2772528726360841, "train_cur_epoch_loss": 1033.310929492116, "train_cur_epoch_avg_loss": 0.4853503661306322, "train_cur_epoch_time": 17.836050033569336, "train_cur_epoch_avg_time": 0.00837766558645812, "epoch": 40, "step": 85160} ################################################## Training, Epoch: 0041, Batch: 000010, Sample Num: 160, Cur Loss: 0.70930970, Cur Avg Loss: 0.59517886, Log Avg loss: 0.59517886, Global Avg Loss: 1.27717279, Time: 0.0075 Steps: 85170, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000020, Sample Num: 320, Cur Loss: 0.53675693, Cur Avg Loss: 0.46921277, Log Avg loss: 0.34324667, Global Avg Loss: 1.27706315, Time: 0.0076 Steps: 85180, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000030, Sample Num: 480, Cur Loss: 0.43191561, Cur Avg Loss: 0.43211395, Log Avg loss: 0.35791633, Global Avg Loss: 1.27695525, Time: 0.0075 Steps: 85190, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000040, Sample Num: 640, Cur Loss: 0.38533765, Cur Avg Loss: 0.43540229, Log Avg loss: 0.44526729, Global Avg Loss: 1.27685764, Time: 0.0076 Steps: 85200, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000050, Sample Num: 800, Cur Loss: 0.37740237, Cur Avg Loss: 0.42584434, Log Avg loss: 0.38761253, Global Avg Loss: 1.27675328, Time: 0.0076 Steps: 85210, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000060, Sample Num: 960, Cur Loss: 0.44541779, Cur Avg Loss: 0.43292547, Log Avg loss: 0.46833114, Global Avg Loss: 1.27665842, Time: 0.0077 Steps: 85220, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000070, Sample Num: 1120, Cur Loss: 0.95572174, Cur Avg Loss: 0.45323230, Log Avg loss: 0.57507326, Global Avg Loss: 1.27657610, Time: 0.0081 Steps: 85230, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000080, Sample Num: 1280, Cur Loss: 0.25911924, Cur Avg Loss: 0.45886844, Log Avg loss: 0.49832145, Global Avg Loss: 1.27648480, Time: 0.0124 Steps: 85240, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000090, Sample Num: 1440, Cur Loss: 0.60701829, Cur Avg Loss: 0.46610599, Log Avg loss: 0.52400635, Global Avg Loss: 1.27639653, Time: 0.0067 Steps: 85250, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000100, Sample Num: 1600, Cur Loss: 0.56115395, Cur Avg Loss: 0.45760822, Log Avg loss: 0.38112837, Global Avg Loss: 1.27629153, Time: 0.0065 Steps: 85260, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000110, Sample Num: 1760, Cur Loss: 0.22143993, Cur Avg Loss: 0.44716248, Log Avg loss: 0.34270499, Global Avg Loss: 1.27618204, Time: 0.0181 Steps: 85270, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000120, Sample Num: 1920, Cur Loss: 0.41064683, Cur Avg Loss: 0.46332012, Log Avg loss: 0.64105423, Global Avg Loss: 1.27610756, Time: 0.0096 Steps: 85280, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000130, Sample Num: 2080, Cur Loss: 0.47564268, Cur Avg Loss: 0.46290351, Log Avg loss: 0.45790423, Global Avg Loss: 1.27601163, Time: 0.0069 Steps: 85290, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000140, Sample Num: 2240, Cur Loss: 0.85605848, Cur Avg Loss: 0.46687862, Log Avg loss: 0.51855501, Global Avg Loss: 1.27592283, Time: 0.0153 Steps: 85300, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000150, Sample Num: 2400, Cur Loss: 0.26767668, Cur Avg Loss: 0.46617736, Log Avg loss: 0.45635977, Global Avg Loss: 1.27582676, Time: 0.0167 Steps: 85310, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000160, Sample Num: 2560, Cur Loss: 0.41080204, Cur Avg Loss: 0.46899357, Log Avg loss: 0.51123657, Global Avg Loss: 1.27573715, Time: 0.0076 Steps: 85320, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000170, Sample Num: 2720, Cur Loss: 0.96084297, Cur Avg Loss: 0.47408734, Log Avg loss: 0.55558777, Global Avg Loss: 1.27565275, Time: 0.0070 Steps: 85330, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000180, Sample Num: 2880, Cur Loss: 0.84156370, Cur Avg Loss: 0.46898940, Log Avg loss: 0.38232435, Global Avg Loss: 1.27554808, Time: 0.0064 Steps: 85340, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000190, Sample Num: 3040, Cur Loss: 0.45573044, Cur Avg Loss: 0.46802770, Log Avg loss: 0.45071722, Global Avg Loss: 1.27545143, Time: 0.0077 Steps: 85350, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000200, Sample Num: 3200, Cur Loss: 0.30439287, Cur Avg Loss: 0.46529116, Log Avg loss: 0.41329684, Global Avg Loss: 1.27535043, Time: 0.0066 Steps: 85360, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000210, Sample Num: 3360, Cur Loss: 0.19149703, Cur Avg Loss: 0.46961581, Log Avg loss: 0.55610884, Global Avg Loss: 1.27526618, Time: 0.0065 Steps: 85370, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000220, Sample Num: 3520, Cur Loss: 0.46800268, Cur Avg Loss: 0.46361241, Log Avg loss: 0.33754103, Global Avg Loss: 1.27515635, Time: 0.0178 Steps: 85380, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000230, Sample Num: 3680, Cur Loss: 0.91662073, Cur Avg Loss: 0.46566370, Log Avg loss: 0.51079205, Global Avg Loss: 1.27506684, Time: 0.0205 Steps: 85390, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000240, Sample Num: 3840, Cur Loss: 0.74247754, Cur Avg Loss: 0.46624343, Log Avg loss: 0.47957718, Global Avg Loss: 1.27497369, Time: 0.0069 Steps: 85400, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000250, Sample Num: 4000, Cur Loss: 0.09692793, Cur Avg Loss: 0.46663812, Log Avg loss: 0.47611061, Global Avg Loss: 1.27488016, Time: 0.0081 Steps: 85410, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000260, Sample Num: 4160, Cur Loss: 0.53656787, Cur Avg Loss: 0.46317782, Log Avg loss: 0.37667028, Global Avg Loss: 1.27477500, Time: 0.0067 Steps: 85420, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000270, Sample Num: 4320, Cur Loss: 0.66755593, Cur Avg Loss: 0.46247388, Log Avg loss: 0.44417163, Global Avg Loss: 1.27467778, Time: 0.0066 Steps: 85430, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000280, Sample Num: 4480, Cur Loss: 0.47659516, Cur Avg Loss: 0.46031093, Log Avg loss: 0.40191120, Global Avg Loss: 1.27457563, Time: 0.0066 Steps: 85440, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000290, Sample Num: 4640, Cur Loss: 0.75740230, Cur Avg Loss: 0.46809317, Log Avg loss: 0.68599588, Global Avg Loss: 1.27450675, Time: 0.0067 Steps: 85450, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000300, Sample Num: 4800, Cur Loss: 0.30655152, Cur Avg Loss: 0.46645363, Log Avg loss: 0.41890694, Global Avg Loss: 1.27440663, Time: 0.0067 Steps: 85460, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000310, Sample Num: 4960, Cur Loss: 0.28002173, Cur Avg Loss: 0.46982911, Log Avg loss: 0.57109347, Global Avg Loss: 1.27432434, Time: 0.0065 Steps: 85470, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000320, Sample Num: 5120, Cur Loss: 0.20681325, Cur Avg Loss: 0.47329119, Log Avg loss: 0.58061568, Global Avg Loss: 1.27424319, Time: 0.0075 Steps: 85480, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000330, Sample Num: 5280, Cur Loss: 0.63780856, Cur Avg Loss: 0.46981063, Log Avg loss: 0.35843275, Global Avg Loss: 1.27413606, Time: 0.0074 Steps: 85490, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000340, Sample Num: 5440, Cur Loss: 0.27761608, Cur Avg Loss: 0.46950649, Log Avg loss: 0.45946982, Global Avg Loss: 1.27404078, Time: 0.0075 Steps: 85500, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000350, Sample Num: 5600, Cur Loss: 0.17746256, Cur Avg Loss: 0.46992891, Log Avg loss: 0.48429122, Global Avg Loss: 1.27394842, Time: 0.0075 Steps: 85510, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000360, Sample Num: 5760, Cur Loss: 0.44611955, Cur Avg Loss: 0.47052614, Log Avg loss: 0.49142937, Global Avg Loss: 1.27385692, Time: 0.0076 Steps: 85520, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000370, Sample Num: 5920, Cur Loss: 0.50447458, Cur Avg Loss: 0.47255352, Log Avg loss: 0.54553901, Global Avg Loss: 1.27377177, Time: 0.0074 Steps: 85530, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000380, Sample Num: 6080, Cur Loss: 0.60232854, Cur Avg Loss: 0.47195215, Log Avg loss: 0.44970137, Global Avg Loss: 1.27367543, Time: 0.0074 Steps: 85540, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000390, Sample Num: 6240, Cur Loss: 0.56839681, Cur Avg Loss: 0.47366741, Log Avg loss: 0.53884760, Global Avg Loss: 1.27358954, Time: 0.0076 Steps: 85550, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000400, Sample Num: 6400, Cur Loss: 0.12838851, Cur Avg Loss: 0.47267142, Log Avg loss: 0.43382784, Global Avg Loss: 1.27349139, Time: 0.0075 Steps: 85560, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000410, Sample Num: 6560, Cur Loss: 0.41678047, Cur Avg Loss: 0.47167724, Log Avg loss: 0.43190971, Global Avg Loss: 1.27339304, Time: 0.0075 Steps: 85570, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000420, Sample Num: 6720, Cur Loss: 0.28372964, Cur Avg Loss: 0.46945851, Log Avg loss: 0.37849082, Global Avg Loss: 1.27328847, Time: 0.0076 Steps: 85580, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000430, Sample Num: 6880, Cur Loss: 0.22873977, Cur Avg Loss: 0.46582976, Log Avg loss: 0.31342227, Global Avg Loss: 1.27317632, Time: 0.0077 Steps: 85590, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000440, Sample Num: 7040, Cur Loss: 0.16048250, Cur Avg Loss: 0.46543830, Log Avg loss: 0.44860547, Global Avg Loss: 1.27307999, Time: 0.0076 Steps: 85600, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000450, Sample Num: 7200, Cur Loss: 0.53060597, Cur Avg Loss: 0.46627530, Log Avg loss: 0.50310304, Global Avg Loss: 1.27299005, Time: 0.0075 Steps: 85610, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000460, Sample Num: 7360, Cur Loss: 0.29386583, Cur Avg Loss: 0.46736986, Log Avg loss: 0.51662508, Global Avg Loss: 1.27290171, Time: 0.0076 Steps: 85620, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000470, Sample Num: 7520, Cur Loss: 0.84112966, Cur Avg Loss: 0.47025160, Log Avg loss: 0.60281183, Global Avg Loss: 1.27282346, Time: 0.0075 Steps: 85630, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000480, Sample Num: 7680, Cur Loss: 0.12018744, Cur Avg Loss: 0.46831367, Log Avg loss: 0.37723095, Global Avg Loss: 1.27271888, Time: 0.0074 Steps: 85640, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000490, Sample Num: 7840, Cur Loss: 0.78069210, Cur Avg Loss: 0.47138140, Log Avg loss: 0.61863255, Global Avg Loss: 1.27264252, Time: 0.0076 Steps: 85650, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000500, Sample Num: 8000, Cur Loss: 0.83453542, Cur Avg Loss: 0.47129111, Log Avg loss: 0.46686678, Global Avg Loss: 1.27254845, Time: 0.0076 Steps: 85660, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000510, Sample Num: 8160, Cur Loss: 0.36064887, Cur Avg Loss: 0.47287877, Log Avg loss: 0.55226175, Global Avg Loss: 1.27246437, Time: 0.0075 Steps: 85670, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000520, Sample Num: 8320, Cur Loss: 0.16812488, Cur Avg Loss: 0.47181866, Log Avg loss: 0.41775290, Global Avg Loss: 1.27236462, Time: 0.0076 Steps: 85680, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000530, Sample Num: 8480, Cur Loss: 0.62083608, Cur Avg Loss: 0.47138111, Log Avg loss: 0.44862895, Global Avg Loss: 1.27226849, Time: 0.0075 Steps: 85690, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000540, Sample Num: 8640, Cur Loss: 0.20338959, Cur Avg Loss: 0.46836807, Log Avg loss: 0.30867677, Global Avg Loss: 1.27215605, Time: 0.0076 Steps: 85700, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000550, Sample Num: 8800, Cur Loss: 0.64699298, Cur Avg Loss: 0.46568592, Log Avg loss: 0.32084977, Global Avg Loss: 1.27204506, Time: 0.0077 Steps: 85710, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000560, Sample Num: 8960, Cur Loss: 0.60681832, Cur Avg Loss: 0.46701811, Log Avg loss: 0.54028860, Global Avg Loss: 1.27195969, Time: 0.0076 Steps: 85720, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000570, Sample Num: 9120, Cur Loss: 0.46468323, Cur Avg Loss: 0.46693649, Log Avg loss: 0.46236577, Global Avg Loss: 1.27186526, Time: 0.0075 Steps: 85730, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000580, Sample Num: 9280, Cur Loss: 0.53197169, Cur Avg Loss: 0.46831414, Log Avg loss: 0.54684034, Global Avg Loss: 1.27178070, Time: 0.0075 Steps: 85740, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000590, Sample Num: 9440, Cur Loss: 0.33806789, Cur Avg Loss: 0.46884801, Log Avg loss: 0.49981210, Global Avg Loss: 1.27169067, Time: 0.0076 Steps: 85750, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000600, Sample Num: 9600, Cur Loss: 0.26038638, Cur Avg Loss: 0.46751611, Log Avg loss: 0.38893433, Global Avg Loss: 1.27158774, Time: 0.0077 Steps: 85760, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000610, Sample Num: 9760, Cur Loss: 0.26591146, Cur Avg Loss: 0.46585324, Log Avg loss: 0.36608074, Global Avg Loss: 1.27148216, Time: 0.0075 Steps: 85770, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000620, Sample Num: 9920, Cur Loss: 0.27487472, Cur Avg Loss: 0.46856340, Log Avg loss: 0.63388313, Global Avg Loss: 1.27140783, Time: 0.0076 Steps: 85780, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000630, Sample Num: 10080, Cur Loss: 0.33443010, Cur Avg Loss: 0.47095904, Log Avg loss: 0.61948889, Global Avg Loss: 1.27133184, Time: 0.0075 Steps: 85790, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000640, Sample Num: 10240, Cur Loss: 0.20214269, Cur Avg Loss: 0.47236923, Log Avg loss: 0.56121111, Global Avg Loss: 1.27124908, Time: 0.0075 Steps: 85800, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000650, Sample Num: 10400, Cur Loss: 0.18654940, Cur Avg Loss: 0.47216261, Log Avg loss: 0.45893906, Global Avg Loss: 1.27115441, Time: 0.0076 Steps: 85810, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000660, Sample Num: 10560, Cur Loss: 0.49203384, Cur Avg Loss: 0.47105884, Log Avg loss: 0.39931387, Global Avg Loss: 1.27105283, Time: 0.0076 Steps: 85820, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000670, Sample Num: 10720, Cur Loss: 0.52961153, Cur Avg Loss: 0.47071257, Log Avg loss: 0.44785890, Global Avg Loss: 1.27095692, Time: 0.0076 Steps: 85830, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000680, Sample Num: 10880, Cur Loss: 0.48737168, Cur Avg Loss: 0.47011650, Log Avg loss: 0.43017968, Global Avg Loss: 1.27085897, Time: 0.0076 Steps: 85840, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000690, Sample Num: 11040, Cur Loss: 0.45326167, Cur Avg Loss: 0.46992794, Log Avg loss: 0.45710550, Global Avg Loss: 1.27076418, Time: 0.0074 Steps: 85850, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000700, Sample Num: 11200, Cur Loss: 0.74203598, Cur Avg Loss: 0.46964430, Log Avg loss: 0.45007306, Global Avg Loss: 1.27066860, Time: 0.0075 Steps: 85860, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000710, Sample Num: 11360, Cur Loss: 0.42857403, Cur Avg Loss: 0.47012862, Log Avg loss: 0.50403110, Global Avg Loss: 1.27057932, Time: 0.0075 Steps: 85870, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000720, Sample Num: 11520, Cur Loss: 0.79757392, Cur Avg Loss: 0.47058049, Log Avg loss: 0.50266323, Global Avg Loss: 1.27048990, Time: 0.0076 Steps: 85880, Updated lr: 0.000020 Training, Epoch: 0041, Batch: 000730, Sample Num: 11680, Cur Loss: 0.32223582, Cur Avg Loss: 0.47126095, Log Avg loss: 0.52025463, Global Avg Loss: 1.27040255, Time: 0.0076 Steps: 85890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000740, Sample Num: 11840, Cur Loss: 0.45335835, Cur Avg Loss: 0.47027762, Log Avg loss: 0.39849457, Global Avg Loss: 1.27030105, Time: 0.0077 Steps: 85900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000750, Sample Num: 12000, Cur Loss: 0.29233381, Cur Avg Loss: 0.46998904, Log Avg loss: 0.44863395, Global Avg Loss: 1.27020541, Time: 0.0074 Steps: 85910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000760, Sample Num: 12160, Cur Loss: 0.59551936, Cur Avg Loss: 0.47031553, Log Avg loss: 0.49480190, Global Avg Loss: 1.27011516, Time: 0.0076 Steps: 85920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000770, Sample Num: 12320, Cur Loss: 0.24344970, Cur Avg Loss: 0.47024612, Log Avg loss: 0.46497108, Global Avg Loss: 1.27002146, Time: 0.0076 Steps: 85930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000780, Sample Num: 12480, Cur Loss: 0.41004771, Cur Avg Loss: 0.46943339, Log Avg loss: 0.40685328, Global Avg Loss: 1.26992102, Time: 0.0077 Steps: 85940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000790, Sample Num: 12640, Cur Loss: 0.33723837, Cur Avg Loss: 0.46836498, Log Avg loss: 0.38502933, Global Avg Loss: 1.26981807, Time: 0.0076 Steps: 85950, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000800, Sample Num: 12800, Cur Loss: 0.31703544, Cur Avg Loss: 0.46777437, Log Avg loss: 0.42111596, Global Avg Loss: 1.26971934, Time: 0.0076 Steps: 85960, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000810, Sample Num: 12960, Cur Loss: 0.46140304, Cur Avg Loss: 0.46723619, Log Avg loss: 0.42418179, Global Avg Loss: 1.26962098, Time: 0.0075 Steps: 85970, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000820, Sample Num: 13120, Cur Loss: 0.59721386, Cur Avg Loss: 0.46825934, Log Avg loss: 0.55113447, Global Avg Loss: 1.26953742, Time: 0.0075 Steps: 85980, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000830, Sample Num: 13280, Cur Loss: 0.33270007, Cur Avg Loss: 0.46863835, Log Avg loss: 0.49971722, Global Avg Loss: 1.26944789, Time: 0.0076 Steps: 85990, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000840, Sample Num: 13440, Cur Loss: 0.64511895, Cur Avg Loss: 0.46854031, Log Avg loss: 0.46040261, Global Avg Loss: 1.26935382, Time: 0.0075 Steps: 86000, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000850, Sample Num: 13600, Cur Loss: 0.39512879, Cur Avg Loss: 0.46968214, Log Avg loss: 0.56559579, Global Avg Loss: 1.26927200, Time: 0.0076 Steps: 86010, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000860, Sample Num: 13760, Cur Loss: 0.46207830, Cur Avg Loss: 0.46938632, Log Avg loss: 0.44424177, Global Avg Loss: 1.26917609, Time: 0.0075 Steps: 86020, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000870, Sample Num: 13920, Cur Loss: 0.44182417, Cur Avg Loss: 0.46907242, Log Avg loss: 0.44207720, Global Avg Loss: 1.26907994, Time: 0.0076 Steps: 86030, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000880, Sample Num: 14080, Cur Loss: 0.37307340, Cur Avg Loss: 0.46951549, Log Avg loss: 0.50806248, Global Avg Loss: 1.26899150, Time: 0.0076 Steps: 86040, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000890, Sample Num: 14240, Cur Loss: 0.26693398, Cur Avg Loss: 0.46984938, Log Avg loss: 0.49923183, Global Avg Loss: 1.26890204, Time: 0.0076 Steps: 86050, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000900, Sample Num: 14400, Cur Loss: 0.29223365, Cur Avg Loss: 0.47010147, Log Avg loss: 0.49253748, Global Avg Loss: 1.26881183, Time: 0.0076 Steps: 86060, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000910, Sample Num: 14560, Cur Loss: 0.38333029, Cur Avg Loss: 0.47179946, Log Avg loss: 0.62461865, Global Avg Loss: 1.26873698, Time: 0.0075 Steps: 86070, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000920, Sample Num: 14720, Cur Loss: 1.00922322, Cur Avg Loss: 0.47193716, Log Avg loss: 0.48446786, Global Avg Loss: 1.26864587, Time: 0.0077 Steps: 86080, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000930, Sample Num: 14880, Cur Loss: 0.47629905, Cur Avg Loss: 0.47205923, Log Avg loss: 0.48328978, Global Avg Loss: 1.26855465, Time: 0.0076 Steps: 86090, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000940, Sample Num: 15040, Cur Loss: 0.23332818, Cur Avg Loss: 0.47280256, Log Avg loss: 0.54193157, Global Avg Loss: 1.26847026, Time: 0.0078 Steps: 86100, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000950, Sample Num: 15200, Cur Loss: 0.62298977, Cur Avg Loss: 0.47284163, Log Avg loss: 0.47651501, Global Avg Loss: 1.26837829, Time: 0.0073 Steps: 86110, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000960, Sample Num: 15360, Cur Loss: 0.41026539, Cur Avg Loss: 0.47426547, Log Avg loss: 0.60952998, Global Avg Loss: 1.26830178, Time: 0.0066 Steps: 86120, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000970, Sample Num: 15520, Cur Loss: 0.16033342, Cur Avg Loss: 0.47393484, Log Avg loss: 0.44219421, Global Avg Loss: 1.26820587, Time: 0.0144 Steps: 86130, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000980, Sample Num: 15680, Cur Loss: 1.08817244, Cur Avg Loss: 0.47626414, Log Avg loss: 0.70220624, Global Avg Loss: 1.26814016, Time: 0.0064 Steps: 86140, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 000990, Sample Num: 15840, Cur Loss: 0.89936119, Cur Avg Loss: 0.47770414, Log Avg loss: 0.61882450, Global Avg Loss: 1.26806479, Time: 0.0169 Steps: 86150, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001000, Sample Num: 16000, Cur Loss: 0.10885988, Cur Avg Loss: 0.47711271, Log Avg loss: 0.41856097, Global Avg Loss: 1.26796619, Time: 0.0132 Steps: 86160, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001010, Sample Num: 16160, Cur Loss: 0.62353224, Cur Avg Loss: 0.47758032, Log Avg loss: 0.52434131, Global Avg Loss: 1.26787990, Time: 0.0144 Steps: 86170, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001020, Sample Num: 16320, Cur Loss: 1.03518248, Cur Avg Loss: 0.47770398, Log Avg loss: 0.49019304, Global Avg Loss: 1.26778966, Time: 0.0199 Steps: 86180, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001030, Sample Num: 16480, Cur Loss: 0.55478746, Cur Avg Loss: 0.47809820, Log Avg loss: 0.51830890, Global Avg Loss: 1.26770270, Time: 0.0066 Steps: 86190, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001040, Sample Num: 16640, Cur Loss: 0.66898400, Cur Avg Loss: 0.47814840, Log Avg loss: 0.48331963, Global Avg Loss: 1.26761171, Time: 0.0198 Steps: 86200, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001050, Sample Num: 16800, Cur Loss: 0.27585560, Cur Avg Loss: 0.47681249, Log Avg loss: 0.33787703, Global Avg Loss: 1.26750386, Time: 0.0066 Steps: 86210, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001060, Sample Num: 16960, Cur Loss: 0.47429669, Cur Avg Loss: 0.47617637, Log Avg loss: 0.40938455, Global Avg Loss: 1.26740433, Time: 0.0064 Steps: 86220, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001070, Sample Num: 17120, Cur Loss: 0.85983586, Cur Avg Loss: 0.47826270, Log Avg loss: 0.69941332, Global Avg Loss: 1.26733846, Time: 0.0233 Steps: 86230, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001080, Sample Num: 17280, Cur Loss: 0.17677660, Cur Avg Loss: 0.47826567, Log Avg loss: 0.47858314, Global Avg Loss: 1.26724700, Time: 0.0192 Steps: 86240, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001090, Sample Num: 17440, Cur Loss: 0.83255082, Cur Avg Loss: 0.47954384, Log Avg loss: 0.61758614, Global Avg Loss: 1.26717168, Time: 0.0067 Steps: 86250, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001100, Sample Num: 17600, Cur Loss: 0.48907322, Cur Avg Loss: 0.48008638, Log Avg loss: 0.53922392, Global Avg Loss: 1.26708729, Time: 0.0073 Steps: 86260, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001110, Sample Num: 17760, Cur Loss: 0.33983487, Cur Avg Loss: 0.47972418, Log Avg loss: 0.43988154, Global Avg Loss: 1.26699140, Time: 0.0072 Steps: 86270, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001120, Sample Num: 17920, Cur Loss: 0.47685933, Cur Avg Loss: 0.47946957, Log Avg loss: 0.45120869, Global Avg Loss: 1.26689685, Time: 0.0192 Steps: 86280, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001130, Sample Num: 18080, Cur Loss: 0.46613419, Cur Avg Loss: 0.48010721, Log Avg loss: 0.55152206, Global Avg Loss: 1.26681395, Time: 0.0066 Steps: 86290, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001140, Sample Num: 18240, Cur Loss: 0.27181679, Cur Avg Loss: 0.47860338, Log Avg loss: 0.30867095, Global Avg Loss: 1.26670293, Time: 0.0187 Steps: 86300, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001150, Sample Num: 18400, Cur Loss: 0.32624876, Cur Avg Loss: 0.47894141, Log Avg loss: 0.51747685, Global Avg Loss: 1.26661612, Time: 0.0065 Steps: 86310, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001160, Sample Num: 18560, Cur Loss: 0.29433483, Cur Avg Loss: 0.47827384, Log Avg loss: 0.40150335, Global Avg Loss: 1.26651590, Time: 0.0115 Steps: 86320, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001170, Sample Num: 18720, Cur Loss: 0.38869709, Cur Avg Loss: 0.47808138, Log Avg loss: 0.45575582, Global Avg Loss: 1.26642198, Time: 0.0064 Steps: 86330, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001180, Sample Num: 18880, Cur Loss: 0.81475770, Cur Avg Loss: 0.47782693, Log Avg loss: 0.44805581, Global Avg Loss: 1.26632720, Time: 0.0063 Steps: 86340, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001190, Sample Num: 19040, Cur Loss: 0.71793222, Cur Avg Loss: 0.47759481, Log Avg loss: 0.45020471, Global Avg Loss: 1.26623269, Time: 0.0181 Steps: 86350, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001200, Sample Num: 19200, Cur Loss: 0.31788397, Cur Avg Loss: 0.47783532, Log Avg loss: 0.50645658, Global Avg Loss: 1.26614471, Time: 0.0066 Steps: 86360, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001210, Sample Num: 19360, Cur Loss: 0.70859718, Cur Avg Loss: 0.47744958, Log Avg loss: 0.43116059, Global Avg Loss: 1.26604803, Time: 0.0067 Steps: 86370, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001220, Sample Num: 19520, Cur Loss: 0.11631903, Cur Avg Loss: 0.47827462, Log Avg loss: 0.57810434, Global Avg Loss: 1.26596839, Time: 0.0066 Steps: 86380, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001230, Sample Num: 19680, Cur Loss: 0.73012620, Cur Avg Loss: 0.47920735, Log Avg loss: 0.59300079, Global Avg Loss: 1.26589049, Time: 0.0066 Steps: 86390, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001240, Sample Num: 19840, Cur Loss: 0.26279458, Cur Avg Loss: 0.47911732, Log Avg loss: 0.46804337, Global Avg Loss: 1.26579815, Time: 0.0066 Steps: 86400, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001250, Sample Num: 20000, Cur Loss: 0.36209086, Cur Avg Loss: 0.47869142, Log Avg loss: 0.42588027, Global Avg Loss: 1.26570095, Time: 0.0067 Steps: 86410, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001260, Sample Num: 20160, Cur Loss: 0.50650012, Cur Avg Loss: 0.47839148, Log Avg loss: 0.44089833, Global Avg Loss: 1.26560551, Time: 0.0076 Steps: 86420, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001270, Sample Num: 20320, Cur Loss: 0.30916569, Cur Avg Loss: 0.47839092, Log Avg loss: 0.47832055, Global Avg Loss: 1.26551442, Time: 0.0075 Steps: 86430, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001280, Sample Num: 20480, Cur Loss: 0.62892425, Cur Avg Loss: 0.47884402, Log Avg loss: 0.53638769, Global Avg Loss: 1.26543007, Time: 0.0083 Steps: 86440, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001290, Sample Num: 20640, Cur Loss: 0.45584983, Cur Avg Loss: 0.47890109, Log Avg loss: 0.48620673, Global Avg Loss: 1.26533993, Time: 0.0081 Steps: 86450, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001300, Sample Num: 20800, Cur Loss: 0.62795949, Cur Avg Loss: 0.47849784, Log Avg loss: 0.42647793, Global Avg Loss: 1.26524291, Time: 0.0081 Steps: 86460, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001310, Sample Num: 20960, Cur Loss: 0.38676339, Cur Avg Loss: 0.47849491, Log Avg loss: 0.47811438, Global Avg Loss: 1.26515188, Time: 0.0081 Steps: 86470, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001320, Sample Num: 21120, Cur Loss: 0.47752100, Cur Avg Loss: 0.47942037, Log Avg loss: 0.60065497, Global Avg Loss: 1.26507504, Time: 0.0081 Steps: 86480, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001330, Sample Num: 21280, Cur Loss: 0.30112958, Cur Avg Loss: 0.47972334, Log Avg loss: 0.51971596, Global Avg Loss: 1.26498886, Time: 0.0081 Steps: 86490, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001340, Sample Num: 21440, Cur Loss: 0.32563424, Cur Avg Loss: 0.47965801, Log Avg loss: 0.47096897, Global Avg Loss: 1.26489707, Time: 0.0081 Steps: 86500, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001350, Sample Num: 21600, Cur Loss: 0.72129363, Cur Avg Loss: 0.47882289, Log Avg loss: 0.36691734, Global Avg Loss: 1.26479327, Time: 0.0082 Steps: 86510, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001360, Sample Num: 21760, Cur Loss: 0.21101066, Cur Avg Loss: 0.47793532, Log Avg loss: 0.35811256, Global Avg Loss: 1.26468847, Time: 0.0081 Steps: 86520, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001370, Sample Num: 21920, Cur Loss: 0.09542239, Cur Avg Loss: 0.47740255, Log Avg loss: 0.40494592, Global Avg Loss: 1.26458912, Time: 0.0081 Steps: 86530, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001380, Sample Num: 22080, Cur Loss: 0.13183431, Cur Avg Loss: 0.47706011, Log Avg loss: 0.43014536, Global Avg Loss: 1.26449269, Time: 0.0081 Steps: 86540, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001390, Sample Num: 22240, Cur Loss: 0.57402205, Cur Avg Loss: 0.47654368, Log Avg loss: 0.40527716, Global Avg Loss: 1.26439342, Time: 0.0081 Steps: 86550, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001400, Sample Num: 22400, Cur Loss: 0.48959428, Cur Avg Loss: 0.47709930, Log Avg loss: 0.55433075, Global Avg Loss: 1.26431139, Time: 0.0081 Steps: 86560, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001410, Sample Num: 22560, Cur Loss: 0.38206542, Cur Avg Loss: 0.47663556, Log Avg loss: 0.41171166, Global Avg Loss: 1.26421290, Time: 0.0081 Steps: 86570, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001420, Sample Num: 22720, Cur Loss: 0.27835655, Cur Avg Loss: 0.47709305, Log Avg loss: 0.54159853, Global Avg Loss: 1.26412944, Time: 0.0081 Steps: 86580, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001430, Sample Num: 22880, Cur Loss: 0.25406602, Cur Avg Loss: 0.47684945, Log Avg loss: 0.44225867, Global Avg Loss: 1.26403452, Time: 0.0081 Steps: 86590, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001440, Sample Num: 23040, Cur Loss: 1.23936069, Cur Avg Loss: 0.47663507, Log Avg loss: 0.44597836, Global Avg Loss: 1.26394006, Time: 0.0081 Steps: 86600, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001450, Sample Num: 23200, Cur Loss: 0.62910879, Cur Avg Loss: 0.47667352, Log Avg loss: 0.48221067, Global Avg Loss: 1.26384980, Time: 0.0081 Steps: 86610, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001460, Sample Num: 23360, Cur Loss: 0.56973487, Cur Avg Loss: 0.47802612, Log Avg loss: 0.67415346, Global Avg Loss: 1.26378172, Time: 0.0081 Steps: 86620, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001470, Sample Num: 23520, Cur Loss: 0.56675851, Cur Avg Loss: 0.47846611, Log Avg loss: 0.54270466, Global Avg Loss: 1.26369849, Time: 0.0080 Steps: 86630, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001480, Sample Num: 23680, Cur Loss: 0.55130333, Cur Avg Loss: 0.47816071, Log Avg loss: 0.43326622, Global Avg Loss: 1.26360264, Time: 0.0081 Steps: 86640, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001490, Sample Num: 23840, Cur Loss: 0.36364728, Cur Avg Loss: 0.47802908, Log Avg loss: 0.45854849, Global Avg Loss: 1.26350973, Time: 0.0081 Steps: 86650, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001500, Sample Num: 24000, Cur Loss: 1.48460817, Cur Avg Loss: 0.47940195, Log Avg loss: 0.68395959, Global Avg Loss: 1.26344285, Time: 0.0081 Steps: 86660, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001510, Sample Num: 24160, Cur Loss: 0.52997422, Cur Avg Loss: 0.47950108, Log Avg loss: 0.49437034, Global Avg Loss: 1.26335412, Time: 0.0081 Steps: 86670, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001520, Sample Num: 24320, Cur Loss: 0.19889185, Cur Avg Loss: 0.47892058, Log Avg loss: 0.39126573, Global Avg Loss: 1.26325351, Time: 0.0081 Steps: 86680, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001530, Sample Num: 24480, Cur Loss: 0.43580970, Cur Avg Loss: 0.47942343, Log Avg loss: 0.55585658, Global Avg Loss: 1.26317191, Time: 0.0080 Steps: 86690, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001540, Sample Num: 24640, Cur Loss: 0.32009512, Cur Avg Loss: 0.47940773, Log Avg loss: 0.47700475, Global Avg Loss: 1.26308123, Time: 0.0082 Steps: 86700, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001550, Sample Num: 24800, Cur Loss: 0.52783346, Cur Avg Loss: 0.47932439, Log Avg loss: 0.46649041, Global Avg Loss: 1.26298936, Time: 0.0076 Steps: 86710, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001560, Sample Num: 24960, Cur Loss: 0.97754163, Cur Avg Loss: 0.47943853, Log Avg loss: 0.49713015, Global Avg Loss: 1.26290105, Time: 0.0076 Steps: 86720, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001570, Sample Num: 25120, Cur Loss: 0.43643928, Cur Avg Loss: 0.47849204, Log Avg loss: 0.33083910, Global Avg Loss: 1.26279358, Time: 0.0076 Steps: 86730, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001580, Sample Num: 25280, Cur Loss: 0.20458415, Cur Avg Loss: 0.47809659, Log Avg loss: 0.41601094, Global Avg Loss: 1.26269596, Time: 0.0076 Steps: 86740, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001590, Sample Num: 25440, Cur Loss: 0.22747886, Cur Avg Loss: 0.47757454, Log Avg loss: 0.39509075, Global Avg Loss: 1.26259594, Time: 0.0076 Steps: 86750, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001600, Sample Num: 25600, Cur Loss: 0.31925991, Cur Avg Loss: 0.47659616, Log Avg loss: 0.32103481, Global Avg Loss: 1.26248742, Time: 0.0077 Steps: 86760, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001610, Sample Num: 25760, Cur Loss: 0.63194323, Cur Avg Loss: 0.47675387, Log Avg loss: 0.50198705, Global Avg Loss: 1.26239977, Time: 0.0075 Steps: 86770, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001620, Sample Num: 25920, Cur Loss: 0.27734858, Cur Avg Loss: 0.47694656, Log Avg loss: 0.50796896, Global Avg Loss: 1.26231284, Time: 0.0076 Steps: 86780, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001630, Sample Num: 26080, Cur Loss: 0.59065330, Cur Avg Loss: 0.47693762, Log Avg loss: 0.47549002, Global Avg Loss: 1.26222218, Time: 0.0076 Steps: 86790, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001640, Sample Num: 26240, Cur Loss: 1.13268447, Cur Avg Loss: 0.47846617, Log Avg loss: 0.72761950, Global Avg Loss: 1.26216059, Time: 0.0076 Steps: 86800, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001650, Sample Num: 26400, Cur Loss: 0.63958108, Cur Avg Loss: 0.47885456, Log Avg loss: 0.54255052, Global Avg Loss: 1.26207769, Time: 0.0076 Steps: 86810, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001660, Sample Num: 26560, Cur Loss: 0.67843390, Cur Avg Loss: 0.47878292, Log Avg loss: 0.46696317, Global Avg Loss: 1.26198611, Time: 0.0075 Steps: 86820, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001670, Sample Num: 26720, Cur Loss: 0.35775021, Cur Avg Loss: 0.47921281, Log Avg loss: 0.55057421, Global Avg Loss: 1.26190418, Time: 0.0076 Steps: 86830, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001680, Sample Num: 26880, Cur Loss: 0.62590033, Cur Avg Loss: 0.47933238, Log Avg loss: 0.49930066, Global Avg Loss: 1.26181636, Time: 0.0075 Steps: 86840, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001690, Sample Num: 27040, Cur Loss: 0.36107010, Cur Avg Loss: 0.47960076, Log Avg loss: 0.52468776, Global Avg Loss: 1.26173149, Time: 0.0075 Steps: 86850, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001700, Sample Num: 27200, Cur Loss: 0.33593985, Cur Avg Loss: 0.48093653, Log Avg loss: 0.70668215, Global Avg Loss: 1.26166759, Time: 0.0076 Steps: 86860, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001710, Sample Num: 27360, Cur Loss: 0.68413430, Cur Avg Loss: 0.48163966, Log Avg loss: 0.60117156, Global Avg Loss: 1.26159156, Time: 0.0076 Steps: 86870, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001720, Sample Num: 27520, Cur Loss: 0.14473933, Cur Avg Loss: 0.48130870, Log Avg loss: 0.42471527, Global Avg Loss: 1.26149523, Time: 0.0075 Steps: 86880, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001730, Sample Num: 27680, Cur Loss: 0.44234690, Cur Avg Loss: 0.48140449, Log Avg loss: 0.49787919, Global Avg Loss: 1.26140735, Time: 0.0077 Steps: 86890, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001740, Sample Num: 27840, Cur Loss: 1.00000107, Cur Avg Loss: 0.48110765, Log Avg loss: 0.42975393, Global Avg Loss: 1.26131164, Time: 0.0076 Steps: 86900, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001750, Sample Num: 28000, Cur Loss: 0.18399701, Cur Avg Loss: 0.48089734, Log Avg loss: 0.44430458, Global Avg Loss: 1.26121764, Time: 0.0076 Steps: 86910, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001760, Sample Num: 28160, Cur Loss: 1.41580939, Cur Avg Loss: 0.48200469, Log Avg loss: 0.67579100, Global Avg Loss: 1.26115029, Time: 0.0076 Steps: 86920, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001770, Sample Num: 28320, Cur Loss: 0.26338413, Cur Avg Loss: 0.48195590, Log Avg loss: 0.47336810, Global Avg Loss: 1.26105966, Time: 0.0076 Steps: 86930, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001780, Sample Num: 28480, Cur Loss: 0.15198432, Cur Avg Loss: 0.48234710, Log Avg loss: 0.55159050, Global Avg Loss: 1.26097806, Time: 0.0075 Steps: 86940, Updated lr: 0.000019 Training, Epoch: 0041, Batch: 001790, Sample Num: 28640, Cur Loss: 0.27846271, Cur Avg Loss: 0.48304994, Log Avg loss: 0.60815421, Global Avg Loss: 1.26090298, Time: 0.0077 Steps: 86950, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001800, Sample Num: 28800, Cur Loss: 0.18380874, Cur Avg Loss: 0.48244881, Log Avg loss: 0.37484805, Global Avg Loss: 1.26080109, Time: 0.0075 Steps: 86960, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001810, Sample Num: 28960, Cur Loss: 0.28691667, Cur Avg Loss: 0.48193194, Log Avg loss: 0.38889534, Global Avg Loss: 1.26070083, Time: 0.0076 Steps: 86970, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001820, Sample Num: 29120, Cur Loss: 0.51573372, Cur Avg Loss: 0.48129277, Log Avg loss: 0.36560174, Global Avg Loss: 1.26059792, Time: 0.0076 Steps: 86980, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001830, Sample Num: 29280, Cur Loss: 0.30381060, Cur Avg Loss: 0.48130815, Log Avg loss: 0.48410724, Global Avg Loss: 1.26050866, Time: 0.0075 Steps: 86990, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001840, Sample Num: 29440, Cur Loss: 0.24005195, Cur Avg Loss: 0.48104738, Log Avg loss: 0.43332693, Global Avg Loss: 1.26041358, Time: 0.0076 Steps: 87000, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001850, Sample Num: 29600, Cur Loss: 0.38475978, Cur Avg Loss: 0.48071385, Log Avg loss: 0.41934501, Global Avg Loss: 1.26031692, Time: 0.0075 Steps: 87010, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001860, Sample Num: 29760, Cur Loss: 0.85436893, Cur Avg Loss: 0.48091618, Log Avg loss: 0.51834630, Global Avg Loss: 1.26023166, Time: 0.0074 Steps: 87020, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001870, Sample Num: 29920, Cur Loss: 0.93262661, Cur Avg Loss: 0.48125831, Log Avg loss: 0.54489513, Global Avg Loss: 1.26014946, Time: 0.0076 Steps: 87030, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001880, Sample Num: 30080, Cur Loss: 0.42173573, Cur Avg Loss: 0.48144120, Log Avg loss: 0.51564217, Global Avg Loss: 1.26006393, Time: 0.0075 Steps: 87040, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001890, Sample Num: 30240, Cur Loss: 0.49881393, Cur Avg Loss: 0.48132659, Log Avg loss: 0.45977823, Global Avg Loss: 1.25997199, Time: 0.0240 Steps: 87050, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001900, Sample Num: 30400, Cur Loss: 0.87823611, Cur Avg Loss: 0.48244819, Log Avg loss: 0.69443138, Global Avg Loss: 1.25990703, Time: 0.0166 Steps: 87060, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001910, Sample Num: 30560, Cur Loss: 1.58135724, Cur Avg Loss: 0.48279493, Log Avg loss: 0.54867629, Global Avg Loss: 1.25982535, Time: 0.0209 Steps: 87070, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001920, Sample Num: 30720, Cur Loss: 0.42825058, Cur Avg Loss: 0.48274527, Log Avg loss: 0.47325930, Global Avg Loss: 1.25973502, Time: 0.0181 Steps: 87080, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001930, Sample Num: 30880, Cur Loss: 0.43664977, Cur Avg Loss: 0.48315791, Log Avg loss: 0.56238555, Global Avg Loss: 1.25965495, Time: 0.0064 Steps: 87090, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001940, Sample Num: 31040, Cur Loss: 0.25349915, Cur Avg Loss: 0.48335614, Log Avg loss: 0.52161374, Global Avg Loss: 1.25957021, Time: 0.0074 Steps: 87100, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001950, Sample Num: 31200, Cur Loss: 0.53824222, Cur Avg Loss: 0.48341925, Log Avg loss: 0.49566285, Global Avg Loss: 1.25948252, Time: 0.0064 Steps: 87110, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001960, Sample Num: 31360, Cur Loss: 0.84385812, Cur Avg Loss: 0.48333636, Log Avg loss: 0.46717206, Global Avg Loss: 1.25939157, Time: 0.0073 Steps: 87120, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001970, Sample Num: 31520, Cur Loss: 0.40850660, Cur Avg Loss: 0.48301680, Log Avg loss: 0.42038342, Global Avg Loss: 1.25929528, Time: 0.0071 Steps: 87130, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001980, Sample Num: 31680, Cur Loss: 0.32582223, Cur Avg Loss: 0.48286948, Log Avg loss: 0.45384805, Global Avg Loss: 1.25920285, Time: 0.0079 Steps: 87140, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 001990, Sample Num: 31840, Cur Loss: 0.55210394, Cur Avg Loss: 0.48264241, Log Avg loss: 0.43768276, Global Avg Loss: 1.25910858, Time: 0.0065 Steps: 87150, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002000, Sample Num: 32000, Cur Loss: 0.54838020, Cur Avg Loss: 0.48217658, Log Avg loss: 0.38947627, Global Avg Loss: 1.25900881, Time: 0.0064 Steps: 87160, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002010, Sample Num: 32160, Cur Loss: 0.13748443, Cur Avg Loss: 0.48177737, Log Avg loss: 0.40193503, Global Avg Loss: 1.25891049, Time: 0.0064 Steps: 87170, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002020, Sample Num: 32320, Cur Loss: 0.41735458, Cur Avg Loss: 0.48219056, Log Avg loss: 0.56524111, Global Avg Loss: 1.25883092, Time: 0.0127 Steps: 87180, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002030, Sample Num: 32480, Cur Loss: 0.18447720, Cur Avg Loss: 0.48225993, Log Avg loss: 0.49627252, Global Avg Loss: 1.25874346, Time: 0.0178 Steps: 87190, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002040, Sample Num: 32640, Cur Loss: 0.23206279, Cur Avg Loss: 0.48184102, Log Avg loss: 0.39680292, Global Avg Loss: 1.25864461, Time: 0.0063 Steps: 87200, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002050, Sample Num: 32800, Cur Loss: 0.31104970, Cur Avg Loss: 0.48155971, Log Avg loss: 0.42417317, Global Avg Loss: 1.25854893, Time: 0.0074 Steps: 87210, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002060, Sample Num: 32960, Cur Loss: 0.16579871, Cur Avg Loss: 0.48138724, Log Avg loss: 0.44603049, Global Avg Loss: 1.25845577, Time: 0.0073 Steps: 87220, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002070, Sample Num: 33120, Cur Loss: 0.36273038, Cur Avg Loss: 0.48145069, Log Avg loss: 0.49452058, Global Avg Loss: 1.25836819, Time: 0.0074 Steps: 87230, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002080, Sample Num: 33280, Cur Loss: 0.49124110, Cur Avg Loss: 0.48149665, Log Avg loss: 0.49101017, Global Avg Loss: 1.25828023, Time: 0.0075 Steps: 87240, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002090, Sample Num: 33440, Cur Loss: 0.63417208, Cur Avg Loss: 0.48089547, Log Avg loss: 0.35585019, Global Avg Loss: 1.25817680, Time: 0.0075 Steps: 87250, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002100, Sample Num: 33600, Cur Loss: 0.12377542, Cur Avg Loss: 0.48024794, Log Avg loss: 0.34491485, Global Avg Loss: 1.25807214, Time: 0.0074 Steps: 87260, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002110, Sample Num: 33760, Cur Loss: 1.03518200, Cur Avg Loss: 0.48043736, Log Avg loss: 0.52021563, Global Avg Loss: 1.25798760, Time: 0.0075 Steps: 87270, Updated lr: 0.000018 Training, Epoch: 0041, Batch: 002120, Sample Num: 33920, Cur Loss: 0.60294461, Cur Avg Loss: 0.48118602, Log Avg loss: 0.63915321, Global Avg Loss: 1.25791669, Time: 0.0077 Steps: 87280, Updated lr: 0.000018 ***** Running evaluation checkpoint-87289 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-87289 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 18.043831, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.575296, "eval_total_loss": 404.433186, "eval_mae": 0.562344, "eval_mse": 0.575458, "eval_r2": 0.634201, "eval_sp_statistic": 0.753126, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.796557, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.868807, "test_total_loss": 436.141071, "test_mae": 0.664786, "test_mse": 0.869033, "test_r2": 0.439119, "test_sp_statistic": 0.59558, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.669218, "test_ps_pvalue": 0.0, "lr": 1.8170697012802277e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2578178732256746, "train_cur_epoch_loss": 1022.8097023069859, "train_cur_epoch_avg_loss": 0.4804178968092935, "train_cur_epoch_time": 18.04383111000061, "train_cur_epoch_avg_time": 0.00847526120713979, "epoch": 41, "step": 87289} ################################################## Training, Epoch: 0042, Batch: 000001, Sample Num: 16, Cur Loss: 0.30543473, Cur Avg Loss: 0.30543473, Log Avg loss: 0.30007740, Global Avg Loss: 1.25780696, Time: 0.0068 Steps: 87290, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000011, Sample Num: 176, Cur Loss: 0.80279994, Cur Avg Loss: 0.43575214, Log Avg loss: 0.44878388, Global Avg Loss: 1.25771429, Time: 0.0067 Steps: 87300, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000021, Sample Num: 336, Cur Loss: 0.46685809, Cur Avg Loss: 0.40975335, Log Avg loss: 0.38115469, Global Avg Loss: 1.25761389, Time: 0.0067 Steps: 87310, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000031, Sample Num: 496, Cur Loss: 0.96876943, Cur Avg Loss: 0.43075666, Log Avg loss: 0.47486361, Global Avg Loss: 1.25752425, Time: 0.0076 Steps: 87320, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000041, Sample Num: 656, Cur Loss: 0.43314961, Cur Avg Loss: 0.43124900, Log Avg loss: 0.43277524, Global Avg Loss: 1.25742981, Time: 0.0076 Steps: 87330, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000051, Sample Num: 816, Cur Loss: 0.34206793, Cur Avg Loss: 0.43061913, Log Avg loss: 0.42803664, Global Avg Loss: 1.25733485, Time: 0.0076 Steps: 87340, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000061, Sample Num: 976, Cur Loss: 0.27263299, Cur Avg Loss: 0.42892722, Log Avg loss: 0.42029849, Global Avg Loss: 1.25723903, Time: 0.0076 Steps: 87350, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000071, Sample Num: 1136, Cur Loss: 0.48937237, Cur Avg Loss: 0.43331727, Log Avg loss: 0.46009657, Global Avg Loss: 1.25714778, Time: 0.0074 Steps: 87360, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000081, Sample Num: 1296, Cur Loss: 0.52628714, Cur Avg Loss: 0.43540207, Log Avg loss: 0.45020414, Global Avg Loss: 1.25705542, Time: 0.0074 Steps: 87370, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000091, Sample Num: 1456, Cur Loss: 0.45505208, Cur Avg Loss: 0.44046697, Log Avg loss: 0.48149270, Global Avg Loss: 1.25696666, Time: 0.0074 Steps: 87380, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000101, Sample Num: 1616, Cur Loss: 0.15960830, Cur Avg Loss: 0.42915618, Log Avg loss: 0.32622795, Global Avg Loss: 1.25686016, Time: 0.0073 Steps: 87390, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000111, Sample Num: 1776, Cur Loss: 0.50735980, Cur Avg Loss: 0.44242787, Log Avg loss: 0.57647196, Global Avg Loss: 1.25678231, Time: 0.0074 Steps: 87400, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000121, Sample Num: 1936, Cur Loss: 0.31511560, Cur Avg Loss: 0.43971086, Log Avg loss: 0.40955203, Global Avg Loss: 1.25668538, Time: 0.0077 Steps: 87410, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000131, Sample Num: 2096, Cur Loss: 0.31021768, Cur Avg Loss: 0.44826582, Log Avg loss: 0.55178080, Global Avg Loss: 1.25660475, Time: 0.0076 Steps: 87420, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000141, Sample Num: 2256, Cur Loss: 0.46697807, Cur Avg Loss: 0.44519591, Log Avg loss: 0.40498020, Global Avg Loss: 1.25650734, Time: 0.0075 Steps: 87430, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000151, Sample Num: 2416, Cur Loss: 0.25752133, Cur Avg Loss: 0.44106829, Log Avg loss: 0.38286873, Global Avg Loss: 1.25640743, Time: 0.0075 Steps: 87440, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000161, Sample Num: 2576, Cur Loss: 0.28517401, Cur Avg Loss: 0.44336103, Log Avg loss: 0.47798151, Global Avg Loss: 1.25631842, Time: 0.0075 Steps: 87450, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000171, Sample Num: 2736, Cur Loss: 0.46996731, Cur Avg Loss: 0.44490218, Log Avg loss: 0.46971471, Global Avg Loss: 1.25622848, Time: 0.0075 Steps: 87460, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000181, Sample Num: 2896, Cur Loss: 0.96284389, Cur Avg Loss: 0.44510795, Log Avg loss: 0.44862652, Global Avg Loss: 1.25613615, Time: 0.0076 Steps: 87470, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000191, Sample Num: 3056, Cur Loss: 0.95325017, Cur Avg Loss: 0.44853287, Log Avg loss: 0.51052400, Global Avg Loss: 1.25605092, Time: 0.0076 Steps: 87480, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000201, Sample Num: 3216, Cur Loss: 1.07933223, Cur Avg Loss: 0.45766612, Log Avg loss: 0.63211123, Global Avg Loss: 1.25597960, Time: 0.0076 Steps: 87490, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000211, Sample Num: 3376, Cur Loss: 0.40627399, Cur Avg Loss: 0.46600174, Log Avg loss: 0.63354771, Global Avg Loss: 1.25590847, Time: 0.0075 Steps: 87500, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000221, Sample Num: 3536, Cur Loss: 0.22700599, Cur Avg Loss: 0.46228909, Log Avg loss: 0.38395200, Global Avg Loss: 1.25580882, Time: 0.0075 Steps: 87510, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000231, Sample Num: 3696, Cur Loss: 0.30651748, Cur Avg Loss: 0.46348470, Log Avg loss: 0.48990773, Global Avg Loss: 1.25572131, Time: 0.0076 Steps: 87520, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000241, Sample Num: 3856, Cur Loss: 1.09670973, Cur Avg Loss: 0.46256798, Log Avg loss: 0.44139192, Global Avg Loss: 1.25562828, Time: 0.0075 Steps: 87530, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000251, Sample Num: 4016, Cur Loss: 0.85125828, Cur Avg Loss: 0.46001468, Log Avg loss: 0.39848004, Global Avg Loss: 1.25553036, Time: 0.0076 Steps: 87540, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000261, Sample Num: 4176, Cur Loss: 0.22410172, Cur Avg Loss: 0.45754728, Log Avg loss: 0.39561559, Global Avg Loss: 1.25543214, Time: 0.0076 Steps: 87550, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000271, Sample Num: 4336, Cur Loss: 0.28707218, Cur Avg Loss: 0.45921345, Log Avg loss: 0.50270033, Global Avg Loss: 1.25534618, Time: 0.0075 Steps: 87560, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000281, Sample Num: 4496, Cur Loss: 0.41711792, Cur Avg Loss: 0.46033752, Log Avg loss: 0.49079980, Global Avg Loss: 1.25525887, Time: 0.0076 Steps: 87570, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000291, Sample Num: 4656, Cur Loss: 0.45035845, Cur Avg Loss: 0.45914362, Log Avg loss: 0.42559520, Global Avg Loss: 1.25516414, Time: 0.0075 Steps: 87580, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000301, Sample Num: 4816, Cur Loss: 0.39387611, Cur Avg Loss: 0.45564375, Log Avg loss: 0.35379757, Global Avg Loss: 1.25506123, Time: 0.0075 Steps: 87590, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000311, Sample Num: 4976, Cur Loss: 0.58759093, Cur Avg Loss: 0.45849819, Log Avg loss: 0.54441668, Global Avg Loss: 1.25498011, Time: 0.0076 Steps: 87600, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000321, Sample Num: 5136, Cur Loss: 0.48171642, Cur Avg Loss: 0.45833285, Log Avg loss: 0.45319093, Global Avg Loss: 1.25488859, Time: 0.0076 Steps: 87610, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000331, Sample Num: 5296, Cur Loss: 0.34869859, Cur Avg Loss: 0.45940369, Log Avg loss: 0.49377752, Global Avg Loss: 1.25480172, Time: 0.0075 Steps: 87620, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000341, Sample Num: 5456, Cur Loss: 0.50882447, Cur Avg Loss: 0.45899393, Log Avg loss: 0.44543083, Global Avg Loss: 1.25470936, Time: 0.0076 Steps: 87630, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000351, Sample Num: 5616, Cur Loss: 0.34638643, Cur Avg Loss: 0.45904548, Log Avg loss: 0.46080330, Global Avg Loss: 1.25461877, Time: 0.0075 Steps: 87640, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000361, Sample Num: 5776, Cur Loss: 0.61830848, Cur Avg Loss: 0.46008009, Log Avg loss: 0.49639511, Global Avg Loss: 1.25453227, Time: 0.0075 Steps: 87650, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000371, Sample Num: 5936, Cur Loss: 0.64097631, Cur Avg Loss: 0.46487983, Log Avg loss: 0.63815040, Global Avg Loss: 1.25446195, Time: 0.0076 Steps: 87660, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000381, Sample Num: 6096, Cur Loss: 0.74063826, Cur Avg Loss: 0.46529644, Log Avg loss: 0.48075260, Global Avg Loss: 1.25437370, Time: 0.0077 Steps: 87670, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000391, Sample Num: 6256, Cur Loss: 0.70066398, Cur Avg Loss: 0.46773145, Log Avg loss: 0.56050542, Global Avg Loss: 1.25429456, Time: 0.0075 Steps: 87680, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000401, Sample Num: 6416, Cur Loss: 0.43821311, Cur Avg Loss: 0.46701587, Log Avg loss: 0.43903648, Global Avg Loss: 1.25420159, Time: 0.0075 Steps: 87690, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000411, Sample Num: 6576, Cur Loss: 0.12094606, Cur Avg Loss: 0.46825791, Log Avg loss: 0.51806376, Global Avg Loss: 1.25411765, Time: 0.0076 Steps: 87700, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000421, Sample Num: 6736, Cur Loss: 0.63960850, Cur Avg Loss: 0.46655865, Log Avg loss: 0.39671920, Global Avg Loss: 1.25401990, Time: 0.0075 Steps: 87710, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000431, Sample Num: 6896, Cur Loss: 0.66198236, Cur Avg Loss: 0.46338414, Log Avg loss: 0.32973716, Global Avg Loss: 1.25391453, Time: 0.0075 Steps: 87720, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000441, Sample Num: 7056, Cur Loss: 0.24317205, Cur Avg Loss: 0.46205345, Log Avg loss: 0.40470099, Global Avg Loss: 1.25381774, Time: 0.0075 Steps: 87730, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000451, Sample Num: 7216, Cur Loss: 0.24208036, Cur Avg Loss: 0.46357323, Log Avg loss: 0.53059539, Global Avg Loss: 1.25373531, Time: 0.0075 Steps: 87740, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000461, Sample Num: 7376, Cur Loss: 0.18271762, Cur Avg Loss: 0.46274547, Log Avg loss: 0.42541326, Global Avg Loss: 1.25364091, Time: 0.0075 Steps: 87750, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000471, Sample Num: 7536, Cur Loss: 0.17888817, Cur Avg Loss: 0.46157873, Log Avg loss: 0.40779207, Global Avg Loss: 1.25354453, Time: 0.0076 Steps: 87760, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000481, Sample Num: 7696, Cur Loss: 1.43599510, Cur Avg Loss: 0.46676661, Log Avg loss: 0.71111583, Global Avg Loss: 1.25348273, Time: 0.0075 Steps: 87770, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000491, Sample Num: 7856, Cur Loss: 0.31208956, Cur Avg Loss: 0.46439787, Log Avg loss: 0.35046169, Global Avg Loss: 1.25337986, Time: 0.0075 Steps: 87780, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000501, Sample Num: 8016, Cur Loss: 0.50901955, Cur Avg Loss: 0.46434925, Log Avg loss: 0.46196175, Global Avg Loss: 1.25328971, Time: 0.0075 Steps: 87790, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000511, Sample Num: 8176, Cur Loss: 0.70274734, Cur Avg Loss: 0.46618926, Log Avg loss: 0.55837397, Global Avg Loss: 1.25321056, Time: 0.0075 Steps: 87800, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000521, Sample Num: 8336, Cur Loss: 0.54059958, Cur Avg Loss: 0.46751014, Log Avg loss: 0.53500720, Global Avg Loss: 1.25312877, Time: 0.0075 Steps: 87810, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000531, Sample Num: 8496, Cur Loss: 0.39136213, Cur Avg Loss: 0.46611357, Log Avg loss: 0.39335217, Global Avg Loss: 1.25303087, Time: 0.0074 Steps: 87820, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000541, Sample Num: 8656, Cur Loss: 0.87379456, Cur Avg Loss: 0.46659286, Log Avg loss: 0.49204319, Global Avg Loss: 1.25294422, Time: 0.0075 Steps: 87830, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000551, Sample Num: 8816, Cur Loss: 0.30971673, Cur Avg Loss: 0.46553635, Log Avg loss: 0.40837894, Global Avg Loss: 1.25284807, Time: 0.0074 Steps: 87840, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000561, Sample Num: 8976, Cur Loss: 0.35039416, Cur Avg Loss: 0.46512411, Log Avg loss: 0.44240992, Global Avg Loss: 1.25275582, Time: 0.0074 Steps: 87850, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000571, Sample Num: 9136, Cur Loss: 0.53456217, Cur Avg Loss: 0.46307145, Log Avg loss: 0.34791711, Global Avg Loss: 1.25265284, Time: 0.0074 Steps: 87860, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000581, Sample Num: 9296, Cur Loss: 0.73524189, Cur Avg Loss: 0.46203976, Log Avg loss: 0.40313005, Global Avg Loss: 1.25255616, Time: 0.0074 Steps: 87870, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000591, Sample Num: 9456, Cur Loss: 0.22677401, Cur Avg Loss: 0.46358763, Log Avg loss: 0.55351882, Global Avg Loss: 1.25247661, Time: 0.0075 Steps: 87880, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000601, Sample Num: 9616, Cur Loss: 0.53812855, Cur Avg Loss: 0.46594257, Log Avg loss: 0.60511945, Global Avg Loss: 1.25240296, Time: 0.0074 Steps: 87890, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000611, Sample Num: 9776, Cur Loss: 0.21750136, Cur Avg Loss: 0.46616575, Log Avg loss: 0.47957917, Global Avg Loss: 1.25231504, Time: 0.0074 Steps: 87900, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000621, Sample Num: 9936, Cur Loss: 0.24215287, Cur Avg Loss: 0.46815830, Log Avg loss: 0.58990306, Global Avg Loss: 1.25223968, Time: 0.0074 Steps: 87910, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000631, Sample Num: 10096, Cur Loss: 0.32410374, Cur Avg Loss: 0.46719319, Log Avg loss: 0.40725973, Global Avg Loss: 1.25214358, Time: 0.0074 Steps: 87920, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000641, Sample Num: 10256, Cur Loss: 0.53959775, Cur Avg Loss: 0.46671724, Log Avg loss: 0.43668516, Global Avg Loss: 1.25205084, Time: 0.0074 Steps: 87930, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000651, Sample Num: 10416, Cur Loss: 0.37070388, Cur Avg Loss: 0.46718743, Log Avg loss: 0.49732661, Global Avg Loss: 1.25196501, Time: 0.0075 Steps: 87940, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000661, Sample Num: 10576, Cur Loss: 0.60136724, Cur Avg Loss: 0.46638579, Log Avg loss: 0.41419874, Global Avg Loss: 1.25186976, Time: 0.0074 Steps: 87950, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000671, Sample Num: 10736, Cur Loss: 0.29713017, Cur Avg Loss: 0.46559062, Log Avg loss: 0.41302976, Global Avg Loss: 1.25177439, Time: 0.0252 Steps: 87960, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000681, Sample Num: 10896, Cur Loss: 0.14699255, Cur Avg Loss: 0.46650521, Log Avg loss: 0.52787422, Global Avg Loss: 1.25169210, Time: 0.0096 Steps: 87970, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000691, Sample Num: 11056, Cur Loss: 0.33540004, Cur Avg Loss: 0.46500740, Log Avg loss: 0.36300667, Global Avg Loss: 1.25159109, Time: 0.0065 Steps: 87980, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000701, Sample Num: 11216, Cur Loss: 0.13243002, Cur Avg Loss: 0.46374739, Log Avg loss: 0.37668062, Global Avg Loss: 1.25149166, Time: 0.0064 Steps: 87990, Updated lr: 0.000018 Training, Epoch: 0042, Batch: 000711, Sample Num: 11376, Cur Loss: 0.49872029, Cur Avg Loss: 0.46324501, Log Avg loss: 0.42802848, Global Avg Loss: 1.25139809, Time: 0.0074 Steps: 88000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000721, Sample Num: 11536, Cur Loss: 0.63193262, Cur Avg Loss: 0.46329679, Log Avg loss: 0.46697853, Global Avg Loss: 1.25130896, Time: 0.0065 Steps: 88010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000731, Sample Num: 11696, Cur Loss: 0.17780998, Cur Avg Loss: 0.46228061, Log Avg loss: 0.38901371, Global Avg Loss: 1.25121099, Time: 0.0075 Steps: 88020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000741, Sample Num: 11856, Cur Loss: 0.98542178, Cur Avg Loss: 0.46373103, Log Avg loss: 0.56975706, Global Avg Loss: 1.25113358, Time: 0.0064 Steps: 88030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000751, Sample Num: 12016, Cur Loss: 0.79498535, Cur Avg Loss: 0.46451012, Log Avg loss: 0.52224052, Global Avg Loss: 1.25105079, Time: 0.0068 Steps: 88040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000761, Sample Num: 12176, Cur Loss: 0.54179054, Cur Avg Loss: 0.46338939, Log Avg loss: 0.37922202, Global Avg Loss: 1.25095177, Time: 0.0064 Steps: 88050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000771, Sample Num: 12336, Cur Loss: 0.18799101, Cur Avg Loss: 0.46361390, Log Avg loss: 0.48069948, Global Avg Loss: 1.25086430, Time: 0.0191 Steps: 88060, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000781, Sample Num: 12496, Cur Loss: 0.95190400, Cur Avg Loss: 0.46392226, Log Avg loss: 0.48769710, Global Avg Loss: 1.25077765, Time: 0.0228 Steps: 88070, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000791, Sample Num: 12656, Cur Loss: 0.39996502, Cur Avg Loss: 0.46321448, Log Avg loss: 0.40793658, Global Avg Loss: 1.25068196, Time: 0.0064 Steps: 88080, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000801, Sample Num: 12816, Cur Loss: 0.62550944, Cur Avg Loss: 0.46419485, Log Avg loss: 0.54174247, Global Avg Loss: 1.25060148, Time: 0.0065 Steps: 88090, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000811, Sample Num: 12976, Cur Loss: 0.18145207, Cur Avg Loss: 0.46312058, Log Avg loss: 0.37707116, Global Avg Loss: 1.25050233, Time: 0.0073 Steps: 88100, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000821, Sample Num: 13136, Cur Loss: 0.54533082, Cur Avg Loss: 0.46076722, Log Avg loss: 0.26990961, Global Avg Loss: 1.25039104, Time: 0.0065 Steps: 88110, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000831, Sample Num: 13296, Cur Loss: 0.30248475, Cur Avg Loss: 0.46123507, Log Avg loss: 0.49964579, Global Avg Loss: 1.25030584, Time: 0.0076 Steps: 88120, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000841, Sample Num: 13456, Cur Loss: 0.64368868, Cur Avg Loss: 0.46283511, Log Avg loss: 0.59579822, Global Avg Loss: 1.25023157, Time: 0.0064 Steps: 88130, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000851, Sample Num: 13616, Cur Loss: 0.20717026, Cur Avg Loss: 0.46142099, Log Avg loss: 0.34249382, Global Avg Loss: 1.25012859, Time: 0.0164 Steps: 88140, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000861, Sample Num: 13776, Cur Loss: 0.56878823, Cur Avg Loss: 0.46165996, Log Avg loss: 0.48199574, Global Avg Loss: 1.25004145, Time: 0.0208 Steps: 88150, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000871, Sample Num: 13936, Cur Loss: 0.15582246, Cur Avg Loss: 0.46069079, Log Avg loss: 0.37724557, Global Avg Loss: 1.24994245, Time: 0.0129 Steps: 88160, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000881, Sample Num: 14096, Cur Loss: 0.70304787, Cur Avg Loss: 0.46153822, Log Avg loss: 0.53534913, Global Avg Loss: 1.24986140, Time: 0.0064 Steps: 88170, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000891, Sample Num: 14256, Cur Loss: 0.55515045, Cur Avg Loss: 0.46325044, Log Avg loss: 0.61409726, Global Avg Loss: 1.24978930, Time: 0.0078 Steps: 88180, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000901, Sample Num: 14416, Cur Loss: 1.09312367, Cur Avg Loss: 0.46354023, Log Avg loss: 0.48936052, Global Avg Loss: 1.24970307, Time: 0.0064 Steps: 88190, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000911, Sample Num: 14576, Cur Loss: 0.15581426, Cur Avg Loss: 0.46462826, Log Avg loss: 0.56265973, Global Avg Loss: 1.24962518, Time: 0.0064 Steps: 88200, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000921, Sample Num: 14736, Cur Loss: 0.32117894, Cur Avg Loss: 0.46465702, Log Avg loss: 0.46727707, Global Avg Loss: 1.24953649, Time: 0.0066 Steps: 88210, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000931, Sample Num: 14896, Cur Loss: 0.62744343, Cur Avg Loss: 0.46545839, Log Avg loss: 0.53926504, Global Avg Loss: 1.24945597, Time: 0.0064 Steps: 88220, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000941, Sample Num: 15056, Cur Loss: 0.31944260, Cur Avg Loss: 0.46501411, Log Avg loss: 0.42365155, Global Avg Loss: 1.24936238, Time: 0.0064 Steps: 88230, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000951, Sample Num: 15216, Cur Loss: 0.56606215, Cur Avg Loss: 0.46694948, Log Avg loss: 0.64906763, Global Avg Loss: 1.24929435, Time: 0.0064 Steps: 88240, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000961, Sample Num: 15376, Cur Loss: 0.21493953, Cur Avg Loss: 0.46768865, Log Avg loss: 0.53798308, Global Avg Loss: 1.24921375, Time: 0.0065 Steps: 88250, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000971, Sample Num: 15536, Cur Loss: 1.67691982, Cur Avg Loss: 0.46901772, Log Avg loss: 0.59674142, Global Avg Loss: 1.24913982, Time: 0.0074 Steps: 88260, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000981, Sample Num: 15696, Cur Loss: 0.44925636, Cur Avg Loss: 0.47019049, Log Avg loss: 0.58406660, Global Avg Loss: 1.24906447, Time: 0.0075 Steps: 88270, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 000991, Sample Num: 15856, Cur Loss: 0.27755862, Cur Avg Loss: 0.47089541, Log Avg loss: 0.54004848, Global Avg Loss: 1.24898416, Time: 0.0073 Steps: 88280, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001001, Sample Num: 16016, Cur Loss: 0.55767506, Cur Avg Loss: 0.47051741, Log Avg loss: 0.43305787, Global Avg Loss: 1.24889175, Time: 0.0073 Steps: 88290, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001011, Sample Num: 16176, Cur Loss: 0.30185345, Cur Avg Loss: 0.47136713, Log Avg loss: 0.55642344, Global Avg Loss: 1.24881332, Time: 0.0073 Steps: 88300, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001021, Sample Num: 16336, Cur Loss: 0.29933256, Cur Avg Loss: 0.47215980, Log Avg loss: 0.55229897, Global Avg Loss: 1.24873445, Time: 0.0073 Steps: 88310, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001031, Sample Num: 16496, Cur Loss: 0.30288053, Cur Avg Loss: 0.47200933, Log Avg loss: 0.45664632, Global Avg Loss: 1.24864477, Time: 0.0073 Steps: 88320, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001041, Sample Num: 16656, Cur Loss: 0.68003047, Cur Avg Loss: 0.47219596, Log Avg loss: 0.49143753, Global Avg Loss: 1.24855904, Time: 0.0073 Steps: 88330, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001051, Sample Num: 16816, Cur Loss: 0.55893749, Cur Avg Loss: 0.47291674, Log Avg loss: 0.54794984, Global Avg Loss: 1.24847974, Time: 0.0074 Steps: 88340, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001061, Sample Num: 16976, Cur Loss: 0.41125095, Cur Avg Loss: 0.47133169, Log Avg loss: 0.30474321, Global Avg Loss: 1.24837292, Time: 0.0074 Steps: 88350, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001071, Sample Num: 17136, Cur Loss: 0.74813342, Cur Avg Loss: 0.47168671, Log Avg loss: 0.50935399, Global Avg Loss: 1.24828928, Time: 0.0076 Steps: 88360, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001081, Sample Num: 17296, Cur Loss: 0.36621761, Cur Avg Loss: 0.47023121, Log Avg loss: 0.31434742, Global Avg Loss: 1.24818359, Time: 0.0075 Steps: 88370, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001091, Sample Num: 17456, Cur Loss: 0.37467879, Cur Avg Loss: 0.46952437, Log Avg loss: 0.39311511, Global Avg Loss: 1.24808685, Time: 0.0076 Steps: 88380, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001101, Sample Num: 17616, Cur Loss: 0.90170395, Cur Avg Loss: 0.46948698, Log Avg loss: 0.46540750, Global Avg Loss: 1.24799830, Time: 0.0076 Steps: 88390, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001111, Sample Num: 17776, Cur Loss: 0.46882039, Cur Avg Loss: 0.46916446, Log Avg loss: 0.43365476, Global Avg Loss: 1.24790618, Time: 0.0077 Steps: 88400, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001121, Sample Num: 17936, Cur Loss: 0.28596383, Cur Avg Loss: 0.46866810, Log Avg loss: 0.41352322, Global Avg Loss: 1.24781180, Time: 0.0076 Steps: 88410, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001131, Sample Num: 18096, Cur Loss: 0.22791080, Cur Avg Loss: 0.46822215, Log Avg loss: 0.41823058, Global Avg Loss: 1.24771798, Time: 0.0076 Steps: 88420, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001141, Sample Num: 18256, Cur Loss: 0.39778286, Cur Avg Loss: 0.46789450, Log Avg loss: 0.43083696, Global Avg Loss: 1.24762560, Time: 0.0076 Steps: 88430, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001151, Sample Num: 18416, Cur Loss: 0.21019909, Cur Avg Loss: 0.46849111, Log Avg loss: 0.53656466, Global Avg Loss: 1.24754520, Time: 0.0076 Steps: 88440, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001161, Sample Num: 18576, Cur Loss: 0.19371572, Cur Avg Loss: 0.46905510, Log Avg loss: 0.53396998, Global Avg Loss: 1.24746453, Time: 0.0077 Steps: 88450, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001171, Sample Num: 18736, Cur Loss: 0.97963858, Cur Avg Loss: 0.47015376, Log Avg loss: 0.59770906, Global Avg Loss: 1.24739107, Time: 0.0076 Steps: 88460, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001181, Sample Num: 18896, Cur Loss: 0.26585776, Cur Avg Loss: 0.47022838, Log Avg loss: 0.47896605, Global Avg Loss: 1.24730422, Time: 0.0076 Steps: 88470, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001191, Sample Num: 19056, Cur Loss: 0.28583178, Cur Avg Loss: 0.47015346, Log Avg loss: 0.46130521, Global Avg Loss: 1.24721538, Time: 0.0075 Steps: 88480, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001201, Sample Num: 19216, Cur Loss: 0.33908927, Cur Avg Loss: 0.47005339, Log Avg loss: 0.45813565, Global Avg Loss: 1.24712621, Time: 0.0075 Steps: 88490, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001211, Sample Num: 19376, Cur Loss: 0.98350215, Cur Avg Loss: 0.47096371, Log Avg loss: 0.58029209, Global Avg Loss: 1.24705086, Time: 0.0075 Steps: 88500, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001221, Sample Num: 19536, Cur Loss: 0.70729411, Cur Avg Loss: 0.47185485, Log Avg loss: 0.57977256, Global Avg Loss: 1.24697547, Time: 0.0076 Steps: 88510, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001231, Sample Num: 19696, Cur Loss: 0.33702487, Cur Avg Loss: 0.47301841, Log Avg loss: 0.61508852, Global Avg Loss: 1.24690409, Time: 0.0075 Steps: 88520, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001241, Sample Num: 19856, Cur Loss: 0.30540410, Cur Avg Loss: 0.47463497, Log Avg loss: 0.67363400, Global Avg Loss: 1.24683934, Time: 0.0076 Steps: 88530, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001251, Sample Num: 20016, Cur Loss: 0.29027927, Cur Avg Loss: 0.47373511, Log Avg loss: 0.36206243, Global Avg Loss: 1.24673941, Time: 0.0077 Steps: 88540, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001261, Sample Num: 20176, Cur Loss: 0.38507339, Cur Avg Loss: 0.47430862, Log Avg loss: 0.54605435, Global Avg Loss: 1.24666028, Time: 0.0076 Steps: 88550, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001271, Sample Num: 20336, Cur Loss: 0.24464765, Cur Avg Loss: 0.47444147, Log Avg loss: 0.49119396, Global Avg Loss: 1.24657497, Time: 0.0075 Steps: 88560, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001281, Sample Num: 20496, Cur Loss: 0.29309398, Cur Avg Loss: 0.47315525, Log Avg loss: 0.30967677, Global Avg Loss: 1.24646919, Time: 0.0076 Steps: 88570, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001291, Sample Num: 20656, Cur Loss: 0.22379726, Cur Avg Loss: 0.47313345, Log Avg loss: 0.47034065, Global Avg Loss: 1.24638157, Time: 0.0076 Steps: 88580, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001301, Sample Num: 20816, Cur Loss: 0.31879526, Cur Avg Loss: 0.47287406, Log Avg loss: 0.43938755, Global Avg Loss: 1.24629048, Time: 0.0074 Steps: 88590, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001311, Sample Num: 20976, Cur Loss: 0.22666863, Cur Avg Loss: 0.47277542, Log Avg loss: 0.45994153, Global Avg Loss: 1.24620173, Time: 0.0074 Steps: 88600, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001321, Sample Num: 21136, Cur Loss: 0.63045794, Cur Avg Loss: 0.47330006, Log Avg loss: 0.54208150, Global Avg Loss: 1.24612226, Time: 0.0074 Steps: 88610, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001331, Sample Num: 21296, Cur Loss: 0.48856729, Cur Avg Loss: 0.47365367, Log Avg loss: 0.52036510, Global Avg Loss: 1.24604037, Time: 0.0074 Steps: 88620, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001341, Sample Num: 21456, Cur Loss: 0.88111645, Cur Avg Loss: 0.47401408, Log Avg loss: 0.52198465, Global Avg Loss: 1.24595867, Time: 0.0075 Steps: 88630, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001351, Sample Num: 21616, Cur Loss: 0.58801031, Cur Avg Loss: 0.47363945, Log Avg loss: 0.42340078, Global Avg Loss: 1.24586588, Time: 0.0075 Steps: 88640, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001361, Sample Num: 21776, Cur Loss: 0.27537483, Cur Avg Loss: 0.47386435, Log Avg loss: 0.50424833, Global Avg Loss: 1.24578222, Time: 0.0076 Steps: 88650, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001371, Sample Num: 21936, Cur Loss: 0.46199805, Cur Avg Loss: 0.47433347, Log Avg loss: 0.53818120, Global Avg Loss: 1.24570241, Time: 0.0076 Steps: 88660, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001381, Sample Num: 22096, Cur Loss: 0.55344152, Cur Avg Loss: 0.47423724, Log Avg loss: 0.46104369, Global Avg Loss: 1.24561392, Time: 0.0076 Steps: 88670, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001391, Sample Num: 22256, Cur Loss: 0.15940800, Cur Avg Loss: 0.47458480, Log Avg loss: 0.52258393, Global Avg Loss: 1.24553238, Time: 0.0076 Steps: 88680, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001401, Sample Num: 22416, Cur Loss: 0.16282251, Cur Avg Loss: 0.47369257, Log Avg loss: 0.34958288, Global Avg Loss: 1.24543136, Time: 0.0076 Steps: 88690, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001411, Sample Num: 22576, Cur Loss: 0.33297920, Cur Avg Loss: 0.47391297, Log Avg loss: 0.50479029, Global Avg Loss: 1.24534786, Time: 0.0075 Steps: 88700, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001421, Sample Num: 22736, Cur Loss: 0.28361481, Cur Avg Loss: 0.47413583, Log Avg loss: 0.50558202, Global Avg Loss: 1.24526447, Time: 0.0077 Steps: 88710, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001431, Sample Num: 22896, Cur Loss: 0.23089509, Cur Avg Loss: 0.47316451, Log Avg loss: 0.33513998, Global Avg Loss: 1.24516189, Time: 0.0075 Steps: 88720, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001441, Sample Num: 23056, Cur Loss: 0.51645744, Cur Avg Loss: 0.47385446, Log Avg loss: 0.57258591, Global Avg Loss: 1.24508609, Time: 0.0075 Steps: 88730, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001451, Sample Num: 23216, Cur Loss: 0.28739920, Cur Avg Loss: 0.47386562, Log Avg loss: 0.47547369, Global Avg Loss: 1.24499936, Time: 0.0075 Steps: 88740, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001461, Sample Num: 23376, Cur Loss: 0.09104468, Cur Avg Loss: 0.47316787, Log Avg loss: 0.37192505, Global Avg Loss: 1.24490099, Time: 0.0075 Steps: 88750, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001471, Sample Num: 23536, Cur Loss: 0.49104434, Cur Avg Loss: 0.47341779, Log Avg loss: 0.50993126, Global Avg Loss: 1.24481818, Time: 0.0076 Steps: 88760, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001481, Sample Num: 23696, Cur Loss: 0.47021616, Cur Avg Loss: 0.47339356, Log Avg loss: 0.46982935, Global Avg Loss: 1.24473088, Time: 0.0077 Steps: 88770, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001491, Sample Num: 23856, Cur Loss: 0.74870181, Cur Avg Loss: 0.47368101, Log Avg loss: 0.51625169, Global Avg Loss: 1.24464883, Time: 0.0075 Steps: 88780, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001501, Sample Num: 24016, Cur Loss: 0.30930510, Cur Avg Loss: 0.47395780, Log Avg loss: 0.51522797, Global Avg Loss: 1.24456667, Time: 0.0075 Steps: 88790, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001511, Sample Num: 24176, Cur Loss: 0.66568476, Cur Avg Loss: 0.47476792, Log Avg loss: 0.59636612, Global Avg Loss: 1.24449368, Time: 0.0077 Steps: 88800, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001521, Sample Num: 24336, Cur Loss: 0.58003408, Cur Avg Loss: 0.47478945, Log Avg loss: 0.47804286, Global Avg Loss: 1.24440738, Time: 0.0075 Steps: 88810, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001531, Sample Num: 24496, Cur Loss: 0.26723987, Cur Avg Loss: 0.47492152, Log Avg loss: 0.49500897, Global Avg Loss: 1.24432300, Time: 0.0075 Steps: 88820, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001541, Sample Num: 24656, Cur Loss: 0.52228034, Cur Avg Loss: 0.47554948, Log Avg loss: 0.57169064, Global Avg Loss: 1.24424728, Time: 0.0074 Steps: 88830, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001551, Sample Num: 24816, Cur Loss: 0.10398254, Cur Avg Loss: 0.47609363, Log Avg loss: 0.55994750, Global Avg Loss: 1.24417026, Time: 0.0074 Steps: 88840, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001561, Sample Num: 24976, Cur Loss: 0.23156592, Cur Avg Loss: 0.47628238, Log Avg loss: 0.50555710, Global Avg Loss: 1.24408713, Time: 0.0074 Steps: 88850, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001571, Sample Num: 25136, Cur Loss: 0.40607783, Cur Avg Loss: 0.47662998, Log Avg loss: 0.53088999, Global Avg Loss: 1.24400687, Time: 0.0080 Steps: 88860, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001581, Sample Num: 25296, Cur Loss: 0.38601667, Cur Avg Loss: 0.47673154, Log Avg loss: 0.49268767, Global Avg Loss: 1.24392232, Time: 0.0065 Steps: 88870, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001591, Sample Num: 25456, Cur Loss: 0.22447173, Cur Avg Loss: 0.47598743, Log Avg loss: 0.35834223, Global Avg Loss: 1.24382269, Time: 0.0214 Steps: 88880, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001601, Sample Num: 25616, Cur Loss: 0.47966430, Cur Avg Loss: 0.47569254, Log Avg loss: 0.42877548, Global Avg Loss: 1.24373099, Time: 0.0065 Steps: 88890, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001611, Sample Num: 25776, Cur Loss: 0.10642292, Cur Avg Loss: 0.47522355, Log Avg loss: 0.40013893, Global Avg Loss: 1.24363610, Time: 0.0085 Steps: 88900, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001621, Sample Num: 25936, Cur Loss: 0.16764307, Cur Avg Loss: 0.47456964, Log Avg loss: 0.36922482, Global Avg Loss: 1.24353775, Time: 0.0064 Steps: 88910, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001631, Sample Num: 26096, Cur Loss: 0.21977191, Cur Avg Loss: 0.47499333, Log Avg loss: 0.54367425, Global Avg Loss: 1.24345905, Time: 0.0065 Steps: 88920, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001641, Sample Num: 26256, Cur Loss: 0.41765195, Cur Avg Loss: 0.47436140, Log Avg loss: 0.37129322, Global Avg Loss: 1.24336097, Time: 0.0239 Steps: 88930, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001651, Sample Num: 26416, Cur Loss: 0.40779251, Cur Avg Loss: 0.47446936, Log Avg loss: 0.49218542, Global Avg Loss: 1.24327652, Time: 0.0064 Steps: 88940, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001661, Sample Num: 26576, Cur Loss: 0.29179186, Cur Avg Loss: 0.47472818, Log Avg loss: 0.51745872, Global Avg Loss: 1.24319492, Time: 0.0079 Steps: 88950, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001671, Sample Num: 26736, Cur Loss: 0.23425314, Cur Avg Loss: 0.47519527, Log Avg loss: 0.55277984, Global Avg Loss: 1.24311731, Time: 0.0065 Steps: 88960, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001681, Sample Num: 26896, Cur Loss: 0.16007382, Cur Avg Loss: 0.47479161, Log Avg loss: 0.40733926, Global Avg Loss: 1.24302337, Time: 0.0066 Steps: 88970, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001691, Sample Num: 27056, Cur Loss: 0.45399177, Cur Avg Loss: 0.47527606, Log Avg loss: 0.55671259, Global Avg Loss: 1.24294624, Time: 0.0073 Steps: 88980, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001701, Sample Num: 27216, Cur Loss: 0.77344310, Cur Avg Loss: 0.47604926, Log Avg loss: 0.60679697, Global Avg Loss: 1.24287475, Time: 0.0126 Steps: 88990, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001711, Sample Num: 27376, Cur Loss: 0.36646903, Cur Avg Loss: 0.47543766, Log Avg loss: 0.37140488, Global Avg Loss: 1.24277683, Time: 0.0229 Steps: 89000, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001721, Sample Num: 27536, Cur Loss: 0.71633828, Cur Avg Loss: 0.47555475, Log Avg loss: 0.49558889, Global Avg Loss: 1.24269289, Time: 0.0221 Steps: 89010, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001731, Sample Num: 27696, Cur Loss: 0.61659658, Cur Avg Loss: 0.47645933, Log Avg loss: 0.63213708, Global Avg Loss: 1.24262430, Time: 0.0064 Steps: 89020, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001741, Sample Num: 27856, Cur Loss: 0.29151118, Cur Avg Loss: 0.47680076, Log Avg loss: 0.53590277, Global Avg Loss: 1.24254492, Time: 0.0064 Steps: 89030, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001751, Sample Num: 28016, Cur Loss: 0.32045919, Cur Avg Loss: 0.47703679, Log Avg loss: 0.51812864, Global Avg Loss: 1.24246356, Time: 0.0230 Steps: 89040, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001761, Sample Num: 28176, Cur Loss: 0.41939485, Cur Avg Loss: 0.47690944, Log Avg loss: 0.45461086, Global Avg Loss: 1.24237509, Time: 0.0064 Steps: 89050, Updated lr: 0.000017 Training, Epoch: 0042, Batch: 001771, Sample Num: 28336, Cur Loss: 0.48435959, Cur Avg Loss: 0.47648376, Log Avg loss: 0.40152140, Global Avg Loss: 1.24228068, Time: 0.0065 Steps: 89060, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001781, Sample Num: 28496, Cur Loss: 1.08651638, Cur Avg Loss: 0.47628873, Log Avg loss: 0.44174880, Global Avg Loss: 1.24219080, Time: 0.0065 Steps: 89070, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001791, Sample Num: 28656, Cur Loss: 0.56867862, Cur Avg Loss: 0.47704969, Log Avg loss: 0.61257807, Global Avg Loss: 1.24212012, Time: 0.0064 Steps: 89080, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001801, Sample Num: 28816, Cur Loss: 0.22994475, Cur Avg Loss: 0.47736209, Log Avg loss: 0.53331190, Global Avg Loss: 1.24204056, Time: 0.0073 Steps: 89090, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001811, Sample Num: 28976, Cur Loss: 0.24936880, Cur Avg Loss: 0.47667353, Log Avg loss: 0.35266325, Global Avg Loss: 1.24194074, Time: 0.0063 Steps: 89100, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001821, Sample Num: 29136, Cur Loss: 0.64307415, Cur Avg Loss: 0.47676265, Log Avg loss: 0.49290247, Global Avg Loss: 1.24185668, Time: 0.0064 Steps: 89110, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001831, Sample Num: 29296, Cur Loss: 0.22568382, Cur Avg Loss: 0.47725739, Log Avg loss: 0.56734958, Global Avg Loss: 1.24178100, Time: 0.0065 Steps: 89120, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001841, Sample Num: 29456, Cur Loss: 1.01938486, Cur Avg Loss: 0.47691328, Log Avg loss: 0.41390669, Global Avg Loss: 1.24168811, Time: 0.0064 Steps: 89130, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001851, Sample Num: 29616, Cur Loss: 0.88310444, Cur Avg Loss: 0.47713452, Log Avg loss: 0.51786485, Global Avg Loss: 1.24160691, Time: 0.0064 Steps: 89140, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001861, Sample Num: 29776, Cur Loss: 0.94190669, Cur Avg Loss: 0.47676018, Log Avg loss: 0.40747012, Global Avg Loss: 1.24151335, Time: 0.0064 Steps: 89150, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001871, Sample Num: 29936, Cur Loss: 0.31746072, Cur Avg Loss: 0.47625570, Log Avg loss: 0.38237299, Global Avg Loss: 1.24141699, Time: 0.0064 Steps: 89160, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001881, Sample Num: 30096, Cur Loss: 0.33077118, Cur Avg Loss: 0.47646803, Log Avg loss: 0.51619433, Global Avg Loss: 1.24133566, Time: 0.0064 Steps: 89170, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001891, Sample Num: 30256, Cur Loss: 0.17228428, Cur Avg Loss: 0.47635071, Log Avg loss: 0.45428254, Global Avg Loss: 1.24124740, Time: 0.0065 Steps: 89180, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001901, Sample Num: 30416, Cur Loss: 0.44901544, Cur Avg Loss: 0.47661457, Log Avg loss: 0.52650982, Global Avg Loss: 1.24116727, Time: 0.0073 Steps: 89190, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001911, Sample Num: 30576, Cur Loss: 0.73353797, Cur Avg Loss: 0.47686481, Log Avg loss: 0.52443653, Global Avg Loss: 1.24108692, Time: 0.0074 Steps: 89200, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001921, Sample Num: 30736, Cur Loss: 0.86998796, Cur Avg Loss: 0.47643783, Log Avg loss: 0.39484099, Global Avg Loss: 1.24099206, Time: 0.0074 Steps: 89210, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001931, Sample Num: 30896, Cur Loss: 0.46325031, Cur Avg Loss: 0.47650014, Log Avg loss: 0.48847102, Global Avg Loss: 1.24090771, Time: 0.0073 Steps: 89220, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001941, Sample Num: 31056, Cur Loss: 0.32900617, Cur Avg Loss: 0.47619087, Log Avg loss: 0.41647003, Global Avg Loss: 1.24081532, Time: 0.0074 Steps: 89230, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001951, Sample Num: 31216, Cur Loss: 0.29724461, Cur Avg Loss: 0.47608328, Log Avg loss: 0.45520142, Global Avg Loss: 1.24072728, Time: 0.0073 Steps: 89240, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001961, Sample Num: 31376, Cur Loss: 0.54690087, Cur Avg Loss: 0.47563039, Log Avg loss: 0.38727048, Global Avg Loss: 1.24063166, Time: 0.0073 Steps: 89250, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001971, Sample Num: 31536, Cur Loss: 0.38986090, Cur Avg Loss: 0.47520372, Log Avg loss: 0.39153294, Global Avg Loss: 1.24053653, Time: 0.0074 Steps: 89260, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001981, Sample Num: 31696, Cur Loss: 0.19366087, Cur Avg Loss: 0.47526491, Log Avg loss: 0.48732606, Global Avg Loss: 1.24045216, Time: 0.0073 Steps: 89270, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 001991, Sample Num: 31856, Cur Loss: 0.99692178, Cur Avg Loss: 0.47539776, Log Avg loss: 0.50171540, Global Avg Loss: 1.24036941, Time: 0.0073 Steps: 89280, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002001, Sample Num: 32016, Cur Loss: 0.19032210, Cur Avg Loss: 0.47524755, Log Avg loss: 0.44534131, Global Avg Loss: 1.24028038, Time: 0.0074 Steps: 89290, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002011, Sample Num: 32176, Cur Loss: 0.42461836, Cur Avg Loss: 0.47572412, Log Avg loss: 0.57108578, Global Avg Loss: 1.24020544, Time: 0.0072 Steps: 89300, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002021, Sample Num: 32336, Cur Loss: 0.67053974, Cur Avg Loss: 0.47602151, Log Avg loss: 0.53582582, Global Avg Loss: 1.24012657, Time: 0.0074 Steps: 89310, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002031, Sample Num: 32496, Cur Loss: 0.90288913, Cur Avg Loss: 0.47588058, Log Avg loss: 0.44739898, Global Avg Loss: 1.24003782, Time: 0.0074 Steps: 89320, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002041, Sample Num: 32656, Cur Loss: 0.46034360, Cur Avg Loss: 0.47623092, Log Avg loss: 0.54738438, Global Avg Loss: 1.23996028, Time: 0.0074 Steps: 89330, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002051, Sample Num: 32816, Cur Loss: 0.55252689, Cur Avg Loss: 0.47661427, Log Avg loss: 0.55485777, Global Avg Loss: 1.23988359, Time: 0.0073 Steps: 89340, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002061, Sample Num: 32976, Cur Loss: 0.18380782, Cur Avg Loss: 0.47688266, Log Avg loss: 0.53192750, Global Avg Loss: 1.23980436, Time: 0.0074 Steps: 89350, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002071, Sample Num: 33136, Cur Loss: 0.38310152, Cur Avg Loss: 0.47641287, Log Avg loss: 0.37959060, Global Avg Loss: 1.23970810, Time: 0.0074 Steps: 89360, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002081, Sample Num: 33296, Cur Loss: 0.39841211, Cur Avg Loss: 0.47588771, Log Avg loss: 0.36712713, Global Avg Loss: 1.23961046, Time: 0.0076 Steps: 89370, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002091, Sample Num: 33456, Cur Loss: 0.23442015, Cur Avg Loss: 0.47603503, Log Avg loss: 0.50669191, Global Avg Loss: 1.23952846, Time: 0.0074 Steps: 89380, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002101, Sample Num: 33616, Cur Loss: 1.06888866, Cur Avg Loss: 0.47679366, Log Avg loss: 0.63542355, Global Avg Loss: 1.23946088, Time: 0.0073 Steps: 89390, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002111, Sample Num: 33776, Cur Loss: 0.46210438, Cur Avg Loss: 0.47653846, Log Avg loss: 0.42292011, Global Avg Loss: 1.23936954, Time: 0.0074 Steps: 89400, Updated lr: 0.000016 Training, Epoch: 0042, Batch: 002121, Sample Num: 33936, Cur Loss: 0.19846176, Cur Avg Loss: 0.47590690, Log Avg loss: 0.34258569, Global Avg Loss: 1.23926924, Time: 0.0074 Steps: 89410, Updated lr: 0.000016 ***** Running evaluation checkpoint-89418 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-89418 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.504706, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.599436, "eval_total_loss": 421.403352, "eval_mae": 0.60414, "eval_mse": 0.599538, "eval_r2": 0.618894, "eval_sp_statistic": 0.749712, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.797387, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.851238, "test_total_loss": 427.321271, "test_mae": 0.668373, "test_mse": 0.851449, "test_r2": 0.450467, "test_sp_statistic": 0.601112, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.672794, "test_ps_pvalue": 0.0, "lr": 1.6151730678046468e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2391857607143708, "train_cur_epoch_loss": 1011.8480155616999, "train_cur_epoch_avg_loss": 0.4752691477509159, "train_cur_epoch_time": 17.504706144332886, "train_cur_epoch_avg_time": 0.008222032007671623, "epoch": 42, "step": 89418} ################################################## Training, Epoch: 0043, Batch: 000002, Sample Num: 32, Cur Loss: 0.19212809, Cur Avg Loss: 0.21237849, Log Avg loss: 0.28742301, Global Avg Loss: 1.23916279, Time: 0.0076 Steps: 89420, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000012, Sample Num: 192, Cur Loss: 1.35316873, Cur Avg Loss: 0.47298826, Log Avg loss: 0.52511022, Global Avg Loss: 1.23908295, Time: 0.0074 Steps: 89430, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000022, Sample Num: 352, Cur Loss: 0.75449610, Cur Avg Loss: 0.44404484, Log Avg loss: 0.40931273, Global Avg Loss: 1.23899018, Time: 0.0074 Steps: 89440, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000032, Sample Num: 512, Cur Loss: 0.70897937, Cur Avg Loss: 0.42682582, Log Avg loss: 0.38894398, Global Avg Loss: 1.23889515, Time: 0.0073 Steps: 89450, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000042, Sample Num: 672, Cur Loss: 0.60891914, Cur Avg Loss: 0.41689142, Log Avg loss: 0.38510136, Global Avg Loss: 1.23879971, Time: 0.0075 Steps: 89460, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000052, Sample Num: 832, Cur Loss: 0.30835792, Cur Avg Loss: 0.41508298, Log Avg loss: 0.40748753, Global Avg Loss: 1.23870679, Time: 0.0075 Steps: 89470, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000062, Sample Num: 992, Cur Loss: 0.29263341, Cur Avg Loss: 0.42032918, Log Avg loss: 0.44760938, Global Avg Loss: 1.23861838, Time: 0.0074 Steps: 89480, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000072, Sample Num: 1152, Cur Loss: 0.31074262, Cur Avg Loss: 0.41541268, Log Avg loss: 0.38493039, Global Avg Loss: 1.23852299, Time: 0.0075 Steps: 89490, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000082, Sample Num: 1312, Cur Loss: 1.40512347, Cur Avg Loss: 0.42439804, Log Avg loss: 0.48909265, Global Avg Loss: 1.23843925, Time: 0.0076 Steps: 89500, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000092, Sample Num: 1472, Cur Loss: 0.58316171, Cur Avg Loss: 0.42331457, Log Avg loss: 0.41443009, Global Avg Loss: 1.23834719, Time: 0.0075 Steps: 89510, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000102, Sample Num: 1632, Cur Loss: 0.26428658, Cur Avg Loss: 0.43794682, Log Avg loss: 0.57256351, Global Avg Loss: 1.23827282, Time: 0.0074 Steps: 89520, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000112, Sample Num: 1792, Cur Loss: 0.25564662, Cur Avg Loss: 0.43454891, Log Avg loss: 0.39989028, Global Avg Loss: 1.23817918, Time: 0.0075 Steps: 89530, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000122, Sample Num: 1952, Cur Loss: 0.48723280, Cur Avg Loss: 0.44508466, Log Avg loss: 0.56308506, Global Avg Loss: 1.23810378, Time: 0.0073 Steps: 89540, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000132, Sample Num: 2112, Cur Loss: 0.13893841, Cur Avg Loss: 0.44636858, Log Avg loss: 0.46203241, Global Avg Loss: 1.23801712, Time: 0.0074 Steps: 89550, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000142, Sample Num: 2272, Cur Loss: 0.19464874, Cur Avg Loss: 0.44670316, Log Avg loss: 0.45111956, Global Avg Loss: 1.23792926, Time: 0.0074 Steps: 89560, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000152, Sample Num: 2432, Cur Loss: 1.08681822, Cur Avg Loss: 0.44935135, Log Avg loss: 0.48695569, Global Avg Loss: 1.23784541, Time: 0.0074 Steps: 89570, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000162, Sample Num: 2592, Cur Loss: 0.29291600, Cur Avg Loss: 0.45403354, Log Avg loss: 0.52520287, Global Avg Loss: 1.23776586, Time: 0.0074 Steps: 89580, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000172, Sample Num: 2752, Cur Loss: 0.53879285, Cur Avg Loss: 0.44984664, Log Avg loss: 0.38201878, Global Avg Loss: 1.23767034, Time: 0.0073 Steps: 89590, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000182, Sample Num: 2912, Cur Loss: 0.39044973, Cur Avg Loss: 0.45180422, Log Avg loss: 0.48547460, Global Avg Loss: 1.23758639, Time: 0.0073 Steps: 89600, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000192, Sample Num: 3072, Cur Loss: 0.34941119, Cur Avg Loss: 0.45702971, Log Avg loss: 0.55213358, Global Avg Loss: 1.23750990, Time: 0.0073 Steps: 89610, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000202, Sample Num: 3232, Cur Loss: 0.39553559, Cur Avg Loss: 0.46103143, Log Avg loss: 0.53786458, Global Avg Loss: 1.23743183, Time: 0.0075 Steps: 89620, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000212, Sample Num: 3392, Cur Loss: 0.35708964, Cur Avg Loss: 0.45461912, Log Avg loss: 0.32509051, Global Avg Loss: 1.23733004, Time: 0.0074 Steps: 89630, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000222, Sample Num: 3552, Cur Loss: 0.33689383, Cur Avg Loss: 0.45390845, Log Avg loss: 0.43884217, Global Avg Loss: 1.23724096, Time: 0.0074 Steps: 89640, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000232, Sample Num: 3712, Cur Loss: 0.34065473, Cur Avg Loss: 0.45246754, Log Avg loss: 0.42047940, Global Avg Loss: 1.23714986, Time: 0.0075 Steps: 89650, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000242, Sample Num: 3872, Cur Loss: 0.50458765, Cur Avg Loss: 0.45186564, Log Avg loss: 0.43790140, Global Avg Loss: 1.23706072, Time: 0.0074 Steps: 89660, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000252, Sample Num: 4032, Cur Loss: 0.32419434, Cur Avg Loss: 0.44910724, Log Avg loss: 0.38235409, Global Avg Loss: 1.23696540, Time: 0.0073 Steps: 89670, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000262, Sample Num: 4192, Cur Loss: 0.31663197, Cur Avg Loss: 0.44731806, Log Avg loss: 0.40223070, Global Avg Loss: 1.23687232, Time: 0.0073 Steps: 89680, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000272, Sample Num: 4352, Cur Loss: 0.66251242, Cur Avg Loss: 0.45644310, Log Avg loss: 0.69551905, Global Avg Loss: 1.23681196, Time: 0.0074 Steps: 89690, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000282, Sample Num: 4512, Cur Loss: 0.34618625, Cur Avg Loss: 0.45934297, Log Avg loss: 0.53821944, Global Avg Loss: 1.23673408, Time: 0.0074 Steps: 89700, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000292, Sample Num: 4672, Cur Loss: 0.22257480, Cur Avg Loss: 0.46184989, Log Avg loss: 0.53254515, Global Avg Loss: 1.23665558, Time: 0.0074 Steps: 89710, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000302, Sample Num: 4832, Cur Loss: 0.38688213, Cur Avg Loss: 0.46237508, Log Avg loss: 0.47771048, Global Avg Loss: 1.23657099, Time: 0.0074 Steps: 89720, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000312, Sample Num: 4992, Cur Loss: 0.32383490, Cur Avg Loss: 0.46145743, Log Avg loss: 0.43374437, Global Avg Loss: 1.23648152, Time: 0.0074 Steps: 89730, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000322, Sample Num: 5152, Cur Loss: 0.49260393, Cur Avg Loss: 0.46041696, Log Avg loss: 0.42795442, Global Avg Loss: 1.23639143, Time: 0.0075 Steps: 89740, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000332, Sample Num: 5312, Cur Loss: 1.00929677, Cur Avg Loss: 0.46091389, Log Avg loss: 0.47691510, Global Avg Loss: 1.23630681, Time: 0.0074 Steps: 89750, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000342, Sample Num: 5472, Cur Loss: 1.04896355, Cur Avg Loss: 0.46261023, Log Avg loss: 0.51892863, Global Avg Loss: 1.23622688, Time: 0.0074 Steps: 89760, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000352, Sample Num: 5632, Cur Loss: 0.28185657, Cur Avg Loss: 0.46099875, Log Avg loss: 0.40588615, Global Avg Loss: 1.23613439, Time: 0.0073 Steps: 89770, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000362, Sample Num: 5792, Cur Loss: 0.30140674, Cur Avg Loss: 0.46040470, Log Avg loss: 0.43949432, Global Avg Loss: 1.23604565, Time: 0.0074 Steps: 89780, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000372, Sample Num: 5952, Cur Loss: 0.54329348, Cur Avg Loss: 0.46535078, Log Avg loss: 0.64439875, Global Avg Loss: 1.23597976, Time: 0.0073 Steps: 89790, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000382, Sample Num: 6112, Cur Loss: 0.48148245, Cur Avg Loss: 0.46391333, Log Avg loss: 0.41044014, Global Avg Loss: 1.23588783, Time: 0.0076 Steps: 89800, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000392, Sample Num: 6272, Cur Loss: 0.62231427, Cur Avg Loss: 0.46354562, Log Avg loss: 0.44949906, Global Avg Loss: 1.23580027, Time: 0.0157 Steps: 89810, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000402, Sample Num: 6432, Cur Loss: 0.45981169, Cur Avg Loss: 0.46358923, Log Avg loss: 0.46529878, Global Avg Loss: 1.23571449, Time: 0.0204 Steps: 89820, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000412, Sample Num: 6592, Cur Loss: 0.48892686, Cur Avg Loss: 0.46244089, Log Avg loss: 0.41627758, Global Avg Loss: 1.23562327, Time: 0.0063 Steps: 89830, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000422, Sample Num: 6752, Cur Loss: 0.23210454, Cur Avg Loss: 0.45881440, Log Avg loss: 0.30940326, Global Avg Loss: 1.23552017, Time: 0.0224 Steps: 89840, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000432, Sample Num: 6912, Cur Loss: 0.66814286, Cur Avg Loss: 0.45677936, Log Avg loss: 0.37090060, Global Avg Loss: 1.23542394, Time: 0.0100 Steps: 89850, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000442, Sample Num: 7072, Cur Loss: 1.01962125, Cur Avg Loss: 0.45985983, Log Avg loss: 0.59293584, Global Avg Loss: 1.23535244, Time: 0.0123 Steps: 89860, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000452, Sample Num: 7232, Cur Loss: 0.20637831, Cur Avg Loss: 0.46113276, Log Avg loss: 0.51739643, Global Avg Loss: 1.23527255, Time: 0.0065 Steps: 89870, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000462, Sample Num: 7392, Cur Loss: 0.35012966, Cur Avg Loss: 0.46250764, Log Avg loss: 0.52465224, Global Avg Loss: 1.23519349, Time: 0.0114 Steps: 89880, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000472, Sample Num: 7552, Cur Loss: 0.16051987, Cur Avg Loss: 0.46292128, Log Avg loss: 0.48203150, Global Avg Loss: 1.23510970, Time: 0.0064 Steps: 89890, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000482, Sample Num: 7712, Cur Loss: 0.65525925, Cur Avg Loss: 0.46409774, Log Avg loss: 0.51962651, Global Avg Loss: 1.23503012, Time: 0.0065 Steps: 89900, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000492, Sample Num: 7872, Cur Loss: 0.53882545, Cur Avg Loss: 0.46516095, Log Avg loss: 0.51640792, Global Avg Loss: 1.23495019, Time: 0.0066 Steps: 89910, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000502, Sample Num: 8032, Cur Loss: 0.57354510, Cur Avg Loss: 0.46430502, Log Avg loss: 0.42219305, Global Avg Loss: 1.23485980, Time: 0.0063 Steps: 89920, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000512, Sample Num: 8192, Cur Loss: 0.21334250, Cur Avg Loss: 0.46208868, Log Avg loss: 0.35082841, Global Avg Loss: 1.23476150, Time: 0.0080 Steps: 89930, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000522, Sample Num: 8352, Cur Loss: 0.29373205, Cur Avg Loss: 0.45967927, Log Avg loss: 0.33631746, Global Avg Loss: 1.23466161, Time: 0.0083 Steps: 89940, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000532, Sample Num: 8512, Cur Loss: 0.69072777, Cur Avg Loss: 0.46039233, Log Avg loss: 0.49761415, Global Avg Loss: 1.23457967, Time: 0.0064 Steps: 89950, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000542, Sample Num: 8672, Cur Loss: 0.51220095, Cur Avg Loss: 0.46058404, Log Avg loss: 0.47078317, Global Avg Loss: 1.23449476, Time: 0.0212 Steps: 89960, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000552, Sample Num: 8832, Cur Loss: 0.31173044, Cur Avg Loss: 0.46109049, Log Avg loss: 0.48853978, Global Avg Loss: 1.23441185, Time: 0.0064 Steps: 89970, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000562, Sample Num: 8992, Cur Loss: 0.89091814, Cur Avg Loss: 0.46136331, Log Avg loss: 0.47642317, Global Avg Loss: 1.23432761, Time: 0.0075 Steps: 89980, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000572, Sample Num: 9152, Cur Loss: 0.48800439, Cur Avg Loss: 0.46046054, Log Avg loss: 0.40972457, Global Avg Loss: 1.23423598, Time: 0.0064 Steps: 89990, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000582, Sample Num: 9312, Cur Loss: 0.22623840, Cur Avg Loss: 0.46002311, Log Avg loss: 0.43500235, Global Avg Loss: 1.23414718, Time: 0.0150 Steps: 90000, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000592, Sample Num: 9472, Cur Loss: 0.44713247, Cur Avg Loss: 0.45832238, Log Avg loss: 0.35933963, Global Avg Loss: 1.23404999, Time: 0.0067 Steps: 90010, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000602, Sample Num: 9632, Cur Loss: 0.53688139, Cur Avg Loss: 0.46248784, Log Avg loss: 0.70908334, Global Avg Loss: 1.23399167, Time: 0.0065 Steps: 90020, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000612, Sample Num: 9792, Cur Loss: 0.18582603, Cur Avg Loss: 0.46358200, Log Avg loss: 0.52945027, Global Avg Loss: 1.23391341, Time: 0.0117 Steps: 90030, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000622, Sample Num: 9952, Cur Loss: 0.11580963, Cur Avg Loss: 0.46172685, Log Avg loss: 0.34819205, Global Avg Loss: 1.23381504, Time: 0.0067 Steps: 90040, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000632, Sample Num: 10112, Cur Loss: 0.20333186, Cur Avg Loss: 0.46122584, Log Avg loss: 0.43006301, Global Avg Loss: 1.23372579, Time: 0.0066 Steps: 90050, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000642, Sample Num: 10272, Cur Loss: 0.50491410, Cur Avg Loss: 0.46005043, Log Avg loss: 0.38576405, Global Avg Loss: 1.23363163, Time: 0.0065 Steps: 90060, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000652, Sample Num: 10432, Cur Loss: 0.44053981, Cur Avg Loss: 0.45989014, Log Avg loss: 0.44959978, Global Avg Loss: 1.23354458, Time: 0.0065 Steps: 90070, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000662, Sample Num: 10592, Cur Loss: 0.43731827, Cur Avg Loss: 0.46075375, Log Avg loss: 0.51706087, Global Avg Loss: 1.23346505, Time: 0.0065 Steps: 90080, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000672, Sample Num: 10752, Cur Loss: 0.45823988, Cur Avg Loss: 0.45910792, Log Avg loss: 0.35015386, Global Avg Loss: 1.23336700, Time: 0.0066 Steps: 90090, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000682, Sample Num: 10912, Cur Loss: 0.18377505, Cur Avg Loss: 0.45959945, Log Avg loss: 0.49263058, Global Avg Loss: 1.23328479, Time: 0.0067 Steps: 90100, Updated lr: 0.000016 Training, Epoch: 0043, Batch: 000692, Sample Num: 11072, Cur Loss: 0.67714959, Cur Avg Loss: 0.45878179, Log Avg loss: 0.40301749, Global Avg Loss: 1.23319265, Time: 0.0074 Steps: 90110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000702, Sample Num: 11232, Cur Loss: 0.12318474, Cur Avg Loss: 0.45927965, Log Avg loss: 0.49373122, Global Avg Loss: 1.23311059, Time: 0.0075 Steps: 90120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000712, Sample Num: 11392, Cur Loss: 0.63012034, Cur Avg Loss: 0.45887888, Log Avg loss: 0.43074484, Global Avg Loss: 1.23302157, Time: 0.0076 Steps: 90130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000722, Sample Num: 11552, Cur Loss: 0.23396155, Cur Avg Loss: 0.45926209, Log Avg loss: 0.48654687, Global Avg Loss: 1.23293876, Time: 0.0076 Steps: 90140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000732, Sample Num: 11712, Cur Loss: 0.47819382, Cur Avg Loss: 0.45900722, Log Avg loss: 0.44060596, Global Avg Loss: 1.23285087, Time: 0.0076 Steps: 90150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000742, Sample Num: 11872, Cur Loss: 0.17208633, Cur Avg Loss: 0.45837489, Log Avg loss: 0.41208769, Global Avg Loss: 1.23275983, Time: 0.0078 Steps: 90160, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000752, Sample Num: 12032, Cur Loss: 0.50883532, Cur Avg Loss: 0.45857452, Log Avg loss: 0.47338763, Global Avg Loss: 1.23267562, Time: 0.0076 Steps: 90170, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000762, Sample Num: 12192, Cur Loss: 0.35941380, Cur Avg Loss: 0.45959854, Log Avg loss: 0.53660475, Global Avg Loss: 1.23259843, Time: 0.0075 Steps: 90180, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000772, Sample Num: 12352, Cur Loss: 0.13084735, Cur Avg Loss: 0.45818940, Log Avg loss: 0.35081236, Global Avg Loss: 1.23250066, Time: 0.0074 Steps: 90190, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000782, Sample Num: 12512, Cur Loss: 0.59440577, Cur Avg Loss: 0.45809432, Log Avg loss: 0.45075468, Global Avg Loss: 1.23241399, Time: 0.0074 Steps: 90200, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000792, Sample Num: 12672, Cur Loss: 0.21147026, Cur Avg Loss: 0.45822331, Log Avg loss: 0.46830987, Global Avg Loss: 1.23232929, Time: 0.0075 Steps: 90210, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000802, Sample Num: 12832, Cur Loss: 0.85867459, Cur Avg Loss: 0.45971211, Log Avg loss: 0.57762524, Global Avg Loss: 1.23225672, Time: 0.0075 Steps: 90220, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000812, Sample Num: 12992, Cur Loss: 0.13601482, Cur Avg Loss: 0.46011754, Log Avg loss: 0.49263278, Global Avg Loss: 1.23217475, Time: 0.0074 Steps: 90230, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000822, Sample Num: 13152, Cur Loss: 1.14469802, Cur Avg Loss: 0.45973435, Log Avg loss: 0.42861997, Global Avg Loss: 1.23208570, Time: 0.0075 Steps: 90240, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000832, Sample Num: 13312, Cur Loss: 0.55981410, Cur Avg Loss: 0.46190644, Log Avg loss: 0.64045208, Global Avg Loss: 1.23202015, Time: 0.0075 Steps: 90250, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000842, Sample Num: 13472, Cur Loss: 0.38305607, Cur Avg Loss: 0.46340465, Log Avg loss: 0.58805589, Global Avg Loss: 1.23194880, Time: 0.0074 Steps: 90260, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000852, Sample Num: 13632, Cur Loss: 0.45116931, Cur Avg Loss: 0.46366314, Log Avg loss: 0.48542737, Global Avg Loss: 1.23186611, Time: 0.0074 Steps: 90270, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000862, Sample Num: 13792, Cur Loss: 0.26414078, Cur Avg Loss: 0.46346267, Log Avg loss: 0.44638295, Global Avg Loss: 1.23177910, Time: 0.0074 Steps: 90280, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000872, Sample Num: 13952, Cur Loss: 0.48090065, Cur Avg Loss: 0.46344707, Log Avg loss: 0.46210252, Global Avg Loss: 1.23169386, Time: 0.0075 Steps: 90290, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000882, Sample Num: 14112, Cur Loss: 0.33488312, Cur Avg Loss: 0.46333271, Log Avg loss: 0.45336075, Global Avg Loss: 1.23160766, Time: 0.0074 Steps: 90300, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000892, Sample Num: 14272, Cur Loss: 0.47073698, Cur Avg Loss: 0.46337599, Log Avg loss: 0.46719330, Global Avg Loss: 1.23152302, Time: 0.0074 Steps: 90310, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000902, Sample Num: 14432, Cur Loss: 0.50723439, Cur Avg Loss: 0.46353981, Log Avg loss: 0.47815185, Global Avg Loss: 1.23143961, Time: 0.0075 Steps: 90320, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000912, Sample Num: 14592, Cur Loss: 0.49276862, Cur Avg Loss: 0.46322810, Log Avg loss: 0.43511240, Global Avg Loss: 1.23135145, Time: 0.0075 Steps: 90330, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000922, Sample Num: 14752, Cur Loss: 0.93931127, Cur Avg Loss: 0.46513498, Log Avg loss: 0.63904213, Global Avg Loss: 1.23128588, Time: 0.0074 Steps: 90340, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000932, Sample Num: 14912, Cur Loss: 0.37908179, Cur Avg Loss: 0.46574373, Log Avg loss: 0.52187021, Global Avg Loss: 1.23120737, Time: 0.0076 Steps: 90350, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000942, Sample Num: 15072, Cur Loss: 0.43598834, Cur Avg Loss: 0.46502164, Log Avg loss: 0.39772277, Global Avg Loss: 1.23111513, Time: 0.0075 Steps: 90360, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000952, Sample Num: 15232, Cur Loss: 0.38102883, Cur Avg Loss: 0.46574698, Log Avg loss: 0.53407434, Global Avg Loss: 1.23103799, Time: 0.0075 Steps: 90370, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000962, Sample Num: 15392, Cur Loss: 0.42923281, Cur Avg Loss: 0.46743019, Log Avg loss: 0.62767216, Global Avg Loss: 1.23097123, Time: 0.0075 Steps: 90380, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000972, Sample Num: 15552, Cur Loss: 0.87627804, Cur Avg Loss: 0.46693035, Log Avg loss: 0.41884496, Global Avg Loss: 1.23088139, Time: 0.0074 Steps: 90390, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000982, Sample Num: 15712, Cur Loss: 0.56112725, Cur Avg Loss: 0.46861708, Log Avg loss: 0.63256724, Global Avg Loss: 1.23081520, Time: 0.0075 Steps: 90400, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 000992, Sample Num: 15872, Cur Loss: 0.44062984, Cur Avg Loss: 0.46733861, Log Avg loss: 0.34179343, Global Avg Loss: 1.23071687, Time: 0.0074 Steps: 90410, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001002, Sample Num: 16032, Cur Loss: 0.23257576, Cur Avg Loss: 0.46726182, Log Avg loss: 0.45964416, Global Avg Loss: 1.23063159, Time: 0.0074 Steps: 90420, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001012, Sample Num: 16192, Cur Loss: 0.20740297, Cur Avg Loss: 0.46702870, Log Avg loss: 0.44366969, Global Avg Loss: 1.23054457, Time: 0.0075 Steps: 90430, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001022, Sample Num: 16352, Cur Loss: 0.26448473, Cur Avg Loss: 0.46730748, Log Avg loss: 0.49552081, Global Avg Loss: 1.23046330, Time: 0.0074 Steps: 90440, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001032, Sample Num: 16512, Cur Loss: 0.21539363, Cur Avg Loss: 0.46772535, Log Avg loss: 0.51043156, Global Avg Loss: 1.23038369, Time: 0.0075 Steps: 90450, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001042, Sample Num: 16672, Cur Loss: 0.50101161, Cur Avg Loss: 0.46757362, Log Avg loss: 0.45191443, Global Avg Loss: 1.23029763, Time: 0.0074 Steps: 90460, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001052, Sample Num: 16832, Cur Loss: 0.97742742, Cur Avg Loss: 0.46887170, Log Avg loss: 0.60413154, Global Avg Loss: 1.23022842, Time: 0.0075 Steps: 90470, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001062, Sample Num: 16992, Cur Loss: 0.37712198, Cur Avg Loss: 0.46978293, Log Avg loss: 0.56564488, Global Avg Loss: 1.23015497, Time: 0.0074 Steps: 90480, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001072, Sample Num: 17152, Cur Loss: 0.37836543, Cur Avg Loss: 0.46931960, Log Avg loss: 0.42011345, Global Avg Loss: 1.23006545, Time: 0.0074 Steps: 90490, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001082, Sample Num: 17312, Cur Loss: 0.39816079, Cur Avg Loss: 0.46943943, Log Avg loss: 0.48228527, Global Avg Loss: 1.22998283, Time: 0.0074 Steps: 90500, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001092, Sample Num: 17472, Cur Loss: 0.53454840, Cur Avg Loss: 0.46893290, Log Avg loss: 0.41412672, Global Avg Loss: 1.22989269, Time: 0.0074 Steps: 90510, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001102, Sample Num: 17632, Cur Loss: 0.14568727, Cur Avg Loss: 0.46924681, Log Avg loss: 0.50352524, Global Avg Loss: 1.22981244, Time: 0.0074 Steps: 90520, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001112, Sample Num: 17792, Cur Loss: 0.42048734, Cur Avg Loss: 0.46872303, Log Avg loss: 0.41100244, Global Avg Loss: 1.22972200, Time: 0.0074 Steps: 90530, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001122, Sample Num: 17952, Cur Loss: 0.85751903, Cur Avg Loss: 0.46773473, Log Avg loss: 0.35783679, Global Avg Loss: 1.22962570, Time: 0.0075 Steps: 90540, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001132, Sample Num: 18112, Cur Loss: 0.14837918, Cur Avg Loss: 0.46734936, Log Avg loss: 0.42411047, Global Avg Loss: 1.22953674, Time: 0.0074 Steps: 90550, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001142, Sample Num: 18272, Cur Loss: 0.56741309, Cur Avg Loss: 0.46730146, Log Avg loss: 0.46187908, Global Avg Loss: 1.22945197, Time: 0.0074 Steps: 90560, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001152, Sample Num: 18432, Cur Loss: 0.59840190, Cur Avg Loss: 0.46741695, Log Avg loss: 0.48060536, Global Avg Loss: 1.22936929, Time: 0.0075 Steps: 90570, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001162, Sample Num: 18592, Cur Loss: 0.31416014, Cur Avg Loss: 0.46805450, Log Avg loss: 0.54150061, Global Avg Loss: 1.22929335, Time: 0.0074 Steps: 90580, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001172, Sample Num: 18752, Cur Loss: 0.46591058, Cur Avg Loss: 0.46799400, Log Avg loss: 0.46096404, Global Avg Loss: 1.22920854, Time: 0.0074 Steps: 90590, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001182, Sample Num: 18912, Cur Loss: 0.22463483, Cur Avg Loss: 0.46718923, Log Avg loss: 0.37287035, Global Avg Loss: 1.22911402, Time: 0.0074 Steps: 90600, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001192, Sample Num: 19072, Cur Loss: 0.27877098, Cur Avg Loss: 0.46716623, Log Avg loss: 0.46444762, Global Avg Loss: 1.22902963, Time: 0.0074 Steps: 90610, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001202, Sample Num: 19232, Cur Loss: 0.59693593, Cur Avg Loss: 0.46705778, Log Avg loss: 0.45413012, Global Avg Loss: 1.22894412, Time: 0.0074 Steps: 90620, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001212, Sample Num: 19392, Cur Loss: 0.25674784, Cur Avg Loss: 0.46738385, Log Avg loss: 0.50657811, Global Avg Loss: 1.22886441, Time: 0.0074 Steps: 90630, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001222, Sample Num: 19552, Cur Loss: 0.86962008, Cur Avg Loss: 0.46725294, Log Avg loss: 0.45138601, Global Avg Loss: 1.22877863, Time: 0.0074 Steps: 90640, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001232, Sample Num: 19712, Cur Loss: 0.51519704, Cur Avg Loss: 0.46730252, Log Avg loss: 0.47336184, Global Avg Loss: 1.22869530, Time: 0.0074 Steps: 90650, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001242, Sample Num: 19872, Cur Loss: 0.44182536, Cur Avg Loss: 0.46783361, Log Avg loss: 0.53326356, Global Avg Loss: 1.22861859, Time: 0.0074 Steps: 90660, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001252, Sample Num: 20032, Cur Loss: 0.39709798, Cur Avg Loss: 0.46774458, Log Avg loss: 0.45668720, Global Avg Loss: 1.22853346, Time: 0.0074 Steps: 90670, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001262, Sample Num: 20192, Cur Loss: 0.59112060, Cur Avg Loss: 0.46861452, Log Avg loss: 0.57753031, Global Avg Loss: 1.22846167, Time: 0.0074 Steps: 90680, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001272, Sample Num: 20352, Cur Loss: 0.54800344, Cur Avg Loss: 0.46870277, Log Avg loss: 0.47984099, Global Avg Loss: 1.22837912, Time: 0.0074 Steps: 90690, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001282, Sample Num: 20512, Cur Loss: 0.43462440, Cur Avg Loss: 0.46904522, Log Avg loss: 0.51260444, Global Avg Loss: 1.22830020, Time: 0.0075 Steps: 90700, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001292, Sample Num: 20672, Cur Loss: 0.30148190, Cur Avg Loss: 0.46861193, Log Avg loss: 0.41306403, Global Avg Loss: 1.22821033, Time: 0.0073 Steps: 90710, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001302, Sample Num: 20832, Cur Loss: 0.50482810, Cur Avg Loss: 0.46823142, Log Avg loss: 0.41906992, Global Avg Loss: 1.22812114, Time: 0.0074 Steps: 90720, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001312, Sample Num: 20992, Cur Loss: 0.37125123, Cur Avg Loss: 0.46800570, Log Avg loss: 0.43861609, Global Avg Loss: 1.22803412, Time: 0.0073 Steps: 90730, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001322, Sample Num: 21152, Cur Loss: 0.44809511, Cur Avg Loss: 0.46742368, Log Avg loss: 0.39106281, Global Avg Loss: 1.22794188, Time: 0.0063 Steps: 90740, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001332, Sample Num: 21312, Cur Loss: 0.48163196, Cur Avg Loss: 0.46609608, Log Avg loss: 0.29058748, Global Avg Loss: 1.22783859, Time: 0.0066 Steps: 90750, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001342, Sample Num: 21472, Cur Loss: 0.32355365, Cur Avg Loss: 0.46583528, Log Avg loss: 0.43109654, Global Avg Loss: 1.22775081, Time: 0.0063 Steps: 90760, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001352, Sample Num: 21632, Cur Loss: 0.22212741, Cur Avg Loss: 0.46528057, Log Avg loss: 0.39083833, Global Avg Loss: 1.22765861, Time: 0.0214 Steps: 90770, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001362, Sample Num: 21792, Cur Loss: 0.77838802, Cur Avg Loss: 0.46633945, Log Avg loss: 0.60950082, Global Avg Loss: 1.22759051, Time: 0.0193 Steps: 90780, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001372, Sample Num: 21952, Cur Loss: 0.72853643, Cur Avg Loss: 0.46774313, Log Avg loss: 0.65892429, Global Avg Loss: 1.22752788, Time: 0.0115 Steps: 90790, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001382, Sample Num: 22112, Cur Loss: 0.49150240, Cur Avg Loss: 0.46747357, Log Avg loss: 0.43049011, Global Avg Loss: 1.22744010, Time: 0.0064 Steps: 90800, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001392, Sample Num: 22272, Cur Loss: 0.47717503, Cur Avg Loss: 0.46750930, Log Avg loss: 0.47244648, Global Avg Loss: 1.22735696, Time: 0.0064 Steps: 90810, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001402, Sample Num: 22432, Cur Loss: 0.40995017, Cur Avg Loss: 0.46676095, Log Avg loss: 0.36259052, Global Avg Loss: 1.22726174, Time: 0.0064 Steps: 90820, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001412, Sample Num: 22592, Cur Loss: 0.71421468, Cur Avg Loss: 0.46622079, Log Avg loss: 0.39049024, Global Avg Loss: 1.22716961, Time: 0.0174 Steps: 90830, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001422, Sample Num: 22752, Cur Loss: 0.37774748, Cur Avg Loss: 0.46706972, Log Avg loss: 0.58693888, Global Avg Loss: 1.22709914, Time: 0.0234 Steps: 90840, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001432, Sample Num: 22912, Cur Loss: 0.17379022, Cur Avg Loss: 0.46669029, Log Avg loss: 0.41273550, Global Avg Loss: 1.22700950, Time: 0.0064 Steps: 90850, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001442, Sample Num: 23072, Cur Loss: 0.33117399, Cur Avg Loss: 0.46670164, Log Avg loss: 0.46832784, Global Avg Loss: 1.22692600, Time: 0.0144 Steps: 90860, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001452, Sample Num: 23232, Cur Loss: 0.88939309, Cur Avg Loss: 0.46785554, Log Avg loss: 0.63424686, Global Avg Loss: 1.22686077, Time: 0.0064 Steps: 90870, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001462, Sample Num: 23392, Cur Loss: 0.12656458, Cur Avg Loss: 0.46866001, Log Avg loss: 0.58546906, Global Avg Loss: 1.22679020, Time: 0.0072 Steps: 90880, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001472, Sample Num: 23552, Cur Loss: 0.23663186, Cur Avg Loss: 0.46742115, Log Avg loss: 0.28629973, Global Avg Loss: 1.22668672, Time: 0.0205 Steps: 90890, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001482, Sample Num: 23712, Cur Loss: 0.96148717, Cur Avg Loss: 0.46965112, Log Avg loss: 0.79790320, Global Avg Loss: 1.22663955, Time: 0.0189 Steps: 90900, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001492, Sample Num: 23872, Cur Loss: 0.69519794, Cur Avg Loss: 0.47083169, Log Avg loss: 0.64579231, Global Avg Loss: 1.22657566, Time: 0.0166 Steps: 90910, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001502, Sample Num: 24032, Cur Loss: 0.38578868, Cur Avg Loss: 0.47061955, Log Avg loss: 0.43896864, Global Avg Loss: 1.22648903, Time: 0.0063 Steps: 90920, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001512, Sample Num: 24192, Cur Loss: 0.37642437, Cur Avg Loss: 0.47070601, Log Avg loss: 0.48369187, Global Avg Loss: 1.22640734, Time: 0.0066 Steps: 90930, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001522, Sample Num: 24352, Cur Loss: 0.83414555, Cur Avg Loss: 0.47045809, Log Avg loss: 0.43297230, Global Avg Loss: 1.22632010, Time: 0.0064 Steps: 90940, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001532, Sample Num: 24512, Cur Loss: 0.59287196, Cur Avg Loss: 0.47112090, Log Avg loss: 0.57200099, Global Avg Loss: 1.22624815, Time: 0.0065 Steps: 90950, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001542, Sample Num: 24672, Cur Loss: 0.72765815, Cur Avg Loss: 0.47138641, Log Avg loss: 0.51206270, Global Avg Loss: 1.22616964, Time: 0.0065 Steps: 90960, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001552, Sample Num: 24832, Cur Loss: 0.18998614, Cur Avg Loss: 0.47176800, Log Avg loss: 0.53060864, Global Avg Loss: 1.22609318, Time: 0.0064 Steps: 90970, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001562, Sample Num: 24992, Cur Loss: 0.25901529, Cur Avg Loss: 0.47129120, Log Avg loss: 0.39729265, Global Avg Loss: 1.22600208, Time: 0.0064 Steps: 90980, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001572, Sample Num: 25152, Cur Loss: 1.05198932, Cur Avg Loss: 0.47140375, Log Avg loss: 0.48898305, Global Avg Loss: 1.22592108, Time: 0.0066 Steps: 90990, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001582, Sample Num: 25312, Cur Loss: 0.34272987, Cur Avg Loss: 0.47093607, Log Avg loss: 0.39741687, Global Avg Loss: 1.22583004, Time: 0.0065 Steps: 91000, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001592, Sample Num: 25472, Cur Loss: 0.45190012, Cur Avg Loss: 0.47114068, Log Avg loss: 0.50351032, Global Avg Loss: 1.22575067, Time: 0.0065 Steps: 91010, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001602, Sample Num: 25632, Cur Loss: 0.21964821, Cur Avg Loss: 0.47109126, Log Avg loss: 0.46322319, Global Avg Loss: 1.22566689, Time: 0.0076 Steps: 91020, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001612, Sample Num: 25792, Cur Loss: 0.20026709, Cur Avg Loss: 0.47058523, Log Avg loss: 0.38952007, Global Avg Loss: 1.22557504, Time: 0.0076 Steps: 91030, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001622, Sample Num: 25952, Cur Loss: 0.96378666, Cur Avg Loss: 0.47095062, Log Avg loss: 0.52985145, Global Avg Loss: 1.22549862, Time: 0.0077 Steps: 91040, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001632, Sample Num: 26112, Cur Loss: 0.35353363, Cur Avg Loss: 0.47148195, Log Avg loss: 0.55766345, Global Avg Loss: 1.22542527, Time: 0.0076 Steps: 91050, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001642, Sample Num: 26272, Cur Loss: 1.07495904, Cur Avg Loss: 0.47140745, Log Avg loss: 0.45924880, Global Avg Loss: 1.22534113, Time: 0.0075 Steps: 91060, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001652, Sample Num: 26432, Cur Loss: 0.68658751, Cur Avg Loss: 0.47145322, Log Avg loss: 0.47896943, Global Avg Loss: 1.22525918, Time: 0.0076 Steps: 91070, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001662, Sample Num: 26592, Cur Loss: 0.45108673, Cur Avg Loss: 0.47152910, Log Avg loss: 0.48406291, Global Avg Loss: 1.22517780, Time: 0.0075 Steps: 91080, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001672, Sample Num: 26752, Cur Loss: 0.39052486, Cur Avg Loss: 0.47197304, Log Avg loss: 0.54575645, Global Avg Loss: 1.22510321, Time: 0.0075 Steps: 91090, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001682, Sample Num: 26912, Cur Loss: 1.03227270, Cur Avg Loss: 0.47180384, Log Avg loss: 0.44351388, Global Avg Loss: 1.22501741, Time: 0.0075 Steps: 91100, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001692, Sample Num: 27072, Cur Loss: 0.38453284, Cur Avg Loss: 0.47174310, Log Avg loss: 0.46152655, Global Avg Loss: 1.22493362, Time: 0.0075 Steps: 91110, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001702, Sample Num: 27232, Cur Loss: 0.81744635, Cur Avg Loss: 0.47217894, Log Avg loss: 0.54592312, Global Avg Loss: 1.22485910, Time: 0.0076 Steps: 91120, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001712, Sample Num: 27392, Cur Loss: 0.75143170, Cur Avg Loss: 0.47356637, Log Avg loss: 0.70970609, Global Avg Loss: 1.22480257, Time: 0.0075 Steps: 91130, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001722, Sample Num: 27552, Cur Loss: 0.35455185, Cur Avg Loss: 0.47411008, Log Avg loss: 0.56719485, Global Avg Loss: 1.22473041, Time: 0.0076 Steps: 91140, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001732, Sample Num: 27712, Cur Loss: 0.27082080, Cur Avg Loss: 0.47365456, Log Avg loss: 0.39521294, Global Avg Loss: 1.22463941, Time: 0.0075 Steps: 91150, Updated lr: 0.000015 Training, Epoch: 0043, Batch: 001742, Sample Num: 27872, Cur Loss: 0.49892500, Cur Avg Loss: 0.47384196, Log Avg loss: 0.50629975, Global Avg Loss: 1.22456061, Time: 0.0075 Steps: 91160, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001752, Sample Num: 28032, Cur Loss: 0.48909354, Cur Avg Loss: 0.47353000, Log Avg loss: 0.41918651, Global Avg Loss: 1.22447227, Time: 0.0075 Steps: 91170, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001762, Sample Num: 28192, Cur Loss: 0.54924244, Cur Avg Loss: 0.47449424, Log Avg loss: 0.64342887, Global Avg Loss: 1.22440855, Time: 0.0075 Steps: 91180, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001772, Sample Num: 28352, Cur Loss: 0.26321504, Cur Avg Loss: 0.47422078, Log Avg loss: 0.42603702, Global Avg Loss: 1.22432100, Time: 0.0075 Steps: 91190, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001782, Sample Num: 28512, Cur Loss: 0.85816687, Cur Avg Loss: 0.47429437, Log Avg loss: 0.48733450, Global Avg Loss: 1.22424019, Time: 0.0075 Steps: 91200, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001792, Sample Num: 28672, Cur Loss: 0.17076781, Cur Avg Loss: 0.47399975, Log Avg loss: 0.42149820, Global Avg Loss: 1.22415218, Time: 0.0084 Steps: 91210, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001802, Sample Num: 28832, Cur Loss: 0.24056485, Cur Avg Loss: 0.47345404, Log Avg loss: 0.37566447, Global Avg Loss: 1.22405916, Time: 0.0076 Steps: 91220, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001812, Sample Num: 28992, Cur Loss: 0.38451096, Cur Avg Loss: 0.47383217, Log Avg loss: 0.54197013, Global Avg Loss: 1.22398439, Time: 0.0076 Steps: 91230, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001822, Sample Num: 29152, Cur Loss: 0.46500012, Cur Avg Loss: 0.47382093, Log Avg loss: 0.47178383, Global Avg Loss: 1.22390195, Time: 0.0076 Steps: 91240, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001832, Sample Num: 29312, Cur Loss: 0.35138899, Cur Avg Loss: 0.47387377, Log Avg loss: 0.48350143, Global Avg Loss: 1.22382081, Time: 0.0075 Steps: 91250, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001842, Sample Num: 29472, Cur Loss: 1.50564468, Cur Avg Loss: 0.47493137, Log Avg loss: 0.66868494, Global Avg Loss: 1.22375998, Time: 0.0074 Steps: 91260, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001852, Sample Num: 29632, Cur Loss: 0.68896985, Cur Avg Loss: 0.47451884, Log Avg loss: 0.39852910, Global Avg Loss: 1.22366957, Time: 0.0075 Steps: 91270, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001862, Sample Num: 29792, Cur Loss: 0.32484645, Cur Avg Loss: 0.47499396, Log Avg loss: 0.56298708, Global Avg Loss: 1.22359719, Time: 0.0074 Steps: 91280, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001872, Sample Num: 29952, Cur Loss: 0.42608932, Cur Avg Loss: 0.47452725, Log Avg loss: 0.38762529, Global Avg Loss: 1.22350561, Time: 0.0077 Steps: 91290, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001882, Sample Num: 30112, Cur Loss: 0.23808517, Cur Avg Loss: 0.47465832, Log Avg loss: 0.49919583, Global Avg Loss: 1.22342628, Time: 0.0076 Steps: 91300, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001892, Sample Num: 30272, Cur Loss: 0.24853195, Cur Avg Loss: 0.47461575, Log Avg loss: 0.46660407, Global Avg Loss: 1.22334339, Time: 0.0075 Steps: 91310, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001902, Sample Num: 30432, Cur Loss: 1.02501559, Cur Avg Loss: 0.47580024, Log Avg loss: 0.69990453, Global Avg Loss: 1.22328608, Time: 0.0075 Steps: 91320, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001912, Sample Num: 30592, Cur Loss: 0.18139043, Cur Avg Loss: 0.47528040, Log Avg loss: 0.37640733, Global Avg Loss: 1.22319335, Time: 0.0075 Steps: 91330, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001922, Sample Num: 30752, Cur Loss: 0.81217498, Cur Avg Loss: 0.47574415, Log Avg loss: 0.56441265, Global Avg Loss: 1.22312122, Time: 0.0077 Steps: 91340, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001932, Sample Num: 30912, Cur Loss: 0.37180692, Cur Avg Loss: 0.47572960, Log Avg loss: 0.47293437, Global Avg Loss: 1.22303910, Time: 0.0075 Steps: 91350, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001942, Sample Num: 31072, Cur Loss: 0.63813490, Cur Avg Loss: 0.47512303, Log Avg loss: 0.35793217, Global Avg Loss: 1.22294441, Time: 0.0075 Steps: 91360, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001952, Sample Num: 31232, Cur Loss: 0.69385320, Cur Avg Loss: 0.47484264, Log Avg loss: 0.42039237, Global Avg Loss: 1.22285657, Time: 0.0076 Steps: 91370, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001962, Sample Num: 31392, Cur Loss: 0.50345784, Cur Avg Loss: 0.47473055, Log Avg loss: 0.45285009, Global Avg Loss: 1.22277231, Time: 0.0076 Steps: 91380, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001972, Sample Num: 31552, Cur Loss: 0.13463047, Cur Avg Loss: 0.47453520, Log Avg loss: 0.43620754, Global Avg Loss: 1.22268624, Time: 0.0076 Steps: 91390, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001982, Sample Num: 31712, Cur Loss: 0.32838809, Cur Avg Loss: 0.47430862, Log Avg loss: 0.42962597, Global Avg Loss: 1.22259948, Time: 0.0075 Steps: 91400, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 001992, Sample Num: 31872, Cur Loss: 0.07800889, Cur Avg Loss: 0.47398998, Log Avg loss: 0.41083608, Global Avg Loss: 1.22251067, Time: 0.0075 Steps: 91410, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002002, Sample Num: 32032, Cur Loss: 0.85159421, Cur Avg Loss: 0.47424128, Log Avg loss: 0.52429997, Global Avg Loss: 1.22243430, Time: 0.0076 Steps: 91420, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002012, Sample Num: 32192, Cur Loss: 0.54998267, Cur Avg Loss: 0.47414080, Log Avg loss: 0.45402508, Global Avg Loss: 1.22235025, Time: 0.0076 Steps: 91430, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002022, Sample Num: 32352, Cur Loss: 0.27321792, Cur Avg Loss: 0.47372891, Log Avg loss: 0.39085689, Global Avg Loss: 1.22225932, Time: 0.0076 Steps: 91440, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002032, Sample Num: 32512, Cur Loss: 0.73568785, Cur Avg Loss: 0.47418775, Log Avg loss: 0.56696450, Global Avg Loss: 1.22218766, Time: 0.0075 Steps: 91450, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002042, Sample Num: 32672, Cur Loss: 0.35350615, Cur Avg Loss: 0.47420552, Log Avg loss: 0.47781795, Global Avg Loss: 1.22210628, Time: 0.0075 Steps: 91460, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002052, Sample Num: 32832, Cur Loss: 0.17072526, Cur Avg Loss: 0.47344867, Log Avg loss: 0.31889969, Global Avg Loss: 1.22200753, Time: 0.0065 Steps: 91470, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002062, Sample Num: 32992, Cur Loss: 0.39554083, Cur Avg Loss: 0.47367803, Log Avg loss: 0.52074139, Global Avg Loss: 1.22193087, Time: 0.0152 Steps: 91480, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002072, Sample Num: 33152, Cur Loss: 0.57058525, Cur Avg Loss: 0.47359711, Log Avg loss: 0.45691218, Global Avg Loss: 1.22184726, Time: 0.0065 Steps: 91490, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002082, Sample Num: 33312, Cur Loss: 0.17845777, Cur Avg Loss: 0.47330722, Log Avg loss: 0.41324209, Global Avg Loss: 1.22175889, Time: 0.0069 Steps: 91500, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002092, Sample Num: 33472, Cur Loss: 0.29674199, Cur Avg Loss: 0.47345567, Log Avg loss: 0.50436378, Global Avg Loss: 1.22168049, Time: 0.0065 Steps: 91510, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002102, Sample Num: 33632, Cur Loss: 0.20354559, Cur Avg Loss: 0.47384900, Log Avg loss: 0.55613353, Global Avg Loss: 1.22160777, Time: 0.0235 Steps: 91520, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002112, Sample Num: 33792, Cur Loss: 0.23960882, Cur Avg Loss: 0.47312166, Log Avg loss: 0.32023467, Global Avg Loss: 1.22150929, Time: 0.0066 Steps: 91530, Updated lr: 0.000014 Training, Epoch: 0043, Batch: 002122, Sample Num: 33952, Cur Loss: 0.39304250, Cur Avg Loss: 0.47310208, Log Avg loss: 0.46896689, Global Avg Loss: 1.22142708, Time: 0.0137 Steps: 91540, Updated lr: 0.000014 ***** Running evaluation checkpoint-91547 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-91547 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.590959, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.582437, "eval_total_loss": 409.453428, "eval_mae": 0.582552, "eval_mse": 0.582561, "eval_r2": 0.629686, "eval_sp_statistic": 0.74791, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.796946, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.872996, "test_total_loss": 438.243939, "test_mae": 0.651724, "test_mse": 0.873236, "test_r2": 0.436405, "test_sp_statistic": 0.605038, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.674989, "test_ps_pvalue": 0.0, "lr": 1.413276434329066e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.22135850768082, "train_cur_epoch_loss": 1006.1949510984123, "train_cur_epoch_avg_loss": 0.4726138802716826, "train_cur_epoch_time": 17.590958833694458, "train_cur_epoch_avg_time": 0.008262545248329948, "epoch": 43, "step": 91547} ################################################## Training, Epoch: 0044, Batch: 000003, Sample Num: 48, Cur Loss: 0.26748592, Cur Avg Loss: 0.51870090, Log Avg loss: 0.38284315, Global Avg Loss: 1.22133548, Time: 0.0075 Steps: 91550, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000013, Sample Num: 208, Cur Loss: 0.58759248, Cur Avg Loss: 0.49391410, Log Avg loss: 0.48647807, Global Avg Loss: 1.22125522, Time: 0.0075 Steps: 91560, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000023, Sample Num: 368, Cur Loss: 0.85883588, Cur Avg Loss: 0.47646044, Log Avg loss: 0.45377067, Global Avg Loss: 1.22117141, Time: 0.0074 Steps: 91570, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000033, Sample Num: 528, Cur Loss: 0.57904989, Cur Avg Loss: 0.47483683, Log Avg loss: 0.47110253, Global Avg Loss: 1.22108951, Time: 0.0074 Steps: 91580, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000043, Sample Num: 688, Cur Loss: 0.35275221, Cur Avg Loss: 0.45466967, Log Avg loss: 0.38811803, Global Avg Loss: 1.22099856, Time: 0.0074 Steps: 91590, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000053, Sample Num: 848, Cur Loss: 0.21468309, Cur Avg Loss: 0.43486372, Log Avg loss: 0.34969814, Global Avg Loss: 1.22090344, Time: 0.0075 Steps: 91600, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000063, Sample Num: 1008, Cur Loss: 0.10592005, Cur Avg Loss: 0.42826745, Log Avg loss: 0.39330725, Global Avg Loss: 1.22081310, Time: 0.0075 Steps: 91610, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000073, Sample Num: 1168, Cur Loss: 0.48482251, Cur Avg Loss: 0.43305879, Log Avg loss: 0.46324419, Global Avg Loss: 1.22073041, Time: 0.0074 Steps: 91620, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000083, Sample Num: 1328, Cur Loss: 0.38887855, Cur Avg Loss: 0.43960606, Log Avg loss: 0.48740114, Global Avg Loss: 1.22065038, Time: 0.0074 Steps: 91630, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000093, Sample Num: 1488, Cur Loss: 0.22405455, Cur Avg Loss: 0.43724317, Log Avg loss: 0.41763117, Global Avg Loss: 1.22056276, Time: 0.0073 Steps: 91640, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000103, Sample Num: 1648, Cur Loss: 0.12851238, Cur Avg Loss: 0.42999073, Log Avg loss: 0.36254302, Global Avg Loss: 1.22046914, Time: 0.0065 Steps: 91650, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000113, Sample Num: 1808, Cur Loss: 0.39907047, Cur Avg Loss: 0.44720753, Log Avg loss: 0.62454065, Global Avg Loss: 1.22040412, Time: 0.0064 Steps: 91660, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000123, Sample Num: 1968, Cur Loss: 0.35189709, Cur Avg Loss: 0.44677805, Log Avg loss: 0.44192491, Global Avg Loss: 1.22031920, Time: 0.0065 Steps: 91670, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000133, Sample Num: 2128, Cur Loss: 0.60825896, Cur Avg Loss: 0.44566771, Log Avg loss: 0.43201055, Global Avg Loss: 1.22023321, Time: 0.0173 Steps: 91680, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000143, Sample Num: 2288, Cur Loss: 0.15225647, Cur Avg Loss: 0.45127715, Log Avg loss: 0.52588261, Global Avg Loss: 1.22015749, Time: 0.0104 Steps: 91690, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000153, Sample Num: 2448, Cur Loss: 0.50232363, Cur Avg Loss: 0.45674202, Log Avg loss: 0.53488972, Global Avg Loss: 1.22008276, Time: 0.0065 Steps: 91700, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000163, Sample Num: 2608, Cur Loss: 0.18909946, Cur Avg Loss: 0.45122392, Log Avg loss: 0.36679703, Global Avg Loss: 1.21998972, Time: 0.0066 Steps: 91710, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000173, Sample Num: 2768, Cur Loss: 0.58460957, Cur Avg Loss: 0.45063509, Log Avg loss: 0.44103720, Global Avg Loss: 1.21990479, Time: 0.0064 Steps: 91720, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000183, Sample Num: 2928, Cur Loss: 0.15827905, Cur Avg Loss: 0.44933131, Log Avg loss: 0.42677586, Global Avg Loss: 1.21981832, Time: 0.0115 Steps: 91730, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000193, Sample Num: 3088, Cur Loss: 0.39004198, Cur Avg Loss: 0.45196590, Log Avg loss: 0.50017897, Global Avg Loss: 1.21973988, Time: 0.0119 Steps: 91740, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000203, Sample Num: 3248, Cur Loss: 0.14665955, Cur Avg Loss: 0.44549239, Log Avg loss: 0.32055358, Global Avg Loss: 1.21964188, Time: 0.0066 Steps: 91750, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000213, Sample Num: 3408, Cur Loss: 0.28558666, Cur Avg Loss: 0.44156896, Log Avg loss: 0.36192328, Global Avg Loss: 1.21954840, Time: 0.0065 Steps: 91760, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000223, Sample Num: 3568, Cur Loss: 0.34574166, Cur Avg Loss: 0.44562260, Log Avg loss: 0.53196523, Global Avg Loss: 1.21947348, Time: 0.0063 Steps: 91770, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000233, Sample Num: 3728, Cur Loss: 0.12463017, Cur Avg Loss: 0.44539471, Log Avg loss: 0.44031274, Global Avg Loss: 1.21938858, Time: 0.0065 Steps: 91780, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000243, Sample Num: 3888, Cur Loss: 0.92019832, Cur Avg Loss: 0.44292102, Log Avg loss: 0.38528412, Global Avg Loss: 1.21929771, Time: 0.0066 Steps: 91790, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000253, Sample Num: 4048, Cur Loss: 0.73889548, Cur Avg Loss: 0.44717112, Log Avg loss: 0.55044845, Global Avg Loss: 1.21922485, Time: 0.0068 Steps: 91800, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000263, Sample Num: 4208, Cur Loss: 0.64832723, Cur Avg Loss: 0.44865053, Log Avg loss: 0.48607962, Global Avg Loss: 1.21914500, Time: 0.0074 Steps: 91810, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000273, Sample Num: 4368, Cur Loss: 0.37151149, Cur Avg Loss: 0.44743107, Log Avg loss: 0.41535912, Global Avg Loss: 1.21905746, Time: 0.0067 Steps: 91820, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000283, Sample Num: 4528, Cur Loss: 0.70855433, Cur Avg Loss: 0.44860188, Log Avg loss: 0.48056515, Global Avg Loss: 1.21897704, Time: 0.0064 Steps: 91830, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000293, Sample Num: 4688, Cur Loss: 0.24878272, Cur Avg Loss: 0.44598971, Log Avg loss: 0.37206530, Global Avg Loss: 1.21888482, Time: 0.0066 Steps: 91840, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000303, Sample Num: 4848, Cur Loss: 0.75407767, Cur Avg Loss: 0.44990892, Log Avg loss: 0.56474166, Global Avg Loss: 1.21881361, Time: 0.0126 Steps: 91850, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000313, Sample Num: 5008, Cur Loss: 0.34546104, Cur Avg Loss: 0.45550855, Log Avg loss: 0.62517755, Global Avg Loss: 1.21874898, Time: 0.0232 Steps: 91860, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000323, Sample Num: 5168, Cur Loss: 0.50535595, Cur Avg Loss: 0.45160052, Log Avg loss: 0.32927917, Global Avg Loss: 1.21865216, Time: 0.0067 Steps: 91870, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000333, Sample Num: 5328, Cur Loss: 0.36831847, Cur Avg Loss: 0.45108714, Log Avg loss: 0.43450475, Global Avg Loss: 1.21856682, Time: 0.0065 Steps: 91880, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000343, Sample Num: 5488, Cur Loss: 0.52893263, Cur Avg Loss: 0.45381994, Log Avg loss: 0.54482219, Global Avg Loss: 1.21849350, Time: 0.0073 Steps: 91890, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000353, Sample Num: 5648, Cur Loss: 0.68970907, Cur Avg Loss: 0.45530710, Log Avg loss: 0.50631676, Global Avg Loss: 1.21841600, Time: 0.0065 Steps: 91900, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000363, Sample Num: 5808, Cur Loss: 0.22025242, Cur Avg Loss: 0.45328214, Log Avg loss: 0.38180103, Global Avg Loss: 1.21832498, Time: 0.0065 Steps: 91910, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000373, Sample Num: 5968, Cur Loss: 0.23390636, Cur Avg Loss: 0.45838384, Log Avg loss: 0.64357551, Global Avg Loss: 1.21826245, Time: 0.0065 Steps: 91920, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000383, Sample Num: 6128, Cur Loss: 0.81036276, Cur Avg Loss: 0.46028178, Log Avg loss: 0.53107494, Global Avg Loss: 1.21818770, Time: 0.0066 Steps: 91930, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000393, Sample Num: 6288, Cur Loss: 0.64512348, Cur Avg Loss: 0.46039797, Log Avg loss: 0.46484827, Global Avg Loss: 1.21810576, Time: 0.0065 Steps: 91940, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000403, Sample Num: 6448, Cur Loss: 0.57207692, Cur Avg Loss: 0.46273091, Log Avg loss: 0.55441528, Global Avg Loss: 1.21803358, Time: 0.0076 Steps: 91950, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000413, Sample Num: 6608, Cur Loss: 0.51311576, Cur Avg Loss: 0.46272191, Log Avg loss: 0.46235941, Global Avg Loss: 1.21795141, Time: 0.0076 Steps: 91960, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000423, Sample Num: 6768, Cur Loss: 0.18197049, Cur Avg Loss: 0.46156159, Log Avg loss: 0.41364036, Global Avg Loss: 1.21786395, Time: 0.0074 Steps: 91970, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000433, Sample Num: 6928, Cur Loss: 0.34913725, Cur Avg Loss: 0.45929854, Log Avg loss: 0.36357145, Global Avg Loss: 1.21777108, Time: 0.0077 Steps: 91980, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000443, Sample Num: 7088, Cur Loss: 0.27036813, Cur Avg Loss: 0.45984795, Log Avg loss: 0.48363721, Global Avg Loss: 1.21769127, Time: 0.0077 Steps: 91990, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000453, Sample Num: 7248, Cur Loss: 0.32734501, Cur Avg Loss: 0.45954246, Log Avg loss: 0.44600949, Global Avg Loss: 1.21760739, Time: 0.0077 Steps: 92000, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000463, Sample Num: 7408, Cur Loss: 0.33325708, Cur Avg Loss: 0.46040817, Log Avg loss: 0.49962451, Global Avg Loss: 1.21752936, Time: 0.0075 Steps: 92010, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000473, Sample Num: 7568, Cur Loss: 0.38282883, Cur Avg Loss: 0.46025301, Log Avg loss: 0.45306909, Global Avg Loss: 1.21744628, Time: 0.0077 Steps: 92020, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000483, Sample Num: 7728, Cur Loss: 0.17526312, Cur Avg Loss: 0.45967176, Log Avg loss: 0.43217901, Global Avg Loss: 1.21736096, Time: 0.0075 Steps: 92030, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000493, Sample Num: 7888, Cur Loss: 0.65973514, Cur Avg Loss: 0.46062880, Log Avg loss: 0.50685354, Global Avg Loss: 1.21728376, Time: 0.0078 Steps: 92040, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000503, Sample Num: 8048, Cur Loss: 0.21944639, Cur Avg Loss: 0.45856684, Log Avg loss: 0.35691232, Global Avg Loss: 1.21719029, Time: 0.0075 Steps: 92050, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000513, Sample Num: 8208, Cur Loss: 0.20251903, Cur Avg Loss: 0.45775372, Log Avg loss: 0.41685377, Global Avg Loss: 1.21710336, Time: 0.0075 Steps: 92060, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000523, Sample Num: 8368, Cur Loss: 0.13423271, Cur Avg Loss: 0.45350407, Log Avg loss: 0.23549719, Global Avg Loss: 1.21699674, Time: 0.0075 Steps: 92070, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000533, Sample Num: 8528, Cur Loss: 0.41155785, Cur Avg Loss: 0.45093633, Log Avg loss: 0.31664323, Global Avg Loss: 1.21689896, Time: 0.0076 Steps: 92080, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000543, Sample Num: 8688, Cur Loss: 0.86803865, Cur Avg Loss: 0.45353853, Log Avg loss: 0.59223577, Global Avg Loss: 1.21683113, Time: 0.0076 Steps: 92090, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000553, Sample Num: 8848, Cur Loss: 0.26288098, Cur Avg Loss: 0.45270156, Log Avg loss: 0.40725430, Global Avg Loss: 1.21674323, Time: 0.0075 Steps: 92100, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000563, Sample Num: 9008, Cur Loss: 0.76555312, Cur Avg Loss: 0.45073101, Log Avg loss: 0.34175929, Global Avg Loss: 1.21664823, Time: 0.0076 Steps: 92110, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000573, Sample Num: 9168, Cur Loss: 0.18252638, Cur Avg Loss: 0.45010562, Log Avg loss: 0.41489625, Global Avg Loss: 1.21656120, Time: 0.0077 Steps: 92120, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000583, Sample Num: 9328, Cur Loss: 0.44294870, Cur Avg Loss: 0.45253233, Log Avg loss: 0.59158306, Global Avg Loss: 1.21649336, Time: 0.0077 Steps: 92130, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000593, Sample Num: 9488, Cur Loss: 0.38909727, Cur Avg Loss: 0.45154226, Log Avg loss: 0.39382091, Global Avg Loss: 1.21640408, Time: 0.0076 Steps: 92140, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000603, Sample Num: 9648, Cur Loss: 0.25346151, Cur Avg Loss: 0.45028804, Log Avg loss: 0.37591299, Global Avg Loss: 1.21631287, Time: 0.0075 Steps: 92150, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000613, Sample Num: 9808, Cur Loss: 0.30937445, Cur Avg Loss: 0.44904037, Log Avg loss: 0.37380596, Global Avg Loss: 1.21622145, Time: 0.0076 Steps: 92160, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000623, Sample Num: 9968, Cur Loss: 0.50869685, Cur Avg Loss: 0.45020949, Log Avg loss: 0.52187617, Global Avg Loss: 1.21614612, Time: 0.0075 Steps: 92170, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000633, Sample Num: 10128, Cur Loss: 0.56958580, Cur Avg Loss: 0.44887869, Log Avg loss: 0.36596998, Global Avg Loss: 1.21605389, Time: 0.0075 Steps: 92180, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000643, Sample Num: 10288, Cur Loss: 0.23776451, Cur Avg Loss: 0.45010959, Log Avg loss: 0.52802570, Global Avg Loss: 1.21597926, Time: 0.0075 Steps: 92190, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000653, Sample Num: 10448, Cur Loss: 0.26237899, Cur Avg Loss: 0.45239881, Log Avg loss: 0.59959577, Global Avg Loss: 1.21591240, Time: 0.0075 Steps: 92200, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000663, Sample Num: 10608, Cur Loss: 0.64983141, Cur Avg Loss: 0.45267448, Log Avg loss: 0.47067556, Global Avg Loss: 1.21583159, Time: 0.0076 Steps: 92210, Updated lr: 0.000014 Training, Epoch: 0044, Batch: 000673, Sample Num: 10768, Cur Loss: 0.32587576, Cur Avg Loss: 0.45430082, Log Avg loss: 0.56212715, Global Avg Loss: 1.21576070, Time: 0.0077 Steps: 92220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000683, Sample Num: 10928, Cur Loss: 0.35085773, Cur Avg Loss: 0.45594113, Log Avg loss: 0.56633412, Global Avg Loss: 1.21569029, Time: 0.0075 Steps: 92230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000693, Sample Num: 11088, Cur Loss: 0.27465224, Cur Avg Loss: 0.45575632, Log Avg loss: 0.44313405, Global Avg Loss: 1.21560653, Time: 0.0075 Steps: 92240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000703, Sample Num: 11248, Cur Loss: 0.30099368, Cur Avg Loss: 0.45459058, Log Avg loss: 0.37380435, Global Avg Loss: 1.21551528, Time: 0.0076 Steps: 92250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000713, Sample Num: 11408, Cur Loss: 0.38790670, Cur Avg Loss: 0.45503277, Log Avg loss: 0.48611924, Global Avg Loss: 1.21543622, Time: 0.0075 Steps: 92260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000723, Sample Num: 11568, Cur Loss: 0.40645272, Cur Avg Loss: 0.45763786, Log Avg loss: 0.64338064, Global Avg Loss: 1.21537422, Time: 0.0075 Steps: 92270, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000733, Sample Num: 11728, Cur Loss: 0.57660121, Cur Avg Loss: 0.45642727, Log Avg loss: 0.36890129, Global Avg Loss: 1.21528249, Time: 0.0075 Steps: 92280, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000743, Sample Num: 11888, Cur Loss: 0.39038515, Cur Avg Loss: 0.45678214, Log Avg loss: 0.48279450, Global Avg Loss: 1.21520313, Time: 0.0075 Steps: 92290, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000753, Sample Num: 12048, Cur Loss: 0.45813167, Cur Avg Loss: 0.45611405, Log Avg loss: 0.40647437, Global Avg Loss: 1.21511551, Time: 0.0074 Steps: 92300, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000763, Sample Num: 12208, Cur Loss: 0.60798961, Cur Avg Loss: 0.45615361, Log Avg loss: 0.45913315, Global Avg Loss: 1.21503361, Time: 0.0077 Steps: 92310, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000773, Sample Num: 12368, Cur Loss: 0.61239249, Cur Avg Loss: 0.45747140, Log Avg loss: 0.55801841, Global Avg Loss: 1.21496244, Time: 0.0075 Steps: 92320, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000783, Sample Num: 12528, Cur Loss: 0.28676748, Cur Avg Loss: 0.45649181, Log Avg loss: 0.38076916, Global Avg Loss: 1.21487209, Time: 0.0074 Steps: 92330, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000793, Sample Num: 12688, Cur Loss: 0.88966846, Cur Avg Loss: 0.45701479, Log Avg loss: 0.49796484, Global Avg Loss: 1.21479446, Time: 0.0075 Steps: 92340, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000803, Sample Num: 12848, Cur Loss: 0.41353565, Cur Avg Loss: 0.45745706, Log Avg loss: 0.49252864, Global Avg Loss: 1.21471625, Time: 0.0074 Steps: 92350, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000813, Sample Num: 13008, Cur Loss: 0.57635266, Cur Avg Loss: 0.45860800, Log Avg loss: 0.55102840, Global Avg Loss: 1.21464439, Time: 0.0073 Steps: 92360, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000823, Sample Num: 13168, Cur Loss: 0.38933843, Cur Avg Loss: 0.45940261, Log Avg loss: 0.52400485, Global Avg Loss: 1.21456962, Time: 0.0074 Steps: 92370, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000833, Sample Num: 13328, Cur Loss: 0.22903104, Cur Avg Loss: 0.45811752, Log Avg loss: 0.35235440, Global Avg Loss: 1.21447628, Time: 0.0074 Steps: 92380, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000843, Sample Num: 13488, Cur Loss: 0.30856210, Cur Avg Loss: 0.45857283, Log Avg loss: 0.49649984, Global Avg Loss: 1.21439857, Time: 0.0074 Steps: 92390, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000853, Sample Num: 13648, Cur Loss: 0.73427922, Cur Avg Loss: 0.45979872, Log Avg loss: 0.56314122, Global Avg Loss: 1.21432809, Time: 0.0074 Steps: 92400, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000863, Sample Num: 13808, Cur Loss: 0.28855979, Cur Avg Loss: 0.45925866, Log Avg loss: 0.41319218, Global Avg Loss: 1.21424140, Time: 0.0073 Steps: 92410, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000873, Sample Num: 13968, Cur Loss: 0.42227924, Cur Avg Loss: 0.45970766, Log Avg loss: 0.49845576, Global Avg Loss: 1.21416395, Time: 0.0073 Steps: 92420, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000883, Sample Num: 14128, Cur Loss: 0.77166545, Cur Avg Loss: 0.45998788, Log Avg loss: 0.48445148, Global Avg Loss: 1.21408500, Time: 0.0075 Steps: 92430, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000893, Sample Num: 14288, Cur Loss: 1.12610602, Cur Avg Loss: 0.46026053, Log Avg loss: 0.48433519, Global Avg Loss: 1.21400606, Time: 0.0074 Steps: 92440, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000903, Sample Num: 14448, Cur Loss: 0.15301332, Cur Avg Loss: 0.46060844, Log Avg loss: 0.49167724, Global Avg Loss: 1.21392793, Time: 0.0073 Steps: 92450, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000913, Sample Num: 14608, Cur Loss: 0.77381217, Cur Avg Loss: 0.46179990, Log Avg loss: 0.56938861, Global Avg Loss: 1.21385822, Time: 0.0075 Steps: 92460, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000923, Sample Num: 14768, Cur Loss: 0.38714460, Cur Avg Loss: 0.46260781, Log Avg loss: 0.53636940, Global Avg Loss: 1.21378495, Time: 0.0074 Steps: 92470, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000933, Sample Num: 14928, Cur Loss: 0.35147047, Cur Avg Loss: 0.46263251, Log Avg loss: 0.46491291, Global Avg Loss: 1.21370397, Time: 0.0074 Steps: 92480, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000943, Sample Num: 15088, Cur Loss: 0.37378001, Cur Avg Loss: 0.46219912, Log Avg loss: 0.42176400, Global Avg Loss: 1.21361835, Time: 0.0074 Steps: 92490, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000953, Sample Num: 15248, Cur Loss: 0.36607891, Cur Avg Loss: 0.46167485, Log Avg loss: 0.41223629, Global Avg Loss: 1.21353171, Time: 0.0073 Steps: 92500, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000963, Sample Num: 15408, Cur Loss: 0.50631738, Cur Avg Loss: 0.46254182, Log Avg loss: 0.54516411, Global Avg Loss: 1.21345946, Time: 0.0074 Steps: 92510, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000973, Sample Num: 15568, Cur Loss: 1.18596792, Cur Avg Loss: 0.46378937, Log Avg loss: 0.58392825, Global Avg Loss: 1.21339142, Time: 0.0074 Steps: 92520, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000983, Sample Num: 15728, Cur Loss: 0.40443105, Cur Avg Loss: 0.46369155, Log Avg loss: 0.45417304, Global Avg Loss: 1.21330937, Time: 0.0073 Steps: 92530, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 000993, Sample Num: 15888, Cur Loss: 0.31789529, Cur Avg Loss: 0.46424707, Log Avg loss: 0.51885495, Global Avg Loss: 1.21323433, Time: 0.0073 Steps: 92540, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001003, Sample Num: 16048, Cur Loss: 0.72474277, Cur Avg Loss: 0.46345679, Log Avg loss: 0.38498205, Global Avg Loss: 1.21314483, Time: 0.0225 Steps: 92550, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001013, Sample Num: 16208, Cur Loss: 0.28394488, Cur Avg Loss: 0.46340446, Log Avg loss: 0.45815560, Global Avg Loss: 1.21306327, Time: 0.0065 Steps: 92560, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001023, Sample Num: 16368, Cur Loss: 0.40143836, Cur Avg Loss: 0.46344358, Log Avg loss: 0.46740688, Global Avg Loss: 1.21298272, Time: 0.0104 Steps: 92570, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001033, Sample Num: 16528, Cur Loss: 0.62447804, Cur Avg Loss: 0.46342798, Log Avg loss: 0.46183217, Global Avg Loss: 1.21290158, Time: 0.0075 Steps: 92580, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001043, Sample Num: 16688, Cur Loss: 0.16703546, Cur Avg Loss: 0.46275951, Log Avg loss: 0.39370644, Global Avg Loss: 1.21281311, Time: 0.0209 Steps: 92590, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001053, Sample Num: 16848, Cur Loss: 0.38305894, Cur Avg Loss: 0.46231207, Log Avg loss: 0.41564409, Global Avg Loss: 1.21272702, Time: 0.0158 Steps: 92600, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001063, Sample Num: 17008, Cur Loss: 0.30981711, Cur Avg Loss: 0.46356521, Log Avg loss: 0.59552043, Global Avg Loss: 1.21266037, Time: 0.0146 Steps: 92610, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001073, Sample Num: 17168, Cur Loss: 0.29193950, Cur Avg Loss: 0.46413780, Log Avg loss: 0.52500438, Global Avg Loss: 1.21258613, Time: 0.0067 Steps: 92620, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001083, Sample Num: 17328, Cur Loss: 0.74662817, Cur Avg Loss: 0.46532081, Log Avg loss: 0.59225768, Global Avg Loss: 1.21251916, Time: 0.0065 Steps: 92630, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001093, Sample Num: 17488, Cur Loss: 0.37331992, Cur Avg Loss: 0.46475310, Log Avg loss: 0.40327050, Global Avg Loss: 1.21243181, Time: 0.0189 Steps: 92640, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001103, Sample Num: 17648, Cur Loss: 0.20717150, Cur Avg Loss: 0.46336044, Log Avg loss: 0.31114204, Global Avg Loss: 1.21233453, Time: 0.0225 Steps: 92650, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001113, Sample Num: 17808, Cur Loss: 0.58156157, Cur Avg Loss: 0.46358263, Log Avg loss: 0.48809101, Global Avg Loss: 1.21225636, Time: 0.0066 Steps: 92660, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001123, Sample Num: 17968, Cur Loss: 0.44787210, Cur Avg Loss: 0.46298181, Log Avg loss: 0.39611025, Global Avg Loss: 1.21216829, Time: 0.0072 Steps: 92670, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001133, Sample Num: 18128, Cur Loss: 0.46658748, Cur Avg Loss: 0.46294988, Log Avg loss: 0.45936398, Global Avg Loss: 1.21208707, Time: 0.0066 Steps: 92680, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001143, Sample Num: 18288, Cur Loss: 0.51489556, Cur Avg Loss: 0.46330164, Log Avg loss: 0.50315614, Global Avg Loss: 1.21201058, Time: 0.0065 Steps: 92690, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001153, Sample Num: 18448, Cur Loss: 1.19198191, Cur Avg Loss: 0.46311135, Log Avg loss: 0.44136081, Global Avg Loss: 1.21192745, Time: 0.0073 Steps: 92700, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001163, Sample Num: 18608, Cur Loss: 0.51086873, Cur Avg Loss: 0.46278159, Log Avg loss: 0.42476007, Global Avg Loss: 1.21184254, Time: 0.0115 Steps: 92710, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001173, Sample Num: 18768, Cur Loss: 0.13288328, Cur Avg Loss: 0.46342686, Log Avg loss: 0.53847256, Global Avg Loss: 1.21176992, Time: 0.0194 Steps: 92720, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001183, Sample Num: 18928, Cur Loss: 0.29369679, Cur Avg Loss: 0.46318366, Log Avg loss: 0.43465612, Global Avg Loss: 1.21168612, Time: 0.0233 Steps: 92730, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001193, Sample Num: 19088, Cur Loss: 0.22714463, Cur Avg Loss: 0.46214156, Log Avg loss: 0.33886151, Global Avg Loss: 1.21159200, Time: 0.0065 Steps: 92740, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001203, Sample Num: 19248, Cur Loss: 0.44341150, Cur Avg Loss: 0.46258562, Log Avg loss: 0.51556132, Global Avg Loss: 1.21151696, Time: 0.0070 Steps: 92750, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001213, Sample Num: 19408, Cur Loss: 0.47857720, Cur Avg Loss: 0.46356229, Log Avg loss: 0.58105554, Global Avg Loss: 1.21144899, Time: 0.0065 Steps: 92760, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001223, Sample Num: 19568, Cur Loss: 0.31551725, Cur Avg Loss: 0.46366671, Log Avg loss: 0.47633311, Global Avg Loss: 1.21136975, Time: 0.0064 Steps: 92770, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001233, Sample Num: 19728, Cur Loss: 0.14659879, Cur Avg Loss: 0.46536101, Log Avg loss: 0.67257355, Global Avg Loss: 1.21131168, Time: 0.0075 Steps: 92780, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001243, Sample Num: 19888, Cur Loss: 0.45399767, Cur Avg Loss: 0.46519087, Log Avg loss: 0.44421322, Global Avg Loss: 1.21122901, Time: 0.0066 Steps: 92790, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001253, Sample Num: 20048, Cur Loss: 0.80615830, Cur Avg Loss: 0.46566974, Log Avg loss: 0.52519340, Global Avg Loss: 1.21115508, Time: 0.0067 Steps: 92800, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001263, Sample Num: 20208, Cur Loss: 0.15204649, Cur Avg Loss: 0.46465352, Log Avg loss: 0.33732132, Global Avg Loss: 1.21106093, Time: 0.0067 Steps: 92810, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001273, Sample Num: 20368, Cur Loss: 0.92050171, Cur Avg Loss: 0.46565164, Log Avg loss: 0.59171362, Global Avg Loss: 1.21099420, Time: 0.0067 Steps: 92820, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001283, Sample Num: 20528, Cur Loss: 0.31998158, Cur Avg Loss: 0.46564836, Log Avg loss: 0.46523051, Global Avg Loss: 1.21091387, Time: 0.0068 Steps: 92830, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001293, Sample Num: 20688, Cur Loss: 0.27388754, Cur Avg Loss: 0.46445098, Log Avg loss: 0.31082728, Global Avg Loss: 1.21081692, Time: 0.0066 Steps: 92840, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001303, Sample Num: 20848, Cur Loss: 0.41529950, Cur Avg Loss: 0.46440301, Log Avg loss: 0.45820112, Global Avg Loss: 1.21073586, Time: 0.0066 Steps: 92850, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001313, Sample Num: 21008, Cur Loss: 0.42584157, Cur Avg Loss: 0.46443171, Log Avg loss: 0.46817053, Global Avg Loss: 1.21065589, Time: 0.0066 Steps: 92860, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001323, Sample Num: 21168, Cur Loss: 0.65490556, Cur Avg Loss: 0.46480188, Log Avg loss: 0.51340499, Global Avg Loss: 1.21058081, Time: 0.0066 Steps: 92870, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001333, Sample Num: 21328, Cur Loss: 0.49825418, Cur Avg Loss: 0.46393208, Log Avg loss: 0.34885781, Global Avg Loss: 1.21048804, Time: 0.0066 Steps: 92880, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001343, Sample Num: 21488, Cur Loss: 0.12678911, Cur Avg Loss: 0.46384807, Log Avg loss: 0.45265016, Global Avg Loss: 1.21040645, Time: 0.0075 Steps: 92890, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001353, Sample Num: 21648, Cur Loss: 0.50661641, Cur Avg Loss: 0.46288902, Log Avg loss: 0.33408864, Global Avg Loss: 1.21031212, Time: 0.0075 Steps: 92900, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001363, Sample Num: 21808, Cur Loss: 0.46645314, Cur Avg Loss: 0.46329335, Log Avg loss: 0.51799929, Global Avg Loss: 1.21023761, Time: 0.0075 Steps: 92910, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001373, Sample Num: 21968, Cur Loss: 0.67580050, Cur Avg Loss: 0.46376596, Log Avg loss: 0.52818197, Global Avg Loss: 1.21016421, Time: 0.0075 Steps: 92920, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001383, Sample Num: 22128, Cur Loss: 0.22820184, Cur Avg Loss: 0.46363006, Log Avg loss: 0.44497095, Global Avg Loss: 1.21008186, Time: 0.0076 Steps: 92930, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001393, Sample Num: 22288, Cur Loss: 0.39515764, Cur Avg Loss: 0.46299563, Log Avg loss: 0.37525454, Global Avg Loss: 1.20999204, Time: 0.0075 Steps: 92940, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001403, Sample Num: 22448, Cur Loss: 0.46542066, Cur Avg Loss: 0.46366232, Log Avg loss: 0.55653246, Global Avg Loss: 1.20992174, Time: 0.0076 Steps: 92950, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001413, Sample Num: 22608, Cur Loss: 0.45206997, Cur Avg Loss: 0.46360191, Log Avg loss: 0.45512572, Global Avg Loss: 1.20984054, Time: 0.0076 Steps: 92960, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001423, Sample Num: 22768, Cur Loss: 1.55551004, Cur Avg Loss: 0.46473350, Log Avg loss: 0.62462717, Global Avg Loss: 1.20977760, Time: 0.0074 Steps: 92970, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001433, Sample Num: 22928, Cur Loss: 0.29075029, Cur Avg Loss: 0.46544968, Log Avg loss: 0.56736233, Global Avg Loss: 1.20970850, Time: 0.0076 Steps: 92980, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001443, Sample Num: 23088, Cur Loss: 0.31669533, Cur Avg Loss: 0.46453139, Log Avg loss: 0.33294037, Global Avg Loss: 1.20961422, Time: 0.0075 Steps: 92990, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001453, Sample Num: 23248, Cur Loss: 1.15798318, Cur Avg Loss: 0.46568025, Log Avg loss: 0.63146095, Global Avg Loss: 1.20955205, Time: 0.0075 Steps: 93000, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001463, Sample Num: 23408, Cur Loss: 0.56865078, Cur Avg Loss: 0.46579069, Log Avg loss: 0.48183710, Global Avg Loss: 1.20947381, Time: 0.0075 Steps: 93010, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001473, Sample Num: 23568, Cur Loss: 0.51076829, Cur Avg Loss: 0.46636594, Log Avg loss: 0.55052464, Global Avg Loss: 1.20940297, Time: 0.0075 Steps: 93020, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001483, Sample Num: 23728, Cur Loss: 0.40955311, Cur Avg Loss: 0.46596301, Log Avg loss: 0.40661261, Global Avg Loss: 1.20931668, Time: 0.0075 Steps: 93030, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001493, Sample Num: 23888, Cur Loss: 0.18707225, Cur Avg Loss: 0.46573625, Log Avg loss: 0.43210732, Global Avg Loss: 1.20923314, Time: 0.0075 Steps: 93040, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001503, Sample Num: 24048, Cur Loss: 0.94940323, Cur Avg Loss: 0.46578644, Log Avg loss: 0.47327897, Global Avg Loss: 1.20915405, Time: 0.0076 Steps: 93050, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001513, Sample Num: 24208, Cur Loss: 0.66613263, Cur Avg Loss: 0.46577733, Log Avg loss: 0.46440950, Global Avg Loss: 1.20907402, Time: 0.0076 Steps: 93060, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001523, Sample Num: 24368, Cur Loss: 0.25397015, Cur Avg Loss: 0.46537625, Log Avg loss: 0.40469146, Global Avg Loss: 1.20898759, Time: 0.0076 Steps: 93070, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001533, Sample Num: 24528, Cur Loss: 0.94224632, Cur Avg Loss: 0.46680596, Log Avg loss: 0.68455105, Global Avg Loss: 1.20893125, Time: 0.0076 Steps: 93080, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001543, Sample Num: 24688, Cur Loss: 0.43938634, Cur Avg Loss: 0.46766347, Log Avg loss: 0.59911955, Global Avg Loss: 1.20886574, Time: 0.0075 Steps: 93090, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001553, Sample Num: 24848, Cur Loss: 0.29248714, Cur Avg Loss: 0.46731321, Log Avg loss: 0.41326913, Global Avg Loss: 1.20878029, Time: 0.0074 Steps: 93100, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001563, Sample Num: 25008, Cur Loss: 0.67345697, Cur Avg Loss: 0.46871294, Log Avg loss: 0.68609079, Global Avg Loss: 1.20872415, Time: 0.0075 Steps: 93110, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001573, Sample Num: 25168, Cur Loss: 0.28049719, Cur Avg Loss: 0.46941595, Log Avg loss: 0.57929564, Global Avg Loss: 1.20865656, Time: 0.0075 Steps: 93120, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001583, Sample Num: 25328, Cur Loss: 0.40394783, Cur Avg Loss: 0.47006481, Log Avg loss: 0.57213124, Global Avg Loss: 1.20858821, Time: 0.0075 Steps: 93130, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001593, Sample Num: 25488, Cur Loss: 0.44903007, Cur Avg Loss: 0.46995378, Log Avg loss: 0.45237838, Global Avg Loss: 1.20850702, Time: 0.0076 Steps: 93140, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001603, Sample Num: 25648, Cur Loss: 0.39817056, Cur Avg Loss: 0.46961500, Log Avg loss: 0.41564671, Global Avg Loss: 1.20842190, Time: 0.0074 Steps: 93150, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001613, Sample Num: 25808, Cur Loss: 0.76816207, Cur Avg Loss: 0.46973098, Log Avg loss: 0.48832213, Global Avg Loss: 1.20834460, Time: 0.0075 Steps: 93160, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001623, Sample Num: 25968, Cur Loss: 0.42581508, Cur Avg Loss: 0.46976678, Log Avg loss: 0.47554154, Global Avg Loss: 1.20826595, Time: 0.0075 Steps: 93170, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001633, Sample Num: 26128, Cur Loss: 0.55167377, Cur Avg Loss: 0.46969252, Log Avg loss: 0.45764084, Global Avg Loss: 1.20818540, Time: 0.0076 Steps: 93180, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001643, Sample Num: 26288, Cur Loss: 0.55702317, Cur Avg Loss: 0.46955906, Log Avg loss: 0.44776498, Global Avg Loss: 1.20810380, Time: 0.0074 Steps: 93190, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001653, Sample Num: 26448, Cur Loss: 0.32949898, Cur Avg Loss: 0.46948367, Log Avg loss: 0.45709723, Global Avg Loss: 1.20802322, Time: 0.0074 Steps: 93200, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001663, Sample Num: 26608, Cur Loss: 0.58407176, Cur Avg Loss: 0.46955234, Log Avg loss: 0.48090303, Global Avg Loss: 1.20794521, Time: 0.0074 Steps: 93210, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001673, Sample Num: 26768, Cur Loss: 0.51085907, Cur Avg Loss: 0.46928776, Log Avg loss: 0.42528858, Global Avg Loss: 1.20786125, Time: 0.0075 Steps: 93220, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001683, Sample Num: 26928, Cur Loss: 0.57034934, Cur Avg Loss: 0.46896171, Log Avg loss: 0.41441307, Global Avg Loss: 1.20777614, Time: 0.0074 Steps: 93230, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001693, Sample Num: 27088, Cur Loss: 0.60439396, Cur Avg Loss: 0.46942846, Log Avg loss: 0.54798256, Global Avg Loss: 1.20770538, Time: 0.0074 Steps: 93240, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001703, Sample Num: 27248, Cur Loss: 0.79849696, Cur Avg Loss: 0.46950114, Log Avg loss: 0.48180621, Global Avg Loss: 1.20762754, Time: 0.0074 Steps: 93250, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001713, Sample Num: 27408, Cur Loss: 0.18256946, Cur Avg Loss: 0.46976427, Log Avg loss: 0.51457395, Global Avg Loss: 1.20755322, Time: 0.0075 Steps: 93260, Updated lr: 0.000013 Training, Epoch: 0044, Batch: 001723, Sample Num: 27568, Cur Loss: 0.42147654, Cur Avg Loss: 0.46975884, Log Avg loss: 0.46882870, Global Avg Loss: 1.20747402, Time: 0.0076 Steps: 93270, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001733, Sample Num: 27728, Cur Loss: 0.35401934, Cur Avg Loss: 0.47010339, Log Avg loss: 0.52947038, Global Avg Loss: 1.20740133, Time: 0.0075 Steps: 93280, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001743, Sample Num: 27888, Cur Loss: 0.42389557, Cur Avg Loss: 0.47016086, Log Avg loss: 0.48012019, Global Avg Loss: 1.20732338, Time: 0.0074 Steps: 93290, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001753, Sample Num: 28048, Cur Loss: 0.71626449, Cur Avg Loss: 0.47044005, Log Avg loss: 0.51910241, Global Avg Loss: 1.20724961, Time: 0.0075 Steps: 93300, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001763, Sample Num: 28208, Cur Loss: 0.16604871, Cur Avg Loss: 0.47015593, Log Avg loss: 0.42035013, Global Avg Loss: 1.20716528, Time: 0.0074 Steps: 93310, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001773, Sample Num: 28368, Cur Loss: 1.06910086, Cur Avg Loss: 0.47000699, Log Avg loss: 0.44374865, Global Avg Loss: 1.20708347, Time: 0.0074 Steps: 93320, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001783, Sample Num: 28528, Cur Loss: 0.47785333, Cur Avg Loss: 0.46937889, Log Avg loss: 0.35801659, Global Avg Loss: 1.20699250, Time: 0.0074 Steps: 93330, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001793, Sample Num: 28688, Cur Loss: 0.23641041, Cur Avg Loss: 0.46916575, Log Avg loss: 0.43116393, Global Avg Loss: 1.20690938, Time: 0.0075 Steps: 93340, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001803, Sample Num: 28848, Cur Loss: 0.27602929, Cur Avg Loss: 0.46951879, Log Avg loss: 0.53281780, Global Avg Loss: 1.20683717, Time: 0.0074 Steps: 93350, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001813, Sample Num: 29008, Cur Loss: 0.22225925, Cur Avg Loss: 0.46963302, Log Avg loss: 0.49022940, Global Avg Loss: 1.20676041, Time: 0.0075 Steps: 93360, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001823, Sample Num: 29168, Cur Loss: 0.44501066, Cur Avg Loss: 0.47008712, Log Avg loss: 0.55241477, Global Avg Loss: 1.20669033, Time: 0.0074 Steps: 93370, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001833, Sample Num: 29328, Cur Loss: 0.75374031, Cur Avg Loss: 0.47004332, Log Avg loss: 0.46205949, Global Avg Loss: 1.20661059, Time: 0.0074 Steps: 93380, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001843, Sample Num: 29488, Cur Loss: 0.51886213, Cur Avg Loss: 0.46943164, Log Avg loss: 0.35730973, Global Avg Loss: 1.20651965, Time: 0.0083 Steps: 93390, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001853, Sample Num: 29648, Cur Loss: 0.42983714, Cur Avg Loss: 0.46939859, Log Avg loss: 0.46330860, Global Avg Loss: 1.20644007, Time: 0.0100 Steps: 93400, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001863, Sample Num: 29808, Cur Loss: 0.16397700, Cur Avg Loss: 0.46920433, Log Avg loss: 0.43320720, Global Avg Loss: 1.20635730, Time: 0.0065 Steps: 93410, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001873, Sample Num: 29968, Cur Loss: 1.14113307, Cur Avg Loss: 0.46913220, Log Avg loss: 0.45569456, Global Avg Loss: 1.20627694, Time: 0.0064 Steps: 93420, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001883, Sample Num: 30128, Cur Loss: 0.19806564, Cur Avg Loss: 0.46875666, Log Avg loss: 0.39841721, Global Avg Loss: 1.20619048, Time: 0.0067 Steps: 93430, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001893, Sample Num: 30288, Cur Loss: 0.29249680, Cur Avg Loss: 0.46827692, Log Avg loss: 0.37794160, Global Avg Loss: 1.20610184, Time: 0.0064 Steps: 93440, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001903, Sample Num: 30448, Cur Loss: 0.32093552, Cur Avg Loss: 0.46832740, Log Avg loss: 0.47788454, Global Avg Loss: 1.20602391, Time: 0.0065 Steps: 93450, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001913, Sample Num: 30608, Cur Loss: 0.59272802, Cur Avg Loss: 0.46789804, Log Avg loss: 0.38618990, Global Avg Loss: 1.20593619, Time: 0.0076 Steps: 93460, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001923, Sample Num: 30768, Cur Loss: 0.35860515, Cur Avg Loss: 0.46761106, Log Avg loss: 0.41271185, Global Avg Loss: 1.20585133, Time: 0.0065 Steps: 93470, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001933, Sample Num: 30928, Cur Loss: 1.02463138, Cur Avg Loss: 0.46761561, Log Avg loss: 0.46849140, Global Avg Loss: 1.20577245, Time: 0.0063 Steps: 93480, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001943, Sample Num: 31088, Cur Loss: 0.36055893, Cur Avg Loss: 0.46804762, Log Avg loss: 0.55155542, Global Avg Loss: 1.20570247, Time: 0.0064 Steps: 93490, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001953, Sample Num: 31248, Cur Loss: 0.71350050, Cur Avg Loss: 0.46746057, Log Avg loss: 0.35339629, Global Avg Loss: 1.20561131, Time: 0.0067 Steps: 93500, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001963, Sample Num: 31408, Cur Loss: 0.33699727, Cur Avg Loss: 0.46771654, Log Avg loss: 0.51770751, Global Avg Loss: 1.20553775, Time: 0.0065 Steps: 93510, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001973, Sample Num: 31568, Cur Loss: 0.25765628, Cur Avg Loss: 0.46802868, Log Avg loss: 0.52930237, Global Avg Loss: 1.20546544, Time: 0.0073 Steps: 93520, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001983, Sample Num: 31728, Cur Loss: 0.31568116, Cur Avg Loss: 0.46814598, Log Avg loss: 0.49128862, Global Avg Loss: 1.20538908, Time: 0.0070 Steps: 93530, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 001993, Sample Num: 31888, Cur Loss: 0.77282536, Cur Avg Loss: 0.46877875, Log Avg loss: 0.59425696, Global Avg Loss: 1.20532375, Time: 0.0221 Steps: 93540, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002003, Sample Num: 32048, Cur Loss: 0.32314321, Cur Avg Loss: 0.46880382, Log Avg loss: 0.47380087, Global Avg Loss: 1.20524555, Time: 0.0091 Steps: 93550, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002013, Sample Num: 32208, Cur Loss: 0.74446660, Cur Avg Loss: 0.46919160, Log Avg loss: 0.54686408, Global Avg Loss: 1.20517518, Time: 0.0067 Steps: 93560, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002023, Sample Num: 32368, Cur Loss: 0.32374203, Cur Avg Loss: 0.46887120, Log Avg loss: 0.40437385, Global Avg Loss: 1.20508960, Time: 0.0065 Steps: 93570, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002033, Sample Num: 32528, Cur Loss: 0.26115763, Cur Avg Loss: 0.46849249, Log Avg loss: 0.39187848, Global Avg Loss: 1.20500270, Time: 0.0064 Steps: 93580, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002043, Sample Num: 32688, Cur Loss: 0.82710177, Cur Avg Loss: 0.46882269, Log Avg loss: 0.53595234, Global Avg Loss: 1.20493121, Time: 0.0065 Steps: 93590, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002053, Sample Num: 32848, Cur Loss: 0.37427631, Cur Avg Loss: 0.46873865, Log Avg loss: 0.45156949, Global Avg Loss: 1.20485072, Time: 0.0075 Steps: 93600, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002063, Sample Num: 33008, Cur Loss: 0.69627237, Cur Avg Loss: 0.46861601, Log Avg loss: 0.44343788, Global Avg Loss: 1.20476938, Time: 0.0074 Steps: 93610, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002073, Sample Num: 33168, Cur Loss: 0.59756374, Cur Avg Loss: 0.46834304, Log Avg loss: 0.41202937, Global Avg Loss: 1.20468471, Time: 0.0074 Steps: 93620, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002083, Sample Num: 33328, Cur Loss: 0.43743783, Cur Avg Loss: 0.46823086, Log Avg loss: 0.44497729, Global Avg Loss: 1.20460357, Time: 0.0074 Steps: 93630, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002093, Sample Num: 33488, Cur Loss: 0.60643119, Cur Avg Loss: 0.46820572, Log Avg loss: 0.46296766, Global Avg Loss: 1.20452437, Time: 0.0074 Steps: 93640, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002103, Sample Num: 33648, Cur Loss: 0.43143159, Cur Avg Loss: 0.46739485, Log Avg loss: 0.29768153, Global Avg Loss: 1.20442754, Time: 0.0074 Steps: 93650, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002113, Sample Num: 33808, Cur Loss: 0.35990387, Cur Avg Loss: 0.46733250, Log Avg loss: 0.45422043, Global Avg Loss: 1.20434744, Time: 0.0074 Steps: 93660, Updated lr: 0.000012 Training, Epoch: 0044, Batch: 002123, Sample Num: 33968, Cur Loss: 0.18369541, Cur Avg Loss: 0.46706392, Log Avg loss: 0.41031093, Global Avg Loss: 1.20426267, Time: 0.0074 Steps: 93670, Updated lr: 0.000012 ***** Running evaluation checkpoint-93676 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-93676 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 18.165018, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.568603, "eval_total_loss": 399.727799, "eval_mae": 0.566837, "eval_mse": 0.568744, "eval_r2": 0.638469, "eval_sp_statistic": 0.756474, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.800315, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.89115, "test_total_loss": 447.357158, "test_mae": 0.657708, "test_mse": 0.891403, "test_r2": 0.42468, "test_sp_statistic": 0.60192, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.671358, "test_ps_pvalue": 0.0, "lr": 1.2113798008534852e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.2042239213403616, "train_cur_epoch_loss": 995.1727528236806, "train_cur_epoch_avg_loss": 0.4674367087006485, "train_cur_epoch_time": 18.16501784324646, "train_cur_epoch_avg_time": 0.008532183110965928, "epoch": 44, "step": 93676} ################################################## Training, Epoch: 0045, Batch: 000004, Sample Num: 64, Cur Loss: 0.65461004, Cur Avg Loss: 0.53685735, Log Avg loss: 0.57434907, Global Avg Loss: 1.20419543, Time: 0.0075 Steps: 93680, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000014, Sample Num: 224, Cur Loss: 0.93598187, Cur Avg Loss: 0.51523798, Log Avg loss: 0.50659023, Global Avg Loss: 1.20412097, Time: 0.0075 Steps: 93690, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000024, Sample Num: 384, Cur Loss: 0.26450974, Cur Avg Loss: 0.47049819, Log Avg loss: 0.40786248, Global Avg Loss: 1.20403599, Time: 0.0075 Steps: 93700, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000034, Sample Num: 544, Cur Loss: 0.45712018, Cur Avg Loss: 0.48347531, Log Avg loss: 0.51462040, Global Avg Loss: 1.20396242, Time: 0.0076 Steps: 93710, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000044, Sample Num: 704, Cur Loss: 0.44903284, Cur Avg Loss: 0.53752197, Log Avg loss: 0.72128064, Global Avg Loss: 1.20391092, Time: 0.0075 Steps: 93720, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000054, Sample Num: 864, Cur Loss: 0.66670895, Cur Avg Loss: 0.51614655, Log Avg loss: 0.42209469, Global Avg Loss: 1.20382750, Time: 0.0074 Steps: 93730, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000064, Sample Num: 1024, Cur Loss: 0.23821969, Cur Avg Loss: 0.50119121, Log Avg loss: 0.42043234, Global Avg Loss: 1.20374393, Time: 0.0074 Steps: 93740, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000074, Sample Num: 1184, Cur Loss: 0.25518325, Cur Avg Loss: 0.48967764, Log Avg loss: 0.41599084, Global Avg Loss: 1.20365991, Time: 0.0079 Steps: 93750, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000084, Sample Num: 1344, Cur Loss: 0.09376513, Cur Avg Loss: 0.48548858, Log Avg loss: 0.45448953, Global Avg Loss: 1.20358000, Time: 0.0075 Steps: 93760, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000094, Sample Num: 1504, Cur Loss: 0.38846996, Cur Avg Loss: 0.47779553, Log Avg loss: 0.41317389, Global Avg Loss: 1.20349571, Time: 0.0075 Steps: 93770, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000104, Sample Num: 1664, Cur Loss: 0.28288275, Cur Avg Loss: 0.47114142, Log Avg loss: 0.40859277, Global Avg Loss: 1.20341095, Time: 0.0073 Steps: 93780, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000114, Sample Num: 1824, Cur Loss: 0.43877357, Cur Avg Loss: 0.46829087, Log Avg loss: 0.43864515, Global Avg Loss: 1.20332941, Time: 0.0080 Steps: 93790, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000124, Sample Num: 1984, Cur Loss: 0.14464617, Cur Avg Loss: 0.46211532, Log Avg loss: 0.39171407, Global Avg Loss: 1.20324288, Time: 0.0075 Steps: 93800, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000134, Sample Num: 2144, Cur Loss: 0.38999528, Cur Avg Loss: 0.46097890, Log Avg loss: 0.44688734, Global Avg Loss: 1.20316226, Time: 0.0074 Steps: 93810, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000144, Sample Num: 2304, Cur Loss: 0.19150990, Cur Avg Loss: 0.46246641, Log Avg loss: 0.48239903, Global Avg Loss: 1.20308543, Time: 0.0075 Steps: 93820, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000154, Sample Num: 2464, Cur Loss: 0.47230497, Cur Avg Loss: 0.45492815, Log Avg loss: 0.34637719, Global Avg Loss: 1.20299413, Time: 0.0078 Steps: 93830, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000164, Sample Num: 2624, Cur Loss: 0.72151017, Cur Avg Loss: 0.46328876, Log Avg loss: 0.59204216, Global Avg Loss: 1.20292902, Time: 0.0074 Steps: 93840, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000174, Sample Num: 2784, Cur Loss: 0.70408660, Cur Avg Loss: 0.45978169, Log Avg loss: 0.40226577, Global Avg Loss: 1.20284371, Time: 0.0074 Steps: 93850, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000184, Sample Num: 2944, Cur Loss: 0.21496442, Cur Avg Loss: 0.45544643, Log Avg loss: 0.38001279, Global Avg Loss: 1.20275604, Time: 0.0074 Steps: 93860, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000194, Sample Num: 3104, Cur Loss: 0.50345600, Cur Avg Loss: 0.45749054, Log Avg loss: 0.49510230, Global Avg Loss: 1.20268066, Time: 0.0074 Steps: 93870, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000204, Sample Num: 3264, Cur Loss: 0.39953169, Cur Avg Loss: 0.45560517, Log Avg loss: 0.41902896, Global Avg Loss: 1.20259718, Time: 0.0079 Steps: 93880, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000214, Sample Num: 3424, Cur Loss: 0.26799929, Cur Avg Loss: 0.45312066, Log Avg loss: 0.40243657, Global Avg Loss: 1.20251196, Time: 0.0074 Steps: 93890, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000224, Sample Num: 3584, Cur Loss: 0.27258372, Cur Avg Loss: 0.45104420, Log Avg loss: 0.40660809, Global Avg Loss: 1.20242720, Time: 0.0074 Steps: 93900, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000234, Sample Num: 3744, Cur Loss: 0.38807470, Cur Avg Loss: 0.44639821, Log Avg loss: 0.34232795, Global Avg Loss: 1.20233561, Time: 0.0079 Steps: 93910, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000244, Sample Num: 3904, Cur Loss: 0.13571882, Cur Avg Loss: 0.44539141, Log Avg loss: 0.42183219, Global Avg Loss: 1.20225251, Time: 0.0074 Steps: 93920, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000254, Sample Num: 4064, Cur Loss: 0.15742786, Cur Avg Loss: 0.44780169, Log Avg loss: 0.50661258, Global Avg Loss: 1.20217845, Time: 0.0074 Steps: 93930, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000264, Sample Num: 4224, Cur Loss: 0.65330982, Cur Avg Loss: 0.45187795, Log Avg loss: 0.55541491, Global Avg Loss: 1.20210960, Time: 0.0079 Steps: 93940, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000274, Sample Num: 4384, Cur Loss: 0.19237751, Cur Avg Loss: 0.44964105, Log Avg loss: 0.39058698, Global Avg Loss: 1.20202322, Time: 0.0079 Steps: 93950, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000284, Sample Num: 4544, Cur Loss: 0.36979502, Cur Avg Loss: 0.45041600, Log Avg loss: 0.47164966, Global Avg Loss: 1.20194549, Time: 0.0079 Steps: 93960, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000294, Sample Num: 4704, Cur Loss: 0.67114860, Cur Avg Loss: 0.45069427, Log Avg loss: 0.45859706, Global Avg Loss: 1.20186638, Time: 0.0079 Steps: 93970, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000304, Sample Num: 4864, Cur Loss: 0.48054749, Cur Avg Loss: 0.45032401, Log Avg loss: 0.43943843, Global Avg Loss: 1.20178526, Time: 0.0079 Steps: 93980, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000314, Sample Num: 5024, Cur Loss: 0.54062217, Cur Avg Loss: 0.45530654, Log Avg loss: 0.60677548, Global Avg Loss: 1.20172195, Time: 0.0079 Steps: 93990, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000324, Sample Num: 5184, Cur Loss: 0.38244230, Cur Avg Loss: 0.45221562, Log Avg loss: 0.35516067, Global Avg Loss: 1.20163189, Time: 0.0079 Steps: 94000, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000334, Sample Num: 5344, Cur Loss: 0.22435397, Cur Avg Loss: 0.45921013, Log Avg loss: 0.68583233, Global Avg Loss: 1.20157703, Time: 0.0078 Steps: 94010, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000344, Sample Num: 5504, Cur Loss: 0.21529958, Cur Avg Loss: 0.45780092, Log Avg loss: 0.41073316, Global Avg Loss: 1.20149291, Time: 0.0079 Steps: 94020, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000354, Sample Num: 5664, Cur Loss: 0.21398956, Cur Avg Loss: 0.45847654, Log Avg loss: 0.48171781, Global Avg Loss: 1.20141636, Time: 0.0079 Steps: 94030, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000364, Sample Num: 5824, Cur Loss: 0.17050631, Cur Avg Loss: 0.45334071, Log Avg loss: 0.27153263, Global Avg Loss: 1.20131748, Time: 0.0078 Steps: 94040, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000374, Sample Num: 5984, Cur Loss: 0.39872190, Cur Avg Loss: 0.45131797, Log Avg loss: 0.37769006, Global Avg Loss: 1.20122991, Time: 0.0078 Steps: 94050, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000384, Sample Num: 6144, Cur Loss: 0.64478070, Cur Avg Loss: 0.45267420, Log Avg loss: 0.50339731, Global Avg Loss: 1.20115572, Time: 0.0078 Steps: 94060, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000394, Sample Num: 6304, Cur Loss: 0.52593827, Cur Avg Loss: 0.45414353, Log Avg loss: 0.51056553, Global Avg Loss: 1.20108231, Time: 0.0078 Steps: 94070, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000404, Sample Num: 6464, Cur Loss: 0.56044155, Cur Avg Loss: 0.45233988, Log Avg loss: 0.38127608, Global Avg Loss: 1.20099517, Time: 0.0079 Steps: 94080, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000414, Sample Num: 6624, Cur Loss: 0.64204293, Cur Avg Loss: 0.45401938, Log Avg loss: 0.52187128, Global Avg Loss: 1.20092299, Time: 0.0079 Steps: 94090, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000424, Sample Num: 6784, Cur Loss: 0.38371977, Cur Avg Loss: 0.45374180, Log Avg loss: 0.44225015, Global Avg Loss: 1.20084237, Time: 0.0079 Steps: 94100, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000434, Sample Num: 6944, Cur Loss: 0.29110068, Cur Avg Loss: 0.45285594, Log Avg loss: 0.41529527, Global Avg Loss: 1.20075889, Time: 0.0079 Steps: 94110, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000444, Sample Num: 7104, Cur Loss: 0.71222252, Cur Avg Loss: 0.45260559, Log Avg loss: 0.44174049, Global Avg Loss: 1.20067825, Time: 0.0079 Steps: 94120, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000454, Sample Num: 7264, Cur Loss: 0.32083899, Cur Avg Loss: 0.45261278, Log Avg loss: 0.45293204, Global Avg Loss: 1.20059881, Time: 0.0221 Steps: 94130, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000464, Sample Num: 7424, Cur Loss: 0.48673123, Cur Avg Loss: 0.45542081, Log Avg loss: 0.58290550, Global Avg Loss: 1.20053320, Time: 0.0222 Steps: 94140, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000474, Sample Num: 7584, Cur Loss: 0.11783066, Cur Avg Loss: 0.45800903, Log Avg loss: 0.57810231, Global Avg Loss: 1.20046709, Time: 0.0068 Steps: 94150, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000484, Sample Num: 7744, Cur Loss: 0.39937073, Cur Avg Loss: 0.45747176, Log Avg loss: 0.43200526, Global Avg Loss: 1.20038548, Time: 0.0068 Steps: 94160, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000494, Sample Num: 7904, Cur Loss: 0.80535507, Cur Avg Loss: 0.45663551, Log Avg loss: 0.41616108, Global Avg Loss: 1.20030220, Time: 0.0080 Steps: 94170, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000504, Sample Num: 8064, Cur Loss: 0.54950035, Cur Avg Loss: 0.45586258, Log Avg loss: 0.41767969, Global Avg Loss: 1.20021910, Time: 0.0081 Steps: 94180, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000514, Sample Num: 8224, Cur Loss: 0.18370533, Cur Avg Loss: 0.45482687, Log Avg loss: 0.40262719, Global Avg Loss: 1.20013442, Time: 0.0231 Steps: 94190, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000524, Sample Num: 8384, Cur Loss: 0.31826669, Cur Avg Loss: 0.45469400, Log Avg loss: 0.44786451, Global Avg Loss: 1.20005456, Time: 0.0071 Steps: 94200, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000534, Sample Num: 8544, Cur Loss: 0.17329831, Cur Avg Loss: 0.45291849, Log Avg loss: 0.35988153, Global Avg Loss: 1.19996538, Time: 0.0068 Steps: 94210, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000544, Sample Num: 8704, Cur Loss: 0.76100564, Cur Avg Loss: 0.45273854, Log Avg loss: 0.44312949, Global Avg Loss: 1.19988505, Time: 0.0070 Steps: 94220, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000554, Sample Num: 8864, Cur Loss: 0.23511221, Cur Avg Loss: 0.45200702, Log Avg loss: 0.41221232, Global Avg Loss: 1.19980146, Time: 0.0188 Steps: 94230, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000564, Sample Num: 9024, Cur Loss: 0.50150096, Cur Avg Loss: 0.45238792, Log Avg loss: 0.47348980, Global Avg Loss: 1.19972439, Time: 0.0196 Steps: 94240, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000574, Sample Num: 9184, Cur Loss: 0.53626651, Cur Avg Loss: 0.45236607, Log Avg loss: 0.45113368, Global Avg Loss: 1.19964497, Time: 0.0150 Steps: 94250, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000584, Sample Num: 9344, Cur Loss: 0.21984760, Cur Avg Loss: 0.45234324, Log Avg loss: 0.45103242, Global Avg Loss: 1.19956555, Time: 0.0069 Steps: 94260, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000594, Sample Num: 9504, Cur Loss: 0.48975566, Cur Avg Loss: 0.45244833, Log Avg loss: 0.45858590, Global Avg Loss: 1.19948695, Time: 0.0105 Steps: 94270, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000604, Sample Num: 9664, Cur Loss: 0.53749239, Cur Avg Loss: 0.45592445, Log Avg loss: 0.66240569, Global Avg Loss: 1.19942998, Time: 0.0067 Steps: 94280, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000614, Sample Num: 9824, Cur Loss: 0.24447805, Cur Avg Loss: 0.45540488, Log Avg loss: 0.42402298, Global Avg Loss: 1.19934774, Time: 0.0064 Steps: 94290, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000624, Sample Num: 9984, Cur Loss: 0.45982718, Cur Avg Loss: 0.45520461, Log Avg loss: 0.44290787, Global Avg Loss: 1.19926753, Time: 0.0188 Steps: 94300, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000634, Sample Num: 10144, Cur Loss: 0.31969190, Cur Avg Loss: 0.45297653, Log Avg loss: 0.31394463, Global Avg Loss: 1.19917365, Time: 0.0173 Steps: 94310, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000644, Sample Num: 10304, Cur Loss: 0.36205381, Cur Avg Loss: 0.45264273, Log Avg loss: 0.43148003, Global Avg Loss: 1.19909226, Time: 0.0119 Steps: 94320, Updated lr: 0.000012 Training, Epoch: 0045, Batch: 000654, Sample Num: 10464, Cur Loss: 0.15387857, Cur Avg Loss: 0.45213651, Log Avg loss: 0.41953534, Global Avg Loss: 1.19900962, Time: 0.0066 Steps: 94330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000664, Sample Num: 10624, Cur Loss: 0.53457773, Cur Avg Loss: 0.45144230, Log Avg loss: 0.40604150, Global Avg Loss: 1.19892556, Time: 0.0189 Steps: 94340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000674, Sample Num: 10784, Cur Loss: 0.34234223, Cur Avg Loss: 0.45296267, Log Avg loss: 0.55391509, Global Avg Loss: 1.19885720, Time: 0.0087 Steps: 94350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000684, Sample Num: 10944, Cur Loss: 0.54224145, Cur Avg Loss: 0.45344614, Log Avg loss: 0.48603184, Global Avg Loss: 1.19878166, Time: 0.0207 Steps: 94360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000694, Sample Num: 11104, Cur Loss: 0.51070112, Cur Avg Loss: 0.45437893, Log Avg loss: 0.51818208, Global Avg Loss: 1.19870954, Time: 0.0065 Steps: 94370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000704, Sample Num: 11264, Cur Loss: 0.36018834, Cur Avg Loss: 0.45448096, Log Avg loss: 0.46156193, Global Avg Loss: 1.19863143, Time: 0.0066 Steps: 94380, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000714, Sample Num: 11424, Cur Loss: 0.73478353, Cur Avg Loss: 0.45393890, Log Avg loss: 0.41577757, Global Avg Loss: 1.19854849, Time: 0.0066 Steps: 94390, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000724, Sample Num: 11584, Cur Loss: 0.61488134, Cur Avg Loss: 0.45442771, Log Avg loss: 0.48932893, Global Avg Loss: 1.19847337, Time: 0.0065 Steps: 94400, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000734, Sample Num: 11744, Cur Loss: 0.51684600, Cur Avg Loss: 0.45484518, Log Avg loss: 0.48506983, Global Avg Loss: 1.19839780, Time: 0.0066 Steps: 94410, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000744, Sample Num: 11904, Cur Loss: 0.67545211, Cur Avg Loss: 0.45583702, Log Avg loss: 0.52863805, Global Avg Loss: 1.19832687, Time: 0.0066 Steps: 94420, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000754, Sample Num: 12064, Cur Loss: 0.44148597, Cur Avg Loss: 0.45399504, Log Avg loss: 0.31695194, Global Avg Loss: 1.19823353, Time: 0.0075 Steps: 94430, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000764, Sample Num: 12224, Cur Loss: 0.36824626, Cur Avg Loss: 0.45284073, Log Avg loss: 0.36580551, Global Avg Loss: 1.19814539, Time: 0.0075 Steps: 94440, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000774, Sample Num: 12384, Cur Loss: 0.13436998, Cur Avg Loss: 0.45231738, Log Avg loss: 0.41233310, Global Avg Loss: 1.19806219, Time: 0.0066 Steps: 94450, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000784, Sample Num: 12544, Cur Loss: 0.71999377, Cur Avg Loss: 0.45243263, Log Avg loss: 0.46135319, Global Avg Loss: 1.19798420, Time: 0.0075 Steps: 94460, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000794, Sample Num: 12704, Cur Loss: 0.76960492, Cur Avg Loss: 0.45301359, Log Avg loss: 0.49856077, Global Avg Loss: 1.19791016, Time: 0.0075 Steps: 94470, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000804, Sample Num: 12864, Cur Loss: 0.48424926, Cur Avg Loss: 0.45405725, Log Avg loss: 0.53692407, Global Avg Loss: 1.19784020, Time: 0.0075 Steps: 94480, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000814, Sample Num: 13024, Cur Loss: 0.29656321, Cur Avg Loss: 0.45329685, Log Avg loss: 0.39216071, Global Avg Loss: 1.19775493, Time: 0.0074 Steps: 94490, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000824, Sample Num: 13184, Cur Loss: 0.89015186, Cur Avg Loss: 0.45442597, Log Avg loss: 0.54633640, Global Avg Loss: 1.19768600, Time: 0.0074 Steps: 94500, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000834, Sample Num: 13344, Cur Loss: 0.32266802, Cur Avg Loss: 0.45523039, Log Avg loss: 0.52151414, Global Avg Loss: 1.19761446, Time: 0.0075 Steps: 94510, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000844, Sample Num: 13504, Cur Loss: 0.35295472, Cur Avg Loss: 0.45622996, Log Avg loss: 0.53959417, Global Avg Loss: 1.19754484, Time: 0.0075 Steps: 94520, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000854, Sample Num: 13664, Cur Loss: 0.20261917, Cur Avg Loss: 0.45680295, Log Avg loss: 0.50516379, Global Avg Loss: 1.19747159, Time: 0.0075 Steps: 94530, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000864, Sample Num: 13824, Cur Loss: 0.44222805, Cur Avg Loss: 0.45662386, Log Avg loss: 0.44132956, Global Avg Loss: 1.19739161, Time: 0.0075 Steps: 94540, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000874, Sample Num: 13984, Cur Loss: 0.26230758, Cur Avg Loss: 0.45677655, Log Avg loss: 0.46996852, Global Avg Loss: 1.19731468, Time: 0.0074 Steps: 94550, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000884, Sample Num: 14144, Cur Loss: 0.18640321, Cur Avg Loss: 0.45824927, Log Avg loss: 0.58696492, Global Avg Loss: 1.19725013, Time: 0.0074 Steps: 94560, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000894, Sample Num: 14304, Cur Loss: 0.62346554, Cur Avg Loss: 0.45849346, Log Avg loss: 0.48008069, Global Avg Loss: 1.19717430, Time: 0.0075 Steps: 94570, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000904, Sample Num: 14464, Cur Loss: 0.28677627, Cur Avg Loss: 0.45795314, Log Avg loss: 0.40964794, Global Avg Loss: 1.19709103, Time: 0.0075 Steps: 94580, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000914, Sample Num: 14624, Cur Loss: 1.13662553, Cur Avg Loss: 0.45901676, Log Avg loss: 0.55516825, Global Avg Loss: 1.19702317, Time: 0.0074 Steps: 94590, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000924, Sample Num: 14784, Cur Loss: 0.21513227, Cur Avg Loss: 0.45785626, Log Avg loss: 0.35178643, Global Avg Loss: 1.19693382, Time: 0.0074 Steps: 94600, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000934, Sample Num: 14944, Cur Loss: 0.53671265, Cur Avg Loss: 0.45790896, Log Avg loss: 0.46277888, Global Avg Loss: 1.19685622, Time: 0.0074 Steps: 94610, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000944, Sample Num: 15104, Cur Loss: 0.98361123, Cur Avg Loss: 0.45877993, Log Avg loss: 0.54012790, Global Avg Loss: 1.19678681, Time: 0.0075 Steps: 94620, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000954, Sample Num: 15264, Cur Loss: 0.34424233, Cur Avg Loss: 0.45871301, Log Avg loss: 0.45239601, Global Avg Loss: 1.19670815, Time: 0.0074 Steps: 94630, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000964, Sample Num: 15424, Cur Loss: 0.45283836, Cur Avg Loss: 0.45864898, Log Avg loss: 0.45254013, Global Avg Loss: 1.19662952, Time: 0.0075 Steps: 94640, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000974, Sample Num: 15584, Cur Loss: 0.87466395, Cur Avg Loss: 0.45962562, Log Avg loss: 0.55377443, Global Avg Loss: 1.19656160, Time: 0.0075 Steps: 94650, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000984, Sample Num: 15744, Cur Loss: 0.98065078, Cur Avg Loss: 0.46030011, Log Avg loss: 0.52599470, Global Avg Loss: 1.19649076, Time: 0.0075 Steps: 94660, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 000994, Sample Num: 15904, Cur Loss: 0.16391192, Cur Avg Loss: 0.46039730, Log Avg loss: 0.46996161, Global Avg Loss: 1.19641402, Time: 0.0075 Steps: 94670, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001004, Sample Num: 16064, Cur Loss: 0.30343264, Cur Avg Loss: 0.46098336, Log Avg loss: 0.51923734, Global Avg Loss: 1.19634249, Time: 0.0075 Steps: 94680, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001014, Sample Num: 16224, Cur Loss: 0.17139423, Cur Avg Loss: 0.46092178, Log Avg loss: 0.45473876, Global Avg Loss: 1.19626418, Time: 0.0075 Steps: 94690, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001024, Sample Num: 16384, Cur Loss: 0.25155157, Cur Avg Loss: 0.46006190, Log Avg loss: 0.37287066, Global Avg Loss: 1.19617723, Time: 0.0084 Steps: 94700, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001034, Sample Num: 16544, Cur Loss: 0.34961495, Cur Avg Loss: 0.46243836, Log Avg loss: 0.70578796, Global Avg Loss: 1.19612545, Time: 0.0075 Steps: 94710, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001044, Sample Num: 16704, Cur Loss: 0.29826367, Cur Avg Loss: 0.46269009, Log Avg loss: 0.48871903, Global Avg Loss: 1.19605077, Time: 0.0075 Steps: 94720, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001054, Sample Num: 16864, Cur Loss: 0.84472889, Cur Avg Loss: 0.46265279, Log Avg loss: 0.45875856, Global Avg Loss: 1.19597293, Time: 0.0075 Steps: 94730, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001064, Sample Num: 17024, Cur Loss: 0.38833734, Cur Avg Loss: 0.46233633, Log Avg loss: 0.42898072, Global Avg Loss: 1.19589198, Time: 0.0075 Steps: 94740, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001074, Sample Num: 17184, Cur Loss: 0.33861426, Cur Avg Loss: 0.46290083, Log Avg loss: 0.52296387, Global Avg Loss: 1.19582096, Time: 0.0075 Steps: 94750, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001084, Sample Num: 17344, Cur Loss: 0.44521743, Cur Avg Loss: 0.46267199, Log Avg loss: 0.43809448, Global Avg Loss: 1.19574099, Time: 0.0075 Steps: 94760, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001094, Sample Num: 17504, Cur Loss: 0.84759688, Cur Avg Loss: 0.46290416, Log Avg loss: 0.48807145, Global Avg Loss: 1.19566632, Time: 0.0075 Steps: 94770, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001104, Sample Num: 17664, Cur Loss: 0.38183019, Cur Avg Loss: 0.46294394, Log Avg loss: 0.46729634, Global Avg Loss: 1.19558947, Time: 0.0075 Steps: 94780, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001114, Sample Num: 17824, Cur Loss: 0.32540631, Cur Avg Loss: 0.46369038, Log Avg loss: 0.54609734, Global Avg Loss: 1.19552095, Time: 0.0075 Steps: 94790, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001124, Sample Num: 17984, Cur Loss: 0.46513492, Cur Avg Loss: 0.46373514, Log Avg loss: 0.46872103, Global Avg Loss: 1.19544429, Time: 0.0075 Steps: 94800, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001134, Sample Num: 18144, Cur Loss: 0.58612394, Cur Avg Loss: 0.46376858, Log Avg loss: 0.46752724, Global Avg Loss: 1.19536751, Time: 0.0076 Steps: 94810, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001144, Sample Num: 18304, Cur Loss: 0.23381163, Cur Avg Loss: 0.46372239, Log Avg loss: 0.45848474, Global Avg Loss: 1.19528980, Time: 0.0076 Steps: 94820, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001154, Sample Num: 18464, Cur Loss: 0.29795170, Cur Avg Loss: 0.46389637, Log Avg loss: 0.48379995, Global Avg Loss: 1.19521477, Time: 0.0075 Steps: 94830, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001164, Sample Num: 18624, Cur Loss: 0.74327505, Cur Avg Loss: 0.46404885, Log Avg loss: 0.48164512, Global Avg Loss: 1.19513953, Time: 0.0077 Steps: 94840, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001174, Sample Num: 18784, Cur Loss: 0.31947121, Cur Avg Loss: 0.46535863, Log Avg loss: 0.61781618, Global Avg Loss: 1.19507866, Time: 0.0075 Steps: 94850, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001184, Sample Num: 18944, Cur Loss: 0.60291636, Cur Avg Loss: 0.46439453, Log Avg loss: 0.35120959, Global Avg Loss: 1.19498970, Time: 0.0075 Steps: 94860, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001194, Sample Num: 19104, Cur Loss: 0.15567602, Cur Avg Loss: 0.46425093, Log Avg loss: 0.44724860, Global Avg Loss: 1.19491089, Time: 0.0076 Steps: 94870, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001204, Sample Num: 19264, Cur Loss: 0.40835404, Cur Avg Loss: 0.46338130, Log Avg loss: 0.35954710, Global Avg Loss: 1.19482284, Time: 0.0078 Steps: 94880, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001214, Sample Num: 19424, Cur Loss: 0.88822562, Cur Avg Loss: 0.46260374, Log Avg loss: 0.36898621, Global Avg Loss: 1.19473581, Time: 0.0076 Steps: 94890, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001224, Sample Num: 19584, Cur Loss: 0.23421562, Cur Avg Loss: 0.46247956, Log Avg loss: 0.44740350, Global Avg Loss: 1.19465706, Time: 0.0075 Steps: 94900, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001234, Sample Num: 19744, Cur Loss: 0.28856072, Cur Avg Loss: 0.46274784, Log Avg loss: 0.49558581, Global Avg Loss: 1.19458340, Time: 0.0075 Steps: 94910, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001244, Sample Num: 19904, Cur Loss: 0.43455881, Cur Avg Loss: 0.46168741, Log Avg loss: 0.33083070, Global Avg Loss: 1.19449241, Time: 0.0076 Steps: 94920, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001254, Sample Num: 20064, Cur Loss: 1.15703464, Cur Avg Loss: 0.46151656, Log Avg loss: 0.44026199, Global Avg Loss: 1.19441296, Time: 0.0075 Steps: 94930, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001264, Sample Num: 20224, Cur Loss: 0.13305728, Cur Avg Loss: 0.46049894, Log Avg loss: 0.33288986, Global Avg Loss: 1.19432221, Time: 0.0075 Steps: 94940, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001274, Sample Num: 20384, Cur Loss: 0.50494969, Cur Avg Loss: 0.46040390, Log Avg loss: 0.44839028, Global Avg Loss: 1.19424365, Time: 0.0076 Steps: 94950, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001284, Sample Num: 20544, Cur Loss: 0.27157474, Cur Avg Loss: 0.46135079, Log Avg loss: 0.58198455, Global Avg Loss: 1.19417918, Time: 0.0075 Steps: 94960, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001294, Sample Num: 20704, Cur Loss: 0.20103566, Cur Avg Loss: 0.46064105, Log Avg loss: 0.36951038, Global Avg Loss: 1.19409234, Time: 0.0074 Steps: 94970, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001304, Sample Num: 20864, Cur Loss: 0.37592208, Cur Avg Loss: 0.46074894, Log Avg loss: 0.47471023, Global Avg Loss: 1.19401660, Time: 0.0075 Steps: 94980, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001314, Sample Num: 21024, Cur Loss: 0.61481816, Cur Avg Loss: 0.45975372, Log Avg loss: 0.32997688, Global Avg Loss: 1.19392564, Time: 0.0075 Steps: 94990, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001324, Sample Num: 21184, Cur Loss: 0.45032740, Cur Avg Loss: 0.45977232, Log Avg loss: 0.46221721, Global Avg Loss: 1.19384862, Time: 0.0075 Steps: 95000, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001334, Sample Num: 21344, Cur Loss: 0.32710582, Cur Avg Loss: 0.46088154, Log Avg loss: 0.60774206, Global Avg Loss: 1.19378693, Time: 0.0076 Steps: 95010, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001344, Sample Num: 21504, Cur Loss: 0.57470149, Cur Avg Loss: 0.46060286, Log Avg loss: 0.42342609, Global Avg Loss: 1.19370585, Time: 0.0077 Steps: 95020, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001354, Sample Num: 21664, Cur Loss: 0.23644947, Cur Avg Loss: 0.46085427, Log Avg loss: 0.49464476, Global Avg Loss: 1.19363229, Time: 0.0239 Steps: 95030, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001364, Sample Num: 21824, Cur Loss: 0.43998975, Cur Avg Loss: 0.46063554, Log Avg loss: 0.43101886, Global Avg Loss: 1.19355205, Time: 0.0072 Steps: 95040, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001374, Sample Num: 21984, Cur Loss: 1.21981466, Cur Avg Loss: 0.46255094, Log Avg loss: 0.72381234, Global Avg Loss: 1.19350263, Time: 0.0066 Steps: 95050, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001384, Sample Num: 22144, Cur Loss: 0.28609160, Cur Avg Loss: 0.46230552, Log Avg loss: 0.42858418, Global Avg Loss: 1.19342216, Time: 0.0067 Steps: 95060, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001394, Sample Num: 22304, Cur Loss: 0.52303398, Cur Avg Loss: 0.46181760, Log Avg loss: 0.39428894, Global Avg Loss: 1.19333811, Time: 0.0198 Steps: 95070, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001404, Sample Num: 22464, Cur Loss: 0.37351793, Cur Avg Loss: 0.46131143, Log Avg loss: 0.39075215, Global Avg Loss: 1.19325369, Time: 0.0174 Steps: 95080, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001414, Sample Num: 22624, Cur Loss: 0.11505556, Cur Avg Loss: 0.46152480, Log Avg loss: 0.49148223, Global Avg Loss: 1.19317989, Time: 0.0070 Steps: 95090, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001424, Sample Num: 22784, Cur Loss: 0.31484860, Cur Avg Loss: 0.46186062, Log Avg loss: 0.50934441, Global Avg Loss: 1.19310799, Time: 0.0111 Steps: 95100, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001434, Sample Num: 22944, Cur Loss: 0.19520640, Cur Avg Loss: 0.46093387, Log Avg loss: 0.32896562, Global Avg Loss: 1.19301713, Time: 0.0064 Steps: 95110, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001444, Sample Num: 23104, Cur Loss: 0.41570413, Cur Avg Loss: 0.46172314, Log Avg loss: 0.57490411, Global Avg Loss: 1.19295215, Time: 0.0070 Steps: 95120, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001454, Sample Num: 23264, Cur Loss: 0.68570763, Cur Avg Loss: 0.46223713, Log Avg loss: 0.53645754, Global Avg Loss: 1.19288314, Time: 0.0200 Steps: 95130, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001464, Sample Num: 23424, Cur Loss: 0.17755213, Cur Avg Loss: 0.46147524, Log Avg loss: 0.35069564, Global Avg Loss: 1.19279462, Time: 0.0116 Steps: 95140, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001474, Sample Num: 23584, Cur Loss: 0.71101677, Cur Avg Loss: 0.46231041, Log Avg loss: 0.58457918, Global Avg Loss: 1.19273069, Time: 0.0073 Steps: 95150, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001484, Sample Num: 23744, Cur Loss: 0.70540279, Cur Avg Loss: 0.46148399, Log Avg loss: 0.33967029, Global Avg Loss: 1.19264105, Time: 0.0172 Steps: 95160, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001494, Sample Num: 23904, Cur Loss: 0.18539070, Cur Avg Loss: 0.46260471, Log Avg loss: 0.62892029, Global Avg Loss: 1.19258182, Time: 0.0213 Steps: 95170, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001504, Sample Num: 24064, Cur Loss: 0.80377942, Cur Avg Loss: 0.46217106, Log Avg loss: 0.39738352, Global Avg Loss: 1.19249827, Time: 0.0166 Steps: 95180, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001514, Sample Num: 24224, Cur Loss: 0.42299223, Cur Avg Loss: 0.46315555, Log Avg loss: 0.61122169, Global Avg Loss: 1.19243721, Time: 0.0070 Steps: 95190, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001524, Sample Num: 24384, Cur Loss: 0.63627964, Cur Avg Loss: 0.46271065, Log Avg loss: 0.39535416, Global Avg Loss: 1.19235348, Time: 0.0089 Steps: 95200, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001534, Sample Num: 24544, Cur Loss: 0.87387484, Cur Avg Loss: 0.46261873, Log Avg loss: 0.44860972, Global Avg Loss: 1.19227536, Time: 0.0066 Steps: 95210, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001544, Sample Num: 24704, Cur Loss: 0.37906110, Cur Avg Loss: 0.46221743, Log Avg loss: 0.40065788, Global Avg Loss: 1.19219223, Time: 0.0113 Steps: 95220, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001554, Sample Num: 24864, Cur Loss: 0.34277725, Cur Avg Loss: 0.46239684, Log Avg loss: 0.49009700, Global Avg Loss: 1.19211850, Time: 0.0066 Steps: 95230, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001564, Sample Num: 25024, Cur Loss: 0.39520109, Cur Avg Loss: 0.46268221, Log Avg loss: 0.50702947, Global Avg Loss: 1.19204657, Time: 0.0066 Steps: 95240, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001574, Sample Num: 25184, Cur Loss: 0.51616973, Cur Avg Loss: 0.46230781, Log Avg loss: 0.40375080, Global Avg Loss: 1.19196381, Time: 0.0066 Steps: 95250, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001584, Sample Num: 25344, Cur Loss: 0.66873395, Cur Avg Loss: 0.46349774, Log Avg loss: 0.65079286, Global Avg Loss: 1.19190700, Time: 0.0065 Steps: 95260, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001594, Sample Num: 25504, Cur Loss: 0.34107855, Cur Avg Loss: 0.46408367, Log Avg loss: 0.55689621, Global Avg Loss: 1.19184034, Time: 0.0066 Steps: 95270, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001604, Sample Num: 25664, Cur Loss: 0.34722346, Cur Avg Loss: 0.46430165, Log Avg loss: 0.49904682, Global Avg Loss: 1.19176763, Time: 0.0065 Steps: 95280, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001614, Sample Num: 25824, Cur Loss: 1.07274473, Cur Avg Loss: 0.46454144, Log Avg loss: 0.50300438, Global Avg Loss: 1.19169535, Time: 0.0074 Steps: 95290, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001624, Sample Num: 25984, Cur Loss: 0.50537920, Cur Avg Loss: 0.46501635, Log Avg loss: 0.54166635, Global Avg Loss: 1.19162714, Time: 0.0074 Steps: 95300, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001634, Sample Num: 26144, Cur Loss: 1.44857168, Cur Avg Loss: 0.46527252, Log Avg loss: 0.50687384, Global Avg Loss: 1.19155530, Time: 0.0075 Steps: 95310, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001644, Sample Num: 26304, Cur Loss: 1.17690229, Cur Avg Loss: 0.46557341, Log Avg loss: 0.51473964, Global Avg Loss: 1.19148429, Time: 0.0075 Steps: 95320, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001654, Sample Num: 26464, Cur Loss: 0.26383907, Cur Avg Loss: 0.46548165, Log Avg loss: 0.45039607, Global Avg Loss: 1.19140655, Time: 0.0074 Steps: 95330, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001664, Sample Num: 26624, Cur Loss: 0.58852524, Cur Avg Loss: 0.46610726, Log Avg loss: 0.56958370, Global Avg Loss: 1.19134133, Time: 0.0076 Steps: 95340, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001674, Sample Num: 26784, Cur Loss: 0.35475665, Cur Avg Loss: 0.46548713, Log Avg loss: 0.36229707, Global Avg Loss: 1.19125438, Time: 0.0075 Steps: 95350, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001684, Sample Num: 26944, Cur Loss: 0.46457532, Cur Avg Loss: 0.46480842, Log Avg loss: 0.35119321, Global Avg Loss: 1.19116629, Time: 0.0075 Steps: 95360, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001694, Sample Num: 27104, Cur Loss: 0.24875546, Cur Avg Loss: 0.46510033, Log Avg loss: 0.51425750, Global Avg Loss: 1.19109531, Time: 0.0075 Steps: 95370, Updated lr: 0.000011 Training, Epoch: 0045, Batch: 001704, Sample Num: 27264, Cur Loss: 0.30213982, Cur Avg Loss: 0.46575835, Log Avg loss: 0.57722740, Global Avg Loss: 1.19103095, Time: 0.0075 Steps: 95380, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001714, Sample Num: 27424, Cur Loss: 0.20209140, Cur Avg Loss: 0.46612469, Log Avg loss: 0.52854856, Global Avg Loss: 1.19096150, Time: 0.0076 Steps: 95390, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001724, Sample Num: 27584, Cur Loss: 0.31721473, Cur Avg Loss: 0.46622589, Log Avg loss: 0.48357159, Global Avg Loss: 1.19088735, Time: 0.0075 Steps: 95400, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001734, Sample Num: 27744, Cur Loss: 0.37510386, Cur Avg Loss: 0.46528654, Log Avg loss: 0.30334194, Global Avg Loss: 1.19079433, Time: 0.0075 Steps: 95410, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001744, Sample Num: 27904, Cur Loss: 0.35595226, Cur Avg Loss: 0.46497222, Log Avg loss: 0.41046904, Global Avg Loss: 1.19071255, Time: 0.0076 Steps: 95420, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001754, Sample Num: 28064, Cur Loss: 0.25321370, Cur Avg Loss: 0.46445535, Log Avg loss: 0.37431294, Global Avg Loss: 1.19062700, Time: 0.0075 Steps: 95430, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001764, Sample Num: 28224, Cur Loss: 0.43230319, Cur Avg Loss: 0.46432911, Log Avg loss: 0.44218760, Global Avg Loss: 1.19054858, Time: 0.0075 Steps: 95440, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001774, Sample Num: 28384, Cur Loss: 0.40772495, Cur Avg Loss: 0.46363692, Log Avg loss: 0.34153468, Global Avg Loss: 1.19045963, Time: 0.0076 Steps: 95450, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001784, Sample Num: 28544, Cur Loss: 0.67022187, Cur Avg Loss: 0.46357837, Log Avg loss: 0.45319168, Global Avg Loss: 1.19038240, Time: 0.0075 Steps: 95460, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001794, Sample Num: 28704, Cur Loss: 0.33608076, Cur Avg Loss: 0.46446602, Log Avg loss: 0.62282187, Global Avg Loss: 1.19032295, Time: 0.0075 Steps: 95470, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001804, Sample Num: 28864, Cur Loss: 0.37144724, Cur Avg Loss: 0.46448659, Log Avg loss: 0.46817686, Global Avg Loss: 1.19024732, Time: 0.0074 Steps: 95480, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001814, Sample Num: 29024, Cur Loss: 0.50713462, Cur Avg Loss: 0.46416114, Log Avg loss: 0.40545018, Global Avg Loss: 1.19016513, Time: 0.0075 Steps: 95490, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001824, Sample Num: 29184, Cur Loss: 0.44231677, Cur Avg Loss: 0.46387635, Log Avg loss: 0.41221614, Global Avg Loss: 1.19008367, Time: 0.0075 Steps: 95500, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001834, Sample Num: 29344, Cur Loss: 0.22410820, Cur Avg Loss: 0.46315668, Log Avg loss: 0.33188825, Global Avg Loss: 1.18999382, Time: 0.0075 Steps: 95510, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001844, Sample Num: 29504, Cur Loss: 0.71218401, Cur Avg Loss: 0.46364847, Log Avg loss: 0.55384229, Global Avg Loss: 1.18992722, Time: 0.0075 Steps: 95520, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001854, Sample Num: 29664, Cur Loss: 0.13046904, Cur Avg Loss: 0.46298198, Log Avg loss: 0.34008137, Global Avg Loss: 1.18983826, Time: 0.0074 Steps: 95530, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001864, Sample Num: 29824, Cur Loss: 0.34308907, Cur Avg Loss: 0.46288896, Log Avg loss: 0.44564374, Global Avg Loss: 1.18976036, Time: 0.0075 Steps: 95540, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001874, Sample Num: 29984, Cur Loss: 1.42108607, Cur Avg Loss: 0.46331358, Log Avg loss: 0.54246303, Global Avg Loss: 1.18969262, Time: 0.0074 Steps: 95550, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001884, Sample Num: 30144, Cur Loss: 0.25534374, Cur Avg Loss: 0.46307351, Log Avg loss: 0.41808438, Global Avg Loss: 1.18961187, Time: 0.0075 Steps: 95560, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001894, Sample Num: 30304, Cur Loss: 0.58549547, Cur Avg Loss: 0.46304070, Log Avg loss: 0.45685893, Global Avg Loss: 1.18953520, Time: 0.0077 Steps: 95570, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001904, Sample Num: 30464, Cur Loss: 0.35458905, Cur Avg Loss: 0.46260484, Log Avg loss: 0.38005257, Global Avg Loss: 1.18945051, Time: 0.0075 Steps: 95580, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001914, Sample Num: 30624, Cur Loss: 0.31963527, Cur Avg Loss: 0.46232568, Log Avg loss: 0.40917449, Global Avg Loss: 1.18936888, Time: 0.0076 Steps: 95590, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001924, Sample Num: 30784, Cur Loss: 0.73254406, Cur Avg Loss: 0.46346505, Log Avg loss: 0.68153950, Global Avg Loss: 1.18931576, Time: 0.0076 Steps: 95600, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001934, Sample Num: 30944, Cur Loss: 1.09531057, Cur Avg Loss: 0.46391606, Log Avg loss: 0.55069081, Global Avg Loss: 1.18924897, Time: 0.0077 Steps: 95610, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001944, Sample Num: 31104, Cur Loss: 0.43834144, Cur Avg Loss: 0.46446687, Log Avg loss: 0.57099351, Global Avg Loss: 1.18918431, Time: 0.0076 Steps: 95620, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001954, Sample Num: 31264, Cur Loss: 0.58263612, Cur Avg Loss: 0.46438205, Log Avg loss: 0.44789219, Global Avg Loss: 1.18910679, Time: 0.0076 Steps: 95630, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001964, Sample Num: 31424, Cur Loss: 0.41769552, Cur Avg Loss: 0.46451497, Log Avg loss: 0.49048818, Global Avg Loss: 1.18903375, Time: 0.0076 Steps: 95640, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001974, Sample Num: 31584, Cur Loss: 0.18746689, Cur Avg Loss: 0.46481385, Log Avg loss: 0.52351347, Global Avg Loss: 1.18896417, Time: 0.0076 Steps: 95650, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001984, Sample Num: 31744, Cur Loss: 0.26583618, Cur Avg Loss: 0.46478432, Log Avg loss: 0.45895554, Global Avg Loss: 1.18888785, Time: 0.0076 Steps: 95660, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 001994, Sample Num: 31904, Cur Loss: 0.53396845, Cur Avg Loss: 0.46471835, Log Avg loss: 0.45162997, Global Avg Loss: 1.18881079, Time: 0.0077 Steps: 95670, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002004, Sample Num: 32064, Cur Loss: 0.23083004, Cur Avg Loss: 0.46456180, Log Avg loss: 0.43334533, Global Avg Loss: 1.18873183, Time: 0.0077 Steps: 95680, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002014, Sample Num: 32224, Cur Loss: 0.37988740, Cur Avg Loss: 0.46437550, Log Avg loss: 0.42704120, Global Avg Loss: 1.18865223, Time: 0.0076 Steps: 95690, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002024, Sample Num: 32384, Cur Loss: 0.52340817, Cur Avg Loss: 0.46397834, Log Avg loss: 0.38398976, Global Avg Loss: 1.18856815, Time: 0.0076 Steps: 95700, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002034, Sample Num: 32544, Cur Loss: 0.54594874, Cur Avg Loss: 0.46374270, Log Avg loss: 0.41604911, Global Avg Loss: 1.18848744, Time: 0.0077 Steps: 95710, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002044, Sample Num: 32704, Cur Loss: 0.35050654, Cur Avg Loss: 0.46431886, Log Avg loss: 0.58151045, Global Avg Loss: 1.18842403, Time: 0.0076 Steps: 95720, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002054, Sample Num: 32864, Cur Loss: 0.20680001, Cur Avg Loss: 0.46406056, Log Avg loss: 0.41126364, Global Avg Loss: 1.18834284, Time: 0.0065 Steps: 95730, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002064, Sample Num: 33024, Cur Loss: 0.95780683, Cur Avg Loss: 0.46405347, Log Avg loss: 0.46259741, Global Avg Loss: 1.18826704, Time: 0.0066 Steps: 95740, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002074, Sample Num: 33184, Cur Loss: 0.18412468, Cur Avg Loss: 0.46410384, Log Avg loss: 0.47450087, Global Avg Loss: 1.18819250, Time: 0.0066 Steps: 95750, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002084, Sample Num: 33344, Cur Loss: 0.37198049, Cur Avg Loss: 0.46415223, Log Avg loss: 0.47418898, Global Avg Loss: 1.18811793, Time: 0.0067 Steps: 95760, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002094, Sample Num: 33504, Cur Loss: 0.35802019, Cur Avg Loss: 0.46465193, Log Avg loss: 0.56878844, Global Avg Loss: 1.18805327, Time: 0.0065 Steps: 95770, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002104, Sample Num: 33664, Cur Loss: 0.14110498, Cur Avg Loss: 0.46441819, Log Avg loss: 0.41547240, Global Avg Loss: 1.18797260, Time: 0.0066 Steps: 95780, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002114, Sample Num: 33824, Cur Loss: 0.48657423, Cur Avg Loss: 0.46389234, Log Avg loss: 0.35325398, Global Avg Loss: 1.18788546, Time: 0.0066 Steps: 95790, Updated lr: 0.000010 Training, Epoch: 0045, Batch: 002124, Sample Num: 33984, Cur Loss: 0.37090805, Cur Avg Loss: 0.46352609, Log Avg loss: 0.38610036, Global Avg Loss: 1.18780177, Time: 0.0075 Steps: 95800, Updated lr: 0.000010 ***** Running evaluation checkpoint-95805 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-95805 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.400843, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.580628, "eval_total_loss": 408.181683, "eval_mae": 0.585441, "eval_mse": 0.580741, "eval_r2": 0.630843, "eval_sp_statistic": 0.752702, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.799721, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.879272, "test_total_loss": 441.394372, "test_mae": 0.66057, "test_mse": 0.879517, "test_r2": 0.432352, "test_sp_statistic": 0.601556, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.669308, "test_ps_pvalue": 0.0, "lr": 1.0094831673779043e-05, "cur_epoch_step": 2129, "train_global_avg_loss": 1.187762044789656, "train_cur_epoch_loss": 986.6626455932856, "train_cur_epoch_avg_loss": 0.4634394765586123, "train_cur_epoch_time": 17.400843381881714, "train_cur_epoch_avg_time": 0.008173247243720861, "epoch": 45, "step": 95805} ################################################## Training, Epoch: 0046, Batch: 000005, Sample Num: 80, Cur Loss: 0.24936321, Cur Avg Loss: 0.40571712, Log Avg loss: 0.41618225, Global Avg Loss: 1.18772123, Time: 0.0067 Steps: 95810, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000015, Sample Num: 240, Cur Loss: 0.50893444, Cur Avg Loss: 0.50178867, Log Avg loss: 0.54982445, Global Avg Loss: 1.18765466, Time: 0.0085 Steps: 95820, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000025, Sample Num: 400, Cur Loss: 0.53899288, Cur Avg Loss: 0.47655132, Log Avg loss: 0.43869530, Global Avg Loss: 1.18757651, Time: 0.0063 Steps: 95830, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000035, Sample Num: 560, Cur Loss: 0.41298220, Cur Avg Loss: 0.47176209, Log Avg loss: 0.45978900, Global Avg Loss: 1.18750057, Time: 0.0064 Steps: 95840, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000045, Sample Num: 720, Cur Loss: 0.44235292, Cur Avg Loss: 0.47993046, Log Avg loss: 0.50851977, Global Avg Loss: 1.18742973, Time: 0.0063 Steps: 95850, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000055, Sample Num: 880, Cur Loss: 1.04615819, Cur Avg Loss: 0.49067481, Log Avg loss: 0.53902438, Global Avg Loss: 1.18736209, Time: 0.0065 Steps: 95860, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000065, Sample Num: 1040, Cur Loss: 0.08439702, Cur Avg Loss: 0.47051449, Log Avg loss: 0.35963273, Global Avg Loss: 1.18727575, Time: 0.0206 Steps: 95870, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000075, Sample Num: 1200, Cur Loss: 0.28090259, Cur Avg Loss: 0.44970425, Log Avg loss: 0.31443772, Global Avg Loss: 1.18718472, Time: 0.0063 Steps: 95880, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000085, Sample Num: 1360, Cur Loss: 0.65933084, Cur Avg Loss: 0.46126323, Log Avg loss: 0.54795553, Global Avg Loss: 1.18711805, Time: 0.0063 Steps: 95890, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000095, Sample Num: 1520, Cur Loss: 0.52804053, Cur Avg Loss: 0.46712856, Log Avg loss: 0.51698385, Global Avg Loss: 1.18704817, Time: 0.0064 Steps: 95900, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000105, Sample Num: 1680, Cur Loss: 0.18059245, Cur Avg Loss: 0.46107026, Log Avg loss: 0.40351649, Global Avg Loss: 1.18696648, Time: 0.0064 Steps: 95910, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000115, Sample Num: 1840, Cur Loss: 0.68026388, Cur Avg Loss: 0.45818006, Log Avg loss: 0.42783289, Global Avg Loss: 1.18688734, Time: 0.0197 Steps: 95920, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000125, Sample Num: 2000, Cur Loss: 0.52541959, Cur Avg Loss: 0.45285318, Log Avg loss: 0.39159406, Global Avg Loss: 1.18680443, Time: 0.0068 Steps: 95930, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000135, Sample Num: 2160, Cur Loss: 0.31972241, Cur Avg Loss: 0.44882232, Log Avg loss: 0.39843661, Global Avg Loss: 1.18672226, Time: 0.0066 Steps: 95940, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000145, Sample Num: 2320, Cur Loss: 0.52340513, Cur Avg Loss: 0.44801769, Log Avg loss: 0.43715518, Global Avg Loss: 1.18664414, Time: 0.0064 Steps: 95950, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000155, Sample Num: 2480, Cur Loss: 0.21525247, Cur Avg Loss: 0.44251817, Log Avg loss: 0.36277508, Global Avg Loss: 1.18655828, Time: 0.0065 Steps: 95960, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000165, Sample Num: 2640, Cur Loss: 0.16582614, Cur Avg Loss: 0.43691062, Log Avg loss: 0.34999356, Global Avg Loss: 1.18647112, Time: 0.0067 Steps: 95970, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000175, Sample Num: 2800, Cur Loss: 0.40204400, Cur Avg Loss: 0.43593903, Log Avg loss: 0.41990789, Global Avg Loss: 1.18639125, Time: 0.0066 Steps: 95980, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000185, Sample Num: 2960, Cur Loss: 0.21695825, Cur Avg Loss: 0.43876399, Log Avg loss: 0.48820082, Global Avg Loss: 1.18631851, Time: 0.0066 Steps: 95990, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000195, Sample Num: 3120, Cur Loss: 0.16228417, Cur Avg Loss: 0.43481463, Log Avg loss: 0.36175148, Global Avg Loss: 1.18623262, Time: 0.0065 Steps: 96000, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000205, Sample Num: 3280, Cur Loss: 0.24689680, Cur Avg Loss: 0.43762723, Log Avg loss: 0.49247282, Global Avg Loss: 1.18616036, Time: 0.0065 Steps: 96010, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000215, Sample Num: 3440, Cur Loss: 0.60866147, Cur Avg Loss: 0.43896545, Log Avg loss: 0.46639896, Global Avg Loss: 1.18608540, Time: 0.0075 Steps: 96020, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000225, Sample Num: 3600, Cur Loss: 0.30124068, Cur Avg Loss: 0.44356533, Log Avg loss: 0.54246269, Global Avg Loss: 1.18601838, Time: 0.0075 Steps: 96030, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000235, Sample Num: 3760, Cur Loss: 0.14496197, Cur Avg Loss: 0.44249506, Log Avg loss: 0.41841406, Global Avg Loss: 1.18593845, Time: 0.0075 Steps: 96040, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000245, Sample Num: 3920, Cur Loss: 0.59809053, Cur Avg Loss: 0.44290915, Log Avg loss: 0.45264035, Global Avg Loss: 1.18586211, Time: 0.0075 Steps: 96050, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000255, Sample Num: 4080, Cur Loss: 0.43157876, Cur Avg Loss: 0.44171170, Log Avg loss: 0.41237415, Global Avg Loss: 1.18578159, Time: 0.0076 Steps: 96060, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000265, Sample Num: 4240, Cur Loss: 0.28686762, Cur Avg Loss: 0.43688059, Log Avg loss: 0.31368731, Global Avg Loss: 1.18569081, Time: 0.0067 Steps: 96070, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000275, Sample Num: 4400, Cur Loss: 0.94862497, Cur Avg Loss: 0.43895417, Log Avg loss: 0.49390406, Global Avg Loss: 1.18561881, Time: 0.0067 Steps: 96080, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000285, Sample Num: 4560, Cur Loss: 0.27893835, Cur Avg Loss: 0.43717200, Log Avg loss: 0.38816227, Global Avg Loss: 1.18553582, Time: 0.0067 Steps: 96090, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000295, Sample Num: 4720, Cur Loss: 0.22004506, Cur Avg Loss: 0.43868222, Log Avg loss: 0.48172339, Global Avg Loss: 1.18546258, Time: 0.0066 Steps: 96100, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000305, Sample Num: 4880, Cur Loss: 0.59516197, Cur Avg Loss: 0.44111102, Log Avg loss: 0.51276073, Global Avg Loss: 1.18539259, Time: 0.0066 Steps: 96110, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000315, Sample Num: 5040, Cur Loss: 0.35319510, Cur Avg Loss: 0.43985588, Log Avg loss: 0.40157412, Global Avg Loss: 1.18531104, Time: 0.0067 Steps: 96120, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000325, Sample Num: 5200, Cur Loss: 0.46787086, Cur Avg Loss: 0.44118691, Log Avg loss: 0.48311424, Global Avg Loss: 1.18523799, Time: 0.0076 Steps: 96130, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000335, Sample Num: 5360, Cur Loss: 0.25116435, Cur Avg Loss: 0.44124648, Log Avg loss: 0.44318260, Global Avg Loss: 1.18516081, Time: 0.0076 Steps: 96140, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000345, Sample Num: 5520, Cur Loss: 0.88610721, Cur Avg Loss: 0.44200259, Log Avg loss: 0.46733229, Global Avg Loss: 1.18508615, Time: 0.0076 Steps: 96150, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000355, Sample Num: 5680, Cur Loss: 0.45712450, Cur Avg Loss: 0.43961018, Log Avg loss: 0.35707196, Global Avg Loss: 1.18500004, Time: 0.0076 Steps: 96160, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000365, Sample Num: 5840, Cur Loss: 0.24043718, Cur Avg Loss: 0.43872832, Log Avg loss: 0.40742222, Global Avg Loss: 1.18491919, Time: 0.0075 Steps: 96170, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000375, Sample Num: 6000, Cur Loss: 0.22578463, Cur Avg Loss: 0.43643232, Log Avg loss: 0.35262825, Global Avg Loss: 1.18483266, Time: 0.0074 Steps: 96180, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000385, Sample Num: 6160, Cur Loss: 0.68605953, Cur Avg Loss: 0.43807277, Log Avg loss: 0.49958978, Global Avg Loss: 1.18476142, Time: 0.0074 Steps: 96190, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000395, Sample Num: 6320, Cur Loss: 0.32935625, Cur Avg Loss: 0.43951887, Log Avg loss: 0.49519379, Global Avg Loss: 1.18468974, Time: 0.0075 Steps: 96200, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000405, Sample Num: 6480, Cur Loss: 0.36513340, Cur Avg Loss: 0.44250368, Log Avg loss: 0.56040356, Global Avg Loss: 1.18462485, Time: 0.0074 Steps: 96210, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000415, Sample Num: 6640, Cur Loss: 0.37882277, Cur Avg Loss: 0.44453227, Log Avg loss: 0.52669030, Global Avg Loss: 1.18455647, Time: 0.0074 Steps: 96220, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000425, Sample Num: 6800, Cur Loss: 0.65665805, Cur Avg Loss: 0.44324201, Log Avg loss: 0.38969625, Global Avg Loss: 1.18447387, Time: 0.0075 Steps: 96230, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000435, Sample Num: 6960, Cur Loss: 0.14879937, Cur Avg Loss: 0.44285219, Log Avg loss: 0.42628482, Global Avg Loss: 1.18439509, Time: 0.0076 Steps: 96240, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000445, Sample Num: 7120, Cur Loss: 0.47294933, Cur Avg Loss: 0.44204794, Log Avg loss: 0.40706297, Global Avg Loss: 1.18431433, Time: 0.0076 Steps: 96250, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000455, Sample Num: 7280, Cur Loss: 0.48945042, Cur Avg Loss: 0.44254973, Log Avg loss: 0.46487964, Global Avg Loss: 1.18423959, Time: 0.0075 Steps: 96260, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000465, Sample Num: 7440, Cur Loss: 0.49112004, Cur Avg Loss: 0.44390937, Log Avg loss: 0.50577298, Global Avg Loss: 1.18416911, Time: 0.0075 Steps: 96270, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000475, Sample Num: 7600, Cur Loss: 0.50460601, Cur Avg Loss: 0.44377520, Log Avg loss: 0.43753588, Global Avg Loss: 1.18409157, Time: 0.0077 Steps: 96280, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000485, Sample Num: 7760, Cur Loss: 1.16380394, Cur Avg Loss: 0.44748104, Log Avg loss: 0.62350843, Global Avg Loss: 1.18403335, Time: 0.0077 Steps: 96290, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000495, Sample Num: 7920, Cur Loss: 0.39486688, Cur Avg Loss: 0.45087188, Log Avg loss: 0.61532774, Global Avg Loss: 1.18397429, Time: 0.0076 Steps: 96300, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000505, Sample Num: 8080, Cur Loss: 0.57067561, Cur Avg Loss: 0.45167151, Log Avg loss: 0.49125319, Global Avg Loss: 1.18390237, Time: 0.0076 Steps: 96310, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000515, Sample Num: 8240, Cur Loss: 0.24388991, Cur Avg Loss: 0.45121544, Log Avg loss: 0.42818421, Global Avg Loss: 1.18382391, Time: 0.0076 Steps: 96320, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000525, Sample Num: 8400, Cur Loss: 0.89413077, Cur Avg Loss: 0.45100082, Log Avg loss: 0.43994788, Global Avg Loss: 1.18374668, Time: 0.0077 Steps: 96330, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000535, Sample Num: 8560, Cur Loss: 1.02490509, Cur Avg Loss: 0.45000455, Log Avg loss: 0.39770008, Global Avg Loss: 1.18366509, Time: 0.0076 Steps: 96340, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000545, Sample Num: 8720, Cur Loss: 0.16701958, Cur Avg Loss: 0.45118058, Log Avg loss: 0.51409833, Global Avg Loss: 1.18359560, Time: 0.0077 Steps: 96350, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000555, Sample Num: 8880, Cur Loss: 0.23794919, Cur Avg Loss: 0.44973244, Log Avg loss: 0.37080901, Global Avg Loss: 1.18351125, Time: 0.0075 Steps: 96360, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000565, Sample Num: 9040, Cur Loss: 0.18108097, Cur Avg Loss: 0.44680387, Log Avg loss: 0.28426824, Global Avg Loss: 1.18341794, Time: 0.0075 Steps: 96370, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000575, Sample Num: 9200, Cur Loss: 0.24915823, Cur Avg Loss: 0.44659863, Log Avg loss: 0.43500259, Global Avg Loss: 1.18334029, Time: 0.0075 Steps: 96380, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000585, Sample Num: 9360, Cur Loss: 0.50780088, Cur Avg Loss: 0.44658052, Log Avg loss: 0.44553894, Global Avg Loss: 1.18326374, Time: 0.0075 Steps: 96390, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000595, Sample Num: 9520, Cur Loss: 0.41607007, Cur Avg Loss: 0.44614699, Log Avg loss: 0.42078560, Global Avg Loss: 1.18318465, Time: 0.0076 Steps: 96400, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000605, Sample Num: 9680, Cur Loss: 0.39680946, Cur Avg Loss: 0.44516555, Log Avg loss: 0.38676987, Global Avg Loss: 1.18310204, Time: 0.0077 Steps: 96410, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000615, Sample Num: 9840, Cur Loss: 0.12865096, Cur Avg Loss: 0.44536868, Log Avg loss: 0.45765806, Global Avg Loss: 1.18302680, Time: 0.0076 Steps: 96420, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000625, Sample Num: 10000, Cur Loss: 1.12108874, Cur Avg Loss: 0.44549887, Log Avg loss: 0.45350545, Global Avg Loss: 1.18295115, Time: 0.0075 Steps: 96430, Updated lr: 0.000010 Training, Epoch: 0046, Batch: 000635, Sample Num: 10160, Cur Loss: 0.76977062, Cur Avg Loss: 0.44597549, Log Avg loss: 0.47576426, Global Avg Loss: 1.18287782, Time: 0.0075 Steps: 96440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000645, Sample Num: 10320, Cur Loss: 0.50980163, Cur Avg Loss: 0.44536625, Log Avg loss: 0.40667941, Global Avg Loss: 1.18279735, Time: 0.0077 Steps: 96450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000655, Sample Num: 10480, Cur Loss: 0.25133228, Cur Avg Loss: 0.44502791, Log Avg loss: 0.42320514, Global Avg Loss: 1.18271860, Time: 0.0076 Steps: 96460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000665, Sample Num: 10640, Cur Loss: 0.55155593, Cur Avg Loss: 0.44457623, Log Avg loss: 0.41499107, Global Avg Loss: 1.18263902, Time: 0.0076 Steps: 96470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000675, Sample Num: 10800, Cur Loss: 0.75800264, Cur Avg Loss: 0.44542133, Log Avg loss: 0.50162046, Global Avg Loss: 1.18256843, Time: 0.0077 Steps: 96480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000685, Sample Num: 10960, Cur Loss: 0.30241239, Cur Avg Loss: 0.44671171, Log Avg loss: 0.53381270, Global Avg Loss: 1.18250119, Time: 0.0076 Steps: 96490, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000695, Sample Num: 11120, Cur Loss: 0.38723063, Cur Avg Loss: 0.44689060, Log Avg loss: 0.45914445, Global Avg Loss: 1.18242623, Time: 0.0076 Steps: 96500, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000705, Sample Num: 11280, Cur Loss: 0.51119232, Cur Avg Loss: 0.44804953, Log Avg loss: 0.52859492, Global Avg Loss: 1.18235849, Time: 0.0076 Steps: 96510, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000715, Sample Num: 11440, Cur Loss: 0.51948279, Cur Avg Loss: 0.44834738, Log Avg loss: 0.46934577, Global Avg Loss: 1.18228462, Time: 0.0077 Steps: 96520, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000725, Sample Num: 11600, Cur Loss: 0.37382263, Cur Avg Loss: 0.44765730, Log Avg loss: 0.39831694, Global Avg Loss: 1.18220340, Time: 0.0076 Steps: 96530, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000735, Sample Num: 11760, Cur Loss: 0.22044161, Cur Avg Loss: 0.44714163, Log Avg loss: 0.40975521, Global Avg Loss: 1.18212339, Time: 0.0076 Steps: 96540, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000745, Sample Num: 11920, Cur Loss: 0.44534826, Cur Avg Loss: 0.44864256, Log Avg loss: 0.55896075, Global Avg Loss: 1.18205884, Time: 0.0076 Steps: 96550, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000755, Sample Num: 12080, Cur Loss: 0.43413600, Cur Avg Loss: 0.45030716, Log Avg loss: 0.57432012, Global Avg Loss: 1.18199591, Time: 0.0078 Steps: 96560, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000765, Sample Num: 12240, Cur Loss: 0.37036467, Cur Avg Loss: 0.45007705, Log Avg loss: 0.43270338, Global Avg Loss: 1.18191831, Time: 0.0076 Steps: 96570, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000775, Sample Num: 12400, Cur Loss: 0.84438848, Cur Avg Loss: 0.45125809, Log Avg loss: 0.54160817, Global Avg Loss: 1.18185202, Time: 0.0075 Steps: 96580, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000785, Sample Num: 12560, Cur Loss: 0.31661582, Cur Avg Loss: 0.45153639, Log Avg loss: 0.47310417, Global Avg Loss: 1.18177864, Time: 0.0075 Steps: 96590, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000795, Sample Num: 12720, Cur Loss: 0.21252668, Cur Avg Loss: 0.45032827, Log Avg loss: 0.35549141, Global Avg Loss: 1.18169310, Time: 0.0076 Steps: 96600, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000805, Sample Num: 12880, Cur Loss: 0.28780639, Cur Avg Loss: 0.44990193, Log Avg loss: 0.41600772, Global Avg Loss: 1.18161385, Time: 0.0075 Steps: 96610, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000815, Sample Num: 13040, Cur Loss: 0.32877278, Cur Avg Loss: 0.44999768, Log Avg loss: 0.45770540, Global Avg Loss: 1.18153892, Time: 0.0074 Steps: 96620, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000825, Sample Num: 13200, Cur Loss: 0.34562641, Cur Avg Loss: 0.44978786, Log Avg loss: 0.43268797, Global Avg Loss: 1.18146143, Time: 0.0076 Steps: 96630, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000835, Sample Num: 13360, Cur Loss: 0.45573792, Cur Avg Loss: 0.44946762, Log Avg loss: 0.42304767, Global Avg Loss: 1.18138295, Time: 0.0075 Steps: 96640, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000845, Sample Num: 13520, Cur Loss: 1.02948165, Cur Avg Loss: 0.44983112, Log Avg loss: 0.48018279, Global Avg Loss: 1.18131040, Time: 0.0075 Steps: 96650, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000855, Sample Num: 13680, Cur Loss: 0.32020366, Cur Avg Loss: 0.44883512, Log Avg loss: 0.36467355, Global Avg Loss: 1.18122591, Time: 0.0074 Steps: 96660, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000865, Sample Num: 13840, Cur Loss: 0.36711088, Cur Avg Loss: 0.44980833, Log Avg loss: 0.53301777, Global Avg Loss: 1.18115886, Time: 0.0066 Steps: 96670, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000875, Sample Num: 14000, Cur Loss: 0.24467072, Cur Avg Loss: 0.45010968, Log Avg loss: 0.47617612, Global Avg Loss: 1.18108594, Time: 0.0067 Steps: 96680, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000885, Sample Num: 14160, Cur Loss: 1.06659508, Cur Avg Loss: 0.45080388, Log Avg loss: 0.51154637, Global Avg Loss: 1.18101669, Time: 0.0112 Steps: 96690, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000895, Sample Num: 14320, Cur Loss: 0.73013055, Cur Avg Loss: 0.45196203, Log Avg loss: 0.55445876, Global Avg Loss: 1.18095190, Time: 0.0065 Steps: 96700, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000905, Sample Num: 14480, Cur Loss: 0.74996448, Cur Avg Loss: 0.45171415, Log Avg loss: 0.42952914, Global Avg Loss: 1.18087420, Time: 0.0064 Steps: 96710, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000915, Sample Num: 14640, Cur Loss: 0.38281664, Cur Avg Loss: 0.45044318, Log Avg loss: 0.33542022, Global Avg Loss: 1.18078679, Time: 0.0065 Steps: 96720, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000925, Sample Num: 14800, Cur Loss: 0.60859054, Cur Avg Loss: 0.45442672, Log Avg loss: 0.81892085, Global Avg Loss: 1.18074938, Time: 0.0065 Steps: 96730, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000935, Sample Num: 14960, Cur Loss: 0.55922246, Cur Avg Loss: 0.45433720, Log Avg loss: 0.44605653, Global Avg Loss: 1.18067343, Time: 0.0065 Steps: 96740, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000945, Sample Num: 15120, Cur Loss: 0.20575462, Cur Avg Loss: 0.45376894, Log Avg loss: 0.40063632, Global Avg Loss: 1.18059281, Time: 0.0065 Steps: 96750, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000955, Sample Num: 15280, Cur Loss: 0.52136713, Cur Avg Loss: 0.45350443, Log Avg loss: 0.42850822, Global Avg Loss: 1.18051508, Time: 0.0064 Steps: 96760, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000965, Sample Num: 15440, Cur Loss: 0.26657039, Cur Avg Loss: 0.45534521, Log Avg loss: 0.63113956, Global Avg Loss: 1.18045831, Time: 0.0064 Steps: 96770, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000975, Sample Num: 15600, Cur Loss: 0.20254239, Cur Avg Loss: 0.45494035, Log Avg loss: 0.41587160, Global Avg Loss: 1.18037931, Time: 0.0072 Steps: 96780, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000985, Sample Num: 15760, Cur Loss: 0.27024242, Cur Avg Loss: 0.45412693, Log Avg loss: 0.37481822, Global Avg Loss: 1.18029608, Time: 0.0065 Steps: 96790, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 000995, Sample Num: 15920, Cur Loss: 0.32208443, Cur Avg Loss: 0.45432060, Log Avg loss: 0.47339680, Global Avg Loss: 1.18022305, Time: 0.0064 Steps: 96800, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001005, Sample Num: 16080, Cur Loss: 0.16101623, Cur Avg Loss: 0.45486069, Log Avg loss: 0.50860022, Global Avg Loss: 1.18015368, Time: 0.0064 Steps: 96810, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001015, Sample Num: 16240, Cur Loss: 0.45274258, Cur Avg Loss: 0.45591408, Log Avg loss: 0.56177987, Global Avg Loss: 1.18008981, Time: 0.0066 Steps: 96820, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001025, Sample Num: 16400, Cur Loss: 1.03965509, Cur Avg Loss: 0.45757208, Log Avg loss: 0.62585871, Global Avg Loss: 1.18003257, Time: 0.0064 Steps: 96830, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001035, Sample Num: 16560, Cur Loss: 0.30625987, Cur Avg Loss: 0.45831982, Log Avg loss: 0.53496318, Global Avg Loss: 1.17996596, Time: 0.0214 Steps: 96840, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001045, Sample Num: 16720, Cur Loss: 0.20165451, Cur Avg Loss: 0.45898495, Log Avg loss: 0.52782578, Global Avg Loss: 1.17989863, Time: 0.0223 Steps: 96850, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001055, Sample Num: 16880, Cur Loss: 0.55858749, Cur Avg Loss: 0.45972739, Log Avg loss: 0.53731295, Global Avg Loss: 1.17983228, Time: 0.0068 Steps: 96860, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001065, Sample Num: 17040, Cur Loss: 0.31314474, Cur Avg Loss: 0.45918827, Log Avg loss: 0.40231044, Global Avg Loss: 1.17975202, Time: 0.0066 Steps: 96870, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001075, Sample Num: 17200, Cur Loss: 0.24097911, Cur Avg Loss: 0.45860995, Log Avg loss: 0.39701921, Global Avg Loss: 1.17967123, Time: 0.0067 Steps: 96880, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001085, Sample Num: 17360, Cur Loss: 0.21763973, Cur Avg Loss: 0.45816487, Log Avg loss: 0.41031836, Global Avg Loss: 1.17959182, Time: 0.0067 Steps: 96890, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001095, Sample Num: 17520, Cur Loss: 1.03024817, Cur Avg Loss: 0.45973625, Log Avg loss: 0.63023183, Global Avg Loss: 1.17953513, Time: 0.0067 Steps: 96900, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001105, Sample Num: 17680, Cur Loss: 0.24212059, Cur Avg Loss: 0.45889495, Log Avg loss: 0.36677158, Global Avg Loss: 1.17945126, Time: 0.0067 Steps: 96910, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001115, Sample Num: 17840, Cur Loss: 0.35141063, Cur Avg Loss: 0.46031909, Log Avg loss: 0.61768743, Global Avg Loss: 1.17939330, Time: 0.0067 Steps: 96920, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001125, Sample Num: 18000, Cur Loss: 0.37112725, Cur Avg Loss: 0.45971990, Log Avg loss: 0.39290959, Global Avg Loss: 1.17931216, Time: 0.0068 Steps: 96930, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001135, Sample Num: 18160, Cur Loss: 0.15389134, Cur Avg Loss: 0.46026328, Log Avg loss: 0.52139433, Global Avg Loss: 1.17924429, Time: 0.0076 Steps: 96940, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001145, Sample Num: 18320, Cur Loss: 0.76605231, Cur Avg Loss: 0.46053194, Log Avg loss: 0.49102441, Global Avg Loss: 1.17917330, Time: 0.0075 Steps: 96950, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001155, Sample Num: 18480, Cur Loss: 0.22156107, Cur Avg Loss: 0.46004988, Log Avg loss: 0.40485357, Global Avg Loss: 1.17909344, Time: 0.0074 Steps: 96960, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001165, Sample Num: 18640, Cur Loss: 0.55302966, Cur Avg Loss: 0.45971074, Log Avg loss: 0.42054040, Global Avg Loss: 1.17901522, Time: 0.0074 Steps: 96970, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001175, Sample Num: 18800, Cur Loss: 0.15578483, Cur Avg Loss: 0.45982224, Log Avg loss: 0.47281259, Global Avg Loss: 1.17894240, Time: 0.0076 Steps: 96980, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001185, Sample Num: 18960, Cur Loss: 0.20856926, Cur Avg Loss: 0.45883287, Log Avg loss: 0.34258148, Global Avg Loss: 1.17885617, Time: 0.0074 Steps: 96990, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001195, Sample Num: 19120, Cur Loss: 0.13293904, Cur Avg Loss: 0.45935310, Log Avg loss: 0.52099978, Global Avg Loss: 1.17878835, Time: 0.0073 Steps: 97000, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001205, Sample Num: 19280, Cur Loss: 0.58483291, Cur Avg Loss: 0.45875719, Log Avg loss: 0.38754687, Global Avg Loss: 1.17870678, Time: 0.0076 Steps: 97010, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001215, Sample Num: 19440, Cur Loss: 0.25016466, Cur Avg Loss: 0.45837345, Log Avg loss: 0.41213235, Global Avg Loss: 1.17862777, Time: 0.0075 Steps: 97020, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001225, Sample Num: 19600, Cur Loss: 0.72973216, Cur Avg Loss: 0.45871266, Log Avg loss: 0.49992674, Global Avg Loss: 1.17855782, Time: 0.0074 Steps: 97030, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001235, Sample Num: 19760, Cur Loss: 0.52095199, Cur Avg Loss: 0.45864300, Log Avg loss: 0.45011013, Global Avg Loss: 1.17848276, Time: 0.0074 Steps: 97040, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001245, Sample Num: 19920, Cur Loss: 0.33954605, Cur Avg Loss: 0.45818349, Log Avg loss: 0.40143325, Global Avg Loss: 1.17840269, Time: 0.0074 Steps: 97050, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001255, Sample Num: 20080, Cur Loss: 0.58504760, Cur Avg Loss: 0.45866868, Log Avg loss: 0.51907497, Global Avg Loss: 1.17833476, Time: 0.0074 Steps: 97060, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001265, Sample Num: 20240, Cur Loss: 0.37556356, Cur Avg Loss: 0.45952733, Log Avg loss: 0.56728825, Global Avg Loss: 1.17827181, Time: 0.0076 Steps: 97070, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001275, Sample Num: 20400, Cur Loss: 0.14388894, Cur Avg Loss: 0.45917084, Log Avg loss: 0.41407454, Global Avg Loss: 1.17819309, Time: 0.0075 Steps: 97080, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001285, Sample Num: 20560, Cur Loss: 0.44157997, Cur Avg Loss: 0.45979330, Log Avg loss: 0.53915706, Global Avg Loss: 1.17812727, Time: 0.0075 Steps: 97090, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001295, Sample Num: 20720, Cur Loss: 0.47220027, Cur Avg Loss: 0.45957652, Log Avg loss: 0.43172025, Global Avg Loss: 1.17805040, Time: 0.0074 Steps: 97100, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001305, Sample Num: 20880, Cur Loss: 0.19863454, Cur Avg Loss: 0.45908639, Log Avg loss: 0.39561469, Global Avg Loss: 1.17796983, Time: 0.0074 Steps: 97110, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001315, Sample Num: 21040, Cur Loss: 0.34976202, Cur Avg Loss: 0.45861400, Log Avg loss: 0.39696642, Global Avg Loss: 1.17788942, Time: 0.0074 Steps: 97120, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001325, Sample Num: 21200, Cur Loss: 0.46334440, Cur Avg Loss: 0.45807709, Log Avg loss: 0.38747450, Global Avg Loss: 1.17780804, Time: 0.0074 Steps: 97130, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001335, Sample Num: 21360, Cur Loss: 1.10323000, Cur Avg Loss: 0.45750525, Log Avg loss: 0.38173554, Global Avg Loss: 1.17772609, Time: 0.0074 Steps: 97140, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001345, Sample Num: 21520, Cur Loss: 0.46954536, Cur Avg Loss: 0.45794651, Log Avg loss: 0.51685545, Global Avg Loss: 1.17765806, Time: 0.0075 Steps: 97150, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001355, Sample Num: 21680, Cur Loss: 0.32142162, Cur Avg Loss: 0.45782112, Log Avg loss: 0.44095610, Global Avg Loss: 1.17758224, Time: 0.0074 Steps: 97160, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001365, Sample Num: 21840, Cur Loss: 0.22026215, Cur Avg Loss: 0.45747702, Log Avg loss: 0.41085127, Global Avg Loss: 1.17750333, Time: 0.0075 Steps: 97170, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001375, Sample Num: 22000, Cur Loss: 0.23914133, Cur Avg Loss: 0.45686067, Log Avg loss: 0.37272842, Global Avg Loss: 1.17742052, Time: 0.0076 Steps: 97180, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001385, Sample Num: 22160, Cur Loss: 0.53073955, Cur Avg Loss: 0.45768974, Log Avg loss: 0.57168727, Global Avg Loss: 1.17735820, Time: 0.0075 Steps: 97190, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001395, Sample Num: 22320, Cur Loss: 0.19866109, Cur Avg Loss: 0.45709344, Log Avg loss: 0.37450573, Global Avg Loss: 1.17727560, Time: 0.0074 Steps: 97200, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001405, Sample Num: 22480, Cur Loss: 0.76351231, Cur Avg Loss: 0.45675998, Log Avg loss: 0.41024193, Global Avg Loss: 1.17719669, Time: 0.0074 Steps: 97210, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001415, Sample Num: 22640, Cur Loss: 0.95982409, Cur Avg Loss: 0.45754133, Log Avg loss: 0.56732156, Global Avg Loss: 1.17713396, Time: 0.0074 Steps: 97220, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001425, Sample Num: 22800, Cur Loss: 0.19286834, Cur Avg Loss: 0.45764660, Log Avg loss: 0.47254235, Global Avg Loss: 1.17706149, Time: 0.0074 Steps: 97230, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001435, Sample Num: 22960, Cur Loss: 0.42622107, Cur Avg Loss: 0.45741443, Log Avg loss: 0.42432951, Global Avg Loss: 1.17698408, Time: 0.0075 Steps: 97240, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001445, Sample Num: 23120, Cur Loss: 0.13473582, Cur Avg Loss: 0.45678545, Log Avg loss: 0.36652743, Global Avg Loss: 1.17690075, Time: 0.0074 Steps: 97250, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001455, Sample Num: 23280, Cur Loss: 0.62048817, Cur Avg Loss: 0.45607125, Log Avg loss: 0.35286887, Global Avg Loss: 1.17681602, Time: 0.0075 Steps: 97260, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001465, Sample Num: 23440, Cur Loss: 0.26718646, Cur Avg Loss: 0.45636373, Log Avg loss: 0.49892035, Global Avg Loss: 1.17674633, Time: 0.0075 Steps: 97270, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001475, Sample Num: 23600, Cur Loss: 0.59334856, Cur Avg Loss: 0.45651997, Log Avg loss: 0.47940811, Global Avg Loss: 1.17667465, Time: 0.0074 Steps: 97280, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001485, Sample Num: 23760, Cur Loss: 0.68168670, Cur Avg Loss: 0.45654094, Log Avg loss: 0.45963524, Global Avg Loss: 1.17660095, Time: 0.0074 Steps: 97290, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001495, Sample Num: 23920, Cur Loss: 0.44686708, Cur Avg Loss: 0.45634094, Log Avg loss: 0.42664082, Global Avg Loss: 1.17652387, Time: 0.0075 Steps: 97300, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001505, Sample Num: 24080, Cur Loss: 0.21218596, Cur Avg Loss: 0.45739115, Log Avg loss: 0.61439649, Global Avg Loss: 1.17646610, Time: 0.0074 Steps: 97310, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001515, Sample Num: 24240, Cur Loss: 0.89402401, Cur Avg Loss: 0.45801051, Log Avg loss: 0.55122441, Global Avg Loss: 1.17640186, Time: 0.0075 Steps: 97320, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001525, Sample Num: 24400, Cur Loss: 0.67022717, Cur Avg Loss: 0.45895231, Log Avg loss: 0.60163608, Global Avg Loss: 1.17634280, Time: 0.0075 Steps: 97330, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001535, Sample Num: 24560, Cur Loss: 0.12668344, Cur Avg Loss: 0.45996887, Log Avg loss: 0.61499396, Global Avg Loss: 1.17628513, Time: 0.0075 Steps: 97340, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001545, Sample Num: 24720, Cur Loss: 0.40733045, Cur Avg Loss: 0.45974982, Log Avg loss: 0.42612568, Global Avg Loss: 1.17620808, Time: 0.0075 Steps: 97350, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001555, Sample Num: 24880, Cur Loss: 0.51854700, Cur Avg Loss: 0.46021411, Log Avg loss: 0.53194596, Global Avg Loss: 1.17614190, Time: 0.0073 Steps: 97360, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001565, Sample Num: 25040, Cur Loss: 0.54161936, Cur Avg Loss: 0.46006823, Log Avg loss: 0.43738470, Global Avg Loss: 1.17606603, Time: 0.0074 Steps: 97370, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001575, Sample Num: 25200, Cur Loss: 0.70756185, Cur Avg Loss: 0.46029975, Log Avg loss: 0.49653245, Global Avg Loss: 1.17599625, Time: 0.0074 Steps: 97380, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001585, Sample Num: 25360, Cur Loss: 0.46827519, Cur Avg Loss: 0.46032516, Log Avg loss: 0.46432713, Global Avg Loss: 1.17592318, Time: 0.0075 Steps: 97390, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001595, Sample Num: 25520, Cur Loss: 0.65972936, Cur Avg Loss: 0.46017174, Log Avg loss: 0.43585429, Global Avg Loss: 1.17584719, Time: 0.0074 Steps: 97400, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001605, Sample Num: 25680, Cur Loss: 0.29926747, Cur Avg Loss: 0.46056755, Log Avg loss: 0.52369933, Global Avg Loss: 1.17578024, Time: 0.0075 Steps: 97410, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001615, Sample Num: 25840, Cur Loss: 0.37449238, Cur Avg Loss: 0.46032249, Log Avg loss: 0.42099026, Global Avg Loss: 1.17570277, Time: 0.0074 Steps: 97420, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001625, Sample Num: 26000, Cur Loss: 0.72967535, Cur Avg Loss: 0.46079881, Log Avg loss: 0.53772432, Global Avg Loss: 1.17563729, Time: 0.0074 Steps: 97430, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001635, Sample Num: 26160, Cur Loss: 0.47312579, Cur Avg Loss: 0.46156600, Log Avg loss: 0.58623550, Global Avg Loss: 1.17557680, Time: 0.0075 Steps: 97440, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001645, Sample Num: 26320, Cur Loss: 0.83007902, Cur Avg Loss: 0.46167666, Log Avg loss: 0.47976942, Global Avg Loss: 1.17550540, Time: 0.0076 Steps: 97450, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001655, Sample Num: 26480, Cur Loss: 0.55205560, Cur Avg Loss: 0.46185783, Log Avg loss: 0.49166045, Global Avg Loss: 1.17543523, Time: 0.0074 Steps: 97460, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001665, Sample Num: 26640, Cur Loss: 0.22663862, Cur Avg Loss: 0.46295352, Log Avg loss: 0.64429030, Global Avg Loss: 1.17538074, Time: 0.0075 Steps: 97470, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001675, Sample Num: 26800, Cur Loss: 0.48486483, Cur Avg Loss: 0.46282513, Log Avg loss: 0.44144724, Global Avg Loss: 1.17530545, Time: 0.0075 Steps: 97480, Updated lr: 0.000009 Training, Epoch: 0046, Batch: 001685, Sample Num: 26960, Cur Loss: 0.15077992, Cur Avg Loss: 0.46267354, Log Avg loss: 0.43728192, Global Avg Loss: 1.17522974, Time: 0.0074 Steps: 97490, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001695, Sample Num: 27120, Cur Loss: 0.68383336, Cur Avg Loss: 0.46184730, Log Avg loss: 0.32262706, Global Avg Loss: 1.17514230, Time: 0.0074 Steps: 97500, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001705, Sample Num: 27280, Cur Loss: 0.63953853, Cur Avg Loss: 0.46193956, Log Avg loss: 0.47757783, Global Avg Loss: 1.17507076, Time: 0.0078 Steps: 97510, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001715, Sample Num: 27440, Cur Loss: 0.87655902, Cur Avg Loss: 0.46174020, Log Avg loss: 0.42774887, Global Avg Loss: 1.17499413, Time: 0.0073 Steps: 97520, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001725, Sample Num: 27600, Cur Loss: 0.74406922, Cur Avg Loss: 0.46219668, Log Avg loss: 0.54048306, Global Avg Loss: 1.17492907, Time: 0.0064 Steps: 97530, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001735, Sample Num: 27760, Cur Loss: 1.18315268, Cur Avg Loss: 0.46328657, Log Avg loss: 0.65129207, Global Avg Loss: 1.17487538, Time: 0.0067 Steps: 97540, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001745, Sample Num: 27920, Cur Loss: 0.34432542, Cur Avg Loss: 0.46306946, Log Avg loss: 0.42540110, Global Avg Loss: 1.17479855, Time: 0.0064 Steps: 97550, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001755, Sample Num: 28080, Cur Loss: 0.53945386, Cur Avg Loss: 0.46274235, Log Avg loss: 0.40566157, Global Avg Loss: 1.17471972, Time: 0.0067 Steps: 97560, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001765, Sample Num: 28240, Cur Loss: 1.09098709, Cur Avg Loss: 0.46253261, Log Avg loss: 0.42572241, Global Avg Loss: 1.17464295, Time: 0.0072 Steps: 97570, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001775, Sample Num: 28400, Cur Loss: 0.24966399, Cur Avg Loss: 0.46250491, Log Avg loss: 0.45761643, Global Avg Loss: 1.17456947, Time: 0.0212 Steps: 97580, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001785, Sample Num: 28560, Cur Loss: 0.36695737, Cur Avg Loss: 0.46242240, Log Avg loss: 0.44777750, Global Avg Loss: 1.17449500, Time: 0.0139 Steps: 97590, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001795, Sample Num: 28720, Cur Loss: 0.22644849, Cur Avg Loss: 0.46209257, Log Avg loss: 0.40321811, Global Avg Loss: 1.17441597, Time: 0.0183 Steps: 97600, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001805, Sample Num: 28880, Cur Loss: 0.24852630, Cur Avg Loss: 0.46270055, Log Avg loss: 0.57183275, Global Avg Loss: 1.17435424, Time: 0.0219 Steps: 97610, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001815, Sample Num: 29040, Cur Loss: 0.72011709, Cur Avg Loss: 0.46248805, Log Avg loss: 0.42413065, Global Avg Loss: 1.17427739, Time: 0.0153 Steps: 97620, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001825, Sample Num: 29200, Cur Loss: 0.21817139, Cur Avg Loss: 0.46213124, Log Avg loss: 0.39737168, Global Avg Loss: 1.17419781, Time: 0.0065 Steps: 97630, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001835, Sample Num: 29360, Cur Loss: 0.36987358, Cur Avg Loss: 0.46235005, Log Avg loss: 0.50228275, Global Avg Loss: 1.17412899, Time: 0.0219 Steps: 97640, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001845, Sample Num: 29520, Cur Loss: 0.20175260, Cur Avg Loss: 0.46205606, Log Avg loss: 0.40810793, Global Avg Loss: 1.17405055, Time: 0.0084 Steps: 97650, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001855, Sample Num: 29680, Cur Loss: 0.32525691, Cur Avg Loss: 0.46208559, Log Avg loss: 0.46753516, Global Avg Loss: 1.17397820, Time: 0.0066 Steps: 97660, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001865, Sample Num: 29840, Cur Loss: 0.48320580, Cur Avg Loss: 0.46171004, Log Avg loss: 0.39204414, Global Avg Loss: 1.17389815, Time: 0.0066 Steps: 97670, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001875, Sample Num: 30000, Cur Loss: 0.19475153, Cur Avg Loss: 0.46134433, Log Avg loss: 0.39313929, Global Avg Loss: 1.17381822, Time: 0.0063 Steps: 97680, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001885, Sample Num: 30160, Cur Loss: 0.25878036, Cur Avg Loss: 0.46071420, Log Avg loss: 0.34256555, Global Avg Loss: 1.17373312, Time: 0.0063 Steps: 97690, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001895, Sample Num: 30320, Cur Loss: 0.55720907, Cur Avg Loss: 0.46154902, Log Avg loss: 0.61891186, Global Avg Loss: 1.17367634, Time: 0.0063 Steps: 97700, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001905, Sample Num: 30480, Cur Loss: 0.45631820, Cur Avg Loss: 0.46153363, Log Avg loss: 0.45861695, Global Avg Loss: 1.17360315, Time: 0.0237 Steps: 97710, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001915, Sample Num: 30640, Cur Loss: 0.47078335, Cur Avg Loss: 0.46100916, Log Avg loss: 0.36109907, Global Avg Loss: 1.17352001, Time: 0.0064 Steps: 97720, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001925, Sample Num: 30800, Cur Loss: 0.86399245, Cur Avg Loss: 0.46120613, Log Avg loss: 0.49892517, Global Avg Loss: 1.17345098, Time: 0.0114 Steps: 97730, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001935, Sample Num: 30960, Cur Loss: 0.65152252, Cur Avg Loss: 0.46121424, Log Avg loss: 0.46277469, Global Avg Loss: 1.17337827, Time: 0.0064 Steps: 97740, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001945, Sample Num: 31120, Cur Loss: 0.63440740, Cur Avg Loss: 0.46151828, Log Avg loss: 0.52035178, Global Avg Loss: 1.17331147, Time: 0.0077 Steps: 97750, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001955, Sample Num: 31280, Cur Loss: 0.41459450, Cur Avg Loss: 0.46241817, Log Avg loss: 0.63744597, Global Avg Loss: 1.17325665, Time: 0.0064 Steps: 97760, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001965, Sample Num: 31440, Cur Loss: 0.61701041, Cur Avg Loss: 0.46233594, Log Avg loss: 0.44626081, Global Avg Loss: 1.17318229, Time: 0.0065 Steps: 97770, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001975, Sample Num: 31600, Cur Loss: 0.73347366, Cur Avg Loss: 0.46206137, Log Avg loss: 0.40810661, Global Avg Loss: 1.17310405, Time: 0.0065 Steps: 97780, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001985, Sample Num: 31760, Cur Loss: 0.20586941, Cur Avg Loss: 0.46198483, Log Avg loss: 0.44686892, Global Avg Loss: 1.17302978, Time: 0.0065 Steps: 97790, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 001995, Sample Num: 31920, Cur Loss: 0.42449248, Cur Avg Loss: 0.46184358, Log Avg loss: 0.43380506, Global Avg Loss: 1.17295420, Time: 0.0065 Steps: 97800, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002005, Sample Num: 32080, Cur Loss: 0.26731187, Cur Avg Loss: 0.46232674, Log Avg loss: 0.55871815, Global Avg Loss: 1.17289140, Time: 0.0065 Steps: 97810, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002015, Sample Num: 32240, Cur Loss: 0.42736667, Cur Avg Loss: 0.46266611, Log Avg loss: 0.53070934, Global Avg Loss: 1.17282575, Time: 0.0065 Steps: 97820, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002025, Sample Num: 32400, Cur Loss: 0.35667050, Cur Avg Loss: 0.46251514, Log Avg loss: 0.43209377, Global Avg Loss: 1.17275003, Time: 0.0066 Steps: 97830, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002035, Sample Num: 32560, Cur Loss: 0.39262092, Cur Avg Loss: 0.46264375, Log Avg loss: 0.48868783, Global Avg Loss: 1.17268012, Time: 0.0074 Steps: 97840, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002045, Sample Num: 32720, Cur Loss: 0.49041456, Cur Avg Loss: 0.46211771, Log Avg loss: 0.35506926, Global Avg Loss: 1.17259656, Time: 0.0074 Steps: 97850, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002055, Sample Num: 32880, Cur Loss: 0.36845064, Cur Avg Loss: 0.46154945, Log Avg loss: 0.34533929, Global Avg Loss: 1.17251203, Time: 0.0074 Steps: 97860, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002065, Sample Num: 33040, Cur Loss: 0.29629856, Cur Avg Loss: 0.46100360, Log Avg loss: 0.34883186, Global Avg Loss: 1.17242786, Time: 0.0074 Steps: 97870, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002075, Sample Num: 33200, Cur Loss: 0.47991192, Cur Avg Loss: 0.46116798, Log Avg loss: 0.49511302, Global Avg Loss: 1.17235867, Time: 0.0074 Steps: 97880, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002085, Sample Num: 33360, Cur Loss: 0.30125678, Cur Avg Loss: 0.46109408, Log Avg loss: 0.44575873, Global Avg Loss: 1.17228444, Time: 0.0075 Steps: 97890, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002095, Sample Num: 33520, Cur Loss: 1.34994614, Cur Avg Loss: 0.46169051, Log Avg loss: 0.58604659, Global Avg Loss: 1.17222456, Time: 0.0075 Steps: 97900, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002105, Sample Num: 33680, Cur Loss: 0.15838635, Cur Avg Loss: 0.46156471, Log Avg loss: 0.43521027, Global Avg Loss: 1.17214928, Time: 0.0075 Steps: 97910, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002115, Sample Num: 33840, Cur Loss: 0.99009407, Cur Avg Loss: 0.46160279, Log Avg loss: 0.46961733, Global Avg Loss: 1.17207754, Time: 0.0079 Steps: 97920, Updated lr: 0.000008 Training, Epoch: 0046, Batch: 002125, Sample Num: 34000, Cur Loss: 0.77613902, Cur Avg Loss: 0.46116242, Log Avg loss: 0.36802460, Global Avg Loss: 1.17199543, Time: 0.0078 Steps: 97930, Updated lr: 0.000008 ***** Running evaluation checkpoint-97934 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-97934 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.897568, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.564394, "eval_total_loss": 396.76914, "eval_mae": 0.553743, "eval_mse": 0.564553, "eval_r2": 0.641133, "eval_sp_statistic": 0.756978, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.800871, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.912234, "test_total_loss": 457.941222, "test_mae": 0.659774, "test_mse": 0.912499, "test_r2": 0.411065, "test_sp_statistic": 0.602648, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.672431, "test_ps_pvalue": 0.0, "lr": 8.075865339023234e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.171956971155961, "train_cur_epoch_loss": 980.8913121148944, "train_cur_epoch_avg_loss": 0.46072865763968734, "train_cur_epoch_time": 17.897567987442017, "train_cur_epoch_avg_time": 0.008406560820780656, "epoch": 46, "step": 97934} ################################################## Training, Epoch: 0047, Batch: 000006, Sample Num: 96, Cur Loss: 0.84698594, Cur Avg Loss: 0.41035912, Log Avg loss: 0.33833297, Global Avg Loss: 1.17191031, Time: 0.0075 Steps: 97940, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000016, Sample Num: 256, Cur Loss: 0.20738685, Cur Avg Loss: 0.41669057, Log Avg loss: 0.42048945, Global Avg Loss: 1.17183360, Time: 0.0074 Steps: 97950, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000026, Sample Num: 416, Cur Loss: 0.34277332, Cur Avg Loss: 0.40037517, Log Avg loss: 0.37427053, Global Avg Loss: 1.17175218, Time: 0.0074 Steps: 97960, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000036, Sample Num: 576, Cur Loss: 0.45195991, Cur Avg Loss: 0.40978443, Log Avg loss: 0.43424851, Global Avg Loss: 1.17167690, Time: 0.0075 Steps: 97970, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000046, Sample Num: 736, Cur Loss: 0.54559726, Cur Avg Loss: 0.43469136, Log Avg loss: 0.52435628, Global Avg Loss: 1.17161084, Time: 0.0074 Steps: 97980, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000056, Sample Num: 896, Cur Loss: 0.30767429, Cur Avg Loss: 0.42951463, Log Avg loss: 0.40570167, Global Avg Loss: 1.17153268, Time: 0.0074 Steps: 97990, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000066, Sample Num: 1056, Cur Loss: 1.12221932, Cur Avg Loss: 0.43384539, Log Avg loss: 0.45809763, Global Avg Loss: 1.17145988, Time: 0.0074 Steps: 98000, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000076, Sample Num: 1216, Cur Loss: 0.65458411, Cur Avg Loss: 0.43355751, Log Avg loss: 0.43165755, Global Avg Loss: 1.17138439, Time: 0.0074 Steps: 98010, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000086, Sample Num: 1376, Cur Loss: 0.63809729, Cur Avg Loss: 0.42541910, Log Avg loss: 0.36356712, Global Avg Loss: 1.17130198, Time: 0.0074 Steps: 98020, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000096, Sample Num: 1536, Cur Loss: 0.36298582, Cur Avg Loss: 0.42768958, Log Avg loss: 0.44721579, Global Avg Loss: 1.17122812, Time: 0.0075 Steps: 98030, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000106, Sample Num: 1696, Cur Loss: 0.65536690, Cur Avg Loss: 0.43261397, Log Avg loss: 0.47988811, Global Avg Loss: 1.17115760, Time: 0.0074 Steps: 98040, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000116, Sample Num: 1856, Cur Loss: 0.92103946, Cur Avg Loss: 0.43733988, Log Avg loss: 0.48743454, Global Avg Loss: 1.17108787, Time: 0.0074 Steps: 98050, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000126, Sample Num: 2016, Cur Loss: 0.30119413, Cur Avg Loss: 0.44153461, Log Avg loss: 0.49019348, Global Avg Loss: 1.17101843, Time: 0.0074 Steps: 98060, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000136, Sample Num: 2176, Cur Loss: 0.62161559, Cur Avg Loss: 0.43592307, Log Avg loss: 0.36521767, Global Avg Loss: 1.17093627, Time: 0.0074 Steps: 98070, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000146, Sample Num: 2336, Cur Loss: 0.24330074, Cur Avg Loss: 0.43024987, Log Avg loss: 0.35309432, Global Avg Loss: 1.17085288, Time: 0.0074 Steps: 98080, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000156, Sample Num: 2496, Cur Loss: 1.07309246, Cur Avg Loss: 0.42439159, Log Avg loss: 0.33886065, Global Avg Loss: 1.17076806, Time: 0.0074 Steps: 98090, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000166, Sample Num: 2656, Cur Loss: 0.35680985, Cur Avg Loss: 0.42963931, Log Avg loss: 0.51150383, Global Avg Loss: 1.17070086, Time: 0.0074 Steps: 98100, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000176, Sample Num: 2816, Cur Loss: 0.48934388, Cur Avg Loss: 0.43269115, Log Avg loss: 0.48335157, Global Avg Loss: 1.17063080, Time: 0.0074 Steps: 98110, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000186, Sample Num: 2976, Cur Loss: 0.61916423, Cur Avg Loss: 0.43183538, Log Avg loss: 0.41677388, Global Avg Loss: 1.17055397, Time: 0.0075 Steps: 98120, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000196, Sample Num: 3136, Cur Loss: 0.28497589, Cur Avg Loss: 0.42505054, Log Avg loss: 0.29885257, Global Avg Loss: 1.17046514, Time: 0.0075 Steps: 98130, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000206, Sample Num: 3296, Cur Loss: 0.57316226, Cur Avg Loss: 0.42493077, Log Avg loss: 0.42258313, Global Avg Loss: 1.17038893, Time: 0.0074 Steps: 98140, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000216, Sample Num: 3456, Cur Loss: 0.16601735, Cur Avg Loss: 0.42530106, Log Avg loss: 0.43292906, Global Avg Loss: 1.17031380, Time: 0.0074 Steps: 98150, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000226, Sample Num: 3616, Cur Loss: 0.27220386, Cur Avg Loss: 0.42174177, Log Avg loss: 0.34486108, Global Avg Loss: 1.17022970, Time: 0.0074 Steps: 98160, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000236, Sample Num: 3776, Cur Loss: 0.27053428, Cur Avg Loss: 0.42440724, Log Avg loss: 0.48464705, Global Avg Loss: 1.17015987, Time: 0.0074 Steps: 98170, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000246, Sample Num: 3936, Cur Loss: 0.41710544, Cur Avg Loss: 0.43071021, Log Avg loss: 0.57946011, Global Avg Loss: 1.17009970, Time: 0.0075 Steps: 98180, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000256, Sample Num: 4096, Cur Loss: 0.33008987, Cur Avg Loss: 0.43491401, Log Avg loss: 0.53832768, Global Avg Loss: 1.17003536, Time: 0.0084 Steps: 98190, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000266, Sample Num: 4256, Cur Loss: 0.20477338, Cur Avg Loss: 0.42926684, Log Avg loss: 0.28469920, Global Avg Loss: 1.16994520, Time: 0.0074 Steps: 98200, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000276, Sample Num: 4416, Cur Loss: 0.55674511, Cur Avg Loss: 0.43319371, Log Avg loss: 0.53764844, Global Avg Loss: 1.16988082, Time: 0.0074 Steps: 98210, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000286, Sample Num: 4576, Cur Loss: 0.84234297, Cur Avg Loss: 0.44043116, Log Avg loss: 0.64018470, Global Avg Loss: 1.16982689, Time: 0.0075 Steps: 98220, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000296, Sample Num: 4736, Cur Loss: 0.15492947, Cur Avg Loss: 0.43922561, Log Avg loss: 0.40474705, Global Avg Loss: 1.16974901, Time: 0.0076 Steps: 98230, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000306, Sample Num: 4896, Cur Loss: 0.31985402, Cur Avg Loss: 0.43839951, Log Avg loss: 0.41394680, Global Avg Loss: 1.16967207, Time: 0.0074 Steps: 98240, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000316, Sample Num: 5056, Cur Loss: 0.62035018, Cur Avg Loss: 0.44090900, Log Avg loss: 0.51769953, Global Avg Loss: 1.16960571, Time: 0.0074 Steps: 98250, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000326, Sample Num: 5216, Cur Loss: 0.19175169, Cur Avg Loss: 0.44055974, Log Avg loss: 0.42952323, Global Avg Loss: 1.16953039, Time: 0.0074 Steps: 98260, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000336, Sample Num: 5376, Cur Loss: 0.35651124, Cur Avg Loss: 0.44012488, Log Avg loss: 0.42594817, Global Avg Loss: 1.16945473, Time: 0.0074 Steps: 98270, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000346, Sample Num: 5536, Cur Loss: 0.60420501, Cur Avg Loss: 0.44283577, Log Avg loss: 0.53392166, Global Avg Loss: 1.16939006, Time: 0.0075 Steps: 98280, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000356, Sample Num: 5696, Cur Loss: 0.44510919, Cur Avg Loss: 0.44334216, Log Avg loss: 0.46086329, Global Avg Loss: 1.16931798, Time: 0.0074 Steps: 98290, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000366, Sample Num: 5856, Cur Loss: 0.52702111, Cur Avg Loss: 0.44381124, Log Avg loss: 0.46051078, Global Avg Loss: 1.16924587, Time: 0.0074 Steps: 98300, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000376, Sample Num: 6016, Cur Loss: 0.30055583, Cur Avg Loss: 0.44310064, Log Avg loss: 0.41709238, Global Avg Loss: 1.16916936, Time: 0.0074 Steps: 98310, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000386, Sample Num: 6176, Cur Loss: 0.30533418, Cur Avg Loss: 0.44252394, Log Avg loss: 0.42084026, Global Avg Loss: 1.16909325, Time: 0.0075 Steps: 98320, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000396, Sample Num: 6336, Cur Loss: 0.15684390, Cur Avg Loss: 0.44081203, Log Avg loss: 0.37473205, Global Avg Loss: 1.16901246, Time: 0.0075 Steps: 98330, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000406, Sample Num: 6496, Cur Loss: 0.16284680, Cur Avg Loss: 0.44563484, Log Avg loss: 0.63661819, Global Avg Loss: 1.16895833, Time: 0.0152 Steps: 98340, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000416, Sample Num: 6656, Cur Loss: 0.65470272, Cur Avg Loss: 0.44766183, Log Avg loss: 0.52995777, Global Avg Loss: 1.16889335, Time: 0.0174 Steps: 98350, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000426, Sample Num: 6816, Cur Loss: 0.25428665, Cur Avg Loss: 0.44664071, Log Avg loss: 0.40416196, Global Avg Loss: 1.16881561, Time: 0.0189 Steps: 98360, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000436, Sample Num: 6976, Cur Loss: 0.44206798, Cur Avg Loss: 0.44700595, Log Avg loss: 0.46256536, Global Avg Loss: 1.16874381, Time: 0.0074 Steps: 98370, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000446, Sample Num: 7136, Cur Loss: 0.59655762, Cur Avg Loss: 0.44587605, Log Avg loss: 0.39661216, Global Avg Loss: 1.16866533, Time: 0.0223 Steps: 98380, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000456, Sample Num: 7296, Cur Loss: 0.22175084, Cur Avg Loss: 0.44769561, Log Avg loss: 0.52884810, Global Avg Loss: 1.16860030, Time: 0.0221 Steps: 98390, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000466, Sample Num: 7456, Cur Loss: 0.51249230, Cur Avg Loss: 0.44960270, Log Avg loss: 0.53656592, Global Avg Loss: 1.16853607, Time: 0.0176 Steps: 98400, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000476, Sample Num: 7616, Cur Loss: 0.82502127, Cur Avg Loss: 0.44785608, Log Avg loss: 0.36646365, Global Avg Loss: 1.16845456, Time: 0.0067 Steps: 98410, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000486, Sample Num: 7776, Cur Loss: 0.86533976, Cur Avg Loss: 0.44904404, Log Avg loss: 0.50559096, Global Avg Loss: 1.16838721, Time: 0.0188 Steps: 98420, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000496, Sample Num: 7936, Cur Loss: 0.19102532, Cur Avg Loss: 0.44963062, Log Avg loss: 0.47813824, Global Avg Loss: 1.16831709, Time: 0.0067 Steps: 98430, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000506, Sample Num: 8096, Cur Loss: 0.27566162, Cur Avg Loss: 0.44832670, Log Avg loss: 0.38365225, Global Avg Loss: 1.16823738, Time: 0.0067 Steps: 98440, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000516, Sample Num: 8256, Cur Loss: 0.51868927, Cur Avg Loss: 0.44746906, Log Avg loss: 0.40407265, Global Avg Loss: 1.16815976, Time: 0.0066 Steps: 98450, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000526, Sample Num: 8416, Cur Loss: 0.19899204, Cur Avg Loss: 0.44792269, Log Avg loss: 0.47132996, Global Avg Loss: 1.16808898, Time: 0.0064 Steps: 98460, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000536, Sample Num: 8576, Cur Loss: 0.59109861, Cur Avg Loss: 0.44766575, Log Avg loss: 0.43415097, Global Avg Loss: 1.16801445, Time: 0.0073 Steps: 98470, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000546, Sample Num: 8736, Cur Loss: 0.24412850, Cur Avg Loss: 0.44716018, Log Avg loss: 0.42006164, Global Avg Loss: 1.16793850, Time: 0.0065 Steps: 98480, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000556, Sample Num: 8896, Cur Loss: 0.62385762, Cur Avg Loss: 0.44588645, Log Avg loss: 0.37634051, Global Avg Loss: 1.16785813, Time: 0.0164 Steps: 98490, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000566, Sample Num: 9056, Cur Loss: 0.81671417, Cur Avg Loss: 0.44557073, Log Avg loss: 0.42801652, Global Avg Loss: 1.16778302, Time: 0.0068 Steps: 98500, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000576, Sample Num: 9216, Cur Loss: 0.50335383, Cur Avg Loss: 0.44355775, Log Avg loss: 0.32962332, Global Avg Loss: 1.16769793, Time: 0.0064 Steps: 98510, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000586, Sample Num: 9376, Cur Loss: 0.43309000, Cur Avg Loss: 0.44519114, Log Avg loss: 0.53927421, Global Avg Loss: 1.16763415, Time: 0.0068 Steps: 98520, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000596, Sample Num: 9536, Cur Loss: 0.44814116, Cur Avg Loss: 0.44459644, Log Avg loss: 0.40974728, Global Avg Loss: 1.16755723, Time: 0.0066 Steps: 98530, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000606, Sample Num: 9696, Cur Loss: 0.69187915, Cur Avg Loss: 0.44380373, Log Avg loss: 0.39655841, Global Avg Loss: 1.16747898, Time: 0.0064 Steps: 98540, Updated lr: 0.000008 Training, Epoch: 0047, Batch: 000616, Sample Num: 9856, Cur Loss: 0.30474657, Cur Avg Loss: 0.44438439, Log Avg loss: 0.47957229, Global Avg Loss: 1.16740918, Time: 0.0228 Steps: 98550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000626, Sample Num: 10016, Cur Loss: 0.62609655, Cur Avg Loss: 0.44665908, Log Avg loss: 0.58677969, Global Avg Loss: 1.16735027, Time: 0.0127 Steps: 98560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000636, Sample Num: 10176, Cur Loss: 0.49968946, Cur Avg Loss: 0.44604950, Log Avg loss: 0.40789007, Global Avg Loss: 1.16727322, Time: 0.0064 Steps: 98570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000646, Sample Num: 10336, Cur Loss: 0.21124107, Cur Avg Loss: 0.44553929, Log Avg loss: 0.41309010, Global Avg Loss: 1.16719672, Time: 0.0067 Steps: 98580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000656, Sample Num: 10496, Cur Loss: 0.26403332, Cur Avg Loss: 0.44488878, Log Avg loss: 0.40286554, Global Avg Loss: 1.16711919, Time: 0.0066 Steps: 98590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000666, Sample Num: 10656, Cur Loss: 0.45703015, Cur Avg Loss: 0.44615169, Log Avg loss: 0.52899863, Global Avg Loss: 1.16705447, Time: 0.0065 Steps: 98600, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000676, Sample Num: 10816, Cur Loss: 0.52288926, Cur Avg Loss: 0.44728281, Log Avg loss: 0.52261530, Global Avg Loss: 1.16698912, Time: 0.0065 Steps: 98610, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000686, Sample Num: 10976, Cur Loss: 0.14406681, Cur Avg Loss: 0.44812827, Log Avg loss: 0.50528161, Global Avg Loss: 1.16692202, Time: 0.0066 Steps: 98620, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000696, Sample Num: 11136, Cur Loss: 0.30672500, Cur Avg Loss: 0.44750726, Log Avg loss: 0.40490559, Global Avg Loss: 1.16684476, Time: 0.0066 Steps: 98630, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000706, Sample Num: 11296, Cur Loss: 0.30808538, Cur Avg Loss: 0.44766694, Log Avg loss: 0.45878079, Global Avg Loss: 1.16677298, Time: 0.0077 Steps: 98640, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000716, Sample Num: 11456, Cur Loss: 0.30053830, Cur Avg Loss: 0.44621491, Log Avg loss: 0.34370189, Global Avg Loss: 1.16668955, Time: 0.0075 Steps: 98650, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000726, Sample Num: 11616, Cur Loss: 0.13360237, Cur Avg Loss: 0.44528384, Log Avg loss: 0.37861914, Global Avg Loss: 1.16660967, Time: 0.0076 Steps: 98660, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000736, Sample Num: 11776, Cur Loss: 0.64074075, Cur Avg Loss: 0.44564642, Log Avg loss: 0.47196969, Global Avg Loss: 1.16653927, Time: 0.0076 Steps: 98670, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000746, Sample Num: 11936, Cur Loss: 0.43573594, Cur Avg Loss: 0.44430779, Log Avg loss: 0.34578472, Global Avg Loss: 1.16645610, Time: 0.0076 Steps: 98680, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000756, Sample Num: 12096, Cur Loss: 0.23755267, Cur Avg Loss: 0.44454296, Log Avg loss: 0.46208680, Global Avg Loss: 1.16638472, Time: 0.0075 Steps: 98690, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000766, Sample Num: 12256, Cur Loss: 0.26522362, Cur Avg Loss: 0.44463430, Log Avg loss: 0.45153922, Global Avg Loss: 1.16631230, Time: 0.0074 Steps: 98700, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000776, Sample Num: 12416, Cur Loss: 0.62940723, Cur Avg Loss: 0.44574716, Log Avg loss: 0.53099231, Global Avg Loss: 1.16624794, Time: 0.0075 Steps: 98710, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000786, Sample Num: 12576, Cur Loss: 0.68229717, Cur Avg Loss: 0.44517353, Log Avg loss: 0.40065980, Global Avg Loss: 1.16617039, Time: 0.0075 Steps: 98720, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000796, Sample Num: 12736, Cur Loss: 0.70064378, Cur Avg Loss: 0.44572845, Log Avg loss: 0.48934531, Global Avg Loss: 1.16610183, Time: 0.0077 Steps: 98730, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000806, Sample Num: 12896, Cur Loss: 0.75120085, Cur Avg Loss: 0.44878856, Log Avg loss: 0.69237300, Global Avg Loss: 1.16605385, Time: 0.0076 Steps: 98740, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000816, Sample Num: 13056, Cur Loss: 0.28568792, Cur Avg Loss: 0.44845403, Log Avg loss: 0.42149087, Global Avg Loss: 1.16597846, Time: 0.0076 Steps: 98750, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000826, Sample Num: 13216, Cur Loss: 1.00910699, Cur Avg Loss: 0.44977202, Log Avg loss: 0.55731982, Global Avg Loss: 1.16591683, Time: 0.0077 Steps: 98760, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000836, Sample Num: 13376, Cur Loss: 0.37802416, Cur Avg Loss: 0.45020815, Log Avg loss: 0.48623281, Global Avg Loss: 1.16584801, Time: 0.0076 Steps: 98770, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000846, Sample Num: 13536, Cur Loss: 0.19154130, Cur Avg Loss: 0.45120433, Log Avg loss: 0.53448498, Global Avg Loss: 1.16578409, Time: 0.0076 Steps: 98780, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000856, Sample Num: 13696, Cur Loss: 0.39152002, Cur Avg Loss: 0.45277461, Log Avg loss: 0.58562011, Global Avg Loss: 1.16572537, Time: 0.0075 Steps: 98790, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000866, Sample Num: 13856, Cur Loss: 0.37528491, Cur Avg Loss: 0.45336467, Log Avg loss: 0.50387437, Global Avg Loss: 1.16565838, Time: 0.0076 Steps: 98800, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000876, Sample Num: 14016, Cur Loss: 1.07888091, Cur Avg Loss: 0.45577856, Log Avg loss: 0.66482142, Global Avg Loss: 1.16560769, Time: 0.0077 Steps: 98810, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000886, Sample Num: 14176, Cur Loss: 0.30605039, Cur Avg Loss: 0.45510685, Log Avg loss: 0.39626451, Global Avg Loss: 1.16552984, Time: 0.0076 Steps: 98820, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000896, Sample Num: 14336, Cur Loss: 0.16997170, Cur Avg Loss: 0.45579720, Log Avg loss: 0.51696235, Global Avg Loss: 1.16546421, Time: 0.0077 Steps: 98830, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000906, Sample Num: 14496, Cur Loss: 0.44442847, Cur Avg Loss: 0.45639019, Log Avg loss: 0.50952245, Global Avg Loss: 1.16539785, Time: 0.0075 Steps: 98840, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000916, Sample Num: 14656, Cur Loss: 0.82916379, Cur Avg Loss: 0.45674965, Log Avg loss: 0.48931619, Global Avg Loss: 1.16532946, Time: 0.0075 Steps: 98850, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000926, Sample Num: 14816, Cur Loss: 0.28354430, Cur Avg Loss: 0.45751042, Log Avg loss: 0.52719712, Global Avg Loss: 1.16526491, Time: 0.0076 Steps: 98860, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000936, Sample Num: 14976, Cur Loss: 0.21852401, Cur Avg Loss: 0.45686166, Log Avg loss: 0.39678702, Global Avg Loss: 1.16518718, Time: 0.0076 Steps: 98870, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000946, Sample Num: 15136, Cur Loss: 0.33467993, Cur Avg Loss: 0.45644380, Log Avg loss: 0.41733141, Global Avg Loss: 1.16511155, Time: 0.0075 Steps: 98880, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000956, Sample Num: 15296, Cur Loss: 0.39506912, Cur Avg Loss: 0.45881035, Log Avg loss: 0.68268647, Global Avg Loss: 1.16506276, Time: 0.0075 Steps: 98890, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000966, Sample Num: 15456, Cur Loss: 0.16456771, Cur Avg Loss: 0.45791219, Log Avg loss: 0.37204835, Global Avg Loss: 1.16498258, Time: 0.0076 Steps: 98900, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000976, Sample Num: 15616, Cur Loss: 0.18036950, Cur Avg Loss: 0.45743146, Log Avg loss: 0.41099201, Global Avg Loss: 1.16490635, Time: 0.0075 Steps: 98910, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000986, Sample Num: 15776, Cur Loss: 0.44766176, Cur Avg Loss: 0.45688361, Log Avg loss: 0.40341392, Global Avg Loss: 1.16482937, Time: 0.0076 Steps: 98920, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 000996, Sample Num: 15936, Cur Loss: 0.20609868, Cur Avg Loss: 0.45613363, Log Avg loss: 0.38218583, Global Avg Loss: 1.16475026, Time: 0.0075 Steps: 98930, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001006, Sample Num: 16096, Cur Loss: 0.76042044, Cur Avg Loss: 0.45591171, Log Avg loss: 0.43380820, Global Avg Loss: 1.16467638, Time: 0.0075 Steps: 98940, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001016, Sample Num: 16256, Cur Loss: 0.12649006, Cur Avg Loss: 0.45497734, Log Avg loss: 0.36097941, Global Avg Loss: 1.16459516, Time: 0.0077 Steps: 98950, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001026, Sample Num: 16416, Cur Loss: 0.59395444, Cur Avg Loss: 0.45380097, Log Avg loss: 0.33428252, Global Avg Loss: 1.16451126, Time: 0.0078 Steps: 98960, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001036, Sample Num: 16576, Cur Loss: 0.29948765, Cur Avg Loss: 0.45410572, Log Avg loss: 0.48537240, Global Avg Loss: 1.16444263, Time: 0.0077 Steps: 98970, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001046, Sample Num: 16736, Cur Loss: 0.31915367, Cur Avg Loss: 0.45412361, Log Avg loss: 0.45597766, Global Avg Loss: 1.16437106, Time: 0.0077 Steps: 98980, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001056, Sample Num: 16896, Cur Loss: 0.24337348, Cur Avg Loss: 0.45475192, Log Avg loss: 0.52047241, Global Avg Loss: 1.16430601, Time: 0.0076 Steps: 98990, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001066, Sample Num: 17056, Cur Loss: 0.24496159, Cur Avg Loss: 0.45547404, Log Avg loss: 0.53172989, Global Avg Loss: 1.16424211, Time: 0.0075 Steps: 99000, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001076, Sample Num: 17216, Cur Loss: 0.71455145, Cur Avg Loss: 0.45592071, Log Avg loss: 0.50353624, Global Avg Loss: 1.16417538, Time: 0.0075 Steps: 99010, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001086, Sample Num: 17376, Cur Loss: 0.30366385, Cur Avg Loss: 0.45474724, Log Avg loss: 0.32848209, Global Avg Loss: 1.16409099, Time: 0.0076 Steps: 99020, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001096, Sample Num: 17536, Cur Loss: 0.31813711, Cur Avg Loss: 0.45421168, Log Avg loss: 0.39604949, Global Avg Loss: 1.16401343, Time: 0.0076 Steps: 99030, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001106, Sample Num: 17696, Cur Loss: 0.26021180, Cur Avg Loss: 0.45368793, Log Avg loss: 0.39628515, Global Avg Loss: 1.16393591, Time: 0.0076 Steps: 99040, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001116, Sample Num: 17856, Cur Loss: 0.47646502, Cur Avg Loss: 0.45309694, Log Avg loss: 0.38773362, Global Avg Loss: 1.16385755, Time: 0.0075 Steps: 99050, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001126, Sample Num: 18016, Cur Loss: 0.39686614, Cur Avg Loss: 0.45234746, Log Avg loss: 0.36870466, Global Avg Loss: 1.16377728, Time: 0.0076 Steps: 99060, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001136, Sample Num: 18176, Cur Loss: 0.35563642, Cur Avg Loss: 0.45227233, Log Avg loss: 0.44381361, Global Avg Loss: 1.16370461, Time: 0.0075 Steps: 99070, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001146, Sample Num: 18336, Cur Loss: 0.36484838, Cur Avg Loss: 0.45279235, Log Avg loss: 0.51186670, Global Avg Loss: 1.16363882, Time: 0.0075 Steps: 99080, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001156, Sample Num: 18496, Cur Loss: 0.71287936, Cur Avg Loss: 0.45303955, Log Avg loss: 0.48136772, Global Avg Loss: 1.16356996, Time: 0.0075 Steps: 99090, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001166, Sample Num: 18656, Cur Loss: 0.45798767, Cur Avg Loss: 0.45311861, Log Avg loss: 0.46225802, Global Avg Loss: 1.16349920, Time: 0.0075 Steps: 99100, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001176, Sample Num: 18816, Cur Loss: 0.57423246, Cur Avg Loss: 0.45366993, Log Avg loss: 0.51795395, Global Avg Loss: 1.16343406, Time: 0.0075 Steps: 99110, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001186, Sample Num: 18976, Cur Loss: 1.10402000, Cur Avg Loss: 0.45502301, Log Avg loss: 0.61414596, Global Avg Loss: 1.16337865, Time: 0.0075 Steps: 99120, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001196, Sample Num: 19136, Cur Loss: 0.19325465, Cur Avg Loss: 0.45507214, Log Avg loss: 0.46089809, Global Avg Loss: 1.16330778, Time: 0.0076 Steps: 99130, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001206, Sample Num: 19296, Cur Loss: 0.35049462, Cur Avg Loss: 0.45545496, Log Avg loss: 0.50124068, Global Avg Loss: 1.16324100, Time: 0.0075 Steps: 99140, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001216, Sample Num: 19456, Cur Loss: 0.41608649, Cur Avg Loss: 0.45547685, Log Avg loss: 0.45811705, Global Avg Loss: 1.16316988, Time: 0.0075 Steps: 99150, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001226, Sample Num: 19616, Cur Loss: 0.40445918, Cur Avg Loss: 0.45569708, Log Avg loss: 0.48247609, Global Avg Loss: 1.16310124, Time: 0.0077 Steps: 99160, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001236, Sample Num: 19776, Cur Loss: 0.13261408, Cur Avg Loss: 0.45643131, Log Avg loss: 0.54644813, Global Avg Loss: 1.16303906, Time: 0.0076 Steps: 99170, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001246, Sample Num: 19936, Cur Loss: 0.33314922, Cur Avg Loss: 0.45560057, Log Avg loss: 0.35292183, Global Avg Loss: 1.16295737, Time: 0.0075 Steps: 99180, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001256, Sample Num: 20096, Cur Loss: 0.26429778, Cur Avg Loss: 0.45596282, Log Avg loss: 0.50109894, Global Avg Loss: 1.16289065, Time: 0.0076 Steps: 99190, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001266, Sample Num: 20256, Cur Loss: 0.27282280, Cur Avg Loss: 0.45603360, Log Avg loss: 0.46492398, Global Avg Loss: 1.16282029, Time: 0.0076 Steps: 99200, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001276, Sample Num: 20416, Cur Loss: 0.21353194, Cur Avg Loss: 0.45557749, Log Avg loss: 0.39783313, Global Avg Loss: 1.16274318, Time: 0.0077 Steps: 99210, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001286, Sample Num: 20576, Cur Loss: 0.33030921, Cur Avg Loss: 0.45587384, Log Avg loss: 0.49368834, Global Avg Loss: 1.16267575, Time: 0.0075 Steps: 99220, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001296, Sample Num: 20736, Cur Loss: 0.42214698, Cur Avg Loss: 0.45639400, Log Avg loss: 0.52328621, Global Avg Loss: 1.16261131, Time: 0.0075 Steps: 99230, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001306, Sample Num: 20896, Cur Loss: 0.53015548, Cur Avg Loss: 0.45601967, Log Avg loss: 0.40750692, Global Avg Loss: 1.16253522, Time: 0.0076 Steps: 99240, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001316, Sample Num: 21056, Cur Loss: 0.26654118, Cur Avg Loss: 0.45539666, Log Avg loss: 0.37403177, Global Avg Loss: 1.16245578, Time: 0.0076 Steps: 99250, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001326, Sample Num: 21216, Cur Loss: 0.17142171, Cur Avg Loss: 0.45582542, Log Avg loss: 0.51225045, Global Avg Loss: 1.16239027, Time: 0.0075 Steps: 99260, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001336, Sample Num: 21376, Cur Loss: 0.44017041, Cur Avg Loss: 0.45818514, Log Avg loss: 0.77108399, Global Avg Loss: 1.16235085, Time: 0.0253 Steps: 99270, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001346, Sample Num: 21536, Cur Loss: 0.35864019, Cur Avg Loss: 0.45760572, Log Avg loss: 0.38019455, Global Avg Loss: 1.16227207, Time: 0.0071 Steps: 99280, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001356, Sample Num: 21696, Cur Loss: 0.28442335, Cur Avg Loss: 0.45745838, Log Avg loss: 0.43762656, Global Avg Loss: 1.16219909, Time: 0.0066 Steps: 99290, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001366, Sample Num: 21856, Cur Loss: 0.57208043, Cur Avg Loss: 0.45809403, Log Avg loss: 0.54428835, Global Avg Loss: 1.16213686, Time: 0.0066 Steps: 99300, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001376, Sample Num: 22016, Cur Loss: 0.23527479, Cur Avg Loss: 0.45770259, Log Avg loss: 0.40423235, Global Avg Loss: 1.16206055, Time: 0.0115 Steps: 99310, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001386, Sample Num: 22176, Cur Loss: 0.24460343, Cur Avg Loss: 0.45732406, Log Avg loss: 0.40523822, Global Avg Loss: 1.16198435, Time: 0.0078 Steps: 99320, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001396, Sample Num: 22336, Cur Loss: 0.10661312, Cur Avg Loss: 0.45704219, Log Avg loss: 0.41797469, Global Avg Loss: 1.16190944, Time: 0.0063 Steps: 99330, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001406, Sample Num: 22496, Cur Loss: 0.22286910, Cur Avg Loss: 0.45750524, Log Avg loss: 0.52214752, Global Avg Loss: 1.16184504, Time: 0.0192 Steps: 99340, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001416, Sample Num: 22656, Cur Loss: 0.20083550, Cur Avg Loss: 0.45674817, Log Avg loss: 0.35030366, Global Avg Loss: 1.16176336, Time: 0.0065 Steps: 99350, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001426, Sample Num: 22816, Cur Loss: 0.33435583, Cur Avg Loss: 0.45652589, Log Avg loss: 0.42505126, Global Avg Loss: 1.16168921, Time: 0.0064 Steps: 99360, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001436, Sample Num: 22976, Cur Loss: 0.49389228, Cur Avg Loss: 0.45576190, Log Avg loss: 0.34681643, Global Avg Loss: 1.16160721, Time: 0.0066 Steps: 99370, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001446, Sample Num: 23136, Cur Loss: 0.47540930, Cur Avg Loss: 0.45573567, Log Avg loss: 0.45196939, Global Avg Loss: 1.16153580, Time: 0.0063 Steps: 99380, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001456, Sample Num: 23296, Cur Loss: 0.34729558, Cur Avg Loss: 0.45621331, Log Avg loss: 0.52528051, Global Avg Loss: 1.16147178, Time: 0.0196 Steps: 99390, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001466, Sample Num: 23456, Cur Loss: 0.55635768, Cur Avg Loss: 0.45634182, Log Avg loss: 0.47505154, Global Avg Loss: 1.16140273, Time: 0.0073 Steps: 99400, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001476, Sample Num: 23616, Cur Loss: 0.18277708, Cur Avg Loss: 0.45565829, Log Avg loss: 0.35545390, Global Avg Loss: 1.16132165, Time: 0.0066 Steps: 99410, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001486, Sample Num: 23776, Cur Loss: 0.97137183, Cur Avg Loss: 0.45559193, Log Avg loss: 0.44579680, Global Avg Loss: 1.16124968, Time: 0.0066 Steps: 99420, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001496, Sample Num: 23936, Cur Loss: 0.34249046, Cur Avg Loss: 0.45585539, Log Avg loss: 0.49500521, Global Avg Loss: 1.16118268, Time: 0.0113 Steps: 99430, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001506, Sample Num: 24096, Cur Loss: 0.87031174, Cur Avg Loss: 0.45604309, Log Avg loss: 0.48412281, Global Avg Loss: 1.16111459, Time: 0.0066 Steps: 99440, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001516, Sample Num: 24256, Cur Loss: 0.36532095, Cur Avg Loss: 0.45588864, Log Avg loss: 0.43262947, Global Avg Loss: 1.16104134, Time: 0.0064 Steps: 99450, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001526, Sample Num: 24416, Cur Loss: 0.43879476, Cur Avg Loss: 0.45632396, Log Avg loss: 0.52231837, Global Avg Loss: 1.16097712, Time: 0.0067 Steps: 99460, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001536, Sample Num: 24576, Cur Loss: 0.45927754, Cur Avg Loss: 0.45659338, Log Avg loss: 0.49770626, Global Avg Loss: 1.16091044, Time: 0.0073 Steps: 99470, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001546, Sample Num: 24736, Cur Loss: 0.23856202, Cur Avg Loss: 0.45607892, Log Avg loss: 0.37705770, Global Avg Loss: 1.16083164, Time: 0.0066 Steps: 99480, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001556, Sample Num: 24896, Cur Loss: 0.21926555, Cur Avg Loss: 0.45631506, Log Avg loss: 0.49282198, Global Avg Loss: 1.16076450, Time: 0.0075 Steps: 99490, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001566, Sample Num: 25056, Cur Loss: 0.40920356, Cur Avg Loss: 0.45614109, Log Avg loss: 0.42907214, Global Avg Loss: 1.16069096, Time: 0.0075 Steps: 99500, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001576, Sample Num: 25216, Cur Loss: 0.42220032, Cur Avg Loss: 0.45580047, Log Avg loss: 0.40245915, Global Avg Loss: 1.16061477, Time: 0.0065 Steps: 99510, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001586, Sample Num: 25376, Cur Loss: 0.51200807, Cur Avg Loss: 0.45583137, Log Avg loss: 0.46070175, Global Avg Loss: 1.16054444, Time: 0.0067 Steps: 99520, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001596, Sample Num: 25536, Cur Loss: 0.15519641, Cur Avg Loss: 0.45684284, Log Avg loss: 0.61726225, Global Avg Loss: 1.16048985, Time: 0.0066 Steps: 99530, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001606, Sample Num: 25696, Cur Loss: 0.13581727, Cur Avg Loss: 0.45720471, Log Avg loss: 0.51495882, Global Avg Loss: 1.16042500, Time: 0.0066 Steps: 99540, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001616, Sample Num: 25856, Cur Loss: 0.15408882, Cur Avg Loss: 0.45740469, Log Avg loss: 0.48952095, Global Avg Loss: 1.16035761, Time: 0.0066 Steps: 99550, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001626, Sample Num: 26016, Cur Loss: 0.25788963, Cur Avg Loss: 0.45771800, Log Avg loss: 0.50834847, Global Avg Loss: 1.16029212, Time: 0.0067 Steps: 99560, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001636, Sample Num: 26176, Cur Loss: 0.58658433, Cur Avg Loss: 0.45756569, Log Avg loss: 0.43280099, Global Avg Loss: 1.16021906, Time: 0.0067 Steps: 99570, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001646, Sample Num: 26336, Cur Loss: 0.36710951, Cur Avg Loss: 0.45774176, Log Avg loss: 0.48654702, Global Avg Loss: 1.16015141, Time: 0.0075 Steps: 99580, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001656, Sample Num: 26496, Cur Loss: 0.26640797, Cur Avg Loss: 0.45742755, Log Avg loss: 0.40570802, Global Avg Loss: 1.16007565, Time: 0.0078 Steps: 99590, Updated lr: 0.000007 Training, Epoch: 0047, Batch: 001666, Sample Num: 26656, Cur Loss: 0.12261321, Cur Avg Loss: 0.45754913, Log Avg loss: 0.47768279, Global Avg Loss: 1.16000714, Time: 0.0076 Steps: 99600, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001676, Sample Num: 26816, Cur Loss: 0.22330572, Cur Avg Loss: 0.45721476, Log Avg loss: 0.40150929, Global Avg Loss: 1.15993099, Time: 0.0076 Steps: 99610, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001686, Sample Num: 26976, Cur Loss: 0.59250975, Cur Avg Loss: 0.45839705, Log Avg loss: 0.65654810, Global Avg Loss: 1.15988046, Time: 0.0077 Steps: 99620, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001696, Sample Num: 27136, Cur Loss: 0.26971930, Cur Avg Loss: 0.45937574, Log Avg loss: 0.62438354, Global Avg Loss: 1.15982671, Time: 0.0076 Steps: 99630, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001706, Sample Num: 27296, Cur Loss: 0.31204647, Cur Avg Loss: 0.45909228, Log Avg loss: 0.41101766, Global Avg Loss: 1.15975156, Time: 0.0075 Steps: 99640, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001716, Sample Num: 27456, Cur Loss: 0.28888914, Cur Avg Loss: 0.45862852, Log Avg loss: 0.37951066, Global Avg Loss: 1.15967326, Time: 0.0075 Steps: 99650, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001726, Sample Num: 27616, Cur Loss: 1.11884379, Cur Avg Loss: 0.45928256, Log Avg loss: 0.57151599, Global Avg Loss: 1.15961425, Time: 0.0076 Steps: 99660, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001736, Sample Num: 27776, Cur Loss: 0.99305815, Cur Avg Loss: 0.45956288, Log Avg loss: 0.50794591, Global Avg Loss: 1.15954886, Time: 0.0076 Steps: 99670, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001746, Sample Num: 27936, Cur Loss: 0.66255462, Cur Avg Loss: 0.45929013, Log Avg loss: 0.41193980, Global Avg Loss: 1.15947386, Time: 0.0078 Steps: 99680, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001756, Sample Num: 28096, Cur Loss: 0.47579855, Cur Avg Loss: 0.45954526, Log Avg loss: 0.50409229, Global Avg Loss: 1.15940812, Time: 0.0078 Steps: 99690, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001766, Sample Num: 28256, Cur Loss: 0.43359020, Cur Avg Loss: 0.45891727, Log Avg loss: 0.34864164, Global Avg Loss: 1.15932680, Time: 0.0078 Steps: 99700, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001776, Sample Num: 28416, Cur Loss: 0.70459116, Cur Avg Loss: 0.45891955, Log Avg loss: 0.45932195, Global Avg Loss: 1.15925660, Time: 0.0075 Steps: 99710, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001786, Sample Num: 28576, Cur Loss: 0.59106147, Cur Avg Loss: 0.45907944, Log Avg loss: 0.48747656, Global Avg Loss: 1.15918923, Time: 0.0075 Steps: 99720, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001796, Sample Num: 28736, Cur Loss: 0.64544505, Cur Avg Loss: 0.45919235, Log Avg loss: 0.47935701, Global Avg Loss: 1.15912106, Time: 0.0076 Steps: 99730, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001806, Sample Num: 28896, Cur Loss: 0.31002644, Cur Avg Loss: 0.45905981, Log Avg loss: 0.43525629, Global Avg Loss: 1.15904849, Time: 0.0076 Steps: 99740, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001816, Sample Num: 29056, Cur Loss: 0.17644304, Cur Avg Loss: 0.45878125, Log Avg loss: 0.40847321, Global Avg Loss: 1.15897324, Time: 0.0077 Steps: 99750, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001826, Sample Num: 29216, Cur Loss: 0.27229363, Cur Avg Loss: 0.45889523, Log Avg loss: 0.47959463, Global Avg Loss: 1.15890514, Time: 0.0076 Steps: 99760, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001836, Sample Num: 29376, Cur Loss: 0.40865320, Cur Avg Loss: 0.45906436, Log Avg loss: 0.48994733, Global Avg Loss: 1.15883809, Time: 0.0076 Steps: 99770, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001846, Sample Num: 29536, Cur Loss: 0.89652216, Cur Avg Loss: 0.45962735, Log Avg loss: 0.56299133, Global Avg Loss: 1.15877837, Time: 0.0076 Steps: 99780, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001856, Sample Num: 29696, Cur Loss: 0.52182269, Cur Avg Loss: 0.45948767, Log Avg loss: 0.43370316, Global Avg Loss: 1.15870571, Time: 0.0077 Steps: 99790, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001866, Sample Num: 29856, Cur Loss: 0.44805849, Cur Avg Loss: 0.45967564, Log Avg loss: 0.49456277, Global Avg Loss: 1.15863917, Time: 0.0076 Steps: 99800, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001876, Sample Num: 30016, Cur Loss: 0.28099185, Cur Avg Loss: 0.45967827, Log Avg loss: 0.46016910, Global Avg Loss: 1.15856919, Time: 0.0076 Steps: 99810, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001886, Sample Num: 30176, Cur Loss: 0.35941884, Cur Avg Loss: 0.45946599, Log Avg loss: 0.41964255, Global Avg Loss: 1.15849516, Time: 0.0077 Steps: 99820, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001896, Sample Num: 30336, Cur Loss: 0.47441387, Cur Avg Loss: 0.45922273, Log Avg loss: 0.41334456, Global Avg Loss: 1.15842052, Time: 0.0077 Steps: 99830, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001906, Sample Num: 30496, Cur Loss: 0.25213313, Cur Avg Loss: 0.45916555, Log Avg loss: 0.44832372, Global Avg Loss: 1.15834939, Time: 0.0076 Steps: 99840, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001916, Sample Num: 30656, Cur Loss: 0.36802414, Cur Avg Loss: 0.45854935, Log Avg loss: 0.34110179, Global Avg Loss: 1.15826755, Time: 0.0076 Steps: 99850, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001926, Sample Num: 30816, Cur Loss: 0.82379067, Cur Avg Loss: 0.45902196, Log Avg loss: 0.54957403, Global Avg Loss: 1.15820659, Time: 0.0076 Steps: 99860, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001936, Sample Num: 30976, Cur Loss: 0.96577245, Cur Avg Loss: 0.45885133, Log Avg loss: 0.42598774, Global Avg Loss: 1.15813328, Time: 0.0076 Steps: 99870, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001946, Sample Num: 31136, Cur Loss: 0.41635334, Cur Avg Loss: 0.45894320, Log Avg loss: 0.47672919, Global Avg Loss: 1.15806505, Time: 0.0076 Steps: 99880, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001956, Sample Num: 31296, Cur Loss: 0.14593586, Cur Avg Loss: 0.45844024, Log Avg loss: 0.36056470, Global Avg Loss: 1.15798522, Time: 0.0077 Steps: 99890, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001966, Sample Num: 31456, Cur Loss: 0.12981199, Cur Avg Loss: 0.45818646, Log Avg loss: 0.40854688, Global Avg Loss: 1.15791020, Time: 0.0076 Steps: 99900, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001976, Sample Num: 31616, Cur Loss: 0.49961326, Cur Avg Loss: 0.45914350, Log Avg loss: 0.64729685, Global Avg Loss: 1.15785909, Time: 0.0075 Steps: 99910, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001986, Sample Num: 31776, Cur Loss: 0.62586182, Cur Avg Loss: 0.45927415, Log Avg loss: 0.48509149, Global Avg Loss: 1.15779176, Time: 0.0076 Steps: 99920, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 001996, Sample Num: 31936, Cur Loss: 1.01811254, Cur Avg Loss: 0.45927527, Log Avg loss: 0.45949799, Global Avg Loss: 1.15772188, Time: 0.0077 Steps: 99930, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002006, Sample Num: 32096, Cur Loss: 0.30412525, Cur Avg Loss: 0.45899588, Log Avg loss: 0.40322827, Global Avg Loss: 1.15764639, Time: 0.0077 Steps: 99940, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002016, Sample Num: 32256, Cur Loss: 0.74618280, Cur Avg Loss: 0.45911083, Log Avg loss: 0.48217123, Global Avg Loss: 1.15757880, Time: 0.0075 Steps: 99950, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002026, Sample Num: 32416, Cur Loss: 0.20917571, Cur Avg Loss: 0.45918197, Log Avg loss: 0.47352338, Global Avg Loss: 1.15751037, Time: 0.0075 Steps: 99960, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002036, Sample Num: 32576, Cur Loss: 0.27470160, Cur Avg Loss: 0.45931320, Log Avg loss: 0.48589907, Global Avg Loss: 1.15744319, Time: 0.0076 Steps: 99970, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002046, Sample Num: 32736, Cur Loss: 0.65767580, Cur Avg Loss: 0.45971718, Log Avg loss: 0.54196753, Global Avg Loss: 1.15738163, Time: 0.0076 Steps: 99980, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002056, Sample Num: 32896, Cur Loss: 0.16490345, Cur Avg Loss: 0.45921058, Log Avg loss: 0.35556155, Global Avg Loss: 1.15730144, Time: 0.0127 Steps: 99990, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002066, Sample Num: 33056, Cur Loss: 0.28399700, Cur Avg Loss: 0.45881872, Log Avg loss: 0.37825210, Global Avg Loss: 1.15722353, Time: 0.0067 Steps: 100000, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002076, Sample Num: 33216, Cur Loss: 0.48218799, Cur Avg Loss: 0.45877328, Log Avg loss: 0.44938587, Global Avg Loss: 1.15715276, Time: 0.0064 Steps: 100010, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002086, Sample Num: 33376, Cur Loss: 0.19607472, Cur Avg Loss: 0.45939653, Log Avg loss: 0.58878187, Global Avg Loss: 1.15709593, Time: 0.0184 Steps: 100020, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002096, Sample Num: 33536, Cur Loss: 0.14644268, Cur Avg Loss: 0.45860548, Log Avg loss: 0.29359273, Global Avg Loss: 1.15700961, Time: 0.0067 Steps: 100030, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002106, Sample Num: 33696, Cur Loss: 0.29286867, Cur Avg Loss: 0.45835526, Log Avg loss: 0.40590847, Global Avg Loss: 1.15693453, Time: 0.0070 Steps: 100040, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002116, Sample Num: 33856, Cur Loss: 0.58420116, Cur Avg Loss: 0.45864833, Log Avg loss: 0.52036954, Global Avg Loss: 1.15687090, Time: 0.0066 Steps: 100050, Updated lr: 0.000006 Training, Epoch: 0047, Batch: 002126, Sample Num: 34016, Cur Loss: 0.56949961, Cur Avg Loss: 0.45839694, Log Avg loss: 0.40520316, Global Avg Loss: 1.15679578, Time: 0.0065 Steps: 100060, Updated lr: 0.000006 ***** Running evaluation checkpoint-100063 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-100063 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.790416, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.570406, "eval_total_loss": 400.99534, "eval_mae": 0.569685, "eval_mse": 0.570537, "eval_r2": 0.637329, "eval_sp_statistic": 0.754046, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.799972, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.916169, "test_total_loss": 459.916972, "test_mae": 0.660614, "test_mse": 0.916439, "test_r2": 0.408522, "test_sp_statistic": 0.604275, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.671164, "test_ps_pvalue": 0.0, "lr": 6.056899004267426e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.1567705574965252, "train_cur_epoch_loss": 975.4982815869153, "train_cur_epoch_avg_loss": 0.4581955291624778, "train_cur_epoch_time": 17.790416479110718, "train_cur_epoch_avg_time": 0.008356231319450784, "epoch": 47, "step": 100063} ################################################## Training, Epoch: 0048, Batch: 000007, Sample Num: 112, Cur Loss: 0.28452659, Cur Avg Loss: 0.42610679, Log Avg loss: 0.39291348, Global Avg Loss: 1.15671945, Time: 0.0074 Steps: 100070, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000017, Sample Num: 272, Cur Loss: 0.32325345, Cur Avg Loss: 0.37695336, Log Avg loss: 0.34254596, Global Avg Loss: 1.15663809, Time: 0.0074 Steps: 100080, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000027, Sample Num: 432, Cur Loss: 0.61432815, Cur Avg Loss: 0.37290019, Log Avg loss: 0.36600979, Global Avg Loss: 1.15655910, Time: 0.0074 Steps: 100090, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000037, Sample Num: 592, Cur Loss: 0.61806858, Cur Avg Loss: 0.36932189, Log Avg loss: 0.35966050, Global Avg Loss: 1.15647949, Time: 0.0074 Steps: 100100, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000047, Sample Num: 752, Cur Loss: 0.49071693, Cur Avg Loss: 0.37010733, Log Avg loss: 0.37301344, Global Avg Loss: 1.15640123, Time: 0.0075 Steps: 100110, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000057, Sample Num: 912, Cur Loss: 0.35026705, Cur Avg Loss: 0.37308232, Log Avg loss: 0.38706475, Global Avg Loss: 1.15632439, Time: 0.0074 Steps: 100120, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000067, Sample Num: 1072, Cur Loss: 0.15557259, Cur Avg Loss: 0.38814100, Log Avg loss: 0.47397548, Global Avg Loss: 1.15625624, Time: 0.0073 Steps: 100130, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000077, Sample Num: 1232, Cur Loss: 0.19759737, Cur Avg Loss: 0.39323816, Log Avg loss: 0.42738919, Global Avg Loss: 1.15618346, Time: 0.0074 Steps: 100140, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000087, Sample Num: 1392, Cur Loss: 0.27538323, Cur Avg Loss: 0.39318396, Log Avg loss: 0.39276657, Global Avg Loss: 1.15610723, Time: 0.0074 Steps: 100150, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000097, Sample Num: 1552, Cur Loss: 1.25105822, Cur Avg Loss: 0.41328932, Log Avg loss: 0.58820601, Global Avg Loss: 1.15605053, Time: 0.0074 Steps: 100160, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000107, Sample Num: 1712, Cur Loss: 0.30369943, Cur Avg Loss: 0.42443926, Log Avg loss: 0.53259363, Global Avg Loss: 1.15598829, Time: 0.0074 Steps: 100170, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000117, Sample Num: 1872, Cur Loss: 0.51367188, Cur Avg Loss: 0.41616305, Log Avg loss: 0.32760764, Global Avg Loss: 1.15590560, Time: 0.0074 Steps: 100180, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000127, Sample Num: 2032, Cur Loss: 0.23296444, Cur Avg Loss: 0.41417088, Log Avg loss: 0.39086244, Global Avg Loss: 1.15582924, Time: 0.0074 Steps: 100190, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000137, Sample Num: 2192, Cur Loss: 0.35814798, Cur Avg Loss: 0.41523583, Log Avg loss: 0.42876071, Global Avg Loss: 1.15575668, Time: 0.0074 Steps: 100200, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000147, Sample Num: 2352, Cur Loss: 0.46582761, Cur Avg Loss: 0.41852612, Log Avg loss: 0.46360315, Global Avg Loss: 1.15568761, Time: 0.0074 Steps: 100210, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000157, Sample Num: 2512, Cur Loss: 0.26689309, Cur Avg Loss: 0.42286568, Log Avg loss: 0.48665716, Global Avg Loss: 1.15562086, Time: 0.0074 Steps: 100220, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000167, Sample Num: 2672, Cur Loss: 0.68664336, Cur Avg Loss: 0.41782863, Log Avg loss: 0.33874694, Global Avg Loss: 1.15553936, Time: 0.0081 Steps: 100230, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000177, Sample Num: 2832, Cur Loss: 0.67716193, Cur Avg Loss: 0.42247787, Log Avg loss: 0.50012015, Global Avg Loss: 1.15547397, Time: 0.0203 Steps: 100240, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000187, Sample Num: 2992, Cur Loss: 0.21280310, Cur Avg Loss: 0.41657839, Log Avg loss: 0.31215753, Global Avg Loss: 1.15538985, Time: 0.0215 Steps: 100250, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000197, Sample Num: 3152, Cur Loss: 0.57983881, Cur Avg Loss: 0.41946840, Log Avg loss: 0.47351159, Global Avg Loss: 1.15532184, Time: 0.0205 Steps: 100260, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000207, Sample Num: 3312, Cur Loss: 1.74361014, Cur Avg Loss: 0.41987391, Log Avg loss: 0.42786245, Global Avg Loss: 1.15524929, Time: 0.0135 Steps: 100270, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000217, Sample Num: 3472, Cur Loss: 0.48128542, Cur Avg Loss: 0.42371949, Log Avg loss: 0.50332297, Global Avg Loss: 1.15518428, Time: 0.0236 Steps: 100280, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000227, Sample Num: 3632, Cur Loss: 1.11343956, Cur Avg Loss: 0.43128704, Log Avg loss: 0.59550304, Global Avg Loss: 1.15512847, Time: 0.0170 Steps: 100290, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000237, Sample Num: 3792, Cur Loss: 0.60354155, Cur Avg Loss: 0.42921993, Log Avg loss: 0.38229656, Global Avg Loss: 1.15505142, Time: 0.0072 Steps: 100300, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000247, Sample Num: 3952, Cur Loss: 0.28447339, Cur Avg Loss: 0.43112515, Log Avg loss: 0.47627890, Global Avg Loss: 1.15498375, Time: 0.0064 Steps: 100310, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000257, Sample Num: 4112, Cur Loss: 0.29960147, Cur Avg Loss: 0.43281605, Log Avg loss: 0.47458106, Global Avg Loss: 1.15491593, Time: 0.0192 Steps: 100320, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000267, Sample Num: 4272, Cur Loss: 0.14161925, Cur Avg Loss: 0.43143180, Log Avg loss: 0.39585656, Global Avg Loss: 1.15484027, Time: 0.0179 Steps: 100330, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000277, Sample Num: 4432, Cur Loss: 0.13123915, Cur Avg Loss: 0.43320612, Log Avg loss: 0.48058070, Global Avg Loss: 1.15477308, Time: 0.0116 Steps: 100340, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000287, Sample Num: 4592, Cur Loss: 0.56337392, Cur Avg Loss: 0.43562747, Log Avg loss: 0.50269885, Global Avg Loss: 1.15470810, Time: 0.0067 Steps: 100350, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000297, Sample Num: 4752, Cur Loss: 0.76965916, Cur Avg Loss: 0.43566730, Log Avg loss: 0.43681018, Global Avg Loss: 1.15463656, Time: 0.0125 Steps: 100360, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000307, Sample Num: 4912, Cur Loss: 1.05458045, Cur Avg Loss: 0.43687028, Log Avg loss: 0.47259896, Global Avg Loss: 1.15456861, Time: 0.0074 Steps: 100370, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000317, Sample Num: 5072, Cur Loss: 0.20398173, Cur Avg Loss: 0.43816512, Log Avg loss: 0.47791653, Global Avg Loss: 1.15450120, Time: 0.0242 Steps: 100380, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000327, Sample Num: 5232, Cur Loss: 0.50308818, Cur Avg Loss: 0.43597157, Log Avg loss: 0.36643605, Global Avg Loss: 1.15442270, Time: 0.0066 Steps: 100390, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000337, Sample Num: 5392, Cur Loss: 0.38410819, Cur Avg Loss: 0.43577543, Log Avg loss: 0.42936188, Global Avg Loss: 1.15435048, Time: 0.0229 Steps: 100400, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000347, Sample Num: 5552, Cur Loss: 0.57276231, Cur Avg Loss: 0.43985492, Log Avg loss: 0.57733362, Global Avg Loss: 1.15429302, Time: 0.0239 Steps: 100410, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000357, Sample Num: 5712, Cur Loss: 0.26202208, Cur Avg Loss: 0.43998487, Log Avg loss: 0.44449409, Global Avg Loss: 1.15422234, Time: 0.0067 Steps: 100420, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000367, Sample Num: 5872, Cur Loss: 0.15184560, Cur Avg Loss: 0.43632687, Log Avg loss: 0.30573619, Global Avg Loss: 1.15413785, Time: 0.0116 Steps: 100430, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000377, Sample Num: 6032, Cur Loss: 0.58074880, Cur Avg Loss: 0.43605162, Log Avg loss: 0.42595002, Global Avg Loss: 1.15406535, Time: 0.0064 Steps: 100440, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000387, Sample Num: 6192, Cur Loss: 0.22896282, Cur Avg Loss: 0.43264953, Log Avg loss: 0.30439068, Global Avg Loss: 1.15398076, Time: 0.0070 Steps: 100450, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000397, Sample Num: 6352, Cur Loss: 0.80266362, Cur Avg Loss: 0.43255851, Log Avg loss: 0.42903621, Global Avg Loss: 1.15390860, Time: 0.0065 Steps: 100460, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000407, Sample Num: 6512, Cur Loss: 0.22433062, Cur Avg Loss: 0.43021695, Log Avg loss: 0.33725714, Global Avg Loss: 1.15382732, Time: 0.0219 Steps: 100470, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000417, Sample Num: 6672, Cur Loss: 0.48241138, Cur Avg Loss: 0.43067902, Log Avg loss: 0.44948490, Global Avg Loss: 1.15375722, Time: 0.0071 Steps: 100480, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000427, Sample Num: 6832, Cur Loss: 0.34620848, Cur Avg Loss: 0.43092106, Log Avg loss: 0.44101427, Global Avg Loss: 1.15368629, Time: 0.0211 Steps: 100490, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000437, Sample Num: 6992, Cur Loss: 0.38097024, Cur Avg Loss: 0.43143907, Log Avg loss: 0.45355797, Global Avg Loss: 1.15361663, Time: 0.0067 Steps: 100500, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000447, Sample Num: 7152, Cur Loss: 0.40003979, Cur Avg Loss: 0.43323301, Log Avg loss: 0.51162831, Global Avg Loss: 1.15355276, Time: 0.0067 Steps: 100510, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000457, Sample Num: 7312, Cur Loss: 0.90471196, Cur Avg Loss: 0.43346077, Log Avg loss: 0.44364178, Global Avg Loss: 1.15348213, Time: 0.0066 Steps: 100520, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000467, Sample Num: 7472, Cur Loss: 0.42764175, Cur Avg Loss: 0.43377344, Log Avg loss: 0.44806253, Global Avg Loss: 1.15341196, Time: 0.0067 Steps: 100530, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000477, Sample Num: 7632, Cur Loss: 0.25655505, Cur Avg Loss: 0.43628655, Log Avg loss: 0.55364881, Global Avg Loss: 1.15335231, Time: 0.0067 Steps: 100540, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000487, Sample Num: 7792, Cur Loss: 0.11917681, Cur Avg Loss: 0.43775908, Log Avg loss: 0.50799865, Global Avg Loss: 1.15328812, Time: 0.0066 Steps: 100550, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000497, Sample Num: 7952, Cur Loss: 0.97612995, Cur Avg Loss: 0.44042098, Log Avg loss: 0.57005564, Global Avg Loss: 1.15323013, Time: 0.0076 Steps: 100560, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000507, Sample Num: 8112, Cur Loss: 0.30421472, Cur Avg Loss: 0.44193794, Log Avg loss: 0.51733049, Global Avg Loss: 1.15316690, Time: 0.0076 Steps: 100570, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000517, Sample Num: 8272, Cur Loss: 0.27685294, Cur Avg Loss: 0.44236059, Log Avg loss: 0.46378917, Global Avg Loss: 1.15309836, Time: 0.0067 Steps: 100580, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000527, Sample Num: 8432, Cur Loss: 0.17878644, Cur Avg Loss: 0.44321709, Log Avg loss: 0.48749806, Global Avg Loss: 1.15303219, Time: 0.0067 Steps: 100590, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000537, Sample Num: 8592, Cur Loss: 0.54935646, Cur Avg Loss: 0.44394219, Log Avg loss: 0.48215517, Global Avg Loss: 1.15296550, Time: 0.0067 Steps: 100600, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000547, Sample Num: 8752, Cur Loss: 0.39977273, Cur Avg Loss: 0.44360464, Log Avg loss: 0.42547788, Global Avg Loss: 1.15289319, Time: 0.0067 Steps: 100610, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000557, Sample Num: 8912, Cur Loss: 0.33034563, Cur Avg Loss: 0.44154040, Log Avg loss: 0.32862657, Global Avg Loss: 1.15281127, Time: 0.0068 Steps: 100620, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000567, Sample Num: 9072, Cur Loss: 0.35822940, Cur Avg Loss: 0.44016008, Log Avg loss: 0.36327614, Global Avg Loss: 1.15273281, Time: 0.0075 Steps: 100630, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000577, Sample Num: 9232, Cur Loss: 0.74448824, Cur Avg Loss: 0.43916132, Log Avg loss: 0.38253169, Global Avg Loss: 1.15265628, Time: 0.0077 Steps: 100640, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000587, Sample Num: 9392, Cur Loss: 0.69639736, Cur Avg Loss: 0.43820774, Log Avg loss: 0.38318639, Global Avg Loss: 1.15257983, Time: 0.0076 Steps: 100650, Updated lr: 0.000006 Training, Epoch: 0048, Batch: 000597, Sample Num: 9552, Cur Loss: 0.36581954, Cur Avg Loss: 0.43650485, Log Avg loss: 0.33654527, Global Avg Loss: 1.15249877, Time: 0.0075 Steps: 100660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000607, Sample Num: 9712, Cur Loss: 0.24915342, Cur Avg Loss: 0.43597466, Log Avg loss: 0.40432211, Global Avg Loss: 1.15242445, Time: 0.0076 Steps: 100670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000617, Sample Num: 9872, Cur Loss: 0.38361627, Cur Avg Loss: 0.43592652, Log Avg loss: 0.43300419, Global Avg Loss: 1.15235299, Time: 0.0076 Steps: 100680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000627, Sample Num: 10032, Cur Loss: 0.25609112, Cur Avg Loss: 0.43609160, Log Avg loss: 0.44627726, Global Avg Loss: 1.15228287, Time: 0.0075 Steps: 100690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000637, Sample Num: 10192, Cur Loss: 0.49994475, Cur Avg Loss: 0.43608370, Log Avg loss: 0.43558830, Global Avg Loss: 1.15221169, Time: 0.0077 Steps: 100700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000647, Sample Num: 10352, Cur Loss: 0.19212551, Cur Avg Loss: 0.43584433, Log Avg loss: 0.42059642, Global Avg Loss: 1.15213905, Time: 0.0076 Steps: 100710, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000657, Sample Num: 10512, Cur Loss: 1.35089970, Cur Avg Loss: 0.43700841, Log Avg loss: 0.51232482, Global Avg Loss: 1.15207552, Time: 0.0077 Steps: 100720, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000667, Sample Num: 10672, Cur Loss: 0.15113807, Cur Avg Loss: 0.43645634, Log Avg loss: 0.40018476, Global Avg Loss: 1.15200088, Time: 0.0076 Steps: 100730, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000677, Sample Num: 10832, Cur Loss: 0.49239150, Cur Avg Loss: 0.43584024, Log Avg loss: 0.39474691, Global Avg Loss: 1.15192571, Time: 0.0076 Steps: 100740, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000687, Sample Num: 10992, Cur Loss: 0.24332774, Cur Avg Loss: 0.43484122, Log Avg loss: 0.36720765, Global Avg Loss: 1.15184782, Time: 0.0076 Steps: 100750, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000697, Sample Num: 11152, Cur Loss: 0.24317004, Cur Avg Loss: 0.43676963, Log Avg loss: 0.56925101, Global Avg Loss: 1.15179000, Time: 0.0075 Steps: 100760, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000707, Sample Num: 11312, Cur Loss: 0.79579359, Cur Avg Loss: 0.43777364, Log Avg loss: 0.50775340, Global Avg Loss: 1.15172609, Time: 0.0077 Steps: 100770, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000717, Sample Num: 11472, Cur Loss: 0.65137863, Cur Avg Loss: 0.43693486, Log Avg loss: 0.37763308, Global Avg Loss: 1.15164928, Time: 0.0077 Steps: 100780, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000727, Sample Num: 11632, Cur Loss: 0.49880475, Cur Avg Loss: 0.44017308, Log Avg loss: 0.67235360, Global Avg Loss: 1.15160173, Time: 0.0076 Steps: 100790, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000737, Sample Num: 11792, Cur Loss: 0.71203536, Cur Avg Loss: 0.44214337, Log Avg loss: 0.58538322, Global Avg Loss: 1.15154556, Time: 0.0077 Steps: 100800, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000747, Sample Num: 11952, Cur Loss: 0.46519625, Cur Avg Loss: 0.44315866, Log Avg loss: 0.51798518, Global Avg Loss: 1.15148271, Time: 0.0077 Steps: 100810, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000757, Sample Num: 12112, Cur Loss: 0.41064805, Cur Avg Loss: 0.44349442, Log Avg loss: 0.46857592, Global Avg Loss: 1.15141497, Time: 0.0076 Steps: 100820, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000767, Sample Num: 12272, Cur Loss: 0.20860007, Cur Avg Loss: 0.44167685, Log Avg loss: 0.30408710, Global Avg Loss: 1.15133094, Time: 0.0077 Steps: 100830, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000777, Sample Num: 12432, Cur Loss: 0.17326885, Cur Avg Loss: 0.44393434, Log Avg loss: 0.61708353, Global Avg Loss: 1.15127796, Time: 0.0075 Steps: 100840, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000787, Sample Num: 12592, Cur Loss: 0.65857512, Cur Avg Loss: 0.44452293, Log Avg loss: 0.49025603, Global Avg Loss: 1.15121241, Time: 0.0076 Steps: 100850, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000797, Sample Num: 12752, Cur Loss: 0.62192965, Cur Avg Loss: 0.44459442, Log Avg loss: 0.45022100, Global Avg Loss: 1.15114291, Time: 0.0077 Steps: 100860, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000807, Sample Num: 12912, Cur Loss: 0.79757357, Cur Avg Loss: 0.44557239, Log Avg loss: 0.52351627, Global Avg Loss: 1.15108069, Time: 0.0077 Steps: 100870, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000817, Sample Num: 13072, Cur Loss: 0.50515378, Cur Avg Loss: 0.44681826, Log Avg loss: 0.54736058, Global Avg Loss: 1.15102084, Time: 0.0076 Steps: 100880, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000827, Sample Num: 13232, Cur Loss: 0.52952886, Cur Avg Loss: 0.44596345, Log Avg loss: 0.37612509, Global Avg Loss: 1.15094404, Time: 0.0076 Steps: 100890, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000837, Sample Num: 13392, Cur Loss: 0.36851490, Cur Avg Loss: 0.44621559, Log Avg loss: 0.46706801, Global Avg Loss: 1.15087626, Time: 0.0075 Steps: 100900, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000847, Sample Num: 13552, Cur Loss: 0.65665793, Cur Avg Loss: 0.44591996, Log Avg loss: 0.42117507, Global Avg Loss: 1.15080395, Time: 0.0075 Steps: 100910, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000857, Sample Num: 13712, Cur Loss: 1.11693871, Cur Avg Loss: 0.44654477, Log Avg loss: 0.49946636, Global Avg Loss: 1.15073941, Time: 0.0076 Steps: 100920, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000867, Sample Num: 13872, Cur Loss: 0.22221625, Cur Avg Loss: 0.44653609, Log Avg loss: 0.44579269, Global Avg Loss: 1.15066956, Time: 0.0077 Steps: 100930, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000877, Sample Num: 14032, Cur Loss: 0.50745034, Cur Avg Loss: 0.44711131, Log Avg loss: 0.49698280, Global Avg Loss: 1.15060480, Time: 0.0075 Steps: 100940, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000887, Sample Num: 14192, Cur Loss: 0.64290816, Cur Avg Loss: 0.44702909, Log Avg loss: 0.43981816, Global Avg Loss: 1.15053439, Time: 0.0075 Steps: 100950, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000897, Sample Num: 14352, Cur Loss: 0.41945997, Cur Avg Loss: 0.44798190, Log Avg loss: 0.53249659, Global Avg Loss: 1.15047318, Time: 0.0076 Steps: 100960, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000907, Sample Num: 14512, Cur Loss: 0.58398712, Cur Avg Loss: 0.44806619, Log Avg loss: 0.45562686, Global Avg Loss: 1.15040436, Time: 0.0076 Steps: 100970, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000917, Sample Num: 14672, Cur Loss: 0.25502339, Cur Avg Loss: 0.44738950, Log Avg loss: 0.38601301, Global Avg Loss: 1.15032866, Time: 0.0077 Steps: 100980, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000927, Sample Num: 14832, Cur Loss: 0.29031312, Cur Avg Loss: 0.44700762, Log Avg loss: 0.41198999, Global Avg Loss: 1.15025555, Time: 0.0075 Steps: 100990, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000937, Sample Num: 14992, Cur Loss: 0.46557635, Cur Avg Loss: 0.44666123, Log Avg loss: 0.41455086, Global Avg Loss: 1.15018271, Time: 0.0075 Steps: 101000, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000947, Sample Num: 15152, Cur Loss: 0.19228917, Cur Avg Loss: 0.44656018, Log Avg loss: 0.43709158, Global Avg Loss: 1.15011212, Time: 0.0077 Steps: 101010, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000957, Sample Num: 15312, Cur Loss: 0.43758002, Cur Avg Loss: 0.44690374, Log Avg loss: 0.47943865, Global Avg Loss: 1.15004573, Time: 0.0076 Steps: 101020, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000967, Sample Num: 15472, Cur Loss: 0.10131769, Cur Avg Loss: 0.44704520, Log Avg loss: 0.46058267, Global Avg Loss: 1.14997748, Time: 0.0076 Steps: 101030, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000977, Sample Num: 15632, Cur Loss: 0.43896016, Cur Avg Loss: 0.44912954, Log Avg loss: 0.65068588, Global Avg Loss: 1.14992807, Time: 0.0076 Steps: 101040, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000987, Sample Num: 15792, Cur Loss: 0.55270207, Cur Avg Loss: 0.44936276, Log Avg loss: 0.47214777, Global Avg Loss: 1.14986099, Time: 0.0075 Steps: 101050, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 000997, Sample Num: 15952, Cur Loss: 0.38203216, Cur Avg Loss: 0.44962997, Log Avg loss: 0.47600432, Global Avg Loss: 1.14979431, Time: 0.0076 Steps: 101060, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001007, Sample Num: 16112, Cur Loss: 0.40955323, Cur Avg Loss: 0.45034168, Log Avg loss: 0.52129878, Global Avg Loss: 1.14973213, Time: 0.0075 Steps: 101070, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001017, Sample Num: 16272, Cur Loss: 0.37189209, Cur Avg Loss: 0.44945968, Log Avg loss: 0.36064183, Global Avg Loss: 1.14965406, Time: 0.0076 Steps: 101080, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001027, Sample Num: 16432, Cur Loss: 0.36639234, Cur Avg Loss: 0.45001613, Log Avg loss: 0.50660707, Global Avg Loss: 1.14959045, Time: 0.0076 Steps: 101090, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001037, Sample Num: 16592, Cur Loss: 0.91719919, Cur Avg Loss: 0.45053285, Log Avg loss: 0.50360025, Global Avg Loss: 1.14952656, Time: 0.0076 Steps: 101100, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001047, Sample Num: 16752, Cur Loss: 0.16530828, Cur Avg Loss: 0.45012278, Log Avg loss: 0.40759885, Global Avg Loss: 1.14945318, Time: 0.0076 Steps: 101110, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001057, Sample Num: 16912, Cur Loss: 0.66265047, Cur Avg Loss: 0.45052499, Log Avg loss: 0.49263636, Global Avg Loss: 1.14938822, Time: 0.0076 Steps: 101120, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001067, Sample Num: 17072, Cur Loss: 0.24920796, Cur Avg Loss: 0.45068988, Log Avg loss: 0.46811883, Global Avg Loss: 1.14932086, Time: 0.0076 Steps: 101130, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001077, Sample Num: 17232, Cur Loss: 0.52956593, Cur Avg Loss: 0.44976613, Log Avg loss: 0.35120182, Global Avg Loss: 1.14924195, Time: 0.0076 Steps: 101140, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001087, Sample Num: 17392, Cur Loss: 0.24964538, Cur Avg Loss: 0.45017206, Log Avg loss: 0.49389056, Global Avg Loss: 1.14917716, Time: 0.0075 Steps: 101150, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001097, Sample Num: 17552, Cur Loss: 0.49935311, Cur Avg Loss: 0.45052543, Log Avg loss: 0.48893736, Global Avg Loss: 1.14911189, Time: 0.0075 Steps: 101160, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001107, Sample Num: 17712, Cur Loss: 0.40337098, Cur Avg Loss: 0.44992622, Log Avg loss: 0.38419235, Global Avg Loss: 1.14903628, Time: 0.0076 Steps: 101170, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001117, Sample Num: 17872, Cur Loss: 0.17751703, Cur Avg Loss: 0.45115145, Log Avg loss: 0.58678419, Global Avg Loss: 1.14898071, Time: 0.0076 Steps: 101180, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001127, Sample Num: 18032, Cur Loss: 0.36296919, Cur Avg Loss: 0.45122060, Log Avg loss: 0.45894532, Global Avg Loss: 1.14891252, Time: 0.0076 Steps: 101190, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001137, Sample Num: 18192, Cur Loss: 0.42163652, Cur Avg Loss: 0.45088972, Log Avg loss: 0.41359962, Global Avg Loss: 1.14883986, Time: 0.0075 Steps: 101200, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001147, Sample Num: 18352, Cur Loss: 0.37444672, Cur Avg Loss: 0.45122353, Log Avg loss: 0.48917782, Global Avg Loss: 1.14877468, Time: 0.0075 Steps: 101210, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001157, Sample Num: 18512, Cur Loss: 0.57258648, Cur Avg Loss: 0.45078727, Log Avg loss: 0.40074799, Global Avg Loss: 1.14870078, Time: 0.0075 Steps: 101220, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001167, Sample Num: 18672, Cur Loss: 0.35519537, Cur Avg Loss: 0.45042887, Log Avg loss: 0.40896186, Global Avg Loss: 1.14862771, Time: 0.0076 Steps: 101230, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001177, Sample Num: 18832, Cur Loss: 0.25770402, Cur Avg Loss: 0.45132316, Log Avg loss: 0.55568607, Global Avg Loss: 1.14856914, Time: 0.0076 Steps: 101240, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001187, Sample Num: 18992, Cur Loss: 0.66698682, Cur Avg Loss: 0.45144039, Log Avg loss: 0.46523837, Global Avg Loss: 1.14850165, Time: 0.0264 Steps: 101250, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001197, Sample Num: 19152, Cur Loss: 0.52124995, Cur Avg Loss: 0.45117275, Log Avg loss: 0.41940424, Global Avg Loss: 1.14842965, Time: 0.0065 Steps: 101260, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001207, Sample Num: 19312, Cur Loss: 0.82849646, Cur Avg Loss: 0.45058981, Log Avg loss: 0.38081220, Global Avg Loss: 1.14835385, Time: 0.0070 Steps: 101270, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001217, Sample Num: 19472, Cur Loss: 0.23975337, Cur Avg Loss: 0.45194190, Log Avg loss: 0.61513958, Global Avg Loss: 1.14830120, Time: 0.0114 Steps: 101280, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001227, Sample Num: 19632, Cur Loss: 0.43198270, Cur Avg Loss: 0.45165447, Log Avg loss: 0.41667343, Global Avg Loss: 1.14822897, Time: 0.0072 Steps: 101290, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001237, Sample Num: 19792, Cur Loss: 0.38529998, Cur Avg Loss: 0.45132229, Log Avg loss: 0.41056426, Global Avg Loss: 1.14815615, Time: 0.0064 Steps: 101300, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001247, Sample Num: 19952, Cur Loss: 0.84724677, Cur Avg Loss: 0.45200828, Log Avg loss: 0.53686542, Global Avg Loss: 1.14809581, Time: 0.0188 Steps: 101310, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001257, Sample Num: 20112, Cur Loss: 0.76490468, Cur Avg Loss: 0.45241993, Log Avg loss: 0.50375237, Global Avg Loss: 1.14803222, Time: 0.0064 Steps: 101320, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001267, Sample Num: 20272, Cur Loss: 0.36661729, Cur Avg Loss: 0.45257569, Log Avg loss: 0.47215417, Global Avg Loss: 1.14796552, Time: 0.0192 Steps: 101330, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001277, Sample Num: 20432, Cur Loss: 0.63997537, Cur Avg Loss: 0.45180564, Log Avg loss: 0.35424138, Global Avg Loss: 1.14788719, Time: 0.0064 Steps: 101340, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001287, Sample Num: 20592, Cur Loss: 0.96595275, Cur Avg Loss: 0.45190984, Log Avg loss: 0.46521601, Global Avg Loss: 1.14781983, Time: 0.0065 Steps: 101350, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001297, Sample Num: 20752, Cur Loss: 0.46443105, Cur Avg Loss: 0.45162629, Log Avg loss: 0.41513277, Global Avg Loss: 1.14774755, Time: 0.0171 Steps: 101360, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001307, Sample Num: 20912, Cur Loss: 0.25585291, Cur Avg Loss: 0.45152500, Log Avg loss: 0.43838747, Global Avg Loss: 1.14767757, Time: 0.0109 Steps: 101370, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001317, Sample Num: 21072, Cur Loss: 1.17313135, Cur Avg Loss: 0.45231960, Log Avg loss: 0.55617418, Global Avg Loss: 1.14761923, Time: 0.0064 Steps: 101380, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001327, Sample Num: 21232, Cur Loss: 0.64935124, Cur Avg Loss: 0.45305798, Log Avg loss: 0.55030253, Global Avg Loss: 1.14756031, Time: 0.0066 Steps: 101390, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001337, Sample Num: 21392, Cur Loss: 0.36833996, Cur Avg Loss: 0.45240302, Log Avg loss: 0.36549068, Global Avg Loss: 1.14748319, Time: 0.0078 Steps: 101400, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001347, Sample Num: 21552, Cur Loss: 0.57039046, Cur Avg Loss: 0.45253737, Log Avg loss: 0.47049889, Global Avg Loss: 1.14741643, Time: 0.0234 Steps: 101410, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001357, Sample Num: 21712, Cur Loss: 0.12834185, Cur Avg Loss: 0.45236577, Log Avg loss: 0.42925193, Global Avg Loss: 1.14734562, Time: 0.0065 Steps: 101420, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001367, Sample Num: 21872, Cur Loss: 0.17531177, Cur Avg Loss: 0.45230681, Log Avg loss: 0.44430614, Global Avg Loss: 1.14727631, Time: 0.0067 Steps: 101430, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001377, Sample Num: 22032, Cur Loss: 0.63841635, Cur Avg Loss: 0.45188802, Log Avg loss: 0.39463951, Global Avg Loss: 1.14720211, Time: 0.0171 Steps: 101440, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001387, Sample Num: 22192, Cur Loss: 0.37306628, Cur Avg Loss: 0.45109605, Log Avg loss: 0.34204125, Global Avg Loss: 1.14712275, Time: 0.0065 Steps: 101450, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001397, Sample Num: 22352, Cur Loss: 0.25940511, Cur Avg Loss: 0.45101379, Log Avg loss: 0.43960398, Global Avg Loss: 1.14705301, Time: 0.0205 Steps: 101460, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001407, Sample Num: 22512, Cur Loss: 0.60095346, Cur Avg Loss: 0.45097403, Log Avg loss: 0.44541920, Global Avg Loss: 1.14698386, Time: 0.0067 Steps: 101470, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001417, Sample Num: 22672, Cur Loss: 1.06012940, Cur Avg Loss: 0.45277581, Log Avg loss: 0.70628678, Global Avg Loss: 1.14694044, Time: 0.0073 Steps: 101480, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001427, Sample Num: 22832, Cur Loss: 0.14600784, Cur Avg Loss: 0.45329277, Log Avg loss: 0.52654581, Global Avg Loss: 1.14687931, Time: 0.0066 Steps: 101490, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001437, Sample Num: 22992, Cur Loss: 0.99379086, Cur Avg Loss: 0.45400201, Log Avg loss: 0.55521048, Global Avg Loss: 1.14682102, Time: 0.0065 Steps: 101500, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001447, Sample Num: 23152, Cur Loss: 0.58289242, Cur Avg Loss: 0.45322231, Log Avg loss: 0.34118017, Global Avg Loss: 1.14674165, Time: 0.0067 Steps: 101510, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001457, Sample Num: 23312, Cur Loss: 0.37974131, Cur Avg Loss: 0.45419873, Log Avg loss: 0.59548571, Global Avg Loss: 1.14668735, Time: 0.0067 Steps: 101520, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001467, Sample Num: 23472, Cur Loss: 0.58470774, Cur Avg Loss: 0.45427980, Log Avg loss: 0.46609275, Global Avg Loss: 1.14662032, Time: 0.0066 Steps: 101530, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001477, Sample Num: 23632, Cur Loss: 1.05563784, Cur Avg Loss: 0.45382176, Log Avg loss: 0.38662746, Global Avg Loss: 1.14654547, Time: 0.0066 Steps: 101540, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001487, Sample Num: 23792, Cur Loss: 0.27340072, Cur Avg Loss: 0.45437868, Log Avg loss: 0.53663567, Global Avg Loss: 1.14648541, Time: 0.0073 Steps: 101550, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001497, Sample Num: 23952, Cur Loss: 0.35192800, Cur Avg Loss: 0.45403642, Log Avg loss: 0.40314230, Global Avg Loss: 1.14641222, Time: 0.0076 Steps: 101560, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001507, Sample Num: 24112, Cur Loss: 0.23663718, Cur Avg Loss: 0.45343358, Log Avg loss: 0.36318769, Global Avg Loss: 1.14633511, Time: 0.0076 Steps: 101570, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001517, Sample Num: 24272, Cur Loss: 0.21828133, Cur Avg Loss: 0.45301467, Log Avg loss: 0.38988509, Global Avg Loss: 1.14626064, Time: 0.0076 Steps: 101580, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001527, Sample Num: 24432, Cur Loss: 0.10904051, Cur Avg Loss: 0.45261785, Log Avg loss: 0.39242044, Global Avg Loss: 1.14618643, Time: 0.0075 Steps: 101590, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001537, Sample Num: 24592, Cur Loss: 0.30104470, Cur Avg Loss: 0.45317113, Log Avg loss: 0.53765645, Global Avg Loss: 1.14612654, Time: 0.0075 Steps: 101600, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001547, Sample Num: 24752, Cur Loss: 0.30111045, Cur Avg Loss: 0.45264975, Log Avg loss: 0.37251431, Global Avg Loss: 1.14605040, Time: 0.0073 Steps: 101610, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001557, Sample Num: 24912, Cur Loss: 0.43608853, Cur Avg Loss: 0.45205709, Log Avg loss: 0.36037208, Global Avg Loss: 1.14597309, Time: 0.0075 Steps: 101620, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001567, Sample Num: 25072, Cur Loss: 0.43529293, Cur Avg Loss: 0.45170493, Log Avg loss: 0.39687452, Global Avg Loss: 1.14589938, Time: 0.0074 Steps: 101630, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001577, Sample Num: 25232, Cur Loss: 0.30993333, Cur Avg Loss: 0.45152282, Log Avg loss: 0.42298509, Global Avg Loss: 1.14582825, Time: 0.0074 Steps: 101640, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001587, Sample Num: 25392, Cur Loss: 0.52113056, Cur Avg Loss: 0.45166684, Log Avg loss: 0.47438003, Global Avg Loss: 1.14576220, Time: 0.0075 Steps: 101650, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001597, Sample Num: 25552, Cur Loss: 1.21287024, Cur Avg Loss: 0.45222633, Log Avg loss: 0.54101742, Global Avg Loss: 1.14570271, Time: 0.0075 Steps: 101660, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001607, Sample Num: 25712, Cur Loss: 0.52929837, Cur Avg Loss: 0.45296671, Log Avg loss: 0.57120433, Global Avg Loss: 1.14564621, Time: 0.0075 Steps: 101670, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001617, Sample Num: 25872, Cur Loss: 0.88538599, Cur Avg Loss: 0.45308086, Log Avg loss: 0.47142492, Global Avg Loss: 1.14557990, Time: 0.0075 Steps: 101680, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001627, Sample Num: 26032, Cur Loss: 0.25523090, Cur Avg Loss: 0.45428928, Log Avg loss: 0.64969170, Global Avg Loss: 1.14553113, Time: 0.0075 Steps: 101690, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001637, Sample Num: 26192, Cur Loss: 0.75746572, Cur Avg Loss: 0.45447493, Log Avg loss: 0.48468035, Global Avg Loss: 1.14546615, Time: 0.0076 Steps: 101700, Updated lr: 0.000005 Training, Epoch: 0048, Batch: 001647, Sample Num: 26352, Cur Loss: 1.00602198, Cur Avg Loss: 0.45483615, Log Avg loss: 0.51396702, Global Avg Loss: 1.14540406, Time: 0.0075 Steps: 101710, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001657, Sample Num: 26512, Cur Loss: 0.30499792, Cur Avg Loss: 0.45501433, Log Avg loss: 0.48435991, Global Avg Loss: 1.14533908, Time: 0.0075 Steps: 101720, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001667, Sample Num: 26672, Cur Loss: 0.35883915, Cur Avg Loss: 0.45509579, Log Avg loss: 0.46859500, Global Avg Loss: 1.14527255, Time: 0.0075 Steps: 101730, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001677, Sample Num: 26832, Cur Loss: 0.87868381, Cur Avg Loss: 0.45478607, Log Avg loss: 0.40315599, Global Avg Loss: 1.14519961, Time: 0.0076 Steps: 101740, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001687, Sample Num: 26992, Cur Loss: 0.24905092, Cur Avg Loss: 0.45412027, Log Avg loss: 0.34246517, Global Avg Loss: 1.14512072, Time: 0.0075 Steps: 101750, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001697, Sample Num: 27152, Cur Loss: 0.17990923, Cur Avg Loss: 0.45364517, Log Avg loss: 0.37349552, Global Avg Loss: 1.14504489, Time: 0.0074 Steps: 101760, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001707, Sample Num: 27312, Cur Loss: 0.61053658, Cur Avg Loss: 0.45365018, Log Avg loss: 0.45450055, Global Avg Loss: 1.14497704, Time: 0.0074 Steps: 101770, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001717, Sample Num: 27472, Cur Loss: 0.33421043, Cur Avg Loss: 0.45420341, Log Avg loss: 0.54863926, Global Avg Loss: 1.14491845, Time: 0.0075 Steps: 101780, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001727, Sample Num: 27632, Cur Loss: 0.75283992, Cur Avg Loss: 0.45463055, Log Avg loss: 0.52797124, Global Avg Loss: 1.14485784, Time: 0.0076 Steps: 101790, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001737, Sample Num: 27792, Cur Loss: 0.57670188, Cur Avg Loss: 0.45479076, Log Avg loss: 0.48245784, Global Avg Loss: 1.14479277, Time: 0.0076 Steps: 101800, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001747, Sample Num: 27952, Cur Loss: 0.20790191, Cur Avg Loss: 0.45485208, Log Avg loss: 0.46550408, Global Avg Loss: 1.14472605, Time: 0.0075 Steps: 101810, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001757, Sample Num: 28112, Cur Loss: 0.52407265, Cur Avg Loss: 0.45475440, Log Avg loss: 0.43768911, Global Avg Loss: 1.14465661, Time: 0.0075 Steps: 101820, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001767, Sample Num: 28272, Cur Loss: 0.64374936, Cur Avg Loss: 0.45475243, Log Avg loss: 0.45440674, Global Avg Loss: 1.14458882, Time: 0.0077 Steps: 101830, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001777, Sample Num: 28432, Cur Loss: 0.31148806, Cur Avg Loss: 0.45508251, Log Avg loss: 0.51340838, Global Avg Loss: 1.14452685, Time: 0.0075 Steps: 101840, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001787, Sample Num: 28592, Cur Loss: 0.32301643, Cur Avg Loss: 0.45543424, Log Avg loss: 0.51793636, Global Avg Loss: 1.14446532, Time: 0.0076 Steps: 101850, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001797, Sample Num: 28752, Cur Loss: 0.29452869, Cur Avg Loss: 0.45545346, Log Avg loss: 0.45888845, Global Avg Loss: 1.14439802, Time: 0.0076 Steps: 101860, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001807, Sample Num: 28912, Cur Loss: 0.70029408, Cur Avg Loss: 0.45571327, Log Avg loss: 0.50240122, Global Avg Loss: 1.14433500, Time: 0.0075 Steps: 101870, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001817, Sample Num: 29072, Cur Loss: 0.13199392, Cur Avg Loss: 0.45588603, Log Avg loss: 0.48710237, Global Avg Loss: 1.14427049, Time: 0.0075 Steps: 101880, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001827, Sample Num: 29232, Cur Loss: 0.35048985, Cur Avg Loss: 0.45604354, Log Avg loss: 0.48466306, Global Avg Loss: 1.14420575, Time: 0.0075 Steps: 101890, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001837, Sample Num: 29392, Cur Loss: 0.21570721, Cur Avg Loss: 0.45556105, Log Avg loss: 0.36741064, Global Avg Loss: 1.14412952, Time: 0.0075 Steps: 101900, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001847, Sample Num: 29552, Cur Loss: 0.71007532, Cur Avg Loss: 0.45580619, Log Avg loss: 0.50083821, Global Avg Loss: 1.14406640, Time: 0.0075 Steps: 101910, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001857, Sample Num: 29712, Cur Loss: 0.53945911, Cur Avg Loss: 0.45567963, Log Avg loss: 0.43230494, Global Avg Loss: 1.14399656, Time: 0.0075 Steps: 101920, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001867, Sample Num: 29872, Cur Loss: 0.46778810, Cur Avg Loss: 0.45569741, Log Avg loss: 0.45899843, Global Avg Loss: 1.14392936, Time: 0.0075 Steps: 101930, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001877, Sample Num: 30032, Cur Loss: 0.41517362, Cur Avg Loss: 0.45558150, Log Avg loss: 0.43394151, Global Avg Loss: 1.14385971, Time: 0.0076 Steps: 101940, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001887, Sample Num: 30192, Cur Loss: 0.56316239, Cur Avg Loss: 0.45598008, Log Avg loss: 0.53079349, Global Avg Loss: 1.14379958, Time: 0.0077 Steps: 101950, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001897, Sample Num: 30352, Cur Loss: 0.29275668, Cur Avg Loss: 0.45608024, Log Avg loss: 0.47497990, Global Avg Loss: 1.14373398, Time: 0.0076 Steps: 101960, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001907, Sample Num: 30512, Cur Loss: 0.28530785, Cur Avg Loss: 0.45630542, Log Avg loss: 0.49902152, Global Avg Loss: 1.14367075, Time: 0.0076 Steps: 101970, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001917, Sample Num: 30672, Cur Loss: 0.64006644, Cur Avg Loss: 0.45598985, Log Avg loss: 0.39581138, Global Avg Loss: 1.14359742, Time: 0.0077 Steps: 101980, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001927, Sample Num: 30832, Cur Loss: 0.26486623, Cur Avg Loss: 0.45621168, Log Avg loss: 0.49873734, Global Avg Loss: 1.14353419, Time: 0.0076 Steps: 101990, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001937, Sample Num: 30992, Cur Loss: 0.35548067, Cur Avg Loss: 0.45654284, Log Avg loss: 0.52035637, Global Avg Loss: 1.14347310, Time: 0.0076 Steps: 102000, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001947, Sample Num: 31152, Cur Loss: 0.92682588, Cur Avg Loss: 0.45665922, Log Avg loss: 0.47920260, Global Avg Loss: 1.14340798, Time: 0.0076 Steps: 102010, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001957, Sample Num: 31312, Cur Loss: 0.29501131, Cur Avg Loss: 0.45692855, Log Avg loss: 0.50936666, Global Avg Loss: 1.14334583, Time: 0.0075 Steps: 102020, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001967, Sample Num: 31472, Cur Loss: 0.05897445, Cur Avg Loss: 0.45653756, Log Avg loss: 0.38002030, Global Avg Loss: 1.14327102, Time: 0.0076 Steps: 102030, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001977, Sample Num: 31632, Cur Loss: 0.21319330, Cur Avg Loss: 0.45599604, Log Avg loss: 0.34948038, Global Avg Loss: 1.14319322, Time: 0.0075 Steps: 102040, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001987, Sample Num: 31792, Cur Loss: 0.32530859, Cur Avg Loss: 0.45644722, Log Avg loss: 0.54564449, Global Avg Loss: 1.14313467, Time: 0.0076 Steps: 102050, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 001997, Sample Num: 31952, Cur Loss: 0.28998730, Cur Avg Loss: 0.45646282, Log Avg loss: 0.45956187, Global Avg Loss: 1.14306769, Time: 0.0076 Steps: 102060, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002007, Sample Num: 32112, Cur Loss: 0.21366984, Cur Avg Loss: 0.45587994, Log Avg loss: 0.33948003, Global Avg Loss: 1.14298896, Time: 0.0075 Steps: 102070, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002017, Sample Num: 32272, Cur Loss: 0.43247193, Cur Avg Loss: 0.45586036, Log Avg loss: 0.45193124, Global Avg Loss: 1.14292126, Time: 0.0076 Steps: 102080, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002027, Sample Num: 32432, Cur Loss: 0.58474350, Cur Avg Loss: 0.45554257, Log Avg loss: 0.39144337, Global Avg Loss: 1.14284765, Time: 0.0076 Steps: 102090, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002037, Sample Num: 32592, Cur Loss: 0.39814150, Cur Avg Loss: 0.45569874, Log Avg loss: 0.48735403, Global Avg Loss: 1.14278345, Time: 0.0075 Steps: 102100, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002047, Sample Num: 32752, Cur Loss: 0.62079698, Cur Avg Loss: 0.45537272, Log Avg loss: 0.38896348, Global Avg Loss: 1.14270963, Time: 0.0075 Steps: 102110, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002057, Sample Num: 32912, Cur Loss: 0.30042660, Cur Avg Loss: 0.45498437, Log Avg loss: 0.37548808, Global Avg Loss: 1.14263450, Time: 0.0064 Steps: 102120, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002067, Sample Num: 33072, Cur Loss: 0.99318826, Cur Avg Loss: 0.45521785, Log Avg loss: 0.50324554, Global Avg Loss: 1.14257189, Time: 0.0064 Steps: 102130, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002077, Sample Num: 33232, Cur Loss: 0.82235038, Cur Avg Loss: 0.45561915, Log Avg loss: 0.53856787, Global Avg Loss: 1.14251276, Time: 0.0065 Steps: 102140, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002087, Sample Num: 33392, Cur Loss: 0.20160757, Cur Avg Loss: 0.45542337, Log Avg loss: 0.41475873, Global Avg Loss: 1.14244152, Time: 0.0065 Steps: 102150, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002097, Sample Num: 33552, Cur Loss: 1.02917981, Cur Avg Loss: 0.45562391, Log Avg loss: 0.49747732, Global Avg Loss: 1.14237838, Time: 0.0065 Steps: 102160, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002107, Sample Num: 33712, Cur Loss: 0.53389770, Cur Avg Loss: 0.45516303, Log Avg loss: 0.35851679, Global Avg Loss: 1.14230166, Time: 0.0065 Steps: 102170, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002117, Sample Num: 33872, Cur Loss: 0.21750766, Cur Avg Loss: 0.45489275, Log Avg loss: 0.39794412, Global Avg Loss: 1.14222881, Time: 0.0074 Steps: 102180, Updated lr: 0.000004 Training, Epoch: 0048, Batch: 002127, Sample Num: 34032, Cur Loss: 0.16291009, Cur Avg Loss: 0.45425959, Log Avg loss: 0.32021955, Global Avg Loss: 1.14214837, Time: 0.0075 Steps: 102190, Updated lr: 0.000004 ***** Running evaluation checkpoint-102192 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-102192 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.644536, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.564167, "eval_total_loss": 396.609228, "eval_mae": 0.561854, "eval_mse": 0.564304, "eval_r2": 0.641291, "eval_sp_statistic": 0.756839, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.801541, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.939891, "test_total_loss": 471.825339, "test_mae": 0.665629, "test_mse": 0.940173, "test_r2": 0.393204, "test_sp_statistic": 0.605097, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.670951, "test_ps_pvalue": 0.0, "lr": 4.037932669511617e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.1421288650960293, "train_cur_epoch_loss": 966.5006871186197, "train_cur_epoch_avg_loss": 0.4539693222727194, "train_cur_epoch_time": 17.644536018371582, "train_cur_epoch_avg_time": 0.008287710670911969, "epoch": 48, "step": 102192} ################################################## Training, Epoch: 0049, Batch: 000008, Sample Num: 128, Cur Loss: 0.32534656, Cur Avg Loss: 0.40045106, Log Avg loss: 0.34941525, Global Avg Loss: 1.14207081, Time: 0.0115 Steps: 102200, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000018, Sample Num: 288, Cur Loss: 0.61852849, Cur Avg Loss: 0.36586588, Log Avg loss: 0.33819774, Global Avg Loss: 1.14199216, Time: 0.0067 Steps: 102210, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000028, Sample Num: 448, Cur Loss: 0.37884200, Cur Avg Loss: 0.40330910, Log Avg loss: 0.47070689, Global Avg Loss: 1.14192649, Time: 0.0189 Steps: 102220, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000038, Sample Num: 608, Cur Loss: 0.08129888, Cur Avg Loss: 0.39002505, Log Avg loss: 0.35282969, Global Avg Loss: 1.14184930, Time: 0.0115 Steps: 102230, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000048, Sample Num: 768, Cur Loss: 0.34256369, Cur Avg Loss: 0.39476839, Log Avg loss: 0.41279309, Global Avg Loss: 1.14177799, Time: 0.0187 Steps: 102240, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000058, Sample Num: 928, Cur Loss: 0.16508859, Cur Avg Loss: 0.39756389, Log Avg loss: 0.41098228, Global Avg Loss: 1.14170652, Time: 0.0067 Steps: 102250, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000068, Sample Num: 1088, Cur Loss: 0.49247038, Cur Avg Loss: 0.39731321, Log Avg loss: 0.39585928, Global Avg Loss: 1.14163358, Time: 0.0064 Steps: 102260, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000078, Sample Num: 1248, Cur Loss: 0.27407968, Cur Avg Loss: 0.39269798, Log Avg loss: 0.36131445, Global Avg Loss: 1.14155728, Time: 0.0186 Steps: 102270, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000088, Sample Num: 1408, Cur Loss: 0.61312783, Cur Avg Loss: 0.39267002, Log Avg loss: 0.39245188, Global Avg Loss: 1.14148404, Time: 0.0200 Steps: 102280, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000098, Sample Num: 1568, Cur Loss: 0.20741259, Cur Avg Loss: 0.39795973, Log Avg loss: 0.44450924, Global Avg Loss: 1.14141591, Time: 0.0071 Steps: 102290, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000108, Sample Num: 1728, Cur Loss: 0.37519974, Cur Avg Loss: 0.41246922, Log Avg loss: 0.55466222, Global Avg Loss: 1.14135855, Time: 0.0067 Steps: 102300, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000118, Sample Num: 1888, Cur Loss: 0.19510245, Cur Avg Loss: 0.41355364, Log Avg loss: 0.42526534, Global Avg Loss: 1.14128856, Time: 0.0166 Steps: 102310, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000128, Sample Num: 2048, Cur Loss: 0.24992560, Cur Avg Loss: 0.41477428, Log Avg loss: 0.42917780, Global Avg Loss: 1.14121896, Time: 0.0064 Steps: 102320, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000138, Sample Num: 2208, Cur Loss: 0.29536679, Cur Avg Loss: 0.41108866, Log Avg loss: 0.36391277, Global Avg Loss: 1.14114300, Time: 0.0186 Steps: 102330, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000148, Sample Num: 2368, Cur Loss: 1.14227939, Cur Avg Loss: 0.41593807, Log Avg loss: 0.48285994, Global Avg Loss: 1.14107868, Time: 0.0071 Steps: 102340, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000158, Sample Num: 2528, Cur Loss: 0.39876661, Cur Avg Loss: 0.42097112, Log Avg loss: 0.49546025, Global Avg Loss: 1.14101560, Time: 0.0064 Steps: 102350, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000168, Sample Num: 2688, Cur Loss: 0.14891572, Cur Avg Loss: 0.42863989, Log Avg loss: 0.54980640, Global Avg Loss: 1.14095784, Time: 0.0070 Steps: 102360, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000178, Sample Num: 2848, Cur Loss: 1.09336877, Cur Avg Loss: 0.42986308, Log Avg loss: 0.45041276, Global Avg Loss: 1.14089038, Time: 0.0065 Steps: 102370, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000188, Sample Num: 3008, Cur Loss: 0.69975877, Cur Avg Loss: 0.43314965, Log Avg loss: 0.49165054, Global Avg Loss: 1.14082697, Time: 0.0065 Steps: 102380, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000198, Sample Num: 3168, Cur Loss: 0.64895898, Cur Avg Loss: 0.43494021, Log Avg loss: 0.46860279, Global Avg Loss: 1.14076132, Time: 0.0066 Steps: 102390, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000208, Sample Num: 3328, Cur Loss: 0.49468517, Cur Avg Loss: 0.43290559, Log Avg loss: 0.39262011, Global Avg Loss: 1.14068826, Time: 0.0066 Steps: 102400, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000218, Sample Num: 3488, Cur Loss: 0.57242870, Cur Avg Loss: 0.43196407, Log Avg loss: 0.41238033, Global Avg Loss: 1.14061714, Time: 0.0067 Steps: 102410, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000228, Sample Num: 3648, Cur Loss: 0.85678232, Cur Avg Loss: 0.43514587, Log Avg loss: 0.50450924, Global Avg Loss: 1.14055503, Time: 0.0066 Steps: 102420, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000238, Sample Num: 3808, Cur Loss: 0.38494849, Cur Avg Loss: 0.43351748, Log Avg loss: 0.39639025, Global Avg Loss: 1.14048238, Time: 0.0076 Steps: 102430, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000248, Sample Num: 3968, Cur Loss: 0.34717256, Cur Avg Loss: 0.43136862, Log Avg loss: 0.38022574, Global Avg Loss: 1.14040816, Time: 0.0076 Steps: 102440, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000258, Sample Num: 4128, Cur Loss: 0.78963673, Cur Avg Loss: 0.43067057, Log Avg loss: 0.41335891, Global Avg Loss: 1.14033720, Time: 0.0076 Steps: 102450, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000268, Sample Num: 4288, Cur Loss: 0.76566017, Cur Avg Loss: 0.43259909, Log Avg loss: 0.48235488, Global Avg Loss: 1.14027298, Time: 0.0076 Steps: 102460, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000278, Sample Num: 4448, Cur Loss: 0.45283929, Cur Avg Loss: 0.43483389, Log Avg loss: 0.49472640, Global Avg Loss: 1.14020998, Time: 0.0074 Steps: 102470, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000288, Sample Num: 4608, Cur Loss: 0.34766424, Cur Avg Loss: 0.43224166, Log Avg loss: 0.36017760, Global Avg Loss: 1.14013387, Time: 0.0074 Steps: 102480, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000298, Sample Num: 4768, Cur Loss: 0.59791934, Cur Avg Loss: 0.43337455, Log Avg loss: 0.46600199, Global Avg Loss: 1.14006809, Time: 0.0074 Steps: 102490, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000308, Sample Num: 4928, Cur Loss: 0.35129607, Cur Avg Loss: 0.43584987, Log Avg loss: 0.50961421, Global Avg Loss: 1.14000658, Time: 0.0074 Steps: 102500, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000318, Sample Num: 5088, Cur Loss: 0.23188347, Cur Avg Loss: 0.43474546, Log Avg loss: 0.40072988, Global Avg Loss: 1.13993447, Time: 0.0075 Steps: 102510, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000328, Sample Num: 5248, Cur Loss: 0.89821017, Cur Avg Loss: 0.43706119, Log Avg loss: 0.51070130, Global Avg Loss: 1.13987309, Time: 0.0074 Steps: 102520, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000338, Sample Num: 5408, Cur Loss: 0.21347925, Cur Avg Loss: 0.43540426, Log Avg loss: 0.38105704, Global Avg Loss: 1.13979908, Time: 0.0075 Steps: 102530, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000348, Sample Num: 5568, Cur Loss: 0.32456881, Cur Avg Loss: 0.43576823, Log Avg loss: 0.44807029, Global Avg Loss: 1.13973162, Time: 0.0076 Steps: 102540, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000358, Sample Num: 5728, Cur Loss: 0.50205618, Cur Avg Loss: 0.43685130, Log Avg loss: 0.47454225, Global Avg Loss: 1.13966676, Time: 0.0076 Steps: 102550, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000368, Sample Num: 5888, Cur Loss: 0.36979398, Cur Avg Loss: 0.43675786, Log Avg loss: 0.43341245, Global Avg Loss: 1.13959789, Time: 0.0074 Steps: 102560, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000378, Sample Num: 6048, Cur Loss: 0.69854790, Cur Avg Loss: 0.43380829, Log Avg loss: 0.32526412, Global Avg Loss: 1.13951850, Time: 0.0076 Steps: 102570, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000388, Sample Num: 6208, Cur Loss: 0.54049039, Cur Avg Loss: 0.43521699, Log Avg loss: 0.48846592, Global Avg Loss: 1.13945503, Time: 0.0076 Steps: 102580, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000398, Sample Num: 6368, Cur Loss: 0.47532874, Cur Avg Loss: 0.43368180, Log Avg loss: 0.37411653, Global Avg Loss: 1.13938043, Time: 0.0076 Steps: 102590, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000408, Sample Num: 6528, Cur Loss: 0.78412288, Cur Avg Loss: 0.43453526, Log Avg loss: 0.46850311, Global Avg Loss: 1.13931504, Time: 0.0075 Steps: 102600, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000418, Sample Num: 6688, Cur Loss: 0.51990229, Cur Avg Loss: 0.43395497, Log Avg loss: 0.41027911, Global Avg Loss: 1.13924399, Time: 0.0077 Steps: 102610, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000428, Sample Num: 6848, Cur Loss: 0.83866554, Cur Avg Loss: 0.43511912, Log Avg loss: 0.48378055, Global Avg Loss: 1.13918012, Time: 0.0077 Steps: 102620, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000438, Sample Num: 7008, Cur Loss: 0.28713265, Cur Avg Loss: 0.43512736, Log Avg loss: 0.43548010, Global Avg Loss: 1.13911155, Time: 0.0075 Steps: 102630, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000448, Sample Num: 7168, Cur Loss: 0.20729806, Cur Avg Loss: 0.43597200, Log Avg loss: 0.47296732, Global Avg Loss: 1.13904665, Time: 0.0075 Steps: 102640, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000458, Sample Num: 7328, Cur Loss: 0.42093050, Cur Avg Loss: 0.43591566, Log Avg loss: 0.43339148, Global Avg Loss: 1.13897791, Time: 0.0076 Steps: 102650, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000468, Sample Num: 7488, Cur Loss: 0.80061662, Cur Avg Loss: 0.43681842, Log Avg loss: 0.47816471, Global Avg Loss: 1.13891354, Time: 0.0075 Steps: 102660, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000478, Sample Num: 7648, Cur Loss: 0.42360818, Cur Avg Loss: 0.43838920, Log Avg loss: 0.51190197, Global Avg Loss: 1.13885247, Time: 0.0077 Steps: 102670, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000488, Sample Num: 7808, Cur Loss: 0.54772049, Cur Avg Loss: 0.43869979, Log Avg loss: 0.45354567, Global Avg Loss: 1.13878573, Time: 0.0075 Steps: 102680, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000498, Sample Num: 7968, Cur Loss: 0.62920541, Cur Avg Loss: 0.44026329, Log Avg loss: 0.51656230, Global Avg Loss: 1.13872513, Time: 0.0077 Steps: 102690, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000508, Sample Num: 8128, Cur Loss: 0.88337255, Cur Avg Loss: 0.44201665, Log Avg loss: 0.52933407, Global Avg Loss: 1.13866580, Time: 0.0075 Steps: 102700, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000518, Sample Num: 8288, Cur Loss: 0.11474615, Cur Avg Loss: 0.44311942, Log Avg loss: 0.49914007, Global Avg Loss: 1.13860353, Time: 0.0075 Steps: 102710, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000528, Sample Num: 8448, Cur Loss: 0.33864003, Cur Avg Loss: 0.44472734, Log Avg loss: 0.52801735, Global Avg Loss: 1.13854409, Time: 0.0074 Steps: 102720, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000538, Sample Num: 8608, Cur Loss: 0.43057558, Cur Avg Loss: 0.44411686, Log Avg loss: 0.41188337, Global Avg Loss: 1.13847336, Time: 0.0074 Steps: 102730, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000548, Sample Num: 8768, Cur Loss: 0.60050648, Cur Avg Loss: 0.44235093, Log Avg loss: 0.34734436, Global Avg Loss: 1.13839635, Time: 0.0075 Steps: 102740, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000558, Sample Num: 8928, Cur Loss: 0.26424924, Cur Avg Loss: 0.44140061, Log Avg loss: 0.38932283, Global Avg Loss: 1.13832345, Time: 0.0075 Steps: 102750, Updated lr: 0.000004 Training, Epoch: 0049, Batch: 000568, Sample Num: 9088, Cur Loss: 0.31710303, Cur Avg Loss: 0.44233255, Log Avg loss: 0.49433477, Global Avg Loss: 1.13826078, Time: 0.0074 Steps: 102760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000578, Sample Num: 9248, Cur Loss: 0.32862020, Cur Avg Loss: 0.44036274, Log Avg loss: 0.32847778, Global Avg Loss: 1.13818199, Time: 0.0073 Steps: 102770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000588, Sample Num: 9408, Cur Loss: 1.25380981, Cur Avg Loss: 0.44253378, Log Avg loss: 0.56801975, Global Avg Loss: 1.13812651, Time: 0.0073 Steps: 102780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000598, Sample Num: 9568, Cur Loss: 0.29414329, Cur Avg Loss: 0.44072644, Log Avg loss: 0.33445487, Global Avg Loss: 1.13804833, Time: 0.0074 Steps: 102790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000608, Sample Num: 9728, Cur Loss: 0.63173592, Cur Avg Loss: 0.44171425, Log Avg loss: 0.50078522, Global Avg Loss: 1.13798634, Time: 0.0074 Steps: 102800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000618, Sample Num: 9888, Cur Loss: 0.31260213, Cur Avg Loss: 0.44049998, Log Avg loss: 0.36667247, Global Avg Loss: 1.13791131, Time: 0.0074 Steps: 102810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000628, Sample Num: 10048, Cur Loss: 0.27808073, Cur Avg Loss: 0.43983547, Log Avg loss: 0.39876841, Global Avg Loss: 1.13783942, Time: 0.0075 Steps: 102820, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000638, Sample Num: 10208, Cur Loss: 0.49721700, Cur Avg Loss: 0.43942099, Log Avg loss: 0.41339162, Global Avg Loss: 1.13776897, Time: 0.0075 Steps: 102830, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000648, Sample Num: 10368, Cur Loss: 0.45762223, Cur Avg Loss: 0.44105593, Log Avg loss: 0.54536561, Global Avg Loss: 1.13771137, Time: 0.0075 Steps: 102840, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000658, Sample Num: 10528, Cur Loss: 0.64528537, Cur Avg Loss: 0.44090402, Log Avg loss: 0.43106021, Global Avg Loss: 1.13764266, Time: 0.0075 Steps: 102850, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000668, Sample Num: 10688, Cur Loss: 0.23094961, Cur Avg Loss: 0.44082791, Log Avg loss: 0.43581996, Global Avg Loss: 1.13757443, Time: 0.0074 Steps: 102860, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000678, Sample Num: 10848, Cur Loss: 0.13822998, Cur Avg Loss: 0.44201382, Log Avg loss: 0.52123245, Global Avg Loss: 1.13751452, Time: 0.0074 Steps: 102870, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000688, Sample Num: 11008, Cur Loss: 1.04356861, Cur Avg Loss: 0.44293062, Log Avg loss: 0.50508974, Global Avg Loss: 1.13745304, Time: 0.0075 Steps: 102880, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000698, Sample Num: 11168, Cur Loss: 0.73446709, Cur Avg Loss: 0.44232137, Log Avg loss: 0.40040492, Global Avg Loss: 1.13738141, Time: 0.0074 Steps: 102890, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000708, Sample Num: 11328, Cur Loss: 0.22044633, Cur Avg Loss: 0.44182927, Log Avg loss: 0.40748062, Global Avg Loss: 1.13731048, Time: 0.0074 Steps: 102900, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000718, Sample Num: 11488, Cur Loss: 0.76162165, Cur Avg Loss: 0.44154900, Log Avg loss: 0.42170603, Global Avg Loss: 1.13724094, Time: 0.0074 Steps: 102910, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000728, Sample Num: 11648, Cur Loss: 0.23422205, Cur Avg Loss: 0.44013216, Log Avg loss: 0.33840262, Global Avg Loss: 1.13716332, Time: 0.0074 Steps: 102920, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000738, Sample Num: 11808, Cur Loss: 0.33863676, Cur Avg Loss: 0.44077976, Log Avg loss: 0.48792540, Global Avg Loss: 1.13710025, Time: 0.0075 Steps: 102930, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000748, Sample Num: 11968, Cur Loss: 0.24785046, Cur Avg Loss: 0.44054399, Log Avg loss: 0.42314363, Global Avg Loss: 1.13703089, Time: 0.0074 Steps: 102940, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000758, Sample Num: 12128, Cur Loss: 0.28533074, Cur Avg Loss: 0.43929632, Log Avg loss: 0.34597076, Global Avg Loss: 1.13695405, Time: 0.0075 Steps: 102950, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000768, Sample Num: 12288, Cur Loss: 0.51380122, Cur Avg Loss: 0.43910001, Log Avg loss: 0.42421968, Global Avg Loss: 1.13688483, Time: 0.0084 Steps: 102960, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000778, Sample Num: 12448, Cur Loss: 0.83456922, Cur Avg Loss: 0.44178525, Log Avg loss: 0.64801170, Global Avg Loss: 1.13683735, Time: 0.0075 Steps: 102970, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000788, Sample Num: 12608, Cur Loss: 0.40314353, Cur Avg Loss: 0.44272503, Log Avg loss: 0.51584045, Global Avg Loss: 1.13677705, Time: 0.0075 Steps: 102980, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000798, Sample Num: 12768, Cur Loss: 0.83697808, Cur Avg Loss: 0.44376481, Log Avg loss: 0.52569950, Global Avg Loss: 1.13671771, Time: 0.0074 Steps: 102990, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000808, Sample Num: 12928, Cur Loss: 0.51737225, Cur Avg Loss: 0.44371370, Log Avg loss: 0.43963495, Global Avg Loss: 1.13665004, Time: 0.0075 Steps: 103000, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000818, Sample Num: 13088, Cur Loss: 0.34750432, Cur Avg Loss: 0.44438073, Log Avg loss: 0.49827652, Global Avg Loss: 1.13658806, Time: 0.0074 Steps: 103010, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000828, Sample Num: 13248, Cur Loss: 0.28759515, Cur Avg Loss: 0.44508492, Log Avg loss: 0.50268796, Global Avg Loss: 1.13652653, Time: 0.0074 Steps: 103020, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000838, Sample Num: 13408, Cur Loss: 0.54253572, Cur Avg Loss: 0.44657726, Log Avg loss: 0.57014291, Global Avg Loss: 1.13647156, Time: 0.0074 Steps: 103030, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000848, Sample Num: 13568, Cur Loss: 0.36224550, Cur Avg Loss: 0.44568429, Log Avg loss: 0.37085362, Global Avg Loss: 1.13639726, Time: 0.0075 Steps: 103040, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000858, Sample Num: 13728, Cur Loss: 0.22457351, Cur Avg Loss: 0.44651202, Log Avg loss: 0.51670276, Global Avg Loss: 1.13633712, Time: 0.0074 Steps: 103050, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000868, Sample Num: 13888, Cur Loss: 0.57369590, Cur Avg Loss: 0.44693891, Log Avg loss: 0.48356650, Global Avg Loss: 1.13627378, Time: 0.0074 Steps: 103060, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000878, Sample Num: 14048, Cur Loss: 0.23076423, Cur Avg Loss: 0.44621114, Log Avg loss: 0.38304087, Global Avg Loss: 1.13620070, Time: 0.0074 Steps: 103070, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000888, Sample Num: 14208, Cur Loss: 0.38836977, Cur Avg Loss: 0.44706008, Log Avg loss: 0.52159694, Global Avg Loss: 1.13614108, Time: 0.0073 Steps: 103080, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000898, Sample Num: 14368, Cur Loss: 0.69201195, Cur Avg Loss: 0.44660895, Log Avg loss: 0.40654832, Global Avg Loss: 1.13607031, Time: 0.0074 Steps: 103090, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000908, Sample Num: 14528, Cur Loss: 0.29426238, Cur Avg Loss: 0.44648787, Log Avg loss: 0.43561522, Global Avg Loss: 1.13600237, Time: 0.0075 Steps: 103100, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000918, Sample Num: 14688, Cur Loss: 0.22570904, Cur Avg Loss: 0.44663492, Log Avg loss: 0.45998726, Global Avg Loss: 1.13593680, Time: 0.0074 Steps: 103110, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000928, Sample Num: 14848, Cur Loss: 0.33072668, Cur Avg Loss: 0.44522049, Log Avg loss: 0.31537522, Global Avg Loss: 1.13585723, Time: 0.0075 Steps: 103120, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000938, Sample Num: 15008, Cur Loss: 0.45296943, Cur Avg Loss: 0.44506672, Log Avg loss: 0.43079674, Global Avg Loss: 1.13578886, Time: 0.0075 Steps: 103130, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000948, Sample Num: 15168, Cur Loss: 0.66195112, Cur Avg Loss: 0.44687147, Log Avg loss: 0.61615768, Global Avg Loss: 1.13573848, Time: 0.0074 Steps: 103140, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000958, Sample Num: 15328, Cur Loss: 0.36053830, Cur Avg Loss: 0.44677325, Log Avg loss: 0.43746133, Global Avg Loss: 1.13567079, Time: 0.0074 Steps: 103150, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000968, Sample Num: 15488, Cur Loss: 0.48023251, Cur Avg Loss: 0.44589193, Log Avg loss: 0.36146220, Global Avg Loss: 1.13559574, Time: 0.0075 Steps: 103160, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000978, Sample Num: 15648, Cur Loss: 0.80351847, Cur Avg Loss: 0.44657332, Log Avg loss: 0.51253140, Global Avg Loss: 1.13553535, Time: 0.0075 Steps: 103170, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000988, Sample Num: 15808, Cur Loss: 0.51213861, Cur Avg Loss: 0.44663959, Log Avg loss: 0.45312080, Global Avg Loss: 1.13546921, Time: 0.0066 Steps: 103180, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 000998, Sample Num: 15968, Cur Loss: 0.42717159, Cur Avg Loss: 0.44743569, Log Avg loss: 0.52609015, Global Avg Loss: 1.13541015, Time: 0.0196 Steps: 103190, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001008, Sample Num: 16128, Cur Loss: 0.10828599, Cur Avg Loss: 0.44657838, Log Avg loss: 0.36101917, Global Avg Loss: 1.13533512, Time: 0.0065 Steps: 103200, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001018, Sample Num: 16288, Cur Loss: 0.40203595, Cur Avg Loss: 0.44618813, Log Avg loss: 0.40685131, Global Avg Loss: 1.13526453, Time: 0.0071 Steps: 103210, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001028, Sample Num: 16448, Cur Loss: 0.44097519, Cur Avg Loss: 0.44652207, Log Avg loss: 0.48051662, Global Avg Loss: 1.13520110, Time: 0.0077 Steps: 103220, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001038, Sample Num: 16608, Cur Loss: 0.09192070, Cur Avg Loss: 0.44526872, Log Avg loss: 0.31642415, Global Avg Loss: 1.13512179, Time: 0.0076 Steps: 103230, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001048, Sample Num: 16768, Cur Loss: 0.30909395, Cur Avg Loss: 0.44544057, Log Avg loss: 0.46327943, Global Avg Loss: 1.13505671, Time: 0.0073 Steps: 103240, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001058, Sample Num: 16928, Cur Loss: 1.03190780, Cur Avg Loss: 0.44641672, Log Avg loss: 0.54871645, Global Avg Loss: 1.13499992, Time: 0.0067 Steps: 103250, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001068, Sample Num: 17088, Cur Loss: 0.26632077, Cur Avg Loss: 0.44644092, Log Avg loss: 0.44900132, Global Avg Loss: 1.13493349, Time: 0.0114 Steps: 103260, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001078, Sample Num: 17248, Cur Loss: 0.84196466, Cur Avg Loss: 0.44659129, Log Avg loss: 0.46265073, Global Avg Loss: 1.13486839, Time: 0.0117 Steps: 103270, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001088, Sample Num: 17408, Cur Loss: 0.43526185, Cur Avg Loss: 0.44690012, Log Avg loss: 0.48019265, Global Avg Loss: 1.13480500, Time: 0.0064 Steps: 103280, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001098, Sample Num: 17568, Cur Loss: 0.30621636, Cur Avg Loss: 0.44738305, Log Avg loss: 0.49992537, Global Avg Loss: 1.13474353, Time: 0.0204 Steps: 103290, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001108, Sample Num: 17728, Cur Loss: 0.28234592, Cur Avg Loss: 0.44661293, Log Avg loss: 0.36205378, Global Avg Loss: 1.13466873, Time: 0.0204 Steps: 103300, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001118, Sample Num: 17888, Cur Loss: 0.67246675, Cur Avg Loss: 0.44705137, Log Avg loss: 0.49563067, Global Avg Loss: 1.13460688, Time: 0.0202 Steps: 103310, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001128, Sample Num: 18048, Cur Loss: 0.43133694, Cur Avg Loss: 0.44724683, Log Avg loss: 0.46909935, Global Avg Loss: 1.13454246, Time: 0.0065 Steps: 103320, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001138, Sample Num: 18208, Cur Loss: 0.46431226, Cur Avg Loss: 0.44720185, Log Avg loss: 0.44212846, Global Avg Loss: 1.13447545, Time: 0.0066 Steps: 103330, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001148, Sample Num: 18368, Cur Loss: 0.50859857, Cur Avg Loss: 0.44748679, Log Avg loss: 0.47991300, Global Avg Loss: 1.13441211, Time: 0.0131 Steps: 103340, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001158, Sample Num: 18528, Cur Loss: 0.35359898, Cur Avg Loss: 0.44678846, Log Avg loss: 0.36661968, Global Avg Loss: 1.13433782, Time: 0.0078 Steps: 103350, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001168, Sample Num: 18688, Cur Loss: 0.39214751, Cur Avg Loss: 0.44647970, Log Avg loss: 0.41072531, Global Avg Loss: 1.13426781, Time: 0.0064 Steps: 103360, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001178, Sample Num: 18848, Cur Loss: 1.26781332, Cur Avg Loss: 0.44725369, Log Avg loss: 0.53765547, Global Avg Loss: 1.13421010, Time: 0.0138 Steps: 103370, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001188, Sample Num: 19008, Cur Loss: 0.40093273, Cur Avg Loss: 0.44718090, Log Avg loss: 0.43860671, Global Avg Loss: 1.13414281, Time: 0.0230 Steps: 103380, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001198, Sample Num: 19168, Cur Loss: 0.66956985, Cur Avg Loss: 0.44688099, Log Avg loss: 0.41125113, Global Avg Loss: 1.13407289, Time: 0.0067 Steps: 103390, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001208, Sample Num: 19328, Cur Loss: 0.98238438, Cur Avg Loss: 0.44711662, Log Avg loss: 0.47534581, Global Avg Loss: 1.13400919, Time: 0.0066 Steps: 103400, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001218, Sample Num: 19488, Cur Loss: 0.87913746, Cur Avg Loss: 0.44837651, Log Avg loss: 0.60057115, Global Avg Loss: 1.13395760, Time: 0.0066 Steps: 103410, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001228, Sample Num: 19648, Cur Loss: 0.46742260, Cur Avg Loss: 0.44877545, Log Avg loss: 0.49736567, Global Avg Loss: 1.13389605, Time: 0.0066 Steps: 103420, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001238, Sample Num: 19808, Cur Loss: 0.29635447, Cur Avg Loss: 0.45052671, Log Avg loss: 0.66558122, Global Avg Loss: 1.13385077, Time: 0.0071 Steps: 103430, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001248, Sample Num: 19968, Cur Loss: 0.24522907, Cur Avg Loss: 0.45024026, Log Avg loss: 0.41477805, Global Avg Loss: 1.13378125, Time: 0.0068 Steps: 103440, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001258, Sample Num: 20128, Cur Loss: 0.24140185, Cur Avg Loss: 0.45063312, Log Avg loss: 0.49966225, Global Avg Loss: 1.13371996, Time: 0.0067 Steps: 103450, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001268, Sample Num: 20288, Cur Loss: 0.15173128, Cur Avg Loss: 0.44985873, Log Avg loss: 0.35244092, Global Avg Loss: 1.13364444, Time: 0.0067 Steps: 103460, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001278, Sample Num: 20448, Cur Loss: 1.07301366, Cur Avg Loss: 0.44950302, Log Avg loss: 0.40439811, Global Avg Loss: 1.13357396, Time: 0.0067 Steps: 103470, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001288, Sample Num: 20608, Cur Loss: 0.32241064, Cur Avg Loss: 0.44930370, Log Avg loss: 0.42383113, Global Avg Loss: 1.13350537, Time: 0.0068 Steps: 103480, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001298, Sample Num: 20768, Cur Loss: 0.45743200, Cur Avg Loss: 0.44994118, Log Avg loss: 0.53204874, Global Avg Loss: 1.13344726, Time: 0.0068 Steps: 103490, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001308, Sample Num: 20928, Cur Loss: 0.75242800, Cur Avg Loss: 0.44912241, Log Avg loss: 0.34284565, Global Avg Loss: 1.13337087, Time: 0.0068 Steps: 103500, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001318, Sample Num: 21088, Cur Loss: 0.32438210, Cur Avg Loss: 0.44789467, Log Avg loss: 0.28730626, Global Avg Loss: 1.13328913, Time: 0.0069 Steps: 103510, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001328, Sample Num: 21248, Cur Loss: 0.42828959, Cur Avg Loss: 0.44788333, Log Avg loss: 0.44638865, Global Avg Loss: 1.13322278, Time: 0.0077 Steps: 103520, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001338, Sample Num: 21408, Cur Loss: 0.28525567, Cur Avg Loss: 0.44759077, Log Avg loss: 0.40873905, Global Avg Loss: 1.13315280, Time: 0.0077 Steps: 103530, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001348, Sample Num: 21568, Cur Loss: 0.42301866, Cur Avg Loss: 0.44768126, Log Avg loss: 0.45978942, Global Avg Loss: 1.13308777, Time: 0.0078 Steps: 103540, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001358, Sample Num: 21728, Cur Loss: 0.26470652, Cur Avg Loss: 0.44811967, Log Avg loss: 0.50721697, Global Avg Loss: 1.13302732, Time: 0.0078 Steps: 103550, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001368, Sample Num: 21888, Cur Loss: 0.23741624, Cur Avg Loss: 0.44725722, Log Avg loss: 0.33013697, Global Avg Loss: 1.13294980, Time: 0.0078 Steps: 103560, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001378, Sample Num: 22048, Cur Loss: 0.64124179, Cur Avg Loss: 0.44906465, Log Avg loss: 0.69632088, Global Avg Loss: 1.13290764, Time: 0.0078 Steps: 103570, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001388, Sample Num: 22208, Cur Loss: 0.67007637, Cur Avg Loss: 0.44848560, Log Avg loss: 0.36869163, Global Avg Loss: 1.13283386, Time: 0.0077 Steps: 103580, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001398, Sample Num: 22368, Cur Loss: 0.82064950, Cur Avg Loss: 0.44936116, Log Avg loss: 0.57088933, Global Avg Loss: 1.13277961, Time: 0.0077 Steps: 103590, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001408, Sample Num: 22528, Cur Loss: 0.41633612, Cur Avg Loss: 0.45013275, Log Avg loss: 0.55800152, Global Avg Loss: 1.13272413, Time: 0.0077 Steps: 103600, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001418, Sample Num: 22688, Cur Loss: 1.20633781, Cur Avg Loss: 0.45150498, Log Avg loss: 0.64471478, Global Avg Loss: 1.13267703, Time: 0.0078 Steps: 103610, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001428, Sample Num: 22848, Cur Loss: 0.97599465, Cur Avg Loss: 0.45162618, Log Avg loss: 0.46881187, Global Avg Loss: 1.13261296, Time: 0.0078 Steps: 103620, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001438, Sample Num: 23008, Cur Loss: 1.16194987, Cur Avg Loss: 0.45333517, Log Avg loss: 0.69737888, Global Avg Loss: 1.13257096, Time: 0.0077 Steps: 103630, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001448, Sample Num: 23168, Cur Loss: 0.33867395, Cur Avg Loss: 0.45310781, Log Avg loss: 0.42041345, Global Avg Loss: 1.13250225, Time: 0.0077 Steps: 103640, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001458, Sample Num: 23328, Cur Loss: 0.19730499, Cur Avg Loss: 0.45241588, Log Avg loss: 0.35222406, Global Avg Loss: 1.13242697, Time: 0.0077 Steps: 103650, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001468, Sample Num: 23488, Cur Loss: 0.18982026, Cur Avg Loss: 0.45206704, Log Avg loss: 0.40120644, Global Avg Loss: 1.13235643, Time: 0.0077 Steps: 103660, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001478, Sample Num: 23648, Cur Loss: 0.11999055, Cur Avg Loss: 0.45230296, Log Avg loss: 0.48693673, Global Avg Loss: 1.13229417, Time: 0.0077 Steps: 103670, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001488, Sample Num: 23808, Cur Loss: 0.37519073, Cur Avg Loss: 0.45133734, Log Avg loss: 0.30861886, Global Avg Loss: 1.13221473, Time: 0.0077 Steps: 103680, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001498, Sample Num: 23968, Cur Loss: 0.39215946, Cur Avg Loss: 0.45086608, Log Avg loss: 0.38074245, Global Avg Loss: 1.13214225, Time: 0.0079 Steps: 103690, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001508, Sample Num: 24128, Cur Loss: 0.09176481, Cur Avg Loss: 0.45099619, Log Avg loss: 0.47048694, Global Avg Loss: 1.13207845, Time: 0.0078 Steps: 103700, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001518, Sample Num: 24288, Cur Loss: 0.22221833, Cur Avg Loss: 0.45105971, Log Avg loss: 0.46063828, Global Avg Loss: 1.13201371, Time: 0.0077 Steps: 103710, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001528, Sample Num: 24448, Cur Loss: 0.13706012, Cur Avg Loss: 0.45144908, Log Avg loss: 0.51055556, Global Avg Loss: 1.13195379, Time: 0.0077 Steps: 103720, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001538, Sample Num: 24608, Cur Loss: 0.37963420, Cur Avg Loss: 0.45089633, Log Avg loss: 0.36643486, Global Avg Loss: 1.13187999, Time: 0.0078 Steps: 103730, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001548, Sample Num: 24768, Cur Loss: 0.64429420, Cur Avg Loss: 0.45192209, Log Avg loss: 0.60968525, Global Avg Loss: 1.13182965, Time: 0.0078 Steps: 103740, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001558, Sample Num: 24928, Cur Loss: 0.35171652, Cur Avg Loss: 0.45190038, Log Avg loss: 0.44853902, Global Avg Loss: 1.13176380, Time: 0.0076 Steps: 103750, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001568, Sample Num: 25088, Cur Loss: 0.21022937, Cur Avg Loss: 0.45203421, Log Avg loss: 0.47288496, Global Avg Loss: 1.13170030, Time: 0.0077 Steps: 103760, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001578, Sample Num: 25248, Cur Loss: 0.42765325, Cur Avg Loss: 0.45214521, Log Avg loss: 0.46954982, Global Avg Loss: 1.13163649, Time: 0.0076 Steps: 103770, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001588, Sample Num: 25408, Cur Loss: 0.44276115, Cur Avg Loss: 0.45215615, Log Avg loss: 0.45388263, Global Avg Loss: 1.13157118, Time: 0.0076 Steps: 103780, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001598, Sample Num: 25568, Cur Loss: 0.29560256, Cur Avg Loss: 0.45121637, Log Avg loss: 0.30197945, Global Avg Loss: 1.13149125, Time: 0.0076 Steps: 103790, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001608, Sample Num: 25728, Cur Loss: 0.73106647, Cur Avg Loss: 0.45179873, Log Avg loss: 0.54485956, Global Avg Loss: 1.13143473, Time: 0.0076 Steps: 103800, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001618, Sample Num: 25888, Cur Loss: 0.42687079, Cur Avg Loss: 0.45254513, Log Avg loss: 0.57256635, Global Avg Loss: 1.13138090, Time: 0.0077 Steps: 103810, Updated lr: 0.000003 Training, Epoch: 0049, Batch: 001628, Sample Num: 26048, Cur Loss: 0.38371587, Cur Avg Loss: 0.45343246, Log Avg loss: 0.59700250, Global Avg Loss: 1.13132943, Time: 0.0077 Steps: 103820, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001638, Sample Num: 26208, Cur Loss: 0.77574587, Cur Avg Loss: 0.45333648, Log Avg loss: 0.43771177, Global Avg Loss: 1.13126262, Time: 0.0076 Steps: 103830, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001648, Sample Num: 26368, Cur Loss: 0.16101091, Cur Avg Loss: 0.45324822, Log Avg loss: 0.43878998, Global Avg Loss: 1.13119594, Time: 0.0077 Steps: 103840, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001658, Sample Num: 26528, Cur Loss: 0.79136497, Cur Avg Loss: 0.45383783, Log Avg loss: 0.55100569, Global Avg Loss: 1.13114007, Time: 0.0076 Steps: 103850, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001668, Sample Num: 26688, Cur Loss: 0.45553818, Cur Avg Loss: 0.45335731, Log Avg loss: 0.37368696, Global Avg Loss: 1.13106714, Time: 0.0076 Steps: 103860, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001678, Sample Num: 26848, Cur Loss: 0.58123922, Cur Avg Loss: 0.45346231, Log Avg loss: 0.47097685, Global Avg Loss: 1.13100359, Time: 0.0078 Steps: 103870, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001688, Sample Num: 27008, Cur Loss: 0.77947605, Cur Avg Loss: 0.45396799, Log Avg loss: 0.53882155, Global Avg Loss: 1.13094658, Time: 0.0076 Steps: 103880, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001698, Sample Num: 27168, Cur Loss: 0.08679914, Cur Avg Loss: 0.45369101, Log Avg loss: 0.40693626, Global Avg Loss: 1.13087689, Time: 0.0076 Steps: 103890, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001708, Sample Num: 27328, Cur Loss: 0.34259188, Cur Avg Loss: 0.45406695, Log Avg loss: 0.51790122, Global Avg Loss: 1.13081790, Time: 0.0076 Steps: 103900, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001718, Sample Num: 27488, Cur Loss: 0.48027658, Cur Avg Loss: 0.45397462, Log Avg loss: 0.43820457, Global Avg Loss: 1.13075124, Time: 0.0076 Steps: 103910, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001728, Sample Num: 27648, Cur Loss: 0.30823821, Cur Avg Loss: 0.45376409, Log Avg loss: 0.41759533, Global Avg Loss: 1.13068261, Time: 0.0077 Steps: 103920, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001738, Sample Num: 27808, Cur Loss: 0.55591023, Cur Avg Loss: 0.45398166, Log Avg loss: 0.49157802, Global Avg Loss: 1.13062112, Time: 0.0077 Steps: 103930, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001748, Sample Num: 27968, Cur Loss: 0.46946144, Cur Avg Loss: 0.45400123, Log Avg loss: 0.45740224, Global Avg Loss: 1.13055635, Time: 0.0076 Steps: 103940, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001758, Sample Num: 28128, Cur Loss: 0.14012243, Cur Avg Loss: 0.45419145, Log Avg loss: 0.48744268, Global Avg Loss: 1.13049448, Time: 0.0076 Steps: 103950, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001768, Sample Num: 28288, Cur Loss: 0.58313316, Cur Avg Loss: 0.45467511, Log Avg loss: 0.53970175, Global Avg Loss: 1.13043765, Time: 0.0077 Steps: 103960, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001778, Sample Num: 28448, Cur Loss: 0.18779422, Cur Avg Loss: 0.45390397, Log Avg loss: 0.31756599, Global Avg Loss: 1.13035947, Time: 0.0076 Steps: 103970, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001788, Sample Num: 28608, Cur Loss: 0.62966317, Cur Avg Loss: 0.45396356, Log Avg loss: 0.46455946, Global Avg Loss: 1.13029544, Time: 0.0076 Steps: 103980, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001798, Sample Num: 28768, Cur Loss: 0.09689572, Cur Avg Loss: 0.45369443, Log Avg loss: 0.40557401, Global Avg Loss: 1.13022575, Time: 0.0075 Steps: 103990, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001808, Sample Num: 28928, Cur Loss: 0.19490547, Cur Avg Loss: 0.45363266, Log Avg loss: 0.44252693, Global Avg Loss: 1.13015962, Time: 0.0075 Steps: 104000, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001818, Sample Num: 29088, Cur Loss: 0.22151139, Cur Avg Loss: 0.45346661, Log Avg loss: 0.42344491, Global Avg Loss: 1.13009168, Time: 0.0076 Steps: 104010, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001828, Sample Num: 29248, Cur Loss: 0.50748354, Cur Avg Loss: 0.45355118, Log Avg loss: 0.46892572, Global Avg Loss: 1.13002812, Time: 0.0075 Steps: 104020, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001838, Sample Num: 29408, Cur Loss: 0.57578605, Cur Avg Loss: 0.45370102, Log Avg loss: 0.48109135, Global Avg Loss: 1.12996574, Time: 0.0074 Steps: 104030, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001848, Sample Num: 29568, Cur Loss: 0.34039006, Cur Avg Loss: 0.45374144, Log Avg loss: 0.46117091, Global Avg Loss: 1.12990145, Time: 0.0075 Steps: 104040, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001858, Sample Num: 29728, Cur Loss: 0.87237507, Cur Avg Loss: 0.45425088, Log Avg loss: 0.54839565, Global Avg Loss: 1.12984557, Time: 0.0075 Steps: 104050, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001868, Sample Num: 29888, Cur Loss: 0.59817356, Cur Avg Loss: 0.45439058, Log Avg loss: 0.48034563, Global Avg Loss: 1.12978315, Time: 0.0075 Steps: 104060, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001878, Sample Num: 30048, Cur Loss: 0.61196750, Cur Avg Loss: 0.45491341, Log Avg loss: 0.55257771, Global Avg Loss: 1.12972769, Time: 0.0076 Steps: 104070, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001888, Sample Num: 30208, Cur Loss: 0.76423591, Cur Avg Loss: 0.45476874, Log Avg loss: 0.42760010, Global Avg Loss: 1.12966023, Time: 0.0075 Steps: 104080, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001898, Sample Num: 30368, Cur Loss: 0.22597523, Cur Avg Loss: 0.45527094, Log Avg loss: 0.55008701, Global Avg Loss: 1.12960455, Time: 0.0075 Steps: 104090, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001908, Sample Num: 30528, Cur Loss: 0.20211101, Cur Avg Loss: 0.45505148, Log Avg loss: 0.41339789, Global Avg Loss: 1.12953575, Time: 0.0075 Steps: 104100, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001918, Sample Num: 30688, Cur Loss: 0.64832342, Cur Avg Loss: 0.45442200, Log Avg loss: 0.33431762, Global Avg Loss: 1.12945936, Time: 0.0200 Steps: 104110, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001928, Sample Num: 30848, Cur Loss: 0.11853772, Cur Avg Loss: 0.45456694, Log Avg loss: 0.48236615, Global Avg Loss: 1.12939722, Time: 0.0066 Steps: 104120, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001938, Sample Num: 31008, Cur Loss: 0.31248432, Cur Avg Loss: 0.45466985, Log Avg loss: 0.47451061, Global Avg Loss: 1.12933432, Time: 0.0203 Steps: 104130, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001948, Sample Num: 31168, Cur Loss: 0.26216656, Cur Avg Loss: 0.45443208, Log Avg loss: 0.40835261, Global Avg Loss: 1.12926509, Time: 0.0065 Steps: 104140, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001958, Sample Num: 31328, Cur Loss: 0.36405402, Cur Avg Loss: 0.45409606, Log Avg loss: 0.38863979, Global Avg Loss: 1.12919398, Time: 0.0066 Steps: 104150, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001968, Sample Num: 31488, Cur Loss: 0.72780430, Cur Avg Loss: 0.45402542, Log Avg loss: 0.44019341, Global Avg Loss: 1.12912783, Time: 0.0067 Steps: 104160, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001978, Sample Num: 31648, Cur Loss: 0.42704982, Cur Avg Loss: 0.45347508, Log Avg loss: 0.34516834, Global Avg Loss: 1.12905257, Time: 0.0099 Steps: 104170, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001988, Sample Num: 31808, Cur Loss: 1.24128175, Cur Avg Loss: 0.45354050, Log Avg loss: 0.46648041, Global Avg Loss: 1.12898898, Time: 0.0234 Steps: 104180, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 001998, Sample Num: 31968, Cur Loss: 0.20777079, Cur Avg Loss: 0.45291566, Log Avg loss: 0.32869821, Global Avg Loss: 1.12891217, Time: 0.0067 Steps: 104190, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002008, Sample Num: 32128, Cur Loss: 0.44051135, Cur Avg Loss: 0.45274163, Log Avg loss: 0.41797051, Global Avg Loss: 1.12884394, Time: 0.0203 Steps: 104200, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002018, Sample Num: 32288, Cur Loss: 0.41015971, Cur Avg Loss: 0.45274941, Log Avg loss: 0.45431091, Global Avg Loss: 1.12877921, Time: 0.0065 Steps: 104210, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002028, Sample Num: 32448, Cur Loss: 0.73771369, Cur Avg Loss: 0.45329831, Log Avg loss: 0.56406591, Global Avg Loss: 1.12872502, Time: 0.0069 Steps: 104220, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002038, Sample Num: 32608, Cur Loss: 0.54544604, Cur Avg Loss: 0.45327678, Log Avg loss: 0.44891079, Global Avg Loss: 1.12865980, Time: 0.0127 Steps: 104230, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002048, Sample Num: 32768, Cur Loss: 0.75459707, Cur Avg Loss: 0.45355673, Log Avg loss: 0.51060960, Global Avg Loss: 1.12860051, Time: 0.0083 Steps: 104240, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002058, Sample Num: 32928, Cur Loss: 1.07128191, Cur Avg Loss: 0.45352844, Log Avg loss: 0.44773655, Global Avg Loss: 1.12853520, Time: 0.0074 Steps: 104250, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002068, Sample Num: 33088, Cur Loss: 0.33002168, Cur Avg Loss: 0.45346714, Log Avg loss: 0.44085054, Global Avg Loss: 1.12846924, Time: 0.0074 Steps: 104260, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002078, Sample Num: 33248, Cur Loss: 0.44066525, Cur Avg Loss: 0.45315467, Log Avg loss: 0.38853662, Global Avg Loss: 1.12839828, Time: 0.0074 Steps: 104270, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002088, Sample Num: 33408, Cur Loss: 0.52250063, Cur Avg Loss: 0.45314471, Log Avg loss: 0.45107478, Global Avg Loss: 1.12833333, Time: 0.0074 Steps: 104280, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002098, Sample Num: 33568, Cur Loss: 0.08786768, Cur Avg Loss: 0.45299884, Log Avg loss: 0.42254042, Global Avg Loss: 1.12826565, Time: 0.0073 Steps: 104290, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002108, Sample Num: 33728, Cur Loss: 0.45987973, Cur Avg Loss: 0.45294899, Log Avg loss: 0.44248976, Global Avg Loss: 1.12819990, Time: 0.0074 Steps: 104300, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002118, Sample Num: 33888, Cur Loss: 0.60908538, Cur Avg Loss: 0.45251947, Log Avg loss: 0.36197666, Global Avg Loss: 1.12812644, Time: 0.0074 Steps: 104310, Updated lr: 0.000002 Training, Epoch: 0049, Batch: 002128, Sample Num: 34048, Cur Loss: 0.31245327, Cur Avg Loss: 0.45238210, Log Avg loss: 0.42328802, Global Avg Loss: 1.12805888, Time: 0.0074 Steps: 104320, Updated lr: 0.000002 ***** Running evaluation checkpoint-104321 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-104321 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 17.912044, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.567383, "eval_total_loss": 398.870064, "eval_mae": 0.570185, "eval_mse": 0.567507, "eval_r2": 0.639255, "eval_sp_statistic": 0.756614, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.801867, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.911621, "test_total_loss": 457.63389, "test_mae": 0.65995, "test_mse": 0.91189, "test_r2": 0.411458, "test_sp_statistic": 0.605252, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.671219, "test_ps_pvalue": 0.0, "lr": 2.0189663347558085e-06, "cur_epoch_step": 2129, "train_global_avg_loss": 1.128050492589986, "train_cur_epoch_loss": 962.922455586493, "train_cur_epoch_avg_loss": 0.4522886122999028, "train_cur_epoch_time": 17.912044286727905, "train_cur_epoch_avg_time": 0.008413360397711556, "epoch": 49, "step": 104321} ################################################## Training, Epoch: 0050, Batch: 000009, Sample Num: 144, Cur Loss: 0.80488104, Cur Avg Loss: 0.48591296, Log Avg loss: 0.46265650, Global Avg Loss: 1.12799510, Time: 0.0066 Steps: 104330, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000019, Sample Num: 304, Cur Loss: 0.58690500, Cur Avg Loss: 0.41920929, Log Avg loss: 0.35917598, Global Avg Loss: 1.12792141, Time: 0.0067 Steps: 104340, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000029, Sample Num: 464, Cur Loss: 0.35206646, Cur Avg Loss: 0.38987152, Log Avg loss: 0.33412977, Global Avg Loss: 1.12784534, Time: 0.0066 Steps: 104350, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000039, Sample Num: 624, Cur Loss: 0.19665816, Cur Avg Loss: 0.36197810, Log Avg loss: 0.28108716, Global Avg Loss: 1.12776421, Time: 0.0066 Steps: 104360, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000049, Sample Num: 784, Cur Loss: 0.78747004, Cur Avg Loss: 0.36435568, Log Avg loss: 0.37362827, Global Avg Loss: 1.12769195, Time: 0.0075 Steps: 104370, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000059, Sample Num: 944, Cur Loss: 0.30406097, Cur Avg Loss: 0.37752084, Log Avg loss: 0.44203013, Global Avg Loss: 1.12762626, Time: 0.0077 Steps: 104380, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000069, Sample Num: 1104, Cur Loss: 0.18922937, Cur Avg Loss: 0.39517822, Log Avg loss: 0.49935676, Global Avg Loss: 1.12756608, Time: 0.0076 Steps: 104390, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000079, Sample Num: 1264, Cur Loss: 0.12595740, Cur Avg Loss: 0.39988996, Log Avg loss: 0.43240094, Global Avg Loss: 1.12749949, Time: 0.0076 Steps: 104400, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000089, Sample Num: 1424, Cur Loss: 0.36811668, Cur Avg Loss: 0.39494306, Log Avg loss: 0.35586254, Global Avg Loss: 1.12742559, Time: 0.0076 Steps: 104410, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000099, Sample Num: 1584, Cur Loss: 0.49449772, Cur Avg Loss: 0.38721209, Log Avg loss: 0.31840644, Global Avg Loss: 1.12734811, Time: 0.0075 Steps: 104420, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000109, Sample Num: 1744, Cur Loss: 0.17657626, Cur Avg Loss: 0.40370977, Log Avg loss: 0.56703682, Global Avg Loss: 1.12729445, Time: 0.0075 Steps: 104430, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000119, Sample Num: 1904, Cur Loss: 0.57675934, Cur Avg Loss: 0.41337307, Log Avg loss: 0.51870303, Global Avg Loss: 1.12723618, Time: 0.0076 Steps: 104440, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000129, Sample Num: 2064, Cur Loss: 0.20218241, Cur Avg Loss: 0.41854044, Log Avg loss: 0.48003210, Global Avg Loss: 1.12717422, Time: 0.0077 Steps: 104450, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000139, Sample Num: 2224, Cur Loss: 0.29840019, Cur Avg Loss: 0.41995319, Log Avg loss: 0.43817775, Global Avg Loss: 1.12710826, Time: 0.0075 Steps: 104460, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000149, Sample Num: 2384, Cur Loss: 0.45313621, Cur Avg Loss: 0.42019136, Log Avg loss: 0.42350187, Global Avg Loss: 1.12704091, Time: 0.0075 Steps: 104470, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000159, Sample Num: 2544, Cur Loss: 0.31775731, Cur Avg Loss: 0.41486191, Log Avg loss: 0.33545318, Global Avg Loss: 1.12696515, Time: 0.0077 Steps: 104480, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000169, Sample Num: 2704, Cur Loss: 0.51012248, Cur Avg Loss: 0.42143972, Log Avg loss: 0.52602686, Global Avg Loss: 1.12690763, Time: 0.0077 Steps: 104490, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000179, Sample Num: 2864, Cur Loss: 0.25710297, Cur Avg Loss: 0.41846874, Log Avg loss: 0.36825912, Global Avg Loss: 1.12683504, Time: 0.0075 Steps: 104500, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000189, Sample Num: 3024, Cur Loss: 0.34445295, Cur Avg Loss: 0.42941052, Log Avg loss: 0.62526840, Global Avg Loss: 1.12678704, Time: 0.0075 Steps: 104510, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000199, Sample Num: 3184, Cur Loss: 0.16869450, Cur Avg Loss: 0.42630703, Log Avg loss: 0.36765116, Global Avg Loss: 1.12671441, Time: 0.0077 Steps: 104520, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000209, Sample Num: 3344, Cur Loss: 0.53969228, Cur Avg Loss: 0.42357143, Log Avg loss: 0.36913293, Global Avg Loss: 1.12664194, Time: 0.0076 Steps: 104530, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000219, Sample Num: 3504, Cur Loss: 0.32226363, Cur Avg Loss: 0.42524568, Log Avg loss: 0.46023757, Global Avg Loss: 1.12657819, Time: 0.0076 Steps: 104540, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000229, Sample Num: 3664, Cur Loss: 0.28466052, Cur Avg Loss: 0.42397254, Log Avg loss: 0.39609062, Global Avg Loss: 1.12650832, Time: 0.0076 Steps: 104550, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000239, Sample Num: 3824, Cur Loss: 0.25717407, Cur Avg Loss: 0.42640607, Log Avg loss: 0.48213396, Global Avg Loss: 1.12644670, Time: 0.0076 Steps: 104560, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000249, Sample Num: 3984, Cur Loss: 0.51736224, Cur Avg Loss: 0.43039022, Log Avg loss: 0.52561132, Global Avg Loss: 1.12638924, Time: 0.0077 Steps: 104570, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000259, Sample Num: 4144, Cur Loss: 0.35608757, Cur Avg Loss: 0.42787437, Log Avg loss: 0.36522978, Global Avg Loss: 1.12631646, Time: 0.0075 Steps: 104580, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000269, Sample Num: 4304, Cur Loss: 0.27682626, Cur Avg Loss: 0.42874157, Log Avg loss: 0.45120219, Global Avg Loss: 1.12625191, Time: 0.0076 Steps: 104590, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000279, Sample Num: 4464, Cur Loss: 0.27715003, Cur Avg Loss: 0.43082702, Log Avg loss: 0.48692560, Global Avg Loss: 1.12619079, Time: 0.0075 Steps: 104600, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000289, Sample Num: 4624, Cur Loss: 0.69619590, Cur Avg Loss: 0.42867370, Log Avg loss: 0.36859603, Global Avg Loss: 1.12611836, Time: 0.0075 Steps: 104610, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000299, Sample Num: 4784, Cur Loss: 1.24702835, Cur Avg Loss: 0.42936891, Log Avg loss: 0.44946032, Global Avg Loss: 1.12605369, Time: 0.0076 Steps: 104620, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000309, Sample Num: 4944, Cur Loss: 0.20939037, Cur Avg Loss: 0.42850146, Log Avg loss: 0.40256480, Global Avg Loss: 1.12598454, Time: 0.0075 Steps: 104630, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000319, Sample Num: 5104, Cur Loss: 0.50113523, Cur Avg Loss: 0.42936996, Log Avg loss: 0.45620669, Global Avg Loss: 1.12592053, Time: 0.0077 Steps: 104640, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000329, Sample Num: 5264, Cur Loss: 0.38865155, Cur Avg Loss: 0.42767284, Log Avg loss: 0.37353460, Global Avg Loss: 1.12584864, Time: 0.0076 Steps: 104650, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000339, Sample Num: 5424, Cur Loss: 1.35767508, Cur Avg Loss: 0.42911606, Log Avg loss: 0.47659798, Global Avg Loss: 1.12578660, Time: 0.0075 Steps: 104660, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000349, Sample Num: 5584, Cur Loss: 1.04502583, Cur Avg Loss: 0.43740862, Log Avg loss: 0.71852650, Global Avg Loss: 1.12574769, Time: 0.0075 Steps: 104670, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000359, Sample Num: 5744, Cur Loss: 0.62075084, Cur Avg Loss: 0.44030169, Log Avg loss: 0.54126983, Global Avg Loss: 1.12569186, Time: 0.0076 Steps: 104680, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000369, Sample Num: 5904, Cur Loss: 0.08612154, Cur Avg Loss: 0.43780967, Log Avg loss: 0.34834629, Global Avg Loss: 1.12561761, Time: 0.0076 Steps: 104690, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000379, Sample Num: 6064, Cur Loss: 0.62422597, Cur Avg Loss: 0.43747968, Log Avg loss: 0.42530281, Global Avg Loss: 1.12555072, Time: 0.0076 Steps: 104700, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000389, Sample Num: 6224, Cur Loss: 0.86436963, Cur Avg Loss: 0.43678514, Log Avg loss: 0.41046210, Global Avg Loss: 1.12548243, Time: 0.0075 Steps: 104710, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000399, Sample Num: 6384, Cur Loss: 0.37307081, Cur Avg Loss: 0.43461017, Log Avg loss: 0.35000380, Global Avg Loss: 1.12540837, Time: 0.0075 Steps: 104720, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000409, Sample Num: 6544, Cur Loss: 0.36591062, Cur Avg Loss: 0.43866751, Log Avg loss: 0.60055529, Global Avg Loss: 1.12535826, Time: 0.0076 Steps: 104730, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000419, Sample Num: 6704, Cur Loss: 0.17621173, Cur Avg Loss: 0.44060844, Log Avg loss: 0.51999278, Global Avg Loss: 1.12530046, Time: 0.0076 Steps: 104740, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000429, Sample Num: 6864, Cur Loss: 0.29112601, Cur Avg Loss: 0.44144520, Log Avg loss: 0.47650518, Global Avg Loss: 1.12523852, Time: 0.0075 Steps: 104750, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000439, Sample Num: 7024, Cur Loss: 0.34776220, Cur Avg Loss: 0.43878765, Log Avg loss: 0.32477890, Global Avg Loss: 1.12516212, Time: 0.0075 Steps: 104760, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000449, Sample Num: 7184, Cur Loss: 0.53654140, Cur Avg Loss: 0.43957099, Log Avg loss: 0.47395971, Global Avg Loss: 1.12509996, Time: 0.0075 Steps: 104770, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000459, Sample Num: 7344, Cur Loss: 0.17340373, Cur Avg Loss: 0.44058646, Log Avg loss: 0.48618077, Global Avg Loss: 1.12503898, Time: 0.0076 Steps: 104780, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000469, Sample Num: 7504, Cur Loss: 0.21687932, Cur Avg Loss: 0.44174505, Log Avg loss: 0.49492453, Global Avg Loss: 1.12497885, Time: 0.0078 Steps: 104790, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000479, Sample Num: 7664, Cur Loss: 0.44623792, Cur Avg Loss: 0.44482288, Log Avg loss: 0.58917317, Global Avg Loss: 1.12492773, Time: 0.0075 Steps: 104800, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000489, Sample Num: 7824, Cur Loss: 0.79498971, Cur Avg Loss: 0.44352021, Log Avg loss: 0.38112232, Global Avg Loss: 1.12485676, Time: 0.0075 Steps: 104810, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000499, Sample Num: 7984, Cur Loss: 0.32221648, Cur Avg Loss: 0.44226101, Log Avg loss: 0.38068590, Global Avg Loss: 1.12478576, Time: 0.0076 Steps: 104820, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000509, Sample Num: 8144, Cur Loss: 0.74384844, Cur Avg Loss: 0.44241113, Log Avg loss: 0.44990208, Global Avg Loss: 1.12472138, Time: 0.0075 Steps: 104830, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000519, Sample Num: 8304, Cur Loss: 0.17918541, Cur Avg Loss: 0.43909655, Log Avg loss: 0.27038459, Global Avg Loss: 1.12463989, Time: 0.0076 Steps: 104840, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000529, Sample Num: 8464, Cur Loss: 0.48536575, Cur Avg Loss: 0.43805705, Log Avg loss: 0.38410686, Global Avg Loss: 1.12456927, Time: 0.0075 Steps: 104850, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000539, Sample Num: 8624, Cur Loss: 0.44843233, Cur Avg Loss: 0.43945097, Log Avg loss: 0.51318933, Global Avg Loss: 1.12451096, Time: 0.0076 Steps: 104860, Updated lr: 0.000002 Training, Epoch: 0050, Batch: 000549, Sample Num: 8784, Cur Loss: 0.26078176, Cur Avg Loss: 0.44066760, Log Avg loss: 0.50624431, Global Avg Loss: 1.12445201, Time: 0.0076 Steps: 104870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000559, Sample Num: 8944, Cur Loss: 0.83090812, Cur Avg Loss: 0.44104732, Log Avg loss: 0.46189368, Global Avg Loss: 1.12438883, Time: 0.0075 Steps: 104880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000569, Sample Num: 9104, Cur Loss: 0.64329469, Cur Avg Loss: 0.43915508, Log Avg loss: 0.33337892, Global Avg Loss: 1.12431342, Time: 0.0076 Steps: 104890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000579, Sample Num: 9264, Cur Loss: 0.33538315, Cur Avg Loss: 0.43771162, Log Avg loss: 0.35557867, Global Avg Loss: 1.12424014, Time: 0.0077 Steps: 104900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000589, Sample Num: 9424, Cur Loss: 0.47271019, Cur Avg Loss: 0.43743111, Log Avg loss: 0.42118994, Global Avg Loss: 1.12417312, Time: 0.0075 Steps: 104910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000599, Sample Num: 9584, Cur Loss: 1.03407836, Cur Avg Loss: 0.43808459, Log Avg loss: 0.47657418, Global Avg Loss: 1.12411140, Time: 0.0075 Steps: 104920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000609, Sample Num: 9744, Cur Loss: 0.11624175, Cur Avg Loss: 0.43977109, Log Avg loss: 0.54079250, Global Avg Loss: 1.12405581, Time: 0.0076 Steps: 104930, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000619, Sample Num: 9904, Cur Loss: 0.49592453, Cur Avg Loss: 0.43960804, Log Avg loss: 0.42967867, Global Avg Loss: 1.12398964, Time: 0.0075 Steps: 104940, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000629, Sample Num: 10064, Cur Loss: 0.41189605, Cur Avg Loss: 0.43868729, Log Avg loss: 0.38169280, Global Avg Loss: 1.12391891, Time: 0.0075 Steps: 104950, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000639, Sample Num: 10224, Cur Loss: 0.48075256, Cur Avg Loss: 0.44127730, Log Avg loss: 0.60418904, Global Avg Loss: 1.12386939, Time: 0.0075 Steps: 104960, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000649, Sample Num: 10384, Cur Loss: 0.26666144, Cur Avg Loss: 0.44092824, Log Avg loss: 0.41862315, Global Avg Loss: 1.12380221, Time: 0.0076 Steps: 104970, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000659, Sample Num: 10544, Cur Loss: 0.21036637, Cur Avg Loss: 0.44192534, Log Avg loss: 0.50663675, Global Avg Loss: 1.12374342, Time: 0.0076 Steps: 104980, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000669, Sample Num: 10704, Cur Loss: 0.30956289, Cur Avg Loss: 0.44236791, Log Avg loss: 0.47153328, Global Avg Loss: 1.12368130, Time: 0.0076 Steps: 104990, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000679, Sample Num: 10864, Cur Loss: 0.50033927, Cur Avg Loss: 0.44172771, Log Avg loss: 0.39889840, Global Avg Loss: 1.12361227, Time: 0.0076 Steps: 105000, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000689, Sample Num: 11024, Cur Loss: 0.80939656, Cur Avg Loss: 0.44363782, Log Avg loss: 0.57333438, Global Avg Loss: 1.12355987, Time: 0.0075 Steps: 105010, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000699, Sample Num: 11184, Cur Loss: 0.16666743, Cur Avg Loss: 0.44159247, Log Avg loss: 0.30066799, Global Avg Loss: 1.12348151, Time: 0.0076 Steps: 105020, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000709, Sample Num: 11344, Cur Loss: 0.73710954, Cur Avg Loss: 0.44193320, Log Avg loss: 0.46574999, Global Avg Loss: 1.12341889, Time: 0.0075 Steps: 105030, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000719, Sample Num: 11504, Cur Loss: 0.37480175, Cur Avg Loss: 0.44193924, Log Avg loss: 0.44236749, Global Avg Loss: 1.12335405, Time: 0.0076 Steps: 105040, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000729, Sample Num: 11664, Cur Loss: 0.23903778, Cur Avg Loss: 0.44173941, Log Avg loss: 0.42737176, Global Avg Loss: 1.12328780, Time: 0.0075 Steps: 105050, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000739, Sample Num: 11824, Cur Loss: 0.57951242, Cur Avg Loss: 0.44202037, Log Avg loss: 0.46250234, Global Avg Loss: 1.12322490, Time: 0.0075 Steps: 105060, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000749, Sample Num: 11984, Cur Loss: 0.52029222, Cur Avg Loss: 0.43962409, Log Avg loss: 0.26253939, Global Avg Loss: 1.12314299, Time: 0.0076 Steps: 105070, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000759, Sample Num: 12144, Cur Loss: 0.33777186, Cur Avg Loss: 0.43801472, Log Avg loss: 0.31747222, Global Avg Loss: 1.12306632, Time: 0.0075 Steps: 105080, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000769, Sample Num: 12304, Cur Loss: 0.15596813, Cur Avg Loss: 0.43868407, Log Avg loss: 0.48948838, Global Avg Loss: 1.12300603, Time: 0.0077 Steps: 105090, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000779, Sample Num: 12464, Cur Loss: 0.46078813, Cur Avg Loss: 0.44073650, Log Avg loss: 0.59856806, Global Avg Loss: 1.12295613, Time: 0.0075 Steps: 105100, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000789, Sample Num: 12624, Cur Loss: 0.27023324, Cur Avg Loss: 0.44041789, Log Avg loss: 0.41559824, Global Avg Loss: 1.12288883, Time: 0.0077 Steps: 105110, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000799, Sample Num: 12784, Cur Loss: 0.33934236, Cur Avg Loss: 0.43966090, Log Avg loss: 0.37993434, Global Avg Loss: 1.12281816, Time: 0.0075 Steps: 105120, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000809, Sample Num: 12944, Cur Loss: 0.50530624, Cur Avg Loss: 0.44122318, Log Avg loss: 0.56604936, Global Avg Loss: 1.12276520, Time: 0.0074 Steps: 105130, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000819, Sample Num: 13104, Cur Loss: 0.18445168, Cur Avg Loss: 0.44233196, Log Avg loss: 0.53203255, Global Avg Loss: 1.12270901, Time: 0.0074 Steps: 105140, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000829, Sample Num: 13264, Cur Loss: 0.29894549, Cur Avg Loss: 0.44277823, Log Avg loss: 0.47932728, Global Avg Loss: 1.12264782, Time: 0.0074 Steps: 105150, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000839, Sample Num: 13424, Cur Loss: 0.27165365, Cur Avg Loss: 0.44303755, Log Avg loss: 0.46453549, Global Avg Loss: 1.12258524, Time: 0.0076 Steps: 105160, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000849, Sample Num: 13584, Cur Loss: 0.90258259, Cur Avg Loss: 0.44311107, Log Avg loss: 0.44927892, Global Avg Loss: 1.12252122, Time: 0.0075 Steps: 105170, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000859, Sample Num: 13744, Cur Loss: 0.78338289, Cur Avg Loss: 0.44352720, Log Avg loss: 0.47885715, Global Avg Loss: 1.12246002, Time: 0.0075 Steps: 105180, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000869, Sample Num: 13904, Cur Loss: 0.45593232, Cur Avg Loss: 0.44387512, Log Avg loss: 0.47376121, Global Avg Loss: 1.12239835, Time: 0.0075 Steps: 105190, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000879, Sample Num: 14064, Cur Loss: 0.21816076, Cur Avg Loss: 0.44347687, Log Avg loss: 0.40886905, Global Avg Loss: 1.12233053, Time: 0.0077 Steps: 105200, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000889, Sample Num: 14224, Cur Loss: 0.28744829, Cur Avg Loss: 0.44287027, Log Avg loss: 0.38954963, Global Avg Loss: 1.12226088, Time: 0.0075 Steps: 105210, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000899, Sample Num: 14384, Cur Loss: 0.24506970, Cur Avg Loss: 0.44495177, Log Avg loss: 0.62999737, Global Avg Loss: 1.12221410, Time: 0.0077 Steps: 105220, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000909, Sample Num: 14544, Cur Loss: 0.67005938, Cur Avg Loss: 0.44549471, Log Avg loss: 0.49430515, Global Avg Loss: 1.12215442, Time: 0.0075 Steps: 105230, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000919, Sample Num: 14704, Cur Loss: 0.54849082, Cur Avg Loss: 0.44600437, Log Avg loss: 0.49233288, Global Avg Loss: 1.12209458, Time: 0.0076 Steps: 105240, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000929, Sample Num: 14864, Cur Loss: 0.62666118, Cur Avg Loss: 0.44648874, Log Avg loss: 0.49100164, Global Avg Loss: 1.12203462, Time: 0.0075 Steps: 105250, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000939, Sample Num: 15024, Cur Loss: 0.30480790, Cur Avg Loss: 0.44727574, Log Avg loss: 0.52038879, Global Avg Loss: 1.12197746, Time: 0.0077 Steps: 105260, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000949, Sample Num: 15184, Cur Loss: 0.27327350, Cur Avg Loss: 0.44674097, Log Avg loss: 0.39652609, Global Avg Loss: 1.12190855, Time: 0.0075 Steps: 105270, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000959, Sample Num: 15344, Cur Loss: 0.54857683, Cur Avg Loss: 0.44791113, Log Avg loss: 0.55895908, Global Avg Loss: 1.12185507, Time: 0.0075 Steps: 105280, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000969, Sample Num: 15504, Cur Loss: 0.38279682, Cur Avg Loss: 0.44766957, Log Avg loss: 0.42450386, Global Avg Loss: 1.12178884, Time: 0.0074 Steps: 105290, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000979, Sample Num: 15664, Cur Loss: 0.21092233, Cur Avg Loss: 0.44853166, Log Avg loss: 0.53206841, Global Avg Loss: 1.12173284, Time: 0.0075 Steps: 105300, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000989, Sample Num: 15824, Cur Loss: 0.43287516, Cur Avg Loss: 0.44819814, Log Avg loss: 0.41554592, Global Avg Loss: 1.12166578, Time: 0.0075 Steps: 105310, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 000999, Sample Num: 15984, Cur Loss: 0.40342298, Cur Avg Loss: 0.44714253, Log Avg loss: 0.34274345, Global Avg Loss: 1.12159182, Time: 0.0075 Steps: 105320, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001009, Sample Num: 16144, Cur Loss: 0.32310277, Cur Avg Loss: 0.44645593, Log Avg loss: 0.37786409, Global Avg Loss: 1.12152121, Time: 0.0074 Steps: 105330, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001019, Sample Num: 16304, Cur Loss: 0.24299365, Cur Avg Loss: 0.44546865, Log Avg loss: 0.34585263, Global Avg Loss: 1.12144758, Time: 0.0075 Steps: 105340, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001029, Sample Num: 16464, Cur Loss: 0.36051005, Cur Avg Loss: 0.44583713, Log Avg loss: 0.48338443, Global Avg Loss: 1.12138701, Time: 0.0077 Steps: 105350, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001039, Sample Num: 16624, Cur Loss: 0.63829899, Cur Avg Loss: 0.44683262, Log Avg loss: 0.54926842, Global Avg Loss: 1.12133271, Time: 0.0077 Steps: 105360, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001049, Sample Num: 16784, Cur Loss: 0.32984623, Cur Avg Loss: 0.44590810, Log Avg loss: 0.34985051, Global Avg Loss: 1.12125950, Time: 0.0075 Steps: 105370, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001059, Sample Num: 16944, Cur Loss: 0.15903157, Cur Avg Loss: 0.44561395, Log Avg loss: 0.41475845, Global Avg Loss: 1.12119245, Time: 0.0076 Steps: 105380, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001069, Sample Num: 17104, Cur Loss: 0.60025787, Cur Avg Loss: 0.44576391, Log Avg loss: 0.46164426, Global Avg Loss: 1.12112987, Time: 0.0076 Steps: 105390, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001079, Sample Num: 17264, Cur Loss: 0.15943690, Cur Avg Loss: 0.44535179, Log Avg loss: 0.40129613, Global Avg Loss: 1.12106158, Time: 0.0075 Steps: 105400, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001089, Sample Num: 17424, Cur Loss: 0.21506494, Cur Avg Loss: 0.44485328, Log Avg loss: 0.39106388, Global Avg Loss: 1.12099232, Time: 0.0076 Steps: 105410, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001099, Sample Num: 17584, Cur Loss: 0.27669790, Cur Avg Loss: 0.44563042, Log Avg loss: 0.53026100, Global Avg Loss: 1.12093629, Time: 0.0075 Steps: 105420, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001109, Sample Num: 17744, Cur Loss: 0.18156977, Cur Avg Loss: 0.44467572, Log Avg loss: 0.33975443, Global Avg Loss: 1.12086219, Time: 0.0075 Steps: 105430, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001119, Sample Num: 17904, Cur Loss: 0.24997076, Cur Avg Loss: 0.44414685, Log Avg loss: 0.38549483, Global Avg Loss: 1.12079245, Time: 0.0077 Steps: 105440, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001129, Sample Num: 18064, Cur Loss: 0.18490329, Cur Avg Loss: 0.44372889, Log Avg loss: 0.39695935, Global Avg Loss: 1.12072381, Time: 0.0076 Steps: 105450, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001139, Sample Num: 18224, Cur Loss: 0.35494065, Cur Avg Loss: 0.44428663, Log Avg loss: 0.50725522, Global Avg Loss: 1.12066564, Time: 0.0075 Steps: 105460, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001149, Sample Num: 18384, Cur Loss: 0.75131893, Cur Avg Loss: 0.44414744, Log Avg loss: 0.42829385, Global Avg Loss: 1.12059999, Time: 0.0075 Steps: 105470, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001159, Sample Num: 18544, Cur Loss: 0.28115112, Cur Avg Loss: 0.44387604, Log Avg loss: 0.41269207, Global Avg Loss: 1.12053288, Time: 0.0075 Steps: 105480, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001169, Sample Num: 18704, Cur Loss: 0.40099445, Cur Avg Loss: 0.44342955, Log Avg loss: 0.39168210, Global Avg Loss: 1.12046378, Time: 0.0076 Steps: 105490, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001179, Sample Num: 18864, Cur Loss: 0.26851213, Cur Avg Loss: 0.44397129, Log Avg loss: 0.50730033, Global Avg Loss: 1.12040566, Time: 0.0075 Steps: 105500, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001189, Sample Num: 19024, Cur Loss: 0.59070998, Cur Avg Loss: 0.44411869, Log Avg loss: 0.46149702, Global Avg Loss: 1.12034321, Time: 0.0076 Steps: 105510, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001199, Sample Num: 19184, Cur Loss: 0.37870249, Cur Avg Loss: 0.44454791, Log Avg loss: 0.49558234, Global Avg Loss: 1.12028401, Time: 0.0075 Steps: 105520, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001209, Sample Num: 19344, Cur Loss: 0.37883443, Cur Avg Loss: 0.44495662, Log Avg loss: 0.49396133, Global Avg Loss: 1.12022466, Time: 0.0076 Steps: 105530, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001219, Sample Num: 19504, Cur Loss: 0.56038398, Cur Avg Loss: 0.44481277, Log Avg loss: 0.42742031, Global Avg Loss: 1.12015901, Time: 0.0076 Steps: 105540, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001229, Sample Num: 19664, Cur Loss: 0.43310812, Cur Avg Loss: 0.44598094, Log Avg loss: 0.58838141, Global Avg Loss: 1.12010863, Time: 0.0074 Steps: 105550, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001239, Sample Num: 19824, Cur Loss: 0.72926044, Cur Avg Loss: 0.44627571, Log Avg loss: 0.48250330, Global Avg Loss: 1.12004823, Time: 0.0075 Steps: 105560, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001249, Sample Num: 19984, Cur Loss: 0.44123441, Cur Avg Loss: 0.44594857, Log Avg loss: 0.40541535, Global Avg Loss: 1.11998054, Time: 0.0076 Steps: 105570, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001259, Sample Num: 20144, Cur Loss: 0.70028752, Cur Avg Loss: 0.44657894, Log Avg loss: 0.52531239, Global Avg Loss: 1.11992421, Time: 0.0076 Steps: 105580, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001269, Sample Num: 20304, Cur Loss: 1.16623259, Cur Avg Loss: 0.44837960, Log Avg loss: 0.67508307, Global Avg Loss: 1.11988208, Time: 0.0077 Steps: 105590, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001279, Sample Num: 20464, Cur Loss: 0.16956034, Cur Avg Loss: 0.44808812, Log Avg loss: 0.41109870, Global Avg Loss: 1.11981496, Time: 0.0075 Steps: 105600, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001289, Sample Num: 20624, Cur Loss: 0.24091984, Cur Avg Loss: 0.44870715, Log Avg loss: 0.52788111, Global Avg Loss: 1.11975891, Time: 0.0076 Steps: 105610, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001299, Sample Num: 20784, Cur Loss: 0.50508052, Cur Avg Loss: 0.44873507, Log Avg loss: 0.45233450, Global Avg Loss: 1.11969572, Time: 0.0075 Steps: 105620, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001309, Sample Num: 20944, Cur Loss: 0.78510237, Cur Avg Loss: 0.44851206, Log Avg loss: 0.41954258, Global Avg Loss: 1.11962944, Time: 0.0076 Steps: 105630, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001319, Sample Num: 21104, Cur Loss: 0.53286386, Cur Avg Loss: 0.44854309, Log Avg loss: 0.45260508, Global Avg Loss: 1.11956630, Time: 0.0076 Steps: 105640, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001329, Sample Num: 21264, Cur Loss: 0.52944684, Cur Avg Loss: 0.44784655, Log Avg loss: 0.35597218, Global Avg Loss: 1.11949402, Time: 0.0076 Steps: 105650, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001339, Sample Num: 21424, Cur Loss: 1.30368376, Cur Avg Loss: 0.44885496, Log Avg loss: 0.58287310, Global Avg Loss: 1.11944324, Time: 0.0076 Steps: 105660, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001349, Sample Num: 21584, Cur Loss: 0.08863421, Cur Avg Loss: 0.44853395, Log Avg loss: 0.40555026, Global Avg Loss: 1.11937568, Time: 0.0075 Steps: 105670, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001359, Sample Num: 21744, Cur Loss: 0.53326207, Cur Avg Loss: 0.44825311, Log Avg loss: 0.41036872, Global Avg Loss: 1.11930859, Time: 0.0075 Steps: 105680, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001369, Sample Num: 21904, Cur Loss: 0.30922523, Cur Avg Loss: 0.44765412, Log Avg loss: 0.36625145, Global Avg Loss: 1.11923733, Time: 0.0075 Steps: 105690, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001379, Sample Num: 22064, Cur Loss: 0.15211332, Cur Avg Loss: 0.44741973, Log Avg loss: 0.41533166, Global Avg Loss: 1.11917074, Time: 0.0075 Steps: 105700, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001389, Sample Num: 22224, Cur Loss: 0.14258806, Cur Avg Loss: 0.44727022, Log Avg loss: 0.42665209, Global Avg Loss: 1.11910523, Time: 0.0075 Steps: 105710, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001399, Sample Num: 22384, Cur Loss: 0.17273408, Cur Avg Loss: 0.44816806, Log Avg loss: 0.57287879, Global Avg Loss: 1.11905356, Time: 0.0075 Steps: 105720, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001409, Sample Num: 22544, Cur Loss: 1.24822342, Cur Avg Loss: 0.44837469, Log Avg loss: 0.47728151, Global Avg Loss: 1.11899286, Time: 0.0076 Steps: 105730, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001419, Sample Num: 22704, Cur Loss: 0.93056953, Cur Avg Loss: 0.44922523, Log Avg loss: 0.56906630, Global Avg Loss: 1.11894086, Time: 0.0075 Steps: 105740, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001429, Sample Num: 22864, Cur Loss: 0.22741541, Cur Avg Loss: 0.44907021, Log Avg loss: 0.42707326, Global Avg Loss: 1.11887543, Time: 0.0077 Steps: 105750, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001439, Sample Num: 23024, Cur Loss: 0.28356755, Cur Avg Loss: 0.44837574, Log Avg loss: 0.34913570, Global Avg Loss: 1.11880265, Time: 0.0076 Steps: 105760, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001449, Sample Num: 23184, Cur Loss: 0.78226042, Cur Avg Loss: 0.44840953, Log Avg loss: 0.45327218, Global Avg Loss: 1.11873973, Time: 0.0076 Steps: 105770, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001459, Sample Num: 23344, Cur Loss: 0.40770009, Cur Avg Loss: 0.44937227, Log Avg loss: 0.58887315, Global Avg Loss: 1.11868963, Time: 0.0075 Steps: 105780, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001469, Sample Num: 23504, Cur Loss: 0.20959374, Cur Avg Loss: 0.44980710, Log Avg loss: 0.51324932, Global Avg Loss: 1.11863240, Time: 0.0076 Steps: 105790, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001479, Sample Num: 23664, Cur Loss: 0.50550199, Cur Avg Loss: 0.44986967, Log Avg loss: 0.45906116, Global Avg Loss: 1.11857006, Time: 0.0075 Steps: 105800, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001489, Sample Num: 23824, Cur Loss: 0.94155437, Cur Avg Loss: 0.45009543, Log Avg loss: 0.48348504, Global Avg Loss: 1.11851004, Time: 0.0076 Steps: 105810, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001499, Sample Num: 23984, Cur Loss: 0.23302525, Cur Avg Loss: 0.45000300, Log Avg loss: 0.43623965, Global Avg Loss: 1.11844557, Time: 0.0075 Steps: 105820, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001509, Sample Num: 24144, Cur Loss: 0.18146680, Cur Avg Loss: 0.44983470, Log Avg loss: 0.42460777, Global Avg Loss: 1.11838001, Time: 0.0084 Steps: 105830, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001519, Sample Num: 24304, Cur Loss: 0.65962362, Cur Avg Loss: 0.44990040, Log Avg loss: 0.45981316, Global Avg Loss: 1.11831778, Time: 0.0132 Steps: 105840, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001529, Sample Num: 24464, Cur Loss: 0.12576877, Cur Avg Loss: 0.45008591, Log Avg loss: 0.47826511, Global Avg Loss: 1.11825731, Time: 0.0231 Steps: 105850, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001539, Sample Num: 24624, Cur Loss: 0.50899869, Cur Avg Loss: 0.45014191, Log Avg loss: 0.45870458, Global Avg Loss: 1.11819501, Time: 0.0074 Steps: 105860, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001549, Sample Num: 24784, Cur Loss: 0.19818503, Cur Avg Loss: 0.45020187, Log Avg loss: 0.45942980, Global Avg Loss: 1.11813279, Time: 0.0074 Steps: 105870, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001559, Sample Num: 24944, Cur Loss: 0.44443721, Cur Avg Loss: 0.45035886, Log Avg loss: 0.47467626, Global Avg Loss: 1.11807201, Time: 0.0074 Steps: 105880, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001569, Sample Num: 25104, Cur Loss: 0.46859890, Cur Avg Loss: 0.44965518, Log Avg loss: 0.33995223, Global Avg Loss: 1.11799853, Time: 0.0074 Steps: 105890, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001579, Sample Num: 25264, Cur Loss: 0.96260369, Cur Avg Loss: 0.44927307, Log Avg loss: 0.38932022, Global Avg Loss: 1.11792972, Time: 0.0074 Steps: 105900, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001589, Sample Num: 25424, Cur Loss: 0.23366654, Cur Avg Loss: 0.44966728, Log Avg loss: 0.51191207, Global Avg Loss: 1.11787250, Time: 0.0074 Steps: 105910, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001599, Sample Num: 25584, Cur Loss: 0.37971058, Cur Avg Loss: 0.45011928, Log Avg loss: 0.52194229, Global Avg Loss: 1.11781624, Time: 0.0074 Steps: 105920, Updated lr: 0.000001 Training, Epoch: 0050, Batch: 001609, Sample Num: 25744, Cur Loss: 0.12288868, Cur Avg Loss: 0.44923670, Log Avg loss: 0.30811265, Global Avg Loss: 1.11773980, Time: 0.0073 Steps: 105930, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001619, Sample Num: 25904, Cur Loss: 0.23800734, Cur Avg Loss: 0.44900987, Log Avg loss: 0.41251260, Global Avg Loss: 1.11767323, Time: 0.0074 Steps: 105940, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001629, Sample Num: 26064, Cur Loss: 0.33809650, Cur Avg Loss: 0.44884318, Log Avg loss: 0.42185602, Global Avg Loss: 1.11760756, Time: 0.0074 Steps: 105950, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001639, Sample Num: 26224, Cur Loss: 0.17417657, Cur Avg Loss: 0.44987125, Log Avg loss: 0.61734367, Global Avg Loss: 1.11756035, Time: 0.0074 Steps: 105960, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001649, Sample Num: 26384, Cur Loss: 0.54420668, Cur Avg Loss: 0.45044971, Log Avg loss: 0.54525941, Global Avg Loss: 1.11750634, Time: 0.0074 Steps: 105970, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001659, Sample Num: 26544, Cur Loss: 0.29665563, Cur Avg Loss: 0.45054441, Log Avg loss: 0.46616067, Global Avg Loss: 1.11744488, Time: 0.0074 Steps: 105980, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001669, Sample Num: 26704, Cur Loss: 1.37164247, Cur Avg Loss: 0.45140065, Log Avg loss: 0.59345076, Global Avg Loss: 1.11739544, Time: 0.0074 Steps: 105990, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001679, Sample Num: 26864, Cur Loss: 0.13466083, Cur Avg Loss: 0.45137729, Log Avg loss: 0.44747846, Global Avg Loss: 1.11733224, Time: 0.0074 Steps: 106000, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001689, Sample Num: 27024, Cur Loss: 1.29659617, Cur Avg Loss: 0.45147916, Log Avg loss: 0.46858325, Global Avg Loss: 1.11727105, Time: 0.0076 Steps: 106010, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001699, Sample Num: 27184, Cur Loss: 0.24914438, Cur Avg Loss: 0.45131399, Log Avg loss: 0.42341705, Global Avg Loss: 1.11720560, Time: 0.0075 Steps: 106020, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001709, Sample Num: 27344, Cur Loss: 0.56949145, Cur Avg Loss: 0.45099054, Log Avg loss: 0.39603687, Global Avg Loss: 1.11713759, Time: 0.0075 Steps: 106030, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001719, Sample Num: 27504, Cur Loss: 0.16873007, Cur Avg Loss: 0.45038377, Log Avg loss: 0.34668528, Global Avg Loss: 1.11706493, Time: 0.0076 Steps: 106040, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001729, Sample Num: 27664, Cur Loss: 0.20437649, Cur Avg Loss: 0.44967346, Log Avg loss: 0.32757183, Global Avg Loss: 1.11699048, Time: 0.0075 Steps: 106050, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001739, Sample Num: 27824, Cur Loss: 0.13559932, Cur Avg Loss: 0.45049225, Log Avg loss: 0.59206171, Global Avg Loss: 1.11694099, Time: 0.0075 Steps: 106060, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001749, Sample Num: 27984, Cur Loss: 0.35902137, Cur Avg Loss: 0.44976215, Log Avg loss: 0.32279665, Global Avg Loss: 1.11686612, Time: 0.0076 Steps: 106070, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001759, Sample Num: 28144, Cur Loss: 0.26802194, Cur Avg Loss: 0.45092521, Log Avg loss: 0.65434564, Global Avg Loss: 1.11682252, Time: 0.0075 Steps: 106080, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001769, Sample Num: 28304, Cur Loss: 0.40111589, Cur Avg Loss: 0.45152857, Log Avg loss: 0.55765877, Global Avg Loss: 1.11676981, Time: 0.0075 Steps: 106090, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001779, Sample Num: 28464, Cur Loss: 0.47154385, Cur Avg Loss: 0.45144885, Log Avg loss: 0.43734695, Global Avg Loss: 1.11670578, Time: 0.0076 Steps: 106100, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001789, Sample Num: 28624, Cur Loss: 0.37854373, Cur Avg Loss: 0.45182009, Log Avg loss: 0.51786346, Global Avg Loss: 1.11664934, Time: 0.0075 Steps: 106110, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001799, Sample Num: 28784, Cur Loss: 0.47948533, Cur Avg Loss: 0.45176314, Log Avg loss: 0.44157436, Global Avg Loss: 1.11658573, Time: 0.0074 Steps: 106120, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001809, Sample Num: 28944, Cur Loss: 0.40160805, Cur Avg Loss: 0.45127612, Log Avg loss: 0.36366214, Global Avg Loss: 1.11651478, Time: 0.0075 Steps: 106130, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001819, Sample Num: 29104, Cur Loss: 0.44829446, Cur Avg Loss: 0.45099562, Log Avg loss: 0.40025151, Global Avg Loss: 1.11644730, Time: 0.0073 Steps: 106140, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001829, Sample Num: 29264, Cur Loss: 0.28235590, Cur Avg Loss: 0.45126044, Log Avg loss: 0.49943151, Global Avg Loss: 1.11638917, Time: 0.0075 Steps: 106150, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001839, Sample Num: 29424, Cur Loss: 0.16081661, Cur Avg Loss: 0.45093155, Log Avg loss: 0.39077767, Global Avg Loss: 1.11632082, Time: 0.0074 Steps: 106160, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001849, Sample Num: 29584, Cur Loss: 0.62670106, Cur Avg Loss: 0.45041793, Log Avg loss: 0.35596329, Global Avg Loss: 1.11624921, Time: 0.0075 Steps: 106170, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001859, Sample Num: 29744, Cur Loss: 0.25489768, Cur Avg Loss: 0.45063799, Log Avg loss: 0.49132785, Global Avg Loss: 1.11619035, Time: 0.0075 Steps: 106180, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001869, Sample Num: 29904, Cur Loss: 0.27198279, Cur Avg Loss: 0.45045613, Log Avg loss: 0.41664788, Global Avg Loss: 1.11612447, Time: 0.0074 Steps: 106190, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001879, Sample Num: 30064, Cur Loss: 0.27932110, Cur Avg Loss: 0.45023296, Log Avg loss: 0.40852308, Global Avg Loss: 1.11605785, Time: 0.0075 Steps: 106200, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001889, Sample Num: 30224, Cur Loss: 0.16420540, Cur Avg Loss: 0.45104440, Log Avg loss: 0.60351304, Global Avg Loss: 1.11600959, Time: 0.0074 Steps: 106210, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001899, Sample Num: 30384, Cur Loss: 0.38947245, Cur Avg Loss: 0.45069015, Log Avg loss: 0.38377258, Global Avg Loss: 1.11594065, Time: 0.0074 Steps: 106220, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001909, Sample Num: 30544, Cur Loss: 0.32896036, Cur Avg Loss: 0.45052223, Log Avg loss: 0.41863382, Global Avg Loss: 1.11587501, Time: 0.0074 Steps: 106230, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001919, Sample Num: 30704, Cur Loss: 0.69243860, Cur Avg Loss: 0.45090164, Log Avg loss: 0.52333073, Global Avg Loss: 1.11581924, Time: 0.0074 Steps: 106240, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001929, Sample Num: 30864, Cur Loss: 0.64170480, Cur Avg Loss: 0.45134232, Log Avg loss: 0.53590967, Global Avg Loss: 1.11576466, Time: 0.0074 Steps: 106250, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001939, Sample Num: 31024, Cur Loss: 0.19229221, Cur Avg Loss: 0.45168521, Log Avg loss: 0.51782942, Global Avg Loss: 1.11570839, Time: 0.0073 Steps: 106260, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001949, Sample Num: 31184, Cur Loss: 0.79308653, Cur Avg Loss: 0.45206656, Log Avg loss: 0.52601037, Global Avg Loss: 1.11565290, Time: 0.0074 Steps: 106270, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001959, Sample Num: 31344, Cur Loss: 0.22523192, Cur Avg Loss: 0.45211263, Log Avg loss: 0.46109067, Global Avg Loss: 1.11559131, Time: 0.0074 Steps: 106280, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001969, Sample Num: 31504, Cur Loss: 0.45704579, Cur Avg Loss: 0.45226087, Log Avg loss: 0.48130065, Global Avg Loss: 1.11553163, Time: 0.0074 Steps: 106290, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001979, Sample Num: 31664, Cur Loss: 0.73490965, Cur Avg Loss: 0.45203353, Log Avg loss: 0.40727164, Global Avg Loss: 1.11546500, Time: 0.0264 Steps: 106300, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001989, Sample Num: 31824, Cur Loss: 0.59630066, Cur Avg Loss: 0.45195220, Log Avg loss: 0.43585553, Global Avg Loss: 1.11540108, Time: 0.0235 Steps: 106310, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 001999, Sample Num: 31984, Cur Loss: 0.54681921, Cur Avg Loss: 0.45190467, Log Avg loss: 0.44245078, Global Avg Loss: 1.11533778, Time: 0.0066 Steps: 106320, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002009, Sample Num: 32144, Cur Loss: 0.38664162, Cur Avg Loss: 0.45163262, Log Avg loss: 0.39725076, Global Avg Loss: 1.11527025, Time: 0.0068 Steps: 106330, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002019, Sample Num: 32304, Cur Loss: 0.66495526, Cur Avg Loss: 0.45109141, Log Avg loss: 0.34236294, Global Avg Loss: 1.11519756, Time: 0.0115 Steps: 106340, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002029, Sample Num: 32464, Cur Loss: 0.40441832, Cur Avg Loss: 0.45182837, Log Avg loss: 0.60062095, Global Avg Loss: 1.11514918, Time: 0.0066 Steps: 106350, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002039, Sample Num: 32624, Cur Loss: 0.35760972, Cur Avg Loss: 0.45206874, Log Avg loss: 0.50083952, Global Avg Loss: 1.11509142, Time: 0.0068 Steps: 106360, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002049, Sample Num: 32784, Cur Loss: 0.43040061, Cur Avg Loss: 0.45203028, Log Avg loss: 0.44418864, Global Avg Loss: 1.11502835, Time: 0.0076 Steps: 106370, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002059, Sample Num: 32944, Cur Loss: 0.20125589, Cur Avg Loss: 0.45173653, Log Avg loss: 0.39154522, Global Avg Loss: 1.11496034, Time: 0.0076 Steps: 106380, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002069, Sample Num: 33104, Cur Loss: 0.08595159, Cur Avg Loss: 0.45094866, Log Avg loss: 0.28872762, Global Avg Loss: 1.11488268, Time: 0.0075 Steps: 106390, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002079, Sample Num: 33264, Cur Loss: 0.38260412, Cur Avg Loss: 0.45116984, Log Avg loss: 0.49693231, Global Avg Loss: 1.11482460, Time: 0.0076 Steps: 106400, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002089, Sample Num: 33424, Cur Loss: 0.50126481, Cur Avg Loss: 0.45059570, Log Avg loss: 0.33123177, Global Avg Loss: 1.11475096, Time: 0.0077 Steps: 106410, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002099, Sample Num: 33584, Cur Loss: 0.10636587, Cur Avg Loss: 0.45068046, Log Avg loss: 0.46838568, Global Avg Loss: 1.11469022, Time: 0.0077 Steps: 106420, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002109, Sample Num: 33744, Cur Loss: 0.13749903, Cur Avg Loss: 0.45011818, Log Avg loss: 0.33209710, Global Avg Loss: 1.11461669, Time: 0.0076 Steps: 106430, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002119, Sample Num: 33904, Cur Loss: 0.51624280, Cur Avg Loss: 0.45045353, Log Avg loss: 0.52117833, Global Avg Loss: 1.11456094, Time: 0.0076 Steps: 106440, Updated lr: 0.000000 Training, Epoch: 0050, Batch: 002129, Sample Num: 34055, Cur Loss: 0.62390184, Cur Avg Loss: 0.45075653, Log Avg loss: 0.51496165, Global Avg Loss: 1.11450461, Time: 0.0042 Steps: 106450, Updated lr: 0.000000 ***** Running evaluation checkpoint-106450 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## ***** Running testing checkpoint-106450 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## Epoch Time: 16.216264, Avg time per batch (s): 0.010000 {"eval_avg_loss": 0.567328, "eval_total_loss": 398.831656, "eval_mae": 0.568587, "eval_mse": 0.567453, "eval_r2": 0.639289, "eval_sp_statistic": 0.756635, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.801548, "eval_ps_pvalue": 0.0, "update_flag": false, "test_avg_loss": 0.921566, "test_total_loss": 462.626176, "test_mae": 0.660738, "test_mse": 0.921841, "test_r2": 0.405035, "test_sp_statistic": 0.606233, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.671476, "test_ps_pvalue": 0.0, "lr": 0.0, "cur_epoch_step": 2129, "train_global_avg_loss": 1.1145046133208913, "train_cur_epoch_loss": 959.6606505289674, "train_cur_epoch_avg_loss": 0.45075652913525943, "train_cur_epoch_time": 16.216264486312866, "train_cur_epoch_avg_time": 0.007616845695778707, "epoch": 50, "step": 106450} ################################################## #########################Best Metric######################### {"epoch": 46, "global_step": 97934, "eval_avg_loss": 0.564394, "eval_total_loss": 396.76914, "eval_mae": 0.553743, "eval_mse": 0.564553, "eval_r2": 0.641133, "eval_sp_statistic": 0.756978, "eval_sp_pvalue": 0.0, "eval_ps_statistic": 0.800871, "eval_ps_pvalue": 0.0, "update_flag": true, "test_avg_loss": 0.912234, "test_total_loss": 457.941222, "test_mae": 0.659774, "test_mse": 0.912499, "test_r2": 0.411065, "test_sp_statistic": 0.602648, "test_sp_pvalue": 0.0, "test_ps_statistic": 0.672431, "test_ps_pvalue": 0.0} ################################################## Total Time: 18966.297788, Avg time per epoch(50 epochs): 379.330000 ++++++++++++Validation+++++++++++++ best sp_statistic global step: 97934 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250409154052/checkpoint-97934 ***** Running evaluation checkpoint-97934 ***** Dev Dataset Instantaneous batch size per GPU = 16 Dev Dataset Num examples = 11243 ################################################## {"evaluation_avg_loss_97934": 0.564394, "evaluation_total_loss_97934": 396.76914, "evaluation_mae_97934": 0.553743, "evaluation_mse_97934": 0.564553, "evaluation_r2_97934": 0.641133, "evaluation_sp_statistic_97934": 0.756978, "evaluation_sp_pvalue_97934": 0.0, "evaluation_ps_statistic_97934": 0.800871, "evaluation_ps_pvalue_97934": 0.0} ++++++++++++Testing+++++++++++++ best sp_statistic global step: 97934 checkpoint path: ../models/DMS_Bind_Reps_Strain/protein/regression/luca_base/matrix/20250409154052/checkpoint-97934 ***** Running testing checkpoint-97934 ***** Test Dataset Instantaneous batch size per GPU = 16 Test Dataset Num examples = [8029] ################################################## {"evaluation_avg_loss_97934": 0.912234, "evaluation_total_loss_97934": 457.941222, "evaluation_mae_97934": 0.659774, "evaluation_mse_97934": 0.912499, "evaluation_r2_97934": 0.411065, "evaluation_sp_statistic_97934": 0.602648, "evaluation_sp_pvalue_97934": 0.0, "evaluation_ps_statistic_97934": 0.672431, "evaluation_ps_pvalue_97934": 0.0}