| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.926829268292683, | |
| "global_step": 840, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.9326, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.899180437875726e-06, | |
| "loss": 0.9014, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 6.1800547775785244e-06, | |
| "loss": 0.8828, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 7.798360875751452e-06, | |
| "loss": 0.8145, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 9.053616605738856e-06, | |
| "loss": 0.7803, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.007923521545425e-05, | |
| "loss": 0.7842, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.0946383394261144e-05, | |
| "loss": 0.7754, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.169754131362718e-05, | |
| "loss": 0.7061, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.2360109555157049e-05, | |
| "loss": 0.7168, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.2952797043614585e-05, | |
| "loss": 0.7148, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.348894809355931e-05, | |
| "loss": 0.7031, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.3978415653329978e-05, | |
| "loss": 0.7236, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.4428682160514114e-05, | |
| "loss": 0.7021, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.484556383213687e-05, | |
| "loss": 0.6875, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.5233671383317381e-05, | |
| "loss": 0.7129, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.5596721751502905e-05, | |
| "loss": 0.6943, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.593775515114726e-05, | |
| "loss": 0.6836, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.6259289993032772e-05, | |
| "loss": 0.668, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.6563435861933304e-05, | |
| "loss": 0.6621, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.685197748149031e-05, | |
| "loss": 0.6455, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.7126438171839668e-05, | |
| "loss": 0.6631, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.7388128531435038e-05, | |
| "loss": 0.6348, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.7638184288576362e-05, | |
| "loss": 0.6602, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.7877596091205705e-05, | |
| "loss": 0.6006, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.810723321147771e-05, | |
| "loss": 0.6777, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.832786259838984e-05, | |
| "loss": 0.6318, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.8540164332735572e-05, | |
| "loss": 0.6406, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.8744744270012596e-05, | |
| "loss": 0.6396, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.8942144463773483e-05, | |
| "loss": 0.6348, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9132851821193106e-05, | |
| "loss": 0.6484, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9317305338856604e-05, | |
| "loss": 0.6133, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9495902189378633e-05, | |
| "loss": 0.6211, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9669002871137834e-05, | |
| "loss": 0.6299, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9836935589022987e-05, | |
| "loss": 0.6162, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9999999999999998e-05, | |
| "loss": 0.6201, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 2e-05, | |
| "loss": 0.6279, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9975786924939468e-05, | |
| "loss": 0.6289, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9951573849878937e-05, | |
| "loss": 0.6152, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9927360774818403e-05, | |
| "loss": 0.6318, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.990314769975787e-05, | |
| "loss": 0.6094, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.987893462469734e-05, | |
| "loss": 0.6416, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9854721549636805e-05, | |
| "loss": 0.6279, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9830508474576275e-05, | |
| "loss": 0.6348, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.980629539951574e-05, | |
| "loss": 0.6318, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9782082324455207e-05, | |
| "loss": 0.6348, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9757869249394673e-05, | |
| "loss": 0.6201, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9733656174334143e-05, | |
| "loss": 0.6318, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.970944309927361e-05, | |
| "loss": 0.6367, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9685230024213075e-05, | |
| "loss": 0.6172, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9661016949152545e-05, | |
| "loss": 0.624, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.963680387409201e-05, | |
| "loss": 0.6348, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.961259079903148e-05, | |
| "loss": 0.6182, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9588377723970946e-05, | |
| "loss": 0.6494, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.9564164648910413e-05, | |
| "loss": 0.6025, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.953995157384988e-05, | |
| "loss": 0.6279, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9515738498789348e-05, | |
| "loss": 0.6387, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9491525423728814e-05, | |
| "loss": 0.6064, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9467312348668284e-05, | |
| "loss": 0.6094, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.944309927360775e-05, | |
| "loss": 0.626, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.9418886198547216e-05, | |
| "loss": 0.6094, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.9394673123486686e-05, | |
| "loss": 0.6182, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9370460048426152e-05, | |
| "loss": 0.6016, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9346246973365618e-05, | |
| "loss": 0.6289, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9322033898305087e-05, | |
| "loss": 0.6289, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9297820823244554e-05, | |
| "loss": 0.5928, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.927360774818402e-05, | |
| "loss": 0.6182, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.924939467312349e-05, | |
| "loss": 0.5918, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9225181598062955e-05, | |
| "loss": 0.6221, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9200968523002425e-05, | |
| "loss": 0.6162, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.917675544794189e-05, | |
| "loss": 0.6162, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9152542372881357e-05, | |
| "loss": 0.6445, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9128329297820823e-05, | |
| "loss": 0.6172, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9104116222760293e-05, | |
| "loss": 0.6152, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.907990314769976e-05, | |
| "loss": 0.6084, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9055690072639225e-05, | |
| "loss": 0.6162, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9031476997578695e-05, | |
| "loss": 0.5996, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.900726392251816e-05, | |
| "loss": 0.6064, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.898305084745763e-05, | |
| "loss": 0.5996, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8958837772397097e-05, | |
| "loss": 0.6328, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8934624697336563e-05, | |
| "loss": 0.6133, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.891041162227603e-05, | |
| "loss": 0.6143, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.88861985472155e-05, | |
| "loss": 0.5996, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8861985472154964e-05, | |
| "loss": 0.6182, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.883777239709443e-05, | |
| "loss": 0.623, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.88135593220339e-05, | |
| "loss": 0.6055, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8789346246973366e-05, | |
| "loss": 0.5967, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8765133171912836e-05, | |
| "loss": 0.5947, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.8740920096852302e-05, | |
| "loss": 0.6182, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.8716707021791768e-05, | |
| "loss": 0.6152, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.8692493946731238e-05, | |
| "loss": 0.5977, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.8668280871670704e-05, | |
| "loss": 0.6348, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.864406779661017e-05, | |
| "loss": 0.6074, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.8619854721549636e-05, | |
| "loss": 0.5898, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.8595641646489106e-05, | |
| "loss": 0.5967, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.8571428571428575e-05, | |
| "loss": 0.6113, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.854721549636804e-05, | |
| "loss": 0.6123, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.8523002421307507e-05, | |
| "loss": 0.6211, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.8498789346246974e-05, | |
| "loss": 0.6006, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.8474576271186443e-05, | |
| "loss": 0.6113, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.845036319612591e-05, | |
| "loss": 0.6006, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.8426150121065375e-05, | |
| "loss": 0.6123, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.8401937046004845e-05, | |
| "loss": 0.5938, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.837772397094431e-05, | |
| "loss": 0.6045, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.835351089588378e-05, | |
| "loss": 0.5928, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.8329297820823247e-05, | |
| "loss": 0.5977, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.8305084745762713e-05, | |
| "loss": 0.6299, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.828087167070218e-05, | |
| "loss": 0.5801, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.825665859564165e-05, | |
| "loss": 0.6035, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.8232445520581115e-05, | |
| "loss": 0.5938, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.820823244552058e-05, | |
| "loss": 0.5977, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.818401937046005e-05, | |
| "loss": 0.6016, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.8159806295399516e-05, | |
| "loss": 0.5957, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.8135593220338986e-05, | |
| "loss": 0.6113, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.8111380145278452e-05, | |
| "loss": 0.6094, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.8087167070217918e-05, | |
| "loss": 0.6016, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.8062953995157388e-05, | |
| "loss": 0.6006, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.8038740920096854e-05, | |
| "loss": 0.6182, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.801452784503632e-05, | |
| "loss": 0.6025, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7990314769975786e-05, | |
| "loss": 0.5938, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7966101694915256e-05, | |
| "loss": 0.6172, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7941888619854725e-05, | |
| "loss": 0.6074, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.791767554479419e-05, | |
| "loss": 0.6055, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7893462469733658e-05, | |
| "loss": 0.583, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7869249394673124e-05, | |
| "loss": 0.5918, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7845036319612593e-05, | |
| "loss": 0.6016, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.782082324455206e-05, | |
| "loss": 0.6123, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7796610169491526e-05, | |
| "loss": 0.6104, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.777239709443099e-05, | |
| "loss": 0.5801, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.774818401937046e-05, | |
| "loss": 0.6035, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.772397094430993e-05, | |
| "loss": 0.5811, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.7699757869249397e-05, | |
| "loss": 0.5908, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.7675544794188863e-05, | |
| "loss": 0.6035, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.765133171912833e-05, | |
| "loss": 0.6094, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.76271186440678e-05, | |
| "loss": 0.5957, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.7602905569007265e-05, | |
| "loss": 0.6152, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.757869249394673e-05, | |
| "loss": 0.5908, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.75544794188862e-05, | |
| "loss": 0.5908, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.7530266343825667e-05, | |
| "loss": 0.5957, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.7506053268765136e-05, | |
| "loss": 0.585, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.7481840193704602e-05, | |
| "loss": 0.6143, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.745762711864407e-05, | |
| "loss": 0.6064, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.7433414043583538e-05, | |
| "loss": 0.6025, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.7409200968523004e-05, | |
| "loss": 0.583, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.738498789346247e-05, | |
| "loss": 0.5967, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.7360774818401936e-05, | |
| "loss": 0.5947, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.7336561743341406e-05, | |
| "loss": 0.6123, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.7312348668280875e-05, | |
| "loss": 0.627, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.728813559322034e-05, | |
| "loss": 0.5986, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.7263922518159808e-05, | |
| "loss": 0.623, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.7239709443099274e-05, | |
| "loss": 0.6016, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.7215496368038743e-05, | |
| "loss": 0.6016, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.719128329297821e-05, | |
| "loss": 0.5928, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.7167070217917676e-05, | |
| "loss": 0.582, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.7142857142857142e-05, | |
| "loss": 0.5938, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.711864406779661e-05, | |
| "loss": 0.6016, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.709443099273608e-05, | |
| "loss": 0.584, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.7070217917675547e-05, | |
| "loss": 0.5674, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.7046004842615013e-05, | |
| "loss": 0.5957, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.702179176755448e-05, | |
| "loss": 0.6035, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.699757869249395e-05, | |
| "loss": 0.5879, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.6973365617433415e-05, | |
| "loss": 0.5898, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.694915254237288e-05, | |
| "loss": 0.5732, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.692493946731235e-05, | |
| "loss": 0.5859, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.6900726392251817e-05, | |
| "loss": 0.5908, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.6876513317191286e-05, | |
| "loss": 0.6133, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.6852300242130752e-05, | |
| "loss": 0.5645, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.682808716707022e-05, | |
| "loss": 0.584, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.6803874092009688e-05, | |
| "loss": 0.6035, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.6779661016949154e-05, | |
| "loss": 0.5938, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.675544794188862e-05, | |
| "loss": 0.5664, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.6731234866828087e-05, | |
| "loss": 0.6064, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.6707021791767556e-05, | |
| "loss": 0.5977, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.6682808716707026e-05, | |
| "loss": 0.5889, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.6658595641646492e-05, | |
| "loss": 0.6016, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.6634382566585958e-05, | |
| "loss": 0.5986, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.6610169491525424e-05, | |
| "loss": 0.5908, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.6585956416464894e-05, | |
| "loss": 0.5811, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.656174334140436e-05, | |
| "loss": 0.5791, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.6537530266343826e-05, | |
| "loss": 0.5967, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.6513317191283292e-05, | |
| "loss": 0.6055, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.648910411622276e-05, | |
| "loss": 0.5977, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.646489104116223e-05, | |
| "loss": 0.583, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.6440677966101697e-05, | |
| "loss": 0.6162, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.6416464891041163e-05, | |
| "loss": 0.6016, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.639225181598063e-05, | |
| "loss": 0.5869, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.63680387409201e-05, | |
| "loss": 0.5576, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.6343825665859565e-05, | |
| "loss": 0.5938, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.631961259079903e-05, | |
| "loss": 0.5879, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.62953995157385e-05, | |
| "loss": 0.583, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6271186440677967e-05, | |
| "loss": 0.583, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6246973365617437e-05, | |
| "loss": 0.5781, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6222760290556903e-05, | |
| "loss": 0.5918, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.619854721549637e-05, | |
| "loss": 0.5859, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.617433414043584e-05, | |
| "loss": 0.6152, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6150121065375304e-05, | |
| "loss": 0.582, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.612590799031477e-05, | |
| "loss": 0.5986, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6101694915254237e-05, | |
| "loss": 0.5752, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6077481840193706e-05, | |
| "loss": 0.5791, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6053268765133176e-05, | |
| "loss": 0.5732, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.6029055690072642e-05, | |
| "loss": 0.6094, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.6004842615012108e-05, | |
| "loss": 0.5996, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5980629539951574e-05, | |
| "loss": 0.6006, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5956416464891044e-05, | |
| "loss": 0.6113, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.593220338983051e-05, | |
| "loss": 0.5889, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5907990314769976e-05, | |
| "loss": 0.5781, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.5883777239709442e-05, | |
| "loss": 0.6113, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.5859564164648912e-05, | |
| "loss": 0.5918, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.583535108958838e-05, | |
| "loss": 0.5693, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.5811138014527847e-05, | |
| "loss": 0.5986, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.5786924939467314e-05, | |
| "loss": 0.5625, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.576271186440678e-05, | |
| "loss": 0.5908, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.573849878934625e-05, | |
| "loss": 0.5938, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.5714285714285715e-05, | |
| "loss": 0.6123, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.569007263922518e-05, | |
| "loss": 0.5879, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.566585956416465e-05, | |
| "loss": 0.583, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.5641646489104117e-05, | |
| "loss": 0.6094, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.5617433414043587e-05, | |
| "loss": 0.624, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.5593220338983053e-05, | |
| "loss": 0.5947, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.556900726392252e-05, | |
| "loss": 0.6084, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.554479418886199e-05, | |
| "loss": 0.5908, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.5520581113801455e-05, | |
| "loss": 0.5938, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.549636803874092e-05, | |
| "loss": 0.5625, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.5472154963680387e-05, | |
| "loss": 0.583, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.5447941888619856e-05, | |
| "loss": 0.5938, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.5423728813559326e-05, | |
| "loss": 0.583, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.5399515738498792e-05, | |
| "loss": 0.5918, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.5375302663438258e-05, | |
| "loss": 0.582, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.5351089588377724e-05, | |
| "loss": 0.5996, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.5326876513317194e-05, | |
| "loss": 0.5684, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.530266343825666e-05, | |
| "loss": 0.5996, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.5278450363196126e-05, | |
| "loss": 0.5791, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.5254237288135594e-05, | |
| "loss": 0.5869, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.523002421307506e-05, | |
| "loss": 0.5625, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.520581113801453e-05, | |
| "loss": 0.5771, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.5181598062953998e-05, | |
| "loss": 0.5869, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.5157384987893464e-05, | |
| "loss": 0.5654, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.5133171912832932e-05, | |
| "loss": 0.5605, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.5108958837772398e-05, | |
| "loss": 0.5967, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.5084745762711865e-05, | |
| "loss": 0.5674, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.5060532687651332e-05, | |
| "loss": 0.5889, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.50363196125908e-05, | |
| "loss": 0.5947, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.5012106537530267e-05, | |
| "loss": 0.5801, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.4987893462469735e-05, | |
| "loss": 0.5771, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.4963680387409203e-05, | |
| "loss": 0.6006, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.493946731234867e-05, | |
| "loss": 0.5693, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.4915254237288137e-05, | |
| "loss": 0.5898, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.4891041162227605e-05, | |
| "loss": 0.5918, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.4866828087167071e-05, | |
| "loss": 0.583, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.4842615012106539e-05, | |
| "loss": 0.585, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.4818401937046005e-05, | |
| "loss": 0.5664, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.4794188861985473e-05, | |
| "loss": 0.5771, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.4769975786924942e-05, | |
| "loss": 0.5771, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.4745762711864408e-05, | |
| "loss": 0.6104, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.4721549636803876e-05, | |
| "loss": 0.5732, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.4697336561743342e-05, | |
| "loss": 0.5859, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.467312348668281e-05, | |
| "loss": 0.6025, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.4648910411622276e-05, | |
| "loss": 0.5869, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.4624697336561744e-05, | |
| "loss": 0.6006, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.460048426150121e-05, | |
| "loss": 0.585, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.4576271186440678e-05, | |
| "loss": 0.5859, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.4552058111380148e-05, | |
| "loss": 0.5654, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.4527845036319614e-05, | |
| "loss": 0.5811, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.4503631961259082e-05, | |
| "loss": 0.6006, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.4479418886198548e-05, | |
| "loss": 0.5674, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.4455205811138016e-05, | |
| "loss": 0.5996, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.4430992736077482e-05, | |
| "loss": 0.5889, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.440677966101695e-05, | |
| "loss": 0.585, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.4382566585956417e-05, | |
| "loss": 0.5625, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.4358353510895885e-05, | |
| "loss": 0.5791, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.4334140435835353e-05, | |
| "loss": 0.5869, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.4309927360774821e-05, | |
| "loss": 0.5918, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.4285714285714287e-05, | |
| "loss": 0.5791, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.4261501210653755e-05, | |
| "loss": 0.5674, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.4237288135593221e-05, | |
| "loss": 0.5947, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.4213075060532689e-05, | |
| "loss": 0.6133, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.4188861985472155e-05, | |
| "loss": 0.5986, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.4164648910411623e-05, | |
| "loss": 0.5996, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.4140435835351092e-05, | |
| "loss": 0.5713, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.4116222760290559e-05, | |
| "loss": 0.5762, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.4092009685230026e-05, | |
| "loss": 0.5908, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.4067796610169493e-05, | |
| "loss": 0.5703, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.404358353510896e-05, | |
| "loss": 0.5596, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.4019370460048427e-05, | |
| "loss": 0.5908, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.3995157384987894e-05, | |
| "loss": 0.5674, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.397094430992736e-05, | |
| "loss": 0.5781, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.3946731234866828e-05, | |
| "loss": 0.5762, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.3922518159806298e-05, | |
| "loss": 0.7271, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.3898305084745764e-05, | |
| "loss": 0.3965, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.3874092009685232e-05, | |
| "loss": 0.3896, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.3849878934624698e-05, | |
| "loss": 0.4058, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.3825665859564166e-05, | |
| "loss": 0.3989, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.3801452784503632e-05, | |
| "loss": 0.4155, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.37772397094431e-05, | |
| "loss": 0.4067, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.3753026634382568e-05, | |
| "loss": 0.4009, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.3728813559322034e-05, | |
| "loss": 0.416, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.3704600484261503e-05, | |
| "loss": 0.4058, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.368038740920097e-05, | |
| "loss": 0.3701, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.3656174334140437e-05, | |
| "loss": 0.3926, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.3631961259079905e-05, | |
| "loss": 0.3955, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.3607748184019371e-05, | |
| "loss": 0.3755, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.3583535108958839e-05, | |
| "loss": 0.3726, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.3559322033898305e-05, | |
| "loss": 0.3911, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.3535108958837773e-05, | |
| "loss": 0.3706, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.351089588377724e-05, | |
| "loss": 0.3696, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.3486682808716709e-05, | |
| "loss": 0.3804, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.3462469733656177e-05, | |
| "loss": 0.373, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.3438256658595643e-05, | |
| "loss": 0.3667, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.341404358353511e-05, | |
| "loss": 0.3643, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.3389830508474577e-05, | |
| "loss": 0.3457, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.3365617433414045e-05, | |
| "loss": 0.3525, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.334140435835351e-05, | |
| "loss": 0.356, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.3317191283292979e-05, | |
| "loss": 0.3545, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.3292978208232448e-05, | |
| "loss": 0.353, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.3268765133171914e-05, | |
| "loss": 0.3447, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.3244552058111382e-05, | |
| "loss": 0.3569, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.3220338983050848e-05, | |
| "loss": 0.3545, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.3196125907990316e-05, | |
| "loss": 0.3438, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.3171912832929782e-05, | |
| "loss": 0.3423, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.314769975786925e-05, | |
| "loss": 0.3594, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.3123486682808718e-05, | |
| "loss": 0.3442, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3099273607748184e-05, | |
| "loss": 0.3662, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3075060532687653e-05, | |
| "loss": 0.3584, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.305084745762712e-05, | |
| "loss": 0.3511, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.3026634382566587e-05, | |
| "loss": 0.3711, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.3002421307506055e-05, | |
| "loss": 0.3486, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.2978208232445521e-05, | |
| "loss": 0.355, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.295399515738499e-05, | |
| "loss": 0.3589, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.2929782082324455e-05, | |
| "loss": 0.3447, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.2905569007263923e-05, | |
| "loss": 0.3442, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.288135593220339e-05, | |
| "loss": 0.335, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.2857142857142859e-05, | |
| "loss": 0.3369, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.2832929782082327e-05, | |
| "loss": 0.3467, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.2808716707021793e-05, | |
| "loss": 0.3428, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.278450363196126e-05, | |
| "loss": 0.3452, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.2760290556900727e-05, | |
| "loss": 0.3428, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.2736077481840195e-05, | |
| "loss": 0.3525, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.2711864406779661e-05, | |
| "loss": 0.3398, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.2687651331719129e-05, | |
| "loss": 0.3477, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.2663438256658595e-05, | |
| "loss": 0.3433, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.2639225181598064e-05, | |
| "loss": 0.3438, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.2615012106537532e-05, | |
| "loss": 0.3442, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.2590799031476998e-05, | |
| "loss": 0.3457, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.2566585956416466e-05, | |
| "loss": 0.3374, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.2542372881355932e-05, | |
| "loss": 0.333, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.25181598062954e-05, | |
| "loss": 0.3486, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.2493946731234868e-05, | |
| "loss": 0.3462, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.2469733656174334e-05, | |
| "loss": 0.3335, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.2445520581113804e-05, | |
| "loss": 0.3379, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.242130750605327e-05, | |
| "loss": 0.3467, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.2397094430992738e-05, | |
| "loss": 0.3481, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.2372881355932205e-05, | |
| "loss": 0.354, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.2348668280871672e-05, | |
| "loss": 0.3433, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.232445520581114e-05, | |
| "loss": 0.3501, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.2300242130750606e-05, | |
| "loss": 0.3369, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.2276029055690073e-05, | |
| "loss": 0.3403, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.225181598062954e-05, | |
| "loss": 0.3535, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.2227602905569009e-05, | |
| "loss": 0.353, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2203389830508477e-05, | |
| "loss": 0.3555, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2179176755447943e-05, | |
| "loss": 0.3389, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2154963680387411e-05, | |
| "loss": 0.3657, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2130750605326877e-05, | |
| "loss": 0.3398, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2106537530266345e-05, | |
| "loss": 0.3374, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2082324455205811e-05, | |
| "loss": 0.3325, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2058111380145279e-05, | |
| "loss": 0.3452, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2033898305084745e-05, | |
| "loss": 0.3281, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2009685230024215e-05, | |
| "loss": 0.3228, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.1985472154963682e-05, | |
| "loss": 0.353, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.1961259079903149e-05, | |
| "loss": 0.3457, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.1937046004842616e-05, | |
| "loss": 0.3506, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.1912832929782082e-05, | |
| "loss": 0.3457, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.188861985472155e-05, | |
| "loss": 0.3413, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.1864406779661018e-05, | |
| "loss": 0.3462, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.1840193704600484e-05, | |
| "loss": 0.3354, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.1815980629539952e-05, | |
| "loss": 0.3452, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.179176755447942e-05, | |
| "loss": 0.3594, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1767554479418888e-05, | |
| "loss": 0.3511, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1743341404358356e-05, | |
| "loss": 0.3521, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1719128329297822e-05, | |
| "loss": 0.3311, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.169491525423729e-05, | |
| "loss": 0.3301, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1670702179176756e-05, | |
| "loss": 0.3564, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1646489104116224e-05, | |
| "loss": 0.3345, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.162227602905569e-05, | |
| "loss": 0.355, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1598062953995158e-05, | |
| "loss": 0.3364, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1573849878934627e-05, | |
| "loss": 0.3335, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1549636803874093e-05, | |
| "loss": 0.3477, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1525423728813561e-05, | |
| "loss": 0.3433, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1501210653753027e-05, | |
| "loss": 0.3467, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1476997578692495e-05, | |
| "loss": 0.3462, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1452784503631961e-05, | |
| "loss": 0.3354, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1428571428571429e-05, | |
| "loss": 0.334, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1404358353510895e-05, | |
| "loss": 0.3354, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1380145278450365e-05, | |
| "loss": 0.3501, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1355932203389833e-05, | |
| "loss": 0.3564, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1331719128329299e-05, | |
| "loss": 0.3442, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1307506053268767e-05, | |
| "loss": 0.333, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1283292978208233e-05, | |
| "loss": 0.3428, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.12590799031477e-05, | |
| "loss": 0.3564, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1234866828087168e-05, | |
| "loss": 0.3359, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1210653753026634e-05, | |
| "loss": 0.3335, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1186440677966102e-05, | |
| "loss": 0.3462, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.116222760290557e-05, | |
| "loss": 0.3408, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1138014527845038e-05, | |
| "loss": 0.3335, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1113801452784506e-05, | |
| "loss": 0.3457, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.1089588377723972e-05, | |
| "loss": 0.3457, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.106537530266344e-05, | |
| "loss": 0.3477, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.1041162227602906e-05, | |
| "loss": 0.3564, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.1016949152542374e-05, | |
| "loss": 0.3135, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.099273607748184e-05, | |
| "loss": 0.3438, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0968523002421308e-05, | |
| "loss": 0.3389, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0944309927360777e-05, | |
| "loss": 0.3447, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0920096852300243e-05, | |
| "loss": 0.3428, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0895883777239711e-05, | |
| "loss": 0.3545, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0871670702179177e-05, | |
| "loss": 0.3394, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0847457627118645e-05, | |
| "loss": 0.3442, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0823244552058111e-05, | |
| "loss": 0.3237, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.079903147699758e-05, | |
| "loss": 0.3286, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0774818401937045e-05, | |
| "loss": 0.3296, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0750605326876513e-05, | |
| "loss": 0.3447, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0726392251815983e-05, | |
| "loss": 0.3301, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0702179176755449e-05, | |
| "loss": 0.3398, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0677966101694917e-05, | |
| "loss": 0.3452, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0653753026634383e-05, | |
| "loss": 0.3608, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.062953995157385e-05, | |
| "loss": 0.3345, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0605326876513318e-05, | |
| "loss": 0.3291, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0581113801452785e-05, | |
| "loss": 0.3667, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0556900726392252e-05, | |
| "loss": 0.3594, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0532687651331719e-05, | |
| "loss": 0.3579, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0508474576271188e-05, | |
| "loss": 0.3345, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0484261501210656e-05, | |
| "loss": 0.3281, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0460048426150122e-05, | |
| "loss": 0.3403, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.043583535108959e-05, | |
| "loss": 0.3604, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.0411622276029056e-05, | |
| "loss": 0.3335, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.0387409200968524e-05, | |
| "loss": 0.3462, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.036319612590799e-05, | |
| "loss": 0.3423, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0338983050847458e-05, | |
| "loss": 0.3364, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0314769975786927e-05, | |
| "loss": 0.3496, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0290556900726394e-05, | |
| "loss": 0.3325, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0266343825665861e-05, | |
| "loss": 0.3413, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0242130750605328e-05, | |
| "loss": 0.335, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0217917675544795e-05, | |
| "loss": 0.3423, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.0193704600484262e-05, | |
| "loss": 0.3335, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.016949152542373e-05, | |
| "loss": 0.3379, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.0145278450363195e-05, | |
| "loss": 0.3506, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.0121065375302663e-05, | |
| "loss": 0.3457, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.0096852300242133e-05, | |
| "loss": 0.334, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.0072639225181599e-05, | |
| "loss": 0.3408, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.0048426150121067e-05, | |
| "loss": 0.3525, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.0024213075060533e-05, | |
| "loss": 0.3491, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1e-05, | |
| "loss": 0.3286, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.975786924939469e-06, | |
| "loss": 0.3159, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.951573849878935e-06, | |
| "loss": 0.3252, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.927360774818403e-06, | |
| "loss": 0.335, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.90314769975787e-06, | |
| "loss": 0.3481, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.878934624697337e-06, | |
| "loss": 0.3403, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.854721549636804e-06, | |
| "loss": 0.3467, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.830508474576272e-06, | |
| "loss": 0.3364, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.80629539951574e-06, | |
| "loss": 0.3589, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.782082324455206e-06, | |
| "loss": 0.3408, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.757869249394674e-06, | |
| "loss": 0.3359, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.733656174334142e-06, | |
| "loss": 0.3403, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.709443099273608e-06, | |
| "loss": 0.3462, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.685230024213076e-06, | |
| "loss": 0.3335, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.661016949152544e-06, | |
| "loss": 0.3516, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.63680387409201e-06, | |
| "loss": 0.3638, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.612590799031478e-06, | |
| "loss": 0.3364, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.588377723970946e-06, | |
| "loss": 0.3364, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.564164648910412e-06, | |
| "loss": 0.3496, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.53995157384988e-06, | |
| "loss": 0.3403, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.515738498789347e-06, | |
| "loss": 0.3481, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.491525423728815e-06, | |
| "loss": 0.3374, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.467312348668281e-06, | |
| "loss": 0.3174, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.44309927360775e-06, | |
| "loss": 0.3633, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.418886198547215e-06, | |
| "loss": 0.3472, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.394673123486683e-06, | |
| "loss": 0.3379, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.370460048426151e-06, | |
| "loss": 0.3428, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.346246973365619e-06, | |
| "loss": 0.3369, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.322033898305085e-06, | |
| "loss": 0.354, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.297820823244553e-06, | |
| "loss": 0.3467, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.27360774818402e-06, | |
| "loss": 0.3379, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.249394673123487e-06, | |
| "loss": 0.3584, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.225181598062955e-06, | |
| "loss": 0.3335, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.200968523002422e-06, | |
| "loss": 0.3247, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.17675544794189e-06, | |
| "loss": 0.3403, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.152542372881356e-06, | |
| "loss": 0.3442, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.128329297820824e-06, | |
| "loss": 0.334, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.10411622276029e-06, | |
| "loss": 0.3369, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.079903147699758e-06, | |
| "loss": 0.3496, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.055690072639226e-06, | |
| "loss": 0.3477, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.031476997578694e-06, | |
| "loss": 0.3447, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.00726392251816e-06, | |
| "loss": 0.3286, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.983050847457628e-06, | |
| "loss": 0.3306, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.958837772397096e-06, | |
| "loss": 0.3379, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.934624697336562e-06, | |
| "loss": 0.3257, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.91041162227603e-06, | |
| "loss": 0.3447, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.886198547215496e-06, | |
| "loss": 0.3237, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.861985472154965e-06, | |
| "loss": 0.3296, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.837772397094432e-06, | |
| "loss": 0.3452, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 8.8135593220339e-06, | |
| "loss": 0.3257, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 8.789346246973365e-06, | |
| "loss": 0.353, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 8.765133171912833e-06, | |
| "loss": 0.3345, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.740920096852301e-06, | |
| "loss": 0.3374, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.716707021791769e-06, | |
| "loss": 0.3418, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.692493946731235e-06, | |
| "loss": 0.3481, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 8.668280871670703e-06, | |
| "loss": 0.3389, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 8.64406779661017e-06, | |
| "loss": 0.3413, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 8.619854721549637e-06, | |
| "loss": 0.3438, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.595641646489105e-06, | |
| "loss": 0.332, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.571428571428571e-06, | |
| "loss": 0.3687, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.54721549636804e-06, | |
| "loss": 0.3315, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 8.523002421307507e-06, | |
| "loss": 0.3442, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 8.498789346246974e-06, | |
| "loss": 0.3467, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 8.47457627118644e-06, | |
| "loss": 0.3325, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 8.450363196125908e-06, | |
| "loss": 0.3213, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 8.426150121065376e-06, | |
| "loss": 0.3325, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 8.401937046004844e-06, | |
| "loss": 0.334, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 8.37772397094431e-06, | |
| "loss": 0.3418, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 8.353510895883778e-06, | |
| "loss": 0.3311, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 8.329297820823246e-06, | |
| "loss": 0.3394, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 8.305084745762712e-06, | |
| "loss": 0.3306, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 8.28087167070218e-06, | |
| "loss": 0.3506, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 8.256658595641646e-06, | |
| "loss": 0.3291, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 8.232445520581116e-06, | |
| "loss": 0.353, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 8.208232445520582e-06, | |
| "loss": 0.354, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 8.18401937046005e-06, | |
| "loss": 0.3364, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 8.159806295399516e-06, | |
| "loss": 0.3394, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 8.135593220338983e-06, | |
| "loss": 0.3457, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 8.111380145278451e-06, | |
| "loss": 0.3296, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 8.08716707021792e-06, | |
| "loss": 0.3569, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 8.062953995157385e-06, | |
| "loss": 0.3389, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 8.038740920096853e-06, | |
| "loss": 0.3394, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 8.014527845036321e-06, | |
| "loss": 0.3364, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 7.990314769975787e-06, | |
| "loss": 0.3467, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.966101694915255e-06, | |
| "loss": 0.3462, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.941888619854721e-06, | |
| "loss": 0.334, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.91767554479419e-06, | |
| "loss": 0.3501, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.893462469733657e-06, | |
| "loss": 0.3296, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.869249394673125e-06, | |
| "loss": 0.3438, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.84503631961259e-06, | |
| "loss": 0.3306, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 7.820823244552059e-06, | |
| "loss": 0.3491, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 7.796610169491526e-06, | |
| "loss": 0.3286, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 7.772397094430994e-06, | |
| "loss": 0.356, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 7.74818401937046e-06, | |
| "loss": 0.3535, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 7.723970944309928e-06, | |
| "loss": 0.3345, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 7.699757869249396e-06, | |
| "loss": 0.3398, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 7.675544794188862e-06, | |
| "loss": 0.3296, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 7.65133171912833e-06, | |
| "loss": 0.3496, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 7.627118644067797e-06, | |
| "loss": 0.3301, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 7.602905569007265e-06, | |
| "loss": 0.3574, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 7.578692493946732e-06, | |
| "loss": 0.3418, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 7.554479418886199e-06, | |
| "loss": 0.3398, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 7.530266343825666e-06, | |
| "loss": 0.354, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 7.506053268765134e-06, | |
| "loss": 0.3247, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 7.4818401937046015e-06, | |
| "loss": 0.353, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 7.4576271186440685e-06, | |
| "loss": 0.3359, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 7.4334140435835355e-06, | |
| "loss": 0.3535, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 7.4092009685230025e-06, | |
| "loss": 0.3276, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 7.384987893462471e-06, | |
| "loss": 0.3457, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 7.360774818401938e-06, | |
| "loss": 0.3423, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 7.336561743341405e-06, | |
| "loss": 0.3408, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 7.312348668280872e-06, | |
| "loss": 0.3467, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 7.288135593220339e-06, | |
| "loss": 0.3291, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 7.263922518159807e-06, | |
| "loss": 0.3423, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 7.239709443099274e-06, | |
| "loss": 0.3325, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 7.215496368038741e-06, | |
| "loss": 0.3413, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 7.191283292978209e-06, | |
| "loss": 0.3472, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 7.1670702179176766e-06, | |
| "loss": 0.3462, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 7.1428571428571436e-06, | |
| "loss": 0.3394, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 7.1186440677966106e-06, | |
| "loss": 0.3315, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 7.0944309927360775e-06, | |
| "loss": 0.3496, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 7.070217917675546e-06, | |
| "loss": 0.3638, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 7.046004842615013e-06, | |
| "loss": 0.3408, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 7.02179176755448e-06, | |
| "loss": 0.3457, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.997578692493947e-06, | |
| "loss": 0.3374, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.973365617433414e-06, | |
| "loss": 0.2184, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.949152542372882e-06, | |
| "loss": 0.1985, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.924939467312349e-06, | |
| "loss": 0.2029, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.900726392251816e-06, | |
| "loss": 0.1963, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.876513317191284e-06, | |
| "loss": 0.2009, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.852300242130752e-06, | |
| "loss": 0.1885, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.828087167070219e-06, | |
| "loss": 0.1909, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.803874092009686e-06, | |
| "loss": 0.1858, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 6.779661016949153e-06, | |
| "loss": 0.1929, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 6.75544794188862e-06, | |
| "loss": 0.1863, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 6.731234866828088e-06, | |
| "loss": 0.1941, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 6.707021791767555e-06, | |
| "loss": 0.1946, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 6.682808716707022e-06, | |
| "loss": 0.198, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.658595641646489e-06, | |
| "loss": 0.1833, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.634382566585957e-06, | |
| "loss": 0.1975, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.610169491525424e-06, | |
| "loss": 0.1831, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.585956416464891e-06, | |
| "loss": 0.1826, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.561743341404359e-06, | |
| "loss": 0.188, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.537530266343827e-06, | |
| "loss": 0.1812, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.513317191283294e-06, | |
| "loss": 0.1882, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.489104116222761e-06, | |
| "loss": 0.1992, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.464891041162228e-06, | |
| "loss": 0.1919, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.440677966101695e-06, | |
| "loss": 0.1775, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.416464891041163e-06, | |
| "loss": 0.1812, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.39225181598063e-06, | |
| "loss": 0.1824, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.368038740920097e-06, | |
| "loss": 0.1648, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.343825665859564e-06, | |
| "loss": 0.1882, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.319612590799032e-06, | |
| "loss": 0.1824, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.295399515738499e-06, | |
| "loss": 0.179, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.271186440677966e-06, | |
| "loss": 0.1785, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.246973365617434e-06, | |
| "loss": 0.1843, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 6.222760290556902e-06, | |
| "loss": 0.1995, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 6.198547215496369e-06, | |
| "loss": 0.179, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 6.174334140435836e-06, | |
| "loss": 0.1758, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 6.150121065375303e-06, | |
| "loss": 0.1804, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 6.12590799031477e-06, | |
| "loss": 0.1895, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 6.1016949152542385e-06, | |
| "loss": 0.1841, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 6.0774818401937055e-06, | |
| "loss": 0.1719, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 6.0532687651331724e-06, | |
| "loss": 0.1838, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 6.0290556900726394e-06, | |
| "loss": 0.179, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 6.004842615012107e-06, | |
| "loss": 0.1711, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.980629539951574e-06, | |
| "loss": 0.1777, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.956416464891041e-06, | |
| "loss": 0.1643, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.932203389830509e-06, | |
| "loss": 0.177, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.907990314769976e-06, | |
| "loss": 0.1804, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.883777239709444e-06, | |
| "loss": 0.187, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.859564164648911e-06, | |
| "loss": 0.1609, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.835351089588378e-06, | |
| "loss": 0.1819, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.811138014527845e-06, | |
| "loss": 0.1743, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.7869249394673135e-06, | |
| "loss": 0.1711, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.7627118644067805e-06, | |
| "loss": 0.1799, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.7384987893462475e-06, | |
| "loss": 0.1843, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 0.1768, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.690072639225182e-06, | |
| "loss": 0.1804, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 5.665859564164649e-06, | |
| "loss": 0.1653, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 5.641646489104116e-06, | |
| "loss": 0.1763, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.617433414043584e-06, | |
| "loss": 0.1841, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.593220338983051e-06, | |
| "loss": 0.1792, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.569007263922519e-06, | |
| "loss": 0.1631, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.544794188861986e-06, | |
| "loss": 0.178, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.520581113801453e-06, | |
| "loss": 0.1785, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.49636803874092e-06, | |
| "loss": 0.1682, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.472154963680389e-06, | |
| "loss": 0.1768, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.447941888619856e-06, | |
| "loss": 0.178, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.423728813559323e-06, | |
| "loss": 0.1777, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.39951573849879e-06, | |
| "loss": 0.176, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.375302663438257e-06, | |
| "loss": 0.178, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.351089588377724e-06, | |
| "loss": 0.1628, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.326876513317191e-06, | |
| "loss": 0.1719, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.302663438256659e-06, | |
| "loss": 0.1687, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.278450363196126e-06, | |
| "loss": 0.1819, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.254237288135594e-06, | |
| "loss": 0.1741, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.230024213075061e-06, | |
| "loss": 0.168, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.205811138014528e-06, | |
| "loss": 0.1782, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.181598062953995e-06, | |
| "loss": 0.1829, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.157384987893464e-06, | |
| "loss": 0.1807, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.133171912832931e-06, | |
| "loss": 0.1775, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.108958837772398e-06, | |
| "loss": 0.1865, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.084745762711865e-06, | |
| "loss": 0.1792, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.060532687651332e-06, | |
| "loss": 0.1807, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.0363196125907995e-06, | |
| "loss": 0.1768, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.0121065375302665e-06, | |
| "loss": 0.1714, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.987893462469734e-06, | |
| "loss": 0.1758, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.963680387409201e-06, | |
| "loss": 0.1794, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.939467312348668e-06, | |
| "loss": 0.1794, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.915254237288136e-06, | |
| "loss": 0.1677, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.891041162227603e-06, | |
| "loss": 0.1672, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.866828087167071e-06, | |
| "loss": 0.1807, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.842615012106538e-06, | |
| "loss": 0.1731, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.818401937046005e-06, | |
| "loss": 0.1721, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.794188861985473e-06, | |
| "loss": 0.1724, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.76997578692494e-06, | |
| "loss": 0.1711, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.745762711864408e-06, | |
| "loss": 0.1826, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.721549636803875e-06, | |
| "loss": 0.1907, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.6973365617433416e-06, | |
| "loss": 0.1743, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.673123486682809e-06, | |
| "loss": 0.1602, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.648910411622276e-06, | |
| "loss": 0.1702, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.624697336561743e-06, | |
| "loss": 0.1914, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.600484261501211e-06, | |
| "loss": 0.1677, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.576271186440678e-06, | |
| "loss": 0.1775, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.552058111380145e-06, | |
| "loss": 0.1685, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.527845036319613e-06, | |
| "loss": 0.1709, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.50363196125908e-06, | |
| "loss": 0.1743, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.479418886198548e-06, | |
| "loss": 0.1753, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.455205811138015e-06, | |
| "loss": 0.1685, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.430992736077483e-06, | |
| "loss": 0.1746, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.40677966101695e-06, | |
| "loss": 0.1851, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.382566585956417e-06, | |
| "loss": 0.179, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.3583535108958845e-06, | |
| "loss": 0.1853, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.3341404358353515e-06, | |
| "loss": 0.1721, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.3099273607748185e-06, | |
| "loss": 0.1702, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.2857142857142855e-06, | |
| "loss": 0.1636, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.261501210653753e-06, | |
| "loss": 0.1816, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.23728813559322e-06, | |
| "loss": 0.1741, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.213075060532688e-06, | |
| "loss": 0.1765, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.188861985472155e-06, | |
| "loss": 0.1731, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.164648910411623e-06, | |
| "loss": 0.1782, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.14043583535109e-06, | |
| "loss": 0.1709, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.116222760290558e-06, | |
| "loss": 0.1758, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.092009685230025e-06, | |
| "loss": 0.1719, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.067796610169492e-06, | |
| "loss": 0.1741, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.04358353510896e-06, | |
| "loss": 0.1731, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 4.0193704600484266e-06, | |
| "loss": 0.1699, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 3.9951573849878936e-06, | |
| "loss": 0.178, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 3.9709443099273605e-06, | |
| "loss": 0.1807, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 3.946731234866828e-06, | |
| "loss": 0.1711, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 3.922518159806295e-06, | |
| "loss": 0.1765, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 3.898305084745763e-06, | |
| "loss": 0.1743, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 3.87409200968523e-06, | |
| "loss": 0.1719, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 3.849878934624698e-06, | |
| "loss": 0.175, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 3.825665859564165e-06, | |
| "loss": 0.1816, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 3.8014527845036324e-06, | |
| "loss": 0.165, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 3.7772397094430994e-06, | |
| "loss": 0.177, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 3.753026634382567e-06, | |
| "loss": 0.1829, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 3.7288135593220342e-06, | |
| "loss": 0.179, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 3.7046004842615012e-06, | |
| "loss": 0.177, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 3.680387409200969e-06, | |
| "loss": 0.1821, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 3.656174334140436e-06, | |
| "loss": 0.1733, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 3.6319612590799035e-06, | |
| "loss": 0.1792, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 3.6077481840193705e-06, | |
| "loss": 0.1785, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 3.5835351089588383e-06, | |
| "loss": 0.1638, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 3.5593220338983053e-06, | |
| "loss": 0.1643, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 3.535108958837773e-06, | |
| "loss": 0.1775, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 3.51089588377724e-06, | |
| "loss": 0.1614, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 3.486682808716707e-06, | |
| "loss": 0.186, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 3.4624697336561745e-06, | |
| "loss": 0.1772, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 3.438256658595642e-06, | |
| "loss": 0.1729, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 3.4140435835351093e-06, | |
| "loss": 0.176, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 3.3898305084745763e-06, | |
| "loss": 0.1738, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 3.365617433414044e-06, | |
| "loss": 0.1626, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 3.341404358353511e-06, | |
| "loss": 0.1792, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 3.3171912832929785e-06, | |
| "loss": 0.1621, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 3.2929782082324455e-06, | |
| "loss": 0.1802, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 3.2687651331719134e-06, | |
| "loss": 0.1675, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 3.2445520581113804e-06, | |
| "loss": 0.1729, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 3.2203389830508473e-06, | |
| "loss": 0.1748, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 3.196125907990315e-06, | |
| "loss": 0.1716, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 3.171912832929782e-06, | |
| "loss": 0.1699, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 3.1476997578692496e-06, | |
| "loss": 0.1841, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 3.123486682808717e-06, | |
| "loss": 0.1943, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 3.0992736077481844e-06, | |
| "loss": 0.1865, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 3.0750605326876514e-06, | |
| "loss": 0.168, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 3.0508474576271192e-06, | |
| "loss": 0.1768, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 3.0266343825665862e-06, | |
| "loss": 0.1689, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 3.0024213075060536e-06, | |
| "loss": 0.1753, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 2.9782082324455206e-06, | |
| "loss": 0.1672, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 2.953995157384988e-06, | |
| "loss": 0.1831, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 2.9297820823244554e-06, | |
| "loss": 0.1697, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 2.9055690072639224e-06, | |
| "loss": 0.1626, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 2.8813559322033903e-06, | |
| "loss": 0.176, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 2.8571428571428573e-06, | |
| "loss": 0.1785, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 2.8329297820823247e-06, | |
| "loss": 0.1716, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 2.808716707021792e-06, | |
| "loss": 0.1831, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 2.7845036319612595e-06, | |
| "loss": 0.1721, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 2.7602905569007265e-06, | |
| "loss": 0.1829, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 2.7360774818401943e-06, | |
| "loss": 0.167, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 2.7118644067796613e-06, | |
| "loss": 0.1804, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 2.6876513317191283e-06, | |
| "loss": 0.179, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 2.6634382566585957e-06, | |
| "loss": 0.1851, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 2.639225181598063e-06, | |
| "loss": 0.1699, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 2.6150121065375305e-06, | |
| "loss": 0.1743, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 2.5907990314769975e-06, | |
| "loss": 0.1763, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 2.5665859564164654e-06, | |
| "loss": 0.1719, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 2.5423728813559323e-06, | |
| "loss": 0.1743, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 2.5181598062953998e-06, | |
| "loss": 0.1714, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 2.493946731234867e-06, | |
| "loss": 0.1689, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 2.469733656174334e-06, | |
| "loss": 0.1753, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 2.4455205811138016e-06, | |
| "loss": 0.1741, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 2.421307506053269e-06, | |
| "loss": 0.1711, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 2.3970944309927364e-06, | |
| "loss": 0.1692, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 2.372881355932204e-06, | |
| "loss": 0.1692, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 2.3486682808716708e-06, | |
| "loss": 0.1729, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 2.324455205811138e-06, | |
| "loss": 0.1804, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 2.3002421307506056e-06, | |
| "loss": 0.1729, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 2.2760290556900726e-06, | |
| "loss": 0.1897, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 2.25181598062954e-06, | |
| "loss": 0.1711, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 2.2276029055690074e-06, | |
| "loss": 0.1726, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 2.203389830508475e-06, | |
| "loss": 0.1794, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 2.1791767554479422e-06, | |
| "loss": 0.1721, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 2.1549636803874092e-06, | |
| "loss": 0.1624, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 2.1307506053268766e-06, | |
| "loss": 0.187, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 2.106537530266344e-06, | |
| "loss": 0.1794, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 2.0823244552058115e-06, | |
| "loss": 0.1689, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 2.058111380145279e-06, | |
| "loss": 0.1614, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 2.033898305084746e-06, | |
| "loss": 0.1846, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 2.0096852300242133e-06, | |
| "loss": 0.1611, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 1.9854721549636803e-06, | |
| "loss": 0.1599, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 1.9612590799031477e-06, | |
| "loss": 0.1733, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 1.937046004842615e-06, | |
| "loss": 0.1794, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 1.9128329297820825e-06, | |
| "loss": 0.1782, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 1.8886198547215497e-06, | |
| "loss": 0.1658, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 1.8644067796610171e-06, | |
| "loss": 0.179, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 1.8401937046004845e-06, | |
| "loss": 0.1738, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 1.8159806295399517e-06, | |
| "loss": 0.1575, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 1.7917675544794191e-06, | |
| "loss": 0.1677, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 1.7675544794188866e-06, | |
| "loss": 0.1699, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 1.7433414043583535e-06, | |
| "loss": 0.177, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 1.719128329297821e-06, | |
| "loss": 0.1982, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 1.6949152542372882e-06, | |
| "loss": 0.1599, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 1.6707021791767556e-06, | |
| "loss": 0.1724, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 1.6464891041162228e-06, | |
| "loss": 0.1694, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 1.6222760290556902e-06, | |
| "loss": 0.1733, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 1.5980629539951576e-06, | |
| "loss": 0.1582, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 1.5738498789346248e-06, | |
| "loss": 0.1714, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 1.5496368038740922e-06, | |
| "loss": 0.177, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 1.5254237288135596e-06, | |
| "loss": 0.1763, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 1.5012106537530268e-06, | |
| "loss": 0.1807, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 1.476997578692494e-06, | |
| "loss": 0.1719, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 1.4527845036319612e-06, | |
| "loss": 0.1707, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 1.4285714285714286e-06, | |
| "loss": 0.1621, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 1.404358353510896e-06, | |
| "loss": 0.1653, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 1.3801452784503632e-06, | |
| "loss": 0.1653, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.3559322033898307e-06, | |
| "loss": 0.1692, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.3317191283292979e-06, | |
| "loss": 0.1707, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.3075060532687653e-06, | |
| "loss": 0.1621, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.2832929782082327e-06, | |
| "loss": 0.177, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.2590799031476999e-06, | |
| "loss": 0.178, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.234866828087167e-06, | |
| "loss": 0.1724, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.2106537530266345e-06, | |
| "loss": 0.1655, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.186440677966102e-06, | |
| "loss": 0.1765, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.162227602905569e-06, | |
| "loss": 0.1646, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.1380145278450363e-06, | |
| "loss": 0.1753, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.1138014527845037e-06, | |
| "loss": 0.1799, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.0895883777239711e-06, | |
| "loss": 0.1743, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.0653753026634383e-06, | |
| "loss": 0.1768, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.0411622276029057e-06, | |
| "loss": 0.1785, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.016949152542373e-06, | |
| "loss": 0.1675, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 9.927360774818401e-07, | |
| "loss": 0.1743, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 9.685230024213075e-07, | |
| "loss": 0.1704, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.443099273607749e-07, | |
| "loss": 0.1687, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.200968523002423e-07, | |
| "loss": 0.1741, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 8.958837772397096e-07, | |
| "loss": 0.1777, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 8.716707021791768e-07, | |
| "loss": 0.1643, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 8.474576271186441e-07, | |
| "loss": 0.1763, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 8.232445520581114e-07, | |
| "loss": 0.1816, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 7.990314769975788e-07, | |
| "loss": 0.1699, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 7.748184019370461e-07, | |
| "loss": 0.1619, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 7.506053268765134e-07, | |
| "loss": 0.1753, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 7.263922518159806e-07, | |
| "loss": 0.1685, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 7.02179176755448e-07, | |
| "loss": 0.1716, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 6.779661016949153e-07, | |
| "loss": 0.168, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 6.537530266343826e-07, | |
| "loss": 0.1729, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 6.295399515738499e-07, | |
| "loss": 0.1743, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 6.053268765133172e-07, | |
| "loss": 0.1646, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 5.811138014527845e-07, | |
| "loss": 0.1628, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 5.569007263922519e-07, | |
| "loss": 0.1658, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 5.326876513317192e-07, | |
| "loss": 0.1746, | |
| "step": 840 | |
| } | |
| ], | |
| "max_steps": 861, | |
| "num_train_epochs": 3, | |
| "total_flos": 819001106104320.0, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |