| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9968782518210197, | |
| "global_step": 1440, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.5454545454545457e-07, | |
| "loss": 0.862, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 9.090909090909091e-07, | |
| "loss": 0.8906, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.3636363636363636e-06, | |
| "loss": 0.8815, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.8181818181818183e-06, | |
| "loss": 0.7889, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.2727272727272728e-06, | |
| "loss": 0.8425, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.7272727272727272e-06, | |
| "loss": 0.8451, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.181818181818182e-06, | |
| "loss": 0.9298, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.6363636363636366e-06, | |
| "loss": 0.8441, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.0909090909090915e-06, | |
| "loss": 0.7971, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.5454545454545455e-06, | |
| "loss": 0.8787, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7311, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 5.4545454545454545e-06, | |
| "loss": 0.8228, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.90909090909091e-06, | |
| "loss": 0.7754, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.363636363636364e-06, | |
| "loss": 0.8758, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.818181818181818e-06, | |
| "loss": 0.7815, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 7.272727272727273e-06, | |
| "loss": 0.8514, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 7.727272727272727e-06, | |
| "loss": 0.855, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.181818181818183e-06, | |
| "loss": 0.8809, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.636363636363637e-06, | |
| "loss": 0.8084, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 0.8265, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.545454545454547e-06, | |
| "loss": 0.8748, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1e-05, | |
| "loss": 0.8812, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.0454545454545455e-05, | |
| "loss": 0.8223, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.0909090909090909e-05, | |
| "loss": 0.7911, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.1363636363636366e-05, | |
| "loss": 0.7482, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.181818181818182e-05, | |
| "loss": 0.8515, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.2272727272727274e-05, | |
| "loss": 0.7573, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.2727272727272728e-05, | |
| "loss": 0.8365, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.3181818181818183e-05, | |
| "loss": 0.831, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.3636363636363637e-05, | |
| "loss": 0.9328, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.4090909090909092e-05, | |
| "loss": 0.8345, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.4545454545454546e-05, | |
| "loss": 0.8601, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.8383, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.5454545454545454e-05, | |
| "loss": 0.8194, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.590909090909091e-05, | |
| "loss": 0.8306, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.6363636363636366e-05, | |
| "loss": 0.8341, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.681818181818182e-05, | |
| "loss": 0.861, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.7272727272727274e-05, | |
| "loss": 0.8035, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.772727272727273e-05, | |
| "loss": 0.9271, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.8181818181818182e-05, | |
| "loss": 0.8661, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.8636363636363638e-05, | |
| "loss": 0.916, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9090909090909094e-05, | |
| "loss": 0.9532, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9545454545454546e-05, | |
| "loss": 0.8439, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 2e-05, | |
| "loss": 0.9363, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9999974677957704e-05, | |
| "loss": 0.8266, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.999989871195906e-05, | |
| "loss": 0.8091, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9999772102388785e-05, | |
| "loss": 0.9211, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9999594849888083e-05, | |
| "loss": 0.878, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9999366955354638e-05, | |
| "loss": 0.9082, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9999088419942598e-05, | |
| "loss": 0.7854, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9998759245062583e-05, | |
| "loss": 0.8015, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.999837943238166e-05, | |
| "loss": 0.8802, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9997948983823363e-05, | |
| "loss": 0.9109, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.999746790156766e-05, | |
| "loss": 0.9564, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9996936188050947e-05, | |
| "loss": 0.9064, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9996353845966033e-05, | |
| "loss": 0.8017, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.999572087826214e-05, | |
| "loss": 0.9049, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.999503728814488e-05, | |
| "loss": 0.94, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.999430307907623e-05, | |
| "loss": 0.7948, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9993518254774517e-05, | |
| "loss": 0.8214, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.999268281921442e-05, | |
| "loss": 0.9716, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.999179677662692e-05, | |
| "loss": 0.8468, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9990860131499306e-05, | |
| "loss": 0.918, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.998987288857513e-05, | |
| "loss": 0.8961, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9988835052854188e-05, | |
| "loss": 0.8048, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9987746629592506e-05, | |
| "loss": 0.9069, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9986607624302305e-05, | |
| "loss": 0.9128, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9985418042751975e-05, | |
| "loss": 0.9182, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.998417789096604e-05, | |
| "loss": 0.9286, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9982887175225136e-05, | |
| "loss": 0.8841, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9981545902065974e-05, | |
| "loss": 0.8429, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.998015407828131e-05, | |
| "loss": 0.8837, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.997871171091991e-05, | |
| "loss": 0.8242, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9977218807286507e-05, | |
| "loss": 0.875, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9975675374941778e-05, | |
| "loss": 0.8317, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9974081421702296e-05, | |
| "loss": 0.846, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9972436955640488e-05, | |
| "loss": 0.8815, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.99707419850846e-05, | |
| "loss": 0.9116, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9968996518618664e-05, | |
| "loss": 0.9262, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9967200565082426e-05, | |
| "loss": 0.7856, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9965354133571334e-05, | |
| "loss": 0.8215, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9963457233436468e-05, | |
| "loss": 0.8985, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9961509874284508e-05, | |
| "loss": 0.8959, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9959512065977673e-05, | |
| "loss": 0.8295, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9957463818633683e-05, | |
| "loss": 0.9238, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9955365142625694e-05, | |
| "loss": 0.9243, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.995321604858227e-05, | |
| "loss": 0.8294, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9951016547387286e-05, | |
| "loss": 0.8749, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.9948766650179924e-05, | |
| "loss": 0.9327, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.994646636835458e-05, | |
| "loss": 0.8618, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.9944115713560825e-05, | |
| "loss": 0.8898, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.9941714697703333e-05, | |
| "loss": 0.822, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.993926333294182e-05, | |
| "loss": 0.8817, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9936761631691007e-05, | |
| "loss": 0.8315, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9934209606620532e-05, | |
| "loss": 0.9069, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.993160727065489e-05, | |
| "loss": 0.8181, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9928954636973373e-05, | |
| "loss": 0.994, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.992625171901e-05, | |
| "loss": 0.863, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.9923498530453456e-05, | |
| "loss": 0.9428, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.9920695085247012e-05, | |
| "loss": 0.9404, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.9917841397588453e-05, | |
| "loss": 0.9322, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.991493748193002e-05, | |
| "loss": 0.8693, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.991198335297834e-05, | |
| "loss": 0.9035, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9908979025694312e-05, | |
| "loss": 0.8077, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.990592451529309e-05, | |
| "loss": 0.8423, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9902819837243954e-05, | |
| "loss": 0.9348, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9899665007270266e-05, | |
| "loss": 0.8275, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.989646004134937e-05, | |
| "loss": 0.9544, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9893204955712525e-05, | |
| "loss": 0.9553, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9889899766844817e-05, | |
| "loss": 0.8593, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9886544491485068e-05, | |
| "loss": 0.9636, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9883139146625763e-05, | |
| "loss": 0.9157, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9879683749512962e-05, | |
| "loss": 0.8436, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9876178317646203e-05, | |
| "loss": 0.917, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.987262286877843e-05, | |
| "loss": 0.8816, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9869017420915888e-05, | |
| "loss": 0.9751, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9865361992318032e-05, | |
| "loss": 0.9259, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9861656601497452e-05, | |
| "loss": 0.9112, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9857901267219756e-05, | |
| "loss": 0.8553, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9854096008503495e-05, | |
| "loss": 0.8797, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9850240844620046e-05, | |
| "loss": 1.0086, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9846335795093547e-05, | |
| "loss": 0.8952, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.984238087970075e-05, | |
| "loss": 0.812, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9838376118470965e-05, | |
| "loss": 0.8831, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9834321531685944e-05, | |
| "loss": 0.9386, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9830217139879768e-05, | |
| "loss": 0.9607, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9826062963838756e-05, | |
| "loss": 0.883, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9821859024601345e-05, | |
| "loss": 0.9199, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9817605343458008e-05, | |
| "loss": 0.9299, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.981330194195112e-05, | |
| "loss": 0.9255, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9808948841874863e-05, | |
| "loss": 1.0294, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9804546065275116e-05, | |
| "loss": 0.9103, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.980009363444934e-05, | |
| "loss": 0.9002, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9795591571946454e-05, | |
| "loss": 0.9831, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9791039900566755e-05, | |
| "loss": 0.9152, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.978643864336176e-05, | |
| "loss": 0.9599, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.978178782363411e-05, | |
| "loss": 0.9653, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9777087464937464e-05, | |
| "loss": 0.9409, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.977233759107635e-05, | |
| "loss": 0.9401, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9767538226106078e-05, | |
| "loss": 0.9049, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9762689394332587e-05, | |
| "loss": 0.8994, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9757791120312344e-05, | |
| "loss": 0.8553, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9752843428852208e-05, | |
| "loss": 0.9076, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9747846345009306e-05, | |
| "loss": 0.9675, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.974279989409092e-05, | |
| "loss": 0.8934, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9737704101654335e-05, | |
| "loss": 0.945, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9732558993506723e-05, | |
| "loss": 0.8367, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9727364595705012e-05, | |
| "loss": 0.9107, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.972212093455576e-05, | |
| "loss": 0.7714, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9716828036615006e-05, | |
| "loss": 0.9865, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9711485928688148e-05, | |
| "loss": 0.9583, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9706094637829797e-05, | |
| "loss": 0.8547, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9700654191343662e-05, | |
| "loss": 0.949, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9695164616782378e-05, | |
| "loss": 0.8695, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9689625941947397e-05, | |
| "loss": 0.922, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9684038194888827e-05, | |
| "loss": 0.9375, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9678401403905307e-05, | |
| "loss": 0.9839, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9672715597543845e-05, | |
| "loss": 0.9473, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.966698080459969e-05, | |
| "loss": 0.9679, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9661197054116165e-05, | |
| "loss": 0.8366, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.965536437538456e-05, | |
| "loss": 0.9778, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.964948279794393e-05, | |
| "loss": 0.9235, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.9643552351581003e-05, | |
| "loss": 0.9808, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.963757306632996e-05, | |
| "loss": 0.9195, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.9631544972472357e-05, | |
| "loss": 0.9478, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.962546810053692e-05, | |
| "loss": 0.9075, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.961934248129941e-05, | |
| "loss": 0.9441, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.9613168145782468e-05, | |
| "loss": 0.8808, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.960694512525545e-05, | |
| "loss": 0.9136, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.960067345123427e-05, | |
| "loss": 0.9336, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.959435315548125e-05, | |
| "loss": 0.9673, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.958798427000495e-05, | |
| "loss": 0.9764, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.958156682706001e-05, | |
| "loss": 0.8567, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9575100859146974e-05, | |
| "loss": 0.899, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9568586399012152e-05, | |
| "loss": 0.9528, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.956202347964743e-05, | |
| "loss": 0.9559, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9555412134290103e-05, | |
| "loss": 0.9491, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.954875239642274e-05, | |
| "loss": 0.9203, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9542044299772963e-05, | |
| "loss": 0.9215, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9535287878313315e-05, | |
| "loss": 0.9526, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.952848316626108e-05, | |
| "loss": 0.9186, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.952163019807809e-05, | |
| "loss": 0.891, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.9514729008470584e-05, | |
| "loss": 0.9222, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.9507779632388997e-05, | |
| "loss": 0.9907, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9500782105027812e-05, | |
| "loss": 0.8791, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9493736461825366e-05, | |
| "loss": 0.9428, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9486642738463675e-05, | |
| "loss": 0.9729, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.947950097086825e-05, | |
| "loss": 0.9668, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9472311195207918e-05, | |
| "loss": 0.972, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.946507344789464e-05, | |
| "loss": 0.8848, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9457787765583326e-05, | |
| "loss": 0.9581, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.945045418517165e-05, | |
| "loss": 0.9263, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9443072743799855e-05, | |
| "loss": 0.885, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9435643478850573e-05, | |
| "loss": 0.9069, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9428166427948643e-05, | |
| "loss": 0.9419, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9420641628960897e-05, | |
| "loss": 0.9211, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9413069119995994e-05, | |
| "loss": 0.9004, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9405448939404215e-05, | |
| "loss": 0.8916, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9397781125777265e-05, | |
| "loss": 0.8475, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9390065717948084e-05, | |
| "loss": 0.9358, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.938230275499065e-05, | |
| "loss": 0.9872, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9374492276219776e-05, | |
| "loss": 0.944, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9366634321190914e-05, | |
| "loss": 0.8526, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9358728929699966e-05, | |
| "loss": 0.9562, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.9350776141783055e-05, | |
| "loss": 1.0946, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.9342775997716357e-05, | |
| "loss": 0.9664, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.933472853801586e-05, | |
| "loss": 1.0158, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.9326633803437197e-05, | |
| "loss": 0.9521, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.9318491834975396e-05, | |
| "loss": 0.8175, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.9310302673864724e-05, | |
| "loss": 0.9353, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.9302066361578433e-05, | |
| "loss": 0.9757, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.929378293982857e-05, | |
| "loss": 0.8769, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.9285452450565772e-05, | |
| "loss": 0.9397, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.9277074935979034e-05, | |
| "loss": 0.9353, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.9268650438495517e-05, | |
| "loss": 0.9003, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.926017900078031e-05, | |
| "loss": 0.8505, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.925166066573624e-05, | |
| "loss": 0.9931, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.924309547650363e-05, | |
| "loss": 0.8249, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.9234483476460102e-05, | |
| "loss": 0.9438, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.922582470922034e-05, | |
| "loss": 0.9678, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.921711921863588e-05, | |
| "loss": 0.9648, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.9208367048794878e-05, | |
| "loss": 0.9831, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.91995682440219e-05, | |
| "loss": 0.9613, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.9190722848877683e-05, | |
| "loss": 0.8756, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.9181830908158926e-05, | |
| "loss": 0.9365, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.9172892466898047e-05, | |
| "loss": 0.9263, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.9163907570362964e-05, | |
| "loss": 1.0134, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.9154876264056863e-05, | |
| "loss": 0.8728, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.9145798593717963e-05, | |
| "loss": 0.8961, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.9136674605319304e-05, | |
| "loss": 0.8716, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.9127504345068484e-05, | |
| "loss": 1.0111, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.911828785940745e-05, | |
| "loss": 0.9168, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.9109025195012244e-05, | |
| "loss": 0.947, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.9099716398792788e-05, | |
| "loss": 1.0074, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.9090361517892623e-05, | |
| "loss": 1.0229, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.908096059968869e-05, | |
| "loss": 0.9255, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.907151369179108e-05, | |
| "loss": 0.9936, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.906202084204279e-05, | |
| "loss": 0.8688, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.905248209851949e-05, | |
| "loss": 0.9057, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.904289750952928e-05, | |
| "loss": 0.9435, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.903326712361242e-05, | |
| "loss": 0.9164, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.9023590989541126e-05, | |
| "loss": 0.9033, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.9013869156319296e-05, | |
| "loss": 0.9085, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.900410167318226e-05, | |
| "loss": 0.9262, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.899428858959654e-05, | |
| "loss": 0.8613, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.8984429955259607e-05, | |
| "loss": 0.9152, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.8974525820099608e-05, | |
| "loss": 0.8743, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8964576234275123e-05, | |
| "loss": 0.9574, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.895458124817493e-05, | |
| "loss": 0.9238, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.894454091241771e-05, | |
| "loss": 0.9502, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8934455277851835e-05, | |
| "loss": 1.0451, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8924324395555066e-05, | |
| "loss": 0.9825, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.891414831683434e-05, | |
| "loss": 0.9385, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8903927093225474e-05, | |
| "loss": 0.8589, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8893660776492913e-05, | |
| "loss": 1.0166, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8883349418629487e-05, | |
| "loss": 0.9775, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8872993071856113e-05, | |
| "loss": 0.9641, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.8862591788621572e-05, | |
| "loss": 0.974, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.8852145621602208e-05, | |
| "loss": 0.9263, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.8841654623701673e-05, | |
| "loss": 0.9005, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.883111884805067e-05, | |
| "loss": 0.9284, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8820538348006666e-05, | |
| "loss": 0.8282, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8809913177153644e-05, | |
| "loss": 1.0034, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8799243389301796e-05, | |
| "loss": 0.9435, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8788529038487297e-05, | |
| "loss": 0.8977, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.877777017897199e-05, | |
| "loss": 0.8815, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.8766966865243138e-05, | |
| "loss": 0.9596, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.8756119152013134e-05, | |
| "loss": 0.9779, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.874522709421922e-05, | |
| "loss": 0.8715, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.873429074702324e-05, | |
| "loss": 0.836, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.872331016581131e-05, | |
| "loss": 0.9368, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8712285406193585e-05, | |
| "loss": 0.9233, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8701216524003956e-05, | |
| "loss": 1.0079, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8690103575299754e-05, | |
| "loss": 0.9296, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8678946616361492e-05, | |
| "loss": 0.9434, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.866774570369257e-05, | |
| "loss": 0.986, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8656500894018986e-05, | |
| "loss": 0.9588, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8645212244289047e-05, | |
| "loss": 0.864, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8633879811673086e-05, | |
| "loss": 0.9181, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8622503653563173e-05, | |
| "loss": 0.8731, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8611083827572818e-05, | |
| "loss": 0.8241, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.8599620391536682e-05, | |
| "loss": 0.9319, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.858811340351029e-05, | |
| "loss": 0.9154, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.8576562921769727e-05, | |
| "loss": 0.9902, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.8564969004811355e-05, | |
| "loss": 0.8483, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8553331711351502e-05, | |
| "loss": 0.9915, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8541651100326175e-05, | |
| "loss": 0.9597, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8529927230890757e-05, | |
| "loss": 1.011, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8518160162419722e-05, | |
| "loss": 0.8831, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.85063499545063e-05, | |
| "loss": 0.9561, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.849449666696221e-05, | |
| "loss": 0.9443, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8482600359817344e-05, | |
| "loss": 0.9633, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8470661093319462e-05, | |
| "loss": 0.9759, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8458678927933884e-05, | |
| "loss": 0.9752, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8446653924343188e-05, | |
| "loss": 0.9684, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.843458614344691e-05, | |
| "loss": 0.8356, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.842247564636121e-05, | |
| "loss": 1.0318, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8410322494418606e-05, | |
| "loss": 0.9808, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8398126749167615e-05, | |
| "loss": 0.873, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8385888472372474e-05, | |
| "loss": 0.9989, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8373607726012816e-05, | |
| "loss": 0.961, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8361284572283356e-05, | |
| "loss": 0.978, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.834891907359358e-05, | |
| "loss": 0.9402, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.833651129256742e-05, | |
| "loss": 0.9715, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8324061292042953e-05, | |
| "loss": 0.9632, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.831156913507206e-05, | |
| "loss": 0.9161, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.829903488492013e-05, | |
| "loss": 0.9706, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.828645860506573e-05, | |
| "loss": 0.9416, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8273840359200273e-05, | |
| "loss": 0.8993, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.826118021122771e-05, | |
| "loss": 0.8861, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.82484782252642e-05, | |
| "loss": 0.8936, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8235734465637794e-05, | |
| "loss": 1.0163, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.822294899688809e-05, | |
| "loss": 0.9154, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.821012188376593e-05, | |
| "loss": 0.9331, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8197253191233054e-05, | |
| "loss": 1.0286, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8184342984461766e-05, | |
| "loss": 0.8369, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.817139132883464e-05, | |
| "loss": 1.0157, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8158398289944145e-05, | |
| "loss": 0.9841, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8145363933592338e-05, | |
| "loss": 0.9957, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.8132288325790518e-05, | |
| "loss": 0.9361, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.811917153275891e-05, | |
| "loss": 0.9471, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.8106013620926312e-05, | |
| "loss": 0.9691, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.809281465692976e-05, | |
| "loss": 1.0162, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.8079574707614202e-05, | |
| "loss": 0.8634, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.8066293840032146e-05, | |
| "loss": 1.0123, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.8052972121443337e-05, | |
| "loss": 0.9297, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.803960961931439e-05, | |
| "loss": 0.8908, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.802620640131848e-05, | |
| "loss": 1.0379, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.8012762535334978e-05, | |
| "loss": 0.8826, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.799927808944911e-05, | |
| "loss": 1.0062, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.7985753131951617e-05, | |
| "loss": 0.9014, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.797218773133841e-05, | |
| "loss": 0.9044, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.795858195631022e-05, | |
| "loss": 0.9804, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7944935875772244e-05, | |
| "loss": 0.9798, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7931249558833815e-05, | |
| "loss": 0.9973, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7917523074808024e-05, | |
| "loss": 0.9562, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.79037564932114e-05, | |
| "loss": 0.9356, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7889949883763532e-05, | |
| "loss": 0.9679, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.787610331638673e-05, | |
| "loss": 0.927, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.786221686120567e-05, | |
| "loss": 0.8715, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.784829058854703e-05, | |
| "loss": 0.9505, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7834324568939137e-05, | |
| "loss": 0.8715, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7820318873111627e-05, | |
| "loss": 0.9385, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7806273571995066e-05, | |
| "loss": 0.9108, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7792188736720593e-05, | |
| "loss": 0.9772, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7778064438619562e-05, | |
| "loss": 0.9675, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7763900749223194e-05, | |
| "loss": 1.0191, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7749697740262197e-05, | |
| "loss": 0.8954, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.7735455483666408e-05, | |
| "loss": 0.9897, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.772117405156443e-05, | |
| "loss": 0.9625, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.770685351628327e-05, | |
| "loss": 0.9269, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.769249395034797e-05, | |
| "loss": 0.8875, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.7678095426481237e-05, | |
| "loss": 0.9461, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.7663658017603073e-05, | |
| "loss": 1.0047, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.7649181796830418e-05, | |
| "loss": 0.9555, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.7634666837476765e-05, | |
| "loss": 0.9941, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.7620113213051798e-05, | |
| "loss": 0.8917, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7605520997261014e-05, | |
| "loss": 0.9091, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7590890264005357e-05, | |
| "loss": 0.8904, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.757622108738083e-05, | |
| "loss": 0.9379, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7561513541678142e-05, | |
| "loss": 0.9434, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.754676770138231e-05, | |
| "loss": 0.9773, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.7531983641172294e-05, | |
| "loss": 0.9735, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.7517161435920606e-05, | |
| "loss": 0.9642, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.7502301160692954e-05, | |
| "loss": 0.8944, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.7487402890747843e-05, | |
| "loss": 0.918, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.7472466701536193e-05, | |
| "loss": 0.9872, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.7457492668700967e-05, | |
| "loss": 0.9949, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.744248086807679e-05, | |
| "loss": 0.8549, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.7427431375689544e-05, | |
| "loss": 0.9385, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.741234426775601e-05, | |
| "loss": 1.1029, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.7397219620683465e-05, | |
| "loss": 0.9848, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.73820575110693e-05, | |
| "loss": 0.9114, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.7366858015700626e-05, | |
| "loss": 1.0012, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.73516212115539e-05, | |
| "loss": 0.9107, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.7336347175794523e-05, | |
| "loss": 0.9813, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.732103598577645e-05, | |
| "loss": 0.9791, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.73056877190418e-05, | |
| "loss": 0.8552, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.7290302453320468e-05, | |
| "loss": 1.006, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.7274880266529716e-05, | |
| "loss": 0.9354, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.7259421236773807e-05, | |
| "loss": 1.0066, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.7243925442343578e-05, | |
| "loss": 0.9847, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.722839296171606e-05, | |
| "loss": 0.8935, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.721282387355408e-05, | |
| "loss": 0.873, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.719721825670586e-05, | |
| "loss": 0.9384, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.718157619020462e-05, | |
| "loss": 0.8872, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.7165897753268168e-05, | |
| "loss": 0.9777, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.715018302529852e-05, | |
| "loss": 0.9095, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.713443208588147e-05, | |
| "loss": 0.9033, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.711864501478622e-05, | |
| "loss": 0.9215, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.7102821891964937e-05, | |
| "loss": 0.9063, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.7086962797552376e-05, | |
| "loss": 0.9396, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.7071067811865477e-05, | |
| "loss": 0.859, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.7055137015402935e-05, | |
| "loss": 0.9462, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.703917048884481e-05, | |
| "loss": 1.0144, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.7023168313052118e-05, | |
| "loss": 0.8752, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.700713056906641e-05, | |
| "loss": 0.9419, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6991057338109376e-05, | |
| "loss": 0.9295, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.697494870158242e-05, | |
| "loss": 0.9291, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6958804741066254e-05, | |
| "loss": 0.9928, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6942625538320495e-05, | |
| "loss": 0.9423, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6926411175283227e-05, | |
| "loss": 1.002, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6910161734070614e-05, | |
| "loss": 0.9062, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.689387729697646e-05, | |
| "loss": 0.9349, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6877557946471808e-05, | |
| "loss": 0.8808, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.686120376520451e-05, | |
| "loss": 0.9542, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6844814835998825e-05, | |
| "loss": 0.9387, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6828391241854983e-05, | |
| "loss": 0.8898, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6811933065948777e-05, | |
| "loss": 0.9248, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6795440391631122e-05, | |
| "loss": 0.9436, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6778913302427668e-05, | |
| "loss": 0.8674, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6762351882038342e-05, | |
| "loss": 0.9529, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6745756214336938e-05, | |
| "loss": 0.8856, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6729126383370696e-05, | |
| "loss": 0.9027, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.671246247335988e-05, | |
| "loss": 0.9386, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.669576456869733e-05, | |
| "loss": 0.8582, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.6679032753948056e-05, | |
| "loss": 0.9375, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.666226711384881e-05, | |
| "loss": 0.8675, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.664546773330763e-05, | |
| "loss": 1.0055, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.6628634697403447e-05, | |
| "loss": 0.9511, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.661176809138563e-05, | |
| "loss": 1.0775, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6594868000673562e-05, | |
| "loss": 0.9636, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6577934510856205e-05, | |
| "loss": 1.0389, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6560967707691663e-05, | |
| "loss": 1.0259, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6543967677106767e-05, | |
| "loss": 0.9391, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.6526934505196605e-05, | |
| "loss": 0.9565, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.6509868278224125e-05, | |
| "loss": 1.0194, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.649276908261967e-05, | |
| "loss": 0.9288, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.6475637004980547e-05, | |
| "loss": 0.9685, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.64584721320706e-05, | |
| "loss": 0.9009, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.6441274550819756e-05, | |
| "loss": 0.9512, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.642404434832358e-05, | |
| "loss": 0.9748, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.6406781611842874e-05, | |
| "loss": 0.9136, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.6389486428803173e-05, | |
| "loss": 0.9356, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.6372158886794348e-05, | |
| "loss": 0.925, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.635479907357016e-05, | |
| "loss": 0.9704, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.6337407077047783e-05, | |
| "loss": 0.9933, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.63199829853074e-05, | |
| "loss": 0.9215, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.630252688659172e-05, | |
| "loss": 0.9324, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.6285038869305565e-05, | |
| "loss": 0.9964, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.6267519022015394e-05, | |
| "loss": 0.9284, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.624996743344887e-05, | |
| "loss": 0.9563, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.6232384192494404e-05, | |
| "loss": 0.8989, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.621476938820071e-05, | |
| "loss": 0.9737, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.6197123109776362e-05, | |
| "loss": 0.9878, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.6179445446589308e-05, | |
| "loss": 1.0366, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.6161736488166463e-05, | |
| "loss": 0.9869, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.6143996324193227e-05, | |
| "loss": 0.9227, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.612622504451303e-05, | |
| "loss": 1.0251, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.6108422739126896e-05, | |
| "loss": 0.9149, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.609058949819297e-05, | |
| "loss": 0.9265, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.6072725412026066e-05, | |
| "loss": 0.9905, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.605483057109722e-05, | |
| "loss": 0.9826, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.6036905066033207e-05, | |
| "loss": 0.9663, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.6018948987616108e-05, | |
| "loss": 0.9652, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.6000962426782844e-05, | |
| "loss": 0.9741, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5982945474624702e-05, | |
| "loss": 0.93, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.596489822238689e-05, | |
| "loss": 0.9246, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.594682076146806e-05, | |
| "loss": 0.9065, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.592871318341986e-05, | |
| "loss": 0.9947, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5910575579946462e-05, | |
| "loss": 0.9181, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5892408042904098e-05, | |
| "loss": 0.9076, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.58742106643006e-05, | |
| "loss": 0.9491, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.585598353629492e-05, | |
| "loss": 0.9746, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.583772675119668e-05, | |
| "loss": 0.9725, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.58194404014657e-05, | |
| "loss": 1.1093, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5801124579711527e-05, | |
| "loss": 0.912, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5782779378692957e-05, | |
| "loss": 0.9891, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5764404891317583e-05, | |
| "loss": 0.9395, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5746001210641316e-05, | |
| "loss": 0.9278, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5727568429867914e-05, | |
| "loss": 1.0108, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.57091066423485e-05, | |
| "loss": 0.8841, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.569061594158112e-05, | |
| "loss": 0.9375, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5672096421210217e-05, | |
| "loss": 0.8543, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5653548175026226e-05, | |
| "loss": 0.9097, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5634971296965027e-05, | |
| "loss": 0.9258, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.561636588110753e-05, | |
| "loss": 0.9161, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5597732021679153e-05, | |
| "loss": 0.8855, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.557906981304937e-05, | |
| "loss": 0.9972, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.5560379349731234e-05, | |
| "loss": 1.0628, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.5541660726380884e-05, | |
| "loss": 0.7049, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.552291403779707e-05, | |
| "loss": 0.4825, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.550413937892069e-05, | |
| "loss": 0.4481, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.5485336844834274e-05, | |
| "loss": 0.4417, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.546650653076154e-05, | |
| "loss": 0.5073, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.544764853206689e-05, | |
| "loss": 0.4679, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.542876294425493e-05, | |
| "loss": 0.4235, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.5409849862969994e-05, | |
| "loss": 0.472, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.5390909383995648e-05, | |
| "loss": 0.4321, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.5371941603254215e-05, | |
| "loss": 0.4257, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.5352946616806284e-05, | |
| "loss": 0.397, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.5333924520850227e-05, | |
| "loss": 0.4081, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.5314875411721705e-05, | |
| "loss": 0.4274, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.529579938589319e-05, | |
| "loss": 0.4633, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.5276696539973466e-05, | |
| "loss": 0.4458, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.5257566970707147e-05, | |
| "loss": 0.4029, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.5238410774974188e-05, | |
| "loss": 0.3795, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.5219228049789388e-05, | |
| "loss": 0.3949, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.5200018892301903e-05, | |
| "loss": 0.4275, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.5180783399794749e-05, | |
| "loss": 0.386, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.5161521669684324e-05, | |
| "loss": 0.4296, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.514223379951989e-05, | |
| "loss": 0.3951, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.5122919886983103e-05, | |
| "loss": 0.3857, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.5103580029887504e-05, | |
| "loss": 0.42, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.5084214326178033e-05, | |
| "loss": 0.4164, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.5064822873930516e-05, | |
| "loss": 0.3741, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.5045405771351194e-05, | |
| "loss": 0.4453, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.5025963116776203e-05, | |
| "loss": 0.4425, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.5006495008671089e-05, | |
| "loss": 0.418, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.49870015456303e-05, | |
| "loss": 0.4239, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.4967482826376698e-05, | |
| "loss": 0.3721, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4947938949761054e-05, | |
| "loss": 0.4399, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4928370014761533e-05, | |
| "loss": 0.4596, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.490877612048322e-05, | |
| "loss": 0.4163, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4889157366157597e-05, | |
| "loss": 0.4167, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4869513851142051e-05, | |
| "loss": 0.4211, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4849845674919365e-05, | |
| "loss": 0.3894, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.483015293709722e-05, | |
| "loss": 0.3833, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4810435737407679e-05, | |
| "loss": 0.4314, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4790694175706698e-05, | |
| "loss": 0.4456, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4770928351973605e-05, | |
| "loss": 0.4001, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4751138366310612e-05, | |
| "loss": 0.4039, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4731324318942284e-05, | |
| "loss": 0.3601, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4711486310215053e-05, | |
| "loss": 0.4098, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4691624440596699e-05, | |
| "loss": 0.4034, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4671738810675838e-05, | |
| "loss": 0.4215, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4651829521161424e-05, | |
| "loss": 0.4936, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4631896672882235e-05, | |
| "loss": 0.4182, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4611940366786353e-05, | |
| "loss": 0.3938, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4591960703940662e-05, | |
| "loss": 0.4405, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.4571957785530342e-05, | |
| "loss": 0.415, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.4551931712858334e-05, | |
| "loss": 0.4081, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.453188258734486e-05, | |
| "loss": 0.4478, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.4511810510526869e-05, | |
| "loss": 0.3728, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.4491715584057562e-05, | |
| "loss": 0.3363, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.4471597909705858e-05, | |
| "loss": 0.4494, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.4451457589355873e-05, | |
| "loss": 0.4371, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.4431294725006415e-05, | |
| "loss": 0.4261, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.4411109418770468e-05, | |
| "loss": 0.4078, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.4390901772874668e-05, | |
| "loss": 0.3529, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.4370671889658783e-05, | |
| "loss": 0.4159, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.435041987157521e-05, | |
| "loss": 0.4073, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.4330145821188437e-05, | |
| "loss": 0.4827, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.4309849841174538e-05, | |
| "loss": 0.4802, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.4289532034320649e-05, | |
| "loss": 0.4123, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.4269192503524435e-05, | |
| "loss": 0.4248, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.4248831351793593e-05, | |
| "loss": 0.4399, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.422844868224531e-05, | |
| "loss": 0.4324, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.4208044598105755e-05, | |
| "loss": 0.3905, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.4187619202709538e-05, | |
| "loss": 0.422, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.4167172599499209e-05, | |
| "loss": 0.4109, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.4146704892024714e-05, | |
| "loss": 0.3866, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.4126216183942889e-05, | |
| "loss": 0.3672, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.4105706579016916e-05, | |
| "loss": 0.3688, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.4085176181115812e-05, | |
| "loss": 0.4385, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.40646250942139e-05, | |
| "loss": 0.3453, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.404405342239028e-05, | |
| "loss": 0.4604, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.4023461269828297e-05, | |
| "loss": 0.4166, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.4002848740815023e-05, | |
| "loss": 0.4709, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3982215939740726e-05, | |
| "loss": 0.3525, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3961562971098337e-05, | |
| "loss": 0.4314, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3940889939482925e-05, | |
| "loss": 0.4369, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3920196949591166e-05, | |
| "loss": 0.3846, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3899484106220816e-05, | |
| "loss": 0.4244, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.387875151427017e-05, | |
| "loss": 0.3922, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3857999278737546e-05, | |
| "loss": 0.4203, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3837227504720739e-05, | |
| "loss": 0.4391, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3816436297416496e-05, | |
| "loss": 0.3467, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3795625762119987e-05, | |
| "loss": 0.3743, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3774796004224258e-05, | |
| "loss": 0.3716, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3753947129219713e-05, | |
| "loss": 0.4121, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3733079242693572e-05, | |
| "loss": 0.331, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3712192450329338e-05, | |
| "loss": 0.4222, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3691286857906254e-05, | |
| "loss": 0.3869, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3670362571298781e-05, | |
| "loss": 0.3986, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3649419696476057e-05, | |
| "loss": 0.4474, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.362845833950135e-05, | |
| "loss": 0.4791, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3607478606531533e-05, | |
| "loss": 0.4392, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3586480603816545e-05, | |
| "loss": 0.3602, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.356546443769885e-05, | |
| "loss": 0.4234, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.3544430214612896e-05, | |
| "loss": 0.4063, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.3523378041084576e-05, | |
| "loss": 0.4235, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.3502308023730703e-05, | |
| "loss": 0.3335, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.3481220269258449e-05, | |
| "loss": 0.3894, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.3460114884464814e-05, | |
| "loss": 0.4579, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.3438991976236087e-05, | |
| "loss": 0.4406, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.3417851651547306e-05, | |
| "loss": 0.46, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.3396694017461708e-05, | |
| "loss": 0.4724, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.3375519181130194e-05, | |
| "loss": 0.4395, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.3354327249790786e-05, | |
| "loss": 0.3886, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.3333118330768082e-05, | |
| "loss": 0.4184, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.3311892531472705e-05, | |
| "loss": 0.4415, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.3290649959400777e-05, | |
| "loss": 0.4539, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.3269390722133358e-05, | |
| "loss": 0.4399, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.3248114927335909e-05, | |
| "loss": 0.3937, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.3226822682757745e-05, | |
| "loss": 0.3891, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.320551409623149e-05, | |
| "loss": 0.3631, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.3184189275672532e-05, | |
| "loss": 0.4561, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.316284832907847e-05, | |
| "loss": 0.3919, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.3141491364528576e-05, | |
| "loss": 0.4112, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.3120118490183238e-05, | |
| "loss": 0.3782, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.3098729814283426e-05, | |
| "loss": 0.4207, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.3077325445150133e-05, | |
| "loss": 0.3905, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.3055905491183822e-05, | |
| "loss": 0.4208, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.3034470060863891e-05, | |
| "loss": 0.3682, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.3013019262748112e-05, | |
| "loss": 0.4489, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2991553205472091e-05, | |
| "loss": 0.4073, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2970071997748712e-05, | |
| "loss": 0.4442, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2948575748367585e-05, | |
| "loss": 0.4569, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2927064566194493e-05, | |
| "loss": 0.3878, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2905538560170855e-05, | |
| "loss": 0.4247, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.288399783931315e-05, | |
| "loss": 0.3893, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2862442512712394e-05, | |
| "loss": 0.3758, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2840872689533562e-05, | |
| "loss": 0.4214, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2819288479015048e-05, | |
| "loss": 0.3903, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2797689990468113e-05, | |
| "loss": 0.3964, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2776077333276326e-05, | |
| "loss": 0.4012, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2754450616895006e-05, | |
| "loss": 0.3732, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2732809950850685e-05, | |
| "loss": 0.4625, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2711155444740529e-05, | |
| "loss": 0.4015, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2689487208231805e-05, | |
| "loss": 0.4655, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2667805351061314e-05, | |
| "loss": 0.4486, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2646109983034833e-05, | |
| "loss": 0.3759, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2624401214026574e-05, | |
| "loss": 0.4325, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2602679153978602e-05, | |
| "loss": 0.4113, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2580943912900309e-05, | |
| "loss": 0.414, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.255919560086783e-05, | |
| "loss": 0.4371, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.2537434328023501e-05, | |
| "loss": 0.3822, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.2515660204575298e-05, | |
| "loss": 0.4596, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.2493873340796271e-05, | |
| "loss": 0.4196, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.2472073847024004e-05, | |
| "loss": 0.4373, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.2450261833660033e-05, | |
| "loss": 0.4279, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.2428437411169306e-05, | |
| "loss": 0.4424, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.2406600690079608e-05, | |
| "loss": 0.4165, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.238475178098102e-05, | |
| "loss": 0.4107, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.2362890794525342e-05, | |
| "loss": 0.3784, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.234101784142553e-05, | |
| "loss": 0.4449, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.2319133032455164e-05, | |
| "loss": 0.3852, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.2297236478447848e-05, | |
| "loss": 0.4102, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.2275328290296677e-05, | |
| "loss": 0.3537, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.2253408578953667e-05, | |
| "loss": 0.3405, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.2231477455429185e-05, | |
| "loss": 0.4722, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.2209535030791404e-05, | |
| "loss": 0.4768, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.2187581416165721e-05, | |
| "loss": 0.3828, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.2165616722734211e-05, | |
| "loss": 0.4437, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.2143641061735048e-05, | |
| "loss": 0.4225, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.212165454446196e-05, | |
| "loss": 0.4594, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.2099657282263651e-05, | |
| "loss": 0.4256, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.2077649386543238e-05, | |
| "loss": 0.4465, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.2055630968757696e-05, | |
| "loss": 0.4011, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.2033602140417288e-05, | |
| "loss": 0.3896, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.2011563013084996e-05, | |
| "loss": 0.395, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1989513698375968e-05, | |
| "loss": 0.4412, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1967454307956933e-05, | |
| "loss": 0.4305, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.194538495354566e-05, | |
| "loss": 0.4026, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1923305746910372e-05, | |
| "loss": 0.3974, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1901216799869188e-05, | |
| "loss": 0.4072, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1879118224289563e-05, | |
| "loss": 0.4573, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1857010132087704e-05, | |
| "loss": 0.4181, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1834892635228024e-05, | |
| "loss": 0.4152, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1812765845722561e-05, | |
| "loss": 0.3975, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1790629875630412e-05, | |
| "loss": 0.3084, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1768484837057176e-05, | |
| "loss": 0.4854, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1746330842154371e-05, | |
| "loss": 0.4309, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1724168003118876e-05, | |
| "loss": 0.4201, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1701996432192363e-05, | |
| "loss": 0.376, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1679816241660719e-05, | |
| "loss": 0.4495, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1657627543853491e-05, | |
| "loss": 0.3998, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1635430451143308e-05, | |
| "loss": 0.3632, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1613225075945316e-05, | |
| "loss": 0.4364, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.15910115307166e-05, | |
| "loss": 0.4696, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.156878992795563e-05, | |
| "loss": 0.4335, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.154656038020168e-05, | |
| "loss": 0.4057, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1524323000034256e-05, | |
| "loss": 0.4275, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1502077900072534e-05, | |
| "loss": 0.3872, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.1479825192974791e-05, | |
| "loss": 0.4417, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.1457564991437825e-05, | |
| "loss": 0.4162, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.1435297408196382e-05, | |
| "loss": 0.3493, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.1413022556022606e-05, | |
| "loss": 0.4046, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.1390740547725443e-05, | |
| "loss": 0.371, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.1368451496150088e-05, | |
| "loss": 0.424, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.13461555141774e-05, | |
| "loss": 0.3573, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.1323852714723335e-05, | |
| "loss": 0.4638, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.1301543210738383e-05, | |
| "loss": 0.3928, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.1279227115206988e-05, | |
| "loss": 0.3628, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.1256904541146966e-05, | |
| "loss": 0.4568, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.1234575601608956e-05, | |
| "loss": 0.3592, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.1212240409675825e-05, | |
| "loss": 0.3669, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.1189899078462107e-05, | |
| "loss": 0.4385, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.1167551721113435e-05, | |
| "loss": 0.3741, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.1145198450805948e-05, | |
| "loss": 0.4703, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.1122839380745738e-05, | |
| "loss": 0.4382, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.1100474624168269e-05, | |
| "loss": 0.3887, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.1078104294337806e-05, | |
| "loss": 0.3969, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.1055728504546835e-05, | |
| "loss": 0.3652, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.1033347368115494e-05, | |
| "loss": 0.3843, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.1010960998391003e-05, | |
| "loss": 0.4162, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0988569508747075e-05, | |
| "loss": 0.3777, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0966173012583367e-05, | |
| "loss": 0.3709, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0943771623324884e-05, | |
| "loss": 0.4504, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0921365454421404e-05, | |
| "loss": 0.374, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0898954619346924e-05, | |
| "loss": 0.3568, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0876539231599069e-05, | |
| "loss": 0.4537, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.085411940469851e-05, | |
| "loss": 0.3287, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0831695252188415e-05, | |
| "loss": 0.4042, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0809266887633849e-05, | |
| "loss": 0.3993, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.078683442462121e-05, | |
| "loss": 0.409, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0764397976757658e-05, | |
| "loss": 0.3566, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0741957657670523e-05, | |
| "loss": 0.4301, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0719513581006751e-05, | |
| "loss": 0.3776, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0697065860432314e-05, | |
| "loss": 0.3645, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0674614609631634e-05, | |
| "loss": 0.3891, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.065215994230702e-05, | |
| "loss": 0.4682, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.062970197217808e-05, | |
| "loss": 0.4094, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0607240812981145e-05, | |
| "loss": 0.3954, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0584776578468698e-05, | |
| "loss": 0.507, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.05623093824088e-05, | |
| "loss": 0.4511, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0539839338584509e-05, | |
| "loss": 0.3269, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0517366560793305e-05, | |
| "loss": 0.3342, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0494891162846515e-05, | |
| "loss": 0.4233, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0472413258568734e-05, | |
| "loss": 0.4296, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0449932961797249e-05, | |
| "loss": 0.3722, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0427450386381463e-05, | |
| "loss": 0.4086, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.040496564618233e-05, | |
| "loss": 0.3875, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.0382478855071751e-05, | |
| "loss": 0.3279, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.0359990126932022e-05, | |
| "loss": 0.4163, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.033749957565525e-05, | |
| "loss": 0.4412, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0315007315142772e-05, | |
| "loss": 0.415, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.029251345930458e-05, | |
| "loss": 0.4251, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0270018122058753e-05, | |
| "loss": 0.4075, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0247521417330864e-05, | |
| "loss": 0.3823, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.0225023459053416e-05, | |
| "loss": 0.4515, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0202524361165255e-05, | |
| "loss": 0.4079, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.018002423761101e-05, | |
| "loss": 0.3785, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0157523202340491e-05, | |
| "loss": 0.3989, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0135021369308138e-05, | |
| "loss": 0.415, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.0112518852472414e-05, | |
| "loss": 0.3894, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.0090015765795265e-05, | |
| "loss": 0.3863, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.0067512223241509e-05, | |
| "loss": 0.3902, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.004500833877828e-05, | |
| "loss": 0.4072, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.002250422637444e-05, | |
| "loss": 0.3688, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1e-05, | |
| "loss": 0.412, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.977495773625562e-06, | |
| "loss": 0.3986, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.954991661221724e-06, | |
| "loss": 0.4599, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.93248777675849e-06, | |
| "loss": 0.4113, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.909984234204738e-06, | |
| "loss": 0.4713, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.887481147527587e-06, | |
| "loss": 0.3793, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.864978630691865e-06, | |
| "loss": 0.4093, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.842476797659509e-06, | |
| "loss": 0.3732, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.819975762388993e-06, | |
| "loss": 0.3746, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.797475638834745e-06, | |
| "loss": 0.3627, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.774976540946589e-06, | |
| "loss": 0.4231, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.752478582669136e-06, | |
| "loss": 0.4466, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.729981877941249e-06, | |
| "loss": 0.4179, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.70748654069542e-06, | |
| "loss": 0.4146, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.684992684857232e-06, | |
| "loss": 0.4035, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.662500424344751e-06, | |
| "loss": 0.3877, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.640009873067981e-06, | |
| "loss": 0.4306, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.617521144928252e-06, | |
| "loss": 0.3997, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.595034353817673e-06, | |
| "loss": 0.4332, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.572549613618537e-06, | |
| "loss": 0.4538, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.550067038202756e-06, | |
| "loss": 0.3556, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.52758674143127e-06, | |
| "loss": 0.4435, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.505108837153489e-06, | |
| "loss": 0.368, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.482633439206697e-06, | |
| "loss": 0.3727, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.460160661415496e-06, | |
| "loss": 0.4113, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.437690617591202e-06, | |
| "loss": 0.3771, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.415223421531308e-06, | |
| "loss": 0.4213, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.392759187018858e-06, | |
| "loss": 0.4006, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.370298027821924e-06, | |
| "loss": 0.3888, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.347840057692981e-06, | |
| "loss": 0.4213, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.325385390368367e-06, | |
| "loss": 0.39, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.30293413956769e-06, | |
| "loss": 0.3883, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.280486418993254e-06, | |
| "loss": 0.3724, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.25804234232948e-06, | |
| "loss": 0.4529, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.23560202324235e-06, | |
| "loss": 0.4267, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.213165575378793e-06, | |
| "loss": 0.4244, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.190733112366158e-06, | |
| "loss": 0.4018, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.168304747811588e-06, | |
| "loss": 0.4573, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.145880595301495e-06, | |
| "loss": 0.4155, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.123460768400935e-06, | |
| "loss": 0.3583, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.101045380653076e-06, | |
| "loss": 0.4034, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.078634545578598e-06, | |
| "loss": 0.4101, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.056228376675118e-06, | |
| "loss": 0.3948, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.033826987416635e-06, | |
| "loss": 0.4181, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.011430491252924e-06, | |
| "loss": 0.4648, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.989039001609002e-06, | |
| "loss": 0.4065, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.966652631884506e-06, | |
| "loss": 0.3337, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.944271495453167e-06, | |
| "loss": 0.4112, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.921895705662194e-06, | |
| "loss": 0.3682, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.899525375831732e-06, | |
| "loss": 0.388, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.877160619254264e-06, | |
| "loss": 0.3774, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.854801549194055e-06, | |
| "loss": 0.3638, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.832448278886567e-06, | |
| "loss": 0.4091, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.810100921537895e-06, | |
| "loss": 0.3823, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.787759590324177e-06, | |
| "loss": 0.4022, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.765424398391047e-06, | |
| "loss": 0.4058, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.743095458853034e-06, | |
| "loss": 0.4222, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.720772884793017e-06, | |
| "loss": 0.4015, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.698456789261617e-06, | |
| "loss": 0.3815, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.676147285276668e-06, | |
| "loss": 0.4609, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.653844485822603e-06, | |
| "loss": 0.3784, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.631548503849915e-06, | |
| "loss": 0.4792, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.609259452274559e-06, | |
| "loss": 0.4391, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.586977443977396e-06, | |
| "loss": 0.3597, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.56470259180362e-06, | |
| "loss": 0.4319, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.54243500856218e-06, | |
| "loss": 0.3945, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.52017480702521e-06, | |
| "loss": 0.3338, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.497922099927469e-06, | |
| "loss": 0.3916, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.475676999965747e-06, | |
| "loss": 0.4171, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.453439619798325e-06, | |
| "loss": 0.3987, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.431210072044371e-06, | |
| "loss": 0.3594, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.408988469283403e-06, | |
| "loss": 0.4404, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.386774924054686e-06, | |
| "loss": 0.3953, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 8.364569548856695e-06, | |
| "loss": 0.422, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 8.342372456146512e-06, | |
| "loss": 0.3908, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 8.320183758339284e-06, | |
| "loss": 0.4302, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 8.29800356780764e-06, | |
| "loss": 0.4296, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 8.275831996881129e-06, | |
| "loss": 0.3956, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.253669157845632e-06, | |
| "loss": 0.3831, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.231515162942824e-06, | |
| "loss": 0.392, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.20937012436959e-06, | |
| "loss": 0.3982, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.18723415427744e-06, | |
| "loss": 0.367, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.165107364771979e-06, | |
| "loss": 0.3784, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.142989867912299e-06, | |
| "loss": 0.3956, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.12088177571044e-06, | |
| "loss": 0.3855, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.098783200130814e-06, | |
| "loss": 0.3981, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.076694253089632e-06, | |
| "loss": 0.3995, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.05461504645434e-06, | |
| "loss": 0.4358, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.032545692043068e-06, | |
| "loss": 0.4232, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.010486301624034e-06, | |
| "loss": 0.3855, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.988436986915005e-06, | |
| "loss": 0.383, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.966397859582713e-06, | |
| "loss": 0.4065, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.944369031242307e-06, | |
| "loss": 0.3865, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.922350613456763e-06, | |
| "loss": 0.4409, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.900342717736354e-06, | |
| "loss": 0.3738, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.878345455538043e-06, | |
| "loss": 0.3814, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.856358938264953e-06, | |
| "loss": 0.4189, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.834383277265794e-06, | |
| "loss": 0.3726, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.812418583834282e-06, | |
| "loss": 0.4467, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.790464969208598e-06, | |
| "loss": 0.4064, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.768522544570818e-06, | |
| "loss": 0.3909, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.746591421046336e-06, | |
| "loss": 0.3512, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.724671709703328e-06, | |
| "loss": 0.3641, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.702763521552154e-06, | |
| "loss": 0.422, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.680866967544841e-06, | |
| "loss": 0.3568, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.658982158574471e-06, | |
| "loss": 0.4292, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.637109205474665e-06, | |
| "loss": 0.3753, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.615248219018982e-06, | |
| "loss": 0.4646, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.5933993099203936e-06, | |
| "loss": 0.3828, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.571562588830698e-06, | |
| "loss": 0.3927, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.5497381663399716e-06, | |
| "loss": 0.4506, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.527926152975999e-06, | |
| "loss": 0.4226, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.506126659203733e-06, | |
| "loss": 0.384, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.484339795424707e-06, | |
| "loss": 0.3545, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.462565671976504e-06, | |
| "loss": 0.3547, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.440804399132173e-06, | |
| "loss": 0.3905, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.419056087099695e-06, | |
| "loss": 0.4131, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.3973208460213985e-06, | |
| "loss": 0.359, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.375598785973429e-06, | |
| "loss": 0.3926, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.35389001696517e-06, | |
| "loss": 0.4022, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.332194648938689e-06, | |
| "loss": 0.3788, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.310512791768199e-06, | |
| "loss": 0.4007, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.288844555259471e-06, | |
| "loss": 0.3935, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.267190049149319e-06, | |
| "loss": 0.3516, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.245549383104993e-06, | |
| "loss": 0.3984, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.223922666723677e-06, | |
| "loss": 0.3763, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.202310009531886e-06, | |
| "loss": 0.4442, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.180711520984953e-06, | |
| "loss": 0.3793, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.159127310466441e-06, | |
| "loss": 0.4404, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.137557487287609e-06, | |
| "loss": 0.3862, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.116002160686851e-06, | |
| "loss": 0.3876, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.09446143982915e-06, | |
| "loss": 0.4251, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.072935433805508e-06, | |
| "loss": 0.4505, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.0514242516324184e-06, | |
| "loss": 0.4018, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.0299280022512875e-06, | |
| "loss": 0.4151, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 7.00844679452791e-06, | |
| "loss": 0.3755, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.986980737251889e-06, | |
| "loss": 0.4132, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.9655299391361156e-06, | |
| "loss": 0.4649, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.944094508816182e-06, | |
| "loss": 0.4246, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.922674554849871e-06, | |
| "loss": 0.431, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.9012701857165755e-06, | |
| "loss": 0.4421, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.879881509816764e-06, | |
| "loss": 0.4143, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.858508635471428e-06, | |
| "loss": 0.3797, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.837151670921534e-06, | |
| "loss": 0.3635, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.815810724327469e-06, | |
| "loss": 0.3592, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.794485903768513e-06, | |
| "loss": 0.4198, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.773177317242257e-06, | |
| "loss": 0.4122, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.7518850726640964e-06, | |
| "loss": 0.3734, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.730609277866644e-06, | |
| "loss": 0.4239, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.709350040599227e-06, | |
| "loss": 0.3756, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.688107468527297e-06, | |
| "loss": 0.3468, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.666881669231922e-06, | |
| "loss": 0.3824, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.645672750209216e-06, | |
| "loss": 0.4066, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.6244808188698075e-06, | |
| "loss": 0.4047, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.603305982538295e-06, | |
| "loss": 0.413, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.582148348452699e-06, | |
| "loss": 0.3811, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.561008023763915e-06, | |
| "loss": 0.3428, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.539885115535187e-06, | |
| "loss": 0.3861, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.518779730741555e-06, | |
| "loss": 0.4412, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.4976919762692965e-06, | |
| "loss": 0.3891, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.476621958915426e-06, | |
| "loss": 0.4305, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.4555697853871055e-06, | |
| "loss": 0.3233, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.434535562301153e-06, | |
| "loss": 0.4035, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.4135193961834565e-06, | |
| "loss": 0.3441, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.392521393468471e-06, | |
| "loss": 0.3551, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.371541660498652e-06, | |
| "loss": 0.3629, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.350580303523947e-06, | |
| "loss": 0.3458, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.329637428701219e-06, | |
| "loss": 0.3772, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.308713142093749e-06, | |
| "loss": 0.3876, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.287807549670663e-06, | |
| "loss": 0.403, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.266920757306429e-06, | |
| "loss": 0.3873, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.246052870780288e-06, | |
| "loss": 0.4317, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.225203995775746e-06, | |
| "loss": 0.3689, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.2043742378800154e-06, | |
| "loss": 0.3835, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.183563702583506e-06, | |
| "loss": 0.4217, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.162772495279265e-06, | |
| "loss": 0.3891, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.1420007212624584e-06, | |
| "loss": 0.3735, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.121248485729832e-06, | |
| "loss": 0.3979, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.100515893779188e-06, | |
| "loss": 0.4062, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.079803050408837e-06, | |
| "loss": 0.4235, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.05911006051708e-06, | |
| "loss": 0.3945, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.038437028901667e-06, | |
| "loss": 0.4598, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.01778406025928e-06, | |
| "loss": 0.39, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.99715125918498e-06, | |
| "loss": 0.3637, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.976538730171708e-06, | |
| "loss": 0.4071, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.955946577609721e-06, | |
| "loss": 0.3802, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.935374905786102e-06, | |
| "loss": 0.3767, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.91482381888419e-06, | |
| "loss": 0.4016, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.89429342098309e-06, | |
| "loss": 0.3565, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.873783816057115e-06, | |
| "loss": 0.3799, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.8532951079752895e-06, | |
| "loss": 0.3815, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.832827400500794e-06, | |
| "loss": 0.4156, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.812380797290465e-06, | |
| "loss": 0.3372, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.791955401894248e-06, | |
| "loss": 0.4424, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.771551317754691e-06, | |
| "loss": 0.3707, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.75116864820641e-06, | |
| "loss": 0.3603, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.730807496475568e-06, | |
| "loss": 0.3933, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.710467965679356e-06, | |
| "loss": 0.4183, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.690150158825462e-06, | |
| "loss": 0.3446, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.6698541788115644e-06, | |
| "loss": 0.4104, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.649580128424792e-06, | |
| "loss": 0.3721, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.629328110341218e-06, | |
| "loss": 0.3887, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.609098227125334e-06, | |
| "loss": 0.4046, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.588890581229531e-06, | |
| "loss": 0.3951, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.568705274993584e-06, | |
| "loss": 0.4098, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.548542410644132e-06, | |
| "loss": 0.4195, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.528402090294142e-06, | |
| "loss": 0.3786, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.508284415942442e-06, | |
| "loss": 0.3708, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.488189489473133e-06, | |
| "loss": 0.3573, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.468117412655147e-06, | |
| "loss": 0.4249, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.448068287141663e-06, | |
| "loss": 0.4163, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.428042214469662e-06, | |
| "loss": 0.4761, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.4080392960593355e-06, | |
| "loss": 0.4418, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.388059633213651e-06, | |
| "loss": 0.3294, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.368103327117768e-06, | |
| "loss": 0.3234, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.348170478838579e-06, | |
| "loss": 0.4148, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.328261189324166e-06, | |
| "loss": 0.4313, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.308375559403307e-06, | |
| "loss": 0.3532, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.288513689784951e-06, | |
| "loss": 0.4148, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.26867568105772e-06, | |
| "loss": 0.3999, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.2488616336893915e-06, | |
| "loss": 0.1375, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.229071648026398e-06, | |
| "loss": 0.1576, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 5.209305824293307e-06, | |
| "loss": 0.155, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 5.189564262592326e-06, | |
| "loss": 0.1728, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 5.1698470629027845e-06, | |
| "loss": 0.156, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 5.150154325080636e-06, | |
| "loss": 0.137, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 5.130486148857952e-06, | |
| "loss": 0.1431, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.110842633842406e-06, | |
| "loss": 0.1087, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.0912238795167845e-06, | |
| "loss": 0.1651, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.071629985238474e-06, | |
| "loss": 0.1251, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.05206105023895e-06, | |
| "loss": 0.1702, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.032517173623306e-06, | |
| "loss": 0.1325, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 5.012998454369701e-06, | |
| "loss": 0.1513, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.993504991328914e-06, | |
| "loss": 0.1285, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.974036883223798e-06, | |
| "loss": 0.1377, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.954594228648807e-06, | |
| "loss": 0.1354, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.935177126069485e-06, | |
| "loss": 0.1333, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.91578567382197e-06, | |
| "loss": 0.1567, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.896419970112499e-06, | |
| "loss": 0.1529, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.877080113016898e-06, | |
| "loss": 0.1289, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.857766200480115e-06, | |
| "loss": 0.1245, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.838478330315677e-06, | |
| "loss": 0.1376, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.819216600205254e-06, | |
| "loss": 0.1443, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.799981107698097e-06, | |
| "loss": 0.1374, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.780771950210616e-06, | |
| "loss": 0.1448, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.761589225025811e-06, | |
| "loss": 0.1322, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.742433029292856e-06, | |
| "loss": 0.1591, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.723303460026538e-06, | |
| "loss": 0.1313, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.704200614106813e-06, | |
| "loss": 0.1467, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.685124588278297e-06, | |
| "loss": 0.1279, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.6660754791497755e-06, | |
| "loss": 0.1491, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.647053383193717e-06, | |
| "loss": 0.1476, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.628058396745787e-06, | |
| "loss": 0.1399, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.609090616004355e-06, | |
| "loss": 0.1502, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.590150137030009e-06, | |
| "loss": 0.1557, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.571237055745074e-06, | |
| "loss": 0.1463, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.552351467933115e-06, | |
| "loss": 0.1479, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.533493469238465e-06, | |
| "loss": 0.1242, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.514663155165731e-06, | |
| "loss": 0.1287, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.495860621079315e-06, | |
| "loss": 0.1331, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.477085962202931e-06, | |
| "loss": 0.1428, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.458339273619119e-06, | |
| "loss": 0.1703, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.439620650268771e-06, | |
| "loss": 0.1218, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.420930186950631e-06, | |
| "loss": 0.1369, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.402267978320854e-06, | |
| "loss": 0.1469, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.383634118892472e-06, | |
| "loss": 0.1457, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.365028703034976e-06, | |
| "loss": 0.1226, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.3464518249737766e-06, | |
| "loss": 0.1276, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.327903578789785e-06, | |
| "loss": 0.123, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.309384058418884e-06, | |
| "loss": 0.1359, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.290893357651502e-06, | |
| "loss": 0.1397, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.272431570132092e-06, | |
| "loss": 0.1304, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.253998789358683e-06, | |
| "loss": 0.138, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.23559510868242e-06, | |
| "loss": 0.1149, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.217220621307043e-06, | |
| "loss": 0.144, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.198875420288477e-06, | |
| "loss": 0.1167, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.180559598534297e-06, | |
| "loss": 0.1324, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.162273248803322e-06, | |
| "loss": 0.1249, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.144016463705081e-06, | |
| "loss": 0.1625, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.125789335699404e-06, | |
| "loss": 0.1343, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.107591957095903e-06, | |
| "loss": 0.1267, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.08942442005354e-06, | |
| "loss": 0.1405, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.071286816580142e-06, | |
| "loss": 0.1118, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.053179238531943e-06, | |
| "loss": 0.133, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.035101777613113e-06, | |
| "loss": 0.1288, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.017054525375299e-06, | |
| "loss": 0.1323, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.999037573217157e-06, | |
| "loss": 0.1009, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.981051012383893e-06, | |
| "loss": 0.1326, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.963094933966797e-06, | |
| "loss": 0.1204, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.945169428902784e-06, | |
| "loss": 0.1299, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.927274587973935e-06, | |
| "loss": 0.1232, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.909410501807033e-06, | |
| "loss": 0.1109, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.8915772608731066e-06, | |
| "loss": 0.1447, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.873774955486973e-06, | |
| "loss": 0.1354, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.856003675806777e-06, | |
| "loss": 0.1197, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.838263511833542e-06, | |
| "loss": 0.1201, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.820554553410693e-06, | |
| "loss": 0.1136, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.8028768902236458e-06, | |
| "loss": 0.1357, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.78523061179929e-06, | |
| "loss": 0.1155, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.7676158075056024e-06, | |
| "loss": 0.1288, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.7500325665511337e-06, | |
| "loss": 0.1269, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.7324809779846116e-06, | |
| "loss": 0.1224, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.7149611306944356e-06, | |
| "loss": 0.151, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.697473113408282e-06, | |
| "loss": 0.1219, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.680017014692604e-06, | |
| "loss": 0.1214, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.662592922952218e-06, | |
| "loss": 0.1232, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.645200926429844e-06, | |
| "loss": 0.1389, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.627841113205652e-06, | |
| "loss": 0.1211, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.610513571196832e-06, | |
| "loss": 0.1298, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.5932183881571305e-06, | |
| "loss": 0.1422, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.5759556516764205e-06, | |
| "loss": 0.1224, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.558725449180247e-06, | |
| "loss": 0.1414, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.541527867929403e-06, | |
| "loss": 0.1035, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.524362995019455e-06, | |
| "loss": 0.1411, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.507230917380332e-06, | |
| "loss": 0.1205, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.4901317217758767e-06, | |
| "loss": 0.1295, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.4730654948033957e-06, | |
| "loss": 0.1374, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.456032322893237e-06, | |
| "loss": 0.1247, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.4390322923083385e-06, | |
| "loss": 0.18, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.422065489143799e-06, | |
| "loss": 0.1515, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.4051319993264397e-06, | |
| "loss": 0.1152, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.388231908614371e-06, | |
| "loss": 0.1433, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.3713653025965544e-06, | |
| "loss": 0.122, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.3545322666923717e-06, | |
| "loss": 0.1623, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.3377328861511927e-06, | |
| "loss": 0.1137, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.3209672460519426e-06, | |
| "loss": 0.1218, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.3042354313026702e-06, | |
| "loss": 0.1392, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.287537526640121e-06, | |
| "loss": 0.1291, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.2708736166293064e-06, | |
| "loss": 0.1232, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.254243785663065e-06, | |
| "loss": 0.142, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.237648117961665e-06, | |
| "loss": 0.1385, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.2210866975723333e-06, | |
| "loss": 0.1123, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.2045596083688814e-06, | |
| "loss": 0.154, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.188066934051226e-06, | |
| "loss": 0.1344, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.1716087581450193e-06, | |
| "loss": 0.1303, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.1551851640011755e-06, | |
| "loss": 0.1296, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.1387962347954936e-06, | |
| "loss": 0.1253, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 3.1224420535281974e-06, | |
| "loss": 0.1329, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 3.1061227030235442e-06, | |
| "loss": 0.1223, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 3.0898382659293902e-06, | |
| "loss": 0.1265, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 3.073588824716777e-06, | |
| "loss": 0.1363, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 3.05737446167951e-06, | |
| "loss": 0.1272, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 3.041195258933749e-06, | |
| "loss": 0.1149, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 3.025051298417585e-06, | |
| "loss": 0.1465, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 3.008942661890627e-06, | |
| "loss": 0.1324, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.9928694309335917e-06, | |
| "loss": 0.1371, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.976831686947884e-06, | |
| "loss": 0.1914, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.960829511155191e-06, | |
| "loss": 0.1327, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.944862984597068e-06, | |
| "loss": 0.1497, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.9289321881345257e-06, | |
| "loss": 0.1337, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.913037202447625e-06, | |
| "loss": 0.1646, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.897178108035067e-06, | |
| "loss": 0.1406, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.8813549852137824e-06, | |
| "loss": 0.1359, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.8655679141185287e-06, | |
| "loss": 0.1446, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.8498169747014824e-06, | |
| "loss": 0.1546, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.8341022467318335e-06, | |
| "loss": 0.1327, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.818423809795384e-06, | |
| "loss": 0.1183, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.802781743294143e-06, | |
| "loss": 0.125, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.787176126445923e-06, | |
| "loss": 0.1312, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.771607038283942e-06, | |
| "loss": 0.1296, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.756074557656424e-06, | |
| "loss": 0.1401, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.7405787632261936e-06, | |
| "loss": 0.1421, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.725119733470284e-06, | |
| "loss": 0.1454, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.709697546679537e-06, | |
| "loss": 0.1125, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.6943122809582e-06, | |
| "loss": 0.1072, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.678964014223553e-06, | |
| "loss": 0.1211, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.663652824205476e-06, | |
| "loss": 0.1639, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.6483787884461022e-06, | |
| "loss": 0.1464, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.6331419842993746e-06, | |
| "loss": 0.1379, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.617942488930705e-06, | |
| "loss": 0.171, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.6027803793165353e-06, | |
| "loss": 0.1443, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.5876557322439922e-06, | |
| "loss": 0.1848, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.5725686243104586e-06, | |
| "loss": 0.0993, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.5575191319232127e-06, | |
| "loss": 0.1149, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.5425073312990334e-06, | |
| "loss": 0.1031, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.52753329846381e-06, | |
| "loss": 0.1051, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.5125971092521607e-06, | |
| "loss": 0.1666, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.4976988393070477e-06, | |
| "loss": 0.121, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.4828385640793974e-06, | |
| "loss": 0.1195, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.468016358827712e-06, | |
| "loss": 0.1772, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.4532322986176926e-06, | |
| "loss": 0.1517, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.4384864583218592e-06, | |
| "loss": 0.1352, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.4237789126191715e-06, | |
| "loss": 0.1263, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.409109735994647e-06, | |
| "loss": 0.1167, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.3944790027389888e-06, | |
| "loss": 0.1322, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.3798867869482047e-06, | |
| "loss": 0.1407, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.3653331625232367e-06, | |
| "loss": 0.1179, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.350818203169586e-06, | |
| "loss": 0.138, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.336341982396928e-06, | |
| "loss": 0.1265, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.321904573518765e-06, | |
| "loss": 0.152, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.307506049652031e-06, | |
| "loss": 0.1226, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.2931464837167303e-06, | |
| "loss": 0.1395, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.2788259484355712e-06, | |
| "loss": 0.1726, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.2645445163335944e-06, | |
| "loss": 0.1148, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.250302259737803e-06, | |
| "loss": 0.141, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.2360992507768055e-06, | |
| "loss": 0.1247, | |
| "step": 1137 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.2219355613804406e-06, | |
| "loss": 0.1256, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.2078112632794092e-06, | |
| "loss": 0.1472, | |
| "step": 1139 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.1937264280049365e-06, | |
| "loss": 0.1645, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.179681126888371e-06, | |
| "loss": 0.125, | |
| "step": 1141 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.1656754310608663e-06, | |
| "loss": 0.1197, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.1517094114529746e-06, | |
| "loss": 0.1462, | |
| "step": 1143 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.137783138794335e-06, | |
| "loss": 0.1217, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.1238966836132693e-06, | |
| "loss": 0.1334, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 2.1100501162364707e-06, | |
| "loss": 0.1265, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 2.096243506788602e-06, | |
| "loss": 0.1675, | |
| "step": 1147 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 2.0824769251919775e-06, | |
| "loss": 0.1263, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 2.0687504411661897e-06, | |
| "loss": 0.1424, | |
| "step": 1149 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 2.055064124227758e-06, | |
| "loss": 0.1503, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 2.0414180436897846e-06, | |
| "loss": 0.1276, | |
| "step": 1151 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 2.027812268661592e-06, | |
| "loss": 0.1288, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 2.014246868048385e-06, | |
| "loss": 0.1315, | |
| "step": 1153 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 2.0007219105508935e-06, | |
| "loss": 0.0959, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.987237464665024e-06, | |
| "loss": 0.1163, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.9737935986815205e-06, | |
| "loss": 0.1613, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.9603903806856106e-06, | |
| "loss": 0.1671, | |
| "step": 1157 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.9470278785566653e-06, | |
| "loss": 0.1055, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.933706159967854e-06, | |
| "loss": 0.1701, | |
| "step": 1159 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.9204252923858003e-06, | |
| "loss": 0.1371, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.9071853430702415e-06, | |
| "loss": 0.1302, | |
| "step": 1161 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.8939863790736923e-06, | |
| "loss": 0.1125, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.8808284672410904e-06, | |
| "loss": 0.1174, | |
| "step": 1163 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.8677116742094858e-06, | |
| "loss": 0.1314, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.8546360664076657e-06, | |
| "loss": 0.1172, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.8416017100558592e-06, | |
| "loss": 0.155, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.8286086711653605e-06, | |
| "loss": 0.1217, | |
| "step": 1167 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.8156570155382357e-06, | |
| "loss": 0.136, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.8027468087669486e-06, | |
| "loss": 0.1284, | |
| "step": 1169 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.7898781162340683e-06, | |
| "loss": 0.1278, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.7770510031119103e-06, | |
| "loss": 0.1354, | |
| "step": 1171 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.764265534362205e-06, | |
| "loss": 0.1142, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.7515217747358015e-06, | |
| "loss": 0.1251, | |
| "step": 1173 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.7388197887722914e-06, | |
| "loss": 0.1586, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.7261596407997306e-06, | |
| "loss": 0.1449, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.7135413949342706e-06, | |
| "loss": 0.1258, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.7009651150798713e-06, | |
| "loss": 0.1236, | |
| "step": 1177 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.6884308649279413e-06, | |
| "loss": 0.1471, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.6759387079570532e-06, | |
| "loss": 0.1411, | |
| "step": 1179 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.6634887074325844e-06, | |
| "loss": 0.1245, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.6510809264064253e-06, | |
| "loss": 0.1505, | |
| "step": 1181 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.6387154277166484e-06, | |
| "loss": 0.1303, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.6263922739871886e-06, | |
| "loss": 0.109, | |
| "step": 1183 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.6141115276275298e-06, | |
| "loss": 0.1146, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.6018732508323886e-06, | |
| "loss": 0.1434, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.5896775055813973e-06, | |
| "loss": 0.1185, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.5775243536387908e-06, | |
| "loss": 0.1584, | |
| "step": 1187 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.565413856553095e-06, | |
| "loss": 0.1305, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.553346075656813e-06, | |
| "loss": 0.1384, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.5413210720661187e-06, | |
| "loss": 0.1438, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.5293389066805397e-06, | |
| "loss": 0.1819, | |
| "step": 1191 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.5173996401826563e-06, | |
| "loss": 0.1646, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.5055033330377912e-06, | |
| "loss": 0.1529, | |
| "step": 1193 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.493650045493703e-06, | |
| "loss": 0.1514, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.4818398375802834e-06, | |
| "loss": 0.1263, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.470072769109242e-06, | |
| "loss": 0.1204, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.4583488996738292e-06, | |
| "loss": 0.1314, | |
| "step": 1197 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.4466682886485007e-06, | |
| "loss": 0.1135, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.4350309951886488e-06, | |
| "loss": 0.1354, | |
| "step": 1199 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.4234370782302742e-06, | |
| "loss": 0.1266, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.4118865964897143e-06, | |
| "loss": 0.1401, | |
| "step": 1201 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.4003796084633204e-06, | |
| "loss": 0.1403, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.388916172427187e-06, | |
| "loss": 0.1738, | |
| "step": 1203 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3774963464368295e-06, | |
| "loss": 0.1419, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3661201883269159e-06, | |
| "loss": 0.1468, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3547877557109546e-06, | |
| "loss": 0.1182, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3434991059810153e-06, | |
| "loss": 0.1221, | |
| "step": 1207 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3322542963074314e-06, | |
| "loss": 0.1207, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.3210533836385086e-06, | |
| "loss": 0.1427, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.3098964247002499e-06, | |
| "loss": 0.1529, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.298783475996046e-06, | |
| "loss": 0.1199, | |
| "step": 1211 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.287714593806415e-06, | |
| "loss": 0.1342, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.2766898341886914e-06, | |
| "loss": 0.1159, | |
| "step": 1213 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2657092529767644e-06, | |
| "loss": 0.1635, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2547729057807811e-06, | |
| "loss": 0.1378, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2438808479868715e-06, | |
| "loss": 0.1098, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2330331347568635e-06, | |
| "loss": 0.1323, | |
| "step": 1217 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2222298210280103e-06, | |
| "loss": 0.1139, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.2114709615127052e-06, | |
| "loss": 0.1151, | |
| "step": 1219 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.200756610698205e-06, | |
| "loss": 0.1329, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.1900868228463603e-06, | |
| "loss": 0.1559, | |
| "step": 1221 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.1794616519933344e-06, | |
| "loss": 0.1805, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.1688811519493326e-06, | |
| "loss": 0.1445, | |
| "step": 1223 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.1583453762983289e-06, | |
| "loss": 0.1545, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.1478543783977948e-06, | |
| "loss": 0.1395, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.137408211378429e-06, | |
| "loss": 0.1438, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.1270069281438867e-06, | |
| "loss": 0.1279, | |
| "step": 1227 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.1166505813705187e-06, | |
| "loss": 0.1738, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.106339223507088e-06, | |
| "loss": 0.139, | |
| "step": 1229 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.09607290677453e-06, | |
| "loss": 0.1378, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.0858516831656595e-06, | |
| "loss": 0.1033, | |
| "step": 1231 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.0756756044449358e-06, | |
| "loss": 0.1385, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0655447221481685e-06, | |
| "loss": 0.1129, | |
| "step": 1233 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0554590875822924e-06, | |
| "loss": 0.1304, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0454187518250735e-06, | |
| "loss": 0.1216, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.035423765724879e-06, | |
| "loss": 0.1086, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0254741799003975e-06, | |
| "loss": 0.133, | |
| "step": 1237 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 1.015570044740396e-06, | |
| "loss": 0.1455, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 1.0057114104034605e-06, | |
| "loss": 0.1397, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.958983268177425e-07, | |
| "loss": 0.1206, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.86130843680706e-07, | |
| "loss": 0.2065, | |
| "step": 1241 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.764090104588741e-07, | |
| "loss": 0.1011, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.667328763875816e-07, | |
| "loss": 0.1228, | |
| "step": 1243 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.571024904707238e-07, | |
| "loss": 0.1419, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.475179014805092e-07, | |
| "loss": 0.1416, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.379791579572118e-07, | |
| "loss": 0.1389, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 9.284863082089223e-07, | |
| "loss": 0.1305, | |
| "step": 1247 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 9.190394003113123e-07, | |
| "loss": 0.1265, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 9.0963848210738e-07, | |
| "loss": 0.1218, | |
| "step": 1249 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 9.002836012072169e-07, | |
| "loss": 0.1164, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.909748049877587e-07, | |
| "loss": 0.1538, | |
| "step": 1251 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.817121405925543e-07, | |
| "loss": 0.1311, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.724956549315177e-07, | |
| "loss": 0.1437, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.633253946806974e-07, | |
| "loss": 0.1079, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.542014062820369e-07, | |
| "loss": 0.1087, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.451237359431397e-07, | |
| "loss": 0.145, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 8.360924296370376e-07, | |
| "loss": 0.1226, | |
| "step": 1257 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 8.271075331019541e-07, | |
| "loss": 0.1472, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 8.181690918410756e-07, | |
| "loss": 0.1189, | |
| "step": 1259 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 8.092771511223185e-07, | |
| "loss": 0.1354, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 8.004317559781049e-07, | |
| "loss": 0.1332, | |
| "step": 1261 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.916329512051235e-07, | |
| "loss": 0.101, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.828807813641226e-07, | |
| "loss": 0.1158, | |
| "step": 1263 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.741752907796584e-07, | |
| "loss": 0.1123, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.655165235398987e-07, | |
| "loss": 0.1483, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.56904523496369e-07, | |
| "loss": 0.1442, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.483393342637635e-07, | |
| "loss": 0.1097, | |
| "step": 1267 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.398209992196914e-07, | |
| "loss": 0.1645, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.313495615044874e-07, | |
| "loss": 0.1222, | |
| "step": 1269 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.229250640209661e-07, | |
| "loss": 0.1379, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.145475494342302e-07, | |
| "loss": 0.1517, | |
| "step": 1271 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.062170601714302e-07, | |
| "loss": 0.1115, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.979336384215695e-07, | |
| "loss": 0.1118, | |
| "step": 1273 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.896973261352779e-07, | |
| "loss": 0.1123, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.815081650246048e-07, | |
| "loss": 0.1162, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.73366196562808e-07, | |
| "loss": 0.1254, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.652714619841405e-07, | |
| "loss": 0.1539, | |
| "step": 1277 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.572240022836451e-07, | |
| "loss": 0.1437, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.492238582169453e-07, | |
| "loss": 0.1425, | |
| "step": 1279 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.412710703000368e-07, | |
| "loss": 0.1366, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.33365678809088e-07, | |
| "loss": 0.0997, | |
| "step": 1281 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.255077237802288e-07, | |
| "loss": 0.1183, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.176972450093543e-07, | |
| "loss": 0.1212, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.099342820519183e-07, | |
| "loss": 0.1736, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.022188742227375e-07, | |
| "loss": 0.126, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.945510605957861e-07, | |
| "loss": 0.1263, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.869308800040075e-07, | |
| "loss": 0.134, | |
| "step": 1287 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.79358371039106e-07, | |
| "loss": 0.1443, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.718335720513602e-07, | |
| "loss": 0.1356, | |
| "step": 1289 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.643565211494285e-07, | |
| "loss": 0.1171, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.569272562001471e-07, | |
| "loss": 0.1389, | |
| "step": 1291 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.495458148283505e-07, | |
| "loss": 0.1538, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.422122344166736e-07, | |
| "loss": 0.0964, | |
| "step": 1293 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.349265521053603e-07, | |
| "loss": 0.142, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.276888047920836e-07, | |
| "loss": 0.1438, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.204990291317535e-07, | |
| "loss": 0.1527, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.13357261536327e-07, | |
| "loss": 0.1116, | |
| "step": 1297 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.062635381746362e-07, | |
| "loss": 0.1337, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.992178949721893e-07, | |
| "loss": 0.1161, | |
| "step": 1299 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.92220367611006e-07, | |
| "loss": 0.132, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.852709915294196e-07, | |
| "loss": 0.1109, | |
| "step": 1301 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.783698019219118e-07, | |
| "loss": 0.1475, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.715168337389231e-07, | |
| "loss": 0.131, | |
| "step": 1303 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.647121216866857e-07, | |
| "loss": 0.139, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.5795570022703964e-07, | |
| "loss": 0.144, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.512476035772628e-07, | |
| "loss": 0.1128, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.4458786570989786e-07, | |
| "loss": 0.1319, | |
| "step": 1307 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.3797652035257544e-07, | |
| "loss": 0.1509, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.314136009878511e-07, | |
| "loss": 0.1287, | |
| "step": 1309 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.248991408530279e-07, | |
| "loss": 0.126, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.1843317293999374e-07, | |
| "loss": 0.1323, | |
| "step": 1311 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.1201572999505e-07, | |
| "loss": 0.1479, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.0564684451875005e-07, | |
| "loss": 0.1047, | |
| "step": 1313 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 3.9932654876573164e-07, | |
| "loss": 0.1266, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.9305487474455285e-07, | |
| "loss": 0.1307, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.8683185421753313e-07, | |
| "loss": 0.114, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.806575187005901e-07, | |
| "loss": 0.1167, | |
| "step": 1317 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.74531899463082e-07, | |
| "loss": 0.1252, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.684550275276455e-07, | |
| "loss": 0.0919, | |
| "step": 1319 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.6242693367004365e-07, | |
| "loss": 0.1265, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.564476484190016e-07, | |
| "loss": 0.1309, | |
| "step": 1321 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.5051720205606877e-07, | |
| "loss": 0.1299, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.4463562461544255e-07, | |
| "loss": 0.1489, | |
| "step": 1323 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.38802945883836e-07, | |
| "loss": 0.1881, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.330191954003159e-07, | |
| "loss": 0.141, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.2728440245615724e-07, | |
| "loss": 0.1507, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.215985960946943e-07, | |
| "loss": 0.1539, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.1596180511117235e-07, | |
| "loss": 0.1264, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.10374058052606e-07, | |
| "loss": 0.1448, | |
| "step": 1329 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.048353832176221e-07, | |
| "loss": 0.1447, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.9934580865634056e-07, | |
| "loss": 0.1394, | |
| "step": 1331 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.939053621702015e-07, | |
| "loss": 0.145, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.8851407131185506e-07, | |
| "loss": 0.1217, | |
| "step": 1333 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.83171963384995e-07, | |
| "loss": 0.1365, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.7787906544424093e-07, | |
| "loss": 0.1375, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.7263540429498747e-07, | |
| "loss": 0.1296, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.6744100649327976e-07, | |
| "loss": 0.1189, | |
| "step": 1337 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.6229589834566807e-07, | |
| "loss": 0.1422, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.572001059090812e-07, | |
| "loss": 0.1135, | |
| "step": 1339 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.5215365499069446e-07, | |
| "loss": 0.1584, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.4715657114779523e-07, | |
| "loss": 0.1633, | |
| "step": 1341 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.4220887968765873e-07, | |
| "loss": 0.1273, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.373106056674146e-07, | |
| "loss": 0.1841, | |
| "step": 1343 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.3246177389392388e-07, | |
| "loss": 0.1159, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.276624089236501e-07, | |
| "loss": 0.149, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.229125350625394e-07, | |
| "loss": 0.1444, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.1821217636589176e-07, | |
| "loss": 0.1503, | |
| "step": 1347 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 2.1356135663824328e-07, | |
| "loss": 0.126, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 2.0896009943324634e-07, | |
| "loss": 0.1204, | |
| "step": 1349 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 2.0440842805354522e-07, | |
| "loss": 0.1353, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.99906365550665e-07, | |
| "loss": 0.1187, | |
| "step": 1351 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.9545393472488738e-07, | |
| "loss": 0.1273, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.9105115812514065e-07, | |
| "loss": 0.1231, | |
| "step": 1353 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.866980580488842e-07, | |
| "loss": 0.1239, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.8239465654199651e-07, | |
| "loss": 0.1326, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.7814097539865626e-07, | |
| "loss": 0.1117, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.7393703616124802e-07, | |
| "loss": 0.1311, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.6978286012023225e-07, | |
| "loss": 0.1222, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.6567846831405664e-07, | |
| "loss": 0.1516, | |
| "step": 1359 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.6162388152903498e-07, | |
| "loss": 0.1456, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.5761912029925387e-07, | |
| "loss": 0.1675, | |
| "step": 1361 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.5366420490645738e-07, | |
| "loss": 0.122, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.4975915537995266e-07, | |
| "loss": 0.1102, | |
| "step": 1363 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.4590399149650769e-07, | |
| "loss": 0.0966, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.4209873278024478e-07, | |
| "loss": 0.0929, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.3834339850254952e-07, | |
| "loss": 0.1417, | |
| "step": 1366 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.3463800768196866e-07, | |
| "loss": 0.0977, | |
| "step": 1367 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.309825790841146e-07, | |
| "loss": 0.1279, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.273771312215699e-07, | |
| "loss": 0.1538, | |
| "step": 1369 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.2382168235379742e-07, | |
| "loss": 0.1314, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.2031625048704144e-07, | |
| "loss": 0.1357, | |
| "step": 1371 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.1686085337423991e-07, | |
| "loss": 0.1181, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.134555085149347e-07, | |
| "loss": 0.1369, | |
| "step": 1373 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.1010023315518592e-07, | |
| "loss": 0.1503, | |
| "step": 1374 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.0679504428747545e-07, | |
| "loss": 0.1277, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.0353995865063138e-07, | |
| "loss": 0.0987, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 1.0033499272973702e-07, | |
| "loss": 0.1382, | |
| "step": 1377 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.718016275604759e-08, | |
| "loss": 0.1172, | |
| "step": 1378 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.407548470691252e-08, | |
| "loss": 0.1409, | |
| "step": 1379 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.10209743056889e-08, | |
| "loss": 0.1493, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 8.801664702166368e-08, | |
| "loss": 0.1508, | |
| "step": 1381 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 8.506251806997934e-08, | |
| "loss": 0.1439, | |
| "step": 1382 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 8.21586024115506e-08, | |
| "loss": 0.121, | |
| "step": 1383 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 7.930491475299229e-08, | |
| "loss": 0.1743, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 7.650146954654491e-08, | |
| "loss": 0.1235, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 7.37482809900003e-08, | |
| "loss": 0.1337, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 7.104536302662834e-08, | |
| "loss": 0.1305, | |
| "step": 1387 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.839272934511143e-08, | |
| "loss": 0.1374, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.579039337946791e-08, | |
| "loss": 0.1184, | |
| "step": 1389 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.323836830899321e-08, | |
| "loss": 0.1536, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.073666705818104e-08, | |
| "loss": 0.1178, | |
| "step": 1391 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.828530229667228e-08, | |
| "loss": 0.1527, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.588428643917509e-08, | |
| "loss": 0.1192, | |
| "step": 1393 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.353363164541825e-08, | |
| "loss": 0.1218, | |
| "step": 1394 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.123334982007566e-08, | |
| "loss": 0.1219, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.898345261271531e-08, | |
| "loss": 0.1588, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.678395141773373e-08, | |
| "loss": 0.1137, | |
| "step": 1397 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.463485737430606e-08, | |
| "loss": 0.1579, | |
| "step": 1398 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.253618136631943e-08, | |
| "loss": 0.136, | |
| "step": 1399 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.0487934022328533e-08, | |
| "loss": 0.1502, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.849012571549349e-08, | |
| "loss": 0.1402, | |
| "step": 1401 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.654276656353206e-08, | |
| "loss": 0.0991, | |
| "step": 1402 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.464586642866752e-08, | |
| "loss": 0.1216, | |
| "step": 1403 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.27994349175742e-08, | |
| "loss": 0.1323, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.1003481381337576e-08, | |
| "loss": 0.1336, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.9258014915399813e-08, | |
| "loss": 0.1631, | |
| "step": 1406 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.756304435951429e-08, | |
| "loss": 0.1292, | |
| "step": 1407 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.591857829770672e-08, | |
| "loss": 0.1323, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.4324625058222974e-08, | |
| "loss": 0.1153, | |
| "step": 1409 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.278119271349466e-08, | |
| "loss": 0.1358, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 2.1288289080092505e-08, | |
| "loss": 0.1219, | |
| "step": 1411 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.984592171869082e-08, | |
| "loss": 0.1123, | |
| "step": 1412 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.8454097934027524e-08, | |
| "loss": 0.1422, | |
| "step": 1413 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.711282477486642e-08, | |
| "loss": 0.1373, | |
| "step": 1414 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.582210903396275e-08, | |
| "loss": 0.1324, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.4581957248026579e-08, | |
| "loss": 0.1436, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.3392375697696136e-08, | |
| "loss": 0.1248, | |
| "step": 1417 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.2253370407495636e-08, | |
| "loss": 0.1508, | |
| "step": 1418 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.1164947145815287e-08, | |
| "loss": 0.1538, | |
| "step": 1419 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.0127111424872437e-08, | |
| "loss": 0.1478, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 9.139868500693816e-09, | |
| "loss": 0.1182, | |
| "step": 1421 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.203223373078883e-09, | |
| "loss": 0.158, | |
| "step": 1422 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.317180785582079e-09, | |
| "loss": 0.157, | |
| "step": 1423 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 6.481745225485059e-09, | |
| "loss": 0.1119, | |
| "step": 1424 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.696920923774496e-09, | |
| "loss": 0.1396, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.962711855120983e-09, | |
| "loss": 0.1264, | |
| "step": 1426 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.279121737859049e-09, | |
| "loss": 0.1356, | |
| "step": 1427 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 3.6461540339682855e-09, | |
| "loss": 0.1512, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 3.063811949056694e-09, | |
| "loss": 0.1287, | |
| "step": 1429 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.532098432341812e-09, | |
| "loss": 0.1599, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.051016176637388e-09, | |
| "loss": 0.1351, | |
| "step": 1431 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 1.6205676183411733e-09, | |
| "loss": 0.1352, | |
| "step": 1432 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 1.2407549374204852e-09, | |
| "loss": 0.1368, | |
| "step": 1433 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 9.115800574022171e-10, | |
| "loss": 0.1116, | |
| "step": 1434 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 6.330446453617356e-10, | |
| "loss": 0.1175, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.0515011191621933e-10, | |
| "loss": 0.1471, | |
| "step": 1436 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.2789761121688736e-10, | |
| "loss": 0.1406, | |
| "step": 1437 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.0128804094233779e-10, | |
| "loss": 0.0964, | |
| "step": 1438 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.5322042296327398e-11, | |
| "loss": 0.103, | |
| "step": 1439 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.1147, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 1440, | |
| "total_flos": 2.2020327372069274e+17, | |
| "train_loss": 0.48756657937645087, | |
| "train_runtime": 36917.0164, | |
| "train_samples_per_second": 2.499, | |
| "train_steps_per_second": 0.039 | |
| } | |
| ], | |
| "max_steps": 1440, | |
| "num_train_epochs": 3, | |
| "total_flos": 2.2020327372069274e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |