| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9998852553069421, | |
| "eval_steps": 500, | |
| "global_step": 4357, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.587155963302753e-07, | |
| "loss": 1.5959, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.2935779816513764e-06, | |
| "loss": 1.6335, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.587155963302753e-06, | |
| "loss": 1.6419, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 6.880733944954129e-06, | |
| "loss": 1.6029, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 9.174311926605506e-06, | |
| "loss": 1.6276, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.1467889908256882e-05, | |
| "loss": 1.6056, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.3761467889908258e-05, | |
| "loss": 1.585, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.6055045871559634e-05, | |
| "loss": 1.5484, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.834862385321101e-05, | |
| "loss": 1.4846, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.0642201834862388e-05, | |
| "loss": 1.4423, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.2935779816513765e-05, | |
| "loss": 1.4069, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.5229357798165138e-05, | |
| "loss": 1.3276, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.7522935779816515e-05, | |
| "loss": 1.3297, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.9816513761467892e-05, | |
| "loss": 1.3042, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.211009174311927e-05, | |
| "loss": 1.297, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.4403669724770645e-05, | |
| "loss": 1.2499, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.669724770642202e-05, | |
| "loss": 1.2709, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.89908256880734e-05, | |
| "loss": 1.2304, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.1284403669724776e-05, | |
| "loss": 1.1782, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.3577981651376146e-05, | |
| "loss": 1.1919, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.587155963302753e-05, | |
| "loss": 1.1513, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.81651376146789e-05, | |
| "loss": 1.1253, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.0458715596330276e-05, | |
| "loss": 1.1109, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.2752293577981646e-05, | |
| "loss": 1.1154, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.504587155963303e-05, | |
| "loss": 1.092, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.733944954128441e-05, | |
| "loss": 1.1048, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.9633027522935784e-05, | |
| "loss": 1.1025, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.192660550458716e-05, | |
| "loss": 1.0869, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.422018348623854e-05, | |
| "loss": 1.039, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.651376146788991e-05, | |
| "loss": 1.0572, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.880733944954129e-05, | |
| "loss": 1.0799, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 7.110091743119265e-05, | |
| "loss": 1.0541, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 7.339449541284404e-05, | |
| "loss": 1.0568, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 7.568807339449542e-05, | |
| "loss": 1.0224, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 7.79816513761468e-05, | |
| "loss": 1.1003, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.027522935779816e-05, | |
| "loss": 1.0502, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.256880733944955e-05, | |
| "loss": 1.0328, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.486238532110093e-05, | |
| "loss": 1.0297, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.715596330275229e-05, | |
| "loss": 1.0504, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.944954128440367e-05, | |
| "loss": 1.0365, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.174311926605506e-05, | |
| "loss": 1.0071, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.403669724770642e-05, | |
| "loss": 1.031, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.63302752293578e-05, | |
| "loss": 0.9967, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.862385321100918e-05, | |
| "loss": 1.012, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00010091743119266055, | |
| "loss": 0.9894, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00010321100917431193, | |
| "loss": 1.0349, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00010550458715596329, | |
| "loss": 1.0207, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.0001077981651376147, | |
| "loss": 1.0286, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00011009174311926606, | |
| "loss": 1.0229, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00011238532110091744, | |
| "loss": 0.9715, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00011467889908256881, | |
| "loss": 0.9984, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00011697247706422019, | |
| "loss": 1.037, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00011926605504587157, | |
| "loss": 0.988, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00012155963302752293, | |
| "loss": 0.9725, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00012385321100917432, | |
| "loss": 0.9794, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0001261467889908257, | |
| "loss": 0.9776, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00012844036697247707, | |
| "loss": 1.0248, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00013073394495412844, | |
| "loss": 1.0274, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00013302752293577983, | |
| "loss": 0.9989, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0001353211009174312, | |
| "loss": 0.9658, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00013761467889908258, | |
| "loss": 0.9641, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00013990825688073395, | |
| "loss": 0.9697, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0001422018348623853, | |
| "loss": 0.9573, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00014449541284403673, | |
| "loss": 0.9763, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.0001467889908256881, | |
| "loss": 0.987, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 0.00014908256880733945, | |
| "loss": 0.9707, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00015137614678899084, | |
| "loss": 1.0025, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0001536697247706422, | |
| "loss": 0.9905, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0001559633027522936, | |
| "loss": 0.9967, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00015825688073394496, | |
| "loss": 0.9689, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00016055045871559632, | |
| "loss": 1.0046, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0001628440366972477, | |
| "loss": 0.9616, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.0001651376146788991, | |
| "loss": 0.9492, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00016743119266055047, | |
| "loss": 1.0198, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00016972477064220186, | |
| "loss": 0.9947, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00017201834862385322, | |
| "loss": 0.9475, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00017431192660550458, | |
| "loss": 1.0068, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00017660550458715597, | |
| "loss": 0.9939, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00017889908256880734, | |
| "loss": 0.9984, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00018119266055045873, | |
| "loss": 0.9561, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00018348623853211012, | |
| "loss": 0.979, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00018577981651376148, | |
| "loss": 0.9584, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00018807339449541284, | |
| "loss": 0.9424, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019036697247706424, | |
| "loss": 0.9636, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0001926605504587156, | |
| "loss": 0.9377, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.000194954128440367, | |
| "loss": 0.9592, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019724770642201835, | |
| "loss": 0.9749, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019954128440366972, | |
| "loss": 0.974, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019999948643469536, | |
| "loss": 0.9534, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019999740008468594, | |
| "loss": 0.9777, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019999370888559804, | |
| "loss": 0.9829, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.0001999884128966714, | |
| "loss": 0.9788, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019998151220290082, | |
| "loss": 0.9425, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019997300691503497, | |
| "loss": 0.9491, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0001999628971695744, | |
| "loss": 0.9619, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019995118312876944, | |
| "loss": 0.9392, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0001999378649806177, | |
| "loss": 0.9862, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019992294293886095, | |
| "loss": 0.9365, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019990641724298156, | |
| "loss": 0.9803, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.000199888288158199, | |
| "loss": 0.9848, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.00019986855597546526, | |
| "loss": 1.0057, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019984722101146029, | |
| "loss": 1.0033, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.000199824283608587, | |
| "loss": 0.9879, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019979974413496566, | |
| "loss": 0.9598, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019977360298442803, | |
| "loss": 0.9508, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019974586057651102, | |
| "loss": 0.9945, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019971651735644995, | |
| "loss": 0.9571, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019968557379517152, | |
| "loss": 0.9827, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00019965303038928608, | |
| "loss": 0.9841, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019961888766107972, | |
| "loss": 0.9818, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019958314615850598, | |
| "loss": 0.9905, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019954580645517697, | |
| "loss": 0.9787, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0001995068691503541, | |
| "loss": 0.9553, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019946633486893865, | |
| "loss": 0.9473, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019942420426146153, | |
| "loss": 0.9429, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019938047800407302, | |
| "loss": 0.9637, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.00019933515679853182, | |
| "loss": 0.9538, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 0.0001992882413721937, | |
| "loss": 0.9558, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0001992397324780001, | |
| "loss": 0.9414, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019918963089446577, | |
| "loss": 0.9441, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019913793742566647, | |
| "loss": 0.9579, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019908465290122585, | |
| "loss": 0.9588, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019902977817630243, | |
| "loss": 0.9563, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019897331413157548, | |
| "loss": 0.9602, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019891526167323145, | |
| "loss": 1.0001, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0001988556217329488, | |
| "loss": 0.9464, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.00019879439526788341, | |
| "loss": 0.9516, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019873158326065327, | |
| "loss": 0.9838, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019866718671932249, | |
| "loss": 0.9449, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019860120667738516, | |
| "loss": 0.9682, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019853364419374902, | |
| "loss": 0.9793, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019846450035271808, | |
| "loss": 0.9388, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019839377626397554, | |
| "loss": 0.9399, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019832147306256576, | |
| "loss": 0.9485, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019824759190887622, | |
| "loss": 0.9451, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019817213398861866, | |
| "loss": 0.9355, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0001980951005128104, | |
| "loss": 0.947, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019801649271775459, | |
| "loss": 0.9964, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019793631186502047, | |
| "loss": 0.9402, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019785455924142318, | |
| "loss": 0.9519, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.0001977712361590031, | |
| "loss": 0.9518, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019768634395500465, | |
| "loss": 0.9448, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019759988399185505, | |
| "loss": 0.9568, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00019751185765714234, | |
| "loss": 0.9516, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019742226636359296, | |
| "loss": 0.9648, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019733111154904943, | |
| "loss": 0.9031, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019723839467644699, | |
| "loss": 0.9436, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019714411723379015, | |
| "loss": 0.9485, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0001970482807341289, | |
| "loss": 0.9723, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0001969508867155345, | |
| "loss": 0.9325, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019685193674107452, | |
| "loss": 0.9666, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019675143239878805, | |
| "loss": 0.9667, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00019664937530166002, | |
| "loss": 0.9436, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019654576708759538, | |
| "loss": 0.945, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019644060941939286, | |
| "loss": 0.9491, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019633390398471817, | |
| "loss": 0.9779, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019622565249607704, | |
| "loss": 0.9211, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.0001961158566907877, | |
| "loss": 0.9419, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019600451833095287, | |
| "loss": 1.0127, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019589163920343163, | |
| "loss": 0.9992, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019577722111981078, | |
| "loss": 0.9306, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 0.00019566126591637558, | |
| "loss": 0.9476, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.0001955437754540805, | |
| "loss": 0.9328, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019542475161851906, | |
| "loss": 0.9261, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019530419631989392, | |
| "loss": 0.927, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019518211149298595, | |
| "loss": 0.9426, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019505849909712332, | |
| "loss": 0.9454, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019493336111615003, | |
| "loss": 0.9581, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019480669955839402, | |
| "loss": 0.9698, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00019467851645663494, | |
| "loss": 0.9472, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019454881386807163, | |
| "loss": 0.931, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019441759387428903, | |
| "loss": 0.9286, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019428485858122472, | |
| "loss": 0.9338, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019415061011913523, | |
| "loss": 0.9526, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019401485064256176, | |
| "loss": 0.9506, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.0001938775823302957, | |
| "loss": 1.0001, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019373880738534358, | |
| "loss": 0.9568, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00019359852803489168, | |
| "loss": 0.9639, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.0001934567465302704, | |
| "loss": 0.9403, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019331346514691813, | |
| "loss": 0.9749, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019316868618434455, | |
| "loss": 0.9602, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019302241196609397, | |
| "loss": 0.9529, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.0001928746448397078, | |
| "loss": 0.9392, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019272538717668715, | |
| "loss": 0.9504, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019257464137245446, | |
| "loss": 0.931, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.0001924224098463153, | |
| "loss": 0.955, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019226869504141943, | |
| "loss": 0.9143, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 0.00019211349942472165, | |
| "loss": 0.9349, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019195682548694208, | |
| "loss": 0.9586, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019179867574252638, | |
| "loss": 0.9752, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019163905272960528, | |
| "loss": 0.9318, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.0001914779590099538, | |
| "loss": 0.9688, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019131539716895024, | |
| "loss": 0.963, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019115136981553464, | |
| "loss": 0.982, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019098587958216688, | |
| "loss": 0.9634, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019081892912478456, | |
| "loss": 0.9546, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00019065052112276018, | |
| "loss": 0.967, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00019048065827885827, | |
| "loss": 0.9423, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.000190309343319192, | |
| "loss": 0.952, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00019013657899317942, | |
| "loss": 0.9357, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.0001899623680734993, | |
| "loss": 0.9554, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00018978671335604665, | |
| "loss": 0.927, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00018960961765988792, | |
| "loss": 0.9476, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00018943108382721562, | |
| "loss": 0.928, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00018925111472330283, | |
| "loss": 0.9455, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018906971323645713, | |
| "loss": 0.9394, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018888688227797432, | |
| "loss": 0.9644, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018870262478209163, | |
| "loss": 0.9554, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018851694370594069, | |
| "loss": 0.945, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018832984202949996, | |
| "loss": 1.032, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018814132275554713, | |
| "loss": 0.9212, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.0001879513889096106, | |
| "loss": 0.948, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018776004353992124, | |
| "loss": 0.9458, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 0.00018756728971736327, | |
| "loss": 0.8995, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018737313053542512, | |
| "loss": 0.9336, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.0001871775691101496, | |
| "loss": 0.9549, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018698060858008403, | |
| "loss": 0.9443, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018678225210622986, | |
| "loss": 0.9268, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018658250287199196, | |
| "loss": 0.9454, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018638136408312728, | |
| "loss": 0.9691, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.0001861788389676939, | |
| "loss": 0.9272, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018597493077599867, | |
| "loss": 0.9535, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018576964278054544, | |
| "loss": 0.9311, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018556297827598242, | |
| "loss": 0.936, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018535494057904915, | |
| "loss": 0.9613, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018514553302852356, | |
| "loss": 0.9282, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018493475898516813, | |
| "loss": 0.9634, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018472262183167614, | |
| "loss": 0.9456, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018450912497261723, | |
| "loss": 0.9744, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018429427183438288, | |
| "loss": 0.9506, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.00018407806586513134, | |
| "loss": 0.9646, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00018386051053473232, | |
| "loss": 0.9508, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00018364160933471134, | |
| "loss": 0.9609, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0001834213657781936, | |
| "loss": 0.9239, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00018319978339984767, | |
| "loss": 0.943, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0001829768657558288, | |
| "loss": 0.9624, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00018275261642372175, | |
| "loss": 0.9136, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.0001825270390024834, | |
| "loss": 0.9307, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00018230013711238513, | |
| "loss": 0.9215, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 0.00018207191439495438, | |
| "loss": 0.9356, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00018184237451291665, | |
| "loss": 0.9342, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00018161152115013637, | |
| "loss": 0.9334, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00018137935801155794, | |
| "loss": 0.9071, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.0001811458888231462, | |
| "loss": 0.9096, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.0001809111173318267, | |
| "loss": 0.9713, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00018067504730542551, | |
| "loss": 0.9104, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.0001804376825326088, | |
| "loss": 0.9535, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.00018019902682282193, | |
| "loss": 0.9333, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.0001799590840062285, | |
| "loss": 0.9262, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017971785793364866, | |
| "loss": 0.9117, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.0001794753524764975, | |
| "loss": 0.9214, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017923157152672278, | |
| "loss": 0.9709, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017898651899674254, | |
| "loss": 0.9311, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017874019881938233, | |
| "loss": 0.9052, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.000178492614947812, | |
| "loss": 0.9493, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017824377135548236, | |
| "loss": 0.971, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017799367203606128, | |
| "loss": 0.9343, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.00017774232100336982, | |
| "loss": 0.9173, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00017748972229131757, | |
| "loss": 0.9286, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.000177235879953838, | |
| "loss": 0.9438, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00017698079806482343, | |
| "loss": 0.9414, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.0001767244807180597, | |
| "loss": 0.9522, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00017646693202716033, | |
| "loss": 0.9349, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.0001762081561255005, | |
| "loss": 0.9318, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00017594815716615093, | |
| "loss": 0.979, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.000175686939321811, | |
| "loss": 0.9571, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017542450678474184, | |
| "loss": 0.943, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017516086376669917, | |
| "loss": 0.9166, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017489601449886547, | |
| "loss": 0.9254, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017462996323178235, | |
| "loss": 0.9543, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017436271423528206, | |
| "loss": 0.9296, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0001740942717984192, | |
| "loss": 0.9449, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017382464022940182, | |
| "loss": 0.9476, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00017355382385552206, | |
| "loss": 0.9573, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0001732818270230871, | |
| "loss": 0.9341, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.000173008654097349, | |
| "loss": 0.9594, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.000172734309462435, | |
| "loss": 0.9488, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00017245879752127692, | |
| "loss": 0.9381, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.0001721821226955405, | |
| "loss": 0.937, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00017190428942555463, | |
| "loss": 0.9265, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.0001716253021702399, | |
| "loss": 0.9543, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.0001713451654070371, | |
| "loss": 0.9239, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.0001710638836318354, | |
| "loss": 0.9253, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 0.00017078146135890014, | |
| "loss": 0.9328, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.0001704979031208004, | |
| "loss": 0.9717, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.0001702132134683363, | |
| "loss": 0.9509, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00016992739697046586, | |
| "loss": 0.9542, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00016964045821423178, | |
| "loss": 0.9394, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00016935240180468775, | |
| "loss": 0.9624, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00016906323236482465, | |
| "loss": 0.9152, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00016877295453549614, | |
| "loss": 0.8982, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00016848157297534453, | |
| "loss": 0.9359, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.0001681890923607256, | |
| "loss": 0.931, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016789551738563384, | |
| "loss": 0.9682, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016760085276162708, | |
| "loss": 0.9452, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016730510321775075, | |
| "loss": 0.9409, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016700827350046206, | |
| "loss": 0.9172, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016671036837355386, | |
| "loss": 0.9517, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016641139261807818, | |
| "loss": 0.93, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016611135103226937, | |
| "loss": 0.9467, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00016581024843146725, | |
| "loss": 0.941, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016550808964803978, | |
| "loss": 0.9522, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016520487953130552, | |
| "loss": 0.9147, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016490062294745571, | |
| "loss": 0.9483, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016459532477947634, | |
| "loss": 0.9665, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016428898992706955, | |
| "loss": 0.9018, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016398162330657533, | |
| "loss": 0.9266, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.0001636732298508922, | |
| "loss": 0.937, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.0001633638145093984, | |
| "loss": 0.9404, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 0.00016305338224787235, | |
| "loss": 0.9352, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.0001627419380484128, | |
| "loss": 0.981, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00016242948690935912, | |
| "loss": 0.9355, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00016211603384521083, | |
| "loss": 0.9549, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00016180158388654742, | |
| "loss": 0.9515, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00016148614207994735, | |
| "loss": 0.9278, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00016116971348790712, | |
| "loss": 0.9441, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.0001608523031887601, | |
| "loss": 0.9572, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00016053391627659505, | |
| "loss": 0.9282, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.0001602145578611742, | |
| "loss": 0.9604, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00015989423306785142, | |
| "loss": 0.9238, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00015957294703748982, | |
| "loss": 0.9498, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00015925070492637944, | |
| "loss": 0.9186, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.0001589275119061542, | |
| "loss": 0.938, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00015860337316370916, | |
| "loss": 0.927, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.0001582782939011173, | |
| "loss": 0.9111, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00015795227933554568, | |
| "loss": 0.9417, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.00015762533469917216, | |
| "loss": 0.9582, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015729746523910113, | |
| "loss": 0.9591, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015696867621727942, | |
| "loss": 0.9703, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015663897291041175, | |
| "loss": 0.9027, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015630836060987624, | |
| "loss": 0.9339, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015597684462163923, | |
| "loss": 0.939, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.0001556444302661704, | |
| "loss": 0.9033, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015531112287835717, | |
| "loss": 0.9203, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00015497692780741908, | |
| "loss": 0.9783, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.0001546418504168222, | |
| "loss": 0.9225, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015430589608419264, | |
| "loss": 0.9399, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015396907020123068, | |
| "loss": 0.98, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015363137817362392, | |
| "loss": 0.9441, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015329282542096064, | |
| "loss": 0.9427, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015295341737664285, | |
| "loss": 0.9883, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.000152613159487799, | |
| "loss": 0.951, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015227205721519675, | |
| "loss": 0.9332, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015193011603315503, | |
| "loss": 0.9358, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00015158734142945644, | |
| "loss": 0.9267, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001512437389052591, | |
| "loss": 0.9338, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001508993139750083, | |
| "loss": 0.9487, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001505540721663481, | |
| "loss": 0.9521, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001502080190200325, | |
| "loss": 0.9292, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00014986116008983664, | |
| "loss": 0.8889, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00014951350094246762, | |
| "loss": 0.9395, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0001491650471574751, | |
| "loss": 0.9381, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00014881580432716182, | |
| "loss": 0.9487, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014846577805649388, | |
| "loss": 0.9345, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014811497396301072, | |
| "loss": 0.9259, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014776339767673491, | |
| "loss": 0.9292, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.000147411054840082, | |
| "loss": 0.923, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014705795110776974, | |
| "loss": 0.9341, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.0001467040921467275, | |
| "loss": 0.9282, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014634948363600518, | |
| "loss": 0.9365, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014599413126668213, | |
| "loss": 0.9484, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 0.00014563804074177588, | |
| "loss": 0.9545, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014528121777615058, | |
| "loss": 0.9604, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.0001449236680964251, | |
| "loss": 0.9648, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014456539744088138, | |
| "loss": 0.9257, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014420641155937224, | |
| "loss": 0.9188, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014384671621322915, | |
| "loss": 0.9493, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014348631717516953, | |
| "loss": 0.9224, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014312522022920444, | |
| "loss": 0.9354, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014276343117054563, | |
| "loss": 0.944, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014240095580551234, | |
| "loss": 0.9511, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.0001420377999514384, | |
| "loss": 0.9234, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00014167396943657874, | |
| "loss": 0.925, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.0001413094701000158, | |
| "loss": 0.9141, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00014094430779156597, | |
| "loss": 0.963, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00014057848837168547, | |
| "loss": 0.9122, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00014021201771137663, | |
| "loss": 0.9329, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00013984490169209333, | |
| "loss": 0.9249, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00013947714620564678, | |
| "loss": 0.935, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.00013910875715411098, | |
| "loss": 0.9623, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.000138739740449728, | |
| "loss": 0.947, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001383701020148129, | |
| "loss": 0.9583, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.000137999847781659, | |
| "loss": 0.9329, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00013762898369244238, | |
| "loss": 0.9346, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00013725751569912682, | |
| "loss": 0.9271, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00013688544976336783, | |
| "loss": 0.9435, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00013651279185641752, | |
| "loss": 0.9548, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001361395479590283, | |
| "loss": 0.9111, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00013576572406135722, | |
| "loss": 0.9306, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00013539132616286956, | |
| "loss": 0.9394, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00013501636027224282, | |
| "loss": 0.9214, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.0001346408324072701, | |
| "loss": 0.9064, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.0001342647485947635, | |
| "loss": 0.9462, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00013388811487045766, | |
| "loss": 0.9566, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.0001335109372789125, | |
| "loss": 0.9488, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00013313322187341652, | |
| "loss": 0.9491, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 0.00013275497471588953, | |
| "loss": 0.899, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0001323762018767854, | |
| "loss": 0.918, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00013199690943499457, | |
| "loss": 0.9125, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0001316171034777466, | |
| "loss": 0.9121, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00013123679010051232, | |
| "loss": 0.9151, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00013085597540690618, | |
| "loss": 0.9364, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00013047466550858812, | |
| "loss": 0.9368, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00013009286652516575, | |
| "loss": 0.959, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00012971058458409576, | |
| "loss": 0.9644, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.00012932782582058584, | |
| "loss": 0.9182, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00012894459637749627, | |
| "loss": 0.9549, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001285609024052411, | |
| "loss": 0.9235, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00012817675006168963, | |
| "loss": 0.9134, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00012779214551206746, | |
| "loss": 0.9379, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001274070949288577, | |
| "loss": 0.9506, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00012702160449170165, | |
| "loss": 0.93, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001266356803873, | |
| "loss": 0.931, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.0001262493288093131, | |
| "loss": 0.9204, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.000125862555958262, | |
| "loss": 0.9263, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00012547536804142858, | |
| "loss": 0.916, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001250877712727561, | |
| "loss": 0.9193, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00012469977187274945, | |
| "loss": 0.9434, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00012431137606837538, | |
| "loss": 0.9104, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.0001239225900929624, | |
| "loss": 0.9395, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00012353342018610084, | |
| "loss": 0.9354, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00012314387259354282, | |
| "loss": 0.9533, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 0.00012275395356710177, | |
| "loss": 0.9188, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001223636693645523, | |
| "loss": 0.9417, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00012197302624952971, | |
| "loss": 0.9422, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00012158203049142947, | |
| "loss": 0.9394, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00012119068836530644, | |
| "loss": 0.914, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00012079900615177449, | |
| "loss": 0.9266, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00012040699013690543, | |
| "loss": 0.9211, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00012001464661212827, | |
| "loss": 0.9532, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001196219818741281, | |
| "loss": 0.9343, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00011922900222474523, | |
| "loss": 0.9257, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011883571397087387, | |
| "loss": 0.9729, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011844212342436112, | |
| "loss": 0.9435, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011804823690190538, | |
| "loss": 0.9536, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011765406072495528, | |
| "loss": 0.9418, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011725960121960806, | |
| "loss": 0.9292, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011686486471650798, | |
| "loss": 0.9274, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.0001164698575507449, | |
| "loss": 0.9358, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011607458606175253, | |
| "loss": 0.9194, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 0.00011567905659320663, | |
| "loss": 0.9107, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011528327549292326, | |
| "loss": 0.9226, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011488724911275694, | |
| "loss": 0.9086, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011449098380849858, | |
| "loss": 0.9317, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011409448593977363, | |
| "loss": 0.933, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011369776186993994, | |
| "loss": 0.9545, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.0001133008179659856, | |
| "loss": 0.9344, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011290366059842683, | |
| "loss": 0.9683, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00011250629614120571, | |
| "loss": 0.9388, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00011210873097158786, | |
| "loss": 0.9207, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00011171097147006013, | |
| "loss": 0.8945, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00011131302402022821, | |
| "loss": 0.9312, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00011091489500871408, | |
| "loss": 0.8946, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00011051659082505366, | |
| "loss": 0.9266, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00011011811786159416, | |
| "loss": 0.9173, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00010971948251339157, | |
| "loss": 0.9378, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00010932069117810787, | |
| "loss": 0.9207, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 0.00010892175025590856, | |
| "loss": 0.9289, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010852266614935982, | |
| "loss": 0.9638, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010812344526332578, | |
| "loss": 0.9314, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010772409400486571, | |
| "loss": 0.9458, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010732461878313125, | |
| "loss": 0.8959, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010692502600926348, | |
| "loss": 0.8879, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010652532209629011, | |
| "loss": 0.9377, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010612551345902244, | |
| "loss": 0.9059, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00010572560651395258, | |
| "loss": 0.9251, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.0001053256076791503, | |
| "loss": 0.9417, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010492552337416007, | |
| "loss": 0.887, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010452536001989815, | |
| "loss": 0.9594, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010412512403854942, | |
| "loss": 0.9297, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010372482185346435, | |
| "loss": 0.9264, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010332445988905586, | |
| "loss": 0.898, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010292404457069631, | |
| "loss": 0.9527, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010252358232461436, | |
| "loss": 0.9603, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.00010212307957779173, | |
| "loss": 0.9433, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00010172254275786017, | |
| "loss": 0.9398, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001013219782929983, | |
| "loss": 0.9016, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.0001009213926118284, | |
| "loss": 0.9427, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00010052079214331318, | |
| "loss": 0.9138, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 0.00010012018331665272, | |
| "loss": 0.9185, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.971957256118129e-05, | |
| "loss": 0.9487, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.931896630626402e-05, | |
| "loss": 0.9152, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.891837098119389e-05, | |
| "loss": 0.9337, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.851779301508842e-05, | |
| "loss": 0.9453, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.811723883678654e-05, | |
| "loss": 0.9351, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.771671487474546e-05, | |
| "loss": 0.9204, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.731622755693737e-05, | |
| "loss": 0.9524, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.691578331074643e-05, | |
| "loss": 0.8976, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.651538856286551e-05, | |
| "loss": 0.9142, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.611504973919311e-05, | |
| "loss": 0.9019, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.571477326473021e-05, | |
| "loss": 0.9658, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.531456556347712e-05, | |
| "loss": 0.9435, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.491443305833043e-05, | |
| "loss": 0.9344, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.451438217097994e-05, | |
| "loss": 0.9224, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.411441932180554e-05, | |
| "loss": 0.9179, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.371455092977423e-05, | |
| "loss": 0.9382, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.331478341233706e-05, | |
| "loss": 0.9192, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.291512318532614e-05, | |
| "loss": 0.9384, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.251557666285174e-05, | |
| "loss": 0.939, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.211615025719919e-05, | |
| "loss": 0.9004, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.17168503787262e-05, | |
| "loss": 0.9569, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.131768343575979e-05, | |
| "loss": 0.9301, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.091865583449351e-05, | |
| "loss": 0.8872, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.051977397888464e-05, | |
| "loss": 0.9157, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.01210442705514e-05, | |
| "loss": 0.9084, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.972247310867027e-05, | |
| "loss": 0.9319, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.932406688987309e-05, | |
| "loss": 0.9006, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.892583200814466e-05, | |
| "loss": 0.9166, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.852777485471997e-05, | |
| "loss": 0.9481, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.81299018179817e-05, | |
| "loss": 0.9336, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.773221928335759e-05, | |
| "loss": 0.9258, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.73347336332181e-05, | |
| "loss": 0.9228, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.693745124677386e-05, | |
| "loss": 0.9629, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.654037849997342e-05, | |
| "loss": 0.9143, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.614352176540067e-05, | |
| "loss": 0.9296, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.57468874121729e-05, | |
| "loss": 0.9131, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.535048180583838e-05, | |
| "loss": 0.9382, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.495431130827422e-05, | |
| "loss": 0.9269, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.455838227758432e-05, | |
| "loss": 0.9828, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.416270106799726e-05, | |
| "loss": 0.9416, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.376727402976447e-05, | |
| "loss": 0.9548, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.3372107509058e-05, | |
| "loss": 0.9451, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.297720784786906e-05, | |
| "loss": 0.94, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.2582581383906e-05, | |
| "loss": 0.9348, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.218823445049265e-05, | |
| "loss": 0.9086, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.179417337646669e-05, | |
| "loss": 0.9297, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.140040448607804e-05, | |
| "loss": 0.9043, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.100693409888748e-05, | |
| "loss": 0.9659, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 8.061376852966495e-05, | |
| "loss": 0.9371, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 8.02209140882886e-05, | |
| "loss": 0.9234, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.982837707964321e-05, | |
| "loss": 0.9456, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.943616380351913e-05, | |
| "loss": 0.9487, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.904428055451118e-05, | |
| "loss": 0.9433, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.865273362191759e-05, | |
| "loss": 0.9324, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.826152928963904e-05, | |
| "loss": 0.9464, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.787067383607796e-05, | |
| "loss": 0.9324, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.748017353403748e-05, | |
| "loss": 0.9352, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.70900346506211e-05, | |
| "loss": 0.9325, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.670026344713189e-05, | |
| "loss": 0.9046, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.631086617897203e-05, | |
| "loss": 0.93, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.592184909554245e-05, | |
| "loss": 0.9312, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.553321844014258e-05, | |
| "loss": 0.9261, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.514498044987009e-05, | |
| "loss": 0.9278, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.475714135552074e-05, | |
| "loss": 0.9262, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.43697073814885e-05, | |
| "loss": 0.9214, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.39826847456656e-05, | |
| "loss": 0.9417, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.359607965934274e-05, | |
| "loss": 0.9129, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.32098983271094e-05, | |
| "loss": 0.9107, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.282414694675426e-05, | |
| "loss": 0.9148, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.243883170916574e-05, | |
| "loss": 0.9254, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.205395879823271e-05, | |
| "loss": 0.9516, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.166953439074504e-05, | |
| "loss": 0.9656, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.128556465629475e-05, | |
| "loss": 0.9507, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.090205575717678e-05, | |
| "loss": 0.9419, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 7.051901384829021e-05, | |
| "loss": 0.9415, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 7.013644507703937e-05, | |
| "loss": 0.9289, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.975435558323532e-05, | |
| "loss": 0.9579, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.937275149899725e-05, | |
| "loss": 0.9232, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.899163894865395e-05, | |
| "loss": 0.9468, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.86110240486457e-05, | |
| "loss": 0.9208, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.823091290742602e-05, | |
| "loss": 0.9138, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.785131162536374e-05, | |
| "loss": 0.9134, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.747222629464484e-05, | |
| "loss": 0.9252, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.709366299917497e-05, | |
| "loss": 0.9494, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.671562781448166e-05, | |
| "loss": 0.9539, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.633812680761684e-05, | |
| "loss": 0.9377, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.59611660370594e-05, | |
| "loss": 0.9295, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.558475155261811e-05, | |
| "loss": 0.943, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.52088893953344e-05, | |
| "loss": 0.9288, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.48335855973855e-05, | |
| "loss": 0.931, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 6.445884618198754e-05, | |
| "loss": 0.9198, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.408467716329894e-05, | |
| "loss": 0.9219, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.371108454632391e-05, | |
| "loss": 0.9367, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.33380743268159e-05, | |
| "loss": 0.9275, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.29656524911817e-05, | |
| "loss": 0.9398, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.259382501638509e-05, | |
| "loss": 0.9081, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.222259786985101e-05, | |
| "loss": 0.9107, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.185197700936982e-05, | |
| "loss": 0.9325, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.14819683830016e-05, | |
| "loss": 0.9389, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.111257792898082e-05, | |
| "loss": 0.9507, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 6.0743811575620846e-05, | |
| "loss": 0.9221, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 6.0375675241219e-05, | |
| "loss": 0.9263, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 6.000817483396148e-05, | |
| "loss": 0.9501, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.96413162518285e-05, | |
| "loss": 0.9477, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.9275105382499694e-05, | |
| "loss": 0.9091, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.890954810325966e-05, | |
| "loss": 0.9166, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.854465028090355e-05, | |
| "loss": 0.942, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.8180417771643006e-05, | |
| "loss": 0.9218, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.781685642101196e-05, | |
| "loss": 0.9067, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.7453972063773184e-05, | |
| "loss": 0.9334, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.7091770523824317e-05, | |
| "loss": 0.9322, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.673025761410462e-05, | |
| "loss": 0.9175, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.636943913650147e-05, | |
| "loss": 0.9245, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.60093208817575e-05, | |
| "loss": 0.9534, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.564990862937744e-05, | |
| "loss": 0.9286, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.5291208147535466e-05, | |
| "loss": 0.903, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.4933225192982586e-05, | |
| "loss": 0.9562, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.457596551095441e-05, | |
| "loss": 0.9073, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.421943483507863e-05, | |
| "loss": 0.9473, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.3863638887283364e-05, | |
| "loss": 0.9362, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.350858337770498e-05, | |
| "loss": 0.95, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.315427400459678e-05, | |
| "loss": 0.9603, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.280071645423726e-05, | |
| "loss": 0.9356, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.244791640083906e-05, | |
| "loss": 0.9236, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.2095879506457736e-05, | |
| "loss": 0.9292, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 5.174461142090111e-05, | |
| "loss": 0.9235, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 5.139411778163832e-05, | |
| "loss": 0.9339, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 5.104440421370962e-05, | |
| "loss": 0.9275, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 5.0695476329635825e-05, | |
| "loss": 0.9257, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 5.034733972932855e-05, | |
| "loss": 0.9501, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 5.000000000000002e-05, | |
| "loss": 0.9274, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.9653462716073594e-05, | |
| "loss": 0.9577, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.930773343909434e-05, | |
| "loss": 0.9107, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.8962817717639555e-05, | |
| "loss": 0.9309, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.8618721087230014e-05, | |
| "loss": 0.9078, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.8275449070240854e-05, | |
| "loss": 0.9725, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.793300717581308e-05, | |
| "loss": 0.9075, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.7591400899765234e-05, | |
| "loss": 0.9491, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.7250635724505e-05, | |
| "loss": 0.9454, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.6910717118941286e-05, | |
| "loss": 0.9204, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.6571650538396615e-05, | |
| "loss": 0.9513, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.6233441424519295e-05, | |
| "loss": 0.9194, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.5896095205196356e-05, | |
| "loss": 0.9324, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.5559617294466176e-05, | |
| "loss": 0.9291, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.522401309243193e-05, | |
| "loss": 0.9232, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.488928798517442e-05, | |
| "loss": 0.9222, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.455544734466624e-05, | |
| "loss": 0.8924, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.422249652868506e-05, | |
| "loss": 0.9431, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.389044088072798e-05, | |
| "loss": 0.9558, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.355928572992547e-05, | |
| "loss": 0.9277, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.322903639095619e-05, | |
| "loss": 0.9506, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.289969816396132e-05, | |
| "loss": 0.916, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.2571276334459895e-05, | |
| "loss": 0.9643, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.224377617326353e-05, | |
| "loss": 0.9275, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.1917202936392265e-05, | |
| "loss": 0.9156, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.15915618649899e-05, | |
| "loss": 0.9141, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.126685818524013e-05, | |
| "loss": 0.9155, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.094309710828236e-05, | |
| "loss": 0.9535, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.0620283830128414e-05, | |
| "loss": 0.9306, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.029842353157888e-05, | |
| "loss": 0.9403, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.9977521378140084e-05, | |
| "loss": 0.9113, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.965758251994115e-05, | |
| "loss": 0.9356, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.933861209165146e-05, | |
| "loss": 0.9075, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.9020615212398016e-05, | |
| "loss": 0.9311, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.8703596985683556e-05, | |
| "loss": 0.9257, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.838756249930439e-05, | |
| "loss": 0.9208, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.807251682526902e-05, | |
| "loss": 0.9287, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.775846501971636e-05, | |
| "loss": 0.9202, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.7445412122835077e-05, | |
| "loss": 0.9251, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.713336315878224e-05, | |
| "loss": 0.9225, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.6822323135603054e-05, | |
| "loss": 0.9223, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.651229704515018e-05, | |
| "loss": 0.9408, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.6203289863003905e-05, | |
| "loss": 0.9287, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.5895306548392005e-05, | |
| "loss": 0.9412, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.558835204411044e-05, | |
| "loss": 0.9387, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.52824312764438e-05, | |
| "loss": 0.9355, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.497754915508632e-05, | |
| "loss": 0.9618, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.467371057306318e-05, | |
| "loss": 0.9666, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.437092040665183e-05, | |
| "loss": 0.9278, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.406918351530376e-05, | |
| "loss": 0.9422, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.3768504741566664e-05, | |
| "loss": 0.9275, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.346888891100649e-05, | |
| "loss": 0.9051, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.3170340832130134e-05, | |
| "loss": 0.8954, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.287286529630832e-05, | |
| "loss": 0.9642, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.2576467077698493e-05, | |
| "loss": 0.9703, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.228115093316848e-05, | |
| "loss": 0.9327, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.198692160221987e-05, | |
| "loss": 0.9617, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.169378380691218e-05, | |
| "loss": 0.9156, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.140174225178692e-05, | |
| "loss": 0.946, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.111080162379215e-05, | |
| "loss": 0.9354, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.082096659220722e-05, | |
| "loss": 0.8934, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.0532241808567966e-05, | |
| "loss": 0.8923, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.0244631906591825e-05, | |
| "loss": 0.9405, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.9958141502103722e-05, | |
| "loss": 0.9512, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.9672775192961756e-05, | |
| "loss": 0.9447, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.938853755898364e-05, | |
| "loss": 0.905, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.910543316187301e-05, | |
| "loss": 0.9764, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.882346654514627e-05, | |
| "loss": 0.9429, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.8542642234059725e-05, | |
| "loss": 0.925, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.826296473553697e-05, | |
| "loss": 0.9438, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.7984438538096392e-05, | |
| "loss": 0.9118, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.7707068111779377e-05, | |
| "loss": 0.9838, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.7430857908078345e-05, | |
| "loss": 0.9473, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.7155812359865517e-05, | |
| "loss": 0.911, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.6881935881321563e-05, | |
| "loss": 0.9431, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.6609232867864896e-05, | |
| "loss": 0.9231, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.6337707696081094e-05, | |
| "loss": 0.9174, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.606736472365272e-05, | |
| "loss": 0.9345, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.5798208289289204e-05, | |
| "loss": 0.8839, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.5530242712657492e-05, | |
| "loss": 0.9211, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.526347229431242e-05, | |
| "loss": 0.9267, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.499790131562797e-05, | |
| "loss": 0.9288, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.4733534038728257e-05, | |
| "loss": 0.9291, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.4470374706419485e-05, | |
| "loss": 0.9016, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.4208427542121504e-05, | |
| "loss": 0.9249, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.394769674980035e-05, | |
| "loss": 0.9389, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.3688186513900455e-05, | |
| "loss": 0.9513, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.34299009992778e-05, | |
| "loss": 0.9435, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.317284435113278e-05, | |
| "loss": 0.9303, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.2917020694944023e-05, | |
| "loss": 0.9088, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.2662434136401722e-05, | |
| "loss": 0.9796, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.2409088761342235e-05, | |
| "loss": 0.9385, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.215698863568213e-05, | |
| "loss": 0.9183, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.1906137805353212e-05, | |
| "loss": 0.9324, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.1656540296237316e-05, | |
| "loss": 0.9309, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.1408200114101985e-05, | |
| "loss": 0.9402, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.116112124453592e-05, | |
| "loss": 0.9777, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.0915307652885164e-05, | |
| "loss": 0.8989, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.067076328418949e-05, | |
| "loss": 0.9326, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.0427492063118935e-05, | |
| "loss": 0.9245, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 2.018549789391102e-05, | |
| "loss": 0.9139, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.994478466030787e-05, | |
| "loss": 0.9229, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.970535622549401e-05, | |
| "loss": 0.9336, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.946721643203443e-05, | |
| "loss": 0.9187, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.923036910181275e-05, | |
| "loss": 0.9073, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.8994818035969975e-05, | |
| "loss": 0.962, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.8760567014843545e-05, | |
| "loss": 0.9457, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.8527619797906494e-05, | |
| "loss": 0.9619, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.8295980123707357e-05, | |
| "loss": 0.9642, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.8065651709809905e-05, | |
| "loss": 0.9517, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.783663825273372e-05, | |
| "loss": 0.9185, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.7608943427894686e-05, | |
| "loss": 0.9393, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.7382570889546124e-05, | |
| "loss": 0.925, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.7157524270720036e-05, | |
| "loss": 0.9554, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6933807183168994e-05, | |
| "loss": 0.9329, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6711423217307885e-05, | |
| "loss": 0.9038, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.64903759421566e-05, | |
| "loss": 0.9563, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.627066890528247e-05, | |
| "loss": 0.9192, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6052305632743592e-05, | |
| "loss": 0.9203, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.583528962903197e-05, | |
| "loss": 0.9128, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.5619624377017537e-05, | |
| "loss": 0.9168, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.540531333789207e-05, | |
| "loss": 0.9275, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.5192359951113755e-05, | |
| "loss": 0.9139, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.4980767634351877e-05, | |
| "loss": 0.9436, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.4770539783432113e-05, | |
| "loss": 0.9219, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.4561679772281877e-05, | |
| "loss": 0.9567, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.4354190952876334e-05, | |
| "loss": 0.9052, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.4148076655184373e-05, | |
| "loss": 0.8918, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.3943340187115494e-05, | |
| "loss": 0.9625, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.373998483446638e-05, | |
| "loss": 0.9461, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.3538013860868436e-05, | |
| "loss": 0.9113, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.3337430507735205e-05, | |
| "loss": 0.9155, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.313823799421051e-05, | |
| "loss": 0.9207, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.2940439517116676e-05, | |
| "loss": 0.9113, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.2744038250903267e-05, | |
| "loss": 0.9277, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.2549037347596115e-05, | |
| "loss": 0.9335, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.2355439936746827e-05, | |
| "loss": 0.9102, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.2163249125382426e-05, | |
| "loss": 0.9435, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.1972467997955595e-05, | |
| "loss": 0.9331, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.1783099616295056e-05, | |
| "loss": 0.9538, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.1595147019556607e-05, | |
| "loss": 0.9138, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.140861322417417e-05, | |
| "loss": 0.8933, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.1223501223811451e-05, | |
| "loss": 0.9213, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.1039813989313951e-05, | |
| "loss": 0.9364, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.085755446866119e-05, | |
| "loss": 0.9287, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.0676725586919457e-05, | |
| "loss": 0.9445, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.0497330246194848e-05, | |
| "loss": 0.9364, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.031937132558668e-05, | |
| "loss": 0.9478, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.014285168114133e-05, | |
| "loss": 0.9348, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.96777414580633e-06, | |
| "loss": 0.9165, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.794141529384915e-06, | |
| "loss": 0.9434, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.621956618491024e-06, | |
| "loss": 0.9343, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.451222176504414e-06, | |
| "loss": 0.9385, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.281940943526491e-06, | |
| "loss": 0.9237, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.114115636336152e-06, | |
| "loss": 0.9513, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.947748948346357e-06, | |
| "loss": 0.9342, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.782843549560771e-06, | |
| "loss": 0.9381, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.61940208653097e-06, | |
| "loss": 0.9332, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.457427182313937e-06, | |
| "loss": 0.9218, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.296921436430071e-06, | |
| "loss": 0.9162, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.137887424821277e-06, | |
| "loss": 0.8901, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 7.980327699809832e-06, | |
| "loss": 0.9368, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.824244790057223e-06, | |
| "loss": 0.9305, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.66964120052377e-06, | |
| "loss": 0.9522, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.516519412428203e-06, | |
| "loss": 0.9225, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.3648818832080745e-06, | |
| "loss": 0.8974, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.214731046480094e-06, | |
| "loss": 0.9047, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.066069312001289e-06, | |
| "loss": 0.9516, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.918899065630113e-06, | |
| "loss": 0.9177, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.773222669288359e-06, | |
| "loss": 0.9338, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 6.629042460923096e-06, | |
| "loss": 0.9083, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 6.486360754469234e-06, | |
| "loss": 0.9496, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 6.345179839812343e-06, | |
| "loss": 0.9246, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 6.205501982751971e-06, | |
| "loss": 0.9335, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 6.067329424965162e-06, | |
| "loss": 0.96, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.930664383970641e-06, | |
| "loss": 0.8985, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.795509053093029e-06, | |
| "loss": 0.9208, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.6618656014278406e-06, | |
| "loss": 0.951, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.5297361738065325e-06, | |
| "loss": 0.9716, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 5.399122890762143e-06, | |
| "loss": 0.9138, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 5.270027848495207e-06, | |
| "loss": 0.9603, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 5.1424531188402405e-06, | |
| "loss": 0.9495, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 5.016400749232297e-06, | |
| "loss": 0.8994, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.89187276267431e-06, | |
| "loss": 0.9471, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.7688711577044354e-06, | |
| "loss": 0.9076, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.647397908364182e-06, | |
| "loss": 0.923, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.5274549641665105e-06, | |
| "loss": 0.9449, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.40904425006472e-06, | |
| "loss": 0.9046, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.2921676664214535e-06, | |
| "loss": 0.9415, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.176827088978297e-06, | |
| "loss": 0.9518, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.0630243688255185e-06, | |
| "loss": 0.916, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.950761332372543e-06, | |
| "loss": 0.9518, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.8400397813185054e-06, | |
| "loss": 0.9258, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.7308614926234165e-06, | |
| "loss": 0.938, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.6232282184795794e-06, | |
| "loss": 0.9097, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.517141686283498e-06, | |
| "loss": 0.9193, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.412603598608188e-06, | |
| "loss": 0.8956, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.3096156331758e-06, | |
| "loss": 0.9154, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.2081794428307278e-06, | |
| "loss": 0.9424, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.1082966555130654e-06, | |
| "loss": 0.9038, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.0099688742324715e-06, | |
| "loss": 0.944, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.913197677042456e-06, | |
| "loss": 0.9392, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.8179846170150903e-06, | |
| "loss": 0.9454, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.7243312222159924e-06, | |
| "loss": 0.9261, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.6322389956799143e-06, | |
| "loss": 0.9495, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.541709415386495e-06, | |
| "loss": 0.9539, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.4527439342366785e-06, | |
| "loss": 0.9329, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.3653439800292556e-06, | |
| "loss": 0.9284, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.2795109554381024e-06, | |
| "loss": 0.9442, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.195246237989479e-06, | |
| "loss": 0.8929, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.1125511800401234e-06, | |
| "loss": 0.9192, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.0314271087554126e-06, | |
| "loss": 0.9404, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.951875326088104e-06, | |
| "loss": 0.8926, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.8738971087574275e-06, | |
| "loss": 0.9077, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.797493708228659e-06, | |
| "loss": 0.9288, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.7226663506929142e-06, | |
| "loss": 0.9176, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.6494162370475852e-06, | |
| "loss": 0.9047, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5777445428770022e-06, | |
| "loss": 0.9637, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5076524184336027e-06, | |
| "loss": 0.921, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.4391409886194474e-06, | |
| "loss": 0.9037, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.3722113529681668e-06, | |
| "loss": 0.9009, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.306864585627332e-06, | |
| "loss": 0.9559, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.2431017353412233e-06, | |
| "loss": 0.9091, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.1809238254339105e-06, | |
| "loss": 0.9251, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.1203318537929996e-06, | |
| "loss": 0.9733, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.0613267928534453e-06, | |
| "loss": 0.955, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.0039095895820639e-06, | |
| "loss": 0.9215, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.480811654622557e-07, | |
| "loss": 0.9578, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 8.938424164792736e-07, | |
| "loss": 0.9271, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 8.411942131058115e-07, | |
| "loss": 0.9399, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 7.90137400288049e-07, | |
| "loss": 0.9239, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 7.406727974320627e-07, | |
| "loss": 0.9071, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 6.928011983907245e-07, | |
| "loss": 0.9487, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 6.465233714509245e-07, | |
| "loss": 0.9517, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 6.01840059321257e-07, | |
| "loss": 0.8902, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 5.587519791200869e-07, | |
| "loss": 0.9174, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 5.172598223640468e-07, | |
| "loss": 0.9388, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.773642549569579e-07, | |
| "loss": 0.9163, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.390659171790934e-07, | |
| "loss": 0.9182, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.023654236769647e-07, | |
| "loss": 0.9126, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.672633634534295e-07, | |
| "loss": 0.902, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.3376029985819903e-07, | |
| "loss": 0.9142, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.0185677057887885e-07, | |
| "loss": 0.9352, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.715532876322646e-07, | |
| "loss": 0.9387, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.428503373561708e-07, | |
| "loss": 0.9471, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.1574838040161473e-07, | |
| "loss": 0.9155, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.9024785172541136e-07, | |
| "loss": 0.9221, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.6634916058319018e-07, | |
| "loss": 0.9257, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.4405269052284455e-07, | |
| "loss": 0.9422, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.2335879937839246e-07, | |
| "loss": 0.9771, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.0426781926416996e-07, | |
| "loss": 0.8896, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 8.678005656957977e-08, | |
| "loss": 0.9573, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 7.089579195409534e-08, | |
| "loss": 0.9849, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 5.6615280342842135e-08, | |
| "loss": 0.9487, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.3938750922412064e-08, | |
| "loss": 0.9108, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.286640713727751e-08, | |
| "loss": 0.9225, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.3398426686471743e-08, | |
| "loss": 0.9169, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5534961520724533e-08, | |
| "loss": 0.9376, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.276137840075194e-09, | |
| "loss": 0.9052, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.6220560918075474e-09, | |
| "loss": 0.9027, | |
| "step": 4345 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.572790968851212e-09, | |
| "loss": 0.9477, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.2839140858256215e-10, | |
| "loss": 0.9126, | |
| "step": 4355 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.9316108822822571, | |
| "eval_runtime": 1883.4384, | |
| "eval_samples_per_second": 8.193, | |
| "eval_steps_per_second": 0.256, | |
| "step": 4357 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 4357, | |
| "total_flos": 2.560373752332288e+16, | |
| "train_loss": 0.9531040664985417, | |
| "train_runtime": 61109.7572, | |
| "train_samples_per_second": 2.282, | |
| "train_steps_per_second": 0.071 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 4357, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "total_flos": 2.560373752332288e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |