{ "best_global_step": 5000, "best_metric": 1.1470744609832764, "best_model_checkpoint": "/workspace/actual_run/checkpoint-5000", "epoch": 0.6012506012506013, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0012025012025012026, "grad_norm": 1.0899138450622559, "learning_rate": 3.6e-06, "loss": 2.3927, "step": 10 }, { "epoch": 0.002405002405002405, "grad_norm": 0.9620404839515686, "learning_rate": 7.6e-06, "loss": 2.2959, "step": 20 }, { "epoch": 0.0036075036075036075, "grad_norm": 0.43678194284439087, "learning_rate": 1.16e-05, "loss": 2.036, "step": 30 }, { "epoch": 0.00481000481000481, "grad_norm": 0.35978421568870544, "learning_rate": 1.56e-05, "loss": 1.898, "step": 40 }, { "epoch": 0.006012506012506013, "grad_norm": 0.2896435856819153, "learning_rate": 1.9600000000000002e-05, "loss": 1.7747, "step": 50 }, { "epoch": 0.007215007215007215, "grad_norm": 0.2754799425601959, "learning_rate": 2.36e-05, "loss": 1.6628, "step": 60 }, { "epoch": 0.008417508417508417, "grad_norm": 0.2863922119140625, "learning_rate": 2.7600000000000003e-05, "loss": 1.6091, "step": 70 }, { "epoch": 0.00962000962000962, "grad_norm": 0.29406264424324036, "learning_rate": 3.16e-05, "loss": 1.5302, "step": 80 }, { "epoch": 0.010822510822510822, "grad_norm": 0.32185423374176025, "learning_rate": 3.56e-05, "loss": 1.498, "step": 90 }, { "epoch": 0.012025012025012025, "grad_norm": 0.34314361214637756, "learning_rate": 3.960000000000001e-05, "loss": 1.4561, "step": 100 }, { "epoch": 0.013227513227513227, "grad_norm": 0.3503110706806183, "learning_rate": 4.36e-05, "loss": 1.4569, "step": 110 }, { "epoch": 0.01443001443001443, "grad_norm": 0.3761644661426544, "learning_rate": 4.76e-05, "loss": 1.4043, "step": 120 }, { "epoch": 0.015632515632515633, "grad_norm": 0.3741130530834198, "learning_rate": 5.16e-05, "loss": 1.4147, "step": 130 }, { "epoch": 0.016835016835016835, "grad_norm": 0.381028950214386, "learning_rate": 5.560000000000001e-05, "loss": 1.4226, "step": 140 }, { "epoch": 0.018037518037518036, "grad_norm": 0.3512011468410492, "learning_rate": 5.96e-05, "loss": 1.4034, "step": 150 }, { "epoch": 0.01924001924001924, "grad_norm": 0.33474254608154297, "learning_rate": 6.36e-05, "loss": 1.3853, "step": 160 }, { "epoch": 0.020442520442520443, "grad_norm": 0.335902601480484, "learning_rate": 6.76e-05, "loss": 1.3865, "step": 170 }, { "epoch": 0.021645021645021644, "grad_norm": 0.31389617919921875, "learning_rate": 7.16e-05, "loss": 1.3739, "step": 180 }, { "epoch": 0.02284752284752285, "grad_norm": 0.3155975043773651, "learning_rate": 7.560000000000001e-05, "loss": 1.3713, "step": 190 }, { "epoch": 0.02405002405002405, "grad_norm": 0.31867676973342896, "learning_rate": 7.960000000000001e-05, "loss": 1.3689, "step": 200 }, { "epoch": 0.025252525252525252, "grad_norm": 0.2963453531265259, "learning_rate": 8.36e-05, "loss": 1.3701, "step": 210 }, { "epoch": 0.026455026455026454, "grad_norm": 0.3254355788230896, "learning_rate": 8.76e-05, "loss": 1.3618, "step": 220 }, { "epoch": 0.02765752765752766, "grad_norm": 0.31009721755981445, "learning_rate": 9.16e-05, "loss": 1.3562, "step": 230 }, { "epoch": 0.02886002886002886, "grad_norm": 0.2769719958305359, "learning_rate": 9.56e-05, "loss": 1.341, "step": 240 }, { "epoch": 0.03006253006253006, "grad_norm": 0.29058948159217834, "learning_rate": 9.960000000000001e-05, "loss": 1.3509, "step": 250 }, { "epoch": 0.031265031265031266, "grad_norm": 0.2969858646392822, "learning_rate": 9.999969280942074e-05, "loss": 1.3533, "step": 260 }, { "epoch": 0.032467532467532464, "grad_norm": 0.27514412999153137, "learning_rate": 9.999863092090635e-05, "loss": 1.3382, "step": 270 }, { "epoch": 0.03367003367003367, "grad_norm": 0.2757578492164612, "learning_rate": 9.99968105580855e-05, "loss": 1.3302, "step": 280 }, { "epoch": 0.034872534872534874, "grad_norm": 0.2729010283946991, "learning_rate": 9.999423174857289e-05, "loss": 1.3325, "step": 290 }, { "epoch": 0.03607503607503607, "grad_norm": 0.26693403720855713, "learning_rate": 9.999089453148882e-05, "loss": 1.3318, "step": 300 }, { "epoch": 0.03727753727753728, "grad_norm": 0.26203256845474243, "learning_rate": 9.998679895745853e-05, "loss": 1.3308, "step": 310 }, { "epoch": 0.03848003848003848, "grad_norm": 0.24240988492965698, "learning_rate": 9.998194508861147e-05, "loss": 1.3211, "step": 320 }, { "epoch": 0.03968253968253968, "grad_norm": 0.23748330771923065, "learning_rate": 9.997633299858036e-05, "loss": 1.317, "step": 330 }, { "epoch": 0.040885040885040885, "grad_norm": 0.27641433477401733, "learning_rate": 9.996996277250007e-05, "loss": 1.3049, "step": 340 }, { "epoch": 0.04208754208754209, "grad_norm": 0.261136531829834, "learning_rate": 9.996283450700628e-05, "loss": 1.3099, "step": 350 }, { "epoch": 0.04329004329004329, "grad_norm": 0.26694369316101074, "learning_rate": 9.995494831023409e-05, "loss": 1.3064, "step": 360 }, { "epoch": 0.04449254449254449, "grad_norm": 0.2695152759552002, "learning_rate": 9.994630430181631e-05, "loss": 1.3085, "step": 370 }, { "epoch": 0.0456950456950457, "grad_norm": 0.26653772592544556, "learning_rate": 9.99369026128817e-05, "loss": 1.3184, "step": 380 }, { "epoch": 0.046897546897546896, "grad_norm": 0.25345709919929504, "learning_rate": 9.992674338605297e-05, "loss": 1.2965, "step": 390 }, { "epoch": 0.0481000481000481, "grad_norm": 0.2470959722995758, "learning_rate": 9.991582677544458e-05, "loss": 1.29, "step": 400 }, { "epoch": 0.0493025493025493, "grad_norm": 0.2398555725812912, "learning_rate": 9.990415294666045e-05, "loss": 1.2991, "step": 410 }, { "epoch": 0.050505050505050504, "grad_norm": 0.24114194512367249, "learning_rate": 9.98917220767914e-05, "loss": 1.294, "step": 420 }, { "epoch": 0.05170755170755171, "grad_norm": 0.24480991065502167, "learning_rate": 9.98785343544125e-05, "loss": 1.2827, "step": 430 }, { "epoch": 0.05291005291005291, "grad_norm": 0.2351623922586441, "learning_rate": 9.986458997958021e-05, "loss": 1.2868, "step": 440 }, { "epoch": 0.05411255411255411, "grad_norm": 0.23396646976470947, "learning_rate": 9.984988916382932e-05, "loss": 1.2894, "step": 450 }, { "epoch": 0.05531505531505532, "grad_norm": 0.24305066466331482, "learning_rate": 9.983443213016975e-05, "loss": 1.2929, "step": 460 }, { "epoch": 0.056517556517556515, "grad_norm": 0.24993565678596497, "learning_rate": 9.981821911308316e-05, "loss": 1.2809, "step": 470 }, { "epoch": 0.05772005772005772, "grad_norm": 0.24196045100688934, "learning_rate": 9.980125035851945e-05, "loss": 1.2771, "step": 480 }, { "epoch": 0.058922558922558925, "grad_norm": 0.2293768674135208, "learning_rate": 9.978352612389298e-05, "loss": 1.2899, "step": 490 }, { "epoch": 0.06012506012506012, "grad_norm": 0.22991864383220673, "learning_rate": 9.97650466780786e-05, "loss": 1.2813, "step": 500 }, { "epoch": 0.06132756132756133, "grad_norm": 0.24029232561588287, "learning_rate": 9.97458123014077e-05, "loss": 1.2816, "step": 510 }, { "epoch": 0.06253006253006253, "grad_norm": 0.2242475003004074, "learning_rate": 9.972582328566387e-05, "loss": 1.2859, "step": 520 }, { "epoch": 0.06373256373256374, "grad_norm": 0.2273736447095871, "learning_rate": 9.97050799340785e-05, "loss": 1.2743, "step": 530 }, { "epoch": 0.06493506493506493, "grad_norm": 0.23436114192008972, "learning_rate": 9.968358256132623e-05, "loss": 1.2841, "step": 540 }, { "epoch": 0.06613756613756613, "grad_norm": 0.22182011604309082, "learning_rate": 9.966133149352006e-05, "loss": 1.2933, "step": 550 }, { "epoch": 0.06734006734006734, "grad_norm": 0.23829922080039978, "learning_rate": 9.963832706820649e-05, "loss": 1.2865, "step": 560 }, { "epoch": 0.06854256854256854, "grad_norm": 0.24821604788303375, "learning_rate": 9.96145696343604e-05, "loss": 1.2835, "step": 570 }, { "epoch": 0.06974506974506975, "grad_norm": 0.22477108240127563, "learning_rate": 9.959005955237972e-05, "loss": 1.2845, "step": 580 }, { "epoch": 0.07094757094757095, "grad_norm": 0.23542803525924683, "learning_rate": 9.956479719407998e-05, "loss": 1.287, "step": 590 }, { "epoch": 0.07215007215007214, "grad_norm": 0.22595302760601044, "learning_rate": 9.953878294268866e-05, "loss": 1.2762, "step": 600 }, { "epoch": 0.07335257335257335, "grad_norm": 0.21597643196582794, "learning_rate": 9.95120171928394e-05, "loss": 1.283, "step": 610 }, { "epoch": 0.07455507455507455, "grad_norm": 0.23060384392738342, "learning_rate": 9.948450035056599e-05, "loss": 1.2847, "step": 620 }, { "epoch": 0.07575757575757576, "grad_norm": 0.21899910271167755, "learning_rate": 9.945623283329621e-05, "loss": 1.2699, "step": 630 }, { "epoch": 0.07696007696007696, "grad_norm": 0.21934176981449127, "learning_rate": 9.942721506984551e-05, "loss": 1.274, "step": 640 }, { "epoch": 0.07816257816257816, "grad_norm": 0.22260645031929016, "learning_rate": 9.939744750041052e-05, "loss": 1.2653, "step": 650 }, { "epoch": 0.07936507936507936, "grad_norm": 0.22758053243160248, "learning_rate": 9.936693057656231e-05, "loss": 1.2712, "step": 660 }, { "epoch": 0.08056758056758057, "grad_norm": 0.21833918988704681, "learning_rate": 9.933566476123964e-05, "loss": 1.2631, "step": 670 }, { "epoch": 0.08177008177008177, "grad_norm": 0.2229488492012024, "learning_rate": 9.930365052874182e-05, "loss": 1.2632, "step": 680 }, { "epoch": 0.08297258297258298, "grad_norm": 0.2153736650943756, "learning_rate": 9.927088836472163e-05, "loss": 1.284, "step": 690 }, { "epoch": 0.08417508417508418, "grad_norm": 0.23731300234794617, "learning_rate": 9.923737876617784e-05, "loss": 1.2629, "step": 700 }, { "epoch": 0.08537758537758537, "grad_norm": 0.22687861323356628, "learning_rate": 9.920312224144779e-05, "loss": 1.2757, "step": 710 }, { "epoch": 0.08658008658008658, "grad_norm": 0.22836917638778687, "learning_rate": 9.916811931019956e-05, "loss": 1.2589, "step": 720 }, { "epoch": 0.08778258778258778, "grad_norm": 0.2113468050956726, "learning_rate": 9.91323705034242e-05, "loss": 1.2515, "step": 730 }, { "epoch": 0.08898508898508899, "grad_norm": 0.2195415198802948, "learning_rate": 9.909587636342756e-05, "loss": 1.2544, "step": 740 }, { "epoch": 0.09018759018759019, "grad_norm": 0.2246798574924469, "learning_rate": 9.905863744382217e-05, "loss": 1.2503, "step": 750 }, { "epoch": 0.0913900913900914, "grad_norm": 0.22869926691055298, "learning_rate": 9.902065430951874e-05, "loss": 1.2626, "step": 760 }, { "epoch": 0.09259259259259259, "grad_norm": 0.2191305309534073, "learning_rate": 9.898192753671774e-05, "loss": 1.2612, "step": 770 }, { "epoch": 0.09379509379509379, "grad_norm": 0.22004824876785278, "learning_rate": 9.894245771290042e-05, "loss": 1.2734, "step": 780 }, { "epoch": 0.094997594997595, "grad_norm": 0.22499217092990875, "learning_rate": 9.890224543682017e-05, "loss": 1.244, "step": 790 }, { "epoch": 0.0962000962000962, "grad_norm": 0.22316116094589233, "learning_rate": 9.886129131849324e-05, "loss": 1.2567, "step": 800 }, { "epoch": 0.09740259740259741, "grad_norm": 0.23548166453838348, "learning_rate": 9.88195959791896e-05, "loss": 1.2649, "step": 810 }, { "epoch": 0.0986050986050986, "grad_norm": 0.22244389355182648, "learning_rate": 9.877716005142339e-05, "loss": 1.2516, "step": 820 }, { "epoch": 0.0998075998075998, "grad_norm": 0.22191107273101807, "learning_rate": 9.873398417894348e-05, "loss": 1.2503, "step": 830 }, { "epoch": 0.10101010101010101, "grad_norm": 0.22165456414222717, "learning_rate": 9.869006901672362e-05, "loss": 1.2466, "step": 840 }, { "epoch": 0.10221260221260221, "grad_norm": 0.2198014259338379, "learning_rate": 9.864541523095247e-05, "loss": 1.246, "step": 850 }, { "epoch": 0.10341510341510342, "grad_norm": 0.22250819206237793, "learning_rate": 9.860002349902358e-05, "loss": 1.2362, "step": 860 }, { "epoch": 0.10461760461760462, "grad_norm": 0.2301010638475418, "learning_rate": 9.855389450952503e-05, "loss": 1.255, "step": 870 }, { "epoch": 0.10582010582010581, "grad_norm": 0.22151581943035126, "learning_rate": 9.850702896222907e-05, "loss": 1.2555, "step": 880 }, { "epoch": 0.10702260702260702, "grad_norm": 0.2172437161207199, "learning_rate": 9.845942756808145e-05, "loss": 1.2519, "step": 890 }, { "epoch": 0.10822510822510822, "grad_norm": 0.2380341738462448, "learning_rate": 9.841109104919057e-05, "loss": 1.2565, "step": 900 }, { "epoch": 0.10942760942760943, "grad_norm": 0.2219957560300827, "learning_rate": 9.836202013881673e-05, "loss": 1.2472, "step": 910 }, { "epoch": 0.11063011063011063, "grad_norm": 0.23652082681655884, "learning_rate": 9.831221558136075e-05, "loss": 1.249, "step": 920 }, { "epoch": 0.11183261183261184, "grad_norm": 0.21968142688274384, "learning_rate": 9.826167813235291e-05, "loss": 1.2451, "step": 930 }, { "epoch": 0.11303511303511303, "grad_norm": 0.22762618958950043, "learning_rate": 9.82104085584413e-05, "loss": 1.2433, "step": 940 }, { "epoch": 0.11423761423761424, "grad_norm": 0.20969653129577637, "learning_rate": 9.815840763738032e-05, "loss": 1.2452, "step": 950 }, { "epoch": 0.11544011544011544, "grad_norm": 0.22823800146579742, "learning_rate": 9.810567615801883e-05, "loss": 1.2482, "step": 960 }, { "epoch": 0.11664261664261664, "grad_norm": 0.22033190727233887, "learning_rate": 9.805221492028815e-05, "loss": 1.2466, "step": 970 }, { "epoch": 0.11784511784511785, "grad_norm": 0.22417309880256653, "learning_rate": 9.799802473519e-05, "loss": 1.2531, "step": 980 }, { "epoch": 0.11904761904761904, "grad_norm": 0.21632517874240875, "learning_rate": 9.794310642478418e-05, "loss": 1.2479, "step": 990 }, { "epoch": 0.12025012025012025, "grad_norm": 0.21755601465702057, "learning_rate": 9.788746082217601e-05, "loss": 1.2575, "step": 1000 }, { "epoch": 0.12025012025012025, "eval_loss": 1.2297765016555786, "eval_runtime": 17.3877, "eval_samples_per_second": 14.723, "eval_steps_per_second": 0.46, "step": 1000 }, { "epoch": 0.12145262145262145, "grad_norm": 0.23050642013549805, "learning_rate": 9.783108877150384e-05, "loss": 1.2438, "step": 1010 }, { "epoch": 0.12265512265512266, "grad_norm": 0.19902385771274567, "learning_rate": 9.77739911279261e-05, "loss": 1.2523, "step": 1020 }, { "epoch": 0.12385762385762386, "grad_norm": 0.27783408761024475, "learning_rate": 9.77161687576084e-05, "loss": 1.2591, "step": 1030 }, { "epoch": 0.12506012506012507, "grad_norm": 0.21347227692604065, "learning_rate": 9.765762253771045e-05, "loss": 1.2519, "step": 1040 }, { "epoch": 0.12626262626262627, "grad_norm": 0.209860697388649, "learning_rate": 9.75983533563726e-05, "loss": 1.229, "step": 1050 }, { "epoch": 0.12746512746512748, "grad_norm": 0.20719142258167267, "learning_rate": 9.753836211270254e-05, "loss": 1.2559, "step": 1060 }, { "epoch": 0.12866762866762868, "grad_norm": 0.22753511369228363, "learning_rate": 9.747764971676146e-05, "loss": 1.2511, "step": 1070 }, { "epoch": 0.12987012987012986, "grad_norm": 0.22132696211338043, "learning_rate": 9.741621708955048e-05, "loss": 1.2591, "step": 1080 }, { "epoch": 0.13107263107263106, "grad_norm": 0.22907616198062897, "learning_rate": 9.735406516299649e-05, "loss": 1.2483, "step": 1090 }, { "epoch": 0.13227513227513227, "grad_norm": 0.20939530432224274, "learning_rate": 9.729119487993809e-05, "loss": 1.2426, "step": 1100 }, { "epoch": 0.13347763347763347, "grad_norm": 0.22089892625808716, "learning_rate": 9.722760719411128e-05, "loss": 1.2527, "step": 1110 }, { "epoch": 0.13468013468013468, "grad_norm": 0.2165553718805313, "learning_rate": 9.716330307013499e-05, "loss": 1.2486, "step": 1120 }, { "epoch": 0.13588263588263588, "grad_norm": 0.22434702515602112, "learning_rate": 9.709828348349645e-05, "loss": 1.244, "step": 1130 }, { "epoch": 0.1370851370851371, "grad_norm": 0.22943449020385742, "learning_rate": 9.703254942053641e-05, "loss": 1.244, "step": 1140 }, { "epoch": 0.1382876382876383, "grad_norm": 0.22439205646514893, "learning_rate": 9.69661018784341e-05, "loss": 1.2324, "step": 1150 }, { "epoch": 0.1394901394901395, "grad_norm": 0.22962580621242523, "learning_rate": 9.689894186519217e-05, "loss": 1.2437, "step": 1160 }, { "epoch": 0.1406926406926407, "grad_norm": 0.22367720305919647, "learning_rate": 9.683107039962144e-05, "loss": 1.2353, "step": 1170 }, { "epoch": 0.1418951418951419, "grad_norm": 0.23102396726608276, "learning_rate": 9.676248851132536e-05, "loss": 1.224, "step": 1180 }, { "epoch": 0.14309764309764308, "grad_norm": 0.20691992342472076, "learning_rate": 9.669319724068441e-05, "loss": 1.2448, "step": 1190 }, { "epoch": 0.1443001443001443, "grad_norm": 0.2133265882730484, "learning_rate": 9.662319763884035e-05, "loss": 1.2562, "step": 1200 }, { "epoch": 0.1455026455026455, "grad_norm": 0.21801921725273132, "learning_rate": 9.655249076768023e-05, "loss": 1.2366, "step": 1210 }, { "epoch": 0.1467051467051467, "grad_norm": 0.22324581444263458, "learning_rate": 9.648107769982033e-05, "loss": 1.2341, "step": 1220 }, { "epoch": 0.1479076479076479, "grad_norm": 0.20417769253253937, "learning_rate": 9.640895951858984e-05, "loss": 1.2339, "step": 1230 }, { "epoch": 0.1491101491101491, "grad_norm": 0.21648766100406647, "learning_rate": 9.63361373180145e-05, "loss": 1.23, "step": 1240 }, { "epoch": 0.15031265031265031, "grad_norm": 0.21471014618873596, "learning_rate": 9.626261220279988e-05, "loss": 1.2445, "step": 1250 }, { "epoch": 0.15151515151515152, "grad_norm": 0.21205437183380127, "learning_rate": 9.618838528831477e-05, "loss": 1.2434, "step": 1260 }, { "epoch": 0.15271765271765272, "grad_norm": 0.20593945682048798, "learning_rate": 9.611345770057414e-05, "loss": 1.2367, "step": 1270 }, { "epoch": 0.15392015392015393, "grad_norm": 0.2065548151731491, "learning_rate": 9.60378305762221e-05, "loss": 1.2386, "step": 1280 }, { "epoch": 0.15512265512265513, "grad_norm": 0.21276754140853882, "learning_rate": 9.596150506251469e-05, "loss": 1.2358, "step": 1290 }, { "epoch": 0.1563251563251563, "grad_norm": 0.22176425158977509, "learning_rate": 9.588448231730245e-05, "loss": 1.235, "step": 1300 }, { "epoch": 0.15752765752765752, "grad_norm": 0.21053512394428253, "learning_rate": 9.580676350901282e-05, "loss": 1.2195, "step": 1310 }, { "epoch": 0.15873015873015872, "grad_norm": 0.22540676593780518, "learning_rate": 9.572834981663246e-05, "loss": 1.2313, "step": 1320 }, { "epoch": 0.15993265993265993, "grad_norm": 0.20599214732646942, "learning_rate": 9.56492424296894e-05, "loss": 1.2177, "step": 1330 }, { "epoch": 0.16113516113516113, "grad_norm": 0.2064780443906784, "learning_rate": 9.556944254823485e-05, "loss": 1.2278, "step": 1340 }, { "epoch": 0.16233766233766234, "grad_norm": 0.2161046266555786, "learning_rate": 9.548895138282521e-05, "loss": 1.2369, "step": 1350 }, { "epoch": 0.16354016354016354, "grad_norm": 0.2143697589635849, "learning_rate": 9.540777015450349e-05, "loss": 1.2163, "step": 1360 }, { "epoch": 0.16474266474266475, "grad_norm": 0.20479600131511688, "learning_rate": 9.532590009478094e-05, "loss": 1.2185, "step": 1370 }, { "epoch": 0.16594516594516595, "grad_norm": 0.2071349024772644, "learning_rate": 9.52433424456183e-05, "loss": 1.2449, "step": 1380 }, { "epoch": 0.16714766714766716, "grad_norm": 0.22588619589805603, "learning_rate": 9.516009845940697e-05, "loss": 1.2395, "step": 1390 }, { "epoch": 0.16835016835016836, "grad_norm": 0.21358510851860046, "learning_rate": 9.507616939895005e-05, "loss": 1.2259, "step": 1400 }, { "epoch": 0.16955266955266957, "grad_norm": 0.21996615827083588, "learning_rate": 9.49915565374431e-05, "loss": 1.2333, "step": 1410 }, { "epoch": 0.17075517075517074, "grad_norm": 0.22476601600646973, "learning_rate": 9.490626115845489e-05, "loss": 1.2375, "step": 1420 }, { "epoch": 0.17195767195767195, "grad_norm": 0.21539901196956635, "learning_rate": 9.482028455590793e-05, "loss": 1.2351, "step": 1430 }, { "epoch": 0.17316017316017315, "grad_norm": 0.21999700367450714, "learning_rate": 9.47336280340588e-05, "loss": 1.2217, "step": 1440 }, { "epoch": 0.17436267436267436, "grad_norm": 0.20949865877628326, "learning_rate": 9.464629290747842e-05, "loss": 1.2099, "step": 1450 }, { "epoch": 0.17556517556517556, "grad_norm": 0.21222126483917236, "learning_rate": 9.455828050103205e-05, "loss": 1.226, "step": 1460 }, { "epoch": 0.17676767676767677, "grad_norm": 0.22125640511512756, "learning_rate": 9.446959214985922e-05, "loss": 1.2223, "step": 1470 }, { "epoch": 0.17797017797017797, "grad_norm": 0.21773898601531982, "learning_rate": 9.438022919935348e-05, "loss": 1.2289, "step": 1480 }, { "epoch": 0.17917267917267918, "grad_norm": 0.22156096994876862, "learning_rate": 9.4290193005142e-05, "loss": 1.2191, "step": 1490 }, { "epoch": 0.18037518037518038, "grad_norm": 0.2058996856212616, "learning_rate": 9.419948493306497e-05, "loss": 1.2111, "step": 1500 }, { "epoch": 0.1815776815776816, "grad_norm": 0.22363074123859406, "learning_rate": 9.410810635915491e-05, "loss": 1.2171, "step": 1510 }, { "epoch": 0.1827801827801828, "grad_norm": 0.20594915747642517, "learning_rate": 9.401605866961579e-05, "loss": 1.2205, "step": 1520 }, { "epoch": 0.18398268398268397, "grad_norm": 0.22017163038253784, "learning_rate": 9.392334326080199e-05, "loss": 1.2306, "step": 1530 }, { "epoch": 0.18518518518518517, "grad_norm": 0.20129500329494476, "learning_rate": 9.382996153919713e-05, "loss": 1.2044, "step": 1540 }, { "epoch": 0.18638768638768638, "grad_norm": 0.21580781042575836, "learning_rate": 9.373591492139273e-05, "loss": 1.2207, "step": 1550 }, { "epoch": 0.18759018759018758, "grad_norm": 0.20488041639328003, "learning_rate": 9.364120483406672e-05, "loss": 1.2311, "step": 1560 }, { "epoch": 0.1887926887926888, "grad_norm": 0.22156283259391785, "learning_rate": 9.35458327139618e-05, "loss": 1.2158, "step": 1570 }, { "epoch": 0.18999518999519, "grad_norm": 0.22037768363952637, "learning_rate": 9.344980000786363e-05, "loss": 1.2246, "step": 1580 }, { "epoch": 0.1911976911976912, "grad_norm": 0.19976097345352173, "learning_rate": 9.33531081725789e-05, "loss": 1.2157, "step": 1590 }, { "epoch": 0.1924001924001924, "grad_norm": 0.2248261421918869, "learning_rate": 9.325575867491323e-05, "loss": 1.2235, "step": 1600 }, { "epoch": 0.1936026936026936, "grad_norm": 0.22592051327228546, "learning_rate": 9.315775299164892e-05, "loss": 1.2271, "step": 1610 }, { "epoch": 0.19480519480519481, "grad_norm": 0.22421322762966156, "learning_rate": 9.305909260952254e-05, "loss": 1.2378, "step": 1620 }, { "epoch": 0.19600769600769602, "grad_norm": 0.2143516093492508, "learning_rate": 9.295977902520235e-05, "loss": 1.2326, "step": 1630 }, { "epoch": 0.1972101972101972, "grad_norm": 0.2073206901550293, "learning_rate": 9.285981374526564e-05, "loss": 1.2241, "step": 1640 }, { "epoch": 0.1984126984126984, "grad_norm": 0.2234421968460083, "learning_rate": 9.275919828617589e-05, "loss": 1.2277, "step": 1650 }, { "epoch": 0.1996151996151996, "grad_norm": 0.218864306807518, "learning_rate": 9.265793417425967e-05, "loss": 1.2256, "step": 1660 }, { "epoch": 0.2008177008177008, "grad_norm": 0.2139173001050949, "learning_rate": 9.25560229456836e-05, "loss": 1.2247, "step": 1670 }, { "epoch": 0.20202020202020202, "grad_norm": 0.2014576643705368, "learning_rate": 9.245346614643096e-05, "loss": 1.2215, "step": 1680 }, { "epoch": 0.20322270322270322, "grad_norm": 0.22390897572040558, "learning_rate": 9.235026533227833e-05, "loss": 1.2273, "step": 1690 }, { "epoch": 0.20442520442520443, "grad_norm": 0.2203395515680313, "learning_rate": 9.224642206877187e-05, "loss": 1.2067, "step": 1700 }, { "epoch": 0.20562770562770563, "grad_norm": 0.22504153847694397, "learning_rate": 9.214193793120368e-05, "loss": 1.2285, "step": 1710 }, { "epoch": 0.20683020683020684, "grad_norm": 0.2131216675043106, "learning_rate": 9.203681450458781e-05, "loss": 1.2239, "step": 1720 }, { "epoch": 0.20803270803270804, "grad_norm": 0.2105296105146408, "learning_rate": 9.19310533836363e-05, "loss": 1.2145, "step": 1730 }, { "epoch": 0.20923520923520925, "grad_norm": 0.2101157158613205, "learning_rate": 9.182465617273494e-05, "loss": 1.2134, "step": 1740 }, { "epoch": 0.21043771043771045, "grad_norm": 0.20046111941337585, "learning_rate": 9.171762448591894e-05, "loss": 1.2207, "step": 1750 }, { "epoch": 0.21164021164021163, "grad_norm": 0.232273668050766, "learning_rate": 9.160995994684845e-05, "loss": 1.2315, "step": 1760 }, { "epoch": 0.21284271284271283, "grad_norm": 0.21907658874988556, "learning_rate": 9.15016641887839e-05, "loss": 1.2183, "step": 1770 }, { "epoch": 0.21404521404521404, "grad_norm": 0.20053844153881073, "learning_rate": 9.139273885456132e-05, "loss": 1.2187, "step": 1780 }, { "epoch": 0.21524771524771524, "grad_norm": 0.2047508955001831, "learning_rate": 9.128318559656725e-05, "loss": 1.2022, "step": 1790 }, { "epoch": 0.21645021645021645, "grad_norm": 0.21334955096244812, "learning_rate": 9.117300607671386e-05, "loss": 1.1995, "step": 1800 }, { "epoch": 0.21765271765271765, "grad_norm": 0.20535485446453094, "learning_rate": 9.106220196641357e-05, "loss": 1.2255, "step": 1810 }, { "epoch": 0.21885521885521886, "grad_norm": 0.20729170739650726, "learning_rate": 9.095077494655388e-05, "loss": 1.2186, "step": 1820 }, { "epoch": 0.22005772005772006, "grad_norm": 0.20496122539043427, "learning_rate": 9.083872670747166e-05, "loss": 1.2024, "step": 1830 }, { "epoch": 0.22126022126022127, "grad_norm": 0.19766771793365479, "learning_rate": 9.07260589489277e-05, "loss": 1.2097, "step": 1840 }, { "epoch": 0.22246272246272247, "grad_norm": 0.21484710276126862, "learning_rate": 9.061277338008077e-05, "loss": 1.2311, "step": 1850 }, { "epoch": 0.22366522366522368, "grad_norm": 0.206639364361763, "learning_rate": 9.049887171946179e-05, "loss": 1.2167, "step": 1860 }, { "epoch": 0.22486772486772486, "grad_norm": 0.2012685239315033, "learning_rate": 9.038435569494773e-05, "loss": 1.2172, "step": 1870 }, { "epoch": 0.22607022607022606, "grad_norm": 0.2186802476644516, "learning_rate": 9.026922704373543e-05, "loss": 1.2147, "step": 1880 }, { "epoch": 0.22727272727272727, "grad_norm": 0.19656942784786224, "learning_rate": 9.015348751231517e-05, "loss": 1.2212, "step": 1890 }, { "epoch": 0.22847522847522847, "grad_norm": 0.20295904576778412, "learning_rate": 9.003713885644425e-05, "loss": 1.2021, "step": 1900 }, { "epoch": 0.22967772967772968, "grad_norm": 0.2020638883113861, "learning_rate": 8.99201828411203e-05, "loss": 1.2098, "step": 1910 }, { "epoch": 0.23088023088023088, "grad_norm": 0.2027895748615265, "learning_rate": 8.980262124055458e-05, "loss": 1.2116, "step": 1920 }, { "epoch": 0.23208273208273208, "grad_norm": 0.1999751627445221, "learning_rate": 8.968445583814497e-05, "loss": 1.2056, "step": 1930 }, { "epoch": 0.2332852332852333, "grad_norm": 0.20799754559993744, "learning_rate": 8.956568842644901e-05, "loss": 1.2088, "step": 1940 }, { "epoch": 0.2344877344877345, "grad_norm": 0.21467134356498718, "learning_rate": 8.944632080715662e-05, "loss": 1.2239, "step": 1950 }, { "epoch": 0.2356902356902357, "grad_norm": 0.2006864845752716, "learning_rate": 8.932635479106283e-05, "loss": 1.203, "step": 1960 }, { "epoch": 0.2368927368927369, "grad_norm": 0.20447371900081635, "learning_rate": 8.920579219804028e-05, "loss": 1.222, "step": 1970 }, { "epoch": 0.23809523809523808, "grad_norm": 0.21055077016353607, "learning_rate": 8.908463485701164e-05, "loss": 1.2173, "step": 1980 }, { "epoch": 0.2392977392977393, "grad_norm": 0.2071334421634674, "learning_rate": 8.896288460592186e-05, "loss": 1.2014, "step": 1990 }, { "epoch": 0.2405002405002405, "grad_norm": 0.21648217737674713, "learning_rate": 8.884054329171021e-05, "loss": 1.2085, "step": 2000 }, { "epoch": 0.2405002405002405, "eval_loss": 1.1967148780822754, "eval_runtime": 14.3237, "eval_samples_per_second": 17.872, "eval_steps_per_second": 0.559, "step": 2000 }, { "epoch": 0.2417027417027417, "grad_norm": 0.20826809108257294, "learning_rate": 8.871761277028243e-05, "loss": 1.2277, "step": 2010 }, { "epoch": 0.2429052429052429, "grad_norm": 0.2014724761247635, "learning_rate": 8.859409490648238e-05, "loss": 1.2081, "step": 2020 }, { "epoch": 0.2441077441077441, "grad_norm": 0.21313191950321198, "learning_rate": 8.84699915740639e-05, "loss": 1.2047, "step": 2030 }, { "epoch": 0.2453102453102453, "grad_norm": 0.21019341051578522, "learning_rate": 8.834530465566233e-05, "loss": 1.2178, "step": 2040 }, { "epoch": 0.24651274651274652, "grad_norm": 0.2147923707962036, "learning_rate": 8.822003604276595e-05, "loss": 1.2189, "step": 2050 }, { "epoch": 0.24771524771524772, "grad_norm": 0.214359313249588, "learning_rate": 8.809418763568725e-05, "loss": 1.2094, "step": 2060 }, { "epoch": 0.24891774891774893, "grad_norm": 0.21459384262561798, "learning_rate": 8.796776134353418e-05, "loss": 1.216, "step": 2070 }, { "epoch": 0.25012025012025013, "grad_norm": 0.2167990505695343, "learning_rate": 8.784075908418115e-05, "loss": 1.2014, "step": 2080 }, { "epoch": 0.25132275132275134, "grad_norm": 0.2073349505662918, "learning_rate": 8.771318278423992e-05, "loss": 1.2033, "step": 2090 }, { "epoch": 0.25252525252525254, "grad_norm": 0.20970456302165985, "learning_rate": 8.758503437903038e-05, "loss": 1.1972, "step": 2100 }, { "epoch": 0.25372775372775375, "grad_norm": 0.19900967180728912, "learning_rate": 8.745631581255125e-05, "loss": 1.1896, "step": 2110 }, { "epoch": 0.25493025493025495, "grad_norm": 0.21769461035728455, "learning_rate": 8.732702903745048e-05, "loss": 1.213, "step": 2120 }, { "epoch": 0.25613275613275616, "grad_norm": 0.21593014895915985, "learning_rate": 8.719717601499571e-05, "loss": 1.2076, "step": 2130 }, { "epoch": 0.25733525733525736, "grad_norm": 0.20195241272449493, "learning_rate": 8.706675871504451e-05, "loss": 1.2142, "step": 2140 }, { "epoch": 0.2585377585377585, "grad_norm": 0.2088335007429123, "learning_rate": 8.693577911601447e-05, "loss": 1.2198, "step": 2150 }, { "epoch": 0.2597402597402597, "grad_norm": 0.21312519907951355, "learning_rate": 8.680423920485321e-05, "loss": 1.2187, "step": 2160 }, { "epoch": 0.2609427609427609, "grad_norm": 0.20363889634609222, "learning_rate": 8.667214097700824e-05, "loss": 1.1993, "step": 2170 }, { "epoch": 0.2621452621452621, "grad_norm": 0.21147218346595764, "learning_rate": 8.653948643639665e-05, "loss": 1.2278, "step": 2180 }, { "epoch": 0.26334776334776333, "grad_norm": 0.1997493952512741, "learning_rate": 8.640627759537475e-05, "loss": 1.2036, "step": 2190 }, { "epoch": 0.26455026455026454, "grad_norm": 0.21293872594833374, "learning_rate": 8.627251647470756e-05, "loss": 1.2096, "step": 2200 }, { "epoch": 0.26575276575276574, "grad_norm": 0.19573794305324554, "learning_rate": 8.613820510353807e-05, "loss": 1.2034, "step": 2210 }, { "epoch": 0.26695526695526695, "grad_norm": 0.21020345389842987, "learning_rate": 8.600334551935658e-05, "loss": 1.2207, "step": 2220 }, { "epoch": 0.26815776815776815, "grad_norm": 0.20261149108409882, "learning_rate": 8.58679397679697e-05, "loss": 1.2128, "step": 2230 }, { "epoch": 0.26936026936026936, "grad_norm": 0.20684467256069183, "learning_rate": 8.573198990346933e-05, "loss": 1.1972, "step": 2240 }, { "epoch": 0.27056277056277056, "grad_norm": 0.19626250863075256, "learning_rate": 8.559549798820153e-05, "loss": 1.1957, "step": 2250 }, { "epoch": 0.27176527176527177, "grad_norm": 0.20869341492652893, "learning_rate": 8.545846609273522e-05, "loss": 1.2164, "step": 2260 }, { "epoch": 0.27296777296777297, "grad_norm": 0.21097266674041748, "learning_rate": 8.532089629583073e-05, "loss": 1.2131, "step": 2270 }, { "epoch": 0.2741702741702742, "grad_norm": 0.2031770646572113, "learning_rate": 8.518279068440835e-05, "loss": 1.206, "step": 2280 }, { "epoch": 0.2753727753727754, "grad_norm": 0.21206524968147278, "learning_rate": 8.50441513535166e-05, "loss": 1.2118, "step": 2290 }, { "epoch": 0.2765752765752766, "grad_norm": 0.19096438586711884, "learning_rate": 8.490498040630048e-05, "loss": 1.2009, "step": 2300 }, { "epoch": 0.2777777777777778, "grad_norm": 0.20884640514850616, "learning_rate": 8.476527995396953e-05, "loss": 1.1908, "step": 2310 }, { "epoch": 0.278980278980279, "grad_norm": 0.20409995317459106, "learning_rate": 8.462505211576588e-05, "loss": 1.1933, "step": 2320 }, { "epoch": 0.2801827801827802, "grad_norm": 0.2128736823797226, "learning_rate": 8.4484299018932e-05, "loss": 1.207, "step": 2330 }, { "epoch": 0.2813852813852814, "grad_norm": 0.2101830691099167, "learning_rate": 8.434302279867852e-05, "loss": 1.1973, "step": 2340 }, { "epoch": 0.2825877825877826, "grad_norm": 0.19666483998298645, "learning_rate": 8.420122559815177e-05, "loss": 1.1878, "step": 2350 }, { "epoch": 0.2837902837902838, "grad_norm": 0.1945558786392212, "learning_rate": 8.405890956840135e-05, "loss": 1.2162, "step": 2360 }, { "epoch": 0.284992784992785, "grad_norm": 0.2024092972278595, "learning_rate": 8.39160768683474e-05, "loss": 1.2096, "step": 2370 }, { "epoch": 0.28619528619528617, "grad_norm": 0.2207275927066803, "learning_rate": 8.377272966474788e-05, "loss": 1.2031, "step": 2380 }, { "epoch": 0.2873977873977874, "grad_norm": 0.1977287232875824, "learning_rate": 8.362887013216578e-05, "loss": 1.1946, "step": 2390 }, { "epoch": 0.2886002886002886, "grad_norm": 0.19938814640045166, "learning_rate": 8.348450045293604e-05, "loss": 1.2117, "step": 2400 }, { "epoch": 0.2898027898027898, "grad_norm": 0.21072271466255188, "learning_rate": 8.333962281713246e-05, "loss": 1.2087, "step": 2410 }, { "epoch": 0.291005291005291, "grad_norm": 0.2154613584280014, "learning_rate": 8.319423942253451e-05, "loss": 1.1915, "step": 2420 }, { "epoch": 0.2922077922077922, "grad_norm": 0.20417426526546478, "learning_rate": 8.304835247459397e-05, "loss": 1.1966, "step": 2430 }, { "epoch": 0.2934102934102934, "grad_norm": 0.21270157396793365, "learning_rate": 8.290196418640148e-05, "loss": 1.1878, "step": 2440 }, { "epoch": 0.2946127946127946, "grad_norm": 0.19687621295452118, "learning_rate": 8.275507677865295e-05, "loss": 1.2036, "step": 2450 }, { "epoch": 0.2958152958152958, "grad_norm": 0.19195468723773956, "learning_rate": 8.260769247961586e-05, "loss": 1.1901, "step": 2460 }, { "epoch": 0.297017797017797, "grad_norm": 0.2017616629600525, "learning_rate": 8.245981352509555e-05, "loss": 1.2007, "step": 2470 }, { "epoch": 0.2982202982202982, "grad_norm": 0.21277114748954773, "learning_rate": 8.231144215840117e-05, "loss": 1.1899, "step": 2480 }, { "epoch": 0.2994227994227994, "grad_norm": 0.21405024826526642, "learning_rate": 8.216258063031175e-05, "loss": 1.2029, "step": 2490 }, { "epoch": 0.30062530062530063, "grad_norm": 0.20833241939544678, "learning_rate": 8.201323119904201e-05, "loss": 1.2087, "step": 2500 }, { "epoch": 0.30182780182780183, "grad_norm": 0.20577466487884521, "learning_rate": 8.186339613020809e-05, "loss": 1.1951, "step": 2510 }, { "epoch": 0.30303030303030304, "grad_norm": 0.20994292199611664, "learning_rate": 8.171307769679326e-05, "loss": 1.2139, "step": 2520 }, { "epoch": 0.30423280423280424, "grad_norm": 0.2054295688867569, "learning_rate": 8.156227817911333e-05, "loss": 1.2085, "step": 2530 }, { "epoch": 0.30543530543530545, "grad_norm": 0.1991153508424759, "learning_rate": 8.141099986478212e-05, "loss": 1.1936, "step": 2540 }, { "epoch": 0.30663780663780665, "grad_norm": 0.20854881405830383, "learning_rate": 8.125924504867679e-05, "loss": 1.204, "step": 2550 }, { "epoch": 0.30784030784030786, "grad_norm": 0.20640797913074493, "learning_rate": 8.110701603290292e-05, "loss": 1.1792, "step": 2560 }, { "epoch": 0.30904280904280906, "grad_norm": 0.20775394141674042, "learning_rate": 8.09543151267597e-05, "loss": 1.1927, "step": 2570 }, { "epoch": 0.31024531024531027, "grad_norm": 0.21143817901611328, "learning_rate": 8.080114464670485e-05, "loss": 1.1949, "step": 2580 }, { "epoch": 0.3114478114478115, "grad_norm": 0.2006131410598755, "learning_rate": 8.064750691631944e-05, "loss": 1.192, "step": 2590 }, { "epoch": 0.3126503126503126, "grad_norm": 0.2078796625137329, "learning_rate": 8.049340426627274e-05, "loss": 1.2008, "step": 2600 }, { "epoch": 0.31385281385281383, "grad_norm": 0.20483560860157013, "learning_rate": 8.033883903428674e-05, "loss": 1.2157, "step": 2610 }, { "epoch": 0.31505531505531503, "grad_norm": 0.2086118459701538, "learning_rate": 8.018381356510079e-05, "loss": 1.1891, "step": 2620 }, { "epoch": 0.31625781625781624, "grad_norm": 0.20505578815937042, "learning_rate": 8.0028330210436e-05, "loss": 1.2005, "step": 2630 }, { "epoch": 0.31746031746031744, "grad_norm": 0.2115529179573059, "learning_rate": 7.987239132895954e-05, "loss": 1.2056, "step": 2640 }, { "epoch": 0.31866281866281865, "grad_norm": 0.19756314158439636, "learning_rate": 7.97159992862489e-05, "loss": 1.2031, "step": 2650 }, { "epoch": 0.31986531986531985, "grad_norm": 0.20881183445453644, "learning_rate": 7.955915645475593e-05, "loss": 1.1852, "step": 2660 }, { "epoch": 0.32106782106782106, "grad_norm": 0.21202848851680756, "learning_rate": 7.940186521377097e-05, "loss": 1.1975, "step": 2670 }, { "epoch": 0.32227032227032226, "grad_norm": 0.20970021188259125, "learning_rate": 7.924412794938664e-05, "loss": 1.1986, "step": 2680 }, { "epoch": 0.32347282347282347, "grad_norm": 0.2039441466331482, "learning_rate": 7.90859470544617e-05, "loss": 1.2239, "step": 2690 }, { "epoch": 0.3246753246753247, "grad_norm": 0.2211122065782547, "learning_rate": 7.892732492858474e-05, "loss": 1.1958, "step": 2700 }, { "epoch": 0.3258778258778259, "grad_norm": 0.20219026505947113, "learning_rate": 7.876826397803781e-05, "loss": 1.2029, "step": 2710 }, { "epoch": 0.3270803270803271, "grad_norm": 0.197218120098114, "learning_rate": 7.860876661575985e-05, "loss": 1.1929, "step": 2720 }, { "epoch": 0.3282828282828283, "grad_norm": 0.20465494692325592, "learning_rate": 7.844883526131013e-05, "loss": 1.198, "step": 2730 }, { "epoch": 0.3294853294853295, "grad_norm": 0.22161713242530823, "learning_rate": 7.828847234083153e-05, "loss": 1.1981, "step": 2740 }, { "epoch": 0.3306878306878307, "grad_norm": 0.20785829424858093, "learning_rate": 7.81276802870138e-05, "loss": 1.1915, "step": 2750 }, { "epoch": 0.3318903318903319, "grad_norm": 0.20283977687358856, "learning_rate": 7.796646153905653e-05, "loss": 1.1856, "step": 2760 }, { "epoch": 0.3330928330928331, "grad_norm": 0.20320013165473938, "learning_rate": 7.780481854263225e-05, "loss": 1.1918, "step": 2770 }, { "epoch": 0.3342953342953343, "grad_norm": 0.20670950412750244, "learning_rate": 7.76427537498493e-05, "loss": 1.1965, "step": 2780 }, { "epoch": 0.3354978354978355, "grad_norm": 0.199970081448555, "learning_rate": 7.748026961921465e-05, "loss": 1.2006, "step": 2790 }, { "epoch": 0.3367003367003367, "grad_norm": 0.2103297859430313, "learning_rate": 7.731736861559651e-05, "loss": 1.1927, "step": 2800 }, { "epoch": 0.3379028379028379, "grad_norm": 0.2028837651014328, "learning_rate": 7.715405321018707e-05, "loss": 1.2001, "step": 2810 }, { "epoch": 0.33910533910533913, "grad_norm": 0.20478375256061554, "learning_rate": 7.699032588046498e-05, "loss": 1.1841, "step": 2820 }, { "epoch": 0.3403078403078403, "grad_norm": 0.20195071399211884, "learning_rate": 7.682618911015767e-05, "loss": 1.1666, "step": 2830 }, { "epoch": 0.3415103415103415, "grad_norm": 0.21081767976284027, "learning_rate": 7.666164538920378e-05, "loss": 1.1951, "step": 2840 }, { "epoch": 0.3427128427128427, "grad_norm": 0.2054494321346283, "learning_rate": 7.649669721371537e-05, "loss": 1.2028, "step": 2850 }, { "epoch": 0.3439153439153439, "grad_norm": 0.20955149829387665, "learning_rate": 7.633134708594001e-05, "loss": 1.2108, "step": 2860 }, { "epoch": 0.3451178451178451, "grad_norm": 0.2094833254814148, "learning_rate": 7.616559751422285e-05, "loss": 1.2022, "step": 2870 }, { "epoch": 0.3463203463203463, "grad_norm": 0.20287872850894928, "learning_rate": 7.599945101296856e-05, "loss": 1.202, "step": 2880 }, { "epoch": 0.3475228475228475, "grad_norm": 0.2033422887325287, "learning_rate": 7.583291010260321e-05, "loss": 1.1895, "step": 2890 }, { "epoch": 0.3487253487253487, "grad_norm": 0.20638012886047363, "learning_rate": 7.566597730953602e-05, "loss": 1.1981, "step": 2900 }, { "epoch": 0.3499278499278499, "grad_norm": 0.19885151088237762, "learning_rate": 7.549865516612098e-05, "loss": 1.1895, "step": 2910 }, { "epoch": 0.3511303511303511, "grad_norm": 0.20345020294189453, "learning_rate": 7.533094621061854e-05, "loss": 1.1855, "step": 2920 }, { "epoch": 0.35233285233285233, "grad_norm": 0.19838982820510864, "learning_rate": 7.516285298715706e-05, "loss": 1.1892, "step": 2930 }, { "epoch": 0.35353535353535354, "grad_norm": 0.1952482908964157, "learning_rate": 7.499437804569416e-05, "loss": 1.174, "step": 2940 }, { "epoch": 0.35473785473785474, "grad_norm": 0.20667259395122528, "learning_rate": 7.482552394197812e-05, "loss": 1.1913, "step": 2950 }, { "epoch": 0.35594035594035595, "grad_norm": 0.2147095650434494, "learning_rate": 7.465629323750905e-05, "loss": 1.1997, "step": 2960 }, { "epoch": 0.35714285714285715, "grad_norm": 0.21083101630210876, "learning_rate": 7.448668849950008e-05, "loss": 1.2023, "step": 2970 }, { "epoch": 0.35834535834535836, "grad_norm": 0.2125108540058136, "learning_rate": 7.431671230083836e-05, "loss": 1.1925, "step": 2980 }, { "epoch": 0.35954785954785956, "grad_norm": 0.20733962953090668, "learning_rate": 7.414636722004614e-05, "loss": 1.1871, "step": 2990 }, { "epoch": 0.36075036075036077, "grad_norm": 0.20381608605384827, "learning_rate": 7.397565584124144e-05, "loss": 1.1913, "step": 3000 }, { "epoch": 0.36075036075036077, "eval_loss": 1.1786988973617554, "eval_runtime": 14.3861, "eval_samples_per_second": 17.795, "eval_steps_per_second": 0.556, "step": 3000 }, { "epoch": 0.36195286195286197, "grad_norm": 0.19937773048877716, "learning_rate": 7.380458075409912e-05, "loss": 1.1869, "step": 3010 }, { "epoch": 0.3631553631553632, "grad_norm": 0.20516780018806458, "learning_rate": 7.363314455381136e-05, "loss": 1.175, "step": 3020 }, { "epoch": 0.3643578643578644, "grad_norm": 0.20108605921268463, "learning_rate": 7.346134984104846e-05, "loss": 1.1788, "step": 3030 }, { "epoch": 0.3655603655603656, "grad_norm": 0.20104803144931793, "learning_rate": 7.328919922191928e-05, "loss": 1.1832, "step": 3040 }, { "epoch": 0.3667628667628668, "grad_norm": 0.21416722238063812, "learning_rate": 7.311669530793176e-05, "loss": 1.1949, "step": 3050 }, { "epoch": 0.36796536796536794, "grad_norm": 0.19773589074611664, "learning_rate": 7.29438407159533e-05, "loss": 1.1887, "step": 3060 }, { "epoch": 0.36916786916786914, "grad_norm": 0.2011365294456482, "learning_rate": 7.277063806817102e-05, "loss": 1.1924, "step": 3070 }, { "epoch": 0.37037037037037035, "grad_norm": 0.19436123967170715, "learning_rate": 7.259708999205203e-05, "loss": 1.1853, "step": 3080 }, { "epoch": 0.37157287157287155, "grad_norm": 0.20969006419181824, "learning_rate": 7.242319912030356e-05, "loss": 1.1869, "step": 3090 }, { "epoch": 0.37277537277537276, "grad_norm": 0.19676007330417633, "learning_rate": 7.224896809083297e-05, "loss": 1.194, "step": 3100 }, { "epoch": 0.37397787397787396, "grad_norm": 0.20664817094802856, "learning_rate": 7.207439954670784e-05, "loss": 1.1919, "step": 3110 }, { "epoch": 0.37518037518037517, "grad_norm": 0.19718950986862183, "learning_rate": 7.189949613611582e-05, "loss": 1.1814, "step": 3120 }, { "epoch": 0.3763828763828764, "grad_norm": 0.19568046927452087, "learning_rate": 7.172426051232437e-05, "loss": 1.1958, "step": 3130 }, { "epoch": 0.3775853775853776, "grad_norm": 0.21798531711101532, "learning_rate": 7.154869533364067e-05, "loss": 1.1898, "step": 3140 }, { "epoch": 0.3787878787878788, "grad_norm": 0.20783917605876923, "learning_rate": 7.13728032633712e-05, "loss": 1.1806, "step": 3150 }, { "epoch": 0.37999037999038, "grad_norm": 0.20036672055721283, "learning_rate": 7.119658696978132e-05, "loss": 1.1902, "step": 3160 }, { "epoch": 0.3811928811928812, "grad_norm": 0.20404507219791412, "learning_rate": 7.102004912605485e-05, "loss": 1.1825, "step": 3170 }, { "epoch": 0.3823953823953824, "grad_norm": 0.21140117943286896, "learning_rate": 7.084319241025351e-05, "loss": 1.1956, "step": 3180 }, { "epoch": 0.3835978835978836, "grad_norm": 0.1950325220823288, "learning_rate": 7.066601950527625e-05, "loss": 1.1872, "step": 3190 }, { "epoch": 0.3848003848003848, "grad_norm": 0.20742125809192657, "learning_rate": 7.048853309881858e-05, "loss": 1.1929, "step": 3200 }, { "epoch": 0.386002886002886, "grad_norm": 0.21259550750255585, "learning_rate": 7.031073588333179e-05, "loss": 1.1935, "step": 3210 }, { "epoch": 0.3872053872053872, "grad_norm": 0.2075534164905548, "learning_rate": 7.013263055598216e-05, "loss": 1.1821, "step": 3220 }, { "epoch": 0.3884078884078884, "grad_norm": 0.19778668880462646, "learning_rate": 6.995421981860994e-05, "loss": 1.178, "step": 3230 }, { "epoch": 0.38961038961038963, "grad_norm": 0.19642433524131775, "learning_rate": 6.977550637768845e-05, "loss": 1.1993, "step": 3240 }, { "epoch": 0.39081289081289083, "grad_norm": 0.2102692574262619, "learning_rate": 6.959649294428301e-05, "loss": 1.192, "step": 3250 }, { "epoch": 0.39201539201539204, "grad_norm": 0.20500397682189941, "learning_rate": 6.941718223400975e-05, "loss": 1.1994, "step": 3260 }, { "epoch": 0.39321789321789324, "grad_norm": 0.21970514953136444, "learning_rate": 6.923757696699446e-05, "loss": 1.1792, "step": 3270 }, { "epoch": 0.3944203944203944, "grad_norm": 0.19493255019187927, "learning_rate": 6.90576798678314e-05, "loss": 1.1938, "step": 3280 }, { "epoch": 0.3956228956228956, "grad_norm": 0.19722020626068115, "learning_rate": 6.887749366554182e-05, "loss": 1.1904, "step": 3290 }, { "epoch": 0.3968253968253968, "grad_norm": 0.21302542090415955, "learning_rate": 6.869702109353264e-05, "loss": 1.1888, "step": 3300 }, { "epoch": 0.398027898027898, "grad_norm": 0.20728740096092224, "learning_rate": 6.851626488955505e-05, "loss": 1.2069, "step": 3310 }, { "epoch": 0.3992303992303992, "grad_norm": 0.2085309624671936, "learning_rate": 6.833522779566281e-05, "loss": 1.184, "step": 3320 }, { "epoch": 0.4004329004329004, "grad_norm": 0.2124137580394745, "learning_rate": 6.815391255817086e-05, "loss": 1.1877, "step": 3330 }, { "epoch": 0.4016354016354016, "grad_norm": 0.2014608383178711, "learning_rate": 6.797232192761345e-05, "loss": 1.1776, "step": 3340 }, { "epoch": 0.40283790283790283, "grad_norm": 0.19843342900276184, "learning_rate": 6.779045865870256e-05, "loss": 1.1966, "step": 3350 }, { "epoch": 0.40404040404040403, "grad_norm": 0.20755374431610107, "learning_rate": 6.760832551028609e-05, "loss": 1.1876, "step": 3360 }, { "epoch": 0.40524290524290524, "grad_norm": 0.19694215059280396, "learning_rate": 6.742592524530595e-05, "loss": 1.1788, "step": 3370 }, { "epoch": 0.40644540644540644, "grad_norm": 0.20904555916786194, "learning_rate": 6.72432606307562e-05, "loss": 1.1818, "step": 3380 }, { "epoch": 0.40764790764790765, "grad_norm": 0.21066156029701233, "learning_rate": 6.706033443764104e-05, "loss": 1.1899, "step": 3390 }, { "epoch": 0.40885040885040885, "grad_norm": 0.20113535225391388, "learning_rate": 6.687714944093282e-05, "loss": 1.1867, "step": 3400 }, { "epoch": 0.41005291005291006, "grad_norm": 0.198592409491539, "learning_rate": 6.669370841952987e-05, "loss": 1.186, "step": 3410 }, { "epoch": 0.41125541125541126, "grad_norm": 0.2045023888349533, "learning_rate": 6.651001415621446e-05, "loss": 1.1737, "step": 3420 }, { "epoch": 0.41245791245791247, "grad_norm": 0.20362548530101776, "learning_rate": 6.632606943761042e-05, "loss": 1.1602, "step": 3430 }, { "epoch": 0.4136604136604137, "grad_norm": 0.20577895641326904, "learning_rate": 6.614187705414105e-05, "loss": 1.1885, "step": 3440 }, { "epoch": 0.4148629148629149, "grad_norm": 0.202180415391922, "learning_rate": 6.595743979998668e-05, "loss": 1.1863, "step": 3450 }, { "epoch": 0.4160654160654161, "grad_norm": 0.19655773043632507, "learning_rate": 6.577276047304222e-05, "loss": 1.1767, "step": 3460 }, { "epoch": 0.4172679172679173, "grad_norm": 0.1869806945323944, "learning_rate": 6.558784187487494e-05, "loss": 1.1911, "step": 3470 }, { "epoch": 0.4184704184704185, "grad_norm": 0.19990621507167816, "learning_rate": 6.540268681068174e-05, "loss": 1.1714, "step": 3480 }, { "epoch": 0.4196729196729197, "grad_norm": 0.2027176469564438, "learning_rate": 6.521729808924668e-05, "loss": 1.1753, "step": 3490 }, { "epoch": 0.4208754208754209, "grad_norm": 0.19918115437030792, "learning_rate": 6.503167852289844e-05, "loss": 1.1862, "step": 3500 }, { "epoch": 0.42207792207792205, "grad_norm": 0.20005355775356293, "learning_rate": 6.484583092746753e-05, "loss": 1.1848, "step": 3510 }, { "epoch": 0.42328042328042326, "grad_norm": 0.19497598707675934, "learning_rate": 6.46597581222437e-05, "loss": 1.1677, "step": 3520 }, { "epoch": 0.42448292448292446, "grad_norm": 0.21364623308181763, "learning_rate": 6.447346292993307e-05, "loss": 1.1889, "step": 3530 }, { "epoch": 0.42568542568542567, "grad_norm": 0.1999025195837021, "learning_rate": 6.428694817661531e-05, "loss": 1.1833, "step": 3540 }, { "epoch": 0.42688792688792687, "grad_norm": 0.20487244427204132, "learning_rate": 6.410021669170093e-05, "loss": 1.189, "step": 3550 }, { "epoch": 0.4280904280904281, "grad_norm": 0.20337916910648346, "learning_rate": 6.391327130788813e-05, "loss": 1.1849, "step": 3560 }, { "epoch": 0.4292929292929293, "grad_norm": 0.20541474223136902, "learning_rate": 6.372611486111994e-05, "loss": 1.1879, "step": 3570 }, { "epoch": 0.4304954304954305, "grad_norm": 0.1998613476753235, "learning_rate": 6.353875019054128e-05, "loss": 1.1795, "step": 3580 }, { "epoch": 0.4316979316979317, "grad_norm": 0.19440466165542603, "learning_rate": 6.335118013845572e-05, "loss": 1.1865, "step": 3590 }, { "epoch": 0.4329004329004329, "grad_norm": 0.2010820358991623, "learning_rate": 6.316340755028247e-05, "loss": 1.184, "step": 3600 }, { "epoch": 0.4341029341029341, "grad_norm": 0.20916162431240082, "learning_rate": 6.297543527451322e-05, "loss": 1.1758, "step": 3610 }, { "epoch": 0.4353054353054353, "grad_norm": 0.20128333568572998, "learning_rate": 6.278726616266884e-05, "loss": 1.1753, "step": 3620 }, { "epoch": 0.4365079365079365, "grad_norm": 0.20334778726100922, "learning_rate": 6.259890306925627e-05, "loss": 1.1811, "step": 3630 }, { "epoch": 0.4377104377104377, "grad_norm": 0.21417754888534546, "learning_rate": 6.241034885172505e-05, "loss": 1.1873, "step": 3640 }, { "epoch": 0.4389129389129389, "grad_norm": 0.2282254844903946, "learning_rate": 6.22216063704241e-05, "loss": 1.1857, "step": 3650 }, { "epoch": 0.4401154401154401, "grad_norm": 0.1929093897342682, "learning_rate": 6.203267848855828e-05, "loss": 1.1864, "step": 3660 }, { "epoch": 0.44131794131794133, "grad_norm": 0.19937632977962494, "learning_rate": 6.184356807214493e-05, "loss": 1.1752, "step": 3670 }, { "epoch": 0.44252044252044254, "grad_norm": 0.20052993297576904, "learning_rate": 6.165427798997046e-05, "loss": 1.1702, "step": 3680 }, { "epoch": 0.44372294372294374, "grad_norm": 0.20837515592575073, "learning_rate": 6.146481111354676e-05, "loss": 1.1869, "step": 3690 }, { "epoch": 0.44492544492544495, "grad_norm": 0.19809991121292114, "learning_rate": 6.127517031706769e-05, "loss": 1.1922, "step": 3700 }, { "epoch": 0.44612794612794615, "grad_norm": 0.21065044403076172, "learning_rate": 6.108535847736546e-05, "loss": 1.1839, "step": 3710 }, { "epoch": 0.44733044733044736, "grad_norm": 0.20427262783050537, "learning_rate": 6.089537847386696e-05, "loss": 1.1687, "step": 3720 }, { "epoch": 0.4485329485329485, "grad_norm": 0.21668106317520142, "learning_rate": 6.070523318855015e-05, "loss": 1.1811, "step": 3730 }, { "epoch": 0.4497354497354497, "grad_norm": 0.20621725916862488, "learning_rate": 6.051492550590029e-05, "loss": 1.1824, "step": 3740 }, { "epoch": 0.4509379509379509, "grad_norm": 0.21493861079216003, "learning_rate": 6.032445831286615e-05, "loss": 1.1799, "step": 3750 }, { "epoch": 0.4521404521404521, "grad_norm": 0.20860455930233002, "learning_rate": 6.013383449881631e-05, "loss": 1.1793, "step": 3760 }, { "epoch": 0.4533429533429533, "grad_norm": 0.19611907005310059, "learning_rate": 5.994305695549526e-05, "loss": 1.1845, "step": 3770 }, { "epoch": 0.45454545454545453, "grad_norm": 0.20090104639530182, "learning_rate": 5.9752128576979514e-05, "loss": 1.1689, "step": 3780 }, { "epoch": 0.45574795574795574, "grad_norm": 0.22553130984306335, "learning_rate": 5.9561052259633796e-05, "loss": 1.191, "step": 3790 }, { "epoch": 0.45695045695045694, "grad_norm": 0.20412489771842957, "learning_rate": 5.936983090206701e-05, "loss": 1.1689, "step": 3800 }, { "epoch": 0.45815295815295815, "grad_norm": 0.20666392147541046, "learning_rate": 5.917846740508832e-05, "loss": 1.175, "step": 3810 }, { "epoch": 0.45935545935545935, "grad_norm": 0.20357099175453186, "learning_rate": 5.8986964671663115e-05, "loss": 1.1599, "step": 3820 }, { "epoch": 0.46055796055796056, "grad_norm": 0.20838573575019836, "learning_rate": 5.8795325606869025e-05, "loss": 1.1801, "step": 3830 }, { "epoch": 0.46176046176046176, "grad_norm": 0.21168790757656097, "learning_rate": 5.8603553117851753e-05, "loss": 1.1941, "step": 3840 }, { "epoch": 0.46296296296296297, "grad_norm": 0.20583811402320862, "learning_rate": 5.8411650113781094e-05, "loss": 1.1712, "step": 3850 }, { "epoch": 0.46416546416546417, "grad_norm": 0.20696823298931122, "learning_rate": 5.82196195058067e-05, "loss": 1.1683, "step": 3860 }, { "epoch": 0.4653679653679654, "grad_norm": 0.19953173398971558, "learning_rate": 5.802746420701399e-05, "loss": 1.1861, "step": 3870 }, { "epoch": 0.4665704665704666, "grad_norm": 0.20347267389297485, "learning_rate": 5.783518713237993e-05, "loss": 1.1963, "step": 3880 }, { "epoch": 0.4677729677729678, "grad_norm": 0.20178109407424927, "learning_rate": 5.7642791198728795e-05, "loss": 1.165, "step": 3890 }, { "epoch": 0.468975468975469, "grad_norm": 0.20376752316951752, "learning_rate": 5.745027932468794e-05, "loss": 1.1659, "step": 3900 }, { "epoch": 0.4701779701779702, "grad_norm": 0.2073439061641693, "learning_rate": 5.725765443064355e-05, "loss": 1.1892, "step": 3910 }, { "epoch": 0.4713804713804714, "grad_norm": 0.19912444055080414, "learning_rate": 5.7064919438696275e-05, "loss": 1.1858, "step": 3920 }, { "epoch": 0.4725829725829726, "grad_norm": 0.20428648591041565, "learning_rate": 5.687207727261699e-05, "loss": 1.1786, "step": 3930 }, { "epoch": 0.4737854737854738, "grad_norm": 0.21749798953533173, "learning_rate": 5.667913085780232e-05, "loss": 1.1752, "step": 3940 }, { "epoch": 0.474987974987975, "grad_norm": 0.2052384912967682, "learning_rate": 5.648608312123039e-05, "loss": 1.1752, "step": 3950 }, { "epoch": 0.47619047619047616, "grad_norm": 0.20432449877262115, "learning_rate": 5.629293699141638e-05, "loss": 1.1871, "step": 3960 }, { "epoch": 0.47739297739297737, "grad_norm": 0.19696982204914093, "learning_rate": 5.609969539836801e-05, "loss": 1.1618, "step": 3970 }, { "epoch": 0.4785954785954786, "grad_norm": 0.21920330822467804, "learning_rate": 5.5906361273541244e-05, "loss": 1.1877, "step": 3980 }, { "epoch": 0.4797979797979798, "grad_norm": 0.21005605161190033, "learning_rate": 5.5712937549795704e-05, "loss": 1.1934, "step": 3990 }, { "epoch": 0.481000481000481, "grad_norm": 0.20428964495658875, "learning_rate": 5.551942716135021e-05, "loss": 1.1762, "step": 4000 }, { "epoch": 0.481000481000481, "eval_loss": 1.1611641645431519, "eval_runtime": 14.3743, "eval_samples_per_second": 17.81, "eval_steps_per_second": 0.557, "step": 4000 }, { "epoch": 0.4822029822029822, "grad_norm": 0.20652806758880615, "learning_rate": 5.532583304373833e-05, "loss": 1.1826, "step": 4010 }, { "epoch": 0.4834054834054834, "grad_norm": 0.2061479687690735, "learning_rate": 5.513215813376373e-05, "loss": 1.1844, "step": 4020 }, { "epoch": 0.4846079846079846, "grad_norm": 0.2116565853357315, "learning_rate": 5.493840536945574e-05, "loss": 1.172, "step": 4030 }, { "epoch": 0.4858104858104858, "grad_norm": 0.20410141348838806, "learning_rate": 5.47445776900247e-05, "loss": 1.1754, "step": 4040 }, { "epoch": 0.487012987012987, "grad_norm": 0.20021560788154602, "learning_rate": 5.4550678035817404e-05, "loss": 1.1713, "step": 4050 }, { "epoch": 0.4882154882154882, "grad_norm": 0.20905666053295135, "learning_rate": 5.435670934827253e-05, "loss": 1.1746, "step": 4060 }, { "epoch": 0.4894179894179894, "grad_norm": 0.2209169864654541, "learning_rate": 5.416267456987596e-05, "loss": 1.1763, "step": 4070 }, { "epoch": 0.4906204906204906, "grad_norm": 0.19423747062683105, "learning_rate": 5.3968576644116153e-05, "loss": 1.1691, "step": 4080 }, { "epoch": 0.49182299182299183, "grad_norm": 0.19502019882202148, "learning_rate": 5.377441851543954e-05, "loss": 1.1726, "step": 4090 }, { "epoch": 0.49302549302549303, "grad_norm": 0.20400142669677734, "learning_rate": 5.358020312920582e-05, "loss": 1.1736, "step": 4100 }, { "epoch": 0.49422799422799424, "grad_norm": 0.2074715495109558, "learning_rate": 5.338593343164324e-05, "loss": 1.1811, "step": 4110 }, { "epoch": 0.49543049543049544, "grad_norm": 0.20175635814666748, "learning_rate": 5.319161236980402e-05, "loss": 1.1653, "step": 4120 }, { "epoch": 0.49663299663299665, "grad_norm": 0.20830634236335754, "learning_rate": 5.299724289151949e-05, "loss": 1.1735, "step": 4130 }, { "epoch": 0.49783549783549785, "grad_norm": 0.19762635231018066, "learning_rate": 5.280282794535552e-05, "loss": 1.1528, "step": 4140 }, { "epoch": 0.49903799903799906, "grad_norm": 0.22000741958618164, "learning_rate": 5.260837048056768e-05, "loss": 1.1677, "step": 4150 }, { "epoch": 0.5002405002405003, "grad_norm": 0.19429528713226318, "learning_rate": 5.241387344705655e-05, "loss": 1.1617, "step": 4160 }, { "epoch": 0.5014430014430015, "grad_norm": 0.20485764741897583, "learning_rate": 5.2219339795322985e-05, "loss": 1.1745, "step": 4170 }, { "epoch": 0.5026455026455027, "grad_norm": 0.20905248820781708, "learning_rate": 5.2024772476423324e-05, "loss": 1.1609, "step": 4180 }, { "epoch": 0.5038480038480039, "grad_norm": 0.2026577889919281, "learning_rate": 5.183017444192462e-05, "loss": 1.1693, "step": 4190 }, { "epoch": 0.5050505050505051, "grad_norm": 0.20207685232162476, "learning_rate": 5.16355486438599e-05, "loss": 1.1686, "step": 4200 }, { "epoch": 0.5062530062530063, "grad_norm": 0.1890622228384018, "learning_rate": 5.144089803468333e-05, "loss": 1.1672, "step": 4210 }, { "epoch": 0.5074555074555075, "grad_norm": 0.19660572707653046, "learning_rate": 5.1246225567225506e-05, "loss": 1.1672, "step": 4220 }, { "epoch": 0.5086580086580087, "grad_norm": 0.20353861153125763, "learning_rate": 5.1051534194648575e-05, "loss": 1.1759, "step": 4230 }, { "epoch": 0.5098605098605099, "grad_norm": 0.20563137531280518, "learning_rate": 5.085682687040148e-05, "loss": 1.1745, "step": 4240 }, { "epoch": 0.5110630110630111, "grad_norm": 0.20082563161849976, "learning_rate": 5.066210654817514e-05, "loss": 1.1726, "step": 4250 }, { "epoch": 0.5122655122655123, "grad_norm": 0.20972830057144165, "learning_rate": 5.0467376181857694e-05, "loss": 1.1588, "step": 4260 }, { "epoch": 0.5134680134680135, "grad_norm": 0.21431012451648712, "learning_rate": 5.0272638725489595e-05, "loss": 1.159, "step": 4270 }, { "epoch": 0.5146705146705147, "grad_norm": 0.20237547159194946, "learning_rate": 5.0077897133218884e-05, "loss": 1.1649, "step": 4280 }, { "epoch": 0.5158730158730159, "grad_norm": 0.19756945967674255, "learning_rate": 4.988315435925636e-05, "loss": 1.1692, "step": 4290 }, { "epoch": 0.517075517075517, "grad_norm": 0.2220674604177475, "learning_rate": 4.968841335783066e-05, "loss": 1.1713, "step": 4300 }, { "epoch": 0.5182780182780182, "grad_norm": 0.20141322910785675, "learning_rate": 4.949367708314365e-05, "loss": 1.1604, "step": 4310 }, { "epoch": 0.5194805194805194, "grad_norm": 0.2010202258825302, "learning_rate": 4.929894848932541e-05, "loss": 1.1758, "step": 4320 }, { "epoch": 0.5206830206830206, "grad_norm": 0.20987989008426666, "learning_rate": 4.910423053038955e-05, "loss": 1.1618, "step": 4330 }, { "epoch": 0.5218855218855218, "grad_norm": 0.20549160242080688, "learning_rate": 4.890952616018831e-05, "loss": 1.1745, "step": 4340 }, { "epoch": 0.523088023088023, "grad_norm": 0.20881348848342896, "learning_rate": 4.871483833236782e-05, "loss": 1.1615, "step": 4350 }, { "epoch": 0.5242905242905243, "grad_norm": 0.19820332527160645, "learning_rate": 4.852017000032322e-05, "loss": 1.1626, "step": 4360 }, { "epoch": 0.5254930254930255, "grad_norm": 0.20698867738246918, "learning_rate": 4.8325524117154003e-05, "loss": 1.1647, "step": 4370 }, { "epoch": 0.5266955266955267, "grad_norm": 0.1959027200937271, "learning_rate": 4.813090363561896e-05, "loss": 1.1777, "step": 4380 }, { "epoch": 0.5278980278980279, "grad_norm": 0.19776858389377594, "learning_rate": 4.79363115080917e-05, "loss": 1.1574, "step": 4390 }, { "epoch": 0.5291005291005291, "grad_norm": 0.20591264963150024, "learning_rate": 4.774175068651558e-05, "loss": 1.1642, "step": 4400 }, { "epoch": 0.5303030303030303, "grad_norm": 0.20426321029663086, "learning_rate": 4.7547224122359144e-05, "loss": 1.1537, "step": 4410 }, { "epoch": 0.5315055315055315, "grad_norm": 0.1917085498571396, "learning_rate": 4.735273476657116e-05, "loss": 1.1725, "step": 4420 }, { "epoch": 0.5327080327080327, "grad_norm": 0.21214643120765686, "learning_rate": 4.715828556953605e-05, "loss": 1.1736, "step": 4430 }, { "epoch": 0.5339105339105339, "grad_norm": 0.20908360183238983, "learning_rate": 4.696387948102893e-05, "loss": 1.1641, "step": 4440 }, { "epoch": 0.5351130351130351, "grad_norm": 0.2076905071735382, "learning_rate": 4.6769519450171074e-05, "loss": 1.1695, "step": 4450 }, { "epoch": 0.5363155363155363, "grad_norm": 0.212940976023674, "learning_rate": 4.657520842538494e-05, "loss": 1.1578, "step": 4460 }, { "epoch": 0.5375180375180375, "grad_norm": 0.20305852591991425, "learning_rate": 4.63809493543497e-05, "loss": 1.1774, "step": 4470 }, { "epoch": 0.5387205387205387, "grad_norm": 0.20637470483779907, "learning_rate": 4.618674518395628e-05, "loss": 1.1579, "step": 4480 }, { "epoch": 0.5399230399230399, "grad_norm": 0.20714348554611206, "learning_rate": 4.599259886026284e-05, "loss": 1.1549, "step": 4490 }, { "epoch": 0.5411255411255411, "grad_norm": 0.20754988491535187, "learning_rate": 4.5798513328449976e-05, "loss": 1.1563, "step": 4500 }, { "epoch": 0.5423280423280423, "grad_norm": 0.20212216675281525, "learning_rate": 4.560449153277611e-05, "loss": 1.1645, "step": 4510 }, { "epoch": 0.5435305435305435, "grad_norm": 0.1956162005662918, "learning_rate": 4.541053641653274e-05, "loss": 1.1674, "step": 4520 }, { "epoch": 0.5447330447330447, "grad_norm": 0.19848354160785675, "learning_rate": 4.521665092199991e-05, "loss": 1.1677, "step": 4530 }, { "epoch": 0.5459355459355459, "grad_norm": 0.19816353917121887, "learning_rate": 4.502283799040145e-05, "loss": 1.1626, "step": 4540 }, { "epoch": 0.5471380471380471, "grad_norm": 0.20949548482894897, "learning_rate": 4.482910056186047e-05, "loss": 1.1752, "step": 4550 }, { "epoch": 0.5483405483405484, "grad_norm": 0.20675860345363617, "learning_rate": 4.463544157535464e-05, "loss": 1.1584, "step": 4560 }, { "epoch": 0.5495430495430496, "grad_norm": 0.2039814293384552, "learning_rate": 4.444186396867175e-05, "loss": 1.1671, "step": 4570 }, { "epoch": 0.5507455507455508, "grad_norm": 0.2231522798538208, "learning_rate": 4.4248370678365e-05, "loss": 1.1648, "step": 4580 }, { "epoch": 0.551948051948052, "grad_norm": 0.2013349086046219, "learning_rate": 4.4054964639708554e-05, "loss": 1.1764, "step": 4590 }, { "epoch": 0.5531505531505532, "grad_norm": 0.20852096378803253, "learning_rate": 4.386164878665297e-05, "loss": 1.1717, "step": 4600 }, { "epoch": 0.5543530543530544, "grad_norm": 0.21214091777801514, "learning_rate": 4.3668426051780704e-05, "loss": 1.1644, "step": 4610 }, { "epoch": 0.5555555555555556, "grad_norm": 0.215733602643013, "learning_rate": 4.3475299366261604e-05, "loss": 1.1631, "step": 4620 }, { "epoch": 0.5567580567580568, "grad_norm": 0.20751333236694336, "learning_rate": 4.328227165980847e-05, "loss": 1.1545, "step": 4630 }, { "epoch": 0.557960557960558, "grad_norm": 0.1947864145040512, "learning_rate": 4.3089345860632594e-05, "loss": 1.1601, "step": 4640 }, { "epoch": 0.5591630591630592, "grad_norm": 0.20173963904380798, "learning_rate": 4.289652489539937e-05, "loss": 1.1521, "step": 4650 }, { "epoch": 0.5603655603655604, "grad_norm": 0.19809898734092712, "learning_rate": 4.2703811689183797e-05, "loss": 1.1608, "step": 4660 }, { "epoch": 0.5615680615680616, "grad_norm": 0.2071085423231125, "learning_rate": 4.2511209165426306e-05, "loss": 1.1748, "step": 4670 }, { "epoch": 0.5627705627705628, "grad_norm": 0.1951216459274292, "learning_rate": 4.2318720245888135e-05, "loss": 1.1685, "step": 4680 }, { "epoch": 0.563973063973064, "grad_norm": 0.1975964456796646, "learning_rate": 4.212634785060728e-05, "loss": 1.1637, "step": 4690 }, { "epoch": 0.5651755651755652, "grad_norm": 0.2012360543012619, "learning_rate": 4.193409489785403e-05, "loss": 1.1811, "step": 4700 }, { "epoch": 0.5663780663780664, "grad_norm": 0.20509780943393707, "learning_rate": 4.174196430408673e-05, "loss": 1.1719, "step": 4710 }, { "epoch": 0.5675805675805676, "grad_norm": 0.19332098960876465, "learning_rate": 4.1549958983907555e-05, "loss": 1.1605, "step": 4720 }, { "epoch": 0.5687830687830688, "grad_norm": 0.20153947174549103, "learning_rate": 4.1358081850018326e-05, "loss": 1.1546, "step": 4730 }, { "epoch": 0.56998556998557, "grad_norm": 0.2031155526638031, "learning_rate": 4.116633581317623e-05, "loss": 1.1708, "step": 4740 }, { "epoch": 0.5711880711880711, "grad_norm": 0.1971137523651123, "learning_rate": 4.097472378214981e-05, "loss": 1.1527, "step": 4750 }, { "epoch": 0.5723905723905723, "grad_norm": 0.20589309930801392, "learning_rate": 4.0783248663674645e-05, "loss": 1.1551, "step": 4760 }, { "epoch": 0.5735930735930735, "grad_norm": 0.21493607759475708, "learning_rate": 4.059191336240947e-05, "loss": 1.166, "step": 4770 }, { "epoch": 0.5747955747955747, "grad_norm": 0.20907354354858398, "learning_rate": 4.0400720780891935e-05, "loss": 1.1576, "step": 4780 }, { "epoch": 0.575998075998076, "grad_norm": 0.20002475380897522, "learning_rate": 4.0209673819494695e-05, "loss": 1.1631, "step": 4790 }, { "epoch": 0.5772005772005772, "grad_norm": 0.1972057819366455, "learning_rate": 4.00187753763813e-05, "loss": 1.1631, "step": 4800 }, { "epoch": 0.5784030784030784, "grad_norm": 0.2116405963897705, "learning_rate": 3.982802834746236e-05, "loss": 1.1727, "step": 4810 }, { "epoch": 0.5796055796055796, "grad_norm": 0.19388829171657562, "learning_rate": 3.963743562635149e-05, "loss": 1.1569, "step": 4820 }, { "epoch": 0.5808080808080808, "grad_norm": 0.2148236185312271, "learning_rate": 3.9447000104321544e-05, "loss": 1.1676, "step": 4830 }, { "epoch": 0.582010582010582, "grad_norm": 0.20140714943408966, "learning_rate": 3.925672467026057e-05, "loss": 1.1517, "step": 4840 }, { "epoch": 0.5832130832130832, "grad_norm": 0.20167213678359985, "learning_rate": 3.9066612210628216e-05, "loss": 1.1651, "step": 4850 }, { "epoch": 0.5844155844155844, "grad_norm": 0.20786848664283752, "learning_rate": 3.887666560941174e-05, "loss": 1.1512, "step": 4860 }, { "epoch": 0.5856180856180856, "grad_norm": 0.19768615067005157, "learning_rate": 3.868688774808239e-05, "loss": 1.1676, "step": 4870 }, { "epoch": 0.5868205868205868, "grad_norm": 0.21151971817016602, "learning_rate": 3.849728150555161e-05, "loss": 1.1466, "step": 4880 }, { "epoch": 0.588023088023088, "grad_norm": 0.21175932884216309, "learning_rate": 3.8307849758127454e-05, "loss": 1.1508, "step": 4890 }, { "epoch": 0.5892255892255892, "grad_norm": 0.20844349265098572, "learning_rate": 3.8118595379470826e-05, "loss": 1.1757, "step": 4900 }, { "epoch": 0.5904280904280904, "grad_norm": 0.19040937721729279, "learning_rate": 3.7929521240552066e-05, "loss": 1.1708, "step": 4910 }, { "epoch": 0.5916305916305916, "grad_norm": 0.20293647050857544, "learning_rate": 3.77406302096072e-05, "loss": 1.1553, "step": 4920 }, { "epoch": 0.5928330928330928, "grad_norm": 0.21380431950092316, "learning_rate": 3.755192515209458e-05, "loss": 1.1618, "step": 4930 }, { "epoch": 0.594035594035594, "grad_norm": 0.2078305035829544, "learning_rate": 3.7363408930651326e-05, "loss": 1.1498, "step": 4940 }, { "epoch": 0.5952380952380952, "grad_norm": 0.19942502677440643, "learning_rate": 3.7175084405049975e-05, "loss": 1.1555, "step": 4950 }, { "epoch": 0.5964405964405964, "grad_norm": 0.2114267349243164, "learning_rate": 3.6986954432155e-05, "loss": 1.1674, "step": 4960 }, { "epoch": 0.5976430976430976, "grad_norm": 0.19917641580104828, "learning_rate": 3.679902186587959e-05, "loss": 1.1687, "step": 4970 }, { "epoch": 0.5988455988455988, "grad_norm": 0.2174818068742752, "learning_rate": 3.6611289557142226e-05, "loss": 1.1681, "step": 4980 }, { "epoch": 0.6000481000481, "grad_norm": 0.193511962890625, "learning_rate": 3.642376035382359e-05, "loss": 1.1653, "step": 4990 }, { "epoch": 0.6012506012506013, "grad_norm": 0.2069484144449234, "learning_rate": 3.6236437100723194e-05, "loss": 1.1723, "step": 5000 }, { "epoch": 0.6012506012506013, "eval_loss": 1.1470744609832764, "eval_runtime": 14.362, "eval_samples_per_second": 17.825, "eval_steps_per_second": 0.557, "step": 5000 } ], "logging_steps": 10, "max_steps": 8316, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4.858009161721446e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }