{ "best_metric": 0.8637903928756714, "best_model_checkpoint": "CTCLLMs_backboneTrain/checkpoints/LongSpeech_CTC-Shrink_augment_data_self_tokenizer_addMLS_projector_restore_2epoch_OpenASQA_LibriSQA_CommonVoice_random/checkpoint-44000", "epoch": 2.0, "eval_steps": 2000, "global_step": 48676, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008217602103706139, "grad_norm": 2.583192825317383, "learning_rate": 2.737850787132101e-06, "loss": 1.7034, "step": 20 }, { "epoch": 0.0016435204207412278, "grad_norm": 1.2284363508224487, "learning_rate": 5.475701574264202e-06, "loss": 1.3595, "step": 40 }, { "epoch": 0.0024652806311118414, "grad_norm": 0.7835968732833862, "learning_rate": 8.213552361396305e-06, "loss": 1.0712, "step": 60 }, { "epoch": 0.0032870408414824555, "grad_norm": 0.7891668081283569, "learning_rate": 1.0951403148528404e-05, "loss": 0.9601, "step": 80 }, { "epoch": 0.004108801051853069, "grad_norm": 0.7853180170059204, "learning_rate": 1.3689253935660506e-05, "loss": 0.9205, "step": 100 }, { "epoch": 0.004930561262223683, "grad_norm": 0.8246210217475891, "learning_rate": 1.642710472279261e-05, "loss": 0.8811, "step": 120 }, { "epoch": 0.005752321472594297, "grad_norm": 0.8358054757118225, "learning_rate": 1.916495550992471e-05, "loss": 0.8942, "step": 140 }, { "epoch": 0.006574081682964911, "grad_norm": 0.8320625424385071, "learning_rate": 2.190280629705681e-05, "loss": 0.8571, "step": 160 }, { "epoch": 0.007395841893335525, "grad_norm": 0.915210485458374, "learning_rate": 2.464065708418891e-05, "loss": 0.8753, "step": 180 }, { "epoch": 0.008217602103706138, "grad_norm": 0.819446861743927, "learning_rate": 2.7378507871321012e-05, "loss": 0.8164, "step": 200 }, { "epoch": 0.009039362314076753, "grad_norm": 0.8161728978157043, "learning_rate": 3.0116358658453113e-05, "loss": 0.8124, "step": 220 }, { "epoch": 0.009861122524447366, "grad_norm": 0.8150123953819275, "learning_rate": 3.285420944558522e-05, "loss": 0.8046, "step": 240 }, { "epoch": 0.01068288273481798, "grad_norm": 0.7666240334510803, "learning_rate": 3.559206023271732e-05, "loss": 0.804, "step": 260 }, { "epoch": 0.011504642945188595, "grad_norm": 0.7685117721557617, "learning_rate": 3.832991101984942e-05, "loss": 0.7959, "step": 280 }, { "epoch": 0.012326403155559208, "grad_norm": 0.7646723985671997, "learning_rate": 4.1067761806981516e-05, "loss": 0.7919, "step": 300 }, { "epoch": 0.013148163365929822, "grad_norm": 0.6962915062904358, "learning_rate": 4.380561259411362e-05, "loss": 0.8057, "step": 320 }, { "epoch": 0.013969923576300435, "grad_norm": 0.7180787324905396, "learning_rate": 4.654346338124572e-05, "loss": 0.7921, "step": 340 }, { "epoch": 0.01479168378667105, "grad_norm": 0.6863545179367065, "learning_rate": 4.928131416837782e-05, "loss": 0.7752, "step": 360 }, { "epoch": 0.015613443997041664, "grad_norm": 0.7281647324562073, "learning_rate": 5.201916495550992e-05, "loss": 0.8258, "step": 380 }, { "epoch": 0.016435204207412277, "grad_norm": 0.6535085439682007, "learning_rate": 5.4757015742642024e-05, "loss": 0.8181, "step": 400 }, { "epoch": 0.01725696441778289, "grad_norm": 0.722362220287323, "learning_rate": 5.7494866529774125e-05, "loss": 0.7875, "step": 420 }, { "epoch": 0.018078724628153506, "grad_norm": 0.6223776340484619, "learning_rate": 6.023271731690623e-05, "loss": 0.7995, "step": 440 }, { "epoch": 0.01890048483852412, "grad_norm": 0.6497051119804382, "learning_rate": 6.297056810403833e-05, "loss": 0.8061, "step": 460 }, { "epoch": 0.01972224504889473, "grad_norm": 0.6080052256584167, "learning_rate": 6.570841889117044e-05, "loss": 0.7808, "step": 480 }, { "epoch": 0.020544005259265346, "grad_norm": 0.6562979221343994, "learning_rate": 6.844626967830253e-05, "loss": 0.8084, "step": 500 }, { "epoch": 0.02136576546963596, "grad_norm": 0.5958985686302185, "learning_rate": 7.118412046543464e-05, "loss": 0.7635, "step": 520 }, { "epoch": 0.022187525680006575, "grad_norm": 0.6067186594009399, "learning_rate": 7.392197125256673e-05, "loss": 0.8001, "step": 540 }, { "epoch": 0.02300928589037719, "grad_norm": 0.5998467803001404, "learning_rate": 7.665982203969884e-05, "loss": 0.7907, "step": 560 }, { "epoch": 0.0238310461007478, "grad_norm": 0.6743142604827881, "learning_rate": 7.939767282683094e-05, "loss": 0.7716, "step": 580 }, { "epoch": 0.024652806311118415, "grad_norm": 0.5700744390487671, "learning_rate": 8.213552361396303e-05, "loss": 0.775, "step": 600 }, { "epoch": 0.02547456652148903, "grad_norm": 0.5921510457992554, "learning_rate": 8.487337440109514e-05, "loss": 0.7874, "step": 620 }, { "epoch": 0.026296326731859644, "grad_norm": 0.6034173965454102, "learning_rate": 8.761122518822724e-05, "loss": 0.7815, "step": 640 }, { "epoch": 0.02711808694223026, "grad_norm": 0.5742852091789246, "learning_rate": 9.034907597535934e-05, "loss": 0.7881, "step": 660 }, { "epoch": 0.02793984715260087, "grad_norm": 0.5726728439331055, "learning_rate": 9.308692676249144e-05, "loss": 0.7905, "step": 680 }, { "epoch": 0.028761607362971484, "grad_norm": 0.5717937350273132, "learning_rate": 9.582477754962355e-05, "loss": 0.7674, "step": 700 }, { "epoch": 0.0295833675733421, "grad_norm": 0.5768669247627258, "learning_rate": 9.856262833675564e-05, "loss": 0.7687, "step": 720 }, { "epoch": 0.030405127783712713, "grad_norm": 0.577836275100708, "learning_rate": 0.00010130047912388776, "loss": 0.775, "step": 740 }, { "epoch": 0.031226887994083328, "grad_norm": 0.5653334259986877, "learning_rate": 0.00010403832991101984, "loss": 0.7531, "step": 760 }, { "epoch": 0.03204864820445394, "grad_norm": 0.6732868552207947, "learning_rate": 0.00010677618069815197, "loss": 0.7961, "step": 780 }, { "epoch": 0.03287040841482455, "grad_norm": 0.5599177479743958, "learning_rate": 0.00010951403148528405, "loss": 0.7617, "step": 800 }, { "epoch": 0.03369216862519517, "grad_norm": 0.6064387559890747, "learning_rate": 0.00011225188227241617, "loss": 0.7279, "step": 820 }, { "epoch": 0.03451392883556578, "grad_norm": 0.5485804080963135, "learning_rate": 0.00011498973305954825, "loss": 0.7499, "step": 840 }, { "epoch": 0.0353356890459364, "grad_norm": 0.5458228588104248, "learning_rate": 0.00011772758384668037, "loss": 0.7726, "step": 860 }, { "epoch": 0.03615744925630701, "grad_norm": 0.5380986928939819, "learning_rate": 0.00012046543463381245, "loss": 0.8033, "step": 880 }, { "epoch": 0.036979209466677626, "grad_norm": 0.6237996220588684, "learning_rate": 0.00012320328542094456, "loss": 0.7961, "step": 900 }, { "epoch": 0.03780096967704824, "grad_norm": 0.561638593673706, "learning_rate": 0.00012594113620807666, "loss": 0.7799, "step": 920 }, { "epoch": 0.03862272988741885, "grad_norm": 0.5861026644706726, "learning_rate": 0.00012867898699520878, "loss": 0.7801, "step": 940 }, { "epoch": 0.03944449009778946, "grad_norm": 0.5767973065376282, "learning_rate": 0.00013141683778234087, "loss": 0.7743, "step": 960 }, { "epoch": 0.04026625030816008, "grad_norm": 0.5575984120368958, "learning_rate": 0.00013415468856947297, "loss": 0.7781, "step": 980 }, { "epoch": 0.04108801051853069, "grad_norm": 0.6257224678993225, "learning_rate": 0.00013689253935660506, "loss": 0.7942, "step": 1000 }, { "epoch": 0.041909770728901306, "grad_norm": 0.5779247283935547, "learning_rate": 0.00013963039014373718, "loss": 0.7781, "step": 1020 }, { "epoch": 0.04273153093927192, "grad_norm": 0.580823540687561, "learning_rate": 0.00014236824093086928, "loss": 0.7821, "step": 1040 }, { "epoch": 0.043553291149642535, "grad_norm": 0.6481038331985474, "learning_rate": 0.00014510609171800137, "loss": 0.7743, "step": 1060 }, { "epoch": 0.04437505136001315, "grad_norm": 0.5689835548400879, "learning_rate": 0.00014784394250513347, "loss": 0.7839, "step": 1080 }, { "epoch": 0.045196811570383764, "grad_norm": 0.6186492443084717, "learning_rate": 0.0001505817932922656, "loss": 0.8148, "step": 1100 }, { "epoch": 0.04601857178075438, "grad_norm": 0.6207014322280884, "learning_rate": 0.00015331964407939769, "loss": 0.7854, "step": 1120 }, { "epoch": 0.04684033199112499, "grad_norm": 0.649277925491333, "learning_rate": 0.00015605749486652978, "loss": 0.792, "step": 1140 }, { "epoch": 0.0476620922014956, "grad_norm": 0.6359161734580994, "learning_rate": 0.00015879534565366188, "loss": 0.7871, "step": 1160 }, { "epoch": 0.048483852411866216, "grad_norm": 0.6189965009689331, "learning_rate": 0.000161533196440794, "loss": 0.8014, "step": 1180 }, { "epoch": 0.04930561262223683, "grad_norm": 0.6509179472923279, "learning_rate": 0.00016427104722792606, "loss": 0.7993, "step": 1200 }, { "epoch": 0.050127372832607445, "grad_norm": 0.70870041847229, "learning_rate": 0.0001670088980150582, "loss": 0.7892, "step": 1220 }, { "epoch": 0.05094913304297806, "grad_norm": 0.6406404376029968, "learning_rate": 0.00016974674880219028, "loss": 0.7823, "step": 1240 }, { "epoch": 0.051770893253348674, "grad_norm": 0.65333092212677, "learning_rate": 0.0001724845995893224, "loss": 0.788, "step": 1260 }, { "epoch": 0.05259265346371929, "grad_norm": 0.6580297946929932, "learning_rate": 0.00017522245037645447, "loss": 0.7818, "step": 1280 }, { "epoch": 0.0534144136740899, "grad_norm": 0.6168191432952881, "learning_rate": 0.0001779603011635866, "loss": 0.8192, "step": 1300 }, { "epoch": 0.05423617388446052, "grad_norm": 0.6828853487968445, "learning_rate": 0.0001806981519507187, "loss": 0.7965, "step": 1320 }, { "epoch": 0.055057934094831125, "grad_norm": 0.6340402960777283, "learning_rate": 0.0001834360027378508, "loss": 0.7994, "step": 1340 }, { "epoch": 0.05587969430520174, "grad_norm": 0.6091774106025696, "learning_rate": 0.00018617385352498288, "loss": 0.8142, "step": 1360 }, { "epoch": 0.056701454515572354, "grad_norm": 0.6250841617584229, "learning_rate": 0.000188911704312115, "loss": 0.7998, "step": 1380 }, { "epoch": 0.05752321472594297, "grad_norm": 0.6069123148918152, "learning_rate": 0.0001916495550992471, "loss": 0.8138, "step": 1400 }, { "epoch": 0.05834497493631358, "grad_norm": 0.6658041477203369, "learning_rate": 0.00019438740588637922, "loss": 0.8082, "step": 1420 }, { "epoch": 0.0591667351466842, "grad_norm": 0.6972244381904602, "learning_rate": 0.00019712525667351128, "loss": 0.8203, "step": 1440 }, { "epoch": 0.05998849535705481, "grad_norm": 0.6969318985939026, "learning_rate": 0.0001998631074606434, "loss": 0.7889, "step": 1460 }, { "epoch": 0.06081025556742543, "grad_norm": 0.6873449087142944, "learning_rate": 0.00019999992008709735, "loss": 0.8483, "step": 1480 }, { "epoch": 0.06163201577779604, "grad_norm": 0.6407928466796875, "learning_rate": 0.0001999996633033991, "loss": 0.8578, "step": 1500 }, { "epoch": 0.062453775988166656, "grad_norm": 0.6195780634880066, "learning_rate": 0.0001999992294279946, "loss": 0.8053, "step": 1520 }, { "epoch": 0.06327553619853726, "grad_norm": 0.7080503702163696, "learning_rate": 0.00019999861846165223, "loss": 0.7953, "step": 1540 }, { "epoch": 0.06409729640890788, "grad_norm": 0.8221389055252075, "learning_rate": 0.0001999978304054539, "loss": 0.831, "step": 1560 }, { "epoch": 0.06491905661927849, "grad_norm": 0.7362856268882751, "learning_rate": 0.00019999686526079525, "loss": 0.8333, "step": 1580 }, { "epoch": 0.0657408168296491, "grad_norm": 0.7318360805511475, "learning_rate": 0.0001999957230293855, "loss": 0.8352, "step": 1600 }, { "epoch": 0.06656257704001972, "grad_norm": 0.7396681904792786, "learning_rate": 0.0001999944037132474, "loss": 0.8108, "step": 1620 }, { "epoch": 0.06738433725039034, "grad_norm": 0.624099850654602, "learning_rate": 0.00019999290731471738, "loss": 0.8049, "step": 1640 }, { "epoch": 0.06820609746076095, "grad_norm": 0.8170691728591919, "learning_rate": 0.00019999123383644544, "loss": 0.8198, "step": 1660 }, { "epoch": 0.06902785767113157, "grad_norm": 0.7529473304748535, "learning_rate": 0.00019998938328139517, "loss": 0.8276, "step": 1680 }, { "epoch": 0.06984961788150218, "grad_norm": 0.6491063833236694, "learning_rate": 0.0001999873556528438, "loss": 0.8363, "step": 1700 }, { "epoch": 0.0706713780918728, "grad_norm": 0.6839701533317566, "learning_rate": 0.00019998515095438207, "loss": 0.8183, "step": 1720 }, { "epoch": 0.07149313830224341, "grad_norm": 0.7682334184646606, "learning_rate": 0.00019998276918991437, "loss": 0.8124, "step": 1740 }, { "epoch": 0.07231489851261402, "grad_norm": 0.7527047395706177, "learning_rate": 0.00019998021036365856, "loss": 0.8009, "step": 1760 }, { "epoch": 0.07313665872298464, "grad_norm": 0.7900727391242981, "learning_rate": 0.00019997747448014615, "loss": 0.8252, "step": 1780 }, { "epoch": 0.07395841893335525, "grad_norm": 0.799541175365448, "learning_rate": 0.0001999745615442222, "loss": 0.8285, "step": 1800 }, { "epoch": 0.07478017914372587, "grad_norm": 0.7471742033958435, "learning_rate": 0.00019997147156104527, "loss": 0.8078, "step": 1820 }, { "epoch": 0.07560193935409648, "grad_norm": 0.8705668449401855, "learning_rate": 0.00019996820453608752, "loss": 0.847, "step": 1840 }, { "epoch": 0.07642369956446708, "grad_norm": 0.6803821921348572, "learning_rate": 0.00019996476047513454, "loss": 0.8152, "step": 1860 }, { "epoch": 0.0772454597748377, "grad_norm": 0.7618655562400818, "learning_rate": 0.00019996113938428555, "loss": 0.8178, "step": 1880 }, { "epoch": 0.07806721998520831, "grad_norm": 0.7413930296897888, "learning_rate": 0.0001999573412699532, "loss": 0.8538, "step": 1900 }, { "epoch": 0.07888898019557893, "grad_norm": 0.7185872197151184, "learning_rate": 0.0001999533661388637, "loss": 0.8205, "step": 1920 }, { "epoch": 0.07971074040594954, "grad_norm": 0.7739173769950867, "learning_rate": 0.0001999492139980566, "loss": 0.8305, "step": 1940 }, { "epoch": 0.08053250061632015, "grad_norm": 0.7355234026908875, "learning_rate": 0.0001999448848548851, "loss": 0.8434, "step": 1960 }, { "epoch": 0.08135426082669077, "grad_norm": 0.6984680891036987, "learning_rate": 0.00019994037871701577, "loss": 0.8307, "step": 1980 }, { "epoch": 0.08217602103706138, "grad_norm": 0.8437952995300293, "learning_rate": 0.00019993569559242864, "loss": 0.8156, "step": 2000 }, { "epoch": 0.08217602103706138, "eval_loss": 1.0192168951034546, "eval_runtime": 16.6004, "eval_samples_per_second": 157.827, "eval_steps_per_second": 4.94, "step": 2000 }, { "epoch": 0.082997781247432, "grad_norm": 0.7162949442863464, "learning_rate": 0.00019993132946350455, "loss": 0.8269, "step": 2020 }, { "epoch": 0.08381954145780261, "grad_norm": 0.8335065841674805, "learning_rate": 0.00019992631008726108, "loss": 0.8296, "step": 2040 }, { "epoch": 0.08464130166817323, "grad_norm": 0.7615776658058167, "learning_rate": 0.00019992111374921422, "loss": 0.8325, "step": 2060 }, { "epoch": 0.08546306187854384, "grad_norm": 0.8376593589782715, "learning_rate": 0.00019991574045856637, "loss": 0.8299, "step": 2080 }, { "epoch": 0.08628482208891446, "grad_norm": 0.7417710423469543, "learning_rate": 0.00019991019022483312, "loss": 0.8419, "step": 2100 }, { "epoch": 0.08710658229928507, "grad_norm": 0.781035840511322, "learning_rate": 0.00019990446305784358, "loss": 0.8179, "step": 2120 }, { "epoch": 0.08792834250965569, "grad_norm": 0.7956770062446594, "learning_rate": 0.0001998985589677401, "loss": 0.8365, "step": 2140 }, { "epoch": 0.0887501027200263, "grad_norm": 0.8096093535423279, "learning_rate": 0.00019989247796497838, "loss": 0.8122, "step": 2160 }, { "epoch": 0.08957186293039691, "grad_norm": 0.9604154825210571, "learning_rate": 0.00019988622006032736, "loss": 0.8284, "step": 2180 }, { "epoch": 0.09039362314076753, "grad_norm": 0.7302993535995483, "learning_rate": 0.00019988011120562424, "loss": 0.8275, "step": 2200 }, { "epoch": 0.09121538335113814, "grad_norm": 0.7435317635536194, "learning_rate": 0.00019987350837444987, "loss": 0.8482, "step": 2220 }, { "epoch": 0.09203714356150876, "grad_norm": 0.7641948461532593, "learning_rate": 0.00019986672867497988, "loss": 0.8277, "step": 2240 }, { "epoch": 0.09285890377187937, "grad_norm": 0.8367336392402649, "learning_rate": 0.00019985977211922068, "loss": 0.8148, "step": 2260 }, { "epoch": 0.09368066398224997, "grad_norm": 0.8000660538673401, "learning_rate": 0.0001998526387194917, "loss": 0.8525, "step": 2280 }, { "epoch": 0.09450242419262059, "grad_norm": 0.7250078320503235, "learning_rate": 0.0001998453284884257, "loss": 0.8394, "step": 2300 }, { "epoch": 0.0953241844029912, "grad_norm": 0.8181660771369934, "learning_rate": 0.00019983784143896854, "loss": 0.8396, "step": 2320 }, { "epoch": 0.09614594461336182, "grad_norm": 0.8068580031394958, "learning_rate": 0.00019983017758437916, "loss": 0.8431, "step": 2340 }, { "epoch": 0.09696770482373243, "grad_norm": 0.8218814134597778, "learning_rate": 0.0001998223369382297, "loss": 0.8335, "step": 2360 }, { "epoch": 0.09778946503410305, "grad_norm": 0.7909825444221497, "learning_rate": 0.00019981431951440537, "loss": 0.8337, "step": 2380 }, { "epoch": 0.09861122524447366, "grad_norm": 0.7333732843399048, "learning_rate": 0.00019980612532710434, "loss": 0.8423, "step": 2400 }, { "epoch": 0.09943298545484427, "grad_norm": 0.8515979647636414, "learning_rate": 0.00019979775439083795, "loss": 0.8397, "step": 2420 }, { "epoch": 0.10025474566521489, "grad_norm": 0.7907799482345581, "learning_rate": 0.0001997892067204304, "loss": 0.8332, "step": 2440 }, { "epoch": 0.1010765058755855, "grad_norm": 0.862369179725647, "learning_rate": 0.00019978048233101903, "loss": 0.8295, "step": 2460 }, { "epoch": 0.10189826608595612, "grad_norm": 0.733180046081543, "learning_rate": 0.00019977158123805403, "loss": 0.8182, "step": 2480 }, { "epoch": 0.10272002629632673, "grad_norm": 0.8313851952552795, "learning_rate": 0.00019976250345729856, "loss": 0.8501, "step": 2500 }, { "epoch": 0.10354178650669735, "grad_norm": 0.7268729209899902, "learning_rate": 0.0001997532490048287, "loss": 0.8188, "step": 2520 }, { "epoch": 0.10436354671706796, "grad_norm": 0.8857830166816711, "learning_rate": 0.0001997438178970333, "loss": 0.8125, "step": 2540 }, { "epoch": 0.10518530692743858, "grad_norm": 0.8165369033813477, "learning_rate": 0.0001997342101506142, "loss": 0.8283, "step": 2560 }, { "epoch": 0.10600706713780919, "grad_norm": 0.8085136413574219, "learning_rate": 0.00019972442578258597, "loss": 0.8499, "step": 2580 }, { "epoch": 0.1068288273481798, "grad_norm": 0.914569079875946, "learning_rate": 0.00019971446481027591, "loss": 0.8235, "step": 2600 }, { "epoch": 0.10765058755855042, "grad_norm": 0.8516772389411926, "learning_rate": 0.0001997043272513242, "loss": 0.8541, "step": 2620 }, { "epoch": 0.10847234776892103, "grad_norm": 0.8908547163009644, "learning_rate": 0.0001996940131236836, "loss": 0.8696, "step": 2640 }, { "epoch": 0.10929410797929165, "grad_norm": 0.7930579781532288, "learning_rate": 0.00019968352244561976, "loss": 0.8295, "step": 2660 }, { "epoch": 0.11011586818966225, "grad_norm": 0.8197824358940125, "learning_rate": 0.00019967285523571075, "loss": 0.8251, "step": 2680 }, { "epoch": 0.11093762840003286, "grad_norm": 0.8889797925949097, "learning_rate": 0.00019966201151284745, "loss": 0.8482, "step": 2700 }, { "epoch": 0.11175938861040348, "grad_norm": 0.8669871091842651, "learning_rate": 0.0001996509912962332, "loss": 0.848, "step": 2720 }, { "epoch": 0.1125811488207741, "grad_norm": 0.8625435829162598, "learning_rate": 0.00019963979460538398, "loss": 0.853, "step": 2740 }, { "epoch": 0.11340290903114471, "grad_norm": 0.9078089594841003, "learning_rate": 0.00019962842146012828, "loss": 0.8474, "step": 2760 }, { "epoch": 0.11422466924151532, "grad_norm": 0.8258838057518005, "learning_rate": 0.00019961687188060708, "loss": 0.8446, "step": 2780 }, { "epoch": 0.11504642945188594, "grad_norm": 0.8453534841537476, "learning_rate": 0.00019960573637644915, "loss": 0.8416, "step": 2800 }, { "epoch": 0.11586818966225655, "grad_norm": 0.8726224899291992, "learning_rate": 0.00019959384280922383, "loss": 0.8436, "step": 2820 }, { "epoch": 0.11668994987262717, "grad_norm": 0.7921317219734192, "learning_rate": 0.00019958177286896915, "loss": 0.8331, "step": 2840 }, { "epoch": 0.11751171008299778, "grad_norm": 1.0417413711547852, "learning_rate": 0.00019956952657706, "loss": 0.8612, "step": 2860 }, { "epoch": 0.1183334702933684, "grad_norm": 0.8408219218254089, "learning_rate": 0.00019955710395518363, "loss": 0.8326, "step": 2880 }, { "epoch": 0.11915523050373901, "grad_norm": 0.9015172123908997, "learning_rate": 0.00019954450502533954, "loss": 0.831, "step": 2900 }, { "epoch": 0.11997699071410962, "grad_norm": 0.8820521831512451, "learning_rate": 0.00019953172980983949, "loss": 0.8383, "step": 2920 }, { "epoch": 0.12079875092448024, "grad_norm": 0.899238646030426, "learning_rate": 0.00019951877833130737, "loss": 0.8387, "step": 2940 }, { "epoch": 0.12162051113485085, "grad_norm": 0.827013373374939, "learning_rate": 0.00019950565061267929, "loss": 0.8421, "step": 2960 }, { "epoch": 0.12244227134522147, "grad_norm": 0.8747543692588806, "learning_rate": 0.00019949234667720336, "loss": 0.819, "step": 2980 }, { "epoch": 0.12326403155559208, "grad_norm": 0.7559501528739929, "learning_rate": 0.00019947886654843991, "loss": 0.8384, "step": 3000 }, { "epoch": 0.1240857917659627, "grad_norm": 0.8533095121383667, "learning_rate": 0.00019946521025026117, "loss": 0.8291, "step": 3020 }, { "epoch": 0.12490755197633331, "grad_norm": 0.9044725894927979, "learning_rate": 0.00019945137780685138, "loss": 0.8311, "step": 3040 }, { "epoch": 0.1257293121867039, "grad_norm": 1.075642466545105, "learning_rate": 0.00019943736924270679, "loss": 0.8177, "step": 3060 }, { "epoch": 0.12655107239707453, "grad_norm": 0.8421041369438171, "learning_rate": 0.0001994231845826354, "loss": 0.8205, "step": 3080 }, { "epoch": 0.12737283260744514, "grad_norm": 0.9026604890823364, "learning_rate": 0.0001994088238517572, "loss": 0.8431, "step": 3100 }, { "epoch": 0.12819459281781576, "grad_norm": 0.882366955280304, "learning_rate": 0.00019939428707550395, "loss": 0.836, "step": 3120 }, { "epoch": 0.12901635302818637, "grad_norm": 0.8641519546508789, "learning_rate": 0.00019937957427961918, "loss": 0.8237, "step": 3140 }, { "epoch": 0.12983811323855698, "grad_norm": 0.9230402708053589, "learning_rate": 0.00019936468549015804, "loss": 0.847, "step": 3160 }, { "epoch": 0.1306598734489276, "grad_norm": 0.8539056181907654, "learning_rate": 0.0001993496207334875, "loss": 0.8347, "step": 3180 }, { "epoch": 0.1314816336592982, "grad_norm": 1.0330878496170044, "learning_rate": 0.00019933438003628604, "loss": 0.8183, "step": 3200 }, { "epoch": 0.13230339386966883, "grad_norm": 0.9674934148788452, "learning_rate": 0.0001993189634255438, "loss": 0.8457, "step": 3220 }, { "epoch": 0.13312515408003944, "grad_norm": 0.7981789708137512, "learning_rate": 0.00019930337092856243, "loss": 0.8404, "step": 3240 }, { "epoch": 0.13394691429041006, "grad_norm": 0.9389087557792664, "learning_rate": 0.00019928760257295494, "loss": 0.8212, "step": 3260 }, { "epoch": 0.13476867450078067, "grad_norm": 0.907522439956665, "learning_rate": 0.00019927165838664598, "loss": 0.829, "step": 3280 }, { "epoch": 0.1355904347111513, "grad_norm": 0.7471171021461487, "learning_rate": 0.00019925553839787147, "loss": 0.8199, "step": 3300 }, { "epoch": 0.1364121949215219, "grad_norm": 0.8877785205841064, "learning_rate": 0.00019923924263517856, "loss": 0.82, "step": 3320 }, { "epoch": 0.13723395513189252, "grad_norm": 0.9058935046195984, "learning_rate": 0.00019922277112742592, "loss": 0.8416, "step": 3340 }, { "epoch": 0.13805571534226313, "grad_norm": 1.0315223932266235, "learning_rate": 0.0001992061239037833, "loss": 0.8356, "step": 3360 }, { "epoch": 0.13887747555263374, "grad_norm": 0.9553551077842712, "learning_rate": 0.00019918930099373157, "loss": 0.8114, "step": 3380 }, { "epoch": 0.13969923576300436, "grad_norm": 0.8403427600860596, "learning_rate": 0.00019917230242706287, "loss": 0.8311, "step": 3400 }, { "epoch": 0.14052099597337497, "grad_norm": 0.8313325643539429, "learning_rate": 0.00019915512823388034, "loss": 0.8412, "step": 3420 }, { "epoch": 0.1413427561837456, "grad_norm": 0.9792637228965759, "learning_rate": 0.00019913777844459814, "loss": 0.8405, "step": 3440 }, { "epoch": 0.1421645163941162, "grad_norm": 0.8566033244132996, "learning_rate": 0.00019912025308994148, "loss": 0.8007, "step": 3460 }, { "epoch": 0.14298627660448682, "grad_norm": 0.83521568775177, "learning_rate": 0.00019910255220094634, "loss": 0.8301, "step": 3480 }, { "epoch": 0.14380803681485743, "grad_norm": 0.9577664136886597, "learning_rate": 0.00019908467580895964, "loss": 0.8332, "step": 3500 }, { "epoch": 0.14462979702522805, "grad_norm": 0.8675875663757324, "learning_rate": 0.00019906662394563913, "loss": 0.8434, "step": 3520 }, { "epoch": 0.14545155723559866, "grad_norm": 0.9066608548164368, "learning_rate": 0.00019904839664295322, "loss": 0.8273, "step": 3540 }, { "epoch": 0.14627331744596928, "grad_norm": 0.8817648887634277, "learning_rate": 0.00019902999393318113, "loss": 0.8201, "step": 3560 }, { "epoch": 0.1470950776563399, "grad_norm": 0.931098997592926, "learning_rate": 0.00019901141584891262, "loss": 0.8631, "step": 3580 }, { "epoch": 0.1479168378667105, "grad_norm": 0.9475566744804382, "learning_rate": 0.00019899266242304797, "loss": 0.8335, "step": 3600 }, { "epoch": 0.14873859807708112, "grad_norm": 0.9767733812332153, "learning_rate": 0.00019897373368879816, "loss": 0.8283, "step": 3620 }, { "epoch": 0.14956035828745173, "grad_norm": 0.9582000374794006, "learning_rate": 0.00019895462967968444, "loss": 0.857, "step": 3640 }, { "epoch": 0.15038211849782235, "grad_norm": 1.07439124584198, "learning_rate": 0.0001989353504295386, "loss": 0.8316, "step": 3660 }, { "epoch": 0.15120387870819296, "grad_norm": 0.7958013415336609, "learning_rate": 0.00019891589597250265, "loss": 0.8285, "step": 3680 }, { "epoch": 0.15202563891856358, "grad_norm": 0.9222463965415955, "learning_rate": 0.000198896266343029, "loss": 0.8262, "step": 3700 }, { "epoch": 0.15284739912893416, "grad_norm": 0.8631669282913208, "learning_rate": 0.00019887646157588015, "loss": 0.8399, "step": 3720 }, { "epoch": 0.15366915933930478, "grad_norm": 0.9066493511199951, "learning_rate": 0.0001988564817061289, "loss": 0.8441, "step": 3740 }, { "epoch": 0.1544909195496754, "grad_norm": 0.8802810907363892, "learning_rate": 0.000198836326769158, "loss": 0.8382, "step": 3760 }, { "epoch": 0.155312679760046, "grad_norm": 0.8617781400680542, "learning_rate": 0.00019881599680066024, "loss": 0.8322, "step": 3780 }, { "epoch": 0.15613443997041662, "grad_norm": 0.9121951460838318, "learning_rate": 0.00019879549183663854, "loss": 0.8103, "step": 3800 }, { "epoch": 0.15695620018078724, "grad_norm": 0.8364813923835754, "learning_rate": 0.0001987748119134056, "loss": 0.8258, "step": 3820 }, { "epoch": 0.15777796039115785, "grad_norm": 0.971182644367218, "learning_rate": 0.00019875395706758388, "loss": 0.8195, "step": 3840 }, { "epoch": 0.15859972060152847, "grad_norm": 0.9081626534461975, "learning_rate": 0.00019873292733610577, "loss": 0.7991, "step": 3860 }, { "epoch": 0.15942148081189908, "grad_norm": 1.0428310632705688, "learning_rate": 0.00019871278713727932, "loss": 0.8337, "step": 3880 }, { "epoch": 0.1602432410222697, "grad_norm": 0.8817542791366577, "learning_rate": 0.0001986914164861707, "loss": 0.8422, "step": 3900 }, { "epoch": 0.1610650012326403, "grad_norm": 0.9848262667655945, "learning_rate": 0.00019867095248177687, "loss": 0.8604, "step": 3920 }, { "epoch": 0.16188676144301092, "grad_norm": 0.8702915906906128, "learning_rate": 0.00019864924105494623, "loss": 0.824, "step": 3940 }, { "epoch": 0.16270852165338154, "grad_norm": 0.8769912123680115, "learning_rate": 0.00019862735492790314, "loss": 0.8125, "step": 3960 }, { "epoch": 0.16353028186375215, "grad_norm": 0.9141522645950317, "learning_rate": 0.00019860529413940633, "loss": 0.8426, "step": 3980 }, { "epoch": 0.16435204207412277, "grad_norm": 0.9294391870498657, "learning_rate": 0.00019858305872852373, "loss": 0.8605, "step": 4000 }, { "epoch": 0.16435204207412277, "eval_loss": 1.0459537506103516, "eval_runtime": 16.5325, "eval_samples_per_second": 158.476, "eval_steps_per_second": 4.96, "step": 4000 }, { "epoch": 0.16517380228449338, "grad_norm": 0.8110593557357788, "learning_rate": 0.00019856064873463252, "loss": 0.839, "step": 4020 }, { "epoch": 0.165995562494864, "grad_norm": 0.8444618582725525, "learning_rate": 0.00019853806419741908, "loss": 0.8355, "step": 4040 }, { "epoch": 0.1668173227052346, "grad_norm": 1.0283797979354858, "learning_rate": 0.00019851530515687892, "loss": 0.8491, "step": 4060 }, { "epoch": 0.16763908291560523, "grad_norm": 1.0117839574813843, "learning_rate": 0.00019849237165331656, "loss": 0.8466, "step": 4080 }, { "epoch": 0.16846084312597584, "grad_norm": 0.8671867847442627, "learning_rate": 0.0001984692637273454, "loss": 0.8684, "step": 4100 }, { "epoch": 0.16928260333634645, "grad_norm": 1.0693740844726562, "learning_rate": 0.00019844598141988782, "loss": 0.8419, "step": 4120 }, { "epoch": 0.17010436354671707, "grad_norm": 1.088922381401062, "learning_rate": 0.00019842252477217503, "loss": 0.8169, "step": 4140 }, { "epoch": 0.17092612375708768, "grad_norm": 0.8914449214935303, "learning_rate": 0.00019839889382574692, "loss": 0.8305, "step": 4160 }, { "epoch": 0.1717478839674583, "grad_norm": 1.0136765241622925, "learning_rate": 0.00019837508862245208, "loss": 0.8481, "step": 4180 }, { "epoch": 0.1725696441778289, "grad_norm": 0.9876135587692261, "learning_rate": 0.00019835110920444772, "loss": 0.8384, "step": 4200 }, { "epoch": 0.17339140438819953, "grad_norm": 0.9133214354515076, "learning_rate": 0.0001983269556141995, "loss": 0.8679, "step": 4220 }, { "epoch": 0.17421316459857014, "grad_norm": 0.9155489802360535, "learning_rate": 0.0001983026278944816, "loss": 0.8546, "step": 4240 }, { "epoch": 0.17503492480894076, "grad_norm": 1.0384551286697388, "learning_rate": 0.0001982781260883765, "loss": 0.8076, "step": 4260 }, { "epoch": 0.17585668501931137, "grad_norm": 0.9135899543762207, "learning_rate": 0.00019825345023927505, "loss": 0.8113, "step": 4280 }, { "epoch": 0.17667844522968199, "grad_norm": 0.8842368125915527, "learning_rate": 0.00019822860039087628, "loss": 0.8315, "step": 4300 }, { "epoch": 0.1775002054400526, "grad_norm": 0.9855481386184692, "learning_rate": 0.00019820357658718738, "loss": 0.8051, "step": 4320 }, { "epoch": 0.17832196565042321, "grad_norm": 0.9533560276031494, "learning_rate": 0.0001981783788725235, "loss": 0.8215, "step": 4340 }, { "epoch": 0.17914372586079383, "grad_norm": 0.9931275248527527, "learning_rate": 0.00019815300729150793, "loss": 0.8463, "step": 4360 }, { "epoch": 0.17996548607116444, "grad_norm": 0.9402872323989868, "learning_rate": 0.00019812746188907173, "loss": 0.8228, "step": 4380 }, { "epoch": 0.18078724628153506, "grad_norm": 0.8862301111221313, "learning_rate": 0.0001981017427104539, "loss": 0.8377, "step": 4400 }, { "epoch": 0.18160900649190567, "grad_norm": 1.069446086883545, "learning_rate": 0.0001980771485720597, "loss": 0.8357, "step": 4420 }, { "epoch": 0.1824307667022763, "grad_norm": 0.9147841334342957, "learning_rate": 0.0001980510906611715, "loss": 0.8329, "step": 4440 }, { "epoch": 0.1832525269126469, "grad_norm": 0.9529426097869873, "learning_rate": 0.0001980248591093492, "loss": 0.8439, "step": 4460 }, { "epoch": 0.18407428712301752, "grad_norm": 0.9662571549415588, "learning_rate": 0.00019799845396304688, "loss": 0.8325, "step": 4480 }, { "epoch": 0.18489604733338813, "grad_norm": 0.9648681879043579, "learning_rate": 0.000197971875269026, "loss": 0.8324, "step": 4500 }, { "epoch": 0.18571780754375875, "grad_norm": 1.0208882093429565, "learning_rate": 0.00019794646480399925, "loss": 0.8536, "step": 4520 }, { "epoch": 0.18653956775412933, "grad_norm": 1.0145078897476196, "learning_rate": 0.00019791954782758873, "loss": 0.8317, "step": 4540 }, { "epoch": 0.18736132796449995, "grad_norm": 1.0626345872879028, "learning_rate": 0.00019789245744319638, "loss": 0.8443, "step": 4560 }, { "epoch": 0.18818308817487056, "grad_norm": 1.0466902256011963, "learning_rate": 0.00019786519369879716, "loss": 0.823, "step": 4580 }, { "epoch": 0.18900484838524118, "grad_norm": 0.9382312893867493, "learning_rate": 0.00019783775664267302, "loss": 0.8326, "step": 4600 }, { "epoch": 0.1898266085956118, "grad_norm": 0.9509291648864746, "learning_rate": 0.00019781014632341292, "loss": 0.8415, "step": 4620 }, { "epoch": 0.1906483688059824, "grad_norm": 1.018986701965332, "learning_rate": 0.0001977823627899126, "loss": 0.8415, "step": 4640 }, { "epoch": 0.19147012901635302, "grad_norm": 1.0662530660629272, "learning_rate": 0.0001977544060913746, "loss": 0.8453, "step": 4660 }, { "epoch": 0.19229188922672363, "grad_norm": 1.0088191032409668, "learning_rate": 0.000197726276277308, "loss": 0.8611, "step": 4680 }, { "epoch": 0.19311364943709425, "grad_norm": 0.9862759709358215, "learning_rate": 0.0001976979733975286, "loss": 0.8222, "step": 4700 }, { "epoch": 0.19393540964746486, "grad_norm": 0.982672929763794, "learning_rate": 0.0001976694975021586, "loss": 0.8109, "step": 4720 }, { "epoch": 0.19475716985783548, "grad_norm": 0.8332194685935974, "learning_rate": 0.0001976408486416266, "loss": 0.826, "step": 4740 }, { "epoch": 0.1955789300682061, "grad_norm": 1.1170557737350464, "learning_rate": 0.00019761202686666756, "loss": 0.8748, "step": 4760 }, { "epoch": 0.1964006902785767, "grad_norm": 0.8429856300354004, "learning_rate": 0.0001975830322283226, "loss": 0.8304, "step": 4780 }, { "epoch": 0.19722245048894732, "grad_norm": 1.002530813217163, "learning_rate": 0.000197553864777939, "loss": 0.8542, "step": 4800 }, { "epoch": 0.19804421069931794, "grad_norm": 1.2244071960449219, "learning_rate": 0.0001975245245671701, "loss": 0.8368, "step": 4820 }, { "epoch": 0.19886597090968855, "grad_norm": 0.9923454523086548, "learning_rate": 0.0001974950116479751, "loss": 0.8236, "step": 4840 }, { "epoch": 0.19968773112005916, "grad_norm": 0.9827476143836975, "learning_rate": 0.00019746532607261915, "loss": 0.8356, "step": 4860 }, { "epoch": 0.20050949133042978, "grad_norm": 0.9938998222351074, "learning_rate": 0.0001974354678936731, "loss": 0.8383, "step": 4880 }, { "epoch": 0.2013312515408004, "grad_norm": 0.9436901807785034, "learning_rate": 0.00019740543716401346, "loss": 0.8464, "step": 4900 }, { "epoch": 0.202153011751171, "grad_norm": 0.8767272233963013, "learning_rate": 0.0001973752339368224, "loss": 0.849, "step": 4920 }, { "epoch": 0.20297477196154162, "grad_norm": 0.9653998613357544, "learning_rate": 0.00019734485826558747, "loss": 0.8356, "step": 4940 }, { "epoch": 0.20379653217191224, "grad_norm": 0.8907719850540161, "learning_rate": 0.00019731431020410167, "loss": 0.8323, "step": 4960 }, { "epoch": 0.20461829238228285, "grad_norm": 1.1022579669952393, "learning_rate": 0.00019728358980646325, "loss": 0.8437, "step": 4980 }, { "epoch": 0.20544005259265347, "grad_norm": 0.9100618958473206, "learning_rate": 0.00019725269712707566, "loss": 0.8502, "step": 5000 }, { "epoch": 0.20626181280302408, "grad_norm": 0.9730123281478882, "learning_rate": 0.0001972216322206475, "loss": 0.8245, "step": 5020 }, { "epoch": 0.2070835730133947, "grad_norm": 0.9601908922195435, "learning_rate": 0.00019719039514219224, "loss": 0.8326, "step": 5040 }, { "epoch": 0.2079053332237653, "grad_norm": 1.0868589878082275, "learning_rate": 0.00019715898594702843, "loss": 0.8378, "step": 5060 }, { "epoch": 0.20872709343413592, "grad_norm": 0.924371600151062, "learning_rate": 0.0001971274046907793, "loss": 0.832, "step": 5080 }, { "epoch": 0.20954885364450654, "grad_norm": 1.1059744358062744, "learning_rate": 0.00019709565142937287, "loss": 0.8093, "step": 5100 }, { "epoch": 0.21037061385487715, "grad_norm": 1.0641423463821411, "learning_rate": 0.00019706372621904164, "loss": 0.8173, "step": 5120 }, { "epoch": 0.21119237406524777, "grad_norm": 1.1420958042144775, "learning_rate": 0.00019703162911632275, "loss": 0.8424, "step": 5140 }, { "epoch": 0.21201413427561838, "grad_norm": 0.9624399542808533, "learning_rate": 0.00019699936017805768, "loss": 0.8347, "step": 5160 }, { "epoch": 0.212835894485989, "grad_norm": 0.9680808186531067, "learning_rate": 0.00019696691946139225, "loss": 0.8207, "step": 5180 }, { "epoch": 0.2136576546963596, "grad_norm": 0.9132868647575378, "learning_rate": 0.00019693430702377647, "loss": 0.8398, "step": 5200 }, { "epoch": 0.21447941490673023, "grad_norm": 0.9321950674057007, "learning_rate": 0.00019690152292296446, "loss": 0.85, "step": 5220 }, { "epoch": 0.21530117511710084, "grad_norm": 0.9252221584320068, "learning_rate": 0.00019686856721701435, "loss": 0.8251, "step": 5240 }, { "epoch": 0.21612293532747145, "grad_norm": 0.9873983860015869, "learning_rate": 0.00019683543996428811, "loss": 0.8092, "step": 5260 }, { "epoch": 0.21694469553784207, "grad_norm": 1.0303717851638794, "learning_rate": 0.0001968021412234516, "loss": 0.8396, "step": 5280 }, { "epoch": 0.21776645574821268, "grad_norm": 0.9478332996368408, "learning_rate": 0.00019676867105347431, "loss": 0.8194, "step": 5300 }, { "epoch": 0.2185882159585833, "grad_norm": 1.01088547706604, "learning_rate": 0.00019673502951362935, "loss": 0.8207, "step": 5320 }, { "epoch": 0.2194099761689539, "grad_norm": 0.9483580589294434, "learning_rate": 0.00019670121666349327, "loss": 0.8452, "step": 5340 }, { "epoch": 0.2202317363793245, "grad_norm": 1.0161420106887817, "learning_rate": 0.00019666723256294604, "loss": 0.8377, "step": 5360 }, { "epoch": 0.22105349658969511, "grad_norm": 1.0933947563171387, "learning_rate": 0.00019663307727217085, "loss": 0.847, "step": 5380 }, { "epoch": 0.22187525680006573, "grad_norm": 1.0978140830993652, "learning_rate": 0.0001965987508516542, "loss": 0.8423, "step": 5400 }, { "epoch": 0.22269701701043634, "grad_norm": 0.9424787163734436, "learning_rate": 0.00019656425336218544, "loss": 0.8106, "step": 5420 }, { "epoch": 0.22351877722080696, "grad_norm": 0.9634792804718018, "learning_rate": 0.00019652958486485696, "loss": 0.8387, "step": 5440 }, { "epoch": 0.22434053743117757, "grad_norm": 1.0137280225753784, "learning_rate": 0.0001964947454210641, "loss": 0.8375, "step": 5460 }, { "epoch": 0.2251622976415482, "grad_norm": 1.0315325260162354, "learning_rate": 0.00019645973509250467, "loss": 0.851, "step": 5480 }, { "epoch": 0.2259840578519188, "grad_norm": 0.978634238243103, "learning_rate": 0.00019642455394117944, "loss": 0.7957, "step": 5500 }, { "epoch": 0.22680581806228942, "grad_norm": 0.9586151838302612, "learning_rate": 0.00019638920202939142, "loss": 0.8423, "step": 5520 }, { "epoch": 0.22762757827266003, "grad_norm": 0.9884860515594482, "learning_rate": 0.00019635367941974615, "loss": 0.85, "step": 5540 }, { "epoch": 0.22844933848303065, "grad_norm": 0.884602963924408, "learning_rate": 0.00019631798617515144, "loss": 0.8204, "step": 5560 }, { "epoch": 0.22927109869340126, "grad_norm": 0.9110316038131714, "learning_rate": 0.0001962821223588173, "loss": 0.8184, "step": 5580 }, { "epoch": 0.23009285890377187, "grad_norm": 0.9724137783050537, "learning_rate": 0.00019624608803425574, "loss": 0.8198, "step": 5600 }, { "epoch": 0.2309146191141425, "grad_norm": 1.004752278327942, "learning_rate": 0.00019620988326528077, "loss": 0.8502, "step": 5620 }, { "epoch": 0.2317363793245131, "grad_norm": 1.0108088254928589, "learning_rate": 0.00019617350811600831, "loss": 0.8367, "step": 5640 }, { "epoch": 0.23255813953488372, "grad_norm": 1.0130361318588257, "learning_rate": 0.00019613696265085591, "loss": 0.8151, "step": 5660 }, { "epoch": 0.23337989974525433, "grad_norm": 1.0931516885757446, "learning_rate": 0.0001961002469345428, "loss": 0.8427, "step": 5680 }, { "epoch": 0.23420165995562495, "grad_norm": 0.9896870851516724, "learning_rate": 0.00019606336103208968, "loss": 0.8365, "step": 5700 }, { "epoch": 0.23502342016599556, "grad_norm": 1.0382894277572632, "learning_rate": 0.0001960263050088186, "loss": 0.8229, "step": 5720 }, { "epoch": 0.23584518037636618, "grad_norm": 0.9921779036521912, "learning_rate": 0.00019598907893035299, "loss": 0.8462, "step": 5740 }, { "epoch": 0.2366669405867368, "grad_norm": 1.02907395362854, "learning_rate": 0.00019595168286261732, "loss": 0.8369, "step": 5760 }, { "epoch": 0.2374887007971074, "grad_norm": 0.962459921836853, "learning_rate": 0.00019591411687183715, "loss": 0.8207, "step": 5780 }, { "epoch": 0.23831046100747802, "grad_norm": 1.0783615112304688, "learning_rate": 0.000195876381024539, "loss": 0.8238, "step": 5800 }, { "epoch": 0.23913222121784863, "grad_norm": 1.0806901454925537, "learning_rate": 0.00019583847538755014, "loss": 0.8596, "step": 5820 }, { "epoch": 0.23995398142821925, "grad_norm": 1.0261567831039429, "learning_rate": 0.00019580040002799848, "loss": 0.835, "step": 5840 }, { "epoch": 0.24077574163858986, "grad_norm": 1.0381710529327393, "learning_rate": 0.0001957621550133126, "loss": 0.8614, "step": 5860 }, { "epoch": 0.24159750184896048, "grad_norm": 1.0195953845977783, "learning_rate": 0.00019572374041122148, "loss": 0.8411, "step": 5880 }, { "epoch": 0.2424192620593311, "grad_norm": 0.9467645287513733, "learning_rate": 0.0001956851562897544, "loss": 0.8375, "step": 5900 }, { "epoch": 0.2432410222697017, "grad_norm": 0.9575105309486389, "learning_rate": 0.0001956464027172409, "loss": 0.8215, "step": 5920 }, { "epoch": 0.24406278248007232, "grad_norm": 1.0752394199371338, "learning_rate": 0.00019560747976231054, "loss": 0.8377, "step": 5940 }, { "epoch": 0.24488454269044294, "grad_norm": 0.9938384890556335, "learning_rate": 0.0001955683874938929, "loss": 0.8152, "step": 5960 }, { "epoch": 0.24570630290081355, "grad_norm": 0.9279704689979553, "learning_rate": 0.00019552912598121735, "loss": 0.8061, "step": 5980 }, { "epoch": 0.24652806311118416, "grad_norm": 0.9615955948829651, "learning_rate": 0.00019548969529381306, "loss": 0.8492, "step": 6000 }, { "epoch": 0.24652806311118416, "eval_loss": 1.0622910261154175, "eval_runtime": 16.4076, "eval_samples_per_second": 159.682, "eval_steps_per_second": 4.998, "step": 6000 }, { "epoch": 0.24734982332155478, "grad_norm": 1.052895188331604, "learning_rate": 0.0001954500955015087, "loss": 0.8352, "step": 6020 }, { "epoch": 0.2481715835319254, "grad_norm": 1.0382367372512817, "learning_rate": 0.00019541032667443243, "loss": 0.837, "step": 6040 }, { "epoch": 0.248993343742296, "grad_norm": 1.0128381252288818, "learning_rate": 0.00019537038888301183, "loss": 0.8292, "step": 6060 }, { "epoch": 0.24981510395266662, "grad_norm": 1.0595107078552246, "learning_rate": 0.00019533028219797366, "loss": 0.8431, "step": 6080 }, { "epoch": 0.2506368641630372, "grad_norm": 1.0556915998458862, "learning_rate": 0.00019529000669034376, "loss": 0.8316, "step": 6100 }, { "epoch": 0.2514586243734078, "grad_norm": 0.9882562160491943, "learning_rate": 0.00019524956243144692, "loss": 0.8377, "step": 6120 }, { "epoch": 0.25228038458377844, "grad_norm": 1.0965570211410522, "learning_rate": 0.00019520894949290684, "loss": 0.8383, "step": 6140 }, { "epoch": 0.25310214479414905, "grad_norm": 1.075129747390747, "learning_rate": 0.0001951681679466459, "loss": 0.8372, "step": 6160 }, { "epoch": 0.25392390500451967, "grad_norm": 1.1094706058502197, "learning_rate": 0.00019512721786488509, "loss": 0.8321, "step": 6180 }, { "epoch": 0.2547456652148903, "grad_norm": 0.9913383722305298, "learning_rate": 0.00019508609932014382, "loss": 0.8238, "step": 6200 }, { "epoch": 0.2555674254252609, "grad_norm": 1.10612154006958, "learning_rate": 0.0001950448123852399, "loss": 0.8637, "step": 6220 }, { "epoch": 0.2563891856356315, "grad_norm": 1.0326836109161377, "learning_rate": 0.00019500335713328932, "loss": 0.84, "step": 6240 }, { "epoch": 0.2572109458460021, "grad_norm": 0.9649391174316406, "learning_rate": 0.00019496173363770615, "loss": 0.8217, "step": 6260 }, { "epoch": 0.25803270605637274, "grad_norm": 0.9617984890937805, "learning_rate": 0.0001949199419722023, "loss": 0.8537, "step": 6280 }, { "epoch": 0.25885446626674335, "grad_norm": 1.1347591876983643, "learning_rate": 0.0001948779822107877, "loss": 0.8131, "step": 6300 }, { "epoch": 0.25967622647711397, "grad_norm": 0.9121894240379333, "learning_rate": 0.00019483585442776983, "loss": 0.8407, "step": 6320 }, { "epoch": 0.2604979866874846, "grad_norm": 1.203627586364746, "learning_rate": 0.00019479355869775374, "loss": 0.8455, "step": 6340 }, { "epoch": 0.2613197468978552, "grad_norm": 1.0639876127243042, "learning_rate": 0.00019475109509564192, "loss": 0.8466, "step": 6360 }, { "epoch": 0.2621415071082258, "grad_norm": 1.0065891742706299, "learning_rate": 0.00019470846369663413, "loss": 0.8502, "step": 6380 }, { "epoch": 0.2629632673185964, "grad_norm": 1.0567289590835571, "learning_rate": 0.00019466566457622734, "loss": 0.8375, "step": 6400 }, { "epoch": 0.26378502752896704, "grad_norm": 1.1206752061843872, "learning_rate": 0.0001946226978102154, "loss": 0.8333, "step": 6420 }, { "epoch": 0.26460678773933766, "grad_norm": 1.0563714504241943, "learning_rate": 0.00019457956347468925, "loss": 0.836, "step": 6440 }, { "epoch": 0.26542854794970827, "grad_norm": 1.051429033279419, "learning_rate": 0.0001945362616460364, "loss": 0.8287, "step": 6460 }, { "epoch": 0.2662503081600789, "grad_norm": 0.9783703088760376, "learning_rate": 0.0001944927924009411, "loss": 0.8148, "step": 6480 }, { "epoch": 0.2670720683704495, "grad_norm": 1.2782011032104492, "learning_rate": 0.00019444915581638404, "loss": 0.866, "step": 6500 }, { "epoch": 0.2678938285808201, "grad_norm": 0.8880527019500732, "learning_rate": 0.0001944053519696422, "loss": 0.8282, "step": 6520 }, { "epoch": 0.26871558879119073, "grad_norm": 1.0727986097335815, "learning_rate": 0.0001943613809382889, "loss": 0.8523, "step": 6540 }, { "epoch": 0.26953734900156134, "grad_norm": 1.0758675336837769, "learning_rate": 0.00019431724280019342, "loss": 0.8381, "step": 6560 }, { "epoch": 0.27035910921193196, "grad_norm": 1.10956609249115, "learning_rate": 0.00019427293763352096, "loss": 0.8159, "step": 6580 }, { "epoch": 0.2711808694223026, "grad_norm": 1.1092921495437622, "learning_rate": 0.00019422846551673262, "loss": 0.8387, "step": 6600 }, { "epoch": 0.2720026296326732, "grad_norm": 1.057029366493225, "learning_rate": 0.00019418382652858506, "loss": 0.8306, "step": 6620 }, { "epoch": 0.2728243898430438, "grad_norm": 0.9744523763656616, "learning_rate": 0.00019413902074813047, "loss": 0.8582, "step": 6640 }, { "epoch": 0.2736461500534144, "grad_norm": 0.9532150626182556, "learning_rate": 0.00019409404825471654, "loss": 0.8531, "step": 6660 }, { "epoch": 0.27446791026378503, "grad_norm": 1.0654603242874146, "learning_rate": 0.00019404890912798597, "loss": 0.8437, "step": 6680 }, { "epoch": 0.27528967047415565, "grad_norm": 1.0381238460540771, "learning_rate": 0.00019400360344787676, "loss": 0.8618, "step": 6700 }, { "epoch": 0.27611143068452626, "grad_norm": 1.1071590185165405, "learning_rate": 0.00019395813129462176, "loss": 0.824, "step": 6720 }, { "epoch": 0.2769331908948969, "grad_norm": 1.1044433116912842, "learning_rate": 0.00019391249274874865, "loss": 0.8332, "step": 6740 }, { "epoch": 0.2777549511052675, "grad_norm": 1.145683765411377, "learning_rate": 0.00019386668789107977, "loss": 0.83, "step": 6760 }, { "epoch": 0.2785767113156381, "grad_norm": 1.196481466293335, "learning_rate": 0.00019382071680273198, "loss": 0.8368, "step": 6780 }, { "epoch": 0.2793984715260087, "grad_norm": 1.0642255544662476, "learning_rate": 0.00019377457956511662, "loss": 0.8439, "step": 6800 }, { "epoch": 0.28022023173637933, "grad_norm": 1.0132989883422852, "learning_rate": 0.0001937282762599391, "loss": 0.8403, "step": 6820 }, { "epoch": 0.28104199194674995, "grad_norm": 1.1021807193756104, "learning_rate": 0.00019368180696919905, "loss": 0.8373, "step": 6840 }, { "epoch": 0.28186375215712056, "grad_norm": 1.0249390602111816, "learning_rate": 0.00019363517177519004, "loss": 0.8246, "step": 6860 }, { "epoch": 0.2826855123674912, "grad_norm": 1.0310267210006714, "learning_rate": 0.0001935883707604993, "loss": 0.8266, "step": 6880 }, { "epoch": 0.2835072725778618, "grad_norm": 1.1064010858535767, "learning_rate": 0.00019354140400800797, "loss": 0.8403, "step": 6900 }, { "epoch": 0.2843290327882324, "grad_norm": 1.0507344007492065, "learning_rate": 0.0001934942716008904, "loss": 0.8365, "step": 6920 }, { "epoch": 0.285150792998603, "grad_norm": 1.2774583101272583, "learning_rate": 0.00019344697362261458, "loss": 0.8394, "step": 6940 }, { "epoch": 0.28597255320897363, "grad_norm": 1.1305222511291504, "learning_rate": 0.0001933995101569415, "loss": 0.8446, "step": 6960 }, { "epoch": 0.28679431341934425, "grad_norm": 1.0519880056381226, "learning_rate": 0.00019335188128792542, "loss": 0.8589, "step": 6980 }, { "epoch": 0.28761607362971486, "grad_norm": 0.9783779978752136, "learning_rate": 0.00019330408709991326, "loss": 0.8364, "step": 7000 }, { "epoch": 0.2884378338400855, "grad_norm": 1.033211350440979, "learning_rate": 0.0001932561276775449, "loss": 0.8412, "step": 7020 }, { "epoch": 0.2892595940504561, "grad_norm": 1.1831096410751343, "learning_rate": 0.00019320800310575288, "loss": 0.8495, "step": 7040 }, { "epoch": 0.2900813542608267, "grad_norm": 1.0064650774002075, "learning_rate": 0.00019315971346976193, "loss": 0.8482, "step": 7060 }, { "epoch": 0.2909031144711973, "grad_norm": 1.0976219177246094, "learning_rate": 0.00019311125885508945, "loss": 0.8173, "step": 7080 }, { "epoch": 0.29172487468156794, "grad_norm": 1.110113263130188, "learning_rate": 0.00019306263934754477, "loss": 0.8285, "step": 7100 }, { "epoch": 0.29254663489193855, "grad_norm": 1.0953800678253174, "learning_rate": 0.0001930138550332292, "loss": 0.8382, "step": 7120 }, { "epoch": 0.29336839510230917, "grad_norm": 1.049208402633667, "learning_rate": 0.0001929649059985362, "loss": 0.8238, "step": 7140 }, { "epoch": 0.2941901553126798, "grad_norm": 1.096807837486267, "learning_rate": 0.0001929157923301506, "loss": 0.8139, "step": 7160 }, { "epoch": 0.2950119155230504, "grad_norm": 1.2268364429473877, "learning_rate": 0.00019286651411504893, "loss": 0.8349, "step": 7180 }, { "epoch": 0.295833675733421, "grad_norm": 1.0595046281814575, "learning_rate": 0.00019281707144049915, "loss": 0.8448, "step": 7200 }, { "epoch": 0.2966554359437916, "grad_norm": 1.0562009811401367, "learning_rate": 0.00019276746439406047, "loss": 0.834, "step": 7220 }, { "epoch": 0.29747719615416224, "grad_norm": 1.0876846313476562, "learning_rate": 0.0001927176930635831, "loss": 0.8201, "step": 7240 }, { "epoch": 0.29829895636453285, "grad_norm": 0.9775159955024719, "learning_rate": 0.00019266775753720822, "loss": 0.8612, "step": 7260 }, { "epoch": 0.29912071657490347, "grad_norm": 1.1259452104568481, "learning_rate": 0.00019261765790336784, "loss": 0.8139, "step": 7280 }, { "epoch": 0.2999424767852741, "grad_norm": 1.00784432888031, "learning_rate": 0.00019256739425078454, "loss": 0.8237, "step": 7300 }, { "epoch": 0.3007642369956447, "grad_norm": 1.0394659042358398, "learning_rate": 0.00019251696666847137, "loss": 0.8692, "step": 7320 }, { "epoch": 0.3015859972060153, "grad_norm": 1.1670759916305542, "learning_rate": 0.00019246637524573173, "loss": 0.8478, "step": 7340 }, { "epoch": 0.3024077574163859, "grad_norm": 1.0558823347091675, "learning_rate": 0.0001924156200721591, "loss": 0.8413, "step": 7360 }, { "epoch": 0.30322951762675654, "grad_norm": 1.0909830331802368, "learning_rate": 0.000192364701237637, "loss": 0.8584, "step": 7380 }, { "epoch": 0.30405127783712715, "grad_norm": 0.9988498091697693, "learning_rate": 0.00019231361883233878, "loss": 0.8212, "step": 7400 }, { "epoch": 0.3048730380474977, "grad_norm": 1.0689078569412231, "learning_rate": 0.00019226237294672744, "loss": 0.8464, "step": 7420 }, { "epoch": 0.3056947982578683, "grad_norm": 1.0881212949752808, "learning_rate": 0.00019221096367155548, "loss": 0.8607, "step": 7440 }, { "epoch": 0.30651655846823894, "grad_norm": 1.1320979595184326, "learning_rate": 0.00019215939109786477, "loss": 0.8359, "step": 7460 }, { "epoch": 0.30733831867860956, "grad_norm": 0.8782603144645691, "learning_rate": 0.0001921076553169864, "loss": 0.8558, "step": 7480 }, { "epoch": 0.30816007888898017, "grad_norm": 1.0550236701965332, "learning_rate": 0.00019205575642054044, "loss": 0.8088, "step": 7500 }, { "epoch": 0.3089818390993508, "grad_norm": 1.1194961071014404, "learning_rate": 0.0001920036945004358, "loss": 0.8233, "step": 7520 }, { "epoch": 0.3098035993097214, "grad_norm": 1.10885751247406, "learning_rate": 0.00019195146964887024, "loss": 0.8341, "step": 7540 }, { "epoch": 0.310625359520092, "grad_norm": 1.015629768371582, "learning_rate": 0.0001918990819583298, "loss": 0.8547, "step": 7560 }, { "epoch": 0.31144711973046263, "grad_norm": 1.142196536064148, "learning_rate": 0.0001918465315215892, "loss": 0.856, "step": 7580 }, { "epoch": 0.31226887994083324, "grad_norm": 0.9691776037216187, "learning_rate": 0.0001917938184317111, "loss": 0.8295, "step": 7600 }, { "epoch": 0.31309064015120386, "grad_norm": 0.9759687781333923, "learning_rate": 0.00019174094278204636, "loss": 0.8333, "step": 7620 }, { "epoch": 0.3139124003615745, "grad_norm": 1.148779034614563, "learning_rate": 0.00019168790466623375, "loss": 0.8263, "step": 7640 }, { "epoch": 0.3147341605719451, "grad_norm": 1.2947983741760254, "learning_rate": 0.00019163470417819963, "loss": 0.8627, "step": 7660 }, { "epoch": 0.3155559207823157, "grad_norm": 1.0895724296569824, "learning_rate": 0.00019158134141215792, "loss": 0.8429, "step": 7680 }, { "epoch": 0.3163776809926863, "grad_norm": 1.070154070854187, "learning_rate": 0.0001915278164626101, "loss": 0.8188, "step": 7700 }, { "epoch": 0.31719944120305693, "grad_norm": 1.00252366065979, "learning_rate": 0.00019147412942434463, "loss": 0.8482, "step": 7720 }, { "epoch": 0.31802120141342755, "grad_norm": 1.273224949836731, "learning_rate": 0.00019142028039243717, "loss": 0.8326, "step": 7740 }, { "epoch": 0.31884296162379816, "grad_norm": 1.2733259201049805, "learning_rate": 0.00019136626946225017, "loss": 0.8377, "step": 7760 }, { "epoch": 0.3196647218341688, "grad_norm": 0.9789584279060364, "learning_rate": 0.00019131209672943288, "loss": 0.8355, "step": 7780 }, { "epoch": 0.3204864820445394, "grad_norm": 1.1093429327011108, "learning_rate": 0.00019125776228992103, "loss": 0.836, "step": 7800 }, { "epoch": 0.32130824225491, "grad_norm": 1.0080488920211792, "learning_rate": 0.00019120326623993668, "loss": 0.8375, "step": 7820 }, { "epoch": 0.3221300024652806, "grad_norm": 1.0703438520431519, "learning_rate": 0.0001911486086759882, "loss": 0.8306, "step": 7840 }, { "epoch": 0.32295176267565123, "grad_norm": 1.0767182111740112, "learning_rate": 0.0001910937896948699, "loss": 0.8483, "step": 7860 }, { "epoch": 0.32377352288602185, "grad_norm": 0.9747923612594604, "learning_rate": 0.00019103880939366197, "loss": 0.8489, "step": 7880 }, { "epoch": 0.32459528309639246, "grad_norm": 1.0351313352584839, "learning_rate": 0.00019098366786973032, "loss": 0.8351, "step": 7900 }, { "epoch": 0.3254170433067631, "grad_norm": 1.049666404724121, "learning_rate": 0.00019092836522072631, "loss": 0.8271, "step": 7920 }, { "epoch": 0.3262388035171337, "grad_norm": 1.0181846618652344, "learning_rate": 0.0001908729015445867, "loss": 0.8352, "step": 7940 }, { "epoch": 0.3270605637275043, "grad_norm": 1.1390068531036377, "learning_rate": 0.00019081727693953337, "loss": 0.8392, "step": 7960 }, { "epoch": 0.3278823239378749, "grad_norm": 1.0242650508880615, "learning_rate": 0.00019076149150407324, "loss": 0.8398, "step": 7980 }, { "epoch": 0.32870408414824553, "grad_norm": 1.0052822828292847, "learning_rate": 0.0001907083464611993, "loss": 0.8257, "step": 8000 }, { "epoch": 0.32870408414824553, "eval_loss": 1.0951544046401978, "eval_runtime": 16.4946, "eval_samples_per_second": 158.84, "eval_steps_per_second": 4.971, "step": 8000 }, { "epoch": 0.32952584435861615, "grad_norm": 1.0948665142059326, "learning_rate": 0.00019065224769085476, "loss": 0.8343, "step": 8020 }, { "epoch": 0.33034760456898676, "grad_norm": 1.1585348844528198, "learning_rate": 0.00019059598838235754, "loss": 0.8272, "step": 8040 }, { "epoch": 0.3311693647793574, "grad_norm": 1.0641188621520996, "learning_rate": 0.00019053956863533854, "loss": 0.8499, "step": 8060 }, { "epoch": 0.331991124989728, "grad_norm": 1.0595240592956543, "learning_rate": 0.00019048298854971272, "loss": 0.8407, "step": 8080 }, { "epoch": 0.3328128852000986, "grad_norm": 1.1425433158874512, "learning_rate": 0.00019042624822567908, "loss": 0.8671, "step": 8100 }, { "epoch": 0.3336346454104692, "grad_norm": 1.1736706495285034, "learning_rate": 0.0001903693477637204, "loss": 0.8115, "step": 8120 }, { "epoch": 0.33445640562083984, "grad_norm": 1.062788724899292, "learning_rate": 0.000190312287264603, "loss": 0.8416, "step": 8140 }, { "epoch": 0.33527816583121045, "grad_norm": 1.0873854160308838, "learning_rate": 0.0001902550668293766, "loss": 0.8513, "step": 8160 }, { "epoch": 0.33609992604158107, "grad_norm": 1.0588126182556152, "learning_rate": 0.00019019768655937423, "loss": 0.8232, "step": 8180 }, { "epoch": 0.3369216862519517, "grad_norm": 1.0401087999343872, "learning_rate": 0.00019014014655621193, "loss": 0.8369, "step": 8200 }, { "epoch": 0.3377434464623223, "grad_norm": 0.988150417804718, "learning_rate": 0.0001900824469217886, "loss": 0.8076, "step": 8220 }, { "epoch": 0.3385652066726929, "grad_norm": 1.2257081270217896, "learning_rate": 0.00019002458775828584, "loss": 0.824, "step": 8240 }, { "epoch": 0.3393869668830635, "grad_norm": 0.9933615922927856, "learning_rate": 0.0001899665691681678, "loss": 0.8391, "step": 8260 }, { "epoch": 0.34020872709343414, "grad_norm": 1.1779851913452148, "learning_rate": 0.0001899083912541809, "loss": 0.8333, "step": 8280 }, { "epoch": 0.34103048730380475, "grad_norm": 1.118120551109314, "learning_rate": 0.0001898500541193538, "loss": 0.8333, "step": 8300 }, { "epoch": 0.34185224751417537, "grad_norm": 1.0865180492401123, "learning_rate": 0.00018979155786699706, "loss": 0.8497, "step": 8320 }, { "epoch": 0.342674007724546, "grad_norm": 0.9836400747299194, "learning_rate": 0.000189732902600703, "loss": 0.8547, "step": 8340 }, { "epoch": 0.3434957679349166, "grad_norm": 1.1521192789077759, "learning_rate": 0.00018967408842434562, "loss": 0.8476, "step": 8360 }, { "epoch": 0.3443175281452872, "grad_norm": 1.1589045524597168, "learning_rate": 0.0001896151154420803, "loss": 0.8479, "step": 8380 }, { "epoch": 0.3451392883556578, "grad_norm": 1.0231435298919678, "learning_rate": 0.00018955598375834364, "loss": 0.863, "step": 8400 }, { "epoch": 0.34596104856602844, "grad_norm": 1.0295898914337158, "learning_rate": 0.00018949669347785328, "loss": 0.8224, "step": 8420 }, { "epoch": 0.34678280877639905, "grad_norm": 1.1238269805908203, "learning_rate": 0.00018943724470560778, "loss": 0.8276, "step": 8440 }, { "epoch": 0.34760456898676967, "grad_norm": 1.0870115756988525, "learning_rate": 0.00018937763754688634, "loss": 0.8372, "step": 8460 }, { "epoch": 0.3484263291971403, "grad_norm": 1.1568728685379028, "learning_rate": 0.0001893178721072486, "loss": 0.862, "step": 8480 }, { "epoch": 0.3492480894075109, "grad_norm": 1.0375559329986572, "learning_rate": 0.00018925794849253462, "loss": 0.843, "step": 8500 }, { "epoch": 0.3500698496178815, "grad_norm": 1.1177926063537598, "learning_rate": 0.00018919786680886443, "loss": 0.8303, "step": 8520 }, { "epoch": 0.3508916098282521, "grad_norm": 1.1874128580093384, "learning_rate": 0.00018913762716263818, "loss": 0.8429, "step": 8540 }, { "epoch": 0.35171337003862274, "grad_norm": 1.2707151174545288, "learning_rate": 0.00018907722966053555, "loss": 0.8257, "step": 8560 }, { "epoch": 0.35253513024899336, "grad_norm": 1.1079628467559814, "learning_rate": 0.00018901667440951586, "loss": 0.849, "step": 8580 }, { "epoch": 0.35335689045936397, "grad_norm": 1.1875925064086914, "learning_rate": 0.0001889559615168179, "loss": 0.8704, "step": 8600 }, { "epoch": 0.3541786506697346, "grad_norm": 1.1461087465286255, "learning_rate": 0.00018889509108995943, "loss": 0.8292, "step": 8620 }, { "epoch": 0.3550004108801052, "grad_norm": 1.3481261730194092, "learning_rate": 0.0001888340632367373, "loss": 0.8163, "step": 8640 }, { "epoch": 0.3558221710904758, "grad_norm": 1.1863452196121216, "learning_rate": 0.00018877287806522722, "loss": 0.8532, "step": 8660 }, { "epoch": 0.35664393130084643, "grad_norm": 1.3977798223495483, "learning_rate": 0.00018871153568378332, "loss": 0.8714, "step": 8680 }, { "epoch": 0.35746569151121704, "grad_norm": 1.1754332780838013, "learning_rate": 0.0001886500362010383, "loss": 0.8243, "step": 8700 }, { "epoch": 0.35828745172158766, "grad_norm": 1.1255104541778564, "learning_rate": 0.000188588379725903, "loss": 0.8025, "step": 8720 }, { "epoch": 0.35910921193195827, "grad_norm": 1.0885831117630005, "learning_rate": 0.00018852656636756627, "loss": 0.8179, "step": 8740 }, { "epoch": 0.3599309721423289, "grad_norm": 1.121172308921814, "learning_rate": 0.00018846459623549482, "loss": 0.831, "step": 8760 }, { "epoch": 0.3607527323526995, "grad_norm": 1.2006275653839111, "learning_rate": 0.000188402469439433, "loss": 0.8451, "step": 8780 }, { "epoch": 0.3615744925630701, "grad_norm": 1.0075160264968872, "learning_rate": 0.00018834018608940257, "loss": 0.8326, "step": 8800 }, { "epoch": 0.36239625277344073, "grad_norm": 1.3210777044296265, "learning_rate": 0.00018827774629570252, "loss": 0.8466, "step": 8820 }, { "epoch": 0.36321801298381134, "grad_norm": 1.157143473625183, "learning_rate": 0.00018821515016890895, "loss": 0.84, "step": 8840 }, { "epoch": 0.36403977319418196, "grad_norm": 1.0349316596984863, "learning_rate": 0.0001881523978198748, "loss": 0.8166, "step": 8860 }, { "epoch": 0.3648615334045526, "grad_norm": 1.1739977598190308, "learning_rate": 0.00018808948935972964, "loss": 0.8154, "step": 8880 }, { "epoch": 0.3656832936149232, "grad_norm": 1.0839564800262451, "learning_rate": 0.00018802642489987946, "loss": 0.8446, "step": 8900 }, { "epoch": 0.3665050538252938, "grad_norm": 1.126232624053955, "learning_rate": 0.0001879632045520066, "loss": 0.8262, "step": 8920 }, { "epoch": 0.3673268140356644, "grad_norm": 1.1430919170379639, "learning_rate": 0.00018789982842806947, "loss": 0.8449, "step": 8940 }, { "epoch": 0.36814857424603503, "grad_norm": 1.1284793615341187, "learning_rate": 0.00018783629664030226, "loss": 0.8482, "step": 8960 }, { "epoch": 0.36897033445640565, "grad_norm": 1.0018378496170044, "learning_rate": 0.00018777260930121487, "loss": 0.8353, "step": 8980 }, { "epoch": 0.36979209466677626, "grad_norm": 1.055388331413269, "learning_rate": 0.00018771515779286891, "loss": 0.8397, "step": 9000 }, { "epoch": 0.3706138548771469, "grad_norm": 1.152448296546936, "learning_rate": 0.00018765117521722443, "loss": 0.8607, "step": 9020 }, { "epoch": 0.3714356150875175, "grad_norm": 1.1177656650543213, "learning_rate": 0.00018758703741809558, "loss": 0.8254, "step": 9040 }, { "epoch": 0.37225737529788805, "grad_norm": 1.336777687072754, "learning_rate": 0.00018752274450906545, "loss": 0.8367, "step": 9060 }, { "epoch": 0.37307913550825866, "grad_norm": 1.19560706615448, "learning_rate": 0.00018745829660399185, "loss": 0.8528, "step": 9080 }, { "epoch": 0.3739008957186293, "grad_norm": 1.2169603109359741, "learning_rate": 0.00018739369381700707, "loss": 0.8425, "step": 9100 }, { "epoch": 0.3747226559289999, "grad_norm": 1.1716234683990479, "learning_rate": 0.00018732893626251766, "loss": 0.8358, "step": 9120 }, { "epoch": 0.3755444161393705, "grad_norm": 1.0265463590621948, "learning_rate": 0.00018726402405520425, "loss": 0.8447, "step": 9140 }, { "epoch": 0.3763661763497411, "grad_norm": 1.2216025590896606, "learning_rate": 0.00018719895731002137, "loss": 0.8247, "step": 9160 }, { "epoch": 0.37718793656011174, "grad_norm": 1.0820845365524292, "learning_rate": 0.0001871337361421972, "loss": 0.8452, "step": 9180 }, { "epoch": 0.37800969677048235, "grad_norm": 1.019952416419983, "learning_rate": 0.00018706836066723347, "loss": 0.8493, "step": 9200 }, { "epoch": 0.37883145698085297, "grad_norm": 1.0316121578216553, "learning_rate": 0.00018700283100090502, "loss": 0.8447, "step": 9220 }, { "epoch": 0.3796532171912236, "grad_norm": 1.1700369119644165, "learning_rate": 0.00018693714725925994, "loss": 0.8337, "step": 9240 }, { "epoch": 0.3804749774015942, "grad_norm": 1.032667636871338, "learning_rate": 0.00018687130955861902, "loss": 0.8325, "step": 9260 }, { "epoch": 0.3812967376119648, "grad_norm": 1.2093219757080078, "learning_rate": 0.0001868053180155758, "loss": 0.8295, "step": 9280 }, { "epoch": 0.3821184978223354, "grad_norm": 1.1522185802459717, "learning_rate": 0.00018673917274699618, "loss": 0.841, "step": 9300 }, { "epoch": 0.38294025803270604, "grad_norm": 1.2028223276138306, "learning_rate": 0.00018667287387001834, "loss": 0.8432, "step": 9320 }, { "epoch": 0.38376201824307665, "grad_norm": 1.2288753986358643, "learning_rate": 0.00018660642150205255, "loss": 0.8586, "step": 9340 }, { "epoch": 0.38458377845344727, "grad_norm": 1.1289194822311401, "learning_rate": 0.00018653981576078075, "loss": 0.8408, "step": 9360 }, { "epoch": 0.3854055386638179, "grad_norm": 1.208264708518982, "learning_rate": 0.00018647305676415665, "loss": 0.8233, "step": 9380 }, { "epoch": 0.3862272988741885, "grad_norm": 1.23066246509552, "learning_rate": 0.00018640949387227146, "loss": 0.8227, "step": 9400 }, { "epoch": 0.3870490590845591, "grad_norm": 1.2344571352005005, "learning_rate": 0.0001863424363680021, "loss": 0.8376, "step": 9420 }, { "epoch": 0.3878708192949297, "grad_norm": 1.1864609718322754, "learning_rate": 0.00018627522595792413, "loss": 0.8391, "step": 9440 }, { "epoch": 0.38869257950530034, "grad_norm": 1.109244465827942, "learning_rate": 0.00018620786276106203, "loss": 0.8256, "step": 9460 }, { "epoch": 0.38951433971567095, "grad_norm": 1.0809723138809204, "learning_rate": 0.00018614034689671082, "loss": 0.8492, "step": 9480 }, { "epoch": 0.39033609992604157, "grad_norm": 1.2196381092071533, "learning_rate": 0.00018607267848443591, "loss": 0.8415, "step": 9500 }, { "epoch": 0.3911578601364122, "grad_norm": 1.1039822101593018, "learning_rate": 0.00018600485764407282, "loss": 0.8464, "step": 9520 }, { "epoch": 0.3919796203467828, "grad_norm": 1.115871548652649, "learning_rate": 0.00018593688449572703, "loss": 0.8368, "step": 9540 }, { "epoch": 0.3928013805571534, "grad_norm": 1.0675318241119385, "learning_rate": 0.0001858687591597738, "loss": 0.8276, "step": 9560 }, { "epoch": 0.393623140767524, "grad_norm": 1.1515909433364868, "learning_rate": 0.00018580048175685784, "loss": 0.8259, "step": 9580 }, { "epoch": 0.39444490097789464, "grad_norm": 1.1502107381820679, "learning_rate": 0.00018573205240789316, "loss": 0.8549, "step": 9600 }, { "epoch": 0.39526666118826526, "grad_norm": 1.0891849994659424, "learning_rate": 0.00018566347123406284, "loss": 0.8396, "step": 9620 }, { "epoch": 0.39608842139863587, "grad_norm": 1.1536388397216797, "learning_rate": 0.00018559473835681896, "loss": 0.8401, "step": 9640 }, { "epoch": 0.3969101816090065, "grad_norm": 1.320541262626648, "learning_rate": 0.00018552585389788203, "loss": 0.8454, "step": 9660 }, { "epoch": 0.3977319418193771, "grad_norm": 0.96424800157547, "learning_rate": 0.00018545681797924125, "loss": 0.8257, "step": 9680 }, { "epoch": 0.3985537020297477, "grad_norm": 1.1451895236968994, "learning_rate": 0.00018538763072315382, "loss": 0.8327, "step": 9700 }, { "epoch": 0.39937546224011833, "grad_norm": 1.1693811416625977, "learning_rate": 0.00018531829225214508, "loss": 0.8342, "step": 9720 }, { "epoch": 0.40019722245048894, "grad_norm": 1.0285801887512207, "learning_rate": 0.00018524880268900812, "loss": 0.8261, "step": 9740 }, { "epoch": 0.40101898266085956, "grad_norm": 1.1713870763778687, "learning_rate": 0.00018517916215680363, "loss": 0.8597, "step": 9760 }, { "epoch": 0.40184074287123017, "grad_norm": 1.117725133895874, "learning_rate": 0.00018510937077885958, "loss": 0.8281, "step": 9780 }, { "epoch": 0.4026625030816008, "grad_norm": 1.2378820180892944, "learning_rate": 0.00018503942867877118, "loss": 0.8619, "step": 9800 }, { "epoch": 0.4034842632919714, "grad_norm": 1.2209067344665527, "learning_rate": 0.00018496933598040048, "loss": 0.8549, "step": 9820 }, { "epoch": 0.404306023502342, "grad_norm": 1.0837441682815552, "learning_rate": 0.00018489909280787627, "loss": 0.8575, "step": 9840 }, { "epoch": 0.40512778371271263, "grad_norm": 1.2987329959869385, "learning_rate": 0.00018482869928559379, "loss": 0.8467, "step": 9860 }, { "epoch": 0.40594954392308324, "grad_norm": 1.216752052307129, "learning_rate": 0.00018475815553821456, "loss": 0.8333, "step": 9880 }, { "epoch": 0.40677130413345386, "grad_norm": 1.2210928201675415, "learning_rate": 0.0001846874616906661, "loss": 0.8377, "step": 9900 }, { "epoch": 0.4075930643438245, "grad_norm": 1.007938027381897, "learning_rate": 0.0001846166178681418, "loss": 0.8422, "step": 9920 }, { "epoch": 0.4084148245541951, "grad_norm": 1.1827200651168823, "learning_rate": 0.00018454562419610058, "loss": 0.8456, "step": 9940 }, { "epoch": 0.4092365847645657, "grad_norm": 1.2097376585006714, "learning_rate": 0.00018447448080026682, "loss": 0.8408, "step": 9960 }, { "epoch": 0.4100583449749363, "grad_norm": 1.0128288269042969, "learning_rate": 0.00018440318780662998, "loss": 0.8517, "step": 9980 }, { "epoch": 0.41088010518530693, "grad_norm": 1.1385818719863892, "learning_rate": 0.00018433174534144444, "loss": 0.8532, "step": 10000 }, { "epoch": 0.41088010518530693, "eval_loss": 1.1022228002548218, "eval_runtime": 16.4762, "eval_samples_per_second": 159.018, "eval_steps_per_second": 4.977, "step": 10000 }, { "epoch": 0.41170186539567755, "grad_norm": 1.2568473815917969, "learning_rate": 0.00018426015353122934, "loss": 0.8304, "step": 10020 }, { "epoch": 0.41252362560604816, "grad_norm": 1.2366386651992798, "learning_rate": 0.00018418841250276825, "loss": 0.8244, "step": 10040 }, { "epoch": 0.4133453858164188, "grad_norm": 1.0323954820632935, "learning_rate": 0.000184116522383109, "loss": 0.8167, "step": 10060 }, { "epoch": 0.4141671460267894, "grad_norm": 1.1676981449127197, "learning_rate": 0.00018404448329956344, "loss": 0.8356, "step": 10080 }, { "epoch": 0.41498890623716, "grad_norm": 1.2117750644683838, "learning_rate": 0.0001839722953797073, "loss": 0.8483, "step": 10100 }, { "epoch": 0.4158106664475306, "grad_norm": 1.1228010654449463, "learning_rate": 0.00018389995875137978, "loss": 0.818, "step": 10120 }, { "epoch": 0.41663242665790123, "grad_norm": 1.03129243850708, "learning_rate": 0.00018382747354268351, "loss": 0.8556, "step": 10140 }, { "epoch": 0.41745418686827185, "grad_norm": 1.239634394645691, "learning_rate": 0.00018375847508876958, "loss": 0.8581, "step": 10160 }, { "epoch": 0.41827594707864246, "grad_norm": 1.046134114265442, "learning_rate": 0.0001836857005178056, "loss": 0.8241, "step": 10180 }, { "epoch": 0.4190977072890131, "grad_norm": 1.3205509185791016, "learning_rate": 0.00018361277774590765, "loss": 0.8409, "step": 10200 }, { "epoch": 0.4199194674993837, "grad_norm": 1.1549128293991089, "learning_rate": 0.00018353970690221646, "loss": 0.8252, "step": 10220 }, { "epoch": 0.4207412277097543, "grad_norm": 1.1062310934066772, "learning_rate": 0.00018346648811613477, "loss": 0.8428, "step": 10240 }, { "epoch": 0.4215629879201249, "grad_norm": 1.2223172187805176, "learning_rate": 0.0001833931215173274, "loss": 0.8118, "step": 10260 }, { "epoch": 0.42238474813049554, "grad_norm": 1.2306350469589233, "learning_rate": 0.00018331960723572105, "loss": 0.8615, "step": 10280 }, { "epoch": 0.42320650834086615, "grad_norm": 1.301054835319519, "learning_rate": 0.0001832459454015038, "loss": 0.8388, "step": 10300 }, { "epoch": 0.42402826855123676, "grad_norm": 1.1266446113586426, "learning_rate": 0.00018317213614512507, "loss": 0.844, "step": 10320 }, { "epoch": 0.4248500287616074, "grad_norm": 1.1231412887573242, "learning_rate": 0.0001830981795972954, "loss": 0.8581, "step": 10340 }, { "epoch": 0.425671788971978, "grad_norm": 1.0722932815551758, "learning_rate": 0.00018302407588898612, "loss": 0.844, "step": 10360 }, { "epoch": 0.4264935491823486, "grad_norm": 1.122125506401062, "learning_rate": 0.0001829498251514292, "loss": 0.817, "step": 10380 }, { "epoch": 0.4273153093927192, "grad_norm": 1.1031594276428223, "learning_rate": 0.00018287542751611703, "loss": 0.8448, "step": 10400 }, { "epoch": 0.42813706960308984, "grad_norm": 1.166870355606079, "learning_rate": 0.00018280088311480201, "loss": 0.8607, "step": 10420 }, { "epoch": 0.42895882981346045, "grad_norm": 1.1672008037567139, "learning_rate": 0.0001827261920794966, "loss": 0.8206, "step": 10440 }, { "epoch": 0.42978059002383107, "grad_norm": 1.7684413194656372, "learning_rate": 0.00018265135454247284, "loss": 0.8503, "step": 10460 }, { "epoch": 0.4306023502342017, "grad_norm": 1.3215503692626953, "learning_rate": 0.00018257637063626226, "loss": 0.8535, "step": 10480 }, { "epoch": 0.4314241104445723, "grad_norm": 1.1748920679092407, "learning_rate": 0.0001825012404936556, "loss": 0.8364, "step": 10500 }, { "epoch": 0.4322458706549429, "grad_norm": 1.0914533138275146, "learning_rate": 0.00018242596424770252, "loss": 0.8144, "step": 10520 }, { "epoch": 0.4330676308653135, "grad_norm": 1.2216529846191406, "learning_rate": 0.0001823505420317115, "loss": 0.8566, "step": 10540 }, { "epoch": 0.43388939107568414, "grad_norm": 1.2634260654449463, "learning_rate": 0.00018227497397924948, "loss": 0.8372, "step": 10560 }, { "epoch": 0.43471115128605475, "grad_norm": 1.244780421257019, "learning_rate": 0.00018219926022414163, "loss": 0.8397, "step": 10580 }, { "epoch": 0.43553291149642537, "grad_norm": 1.177216649055481, "learning_rate": 0.00018212340090047118, "loss": 0.8122, "step": 10600 }, { "epoch": 0.436354671706796, "grad_norm": 1.1915186643600464, "learning_rate": 0.0001820473961425792, "loss": 0.829, "step": 10620 }, { "epoch": 0.4371764319171666, "grad_norm": 0.9915897250175476, "learning_rate": 0.00018197124608506423, "loss": 0.8401, "step": 10640 }, { "epoch": 0.4379981921275372, "grad_norm": 1.1870684623718262, "learning_rate": 0.00018189495086278214, "loss": 0.8481, "step": 10660 }, { "epoch": 0.4388199523379078, "grad_norm": 1.1735584735870361, "learning_rate": 0.00018181851061084596, "loss": 0.8539, "step": 10680 }, { "epoch": 0.4396417125482784, "grad_norm": 1.2218267917633057, "learning_rate": 0.00018174192546462542, "loss": 0.8416, "step": 10700 }, { "epoch": 0.440463472758649, "grad_norm": 1.2126885652542114, "learning_rate": 0.000181665195559747, "loss": 0.8353, "step": 10720 }, { "epoch": 0.4412852329690196, "grad_norm": 1.1790574789047241, "learning_rate": 0.00018158832103209335, "loss": 0.8235, "step": 10740 }, { "epoch": 0.44210699317939023, "grad_norm": 1.153855323791504, "learning_rate": 0.00018151130201780343, "loss": 0.8415, "step": 10760 }, { "epoch": 0.44292875338976084, "grad_norm": 1.0696359872817993, "learning_rate": 0.00018143413865327198, "loss": 0.8514, "step": 10780 }, { "epoch": 0.44375051360013146, "grad_norm": 1.2022265195846558, "learning_rate": 0.00018135683107514932, "loss": 0.8102, "step": 10800 }, { "epoch": 0.44457227381050207, "grad_norm": 1.1839139461517334, "learning_rate": 0.00018127937942034127, "loss": 0.8243, "step": 10820 }, { "epoch": 0.4453940340208727, "grad_norm": 1.1672074794769287, "learning_rate": 0.00018120178382600874, "loss": 0.8707, "step": 10840 }, { "epoch": 0.4462157942312433, "grad_norm": 1.1144109964370728, "learning_rate": 0.00018112404442956754, "loss": 0.833, "step": 10860 }, { "epoch": 0.4470375544416139, "grad_norm": 1.2560542821884155, "learning_rate": 0.00018104616136868816, "loss": 0.8164, "step": 10880 }, { "epoch": 0.44785931465198453, "grad_norm": 1.1620038747787476, "learning_rate": 0.00018096813478129552, "loss": 0.857, "step": 10900 }, { "epoch": 0.44868107486235514, "grad_norm": 1.0956732034683228, "learning_rate": 0.0001808899648055687, "loss": 0.8069, "step": 10920 }, { "epoch": 0.44950283507272576, "grad_norm": 1.153667688369751, "learning_rate": 0.00018081165157994068, "loss": 0.8228, "step": 10940 }, { "epoch": 0.4503245952830964, "grad_norm": 1.1663360595703125, "learning_rate": 0.00018073319524309822, "loss": 0.8438, "step": 10960 }, { "epoch": 0.451146355493467, "grad_norm": 1.0581029653549194, "learning_rate": 0.00018065459593398137, "loss": 0.8245, "step": 10980 }, { "epoch": 0.4519681157038376, "grad_norm": 1.3234822750091553, "learning_rate": 0.0001805758537917835, "loss": 0.8462, "step": 11000 }, { "epoch": 0.4527898759142082, "grad_norm": 1.1863617897033691, "learning_rate": 0.0001804969689559509, "loss": 0.815, "step": 11020 }, { "epoch": 0.45361163612457883, "grad_norm": 1.2528035640716553, "learning_rate": 0.00018041794156618252, "loss": 0.8239, "step": 11040 }, { "epoch": 0.45443339633494945, "grad_norm": 2.205885887145996, "learning_rate": 0.00018033877176242975, "loss": 0.8061, "step": 11060 }, { "epoch": 0.45525515654532006, "grad_norm": 1.2284212112426758, "learning_rate": 0.00018025945968489626, "loss": 0.837, "step": 11080 }, { "epoch": 0.4560769167556907, "grad_norm": 1.1616712808609009, "learning_rate": 0.00018018000547403765, "loss": 0.8502, "step": 11100 }, { "epoch": 0.4568986769660613, "grad_norm": 1.188586711883545, "learning_rate": 0.00018010040927056117, "loss": 0.842, "step": 11120 }, { "epoch": 0.4577204371764319, "grad_norm": 1.1483261585235596, "learning_rate": 0.00018002067121542558, "loss": 0.8503, "step": 11140 }, { "epoch": 0.4585421973868025, "grad_norm": 1.1956408023834229, "learning_rate": 0.00017994079144984087, "loss": 0.8452, "step": 11160 }, { "epoch": 0.45936395759717313, "grad_norm": 1.281135082244873, "learning_rate": 0.00017986077011526792, "loss": 0.8641, "step": 11180 }, { "epoch": 0.46018571780754375, "grad_norm": 1.3326410055160522, "learning_rate": 0.00017978060735341836, "loss": 0.81, "step": 11200 }, { "epoch": 0.46100747801791436, "grad_norm": 1.331778645515442, "learning_rate": 0.0001797003033062543, "loss": 0.8401, "step": 11220 }, { "epoch": 0.461829238228285, "grad_norm": 1.1928249597549438, "learning_rate": 0.000179619858115988, "loss": 0.8165, "step": 11240 }, { "epoch": 0.4626509984386556, "grad_norm": 1.3430322408676147, "learning_rate": 0.0001795392719250817, "loss": 0.8581, "step": 11260 }, { "epoch": 0.4634727586490262, "grad_norm": 1.142386555671692, "learning_rate": 0.00017945854487624733, "loss": 0.8397, "step": 11280 }, { "epoch": 0.4642945188593968, "grad_norm": 1.1435790061950684, "learning_rate": 0.0001793776771124463, "loss": 0.8781, "step": 11300 }, { "epoch": 0.46511627906976744, "grad_norm": 1.1941109895706177, "learning_rate": 0.00017929666877688919, "loss": 0.8508, "step": 11320 }, { "epoch": 0.46593803928013805, "grad_norm": 1.1792101860046387, "learning_rate": 0.00017921552001303552, "loss": 0.8559, "step": 11340 }, { "epoch": 0.46675979949050866, "grad_norm": 1.2072590589523315, "learning_rate": 0.00017913423096459354, "loss": 0.8416, "step": 11360 }, { "epoch": 0.4675815597008793, "grad_norm": 1.220444679260254, "learning_rate": 0.00017905280177551983, "loss": 0.8261, "step": 11380 }, { "epoch": 0.4684033199112499, "grad_norm": 1.1701765060424805, "learning_rate": 0.00017897123259001926, "loss": 0.8509, "step": 11400 }, { "epoch": 0.4692250801216205, "grad_norm": 1.2121479511260986, "learning_rate": 0.00017888952355254455, "loss": 0.8424, "step": 11420 }, { "epoch": 0.4700468403319911, "grad_norm": 1.1610180139541626, "learning_rate": 0.0001788076748077962, "loss": 0.8433, "step": 11440 }, { "epoch": 0.47086860054236174, "grad_norm": 1.0766757726669312, "learning_rate": 0.0001787256865007219, "loss": 0.8175, "step": 11460 }, { "epoch": 0.47169036075273235, "grad_norm": 1.2801762819290161, "learning_rate": 0.00017864355877651676, "loss": 0.8416, "step": 11480 }, { "epoch": 0.47251212096310297, "grad_norm": 1.167050838470459, "learning_rate": 0.00017856129178062257, "loss": 0.842, "step": 11500 }, { "epoch": 0.4733338811734736, "grad_norm": 1.0567585229873657, "learning_rate": 0.0001784788856587279, "loss": 0.8467, "step": 11520 }, { "epoch": 0.4741556413838442, "grad_norm": 1.101453185081482, "learning_rate": 0.00017839634055676762, "loss": 0.836, "step": 11540 }, { "epoch": 0.4749774015942148, "grad_norm": 1.2371095418930054, "learning_rate": 0.00017831365662092274, "loss": 0.8623, "step": 11560 }, { "epoch": 0.4757991618045854, "grad_norm": 1.2750577926635742, "learning_rate": 0.00017823083399762018, "loss": 0.8413, "step": 11580 }, { "epoch": 0.47662092201495604, "grad_norm": 1.2494827508926392, "learning_rate": 0.00017814787283353245, "loss": 0.8393, "step": 11600 }, { "epoch": 0.47744268222532665, "grad_norm": 1.1090929508209229, "learning_rate": 0.0001780647732755773, "loss": 0.8224, "step": 11620 }, { "epoch": 0.47826444243569727, "grad_norm": 1.2686065435409546, "learning_rate": 0.00017798153547091773, "loss": 0.8468, "step": 11640 }, { "epoch": 0.4790862026460679, "grad_norm": 1.0384360551834106, "learning_rate": 0.0001778981595669615, "loss": 0.8346, "step": 11660 }, { "epoch": 0.4799079628564385, "grad_norm": 1.1490122079849243, "learning_rate": 0.0001778146457113608, "loss": 0.8565, "step": 11680 }, { "epoch": 0.4807297230668091, "grad_norm": 1.1408191919326782, "learning_rate": 0.00017773099405201236, "loss": 0.8337, "step": 11700 }, { "epoch": 0.4815514832771797, "grad_norm": 1.1228752136230469, "learning_rate": 0.00017764720473705675, "loss": 0.8236, "step": 11720 }, { "epoch": 0.48237324348755034, "grad_norm": 1.0747302770614624, "learning_rate": 0.00017756327791487847, "loss": 0.8439, "step": 11740 }, { "epoch": 0.48319500369792096, "grad_norm": 1.1848806142807007, "learning_rate": 0.0001774792137341054, "loss": 0.8433, "step": 11760 }, { "epoch": 0.48401676390829157, "grad_norm": 1.1458165645599365, "learning_rate": 0.00017739501234360875, "loss": 0.8513, "step": 11780 }, { "epoch": 0.4848385241186622, "grad_norm": 1.2221580743789673, "learning_rate": 0.00017731067389250272, "loss": 0.8345, "step": 11800 }, { "epoch": 0.4856602843290328, "grad_norm": 1.2535064220428467, "learning_rate": 0.00017722619853014423, "loss": 0.842, "step": 11820 }, { "epoch": 0.4864820445394034, "grad_norm": 1.2956807613372803, "learning_rate": 0.0001771415864061326, "loss": 0.833, "step": 11840 }, { "epoch": 0.487303804749774, "grad_norm": 1.3538552522659302, "learning_rate": 0.0001770568376703094, "loss": 0.8389, "step": 11860 }, { "epoch": 0.48812556496014464, "grad_norm": 1.325257658958435, "learning_rate": 0.00017697195247275813, "loss": 0.8041, "step": 11880 }, { "epoch": 0.48894732517051526, "grad_norm": 1.1765714883804321, "learning_rate": 0.00017688693096380392, "loss": 0.8406, "step": 11900 }, { "epoch": 0.48976908538088587, "grad_norm": 1.1218301057815552, "learning_rate": 0.00017680177329401333, "loss": 0.8562, "step": 11920 }, { "epoch": 0.4905908455912565, "grad_norm": 1.1272341012954712, "learning_rate": 0.00017671647961419406, "loss": 0.8638, "step": 11940 }, { "epoch": 0.4914126058016271, "grad_norm": 1.0992316007614136, "learning_rate": 0.00017663105007539463, "loss": 0.8156, "step": 11960 }, { "epoch": 0.4922343660119977, "grad_norm": 1.1406649351119995, "learning_rate": 0.00017654548482890414, "loss": 0.8444, "step": 11980 }, { "epoch": 0.49305612622236833, "grad_norm": 1.251295566558838, "learning_rate": 0.00017645978402625214, "loss": 0.8549, "step": 12000 }, { "epoch": 0.49305612622236833, "eval_loss": 1.1104093790054321, "eval_runtime": 16.7503, "eval_samples_per_second": 156.415, "eval_steps_per_second": 4.895, "step": 12000 }, { "epoch": 0.49387788643273894, "grad_norm": 1.153436303138733, "learning_rate": 0.00017637394781920812, "loss": 0.8262, "step": 12020 }, { "epoch": 0.49469964664310956, "grad_norm": 1.200875997543335, "learning_rate": 0.00017628797635978134, "loss": 0.8644, "step": 12040 }, { "epoch": 0.4955214068534802, "grad_norm": 1.2446619272232056, "learning_rate": 0.00017620186980022072, "loss": 0.8485, "step": 12060 }, { "epoch": 0.4963431670638508, "grad_norm": 1.1602336168289185, "learning_rate": 0.00017611562829301429, "loss": 0.8318, "step": 12080 }, { "epoch": 0.4971649272742214, "grad_norm": 1.4564729928970337, "learning_rate": 0.00017602925199088917, "loss": 0.8375, "step": 12100 }, { "epoch": 0.497986687484592, "grad_norm": 1.107946753501892, "learning_rate": 0.00017594274104681108, "loss": 0.8292, "step": 12120 }, { "epoch": 0.49880844769496263, "grad_norm": 1.250048279762268, "learning_rate": 0.00017585609561398426, "loss": 0.817, "step": 12140 }, { "epoch": 0.49963020790533325, "grad_norm": 0.956064760684967, "learning_rate": 0.00017576931584585117, "loss": 0.8311, "step": 12160 }, { "epoch": 0.5004519681157038, "grad_norm": 1.1825581789016724, "learning_rate": 0.000175682401896092, "loss": 0.8232, "step": 12180 }, { "epoch": 0.5012737283260744, "grad_norm": 1.26679265499115, "learning_rate": 0.00017559535391862476, "loss": 0.7985, "step": 12200 }, { "epoch": 0.502095488536445, "grad_norm": 1.1247515678405762, "learning_rate": 0.00017550817206760463, "loss": 0.8177, "step": 12220 }, { "epoch": 0.5029172487468156, "grad_norm": 1.2112337350845337, "learning_rate": 0.00017542085649742403, "loss": 0.851, "step": 12240 }, { "epoch": 0.5037390089571863, "grad_norm": 1.3968725204467773, "learning_rate": 0.00017533340736271207, "loss": 0.8422, "step": 12260 }, { "epoch": 0.5045607691675569, "grad_norm": 1.2992043495178223, "learning_rate": 0.00017524582481833444, "loss": 0.8492, "step": 12280 }, { "epoch": 0.5053825293779275, "grad_norm": 1.1721656322479248, "learning_rate": 0.0001751581090193931, "loss": 0.8321, "step": 12300 }, { "epoch": 0.5062042895882981, "grad_norm": 1.3727058172225952, "learning_rate": 0.00017507026012122595, "loss": 0.8666, "step": 12320 }, { "epoch": 0.5070260497986687, "grad_norm": 1.27950119972229, "learning_rate": 0.0001749822782794067, "loss": 0.8643, "step": 12340 }, { "epoch": 0.5078478100090393, "grad_norm": 0.9998101592063904, "learning_rate": 0.00017489416364974432, "loss": 0.8319, "step": 12360 }, { "epoch": 0.50866957021941, "grad_norm": 1.210250973701477, "learning_rate": 0.0001748059163882831, "loss": 0.8183, "step": 12380 }, { "epoch": 0.5094913304297806, "grad_norm": 1.2826182842254639, "learning_rate": 0.00017471753665130213, "loss": 0.8421, "step": 12400 }, { "epoch": 0.5103130906401512, "grad_norm": 1.2036994695663452, "learning_rate": 0.00017462902459531508, "loss": 0.8363, "step": 12420 }, { "epoch": 0.5111348508505218, "grad_norm": 1.0195825099945068, "learning_rate": 0.00017454038037707008, "loss": 0.8338, "step": 12440 }, { "epoch": 0.5119566110608924, "grad_norm": 1.2347939014434814, "learning_rate": 0.00017445160415354916, "loss": 0.839, "step": 12460 }, { "epoch": 0.512778371271263, "grad_norm": 1.196473240852356, "learning_rate": 0.00017436269608196817, "loss": 0.85, "step": 12480 }, { "epoch": 0.5136001314816336, "grad_norm": 1.4037846326828003, "learning_rate": 0.00017427365631977648, "loss": 0.84, "step": 12500 }, { "epoch": 0.5144218916920043, "grad_norm": 1.2222518920898438, "learning_rate": 0.00017418448502465667, "loss": 0.855, "step": 12520 }, { "epoch": 0.5152436519023749, "grad_norm": 1.2747788429260254, "learning_rate": 0.0001740951823545242, "loss": 0.8545, "step": 12540 }, { "epoch": 0.5160654121127455, "grad_norm": 1.2375946044921875, "learning_rate": 0.00017400574846752724, "loss": 0.8351, "step": 12560 }, { "epoch": 0.5168871723231161, "grad_norm": 1.084808588027954, "learning_rate": 0.00017391618352204633, "loss": 0.8344, "step": 12580 }, { "epoch": 0.5177089325334867, "grad_norm": 1.228043794631958, "learning_rate": 0.00017382648767669408, "loss": 0.8321, "step": 12600 }, { "epoch": 0.5185306927438573, "grad_norm": 1.125532865524292, "learning_rate": 0.00017373666109031497, "loss": 0.8383, "step": 12620 }, { "epoch": 0.5193524529542279, "grad_norm": 1.1619880199432373, "learning_rate": 0.00017364670392198492, "loss": 0.8318, "step": 12640 }, { "epoch": 0.5201742131645986, "grad_norm": 1.1036595106124878, "learning_rate": 0.00017355661633101116, "loss": 0.8265, "step": 12660 }, { "epoch": 0.5209959733749692, "grad_norm": 1.084410548210144, "learning_rate": 0.0001734663984769319, "loss": 0.8293, "step": 12680 }, { "epoch": 0.5218177335853398, "grad_norm": 1.3622374534606934, "learning_rate": 0.000173376050519516, "loss": 0.8496, "step": 12700 }, { "epoch": 0.5226394937957104, "grad_norm": 1.2676513195037842, "learning_rate": 0.00017328557261876273, "loss": 0.8357, "step": 12720 }, { "epoch": 0.523461254006081, "grad_norm": 1.1944401264190674, "learning_rate": 0.00017319496493490148, "loss": 0.8162, "step": 12740 }, { "epoch": 0.5242830142164516, "grad_norm": 1.2380664348602295, "learning_rate": 0.00017310422762839155, "loss": 0.8319, "step": 12760 }, { "epoch": 0.5251047744268222, "grad_norm": 1.3865251541137695, "learning_rate": 0.00017301336085992163, "loss": 0.8181, "step": 12780 }, { "epoch": 0.5259265346371929, "grad_norm": 1.1436952352523804, "learning_rate": 0.00017292236479040984, "loss": 0.827, "step": 12800 }, { "epoch": 0.5267482948475635, "grad_norm": 1.4181216955184937, "learning_rate": 0.0001728312395810032, "loss": 0.8297, "step": 12820 }, { "epoch": 0.5275700550579341, "grad_norm": 1.4452232122421265, "learning_rate": 0.00017273998539307742, "loss": 0.8397, "step": 12840 }, { "epoch": 0.5283918152683047, "grad_norm": 1.1824501752853394, "learning_rate": 0.00017264860238823667, "loss": 0.8088, "step": 12860 }, { "epoch": 0.5292135754786753, "grad_norm": 1.1516762971878052, "learning_rate": 0.0001725570907283132, "loss": 0.8235, "step": 12880 }, { "epoch": 0.5300353356890459, "grad_norm": 1.0846797227859497, "learning_rate": 0.00017246545057536712, "loss": 0.8149, "step": 12900 }, { "epoch": 0.5308570958994165, "grad_norm": 1.244383692741394, "learning_rate": 0.00017237368209168608, "loss": 0.8488, "step": 12920 }, { "epoch": 0.5316788561097872, "grad_norm": 1.0443450212478638, "learning_rate": 0.000172281785439785, "loss": 0.8089, "step": 12940 }, { "epoch": 0.5325006163201578, "grad_norm": 1.282185673713684, "learning_rate": 0.00017218976078240582, "loss": 0.8164, "step": 12960 }, { "epoch": 0.5333223765305284, "grad_norm": 1.264277696609497, "learning_rate": 0.0001720976082825171, "loss": 0.8512, "step": 12980 }, { "epoch": 0.534144136740899, "grad_norm": 1.1357461214065552, "learning_rate": 0.00017200532810331378, "loss": 0.8368, "step": 13000 }, { "epoch": 0.5349658969512696, "grad_norm": 1.066361427307129, "learning_rate": 0.00017191292040821696, "loss": 0.8273, "step": 13020 }, { "epoch": 0.5357876571616402, "grad_norm": 1.2606313228607178, "learning_rate": 0.00017182038536087363, "loss": 0.8513, "step": 13040 }, { "epoch": 0.5366094173720108, "grad_norm": 1.1224803924560547, "learning_rate": 0.00017172772312515618, "loss": 0.8417, "step": 13060 }, { "epoch": 0.5374311775823815, "grad_norm": 1.2137328386306763, "learning_rate": 0.00017163957634245275, "loss": 0.8264, "step": 13080 }, { "epoch": 0.5382529377927521, "grad_norm": 1.1143057346343994, "learning_rate": 0.00017154666656159665, "loss": 0.8317, "step": 13100 }, { "epoch": 0.5390746980031227, "grad_norm": 1.087512493133545, "learning_rate": 0.00017145363007710135, "loss": 0.8194, "step": 13120 }, { "epoch": 0.5398964582134933, "grad_norm": 1.2861449718475342, "learning_rate": 0.0001713604670537273, "loss": 0.843, "step": 13140 }, { "epoch": 0.5407182184238639, "grad_norm": 1.2231301069259644, "learning_rate": 0.00017126717765645908, "loss": 0.8192, "step": 13160 }, { "epoch": 0.5415399786342345, "grad_norm": 1.2475714683532715, "learning_rate": 0.00017117376205050502, "loss": 0.8524, "step": 13180 }, { "epoch": 0.5423617388446051, "grad_norm": 1.1694715023040771, "learning_rate": 0.00017108022040129695, "loss": 0.8381, "step": 13200 }, { "epoch": 0.5431834990549758, "grad_norm": 1.29911470413208, "learning_rate": 0.00017098655287448993, "loss": 0.8056, "step": 13220 }, { "epoch": 0.5440052592653464, "grad_norm": 1.063346028327942, "learning_rate": 0.00017089275963596195, "loss": 0.8328, "step": 13240 }, { "epoch": 0.544827019475717, "grad_norm": 1.6195141077041626, "learning_rate": 0.0001707988408518136, "loss": 0.8597, "step": 13260 }, { "epoch": 0.5456487796860876, "grad_norm": 1.2005921602249146, "learning_rate": 0.00017070479668836785, "loss": 0.8415, "step": 13280 }, { "epoch": 0.5464705398964582, "grad_norm": 1.1937131881713867, "learning_rate": 0.00017061533875220887, "loss": 0.8171, "step": 13300 }, { "epoch": 0.5472923001068288, "grad_norm": 1.1822235584259033, "learning_rate": 0.0001705210505783601, "loss": 0.8596, "step": 13320 }, { "epoch": 0.5481140603171994, "grad_norm": 1.0105253458023071, "learning_rate": 0.00017042663751715912, "loss": 0.8266, "step": 13340 }, { "epoch": 0.5489358205275701, "grad_norm": 1.20473051071167, "learning_rate": 0.00017033209973580418, "loss": 0.8437, "step": 13360 }, { "epoch": 0.5497575807379407, "grad_norm": 1.237752914428711, "learning_rate": 0.00017023743740171438, "loss": 0.8278, "step": 13380 }, { "epoch": 0.5505793409483113, "grad_norm": 1.2165151834487915, "learning_rate": 0.00017014265068252948, "loss": 0.8494, "step": 13400 }, { "epoch": 0.5514011011586819, "grad_norm": 1.2971493005752563, "learning_rate": 0.00017004773974610941, "loss": 0.8312, "step": 13420 }, { "epoch": 0.5522228613690525, "grad_norm": 1.091404676437378, "learning_rate": 0.0001699527047605342, "loss": 0.8247, "step": 13440 }, { "epoch": 0.5530446215794231, "grad_norm": 1.1684538125991821, "learning_rate": 0.00016985754589410342, "loss": 0.8149, "step": 13460 }, { "epoch": 0.5538663817897937, "grad_norm": 1.201493740081787, "learning_rate": 0.00016976226331533617, "loss": 0.8288, "step": 13480 }, { "epoch": 0.5546881420001644, "grad_norm": 1.2418882846832275, "learning_rate": 0.0001696668571929705, "loss": 0.8286, "step": 13500 }, { "epoch": 0.555509902210535, "grad_norm": 1.073002576828003, "learning_rate": 0.00016957132769596336, "loss": 0.8178, "step": 13520 }, { "epoch": 0.5563316624209056, "grad_norm": 1.156518816947937, "learning_rate": 0.0001694756749934901, "loss": 0.8322, "step": 13540 }, { "epoch": 0.5571534226312762, "grad_norm": 1.2362408638000488, "learning_rate": 0.00016937989925494432, "loss": 0.8338, "step": 13560 }, { "epoch": 0.5579751828416468, "grad_norm": 1.1928801536560059, "learning_rate": 0.00016928400064993745, "loss": 0.8243, "step": 13580 }, { "epoch": 0.5587969430520174, "grad_norm": 1.1574454307556152, "learning_rate": 0.0001691879793482986, "loss": 0.8124, "step": 13600 }, { "epoch": 0.559618703262388, "grad_norm": 1.0951565504074097, "learning_rate": 0.00016909183552007398, "loss": 0.8575, "step": 13620 }, { "epoch": 0.5604404634727587, "grad_norm": 1.2538108825683594, "learning_rate": 0.00016899556933552704, "loss": 0.8149, "step": 13640 }, { "epoch": 0.5612622236831293, "grad_norm": 1.2423375844955444, "learning_rate": 0.0001688991809651377, "loss": 0.8114, "step": 13660 }, { "epoch": 0.5620839838934999, "grad_norm": 1.0723458528518677, "learning_rate": 0.00016880267057960239, "loss": 0.8384, "step": 13680 }, { "epoch": 0.5629057441038705, "grad_norm": 1.0973854064941406, "learning_rate": 0.00016870603834983356, "loss": 0.8373, "step": 13700 }, { "epoch": 0.5637275043142411, "grad_norm": 1.1496849060058594, "learning_rate": 0.00016860928444695943, "loss": 0.831, "step": 13720 }, { "epoch": 0.5645492645246117, "grad_norm": 1.175132393836975, "learning_rate": 0.00016851240904232375, "loss": 0.8486, "step": 13740 }, { "epoch": 0.5653710247349824, "grad_norm": 1.018654227256775, "learning_rate": 0.0001684154123074854, "loss": 0.8332, "step": 13760 }, { "epoch": 0.566192784945353, "grad_norm": 1.3159968852996826, "learning_rate": 0.00016831829441421809, "loss": 0.8336, "step": 13780 }, { "epoch": 0.5670145451557236, "grad_norm": 1.1875556707382202, "learning_rate": 0.0001682210555345102, "loss": 0.8209, "step": 13800 }, { "epoch": 0.5678363053660942, "grad_norm": 1.2860361337661743, "learning_rate": 0.00016812369584056424, "loss": 0.8453, "step": 13820 }, { "epoch": 0.5686580655764648, "grad_norm": 1.2040901184082031, "learning_rate": 0.00016802621550479675, "loss": 0.8204, "step": 13840 }, { "epoch": 0.5694798257868354, "grad_norm": 1.1987425088882446, "learning_rate": 0.00016792861469983793, "loss": 0.8347, "step": 13860 }, { "epoch": 0.570301585997206, "grad_norm": 1.1014000177383423, "learning_rate": 0.00016783089359853127, "loss": 0.8142, "step": 13880 }, { "epoch": 0.5711233462075767, "grad_norm": 1.1898833513259888, "learning_rate": 0.00016773305237393328, "loss": 0.8198, "step": 13900 }, { "epoch": 0.5719451064179473, "grad_norm": 1.2249850034713745, "learning_rate": 0.00016763509119931334, "loss": 0.8039, "step": 13920 }, { "epoch": 0.5727668666283179, "grad_norm": 1.2241109609603882, "learning_rate": 0.00016753701024815304, "loss": 0.8458, "step": 13940 }, { "epoch": 0.5735886268386885, "grad_norm": 1.2025922536849976, "learning_rate": 0.00016743880969414624, "loss": 0.8103, "step": 13960 }, { "epoch": 0.5744103870490591, "grad_norm": 1.1043455600738525, "learning_rate": 0.0001673404897111986, "loss": 0.834, "step": 13980 }, { "epoch": 0.5752321472594297, "grad_norm": 1.2452826499938965, "learning_rate": 0.00016724205047342715, "loss": 0.8216, "step": 14000 }, { "epoch": 0.5752321472594297, "eval_loss": 1.1319005489349365, "eval_runtime": 16.4046, "eval_samples_per_second": 159.712, "eval_steps_per_second": 4.999, "step": 14000 }, { "epoch": 0.5760539074698003, "grad_norm": 1.248207688331604, "learning_rate": 0.00016714349215516032, "loss": 0.824, "step": 14020 }, { "epoch": 0.576875667680171, "grad_norm": 1.3563847541809082, "learning_rate": 0.0001670448149309372, "loss": 0.8587, "step": 14040 }, { "epoch": 0.5776974278905416, "grad_norm": 1.2468847036361694, "learning_rate": 0.00016694601897550762, "loss": 0.8481, "step": 14060 }, { "epoch": 0.5785191881009122, "grad_norm": 1.139793872833252, "learning_rate": 0.0001668471044638316, "loss": 0.8353, "step": 14080 }, { "epoch": 0.5793409483112828, "grad_norm": 1.0366131067276, "learning_rate": 0.0001667480715710791, "loss": 0.8091, "step": 14100 }, { "epoch": 0.5801627085216534, "grad_norm": 1.1087323427200317, "learning_rate": 0.00016664892047262977, "loss": 0.8198, "step": 14120 }, { "epoch": 0.580984468732024, "grad_norm": 1.137624740600586, "learning_rate": 0.0001665496513440726, "loss": 0.832, "step": 14140 }, { "epoch": 0.5818062289423946, "grad_norm": 1.1337158679962158, "learning_rate": 0.00016645026436120551, "loss": 0.8438, "step": 14160 }, { "epoch": 0.5826279891527653, "grad_norm": 1.1277921199798584, "learning_rate": 0.00016635075970003525, "loss": 0.8158, "step": 14180 }, { "epoch": 0.5834497493631359, "grad_norm": 1.1939393281936646, "learning_rate": 0.00016625113753677693, "loss": 0.835, "step": 14200 }, { "epoch": 0.5842715095735065, "grad_norm": 1.343687891960144, "learning_rate": 0.00016615139804785367, "loss": 0.8231, "step": 14220 }, { "epoch": 0.5850932697838771, "grad_norm": 1.1290326118469238, "learning_rate": 0.00016605154140989647, "loss": 0.8456, "step": 14240 }, { "epoch": 0.5859150299942477, "grad_norm": 1.274527668952942, "learning_rate": 0.00016595156779974376, "loss": 0.85, "step": 14260 }, { "epoch": 0.5867367902046183, "grad_norm": 1.2629293203353882, "learning_rate": 0.00016585147739444104, "loss": 0.8245, "step": 14280 }, { "epoch": 0.587558550414989, "grad_norm": 1.1809213161468506, "learning_rate": 0.0001657512703712408, "loss": 0.8063, "step": 14300 }, { "epoch": 0.5883803106253596, "grad_norm": 1.0857917070388794, "learning_rate": 0.00016565094690760193, "loss": 0.8158, "step": 14320 }, { "epoch": 0.5892020708357302, "grad_norm": 1.1015921831130981, "learning_rate": 0.00016555050718118953, "loss": 0.8501, "step": 14340 }, { "epoch": 0.5900238310461008, "grad_norm": 1.3138587474822998, "learning_rate": 0.00016544995136987467, "loss": 0.8416, "step": 14360 }, { "epoch": 0.5908455912564714, "grad_norm": 1.476505994796753, "learning_rate": 0.00016534927965173392, "loss": 0.8526, "step": 14380 }, { "epoch": 0.591667351466842, "grad_norm": 1.1754002571105957, "learning_rate": 0.00016524849220504918, "loss": 0.8136, "step": 14400 }, { "epoch": 0.5924891116772126, "grad_norm": 1.2151134014129639, "learning_rate": 0.00016514758920830724, "loss": 0.821, "step": 14420 }, { "epoch": 0.5933108718875832, "grad_norm": 1.2440420389175415, "learning_rate": 0.0001650465708401995, "loss": 0.8274, "step": 14440 }, { "epoch": 0.5941326320979539, "grad_norm": 1.1762789487838745, "learning_rate": 0.0001649454372796218, "loss": 0.8379, "step": 14460 }, { "epoch": 0.5949543923083245, "grad_norm": 1.1778429746627808, "learning_rate": 0.0001648441887056738, "loss": 0.7901, "step": 14480 }, { "epoch": 0.5957761525186951, "grad_norm": 1.1886394023895264, "learning_rate": 0.0001647478961925985, "loss": 0.8431, "step": 14500 }, { "epoch": 0.5965979127290657, "grad_norm": 1.256296992301941, "learning_rate": 0.0001646515001956767, "loss": 0.8181, "step": 14520 }, { "epoch": 0.5974196729394363, "grad_norm": 1.2841947078704834, "learning_rate": 0.00016454991909764397, "loss": 0.8255, "step": 14540 }, { "epoch": 0.5982414331498069, "grad_norm": 1.1896756887435913, "learning_rate": 0.0001644482236866732, "loss": 0.8304, "step": 14560 }, { "epoch": 0.5990631933601775, "grad_norm": 1.4413669109344482, "learning_rate": 0.00016434641414285922, "loss": 0.8374, "step": 14580 }, { "epoch": 0.5998849535705482, "grad_norm": 1.2350515127182007, "learning_rate": 0.00016424449064649882, "loss": 0.8342, "step": 14600 }, { "epoch": 0.6007067137809188, "grad_norm": 1.3465436697006226, "learning_rate": 0.0001641424533780907, "loss": 0.8066, "step": 14620 }, { "epoch": 0.6015284739912894, "grad_norm": 1.351413369178772, "learning_rate": 0.00016404030251833502, "loss": 0.7952, "step": 14640 }, { "epoch": 0.60235023420166, "grad_norm": 1.1581010818481445, "learning_rate": 0.00016393803824813304, "loss": 0.8005, "step": 14660 }, { "epoch": 0.6031719944120306, "grad_norm": 1.218668818473816, "learning_rate": 0.00016383566074858695, "loss": 0.8293, "step": 14680 }, { "epoch": 0.6039937546224012, "grad_norm": 1.1792479753494263, "learning_rate": 0.00016373317020099943, "loss": 0.8247, "step": 14700 }, { "epoch": 0.6048155148327718, "grad_norm": 1.2358107566833496, "learning_rate": 0.00016363056678687335, "loss": 0.8418, "step": 14720 }, { "epoch": 0.6056372750431425, "grad_norm": 1.2904804944992065, "learning_rate": 0.00016352785068791142, "loss": 0.8317, "step": 14740 }, { "epoch": 0.6064590352535131, "grad_norm": 1.1909780502319336, "learning_rate": 0.000163425022086016, "loss": 0.8221, "step": 14760 }, { "epoch": 0.6072807954638837, "grad_norm": 1.199325442314148, "learning_rate": 0.00016332208116328853, "loss": 0.819, "step": 14780 }, { "epoch": 0.6081025556742543, "grad_norm": 1.4012378454208374, "learning_rate": 0.00016321902810202956, "loss": 0.8104, "step": 14800 }, { "epoch": 0.6089243158846248, "grad_norm": 1.1930865049362183, "learning_rate": 0.00016311586308473812, "loss": 0.8227, "step": 14820 }, { "epoch": 0.6097460760949954, "grad_norm": 1.055925965309143, "learning_rate": 0.00016301258629411144, "loss": 0.8337, "step": 14840 }, { "epoch": 0.610567836305366, "grad_norm": 1.1309912204742432, "learning_rate": 0.00016290919791304487, "loss": 0.8154, "step": 14860 }, { "epoch": 0.6113895965157367, "grad_norm": 1.272495985031128, "learning_rate": 0.0001628056981246312, "loss": 0.8277, "step": 14880 }, { "epoch": 0.6122113567261073, "grad_norm": 1.4250491857528687, "learning_rate": 0.00016270208711216059, "loss": 0.8606, "step": 14900 }, { "epoch": 0.6130331169364779, "grad_norm": 1.079211950302124, "learning_rate": 0.00016259836505912024, "loss": 0.8485, "step": 14920 }, { "epoch": 0.6138548771468485, "grad_norm": 1.2047349214553833, "learning_rate": 0.00016249453214919383, "loss": 0.8123, "step": 14940 }, { "epoch": 0.6146766373572191, "grad_norm": 1.344088077545166, "learning_rate": 0.0001623905885662615, "loss": 0.8155, "step": 14960 }, { "epoch": 0.6154983975675897, "grad_norm": 1.3220460414886475, "learning_rate": 0.0001622865344943993, "loss": 0.8544, "step": 14980 }, { "epoch": 0.6163201577779603, "grad_norm": 1.2667044401168823, "learning_rate": 0.00016218237011787905, "loss": 0.8049, "step": 15000 }, { "epoch": 0.617141917988331, "grad_norm": 1.2205132246017456, "learning_rate": 0.0001620780956211678, "loss": 0.8248, "step": 15020 }, { "epoch": 0.6179636781987016, "grad_norm": 1.2170429229736328, "learning_rate": 0.00016197371118892762, "loss": 0.8176, "step": 15040 }, { "epoch": 0.6187854384090722, "grad_norm": 1.1776739358901978, "learning_rate": 0.00016186921700601535, "loss": 0.8358, "step": 15060 }, { "epoch": 0.6196071986194428, "grad_norm": 1.305285096168518, "learning_rate": 0.00016176461325748215, "loss": 0.8279, "step": 15080 }, { "epoch": 0.6204289588298134, "grad_norm": 1.4242192506790161, "learning_rate": 0.00016165990012857316, "loss": 0.8331, "step": 15100 }, { "epoch": 0.621250719040184, "grad_norm": 1.2353391647338867, "learning_rate": 0.00016155507780472734, "loss": 0.833, "step": 15120 }, { "epoch": 0.6220724792505546, "grad_norm": 1.0569748878479004, "learning_rate": 0.0001614501464715769, "loss": 0.8491, "step": 15140 }, { "epoch": 0.6228942394609253, "grad_norm": 1.0396721363067627, "learning_rate": 0.00016134510631494718, "loss": 0.7974, "step": 15160 }, { "epoch": 0.6237159996712959, "grad_norm": 1.1455570459365845, "learning_rate": 0.00016123995752085623, "loss": 0.8384, "step": 15180 }, { "epoch": 0.6245377598816665, "grad_norm": 1.1126292943954468, "learning_rate": 0.00016113470027551442, "loss": 0.8595, "step": 15200 }, { "epoch": 0.6253595200920371, "grad_norm": 1.1759933233261108, "learning_rate": 0.00016102933476532427, "loss": 0.8161, "step": 15220 }, { "epoch": 0.6261812803024077, "grad_norm": 1.1573827266693115, "learning_rate": 0.00016092386117687996, "loss": 0.8321, "step": 15240 }, { "epoch": 0.6270030405127783, "grad_norm": 1.1900019645690918, "learning_rate": 0.00016081827969696713, "loss": 0.825, "step": 15260 }, { "epoch": 0.627824800723149, "grad_norm": 1.1662086248397827, "learning_rate": 0.0001607125905125624, "loss": 0.8244, "step": 15280 }, { "epoch": 0.6286465609335196, "grad_norm": 1.0952377319335938, "learning_rate": 0.00016060679381083324, "loss": 0.8183, "step": 15300 }, { "epoch": 0.6294683211438902, "grad_norm": 1.1535894870758057, "learning_rate": 0.00016050088977913744, "loss": 0.8147, "step": 15320 }, { "epoch": 0.6302900813542608, "grad_norm": 1.3305004835128784, "learning_rate": 0.0001603948786050229, "loss": 0.8172, "step": 15340 }, { "epoch": 0.6311118415646314, "grad_norm": 1.3150241374969482, "learning_rate": 0.0001602887604762272, "loss": 0.8058, "step": 15360 }, { "epoch": 0.631933601775002, "grad_norm": 1.2193336486816406, "learning_rate": 0.00016018253558067744, "loss": 0.8165, "step": 15380 }, { "epoch": 0.6327553619853726, "grad_norm": 1.2560005187988281, "learning_rate": 0.0001600762041064897, "loss": 0.8206, "step": 15400 }, { "epoch": 0.6335771221957432, "grad_norm": 1.3278276920318604, "learning_rate": 0.00015996976624196884, "loss": 0.8178, "step": 15420 }, { "epoch": 0.6343988824061139, "grad_norm": 1.2090293169021606, "learning_rate": 0.00015986322217560815, "loss": 0.8443, "step": 15440 }, { "epoch": 0.6352206426164845, "grad_norm": 1.2651256322860718, "learning_rate": 0.00015975657209608895, "loss": 0.7986, "step": 15460 }, { "epoch": 0.6360424028268551, "grad_norm": 1.229529619216919, "learning_rate": 0.00015964981619228028, "loss": 0.8127, "step": 15480 }, { "epoch": 0.6368641630372257, "grad_norm": 1.235855221748352, "learning_rate": 0.00015954295465323866, "loss": 0.8184, "step": 15500 }, { "epoch": 0.6376859232475963, "grad_norm": 1.1475471258163452, "learning_rate": 0.00015943598766820772, "loss": 0.8258, "step": 15520 }, { "epoch": 0.6385076834579669, "grad_norm": 1.4335112571716309, "learning_rate": 0.00015932891542661768, "loss": 0.8153, "step": 15540 }, { "epoch": 0.6393294436683375, "grad_norm": 1.1948201656341553, "learning_rate": 0.00015922173811808526, "loss": 0.8501, "step": 15560 }, { "epoch": 0.6401512038787082, "grad_norm": 1.1850942373275757, "learning_rate": 0.00015911445593241318, "loss": 0.8307, "step": 15580 }, { "epoch": 0.6409729640890788, "grad_norm": 1.2479064464569092, "learning_rate": 0.00015900706905959, "loss": 0.812, "step": 15600 }, { "epoch": 0.6417947242994494, "grad_norm": 1.0971261262893677, "learning_rate": 0.00015889957768978956, "loss": 0.8059, "step": 15620 }, { "epoch": 0.64261648450982, "grad_norm": 1.409441351890564, "learning_rate": 0.0001587919820133708, "loss": 0.806, "step": 15640 }, { "epoch": 0.6434382447201906, "grad_norm": 1.1296477317810059, "learning_rate": 0.00015868428222087745, "loss": 0.8279, "step": 15660 }, { "epoch": 0.6442600049305612, "grad_norm": 1.2360674142837524, "learning_rate": 0.00015857647850303744, "loss": 0.8098, "step": 15680 }, { "epoch": 0.6450817651409319, "grad_norm": 1.2544413805007935, "learning_rate": 0.00015846857105076295, "loss": 0.8236, "step": 15700 }, { "epoch": 0.6459035253513025, "grad_norm": 0.9945731163024902, "learning_rate": 0.0001583605600551497, "loss": 0.8175, "step": 15720 }, { "epoch": 0.6467252855616731, "grad_norm": 1.3537676334381104, "learning_rate": 0.00015825244570747695, "loss": 0.807, "step": 15740 }, { "epoch": 0.6475470457720437, "grad_norm": 1.1799793243408203, "learning_rate": 0.0001581442281992068, "loss": 0.8446, "step": 15760 }, { "epoch": 0.6483688059824143, "grad_norm": 1.1959508657455444, "learning_rate": 0.00015803590772198417, "loss": 0.8156, "step": 15780 }, { "epoch": 0.6491905661927849, "grad_norm": 1.1003177165985107, "learning_rate": 0.0001579274844676362, "loss": 0.8295, "step": 15800 }, { "epoch": 0.6500123264031555, "grad_norm": 1.1961150169372559, "learning_rate": 0.00015781895862817226, "loss": 0.8428, "step": 15820 }, { "epoch": 0.6508340866135262, "grad_norm": 1.3239089250564575, "learning_rate": 0.0001577103303957832, "loss": 0.8322, "step": 15840 }, { "epoch": 0.6516558468238968, "grad_norm": 1.190565586090088, "learning_rate": 0.0001576015999628412, "loss": 0.8332, "step": 15860 }, { "epoch": 0.6524776070342674, "grad_norm": 1.158389925956726, "learning_rate": 0.0001574927675218996, "loss": 0.8065, "step": 15880 }, { "epoch": 0.653299367244638, "grad_norm": 1.1195902824401855, "learning_rate": 0.00015738383326569218, "loss": 0.8145, "step": 15900 }, { "epoch": 0.6541211274550086, "grad_norm": 1.1594524383544922, "learning_rate": 0.00015727479738713315, "loss": 0.8094, "step": 15920 }, { "epoch": 0.6549428876653792, "grad_norm": 1.176584005355835, "learning_rate": 0.0001571656600793167, "loss": 0.8383, "step": 15940 }, { "epoch": 0.6557646478757498, "grad_norm": 1.188091516494751, "learning_rate": 0.00015705642153551654, "loss": 0.8155, "step": 15960 }, { "epoch": 0.6565864080861205, "grad_norm": 1.1001297235488892, "learning_rate": 0.00015694708194918575, "loss": 0.8627, "step": 15980 }, { "epoch": 0.6574081682964911, "grad_norm": 1.2960819005966187, "learning_rate": 0.00015683764151395635, "loss": 0.8426, "step": 16000 }, { "epoch": 0.6574081682964911, "eval_loss": 1.121155858039856, "eval_runtime": 16.4628, "eval_samples_per_second": 159.147, "eval_steps_per_second": 4.981, "step": 16000 }, { "epoch": 0.6582299285068617, "grad_norm": 1.2257232666015625, "learning_rate": 0.00015672810042363885, "loss": 0.8271, "step": 16020 }, { "epoch": 0.6590516887172323, "grad_norm": 1.309381127357483, "learning_rate": 0.0001566239433327474, "loss": 0.8404, "step": 16040 }, { "epoch": 0.6598734489276029, "grad_norm": 1.1223074197769165, "learning_rate": 0.00015651420652313064, "loss": 0.8329, "step": 16060 }, { "epoch": 0.6606952091379735, "grad_norm": 1.2301430702209473, "learning_rate": 0.00015640436963120435, "loss": 0.8332, "step": 16080 }, { "epoch": 0.6615169693483441, "grad_norm": 1.0917423963546753, "learning_rate": 0.00015629443285148113, "loss": 0.8068, "step": 16100 }, { "epoch": 0.6623387295587148, "grad_norm": 1.2451525926589966, "learning_rate": 0.0001561843963786506, "loss": 0.8214, "step": 16120 }, { "epoch": 0.6631604897690854, "grad_norm": 1.3091555833816528, "learning_rate": 0.00015607426040757885, "loss": 0.8122, "step": 16140 }, { "epoch": 0.663982249979456, "grad_norm": 1.1900159120559692, "learning_rate": 0.00015596402513330822, "loss": 0.8296, "step": 16160 }, { "epoch": 0.6648040101898266, "grad_norm": 1.2061822414398193, "learning_rate": 0.00015585369075105693, "loss": 0.8226, "step": 16180 }, { "epoch": 0.6656257704001972, "grad_norm": 1.3246262073516846, "learning_rate": 0.00015574325745621866, "loss": 0.8141, "step": 16200 }, { "epoch": 0.6664475306105678, "grad_norm": 1.1857435703277588, "learning_rate": 0.00015563272544436222, "loss": 0.8213, "step": 16220 }, { "epoch": 0.6672692908209384, "grad_norm": 1.2091962099075317, "learning_rate": 0.00015552209491123136, "loss": 0.8306, "step": 16240 }, { "epoch": 0.6680910510313091, "grad_norm": 1.2190937995910645, "learning_rate": 0.00015541136605274423, "loss": 0.804, "step": 16260 }, { "epoch": 0.6689128112416797, "grad_norm": 1.2466769218444824, "learning_rate": 0.00015530053906499306, "loss": 0.8405, "step": 16280 }, { "epoch": 0.6697345714520503, "grad_norm": 1.1477371454238892, "learning_rate": 0.000155189614144244, "loss": 0.8193, "step": 16300 }, { "epoch": 0.6705563316624209, "grad_norm": 1.3594835996627808, "learning_rate": 0.00015508414493801062, "loss": 0.8351, "step": 16320 }, { "epoch": 0.6713780918727915, "grad_norm": 1.2145652770996094, "learning_rate": 0.00015497302961308234, "loss": 0.8394, "step": 16340 }, { "epoch": 0.6721998520831621, "grad_norm": 1.1144131422042847, "learning_rate": 0.00015486181693515012, "loss": 0.8403, "step": 16360 }, { "epoch": 0.6730216122935327, "grad_norm": 1.1824201345443726, "learning_rate": 0.0001547505071011631, "loss": 0.7953, "step": 16380 }, { "epoch": 0.6738433725039034, "grad_norm": 1.1716006994247437, "learning_rate": 0.00015463910030824243, "loss": 0.8262, "step": 16400 }, { "epoch": 0.674665132714274, "grad_norm": 1.243807315826416, "learning_rate": 0.00015452759675368093, "loss": 0.8239, "step": 16420 }, { "epoch": 0.6754868929246446, "grad_norm": 1.235845923423767, "learning_rate": 0.00015441599663494287, "loss": 0.8226, "step": 16440 }, { "epoch": 0.6763086531350152, "grad_norm": 1.1748180389404297, "learning_rate": 0.0001543043001496634, "loss": 0.8161, "step": 16460 }, { "epoch": 0.6771304133453858, "grad_norm": 1.1699196100234985, "learning_rate": 0.00015419250749564841, "loss": 0.8131, "step": 16480 }, { "epoch": 0.6779521735557564, "grad_norm": 1.238051176071167, "learning_rate": 0.00015408061887087416, "loss": 0.8475, "step": 16500 }, { "epoch": 0.678773933766127, "grad_norm": 1.3277729749679565, "learning_rate": 0.0001539686344734867, "loss": 0.8217, "step": 16520 }, { "epoch": 0.6795956939764977, "grad_norm": 1.1437911987304688, "learning_rate": 0.00015385655450180185, "loss": 0.8305, "step": 16540 }, { "epoch": 0.6804174541868683, "grad_norm": 1.2318732738494873, "learning_rate": 0.00015374437915430456, "loss": 0.8295, "step": 16560 }, { "epoch": 0.6812392143972389, "grad_norm": 1.1004635095596313, "learning_rate": 0.00015363210862964874, "loss": 0.8028, "step": 16580 }, { "epoch": 0.6820609746076095, "grad_norm": 0.9944893717765808, "learning_rate": 0.00015351974312665685, "loss": 0.8375, "step": 16600 }, { "epoch": 0.6828827348179801, "grad_norm": 1.3072353601455688, "learning_rate": 0.00015340728284431957, "loss": 0.8113, "step": 16620 }, { "epoch": 0.6837044950283507, "grad_norm": 1.173128604888916, "learning_rate": 0.0001532947279817954, "loss": 0.8208, "step": 16640 }, { "epoch": 0.6845262552387213, "grad_norm": 1.1952176094055176, "learning_rate": 0.0001531820787384103, "loss": 0.8139, "step": 16660 }, { "epoch": 0.685348015449092, "grad_norm": 1.189226508140564, "learning_rate": 0.00015306933531365746, "loss": 0.837, "step": 16680 }, { "epoch": 0.6861697756594626, "grad_norm": 1.0701826810836792, "learning_rate": 0.0001529564979071968, "loss": 0.8424, "step": 16700 }, { "epoch": 0.6869915358698332, "grad_norm": 1.2586934566497803, "learning_rate": 0.00015284356671885465, "loss": 0.81, "step": 16720 }, { "epoch": 0.6878132960802038, "grad_norm": 1.2510169744491577, "learning_rate": 0.00015273054194862344, "loss": 0.8353, "step": 16740 }, { "epoch": 0.6886350562905744, "grad_norm": 1.2519487142562866, "learning_rate": 0.00015261742379666138, "loss": 0.8164, "step": 16760 }, { "epoch": 0.689456816500945, "grad_norm": 1.2734975814819336, "learning_rate": 0.000152504212463292, "loss": 0.8136, "step": 16780 }, { "epoch": 0.6902785767113157, "grad_norm": 1.4471458196640015, "learning_rate": 0.00015239090814900386, "loss": 0.814, "step": 16800 }, { "epoch": 0.6911003369216863, "grad_norm": 1.1526660919189453, "learning_rate": 0.00015227751105445017, "loss": 0.8106, "step": 16820 }, { "epoch": 0.6919220971320569, "grad_norm": 1.2432206869125366, "learning_rate": 0.00015216402138044843, "loss": 0.7911, "step": 16840 }, { "epoch": 0.6927438573424275, "grad_norm": 1.3356374502182007, "learning_rate": 0.00015205043932798015, "loss": 0.8157, "step": 16860 }, { "epoch": 0.6935656175527981, "grad_norm": 1.105976939201355, "learning_rate": 0.00015193676509819043, "loss": 0.8088, "step": 16880 }, { "epoch": 0.6943873777631687, "grad_norm": 1.2425799369812012, "learning_rate": 0.0001518229988923875, "loss": 0.8318, "step": 16900 }, { "epoch": 0.6952091379735393, "grad_norm": 1.1508738994598389, "learning_rate": 0.00015171483598759317, "loss": 0.84, "step": 16920 }, { "epoch": 0.69603089818391, "grad_norm": 1.237749695777893, "learning_rate": 0.00015160089100819412, "loss": 0.823, "step": 16940 }, { "epoch": 0.6968526583942806, "grad_norm": 1.1718213558197021, "learning_rate": 0.00015148685464758913, "loss": 0.7944, "step": 16960 }, { "epoch": 0.6976744186046512, "grad_norm": 1.207578420639038, "learning_rate": 0.0001513727271077277, "loss": 0.8068, "step": 16980 }, { "epoch": 0.6984961788150218, "grad_norm": 1.180071473121643, "learning_rate": 0.00015125850859072098, "loss": 0.8302, "step": 17000 }, { "epoch": 0.6993179390253924, "grad_norm": 1.2875052690505981, "learning_rate": 0.00015114419929884116, "loss": 0.8382, "step": 17020 }, { "epoch": 0.700139699235763, "grad_norm": 1.3213404417037964, "learning_rate": 0.0001510297994345212, "loss": 0.8196, "step": 17040 }, { "epoch": 0.7009614594461336, "grad_norm": 1.2812501192092896, "learning_rate": 0.00015091530920035445, "loss": 0.85, "step": 17060 }, { "epoch": 0.7017832196565043, "grad_norm": 1.122597336769104, "learning_rate": 0.0001508007287990943, "loss": 0.8363, "step": 17080 }, { "epoch": 0.7026049798668749, "grad_norm": 1.1828125715255737, "learning_rate": 0.0001506860584336538, "loss": 0.8217, "step": 17100 }, { "epoch": 0.7034267400772455, "grad_norm": 1.187536358833313, "learning_rate": 0.00015057129830710542, "loss": 0.826, "step": 17120 }, { "epoch": 0.7042485002876161, "grad_norm": 1.1405049562454224, "learning_rate": 0.00015045644862268044, "loss": 0.8035, "step": 17140 }, { "epoch": 0.7050702604979867, "grad_norm": 1.255892276763916, "learning_rate": 0.0001503415095837688, "loss": 0.8251, "step": 17160 }, { "epoch": 0.7058920207083573, "grad_norm": 1.1409162282943726, "learning_rate": 0.00015022648139391875, "loss": 0.8049, "step": 17180 }, { "epoch": 0.7067137809187279, "grad_norm": 1.1684703826904297, "learning_rate": 0.00015011136425683628, "loss": 0.8166, "step": 17200 }, { "epoch": 0.7075355411290986, "grad_norm": 1.2653696537017822, "learning_rate": 0.00014999615837638506, "loss": 0.779, "step": 17220 }, { "epoch": 0.7083573013394692, "grad_norm": 1.2622226476669312, "learning_rate": 0.0001498808639565858, "loss": 0.8285, "step": 17240 }, { "epoch": 0.7091790615498398, "grad_norm": 1.152935266494751, "learning_rate": 0.00014976548120161607, "loss": 0.7834, "step": 17260 }, { "epoch": 0.7100008217602104, "grad_norm": 1.2912839651107788, "learning_rate": 0.0001496500103158098, "loss": 0.8015, "step": 17280 }, { "epoch": 0.710822581970581, "grad_norm": 1.2247825860977173, "learning_rate": 0.00014953445150365705, "loss": 0.8121, "step": 17300 }, { "epoch": 0.7116443421809516, "grad_norm": 1.2684624195098877, "learning_rate": 0.00014941880496980358, "loss": 0.8203, "step": 17320 }, { "epoch": 0.7124661023913222, "grad_norm": 1.2209104299545288, "learning_rate": 0.00014930307091905057, "loss": 0.8254, "step": 17340 }, { "epoch": 0.7132878626016929, "grad_norm": 1.3020516633987427, "learning_rate": 0.000149187249556354, "loss": 0.8342, "step": 17360 }, { "epoch": 0.7141096228120635, "grad_norm": 1.1557644605636597, "learning_rate": 0.00014907134108682466, "loss": 0.8199, "step": 17380 }, { "epoch": 0.7149313830224341, "grad_norm": 1.177384853363037, "learning_rate": 0.00014895534571572754, "loss": 0.8406, "step": 17400 }, { "epoch": 0.7157531432328047, "grad_norm": 1.2047346830368042, "learning_rate": 0.0001488392636484815, "loss": 0.8091, "step": 17420 }, { "epoch": 0.7165749034431753, "grad_norm": 1.171870231628418, "learning_rate": 0.00014872309509065886, "loss": 0.7942, "step": 17440 }, { "epoch": 0.7173966636535459, "grad_norm": 1.1541531085968018, "learning_rate": 0.00014860684024798536, "loss": 0.7921, "step": 17460 }, { "epoch": 0.7182184238639165, "grad_norm": 1.1826390027999878, "learning_rate": 0.0001484904993263392, "loss": 0.8068, "step": 17480 }, { "epoch": 0.7190401840742872, "grad_norm": 1.1863442659378052, "learning_rate": 0.0001483740725317513, "loss": 0.8347, "step": 17500 }, { "epoch": 0.7198619442846578, "grad_norm": 1.2402416467666626, "learning_rate": 0.00014825756007040458, "loss": 0.8276, "step": 17520 }, { "epoch": 0.7206837044950284, "grad_norm": 1.2695879936218262, "learning_rate": 0.00014814096214863355, "loss": 0.8035, "step": 17540 }, { "epoch": 0.721505464705399, "grad_norm": 1.1038098335266113, "learning_rate": 0.0001480242789729242, "loss": 0.8131, "step": 17560 }, { "epoch": 0.7223272249157696, "grad_norm": 1.3015015125274658, "learning_rate": 0.0001479075107499135, "loss": 0.7688, "step": 17580 }, { "epoch": 0.7231489851261402, "grad_norm": 1.1493345499038696, "learning_rate": 0.00014779065768638888, "loss": 0.8167, "step": 17600 }, { "epoch": 0.7239707453365108, "grad_norm": 1.1382313966751099, "learning_rate": 0.0001476737199892882, "loss": 0.8435, "step": 17620 }, { "epoch": 0.7247925055468815, "grad_norm": 1.2679277658462524, "learning_rate": 0.0001475566978656991, "loss": 0.8304, "step": 17640 }, { "epoch": 0.7256142657572521, "grad_norm": 1.2502251863479614, "learning_rate": 0.0001474395915228587, "loss": 0.8252, "step": 17660 }, { "epoch": 0.7264360259676227, "grad_norm": 1.1250088214874268, "learning_rate": 0.00014732240116815343, "loss": 0.8166, "step": 17680 }, { "epoch": 0.7272577861779933, "grad_norm": 1.1370155811309814, "learning_rate": 0.0001472051270091183, "loss": 0.7896, "step": 17700 }, { "epoch": 0.7280795463883639, "grad_norm": 1.2608290910720825, "learning_rate": 0.00014708776925343684, "loss": 0.8028, "step": 17720 }, { "epoch": 0.7289013065987345, "grad_norm": 1.3024847507476807, "learning_rate": 0.00014697032810894064, "loss": 0.7981, "step": 17740 }, { "epoch": 0.7297230668091051, "grad_norm": 1.1868743896484375, "learning_rate": 0.00014685280378360884, "loss": 0.7949, "step": 17760 }, { "epoch": 0.7305448270194758, "grad_norm": 1.253355622291565, "learning_rate": 0.00014673519648556805, "loss": 0.8107, "step": 17780 }, { "epoch": 0.7313665872298464, "grad_norm": 1.2940136194229126, "learning_rate": 0.00014661750642309173, "loss": 0.8017, "step": 17800 }, { "epoch": 0.732188347440217, "grad_norm": 1.2385964393615723, "learning_rate": 0.0001464997338045999, "loss": 0.8055, "step": 17820 }, { "epoch": 0.7330101076505876, "grad_norm": 1.2589930295944214, "learning_rate": 0.0001463818788386588, "loss": 0.7795, "step": 17840 }, { "epoch": 0.7338318678609582, "grad_norm": 1.0926438570022583, "learning_rate": 0.00014626394173398056, "loss": 0.8202, "step": 17860 }, { "epoch": 0.7346536280713288, "grad_norm": 1.1935014724731445, "learning_rate": 0.00014614592269942262, "loss": 0.7994, "step": 17880 }, { "epoch": 0.7354753882816994, "grad_norm": 1.1052745580673218, "learning_rate": 0.0001460278219439877, "loss": 0.8283, "step": 17900 }, { "epoch": 0.7362971484920701, "grad_norm": 1.2212902307510376, "learning_rate": 0.00014590963967682304, "loss": 0.7937, "step": 17920 }, { "epoch": 0.7371189087024407, "grad_norm": 1.2168195247650146, "learning_rate": 0.00014579137610722044, "loss": 0.7938, "step": 17940 }, { "epoch": 0.7379406689128113, "grad_norm": 1.1712826490402222, "learning_rate": 0.00014567303144461552, "loss": 0.8279, "step": 17960 }, { "epoch": 0.7387624291231819, "grad_norm": 1.2737464904785156, "learning_rate": 0.0001455546058985876, "loss": 0.8144, "step": 17980 }, { "epoch": 0.7395841893335525, "grad_norm": 1.2962466478347778, "learning_rate": 0.0001454360996788592, "loss": 0.8398, "step": 18000 }, { "epoch": 0.7395841893335525, "eval_loss": 1.0920464992523193, "eval_runtime": 16.6449, "eval_samples_per_second": 157.406, "eval_steps_per_second": 4.926, "step": 18000 }, { "epoch": 0.7404059495439231, "grad_norm": 1.1804122924804688, "learning_rate": 0.0001453175129952957, "loss": 0.8203, "step": 18020 }, { "epoch": 0.7412277097542938, "grad_norm": 1.0877560377120972, "learning_rate": 0.00014519884605790497, "loss": 0.7971, "step": 18040 }, { "epoch": 0.7420494699646644, "grad_norm": 1.2490891218185425, "learning_rate": 0.00014508009907683705, "loss": 0.8479, "step": 18060 }, { "epoch": 0.742871230175035, "grad_norm": 1.1328068971633911, "learning_rate": 0.0001449612722623837, "loss": 0.8062, "step": 18080 }, { "epoch": 0.7436929903854055, "grad_norm": 1.2985849380493164, "learning_rate": 0.00014484236582497806, "loss": 0.8267, "step": 18100 }, { "epoch": 0.7445147505957761, "grad_norm": 1.1852083206176758, "learning_rate": 0.00014472337997519432, "loss": 0.8416, "step": 18120 }, { "epoch": 0.7453365108061467, "grad_norm": 1.0732208490371704, "learning_rate": 0.0001446043149237472, "loss": 0.7889, "step": 18140 }, { "epoch": 0.7461582710165173, "grad_norm": 1.331568717956543, "learning_rate": 0.00014448517088149176, "loss": 0.7988, "step": 18160 }, { "epoch": 0.7469800312268879, "grad_norm": 1.2419168949127197, "learning_rate": 0.00014436594805942288, "loss": 0.8073, "step": 18180 }, { "epoch": 0.7478017914372586, "grad_norm": 1.2016359567642212, "learning_rate": 0.0001442466466686751, "loss": 0.8104, "step": 18200 }, { "epoch": 0.7486235516476292, "grad_norm": 1.0315485000610352, "learning_rate": 0.00014412726692052195, "loss": 0.8303, "step": 18220 }, { "epoch": 0.7494453118579998, "grad_norm": 1.1660343408584595, "learning_rate": 0.00014400780902637574, "loss": 0.8143, "step": 18240 }, { "epoch": 0.7502670720683704, "grad_norm": 1.1398091316223145, "learning_rate": 0.00014388827319778723, "loss": 0.8144, "step": 18260 }, { "epoch": 0.751088832278741, "grad_norm": 1.2301084995269775, "learning_rate": 0.00014376865964644522, "loss": 0.8116, "step": 18280 }, { "epoch": 0.7519105924891116, "grad_norm": 1.159977674484253, "learning_rate": 0.00014364896858417607, "loss": 0.8183, "step": 18300 }, { "epoch": 0.7527323526994822, "grad_norm": 1.3030027151107788, "learning_rate": 0.0001435292002229434, "loss": 0.8133, "step": 18320 }, { "epoch": 0.7535541129098529, "grad_norm": 1.2161463499069214, "learning_rate": 0.00014340935477484781, "loss": 0.8049, "step": 18340 }, { "epoch": 0.7543758731202235, "grad_norm": 1.2423648834228516, "learning_rate": 0.0001432894324521263, "loss": 0.8061, "step": 18360 }, { "epoch": 0.7551976333305941, "grad_norm": 1.2223330736160278, "learning_rate": 0.00014316943346715216, "loss": 0.8098, "step": 18380 }, { "epoch": 0.7560193935409647, "grad_norm": 1.2927947044372559, "learning_rate": 0.00014304935803243429, "loss": 0.7944, "step": 18400 }, { "epoch": 0.7568411537513353, "grad_norm": 1.2604374885559082, "learning_rate": 0.00014292920636061704, "loss": 0.8082, "step": 18420 }, { "epoch": 0.7576629139617059, "grad_norm": 1.2504435777664185, "learning_rate": 0.00014280897866447985, "loss": 0.8105, "step": 18440 }, { "epoch": 0.7584846741720765, "grad_norm": 1.2936044931411743, "learning_rate": 0.00014268867515693662, "loss": 0.7844, "step": 18460 }, { "epoch": 0.7593064343824472, "grad_norm": 1.1824711561203003, "learning_rate": 0.00014256829605103564, "loss": 0.8227, "step": 18480 }, { "epoch": 0.7601281945928178, "grad_norm": 1.300419807434082, "learning_rate": 0.00014244784155995906, "loss": 0.8513, "step": 18500 }, { "epoch": 0.7609499548031884, "grad_norm": 1.376930832862854, "learning_rate": 0.00014232731189702249, "loss": 0.8247, "step": 18520 }, { "epoch": 0.761771715013559, "grad_norm": 1.229896903038025, "learning_rate": 0.00014220670727567466, "loss": 0.7918, "step": 18540 }, { "epoch": 0.7625934752239296, "grad_norm": 1.1824287176132202, "learning_rate": 0.00014208602790949715, "loss": 0.8236, "step": 18560 }, { "epoch": 0.7634152354343002, "grad_norm": 1.1316356658935547, "learning_rate": 0.00014196527401220374, "loss": 0.8122, "step": 18580 }, { "epoch": 0.7642369956446708, "grad_norm": 1.1641656160354614, "learning_rate": 0.00014184444579764036, "loss": 0.7887, "step": 18600 }, { "epoch": 0.7650587558550415, "grad_norm": 1.2354685068130493, "learning_rate": 0.0001417235434797844, "loss": 0.8223, "step": 18620 }, { "epoch": 0.7658805160654121, "grad_norm": 1.108034372329712, "learning_rate": 0.00014160256727274462, "loss": 0.8055, "step": 18640 }, { "epoch": 0.7667022762757827, "grad_norm": 1.3176486492156982, "learning_rate": 0.00014148151739076055, "loss": 0.829, "step": 18660 }, { "epoch": 0.7675240364861533, "grad_norm": 1.0911662578582764, "learning_rate": 0.0001413603940482022, "loss": 0.7978, "step": 18680 }, { "epoch": 0.7683457966965239, "grad_norm": 1.0955551862716675, "learning_rate": 0.0001412391974595697, "loss": 0.8174, "step": 18700 }, { "epoch": 0.7691675569068945, "grad_norm": 1.1460391283035278, "learning_rate": 0.0001411179278394929, "loss": 0.8279, "step": 18720 }, { "epoch": 0.7699893171172651, "grad_norm": 1.1242876052856445, "learning_rate": 0.00014099658540273096, "loss": 0.8248, "step": 18740 }, { "epoch": 0.7708110773276358, "grad_norm": 1.1270663738250732, "learning_rate": 0.00014087517036417196, "loss": 0.8211, "step": 18760 }, { "epoch": 0.7716328375380064, "grad_norm": 1.1673957109451294, "learning_rate": 0.0001407536829388326, "loss": 0.8119, "step": 18780 }, { "epoch": 0.772454597748377, "grad_norm": 1.123095989227295, "learning_rate": 0.00014063212334185774, "loss": 0.8095, "step": 18800 }, { "epoch": 0.7732763579587476, "grad_norm": 1.0859897136688232, "learning_rate": 0.0001405104917885201, "loss": 0.7797, "step": 18820 }, { "epoch": 0.7740981181691182, "grad_norm": 1.2377300262451172, "learning_rate": 0.0001403887884942198, "loss": 0.8273, "step": 18840 }, { "epoch": 0.7749198783794888, "grad_norm": 1.2705551385879517, "learning_rate": 0.000140267013674484, "loss": 0.8051, "step": 18860 }, { "epoch": 0.7757416385898595, "grad_norm": 1.319014072418213, "learning_rate": 0.00014014516754496656, "loss": 0.8107, "step": 18880 }, { "epoch": 0.7765633988002301, "grad_norm": 1.3769927024841309, "learning_rate": 0.00014002325032144754, "loss": 0.7844, "step": 18900 }, { "epoch": 0.7773851590106007, "grad_norm": 1.3221933841705322, "learning_rate": 0.00013990126221983298, "loss": 0.8131, "step": 18920 }, { "epoch": 0.7782069192209713, "grad_norm": 1.2304918766021729, "learning_rate": 0.00013977920345615444, "loss": 0.8039, "step": 18940 }, { "epoch": 0.7790286794313419, "grad_norm": 1.385318398475647, "learning_rate": 0.00013965707424656856, "loss": 0.7985, "step": 18960 }, { "epoch": 0.7798504396417125, "grad_norm": 1.1849255561828613, "learning_rate": 0.00013953487480735679, "loss": 0.7902, "step": 18980 }, { "epoch": 0.7806721998520831, "grad_norm": 1.2645635604858398, "learning_rate": 0.0001394126053549249, "loss": 0.7931, "step": 19000 }, { "epoch": 0.7814939600624538, "grad_norm": 1.2201372385025024, "learning_rate": 0.00013929026610580276, "loss": 0.8244, "step": 19020 }, { "epoch": 0.7823157202728244, "grad_norm": 1.1765888929367065, "learning_rate": 0.0001391739793672815, "loss": 0.8012, "step": 19040 }, { "epoch": 0.783137480483195, "grad_norm": 1.0896601676940918, "learning_rate": 0.0001390515046378748, "loss": 0.8107, "step": 19060 }, { "epoch": 0.7839592406935656, "grad_norm": 1.125887393951416, "learning_rate": 0.00013892896075125936, "loss": 0.8233, "step": 19080 }, { "epoch": 0.7847810009039362, "grad_norm": 1.1168248653411865, "learning_rate": 0.00013880634792445097, "loss": 0.8036, "step": 19100 }, { "epoch": 0.7856027611143068, "grad_norm": 1.3335416316986084, "learning_rate": 0.00013868366637458755, "loss": 0.8174, "step": 19120 }, { "epoch": 0.7864245213246774, "grad_norm": 1.281420111656189, "learning_rate": 0.0001385609163189286, "loss": 0.8116, "step": 19140 }, { "epoch": 0.787246281535048, "grad_norm": 1.2268955707550049, "learning_rate": 0.00013843809797485515, "loss": 0.8147, "step": 19160 }, { "epoch": 0.7880680417454187, "grad_norm": 1.2544898986816406, "learning_rate": 0.0001383152115598689, "loss": 0.826, "step": 19180 }, { "epoch": 0.7888898019557893, "grad_norm": 1.3066948652267456, "learning_rate": 0.00013819225729159228, "loss": 0.8072, "step": 19200 }, { "epoch": 0.7897115621661599, "grad_norm": 1.2460455894470215, "learning_rate": 0.00013806923538776787, "loss": 0.7989, "step": 19220 }, { "epoch": 0.7905333223765305, "grad_norm": 1.155281662940979, "learning_rate": 0.00013794614606625792, "loss": 0.8405, "step": 19240 }, { "epoch": 0.7913550825869011, "grad_norm": 1.054686427116394, "learning_rate": 0.00013782914896373142, "loss": 0.8056, "step": 19260 }, { "epoch": 0.7921768427972717, "grad_norm": 1.3717306852340698, "learning_rate": 0.00013770592880481307, "loss": 0.802, "step": 19280 }, { "epoch": 0.7929986030076424, "grad_norm": 1.2226320505142212, "learning_rate": 0.0001375826418715972, "loss": 0.813, "step": 19300 }, { "epoch": 0.793820363218013, "grad_norm": 1.250246524810791, "learning_rate": 0.00013745928838241556, "loss": 0.8225, "step": 19320 }, { "epoch": 0.7946421234283836, "grad_norm": 1.0490273237228394, "learning_rate": 0.00013733586855571754, "loss": 0.7996, "step": 19340 }, { "epoch": 0.7954638836387542, "grad_norm": 1.1633754968643188, "learning_rate": 0.00013721238261007023, "loss": 0.8224, "step": 19360 }, { "epoch": 0.7962856438491248, "grad_norm": 1.079055666923523, "learning_rate": 0.0001370888307641578, "loss": 0.8207, "step": 19380 }, { "epoch": 0.7971074040594954, "grad_norm": 1.3664897680282593, "learning_rate": 0.00013696521323678092, "loss": 0.8015, "step": 19400 }, { "epoch": 0.797929164269866, "grad_norm": 1.1427098512649536, "learning_rate": 0.00013684153024685685, "loss": 0.8093, "step": 19420 }, { "epoch": 0.7987509244802367, "grad_norm": 1.1648412942886353, "learning_rate": 0.00013671778201341855, "loss": 0.7898, "step": 19440 }, { "epoch": 0.7995726846906073, "grad_norm": 1.1498785018920898, "learning_rate": 0.00013659396875561467, "loss": 0.8113, "step": 19460 }, { "epoch": 0.8003944449009779, "grad_norm": 1.1189064979553223, "learning_rate": 0.00013647009069270893, "loss": 0.8062, "step": 19480 }, { "epoch": 0.8012162051113485, "grad_norm": 1.225690245628357, "learning_rate": 0.00013634614804407984, "loss": 0.7994, "step": 19500 }, { "epoch": 0.8020379653217191, "grad_norm": 1.1157780885696411, "learning_rate": 0.0001362221410292203, "loss": 0.7882, "step": 19520 }, { "epoch": 0.8028597255320897, "grad_norm": 1.0617575645446777, "learning_rate": 0.00013609806986773722, "loss": 0.8035, "step": 19540 }, { "epoch": 0.8036814857424603, "grad_norm": 1.128070592880249, "learning_rate": 0.00013597393477935102, "loss": 0.8095, "step": 19560 }, { "epoch": 0.804503245952831, "grad_norm": 1.1231319904327393, "learning_rate": 0.00013584973598389544, "loss": 0.799, "step": 19580 }, { "epoch": 0.8053250061632016, "grad_norm": 1.4038573503494263, "learning_rate": 0.00013572547370131695, "loss": 0.7849, "step": 19600 }, { "epoch": 0.8061467663735722, "grad_norm": 1.208066463470459, "learning_rate": 0.00013560114815167447, "loss": 0.8127, "step": 19620 }, { "epoch": 0.8069685265839428, "grad_norm": 1.243054747581482, "learning_rate": 0.00013547675955513904, "loss": 0.8139, "step": 19640 }, { "epoch": 0.8077902867943134, "grad_norm": 1.1254897117614746, "learning_rate": 0.00013535230813199324, "loss": 0.7909, "step": 19660 }, { "epoch": 0.808612047004684, "grad_norm": 1.3242278099060059, "learning_rate": 0.00013522779410263098, "loss": 0.8242, "step": 19680 }, { "epoch": 0.8094338072150546, "grad_norm": 1.042738437652588, "learning_rate": 0.00013510321768755702, "loss": 0.7948, "step": 19700 }, { "epoch": 0.8102555674254253, "grad_norm": 1.1268333196640015, "learning_rate": 0.00013497857910738662, "loss": 0.7923, "step": 19720 }, { "epoch": 0.8110773276357959, "grad_norm": 1.115422010421753, "learning_rate": 0.00013485387858284506, "loss": 0.7998, "step": 19740 }, { "epoch": 0.8118990878461665, "grad_norm": 1.2183444499969482, "learning_rate": 0.0001347291163347674, "loss": 0.8054, "step": 19760 }, { "epoch": 0.8127208480565371, "grad_norm": 1.158079981803894, "learning_rate": 0.000134604292584098, "loss": 0.7716, "step": 19780 }, { "epoch": 0.8135426082669077, "grad_norm": 1.0598976612091064, "learning_rate": 0.0001344794075518901, "loss": 0.7967, "step": 19800 }, { "epoch": 0.8143643684772783, "grad_norm": 1.1043282747268677, "learning_rate": 0.00013435446145930544, "loss": 0.7848, "step": 19820 }, { "epoch": 0.815186128687649, "grad_norm": 1.2107045650482178, "learning_rate": 0.00013422945452761398, "loss": 0.7996, "step": 19840 }, { "epoch": 0.8160078888980196, "grad_norm": 1.1288707256317139, "learning_rate": 0.00013410438697819337, "loss": 0.8172, "step": 19860 }, { "epoch": 0.8168296491083902, "grad_norm": 1.215134859085083, "learning_rate": 0.0001339792590325286, "loss": 0.8155, "step": 19880 }, { "epoch": 0.8176514093187608, "grad_norm": 1.3704839944839478, "learning_rate": 0.00013385407091221163, "loss": 0.7831, "step": 19900 }, { "epoch": 0.8184731695291314, "grad_norm": 1.2613569498062134, "learning_rate": 0.000133728822838941, "loss": 0.8154, "step": 19920 }, { "epoch": 0.819294929739502, "grad_norm": 1.2519367933273315, "learning_rate": 0.00013360351503452137, "loss": 0.8089, "step": 19940 }, { "epoch": 0.8201166899498726, "grad_norm": 1.259395956993103, "learning_rate": 0.0001334781477208632, "loss": 0.8161, "step": 19960 }, { "epoch": 0.8209384501602432, "grad_norm": 1.1715264320373535, "learning_rate": 0.00013335272111998246, "loss": 0.8009, "step": 19980 }, { "epoch": 0.8217602103706139, "grad_norm": 1.1847702264785767, "learning_rate": 0.0001332272354539999, "loss": 0.774, "step": 20000 }, { "epoch": 0.8217602103706139, "eval_loss": 1.0668244361877441, "eval_runtime": 16.5724, "eval_samples_per_second": 158.094, "eval_steps_per_second": 4.948, "step": 20000 }, { "epoch": 0.8225819705809845, "grad_norm": 1.224970817565918, "learning_rate": 0.00013310169094514103, "loss": 0.8, "step": 20020 }, { "epoch": 0.8234037307913551, "grad_norm": 1.1203746795654297, "learning_rate": 0.0001329760878157355, "loss": 0.8158, "step": 20040 }, { "epoch": 0.8242254910017257, "grad_norm": 1.1533716917037964, "learning_rate": 0.00013285042628821675, "loss": 0.7896, "step": 20060 }, { "epoch": 0.8250472512120963, "grad_norm": 1.213652491569519, "learning_rate": 0.00013272470658512174, "loss": 0.8121, "step": 20080 }, { "epoch": 0.8258690114224669, "grad_norm": 1.2149641513824463, "learning_rate": 0.00013259892892909033, "loss": 0.7926, "step": 20100 }, { "epoch": 0.8266907716328376, "grad_norm": 1.3824020624160767, "learning_rate": 0.00013247309354286503, "loss": 0.7909, "step": 20120 }, { "epoch": 0.8275125318432082, "grad_norm": 1.0057183504104614, "learning_rate": 0.00013234720064929075, "loss": 0.819, "step": 20140 }, { "epoch": 0.8283342920535788, "grad_norm": 1.1706757545471191, "learning_rate": 0.000132221250471314, "loss": 0.8248, "step": 20160 }, { "epoch": 0.8291560522639494, "grad_norm": 1.0104855298995972, "learning_rate": 0.00013209524323198294, "loss": 0.7992, "step": 20180 }, { "epoch": 0.82997781247432, "grad_norm": 1.2821381092071533, "learning_rate": 0.0001319691791544466, "loss": 0.802, "step": 20200 }, { "epoch": 0.8307995726846906, "grad_norm": 1.106156826019287, "learning_rate": 0.0001318430584619548, "loss": 0.7799, "step": 20220 }, { "epoch": 0.8316213328950612, "grad_norm": 1.1786506175994873, "learning_rate": 0.0001317168813778576, "loss": 0.7927, "step": 20240 }, { "epoch": 0.8324430931054319, "grad_norm": 1.280644178390503, "learning_rate": 0.00013159064812560484, "loss": 0.7922, "step": 20260 }, { "epoch": 0.8332648533158025, "grad_norm": 1.2000395059585571, "learning_rate": 0.00013146435892874596, "loss": 0.8024, "step": 20280 }, { "epoch": 0.8340866135261731, "grad_norm": 1.384141206741333, "learning_rate": 0.00013134433257673927, "loss": 0.8428, "step": 20300 }, { "epoch": 0.8349083737365437, "grad_norm": 1.1807432174682617, "learning_rate": 0.00013121793493125713, "loss": 0.8039, "step": 20320 }, { "epoch": 0.8357301339469143, "grad_norm": 1.1562411785125732, "learning_rate": 0.00013109148200121524, "loss": 0.8043, "step": 20340 }, { "epoch": 0.8365518941572849, "grad_norm": 1.0757108926773071, "learning_rate": 0.00013096497401055197, "loss": 0.7933, "step": 20360 }, { "epoch": 0.8373736543676555, "grad_norm": 1.1456866264343262, "learning_rate": 0.00013083841118330326, "loss": 0.7793, "step": 20380 }, { "epoch": 0.8381954145780262, "grad_norm": 1.3488410711288452, "learning_rate": 0.0001307117937436021, "loss": 0.7775, "step": 20400 }, { "epoch": 0.8390171747883968, "grad_norm": 1.2938168048858643, "learning_rate": 0.0001305851219156783, "loss": 0.8144, "step": 20420 }, { "epoch": 0.8398389349987674, "grad_norm": 1.131605625152588, "learning_rate": 0.00013045839592385787, "loss": 0.7883, "step": 20440 }, { "epoch": 0.840660695209138, "grad_norm": 1.1760270595550537, "learning_rate": 0.00013033161599256275, "loss": 0.7984, "step": 20460 }, { "epoch": 0.8414824554195086, "grad_norm": 1.1592392921447754, "learning_rate": 0.00013020478234631049, "loss": 0.8142, "step": 20480 }, { "epoch": 0.8423042156298792, "grad_norm": 1.3598871231079102, "learning_rate": 0.00013007789520971374, "loss": 0.7861, "step": 20500 }, { "epoch": 0.8431259758402498, "grad_norm": 1.2098135948181152, "learning_rate": 0.00012995095480747976, "loss": 0.8069, "step": 20520 }, { "epoch": 0.8439477360506205, "grad_norm": 1.0988825559616089, "learning_rate": 0.00012982396136441033, "loss": 0.7971, "step": 20540 }, { "epoch": 0.8447694962609911, "grad_norm": 0.9486870765686035, "learning_rate": 0.000129696915105401, "loss": 0.7807, "step": 20560 }, { "epoch": 0.8455912564713617, "grad_norm": 1.1476234197616577, "learning_rate": 0.00012956981625544093, "loss": 0.7884, "step": 20580 }, { "epoch": 0.8464130166817323, "grad_norm": 1.1225764751434326, "learning_rate": 0.00012944266503961242, "loss": 0.7966, "step": 20600 }, { "epoch": 0.8472347768921029, "grad_norm": 1.316215991973877, "learning_rate": 0.00012931546168309046, "loss": 0.8115, "step": 20620 }, { "epoch": 0.8480565371024735, "grad_norm": 1.1493761539459229, "learning_rate": 0.0001291882064111424, "loss": 0.8094, "step": 20640 }, { "epoch": 0.8488782973128441, "grad_norm": 1.1531728506088257, "learning_rate": 0.00012906089944912755, "loss": 0.8158, "step": 20660 }, { "epoch": 0.8497000575232148, "grad_norm": 1.1480746269226074, "learning_rate": 0.00012893354102249673, "loss": 0.8085, "step": 20680 }, { "epoch": 0.8505218177335854, "grad_norm": 1.265832543373108, "learning_rate": 0.00012880613135679193, "loss": 0.803, "step": 20700 }, { "epoch": 0.851343577943956, "grad_norm": 1.0262377262115479, "learning_rate": 0.00012867867067764584, "loss": 0.774, "step": 20720 }, { "epoch": 0.8521653381543266, "grad_norm": 1.136478066444397, "learning_rate": 0.00012855115921078152, "loss": 0.779, "step": 20740 }, { "epoch": 0.8529870983646972, "grad_norm": 1.1349806785583496, "learning_rate": 0.00012842359718201194, "loss": 0.7922, "step": 20760 }, { "epoch": 0.8538088585750678, "grad_norm": 1.2569667100906372, "learning_rate": 0.00012829598481723964, "loss": 0.7739, "step": 20780 }, { "epoch": 0.8546306187854384, "grad_norm": 1.2146037817001343, "learning_rate": 0.00012816832234245634, "loss": 0.8221, "step": 20800 }, { "epoch": 0.8554523789958091, "grad_norm": 1.300690770149231, "learning_rate": 0.00012804060998374245, "loss": 0.7928, "step": 20820 }, { "epoch": 0.8562741392061797, "grad_norm": 1.0861823558807373, "learning_rate": 0.00012791284796726663, "loss": 0.8043, "step": 20840 }, { "epoch": 0.8570958994165503, "grad_norm": 1.2741750478744507, "learning_rate": 0.0001277850365192857, "loss": 0.8065, "step": 20860 }, { "epoch": 0.8579176596269209, "grad_norm": 1.0767971277236938, "learning_rate": 0.00012765717586614382, "loss": 0.8153, "step": 20880 }, { "epoch": 0.8587394198372915, "grad_norm": 1.1135772466659546, "learning_rate": 0.0001275292662342724, "loss": 0.7685, "step": 20900 }, { "epoch": 0.8595611800476621, "grad_norm": 1.2278940677642822, "learning_rate": 0.0001274013078501895, "loss": 0.7635, "step": 20920 }, { "epoch": 0.8603829402580327, "grad_norm": 1.0262848138809204, "learning_rate": 0.00012727330094049967, "loss": 0.7907, "step": 20940 }, { "epoch": 0.8612047004684034, "grad_norm": 1.2303491830825806, "learning_rate": 0.0001271452457318932, "loss": 0.8026, "step": 20960 }, { "epoch": 0.862026460678774, "grad_norm": 1.1088433265686035, "learning_rate": 0.00012701714245114603, "loss": 0.7966, "step": 20980 }, { "epoch": 0.8628482208891446, "grad_norm": 1.1193281412124634, "learning_rate": 0.00012688899132511924, "loss": 0.7924, "step": 21000 }, { "epoch": 0.8636699810995152, "grad_norm": 1.308245301246643, "learning_rate": 0.00012676079258075858, "loss": 0.7885, "step": 21020 }, { "epoch": 0.8644917413098858, "grad_norm": 1.2952789068222046, "learning_rate": 0.0001266325464450942, "loss": 0.7913, "step": 21040 }, { "epoch": 0.8653135015202564, "grad_norm": 1.1136795282363892, "learning_rate": 0.00012650425314524006, "loss": 0.7927, "step": 21060 }, { "epoch": 0.866135261730627, "grad_norm": 1.222264051437378, "learning_rate": 0.00012637591290839376, "loss": 0.8171, "step": 21080 }, { "epoch": 0.8669570219409977, "grad_norm": 1.3129292726516724, "learning_rate": 0.000126247525961836, "loss": 0.786, "step": 21100 }, { "epoch": 0.8677787821513683, "grad_norm": 1.15342116355896, "learning_rate": 0.00012611909253293016, "loss": 0.7821, "step": 21120 }, { "epoch": 0.8686005423617389, "grad_norm": 1.2096205949783325, "learning_rate": 0.00012599061284912193, "loss": 0.7886, "step": 21140 }, { "epoch": 0.8694223025721095, "grad_norm": 1.0738475322723389, "learning_rate": 0.00012586208713793898, "loss": 0.7813, "step": 21160 }, { "epoch": 0.8702440627824801, "grad_norm": 1.1030628681182861, "learning_rate": 0.00012573351562699048, "loss": 0.7859, "step": 21180 }, { "epoch": 0.8710658229928507, "grad_norm": 1.0959700345993042, "learning_rate": 0.0001256048985439666, "loss": 0.8055, "step": 21200 }, { "epoch": 0.8718875832032214, "grad_norm": 1.2536059617996216, "learning_rate": 0.00012547623611663836, "loss": 0.7837, "step": 21220 }, { "epoch": 0.872709343413592, "grad_norm": 1.2621644735336304, "learning_rate": 0.00012534752857285692, "loss": 0.7977, "step": 21240 }, { "epoch": 0.8735311036239626, "grad_norm": 1.1967788934707642, "learning_rate": 0.00012521877614055357, "loss": 0.8035, "step": 21260 }, { "epoch": 0.8743528638343332, "grad_norm": 1.2444006204605103, "learning_rate": 0.00012508997904773884, "loss": 0.771, "step": 21280 }, { "epoch": 0.8751746240447038, "grad_norm": 1.4090861082077026, "learning_rate": 0.00012497402366798297, "loss": 0.7657, "step": 21300 }, { "epoch": 0.8759963842550744, "grad_norm": 1.2610723972320557, "learning_rate": 0.00012484514234864898, "loss": 0.7928, "step": 21320 }, { "epoch": 0.876818144465445, "grad_norm": 1.2457926273345947, "learning_rate": 0.0001247226643378368, "loss": 0.7985, "step": 21340 }, { "epoch": 0.8776399046758157, "grad_norm": 1.2153024673461914, "learning_rate": 0.0001245936974322525, "loss": 0.8184, "step": 21360 }, { "epoch": 0.8784616648861862, "grad_norm": 1.244173526763916, "learning_rate": 0.00012446468697312316, "loss": 0.7855, "step": 21380 }, { "epoch": 0.8792834250965568, "grad_norm": 1.2349773645401, "learning_rate": 0.00012433563318891633, "loss": 0.7915, "step": 21400 }, { "epoch": 0.8801051853069274, "grad_norm": 1.3552359342575073, "learning_rate": 0.0001242065363081764, "loss": 0.7878, "step": 21420 }, { "epoch": 0.880926945517298, "grad_norm": 1.071760654449463, "learning_rate": 0.00012407739655952393, "loss": 0.7836, "step": 21440 }, { "epoch": 0.8817487057276686, "grad_norm": 1.2500207424163818, "learning_rate": 0.0001239482141716555, "loss": 0.7781, "step": 21460 }, { "epoch": 0.8825704659380392, "grad_norm": 1.2251540422439575, "learning_rate": 0.00012381898937334322, "loss": 0.7826, "step": 21480 }, { "epoch": 0.8833922261484098, "grad_norm": 1.0519969463348389, "learning_rate": 0.0001236897223934342, "loss": 0.7764, "step": 21500 }, { "epoch": 0.8842139863587805, "grad_norm": 0.9632487297058105, "learning_rate": 0.00012356041346085032, "loss": 0.8034, "step": 21520 }, { "epoch": 0.8850357465691511, "grad_norm": 1.2027411460876465, "learning_rate": 0.00012343106280458777, "loss": 0.8176, "step": 21540 }, { "epoch": 0.8858575067795217, "grad_norm": 1.1366872787475586, "learning_rate": 0.00012330814124322003, "loss": 0.7884, "step": 21560 }, { "epoch": 0.8866792669898923, "grad_norm": 1.137412667274475, "learning_rate": 0.00012317870988471372, "loss": 0.7777, "step": 21580 }, { "epoch": 0.8875010272002629, "grad_norm": 1.1433818340301514, "learning_rate": 0.00012304923747849642, "loss": 0.7818, "step": 21600 }, { "epoch": 0.8883227874106335, "grad_norm": 1.1305466890335083, "learning_rate": 0.00012291972425385373, "loss": 0.7829, "step": 21620 }, { "epoch": 0.8891445476210041, "grad_norm": 1.1296701431274414, "learning_rate": 0.00012279017044014364, "loss": 0.8018, "step": 21640 }, { "epoch": 0.8899663078313748, "grad_norm": 1.1467466354370117, "learning_rate": 0.00012266057626679594, "loss": 0.8009, "step": 21660 }, { "epoch": 0.8907880680417454, "grad_norm": 1.2313017845153809, "learning_rate": 0.00012253094196331202, "loss": 0.7924, "step": 21680 }, { "epoch": 0.891609828252116, "grad_norm": 1.1884132623672485, "learning_rate": 0.00012240126775926418, "loss": 0.8, "step": 21700 }, { "epoch": 0.8924315884624866, "grad_norm": 1.1451033353805542, "learning_rate": 0.00012227155388429547, "loss": 0.8158, "step": 21720 }, { "epoch": 0.8932533486728572, "grad_norm": 1.2778671979904175, "learning_rate": 0.00012214180056811916, "loss": 0.773, "step": 21740 }, { "epoch": 0.8940751088832278, "grad_norm": 1.0006611347198486, "learning_rate": 0.00012201200804051842, "loss": 0.7859, "step": 21760 }, { "epoch": 0.8948968690935984, "grad_norm": 1.270612120628357, "learning_rate": 0.0001218821765313458, "loss": 0.7977, "step": 21780 }, { "epoch": 0.8957186293039691, "grad_norm": 1.4604487419128418, "learning_rate": 0.00012175230627052294, "loss": 0.8046, "step": 21800 }, { "epoch": 0.8965403895143397, "grad_norm": 1.153067946434021, "learning_rate": 0.00012162239748804005, "loss": 0.7672, "step": 21820 }, { "epoch": 0.8973621497247103, "grad_norm": 1.1451218128204346, "learning_rate": 0.00012149245041395559, "loss": 0.7758, "step": 21840 }, { "epoch": 0.8981839099350809, "grad_norm": 1.2750080823898315, "learning_rate": 0.00012136246527839583, "loss": 0.7816, "step": 21860 }, { "epoch": 0.8990056701454515, "grad_norm": 1.0611047744750977, "learning_rate": 0.00012123244231155442, "loss": 0.7668, "step": 21880 }, { "epoch": 0.8998274303558221, "grad_norm": 1.098158359527588, "learning_rate": 0.00012110238174369202, "loss": 0.8227, "step": 21900 }, { "epoch": 0.9006491905661927, "grad_norm": 1.0489616394042969, "learning_rate": 0.00012097228380513595, "loss": 0.8008, "step": 21920 }, { "epoch": 0.9014709507765634, "grad_norm": 1.1194531917572021, "learning_rate": 0.00012084214872627955, "loss": 0.8118, "step": 21940 }, { "epoch": 0.902292710986934, "grad_norm": 1.1527105569839478, "learning_rate": 0.00012071197673758212, "loss": 0.769, "step": 21960 }, { "epoch": 0.9031144711973046, "grad_norm": 1.1321651935577393, "learning_rate": 0.00012058176806956818, "loss": 0.7789, "step": 21980 }, { "epoch": 0.9039362314076752, "grad_norm": 1.25917387008667, "learning_rate": 0.00012045152295282727, "loss": 0.7828, "step": 22000 }, { "epoch": 0.9039362314076752, "eval_loss": 1.0472674369812012, "eval_runtime": 23.5948, "eval_samples_per_second": 111.042, "eval_steps_per_second": 3.475, "step": 22000 }, { "epoch": 0.9047579916180458, "grad_norm": 1.260465145111084, "learning_rate": 0.00012032124161801345, "loss": 0.761, "step": 22020 }, { "epoch": 0.9055797518284164, "grad_norm": 1.0791617631912231, "learning_rate": 0.00012019092429584495, "loss": 0.7737, "step": 22040 }, { "epoch": 0.906401512038787, "grad_norm": 1.277144193649292, "learning_rate": 0.0001200605712171037, "loss": 0.8014, "step": 22060 }, { "epoch": 0.9072232722491577, "grad_norm": 1.1210474967956543, "learning_rate": 0.00011993018261263503, "loss": 0.8066, "step": 22080 }, { "epoch": 0.9080450324595283, "grad_norm": 1.2601996660232544, "learning_rate": 0.00011979975871334703, "loss": 0.7787, "step": 22100 }, { "epoch": 0.9088667926698989, "grad_norm": 1.1547856330871582, "learning_rate": 0.00011966929975021046, "loss": 0.7936, "step": 22120 }, { "epoch": 0.9096885528802695, "grad_norm": 1.1628038883209229, "learning_rate": 0.00011953880595425808, "loss": 0.7803, "step": 22140 }, { "epoch": 0.9105103130906401, "grad_norm": 1.1185588836669922, "learning_rate": 0.00011940827755658433, "loss": 0.8004, "step": 22160 }, { "epoch": 0.9113320733010107, "grad_norm": 1.0631072521209717, "learning_rate": 0.00011927771478834496, "loss": 0.7898, "step": 22180 }, { "epoch": 0.9121538335113814, "grad_norm": 1.2936205863952637, "learning_rate": 0.00011914711788075663, "loss": 0.7809, "step": 22200 }, { "epoch": 0.912975593721752, "grad_norm": 1.084246039390564, "learning_rate": 0.00011901648706509636, "loss": 0.795, "step": 22220 }, { "epoch": 0.9137973539321226, "grad_norm": 1.3070451021194458, "learning_rate": 0.00011888582257270132, "loss": 0.793, "step": 22240 }, { "epoch": 0.9146191141424932, "grad_norm": 1.2026519775390625, "learning_rate": 0.00011875512463496822, "loss": 0.8013, "step": 22260 }, { "epoch": 0.9154408743528638, "grad_norm": 1.2209608554840088, "learning_rate": 0.00011862439348335306, "loss": 0.7708, "step": 22280 }, { "epoch": 0.9162626345632344, "grad_norm": 1.1526132822036743, "learning_rate": 0.00011849362934937065, "loss": 0.7822, "step": 22300 }, { "epoch": 0.917084394773605, "grad_norm": 1.0761892795562744, "learning_rate": 0.00011836283246459421, "loss": 0.8061, "step": 22320 }, { "epoch": 0.9179061549839757, "grad_norm": 1.2090603113174438, "learning_rate": 0.00011823200306065494, "loss": 0.7692, "step": 22340 }, { "epoch": 0.9187279151943463, "grad_norm": 1.1188862323760986, "learning_rate": 0.00011810114136924166, "loss": 0.8228, "step": 22360 }, { "epoch": 0.9195496754047169, "grad_norm": 1.138457179069519, "learning_rate": 0.00011797024762210034, "loss": 0.7883, "step": 22380 }, { "epoch": 0.9203714356150875, "grad_norm": 1.0787718296051025, "learning_rate": 0.00011783932205103376, "loss": 0.7743, "step": 22400 }, { "epoch": 0.9211931958254581, "grad_norm": 1.0644888877868652, "learning_rate": 0.00011770836488790103, "loss": 0.8028, "step": 22420 }, { "epoch": 0.9220149560358287, "grad_norm": 1.1350173950195312, "learning_rate": 0.00011757737636461711, "loss": 0.769, "step": 22440 }, { "epoch": 0.9228367162461993, "grad_norm": 1.2350654602050781, "learning_rate": 0.00011744635671315274, "loss": 0.7614, "step": 22460 }, { "epoch": 0.92365847645657, "grad_norm": 1.1380183696746826, "learning_rate": 0.0001173153061655335, "loss": 0.7789, "step": 22480 }, { "epoch": 0.9244802366669406, "grad_norm": 1.2965232133865356, "learning_rate": 0.00011718422495383992, "loss": 0.8028, "step": 22500 }, { "epoch": 0.9253019968773112, "grad_norm": 1.2787531614303589, "learning_rate": 0.00011705311331020667, "loss": 0.7729, "step": 22520 }, { "epoch": 0.9261237570876818, "grad_norm": 1.1900768280029297, "learning_rate": 0.00011692197146682242, "loss": 0.7856, "step": 22540 }, { "epoch": 0.9269455172980524, "grad_norm": 1.2483367919921875, "learning_rate": 0.0001167907996559293, "loss": 0.7906, "step": 22560 }, { "epoch": 0.927767277508423, "grad_norm": 1.0251802206039429, "learning_rate": 0.00011665959810982245, "loss": 0.7756, "step": 22580 }, { "epoch": 0.9285890377187936, "grad_norm": 1.267059087753296, "learning_rate": 0.00011652836706084969, "loss": 0.8083, "step": 22600 }, { "epoch": 0.9294107979291643, "grad_norm": 1.1620546579360962, "learning_rate": 0.00011639710674141114, "loss": 0.7737, "step": 22620 }, { "epoch": 0.9302325581395349, "grad_norm": 1.09406578540802, "learning_rate": 0.00011626581738395872, "loss": 0.8101, "step": 22640 }, { "epoch": 0.9310543183499055, "grad_norm": 1.101579189300537, "learning_rate": 0.00011613449922099576, "loss": 0.7853, "step": 22660 }, { "epoch": 0.9318760785602761, "grad_norm": 1.2870060205459595, "learning_rate": 0.00011600315248507666, "loss": 0.7786, "step": 22680 }, { "epoch": 0.9326978387706467, "grad_norm": 1.1342302560806274, "learning_rate": 0.00011587177740880633, "loss": 0.7688, "step": 22700 }, { "epoch": 0.9335195989810173, "grad_norm": 1.2600188255310059, "learning_rate": 0.00011574037422483995, "loss": 0.7672, "step": 22720 }, { "epoch": 0.9343413591913879, "grad_norm": 1.1466922760009766, "learning_rate": 0.00011560894316588243, "loss": 0.7855, "step": 22740 }, { "epoch": 0.9351631194017586, "grad_norm": 1.2696952819824219, "learning_rate": 0.00011547748446468802, "loss": 0.7869, "step": 22760 }, { "epoch": 0.9359848796121292, "grad_norm": 1.2031900882720947, "learning_rate": 0.00011534599835406001, "loss": 0.784, "step": 22780 }, { "epoch": 0.9368066398224998, "grad_norm": 1.1982570886611938, "learning_rate": 0.00011521448506685022, "loss": 0.7838, "step": 22800 }, { "epoch": 0.9376284000328704, "grad_norm": 1.0865845680236816, "learning_rate": 0.00011508294483595845, "loss": 0.7608, "step": 22820 }, { "epoch": 0.938450160243241, "grad_norm": 1.1527810096740723, "learning_rate": 0.00011495137789433243, "loss": 0.7637, "step": 22840 }, { "epoch": 0.9392719204536116, "grad_norm": 1.3062829971313477, "learning_rate": 0.00011481978447496704, "loss": 0.7919, "step": 22860 }, { "epoch": 0.9400936806639822, "grad_norm": 1.1722383499145508, "learning_rate": 0.00011468816481090406, "loss": 0.7742, "step": 22880 }, { "epoch": 0.9409154408743529, "grad_norm": 1.1022741794586182, "learning_rate": 0.00011455651913523184, "loss": 0.7798, "step": 22900 }, { "epoch": 0.9417372010847235, "grad_norm": 1.2316502332687378, "learning_rate": 0.0001144248476810847, "loss": 0.7957, "step": 22920 }, { "epoch": 0.9425589612950941, "grad_norm": 1.0445023775100708, "learning_rate": 0.00011429315068164269, "loss": 0.7784, "step": 22940 }, { "epoch": 0.9433807215054647, "grad_norm": 1.1832512617111206, "learning_rate": 0.00011416142837013101, "loss": 0.7939, "step": 22960 }, { "epoch": 0.9442024817158353, "grad_norm": 1.1852409839630127, "learning_rate": 0.00011402968097981976, "loss": 0.7692, "step": 22980 }, { "epoch": 0.9450242419262059, "grad_norm": 1.3063126802444458, "learning_rate": 0.00011389790874402347, "loss": 0.7647, "step": 23000 }, { "epoch": 0.9458460021365765, "grad_norm": 1.3952887058258057, "learning_rate": 0.00011376611189610056, "loss": 0.7942, "step": 23020 }, { "epoch": 0.9466677623469472, "grad_norm": 1.1620287895202637, "learning_rate": 0.0001136342906694531, "loss": 0.791, "step": 23040 }, { "epoch": 0.9474895225573178, "grad_norm": 1.1343626976013184, "learning_rate": 0.0001135024452975264, "loss": 0.7983, "step": 23060 }, { "epoch": 0.9483112827676884, "grad_norm": 1.3010191917419434, "learning_rate": 0.00011337057601380841, "loss": 0.7934, "step": 23080 }, { "epoch": 0.949133042978059, "grad_norm": 1.0730324983596802, "learning_rate": 0.0001132386830518295, "loss": 0.7907, "step": 23100 }, { "epoch": 0.9499548031884296, "grad_norm": 1.1792380809783936, "learning_rate": 0.00011310676664516196, "loss": 0.7756, "step": 23120 }, { "epoch": 0.9507765633988002, "grad_norm": 1.037816047668457, "learning_rate": 0.00011297482702741958, "loss": 0.7706, "step": 23140 }, { "epoch": 0.9515983236091708, "grad_norm": 1.2920358180999756, "learning_rate": 0.00011284286443225725, "loss": 0.7788, "step": 23160 }, { "epoch": 0.9524200838195415, "grad_norm": 1.3024156093597412, "learning_rate": 0.00011271087909337059, "loss": 0.7726, "step": 23180 }, { "epoch": 0.9532418440299121, "grad_norm": 1.0410034656524658, "learning_rate": 0.00011257887124449549, "loss": 0.7358, "step": 23200 }, { "epoch": 0.9540636042402827, "grad_norm": 1.1379297971725464, "learning_rate": 0.00011244684111940765, "loss": 0.7654, "step": 23220 }, { "epoch": 0.9548853644506533, "grad_norm": 1.1472039222717285, "learning_rate": 0.00011231478895192232, "loss": 0.7398, "step": 23240 }, { "epoch": 0.9557071246610239, "grad_norm": 1.1397725343704224, "learning_rate": 0.00011218271497589364, "loss": 0.802, "step": 23260 }, { "epoch": 0.9565288848713945, "grad_norm": 1.2153700590133667, "learning_rate": 0.00011205061942521453, "loss": 0.8053, "step": 23280 }, { "epoch": 0.9573506450817652, "grad_norm": 1.1267844438552856, "learning_rate": 0.00011191850253381601, "loss": 0.7778, "step": 23300 }, { "epoch": 0.9581724052921358, "grad_norm": 1.2119829654693604, "learning_rate": 0.00011178636453566691, "loss": 0.8014, "step": 23320 }, { "epoch": 0.9589941655025064, "grad_norm": 1.1640605926513672, "learning_rate": 0.00011165420566477351, "loss": 0.7643, "step": 23340 }, { "epoch": 0.959815925712877, "grad_norm": 1.2424917221069336, "learning_rate": 0.0001115220261551789, "loss": 0.7769, "step": 23360 }, { "epoch": 0.9606376859232476, "grad_norm": 1.1949933767318726, "learning_rate": 0.0001113898262409629, "loss": 0.7647, "step": 23380 }, { "epoch": 0.9614594461336182, "grad_norm": 1.0828710794448853, "learning_rate": 0.0001112576061562414, "loss": 0.7729, "step": 23400 }, { "epoch": 0.9622812063439888, "grad_norm": 1.1788461208343506, "learning_rate": 0.00011112536613516589, "loss": 0.7833, "step": 23420 }, { "epoch": 0.9631029665543595, "grad_norm": 1.1178921461105347, "learning_rate": 0.00011099310641192335, "loss": 0.7873, "step": 23440 }, { "epoch": 0.9639247267647301, "grad_norm": 1.131232738494873, "learning_rate": 0.00011086082722073556, "loss": 0.7622, "step": 23460 }, { "epoch": 0.9647464869751007, "grad_norm": 1.1513261795043945, "learning_rate": 0.00011072852879585876, "loss": 0.7781, "step": 23480 }, { "epoch": 0.9655682471854713, "grad_norm": 1.185996413230896, "learning_rate": 0.00011059621137158332, "loss": 0.7987, "step": 23500 }, { "epoch": 0.9663900073958419, "grad_norm": 1.2174021005630493, "learning_rate": 0.00011046387518223314, "loss": 0.7732, "step": 23520 }, { "epoch": 0.9672117676062125, "grad_norm": 1.2955466508865356, "learning_rate": 0.00011033152046216546, "loss": 0.772, "step": 23540 }, { "epoch": 0.9680335278165831, "grad_norm": 1.1724337339401245, "learning_rate": 0.00011019914744577034, "loss": 0.7811, "step": 23560 }, { "epoch": 0.9688552880269538, "grad_norm": 1.1841801404953003, "learning_rate": 0.00011006675636747017, "loss": 0.7893, "step": 23580 }, { "epoch": 0.9696770482373244, "grad_norm": 1.1488229036331177, "learning_rate": 0.00010993434746171933, "loss": 0.759, "step": 23600 }, { "epoch": 0.970498808447695, "grad_norm": 1.2947933673858643, "learning_rate": 0.00010980192096300389, "loss": 0.7821, "step": 23620 }, { "epoch": 0.9713205686580656, "grad_norm": 1.1167196035385132, "learning_rate": 0.00010966947710584086, "loss": 0.7765, "step": 23640 }, { "epoch": 0.9721423288684362, "grad_norm": 1.114274024963379, "learning_rate": 0.00010953701612477821, "loss": 0.7563, "step": 23660 }, { "epoch": 0.9729640890788068, "grad_norm": 1.1539628505706787, "learning_rate": 0.00010940453825439411, "loss": 0.7754, "step": 23680 }, { "epoch": 0.9737858492891774, "grad_norm": 1.1876133680343628, "learning_rate": 0.00010927204372929667, "loss": 0.7681, "step": 23700 }, { "epoch": 0.974607609499548, "grad_norm": 1.0904812812805176, "learning_rate": 0.00010913953278412353, "loss": 0.7726, "step": 23720 }, { "epoch": 0.9754293697099187, "grad_norm": 1.1562939882278442, "learning_rate": 0.00010900700565354131, "loss": 0.7774, "step": 23740 }, { "epoch": 0.9762511299202893, "grad_norm": 1.227735161781311, "learning_rate": 0.0001088744625722454, "loss": 0.7907, "step": 23760 }, { "epoch": 0.9770728901306599, "grad_norm": 1.1245887279510498, "learning_rate": 0.00010874190377495938, "loss": 0.7882, "step": 23780 }, { "epoch": 0.9778946503410305, "grad_norm": 1.1092678308486938, "learning_rate": 0.0001086093294964347, "loss": 0.759, "step": 23800 }, { "epoch": 0.9787164105514011, "grad_norm": 1.1554052829742432, "learning_rate": 0.00010847673997145016, "loss": 0.7665, "step": 23820 }, { "epoch": 0.9795381707617717, "grad_norm": 1.1881402730941772, "learning_rate": 0.00010834413543481163, "loss": 0.7899, "step": 23840 }, { "epoch": 0.9803599309721424, "grad_norm": 1.209314227104187, "learning_rate": 0.00010821151612135155, "loss": 0.7478, "step": 23860 }, { "epoch": 0.981181691182513, "grad_norm": 1.184097409248352, "learning_rate": 0.00010807888226592848, "loss": 0.7626, "step": 23880 }, { "epoch": 0.9820034513928836, "grad_norm": 1.109991431236267, "learning_rate": 0.00010794623410342682, "loss": 0.7748, "step": 23900 }, { "epoch": 0.9828252116032542, "grad_norm": 1.1060677766799927, "learning_rate": 0.00010781357186875619, "loss": 0.7873, "step": 23920 }, { "epoch": 0.9836469718136248, "grad_norm": 1.1834135055541992, "learning_rate": 0.00010768089579685126, "loss": 0.7814, "step": 23940 }, { "epoch": 0.9844687320239954, "grad_norm": 1.1708906888961792, "learning_rate": 0.00010754820612267106, "loss": 0.753, "step": 23960 }, { "epoch": 0.985290492234366, "grad_norm": 1.2200385332107544, "learning_rate": 0.00010741550308119885, "loss": 0.7651, "step": 23980 }, { "epoch": 0.9861122524447367, "grad_norm": 1.0666234493255615, "learning_rate": 0.00010728278690744153, "loss": 0.7795, "step": 24000 }, { "epoch": 0.9861122524447367, "eval_loss": 1.02431058883667, "eval_runtime": 16.5783, "eval_samples_per_second": 158.038, "eval_steps_per_second": 4.946, "step": 24000 }, { "epoch": 0.9869340126551073, "grad_norm": 1.1617413759231567, "learning_rate": 0.00010715005783642917, "loss": 0.7878, "step": 24020 }, { "epoch": 0.9877557728654779, "grad_norm": 1.11324942111969, "learning_rate": 0.00010701731610321475, "loss": 0.7969, "step": 24040 }, { "epoch": 0.9885775330758485, "grad_norm": 1.1129649877548218, "learning_rate": 0.00010688456194287368, "loss": 0.7595, "step": 24060 }, { "epoch": 0.9893992932862191, "grad_norm": 1.186141848564148, "learning_rate": 0.00010675179559050332, "loss": 0.8078, "step": 24080 }, { "epoch": 0.9902210534965897, "grad_norm": 1.205206036567688, "learning_rate": 0.00010661901728122272, "loss": 0.7449, "step": 24100 }, { "epoch": 0.9910428137069603, "grad_norm": 1.2314426898956299, "learning_rate": 0.00010648622725017199, "loss": 0.768, "step": 24120 }, { "epoch": 0.991864573917331, "grad_norm": 1.1358891725540161, "learning_rate": 0.00010635342573251209, "loss": 0.7753, "step": 24140 }, { "epoch": 0.9926863341277016, "grad_norm": 1.3078477382659912, "learning_rate": 0.00010622061296342425, "loss": 0.7845, "step": 24160 }, { "epoch": 0.9935080943380722, "grad_norm": 1.105637550354004, "learning_rate": 0.0001060877891781097, "loss": 0.7741, "step": 24180 }, { "epoch": 0.9943298545484428, "grad_norm": 1.1030614376068115, "learning_rate": 0.00010595495461178912, "loss": 0.7611, "step": 24200 }, { "epoch": 0.9951516147588134, "grad_norm": 1.0990793704986572, "learning_rate": 0.00010582210949970233, "loss": 0.7711, "step": 24220 }, { "epoch": 0.995973374969184, "grad_norm": 1.1833900213241577, "learning_rate": 0.00010568925407710773, "loss": 0.7684, "step": 24240 }, { "epoch": 0.9967951351795546, "grad_norm": 1.0902843475341797, "learning_rate": 0.00010555638857928209, "loss": 0.7709, "step": 24260 }, { "epoch": 0.9976168953899253, "grad_norm": 1.1372528076171875, "learning_rate": 0.00010542351324152, "loss": 0.7764, "step": 24280 }, { "epoch": 0.9984386556002959, "grad_norm": 1.0270088911056519, "learning_rate": 0.00010529062829913343, "loss": 0.7638, "step": 24300 }, { "epoch": 0.9992604158106665, "grad_norm": 1.0260668992996216, "learning_rate": 0.0001051577339874514, "loss": 0.781, "step": 24320 }, { "epoch": 1.000082176021037, "grad_norm": 1.0618846416473389, "learning_rate": 0.00010502483054181948, "loss": 0.7616, "step": 24340 }, { "epoch": 1.0009039362314076, "grad_norm": 1.0496258735656738, "learning_rate": 0.00010489191819759945, "loss": 0.6874, "step": 24360 }, { "epoch": 1.0017256964417782, "grad_norm": 1.304571270942688, "learning_rate": 0.00010475899719016886, "loss": 0.6708, "step": 24380 }, { "epoch": 1.0025474566521488, "grad_norm": 1.4253283739089966, "learning_rate": 0.00010462606775492054, "loss": 0.6789, "step": 24400 }, { "epoch": 1.0033692168625195, "grad_norm": 1.1459177732467651, "learning_rate": 0.00010449313012726234, "loss": 0.6678, "step": 24420 }, { "epoch": 1.00419097707289, "grad_norm": 1.2071423530578613, "learning_rate": 0.00010436018454261654, "loss": 0.6639, "step": 24440 }, { "epoch": 1.0050127372832607, "grad_norm": 1.1461540460586548, "learning_rate": 0.00010422723123641956, "loss": 0.6722, "step": 24460 }, { "epoch": 1.0058344974936313, "grad_norm": 1.2865045070648193, "learning_rate": 0.00010409427044412141, "loss": 0.6756, "step": 24480 }, { "epoch": 1.006656257704002, "grad_norm": 1.0493274927139282, "learning_rate": 0.00010396130240118549, "loss": 0.677, "step": 24500 }, { "epoch": 1.0074780179143725, "grad_norm": 1.171351671218872, "learning_rate": 0.00010382832734308792, "loss": 0.6736, "step": 24520 }, { "epoch": 1.0082997781247431, "grad_norm": 1.1830470561981201, "learning_rate": 0.00010369534550531734, "loss": 0.6575, "step": 24540 }, { "epoch": 1.0091215383351138, "grad_norm": 1.1490957736968994, "learning_rate": 0.00010357565624934433, "loss": 0.6628, "step": 24560 }, { "epoch": 1.0099432985454844, "grad_norm": 1.0752952098846436, "learning_rate": 0.00010344266217900869, "loss": 0.6489, "step": 24580 }, { "epoch": 1.010765058755855, "grad_norm": 1.1401771306991577, "learning_rate": 0.00010330966201198336, "loss": 0.6436, "step": 24600 }, { "epoch": 1.0115868189662256, "grad_norm": 1.0942751169204712, "learning_rate": 0.00010317665598380131, "loss": 0.6908, "step": 24620 }, { "epoch": 1.0124085791765962, "grad_norm": 1.23777437210083, "learning_rate": 0.00010304364433000604, "loss": 0.676, "step": 24640 }, { "epoch": 1.0132303393869668, "grad_norm": 1.3880153894424438, "learning_rate": 0.00010291062728615099, "loss": 0.6838, "step": 24660 }, { "epoch": 1.0140520995973374, "grad_norm": 1.090610384941101, "learning_rate": 0.00010277760508779903, "loss": 0.6505, "step": 24680 }, { "epoch": 1.014873859807708, "grad_norm": 1.1579679250717163, "learning_rate": 0.00010264457797052227, "loss": 0.679, "step": 24700 }, { "epoch": 1.0156956200180787, "grad_norm": 1.1469173431396484, "learning_rate": 0.00010251154616990151, "loss": 0.7002, "step": 24720 }, { "epoch": 1.0165173802284493, "grad_norm": 1.1271486282348633, "learning_rate": 0.00010237850992152578, "loss": 0.6774, "step": 24740 }, { "epoch": 1.01733914043882, "grad_norm": 1.2221741676330566, "learning_rate": 0.00010224546946099209, "loss": 0.6912, "step": 24760 }, { "epoch": 1.0181609006491905, "grad_norm": 1.0773913860321045, "learning_rate": 0.00010211242502390481, "loss": 0.6631, "step": 24780 }, { "epoch": 1.0189826608595611, "grad_norm": 1.3026174306869507, "learning_rate": 0.0001019793768458754, "loss": 0.6731, "step": 24800 }, { "epoch": 1.0198044210699317, "grad_norm": 1.1742732524871826, "learning_rate": 0.00010184632516252199, "loss": 0.6562, "step": 24820 }, { "epoch": 1.0206261812803024, "grad_norm": 1.1922106742858887, "learning_rate": 0.00010171992303113935, "loss": 0.6519, "step": 24840 }, { "epoch": 1.021447941490673, "grad_norm": 1.0810050964355469, "learning_rate": 0.0001015868651901237, "loss": 0.6625, "step": 24860 }, { "epoch": 1.0222697017010436, "grad_norm": 1.4511394500732422, "learning_rate": 0.00010145380453889195, "loss": 0.6955, "step": 24880 }, { "epoch": 1.0230914619114142, "grad_norm": 1.1028388738632202, "learning_rate": 0.0001013207413130843, "loss": 0.6576, "step": 24900 }, { "epoch": 1.0239132221217848, "grad_norm": 1.169084072113037, "learning_rate": 0.0001011876757483454, "loss": 0.6678, "step": 24920 }, { "epoch": 1.0247349823321554, "grad_norm": 1.1536842584609985, "learning_rate": 0.00010105460808032418, "loss": 0.6477, "step": 24940 }, { "epoch": 1.025556742542526, "grad_norm": 1.123879313468933, "learning_rate": 0.00010092153854467326, "loss": 0.679, "step": 24960 }, { "epoch": 1.0263785027528967, "grad_norm": 1.019872784614563, "learning_rate": 0.00010078846737704848, "loss": 0.6854, "step": 24980 }, { "epoch": 1.0272002629632673, "grad_norm": 1.3068866729736328, "learning_rate": 0.00010065539481310871, "loss": 0.6717, "step": 25000 }, { "epoch": 1.0280220231736379, "grad_norm": 1.2124892473220825, "learning_rate": 0.00010052232108851513, "loss": 0.649, "step": 25020 }, { "epoch": 1.0288437833840085, "grad_norm": 1.2036775350570679, "learning_rate": 0.00010038924643893113, "loss": 0.666, "step": 25040 }, { "epoch": 1.0296655435943791, "grad_norm": 1.244070053100586, "learning_rate": 0.00010025617110002165, "loss": 0.6654, "step": 25060 }, { "epoch": 1.0304873038047497, "grad_norm": 1.1506962776184082, "learning_rate": 0.00010012309530745285, "loss": 0.6764, "step": 25080 }, { "epoch": 1.0313090640151203, "grad_norm": 1.1000773906707764, "learning_rate": 9.999001929689177e-05, "loss": 0.6574, "step": 25100 }, { "epoch": 1.032130824225491, "grad_norm": 1.0797061920166016, "learning_rate": 9.985694330400571e-05, "loss": 0.6618, "step": 25120 }, { "epoch": 1.0329525844358616, "grad_norm": 1.1669949293136597, "learning_rate": 9.972386756446208e-05, "loss": 0.6731, "step": 25140 }, { "epoch": 1.0337743446462322, "grad_norm": 1.1310721635818481, "learning_rate": 9.959079231392771e-05, "loss": 0.6773, "step": 25160 }, { "epoch": 1.0345961048566028, "grad_norm": 1.1668230295181274, "learning_rate": 9.945771778806865e-05, "loss": 0.6772, "step": 25180 }, { "epoch": 1.0354178650669734, "grad_norm": 1.1975557804107666, "learning_rate": 9.93246442225497e-05, "loss": 0.6801, "step": 25200 }, { "epoch": 1.036239625277344, "grad_norm": 1.1820405721664429, "learning_rate": 9.919157185303379e-05, "loss": 0.6581, "step": 25220 }, { "epoch": 1.0370613854877146, "grad_norm": 1.1540806293487549, "learning_rate": 9.90585009151819e-05, "loss": 0.6732, "step": 25240 }, { "epoch": 1.0378831456980853, "grad_norm": 1.064178228378296, "learning_rate": 9.892543164465243e-05, "loss": 0.6732, "step": 25260 }, { "epoch": 1.0387049059084559, "grad_norm": 1.107135534286499, "learning_rate": 9.879236427710082e-05, "loss": 0.6708, "step": 25280 }, { "epoch": 1.0395266661188265, "grad_norm": 1.172105312347412, "learning_rate": 9.865929904817909e-05, "loss": 0.6818, "step": 25300 }, { "epoch": 1.040348426329197, "grad_norm": 1.1912182569503784, "learning_rate": 9.85262361935356e-05, "loss": 0.7184, "step": 25320 }, { "epoch": 1.0411701865395677, "grad_norm": 1.1023154258728027, "learning_rate": 9.83931759488143e-05, "loss": 0.6638, "step": 25340 }, { "epoch": 1.0419919467499383, "grad_norm": 1.076657772064209, "learning_rate": 9.826011854965474e-05, "loss": 0.6727, "step": 25360 }, { "epoch": 1.042813706960309, "grad_norm": 1.2090641260147095, "learning_rate": 9.812706423169129e-05, "loss": 0.6725, "step": 25380 }, { "epoch": 1.0436354671706796, "grad_norm": 1.1991325616836548, "learning_rate": 9.799401323055292e-05, "loss": 0.6793, "step": 25400 }, { "epoch": 1.0444572273810502, "grad_norm": 1.0969122648239136, "learning_rate": 9.786096578186273e-05, "loss": 0.6803, "step": 25420 }, { "epoch": 1.0452789875914208, "grad_norm": 1.2352911233901978, "learning_rate": 9.772792212123748e-05, "loss": 0.6789, "step": 25440 }, { "epoch": 1.0461007478017914, "grad_norm": 1.2759590148925781, "learning_rate": 9.759488248428725e-05, "loss": 0.6728, "step": 25460 }, { "epoch": 1.046922508012162, "grad_norm": 1.3433208465576172, "learning_rate": 9.7461847106615e-05, "loss": 0.6835, "step": 25480 }, { "epoch": 1.0477442682225326, "grad_norm": 1.2637454271316528, "learning_rate": 9.732881622381616e-05, "loss": 0.6738, "step": 25500 }, { "epoch": 1.0485660284329033, "grad_norm": 1.1877262592315674, "learning_rate": 9.719579007147815e-05, "loss": 0.6633, "step": 25520 }, { "epoch": 1.0493877886432739, "grad_norm": 1.1830005645751953, "learning_rate": 9.706276888518013e-05, "loss": 0.6559, "step": 25540 }, { "epoch": 1.0502095488536445, "grad_norm": 1.1390386819839478, "learning_rate": 9.692975290049228e-05, "loss": 0.6782, "step": 25560 }, { "epoch": 1.051031309064015, "grad_norm": 1.1353988647460938, "learning_rate": 9.679674235297572e-05, "loss": 0.6781, "step": 25580 }, { "epoch": 1.0518530692743857, "grad_norm": 1.1977840662002563, "learning_rate": 9.666373747818187e-05, "loss": 0.6697, "step": 25600 }, { "epoch": 1.0526748294847563, "grad_norm": 1.0405921936035156, "learning_rate": 9.653073851165214e-05, "loss": 0.6715, "step": 25620 }, { "epoch": 1.053496589695127, "grad_norm": 1.1129424571990967, "learning_rate": 9.63977456889175e-05, "loss": 0.6806, "step": 25640 }, { "epoch": 1.0543183499054976, "grad_norm": 1.163406491279602, "learning_rate": 9.626475924549792e-05, "loss": 0.6457, "step": 25660 }, { "epoch": 1.0551401101158682, "grad_norm": 1.1244803667068481, "learning_rate": 9.613177941690219e-05, "loss": 0.6664, "step": 25680 }, { "epoch": 1.0559618703262388, "grad_norm": 1.1291669607162476, "learning_rate": 9.599880643862737e-05, "loss": 0.6642, "step": 25700 }, { "epoch": 1.0567836305366094, "grad_norm": 1.3684684038162231, "learning_rate": 9.586584054615836e-05, "loss": 0.6833, "step": 25720 }, { "epoch": 1.05760539074698, "grad_norm": 1.0214548110961914, "learning_rate": 9.57328819749675e-05, "loss": 0.6693, "step": 25740 }, { "epoch": 1.0584271509573506, "grad_norm": 1.0439046621322632, "learning_rate": 9.559993096051425e-05, "loss": 0.6563, "step": 25760 }, { "epoch": 1.0592489111677212, "grad_norm": 1.1112405061721802, "learning_rate": 9.546698773824453e-05, "loss": 0.6683, "step": 25780 }, { "epoch": 1.0600706713780919, "grad_norm": 1.2522891759872437, "learning_rate": 9.53340525435906e-05, "loss": 0.656, "step": 25800 }, { "epoch": 1.0608924315884625, "grad_norm": 1.133664846420288, "learning_rate": 9.520112561197045e-05, "loss": 0.6891, "step": 25820 }, { "epoch": 1.061714191798833, "grad_norm": 1.3006486892700195, "learning_rate": 9.506820717878745e-05, "loss": 0.6536, "step": 25840 }, { "epoch": 1.0625359520092037, "grad_norm": 1.2655730247497559, "learning_rate": 9.493529747942996e-05, "loss": 0.687, "step": 25860 }, { "epoch": 1.0633577122195743, "grad_norm": 1.1235599517822266, "learning_rate": 9.480239674927074e-05, "loss": 0.665, "step": 25880 }, { "epoch": 1.064179472429945, "grad_norm": 1.250252604484558, "learning_rate": 9.46695052236668e-05, "loss": 0.6604, "step": 25900 }, { "epoch": 1.0650012326403155, "grad_norm": 1.0628433227539062, "learning_rate": 9.453662313795879e-05, "loss": 0.6565, "step": 25920 }, { "epoch": 1.0658229928506862, "grad_norm": 1.1810961961746216, "learning_rate": 9.440375072747064e-05, "loss": 0.6722, "step": 25940 }, { "epoch": 1.0666447530610568, "grad_norm": 1.2217936515808105, "learning_rate": 9.427753111349993e-05, "loss": 0.6913, "step": 25960 }, { "epoch": 1.0674665132714274, "grad_norm": 1.0944212675094604, "learning_rate": 9.414467824647579e-05, "loss": 0.6668, "step": 25980 }, { "epoch": 1.068288273481798, "grad_norm": 1.083652377128601, "learning_rate": 9.401183574877579e-05, "loss": 0.6713, "step": 26000 }, { "epoch": 1.068288273481798, "eval_loss": 1.0167440176010132, "eval_runtime": 16.6957, "eval_samples_per_second": 156.927, "eval_steps_per_second": 4.911, "step": 26000 }, { "epoch": 1.0691100336921686, "grad_norm": 1.117463231086731, "learning_rate": 9.387900385565371e-05, "loss": 0.6435, "step": 26020 }, { "epoch": 1.0699317939025392, "grad_norm": 1.027099847793579, "learning_rate": 9.374618280234465e-05, "loss": 0.6909, "step": 26040 }, { "epoch": 1.0707535541129098, "grad_norm": 1.3012546300888062, "learning_rate": 9.36133728240645e-05, "loss": 0.6741, "step": 26060 }, { "epoch": 1.0715753143232805, "grad_norm": 1.179627776145935, "learning_rate": 9.348057415600942e-05, "loss": 0.687, "step": 26080 }, { "epoch": 1.072397074533651, "grad_norm": 1.1117219924926758, "learning_rate": 9.334778703335568e-05, "loss": 0.6616, "step": 26100 }, { "epoch": 1.0732188347440217, "grad_norm": 1.1664800643920898, "learning_rate": 9.321501169125905e-05, "loss": 0.6648, "step": 26120 }, { "epoch": 1.0740405949543923, "grad_norm": 1.1486011743545532, "learning_rate": 9.308224836485447e-05, "loss": 0.684, "step": 26140 }, { "epoch": 1.074862355164763, "grad_norm": 1.1502150297164917, "learning_rate": 9.294949728925554e-05, "loss": 0.6682, "step": 26160 }, { "epoch": 1.0756841153751335, "grad_norm": 1.0855724811553955, "learning_rate": 9.281675869955419e-05, "loss": 0.671, "step": 26180 }, { "epoch": 1.0765058755855041, "grad_norm": 1.267439365386963, "learning_rate": 9.268403283082024e-05, "loss": 0.6935, "step": 26200 }, { "epoch": 1.0773276357958748, "grad_norm": 1.19661545753479, "learning_rate": 9.255131991810099e-05, "loss": 0.6611, "step": 26220 }, { "epoch": 1.0781493960062454, "grad_norm": 1.1992172002792358, "learning_rate": 9.241862019642083e-05, "loss": 0.6527, "step": 26240 }, { "epoch": 1.078971156216616, "grad_norm": 1.117606520652771, "learning_rate": 9.228593390078073e-05, "loss": 0.6792, "step": 26260 }, { "epoch": 1.0797929164269866, "grad_norm": 1.1372742652893066, "learning_rate": 9.21532612661579e-05, "loss": 0.6742, "step": 26280 }, { "epoch": 1.0806146766373572, "grad_norm": 1.1025826930999756, "learning_rate": 9.202060252750539e-05, "loss": 0.6559, "step": 26300 }, { "epoch": 1.0814364368477278, "grad_norm": 1.222730278968811, "learning_rate": 9.188795791975155e-05, "loss": 0.6711, "step": 26320 }, { "epoch": 1.0822581970580984, "grad_norm": 1.1559460163116455, "learning_rate": 9.17553276777998e-05, "loss": 0.6432, "step": 26340 }, { "epoch": 1.083079957268469, "grad_norm": 1.0200605392456055, "learning_rate": 9.162271203652811e-05, "loss": 0.6606, "step": 26360 }, { "epoch": 1.0839017174788397, "grad_norm": 1.1216548681259155, "learning_rate": 9.149011123078861e-05, "loss": 0.6608, "step": 26380 }, { "epoch": 1.0847234776892103, "grad_norm": 1.0914236307144165, "learning_rate": 9.135752549540704e-05, "loss": 0.6644, "step": 26400 }, { "epoch": 1.085545237899581, "grad_norm": 1.2880839109420776, "learning_rate": 9.122495506518254e-05, "loss": 0.6699, "step": 26420 }, { "epoch": 1.0863669981099515, "grad_norm": 1.3740962743759155, "learning_rate": 9.10924001748872e-05, "loss": 0.661, "step": 26440 }, { "epoch": 1.0871887583203221, "grad_norm": 1.2001346349716187, "learning_rate": 9.095986105926547e-05, "loss": 0.6806, "step": 26460 }, { "epoch": 1.0880105185306927, "grad_norm": 1.1882902383804321, "learning_rate": 9.082733795303393e-05, "loss": 0.6581, "step": 26480 }, { "epoch": 1.0888322787410634, "grad_norm": 1.1991751194000244, "learning_rate": 9.069483109088086e-05, "loss": 0.6684, "step": 26500 }, { "epoch": 1.089654038951434, "grad_norm": 1.2505358457565308, "learning_rate": 9.056234070746562e-05, "loss": 0.6719, "step": 26520 }, { "epoch": 1.0904757991618046, "grad_norm": 1.1861186027526855, "learning_rate": 9.042986703741853e-05, "loss": 0.6723, "step": 26540 }, { "epoch": 1.0912975593721752, "grad_norm": 1.2916886806488037, "learning_rate": 9.029741031534025e-05, "loss": 0.6649, "step": 26560 }, { "epoch": 1.0921193195825458, "grad_norm": 1.1472935676574707, "learning_rate": 9.016497077580147e-05, "loss": 0.6717, "step": 26580 }, { "epoch": 1.0929410797929164, "grad_norm": 1.1004657745361328, "learning_rate": 9.00325486533424e-05, "loss": 0.659, "step": 26600 }, { "epoch": 1.093762840003287, "grad_norm": 1.374177098274231, "learning_rate": 8.990014418247242e-05, "loss": 0.663, "step": 26620 }, { "epoch": 1.0945846002136577, "grad_norm": 1.1015247106552124, "learning_rate": 8.976775759766963e-05, "loss": 0.6565, "step": 26640 }, { "epoch": 1.0954063604240283, "grad_norm": 1.4717910289764404, "learning_rate": 8.963538913338051e-05, "loss": 0.6628, "step": 26660 }, { "epoch": 1.096228120634399, "grad_norm": 1.180732011795044, "learning_rate": 8.950303902401942e-05, "loss": 0.6674, "step": 26680 }, { "epoch": 1.0970498808447695, "grad_norm": 1.168150782585144, "learning_rate": 8.937070750396826e-05, "loss": 0.686, "step": 26700 }, { "epoch": 1.0978716410551401, "grad_norm": 1.3096245527267456, "learning_rate": 8.923839480757589e-05, "loss": 0.6577, "step": 26720 }, { "epoch": 1.0986934012655107, "grad_norm": 1.1966016292572021, "learning_rate": 8.910610116915797e-05, "loss": 0.6614, "step": 26740 }, { "epoch": 1.0995151614758814, "grad_norm": 1.2684301137924194, "learning_rate": 8.897382682299628e-05, "loss": 0.6696, "step": 26760 }, { "epoch": 1.100336921686252, "grad_norm": 1.1399879455566406, "learning_rate": 8.884157200333856e-05, "loss": 0.6443, "step": 26780 }, { "epoch": 1.1011586818966226, "grad_norm": 1.2215358018875122, "learning_rate": 8.870933694439789e-05, "loss": 0.6749, "step": 26800 }, { "epoch": 1.1019804421069932, "grad_norm": 1.1253427267074585, "learning_rate": 8.857712188035245e-05, "loss": 0.6739, "step": 26820 }, { "epoch": 1.1028022023173638, "grad_norm": 1.0373494625091553, "learning_rate": 8.844492704534485e-05, "loss": 0.6877, "step": 26840 }, { "epoch": 1.1036239625277344, "grad_norm": 1.2346014976501465, "learning_rate": 8.831275267348199e-05, "loss": 0.6754, "step": 26860 }, { "epoch": 1.104445722738105, "grad_norm": 0.9985619187355042, "learning_rate": 8.818059899883456e-05, "loss": 0.6806, "step": 26880 }, { "epoch": 1.1052674829484757, "grad_norm": 1.1723840236663818, "learning_rate": 8.804846625543646e-05, "loss": 0.6757, "step": 26900 }, { "epoch": 1.1060892431588463, "grad_norm": 1.1545828580856323, "learning_rate": 8.791635467728471e-05, "loss": 0.6533, "step": 26920 }, { "epoch": 1.1069110033692169, "grad_norm": 1.0837960243225098, "learning_rate": 8.778426449833865e-05, "loss": 0.6614, "step": 26940 }, { "epoch": 1.1077327635795875, "grad_norm": 1.1886368989944458, "learning_rate": 8.765219595251986e-05, "loss": 0.6571, "step": 26960 }, { "epoch": 1.1085545237899581, "grad_norm": 1.2560697793960571, "learning_rate": 8.753335295091204e-05, "loss": 0.6819, "step": 26980 }, { "epoch": 1.1093762840003287, "grad_norm": 1.0834345817565918, "learning_rate": 8.74013261523514e-05, "loss": 0.6838, "step": 27000 }, { "epoch": 1.1101980442106993, "grad_norm": 1.2474608421325684, "learning_rate": 8.72693216650723e-05, "loss": 0.6479, "step": 27020 }, { "epoch": 1.11101980442107, "grad_norm": 1.2082535028457642, "learning_rate": 8.713733972284461e-05, "loss": 0.6479, "step": 27040 }, { "epoch": 1.1118415646314406, "grad_norm": 1.2836271524429321, "learning_rate": 8.700538055939816e-05, "loss": 0.6927, "step": 27060 }, { "epoch": 1.1126633248418112, "grad_norm": 1.235249400138855, "learning_rate": 8.687344440842249e-05, "loss": 0.6672, "step": 27080 }, { "epoch": 1.1134850850521818, "grad_norm": 1.1937826871871948, "learning_rate": 8.67415315035664e-05, "loss": 0.6775, "step": 27100 }, { "epoch": 1.1143068452625524, "grad_norm": 1.210194706916809, "learning_rate": 8.66096420784374e-05, "loss": 0.6656, "step": 27120 }, { "epoch": 1.115128605472923, "grad_norm": 1.1403838396072388, "learning_rate": 8.647777636660159e-05, "loss": 0.6406, "step": 27140 }, { "epoch": 1.1159503656832936, "grad_norm": 1.0807080268859863, "learning_rate": 8.634593460158293e-05, "loss": 0.6712, "step": 27160 }, { "epoch": 1.1167721258936643, "grad_norm": 1.2173619270324707, "learning_rate": 8.621411701686309e-05, "loss": 0.6671, "step": 27180 }, { "epoch": 1.1175938861040349, "grad_norm": 1.2164521217346191, "learning_rate": 8.608232384588086e-05, "loss": 0.6551, "step": 27200 }, { "epoch": 1.1184156463144055, "grad_norm": 1.3138072490692139, "learning_rate": 8.595055532203177e-05, "loss": 0.681, "step": 27220 }, { "epoch": 1.119237406524776, "grad_norm": 1.1198703050613403, "learning_rate": 8.581881167866774e-05, "loss": 0.6533, "step": 27240 }, { "epoch": 1.1200591667351467, "grad_norm": 1.1022275686264038, "learning_rate": 8.568709314909663e-05, "loss": 0.6441, "step": 27260 }, { "epoch": 1.1208809269455173, "grad_norm": 1.2056939601898193, "learning_rate": 8.555539996658184e-05, "loss": 0.6768, "step": 27280 }, { "epoch": 1.121702687155888, "grad_norm": 1.2110202312469482, "learning_rate": 8.542373236434182e-05, "loss": 0.6646, "step": 27300 }, { "epoch": 1.1225244473662586, "grad_norm": 1.0746920108795166, "learning_rate": 8.529209057554984e-05, "loss": 0.679, "step": 27320 }, { "epoch": 1.1233462075766292, "grad_norm": 1.1810733079910278, "learning_rate": 8.516047483333325e-05, "loss": 0.6643, "step": 27340 }, { "epoch": 1.1241679677869998, "grad_norm": 1.1760369539260864, "learning_rate": 8.502888537077345e-05, "loss": 0.6543, "step": 27360 }, { "epoch": 1.1249897279973704, "grad_norm": 1.2052829265594482, "learning_rate": 8.489732242090527e-05, "loss": 0.6448, "step": 27380 }, { "epoch": 1.125811488207741, "grad_norm": 1.2746154069900513, "learning_rate": 8.476578621671659e-05, "loss": 0.6622, "step": 27400 }, { "epoch": 1.1266332484181116, "grad_norm": 1.186718225479126, "learning_rate": 8.463427699114785e-05, "loss": 0.6626, "step": 27420 }, { "epoch": 1.1274550086284822, "grad_norm": 1.299566388130188, "learning_rate": 8.450279497709179e-05, "loss": 0.677, "step": 27440 }, { "epoch": 1.1282767688388529, "grad_norm": 1.2594420909881592, "learning_rate": 8.43713404073929e-05, "loss": 0.6743, "step": 27460 }, { "epoch": 1.1290985290492235, "grad_norm": 1.1725876331329346, "learning_rate": 8.423991351484716e-05, "loss": 0.6861, "step": 27480 }, { "epoch": 1.129920289259594, "grad_norm": 1.1445516347885132, "learning_rate": 8.410851453220143e-05, "loss": 0.6626, "step": 27500 }, { "epoch": 1.1307420494699647, "grad_norm": 1.0811272859573364, "learning_rate": 8.397714369215324e-05, "loss": 0.6527, "step": 27520 }, { "epoch": 1.1315638096803353, "grad_norm": 1.0390878915786743, "learning_rate": 8.384580122735024e-05, "loss": 0.6966, "step": 27540 }, { "epoch": 1.132385569890706, "grad_norm": 1.1482422351837158, "learning_rate": 8.371448737038976e-05, "loss": 0.6674, "step": 27560 }, { "epoch": 1.1332073301010765, "grad_norm": 1.1507283449172974, "learning_rate": 8.35832023538186e-05, "loss": 0.6776, "step": 27580 }, { "epoch": 1.1340290903114472, "grad_norm": 1.141131043434143, "learning_rate": 8.345194641013236e-05, "loss": 0.6824, "step": 27600 }, { "epoch": 1.1348508505218178, "grad_norm": 1.1057684421539307, "learning_rate": 8.332071977177526e-05, "loss": 0.687, "step": 27620 }, { "epoch": 1.1356726107321884, "grad_norm": 1.4246220588684082, "learning_rate": 8.318952267113958e-05, "loss": 0.6624, "step": 27640 }, { "epoch": 1.136494370942559, "grad_norm": 0.9534235000610352, "learning_rate": 8.305835534056524e-05, "loss": 0.668, "step": 27660 }, { "epoch": 1.1373161311529296, "grad_norm": 1.2411669492721558, "learning_rate": 8.29272180123395e-05, "loss": 0.6585, "step": 27680 }, { "epoch": 1.1381378913633002, "grad_norm": 1.137723684310913, "learning_rate": 8.279611091869651e-05, "loss": 0.6684, "step": 27700 }, { "epoch": 1.1389596515736709, "grad_norm": 1.1483657360076904, "learning_rate": 8.26650342918168e-05, "loss": 0.668, "step": 27720 }, { "epoch": 1.1397814117840415, "grad_norm": 1.1523141860961914, "learning_rate": 8.253398836382702e-05, "loss": 0.6957, "step": 27740 }, { "epoch": 1.140603171994412, "grad_norm": 1.1631639003753662, "learning_rate": 8.240297336679942e-05, "loss": 0.66, "step": 27760 }, { "epoch": 1.1414249322047827, "grad_norm": 1.2860993146896362, "learning_rate": 8.227198953275142e-05, "loss": 0.685, "step": 27780 }, { "epoch": 1.1422466924151533, "grad_norm": 1.236307144165039, "learning_rate": 8.214103709364535e-05, "loss": 0.663, "step": 27800 }, { "epoch": 1.143068452625524, "grad_norm": 1.1525200605392456, "learning_rate": 8.201011628138789e-05, "loss": 0.6581, "step": 27820 }, { "epoch": 1.1438902128358945, "grad_norm": 1.1676980257034302, "learning_rate": 8.187922732782976e-05, "loss": 0.6768, "step": 27840 }, { "epoch": 1.1447119730462652, "grad_norm": 1.0874487161636353, "learning_rate": 8.174837046476518e-05, "loss": 0.6336, "step": 27860 }, { "epoch": 1.1455337332566358, "grad_norm": 0.9278498291969299, "learning_rate": 8.16175459239316e-05, "loss": 0.6767, "step": 27880 }, { "epoch": 1.1463554934670064, "grad_norm": 1.039865255355835, "learning_rate": 8.148675393700918e-05, "loss": 0.6686, "step": 27900 }, { "epoch": 1.147177253677377, "grad_norm": 1.0158859491348267, "learning_rate": 8.135599473562048e-05, "loss": 0.685, "step": 27920 }, { "epoch": 1.1479990138877476, "grad_norm": 1.2587593793869019, "learning_rate": 8.122526855132997e-05, "loss": 0.6606, "step": 27940 }, { "epoch": 1.1488207740981182, "grad_norm": 1.2706636190414429, "learning_rate": 8.109457561564373e-05, "loss": 0.6761, "step": 27960 }, { "epoch": 1.1496425343084888, "grad_norm": 1.0990588665008545, "learning_rate": 8.09639161600088e-05, "loss": 0.6512, "step": 27980 }, { "epoch": 1.1504642945188595, "grad_norm": 1.1541229486465454, "learning_rate": 8.084635146662451e-05, "loss": 0.663, "step": 28000 }, { "epoch": 1.1504642945188595, "eval_loss": 0.9906555414199829, "eval_runtime": 16.5563, "eval_samples_per_second": 158.248, "eval_steps_per_second": 4.953, "step": 28000 }, { "epoch": 1.15128605472923, "grad_norm": 1.1189225912094116, "learning_rate": 8.071575626051133e-05, "loss": 0.6602, "step": 28020 }, { "epoch": 1.1521078149396007, "grad_norm": 1.2261697053909302, "learning_rate": 8.058519520530936e-05, "loss": 0.6556, "step": 28040 }, { "epoch": 1.1529295751499713, "grad_norm": 1.10615074634552, "learning_rate": 8.045466853223223e-05, "loss": 0.6632, "step": 28060 }, { "epoch": 1.153751335360342, "grad_norm": 1.2373408079147339, "learning_rate": 8.032417647243263e-05, "loss": 0.6697, "step": 28080 }, { "epoch": 1.1545730955707125, "grad_norm": 1.282421588897705, "learning_rate": 8.019371925700199e-05, "loss": 0.6827, "step": 28100 }, { "epoch": 1.1553948557810831, "grad_norm": 1.3164217472076416, "learning_rate": 8.006329711696991e-05, "loss": 0.6604, "step": 28120 }, { "epoch": 1.1562166159914538, "grad_norm": 1.0808019638061523, "learning_rate": 7.993291028330409e-05, "loss": 0.6437, "step": 28140 }, { "epoch": 1.1570383762018244, "grad_norm": 1.116471767425537, "learning_rate": 7.980255898690942e-05, "loss": 0.6599, "step": 28160 }, { "epoch": 1.157860136412195, "grad_norm": 1.13624906539917, "learning_rate": 7.967224345862811e-05, "loss": 0.6707, "step": 28180 }, { "epoch": 1.1586818966225656, "grad_norm": 1.1284352540969849, "learning_rate": 7.954196392923891e-05, "loss": 0.6627, "step": 28200 }, { "epoch": 1.1595036568329362, "grad_norm": 0.9575105905532837, "learning_rate": 7.941172062945683e-05, "loss": 0.669, "step": 28220 }, { "epoch": 1.1603254170433068, "grad_norm": 1.4857319593429565, "learning_rate": 7.928151378993277e-05, "loss": 0.684, "step": 28240 }, { "epoch": 1.1611471772536774, "grad_norm": 1.1151981353759766, "learning_rate": 7.915134364125295e-05, "loss": 0.6515, "step": 28260 }, { "epoch": 1.161968937464048, "grad_norm": 1.0837492942810059, "learning_rate": 7.902121041393876e-05, "loss": 0.6808, "step": 28280 }, { "epoch": 1.1627906976744187, "grad_norm": 1.3104796409606934, "learning_rate": 7.889111433844603e-05, "loss": 0.6816, "step": 28300 }, { "epoch": 1.1636124578847893, "grad_norm": 1.266835331916809, "learning_rate": 7.876105564516498e-05, "loss": 0.6494, "step": 28320 }, { "epoch": 1.16443421809516, "grad_norm": 1.1202675104141235, "learning_rate": 7.863103456441951e-05, "loss": 0.6517, "step": 28340 }, { "epoch": 1.1652559783055305, "grad_norm": 1.1493626832962036, "learning_rate": 7.850105132646699e-05, "loss": 0.6545, "step": 28360 }, { "epoch": 1.1660777385159011, "grad_norm": 1.259341835975647, "learning_rate": 7.837110616149767e-05, "loss": 0.6606, "step": 28380 }, { "epoch": 1.1668994987262717, "grad_norm": 1.210664987564087, "learning_rate": 7.824119929963444e-05, "loss": 0.645, "step": 28400 }, { "epoch": 1.1677212589366424, "grad_norm": 1.1816967725753784, "learning_rate": 7.81113309709324e-05, "loss": 0.6413, "step": 28420 }, { "epoch": 1.168543019147013, "grad_norm": 1.0502477884292603, "learning_rate": 7.79815014053783e-05, "loss": 0.6552, "step": 28440 }, { "epoch": 1.1693647793573836, "grad_norm": 1.1239842176437378, "learning_rate": 7.785171083289039e-05, "loss": 0.6106, "step": 28460 }, { "epoch": 1.1701865395677542, "grad_norm": 1.163809061050415, "learning_rate": 7.772195948331769e-05, "loss": 0.6831, "step": 28480 }, { "epoch": 1.1710082997781248, "grad_norm": 1.0698950290679932, "learning_rate": 7.759873224073584e-05, "loss": 0.6741, "step": 28500 }, { "epoch": 1.1718300599884954, "grad_norm": 1.140785574913025, "learning_rate": 7.7469058036688e-05, "loss": 0.6761, "step": 28520 }, { "epoch": 1.172651820198866, "grad_norm": 1.134072184562683, "learning_rate": 7.733942373320407e-05, "loss": 0.6785, "step": 28540 }, { "epoch": 1.1734735804092367, "grad_norm": 1.0939021110534668, "learning_rate": 7.720982955985641e-05, "loss": 0.6478, "step": 28560 }, { "epoch": 1.1742953406196073, "grad_norm": 1.268872618675232, "learning_rate": 7.708027574614631e-05, "loss": 0.6623, "step": 28580 }, { "epoch": 1.175117100829978, "grad_norm": 1.1508769989013672, "learning_rate": 7.695076252150361e-05, "loss": 0.6663, "step": 28600 }, { "epoch": 1.1759388610403485, "grad_norm": 1.2125996351242065, "learning_rate": 7.682129011528627e-05, "loss": 0.6473, "step": 28620 }, { "epoch": 1.1767606212507191, "grad_norm": 1.1959916353225708, "learning_rate": 7.669185875677998e-05, "loss": 0.6421, "step": 28640 }, { "epoch": 1.1775823814610897, "grad_norm": 1.1845420598983765, "learning_rate": 7.656246867519772e-05, "loss": 0.6756, "step": 28660 }, { "epoch": 1.1784041416714603, "grad_norm": 1.0986628532409668, "learning_rate": 7.643312009967928e-05, "loss": 0.6591, "step": 28680 }, { "epoch": 1.179225901881831, "grad_norm": 1.0492193698883057, "learning_rate": 7.630381325929113e-05, "loss": 0.6567, "step": 28700 }, { "epoch": 1.1800476620922016, "grad_norm": 1.1858165264129639, "learning_rate": 7.617454838302567e-05, "loss": 0.6644, "step": 28720 }, { "epoch": 1.1808694223025722, "grad_norm": 1.107640266418457, "learning_rate": 7.604532569980105e-05, "loss": 0.6514, "step": 28740 }, { "epoch": 1.1816911825129428, "grad_norm": 1.2059592008590698, "learning_rate": 7.591614543846075e-05, "loss": 0.6607, "step": 28760 }, { "epoch": 1.1825129427233134, "grad_norm": 1.0413446426391602, "learning_rate": 7.578700782777299e-05, "loss": 0.6722, "step": 28780 }, { "epoch": 1.183334702933684, "grad_norm": 1.18887197971344, "learning_rate": 7.565791309643058e-05, "loss": 0.6796, "step": 28800 }, { "epoch": 1.1841564631440546, "grad_norm": 1.1347589492797852, "learning_rate": 7.552886147305034e-05, "loss": 0.6652, "step": 28820 }, { "epoch": 1.1849782233544253, "grad_norm": 1.0669459104537964, "learning_rate": 7.53998531861728e-05, "loss": 0.6557, "step": 28840 }, { "epoch": 1.1857999835647959, "grad_norm": 1.1914751529693604, "learning_rate": 7.527088846426164e-05, "loss": 0.6631, "step": 28860 }, { "epoch": 1.1866217437751665, "grad_norm": 1.1654757261276245, "learning_rate": 7.514196753570354e-05, "loss": 0.6392, "step": 28880 }, { "epoch": 1.187443503985537, "grad_norm": 1.1120076179504395, "learning_rate": 7.501309062880745e-05, "loss": 0.6515, "step": 28900 }, { "epoch": 1.1882652641959077, "grad_norm": 1.3329730033874512, "learning_rate": 7.488425797180449e-05, "loss": 0.6495, "step": 28920 }, { "epoch": 1.1890870244062783, "grad_norm": 1.0386924743652344, "learning_rate": 7.475546979284738e-05, "loss": 0.6633, "step": 28940 }, { "epoch": 1.189908784616649, "grad_norm": 1.130223035812378, "learning_rate": 7.46267263200101e-05, "loss": 0.6608, "step": 28960 }, { "epoch": 1.1907305448270196, "grad_norm": 1.205684781074524, "learning_rate": 7.44980277812874e-05, "loss": 0.6873, "step": 28980 }, { "epoch": 1.1915523050373902, "grad_norm": 1.1943809986114502, "learning_rate": 7.436937440459448e-05, "loss": 0.6384, "step": 29000 }, { "epoch": 1.1923740652477608, "grad_norm": 1.1930698156356812, "learning_rate": 7.424076641776657e-05, "loss": 0.6562, "step": 29020 }, { "epoch": 1.1931958254581314, "grad_norm": 1.1172430515289307, "learning_rate": 7.411220404855852e-05, "loss": 0.6548, "step": 29040 }, { "epoch": 1.194017585668502, "grad_norm": 1.2052860260009766, "learning_rate": 7.398368752464438e-05, "loss": 0.6368, "step": 29060 }, { "epoch": 1.1948393458788726, "grad_norm": 1.2752981185913086, "learning_rate": 7.385521707361705e-05, "loss": 0.6425, "step": 29080 }, { "epoch": 1.1956611060892433, "grad_norm": 1.2396671772003174, "learning_rate": 7.372679292298781e-05, "loss": 0.6536, "step": 29100 }, { "epoch": 1.1964828662996139, "grad_norm": 1.0870732069015503, "learning_rate": 7.359841530018589e-05, "loss": 0.6926, "step": 29120 }, { "epoch": 1.1973046265099845, "grad_norm": 1.238871455192566, "learning_rate": 7.347008443255825e-05, "loss": 0.6733, "step": 29140 }, { "epoch": 1.198126386720355, "grad_norm": 0.9936127662658691, "learning_rate": 7.334180054736892e-05, "loss": 0.6675, "step": 29160 }, { "epoch": 1.1989481469307257, "grad_norm": 1.1829917430877686, "learning_rate": 7.321356387179881e-05, "loss": 0.6501, "step": 29180 }, { "epoch": 1.1997699071410963, "grad_norm": 1.218711256980896, "learning_rate": 7.308537463294525e-05, "loss": 0.6366, "step": 29200 }, { "epoch": 1.2005916673514667, "grad_norm": 1.1491694450378418, "learning_rate": 7.29572330578214e-05, "loss": 0.6563, "step": 29220 }, { "epoch": 1.2014134275618376, "grad_norm": 1.1249312162399292, "learning_rate": 7.28291393733562e-05, "loss": 0.6522, "step": 29240 }, { "epoch": 1.202235187772208, "grad_norm": 1.1913214921951294, "learning_rate": 7.270109380639374e-05, "loss": 0.6663, "step": 29260 }, { "epoch": 1.2030569479825788, "grad_norm": 1.059322714805603, "learning_rate": 7.257309658369278e-05, "loss": 0.6719, "step": 29280 }, { "epoch": 1.2038787081929492, "grad_norm": 1.2595094442367554, "learning_rate": 7.244514793192658e-05, "loss": 0.6652, "step": 29300 }, { "epoch": 1.20470046840332, "grad_norm": 1.1310940980911255, "learning_rate": 7.231724807768243e-05, "loss": 0.6286, "step": 29320 }, { "epoch": 1.2055222286136904, "grad_norm": 1.1567174196243286, "learning_rate": 7.2189397247461e-05, "loss": 0.6545, "step": 29340 }, { "epoch": 1.2063439888240612, "grad_norm": 1.2142544984817505, "learning_rate": 7.206159566767633e-05, "loss": 0.6717, "step": 29360 }, { "epoch": 1.2071657490344316, "grad_norm": 1.1306164264678955, "learning_rate": 7.193384356465518e-05, "loss": 0.6589, "step": 29380 }, { "epoch": 1.2079875092448025, "grad_norm": 1.1609306335449219, "learning_rate": 7.180614116463671e-05, "loss": 0.6393, "step": 29400 }, { "epoch": 1.2088092694551729, "grad_norm": 1.0231740474700928, "learning_rate": 7.167848869377201e-05, "loss": 0.6623, "step": 29420 }, { "epoch": 1.2096310296655437, "grad_norm": 1.1379834413528442, "learning_rate": 7.15508863781238e-05, "loss": 0.6451, "step": 29440 }, { "epoch": 1.210452789875914, "grad_norm": 1.1678614616394043, "learning_rate": 7.142333444366593e-05, "loss": 0.6275, "step": 29460 }, { "epoch": 1.211274550086285, "grad_norm": 1.2142267227172852, "learning_rate": 7.129583311628307e-05, "loss": 0.6486, "step": 29480 }, { "epoch": 1.2120963102966553, "grad_norm": 1.1522984504699707, "learning_rate": 7.116838262177025e-05, "loss": 0.6692, "step": 29500 }, { "epoch": 1.2129180705070262, "grad_norm": 1.2733930349349976, "learning_rate": 7.104098318583256e-05, "loss": 0.6467, "step": 29520 }, { "epoch": 1.2137398307173966, "grad_norm": 1.1900440454483032, "learning_rate": 7.09136350340845e-05, "loss": 0.6534, "step": 29540 }, { "epoch": 1.2145615909277674, "grad_norm": 1.1477757692337036, "learning_rate": 7.078633839204985e-05, "loss": 0.6601, "step": 29560 }, { "epoch": 1.2153833511381378, "grad_norm": 1.284244418144226, "learning_rate": 7.065909348516122e-05, "loss": 0.6659, "step": 29580 }, { "epoch": 1.2162051113485086, "grad_norm": 1.0393390655517578, "learning_rate": 7.05319005387595e-05, "loss": 0.6428, "step": 29600 }, { "epoch": 1.217026871558879, "grad_norm": 1.0445849895477295, "learning_rate": 7.040475977809362e-05, "loss": 0.6619, "step": 29620 }, { "epoch": 1.2178486317692498, "grad_norm": 1.2404284477233887, "learning_rate": 7.027767142832012e-05, "loss": 0.6424, "step": 29640 }, { "epoch": 1.2186703919796202, "grad_norm": 1.1194926500320435, "learning_rate": 7.015063571450262e-05, "loss": 0.6492, "step": 29660 }, { "epoch": 1.219492152189991, "grad_norm": 1.1228058338165283, "learning_rate": 7.00236528616116e-05, "loss": 0.6569, "step": 29680 }, { "epoch": 1.2203139124003615, "grad_norm": 1.106655240058899, "learning_rate": 6.989672309452398e-05, "loss": 0.6606, "step": 29700 }, { "epoch": 1.2211356726107323, "grad_norm": 1.1909033060073853, "learning_rate": 6.976984663802252e-05, "loss": 0.6867, "step": 29720 }, { "epoch": 1.2219574328211027, "grad_norm": 1.1197348833084106, "learning_rate": 6.964302371679578e-05, "loss": 0.65, "step": 29740 }, { "epoch": 1.2227791930314735, "grad_norm": 1.0467857122421265, "learning_rate": 6.951625455543724e-05, "loss": 0.6434, "step": 29760 }, { "epoch": 1.223600953241844, "grad_norm": 1.1986229419708252, "learning_rate": 6.938953937844541e-05, "loss": 0.6278, "step": 29780 }, { "epoch": 1.2244227134522148, "grad_norm": 1.1179389953613281, "learning_rate": 6.926287841022312e-05, "loss": 0.6584, "step": 29800 }, { "epoch": 1.2252444736625852, "grad_norm": 1.0440237522125244, "learning_rate": 6.913627187507716e-05, "loss": 0.6305, "step": 29820 }, { "epoch": 1.226066233872956, "grad_norm": 1.2614185810089111, "learning_rate": 6.9009719997218e-05, "loss": 0.6582, "step": 29840 }, { "epoch": 1.2268879940833264, "grad_norm": 1.0907703638076782, "learning_rate": 6.888322300075927e-05, "loss": 0.655, "step": 29860 }, { "epoch": 1.2277097542936972, "grad_norm": 1.0902374982833862, "learning_rate": 6.875678110971738e-05, "loss": 0.6338, "step": 29880 }, { "epoch": 1.2285315145040676, "grad_norm": 1.2563419342041016, "learning_rate": 6.863039454801119e-05, "loss": 0.6591, "step": 29900 }, { "epoch": 1.2293532747144384, "grad_norm": 1.1033470630645752, "learning_rate": 6.850406353946158e-05, "loss": 0.654, "step": 29920 }, { "epoch": 1.2301750349248088, "grad_norm": 1.1465051174163818, "learning_rate": 6.837778830779105e-05, "loss": 0.6621, "step": 29940 }, { "epoch": 1.2309967951351797, "grad_norm": 1.1070022583007812, "learning_rate": 6.825156907662336e-05, "loss": 0.6426, "step": 29960 }, { "epoch": 1.23181855534555, "grad_norm": 1.1090333461761475, "learning_rate": 6.812540606948296e-05, "loss": 0.6625, "step": 29980 }, { "epoch": 1.232640315555921, "grad_norm": 1.1961784362792969, "learning_rate": 6.799929950979487e-05, "loss": 0.6424, "step": 30000 }, { "epoch": 1.232640315555921, "eval_loss": 0.9679434895515442, "eval_runtime": 16.6198, "eval_samples_per_second": 157.644, "eval_steps_per_second": 4.934, "step": 30000 }, { "epoch": 1.2334620757662913, "grad_norm": 1.2960805892944336, "learning_rate": 6.787324962088411e-05, "loss": 0.6821, "step": 30020 }, { "epoch": 1.234283835976662, "grad_norm": 1.1521825790405273, "learning_rate": 6.77472566259753e-05, "loss": 0.6593, "step": 30040 }, { "epoch": 1.2351055961870325, "grad_norm": 1.1858956813812256, "learning_rate": 6.762132074819236e-05, "loss": 0.662, "step": 30060 }, { "epoch": 1.2359273563974031, "grad_norm": 1.1205803155899048, "learning_rate": 6.749544221055808e-05, "loss": 0.6815, "step": 30080 }, { "epoch": 1.2367491166077738, "grad_norm": 0.9257388710975647, "learning_rate": 6.736962123599355e-05, "loss": 0.6531, "step": 30100 }, { "epoch": 1.2375708768181444, "grad_norm": 0.9369202256202698, "learning_rate": 6.724385804731811e-05, "loss": 0.6305, "step": 30120 }, { "epoch": 1.238392637028515, "grad_norm": 1.1999167203903198, "learning_rate": 6.71181528672487e-05, "loss": 0.6468, "step": 30140 }, { "epoch": 1.2392143972388856, "grad_norm": 1.1772797107696533, "learning_rate": 6.699250591839946e-05, "loss": 0.654, "step": 30160 }, { "epoch": 1.2400361574492562, "grad_norm": 1.1733454465866089, "learning_rate": 6.68669174232815e-05, "loss": 0.6439, "step": 30180 }, { "epoch": 1.2408579176596268, "grad_norm": 1.0485635995864868, "learning_rate": 6.674138760430236e-05, "loss": 0.6505, "step": 30200 }, { "epoch": 1.2416796778699974, "grad_norm": 1.1465091705322266, "learning_rate": 6.66159166837657e-05, "loss": 0.6591, "step": 30220 }, { "epoch": 1.242501438080368, "grad_norm": 1.1369998455047607, "learning_rate": 6.649050488387086e-05, "loss": 0.6399, "step": 30240 }, { "epoch": 1.2433231982907387, "grad_norm": 1.2065318822860718, "learning_rate": 6.636515242671247e-05, "loss": 0.6695, "step": 30260 }, { "epoch": 1.2441449585011093, "grad_norm": 1.2178568840026855, "learning_rate": 6.62398595342801e-05, "loss": 0.6448, "step": 30280 }, { "epoch": 1.24496671871148, "grad_norm": 1.1589614152908325, "learning_rate": 6.611462642845782e-05, "loss": 0.6308, "step": 30300 }, { "epoch": 1.2457884789218505, "grad_norm": 1.136525273323059, "learning_rate": 6.598945333102384e-05, "loss": 0.6525, "step": 30320 }, { "epoch": 1.2466102391322211, "grad_norm": 1.2322630882263184, "learning_rate": 6.586434046365002e-05, "loss": 0.639, "step": 30340 }, { "epoch": 1.2474319993425917, "grad_norm": 1.171217679977417, "learning_rate": 6.573928804790165e-05, "loss": 0.6624, "step": 30360 }, { "epoch": 1.2482537595529624, "grad_norm": 1.0731596946716309, "learning_rate": 6.561429630523694e-05, "loss": 0.658, "step": 30380 }, { "epoch": 1.249075519763333, "grad_norm": 0.9198722839355469, "learning_rate": 6.548936545700665e-05, "loss": 0.656, "step": 30400 }, { "epoch": 1.2498972799737036, "grad_norm": 1.2230961322784424, "learning_rate": 6.53644957244537e-05, "loss": 0.6415, "step": 30420 }, { "epoch": 1.2507190401840742, "grad_norm": 1.3008650541305542, "learning_rate": 6.52396873287127e-05, "loss": 0.649, "step": 30440 }, { "epoch": 1.2515408003944448, "grad_norm": 1.1299681663513184, "learning_rate": 6.511494049080982e-05, "loss": 0.6632, "step": 30460 }, { "epoch": 1.2523625606048154, "grad_norm": 1.0429800748825073, "learning_rate": 6.499025543166205e-05, "loss": 0.6435, "step": 30480 }, { "epoch": 1.253184320815186, "grad_norm": 1.313433051109314, "learning_rate": 6.486563237207704e-05, "loss": 0.6573, "step": 30500 }, { "epoch": 1.2540060810255567, "grad_norm": 1.0738648176193237, "learning_rate": 6.47472980935819e-05, "loss": 0.6508, "step": 30520 }, { "epoch": 1.2548278412359273, "grad_norm": 1.1511154174804688, "learning_rate": 6.462279656782608e-05, "loss": 0.6244, "step": 30540 }, { "epoch": 1.255649601446298, "grad_norm": 1.1229740381240845, "learning_rate": 6.450457814479982e-05, "loss": 0.6303, "step": 30560 }, { "epoch": 1.2564713616566685, "grad_norm": 1.1822351217269897, "learning_rate": 6.438019899125807e-05, "loss": 0.6296, "step": 30580 }, { "epoch": 1.2572931218670391, "grad_norm": 1.127118706703186, "learning_rate": 6.425588291764203e-05, "loss": 0.6664, "step": 30600 }, { "epoch": 1.2581148820774097, "grad_norm": 1.275978684425354, "learning_rate": 6.413163014410595e-05, "loss": 0.651, "step": 30620 }, { "epoch": 1.2589366422877803, "grad_norm": 1.1102421283721924, "learning_rate": 6.400744089069191e-05, "loss": 0.6459, "step": 30640 }, { "epoch": 1.259758402498151, "grad_norm": 1.2421194314956665, "learning_rate": 6.388331537732954e-05, "loss": 0.6464, "step": 30660 }, { "epoch": 1.2605801627085216, "grad_norm": 1.248716950416565, "learning_rate": 6.37592538238356e-05, "loss": 0.6495, "step": 30680 }, { "epoch": 1.2614019229188922, "grad_norm": 1.144371509552002, "learning_rate": 6.363525644991348e-05, "loss": 0.6405, "step": 30700 }, { "epoch": 1.2622236831292628, "grad_norm": 1.2534958124160767, "learning_rate": 6.351132347515303e-05, "loss": 0.6581, "step": 30720 }, { "epoch": 1.2630454433396334, "grad_norm": 1.2671653032302856, "learning_rate": 6.338745511902997e-05, "loss": 0.6547, "step": 30740 }, { "epoch": 1.263867203550004, "grad_norm": 1.1555167436599731, "learning_rate": 6.326365160090566e-05, "loss": 0.6461, "step": 30760 }, { "epoch": 1.2646889637603747, "grad_norm": 1.1524670124053955, "learning_rate": 6.313991314002663e-05, "loss": 0.6622, "step": 30780 }, { "epoch": 1.2655107239707453, "grad_norm": 1.204883337020874, "learning_rate": 6.301623995552409e-05, "loss": 0.6624, "step": 30800 }, { "epoch": 1.2663324841811159, "grad_norm": 1.0530204772949219, "learning_rate": 6.289263226641375e-05, "loss": 0.6655, "step": 30820 }, { "epoch": 1.2671542443914865, "grad_norm": 1.1731172800064087, "learning_rate": 6.276909029159536e-05, "loss": 0.6524, "step": 30840 }, { "epoch": 1.267976004601857, "grad_norm": 1.1701477766036987, "learning_rate": 6.26456142498522e-05, "loss": 0.6265, "step": 30860 }, { "epoch": 1.2687977648122277, "grad_norm": 1.3820420503616333, "learning_rate": 6.252220435985088e-05, "loss": 0.6387, "step": 30880 }, { "epoch": 1.2696195250225983, "grad_norm": 1.1477141380310059, "learning_rate": 6.239886084014081e-05, "loss": 0.6521, "step": 30900 }, { "epoch": 1.270441285232969, "grad_norm": 1.2501583099365234, "learning_rate": 6.22755839091538e-05, "loss": 0.6452, "step": 30920 }, { "epoch": 1.2712630454433396, "grad_norm": 1.110160231590271, "learning_rate": 6.215237378520388e-05, "loss": 0.6287, "step": 30940 }, { "epoch": 1.2720848056537102, "grad_norm": 1.2295660972595215, "learning_rate": 6.202923068648665e-05, "loss": 0.6611, "step": 30960 }, { "epoch": 1.2729065658640808, "grad_norm": 1.2019611597061157, "learning_rate": 6.190615483107911e-05, "loss": 0.6363, "step": 30980 }, { "epoch": 1.2737283260744514, "grad_norm": 1.2087301015853882, "learning_rate": 6.17831464369391e-05, "loss": 0.6613, "step": 31000 }, { "epoch": 1.274550086284822, "grad_norm": 1.179447889328003, "learning_rate": 6.166020572190501e-05, "loss": 0.6497, "step": 31020 }, { "epoch": 1.2753718464951926, "grad_norm": 1.183194637298584, "learning_rate": 6.153733290369536e-05, "loss": 0.6464, "step": 31040 }, { "epoch": 1.2761936067055633, "grad_norm": 1.2545372247695923, "learning_rate": 6.14145281999085e-05, "loss": 0.6517, "step": 31060 }, { "epoch": 1.2770153669159339, "grad_norm": 1.143701434135437, "learning_rate": 6.129179182802208e-05, "loss": 0.629, "step": 31080 }, { "epoch": 1.2778371271263045, "grad_norm": 1.2404268980026245, "learning_rate": 6.116912400539277e-05, "loss": 0.6209, "step": 31100 }, { "epoch": 1.278658887336675, "grad_norm": 1.0369129180908203, "learning_rate": 6.104652494925587e-05, "loss": 0.637, "step": 31120 }, { "epoch": 1.2794806475470457, "grad_norm": 1.1535693407058716, "learning_rate": 6.092399487672482e-05, "loss": 0.6406, "step": 31140 }, { "epoch": 1.2803024077574163, "grad_norm": 1.143570899963379, "learning_rate": 6.0801534004791005e-05, "loss": 0.6631, "step": 31160 }, { "epoch": 1.281124167967787, "grad_norm": 1.2489653825759888, "learning_rate": 6.067914255032314e-05, "loss": 0.6433, "step": 31180 }, { "epoch": 1.2819459281781576, "grad_norm": 1.0466420650482178, "learning_rate": 6.05568207300671e-05, "loss": 0.6404, "step": 31200 }, { "epoch": 1.2827676883885282, "grad_norm": 1.2191749811172485, "learning_rate": 6.043456876064546e-05, "loss": 0.6197, "step": 31220 }, { "epoch": 1.2835894485988988, "grad_norm": 1.1350699663162231, "learning_rate": 6.0312386858556956e-05, "loss": 0.6605, "step": 31240 }, { "epoch": 1.2844112088092694, "grad_norm": 1.1058688163757324, "learning_rate": 6.0190275240176386e-05, "loss": 0.6577, "step": 31260 }, { "epoch": 1.28523296901964, "grad_norm": 1.0823215246200562, "learning_rate": 6.006823412175404e-05, "loss": 0.6389, "step": 31280 }, { "epoch": 1.2860547292300106, "grad_norm": 1.1849805116653442, "learning_rate": 5.9946263719415295e-05, "loss": 0.665, "step": 31300 }, { "epoch": 1.2868764894403812, "grad_norm": 1.1973915100097656, "learning_rate": 5.982436424916037e-05, "loss": 0.6612, "step": 31320 }, { "epoch": 1.2876982496507519, "grad_norm": 1.1198099851608276, "learning_rate": 5.970253592686389e-05, "loss": 0.6371, "step": 31340 }, { "epoch": 1.2885200098611225, "grad_norm": 0.9331278204917908, "learning_rate": 5.9580778968274354e-05, "loss": 0.6204, "step": 31360 }, { "epoch": 1.289341770071493, "grad_norm": 1.0500894784927368, "learning_rate": 5.9459093589014e-05, "loss": 0.6481, "step": 31380 }, { "epoch": 1.2901635302818637, "grad_norm": 1.1697520017623901, "learning_rate": 5.9337480004578285e-05, "loss": 0.6694, "step": 31400 }, { "epoch": 1.2909852904922343, "grad_norm": 1.1045079231262207, "learning_rate": 5.921593843033548e-05, "loss": 0.6437, "step": 31420 }, { "epoch": 1.291807050702605, "grad_norm": 1.0866926908493042, "learning_rate": 5.9094469081526396e-05, "loss": 0.6278, "step": 31440 }, { "epoch": 1.2926288109129755, "grad_norm": 1.1703931093215942, "learning_rate": 5.8973072173263865e-05, "loss": 0.6523, "step": 31460 }, { "epoch": 1.2934505711233462, "grad_norm": 1.2150297164916992, "learning_rate": 5.885174792053245e-05, "loss": 0.648, "step": 31480 }, { "epoch": 1.2942723313337168, "grad_norm": 1.2517179250717163, "learning_rate": 5.8730496538188084e-05, "loss": 0.6439, "step": 31500 }, { "epoch": 1.2950940915440874, "grad_norm": 1.0839895009994507, "learning_rate": 5.8609318240957635e-05, "loss": 0.6365, "step": 31520 }, { "epoch": 1.295915851754458, "grad_norm": 1.081815481185913, "learning_rate": 5.848821324343854e-05, "loss": 0.6467, "step": 31540 }, { "epoch": 1.2967376119648286, "grad_norm": 1.0636696815490723, "learning_rate": 5.836718176009848e-05, "loss": 0.6506, "step": 31560 }, { "epoch": 1.2975593721751992, "grad_norm": 1.1643255949020386, "learning_rate": 5.825227013865635e-05, "loss": 0.6489, "step": 31580 }, { "epoch": 1.2983811323855698, "grad_norm": 1.1235414743423462, "learning_rate": 5.813742524085972e-05, "loss": 0.652, "step": 31600 }, { "epoch": 1.2992028925959405, "grad_norm": 1.1215312480926514, "learning_rate": 5.801660816024896e-05, "loss": 0.6475, "step": 31620 }, { "epoch": 1.300024652806311, "grad_norm": 1.1635875701904297, "learning_rate": 5.7895865428993504e-05, "loss": 0.6325, "step": 31640 }, { "epoch": 1.3008464130166817, "grad_norm": 1.1717442274093628, "learning_rate": 5.777519726091938e-05, "loss": 0.6301, "step": 31660 }, { "epoch": 1.3016681732270523, "grad_norm": 1.0808964967727661, "learning_rate": 5.765460386972068e-05, "loss": 0.636, "step": 31680 }, { "epoch": 1.302489933437423, "grad_norm": 1.1285815238952637, "learning_rate": 5.7534085468958965e-05, "loss": 0.6405, "step": 31700 }, { "epoch": 1.3033116936477935, "grad_norm": 1.0786865949630737, "learning_rate": 5.7413642272063164e-05, "loss": 0.6418, "step": 31720 }, { "epoch": 1.3041334538581641, "grad_norm": 1.243355631828308, "learning_rate": 5.729327449232873e-05, "loss": 0.6563, "step": 31740 }, { "epoch": 1.3049552140685348, "grad_norm": 1.1835846900939941, "learning_rate": 5.717298234291786e-05, "loss": 0.6447, "step": 31760 }, { "epoch": 1.3057769742789054, "grad_norm": 1.239445686340332, "learning_rate": 5.705276603685868e-05, "loss": 0.6445, "step": 31780 }, { "epoch": 1.306598734489276, "grad_norm": 1.0167465209960938, "learning_rate": 5.693262578704492e-05, "loss": 0.6587, "step": 31800 }, { "epoch": 1.3074204946996466, "grad_norm": 1.2314865589141846, "learning_rate": 5.6812561806235855e-05, "loss": 0.6464, "step": 31820 }, { "epoch": 1.3082422549100172, "grad_norm": 1.1035876274108887, "learning_rate": 5.6692574307055476e-05, "loss": 0.6549, "step": 31840 }, { "epoch": 1.3090640151203878, "grad_norm": 1.2363945245742798, "learning_rate": 5.6572663501992416e-05, "loss": 0.6621, "step": 31860 }, { "epoch": 1.3098857753307585, "grad_norm": 0.9777959585189819, "learning_rate": 5.645282960339944e-05, "loss": 0.6054, "step": 31880 }, { "epoch": 1.310707535541129, "grad_norm": 1.1316789388656616, "learning_rate": 5.633307282349325e-05, "loss": 0.6553, "step": 31900 }, { "epoch": 1.3115292957514997, "grad_norm": 1.0634005069732666, "learning_rate": 5.6213393374353814e-05, "loss": 0.619, "step": 31920 }, { "epoch": 1.3123510559618703, "grad_norm": 1.0504931211471558, "learning_rate": 5.609379146792426e-05, "loss": 0.6504, "step": 31940 }, { "epoch": 1.313172816172241, "grad_norm": 1.1487746238708496, "learning_rate": 5.597426731601034e-05, "loss": 0.6555, "step": 31960 }, { "epoch": 1.3139945763826115, "grad_norm": 1.0530478954315186, "learning_rate": 5.585482113028009e-05, "loss": 0.6315, "step": 31980 }, { "epoch": 1.3148163365929821, "grad_norm": 1.161407709121704, "learning_rate": 5.5735453122263595e-05, "loss": 0.6467, "step": 32000 }, { "epoch": 1.3148163365929821, "eval_loss": 0.9416676163673401, "eval_runtime": 16.767, "eval_samples_per_second": 156.259, "eval_steps_per_second": 4.891, "step": 32000 }, { "epoch": 1.3156380968033528, "grad_norm": 1.104887843132019, "learning_rate": 5.5616163503352314e-05, "loss": 0.6367, "step": 32020 }, { "epoch": 1.3164598570137234, "grad_norm": 0.9899182319641113, "learning_rate": 5.5496952484799114e-05, "loss": 0.6521, "step": 32040 }, { "epoch": 1.317281617224094, "grad_norm": 1.3830389976501465, "learning_rate": 5.537782027771736e-05, "loss": 0.6269, "step": 32060 }, { "epoch": 1.3181033774344646, "grad_norm": 1.158103108406067, "learning_rate": 5.5258767093081165e-05, "loss": 0.6405, "step": 32080 }, { "epoch": 1.3189251376448352, "grad_norm": 1.2154991626739502, "learning_rate": 5.513979314172449e-05, "loss": 0.626, "step": 32100 }, { "epoch": 1.3197468978552058, "grad_norm": 1.2084523439407349, "learning_rate": 5.502089863434101e-05, "loss": 0.6438, "step": 32120 }, { "epoch": 1.3205686580655764, "grad_norm": 1.1323741674423218, "learning_rate": 5.490208378148385e-05, "loss": 0.6553, "step": 32140 }, { "epoch": 1.321390418275947, "grad_norm": 1.1183146238327026, "learning_rate": 5.478334879356488e-05, "loss": 0.6435, "step": 32160 }, { "epoch": 1.3222121784863177, "grad_norm": 1.035343885421753, "learning_rate": 5.466469388085467e-05, "loss": 0.6335, "step": 32180 }, { "epoch": 1.3230339386966883, "grad_norm": 0.9998947381973267, "learning_rate": 5.454611925348191e-05, "loss": 0.6473, "step": 32200 }, { "epoch": 1.323855698907059, "grad_norm": 1.1686418056488037, "learning_rate": 5.442762512143311e-05, "loss": 0.6328, "step": 32220 }, { "epoch": 1.3246774591174295, "grad_norm": 1.1200451850891113, "learning_rate": 5.4309211694552334e-05, "loss": 0.6295, "step": 32240 }, { "epoch": 1.3254992193278001, "grad_norm": 1.2187868356704712, "learning_rate": 5.41908791825406e-05, "loss": 0.6223, "step": 32260 }, { "epoch": 1.3263209795381707, "grad_norm": 1.0789508819580078, "learning_rate": 5.4072627794955697e-05, "loss": 0.6311, "step": 32280 }, { "epoch": 1.3271427397485414, "grad_norm": 1.2329927682876587, "learning_rate": 5.395445774121166e-05, "loss": 0.6376, "step": 32300 }, { "epoch": 1.327964499958912, "grad_norm": 1.2252501249313354, "learning_rate": 5.3836369230578665e-05, "loss": 0.6175, "step": 32320 }, { "epoch": 1.3287862601692826, "grad_norm": 1.0544497966766357, "learning_rate": 5.371836247218232e-05, "loss": 0.6348, "step": 32340 }, { "epoch": 1.3296080203796532, "grad_norm": 1.1359783411026, "learning_rate": 5.360043767500348e-05, "loss": 0.6333, "step": 32360 }, { "epoch": 1.3304297805900238, "grad_norm": 1.1277779340744019, "learning_rate": 5.3482595047878004e-05, "loss": 0.64, "step": 32380 }, { "epoch": 1.3312515408003944, "grad_norm": 1.1540424823760986, "learning_rate": 5.3364834799495934e-05, "loss": 0.6403, "step": 32400 }, { "epoch": 1.332073301010765, "grad_norm": 1.007522702217102, "learning_rate": 5.324715713840174e-05, "loss": 0.6374, "step": 32420 }, { "epoch": 1.3328950612211357, "grad_norm": 1.1693965196609497, "learning_rate": 5.3129562272993437e-05, "loss": 0.6387, "step": 32440 }, { "epoch": 1.3337168214315063, "grad_norm": 1.1591569185256958, "learning_rate": 5.301205041152253e-05, "loss": 0.6613, "step": 32460 }, { "epoch": 1.3345385816418769, "grad_norm": 1.1624069213867188, "learning_rate": 5.2900491215067996e-05, "loss": 0.6349, "step": 32480 }, { "epoch": 1.3353603418522475, "grad_norm": 1.181348204612732, "learning_rate": 5.2783141809701195e-05, "loss": 0.6565, "step": 32500 }, { "epoch": 1.3361821020626181, "grad_norm": 1.1376749277114868, "learning_rate": 5.266587602175571e-05, "loss": 0.642, "step": 32520 }, { "epoch": 1.3370038622729887, "grad_norm": 1.161271572113037, "learning_rate": 5.25486940589003e-05, "loss": 0.6378, "step": 32540 }, { "epoch": 1.3378256224833593, "grad_norm": 1.0550984144210815, "learning_rate": 5.243159612865513e-05, "loss": 0.6456, "step": 32560 }, { "epoch": 1.33864738269373, "grad_norm": 1.299521803855896, "learning_rate": 5.2314582438391666e-05, "loss": 0.6495, "step": 32580 }, { "epoch": 1.3394691429041006, "grad_norm": 1.115898847579956, "learning_rate": 5.2197653195332094e-05, "loss": 0.6439, "step": 32600 }, { "epoch": 1.3402909031144712, "grad_norm": 1.1047664880752563, "learning_rate": 5.208080860654916e-05, "loss": 0.6428, "step": 32620 }, { "epoch": 1.3411126633248418, "grad_norm": 1.0742267370224, "learning_rate": 5.196404887896562e-05, "loss": 0.633, "step": 32640 }, { "epoch": 1.3419344235352124, "grad_norm": 1.079347014427185, "learning_rate": 5.18473742193539e-05, "loss": 0.628, "step": 32660 }, { "epoch": 1.342756183745583, "grad_norm": 1.0724397897720337, "learning_rate": 5.1730784834336e-05, "loss": 0.6421, "step": 32680 }, { "epoch": 1.3435779439559536, "grad_norm": 1.193382740020752, "learning_rate": 5.161428093038255e-05, "loss": 0.6263, "step": 32700 }, { "epoch": 1.3443997041663243, "grad_norm": 1.219397783279419, "learning_rate": 5.149786271381314e-05, "loss": 0.637, "step": 32720 }, { "epoch": 1.3452214643766949, "grad_norm": 1.2231544256210327, "learning_rate": 5.1381530390795365e-05, "loss": 0.6491, "step": 32740 }, { "epoch": 1.3460432245870655, "grad_norm": 1.1197011470794678, "learning_rate": 5.1265284167344906e-05, "loss": 0.6405, "step": 32760 }, { "epoch": 1.346864984797436, "grad_norm": 1.1939942836761475, "learning_rate": 5.114912424932485e-05, "loss": 0.6389, "step": 32780 }, { "epoch": 1.3476867450078067, "grad_norm": 1.029039740562439, "learning_rate": 5.103305084244545e-05, "loss": 0.6434, "step": 32800 }, { "epoch": 1.3485085052181773, "grad_norm": 1.1776632070541382, "learning_rate": 5.0917064152263804e-05, "loss": 0.6287, "step": 32820 }, { "epoch": 1.349330265428548, "grad_norm": 1.1028509140014648, "learning_rate": 5.080116438418334e-05, "loss": 0.6407, "step": 32840 }, { "epoch": 1.3501520256389186, "grad_norm": 1.1051464080810547, "learning_rate": 5.068535174345373e-05, "loss": 0.6248, "step": 32860 }, { "epoch": 1.3509737858492892, "grad_norm": 1.0466769933700562, "learning_rate": 5.056962643517014e-05, "loss": 0.6292, "step": 32880 }, { "epoch": 1.3517955460596598, "grad_norm": 1.222561001777649, "learning_rate": 5.045398866427331e-05, "loss": 0.6223, "step": 32900 }, { "epoch": 1.3526173062700304, "grad_norm": 1.2215903997421265, "learning_rate": 5.03384386355487e-05, "loss": 0.6381, "step": 32920 }, { "epoch": 1.353439066480401, "grad_norm": 1.1650848388671875, "learning_rate": 5.02229765536266e-05, "loss": 0.6543, "step": 32940 }, { "epoch": 1.3542608266907716, "grad_norm": 1.1978867053985596, "learning_rate": 5.010760262298145e-05, "loss": 0.6255, "step": 32960 }, { "epoch": 1.3550825869011422, "grad_norm": 1.2988346815109253, "learning_rate": 4.999231704793156e-05, "loss": 0.6388, "step": 32980 }, { "epoch": 1.3559043471115129, "grad_norm": 0.9702379107475281, "learning_rate": 4.987712003263892e-05, "loss": 0.6343, "step": 33000 }, { "epoch": 1.3567261073218835, "grad_norm": 1.037627100944519, "learning_rate": 4.976201178110843e-05, "loss": 0.6351, "step": 33020 }, { "epoch": 1.357547867532254, "grad_norm": 1.2792913913726807, "learning_rate": 4.964699249718805e-05, "loss": 0.6324, "step": 33040 }, { "epoch": 1.3583696277426247, "grad_norm": 1.1782094240188599, "learning_rate": 4.953206238456804e-05, "loss": 0.6144, "step": 33060 }, { "epoch": 1.3591913879529953, "grad_norm": 1.1542799472808838, "learning_rate": 4.941722164678074e-05, "loss": 0.6236, "step": 33080 }, { "epoch": 1.360013148163366, "grad_norm": 1.2403596639633179, "learning_rate": 4.930247048720035e-05, "loss": 0.6325, "step": 33100 }, { "epoch": 1.3608349083737366, "grad_norm": 1.1392772197723389, "learning_rate": 4.918780910904229e-05, "loss": 0.6363, "step": 33120 }, { "epoch": 1.3616566685841072, "grad_norm": 1.0998027324676514, "learning_rate": 4.907323771536304e-05, "loss": 0.6461, "step": 33140 }, { "epoch": 1.3624784287944778, "grad_norm": 1.139052391052246, "learning_rate": 4.895875650905967e-05, "loss": 0.6196, "step": 33160 }, { "epoch": 1.3633001890048484, "grad_norm": 1.0922917127609253, "learning_rate": 4.884436569286968e-05, "loss": 0.6418, "step": 33180 }, { "epoch": 1.364121949215219, "grad_norm": 1.2731437683105469, "learning_rate": 4.8730065469370345e-05, "loss": 0.6584, "step": 33200 }, { "epoch": 1.3649437094255896, "grad_norm": 1.07999849319458, "learning_rate": 4.861585604097857e-05, "loss": 0.6337, "step": 33220 }, { "epoch": 1.3657654696359602, "grad_norm": 1.270757794380188, "learning_rate": 4.8501737609950456e-05, "loss": 0.6189, "step": 33240 }, { "epoch": 1.3665872298463309, "grad_norm": 1.1716457605361938, "learning_rate": 4.8387710378380925e-05, "loss": 0.6443, "step": 33260 }, { "epoch": 1.3674089900567015, "grad_norm": 1.2344413995742798, "learning_rate": 4.827377454820351e-05, "loss": 0.648, "step": 33280 }, { "epoch": 1.368230750267072, "grad_norm": 1.134440541267395, "learning_rate": 4.815993032118972e-05, "loss": 0.6276, "step": 33300 }, { "epoch": 1.3690525104774427, "grad_norm": 1.1542905569076538, "learning_rate": 4.804617789894898e-05, "loss": 0.632, "step": 33320 }, { "epoch": 1.3698742706878133, "grad_norm": 1.0746182203292847, "learning_rate": 4.7932517482928044e-05, "loss": 0.6425, "step": 33340 }, { "epoch": 1.370696030898184, "grad_norm": 1.2313367128372192, "learning_rate": 4.7818949274410755e-05, "loss": 0.6359, "step": 33360 }, { "epoch": 1.3715177911085545, "grad_norm": 1.1803609132766724, "learning_rate": 4.770547347451767e-05, "loss": 0.6255, "step": 33380 }, { "epoch": 1.3723395513189252, "grad_norm": 1.082554578781128, "learning_rate": 4.759209028420567e-05, "loss": 0.6194, "step": 33400 }, { "epoch": 1.3731613115292958, "grad_norm": 1.1176568269729614, "learning_rate": 4.747879990426772e-05, "loss": 0.626, "step": 33420 }, { "epoch": 1.3739830717396664, "grad_norm": 1.2282779216766357, "learning_rate": 4.736560253533233e-05, "loss": 0.6334, "step": 33440 }, { "epoch": 1.374804831950037, "grad_norm": 1.1532173156738281, "learning_rate": 4.725249837786333e-05, "loss": 0.6206, "step": 33460 }, { "epoch": 1.3756265921604076, "grad_norm": 1.1878107786178589, "learning_rate": 4.713948763215943e-05, "loss": 0.6585, "step": 33480 }, { "epoch": 1.3764483523707782, "grad_norm": 1.0733743906021118, "learning_rate": 4.7026570498354036e-05, "loss": 0.6409, "step": 33500 }, { "epoch": 1.3772701125811488, "grad_norm": 1.0726110935211182, "learning_rate": 4.691374717641468e-05, "loss": 0.6188, "step": 33520 }, { "epoch": 1.3780918727915195, "grad_norm": 1.0967559814453125, "learning_rate": 4.6801017866142716e-05, "loss": 0.624, "step": 33540 }, { "epoch": 1.37891363300189, "grad_norm": 1.1548309326171875, "learning_rate": 4.668838276717321e-05, "loss": 0.6432, "step": 33560 }, { "epoch": 1.3797353932122607, "grad_norm": 1.1636457443237305, "learning_rate": 4.6575842078974096e-05, "loss": 0.6232, "step": 33580 }, { "epoch": 1.3805571534226313, "grad_norm": 1.2843453884124756, "learning_rate": 4.6463396000846385e-05, "loss": 0.6349, "step": 33600 }, { "epoch": 1.381378913633002, "grad_norm": 1.131871223449707, "learning_rate": 4.635104473192334e-05, "loss": 0.6477, "step": 33620 }, { "epoch": 1.3822006738433725, "grad_norm": 1.2385400533676147, "learning_rate": 4.62387884711705e-05, "loss": 0.6233, "step": 33640 }, { "epoch": 1.3830224340537431, "grad_norm": 1.1033365726470947, "learning_rate": 4.612662741738501e-05, "loss": 0.6398, "step": 33660 }, { "epoch": 1.3838441942641138, "grad_norm": 1.232216477394104, "learning_rate": 4.6014561769195476e-05, "loss": 0.6356, "step": 33680 }, { "epoch": 1.3846659544744844, "grad_norm": 1.1851789951324463, "learning_rate": 4.5902591725061516e-05, "loss": 0.5995, "step": 33700 }, { "epoch": 1.385487714684855, "grad_norm": 1.0367991924285889, "learning_rate": 4.5796308916997966e-05, "loss": 0.6456, "step": 33720 }, { "epoch": 1.3863094748952256, "grad_norm": 1.2613193988800049, "learning_rate": 4.568452587095045e-05, "loss": 0.5976, "step": 33740 }, { "epoch": 1.3871312351055962, "grad_norm": 1.056667685508728, "learning_rate": 4.557283901342667e-05, "loss": 0.6341, "step": 33760 }, { "epoch": 1.3879529953159668, "grad_norm": 0.9964428544044495, "learning_rate": 4.546124854221549e-05, "loss": 0.6187, "step": 33780 }, { "epoch": 1.3887747555263374, "grad_norm": 1.2256003618240356, "learning_rate": 4.5349754654934994e-05, "loss": 0.6263, "step": 33800 }, { "epoch": 1.389596515736708, "grad_norm": 1.1462249755859375, "learning_rate": 4.523835754903235e-05, "loss": 0.6133, "step": 33820 }, { "epoch": 1.3904182759470787, "grad_norm": 1.1194911003112793, "learning_rate": 4.512705742178317e-05, "loss": 0.6437, "step": 33840 }, { "epoch": 1.3912400361574493, "grad_norm": 1.1602319478988647, "learning_rate": 4.501585447029154e-05, "loss": 0.6323, "step": 33860 }, { "epoch": 1.39206179636782, "grad_norm": 1.1044458150863647, "learning_rate": 4.490474889148918e-05, "loss": 0.6317, "step": 33880 }, { "epoch": 1.3928835565781905, "grad_norm": 1.1019541025161743, "learning_rate": 4.479374088213561e-05, "loss": 0.6154, "step": 33900 }, { "epoch": 1.3937053167885611, "grad_norm": 1.163619875907898, "learning_rate": 4.468283063881745e-05, "loss": 0.64, "step": 33920 }, { "epoch": 1.3945270769989317, "grad_norm": 1.2417570352554321, "learning_rate": 4.4572018357948163e-05, "loss": 0.6219, "step": 33940 }, { "epoch": 1.3953488372093024, "grad_norm": 1.0506731271743774, "learning_rate": 4.446130423576788e-05, "loss": 0.6288, "step": 33960 }, { "epoch": 1.396170597419673, "grad_norm": 1.1237679719924927, "learning_rate": 4.4350688468342625e-05, "loss": 0.6271, "step": 33980 }, { "epoch": 1.3969923576300436, "grad_norm": 1.254115343093872, "learning_rate": 4.424017125156454e-05, "loss": 0.616, "step": 34000 }, { "epoch": 1.3969923576300436, "eval_loss": 0.924524188041687, "eval_runtime": 16.5191, "eval_samples_per_second": 158.604, "eval_steps_per_second": 4.964, "step": 34000 }, { "epoch": 1.3978141178404142, "grad_norm": 1.0411611795425415, "learning_rate": 4.412975278115104e-05, "loss": 0.6529, "step": 34020 }, { "epoch": 1.3986358780507848, "grad_norm": 1.1783146858215332, "learning_rate": 4.401943325264478e-05, "loss": 0.6594, "step": 34040 }, { "epoch": 1.3994576382611554, "grad_norm": 1.1332125663757324, "learning_rate": 4.390921286141314e-05, "loss": 0.6194, "step": 34060 }, { "epoch": 1.400279398471526, "grad_norm": 1.1602998971939087, "learning_rate": 4.3799091802647954e-05, "loss": 0.6178, "step": 34080 }, { "epoch": 1.4011011586818967, "grad_norm": 1.1415718793869019, "learning_rate": 4.368907027136512e-05, "loss": 0.6115, "step": 34100 }, { "epoch": 1.4019229188922673, "grad_norm": 1.0940213203430176, "learning_rate": 4.3579148462404273e-05, "loss": 0.6163, "step": 34120 }, { "epoch": 1.402744679102638, "grad_norm": 1.0750993490219116, "learning_rate": 4.346932657042855e-05, "loss": 0.6368, "step": 34140 }, { "epoch": 1.4035664393130085, "grad_norm": 1.1632193326950073, "learning_rate": 4.335960478992399e-05, "loss": 0.6357, "step": 34160 }, { "epoch": 1.4043881995233791, "grad_norm": 0.9799935221672058, "learning_rate": 4.324998331519954e-05, "loss": 0.6241, "step": 34180 }, { "epoch": 1.4052099597337497, "grad_norm": 1.2228275537490845, "learning_rate": 4.314046234038624e-05, "loss": 0.6131, "step": 34200 }, { "epoch": 1.4060317199441204, "grad_norm": 1.1572198867797852, "learning_rate": 4.3031042059437423e-05, "loss": 0.6143, "step": 34220 }, { "epoch": 1.406853480154491, "grad_norm": 1.0190701484680176, "learning_rate": 4.292172266612794e-05, "loss": 0.6292, "step": 34240 }, { "epoch": 1.4076752403648616, "grad_norm": 1.032221794128418, "learning_rate": 4.2812504354053986e-05, "loss": 0.6137, "step": 34260 }, { "epoch": 1.4084970005752322, "grad_norm": 1.2286864519119263, "learning_rate": 4.270338731663285e-05, "loss": 0.6144, "step": 34280 }, { "epoch": 1.4093187607856028, "grad_norm": 1.0505238771438599, "learning_rate": 4.259437174710239e-05, "loss": 0.632, "step": 34300 }, { "epoch": 1.4101405209959734, "grad_norm": 1.101257562637329, "learning_rate": 4.248545783852077e-05, "loss": 0.6191, "step": 34320 }, { "epoch": 1.410962281206344, "grad_norm": 1.1456806659698486, "learning_rate": 4.237664578376611e-05, "loss": 0.6128, "step": 34340 }, { "epoch": 1.4117840414167147, "grad_norm": 1.186767339706421, "learning_rate": 4.226793577553626e-05, "loss": 0.6076, "step": 34360 }, { "epoch": 1.4126058016270853, "grad_norm": 1.081678867340088, "learning_rate": 4.215932800634823e-05, "loss": 0.5981, "step": 34380 }, { "epoch": 1.4134275618374559, "grad_norm": 1.1703848838806152, "learning_rate": 4.205082266853803e-05, "loss": 0.6327, "step": 34400 }, { "epoch": 1.4142493220478265, "grad_norm": 1.0651689767837524, "learning_rate": 4.194241995426025e-05, "loss": 0.6534, "step": 34420 }, { "epoch": 1.415071082258197, "grad_norm": 0.9564984440803528, "learning_rate": 4.183412005548771e-05, "loss": 0.6247, "step": 34440 }, { "epoch": 1.4158928424685677, "grad_norm": 1.0257432460784912, "learning_rate": 4.172592316401129e-05, "loss": 0.6221, "step": 34460 }, { "epoch": 1.4167146026789383, "grad_norm": 1.0298326015472412, "learning_rate": 4.1617829471439286e-05, "loss": 0.6157, "step": 34480 }, { "epoch": 1.417536362889309, "grad_norm": 1.0720484256744385, "learning_rate": 4.150983916919735e-05, "loss": 0.6086, "step": 34500 }, { "epoch": 1.4183581230996796, "grad_norm": 1.0548421144485474, "learning_rate": 4.140195244852797e-05, "loss": 0.6133, "step": 34520 }, { "epoch": 1.4191798833100502, "grad_norm": 1.1134650707244873, "learning_rate": 4.1294169500490244e-05, "loss": 0.613, "step": 34540 }, { "epoch": 1.4200016435204208, "grad_norm": 1.135111689567566, "learning_rate": 4.1186490515959456e-05, "loss": 0.6121, "step": 34560 }, { "epoch": 1.4208234037307914, "grad_norm": 1.080965280532837, "learning_rate": 4.107891568562675e-05, "loss": 0.6202, "step": 34580 }, { "epoch": 1.421645163941162, "grad_norm": 1.1884208917617798, "learning_rate": 4.097144519999898e-05, "loss": 0.64, "step": 34600 }, { "epoch": 1.4224669241515326, "grad_norm": 1.1158133745193481, "learning_rate": 4.086407924939803e-05, "loss": 0.6285, "step": 34620 }, { "epoch": 1.4232886843619033, "grad_norm": 1.0862083435058594, "learning_rate": 4.0756818023960765e-05, "loss": 0.6368, "step": 34640 }, { "epoch": 1.4241104445722739, "grad_norm": 1.0684436559677124, "learning_rate": 4.0649661713638544e-05, "loss": 0.6247, "step": 34660 }, { "epoch": 1.4249322047826445, "grad_norm": 1.1276649236679077, "learning_rate": 4.054261050819691e-05, "loss": 0.6464, "step": 34680 }, { "epoch": 1.425753964993015, "grad_norm": 1.162343144416809, "learning_rate": 4.043566459721537e-05, "loss": 0.6193, "step": 34700 }, { "epoch": 1.4265757252033857, "grad_norm": 0.9801756143569946, "learning_rate": 4.033416368328017e-05, "loss": 0.6304, "step": 34720 }, { "epoch": 1.4273974854137563, "grad_norm": 1.1881067752838135, "learning_rate": 4.022742364106714e-05, "loss": 0.6233, "step": 34740 }, { "epoch": 1.428219245624127, "grad_norm": 1.0229750871658325, "learning_rate": 4.012078945148589e-05, "loss": 0.6281, "step": 34760 }, { "epoch": 1.4290410058344976, "grad_norm": 1.2104301452636719, "learning_rate": 4.001426130337733e-05, "loss": 0.616, "step": 34780 }, { "epoch": 1.4298627660448682, "grad_norm": 1.0440176725387573, "learning_rate": 3.990783938539456e-05, "loss": 0.6205, "step": 34800 }, { "epoch": 1.4306845262552388, "grad_norm": 1.0981454849243164, "learning_rate": 3.98068371306235e-05, "loss": 0.6387, "step": 34820 }, { "epoch": 1.4315062864656094, "grad_norm": 1.1962534189224243, "learning_rate": 3.970062290328703e-05, "loss": 0.6029, "step": 34840 }, { "epoch": 1.43232804667598, "grad_norm": 1.0792949199676514, "learning_rate": 3.959451546150584e-05, "loss": 0.6446, "step": 34860 }, { "epoch": 1.4331498068863506, "grad_norm": 1.2667758464813232, "learning_rate": 3.948851499318811e-05, "loss": 0.6233, "step": 34880 }, { "epoch": 1.4339715670967212, "grad_norm": 0.9692990779876709, "learning_rate": 3.9382621686052454e-05, "loss": 0.6101, "step": 34900 }, { "epoch": 1.4347933273070919, "grad_norm": 1.0378152132034302, "learning_rate": 3.927683572762778e-05, "loss": 0.6223, "step": 34920 }, { "epoch": 1.4356150875174625, "grad_norm": 1.0101780891418457, "learning_rate": 3.9171157305252884e-05, "loss": 0.6109, "step": 34940 }, { "epoch": 1.436436847727833, "grad_norm": 1.1455594301223755, "learning_rate": 3.9065586606076064e-05, "loss": 0.6373, "step": 34960 }, { "epoch": 1.4372586079382037, "grad_norm": 0.9811462759971619, "learning_rate": 3.8960123817055e-05, "loss": 0.6111, "step": 34980 }, { "epoch": 1.4380803681485743, "grad_norm": 1.0883426666259766, "learning_rate": 3.8854769124956104e-05, "loss": 0.6136, "step": 35000 }, { "epoch": 1.438902128358945, "grad_norm": 1.2005125284194946, "learning_rate": 3.874952271635444e-05, "loss": 0.624, "step": 35020 }, { "epoch": 1.4397238885693155, "grad_norm": 1.1082879304885864, "learning_rate": 3.864438477763327e-05, "loss": 0.6224, "step": 35040 }, { "epoch": 1.4405456487796862, "grad_norm": 1.1191489696502686, "learning_rate": 3.8539355494983865e-05, "loss": 0.6174, "step": 35060 }, { "epoch": 1.4413674089900568, "grad_norm": 1.131309986114502, "learning_rate": 3.843443505440494e-05, "loss": 0.6212, "step": 35080 }, { "epoch": 1.4421891692004274, "grad_norm": 1.216101884841919, "learning_rate": 3.832962364170251e-05, "loss": 0.6228, "step": 35100 }, { "epoch": 1.443010929410798, "grad_norm": 1.043578863143921, "learning_rate": 3.82249214424896e-05, "loss": 0.6225, "step": 35120 }, { "epoch": 1.4438326896211686, "grad_norm": 1.0260852575302124, "learning_rate": 3.812032864218563e-05, "loss": 0.6249, "step": 35140 }, { "epoch": 1.4446544498315392, "grad_norm": 1.1600069999694824, "learning_rate": 3.8015845426016494e-05, "loss": 0.6072, "step": 35160 }, { "epoch": 1.4454762100419098, "grad_norm": 1.0437159538269043, "learning_rate": 3.7911471979013845e-05, "loss": 0.619, "step": 35180 }, { "epoch": 1.4462979702522805, "grad_norm": 1.074245572090149, "learning_rate": 3.78072084860151e-05, "loss": 0.6135, "step": 35200 }, { "epoch": 1.447119730462651, "grad_norm": 1.021990180015564, "learning_rate": 3.7703055131662854e-05, "loss": 0.614, "step": 35220 }, { "epoch": 1.4479414906730217, "grad_norm": 0.9545276165008545, "learning_rate": 3.759901210040466e-05, "loss": 0.5856, "step": 35240 }, { "epoch": 1.4487632508833923, "grad_norm": 1.1442620754241943, "learning_rate": 3.749507957649274e-05, "loss": 0.6096, "step": 35260 }, { "epoch": 1.449585011093763, "grad_norm": 1.0844465494155884, "learning_rate": 3.7391257743983554e-05, "loss": 0.6105, "step": 35280 }, { "epoch": 1.4504067713041335, "grad_norm": 1.13474702835083, "learning_rate": 3.728754678673762e-05, "loss": 0.6214, "step": 35300 }, { "epoch": 1.4512285315145041, "grad_norm": 1.0782667398452759, "learning_rate": 3.7183946888419066e-05, "loss": 0.6417, "step": 35320 }, { "epoch": 1.4520502917248748, "grad_norm": 1.2489984035491943, "learning_rate": 3.708045823249531e-05, "loss": 0.6105, "step": 35340 }, { "epoch": 1.4528720519352452, "grad_norm": 1.113853096961975, "learning_rate": 3.69770810022368e-05, "loss": 0.6243, "step": 35360 }, { "epoch": 1.453693812145616, "grad_norm": 1.1300607919692993, "learning_rate": 3.6873815380716624e-05, "loss": 0.6248, "step": 35380 }, { "epoch": 1.4545155723559864, "grad_norm": 1.0935343503952026, "learning_rate": 3.6770661550810316e-05, "loss": 0.6188, "step": 35400 }, { "epoch": 1.4553373325663572, "grad_norm": 1.2234795093536377, "learning_rate": 3.6667619695195285e-05, "loss": 0.6157, "step": 35420 }, { "epoch": 1.4561590927767276, "grad_norm": 1.050308346748352, "learning_rate": 3.656468999635085e-05, "loss": 0.6089, "step": 35440 }, { "epoch": 1.4569808529870985, "grad_norm": 1.116862416267395, "learning_rate": 3.646187263655745e-05, "loss": 0.6126, "step": 35460 }, { "epoch": 1.4578026131974688, "grad_norm": 1.2579872608184814, "learning_rate": 3.6359167797896795e-05, "loss": 0.6189, "step": 35480 }, { "epoch": 1.4586243734078397, "grad_norm": 1.2237604856491089, "learning_rate": 3.625657566225124e-05, "loss": 0.6248, "step": 35500 }, { "epoch": 1.45944613361821, "grad_norm": 1.2082823514938354, "learning_rate": 3.615409641130351e-05, "loss": 0.6277, "step": 35520 }, { "epoch": 1.460267893828581, "grad_norm": 1.17978835105896, "learning_rate": 3.605173022653653e-05, "loss": 0.6119, "step": 35540 }, { "epoch": 1.4610896540389513, "grad_norm": 1.1109099388122559, "learning_rate": 3.5949477289232914e-05, "loss": 0.6391, "step": 35560 }, { "epoch": 1.4619114142493221, "grad_norm": 1.109529733657837, "learning_rate": 3.5847337780474744e-05, "loss": 0.6211, "step": 35580 }, { "epoch": 1.4627331744596925, "grad_norm": 1.159828543663025, "learning_rate": 3.5745311881143196e-05, "loss": 0.6182, "step": 35600 }, { "epoch": 1.4635549346700634, "grad_norm": 1.047654390335083, "learning_rate": 3.564339977191834e-05, "loss": 0.6166, "step": 35620 }, { "epoch": 1.4643766948804338, "grad_norm": 1.009979248046875, "learning_rate": 3.554160163327864e-05, "loss": 0.6, "step": 35640 }, { "epoch": 1.4651984550908046, "grad_norm": 1.0264675617218018, "learning_rate": 3.543991764550079e-05, "loss": 0.6166, "step": 35660 }, { "epoch": 1.466020215301175, "grad_norm": 1.2139075994491577, "learning_rate": 3.533834798865927e-05, "loss": 0.6369, "step": 35680 }, { "epoch": 1.4668419755115458, "grad_norm": 1.1369953155517578, "learning_rate": 3.523689284262611e-05, "loss": 0.6183, "step": 35700 }, { "epoch": 1.4676637357219162, "grad_norm": 1.1860294342041016, "learning_rate": 3.5135552387070636e-05, "loss": 0.6152, "step": 35720 }, { "epoch": 1.468485495932287, "grad_norm": 1.0591294765472412, "learning_rate": 3.503432680145892e-05, "loss": 0.6152, "step": 35740 }, { "epoch": 1.4693072561426574, "grad_norm": 1.112328290939331, "learning_rate": 3.493321626505375e-05, "loss": 0.6106, "step": 35760 }, { "epoch": 1.4701290163530283, "grad_norm": 1.0694029331207275, "learning_rate": 3.483222095691406e-05, "loss": 0.609, "step": 35780 }, { "epoch": 1.4709507765633987, "grad_norm": 1.0409561395645142, "learning_rate": 3.4731341055894785e-05, "loss": 0.6265, "step": 35800 }, { "epoch": 1.4717725367737695, "grad_norm": 1.130654215812683, "learning_rate": 3.463057674064646e-05, "loss": 0.6058, "step": 35820 }, { "epoch": 1.47259429698414, "grad_norm": 1.0501320362091064, "learning_rate": 3.4534957865013894e-05, "loss": 0.588, "step": 35840 }, { "epoch": 1.4734160571945107, "grad_norm": 1.1128648519515991, "learning_rate": 3.443441945508704e-05, "loss": 0.6212, "step": 35860 }, { "epoch": 1.4742378174048811, "grad_norm": 1.1255168914794922, "learning_rate": 3.433399715675645e-05, "loss": 0.6282, "step": 35880 }, { "epoch": 1.475059577615252, "grad_norm": 1.3319727182388306, "learning_rate": 3.4238703683689755e-05, "loss": 0.6392, "step": 35900 }, { "epoch": 1.4758813378256224, "grad_norm": 1.1387158632278442, "learning_rate": 3.414351531414453e-05, "loss": 0.6058, "step": 35920 }, { "epoch": 1.4767030980359932, "grad_norm": 1.0003697872161865, "learning_rate": 3.404343074439131e-05, "loss": 0.6186, "step": 35940 }, { "epoch": 1.4775248582463636, "grad_norm": 1.2400519847869873, "learning_rate": 3.3943462978645225e-05, "loss": 0.6088, "step": 35960 }, { "epoch": 1.4783466184567344, "grad_norm": 1.048429250717163, "learning_rate": 3.384361219394153e-05, "loss": 0.6101, "step": 35980 }, { "epoch": 1.4791683786671048, "grad_norm": 1.1433545351028442, "learning_rate": 3.374387856710828e-05, "loss": 0.6407, "step": 36000 }, { "epoch": 1.4791683786671048, "eval_loss": 0.9071117639541626, "eval_runtime": 16.6266, "eval_samples_per_second": 157.579, "eval_steps_per_second": 4.932, "step": 36000 }, { "epoch": 1.4799901388774757, "grad_norm": 1.257107138633728, "learning_rate": 3.3644262274766024e-05, "loss": 0.6129, "step": 36020 }, { "epoch": 1.480811899087846, "grad_norm": 1.07807457447052, "learning_rate": 3.354476349332756e-05, "loss": 0.619, "step": 36040 }, { "epoch": 1.4816336592982169, "grad_norm": 1.1075770854949951, "learning_rate": 3.344538239899754e-05, "loss": 0.6148, "step": 36060 }, { "epoch": 1.4824554195085873, "grad_norm": 1.2136929035186768, "learning_rate": 3.3346119167772285e-05, "loss": 0.6148, "step": 36080 }, { "epoch": 1.4832771797189581, "grad_norm": 1.0873780250549316, "learning_rate": 3.3246973975439274e-05, "loss": 0.6213, "step": 36100 }, { "epoch": 1.4840989399293285, "grad_norm": 1.09003746509552, "learning_rate": 3.314794699757713e-05, "loss": 0.6244, "step": 36120 }, { "epoch": 1.4849207001396993, "grad_norm": 1.0032758712768555, "learning_rate": 3.3049038409554855e-05, "loss": 0.6265, "step": 36140 }, { "epoch": 1.4857424603500697, "grad_norm": 1.10032057762146, "learning_rate": 3.295024838653205e-05, "loss": 0.6375, "step": 36160 }, { "epoch": 1.4865642205604406, "grad_norm": 1.0320249795913696, "learning_rate": 3.2851577103458196e-05, "loss": 0.6101, "step": 36180 }, { "epoch": 1.487385980770811, "grad_norm": 1.2705104351043701, "learning_rate": 3.2753024735072534e-05, "loss": 0.6337, "step": 36200 }, { "epoch": 1.4882077409811818, "grad_norm": 1.1513526439666748, "learning_rate": 3.2654591455903774e-05, "loss": 0.6179, "step": 36220 }, { "epoch": 1.4890295011915522, "grad_norm": 1.231684923171997, "learning_rate": 3.2556277440269636e-05, "loss": 0.619, "step": 36240 }, { "epoch": 1.489851261401923, "grad_norm": 1.1558480262756348, "learning_rate": 3.2458082862276685e-05, "loss": 0.609, "step": 36260 }, { "epoch": 1.4906730216122934, "grad_norm": 1.005265235900879, "learning_rate": 3.236000789581992e-05, "loss": 0.5977, "step": 36280 }, { "epoch": 1.4914947818226643, "grad_norm": 1.1663181781768799, "learning_rate": 3.2262052714582635e-05, "loss": 0.6104, "step": 36300 }, { "epoch": 1.4923165420330347, "grad_norm": 1.2102375030517578, "learning_rate": 3.216421749203586e-05, "loss": 0.6201, "step": 36320 }, { "epoch": 1.4931383022434055, "grad_norm": 1.1391383409500122, "learning_rate": 3.206650240143827e-05, "loss": 0.6063, "step": 36340 }, { "epoch": 1.4939600624537759, "grad_norm": 1.1457056999206543, "learning_rate": 3.1968907615835756e-05, "loss": 0.6111, "step": 36360 }, { "epoch": 1.4947818226641467, "grad_norm": 1.105281114578247, "learning_rate": 3.187143330806114e-05, "loss": 0.6129, "step": 36380 }, { "epoch": 1.4956035828745171, "grad_norm": 1.1415950059890747, "learning_rate": 3.177407965073398e-05, "loss": 0.6049, "step": 36400 }, { "epoch": 1.496425343084888, "grad_norm": 1.0731744766235352, "learning_rate": 3.1676846816260044e-05, "loss": 0.6257, "step": 36420 }, { "epoch": 1.4972471032952583, "grad_norm": 1.0916352272033691, "learning_rate": 3.1579734976831265e-05, "loss": 0.6219, "step": 36440 }, { "epoch": 1.4980688635056292, "grad_norm": 1.1694599390029907, "learning_rate": 3.14827443044252e-05, "loss": 0.6068, "step": 36460 }, { "epoch": 1.4988906237159996, "grad_norm": 1.0736804008483887, "learning_rate": 3.1385874970804874e-05, "loss": 0.6128, "step": 36480 }, { "epoch": 1.4997123839263704, "grad_norm": 1.2031406164169312, "learning_rate": 3.12891271475184e-05, "loss": 0.6196, "step": 36500 }, { "epoch": 1.5005341441367408, "grad_norm": 1.1104577779769897, "learning_rate": 3.119250100589872e-05, "loss": 0.6187, "step": 36520 }, { "epoch": 1.5013559043471116, "grad_norm": 1.2821518182754517, "learning_rate": 3.109599671706335e-05, "loss": 0.6149, "step": 36540 }, { "epoch": 1.502177664557482, "grad_norm": 1.0993192195892334, "learning_rate": 3.09996144519139e-05, "loss": 0.5835, "step": 36560 }, { "epoch": 1.5029994247678529, "grad_norm": 1.0320968627929688, "learning_rate": 3.090335438113597e-05, "loss": 0.6023, "step": 36580 }, { "epoch": 1.5038211849782233, "grad_norm": 0.9988365769386292, "learning_rate": 3.08072166751987e-05, "loss": 0.6205, "step": 36600 }, { "epoch": 1.504642945188594, "grad_norm": 1.0576531887054443, "learning_rate": 3.071120150435462e-05, "loss": 0.6087, "step": 36620 }, { "epoch": 1.5054647053989645, "grad_norm": 1.0885626077651978, "learning_rate": 3.0615309038639186e-05, "loss": 0.6034, "step": 36640 }, { "epoch": 1.5062864656093353, "grad_norm": 1.0492252111434937, "learning_rate": 3.051953944787054e-05, "loss": 0.5865, "step": 36660 }, { "epoch": 1.5071082258197057, "grad_norm": 1.0457483530044556, "learning_rate": 3.0423892901649344e-05, "loss": 0.6097, "step": 36680 }, { "epoch": 1.5079299860300766, "grad_norm": 1.0300101041793823, "learning_rate": 3.032836956935814e-05, "loss": 0.5837, "step": 36700 }, { "epoch": 1.508751746240447, "grad_norm": 1.1736280918121338, "learning_rate": 3.023296962016151e-05, "loss": 0.5926, "step": 36720 }, { "epoch": 1.5095735064508178, "grad_norm": 1.0206154584884644, "learning_rate": 3.0137693223005335e-05, "loss": 0.6162, "step": 36740 }, { "epoch": 1.5103952666611882, "grad_norm": 1.1228256225585938, "learning_rate": 3.004254054661686e-05, "loss": 0.5999, "step": 36760 }, { "epoch": 1.511217026871559, "grad_norm": 1.0171713829040527, "learning_rate": 2.994751175950411e-05, "loss": 0.6092, "step": 36780 }, { "epoch": 1.5120387870819294, "grad_norm": 1.017462968826294, "learning_rate": 2.985260702995575e-05, "loss": 0.6201, "step": 36800 }, { "epoch": 1.5128605472923002, "grad_norm": 1.2459691762924194, "learning_rate": 2.9757826526040755e-05, "loss": 0.6189, "step": 36820 }, { "epoch": 1.5136823075026706, "grad_norm": 1.027414321899414, "learning_rate": 2.9663170415608078e-05, "loss": 0.6104, "step": 36840 }, { "epoch": 1.5145040677130415, "grad_norm": 1.0714311599731445, "learning_rate": 2.9568638866286458e-05, "loss": 0.6201, "step": 36860 }, { "epoch": 1.5153258279234119, "grad_norm": 1.1437265872955322, "learning_rate": 2.9474232045483952e-05, "loss": 0.6102, "step": 36880 }, { "epoch": 1.5161475881337827, "grad_norm": 1.2031759023666382, "learning_rate": 2.93799501203878e-05, "loss": 0.5923, "step": 36900 }, { "epoch": 1.516969348344153, "grad_norm": 0.992753267288208, "learning_rate": 2.928579325796401e-05, "loss": 0.6025, "step": 36920 }, { "epoch": 1.517791108554524, "grad_norm": 1.1709444522857666, "learning_rate": 2.9191761624957115e-05, "loss": 0.6125, "step": 36940 }, { "epoch": 1.5186128687648943, "grad_norm": 1.1847578287124634, "learning_rate": 2.909785538788995e-05, "loss": 0.6141, "step": 36960 }, { "epoch": 1.5194346289752652, "grad_norm": 1.032343864440918, "learning_rate": 2.900407471306319e-05, "loss": 0.6172, "step": 36980 }, { "epoch": 1.5202563891856355, "grad_norm": 1.2214970588684082, "learning_rate": 2.8910419766555275e-05, "loss": 0.618, "step": 37000 }, { "epoch": 1.5210781493960064, "grad_norm": 1.1395217180252075, "learning_rate": 2.881689071422179e-05, "loss": 0.6024, "step": 37020 }, { "epoch": 1.5218999096063768, "grad_norm": 1.1765543222427368, "learning_rate": 2.8723487721695562e-05, "loss": 0.6021, "step": 37040 }, { "epoch": 1.5227216698167476, "grad_norm": 1.1648917198181152, "learning_rate": 2.8630210954386082e-05, "loss": 0.6181, "step": 37060 }, { "epoch": 1.523543430027118, "grad_norm": 1.1894433498382568, "learning_rate": 2.853706057747929e-05, "loss": 0.6272, "step": 37080 }, { "epoch": 1.5243651902374888, "grad_norm": 1.1555765867233276, "learning_rate": 2.84440367559374e-05, "loss": 0.6187, "step": 37100 }, { "epoch": 1.5251869504478592, "grad_norm": 1.1109102964401245, "learning_rate": 2.8351139654498405e-05, "loss": 0.611, "step": 37120 }, { "epoch": 1.52600871065823, "grad_norm": 1.1215410232543945, "learning_rate": 2.8258369437675926e-05, "loss": 0.5798, "step": 37140 }, { "epoch": 1.5268304708686005, "grad_norm": 1.1160521507263184, "learning_rate": 2.816572626975884e-05, "loss": 0.6054, "step": 37160 }, { "epoch": 1.5276522310789713, "grad_norm": 1.0489914417266846, "learning_rate": 2.8073210314811126e-05, "loss": 0.6054, "step": 37180 }, { "epoch": 1.5284739912893417, "grad_norm": 1.0901767015457153, "learning_rate": 2.798082173667139e-05, "loss": 0.6074, "step": 37200 }, { "epoch": 1.5292957514997125, "grad_norm": 1.1401610374450684, "learning_rate": 2.7888560698952702e-05, "loss": 0.6087, "step": 37220 }, { "epoch": 1.530117511710083, "grad_norm": 1.136517882347107, "learning_rate": 2.7796427365042243e-05, "loss": 0.6071, "step": 37240 }, { "epoch": 1.5309392719204538, "grad_norm": 1.1060153245925903, "learning_rate": 2.770442189810103e-05, "loss": 0.6142, "step": 37260 }, { "epoch": 1.5317610321308242, "grad_norm": 1.029160737991333, "learning_rate": 2.7612544461063727e-05, "loss": 0.6187, "step": 37280 }, { "epoch": 1.532582792341195, "grad_norm": 1.1750712394714355, "learning_rate": 2.752079521663814e-05, "loss": 0.6107, "step": 37300 }, { "epoch": 1.5334045525515654, "grad_norm": 1.0686962604522705, "learning_rate": 2.7429174327305186e-05, "loss": 0.6217, "step": 37320 }, { "epoch": 1.5342263127619362, "grad_norm": 1.0369669198989868, "learning_rate": 2.7337681955318363e-05, "loss": 0.6108, "step": 37340 }, { "epoch": 1.5350480729723066, "grad_norm": 1.0347490310668945, "learning_rate": 2.724631826270362e-05, "loss": 0.6055, "step": 37360 }, { "epoch": 1.5358698331826774, "grad_norm": 1.0429108142852783, "learning_rate": 2.715508341125904e-05, "loss": 0.6004, "step": 37380 }, { "epoch": 1.5366915933930478, "grad_norm": 1.1569420099258423, "learning_rate": 2.7063977562554476e-05, "loss": 0.613, "step": 37400 }, { "epoch": 1.5375133536034187, "grad_norm": 1.1140472888946533, "learning_rate": 2.6973000877931443e-05, "loss": 0.6267, "step": 37420 }, { "epoch": 1.538335113813789, "grad_norm": 1.0750665664672852, "learning_rate": 2.6882153518502616e-05, "loss": 0.6119, "step": 37440 }, { "epoch": 1.53915687402416, "grad_norm": 1.0947927236557007, "learning_rate": 2.6791435645151675e-05, "loss": 0.5913, "step": 37460 }, { "epoch": 1.5399786342345303, "grad_norm": 1.1905947923660278, "learning_rate": 2.670084741853296e-05, "loss": 0.6153, "step": 37480 }, { "epoch": 1.5408003944449011, "grad_norm": 1.0356147289276123, "learning_rate": 2.661038899907129e-05, "loss": 0.6068, "step": 37500 }, { "epoch": 1.5416221546552715, "grad_norm": 1.0834916830062866, "learning_rate": 2.6520060546961566e-05, "loss": 0.6017, "step": 37520 }, { "epoch": 1.5424439148656424, "grad_norm": 1.0963350534439087, "learning_rate": 2.6429862222168467e-05, "loss": 0.6211, "step": 37540 }, { "epoch": 1.5432656750760128, "grad_norm": 1.0668399333953857, "learning_rate": 2.6339794184426393e-05, "loss": 0.6182, "step": 37560 }, { "epoch": 1.5440874352863836, "grad_norm": 1.169129490852356, "learning_rate": 2.6249856593238763e-05, "loss": 0.6078, "step": 37580 }, { "epoch": 1.544909195496754, "grad_norm": 1.1952544450759888, "learning_rate": 2.6160049607878234e-05, "loss": 0.6056, "step": 37600 }, { "epoch": 1.5457309557071248, "grad_norm": 1.1417872905731201, "learning_rate": 2.6070373387386005e-05, "loss": 0.6069, "step": 37620 }, { "epoch": 1.5465527159174952, "grad_norm": 1.152288556098938, "learning_rate": 2.5980828090571817e-05, "loss": 0.6084, "step": 37640 }, { "epoch": 1.547374476127866, "grad_norm": 1.1022766828536987, "learning_rate": 2.589141387601346e-05, "loss": 0.6145, "step": 37660 }, { "epoch": 1.5481962363382364, "grad_norm": 1.0760823488235474, "learning_rate": 2.580213090205663e-05, "loss": 0.5979, "step": 37680 }, { "epoch": 1.5490179965486073, "grad_norm": 1.1265369653701782, "learning_rate": 2.5712979326814613e-05, "loss": 0.6069, "step": 37700 }, { "epoch": 1.5498397567589777, "grad_norm": 1.0889846086502075, "learning_rate": 2.5623959308167945e-05, "loss": 0.6006, "step": 37720 }, { "epoch": 1.5506615169693485, "grad_norm": 1.110885739326477, "learning_rate": 2.553507100376428e-05, "loss": 0.6128, "step": 37740 }, { "epoch": 1.551483277179719, "grad_norm": 0.9926326870918274, "learning_rate": 2.5446314571017936e-05, "loss": 0.6145, "step": 37760 }, { "epoch": 1.5523050373900897, "grad_norm": 0.9480810165405273, "learning_rate": 2.535769016710975e-05, "loss": 0.6213, "step": 37780 }, { "epoch": 1.5531267976004601, "grad_norm": 1.1244728565216064, "learning_rate": 2.5269197948986678e-05, "loss": 0.5935, "step": 37800 }, { "epoch": 1.553948557810831, "grad_norm": 1.1508769989013672, "learning_rate": 2.5180838073361624e-05, "loss": 0.6349, "step": 37820 }, { "epoch": 1.5547703180212014, "grad_norm": 1.137568473815918, "learning_rate": 2.509261069671318e-05, "loss": 0.6057, "step": 37840 }, { "epoch": 1.555592078231572, "grad_norm": 1.1233346462249756, "learning_rate": 2.5004515975285183e-05, "loss": 0.5844, "step": 37860 }, { "epoch": 1.5564138384419426, "grad_norm": 1.188909649848938, "learning_rate": 2.491655406508667e-05, "loss": 0.6043, "step": 37880 }, { "epoch": 1.5572355986523132, "grad_norm": 1.0557928085327148, "learning_rate": 2.4828725121891328e-05, "loss": 0.5953, "step": 37900 }, { "epoch": 1.5580573588626838, "grad_norm": 1.2361866235733032, "learning_rate": 2.4745410928211422e-05, "loss": 0.6316, "step": 37920 }, { "epoch": 1.5588791190730544, "grad_norm": 1.1780842542648315, "learning_rate": 2.4662217010784527e-05, "loss": 0.614, "step": 37940 }, { "epoch": 1.559700879283425, "grad_norm": 1.0414899587631226, "learning_rate": 2.4574774550623027e-05, "loss": 0.5905, "step": 37960 }, { "epoch": 1.5605226394937957, "grad_norm": 1.1471487283706665, "learning_rate": 2.448746566272997e-05, "loss": 0.5866, "step": 37980 }, { "epoch": 1.5613443997041663, "grad_norm": 1.048036813735962, "learning_rate": 2.4400290501722623e-05, "loss": 0.6068, "step": 38000 }, { "epoch": 1.5613443997041663, "eval_loss": 0.8905351758003235, "eval_runtime": 16.5754, "eval_samples_per_second": 158.066, "eval_steps_per_second": 4.947, "step": 38000 }, { "epoch": 1.562166159914537, "grad_norm": 1.0737829208374023, "learning_rate": 2.431324922198156e-05, "loss": 0.6073, "step": 38020 }, { "epoch": 1.5629879201249075, "grad_norm": 1.0627774000167847, "learning_rate": 2.4226341977650145e-05, "loss": 0.6057, "step": 38040 }, { "epoch": 1.5638096803352781, "grad_norm": 1.09597647190094, "learning_rate": 2.4139568922634427e-05, "loss": 0.6003, "step": 38060 }, { "epoch": 1.5646314405456487, "grad_norm": 1.0973150730133057, "learning_rate": 2.40529302106028e-05, "loss": 0.6181, "step": 38080 }, { "epoch": 1.5654532007560193, "grad_norm": 1.163824200630188, "learning_rate": 2.396642599498573e-05, "loss": 0.5985, "step": 38100 }, { "epoch": 1.56627496096639, "grad_norm": 1.2081292867660522, "learning_rate": 2.3880056428975572e-05, "loss": 0.5934, "step": 38120 }, { "epoch": 1.5670967211767606, "grad_norm": 1.1347095966339111, "learning_rate": 2.379382166552614e-05, "loss": 0.6004, "step": 38140 }, { "epoch": 1.5679184813871312, "grad_norm": 1.0734606981277466, "learning_rate": 2.3707721857352628e-05, "loss": 0.6167, "step": 38160 }, { "epoch": 1.5687402415975018, "grad_norm": 1.0315816402435303, "learning_rate": 2.362175715693106e-05, "loss": 0.5854, "step": 38180 }, { "epoch": 1.5695620018078724, "grad_norm": 1.1294316053390503, "learning_rate": 2.3535927716498397e-05, "loss": 0.5945, "step": 38200 }, { "epoch": 1.570383762018243, "grad_norm": 1.1472307443618774, "learning_rate": 2.3450233688051936e-05, "loss": 0.5976, "step": 38220 }, { "epoch": 1.5712055222286136, "grad_norm": 1.253547191619873, "learning_rate": 2.3364675223349186e-05, "loss": 0.6084, "step": 38240 }, { "epoch": 1.5720272824389843, "grad_norm": 1.0203603506088257, "learning_rate": 2.3279252473907674e-05, "loss": 0.5904, "step": 38260 }, { "epoch": 1.5728490426493549, "grad_norm": 1.0900869369506836, "learning_rate": 2.3193965591004408e-05, "loss": 0.6092, "step": 38280 }, { "epoch": 1.5736708028597255, "grad_norm": 1.003138542175293, "learning_rate": 2.3108814725675975e-05, "loss": 0.5982, "step": 38300 }, { "epoch": 1.574492563070096, "grad_norm": 1.138856291770935, "learning_rate": 2.3023800028717956e-05, "loss": 0.6178, "step": 38320 }, { "epoch": 1.5753143232804667, "grad_norm": 1.1773872375488281, "learning_rate": 2.2943162329690658e-05, "loss": 0.6051, "step": 38340 }, { "epoch": 1.5761360834908373, "grad_norm": 1.2317793369293213, "learning_rate": 2.2858413593867434e-05, "loss": 0.6073, "step": 38360 }, { "epoch": 1.576957843701208, "grad_norm": 1.0834593772888184, "learning_rate": 2.2773801469855805e-05, "loss": 0.6154, "step": 38380 }, { "epoch": 1.5777796039115786, "grad_norm": 1.1221275329589844, "learning_rate": 2.2689326107497267e-05, "loss": 0.5776, "step": 38400 }, { "epoch": 1.5786013641219492, "grad_norm": 1.0579743385314941, "learning_rate": 2.260498765639125e-05, "loss": 0.5986, "step": 38420 }, { "epoch": 1.5794231243323198, "grad_norm": 1.0688302516937256, "learning_rate": 2.252078626589462e-05, "loss": 0.5839, "step": 38440 }, { "epoch": 1.5802448845426904, "grad_norm": 1.1622627973556519, "learning_rate": 2.2436722085121565e-05, "loss": 0.6065, "step": 38460 }, { "epoch": 1.581066644753061, "grad_norm": 1.1227208375930786, "learning_rate": 2.2352795262943272e-05, "loss": 0.6048, "step": 38480 }, { "epoch": 1.5818884049634316, "grad_norm": 0.9703273177146912, "learning_rate": 2.2269005947987664e-05, "loss": 0.6037, "step": 38500 }, { "epoch": 1.5827101651738023, "grad_norm": 1.0502504110336304, "learning_rate": 2.2185354288639216e-05, "loss": 0.6096, "step": 38520 }, { "epoch": 1.5835319253841729, "grad_norm": 1.1714296340942383, "learning_rate": 2.210184043303852e-05, "loss": 0.612, "step": 38540 }, { "epoch": 1.5843536855945435, "grad_norm": 1.051988124847412, "learning_rate": 2.2018464529082282e-05, "loss": 0.6334, "step": 38560 }, { "epoch": 1.585175445804914, "grad_norm": 1.1384596824645996, "learning_rate": 2.1935226724422686e-05, "loss": 0.6027, "step": 38580 }, { "epoch": 1.5859972060152847, "grad_norm": 1.1457267999649048, "learning_rate": 2.1852127166467572e-05, "loss": 0.5929, "step": 38600 }, { "epoch": 1.5868189662256553, "grad_norm": 1.1787493228912354, "learning_rate": 2.1769166002379826e-05, "loss": 0.5897, "step": 38620 }, { "epoch": 1.587640726436026, "grad_norm": 1.1157082319259644, "learning_rate": 2.1686343379077246e-05, "loss": 0.5987, "step": 38640 }, { "epoch": 1.5884624866463966, "grad_norm": 1.0821255445480347, "learning_rate": 2.1603659443232394e-05, "loss": 0.6027, "step": 38660 }, { "epoch": 1.5892842468567672, "grad_norm": 1.0201191902160645, "learning_rate": 2.152111434127212e-05, "loss": 0.6001, "step": 38680 }, { "epoch": 1.5901060070671378, "grad_norm": 1.0591297149658203, "learning_rate": 2.1438708219377444e-05, "loss": 0.6132, "step": 38700 }, { "epoch": 1.5909277672775084, "grad_norm": 1.0407557487487793, "learning_rate": 2.1356441223483246e-05, "loss": 0.5904, "step": 38720 }, { "epoch": 1.591749527487879, "grad_norm": 1.0210407972335815, "learning_rate": 2.12743134992781e-05, "loss": 0.5876, "step": 38740 }, { "epoch": 1.5925712876982496, "grad_norm": 1.0307002067565918, "learning_rate": 2.1192325192203843e-05, "loss": 0.6178, "step": 38760 }, { "epoch": 1.5933930479086202, "grad_norm": 0.9952294826507568, "learning_rate": 2.1110476447455453e-05, "loss": 0.59, "step": 38780 }, { "epoch": 1.5942148081189909, "grad_norm": 1.145140528678894, "learning_rate": 2.1028767409980776e-05, "loss": 0.5751, "step": 38800 }, { "epoch": 1.5950365683293615, "grad_norm": 1.0947434902191162, "learning_rate": 2.094719822448019e-05, "loss": 0.5838, "step": 38820 }, { "epoch": 1.595858328539732, "grad_norm": 1.2791152000427246, "learning_rate": 2.086576903540649e-05, "loss": 0.6055, "step": 38840 }, { "epoch": 1.5966800887501027, "grad_norm": 1.142967700958252, "learning_rate": 2.0784479986964467e-05, "loss": 0.5962, "step": 38860 }, { "epoch": 1.5975018489604733, "grad_norm": 1.1335628032684326, "learning_rate": 2.070333122311081e-05, "loss": 0.6056, "step": 38880 }, { "epoch": 1.598323609170844, "grad_norm": 1.2280479669570923, "learning_rate": 2.0622322887553703e-05, "loss": 0.6012, "step": 38900 }, { "epoch": 1.5991453693812145, "grad_norm": 1.0319156646728516, "learning_rate": 2.0541455123752686e-05, "loss": 0.5985, "step": 38920 }, { "epoch": 1.5999671295915852, "grad_norm": 0.9951415061950684, "learning_rate": 2.046072807491832e-05, "loss": 0.594, "step": 38940 }, { "epoch": 1.6007888898019558, "grad_norm": 1.0593833923339844, "learning_rate": 2.0380141884012004e-05, "loss": 0.5987, "step": 38960 }, { "epoch": 1.6016106500123264, "grad_norm": 1.0992417335510254, "learning_rate": 2.0299696693745697e-05, "loss": 0.5815, "step": 38980 }, { "epoch": 1.602432410222697, "grad_norm": 0.9815024137496948, "learning_rate": 2.0219392646581638e-05, "loss": 0.5727, "step": 39000 }, { "epoch": 1.6032541704330676, "grad_norm": 1.1512722969055176, "learning_rate": 2.013922988473209e-05, "loss": 0.6208, "step": 39020 }, { "epoch": 1.6040759306434382, "grad_norm": 1.1306536197662354, "learning_rate": 2.0059208550159125e-05, "loss": 0.6162, "step": 39040 }, { "epoch": 1.6048976908538088, "grad_norm": 1.1175142526626587, "learning_rate": 1.9979328784574415e-05, "loss": 0.5983, "step": 39060 }, { "epoch": 1.6057194510641795, "grad_norm": 1.3080164194107056, "learning_rate": 1.9899590729438856e-05, "loss": 0.6112, "step": 39080 }, { "epoch": 1.60654121127455, "grad_norm": 1.130448579788208, "learning_rate": 1.981999452596236e-05, "loss": 0.5753, "step": 39100 }, { "epoch": 1.6073629714849207, "grad_norm": 1.0560057163238525, "learning_rate": 1.9740540315103772e-05, "loss": 0.593, "step": 39120 }, { "epoch": 1.6081847316952913, "grad_norm": 1.0870985984802246, "learning_rate": 1.9661228237570272e-05, "loss": 0.606, "step": 39140 }, { "epoch": 1.609006491905662, "grad_norm": 1.1902962923049927, "learning_rate": 1.9582058433817528e-05, "loss": 0.6184, "step": 39160 }, { "epoch": 1.6098282521160325, "grad_norm": 1.0945310592651367, "learning_rate": 1.9503031044049136e-05, "loss": 0.5869, "step": 39180 }, { "epoch": 1.6106500123264031, "grad_norm": 1.018189549446106, "learning_rate": 1.942414620821651e-05, "loss": 0.605, "step": 39200 }, { "epoch": 1.6114717725367738, "grad_norm": 1.0882275104522705, "learning_rate": 1.934540406601867e-05, "loss": 0.604, "step": 39220 }, { "epoch": 1.6122935327471444, "grad_norm": 1.091739296913147, "learning_rate": 1.9266804756901812e-05, "loss": 0.588, "step": 39240 }, { "epoch": 1.613115292957515, "grad_norm": 1.0217018127441406, "learning_rate": 1.918834842005933e-05, "loss": 0.5712, "step": 39260 }, { "epoch": 1.6139370531678856, "grad_norm": 1.0229742527008057, "learning_rate": 1.9110035194431298e-05, "loss": 0.5956, "step": 39280 }, { "epoch": 1.6147588133782562, "grad_norm": 1.082470417022705, "learning_rate": 1.903186521870448e-05, "loss": 0.5962, "step": 39300 }, { "epoch": 1.6155805735886268, "grad_norm": 1.0316046476364136, "learning_rate": 1.895383863131185e-05, "loss": 0.6197, "step": 39320 }, { "epoch": 1.6164023337989974, "grad_norm": 1.150221347808838, "learning_rate": 1.887595557043248e-05, "loss": 0.6, "step": 39340 }, { "epoch": 1.617224094009368, "grad_norm": 1.0211883783340454, "learning_rate": 1.879821617399129e-05, "loss": 0.5957, "step": 39360 }, { "epoch": 1.6180458542197387, "grad_norm": 1.1843931674957275, "learning_rate": 1.8720620579658733e-05, "loss": 0.5956, "step": 39380 }, { "epoch": 1.6188676144301093, "grad_norm": 1.0743961334228516, "learning_rate": 1.8643168924850695e-05, "loss": 0.6038, "step": 39400 }, { "epoch": 1.61968937464048, "grad_norm": 0.9900259971618652, "learning_rate": 1.8565861346728032e-05, "loss": 0.5963, "step": 39420 }, { "epoch": 1.6205111348508505, "grad_norm": 1.045684814453125, "learning_rate": 1.848869798219659e-05, "loss": 0.6107, "step": 39440 }, { "epoch": 1.6213328950612211, "grad_norm": 1.1444482803344727, "learning_rate": 1.8411678967906655e-05, "loss": 0.6047, "step": 39460 }, { "epoch": 1.6221546552715917, "grad_norm": 1.0717675685882568, "learning_rate": 1.833480444025304e-05, "loss": 0.6126, "step": 39480 }, { "epoch": 1.6229764154819624, "grad_norm": 1.04216468334198, "learning_rate": 1.8258074535374604e-05, "loss": 0.584, "step": 39500 }, { "epoch": 1.623798175692333, "grad_norm": 1.0672125816345215, "learning_rate": 1.818148938915406e-05, "loss": 0.5991, "step": 39520 }, { "epoch": 1.6246199359027036, "grad_norm": 1.1428139209747314, "learning_rate": 1.81050491372179e-05, "loss": 0.5991, "step": 39540 }, { "epoch": 1.6254416961130742, "grad_norm": 1.1530945301055908, "learning_rate": 1.80287539149358e-05, "loss": 0.6207, "step": 39560 }, { "epoch": 1.6262634563234448, "grad_norm": 1.0400162935256958, "learning_rate": 1.7952603857420837e-05, "loss": 0.6112, "step": 39580 }, { "epoch": 1.6270852165338154, "grad_norm": 1.172669768333435, "learning_rate": 1.7876599099528822e-05, "loss": 0.5991, "step": 39600 }, { "epoch": 1.627906976744186, "grad_norm": 0.9719765782356262, "learning_rate": 1.78007397758584e-05, "loss": 0.6059, "step": 39620 }, { "epoch": 1.6287287369545567, "grad_norm": 1.0943204164505005, "learning_rate": 1.7725026020750547e-05, "loss": 0.5868, "step": 39640 }, { "epoch": 1.6295504971649273, "grad_norm": 1.281966209411621, "learning_rate": 1.764945796828852e-05, "loss": 0.6005, "step": 39660 }, { "epoch": 1.630372257375298, "grad_norm": 1.1267539262771606, "learning_rate": 1.7574035752297503e-05, "loss": 0.5725, "step": 39680 }, { "epoch": 1.6311940175856685, "grad_norm": 1.1680996417999268, "learning_rate": 1.749875950634442e-05, "loss": 0.5967, "step": 39700 }, { "epoch": 1.6320157777960391, "grad_norm": 1.0391075611114502, "learning_rate": 1.742362936373776e-05, "loss": 0.5903, "step": 39720 }, { "epoch": 1.6328375380064097, "grad_norm": 1.166467547416687, "learning_rate": 1.734864545752716e-05, "loss": 0.6019, "step": 39740 }, { "epoch": 1.6336592982167804, "grad_norm": 1.027796983718872, "learning_rate": 1.7273807920503436e-05, "loss": 0.5935, "step": 39760 }, { "epoch": 1.634481058427151, "grad_norm": 1.0871942043304443, "learning_rate": 1.7199116885197995e-05, "loss": 0.6119, "step": 39780 }, { "epoch": 1.6353028186375216, "grad_norm": 1.0691869258880615, "learning_rate": 1.7124572483882996e-05, "loss": 0.6061, "step": 39800 }, { "epoch": 1.6361245788478922, "grad_norm": 1.0936367511749268, "learning_rate": 1.70501748485708e-05, "loss": 0.5716, "step": 39820 }, { "epoch": 1.6369463390582628, "grad_norm": 1.1092150211334229, "learning_rate": 1.6975924111013873e-05, "loss": 0.5975, "step": 39840 }, { "epoch": 1.6377680992686334, "grad_norm": 0.9624285697937012, "learning_rate": 1.6901820402704606e-05, "loss": 0.6096, "step": 39860 }, { "epoch": 1.638589859479004, "grad_norm": 1.1558884382247925, "learning_rate": 1.6827863854874938e-05, "loss": 0.6157, "step": 39880 }, { "epoch": 1.6394116196893747, "grad_norm": 1.3336347341537476, "learning_rate": 1.6754054598496215e-05, "loss": 0.601, "step": 39900 }, { "epoch": 1.6402333798997453, "grad_norm": 0.9931703209877014, "learning_rate": 1.668039276427894e-05, "loss": 0.5828, "step": 39920 }, { "epoch": 1.6410551401101159, "grad_norm": 1.170390248298645, "learning_rate": 1.6606878482672582e-05, "loss": 0.5879, "step": 39940 }, { "epoch": 1.6418769003204865, "grad_norm": 1.0788556337356567, "learning_rate": 1.653351188386526e-05, "loss": 0.603, "step": 39960 }, { "epoch": 1.6426986605308571, "grad_norm": 1.1531950235366821, "learning_rate": 1.6460293097783574e-05, "loss": 0.6071, "step": 39980 }, { "epoch": 1.6435204207412277, "grad_norm": 1.1545748710632324, "learning_rate": 1.638722225409236e-05, "loss": 0.5967, "step": 40000 }, { "epoch": 1.6435204207412277, "eval_loss": 0.8823444247245789, "eval_runtime": 16.6502, "eval_samples_per_second": 157.355, "eval_steps_per_second": 4.925, "step": 40000 }, { "epoch": 1.6443421809515983, "grad_norm": 1.1601914167404175, "learning_rate": 1.6314299482194418e-05, "loss": 0.6032, "step": 40020 }, { "epoch": 1.645163941161969, "grad_norm": 1.0907021760940552, "learning_rate": 1.624152491123043e-05, "loss": 0.6075, "step": 40040 }, { "epoch": 1.6459857013723396, "grad_norm": 1.2451189756393433, "learning_rate": 1.6168898670078537e-05, "loss": 0.5927, "step": 40060 }, { "epoch": 1.6468074615827102, "grad_norm": 1.2181775569915771, "learning_rate": 1.609642088735418e-05, "loss": 0.5866, "step": 40080 }, { "epoch": 1.6476292217930808, "grad_norm": 1.0169905424118042, "learning_rate": 1.6024091691410013e-05, "loss": 0.5901, "step": 40100 }, { "epoch": 1.6484509820034514, "grad_norm": 1.0728631019592285, "learning_rate": 1.595191121033538e-05, "loss": 0.5929, "step": 40120 }, { "epoch": 1.649272742213822, "grad_norm": 1.216049313545227, "learning_rate": 1.5879879571956436e-05, "loss": 0.5836, "step": 40140 }, { "epoch": 1.6500945024241926, "grad_norm": 1.093613862991333, "learning_rate": 1.5807996903835608e-05, "loss": 0.5816, "step": 40160 }, { "epoch": 1.6509162626345633, "grad_norm": 1.1138787269592285, "learning_rate": 1.5736263333271618e-05, "loss": 0.5954, "step": 40180 }, { "epoch": 1.6517380228449339, "grad_norm": 1.0472607612609863, "learning_rate": 1.5664678987299085e-05, "loss": 0.6054, "step": 40200 }, { "epoch": 1.6525597830553045, "grad_norm": 1.0426794290542603, "learning_rate": 1.5593243992688356e-05, "loss": 0.5902, "step": 40220 }, { "epoch": 1.653381543265675, "grad_norm": 1.1743807792663574, "learning_rate": 1.552195847594533e-05, "loss": 0.5822, "step": 40240 }, { "epoch": 1.6542033034760457, "grad_norm": 1.111167550086975, "learning_rate": 1.5450822563311128e-05, "loss": 0.628, "step": 40260 }, { "epoch": 1.6550250636864163, "grad_norm": 1.124616265296936, "learning_rate": 1.5379836380762035e-05, "loss": 0.5867, "step": 40280 }, { "epoch": 1.655846823896787, "grad_norm": 1.0396867990493774, "learning_rate": 1.530900005400906e-05, "loss": 0.6029, "step": 40300 }, { "epoch": 1.6566685841071576, "grad_norm": 1.176413893699646, "learning_rate": 1.5238313708497964e-05, "loss": 0.6044, "step": 40320 }, { "epoch": 1.6574903443175282, "grad_norm": 1.1538265943527222, "learning_rate": 1.5167777469408728e-05, "loss": 0.591, "step": 40340 }, { "epoch": 1.6583121045278988, "grad_norm": 1.1776963472366333, "learning_rate": 1.5097391461655663e-05, "loss": 0.5988, "step": 40360 }, { "epoch": 1.6591338647382694, "grad_norm": 1.0726710557937622, "learning_rate": 1.5027155809886962e-05, "loss": 0.6071, "step": 40380 }, { "epoch": 1.65995562494864, "grad_norm": 1.0493417978286743, "learning_rate": 1.4957070638484515e-05, "loss": 0.5879, "step": 40400 }, { "epoch": 1.6607773851590106, "grad_norm": 1.1617422103881836, "learning_rate": 1.4887136071563856e-05, "loss": 0.5913, "step": 40420 }, { "epoch": 1.6615991453693812, "grad_norm": 1.0822564363479614, "learning_rate": 1.4817352232973626e-05, "loss": 0.601, "step": 40440 }, { "epoch": 1.6624209055797519, "grad_norm": 1.1444182395935059, "learning_rate": 1.4747719246295676e-05, "loss": 0.602, "step": 40460 }, { "epoch": 1.6632426657901225, "grad_norm": 1.0876108407974243, "learning_rate": 1.4678237234844649e-05, "loss": 0.6045, "step": 40480 }, { "epoch": 1.664064426000493, "grad_norm": 1.1832387447357178, "learning_rate": 1.460890632166787e-05, "loss": 0.5967, "step": 40500 }, { "epoch": 1.6648861862108637, "grad_norm": 0.9234125018119812, "learning_rate": 1.4543182020758783e-05, "loss": 0.5895, "step": 40520 }, { "epoch": 1.6657079464212343, "grad_norm": 1.1049689054489136, "learning_rate": 1.4474146102117969e-05, "loss": 0.6013, "step": 40540 }, { "epoch": 1.666529706631605, "grad_norm": 1.0653325319290161, "learning_rate": 1.4405261643181056e-05, "loss": 0.6025, "step": 40560 }, { "epoch": 1.6673514668419755, "grad_norm": 1.2282037734985352, "learning_rate": 1.4336528765937151e-05, "loss": 0.614, "step": 40580 }, { "epoch": 1.6681732270523462, "grad_norm": 1.1362701654434204, "learning_rate": 1.4267947592106845e-05, "loss": 0.6083, "step": 40600 }, { "epoch": 1.6689949872627168, "grad_norm": 1.0828361511230469, "learning_rate": 1.4199518243142196e-05, "loss": 0.5823, "step": 40620 }, { "epoch": 1.6698167474730874, "grad_norm": 1.042107105255127, "learning_rate": 1.4131240840226201e-05, "loss": 0.5848, "step": 40640 }, { "epoch": 1.670638507683458, "grad_norm": 1.00413978099823, "learning_rate": 1.4063115504272973e-05, "loss": 0.6082, "step": 40660 }, { "epoch": 1.6714602678938286, "grad_norm": 1.0120972394943237, "learning_rate": 1.3995142355927216e-05, "loss": 0.582, "step": 40680 }, { "epoch": 1.6722820281041992, "grad_norm": 1.2198662757873535, "learning_rate": 1.3927321515564107e-05, "loss": 0.5908, "step": 40700 }, { "epoch": 1.6731037883145699, "grad_norm": 1.1000231504440308, "learning_rate": 1.3859653103289205e-05, "loss": 0.585, "step": 40720 }, { "epoch": 1.6739255485249405, "grad_norm": 1.0395424365997314, "learning_rate": 1.3792137238937975e-05, "loss": 0.5767, "step": 40740 }, { "epoch": 1.674747308735311, "grad_norm": 1.1138056516647339, "learning_rate": 1.3724774042075882e-05, "loss": 0.5843, "step": 40760 }, { "epoch": 1.6755690689456817, "grad_norm": 1.0291800498962402, "learning_rate": 1.3657563631997928e-05, "loss": 0.583, "step": 40780 }, { "epoch": 1.6763908291560523, "grad_norm": 1.1663140058517456, "learning_rate": 1.3590506127728575e-05, "loss": 0.5913, "step": 40800 }, { "epoch": 1.677212589366423, "grad_norm": 1.0701870918273926, "learning_rate": 1.3523601648021533e-05, "loss": 0.5924, "step": 40820 }, { "epoch": 1.6780343495767935, "grad_norm": 1.1024507284164429, "learning_rate": 1.345685031135947e-05, "loss": 0.5786, "step": 40840 }, { "epoch": 1.6788561097871642, "grad_norm": 0.9765409231185913, "learning_rate": 1.3393578497947468e-05, "loss": 0.5825, "step": 40860 }, { "epoch": 1.6796778699975348, "grad_norm": 1.2072242498397827, "learning_rate": 1.332712612998166e-05, "loss": 0.6016, "step": 40880 }, { "epoch": 1.6804996302079054, "grad_norm": 1.1122468709945679, "learning_rate": 1.326082725300385e-05, "loss": 0.6234, "step": 40900 }, { "epoch": 1.681321390418276, "grad_norm": 1.0995004177093506, "learning_rate": 1.319468198442424e-05, "loss": 0.5913, "step": 40920 }, { "epoch": 1.6821431506286466, "grad_norm": 1.0409519672393799, "learning_rate": 1.3128690441380998e-05, "loss": 0.5948, "step": 40940 }, { "epoch": 1.6829649108390172, "grad_norm": 1.1031830310821533, "learning_rate": 1.3062852740740072e-05, "loss": 0.5988, "step": 40960 }, { "epoch": 1.6837866710493878, "grad_norm": 1.1113379001617432, "learning_rate": 1.2997168999094978e-05, "loss": 0.5822, "step": 40980 }, { "epoch": 1.6846084312597585, "grad_norm": 1.0048646926879883, "learning_rate": 1.2931639332766543e-05, "loss": 0.5817, "step": 41000 }, { "epoch": 1.685430191470129, "grad_norm": 1.159192681312561, "learning_rate": 1.2866263857802818e-05, "loss": 0.5832, "step": 41020 }, { "epoch": 1.6862519516804997, "grad_norm": 1.1091160774230957, "learning_rate": 1.280104268997865e-05, "loss": 0.5858, "step": 41040 }, { "epoch": 1.6870737118908703, "grad_norm": 1.1325902938842773, "learning_rate": 1.2735975944795775e-05, "loss": 0.5843, "step": 41060 }, { "epoch": 1.687895472101241, "grad_norm": 1.0565228462219238, "learning_rate": 1.267106373748237e-05, "loss": 0.5852, "step": 41080 }, { "epoch": 1.6887172323116115, "grad_norm": 1.1818876266479492, "learning_rate": 1.2606306182992933e-05, "loss": 0.588, "step": 41100 }, { "epoch": 1.6895389925219821, "grad_norm": 0.8950326442718506, "learning_rate": 1.2541703396008142e-05, "loss": 0.5963, "step": 41120 }, { "epoch": 1.6903607527323525, "grad_norm": 1.1267063617706299, "learning_rate": 1.2477255490934559e-05, "loss": 0.5758, "step": 41140 }, { "epoch": 1.6911825129427234, "grad_norm": 0.9466457962989807, "learning_rate": 1.241296258190444e-05, "loss": 0.5963, "step": 41160 }, { "epoch": 1.6920042731530938, "grad_norm": 1.0484158992767334, "learning_rate": 1.2348824782775581e-05, "loss": 0.586, "step": 41180 }, { "epoch": 1.6928260333634646, "grad_norm": 1.1975510120391846, "learning_rate": 1.2284842207131109e-05, "loss": 0.5775, "step": 41200 }, { "epoch": 1.693647793573835, "grad_norm": 1.1231242418289185, "learning_rate": 1.2221014968279233e-05, "loss": 0.5915, "step": 41220 }, { "epoch": 1.6944695537842058, "grad_norm": 1.1098995208740234, "learning_rate": 1.2157343179253079e-05, "loss": 0.5886, "step": 41240 }, { "epoch": 1.6952913139945762, "grad_norm": 1.053568720817566, "learning_rate": 1.2093826952810471e-05, "loss": 0.5961, "step": 41260 }, { "epoch": 1.696113074204947, "grad_norm": 1.1454240083694458, "learning_rate": 1.2030466401433748e-05, "loss": 0.5888, "step": 41280 }, { "epoch": 1.6969348344153175, "grad_norm": 1.0990582704544067, "learning_rate": 1.1967261637329607e-05, "loss": 0.5945, "step": 41300 }, { "epoch": 1.6977565946256883, "grad_norm": 1.2519744634628296, "learning_rate": 1.190421277242878e-05, "loss": 0.5782, "step": 41320 }, { "epoch": 1.6985783548360587, "grad_norm": 0.9864106178283691, "learning_rate": 1.1841319918385996e-05, "loss": 0.5856, "step": 41340 }, { "epoch": 1.6994001150464295, "grad_norm": 1.0756564140319824, "learning_rate": 1.1778583186579628e-05, "loss": 0.5893, "step": 41360 }, { "epoch": 1.7002218752568, "grad_norm": 1.163355827331543, "learning_rate": 1.1716002688111616e-05, "loss": 0.6051, "step": 41380 }, { "epoch": 1.7010436354671707, "grad_norm": 1.259600281715393, "learning_rate": 1.1653578533807186e-05, "loss": 0.6031, "step": 41400 }, { "epoch": 1.7018653956775411, "grad_norm": 1.1272526979446411, "learning_rate": 1.1591310834214709e-05, "loss": 0.584, "step": 41420 }, { "epoch": 1.702687155887912, "grad_norm": 1.1468937397003174, "learning_rate": 1.152919969960552e-05, "loss": 0.6068, "step": 41440 }, { "epoch": 1.7035089160982824, "grad_norm": 1.1517606973648071, "learning_rate": 1.1467245239973633e-05, "loss": 0.5757, "step": 41460 }, { "epoch": 1.7043306763086532, "grad_norm": 1.0870920419692993, "learning_rate": 1.1405447565035631e-05, "loss": 0.6043, "step": 41480 }, { "epoch": 1.7051524365190236, "grad_norm": 1.0770379304885864, "learning_rate": 1.1343806784230426e-05, "loss": 0.5905, "step": 41500 }, { "epoch": 1.7059741967293944, "grad_norm": 1.0060986280441284, "learning_rate": 1.128232300671912e-05, "loss": 0.6047, "step": 41520 }, { "epoch": 1.7067959569397648, "grad_norm": 1.0191991329193115, "learning_rate": 1.1220996341384748e-05, "loss": 0.5647, "step": 41540 }, { "epoch": 1.7076177171501357, "grad_norm": 1.191707730293274, "learning_rate": 1.1159826896832082e-05, "loss": 0.5875, "step": 41560 }, { "epoch": 1.708439477360506, "grad_norm": 1.2851048707962036, "learning_rate": 1.1098814781387568e-05, "loss": 0.5908, "step": 41580 }, { "epoch": 1.709261237570877, "grad_norm": 1.121020793914795, "learning_rate": 1.1037960103098877e-05, "loss": 0.6084, "step": 41600 }, { "epoch": 1.7100829977812473, "grad_norm": 1.1154911518096924, "learning_rate": 1.0977262969735014e-05, "loss": 0.5814, "step": 41620 }, { "epoch": 1.7109047579916181, "grad_norm": 1.1276777982711792, "learning_rate": 1.091672348878594e-05, "loss": 0.5853, "step": 41640 }, { "epoch": 1.7117265182019885, "grad_norm": 1.080946922302246, "learning_rate": 1.0856341767462364e-05, "loss": 0.605, "step": 41660 }, { "epoch": 1.7125482784123593, "grad_norm": 1.1074481010437012, "learning_rate": 1.0796117912695736e-05, "loss": 0.5711, "step": 41680 }, { "epoch": 1.7133700386227297, "grad_norm": 1.1239150762557983, "learning_rate": 1.07360520311378e-05, "loss": 0.5841, "step": 41700 }, { "epoch": 1.7141917988331006, "grad_norm": 1.0257426500320435, "learning_rate": 1.0676144229160655e-05, "loss": 0.5723, "step": 41720 }, { "epoch": 1.715013559043471, "grad_norm": 1.100321888923645, "learning_rate": 1.0616394612856361e-05, "loss": 0.5999, "step": 41740 }, { "epoch": 1.7158353192538418, "grad_norm": 1.0664868354797363, "learning_rate": 1.0556803288036954e-05, "loss": 0.6029, "step": 41760 }, { "epoch": 1.7166570794642122, "grad_norm": 1.017471194267273, "learning_rate": 1.0497370360234037e-05, "loss": 0.5844, "step": 41780 }, { "epoch": 1.717478839674583, "grad_norm": 1.0937682390213013, "learning_rate": 1.0438095934698766e-05, "loss": 0.5844, "step": 41800 }, { "epoch": 1.7183005998849534, "grad_norm": 1.0297489166259766, "learning_rate": 1.037898011640157e-05, "loss": 0.5882, "step": 41820 }, { "epoch": 1.7191223600953243, "grad_norm": 1.0966808795928955, "learning_rate": 1.032002301003202e-05, "loss": 0.5827, "step": 41840 }, { "epoch": 1.7199441203056947, "grad_norm": 1.1811338663101196, "learning_rate": 1.026122471999863e-05, "loss": 0.6009, "step": 41860 }, { "epoch": 1.7207658805160655, "grad_norm": 1.0118420124053955, "learning_rate": 1.0202585350428606e-05, "loss": 0.5921, "step": 41880 }, { "epoch": 1.7215876407264359, "grad_norm": 1.0821563005447388, "learning_rate": 1.0144105005167836e-05, "loss": 0.5956, "step": 41900 }, { "epoch": 1.7224094009368067, "grad_norm": 1.0998246669769287, "learning_rate": 1.0085783787780412e-05, "loss": 0.5851, "step": 41920 }, { "epoch": 1.7232311611471771, "grad_norm": 1.2186909914016724, "learning_rate": 1.0027621801548792e-05, "loss": 0.6043, "step": 41940 }, { "epoch": 1.724052921357548, "grad_norm": 1.1629652976989746, "learning_rate": 9.972515496304035e-06, "loss": 0.5626, "step": 41960 }, { "epoch": 1.7248746815679183, "grad_norm": 1.0134592056274414, "learning_rate": 9.914664306824105e-06, "loss": 0.5605, "step": 41980 }, { "epoch": 1.7256964417782892, "grad_norm": 1.134350061416626, "learning_rate": 9.856972651539343e-06, "loss": 0.5788, "step": 42000 }, { "epoch": 1.7256964417782892, "eval_loss": 0.8728025555610657, "eval_runtime": 16.6626, "eval_samples_per_second": 157.238, "eval_steps_per_second": 4.921, "step": 42000 }, { "epoch": 1.7265182019886596, "grad_norm": 1.1034976243972778, "learning_rate": 9.799440632617284e-06, "loss": 0.583, "step": 42020 }, { "epoch": 1.7273399621990304, "grad_norm": 1.0975291728973389, "learning_rate": 9.742068351942668e-06, "loss": 0.5831, "step": 42040 }, { "epoch": 1.7281617224094008, "grad_norm": 1.1249350309371948, "learning_rate": 9.68485591111744e-06, "loss": 0.5758, "step": 42060 }, { "epoch": 1.7289834826197716, "grad_norm": 1.0432801246643066, "learning_rate": 9.6278034114604e-06, "loss": 0.5898, "step": 42080 }, { "epoch": 1.729805242830142, "grad_norm": 1.0415185689926147, "learning_rate": 9.570910954007128e-06, "loss": 0.5781, "step": 42100 }, { "epoch": 1.7306270030405129, "grad_norm": 1.0248029232025146, "learning_rate": 9.517011450287305e-06, "loss": 0.588, "step": 42120 }, { "epoch": 1.7314487632508833, "grad_norm": 1.0104538202285767, "learning_rate": 9.460431364661492e-06, "loss": 0.5787, "step": 42140 }, { "epoch": 1.732270523461254, "grad_norm": 1.0787935256958008, "learning_rate": 9.404011617642439e-06, "loss": 0.6012, "step": 42160 }, { "epoch": 1.7330922836716245, "grad_norm": 1.10727059841156, "learning_rate": 9.347752309145241e-06, "loss": 0.5939, "step": 42180 }, { "epoch": 1.7339140438819953, "grad_norm": 1.0951191186904907, "learning_rate": 9.291653538800727e-06, "loss": 0.5707, "step": 42200 }, { "epoch": 1.7347358040923657, "grad_norm": 0.9968528151512146, "learning_rate": 9.235715405955558e-06, "loss": 0.6074, "step": 42220 }, { "epoch": 1.7355575643027366, "grad_norm": 1.2177423238754272, "learning_rate": 9.17993800967183e-06, "loss": 0.5915, "step": 42240 }, { "epoch": 1.736379324513107, "grad_norm": 1.041279673576355, "learning_rate": 9.124321448727014e-06, "loss": 0.5841, "step": 42260 }, { "epoch": 1.7372010847234778, "grad_norm": 1.0093954801559448, "learning_rate": 9.068865821613803e-06, "loss": 0.5966, "step": 42280 }, { "epoch": 1.7380228449338482, "grad_norm": 0.9137701392173767, "learning_rate": 9.013571226539773e-06, "loss": 0.5792, "step": 42300 }, { "epoch": 1.738844605144219, "grad_norm": 1.1452417373657227, "learning_rate": 8.958437761427452e-06, "loss": 0.5968, "step": 42320 }, { "epoch": 1.7396663653545894, "grad_norm": 1.0660525560379028, "learning_rate": 8.903465523913957e-06, "loss": 0.5883, "step": 42340 }, { "epoch": 1.7404881255649602, "grad_norm": 1.2415095567703247, "learning_rate": 8.848654611350849e-06, "loss": 0.5766, "step": 42360 }, { "epoch": 1.7413098857753306, "grad_norm": 1.097631812095642, "learning_rate": 8.794005120804082e-06, "loss": 0.5838, "step": 42380 }, { "epoch": 1.7421316459857015, "grad_norm": 0.9962956309318542, "learning_rate": 8.739517149053689e-06, "loss": 0.5873, "step": 42400 }, { "epoch": 1.7429534061960719, "grad_norm": 1.1835882663726807, "learning_rate": 8.685190792593656e-06, "loss": 0.5863, "step": 42420 }, { "epoch": 1.7437751664064427, "grad_norm": 1.101272702217102, "learning_rate": 8.631026147631772e-06, "loss": 0.5901, "step": 42440 }, { "epoch": 1.744596926616813, "grad_norm": 1.1990203857421875, "learning_rate": 8.577023310089483e-06, "loss": 0.6065, "step": 42460 }, { "epoch": 1.745418686827184, "grad_norm": 1.1126364469528198, "learning_rate": 8.523182375601635e-06, "loss": 0.5755, "step": 42480 }, { "epoch": 1.7462404470375543, "grad_norm": 1.124306082725525, "learning_rate": 8.469503439516402e-06, "loss": 0.6036, "step": 42500 }, { "epoch": 1.7470622072479252, "grad_norm": 1.1404842138290405, "learning_rate": 8.41598659689502e-06, "loss": 0.5996, "step": 42520 }, { "epoch": 1.7478839674582956, "grad_norm": 1.0869922637939453, "learning_rate": 8.36263194251169e-06, "loss": 0.5756, "step": 42540 }, { "epoch": 1.7487057276686664, "grad_norm": 1.0685384273529053, "learning_rate": 8.309439570853439e-06, "loss": 0.5936, "step": 42560 }, { "epoch": 1.7495274878790368, "grad_norm": 1.1454116106033325, "learning_rate": 8.256409576119827e-06, "loss": 0.5871, "step": 42580 }, { "epoch": 1.7503492480894076, "grad_norm": 1.119585633277893, "learning_rate": 8.203542052222924e-06, "loss": 0.5948, "step": 42600 }, { "epoch": 1.751171008299778, "grad_norm": 1.0509191751480103, "learning_rate": 8.150837092787034e-06, "loss": 0.5856, "step": 42620 }, { "epoch": 1.7519927685101488, "grad_norm": 0.9895453453063965, "learning_rate": 8.098294791148565e-06, "loss": 0.5877, "step": 42640 }, { "epoch": 1.7528145287205192, "grad_norm": 1.1008808612823486, "learning_rate": 8.045915240355917e-06, "loss": 0.59, "step": 42660 }, { "epoch": 1.75363628893089, "grad_norm": 0.9279462695121765, "learning_rate": 7.993698533169192e-06, "loss": 0.5911, "step": 42680 }, { "epoch": 1.7544580491412605, "grad_norm": 1.1394389867782593, "learning_rate": 7.941644762060229e-06, "loss": 0.5756, "step": 42700 }, { "epoch": 1.7552798093516313, "grad_norm": 1.0450705289840698, "learning_rate": 7.889754019212203e-06, "loss": 0.6016, "step": 42720 }, { "epoch": 1.7561015695620017, "grad_norm": 1.2323449850082397, "learning_rate": 7.838026396519638e-06, "loss": 0.5808, "step": 42740 }, { "epoch": 1.7569233297723725, "grad_norm": 1.1465022563934326, "learning_rate": 7.786461985588156e-06, "loss": 0.5987, "step": 42760 }, { "epoch": 1.757745089982743, "grad_norm": 0.9561547636985779, "learning_rate": 7.73506087773439e-06, "loss": 0.563, "step": 42780 }, { "epoch": 1.7585668501931138, "grad_norm": 1.154842734336853, "learning_rate": 7.683823163985737e-06, "loss": 0.5682, "step": 42800 }, { "epoch": 1.7593886104034842, "grad_norm": 1.1190966367721558, "learning_rate": 7.632748935080213e-06, "loss": 0.5896, "step": 42820 }, { "epoch": 1.760210370613855, "grad_norm": 1.064261794090271, "learning_rate": 7.581838281466414e-06, "loss": 0.5778, "step": 42840 }, { "epoch": 1.7610321308242254, "grad_norm": 1.095628261566162, "learning_rate": 7.531091293303094e-06, "loss": 0.5657, "step": 42860 }, { "epoch": 1.7618538910345962, "grad_norm": 1.1514066457748413, "learning_rate": 7.480508060459346e-06, "loss": 0.579, "step": 42880 }, { "epoch": 1.7626756512449666, "grad_norm": 1.1395295858383179, "learning_rate": 7.430088672514124e-06, "loss": 0.6007, "step": 42900 }, { "epoch": 1.7634974114553374, "grad_norm": 1.2230483293533325, "learning_rate": 7.379833218756338e-06, "loss": 0.5996, "step": 42920 }, { "epoch": 1.7643191716657078, "grad_norm": 1.0743821859359741, "learning_rate": 7.329741788184485e-06, "loss": 0.5863, "step": 42940 }, { "epoch": 1.7651409318760787, "grad_norm": 1.1711527109146118, "learning_rate": 7.279814469506652e-06, "loss": 0.5864, "step": 42960 }, { "epoch": 1.765962692086449, "grad_norm": 1.1039119958877563, "learning_rate": 7.230051351140266e-06, "loss": 0.5763, "step": 42980 }, { "epoch": 1.76678445229682, "grad_norm": 1.0589841604232788, "learning_rate": 7.180452521211978e-06, "loss": 0.5857, "step": 43000 }, { "epoch": 1.7676062125071903, "grad_norm": 1.1599675416946411, "learning_rate": 7.131018067557516e-06, "loss": 0.5798, "step": 43020 }, { "epoch": 1.7684279727175611, "grad_norm": 1.10663640499115, "learning_rate": 7.081748077721462e-06, "loss": 0.5749, "step": 43040 }, { "epoch": 1.7692497329279315, "grad_norm": 1.0838004350662231, "learning_rate": 7.032642638957232e-06, "loss": 0.5767, "step": 43060 }, { "epoch": 1.7700714931383024, "grad_norm": 1.0529030561447144, "learning_rate": 6.983701838226708e-06, "loss": 0.6105, "step": 43080 }, { "epoch": 1.7708932533486728, "grad_norm": 1.0537713766098022, "learning_rate": 6.934925762200328e-06, "loss": 0.5857, "step": 43100 }, { "epoch": 1.7717150135590436, "grad_norm": 1.105526328086853, "learning_rate": 6.886314497256752e-06, "loss": 0.5677, "step": 43120 }, { "epoch": 1.772536773769414, "grad_norm": 1.0443006753921509, "learning_rate": 6.837868129482772e-06, "loss": 0.5791, "step": 43140 }, { "epoch": 1.7733585339797848, "grad_norm": 1.0910414457321167, "learning_rate": 6.789586744673226e-06, "loss": 0.5779, "step": 43160 }, { "epoch": 1.7741802941901552, "grad_norm": 1.2543234825134277, "learning_rate": 6.741470428330676e-06, "loss": 0.5898, "step": 43180 }, { "epoch": 1.775002054400526, "grad_norm": 0.9940236806869507, "learning_rate": 6.693519265665449e-06, "loss": 0.5807, "step": 43200 }, { "epoch": 1.7758238146108964, "grad_norm": 1.129135012626648, "learning_rate": 6.645733341595339e-06, "loss": 0.5931, "step": 43220 }, { "epoch": 1.7766455748212673, "grad_norm": 1.0876851081848145, "learning_rate": 6.598112740745544e-06, "loss": 0.587, "step": 43240 }, { "epoch": 1.7774673350316377, "grad_norm": 1.1217765808105469, "learning_rate": 6.550657547448513e-06, "loss": 0.5664, "step": 43260 }, { "epoch": 1.7782890952420085, "grad_norm": 1.006568431854248, "learning_rate": 6.503367845743702e-06, "loss": 0.572, "step": 43280 }, { "epoch": 1.779110855452379, "grad_norm": 1.0781569480895996, "learning_rate": 6.456243719377553e-06, "loss": 0.5824, "step": 43300 }, { "epoch": 1.7799326156627497, "grad_norm": 1.0513370037078857, "learning_rate": 6.4116292395006935e-06, "loss": 0.5924, "step": 43320 }, { "epoch": 1.7807543758731201, "grad_norm": 1.1264588832855225, "learning_rate": 6.364828224809993e-06, "loss": 0.609, "step": 43340 }, { "epoch": 1.781576136083491, "grad_norm": 1.1192883253097534, "learning_rate": 6.318193030800956e-06, "loss": 0.5814, "step": 43360 }, { "epoch": 1.7823978962938614, "grad_norm": 1.1564245223999023, "learning_rate": 6.271723740060908e-06, "loss": 0.5825, "step": 43380 }, { "epoch": 1.7832196565042322, "grad_norm": 1.1175339221954346, "learning_rate": 6.227731656718094e-06, "loss": 0.6158, "step": 43400 }, { "epoch": 1.7840414167146026, "grad_norm": 1.1350170373916626, "learning_rate": 6.1815861137816456e-06, "loss": 0.584, "step": 43420 }, { "epoch": 1.7848631769249734, "grad_norm": 1.0391989946365356, "learning_rate": 6.1356067160345695e-06, "loss": 0.5725, "step": 43440 }, { "epoch": 1.7856849371353438, "grad_norm": 1.0166730880737305, "learning_rate": 6.089793544902756e-06, "loss": 0.5822, "step": 43460 }, { "epoch": 1.7865066973457147, "grad_norm": 1.161569595336914, "learning_rate": 6.0441466815178705e-06, "loss": 0.6067, "step": 43480 }, { "epoch": 1.787328457556085, "grad_norm": 1.1493220329284668, "learning_rate": 5.998666206716985e-06, "loss": 0.5804, "step": 43500 }, { "epoch": 1.7881502177664559, "grad_norm": 1.005832552909851, "learning_rate": 5.953352201042484e-06, "loss": 0.5852, "step": 43520 }, { "epoch": 1.7889719779768263, "grad_norm": 1.1422655582427979, "learning_rate": 5.9082047447420405e-06, "loss": 0.5935, "step": 43540 }, { "epoch": 1.7897937381871971, "grad_norm": 1.0794512033462524, "learning_rate": 5.863223917768268e-06, "loss": 0.5841, "step": 43560 }, { "epoch": 1.7906154983975675, "grad_norm": 0.9649259448051453, "learning_rate": 5.818409799778779e-06, "loss": 0.5813, "step": 43580 }, { "epoch": 1.7914372586079383, "grad_norm": 1.0412386655807495, "learning_rate": 5.7737624701359125e-06, "loss": 0.5912, "step": 43600 }, { "epoch": 1.7922590188183087, "grad_norm": 0.9937067031860352, "learning_rate": 5.729282007906678e-06, "loss": 0.5652, "step": 43620 }, { "epoch": 1.7930807790286796, "grad_norm": 1.1072604656219482, "learning_rate": 5.68496849186253e-06, "loss": 0.5813, "step": 43640 }, { "epoch": 1.79390253923905, "grad_norm": 1.1642615795135498, "learning_rate": 5.640822000479307e-06, "loss": 0.5769, "step": 43660 }, { "epoch": 1.7947242994494208, "grad_norm": 1.1445039510726929, "learning_rate": 5.596842611937025e-06, "loss": 0.5789, "step": 43680 }, { "epoch": 1.7955460596597912, "grad_norm": 1.1920628547668457, "learning_rate": 5.5530304041198075e-06, "loss": 0.5743, "step": 43700 }, { "epoch": 1.796367819870162, "grad_norm": 1.0678682327270508, "learning_rate": 5.509385454615712e-06, "loss": 0.5811, "step": 43720 }, { "epoch": 1.7971895800805324, "grad_norm": 1.15229070186615, "learning_rate": 5.465907840716555e-06, "loss": 0.5835, "step": 43740 }, { "epoch": 1.7980113402909033, "grad_norm": 1.0926685333251953, "learning_rate": 5.422597639417903e-06, "loss": 0.5952, "step": 43760 }, { "epoch": 1.7988331005012737, "grad_norm": 1.0511040687561035, "learning_rate": 5.379454927418714e-06, "loss": 0.5803, "step": 43780 }, { "epoch": 1.7996548607116445, "grad_norm": 1.1438969373703003, "learning_rate": 5.336479781121473e-06, "loss": 0.5866, "step": 43800 }, { "epoch": 1.8004766209220149, "grad_norm": 1.0851142406463623, "learning_rate": 5.293672276631823e-06, "loss": 0.5861, "step": 43820 }, { "epoch": 1.8012983811323857, "grad_norm": 1.071152925491333, "learning_rate": 5.251032489758545e-06, "loss": 0.5965, "step": 43840 }, { "epoch": 1.802120141342756, "grad_norm": 1.1403220891952515, "learning_rate": 5.208560496013471e-06, "loss": 0.5796, "step": 43860 }, { "epoch": 1.802941901553127, "grad_norm": 0.9966292977333069, "learning_rate": 5.166256370611189e-06, "loss": 0.5664, "step": 43880 }, { "epoch": 1.8037636617634973, "grad_norm": 1.014594316482544, "learning_rate": 5.124120188469061e-06, "loss": 0.5889, "step": 43900 }, { "epoch": 1.8045854219738682, "grad_norm": 1.1401876211166382, "learning_rate": 5.082152024207032e-06, "loss": 0.5886, "step": 43920 }, { "epoch": 1.8054071821842386, "grad_norm": 1.0322624444961548, "learning_rate": 5.04035195214747e-06, "loss": 0.5951, "step": 43940 }, { "epoch": 1.8062289423946094, "grad_norm": 1.0889531373977661, "learning_rate": 4.998720046315097e-06, "loss": 0.5795, "step": 43960 }, { "epoch": 1.8070507026049798, "grad_norm": 1.1183452606201172, "learning_rate": 4.957256380436826e-06, "loss": 0.5964, "step": 43980 }, { "epoch": 1.8078724628153506, "grad_norm": 1.037669062614441, "learning_rate": 4.915961027941596e-06, "loss": 0.578, "step": 44000 }, { "epoch": 1.8078724628153506, "eval_loss": 0.8637903928756714, "eval_runtime": 21.3438, "eval_samples_per_second": 122.752, "eval_steps_per_second": 3.842, "step": 44000 }, { "epoch": 1.808694223025721, "grad_norm": 0.8755192160606384, "learning_rate": 4.874834061960298e-06, "loss": 0.4632, "step": 44020 }, { "epoch": 1.8095159832360919, "grad_norm": 0.953360915184021, "learning_rate": 4.83387555532564e-06, "loss": 0.4426, "step": 44040 }, { "epoch": 1.8103377434464623, "grad_norm": 0.9059198498725891, "learning_rate": 4.7930855805719875e-06, "loss": 0.4451, "step": 44060 }, { "epoch": 1.811159503656833, "grad_norm": 0.9322590231895447, "learning_rate": 4.752464209935215e-06, "loss": 0.4425, "step": 44080 }, { "epoch": 1.8119812638672035, "grad_norm": 0.9682320952415466, "learning_rate": 4.712011515352688e-06, "loss": 0.4406, "step": 44100 }, { "epoch": 1.8128030240775743, "grad_norm": 0.9625361561775208, "learning_rate": 4.671727568462958e-06, "loss": 0.4296, "step": 44120 }, { "epoch": 1.8136247842879447, "grad_norm": 0.8894620537757874, "learning_rate": 4.631612440605837e-06, "loss": 0.4459, "step": 44140 }, { "epoch": 1.8144465444983155, "grad_norm": 0.9132700562477112, "learning_rate": 4.5916662028221094e-06, "loss": 0.438, "step": 44160 }, { "epoch": 1.815268304708686, "grad_norm": 1.0488115549087524, "learning_rate": 4.551888925853509e-06, "loss": 0.4485, "step": 44180 }, { "epoch": 1.8160900649190568, "grad_norm": 0.9608358144760132, "learning_rate": 4.512280680142522e-06, "loss": 0.4416, "step": 44200 }, { "epoch": 1.8169118251294272, "grad_norm": 1.0176913738250732, "learning_rate": 4.472841535832295e-06, "loss": 0.4422, "step": 44220 }, { "epoch": 1.817733585339798, "grad_norm": 0.9737944602966309, "learning_rate": 4.433571562766514e-06, "loss": 0.4217, "step": 44240 }, { "epoch": 1.8185553455501684, "grad_norm": 0.9006738662719727, "learning_rate": 4.394470830489272e-06, "loss": 0.4589, "step": 44260 }, { "epoch": 1.8193771057605392, "grad_norm": 1.0624366998672485, "learning_rate": 4.355539408244991e-06, "loss": 0.4506, "step": 44280 }, { "epoch": 1.8201988659709096, "grad_norm": 1.1879520416259766, "learning_rate": 4.316777364978175e-06, "loss": 0.4532, "step": 44300 }, { "epoch": 1.8210206261812805, "grad_norm": 1.1148782968521118, "learning_rate": 4.278184769333482e-06, "loss": 0.4299, "step": 44320 }, { "epoch": 1.8218423863916509, "grad_norm": 0.9817942380905151, "learning_rate": 4.239761689655364e-06, "loss": 0.4249, "step": 44340 }, { "epoch": 1.8226641466020217, "grad_norm": 0.9905684590339661, "learning_rate": 4.201508193988168e-06, "loss": 0.4374, "step": 44360 }, { "epoch": 1.823485906812392, "grad_norm": 1.0558414459228516, "learning_rate": 4.163424350075895e-06, "loss": 0.4501, "step": 44380 }, { "epoch": 1.824307667022763, "grad_norm": 1.105684757232666, "learning_rate": 4.1274018997290775e-06, "loss": 0.438, "step": 44400 }, { "epoch": 1.8251294272331333, "grad_norm": 0.9954769611358643, "learning_rate": 4.089649070449642e-06, "loss": 0.4404, "step": 44420 }, { "epoch": 1.825951187443504, "grad_norm": 1.0505714416503906, "learning_rate": 4.052066091019047e-06, "loss": 0.4338, "step": 44440 }, { "epoch": 1.8267729476538745, "grad_norm": 1.083191990852356, "learning_rate": 4.014653027993797e-06, "loss": 0.4334, "step": 44460 }, { "epoch": 1.8275947078642452, "grad_norm": 1.1424214839935303, "learning_rate": 3.977409947629595e-06, "loss": 0.4414, "step": 44480 }, { "epoch": 1.8284164680746158, "grad_norm": 0.9991270899772644, "learning_rate": 3.9403369158810595e-06, "loss": 0.4419, "step": 44500 }, { "epoch": 1.8292382282849864, "grad_norm": 1.0082736015319824, "learning_rate": 3.903433998401662e-06, "loss": 0.4346, "step": 44520 }, { "epoch": 1.830059988495357, "grad_norm": 1.122463345527649, "learning_rate": 3.866701260543637e-06, "loss": 0.4286, "step": 44540 }, { "epoch": 1.8308817487057276, "grad_norm": 1.15240478515625, "learning_rate": 3.830138767357827e-06, "loss": 0.4248, "step": 44560 }, { "epoch": 1.8317035089160982, "grad_norm": 1.1562724113464355, "learning_rate": 3.7955621469369153e-06, "loss": 0.4267, "step": 44580 }, { "epoch": 1.8325252691264688, "grad_norm": 0.9919096827507019, "learning_rate": 3.7593318168225867e-06, "loss": 0.4265, "step": 44600 }, { "epoch": 1.8333470293368395, "grad_norm": 0.9958898425102234, "learning_rate": 3.723271921523508e-06, "loss": 0.4438, "step": 44620 }, { "epoch": 1.83416878954721, "grad_norm": 1.0441087484359741, "learning_rate": 3.6873825248989524e-06, "loss": 0.4661, "step": 44640 }, { "epoch": 1.8349905497575807, "grad_norm": 1.0398614406585693, "learning_rate": 3.651663690506313e-06, "loss": 0.4385, "step": 44660 }, { "epoch": 1.8358123099679513, "grad_norm": 1.0855858325958252, "learning_rate": 3.616115481600857e-06, "loss": 0.4385, "step": 44680 }, { "epoch": 1.836634070178322, "grad_norm": 1.2031474113464355, "learning_rate": 3.5807379611357826e-06, "loss": 0.4322, "step": 44700 }, { "epoch": 1.8374558303886925, "grad_norm": 1.0220446586608887, "learning_rate": 3.545531191761897e-06, "loss": 0.419, "step": 44720 }, { "epoch": 1.8382775905990631, "grad_norm": 1.1677261590957642, "learning_rate": 3.5104952358277154e-06, "loss": 0.428, "step": 44740 }, { "epoch": 1.8390993508094338, "grad_norm": 1.097439169883728, "learning_rate": 3.475630155379206e-06, "loss": 0.4484, "step": 44760 }, { "epoch": 1.8399211110198044, "grad_norm": 1.115867257118225, "learning_rate": 3.4409360121597235e-06, "loss": 0.4268, "step": 44780 }, { "epoch": 1.840742871230175, "grad_norm": 1.5234203338623047, "learning_rate": 3.406412867609976e-06, "loss": 0.4448, "step": 44800 }, { "epoch": 1.8415646314405456, "grad_norm": 1.0893431901931763, "learning_rate": 3.3720607828677362e-06, "loss": 0.4382, "step": 44820 }, { "epoch": 1.8423863916509162, "grad_norm": 1.0602861642837524, "learning_rate": 3.3378798187679418e-06, "loss": 0.4308, "step": 44840 }, { "epoch": 1.8432081518612868, "grad_norm": 1.249525547027588, "learning_rate": 3.303870035842427e-06, "loss": 0.4261, "step": 44860 }, { "epoch": 1.8440299120716575, "grad_norm": 1.2411805391311646, "learning_rate": 3.270031494319925e-06, "loss": 0.4362, "step": 44880 }, { "epoch": 1.844851672282028, "grad_norm": 1.0888171195983887, "learning_rate": 3.2363642541258676e-06, "loss": 0.4348, "step": 44900 }, { "epoch": 1.8456734324923987, "grad_norm": 1.056647539138794, "learning_rate": 3.2028683748823505e-06, "loss": 0.4356, "step": 44920 }, { "epoch": 1.8464951927027693, "grad_norm": 1.0287542343139648, "learning_rate": 3.169543915907991e-06, "loss": 0.4384, "step": 44940 }, { "epoch": 1.84731695291314, "grad_norm": 1.0829901695251465, "learning_rate": 3.136390936217848e-06, "loss": 0.4452, "step": 44960 }, { "epoch": 1.8481387131235105, "grad_norm": 0.9647024273872375, "learning_rate": 3.1034094945233018e-06, "loss": 0.4303, "step": 44980 }, { "epoch": 1.8489604733338811, "grad_norm": 1.252030611038208, "learning_rate": 3.07059964923192e-06, "loss": 0.454, "step": 45000 }, { "epoch": 1.8497822335442518, "grad_norm": 1.0106678009033203, "learning_rate": 3.037961458447469e-06, "loss": 0.4371, "step": 45020 }, { "epoch": 1.8506039937546224, "grad_norm": 1.0734151601791382, "learning_rate": 3.0054949799696142e-06, "loss": 0.4328, "step": 45040 }, { "epoch": 1.851425753964993, "grad_norm": 1.1433912515640259, "learning_rate": 2.9732002712940187e-06, "loss": 0.4195, "step": 45060 }, { "epoch": 1.8522475141753636, "grad_norm": 1.105094313621521, "learning_rate": 2.9410773896121237e-06, "loss": 0.4369, "step": 45080 }, { "epoch": 1.8530692743857342, "grad_norm": 1.1068464517593384, "learning_rate": 2.909126391811068e-06, "loss": 0.4287, "step": 45100 }, { "epoch": 1.8538910345961048, "grad_norm": 1.1528667211532593, "learning_rate": 2.8773473344736235e-06, "loss": 0.4261, "step": 45120 }, { "epoch": 1.8547127948064754, "grad_norm": 1.1072239875793457, "learning_rate": 2.8457402738780504e-06, "loss": 0.4389, "step": 45140 }, { "epoch": 1.855534555016846, "grad_norm": 0.9646241664886475, "learning_rate": 2.8143052659980185e-06, "loss": 0.4193, "step": 45160 }, { "epoch": 1.8563563152272167, "grad_norm": 1.1033663749694824, "learning_rate": 2.7830423665024862e-06, "loss": 0.4257, "step": 45180 }, { "epoch": 1.8571780754375873, "grad_norm": 1.1620920896530151, "learning_rate": 2.751951630755689e-06, "loss": 0.4328, "step": 45200 }, { "epoch": 1.857999835647958, "grad_norm": 0.9689277410507202, "learning_rate": 2.7210331138168955e-06, "loss": 0.4262, "step": 45220 }, { "epoch": 1.8588215958583285, "grad_norm": 1.0218113660812378, "learning_rate": 2.6902868704404172e-06, "loss": 0.4227, "step": 45240 }, { "epoch": 1.8596433560686991, "grad_norm": 1.1273281574249268, "learning_rate": 2.6597129550754997e-06, "loss": 0.4125, "step": 45260 }, { "epoch": 1.8604651162790697, "grad_norm": 1.068606972694397, "learning_rate": 2.629311421866165e-06, "loss": 0.4241, "step": 45280 }, { "epoch": 1.8612868764894404, "grad_norm": 1.105440616607666, "learning_rate": 2.5990823246512253e-06, "loss": 0.4358, "step": 45300 }, { "epoch": 1.862108636699811, "grad_norm": 1.0128626823425293, "learning_rate": 2.5690257169640688e-06, "loss": 0.4277, "step": 45320 }, { "epoch": 1.8629303969101816, "grad_norm": 0.9848488569259644, "learning_rate": 2.5391416520326284e-06, "loss": 0.4298, "step": 45340 }, { "epoch": 1.8637521571205522, "grad_norm": 1.1279760599136353, "learning_rate": 2.5094301827792933e-06, "loss": 0.4312, "step": 45360 }, { "epoch": 1.8645739173309228, "grad_norm": 1.1564669609069824, "learning_rate": 2.479891361820785e-06, "loss": 0.4386, "step": 45380 }, { "epoch": 1.8653956775412934, "grad_norm": 1.040480375289917, "learning_rate": 2.4505252414680713e-06, "loss": 0.4208, "step": 45400 }, { "epoch": 1.866217437751664, "grad_norm": 1.0698386430740356, "learning_rate": 2.421331873726296e-06, "loss": 0.4353, "step": 45420 }, { "epoch": 1.8670391979620347, "grad_norm": 1.1802546977996826, "learning_rate": 2.3923113102946816e-06, "loss": 0.4328, "step": 45440 }, { "epoch": 1.8678609581724053, "grad_norm": 1.0874077081680298, "learning_rate": 2.363463602566396e-06, "loss": 0.4199, "step": 45460 }, { "epoch": 1.8686827183827759, "grad_norm": 1.2014697790145874, "learning_rate": 2.334788801628518e-06, "loss": 0.4279, "step": 45480 }, { "epoch": 1.8695044785931465, "grad_norm": 1.140428066253662, "learning_rate": 2.3062869582619053e-06, "loss": 0.4294, "step": 45500 }, { "epoch": 1.8703262388035171, "grad_norm": 1.1033835411071777, "learning_rate": 2.277958122941115e-06, "loss": 0.4159, "step": 45520 }, { "epoch": 1.8711479990138877, "grad_norm": 1.0941648483276367, "learning_rate": 2.249802345834373e-06, "loss": 0.4241, "step": 45540 }, { "epoch": 1.8719697592242583, "grad_norm": 1.1285514831542969, "learning_rate": 2.2218196768033496e-06, "loss": 0.421, "step": 45560 }, { "epoch": 1.872791519434629, "grad_norm": 1.0075688362121582, "learning_rate": 2.1940101654032487e-06, "loss": 0.4261, "step": 45580 }, { "epoch": 1.8736132796449996, "grad_norm": 1.2179702520370483, "learning_rate": 2.166373860882509e-06, "loss": 0.4311, "step": 45600 }, { "epoch": 1.8744350398553702, "grad_norm": 1.1928651332855225, "learning_rate": 2.1389108121829593e-06, "loss": 0.413, "step": 45620 }, { "epoch": 1.8752568000657408, "grad_norm": 1.330972671508789, "learning_rate": 2.1116210679395066e-06, "loss": 0.4281, "step": 45640 }, { "epoch": 1.8760785602761114, "grad_norm": 1.1296700239181519, "learning_rate": 2.0845046764801924e-06, "loss": 0.4361, "step": 45660 }, { "epoch": 1.876900320486482, "grad_norm": 1.0228627920150757, "learning_rate": 2.057561685826093e-06, "loss": 0.4562, "step": 45680 }, { "epoch": 1.8777220806968526, "grad_norm": 1.1409462690353394, "learning_rate": 2.030792143691118e-06, "loss": 0.4378, "step": 45700 }, { "epoch": 1.8785438409072233, "grad_norm": 1.078661322593689, "learning_rate": 2.0041960974821027e-06, "loss": 0.4299, "step": 45720 }, { "epoch": 1.8793656011175939, "grad_norm": 1.0520544052124023, "learning_rate": 1.977773594298582e-06, "loss": 0.434, "step": 45740 }, { "epoch": 1.8801873613279645, "grad_norm": 1.0248843431472778, "learning_rate": 1.9515246809327815e-06, "loss": 0.4227, "step": 45760 }, { "epoch": 1.881009121538335, "grad_norm": 0.9711484909057617, "learning_rate": 1.926749043146392e-06, "loss": 0.4221, "step": 45780 }, { "epoch": 1.8818308817487057, "grad_norm": 1.135709524154663, "learning_rate": 1.9008387633469904e-06, "loss": 0.4176, "step": 45800 }, { "epoch": 1.8826526419590763, "grad_norm": 1.0329680442810059, "learning_rate": 1.8763849105989163e-06, "loss": 0.4137, "step": 45820 }, { "epoch": 1.883474402169447, "grad_norm": 1.0853397846221924, "learning_rate": 1.8508134388439902e-06, "loss": 0.4235, "step": 45840 }, { "epoch": 1.8842961623798176, "grad_norm": 1.2208008766174316, "learning_rate": 1.8254157817434447e-06, "loss": 0.4487, "step": 45860 }, { "epoch": 1.8851179225901882, "grad_norm": 0.9802207946777344, "learning_rate": 1.8001919842745686e-06, "loss": 0.442, "step": 45880 }, { "epoch": 1.8859396828005588, "grad_norm": 1.0660961866378784, "learning_rate": 1.775142091106774e-06, "loss": 0.4303, "step": 45900 }, { "epoch": 1.8867614430109294, "grad_norm": 1.0748445987701416, "learning_rate": 1.750266146601498e-06, "loss": 0.429, "step": 45920 }, { "epoch": 1.8875832032213, "grad_norm": 1.2010825872421265, "learning_rate": 1.7255641948121127e-06, "loss": 0.4155, "step": 45940 }, { "epoch": 1.8884049634316706, "grad_norm": 1.149553656578064, "learning_rate": 1.7010362794838918e-06, "loss": 0.4232, "step": 45960 }, { "epoch": 1.8892267236420413, "grad_norm": 1.208585262298584, "learning_rate": 1.6766824440538565e-06, "loss": 0.4242, "step": 45980 }, { "epoch": 1.8900484838524119, "grad_norm": 1.1593657732009888, "learning_rate": 1.6525027316507957e-06, "loss": 0.4376, "step": 46000 }, { "epoch": 1.8900484838524119, "eval_loss": 0.9416248798370361, "eval_runtime": 16.5465, "eval_samples_per_second": 158.341, "eval_steps_per_second": 4.956, "step": 46000 }, { "epoch": 1.8908702440627825, "grad_norm": 0.9449489712715149, "learning_rate": 1.6284971850951125e-06, "loss": 0.4212, "step": 46020 }, { "epoch": 1.891692004273153, "grad_norm": 1.046122670173645, "learning_rate": 1.6046658468987897e-06, "loss": 0.4332, "step": 46040 }, { "epoch": 1.8925137644835237, "grad_norm": 1.1397738456726074, "learning_rate": 1.58100875926529e-06, "loss": 0.4395, "step": 46060 }, { "epoch": 1.8933355246938943, "grad_norm": 1.2659273147583008, "learning_rate": 1.557525964089479e-06, "loss": 0.414, "step": 46080 }, { "epoch": 1.894157284904265, "grad_norm": 1.2042860984802246, "learning_rate": 1.5342175029576023e-06, "loss": 0.4244, "step": 46100 }, { "epoch": 1.8949790451146356, "grad_norm": 1.1012762784957886, "learning_rate": 1.51108341714713e-06, "loss": 0.4324, "step": 46120 }, { "epoch": 1.8958008053250062, "grad_norm": 1.114878535270691, "learning_rate": 1.4881237476267574e-06, "loss": 0.4369, "step": 46140 }, { "epoch": 1.8966225655353768, "grad_norm": 1.1757618188858032, "learning_rate": 1.4653385350563043e-06, "loss": 0.4101, "step": 46160 }, { "epoch": 1.8974443257457474, "grad_norm": 1.1672312021255493, "learning_rate": 1.4427278197866045e-06, "loss": 0.4138, "step": 46180 }, { "epoch": 1.898266085956118, "grad_norm": 1.184259057044983, "learning_rate": 1.4202916418595058e-06, "loss": 0.4153, "step": 46200 }, { "epoch": 1.8990878461664886, "grad_norm": 1.2748756408691406, "learning_rate": 1.3980300410077584e-06, "loss": 0.4031, "step": 46220 }, { "epoch": 1.8999096063768592, "grad_norm": 1.1279730796813965, "learning_rate": 1.3759430566549269e-06, "loss": 0.4296, "step": 46240 }, { "epoch": 1.9007313665872299, "grad_norm": 1.1729366779327393, "learning_rate": 1.3540307279153674e-06, "loss": 0.4261, "step": 46260 }, { "epoch": 1.9015531267976005, "grad_norm": 1.0671958923339844, "learning_rate": 1.3322930935941502e-06, "loss": 0.4218, "step": 46280 }, { "epoch": 1.902374887007971, "grad_norm": 1.169188380241394, "learning_rate": 1.3107301921869152e-06, "loss": 0.4216, "step": 46300 }, { "epoch": 1.9031966472183417, "grad_norm": 1.2280913591384888, "learning_rate": 1.289342061879928e-06, "loss": 0.4149, "step": 46320 }, { "epoch": 1.9040184074287123, "grad_norm": 1.1291908025741577, "learning_rate": 1.268128740549912e-06, "loss": 0.4208, "step": 46340 }, { "epoch": 1.904840167639083, "grad_norm": 1.114375352859497, "learning_rate": 1.2470902657640172e-06, "loss": 0.4061, "step": 46360 }, { "epoch": 1.9056619278494535, "grad_norm": 1.1792467832565308, "learning_rate": 1.2262266747797847e-06, "loss": 0.4135, "step": 46380 }, { "epoch": 1.9064836880598242, "grad_norm": 1.1172113418579102, "learning_rate": 1.2055380045450038e-06, "loss": 0.4259, "step": 46400 }, { "epoch": 1.9073054482701948, "grad_norm": 1.136423945426941, "learning_rate": 1.1850242916977449e-06, "loss": 0.4274, "step": 46420 }, { "epoch": 1.9081272084805654, "grad_norm": 1.257373332977295, "learning_rate": 1.164685572566182e-06, "loss": 0.421, "step": 46440 }, { "epoch": 1.908948968690936, "grad_norm": 1.100024938583374, "learning_rate": 1.1445218831686367e-06, "loss": 0.4245, "step": 46460 }, { "epoch": 1.9097707289013066, "grad_norm": 1.2640262842178345, "learning_rate": 1.1245332592134562e-06, "loss": 0.4188, "step": 46480 }, { "epoch": 1.9105924891116772, "grad_norm": 1.1880619525909424, "learning_rate": 1.104719736098958e-06, "loss": 0.4258, "step": 46500 }, { "epoch": 1.9114142493220478, "grad_norm": 1.1756030321121216, "learning_rate": 1.0850813489133528e-06, "loss": 0.4242, "step": 46520 }, { "epoch": 1.9122360095324185, "grad_norm": 1.2056884765625, "learning_rate": 1.0656181324347093e-06, "loss": 0.4232, "step": 46540 }, { "epoch": 1.913057769742789, "grad_norm": 1.0848861932754517, "learning_rate": 1.0463301211309006e-06, "loss": 0.422, "step": 46560 }, { "epoch": 1.9138795299531597, "grad_norm": 1.2697360515594482, "learning_rate": 1.027217349159504e-06, "loss": 0.4369, "step": 46580 }, { "epoch": 1.9147012901635303, "grad_norm": 1.1444509029388428, "learning_rate": 1.008279850367766e-06, "loss": 0.4231, "step": 46600 }, { "epoch": 1.915523050373901, "grad_norm": 1.225815773010254, "learning_rate": 9.895176582925492e-07, "loss": 0.4063, "step": 46620 }, { "epoch": 1.9163448105842715, "grad_norm": 1.1986970901489258, "learning_rate": 9.709308061602417e-07, "loss": 0.4194, "step": 46640 }, { "epoch": 1.9171665707946421, "grad_norm": 1.0552966594696045, "learning_rate": 9.525193268867249e-07, "loss": 0.4368, "step": 46660 }, { "epoch": 1.9179883310050128, "grad_norm": 1.2278395891189575, "learning_rate": 9.342832530773061e-07, "loss": 0.4247, "step": 46680 }, { "epoch": 1.9188100912153834, "grad_norm": 1.1720478534698486, "learning_rate": 9.162226170266964e-07, "loss": 0.4484, "step": 46700 }, { "epoch": 1.919631851425754, "grad_norm": 1.0597708225250244, "learning_rate": 8.983374507188669e-07, "loss": 0.4158, "step": 46720 }, { "epoch": 1.9204536116361246, "grad_norm": 1.2956316471099854, "learning_rate": 8.806277858270706e-07, "loss": 0.4274, "step": 46740 }, { "epoch": 1.9212753718464952, "grad_norm": 0.9802606701850891, "learning_rate": 8.630936537137757e-07, "loss": 0.4307, "step": 46760 }, { "epoch": 1.9220971320568658, "grad_norm": 1.118882417678833, "learning_rate": 8.457350854305657e-07, "loss": 0.415, "step": 46780 }, { "epoch": 1.9229188922672364, "grad_norm": 1.2108888626098633, "learning_rate": 8.285521117181394e-07, "loss": 0.412, "step": 46800 }, { "epoch": 1.923740652477607, "grad_norm": 1.1903843879699707, "learning_rate": 8.115447630062112e-07, "loss": 0.4167, "step": 46820 }, { "epoch": 1.9245624126879777, "grad_norm": 1.017788052558899, "learning_rate": 7.947130694134996e-07, "loss": 0.4249, "step": 46840 }, { "epoch": 1.9253841728983483, "grad_norm": 1.0981863737106323, "learning_rate": 7.780570607476278e-07, "loss": 0.4072, "step": 46860 }, { "epoch": 1.926205933108719, "grad_norm": 1.1456172466278076, "learning_rate": 7.615767665051122e-07, "loss": 0.4233, "step": 46880 }, { "epoch": 1.9270276933190895, "grad_norm": 1.1544848680496216, "learning_rate": 7.452722158712732e-07, "loss": 0.4324, "step": 46900 }, { "epoch": 1.9278494535294601, "grad_norm": 1.125557780265808, "learning_rate": 7.291434377202255e-07, "loss": 0.4248, "step": 46920 }, { "epoch": 1.9286712137398307, "grad_norm": 0.9662781357765198, "learning_rate": 7.139839337578336e-07, "loss": 0.4405, "step": 46940 }, { "epoch": 1.9294929739502014, "grad_norm": 1.2945139408111572, "learning_rate": 6.981979938184835e-07, "loss": 0.4191, "step": 46960 }, { "epoch": 1.930314734160572, "grad_norm": 1.1728248596191406, "learning_rate": 6.825879097267507e-07, "loss": 0.4314, "step": 46980 }, { "epoch": 1.9311364943709426, "grad_norm": 1.1901942491531372, "learning_rate": 6.671537091268775e-07, "loss": 0.4168, "step": 47000 }, { "epoch": 1.9319582545813132, "grad_norm": 1.1235634088516235, "learning_rate": 6.526541555399446e-07, "loss": 0.422, "step": 47020 }, { "epoch": 1.9327800147916838, "grad_norm": 1.1675294637680054, "learning_rate": 6.375630060813875e-07, "loss": 0.4172, "step": 47040 }, { "epoch": 1.9336017750020544, "grad_norm": 1.1512994766235352, "learning_rate": 6.226478198503527e-07, "loss": 0.4234, "step": 47060 }, { "epoch": 1.934423535212425, "grad_norm": 1.2757488489151, "learning_rate": 6.07908623260467e-07, "loss": 0.4237, "step": 47080 }, { "epoch": 1.9352452954227957, "grad_norm": 1.087795376777649, "learning_rate": 5.933454424137175e-07, "loss": 0.4213, "step": 47100 }, { "epoch": 1.9360670556331663, "grad_norm": 1.1851129531860352, "learning_rate": 5.789583031003743e-07, "loss": 0.4231, "step": 47120 }, { "epoch": 1.936888815843537, "grad_norm": 1.0666377544403076, "learning_rate": 5.647472307989676e-07, "loss": 0.4154, "step": 47140 }, { "epoch": 1.9377105760539075, "grad_norm": 1.0941452980041504, "learning_rate": 5.507122506761886e-07, "loss": 0.4094, "step": 47160 }, { "epoch": 1.9385323362642781, "grad_norm": 1.3112642765045166, "learning_rate": 5.368533875869331e-07, "loss": 0.4098, "step": 47180 }, { "epoch": 1.9393540964746487, "grad_norm": 0.9906901121139526, "learning_rate": 5.231706660741575e-07, "loss": 0.4294, "step": 47200 }, { "epoch": 1.9401758566850194, "grad_norm": 1.081827998161316, "learning_rate": 5.096641103689348e-07, "loss": 0.4079, "step": 47220 }, { "epoch": 1.94099761689539, "grad_norm": 1.2195242643356323, "learning_rate": 4.963337443902982e-07, "loss": 0.4213, "step": 47240 }, { "epoch": 1.9418193771057606, "grad_norm": 1.2033476829528809, "learning_rate": 4.831795917453418e-07, "loss": 0.4276, "step": 47260 }, { "epoch": 1.9426411373161312, "grad_norm": 1.0639315843582153, "learning_rate": 4.702016757290206e-07, "loss": 0.411, "step": 47280 }, { "epoch": 1.9434628975265018, "grad_norm": 1.079074501991272, "learning_rate": 4.5740001932425e-07, "loss": 0.4235, "step": 47300 }, { "epoch": 1.9442846577368724, "grad_norm": 1.1713237762451172, "learning_rate": 4.44774645201762e-07, "loss": 0.4091, "step": 47320 }, { "epoch": 1.945106417947243, "grad_norm": 1.232969045639038, "learning_rate": 4.323255757201383e-07, "loss": 0.4075, "step": 47340 }, { "epoch": 1.9459281781576137, "grad_norm": 1.1638152599334717, "learning_rate": 4.200528329257103e-07, "loss": 0.4281, "step": 47360 }, { "epoch": 1.9467499383679843, "grad_norm": 1.2757538557052612, "learning_rate": 4.0795643855255914e-07, "loss": 0.4385, "step": 47380 }, { "epoch": 1.9475716985783549, "grad_norm": 1.08627450466156, "learning_rate": 3.960364140224826e-07, "loss": 0.429, "step": 47400 }, { "epoch": 1.9483934587887255, "grad_norm": 1.2395083904266357, "learning_rate": 3.842927804449059e-07, "loss": 0.4194, "step": 47420 }, { "epoch": 1.949215218999096, "grad_norm": 1.2743449211120605, "learning_rate": 3.727255586169265e-07, "loss": 0.4343, "step": 47440 }, { "epoch": 1.9500369792094667, "grad_norm": 1.1402217149734497, "learning_rate": 3.6133476902318055e-07, "loss": 0.4113, "step": 47460 }, { "epoch": 1.9508587394198373, "grad_norm": 1.2253342866897583, "learning_rate": 3.501204318358875e-07, "loss": 0.4159, "step": 47480 }, { "epoch": 1.951680499630208, "grad_norm": 1.1899055242538452, "learning_rate": 3.390825669147724e-07, "loss": 0.4185, "step": 47500 }, { "epoch": 1.9525022598405786, "grad_norm": 1.0581785440444946, "learning_rate": 3.282211938070545e-07, "loss": 0.407, "step": 47520 }, { "epoch": 1.9533240200509492, "grad_norm": 1.143174171447754, "learning_rate": 3.1753633174738116e-07, "loss": 0.3927, "step": 47540 }, { "epoch": 1.9541457802613198, "grad_norm": 1.1011602878570557, "learning_rate": 3.070279996578163e-07, "loss": 0.4089, "step": 47560 }, { "epoch": 1.9549675404716904, "grad_norm": 1.1266615390777588, "learning_rate": 2.966962161478182e-07, "loss": 0.4117, "step": 47580 }, { "epoch": 1.955789300682061, "grad_norm": 1.152133584022522, "learning_rate": 2.8654099951417324e-07, "loss": 0.4287, "step": 47600 }, { "epoch": 1.9566110608924316, "grad_norm": 1.1630717515945435, "learning_rate": 2.7656236774099564e-07, "loss": 0.4353, "step": 47620 }, { "epoch": 1.9574328211028023, "grad_norm": 1.036206603050232, "learning_rate": 2.6676033849966087e-07, "loss": 0.413, "step": 47640 }, { "epoch": 1.9582545813131729, "grad_norm": 1.1259959936141968, "learning_rate": 2.5713492914881677e-07, "loss": 0.433, "step": 47660 }, { "epoch": 1.9590763415235435, "grad_norm": 1.2203707695007324, "learning_rate": 2.4768615673432806e-07, "loss": 0.4102, "step": 47680 }, { "epoch": 1.959898101733914, "grad_norm": 1.2147117853164673, "learning_rate": 2.384140379892319e-07, "loss": 0.4058, "step": 47700 }, { "epoch": 1.9607198619442845, "grad_norm": 1.1942362785339355, "learning_rate": 2.2931858933374907e-07, "loss": 0.418, "step": 47720 }, { "epoch": 1.9615416221546553, "grad_norm": 1.0480471849441528, "learning_rate": 2.203998268752061e-07, "loss": 0.4061, "step": 47740 }, { "epoch": 1.9623633823650257, "grad_norm": 1.2805067300796509, "learning_rate": 2.1165776640804658e-07, "loss": 0.4289, "step": 47760 }, { "epoch": 1.9631851425753966, "grad_norm": 1.2007514238357544, "learning_rate": 2.0309242341379765e-07, "loss": 0.4283, "step": 47780 }, { "epoch": 1.964006902785767, "grad_norm": 1.1871360540390015, "learning_rate": 1.947038130610146e-07, "loss": 0.4089, "step": 47800 }, { "epoch": 1.9648286629961378, "grad_norm": 1.1369190216064453, "learning_rate": 1.8649195020528087e-07, "loss": 0.4045, "step": 47820 }, { "epoch": 1.9656504232065082, "grad_norm": 1.255011796951294, "learning_rate": 1.7845684938916363e-07, "loss": 0.4226, "step": 47840 }, { "epoch": 1.966472183416879, "grad_norm": 1.2561933994293213, "learning_rate": 1.7059852484223593e-07, "loss": 0.4188, "step": 47860 }, { "epoch": 1.9672939436272494, "grad_norm": 1.1774425506591797, "learning_rate": 1.6291699048095464e-07, "loss": 0.42, "step": 47880 }, { "epoch": 1.9681157038376202, "grad_norm": 1.0992674827575684, "learning_rate": 1.554122599087493e-07, "loss": 0.4136, "step": 47900 }, { "epoch": 1.9689374640479906, "grad_norm": 1.1646844148635864, "learning_rate": 1.4808434641589985e-07, "loss": 0.4238, "step": 47920 }, { "epoch": 1.9697592242583615, "grad_norm": 1.1003245115280151, "learning_rate": 1.409332629795923e-07, "loss": 0.3955, "step": 47940 }, { "epoch": 1.9705809844687319, "grad_norm": 1.0711121559143066, "learning_rate": 1.3395902226384093e-07, "loss": 0.4218, "step": 47960 }, { "epoch": 1.9714027446791027, "grad_norm": 1.0575335025787354, "learning_rate": 1.2716163661948833e-07, "loss": 0.4218, "step": 47980 }, { "epoch": 1.972224504889473, "grad_norm": 1.1879905462265015, "learning_rate": 1.205411180841831e-07, "loss": 0.4031, "step": 48000 }, { "epoch": 1.972224504889473, "eval_loss": 0.9563899636268616, "eval_runtime": 16.5443, "eval_samples_per_second": 158.363, "eval_steps_per_second": 4.956, "step": 48000 }, { "epoch": 1.973046265099844, "grad_norm": 1.0687421560287476, "learning_rate": 1.1409747838234675e-07, "loss": 0.4245, "step": 48020 }, { "epoch": 1.9738680253102143, "grad_norm": 1.2620290517807007, "learning_rate": 1.0783072892518454e-07, "loss": 0.4032, "step": 48040 }, { "epoch": 1.9746897855205852, "grad_norm": 1.201851487159729, "learning_rate": 1.0174088081060795e-07, "loss": 0.416, "step": 48060 }, { "epoch": 1.9755115457309556, "grad_norm": 1.2505393028259277, "learning_rate": 9.611938979670054e-08, "loss": 0.4171, "step": 48080 }, { "epoch": 1.9763333059413264, "grad_norm": 1.2700996398925781, "learning_rate": 9.037453003418739e-08, "loss": 0.4151, "step": 48100 }, { "epoch": 1.9771550661516968, "grad_norm": 1.1103131771087646, "learning_rate": 8.480660252785378e-08, "loss": 0.4188, "step": 48120 }, { "epoch": 1.9779768263620676, "grad_norm": 1.0756477117538452, "learning_rate": 7.941561713806777e-08, "loss": 0.4078, "step": 48140 }, { "epoch": 1.978798586572438, "grad_norm": 1.1716761589050293, "learning_rate": 7.420158341185924e-08, "loss": 0.4149, "step": 48160 }, { "epoch": 1.9796203467828088, "grad_norm": 1.0660831928253174, "learning_rate": 6.916451058286422e-08, "loss": 0.4208, "step": 48180 }, { "epoch": 1.9804421069931792, "grad_norm": 1.1756824254989624, "learning_rate": 6.430440757136946e-08, "loss": 0.4029, "step": 48200 }, { "epoch": 1.98126386720355, "grad_norm": 1.0613312721252441, "learning_rate": 5.962128298423463e-08, "loss": 0.4096, "step": 48220 }, { "epoch": 1.9820856274139205, "grad_norm": 1.1064634323120117, "learning_rate": 5.533624844936691e-08, "loss": 0.4193, "step": 48240 }, { "epoch": 1.9829073876242913, "grad_norm": 1.0227088928222656, "learning_rate": 5.099825535829217e-08, "loss": 0.4216, "step": 48260 }, { "epoch": 1.9837291478346617, "grad_norm": 1.2082315683364868, "learning_rate": 4.683726425576129e-08, "loss": 0.4266, "step": 48280 }, { "epoch": 1.9845509080450325, "grad_norm": 1.2815755605697632, "learning_rate": 4.2853282510546545e-08, "loss": 0.3935, "step": 48300 }, { "epoch": 1.985372668255403, "grad_norm": 1.1386293172836304, "learning_rate": 3.904631717798202e-08, "loss": 0.4065, "step": 48320 }, { "epoch": 1.9861944284657738, "grad_norm": 1.1509379148483276, "learning_rate": 3.5416374999919235e-08, "loss": 0.4228, "step": 48340 }, { "epoch": 1.9870161886761442, "grad_norm": 1.1019117832183838, "learning_rate": 3.1963462404693835e-08, "loss": 0.4147, "step": 48360 }, { "epoch": 1.987837948886515, "grad_norm": 1.0444250106811523, "learning_rate": 2.8687585507147784e-08, "loss": 0.4276, "step": 48380 }, { "epoch": 1.9886597090968854, "grad_norm": 1.1302378177642822, "learning_rate": 2.5588750108618275e-08, "loss": 0.4149, "step": 48400 }, { "epoch": 1.9894814693072562, "grad_norm": 1.082972764968872, "learning_rate": 2.2666961696893308e-08, "loss": 0.4388, "step": 48420 }, { "epoch": 1.9903032295176266, "grad_norm": 1.1898425817489624, "learning_rate": 1.9922225446245e-08, "loss": 0.3988, "step": 48440 }, { "epoch": 1.9911249897279975, "grad_norm": 1.2624133825302124, "learning_rate": 1.7354546217385192e-08, "loss": 0.4062, "step": 48460 }, { "epoch": 1.9919467499383678, "grad_norm": 1.1594356298446655, "learning_rate": 1.4963928557465425e-08, "loss": 0.4142, "step": 48480 }, { "epoch": 1.9927685101487387, "grad_norm": 1.2006750106811523, "learning_rate": 1.2750376700099153e-08, "loss": 0.4203, "step": 48500 }, { "epoch": 1.993590270359109, "grad_norm": 1.0823673009872437, "learning_rate": 1.0713894565317351e-08, "loss": 0.421, "step": 48520 }, { "epoch": 1.99441203056948, "grad_norm": 1.067572832107544, "learning_rate": 8.854485759568487e-09, "loss": 0.4103, "step": 48540 }, { "epoch": 1.9952337907798503, "grad_norm": 1.3517385721206665, "learning_rate": 7.172153575718543e-09, "loss": 0.42, "step": 48560 }, { "epoch": 1.9960555509902211, "grad_norm": 1.1812193393707275, "learning_rate": 5.666900993062107e-09, "loss": 0.4011, "step": 48580 }, { "epoch": 1.9968773112005915, "grad_norm": 1.1128283739089966, "learning_rate": 4.338730677266867e-09, "loss": 0.4091, "step": 48600 }, { "epoch": 1.9976990714109624, "grad_norm": 1.2038655281066895, "learning_rate": 3.1876449804291163e-09, "loss": 0.4236, "step": 48620 }, { "epoch": 1.9985208316213328, "grad_norm": 1.2628865242004395, "learning_rate": 2.213645941029352e-09, "loss": 0.4098, "step": 48640 }, { "epoch": 1.9993425918317036, "grad_norm": 0.9929442405700684, "learning_rate": 1.4167352839544735e-09, "loss": 0.4198, "step": 48660 } ], "logging_steps": 20, "max_steps": 48676, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.4754061349278974e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }