{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.978768577494693, "eval_steps": 500, "global_step": 2350, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004246284501061571, "grad_norm": 3.09375, "learning_rate": 8.510638297872341e-07, "loss": 1.9544, "step": 1 }, { "epoch": 0.021231422505307854, "grad_norm": 4.03125, "learning_rate": 4.255319148936171e-06, "loss": 1.9708, "step": 5 }, { "epoch": 0.04246284501061571, "grad_norm": 2.09375, "learning_rate": 8.510638297872341e-06, "loss": 1.9597, "step": 10 }, { "epoch": 0.06369426751592357, "grad_norm": 8.25, "learning_rate": 1.2765957446808511e-05, "loss": 1.9264, "step": 15 }, { "epoch": 0.08492569002123142, "grad_norm": 1.6953125, "learning_rate": 1.7021276595744682e-05, "loss": 1.8338, "step": 20 }, { "epoch": 0.10615711252653928, "grad_norm": 2.453125, "learning_rate": 2.1276595744680852e-05, "loss": 1.7511, "step": 25 }, { "epoch": 0.12738853503184713, "grad_norm": 2.609375, "learning_rate": 2.5531914893617022e-05, "loss": 1.6446, "step": 30 }, { "epoch": 0.14861995753715498, "grad_norm": 2.4375, "learning_rate": 2.9787234042553192e-05, "loss": 1.5726, "step": 35 }, { "epoch": 0.16985138004246284, "grad_norm": 6.21875, "learning_rate": 3.4042553191489365e-05, "loss": 1.4611, "step": 40 }, { "epoch": 0.1910828025477707, "grad_norm": 5.03125, "learning_rate": 3.829787234042553e-05, "loss": 1.3599, "step": 45 }, { "epoch": 0.21231422505307856, "grad_norm": 1.03125, "learning_rate": 4.2553191489361704e-05, "loss": 1.2565, "step": 50 }, { "epoch": 0.23354564755838642, "grad_norm": 0.75, "learning_rate": 4.680851063829788e-05, "loss": 1.1707, "step": 55 }, { "epoch": 0.25477707006369427, "grad_norm": 1.34375, "learning_rate": 5.1063829787234044e-05, "loss": 1.0829, "step": 60 }, { "epoch": 0.2760084925690021, "grad_norm": 0.578125, "learning_rate": 5.531914893617022e-05, "loss": 0.9955, "step": 65 }, { "epoch": 0.29723991507430997, "grad_norm": 0.3046875, "learning_rate": 5.9574468085106384e-05, "loss": 0.9635, "step": 70 }, { "epoch": 0.3184713375796178, "grad_norm": 0.29296875, "learning_rate": 6.382978723404256e-05, "loss": 0.9258, "step": 75 }, { "epoch": 0.33970276008492567, "grad_norm": 0.8515625, "learning_rate": 6.808510638297873e-05, "loss": 0.8948, "step": 80 }, { "epoch": 0.3609341825902335, "grad_norm": 0.6484375, "learning_rate": 7.23404255319149e-05, "loss": 0.86, "step": 85 }, { "epoch": 0.3821656050955414, "grad_norm": 0.77734375, "learning_rate": 7.659574468085106e-05, "loss": 0.8485, "step": 90 }, { "epoch": 0.4033970276008493, "grad_norm": 0.44921875, "learning_rate": 8.085106382978723e-05, "loss": 0.8287, "step": 95 }, { "epoch": 0.42462845010615713, "grad_norm": 0.2734375, "learning_rate": 8.510638297872341e-05, "loss": 0.8144, "step": 100 }, { "epoch": 0.445859872611465, "grad_norm": 0.4453125, "learning_rate": 8.936170212765958e-05, "loss": 0.7988, "step": 105 }, { "epoch": 0.46709129511677283, "grad_norm": 0.58203125, "learning_rate": 9.361702127659576e-05, "loss": 0.7956, "step": 110 }, { "epoch": 0.4883227176220807, "grad_norm": 0.53125, "learning_rate": 9.787234042553192e-05, "loss": 0.7839, "step": 115 }, { "epoch": 0.5095541401273885, "grad_norm": 0.51953125, "learning_rate": 0.00010212765957446809, "loss": 0.7604, "step": 120 }, { "epoch": 0.5307855626326964, "grad_norm": 0.310546875, "learning_rate": 0.00010638297872340425, "loss": 0.7615, "step": 125 }, { "epoch": 0.5520169851380042, "grad_norm": 1.421875, "learning_rate": 0.00011063829787234043, "loss": 0.7599, "step": 130 }, { "epoch": 0.5732484076433121, "grad_norm": 1.203125, "learning_rate": 0.00011489361702127661, "loss": 0.7407, "step": 135 }, { "epoch": 0.5944798301486199, "grad_norm": 1.1640625, "learning_rate": 0.00011914893617021277, "loss": 0.7518, "step": 140 }, { "epoch": 0.6157112526539278, "grad_norm": 0.380859375, "learning_rate": 0.00012340425531914893, "loss": 0.7483, "step": 145 }, { "epoch": 0.6369426751592356, "grad_norm": 0.546875, "learning_rate": 0.00012765957446808513, "loss": 0.7278, "step": 150 }, { "epoch": 0.6581740976645435, "grad_norm": 1.0390625, "learning_rate": 0.00013191489361702127, "loss": 0.7319, "step": 155 }, { "epoch": 0.6794055201698513, "grad_norm": 1.109375, "learning_rate": 0.00013617021276595746, "loss": 0.731, "step": 160 }, { "epoch": 0.7006369426751592, "grad_norm": 0.578125, "learning_rate": 0.00014042553191489363, "loss": 0.7207, "step": 165 }, { "epoch": 0.721868365180467, "grad_norm": 0.3359375, "learning_rate": 0.0001446808510638298, "loss": 0.7271, "step": 170 }, { "epoch": 0.7430997876857749, "grad_norm": 0.462890625, "learning_rate": 0.00014893617021276596, "loss": 0.7111, "step": 175 }, { "epoch": 0.7643312101910829, "grad_norm": 0.59765625, "learning_rate": 0.00015319148936170213, "loss": 0.7099, "step": 180 }, { "epoch": 0.7855626326963907, "grad_norm": 0.5, "learning_rate": 0.00015744680851063832, "loss": 0.7114, "step": 185 }, { "epoch": 0.8067940552016986, "grad_norm": 0.267578125, "learning_rate": 0.00016170212765957446, "loss": 0.7038, "step": 190 }, { "epoch": 0.8280254777070064, "grad_norm": 0.498046875, "learning_rate": 0.00016595744680851065, "loss": 0.7097, "step": 195 }, { "epoch": 0.8492569002123143, "grad_norm": 0.75390625, "learning_rate": 0.00017021276595744682, "loss": 0.7065, "step": 200 }, { "epoch": 0.8704883227176221, "grad_norm": 0.267578125, "learning_rate": 0.00017446808510638298, "loss": 0.6975, "step": 205 }, { "epoch": 0.89171974522293, "grad_norm": 0.74609375, "learning_rate": 0.00017872340425531915, "loss": 0.7038, "step": 210 }, { "epoch": 0.9129511677282378, "grad_norm": 0.8203125, "learning_rate": 0.00018297872340425532, "loss": 0.6929, "step": 215 }, { "epoch": 0.9341825902335457, "grad_norm": 0.265625, "learning_rate": 0.0001872340425531915, "loss": 0.7061, "step": 220 }, { "epoch": 0.9554140127388535, "grad_norm": 0.765625, "learning_rate": 0.00019148936170212768, "loss": 0.7031, "step": 225 }, { "epoch": 0.9766454352441614, "grad_norm": 0.66015625, "learning_rate": 0.00019574468085106384, "loss": 0.6968, "step": 230 }, { "epoch": 0.9978768577494692, "grad_norm": 0.71875, "learning_rate": 0.0002, "loss": 0.6871, "step": 235 }, { "epoch": 0.9978768577494692, "eval_loss": 1.4013111591339111, "eval_runtime": 0.536, "eval_samples_per_second": 9.328, "eval_steps_per_second": 1.866, "step": 235 }, { "epoch": 1.019108280254777, "grad_norm": 0.78125, "learning_rate": 0.00019999724204599747, "loss": 0.6759, "step": 240 }, { "epoch": 1.040339702760085, "grad_norm": 0.54296875, "learning_rate": 0.00019998896833611603, "loss": 0.673, "step": 245 }, { "epoch": 1.0615711252653928, "grad_norm": 0.294921875, "learning_rate": 0.0001999751793267259, "loss": 0.6671, "step": 250 }, { "epoch": 1.0828025477707006, "grad_norm": 0.298828125, "learning_rate": 0.0001999558757784162, "loss": 0.6747, "step": 255 }, { "epoch": 1.1040339702760085, "grad_norm": 0.53515625, "learning_rate": 0.0001999310587559529, "loss": 0.669, "step": 260 }, { "epoch": 1.1252653927813163, "grad_norm": 0.298828125, "learning_rate": 0.00019990072962822007, "loss": 0.6718, "step": 265 }, { "epoch": 1.1464968152866242, "grad_norm": 0.326171875, "learning_rate": 0.00019986489006814452, "loss": 0.6729, "step": 270 }, { "epoch": 1.167728237791932, "grad_norm": 0.28125, "learning_rate": 0.00019982354205260347, "loss": 0.6658, "step": 275 }, { "epoch": 1.1889596602972399, "grad_norm": 0.2431640625, "learning_rate": 0.00019977668786231534, "loss": 0.6694, "step": 280 }, { "epoch": 1.2101910828025477, "grad_norm": 0.2890625, "learning_rate": 0.00019972433008171416, "loss": 0.6727, "step": 285 }, { "epoch": 1.2314225053078556, "grad_norm": 0.306640625, "learning_rate": 0.00019966647159880703, "loss": 0.6608, "step": 290 }, { "epoch": 1.2526539278131634, "grad_norm": 0.765625, "learning_rate": 0.00019960311560501454, "loss": 0.6662, "step": 295 }, { "epoch": 1.2738853503184713, "grad_norm": 0.7421875, "learning_rate": 0.0001995342655949951, "loss": 0.6639, "step": 300 }, { "epoch": 1.2951167728237791, "grad_norm": 0.78125, "learning_rate": 0.00019945992536645187, "loss": 0.6639, "step": 305 }, { "epoch": 1.316348195329087, "grad_norm": 0.4140625, "learning_rate": 0.0001993800990199235, "loss": 0.6651, "step": 310 }, { "epoch": 1.3375796178343948, "grad_norm": 0.421875, "learning_rate": 0.0001992947909585578, "loss": 0.6534, "step": 315 }, { "epoch": 1.3588110403397027, "grad_norm": 0.216796875, "learning_rate": 0.000199204005887869, "loss": 0.6569, "step": 320 }, { "epoch": 1.3800424628450108, "grad_norm": 0.29296875, "learning_rate": 0.000199107748815478, "loss": 0.6577, "step": 325 }, { "epoch": 1.4012738853503186, "grad_norm": 0.21875, "learning_rate": 0.00019900602505083648, "loss": 0.6524, "step": 330 }, { "epoch": 1.4225053078556265, "grad_norm": 0.23046875, "learning_rate": 0.0001988988402049336, "loss": 0.6531, "step": 335 }, { "epoch": 1.4437367303609343, "grad_norm": 0.216796875, "learning_rate": 0.00019878620018998696, "loss": 0.6544, "step": 340 }, { "epoch": 1.4649681528662422, "grad_norm": 0.25390625, "learning_rate": 0.00019866811121911607, "loss": 0.6462, "step": 345 }, { "epoch": 1.48619957537155, "grad_norm": 0.3984375, "learning_rate": 0.000198544579806, "loss": 0.6544, "step": 350 }, { "epoch": 1.5074309978768579, "grad_norm": 0.51171875, "learning_rate": 0.0001984156127645178, "loss": 0.6337, "step": 355 }, { "epoch": 1.5286624203821657, "grad_norm": 0.412109375, "learning_rate": 0.00019828121720837286, "loss": 0.6543, "step": 360 }, { "epoch": 1.5498938428874736, "grad_norm": 0.36328125, "learning_rate": 0.00019814140055070042, "loss": 0.647, "step": 365 }, { "epoch": 1.5711252653927814, "grad_norm": 0.271484375, "learning_rate": 0.0001979961705036587, "loss": 0.6368, "step": 370 }, { "epoch": 1.5923566878980893, "grad_norm": 0.33984375, "learning_rate": 0.00019784553507800349, "loss": 0.6504, "step": 375 }, { "epoch": 1.6135881104033971, "grad_norm": 0.65234375, "learning_rate": 0.00019768950258264623, "loss": 0.6427, "step": 380 }, { "epoch": 1.634819532908705, "grad_norm": 0.359375, "learning_rate": 0.0001975280816241959, "loss": 0.6411, "step": 385 }, { "epoch": 1.6560509554140128, "grad_norm": 0.259765625, "learning_rate": 0.00019736128110648407, "loss": 0.6412, "step": 390 }, { "epoch": 1.6772823779193207, "grad_norm": 0.23046875, "learning_rate": 0.0001971891102300738, "loss": 0.6497, "step": 395 }, { "epoch": 1.6985138004246285, "grad_norm": 0.24609375, "learning_rate": 0.00019701157849175228, "loss": 0.6419, "step": 400 }, { "epoch": 1.7197452229299364, "grad_norm": 0.2294921875, "learning_rate": 0.00019682869568400684, "loss": 0.6521, "step": 405 }, { "epoch": 1.7409766454352442, "grad_norm": 0.2470703125, "learning_rate": 0.00019664047189448493, "loss": 0.641, "step": 410 }, { "epoch": 1.762208067940552, "grad_norm": 0.431640625, "learning_rate": 0.00019644691750543767, "loss": 0.6522, "step": 415 }, { "epoch": 1.78343949044586, "grad_norm": 0.46875, "learning_rate": 0.00019624804319314705, "loss": 0.6581, "step": 420 }, { "epoch": 1.8046709129511678, "grad_norm": 0.2275390625, "learning_rate": 0.00019604385992733715, "loss": 0.6452, "step": 425 }, { "epoch": 1.8259023354564756, "grad_norm": 0.546875, "learning_rate": 0.00019583437897056915, "loss": 0.6368, "step": 430 }, { "epoch": 1.8471337579617835, "grad_norm": 0.447265625, "learning_rate": 0.00019561961187761985, "loss": 0.6457, "step": 435 }, { "epoch": 1.8683651804670913, "grad_norm": 0.48046875, "learning_rate": 0.00019539957049484458, "loss": 0.6277, "step": 440 }, { "epoch": 1.8895966029723992, "grad_norm": 0.205078125, "learning_rate": 0.00019517426695952358, "loss": 0.6305, "step": 445 }, { "epoch": 1.910828025477707, "grad_norm": 0.66796875, "learning_rate": 0.0001949437136991925, "loss": 0.6329, "step": 450 }, { "epoch": 1.9320594479830149, "grad_norm": 0.34765625, "learning_rate": 0.00019470792343095718, "loss": 0.6454, "step": 455 }, { "epoch": 1.9532908704883227, "grad_norm": 0.55078125, "learning_rate": 0.0001944669091607919, "loss": 0.6292, "step": 460 }, { "epoch": 1.9745222929936306, "grad_norm": 2.125, "learning_rate": 0.00019422068418282202, "loss": 0.6342, "step": 465 }, { "epoch": 1.9957537154989384, "grad_norm": 3.875, "learning_rate": 0.00019396926207859084, "loss": 0.6707, "step": 470 }, { "epoch": 2.0, "eval_loss": 1.399332880973816, "eval_runtime": 0.4998, "eval_samples_per_second": 10.003, "eval_steps_per_second": 2.001, "step": 471 }, { "epoch": 2.0169851380042463, "grad_norm": 0.33203125, "learning_rate": 0.00019371265671631037, "loss": 0.6274, "step": 475 }, { "epoch": 2.038216560509554, "grad_norm": 0.4453125, "learning_rate": 0.00019345088225009626, "loss": 0.6217, "step": 480 }, { "epoch": 2.059447983014862, "grad_norm": 0.2314453125, "learning_rate": 0.0001931839531191873, "loss": 0.6253, "step": 485 }, { "epoch": 2.08067940552017, "grad_norm": 0.2421875, "learning_rate": 0.00019291188404714878, "loss": 0.619, "step": 490 }, { "epoch": 2.1019108280254777, "grad_norm": 0.24609375, "learning_rate": 0.0001926346900410604, "loss": 0.6275, "step": 495 }, { "epoch": 2.1231422505307855, "grad_norm": 0.212890625, "learning_rate": 0.00019235238639068856, "loss": 0.6175, "step": 500 }, { "epoch": 2.1443736730360934, "grad_norm": 0.1953125, "learning_rate": 0.00019206498866764288, "loss": 0.6188, "step": 505 }, { "epoch": 2.1656050955414012, "grad_norm": 0.224609375, "learning_rate": 0.0001917725127245174, "loss": 0.6104, "step": 510 }, { "epoch": 2.186836518046709, "grad_norm": 0.265625, "learning_rate": 0.0001914749746940161, "loss": 0.6075, "step": 515 }, { "epoch": 2.208067940552017, "grad_norm": 0.201171875, "learning_rate": 0.00019117239098806295, "loss": 0.6254, "step": 520 }, { "epoch": 2.229299363057325, "grad_norm": 0.337890625, "learning_rate": 0.00019086477829689685, "loss": 0.617, "step": 525 }, { "epoch": 2.2505307855626326, "grad_norm": 0.271484375, "learning_rate": 0.0001905521535881509, "loss": 0.6188, "step": 530 }, { "epoch": 2.2717622080679405, "grad_norm": 0.22265625, "learning_rate": 0.00019023453410591635, "loss": 0.6167, "step": 535 }, { "epoch": 2.2929936305732483, "grad_norm": 0.296875, "learning_rate": 0.00018991193736979175, "loss": 0.6048, "step": 540 }, { "epoch": 2.314225053078556, "grad_norm": 0.185546875, "learning_rate": 0.00018958438117391618, "loss": 0.6092, "step": 545 }, { "epoch": 2.335456475583864, "grad_norm": 0.349609375, "learning_rate": 0.00018925188358598813, "loss": 0.6079, "step": 550 }, { "epoch": 2.356687898089172, "grad_norm": 0.306640625, "learning_rate": 0.00018891446294626866, "loss": 0.6136, "step": 555 }, { "epoch": 2.3779193205944797, "grad_norm": 0.1884765625, "learning_rate": 0.00018857213786656985, "loss": 0.6075, "step": 560 }, { "epoch": 2.3991507430997876, "grad_norm": 0.205078125, "learning_rate": 0.0001882249272292282, "loss": 0.6097, "step": 565 }, { "epoch": 2.4203821656050954, "grad_norm": 0.208984375, "learning_rate": 0.00018787285018606297, "loss": 0.6178, "step": 570 }, { "epoch": 2.4416135881104033, "grad_norm": 0.2353515625, "learning_rate": 0.00018751592615732005, "loss": 0.6099, "step": 575 }, { "epoch": 2.462845010615711, "grad_norm": 0.30859375, "learning_rate": 0.0001871541748306005, "loss": 0.6161, "step": 580 }, { "epoch": 2.484076433121019, "grad_norm": 0.208984375, "learning_rate": 0.00018678761615977468, "loss": 0.616, "step": 585 }, { "epoch": 2.505307855626327, "grad_norm": 0.326171875, "learning_rate": 0.00018641627036388169, "loss": 0.611, "step": 590 }, { "epoch": 2.5265392781316347, "grad_norm": 0.23828125, "learning_rate": 0.00018604015792601396, "loss": 0.6063, "step": 595 }, { "epoch": 2.5477707006369426, "grad_norm": 0.390625, "learning_rate": 0.00018565929959218758, "loss": 0.5991, "step": 600 }, { "epoch": 2.5690021231422504, "grad_norm": 0.255859375, "learning_rate": 0.0001852737163701979, "loss": 0.6085, "step": 605 }, { "epoch": 2.5902335456475583, "grad_norm": 0.2294921875, "learning_rate": 0.00018488342952846073, "loss": 0.6027, "step": 610 }, { "epoch": 2.611464968152866, "grad_norm": 0.2236328125, "learning_rate": 0.0001844884605948392, "loss": 0.6108, "step": 615 }, { "epoch": 2.632696390658174, "grad_norm": 0.33984375, "learning_rate": 0.00018408883135545632, "loss": 0.6156, "step": 620 }, { "epoch": 2.653927813163482, "grad_norm": 0.228515625, "learning_rate": 0.00018368456385349334, "loss": 0.6135, "step": 625 }, { "epoch": 2.6751592356687897, "grad_norm": 0.2177734375, "learning_rate": 0.0001832756803879737, "loss": 0.6064, "step": 630 }, { "epoch": 2.6963906581740975, "grad_norm": 0.2236328125, "learning_rate": 0.0001828622035125332, "loss": 0.6026, "step": 635 }, { "epoch": 2.7176220806794054, "grad_norm": 0.2197265625, "learning_rate": 0.00018244415603417603, "loss": 0.614, "step": 640 }, { "epoch": 2.738853503184713, "grad_norm": 0.26953125, "learning_rate": 0.00018202156101201645, "loss": 0.6176, "step": 645 }, { "epoch": 2.7600849256900215, "grad_norm": 0.2021484375, "learning_rate": 0.00018159444175600703, "loss": 0.6057, "step": 650 }, { "epoch": 2.781316348195329, "grad_norm": 0.2490234375, "learning_rate": 0.00018116282182565311, "loss": 0.6044, "step": 655 }, { "epoch": 2.802547770700637, "grad_norm": 0.19921875, "learning_rate": 0.00018072672502871296, "loss": 0.5988, "step": 660 }, { "epoch": 2.8237791932059446, "grad_norm": 0.20703125, "learning_rate": 0.00018028617541988472, "loss": 0.5974, "step": 665 }, { "epoch": 2.845010615711253, "grad_norm": 0.279296875, "learning_rate": 0.00017984119729947944, "loss": 0.6061, "step": 670 }, { "epoch": 2.8662420382165603, "grad_norm": 0.23046875, "learning_rate": 0.000179391815212081, "loss": 0.6102, "step": 675 }, { "epoch": 2.8874734607218686, "grad_norm": 0.2314453125, "learning_rate": 0.0001789380539451919, "loss": 0.6052, "step": 680 }, { "epoch": 2.908704883227176, "grad_norm": 0.34765625, "learning_rate": 0.0001784799385278661, "loss": 0.6062, "step": 685 }, { "epoch": 2.9299363057324843, "grad_norm": 0.2197265625, "learning_rate": 0.0001780174942293287, "loss": 0.6058, "step": 690 }, { "epoch": 2.9511677282377917, "grad_norm": 0.2109375, "learning_rate": 0.00017755074655758174, "loss": 0.6051, "step": 695 }, { "epoch": 2.9723991507431, "grad_norm": 0.283203125, "learning_rate": 0.00017707972125799735, "loss": 0.6019, "step": 700 }, { "epoch": 2.9936305732484074, "grad_norm": 0.2099609375, "learning_rate": 0.0001766044443118978, "loss": 0.6047, "step": 705 }, { "epoch": 2.9978768577494694, "eval_loss": 1.4090731143951416, "eval_runtime": 0.6234, "eval_samples_per_second": 8.021, "eval_steps_per_second": 1.604, "step": 706 }, { "epoch": 3.0148619957537157, "grad_norm": 0.259765625, "learning_rate": 0.0001761249419351222, "loss": 0.5844, "step": 710 }, { "epoch": 3.0360934182590236, "grad_norm": 0.1923828125, "learning_rate": 0.00017564124057658056, "loss": 0.5729, "step": 715 }, { "epoch": 3.0573248407643314, "grad_norm": 0.197265625, "learning_rate": 0.00017515336691679477, "loss": 0.571, "step": 720 }, { "epoch": 3.0785562632696393, "grad_norm": 0.2109375, "learning_rate": 0.0001746613478664271, "loss": 0.5809, "step": 725 }, { "epoch": 3.099787685774947, "grad_norm": 0.1923828125, "learning_rate": 0.00017416521056479577, "loss": 0.5789, "step": 730 }, { "epoch": 3.121019108280255, "grad_norm": 0.193359375, "learning_rate": 0.0001736649823783779, "loss": 0.5725, "step": 735 }, { "epoch": 3.142250530785563, "grad_norm": 0.23046875, "learning_rate": 0.00017316069089930007, "loss": 0.5729, "step": 740 }, { "epoch": 3.1634819532908707, "grad_norm": 0.251953125, "learning_rate": 0.00017265236394381633, "loss": 0.5839, "step": 745 }, { "epoch": 3.1847133757961785, "grad_norm": 0.2158203125, "learning_rate": 0.00017214002955077393, "loss": 0.5877, "step": 750 }, { "epoch": 3.2059447983014864, "grad_norm": 0.208984375, "learning_rate": 0.00017162371598006666, "loss": 0.5787, "step": 755 }, { "epoch": 3.2271762208067942, "grad_norm": 0.31640625, "learning_rate": 0.0001711034517110761, "loss": 0.5841, "step": 760 }, { "epoch": 3.248407643312102, "grad_norm": 0.224609375, "learning_rate": 0.0001705792654411007, "loss": 0.5784, "step": 765 }, { "epoch": 3.26963906581741, "grad_norm": 0.265625, "learning_rate": 0.00017005118608377288, "loss": 0.5898, "step": 770 }, { "epoch": 3.290870488322718, "grad_norm": 0.236328125, "learning_rate": 0.00016951924276746425, "loss": 0.5793, "step": 775 }, { "epoch": 3.3121019108280256, "grad_norm": 0.2041015625, "learning_rate": 0.00016898346483367867, "loss": 0.5716, "step": 780 }, { "epoch": 3.3333333333333335, "grad_norm": 0.2158203125, "learning_rate": 0.00016844388183543418, "loss": 0.5862, "step": 785 }, { "epoch": 3.3545647558386413, "grad_norm": 0.4375, "learning_rate": 0.00016790052353563253, "loss": 0.5867, "step": 790 }, { "epoch": 3.375796178343949, "grad_norm": 0.30078125, "learning_rate": 0.00016735341990541764, "loss": 0.5883, "step": 795 }, { "epoch": 3.397027600849257, "grad_norm": 0.287109375, "learning_rate": 0.0001668026011225225, "loss": 0.5884, "step": 800 }, { "epoch": 3.418259023354565, "grad_norm": 0.3046875, "learning_rate": 0.00016624809756960444, "loss": 0.5748, "step": 805 }, { "epoch": 3.4394904458598727, "grad_norm": 0.2333984375, "learning_rate": 0.0001656899398325693, "loss": 0.5826, "step": 810 }, { "epoch": 3.4607218683651806, "grad_norm": 0.26953125, "learning_rate": 0.0001651281586988844, "loss": 0.5771, "step": 815 }, { "epoch": 3.4819532908704884, "grad_norm": 0.24609375, "learning_rate": 0.00016456278515588024, "loss": 0.5772, "step": 820 }, { "epoch": 3.5031847133757963, "grad_norm": 0.2119140625, "learning_rate": 0.00016399385038904138, "loss": 0.5811, "step": 825 }, { "epoch": 3.524416135881104, "grad_norm": 0.265625, "learning_rate": 0.00016342138578028613, "loss": 0.5806, "step": 830 }, { "epoch": 3.545647558386412, "grad_norm": 0.2041015625, "learning_rate": 0.00016284542290623567, "loss": 0.5873, "step": 835 }, { "epoch": 3.56687898089172, "grad_norm": 0.259765625, "learning_rate": 0.00016226599353647228, "loss": 0.5766, "step": 840 }, { "epoch": 3.5881104033970277, "grad_norm": 0.2578125, "learning_rate": 0.00016168312963178697, "loss": 0.5819, "step": 845 }, { "epoch": 3.6093418259023355, "grad_norm": 0.19921875, "learning_rate": 0.00016109686334241655, "loss": 0.5832, "step": 850 }, { "epoch": 3.6305732484076434, "grad_norm": 0.2021484375, "learning_rate": 0.00016050722700627012, "loss": 0.5846, "step": 855 }, { "epoch": 3.6518046709129512, "grad_norm": 0.2158203125, "learning_rate": 0.0001599142531471456, "loss": 0.5818, "step": 860 }, { "epoch": 3.673036093418259, "grad_norm": 0.19921875, "learning_rate": 0.00015931797447293552, "loss": 0.5826, "step": 865 }, { "epoch": 3.694267515923567, "grad_norm": 0.236328125, "learning_rate": 0.00015871842387382305, "loss": 0.5823, "step": 870 }, { "epoch": 3.715498938428875, "grad_norm": 0.2021484375, "learning_rate": 0.00015811563442046767, "loss": 0.5813, "step": 875 }, { "epoch": 3.7367303609341826, "grad_norm": 0.203125, "learning_rate": 0.00015750963936218105, "loss": 0.581, "step": 880 }, { "epoch": 3.7579617834394905, "grad_norm": 0.19921875, "learning_rate": 0.00015690047212509316, "loss": 0.5763, "step": 885 }, { "epoch": 3.7791932059447984, "grad_norm": 0.2216796875, "learning_rate": 0.00015628816631030836, "loss": 0.5776, "step": 890 }, { "epoch": 3.800424628450106, "grad_norm": 0.2177734375, "learning_rate": 0.00015567275569205218, "loss": 0.5785, "step": 895 }, { "epoch": 3.821656050955414, "grad_norm": 0.193359375, "learning_rate": 0.00015505427421580808, "loss": 0.581, "step": 900 }, { "epoch": 3.842887473460722, "grad_norm": 0.3984375, "learning_rate": 0.00015443275599644538, "loss": 0.5765, "step": 905 }, { "epoch": 3.8641188959660298, "grad_norm": 0.208984375, "learning_rate": 0.00015380823531633729, "loss": 0.5751, "step": 910 }, { "epoch": 3.8853503184713376, "grad_norm": 0.248046875, "learning_rate": 0.00015318074662346994, "loss": 0.5759, "step": 915 }, { "epoch": 3.9065817409766455, "grad_norm": 0.2001953125, "learning_rate": 0.00015255032452954245, "loss": 0.5856, "step": 920 }, { "epoch": 3.9278131634819533, "grad_norm": 0.21875, "learning_rate": 0.00015191700380805752, "loss": 0.5848, "step": 925 }, { "epoch": 3.949044585987261, "grad_norm": 0.20703125, "learning_rate": 0.00015128081939240357, "loss": 0.5871, "step": 930 }, { "epoch": 3.970276008492569, "grad_norm": 0.220703125, "learning_rate": 0.00015064180637392764, "loss": 0.5803, "step": 935 }, { "epoch": 3.991507430997877, "grad_norm": 0.220703125, "learning_rate": 0.00015000000000000001, "loss": 0.5773, "step": 940 }, { "epoch": 4.0, "eval_loss": 1.442795753479004, "eval_runtime": 0.4901, "eval_samples_per_second": 10.203, "eval_steps_per_second": 2.041, "step": 942 }, { "epoch": 4.012738853503185, "grad_norm": 0.2255859375, "learning_rate": 0.00014935543567206984, "loss": 0.5596, "step": 945 }, { "epoch": 4.033970276008493, "grad_norm": 0.287109375, "learning_rate": 0.00014870814894371245, "loss": 0.5489, "step": 950 }, { "epoch": 4.055201698513801, "grad_norm": 0.240234375, "learning_rate": 0.00014805817551866838, "loss": 0.5555, "step": 955 }, { "epoch": 4.076433121019108, "grad_norm": 0.267578125, "learning_rate": 0.00014740555124887375, "loss": 0.5524, "step": 960 }, { "epoch": 4.097664543524417, "grad_norm": 0.26171875, "learning_rate": 0.00014675031213248296, "loss": 0.5603, "step": 965 }, { "epoch": 4.118895966029724, "grad_norm": 0.2041015625, "learning_rate": 0.00014609249431188278, "loss": 0.5587, "step": 970 }, { "epoch": 4.140127388535032, "grad_norm": 0.2578125, "learning_rate": 0.0001454321340716992, "loss": 0.5552, "step": 975 }, { "epoch": 4.16135881104034, "grad_norm": 0.25390625, "learning_rate": 0.00014476926783679538, "loss": 0.5542, "step": 980 }, { "epoch": 4.182590233545648, "grad_norm": 0.28125, "learning_rate": 0.00014410393217026318, "loss": 0.5554, "step": 985 }, { "epoch": 4.203821656050955, "grad_norm": 0.255859375, "learning_rate": 0.00014343616377140582, "loss": 0.5477, "step": 990 }, { "epoch": 4.225053078556264, "grad_norm": 0.34375, "learning_rate": 0.00014276599947371388, "loss": 0.5477, "step": 995 }, { "epoch": 4.246284501061571, "grad_norm": 0.234375, "learning_rate": 0.0001420934762428335, "loss": 0.5563, "step": 1000 }, { "epoch": 4.267515923566879, "grad_norm": 0.2001953125, "learning_rate": 0.00014141863117452745, "loss": 0.5552, "step": 1005 }, { "epoch": 4.288747346072187, "grad_norm": 0.259765625, "learning_rate": 0.0001407415014926288, "loss": 0.553, "step": 1010 }, { "epoch": 4.309978768577495, "grad_norm": 0.2119140625, "learning_rate": 0.00014006212454698797, "loss": 0.5575, "step": 1015 }, { "epoch": 4.3312101910828025, "grad_norm": 0.228515625, "learning_rate": 0.00013938053781141222, "loss": 0.5604, "step": 1020 }, { "epoch": 4.352441613588111, "grad_norm": 0.234375, "learning_rate": 0.00013869677888159887, "loss": 0.5556, "step": 1025 }, { "epoch": 4.373673036093418, "grad_norm": 0.25390625, "learning_rate": 0.00013801088547306148, "loss": 0.5557, "step": 1030 }, { "epoch": 4.3949044585987265, "grad_norm": 0.224609375, "learning_rate": 0.00013732289541904948, "loss": 0.5595, "step": 1035 }, { "epoch": 4.416135881104034, "grad_norm": 0.2109375, "learning_rate": 0.00013663284666846134, "loss": 0.5605, "step": 1040 }, { "epoch": 4.437367303609342, "grad_norm": 0.205078125, "learning_rate": 0.00013594077728375128, "loss": 0.5598, "step": 1045 }, { "epoch": 4.45859872611465, "grad_norm": 0.23046875, "learning_rate": 0.00013524672543882996, "loss": 0.559, "step": 1050 }, { "epoch": 4.479830148619958, "grad_norm": 0.2080078125, "learning_rate": 0.00013455072941695863, "loss": 0.5612, "step": 1055 }, { "epoch": 4.501061571125265, "grad_norm": 0.21484375, "learning_rate": 0.00013385282760863758, "loss": 0.556, "step": 1060 }, { "epoch": 4.522292993630574, "grad_norm": 0.2275390625, "learning_rate": 0.00013315305850948846, "loss": 0.5551, "step": 1065 }, { "epoch": 4.543524416135881, "grad_norm": 0.20703125, "learning_rate": 0.00013245146071813114, "loss": 0.5602, "step": 1070 }, { "epoch": 4.564755838641189, "grad_norm": 0.208984375, "learning_rate": 0.00013174807293405428, "loss": 0.5467, "step": 1075 }, { "epoch": 4.585987261146497, "grad_norm": 0.2177734375, "learning_rate": 0.00013104293395548098, "loss": 0.5551, "step": 1080 }, { "epoch": 4.607218683651805, "grad_norm": 0.21484375, "learning_rate": 0.00013033608267722858, "loss": 0.5533, "step": 1085 }, { "epoch": 4.628450106157112, "grad_norm": 0.208984375, "learning_rate": 0.00012962755808856342, "loss": 0.5531, "step": 1090 }, { "epoch": 4.649681528662421, "grad_norm": 0.22265625, "learning_rate": 0.0001289173992710499, "loss": 0.5585, "step": 1095 }, { "epoch": 4.670912951167728, "grad_norm": 0.25, "learning_rate": 0.00012820564539639512, "loss": 0.5601, "step": 1100 }, { "epoch": 4.692144373673036, "grad_norm": 0.2236328125, "learning_rate": 0.00012749233572428804, "loss": 0.5586, "step": 1105 }, { "epoch": 4.713375796178344, "grad_norm": 0.2099609375, "learning_rate": 0.00012677750960023396, "loss": 0.5584, "step": 1110 }, { "epoch": 4.734607218683652, "grad_norm": 0.259765625, "learning_rate": 0.0001260612064533843, "loss": 0.5655, "step": 1115 }, { "epoch": 4.7558386411889595, "grad_norm": 0.220703125, "learning_rate": 0.0001253434657943616, "loss": 0.5516, "step": 1120 }, { "epoch": 4.777070063694268, "grad_norm": 0.19921875, "learning_rate": 0.0001246243272130804, "loss": 0.552, "step": 1125 }, { "epoch": 4.798301486199575, "grad_norm": 0.205078125, "learning_rate": 0.00012390383037656327, "loss": 0.5549, "step": 1130 }, { "epoch": 4.8195329087048835, "grad_norm": 0.228515625, "learning_rate": 0.00012318201502675285, "loss": 0.5564, "step": 1135 }, { "epoch": 4.840764331210191, "grad_norm": 0.21875, "learning_rate": 0.00012245892097831982, "loss": 0.5617, "step": 1140 }, { "epoch": 4.861995753715499, "grad_norm": 0.2734375, "learning_rate": 0.0001217345881164667, "loss": 0.564, "step": 1145 }, { "epoch": 4.883227176220807, "grad_norm": 0.296875, "learning_rate": 0.00012100905639472779, "loss": 0.5534, "step": 1150 }, { "epoch": 4.904458598726115, "grad_norm": 0.296875, "learning_rate": 0.00012028236583276542, "loss": 0.5556, "step": 1155 }, { "epoch": 4.925690021231422, "grad_norm": 0.224609375, "learning_rate": 0.00011955455651416246, "loss": 0.5656, "step": 1160 }, { "epoch": 4.946921443736731, "grad_norm": 0.21875, "learning_rate": 0.00011882566858421135, "loss": 0.5595, "step": 1165 }, { "epoch": 4.968152866242038, "grad_norm": 0.2373046875, "learning_rate": 0.00011809574224769981, "loss": 0.56, "step": 1170 }, { "epoch": 4.989384288747346, "grad_norm": 0.2265625, "learning_rate": 0.00011736481776669306, "loss": 0.5548, "step": 1175 }, { "epoch": 4.997876857749469, "eval_loss": 1.4904053211212158, "eval_runtime": 0.6021, "eval_samples_per_second": 8.304, "eval_steps_per_second": 1.661, "step": 1177 }, { "epoch": 5.010615711252654, "grad_norm": 0.21484375, "learning_rate": 0.00011663293545831302, "loss": 0.5538, "step": 1180 }, { "epoch": 5.031847133757962, "grad_norm": 0.255859375, "learning_rate": 0.00011590013569251457, "loss": 0.5333, "step": 1185 }, { "epoch": 5.053078556263269, "grad_norm": 0.2265625, "learning_rate": 0.0001151664588898586, "loss": 0.5372, "step": 1190 }, { "epoch": 5.074309978768578, "grad_norm": 0.27734375, "learning_rate": 0.00011443194551928266, "loss": 0.5291, "step": 1195 }, { "epoch": 5.095541401273885, "grad_norm": 0.2197265625, "learning_rate": 0.00011369663609586854, "loss": 0.5301, "step": 1200 }, { "epoch": 5.116772823779193, "grad_norm": 0.21484375, "learning_rate": 0.00011296057117860759, "loss": 0.5246, "step": 1205 }, { "epoch": 5.138004246284501, "grad_norm": 0.267578125, "learning_rate": 0.00011222379136816345, "loss": 0.5341, "step": 1210 }, { "epoch": 5.159235668789809, "grad_norm": 0.2578125, "learning_rate": 0.00011148633730463273, "loss": 0.5391, "step": 1215 }, { "epoch": 5.1804670912951165, "grad_norm": 0.220703125, "learning_rate": 0.00011074824966530312, "loss": 0.5354, "step": 1220 }, { "epoch": 5.201698513800425, "grad_norm": 0.2177734375, "learning_rate": 0.00011000956916240985, "loss": 0.5315, "step": 1225 }, { "epoch": 5.222929936305732, "grad_norm": 0.2216796875, "learning_rate": 0.00010927033654088983, "loss": 0.5396, "step": 1230 }, { "epoch": 5.2441613588110405, "grad_norm": 0.26953125, "learning_rate": 0.00010853059257613448, "loss": 0.5375, "step": 1235 }, { "epoch": 5.265392781316348, "grad_norm": 0.28125, "learning_rate": 0.00010779037807174033, "loss": 0.5332, "step": 1240 }, { "epoch": 5.286624203821656, "grad_norm": 0.244140625, "learning_rate": 0.00010704973385725851, "loss": 0.5324, "step": 1245 }, { "epoch": 5.307855626326964, "grad_norm": 0.2333984375, "learning_rate": 0.00010630870078594249, "loss": 0.5344, "step": 1250 }, { "epoch": 5.329087048832272, "grad_norm": 0.373046875, "learning_rate": 0.00010556731973249485, "loss": 0.5272, "step": 1255 }, { "epoch": 5.350318471337579, "grad_norm": 0.287109375, "learning_rate": 0.00010482563159081238, "loss": 0.5331, "step": 1260 }, { "epoch": 5.371549893842888, "grad_norm": 0.24609375, "learning_rate": 0.00010408367727173067, "loss": 0.5327, "step": 1265 }, { "epoch": 5.392781316348195, "grad_norm": 0.23046875, "learning_rate": 0.00010334149770076747, "loss": 0.5319, "step": 1270 }, { "epoch": 5.414012738853503, "grad_norm": 0.240234375, "learning_rate": 0.0001025991338158651, "loss": 0.5439, "step": 1275 }, { "epoch": 5.435244161358811, "grad_norm": 0.2353515625, "learning_rate": 0.00010185662656513251, "loss": 0.5419, "step": 1280 }, { "epoch": 5.456475583864119, "grad_norm": 0.224609375, "learning_rate": 0.00010111401690458654, "loss": 0.5375, "step": 1285 }, { "epoch": 5.477707006369426, "grad_norm": 0.2421875, "learning_rate": 0.00010037134579589302, "loss": 0.5351, "step": 1290 }, { "epoch": 5.498938428874735, "grad_norm": 0.26953125, "learning_rate": 9.962865420410701e-05, "loss": 0.53, "step": 1295 }, { "epoch": 5.520169851380042, "grad_norm": 0.291015625, "learning_rate": 9.888598309541347e-05, "loss": 0.5318, "step": 1300 }, { "epoch": 5.54140127388535, "grad_norm": 0.2314453125, "learning_rate": 9.814337343486754e-05, "loss": 0.5348, "step": 1305 }, { "epoch": 5.562632696390658, "grad_norm": 0.33203125, "learning_rate": 9.740086618413495e-05, "loss": 0.5301, "step": 1310 }, { "epoch": 5.583864118895966, "grad_norm": 0.2578125, "learning_rate": 9.665850229923258e-05, "loss": 0.5333, "step": 1315 }, { "epoch": 5.6050955414012735, "grad_norm": 0.2373046875, "learning_rate": 9.591632272826934e-05, "loss": 0.5344, "step": 1320 }, { "epoch": 5.626326963906582, "grad_norm": 0.2138671875, "learning_rate": 9.517436840918766e-05, "loss": 0.529, "step": 1325 }, { "epoch": 5.647558386411889, "grad_norm": 0.2216796875, "learning_rate": 9.44326802675052e-05, "loss": 0.5346, "step": 1330 }, { "epoch": 5.6687898089171975, "grad_norm": 0.279296875, "learning_rate": 9.369129921405754e-05, "loss": 0.5354, "step": 1335 }, { "epoch": 5.690021231422505, "grad_norm": 0.2431640625, "learning_rate": 9.295026614274152e-05, "loss": 0.5303, "step": 1340 }, { "epoch": 5.711252653927813, "grad_norm": 0.2578125, "learning_rate": 9.220962192825968e-05, "loss": 0.5366, "step": 1345 }, { "epoch": 5.732484076433121, "grad_norm": 0.248046875, "learning_rate": 9.146940742386553e-05, "loss": 0.542, "step": 1350 }, { "epoch": 5.753715498938429, "grad_norm": 0.21875, "learning_rate": 9.072966345911019e-05, "loss": 0.5402, "step": 1355 }, { "epoch": 5.774946921443737, "grad_norm": 0.2158203125, "learning_rate": 8.999043083759017e-05, "loss": 0.5336, "step": 1360 }, { "epoch": 5.796178343949045, "grad_norm": 0.25, "learning_rate": 8.925175033469688e-05, "loss": 0.5389, "step": 1365 }, { "epoch": 5.817409766454352, "grad_norm": 0.2119140625, "learning_rate": 8.851366269536729e-05, "loss": 0.5355, "step": 1370 }, { "epoch": 5.83864118895966, "grad_norm": 0.251953125, "learning_rate": 8.777620863183657e-05, "loss": 0.5234, "step": 1375 }, { "epoch": 5.859872611464969, "grad_norm": 0.255859375, "learning_rate": 8.703942882139245e-05, "loss": 0.5318, "step": 1380 }, { "epoch": 5.881104033970276, "grad_norm": 0.2294921875, "learning_rate": 8.630336390413147e-05, "loss": 0.5371, "step": 1385 }, { "epoch": 5.902335456475583, "grad_norm": 0.22265625, "learning_rate": 8.556805448071735e-05, "loss": 0.5317, "step": 1390 }, { "epoch": 5.923566878980892, "grad_norm": 0.2255859375, "learning_rate": 8.483354111014141e-05, "loss": 0.5331, "step": 1395 }, { "epoch": 5.9447983014862, "grad_norm": 0.220703125, "learning_rate": 8.409986430748545e-05, "loss": 0.5373, "step": 1400 }, { "epoch": 5.966029723991507, "grad_norm": 0.240234375, "learning_rate": 8.336706454168701e-05, "loss": 0.5328, "step": 1405 }, { "epoch": 5.987261146496815, "grad_norm": 0.2578125, "learning_rate": 8.263518223330697e-05, "loss": 0.5409, "step": 1410 }, { "epoch": 6.0, "eval_loss": 1.5479737520217896, "eval_runtime": 0.4964, "eval_samples_per_second": 10.072, "eval_steps_per_second": 2.014, "step": 1413 }, { "epoch": 6.008492569002123, "grad_norm": 0.220703125, "learning_rate": 8.190425775230021e-05, "loss": 0.5317, "step": 1415 }, { "epoch": 6.029723991507431, "grad_norm": 0.228515625, "learning_rate": 8.117433141578866e-05, "loss": 0.5111, "step": 1420 }, { "epoch": 6.050955414012739, "grad_norm": 0.2275390625, "learning_rate": 8.044544348583755e-05, "loss": 0.5147, "step": 1425 }, { "epoch": 6.072186836518047, "grad_norm": 0.2197265625, "learning_rate": 7.971763416723459e-05, "loss": 0.5099, "step": 1430 }, { "epoch": 6.0934182590233545, "grad_norm": 0.251953125, "learning_rate": 7.89909436052722e-05, "loss": 0.511, "step": 1435 }, { "epoch": 6.114649681528663, "grad_norm": 0.23046875, "learning_rate": 7.826541188353329e-05, "loss": 0.5043, "step": 1440 }, { "epoch": 6.13588110403397, "grad_norm": 0.232421875, "learning_rate": 7.754107902168019e-05, "loss": 0.5112, "step": 1445 }, { "epoch": 6.1571125265392785, "grad_norm": 0.26953125, "learning_rate": 7.681798497324716e-05, "loss": 0.5119, "step": 1450 }, { "epoch": 6.178343949044586, "grad_norm": 0.234375, "learning_rate": 7.609616962343675e-05, "loss": 0.5177, "step": 1455 }, { "epoch": 6.199575371549894, "grad_norm": 0.232421875, "learning_rate": 7.537567278691964e-05, "loss": 0.5208, "step": 1460 }, { "epoch": 6.220806794055202, "grad_norm": 0.23046875, "learning_rate": 7.465653420563845e-05, "loss": 0.5126, "step": 1465 }, { "epoch": 6.24203821656051, "grad_norm": 0.2412109375, "learning_rate": 7.393879354661577e-05, "loss": 0.5185, "step": 1470 }, { "epoch": 6.263269639065817, "grad_norm": 0.26171875, "learning_rate": 7.322249039976608e-05, "loss": 0.5171, "step": 1475 }, { "epoch": 6.284501061571126, "grad_norm": 0.24609375, "learning_rate": 7.2507664275712e-05, "loss": 0.5096, "step": 1480 }, { "epoch": 6.305732484076433, "grad_norm": 0.2578125, "learning_rate": 7.179435460360491e-05, "loss": 0.5116, "step": 1485 }, { "epoch": 6.326963906581741, "grad_norm": 0.224609375, "learning_rate": 7.108260072895013e-05, "loss": 0.5104, "step": 1490 }, { "epoch": 6.348195329087049, "grad_norm": 0.2353515625, "learning_rate": 7.037244191143661e-05, "loss": 0.5165, "step": 1495 }, { "epoch": 6.369426751592357, "grad_norm": 0.275390625, "learning_rate": 6.966391732277143e-05, "loss": 0.5186, "step": 1500 }, { "epoch": 6.3906581740976645, "grad_norm": 0.2294921875, "learning_rate": 6.895706604451905e-05, "loss": 0.5124, "step": 1505 }, { "epoch": 6.411889596602973, "grad_norm": 0.2431640625, "learning_rate": 6.825192706594575e-05, "loss": 0.5181, "step": 1510 }, { "epoch": 6.43312101910828, "grad_norm": 0.2353515625, "learning_rate": 6.75485392818689e-05, "loss": 0.5156, "step": 1515 }, { "epoch": 6.4543524416135885, "grad_norm": 0.26953125, "learning_rate": 6.684694149051156e-05, "loss": 0.5164, "step": 1520 }, { "epoch": 6.475583864118896, "grad_norm": 0.23828125, "learning_rate": 6.614717239136246e-05, "loss": 0.5055, "step": 1525 }, { "epoch": 6.496815286624204, "grad_norm": 0.2431640625, "learning_rate": 6.54492705830414e-05, "loss": 0.5179, "step": 1530 }, { "epoch": 6.518046709129512, "grad_norm": 0.2412109375, "learning_rate": 6.475327456117005e-05, "loss": 0.5144, "step": 1535 }, { "epoch": 6.53927813163482, "grad_norm": 0.251953125, "learning_rate": 6.405922271624874e-05, "loss": 0.5117, "step": 1540 }, { "epoch": 6.560509554140127, "grad_norm": 0.2490234375, "learning_rate": 6.336715333153869e-05, "loss": 0.5143, "step": 1545 }, { "epoch": 6.581740976645436, "grad_norm": 0.25, "learning_rate": 6.267710458095053e-05, "loss": 0.5136, "step": 1550 }, { "epoch": 6.602972399150743, "grad_norm": 0.23046875, "learning_rate": 6.198911452693853e-05, "loss": 0.5133, "step": 1555 }, { "epoch": 6.624203821656051, "grad_norm": 0.2294921875, "learning_rate": 6.130322111840114e-05, "loss": 0.5162, "step": 1560 }, { "epoch": 6.645435244161359, "grad_norm": 0.2431640625, "learning_rate": 6.0619462188587793e-05, "loss": 0.5171, "step": 1565 }, { "epoch": 6.666666666666667, "grad_norm": 0.228515625, "learning_rate": 5.993787545301204e-05, "loss": 0.519, "step": 1570 }, { "epoch": 6.687898089171974, "grad_norm": 0.232421875, "learning_rate": 5.9258498507371194e-05, "loss": 0.5143, "step": 1575 }, { "epoch": 6.709129511677283, "grad_norm": 0.251953125, "learning_rate": 5.8581368825472585e-05, "loss": 0.5238, "step": 1580 }, { "epoch": 6.73036093418259, "grad_norm": 0.2578125, "learning_rate": 5.790652375716652e-05, "loss": 0.513, "step": 1585 }, { "epoch": 6.751592356687898, "grad_norm": 0.234375, "learning_rate": 5.7234000526286156e-05, "loss": 0.5129, "step": 1590 }, { "epoch": 6.772823779193206, "grad_norm": 0.25, "learning_rate": 5.656383622859418e-05, "loss": 0.5251, "step": 1595 }, { "epoch": 6.794055201698514, "grad_norm": 0.224609375, "learning_rate": 5.589606782973683e-05, "loss": 0.5158, "step": 1600 }, { "epoch": 6.8152866242038215, "grad_norm": 0.2392578125, "learning_rate": 5.5230732163204615e-05, "loss": 0.5169, "step": 1605 }, { "epoch": 6.83651804670913, "grad_norm": 0.275390625, "learning_rate": 5.456786592830083e-05, "loss": 0.5203, "step": 1610 }, { "epoch": 6.857749469214437, "grad_norm": 0.2275390625, "learning_rate": 5.39075056881172e-05, "loss": 0.5127, "step": 1615 }, { "epoch": 6.8789808917197455, "grad_norm": 0.2333984375, "learning_rate": 5.3249687867517095e-05, "loss": 0.5176, "step": 1620 }, { "epoch": 6.900212314225053, "grad_norm": 0.26171875, "learning_rate": 5.259444875112624e-05, "loss": 0.5163, "step": 1625 }, { "epoch": 6.921443736730361, "grad_norm": 0.296875, "learning_rate": 5.1941824481331626e-05, "loss": 0.5256, "step": 1630 }, { "epoch": 6.942675159235669, "grad_norm": 0.234375, "learning_rate": 5.129185105628756e-05, "loss": 0.5199, "step": 1635 }, { "epoch": 6.963906581740977, "grad_norm": 0.275390625, "learning_rate": 5.064456432793019e-05, "loss": 0.5166, "step": 1640 }, { "epoch": 6.985138004246284, "grad_norm": 0.240234375, "learning_rate": 5.000000000000002e-05, "loss": 0.5151, "step": 1645 }, { "epoch": 6.997876857749469, "eval_loss": 1.6102020740509033, "eval_runtime": 0.6273, "eval_samples_per_second": 7.971, "eval_steps_per_second": 1.594, "step": 1648 }, { "epoch": 7.006369426751593, "grad_norm": 0.2275390625, "learning_rate": 4.93581936260724e-05, "loss": 0.504, "step": 1650 }, { "epoch": 7.0276008492569, "grad_norm": 0.2412109375, "learning_rate": 4.8719180607596484e-05, "loss": 0.4887, "step": 1655 }, { "epoch": 7.048832271762208, "grad_norm": 0.2451171875, "learning_rate": 4.808299619194251e-05, "loss": 0.502, "step": 1660 }, { "epoch": 7.070063694267516, "grad_norm": 0.2333984375, "learning_rate": 4.744967547045754e-05, "loss": 0.496, "step": 1665 }, { "epoch": 7.091295116772824, "grad_norm": 0.2275390625, "learning_rate": 4.681925337653006e-05, "loss": 0.4983, "step": 1670 }, { "epoch": 7.112526539278131, "grad_norm": 0.2392578125, "learning_rate": 4.6191764683662744e-05, "loss": 0.4982, "step": 1675 }, { "epoch": 7.13375796178344, "grad_norm": 0.248046875, "learning_rate": 4.5567244003554645e-05, "loss": 0.5054, "step": 1680 }, { "epoch": 7.154989384288747, "grad_norm": 0.240234375, "learning_rate": 4.494572578419194e-05, "loss": 0.5019, "step": 1685 }, { "epoch": 7.176220806794055, "grad_norm": 0.2431640625, "learning_rate": 4.432724430794786e-05, "loss": 0.4981, "step": 1690 }, { "epoch": 7.197452229299363, "grad_norm": 0.26171875, "learning_rate": 4.371183368969165e-05, "loss": 0.5, "step": 1695 }, { "epoch": 7.218683651804671, "grad_norm": 0.251953125, "learning_rate": 4.309952787490689e-05, "loss": 0.501, "step": 1700 }, { "epoch": 7.2399150743099785, "grad_norm": 0.25390625, "learning_rate": 4.249036063781896e-05, "loss": 0.5, "step": 1705 }, { "epoch": 7.261146496815287, "grad_norm": 0.2421875, "learning_rate": 4.1884365579532346e-05, "loss": 0.4991, "step": 1710 }, { "epoch": 7.282377919320594, "grad_norm": 0.255859375, "learning_rate": 4.128157612617696e-05, "loss": 0.4962, "step": 1715 }, { "epoch": 7.3036093418259025, "grad_norm": 0.2392578125, "learning_rate": 4.0682025527064486e-05, "loss": 0.5008, "step": 1720 }, { "epoch": 7.32484076433121, "grad_norm": 0.236328125, "learning_rate": 4.008574685285442e-05, "loss": 0.5016, "step": 1725 }, { "epoch": 7.346072186836518, "grad_norm": 0.23828125, "learning_rate": 3.94927729937299e-05, "loss": 0.4939, "step": 1730 }, { "epoch": 7.367303609341826, "grad_norm": 0.248046875, "learning_rate": 3.890313665758348e-05, "loss": 0.5046, "step": 1735 }, { "epoch": 7.388535031847134, "grad_norm": 0.2373046875, "learning_rate": 3.8316870368213e-05, "loss": 0.4882, "step": 1740 }, { "epoch": 7.409766454352441, "grad_norm": 0.2470703125, "learning_rate": 3.773400646352769e-05, "loss": 0.5053, "step": 1745 }, { "epoch": 7.43099787685775, "grad_norm": 0.2412109375, "learning_rate": 3.7154577093764334e-05, "loss": 0.5065, "step": 1750 }, { "epoch": 7.452229299363057, "grad_norm": 0.2392578125, "learning_rate": 3.657861421971388e-05, "loss": 0.5005, "step": 1755 }, { "epoch": 7.473460721868365, "grad_norm": 0.263671875, "learning_rate": 3.6006149610958625e-05, "loss": 0.5066, "step": 1760 }, { "epoch": 7.494692144373673, "grad_norm": 0.2451171875, "learning_rate": 3.543721484411976e-05, "loss": 0.5019, "step": 1765 }, { "epoch": 7.515923566878981, "grad_norm": 0.287109375, "learning_rate": 3.487184130111562e-05, "loss": 0.493, "step": 1770 }, { "epoch": 7.537154989384288, "grad_norm": 0.251953125, "learning_rate": 3.4310060167430725e-05, "loss": 0.5087, "step": 1775 }, { "epoch": 7.558386411889597, "grad_norm": 0.26171875, "learning_rate": 3.375190243039556e-05, "loss": 0.5032, "step": 1780 }, { "epoch": 7.579617834394904, "grad_norm": 0.2431640625, "learning_rate": 3.319739887747752e-05, "loss": 0.5004, "step": 1785 }, { "epoch": 7.600849256900212, "grad_norm": 0.2392578125, "learning_rate": 3.264658009458239e-05, "loss": 0.5014, "step": 1790 }, { "epoch": 7.62208067940552, "grad_norm": 0.2578125, "learning_rate": 3.209947646436752e-05, "loss": 0.5044, "step": 1795 }, { "epoch": 7.643312101910828, "grad_norm": 0.2392578125, "learning_rate": 3.155611816456586e-05, "loss": 0.5027, "step": 1800 }, { "epoch": 7.6645435244161355, "grad_norm": 0.2431640625, "learning_rate": 3.1016535166321356e-05, "loss": 0.5004, "step": 1805 }, { "epoch": 7.685774946921444, "grad_norm": 0.2490234375, "learning_rate": 3.0480757232535772e-05, "loss": 0.5016, "step": 1810 }, { "epoch": 7.707006369426751, "grad_norm": 0.2490234375, "learning_rate": 2.9948813916227115e-05, "loss": 0.5026, "step": 1815 }, { "epoch": 7.7282377919320595, "grad_norm": 0.2470703125, "learning_rate": 2.9420734558899322e-05, "loss": 0.5047, "step": 1820 }, { "epoch": 7.749469214437367, "grad_norm": 0.2333984375, "learning_rate": 2.889654828892393e-05, "loss": 0.5015, "step": 1825 }, { "epoch": 7.770700636942675, "grad_norm": 0.2490234375, "learning_rate": 2.8376284019933373e-05, "loss": 0.505, "step": 1830 }, { "epoch": 7.7919320594479835, "grad_norm": 0.25, "learning_rate": 2.7859970449226104e-05, "loss": 0.5035, "step": 1835 }, { "epoch": 7.813163481953291, "grad_norm": 0.248046875, "learning_rate": 2.73476360561837e-05, "loss": 0.5021, "step": 1840 }, { "epoch": 7.834394904458598, "grad_norm": 0.240234375, "learning_rate": 2.6839309100699973e-05, "loss": 0.4994, "step": 1845 }, { "epoch": 7.855626326963907, "grad_norm": 0.24609375, "learning_rate": 2.6335017621622116e-05, "loss": 0.5017, "step": 1850 }, { "epoch": 7.876857749469215, "grad_norm": 0.23046875, "learning_rate": 2.5834789435204243e-05, "loss": 0.4948, "step": 1855 }, { "epoch": 7.898089171974522, "grad_norm": 0.23828125, "learning_rate": 2.5338652133572915e-05, "loss": 0.4898, "step": 1860 }, { "epoch": 7.91932059447983, "grad_norm": 0.2412109375, "learning_rate": 2.4846633083205263e-05, "loss": 0.5079, "step": 1865 }, { "epoch": 7.940552016985138, "grad_norm": 0.279296875, "learning_rate": 2.4358759423419474e-05, "loss": 0.4995, "step": 1870 }, { "epoch": 7.961783439490446, "grad_norm": 0.2431640625, "learning_rate": 2.3875058064877807e-05, "loss": 0.4999, "step": 1875 }, { "epoch": 7.983014861995754, "grad_norm": 0.23828125, "learning_rate": 2.339555568810221e-05, "loss": 0.4987, "step": 1880 }, { "epoch": 8.0, "eval_loss": 1.6578315496444702, "eval_runtime": 0.4956, "eval_samples_per_second": 10.088, "eval_steps_per_second": 2.018, "step": 1884 }, { "epoch": 8.004246284501061, "grad_norm": 0.2578125, "learning_rate": 2.2920278742002676e-05, "loss": 0.4963, "step": 1885 }, { "epoch": 8.02547770700637, "grad_norm": 0.2412109375, "learning_rate": 2.244925344241828e-05, "loss": 0.4973, "step": 1890 }, { "epoch": 8.046709129511678, "grad_norm": 0.234375, "learning_rate": 2.1982505770671303e-05, "loss": 0.491, "step": 1895 }, { "epoch": 8.067940552016985, "grad_norm": 0.26171875, "learning_rate": 2.1520061472133902e-05, "loss": 0.4888, "step": 1900 }, { "epoch": 8.089171974522293, "grad_norm": 0.2392578125, "learning_rate": 2.1061946054808146e-05, "loss": 0.4977, "step": 1905 }, { "epoch": 8.110403397027602, "grad_norm": 0.2412109375, "learning_rate": 2.0608184787919026e-05, "loss": 0.4917, "step": 1910 }, { "epoch": 8.13163481953291, "grad_norm": 0.2578125, "learning_rate": 2.0158802700520574e-05, "loss": 0.4969, "step": 1915 }, { "epoch": 8.152866242038217, "grad_norm": 0.2421875, "learning_rate": 1.9713824580115335e-05, "loss": 0.4918, "step": 1920 }, { "epoch": 8.174097664543524, "grad_norm": 0.2421875, "learning_rate": 1.927327497128706e-05, "loss": 0.4919, "step": 1925 }, { "epoch": 8.195329087048833, "grad_norm": 0.2421875, "learning_rate": 1.883717817434688e-05, "loss": 0.4926, "step": 1930 }, { "epoch": 8.21656050955414, "grad_norm": 0.2431640625, "learning_rate": 1.840555824399296e-05, "loss": 0.4913, "step": 1935 }, { "epoch": 8.237791932059448, "grad_norm": 0.255859375, "learning_rate": 1.797843898798358e-05, "loss": 0.4879, "step": 1940 }, { "epoch": 8.259023354564755, "grad_norm": 0.2421875, "learning_rate": 1.7555843965823992e-05, "loss": 0.5012, "step": 1945 }, { "epoch": 8.280254777070065, "grad_norm": 0.2412109375, "learning_rate": 1.7137796487466797e-05, "loss": 0.4956, "step": 1950 }, { "epoch": 8.301486199575372, "grad_norm": 0.25, "learning_rate": 1.672431961202635e-05, "loss": 0.4959, "step": 1955 }, { "epoch": 8.32271762208068, "grad_norm": 0.2451171875, "learning_rate": 1.6315436146506703e-05, "loss": 0.4888, "step": 1960 }, { "epoch": 8.343949044585987, "grad_norm": 0.2421875, "learning_rate": 1.5911168644543707e-05, "loss": 0.4824, "step": 1965 }, { "epoch": 8.365180467091296, "grad_norm": 0.2431640625, "learning_rate": 1.5511539405160825e-05, "loss": 0.501, "step": 1970 }, { "epoch": 8.386411889596603, "grad_norm": 0.2490234375, "learning_rate": 1.5116570471539293e-05, "loss": 0.4885, "step": 1975 }, { "epoch": 8.40764331210191, "grad_norm": 0.24609375, "learning_rate": 1.4726283629802107e-05, "loss": 0.4921, "step": 1980 }, { "epoch": 8.428874734607218, "grad_norm": 0.259765625, "learning_rate": 1.4340700407812435e-05, "loss": 0.4962, "step": 1985 }, { "epoch": 8.450106157112527, "grad_norm": 0.236328125, "learning_rate": 1.3959842073986085e-05, "loss": 0.4892, "step": 1990 }, { "epoch": 8.471337579617835, "grad_norm": 0.2412109375, "learning_rate": 1.3583729636118358e-05, "loss": 0.497, "step": 1995 }, { "epoch": 8.492569002123142, "grad_norm": 0.251953125, "learning_rate": 1.3212383840225329e-05, "loss": 0.4964, "step": 2000 }, { "epoch": 8.51380042462845, "grad_norm": 0.2392578125, "learning_rate": 1.2845825169399507e-05, "loss": 0.4892, "step": 2005 }, { "epoch": 8.535031847133759, "grad_norm": 0.2578125, "learning_rate": 1.2484073842679944e-05, "loss": 0.4951, "step": 2010 }, { "epoch": 8.556263269639066, "grad_norm": 0.2392578125, "learning_rate": 1.2127149813937022e-05, "loss": 0.4949, "step": 2015 }, { "epoch": 8.577494692144374, "grad_norm": 0.240234375, "learning_rate": 1.1775072770771834e-05, "loss": 0.4906, "step": 2020 }, { "epoch": 8.598726114649681, "grad_norm": 0.2373046875, "learning_rate": 1.1427862133430156e-05, "loss": 0.4836, "step": 2025 }, { "epoch": 8.61995753715499, "grad_norm": 0.2412109375, "learning_rate": 1.1085537053731354e-05, "loss": 0.4944, "step": 2030 }, { "epoch": 8.641188959660298, "grad_norm": 0.2373046875, "learning_rate": 1.0748116414011888e-05, "loss": 0.4908, "step": 2035 }, { "epoch": 8.662420382165605, "grad_norm": 0.2412109375, "learning_rate": 1.0415618826083828e-05, "loss": 0.4969, "step": 2040 }, { "epoch": 8.683651804670912, "grad_norm": 0.251953125, "learning_rate": 1.0088062630208273e-05, "loss": 0.4904, "step": 2045 }, { "epoch": 8.704883227176222, "grad_norm": 0.26171875, "learning_rate": 9.765465894083636e-06, "loss": 0.4991, "step": 2050 }, { "epoch": 8.726114649681529, "grad_norm": 0.2470703125, "learning_rate": 9.447846411849115e-06, "loss": 0.4971, "step": 2055 }, { "epoch": 8.747346072186836, "grad_norm": 0.25, "learning_rate": 9.135221703103136e-06, "loss": 0.4914, "step": 2060 }, { "epoch": 8.768577494692144, "grad_norm": 0.2451171875, "learning_rate": 8.827609011937066e-06, "loss": 0.4919, "step": 2065 }, { "epoch": 8.789808917197453, "grad_norm": 0.259765625, "learning_rate": 8.525025305983936e-06, "loss": 0.4997, "step": 2070 }, { "epoch": 8.81104033970276, "grad_norm": 0.2412109375, "learning_rate": 8.227487275482592e-06, "loss": 0.4879, "step": 2075 }, { "epoch": 8.832271762208068, "grad_norm": 0.23828125, "learning_rate": 7.935011332357112e-06, "loss": 0.4957, "step": 2080 }, { "epoch": 8.853503184713375, "grad_norm": 0.244140625, "learning_rate": 7.647613609311455e-06, "loss": 0.492, "step": 2085 }, { "epoch": 8.874734607218684, "grad_norm": 0.2353515625, "learning_rate": 7.365309958939615e-06, "loss": 0.4913, "step": 2090 }, { "epoch": 8.895966029723992, "grad_norm": 0.2431640625, "learning_rate": 7.088115952851238e-06, "loss": 0.4872, "step": 2095 }, { "epoch": 8.9171974522293, "grad_norm": 0.2490234375, "learning_rate": 6.81604688081271e-06, "loss": 0.49, "step": 2100 }, { "epoch": 8.938428874734607, "grad_norm": 0.267578125, "learning_rate": 6.549117749903755e-06, "loss": 0.4884, "step": 2105 }, { "epoch": 8.959660297239916, "grad_norm": 0.248046875, "learning_rate": 6.287343283689661e-06, "loss": 0.4897, "step": 2110 }, { "epoch": 8.980891719745223, "grad_norm": 0.234375, "learning_rate": 6.030737921409169e-06, "loss": 0.4875, "step": 2115 }, { "epoch": 8.99787685774947, "eval_loss": 1.6812984943389893, "eval_runtime": 0.4908, "eval_samples_per_second": 10.188, "eval_steps_per_second": 2.038, "step": 2119 }, { "epoch": 9.00212314225053, "grad_norm": 0.2373046875, "learning_rate": 5.779315817178e-06, "loss": 0.4972, "step": 2120 }, { "epoch": 9.023354564755838, "grad_norm": 0.2392578125, "learning_rate": 5.533090839208133e-06, "loss": 0.4875, "step": 2125 }, { "epoch": 9.044585987261147, "grad_norm": 0.244140625, "learning_rate": 5.292076569042826e-06, "loss": 0.4966, "step": 2130 }, { "epoch": 9.065817409766455, "grad_norm": 0.2431640625, "learning_rate": 5.056286300807511e-06, "loss": 0.4931, "step": 2135 }, { "epoch": 9.087048832271762, "grad_norm": 0.2353515625, "learning_rate": 4.825733040476465e-06, "loss": 0.4863, "step": 2140 }, { "epoch": 9.10828025477707, "grad_norm": 0.2353515625, "learning_rate": 4.600429505155424e-06, "loss": 0.4945, "step": 2145 }, { "epoch": 9.129511677282379, "grad_norm": 0.26171875, "learning_rate": 4.380388122380141e-06, "loss": 0.4945, "step": 2150 }, { "epoch": 9.150743099787686, "grad_norm": 0.240234375, "learning_rate": 4.165621029430855e-06, "loss": 0.4925, "step": 2155 }, { "epoch": 9.171974522292993, "grad_norm": 0.2421875, "learning_rate": 3.9561400726628505e-06, "loss": 0.4971, "step": 2160 }, { "epoch": 9.1932059447983, "grad_norm": 0.240234375, "learning_rate": 3.7519568068529855e-06, "loss": 0.4901, "step": 2165 }, { "epoch": 9.21443736730361, "grad_norm": 0.2412109375, "learning_rate": 3.5530824945623542e-06, "loss": 0.4841, "step": 2170 }, { "epoch": 9.235668789808917, "grad_norm": 0.236328125, "learning_rate": 3.359528105515064e-06, "loss": 0.4937, "step": 2175 }, { "epoch": 9.256900212314225, "grad_norm": 0.2470703125, "learning_rate": 3.1713043159931734e-06, "loss": 0.4952, "step": 2180 }, { "epoch": 9.278131634819532, "grad_norm": 0.240234375, "learning_rate": 2.9884215082477408e-06, "loss": 0.487, "step": 2185 }, { "epoch": 9.299363057324841, "grad_norm": 0.2353515625, "learning_rate": 2.810889769926217e-06, "loss": 0.489, "step": 2190 }, { "epoch": 9.320594479830149, "grad_norm": 0.24609375, "learning_rate": 2.6387188935159456e-06, "loss": 0.4974, "step": 2195 }, { "epoch": 9.341825902335456, "grad_norm": 0.2431640625, "learning_rate": 2.471918375804105e-06, "loss": 0.4921, "step": 2200 }, { "epoch": 9.363057324840764, "grad_norm": 0.2470703125, "learning_rate": 2.3104974173537743e-06, "loss": 0.4915, "step": 2205 }, { "epoch": 9.384288747346073, "grad_norm": 0.232421875, "learning_rate": 2.1544649219965575e-06, "loss": 0.4812, "step": 2210 }, { "epoch": 9.40552016985138, "grad_norm": 0.2421875, "learning_rate": 2.003829496341325e-06, "loss": 0.4857, "step": 2215 }, { "epoch": 9.426751592356688, "grad_norm": 0.24609375, "learning_rate": 1.8585994492995916e-06, "loss": 0.4962, "step": 2220 }, { "epoch": 9.447983014861995, "grad_norm": 0.244140625, "learning_rate": 1.7187827916271382e-06, "loss": 0.4935, "step": 2225 }, { "epoch": 9.469214437367304, "grad_norm": 0.2412109375, "learning_rate": 1.5843872354822097e-06, "loss": 0.4898, "step": 2230 }, { "epoch": 9.490445859872612, "grad_norm": 0.2421875, "learning_rate": 1.4554201940000123e-06, "loss": 0.489, "step": 2235 }, { "epoch": 9.511677282377919, "grad_norm": 0.2412109375, "learning_rate": 1.3318887808839274e-06, "loss": 0.4964, "step": 2240 }, { "epoch": 9.532908704883226, "grad_norm": 0.2412109375, "learning_rate": 1.21379981001305e-06, "loss": 0.4868, "step": 2245 }, { "epoch": 9.554140127388536, "grad_norm": 0.236328125, "learning_rate": 1.1011597950663865e-06, "loss": 0.495, "step": 2250 }, { "epoch": 9.575371549893843, "grad_norm": 0.23828125, "learning_rate": 9.939749491635341e-07, "loss": 0.4956, "step": 2255 }, { "epoch": 9.59660297239915, "grad_norm": 0.24609375, "learning_rate": 8.922511845219971e-07, "loss": 0.4862, "step": 2260 }, { "epoch": 9.617834394904458, "grad_norm": 0.2373046875, "learning_rate": 7.959941121310266e-07, "loss": 0.4904, "step": 2265 }, { "epoch": 9.639065817409767, "grad_norm": 0.2470703125, "learning_rate": 7.052090414422119e-07, "loss": 0.4892, "step": 2270 }, { "epoch": 9.660297239915074, "grad_norm": 0.2451171875, "learning_rate": 6.199009800765265e-07, "loss": 0.5003, "step": 2275 }, { "epoch": 9.681528662420382, "grad_norm": 0.2412109375, "learning_rate": 5.400746335481488e-07, "loss": 0.485, "step": 2280 }, { "epoch": 9.70276008492569, "grad_norm": 0.248046875, "learning_rate": 4.6573440500492504e-07, "loss": 0.4933, "step": 2285 }, { "epoch": 9.723991507430998, "grad_norm": 0.2421875, "learning_rate": 3.96884394985475e-07, "loss": 0.4817, "step": 2290 }, { "epoch": 9.745222929936306, "grad_norm": 0.25, "learning_rate": 3.335284011929951e-07, "loss": 0.4828, "step": 2295 }, { "epoch": 9.766454352441613, "grad_norm": 0.2578125, "learning_rate": 2.756699182858369e-07, "loss": 0.4922, "step": 2300 }, { "epoch": 9.787685774946922, "grad_norm": 0.2470703125, "learning_rate": 2.2331213768468363e-07, "loss": 0.493, "step": 2305 }, { "epoch": 9.80891719745223, "grad_norm": 0.251953125, "learning_rate": 1.7645794739654665e-07, "loss": 0.4972, "step": 2310 }, { "epoch": 9.830148619957537, "grad_norm": 0.2431640625, "learning_rate": 1.351099318554705e-07, "loss": 0.4881, "step": 2315 }, { "epoch": 9.851380042462845, "grad_norm": 0.251953125, "learning_rate": 9.927037177993592e-08, "loss": 0.49, "step": 2320 }, { "epoch": 9.872611464968152, "grad_norm": 0.240234375, "learning_rate": 6.894124404711599e-08, "loss": 0.4973, "step": 2325 }, { "epoch": 9.893842887473461, "grad_norm": 0.236328125, "learning_rate": 4.4124221583785595e-08, "loss": 0.4918, "step": 2330 }, { "epoch": 9.915074309978769, "grad_norm": 0.244140625, "learning_rate": 2.482067327409521e-08, "loss": 0.4923, "step": 2335 }, { "epoch": 9.936305732484076, "grad_norm": 0.2392578125, "learning_rate": 1.103166388398691e-08, "loss": 0.4791, "step": 2340 }, { "epoch": 9.957537154989385, "grad_norm": 0.2470703125, "learning_rate": 2.7579540025524097e-09, "loss": 0.4883, "step": 2345 }, { "epoch": 9.978768577494693, "grad_norm": 0.2451171875, "learning_rate": 0.0, "loss": 0.4904, "step": 2350 }, { "epoch": 9.978768577494693, "eval_loss": 1.6824607849121094, "eval_runtime": 0.4916, "eval_samples_per_second": 10.17, "eval_steps_per_second": 2.034, "step": 2350 }, { "epoch": 9.978768577494693, "step": 2350, "total_flos": 1.8676199416922112e+18, "train_loss": 0.5910337928000917, "train_runtime": 14730.3274, "train_samples_per_second": 10.216, "train_steps_per_second": 0.16 } ], "logging_steps": 5, "max_steps": 2350, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.8676199416922112e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }