| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.1912416663846492, | |
| "eval_steps": 100, | |
| "global_step": 8800, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0001353683711800738, | |
| "eval_loss": 0.05271552875638008, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 121.9881, | |
| "eval_samples_per_second": 4.099, | |
| "eval_steps_per_second": 1.025, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0027073674236014755, | |
| "grad_norm": 4.584434986114502, | |
| "learning_rate": 1e-05, | |
| "loss": 0.1125, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.005414734847202951, | |
| "grad_norm": 8.236103057861328, | |
| "learning_rate": 2e-05, | |
| "loss": 0.1108, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.008122102270804427, | |
| "grad_norm": 29.931549072265625, | |
| "learning_rate": 3e-05, | |
| "loss": 0.1274, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.010829469694405902, | |
| "grad_norm": 0.10951080918312073, | |
| "learning_rate": 4e-05, | |
| "loss": 0.096, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.013536837118007378, | |
| "grad_norm": 3.041987419128418, | |
| "learning_rate": 5e-05, | |
| "loss": 0.1144, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.013536837118007378, | |
| "eval_loss": 0.05268385633826256, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 126.9017, | |
| "eval_samples_per_second": 3.94, | |
| "eval_steps_per_second": 0.985, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.016244204541608853, | |
| "grad_norm": 5.450573444366455, | |
| "learning_rate": 4.997285190715352e-05, | |
| "loss": 0.1957, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.01895157196521033, | |
| "grad_norm": 10.190834999084473, | |
| "learning_rate": 4.994570381430705e-05, | |
| "loss": 0.0732, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.021658939388811804, | |
| "grad_norm": 0.6384797096252441, | |
| "learning_rate": 4.991855572146057e-05, | |
| "loss": 0.2031, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.02436630681241328, | |
| "grad_norm": 33.95022964477539, | |
| "learning_rate": 4.989140762861409e-05, | |
| "loss": 0.3639, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.027073674236014755, | |
| "grad_norm": 0.5610571503639221, | |
| "learning_rate": 4.9864259535767615e-05, | |
| "loss": 0.2497, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.027073674236014755, | |
| "eval_loss": 0.05859573185443878, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 126.668, | |
| "eval_samples_per_second": 3.947, | |
| "eval_steps_per_second": 0.987, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.02978104165961623, | |
| "grad_norm": 27.80652618408203, | |
| "learning_rate": 4.983711144292114e-05, | |
| "loss": 0.2845, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.032488409083217706, | |
| "grad_norm": 0.5620715022087097, | |
| "learning_rate": 4.980996335007466e-05, | |
| "loss": 0.11, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.03519577650681918, | |
| "grad_norm": 0.33994266390800476, | |
| "learning_rate": 4.978281525722818e-05, | |
| "loss": 0.3338, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.03790314393042066, | |
| "grad_norm": 0.8733130097389221, | |
| "learning_rate": 4.9755667164381704e-05, | |
| "loss": 0.2284, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.04061051135402213, | |
| "grad_norm": 6.556461811065674, | |
| "learning_rate": 4.972851907153523e-05, | |
| "loss": 0.1527, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.04061051135402213, | |
| "eval_loss": 0.053429730236530304, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 127.1913, | |
| "eval_samples_per_second": 3.931, | |
| "eval_steps_per_second": 0.983, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.04331787877762361, | |
| "grad_norm": 0.0169974397867918, | |
| "learning_rate": 4.970137097868875e-05, | |
| "loss": 0.0819, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.046025246201225084, | |
| "grad_norm": 10.627715110778809, | |
| "learning_rate": 4.967422288584227e-05, | |
| "loss": 0.2767, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.04873261362482656, | |
| "grad_norm": 3.286104440689087, | |
| "learning_rate": 4.964707479299579e-05, | |
| "loss": 0.1822, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.051439981048428035, | |
| "grad_norm": 9.183119773864746, | |
| "learning_rate": 4.9619926700149316e-05, | |
| "loss": 0.2516, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.05414734847202951, | |
| "grad_norm": 31.08264923095703, | |
| "learning_rate": 4.9592778607302834e-05, | |
| "loss": 0.1377, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.05414734847202951, | |
| "eval_loss": 0.06496748328208923, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 126.6592, | |
| "eval_samples_per_second": 3.948, | |
| "eval_steps_per_second": 0.987, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.056854715895630986, | |
| "grad_norm": 0.8989093899726868, | |
| "learning_rate": 4.9565630514456364e-05, | |
| "loss": 0.0798, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.05956208331923246, | |
| "grad_norm": 5.862271785736084, | |
| "learning_rate": 4.953848242160988e-05, | |
| "loss": 0.0596, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.06226945074283394, | |
| "grad_norm": 29.281267166137695, | |
| "learning_rate": 4.951133432876341e-05, | |
| "loss": 0.2643, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.06497681816643541, | |
| "grad_norm": 1.005914330482483, | |
| "learning_rate": 4.948418623591693e-05, | |
| "loss": 0.4338, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.06768418559003689, | |
| "grad_norm": 1.3645344972610474, | |
| "learning_rate": 4.945703814307045e-05, | |
| "loss": 0.0815, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.06768418559003689, | |
| "eval_loss": 0.06759323924779892, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 135.4497, | |
| "eval_samples_per_second": 3.691, | |
| "eval_steps_per_second": 0.923, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.07039155301363836, | |
| "grad_norm": 16.54131317138672, | |
| "learning_rate": 4.942989005022398e-05, | |
| "loss": 0.156, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.07309892043723984, | |
| "grad_norm": 1.9615036249160767, | |
| "learning_rate": 4.9402741957377494e-05, | |
| "loss": 0.2292, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.07580628786084131, | |
| "grad_norm": 1.1833112239837646, | |
| "learning_rate": 4.937559386453102e-05, | |
| "loss": 0.1009, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.07851365528444279, | |
| "grad_norm": 0.9788858294487, | |
| "learning_rate": 4.934844577168454e-05, | |
| "loss": 0.047, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.08122102270804427, | |
| "grad_norm": 0.8096566200256348, | |
| "learning_rate": 4.9321297678838066e-05, | |
| "loss": 0.1603, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.08122102270804427, | |
| "eval_loss": 0.0586598739027977, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 126.1816, | |
| "eval_samples_per_second": 3.963, | |
| "eval_steps_per_second": 0.991, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.08392839013164574, | |
| "grad_norm": 0.018350690603256226, | |
| "learning_rate": 4.929414958599159e-05, | |
| "loss": 0.1014, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.08663575755524722, | |
| "grad_norm": 1.8620424270629883, | |
| "learning_rate": 4.9267001493145107e-05, | |
| "loss": 0.2682, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.08934312497884869, | |
| "grad_norm": 22.248580932617188, | |
| "learning_rate": 4.923985340029863e-05, | |
| "loss": 0.1485, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.09205049240245017, | |
| "grad_norm": 0.05200627073645592, | |
| "learning_rate": 4.9212705307452154e-05, | |
| "loss": 0.1227, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.09475785982605164, | |
| "grad_norm": 1.1178234815597534, | |
| "learning_rate": 4.918555721460568e-05, | |
| "loss": 0.1507, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.09475785982605164, | |
| "eval_loss": 0.05151496082544327, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 126.9972, | |
| "eval_samples_per_second": 3.937, | |
| "eval_steps_per_second": 0.984, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.09746522724965312, | |
| "grad_norm": 1.1798577308654785, | |
| "learning_rate": 4.9158409121759195e-05, | |
| "loss": 0.2476, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.1001725946732546, | |
| "grad_norm": 21.11642074584961, | |
| "learning_rate": 4.9131261028912726e-05, | |
| "loss": 0.2516, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.10287996209685607, | |
| "grad_norm": 21.654647827148438, | |
| "learning_rate": 4.910411293606624e-05, | |
| "loss": 0.2303, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.10558732952045755, | |
| "grad_norm": 27.137983322143555, | |
| "learning_rate": 4.907696484321977e-05, | |
| "loss": 0.1943, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.10829469694405902, | |
| "grad_norm": 12.374953269958496, | |
| "learning_rate": 4.904981675037329e-05, | |
| "loss": 0.2747, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.10829469694405902, | |
| "eval_loss": 0.04837380722165108, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 120.5107, | |
| "eval_samples_per_second": 4.149, | |
| "eval_steps_per_second": 1.037, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.1110020643676605, | |
| "grad_norm": 22.375734329223633, | |
| "learning_rate": 4.902266865752681e-05, | |
| "loss": 0.3168, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.11370943179126197, | |
| "grad_norm": 0.05682824179530144, | |
| "learning_rate": 4.899552056468034e-05, | |
| "loss": 0.065, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.11641679921486345, | |
| "grad_norm": 8.457867622375488, | |
| "learning_rate": 4.8968372471833856e-05, | |
| "loss": 0.1069, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.11912416663846492, | |
| "grad_norm": 0.8650844693183899, | |
| "learning_rate": 4.894122437898738e-05, | |
| "loss": 0.0799, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.1218315340620664, | |
| "grad_norm": 10.979653358459473, | |
| "learning_rate": 4.8914076286140903e-05, | |
| "loss": 0.3638, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.1218315340620664, | |
| "eval_loss": 0.048302676528692245, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 125.5881, | |
| "eval_samples_per_second": 3.981, | |
| "eval_steps_per_second": 0.995, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.12453890148566787, | |
| "grad_norm": 1.0572680234909058, | |
| "learning_rate": 4.888692819329442e-05, | |
| "loss": 0.1154, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.12724626890926935, | |
| "grad_norm": 0.3924155533313751, | |
| "learning_rate": 4.8859780100447944e-05, | |
| "loss": 0.1085, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.12995363633287083, | |
| "grad_norm": 0.07971794903278351, | |
| "learning_rate": 4.883263200760147e-05, | |
| "loss": 0.1827, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.1326610037564723, | |
| "grad_norm": 1.397789478302002, | |
| "learning_rate": 4.880548391475499e-05, | |
| "loss": 0.0695, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.13536837118007378, | |
| "grad_norm": 3.120443105697632, | |
| "learning_rate": 4.877833582190851e-05, | |
| "loss": 0.149, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.13536837118007378, | |
| "eval_loss": 0.057031989097595215, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.6549, | |
| "eval_samples_per_second": 4.323, | |
| "eval_steps_per_second": 1.081, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.13807573860367525, | |
| "grad_norm": 1.3726035356521606, | |
| "learning_rate": 4.875118772906204e-05, | |
| "loss": 0.0567, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.14078310602727673, | |
| "grad_norm": 0.16208282113075256, | |
| "learning_rate": 4.872403963621556e-05, | |
| "loss": 0.2078, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.1434904734508782, | |
| "grad_norm": 1.425825834274292, | |
| "learning_rate": 4.869689154336908e-05, | |
| "loss": 0.353, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.14619784087447968, | |
| "grad_norm": 16.413312911987305, | |
| "learning_rate": 4.8669743450522605e-05, | |
| "loss": 0.0656, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.14890520829808115, | |
| "grad_norm": 9.234221458435059, | |
| "learning_rate": 4.864259535767612e-05, | |
| "loss": 0.2405, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.14890520829808115, | |
| "eval_loss": 0.04857669770717621, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.5724, | |
| "eval_samples_per_second": 4.326, | |
| "eval_steps_per_second": 1.082, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.15161257572168263, | |
| "grad_norm": 39.68949890136719, | |
| "learning_rate": 4.861544726482965e-05, | |
| "loss": 0.2318, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.1543199431452841, | |
| "grad_norm": 0.022096823900938034, | |
| "learning_rate": 4.858829917198317e-05, | |
| "loss": 0.1425, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.15702731056888558, | |
| "grad_norm": 23.582176208496094, | |
| "learning_rate": 4.8561151079136694e-05, | |
| "loss": 0.2259, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.15973467799248706, | |
| "grad_norm": 17.384206771850586, | |
| "learning_rate": 4.853400298629022e-05, | |
| "loss": 0.1154, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.16244204541608853, | |
| "grad_norm": 14.885477066040039, | |
| "learning_rate": 4.8506854893443735e-05, | |
| "loss": 0.2201, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.16244204541608853, | |
| "eval_loss": 0.043531883507966995, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2235, | |
| "eval_samples_per_second": 4.339, | |
| "eval_steps_per_second": 1.085, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.16514941283969, | |
| "grad_norm": 4.196330547332764, | |
| "learning_rate": 4.847970680059726e-05, | |
| "loss": 0.1639, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.16785678026329148, | |
| "grad_norm": 27.09263038635254, | |
| "learning_rate": 4.845255870775078e-05, | |
| "loss": 0.1695, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.17056414768689296, | |
| "grad_norm": 31.63964080810547, | |
| "learning_rate": 4.8425410614904306e-05, | |
| "loss": 0.1279, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.17327151511049443, | |
| "grad_norm": 0.0727367028594017, | |
| "learning_rate": 4.839826252205783e-05, | |
| "loss": 0.1283, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.1759788825340959, | |
| "grad_norm": 5.623673915863037, | |
| "learning_rate": 4.837111442921135e-05, | |
| "loss": 0.195, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.1759788825340959, | |
| "eval_loss": 0.040112994611263275, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.6438, | |
| "eval_samples_per_second": 4.324, | |
| "eval_steps_per_second": 1.081, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.17868624995769739, | |
| "grad_norm": 6.618963241577148, | |
| "learning_rate": 4.834396633636487e-05, | |
| "loss": 0.1477, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.18139361738129886, | |
| "grad_norm": 0.15700267255306244, | |
| "learning_rate": 4.8316818243518395e-05, | |
| "loss": 0.082, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.18410098480490034, | |
| "grad_norm": 0.07716820389032364, | |
| "learning_rate": 4.828967015067192e-05, | |
| "loss": 0.1526, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.1868083522285018, | |
| "grad_norm": 0.1871941089630127, | |
| "learning_rate": 4.8262522057825436e-05, | |
| "loss": 0.2056, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.1895157196521033, | |
| "grad_norm": 7.145862579345703, | |
| "learning_rate": 4.823537396497897e-05, | |
| "loss": 0.186, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.1895157196521033, | |
| "eval_loss": 0.04851188883185387, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2328, | |
| "eval_samples_per_second": 4.339, | |
| "eval_steps_per_second": 1.085, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.19222308707570476, | |
| "grad_norm": 0.10455508530139923, | |
| "learning_rate": 4.8208225872132484e-05, | |
| "loss": 0.3346, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.19493045449930624, | |
| "grad_norm": 6.885641098022461, | |
| "learning_rate": 4.818107777928601e-05, | |
| "loss": 0.2167, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.1976378219229077, | |
| "grad_norm": 12.304588317871094, | |
| "learning_rate": 4.815392968643953e-05, | |
| "loss": 0.1419, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.2003451893465092, | |
| "grad_norm": 12.884013175964355, | |
| "learning_rate": 4.812678159359305e-05, | |
| "loss": 0.2035, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.20305255677011066, | |
| "grad_norm": 22.836992263793945, | |
| "learning_rate": 4.809963350074658e-05, | |
| "loss": 0.1116, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.20305255677011066, | |
| "eval_loss": 0.055405810475349426, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.8109, | |
| "eval_samples_per_second": 4.317, | |
| "eval_steps_per_second": 1.079, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.20575992419371214, | |
| "grad_norm": 2.097397565841675, | |
| "learning_rate": 4.8072485407900096e-05, | |
| "loss": 0.0493, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.20846729161731362, | |
| "grad_norm": 8.2213773727417, | |
| "learning_rate": 4.804533731505362e-05, | |
| "loss": 0.0844, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.2111746590409151, | |
| "grad_norm": 0.014365073293447495, | |
| "learning_rate": 4.8018189222207144e-05, | |
| "loss": 0.2243, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.21388202646451657, | |
| "grad_norm": 0.2695819139480591, | |
| "learning_rate": 4.799104112936066e-05, | |
| "loss": 0.1934, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.21658939388811804, | |
| "grad_norm": 0.25374430418014526, | |
| "learning_rate": 4.7963893036514185e-05, | |
| "loss": 0.2522, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.21658939388811804, | |
| "eval_loss": 0.059707771986722946, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.274, | |
| "eval_samples_per_second": 4.375, | |
| "eval_steps_per_second": 1.094, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.21929676131171952, | |
| "grad_norm": 0.003937189467251301, | |
| "learning_rate": 4.793674494366771e-05, | |
| "loss": 0.0801, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.222004128735321, | |
| "grad_norm": 0.9062768220901489, | |
| "learning_rate": 4.790959685082123e-05, | |
| "loss": 0.2022, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.22471149615892247, | |
| "grad_norm": 3.008561849594116, | |
| "learning_rate": 4.788244875797476e-05, | |
| "loss": 0.06, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.22741886358252394, | |
| "grad_norm": 0.8405902981758118, | |
| "learning_rate": 4.785530066512828e-05, | |
| "loss": 0.2472, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.23012623100612542, | |
| "grad_norm": 8.165879249572754, | |
| "learning_rate": 4.78281525722818e-05, | |
| "loss": 0.1708, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.23012623100612542, | |
| "eval_loss": 0.057815808802843094, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0717, | |
| "eval_samples_per_second": 4.345, | |
| "eval_steps_per_second": 1.086, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.2328335984297269, | |
| "grad_norm": 0.13060545921325684, | |
| "learning_rate": 4.780100447943532e-05, | |
| "loss": 0.117, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.23554096585332837, | |
| "grad_norm": 12.249765396118164, | |
| "learning_rate": 4.7773856386588846e-05, | |
| "loss": 0.1507, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.23824833327692985, | |
| "grad_norm": 1.8489207029342651, | |
| "learning_rate": 4.774670829374236e-05, | |
| "loss": 0.3488, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.24095570070053132, | |
| "grad_norm": 0.9456131458282471, | |
| "learning_rate": 4.771956020089589e-05, | |
| "loss": 0.3084, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.2436630681241328, | |
| "grad_norm": 24.736726760864258, | |
| "learning_rate": 4.769241210804941e-05, | |
| "loss": 0.2627, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.2436630681241328, | |
| "eval_loss": 0.06722074747085571, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.4272, | |
| "eval_samples_per_second": 4.332, | |
| "eval_steps_per_second": 1.083, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.24637043554773427, | |
| "grad_norm": 2.1300015449523926, | |
| "learning_rate": 4.7665264015202934e-05, | |
| "loss": 0.1497, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.24907780297133575, | |
| "grad_norm": 7.431568145751953, | |
| "learning_rate": 4.763811592235646e-05, | |
| "loss": 0.1287, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.2517851703949372, | |
| "grad_norm": 0.017447171732783318, | |
| "learning_rate": 4.7610967829509975e-05, | |
| "loss": 0.0981, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.2544925378185387, | |
| "grad_norm": 0.7597976922988892, | |
| "learning_rate": 4.7583819736663506e-05, | |
| "loss": 0.1814, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.25719990524214015, | |
| "grad_norm": 2.8532488346099854, | |
| "learning_rate": 4.755667164381702e-05, | |
| "loss": 0.2901, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.25719990524214015, | |
| "eval_loss": 0.05738656595349312, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.4019, | |
| "eval_samples_per_second": 4.371, | |
| "eval_steps_per_second": 1.093, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.25990727266574165, | |
| "grad_norm": 6.662034511566162, | |
| "learning_rate": 4.752952355097055e-05, | |
| "loss": 0.1815, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.2626146400893431, | |
| "grad_norm": 0.4220564365386963, | |
| "learning_rate": 4.750237545812407e-05, | |
| "loss": 0.0649, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.2653220075129446, | |
| "grad_norm": 7.456377029418945, | |
| "learning_rate": 4.7475227365277595e-05, | |
| "loss": 0.1509, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.26802937493654605, | |
| "grad_norm": 0.16843333840370178, | |
| "learning_rate": 4.744807927243111e-05, | |
| "loss": 0.1159, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.27073674236014755, | |
| "grad_norm": 0.038046594709157944, | |
| "learning_rate": 4.7420931179584636e-05, | |
| "loss": 0.0354, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.27073674236014755, | |
| "eval_loss": 0.09768073260784149, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.1145, | |
| "eval_samples_per_second": 4.344, | |
| "eval_steps_per_second": 1.086, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.273444109783749, | |
| "grad_norm": 19.501468658447266, | |
| "learning_rate": 4.739378308673816e-05, | |
| "loss": 0.1081, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.2761514772073505, | |
| "grad_norm": 11.927206993103027, | |
| "learning_rate": 4.736663499389168e-05, | |
| "loss": 0.0767, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.27885884463095195, | |
| "grad_norm": 1.6448673009872437, | |
| "learning_rate": 4.733948690104521e-05, | |
| "loss": 0.1237, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.28156621205455346, | |
| "grad_norm": 8.10880184173584, | |
| "learning_rate": 4.7312338808198724e-05, | |
| "loss": 0.0313, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.2842735794781549, | |
| "grad_norm": 10.024765968322754, | |
| "learning_rate": 4.728519071535225e-05, | |
| "loss": 0.156, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.2842735794781549, | |
| "eval_loss": 0.08133851736783981, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2526, | |
| "eval_samples_per_second": 4.338, | |
| "eval_steps_per_second": 1.085, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.2869809469017564, | |
| "grad_norm": 0.131832554936409, | |
| "learning_rate": 4.725804262250577e-05, | |
| "loss": 0.1318, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.28968831432535785, | |
| "grad_norm": 12.999658584594727, | |
| "learning_rate": 4.723089452965929e-05, | |
| "loss": 0.2504, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.29239568174895936, | |
| "grad_norm": 12.960088729858398, | |
| "learning_rate": 4.720374643681282e-05, | |
| "loss": 0.2561, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.2951030491725608, | |
| "grad_norm": 2.730931043624878, | |
| "learning_rate": 4.717659834396634e-05, | |
| "loss": 0.2609, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.2978104165961623, | |
| "grad_norm": 28.357595443725586, | |
| "learning_rate": 4.714945025111986e-05, | |
| "loss": 0.1156, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.2978104165961623, | |
| "eval_loss": 0.06201322376728058, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.4336, | |
| "eval_samples_per_second": 4.331, | |
| "eval_steps_per_second": 1.083, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.30051778401976376, | |
| "grad_norm": 21.61467170715332, | |
| "learning_rate": 4.7122302158273385e-05, | |
| "loss": 0.139, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.30322515144336526, | |
| "grad_norm": 1.2531449794769287, | |
| "learning_rate": 4.70951540654269e-05, | |
| "loss": 0.1351, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.3059325188669667, | |
| "grad_norm": 0.1177072823047638, | |
| "learning_rate": 4.706800597258043e-05, | |
| "loss": 0.2042, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.3086398862905682, | |
| "grad_norm": 0.035650696605443954, | |
| "learning_rate": 4.704085787973395e-05, | |
| "loss": 0.165, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.31134725371416966, | |
| "grad_norm": 0.03751590847969055, | |
| "learning_rate": 4.7013709786887474e-05, | |
| "loss": 0.1089, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.31134725371416966, | |
| "eval_loss": 0.06733442097902298, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.3253, | |
| "eval_samples_per_second": 4.336, | |
| "eval_steps_per_second": 1.084, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.31405462113777116, | |
| "grad_norm": 61.547889709472656, | |
| "learning_rate": 4.6986561694041e-05, | |
| "loss": 0.319, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.3167619885613726, | |
| "grad_norm": 20.737930297851562, | |
| "learning_rate": 4.695941360119452e-05, | |
| "loss": 0.1578, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.3194693559849741, | |
| "grad_norm": 0.3770120143890381, | |
| "learning_rate": 4.693226550834804e-05, | |
| "loss": 0.1038, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.32217672340857556, | |
| "grad_norm": 12.355575561523438, | |
| "learning_rate": 4.690511741550156e-05, | |
| "loss": 0.1896, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.32488409083217706, | |
| "grad_norm": 7.065931797027588, | |
| "learning_rate": 4.6877969322655086e-05, | |
| "loss": 0.1368, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.32488409083217706, | |
| "eval_loss": 0.058929502964019775, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0859, | |
| "eval_samples_per_second": 4.345, | |
| "eval_steps_per_second": 1.086, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.3275914582557785, | |
| "grad_norm": 1.2074625492095947, | |
| "learning_rate": 4.68508212298086e-05, | |
| "loss": 0.2128, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.33029882567938, | |
| "grad_norm": 24.548015594482422, | |
| "learning_rate": 4.6823673136962134e-05, | |
| "loss": 0.1701, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.33300619310298146, | |
| "grad_norm": 1.5518879890441895, | |
| "learning_rate": 4.679652504411565e-05, | |
| "loss": 0.2455, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.33571356052658297, | |
| "grad_norm": 0.4867059886455536, | |
| "learning_rate": 4.6769376951269175e-05, | |
| "loss": 0.2979, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.3384209279501844, | |
| "grad_norm": 2.8492958545684814, | |
| "learning_rate": 4.67422288584227e-05, | |
| "loss": 0.1853, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.3384209279501844, | |
| "eval_loss": 0.06687568873167038, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.1592, | |
| "eval_samples_per_second": 4.342, | |
| "eval_steps_per_second": 1.085, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.3411282953737859, | |
| "grad_norm": 1.295285940170288, | |
| "learning_rate": 4.6715080765576216e-05, | |
| "loss": 0.0603, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.34383566279738736, | |
| "grad_norm": 0.0439661480486393, | |
| "learning_rate": 4.6687932672729747e-05, | |
| "loss": 0.0869, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.34654303022098887, | |
| "grad_norm": 42.599430084228516, | |
| "learning_rate": 4.6660784579883264e-05, | |
| "loss": 0.0603, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.3492503976445903, | |
| "grad_norm": 0.09535373747348785, | |
| "learning_rate": 4.663363648703679e-05, | |
| "loss": 0.3583, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.3519577650681918, | |
| "grad_norm": 0.08492514491081238, | |
| "learning_rate": 4.660648839419031e-05, | |
| "loss": 0.1109, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.3519577650681918, | |
| "eval_loss": 0.07154419273138046, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.7194, | |
| "eval_samples_per_second": 4.321, | |
| "eval_steps_per_second": 1.08, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.35466513249179327, | |
| "grad_norm": 0.145077183842659, | |
| "learning_rate": 4.6579340301343835e-05, | |
| "loss": 0.0875, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.35737249991539477, | |
| "grad_norm": 9.414113998413086, | |
| "learning_rate": 4.655219220849735e-05, | |
| "loss": 0.2708, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.3600798673389962, | |
| "grad_norm": 0.012831827625632286, | |
| "learning_rate": 4.6525044115650876e-05, | |
| "loss": 0.3477, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.3627872347625977, | |
| "grad_norm": 0.06755544990301132, | |
| "learning_rate": 4.64978960228044e-05, | |
| "loss": 0.2059, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.36549460218619917, | |
| "grad_norm": 0.40635791420936584, | |
| "learning_rate": 4.6470747929957924e-05, | |
| "loss": 0.1477, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.36549460218619917, | |
| "eval_loss": 0.08114828914403915, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.4236, | |
| "eval_samples_per_second": 4.332, | |
| "eval_steps_per_second": 1.083, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.3682019696098007, | |
| "grad_norm": 21.508838653564453, | |
| "learning_rate": 4.644359983711145e-05, | |
| "loss": 0.1263, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.3709093370334021, | |
| "grad_norm": 0.01996573433279991, | |
| "learning_rate": 4.6416451744264965e-05, | |
| "loss": 0.0615, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.3736167044570036, | |
| "grad_norm": 0.048173803836107254, | |
| "learning_rate": 4.638930365141849e-05, | |
| "loss": 0.1578, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.37632407188060507, | |
| "grad_norm": 9.07430362701416, | |
| "learning_rate": 4.636215555857201e-05, | |
| "loss": 0.1826, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.3790314393042066, | |
| "grad_norm": 0.015813810750842094, | |
| "learning_rate": 4.633500746572553e-05, | |
| "loss": 0.0999, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.3790314393042066, | |
| "eval_loss": 0.07192689180374146, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7893, | |
| "eval_samples_per_second": 4.356, | |
| "eval_steps_per_second": 1.089, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.381738806727808, | |
| "grad_norm": 17.83676528930664, | |
| "learning_rate": 4.630785937287906e-05, | |
| "loss": 0.3131, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.3844461741514095, | |
| "grad_norm": 0.2090553194284439, | |
| "learning_rate": 4.628071128003258e-05, | |
| "loss": 0.1114, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.387153541575011, | |
| "grad_norm": 15.477974891662598, | |
| "learning_rate": 4.62535631871861e-05, | |
| "loss": 0.1856, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.3898609089986125, | |
| "grad_norm": 2.8544585704803467, | |
| "learning_rate": 4.6226415094339625e-05, | |
| "loss": 0.1231, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.3925682764222139, | |
| "grad_norm": 4.374979019165039, | |
| "learning_rate": 4.619926700149315e-05, | |
| "loss": 0.0525, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.3925682764222139, | |
| "eval_loss": 0.07568126916885376, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.4166, | |
| "eval_samples_per_second": 4.332, | |
| "eval_steps_per_second": 1.083, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.3952756438458154, | |
| "grad_norm": 17.247142791748047, | |
| "learning_rate": 4.617211890864667e-05, | |
| "loss": 0.1124, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.3979830112694169, | |
| "grad_norm": 0.016074227169156075, | |
| "learning_rate": 4.614497081580019e-05, | |
| "loss": 0.0538, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.4006903786930184, | |
| "grad_norm": 0.5515703558921814, | |
| "learning_rate": 4.6117822722953714e-05, | |
| "loss": 0.1552, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.4033977461166198, | |
| "grad_norm": 0.03307110071182251, | |
| "learning_rate": 4.609067463010724e-05, | |
| "loss": 0.0419, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.40610511354022133, | |
| "grad_norm": 0.1128387525677681, | |
| "learning_rate": 4.606352653726076e-05, | |
| "loss": 0.0513, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.40610511354022133, | |
| "eval_loss": 0.0807776153087616, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7987, | |
| "eval_samples_per_second": 4.355, | |
| "eval_steps_per_second": 1.089, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.4088124809638228, | |
| "grad_norm": 7.236224174499512, | |
| "learning_rate": 4.603637844441428e-05, | |
| "loss": 0.1509, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.4115198483874243, | |
| "grad_norm": 13.330012321472168, | |
| "learning_rate": 4.60092303515678e-05, | |
| "loss": 0.2394, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.41422721581102573, | |
| "grad_norm": 16.844871520996094, | |
| "learning_rate": 4.598208225872133e-05, | |
| "loss": 0.1604, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.41693458323462723, | |
| "grad_norm": 0.01411959808319807, | |
| "learning_rate": 4.595493416587485e-05, | |
| "loss": 0.0778, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.4196419506582287, | |
| "grad_norm": 5.0137529373168945, | |
| "learning_rate": 4.5927786073028375e-05, | |
| "loss": 0.1562, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.4196419506582287, | |
| "eval_loss": 0.06787914782762527, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2644, | |
| "eval_samples_per_second": 4.338, | |
| "eval_steps_per_second": 1.084, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.4223493180818302, | |
| "grad_norm": 2.870209217071533, | |
| "learning_rate": 4.590063798018189e-05, | |
| "loss": 0.1097, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.42505668550543163, | |
| "grad_norm": 0.18988296389579773, | |
| "learning_rate": 4.587348988733542e-05, | |
| "loss": 0.1236, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.42776405292903313, | |
| "grad_norm": 2.6342244148254395, | |
| "learning_rate": 4.584634179448894e-05, | |
| "loss": 0.1644, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.4304714203526346, | |
| "grad_norm": 1.666049838066101, | |
| "learning_rate": 4.581919370164246e-05, | |
| "loss": 0.0629, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.4331787877762361, | |
| "grad_norm": 0.0062801106832921505, | |
| "learning_rate": 4.579204560879599e-05, | |
| "loss": 0.1523, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.4331787877762361, | |
| "eval_loss": 0.07749966531991959, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.8371, | |
| "eval_samples_per_second": 4.354, | |
| "eval_steps_per_second": 1.088, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.43588615519983753, | |
| "grad_norm": 0.9460217356681824, | |
| "learning_rate": 4.5764897515949504e-05, | |
| "loss": 0.1426, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.43859352262343904, | |
| "grad_norm": 6.54037618637085, | |
| "learning_rate": 4.573774942310303e-05, | |
| "loss": 0.1225, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.4413008900470405, | |
| "grad_norm": 6.7528395652771, | |
| "learning_rate": 4.571060133025655e-05, | |
| "loss": 0.1678, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.444008257470642, | |
| "grad_norm": 0.906831681728363, | |
| "learning_rate": 4.5683453237410076e-05, | |
| "loss": 0.0716, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.44671562489424343, | |
| "grad_norm": 0.9793009161949158, | |
| "learning_rate": 4.56563051445636e-05, | |
| "loss": 0.1616, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.44671562489424343, | |
| "eval_loss": 0.08035814017057419, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7485, | |
| "eval_samples_per_second": 4.357, | |
| "eval_steps_per_second": 1.089, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.44942299231784494, | |
| "grad_norm": 0.014806450344622135, | |
| "learning_rate": 4.562915705171712e-05, | |
| "loss": 0.1744, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.4521303597414464, | |
| "grad_norm": 20.00578498840332, | |
| "learning_rate": 4.560200895887064e-05, | |
| "loss": 0.0343, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.4548377271650479, | |
| "grad_norm": 0.03944426402449608, | |
| "learning_rate": 4.5574860866024165e-05, | |
| "loss": 0.1867, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.45754509458864934, | |
| "grad_norm": 5.614328861236572, | |
| "learning_rate": 4.554771277317769e-05, | |
| "loss": 0.0663, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.46025246201225084, | |
| "grad_norm": 0.06043994799256325, | |
| "learning_rate": 4.5520564680331206e-05, | |
| "loss": 0.1641, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.46025246201225084, | |
| "eval_loss": 0.08000614494085312, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.1794, | |
| "eval_samples_per_second": 4.341, | |
| "eval_steps_per_second": 1.085, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.4629598294358523, | |
| "grad_norm": 24.133304595947266, | |
| "learning_rate": 4.549341658748473e-05, | |
| "loss": 0.1472, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.4656671968594538, | |
| "grad_norm": 0.7912468314170837, | |
| "learning_rate": 4.5466268494638254e-05, | |
| "loss": 0.2045, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.46837456428305524, | |
| "grad_norm": 21.022037506103516, | |
| "learning_rate": 4.543912040179177e-05, | |
| "loss": 0.242, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.47108193170665674, | |
| "grad_norm": 3.1126956939697266, | |
| "learning_rate": 4.54119723089453e-05, | |
| "loss": 0.1122, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.4737892991302582, | |
| "grad_norm": 0.03333674371242523, | |
| "learning_rate": 4.538482421609882e-05, | |
| "loss": 0.1162, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.4737892991302582, | |
| "eval_loss": 0.06670122593641281, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.6776, | |
| "eval_samples_per_second": 4.322, | |
| "eval_steps_per_second": 1.081, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.4764966665538597, | |
| "grad_norm": 0.3281305134296417, | |
| "learning_rate": 4.535767612325235e-05, | |
| "loss": 0.0489, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.47920403397746114, | |
| "grad_norm": 2.830657720565796, | |
| "learning_rate": 4.5330528030405866e-05, | |
| "loss": 0.1212, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.48191140140106264, | |
| "grad_norm": 1.2127472162246704, | |
| "learning_rate": 4.530337993755939e-05, | |
| "loss": 0.0906, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.4846187688246641, | |
| "grad_norm": 33.909671783447266, | |
| "learning_rate": 4.5276231844712914e-05, | |
| "loss": 0.2042, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.4873261362482656, | |
| "grad_norm": 0.00924315769225359, | |
| "learning_rate": 4.524908375186643e-05, | |
| "loss": 0.1062, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.4873261362482656, | |
| "eval_loss": 0.06282315403223038, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.5799, | |
| "eval_samples_per_second": 4.326, | |
| "eval_steps_per_second": 1.082, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.49003350367186704, | |
| "grad_norm": 0.005617950111627579, | |
| "learning_rate": 4.5221935659019955e-05, | |
| "loss": 0.1927, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.49274087109546855, | |
| "grad_norm": 0.00801560003310442, | |
| "learning_rate": 4.519478756617348e-05, | |
| "loss": 0.1406, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.49544823851907, | |
| "grad_norm": 23.532512664794922, | |
| "learning_rate": 4.5167639473327e-05, | |
| "loss": 0.0821, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.4981556059426715, | |
| "grad_norm": 3.598512649536133, | |
| "learning_rate": 4.514049138048052e-05, | |
| "loss": 0.0751, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.500862973366273, | |
| "grad_norm": 0.09781412780284882, | |
| "learning_rate": 4.5113343287634044e-05, | |
| "loss": 0.2411, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.500862973366273, | |
| "eval_loss": 0.05527465417981148, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.763, | |
| "eval_samples_per_second": 4.319, | |
| "eval_steps_per_second": 1.08, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.5035703407898744, | |
| "grad_norm": 0.02711186371743679, | |
| "learning_rate": 4.508619519478757e-05, | |
| "loss": 0.1311, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.506277708213476, | |
| "grad_norm": 0.5090947151184082, | |
| "learning_rate": 4.505904710194109e-05, | |
| "loss": 0.1294, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.5089850756370774, | |
| "grad_norm": 0.09400524944067001, | |
| "learning_rate": 4.5031899009094615e-05, | |
| "loss": 0.3008, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.5116924430606788, | |
| "grad_norm": 0.27078312635421753, | |
| "learning_rate": 4.500475091624813e-05, | |
| "loss": 0.1703, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.5143998104842803, | |
| "grad_norm": 0.035410646349191666, | |
| "learning_rate": 4.497760282340166e-05, | |
| "loss": 0.1955, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.5143998104842803, | |
| "eval_loss": 0.06636335700750351, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.6171, | |
| "eval_samples_per_second": 4.325, | |
| "eval_steps_per_second": 1.081, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.5171071779078819, | |
| "grad_norm": 2.16019344329834, | |
| "learning_rate": 4.495045473055518e-05, | |
| "loss": 0.3026, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.5198145453314833, | |
| "grad_norm": 0.03840683028101921, | |
| "learning_rate": 4.4923306637708704e-05, | |
| "loss": 0.1793, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.5225219127550847, | |
| "grad_norm": 23.82236099243164, | |
| "learning_rate": 4.489615854486223e-05, | |
| "loss": 0.1801, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.5252292801786862, | |
| "grad_norm": 1.3276259899139404, | |
| "learning_rate": 4.4869010452015745e-05, | |
| "loss": 0.1351, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.5279366476022878, | |
| "grad_norm": 0.8841078877449036, | |
| "learning_rate": 4.4841862359169276e-05, | |
| "loss": 0.1596, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.5279366476022878, | |
| "eval_loss": 0.06823331862688065, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.4954, | |
| "eval_samples_per_second": 4.329, | |
| "eval_steps_per_second": 1.082, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.5306440150258892, | |
| "grad_norm": 0.6198431253433228, | |
| "learning_rate": 4.481471426632279e-05, | |
| "loss": 0.0834, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.5333513824494907, | |
| "grad_norm": 0.3783263564109802, | |
| "learning_rate": 4.478756617347632e-05, | |
| "loss": 0.171, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.5360587498730921, | |
| "grad_norm": 0.045466240495443344, | |
| "learning_rate": 4.476041808062984e-05, | |
| "loss": 0.049, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.5387661172966937, | |
| "grad_norm": 4.255916118621826, | |
| "learning_rate": 4.473326998778336e-05, | |
| "loss": 0.1866, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.5414734847202951, | |
| "grad_norm": 0.548932671546936, | |
| "learning_rate": 4.470612189493688e-05, | |
| "loss": 0.1338, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.5414734847202951, | |
| "eval_loss": 0.05355897173285484, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7915, | |
| "eval_samples_per_second": 4.356, | |
| "eval_steps_per_second": 1.089, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.5441808521438966, | |
| "grad_norm": 3.635627508163452, | |
| "learning_rate": 4.4678973802090405e-05, | |
| "loss": 0.2088, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.546888219567498, | |
| "grad_norm": 1.3527976274490356, | |
| "learning_rate": 4.465182570924393e-05, | |
| "loss": 0.1462, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.5495955869910996, | |
| "grad_norm": 35.63922882080078, | |
| "learning_rate": 4.4624677616397446e-05, | |
| "loss": 0.1028, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.552302954414701, | |
| "grad_norm": 1.668758511543274, | |
| "learning_rate": 4.459752952355098e-05, | |
| "loss": 0.1783, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.5550103218383025, | |
| "grad_norm": 0.10880416631698608, | |
| "learning_rate": 4.4570381430704494e-05, | |
| "loss": 0.1616, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.5550103218383025, | |
| "eval_loss": 0.05879218131303787, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.8074, | |
| "eval_samples_per_second": 4.355, | |
| "eval_steps_per_second": 1.089, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.5577176892619039, | |
| "grad_norm": 4.690010070800781, | |
| "learning_rate": 4.454323333785802e-05, | |
| "loss": 0.122, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.5604250566855055, | |
| "grad_norm": 0.06354126334190369, | |
| "learning_rate": 4.451608524501154e-05, | |
| "loss": 0.1858, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.5631324241091069, | |
| "grad_norm": 7.339563846588135, | |
| "learning_rate": 4.448893715216506e-05, | |
| "loss": 0.092, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.5658397915327084, | |
| "grad_norm": 0.001140138367190957, | |
| "learning_rate": 4.446178905931859e-05, | |
| "loss": 0.1959, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.5685471589563098, | |
| "grad_norm": 0.7119511365890503, | |
| "learning_rate": 4.443464096647211e-05, | |
| "loss": 0.134, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.5685471589563098, | |
| "eval_loss": 0.052126504480838776, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.3761, | |
| "eval_samples_per_second": 4.334, | |
| "eval_steps_per_second": 1.083, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.5712545263799114, | |
| "grad_norm": 0.4891607463359833, | |
| "learning_rate": 4.440749287362563e-05, | |
| "loss": 0.0518, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.5739618938035128, | |
| "grad_norm": 16.58160400390625, | |
| "learning_rate": 4.4380344780779155e-05, | |
| "loss": 0.1553, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.5766692612271143, | |
| "grad_norm": 10.971266746520996, | |
| "learning_rate": 4.435319668793267e-05, | |
| "loss": 0.1916, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.5793766286507157, | |
| "grad_norm": 0.2085973471403122, | |
| "learning_rate": 4.4326048595086196e-05, | |
| "loss": 0.1525, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.5820839960743173, | |
| "grad_norm": 2.1405222415924072, | |
| "learning_rate": 4.429890050223972e-05, | |
| "loss": 0.1593, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.5820839960743173, | |
| "eval_loss": 0.06334167718887329, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 112.7876, | |
| "eval_samples_per_second": 4.433, | |
| "eval_steps_per_second": 1.108, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.5847913634979187, | |
| "grad_norm": 0.03152284771203995, | |
| "learning_rate": 4.427175240939324e-05, | |
| "loss": 0.2384, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.5874987309215202, | |
| "grad_norm": 1.505940556526184, | |
| "learning_rate": 4.424460431654677e-05, | |
| "loss": 0.1345, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.5902060983451216, | |
| "grad_norm": 2.4096856117248535, | |
| "learning_rate": 4.4217456223700284e-05, | |
| "loss": 0.1783, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.5929134657687232, | |
| "grad_norm": 9.761833190917969, | |
| "learning_rate": 4.419030813085381e-05, | |
| "loss": 0.1601, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.5956208331923246, | |
| "grad_norm": 0.4443033039569855, | |
| "learning_rate": 4.416316003800733e-05, | |
| "loss": 0.1302, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.5956208331923246, | |
| "eval_loss": 0.08022167533636093, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.5736, | |
| "eval_samples_per_second": 4.364, | |
| "eval_steps_per_second": 1.091, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.5983282006159261, | |
| "grad_norm": 2.908780574798584, | |
| "learning_rate": 4.4136011945160856e-05, | |
| "loss": 0.2055, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.6010355680395275, | |
| "grad_norm": 48.880680084228516, | |
| "learning_rate": 4.410886385231437e-05, | |
| "loss": 0.2362, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.6037429354631291, | |
| "grad_norm": 5.180681228637695, | |
| "learning_rate": 4.4081715759467904e-05, | |
| "loss": 0.2119, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.6064503028867305, | |
| "grad_norm": 19.13489532470703, | |
| "learning_rate": 4.405456766662142e-05, | |
| "loss": 0.185, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.609157670310332, | |
| "grad_norm": 11.651802062988281, | |
| "learning_rate": 4.4027419573774945e-05, | |
| "loss": 0.0693, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.609157670310332, | |
| "eval_loss": 0.06786096096038818, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.8317, | |
| "eval_samples_per_second": 4.317, | |
| "eval_steps_per_second": 1.079, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.6118650377339334, | |
| "grad_norm": 0.008749375119805336, | |
| "learning_rate": 4.400027148092847e-05, | |
| "loss": 0.1525, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.614572405157535, | |
| "grad_norm": 1.9322203397750854, | |
| "learning_rate": 4.3973123388081986e-05, | |
| "loss": 0.1135, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.6172797725811364, | |
| "grad_norm": 7.96317195892334, | |
| "learning_rate": 4.3945975295235516e-05, | |
| "loss": 0.1073, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.6199871400047379, | |
| "grad_norm": 0.061754632741212845, | |
| "learning_rate": 4.3918827202389034e-05, | |
| "loss": 0.1121, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.6226945074283393, | |
| "grad_norm": 20.79821014404297, | |
| "learning_rate": 4.389167910954256e-05, | |
| "loss": 0.2335, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.6226945074283393, | |
| "eval_loss": 0.05240228772163391, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 113.8695, | |
| "eval_samples_per_second": 4.391, | |
| "eval_steps_per_second": 1.098, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.6254018748519409, | |
| "grad_norm": 0.7108599543571472, | |
| "learning_rate": 4.386453101669608e-05, | |
| "loss": 0.1733, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.6281092422755423, | |
| "grad_norm": 5.500114917755127, | |
| "learning_rate": 4.38373829238496e-05, | |
| "loss": 0.0439, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.6308166096991438, | |
| "grad_norm": 0.4140441119670868, | |
| "learning_rate": 4.381023483100312e-05, | |
| "loss": 0.1013, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.6335239771227452, | |
| "grad_norm": 27.898061752319336, | |
| "learning_rate": 4.3783086738156646e-05, | |
| "loss": 0.1334, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.6362313445463468, | |
| "grad_norm": 0.002663947641849518, | |
| "learning_rate": 4.375593864531017e-05, | |
| "loss": 0.0407, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.6362313445463468, | |
| "eval_loss": 0.08497549593448639, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0355, | |
| "eval_samples_per_second": 4.346, | |
| "eval_steps_per_second": 1.087, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.6389387119699482, | |
| "grad_norm": 0.06536103039979935, | |
| "learning_rate": 4.3728790552463694e-05, | |
| "loss": 0.064, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.6416460793935497, | |
| "grad_norm": 1.3059289455413818, | |
| "learning_rate": 4.370164245961722e-05, | |
| "loss": 0.2077, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.6443534468171511, | |
| "grad_norm": 0.029632871970534325, | |
| "learning_rate": 4.3674494366770735e-05, | |
| "loss": 0.1241, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.6470608142407527, | |
| "grad_norm": 0.18283845484256744, | |
| "learning_rate": 4.364734627392426e-05, | |
| "loss": 0.1071, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.6497681816643541, | |
| "grad_norm": 0.02363455854356289, | |
| "learning_rate": 4.362019818107778e-05, | |
| "loss": 0.0742, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.6497681816643541, | |
| "eval_loss": 0.08371000736951828, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0517, | |
| "eval_samples_per_second": 4.346, | |
| "eval_steps_per_second": 1.086, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.6524755490879556, | |
| "grad_norm": 15.434126853942871, | |
| "learning_rate": 4.35930500882313e-05, | |
| "loss": 0.3661, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.655182916511557, | |
| "grad_norm": 0.0013404142810031772, | |
| "learning_rate": 4.356590199538483e-05, | |
| "loss": 0.0634, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.6578902839351586, | |
| "grad_norm": 9.277833938598633, | |
| "learning_rate": 4.353875390253835e-05, | |
| "loss": 0.2201, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.66059765135876, | |
| "grad_norm": 9.396682739257812, | |
| "learning_rate": 4.351160580969187e-05, | |
| "loss": 0.2939, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.6633050187823615, | |
| "grad_norm": 13.4754638671875, | |
| "learning_rate": 4.3484457716845395e-05, | |
| "loss": 0.1117, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.6633050187823615, | |
| "eval_loss": 0.06177794933319092, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 116.1116, | |
| "eval_samples_per_second": 4.306, | |
| "eval_steps_per_second": 1.077, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.6660123862059629, | |
| "grad_norm": 0.6623813509941101, | |
| "learning_rate": 4.345730962399891e-05, | |
| "loss": 0.1209, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.6687197536295645, | |
| "grad_norm": 0.0009588833199813962, | |
| "learning_rate": 4.343016153115244e-05, | |
| "loss": 0.1701, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.6714271210531659, | |
| "grad_norm": 0.20336271822452545, | |
| "learning_rate": 4.340301343830596e-05, | |
| "loss": 0.28, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.6741344884767674, | |
| "grad_norm": 3.593808174133301, | |
| "learning_rate": 4.3375865345459484e-05, | |
| "loss": 0.1269, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.6768418559003688, | |
| "grad_norm": 0.030866390094161034, | |
| "learning_rate": 4.334871725261301e-05, | |
| "loss": 0.0896, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.6768418559003688, | |
| "eval_loss": 0.05724571645259857, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0334, | |
| "eval_samples_per_second": 4.347, | |
| "eval_steps_per_second": 1.087, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.6795492233239704, | |
| "grad_norm": 0.6822009086608887, | |
| "learning_rate": 4.332156915976653e-05, | |
| "loss": 0.1471, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.6822565907475718, | |
| "grad_norm": 0.5282585620880127, | |
| "learning_rate": 4.329442106692005e-05, | |
| "loss": 0.1834, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.6849639581711733, | |
| "grad_norm": 1.8356233835220337, | |
| "learning_rate": 4.326727297407357e-05, | |
| "loss": 0.1108, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.6876713255947747, | |
| "grad_norm": 22.017763137817383, | |
| "learning_rate": 4.32401248812271e-05, | |
| "loss": 0.1741, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.6903786930183763, | |
| "grad_norm": 22.150096893310547, | |
| "learning_rate": 4.3212976788380614e-05, | |
| "loss": 0.1629, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.6903786930183763, | |
| "eval_loss": 0.06724371761083603, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.1009, | |
| "eval_samples_per_second": 4.344, | |
| "eval_steps_per_second": 1.086, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.6930860604419777, | |
| "grad_norm": 1.9554486274719238, | |
| "learning_rate": 4.3185828695534144e-05, | |
| "loss": 0.2799, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.6957934278655792, | |
| "grad_norm": 2.001250743865967, | |
| "learning_rate": 4.315868060268766e-05, | |
| "loss": 0.1485, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.6985007952891806, | |
| "grad_norm": 13.129595756530762, | |
| "learning_rate": 4.3131532509841185e-05, | |
| "loss": 0.1245, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.7012081627127822, | |
| "grad_norm": 0.695175051689148, | |
| "learning_rate": 4.310438441699471e-05, | |
| "loss": 0.2419, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.7039155301363836, | |
| "grad_norm": 1.0608196258544922, | |
| "learning_rate": 4.3077236324148226e-05, | |
| "loss": 0.1922, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.7039155301363836, | |
| "eval_loss": 0.06558438390493393, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 121.1736, | |
| "eval_samples_per_second": 4.126, | |
| "eval_steps_per_second": 1.032, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.7066228975599851, | |
| "grad_norm": 17.941152572631836, | |
| "learning_rate": 4.305008823130176e-05, | |
| "loss": 0.2501, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 0.7093302649835865, | |
| "grad_norm": 0.3917510509490967, | |
| "learning_rate": 4.3022940138455274e-05, | |
| "loss": 0.0669, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 0.7120376324071881, | |
| "grad_norm": 2.205418825149536, | |
| "learning_rate": 4.29957920456088e-05, | |
| "loss": 0.1968, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 0.7147449998307895, | |
| "grad_norm": 1.7837224006652832, | |
| "learning_rate": 4.296864395276232e-05, | |
| "loss": 0.1016, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 0.717452367254391, | |
| "grad_norm": 0.15405257046222687, | |
| "learning_rate": 4.294149585991584e-05, | |
| "loss": 0.2557, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.717452367254391, | |
| "eval_loss": 0.07933684438467026, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.6706, | |
| "eval_samples_per_second": 4.36, | |
| "eval_steps_per_second": 1.09, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.7201597346779924, | |
| "grad_norm": 16.733394622802734, | |
| "learning_rate": 4.291434776706936e-05, | |
| "loss": 0.079, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 0.722867102101594, | |
| "grad_norm": 0.18252775073051453, | |
| "learning_rate": 4.288719967422289e-05, | |
| "loss": 0.1165, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 0.7255744695251954, | |
| "grad_norm": 0.0028702563140541315, | |
| "learning_rate": 4.286005158137641e-05, | |
| "loss": 0.2177, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 0.7282818369487969, | |
| "grad_norm": 0.00213778973557055, | |
| "learning_rate": 4.2832903488529935e-05, | |
| "loss": 0.0568, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 0.7309892043723983, | |
| "grad_norm": 0.09297258406877518, | |
| "learning_rate": 4.280575539568346e-05, | |
| "loss": 0.0597, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.7309892043723983, | |
| "eval_loss": 0.07073411345481873, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2895, | |
| "eval_samples_per_second": 4.337, | |
| "eval_steps_per_second": 1.084, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.7336965717959999, | |
| "grad_norm": 0.8263906836509705, | |
| "learning_rate": 4.2778607302836976e-05, | |
| "loss": 0.2522, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 0.7364039392196013, | |
| "grad_norm": 0.753628671169281, | |
| "learning_rate": 4.27514592099905e-05, | |
| "loss": 0.1264, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 0.7391113066432028, | |
| "grad_norm": 20.244163513183594, | |
| "learning_rate": 4.272431111714402e-05, | |
| "loss": 0.1499, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 0.7418186740668042, | |
| "grad_norm": 0.2985801100730896, | |
| "learning_rate": 4.269716302429754e-05, | |
| "loss": 0.0601, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 0.7445260414904058, | |
| "grad_norm": 9.927957534790039, | |
| "learning_rate": 4.267001493145107e-05, | |
| "loss": 0.2035, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.7445260414904058, | |
| "eval_loss": 0.082557812333107, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 116.3116, | |
| "eval_samples_per_second": 4.299, | |
| "eval_steps_per_second": 1.075, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.7472334089140072, | |
| "grad_norm": 1.3682934045791626, | |
| "learning_rate": 4.264286683860459e-05, | |
| "loss": 0.1115, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 0.7499407763376087, | |
| "grad_norm": 1.9332202672958374, | |
| "learning_rate": 4.261571874575811e-05, | |
| "loss": 0.2087, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 0.7526481437612101, | |
| "grad_norm": 1.8556005954742432, | |
| "learning_rate": 4.2588570652911636e-05, | |
| "loss": 0.2302, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 0.7553555111848117, | |
| "grad_norm": 0.06130724772810936, | |
| "learning_rate": 4.256142256006515e-05, | |
| "loss": 0.1079, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 0.7580628786084131, | |
| "grad_norm": 0.0025592364836484194, | |
| "learning_rate": 4.2534274467218684e-05, | |
| "loss": 0.2651, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.7580628786084131, | |
| "eval_loss": 0.07107679545879364, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.1747, | |
| "eval_samples_per_second": 4.341, | |
| "eval_steps_per_second": 1.085, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.7607702460320146, | |
| "grad_norm": 0.1967998445034027, | |
| "learning_rate": 4.25071263743722e-05, | |
| "loss": 0.1033, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 0.763477613455616, | |
| "grad_norm": 0.05981509014964104, | |
| "learning_rate": 4.2479978281525725e-05, | |
| "loss": 0.0682, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 0.7661849808792176, | |
| "grad_norm": 0.013984935358166695, | |
| "learning_rate": 4.245283018867925e-05, | |
| "loss": 0.1213, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 0.768892348302819, | |
| "grad_norm": 0.6655880808830261, | |
| "learning_rate": 4.242568209583277e-05, | |
| "loss": 0.3236, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 0.7715997157264205, | |
| "grad_norm": 3.138312816619873, | |
| "learning_rate": 4.239853400298629e-05, | |
| "loss": 0.0734, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.7715997157264205, | |
| "eval_loss": 0.09912095963954926, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2143, | |
| "eval_samples_per_second": 4.34, | |
| "eval_steps_per_second": 1.085, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.774307083150022, | |
| "grad_norm": 0.02605035901069641, | |
| "learning_rate": 4.2371385910139813e-05, | |
| "loss": 0.2264, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 0.7770144505736235, | |
| "grad_norm": 0.10319148004055023, | |
| "learning_rate": 4.234423781729334e-05, | |
| "loss": 0.0924, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 0.779721817997225, | |
| "grad_norm": 1.5601540803909302, | |
| "learning_rate": 4.231708972444686e-05, | |
| "loss": 0.2175, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 0.7824291854208264, | |
| "grad_norm": 1.938568115234375, | |
| "learning_rate": 4.2289941631600385e-05, | |
| "loss": 0.2074, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 0.7851365528444278, | |
| "grad_norm": 0.013515871949493885, | |
| "learning_rate": 4.22627935387539e-05, | |
| "loss": 0.1029, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.7851365528444278, | |
| "eval_loss": 0.07735220342874527, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.5718, | |
| "eval_samples_per_second": 4.326, | |
| "eval_steps_per_second": 1.082, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.7878439202680294, | |
| "grad_norm": 0.0078099388629198074, | |
| "learning_rate": 4.2235645445907426e-05, | |
| "loss": 0.2075, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 0.7905512876916309, | |
| "grad_norm": 0.2638203203678131, | |
| "learning_rate": 4.220849735306095e-05, | |
| "loss": 0.1487, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 0.7932586551152323, | |
| "grad_norm": 0.005258066579699516, | |
| "learning_rate": 4.218134926021447e-05, | |
| "loss": 0.1111, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 0.7959660225388338, | |
| "grad_norm": 0.22548271715641022, | |
| "learning_rate": 4.2154201167368e-05, | |
| "loss": 0.0303, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 0.7986733899624353, | |
| "grad_norm": 0.26001954078674316, | |
| "learning_rate": 4.2127053074521515e-05, | |
| "loss": 0.1185, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.7986733899624353, | |
| "eval_loss": 0.0766243264079094, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.8216, | |
| "eval_samples_per_second": 4.317, | |
| "eval_steps_per_second": 1.079, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.8013807573860368, | |
| "grad_norm": 2.299499988555908, | |
| "learning_rate": 4.209990498167504e-05, | |
| "loss": 0.4069, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 0.8040881248096382, | |
| "grad_norm": 0.04289079084992409, | |
| "learning_rate": 4.207275688882856e-05, | |
| "loss": 0.0866, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 0.8067954922332397, | |
| "grad_norm": 0.13593335449695587, | |
| "learning_rate": 4.2045608795982086e-05, | |
| "loss": 0.11, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 0.8095028596568412, | |
| "grad_norm": 0.02351505681872368, | |
| "learning_rate": 4.201846070313561e-05, | |
| "loss": 0.2065, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 0.8122102270804427, | |
| "grad_norm": 0.8734444975852966, | |
| "learning_rate": 4.199131261028913e-05, | |
| "loss": 0.1191, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.8122102270804427, | |
| "eval_loss": 0.0750957652926445, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.1813, | |
| "eval_samples_per_second": 4.341, | |
| "eval_steps_per_second": 1.085, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.8149175945040441, | |
| "grad_norm": 21.489639282226562, | |
| "learning_rate": 4.196416451744265e-05, | |
| "loss": 0.2108, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 0.8176249619276456, | |
| "grad_norm": 14.320918083190918, | |
| "learning_rate": 4.1937016424596175e-05, | |
| "loss": 0.0872, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 0.8203323293512471, | |
| "grad_norm": 0.14705882966518402, | |
| "learning_rate": 4.19098683317497e-05, | |
| "loss": 0.1089, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 0.8230396967748486, | |
| "grad_norm": 0.018130503594875336, | |
| "learning_rate": 4.1882720238903216e-05, | |
| "loss": 0.0833, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 0.82574706419845, | |
| "grad_norm": 0.07059603184461594, | |
| "learning_rate": 4.185557214605674e-05, | |
| "loss": 0.0807, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.82574706419845, | |
| "eval_loss": 0.07839541137218475, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0296, | |
| "eval_samples_per_second": 4.347, | |
| "eval_steps_per_second": 1.087, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.8284544316220515, | |
| "grad_norm": 2.696398973464966, | |
| "learning_rate": 4.1828424053210264e-05, | |
| "loss": 0.1347, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 0.831161799045653, | |
| "grad_norm": 0.7686747908592224, | |
| "learning_rate": 4.180127596036378e-05, | |
| "loss": 0.1131, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 0.8338691664692545, | |
| "grad_norm": 0.569398045539856, | |
| "learning_rate": 4.177412786751731e-05, | |
| "loss": 0.103, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 0.8365765338928559, | |
| "grad_norm": 17.672176361083984, | |
| "learning_rate": 4.174697977467083e-05, | |
| "loss": 0.1845, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 0.8392839013164574, | |
| "grad_norm": 3.657409906387329, | |
| "learning_rate": 4.171983168182436e-05, | |
| "loss": 0.1087, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.8392839013164574, | |
| "eval_loss": 0.09643953293561935, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 116.3381, | |
| "eval_samples_per_second": 4.298, | |
| "eval_steps_per_second": 1.074, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.8419912687400589, | |
| "grad_norm": 0.23462818562984467, | |
| "learning_rate": 4.169268358897788e-05, | |
| "loss": 0.1078, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 0.8446986361636604, | |
| "grad_norm": 6.32927131652832, | |
| "learning_rate": 4.1665535496131394e-05, | |
| "loss": 0.1914, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 0.8474060035872618, | |
| "grad_norm": 1.2845321893692017, | |
| "learning_rate": 4.1638387403284924e-05, | |
| "loss": 0.0184, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 0.8501133710108633, | |
| "grad_norm": 0.045799922198057175, | |
| "learning_rate": 4.161123931043844e-05, | |
| "loss": 0.0568, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 0.8528207384344648, | |
| "grad_norm": 0.01587361842393875, | |
| "learning_rate": 4.1584091217591965e-05, | |
| "loss": 0.1459, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.8528207384344648, | |
| "eval_loss": 0.0922488272190094, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.5988, | |
| "eval_samples_per_second": 4.325, | |
| "eval_steps_per_second": 1.081, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.8555281058580663, | |
| "grad_norm": 0.11717058718204498, | |
| "learning_rate": 4.155694312474549e-05, | |
| "loss": 0.1182, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 0.8582354732816677, | |
| "grad_norm": 1.3142938613891602, | |
| "learning_rate": 4.152979503189901e-05, | |
| "loss": 0.1918, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 0.8609428407052692, | |
| "grad_norm": 0.06714503467082977, | |
| "learning_rate": 4.150264693905254e-05, | |
| "loss": 0.0725, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 0.8636502081288707, | |
| "grad_norm": 12.575489044189453, | |
| "learning_rate": 4.1475498846206054e-05, | |
| "loss": 0.201, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 0.8663575755524722, | |
| "grad_norm": 0.9065766930580139, | |
| "learning_rate": 4.144835075335958e-05, | |
| "loss": 0.1276, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.8663575755524722, | |
| "eval_loss": 0.08115123957395554, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 117.9244, | |
| "eval_samples_per_second": 4.24, | |
| "eval_steps_per_second": 1.06, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.8690649429760736, | |
| "grad_norm": 0.10184822231531143, | |
| "learning_rate": 4.14212026605131e-05, | |
| "loss": 0.1413, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 0.8717723103996751, | |
| "grad_norm": 0.0267459936439991, | |
| "learning_rate": 4.1394054567666626e-05, | |
| "loss": 0.1017, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 0.8744796778232766, | |
| "grad_norm": 0.059938717633485794, | |
| "learning_rate": 4.136690647482014e-05, | |
| "loss": 0.0481, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 0.8771870452468781, | |
| "grad_norm": 24.361602783203125, | |
| "learning_rate": 4.133975838197367e-05, | |
| "loss": 0.1917, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 0.8798944126704795, | |
| "grad_norm": 0.9935716986656189, | |
| "learning_rate": 4.131261028912719e-05, | |
| "loss": 0.1553, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.8798944126704795, | |
| "eval_loss": 0.10655944794416428, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.5718, | |
| "eval_samples_per_second": 4.364, | |
| "eval_steps_per_second": 1.091, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.882601780094081, | |
| "grad_norm": 17.362058639526367, | |
| "learning_rate": 4.128546219628071e-05, | |
| "loss": 0.2953, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 0.8853091475176825, | |
| "grad_norm": 5.4117865562438965, | |
| "learning_rate": 4.125831410343424e-05, | |
| "loss": 0.1114, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 0.888016514941284, | |
| "grad_norm": 0.013206146657466888, | |
| "learning_rate": 4.1231166010587756e-05, | |
| "loss": 0.2183, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 0.8907238823648854, | |
| "grad_norm": 0.501082718372345, | |
| "learning_rate": 4.1204017917741286e-05, | |
| "loss": 0.0674, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 0.8934312497884869, | |
| "grad_norm": 0.08269740641117096, | |
| "learning_rate": 4.11768698248948e-05, | |
| "loss": 0.0716, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.8934312497884869, | |
| "eval_loss": 0.08250731229782104, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7008, | |
| "eval_samples_per_second": 4.359, | |
| "eval_steps_per_second": 1.09, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.8961386172120884, | |
| "grad_norm": 0.02271423488855362, | |
| "learning_rate": 4.114972173204833e-05, | |
| "loss": 0.1032, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 0.8988459846356899, | |
| "grad_norm": 25.167842864990234, | |
| "learning_rate": 4.112257363920185e-05, | |
| "loss": 0.0961, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 0.9015533520592913, | |
| "grad_norm": 0.049321677535772324, | |
| "learning_rate": 4.109542554635537e-05, | |
| "loss": 0.2755, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 0.9042607194828928, | |
| "grad_norm": 18.843111038208008, | |
| "learning_rate": 4.106827745350889e-05, | |
| "loss": 0.1792, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 0.9069680869064943, | |
| "grad_norm": 1.2584052085876465, | |
| "learning_rate": 4.1041129360662416e-05, | |
| "loss": 0.0299, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.9069680869064943, | |
| "eval_loss": 0.07415125519037247, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.619, | |
| "eval_samples_per_second": 4.325, | |
| "eval_steps_per_second": 1.081, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.9096754543300958, | |
| "grad_norm": 0.1885201781988144, | |
| "learning_rate": 4.101398126781594e-05, | |
| "loss": 0.1009, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 0.9123828217536972, | |
| "grad_norm": 0.022826386615633965, | |
| "learning_rate": 4.098683317496946e-05, | |
| "loss": 0.0213, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 0.9150901891772987, | |
| "grad_norm": 0.9626327157020569, | |
| "learning_rate": 4.095968508212298e-05, | |
| "loss": 0.2421, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 0.9177975566009002, | |
| "grad_norm": 0.1741826981306076, | |
| "learning_rate": 4.0932536989276505e-05, | |
| "loss": 0.0697, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 0.9205049240245017, | |
| "grad_norm": 12.775446891784668, | |
| "learning_rate": 4.090538889643003e-05, | |
| "loss": 0.0762, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.9205049240245017, | |
| "eval_loss": 0.07394778728485107, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0305, | |
| "eval_samples_per_second": 4.347, | |
| "eval_steps_per_second": 1.087, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.9232122914481031, | |
| "grad_norm": 17.907012939453125, | |
| "learning_rate": 4.087824080358355e-05, | |
| "loss": 0.2498, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 0.9259196588717046, | |
| "grad_norm": 0.01809736154973507, | |
| "learning_rate": 4.085109271073707e-05, | |
| "loss": 0.1663, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 0.9286270262953061, | |
| "grad_norm": 0.7653181552886963, | |
| "learning_rate": 4.08239446178906e-05, | |
| "loss": 0.2626, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 0.9313343937189076, | |
| "grad_norm": 0.08846959471702576, | |
| "learning_rate": 4.079679652504412e-05, | |
| "loss": 0.0397, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 0.934041761142509, | |
| "grad_norm": 0.0012274185428395867, | |
| "learning_rate": 4.076964843219764e-05, | |
| "loss": 0.069, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.934041761142509, | |
| "eval_loss": 0.05653444677591324, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7193, | |
| "eval_samples_per_second": 4.358, | |
| "eval_steps_per_second": 1.09, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.9367491285661105, | |
| "grad_norm": 0.34094715118408203, | |
| "learning_rate": 4.0742500339351165e-05, | |
| "loss": 0.0899, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 0.939456495989712, | |
| "grad_norm": 0.6107985377311707, | |
| "learning_rate": 4.071535224650468e-05, | |
| "loss": 0.193, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 0.9421638634133135, | |
| "grad_norm": 29.315858840942383, | |
| "learning_rate": 4.0688204153658206e-05, | |
| "loss": 0.2523, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 0.9448712308369149, | |
| "grad_norm": 0.030056612566113472, | |
| "learning_rate": 4.066105606081173e-05, | |
| "loss": 0.042, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 0.9475785982605164, | |
| "grad_norm": 3.3578696250915527, | |
| "learning_rate": 4.0633907967965254e-05, | |
| "loss": 0.1456, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.9475785982605164, | |
| "eval_loss": 0.07208700478076935, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.5581, | |
| "eval_samples_per_second": 4.327, | |
| "eval_steps_per_second": 1.082, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.9502859656841179, | |
| "grad_norm": 0.08940195292234421, | |
| "learning_rate": 4.060675987511878e-05, | |
| "loss": 0.1813, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 0.9529933331077194, | |
| "grad_norm": 0.05165726691484451, | |
| "learning_rate": 4.0579611782272295e-05, | |
| "loss": 0.0218, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 0.9557007005313208, | |
| "grad_norm": 4.932803153991699, | |
| "learning_rate": 4.055246368942582e-05, | |
| "loss": 0.1261, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 0.9584080679549223, | |
| "grad_norm": 0.024785397574305534, | |
| "learning_rate": 4.052531559657934e-05, | |
| "loss": 0.1756, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 0.9611154353785238, | |
| "grad_norm": 4.420279026031494, | |
| "learning_rate": 4.0498167503732866e-05, | |
| "loss": 0.1332, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.9611154353785238, | |
| "eval_loss": 0.06981143355369568, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.8045, | |
| "eval_samples_per_second": 4.355, | |
| "eval_steps_per_second": 1.089, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.9638228028021253, | |
| "grad_norm": 0.11680381745100021, | |
| "learning_rate": 4.0471019410886384e-05, | |
| "loss": 0.1575, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 0.9665301702257267, | |
| "grad_norm": 0.8033447265625, | |
| "learning_rate": 4.0443871318039914e-05, | |
| "loss": 0.1011, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 0.9692375376493282, | |
| "grad_norm": 0.29050976037979126, | |
| "learning_rate": 4.041672322519343e-05, | |
| "loss": 0.192, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 0.9719449050729297, | |
| "grad_norm": 0.09221057593822479, | |
| "learning_rate": 4.0389575132346955e-05, | |
| "loss": 0.1809, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 0.9746522724965312, | |
| "grad_norm": 0.0891430452466011, | |
| "learning_rate": 4.036242703950048e-05, | |
| "loss": 0.233, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.9746522724965312, | |
| "eval_loss": 0.04730875417590141, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.9057, | |
| "eval_samples_per_second": 4.351, | |
| "eval_steps_per_second": 1.088, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.9773596399201326, | |
| "grad_norm": 0.8720383644104004, | |
| "learning_rate": 4.0335278946653996e-05, | |
| "loss": 0.1189, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 0.9800670073437341, | |
| "grad_norm": 0.02585684508085251, | |
| "learning_rate": 4.030813085380753e-05, | |
| "loss": 0.1412, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 0.9827743747673356, | |
| "grad_norm": 0.012327141128480434, | |
| "learning_rate": 4.0280982760961044e-05, | |
| "loss": 0.1259, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 0.9854817421909371, | |
| "grad_norm": 15.601494789123535, | |
| "learning_rate": 4.025383466811457e-05, | |
| "loss": 0.2373, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 0.9881891096145385, | |
| "grad_norm": 2.5966286659240723, | |
| "learning_rate": 4.022668657526809e-05, | |
| "loss": 0.1293, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.9881891096145385, | |
| "eval_loss": 0.06829043477773666, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.6143, | |
| "eval_samples_per_second": 4.325, | |
| "eval_steps_per_second": 1.081, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.99089647703814, | |
| "grad_norm": 0.3134647607803345, | |
| "learning_rate": 4.019953848242161e-05, | |
| "loss": 0.0711, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 0.9936038444617415, | |
| "grad_norm": 32.37504959106445, | |
| "learning_rate": 4.017239038957513e-05, | |
| "loss": 0.1052, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 0.996311211885343, | |
| "grad_norm": 0.11635546386241913, | |
| "learning_rate": 4.0145242296728657e-05, | |
| "loss": 0.2258, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 0.9990185793089444, | |
| "grad_norm": 0.011782504618167877, | |
| "learning_rate": 4.011809420388218e-05, | |
| "loss": 0.0385, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 1.001725946732546, | |
| "grad_norm": 0.7141312956809998, | |
| "learning_rate": 4.0090946111035704e-05, | |
| "loss": 0.416, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.001725946732546, | |
| "eval_loss": 0.07393650710582733, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.9306, | |
| "eval_samples_per_second": 4.35, | |
| "eval_steps_per_second": 1.088, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.0044333141561474, | |
| "grad_norm": 0.0028216764330863953, | |
| "learning_rate": 4.006379801818922e-05, | |
| "loss": 0.103, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 1.0071406815797488, | |
| "grad_norm": 0.07318446040153503, | |
| "learning_rate": 4.0036649925342745e-05, | |
| "loss": 0.046, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 1.0098480490033503, | |
| "grad_norm": 0.06094740703701973, | |
| "learning_rate": 4.000950183249627e-05, | |
| "loss": 0.1924, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 1.012555416426952, | |
| "grad_norm": 0.0017843634122982621, | |
| "learning_rate": 3.998235373964979e-05, | |
| "loss": 0.1091, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 1.0152627838505532, | |
| "grad_norm": 0.830274760723114, | |
| "learning_rate": 3.995520564680331e-05, | |
| "loss": 0.0678, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.0152627838505532, | |
| "eval_loss": 0.08040212839841843, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.4828, | |
| "eval_samples_per_second": 4.33, | |
| "eval_steps_per_second": 1.082, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.0179701512741548, | |
| "grad_norm": 0.16321413218975067, | |
| "learning_rate": 3.992805755395684e-05, | |
| "loss": 0.0646, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 1.0206775186977564, | |
| "grad_norm": 0.03706735000014305, | |
| "learning_rate": 3.990090946111036e-05, | |
| "loss": 0.1927, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 1.0233848861213577, | |
| "grad_norm": 0.01869605854153633, | |
| "learning_rate": 3.987376136826388e-05, | |
| "loss": 0.0437, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 1.0260922535449593, | |
| "grad_norm": 5.068511486053467, | |
| "learning_rate": 3.9846613275417406e-05, | |
| "loss": 0.1424, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 1.0287996209685606, | |
| "grad_norm": 0.08981607854366302, | |
| "learning_rate": 3.981946518257092e-05, | |
| "loss": 0.2282, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.0287996209685606, | |
| "eval_loss": 0.07650502771139145, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 114.7492, | |
| "eval_samples_per_second": 4.357, | |
| "eval_steps_per_second": 1.089, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.0315069883921621, | |
| "grad_norm": 1.7685487270355225, | |
| "learning_rate": 3.9792317089724454e-05, | |
| "loss": 0.0231, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 1.0342143558157637, | |
| "grad_norm": 0.3977906107902527, | |
| "learning_rate": 3.976516899687797e-05, | |
| "loss": 0.0819, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 1.036921723239365, | |
| "grad_norm": 4.576267719268799, | |
| "learning_rate": 3.9738020904031495e-05, | |
| "loss": 0.1249, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 1.0396290906629666, | |
| "grad_norm": 22.835840225219727, | |
| "learning_rate": 3.971087281118502e-05, | |
| "loss": 0.1024, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 1.0423364580865682, | |
| "grad_norm": 2.8661038875579834, | |
| "learning_rate": 3.9683724718338535e-05, | |
| "loss": 0.1666, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.0423364580865682, | |
| "eval_loss": 0.08426863700151443, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.3533, | |
| "eval_samples_per_second": 4.335, | |
| "eval_steps_per_second": 1.084, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.0450438255101695, | |
| "grad_norm": 0.08293268084526062, | |
| "learning_rate": 3.965657662549206e-05, | |
| "loss": 0.0538, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 1.047751192933771, | |
| "grad_norm": 0.09539427608251572, | |
| "learning_rate": 3.962942853264558e-05, | |
| "loss": 0.2608, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 1.0504585603573724, | |
| "grad_norm": 8.918827056884766, | |
| "learning_rate": 3.960228043979911e-05, | |
| "loss": 0.2025, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 1.053165927780974, | |
| "grad_norm": 1.5636457204818726, | |
| "learning_rate": 3.9575132346952624e-05, | |
| "loss": 0.134, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 1.0558732952045755, | |
| "grad_norm": 1.9283950328826904, | |
| "learning_rate": 3.9547984254106155e-05, | |
| "loss": 0.1059, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.0558732952045755, | |
| "eval_loss": 0.09331218898296356, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 116.4095, | |
| "eval_samples_per_second": 4.295, | |
| "eval_steps_per_second": 1.074, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.0585806626281768, | |
| "grad_norm": 3.447388172149658, | |
| "learning_rate": 3.952083616125967e-05, | |
| "loss": 0.2881, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 1.0612880300517784, | |
| "grad_norm": 0.010725182481110096, | |
| "learning_rate": 3.9493688068413196e-05, | |
| "loss": 0.1261, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 1.06399539747538, | |
| "grad_norm": 0.0917816236615181, | |
| "learning_rate": 3.946653997556672e-05, | |
| "loss": 0.246, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 1.0667027648989813, | |
| "grad_norm": 1.0219537019729614, | |
| "learning_rate": 3.943939188272024e-05, | |
| "loss": 0.0927, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 1.0694101323225829, | |
| "grad_norm": 0.059237148612737656, | |
| "learning_rate": 3.941224378987377e-05, | |
| "loss": 0.1514, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.0694101323225829, | |
| "eval_loss": 0.08446906507015228, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.3793, | |
| "eval_samples_per_second": 4.334, | |
| "eval_steps_per_second": 1.083, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.0721174997461844, | |
| "grad_norm": 34.07741165161133, | |
| "learning_rate": 3.9385095697027285e-05, | |
| "loss": 0.0838, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 1.0748248671697858, | |
| "grad_norm": 0.4585834741592407, | |
| "learning_rate": 3.935794760418081e-05, | |
| "loss": 0.1584, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 1.0775322345933873, | |
| "grad_norm": 27.79207420349121, | |
| "learning_rate": 3.933079951133433e-05, | |
| "loss": 0.0999, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 1.0802396020169887, | |
| "grad_norm": 0.5480608344078064, | |
| "learning_rate": 3.930365141848785e-05, | |
| "loss": 0.0918, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 1.0829469694405902, | |
| "grad_norm": 17.493274688720703, | |
| "learning_rate": 3.9276503325641373e-05, | |
| "loss": 0.1292, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.0829469694405902, | |
| "eval_loss": 0.06774063408374786, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.224, | |
| "eval_samples_per_second": 4.339, | |
| "eval_steps_per_second": 1.085, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.0856543368641918, | |
| "grad_norm": 9.466602325439453, | |
| "learning_rate": 3.92493552327949e-05, | |
| "loss": 0.1294, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 1.088361704287793, | |
| "grad_norm": 0.09533428400754929, | |
| "learning_rate": 3.922220713994842e-05, | |
| "loss": 0.0924, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 1.0910690717113947, | |
| "grad_norm": 0.18998439610004425, | |
| "learning_rate": 3.9195059047101945e-05, | |
| "loss": 0.0505, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 1.093776439134996, | |
| "grad_norm": 0.0072032492607831955, | |
| "learning_rate": 3.916791095425547e-05, | |
| "loss": 0.0635, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 1.0964838065585976, | |
| "grad_norm": 0.26891425251960754, | |
| "learning_rate": 3.9140762861408986e-05, | |
| "loss": 0.0963, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.0964838065585976, | |
| "eval_loss": 0.07367356866598129, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.0833, | |
| "eval_samples_per_second": 4.345, | |
| "eval_steps_per_second": 1.086, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.0991911739821991, | |
| "grad_norm": 11.478422164916992, | |
| "learning_rate": 3.911361476856251e-05, | |
| "loss": 0.1339, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 1.1018985414058005, | |
| "grad_norm": 0.5796290636062622, | |
| "learning_rate": 3.9086466675716034e-05, | |
| "loss": 0.2033, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 1.104605908829402, | |
| "grad_norm": 0.02336627058684826, | |
| "learning_rate": 3.905931858286955e-05, | |
| "loss": 0.0583, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 1.1073132762530036, | |
| "grad_norm": 1.0814790725708008, | |
| "learning_rate": 3.903217049002308e-05, | |
| "loss": 0.124, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 1.110020643676605, | |
| "grad_norm": 0.42775824666023254, | |
| "learning_rate": 3.90050223971766e-05, | |
| "loss": 0.1272, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.110020643676605, | |
| "eval_loss": 0.06376690417528152, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 115.2277, | |
| "eval_samples_per_second": 4.339, | |
| "eval_steps_per_second": 1.085, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.1127280111002065, | |
| "grad_norm": 0.19235482811927795, | |
| "learning_rate": 3.897787430433012e-05, | |
| "loss": 0.1596, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 1.115435378523808, | |
| "grad_norm": 0.5556570291519165, | |
| "learning_rate": 3.8950726211483646e-05, | |
| "loss": 0.2195, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 1.1181427459474094, | |
| "grad_norm": 0.026464879512786865, | |
| "learning_rate": 3.8923578118637164e-05, | |
| "loss": 0.0748, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 1.120850113371011, | |
| "grad_norm": 7.5615034103393555, | |
| "learning_rate": 3.8896430025790694e-05, | |
| "loss": 0.047, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 1.1235574807946123, | |
| "grad_norm": 0.41822749376296997, | |
| "learning_rate": 3.886928193294421e-05, | |
| "loss": 0.0948, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.1235574807946123, | |
| "eval_loss": 0.060398317873477936, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 104.2866, | |
| "eval_samples_per_second": 4.794, | |
| "eval_steps_per_second": 1.199, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.1262648482182138, | |
| "grad_norm": 0.034160107374191284, | |
| "learning_rate": 3.8842133840097735e-05, | |
| "loss": 0.1445, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 1.1289722156418154, | |
| "grad_norm": 0.40724292397499084, | |
| "learning_rate": 3.881498574725126e-05, | |
| "loss": 0.0616, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 1.1316795830654167, | |
| "grad_norm": 0.1533338874578476, | |
| "learning_rate": 3.8787837654404776e-05, | |
| "loss": 0.0982, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 1.1343869504890183, | |
| "grad_norm": 0.003590109059587121, | |
| "learning_rate": 3.87606895615583e-05, | |
| "loss": 0.1408, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 1.1370943179126196, | |
| "grad_norm": 0.04526267945766449, | |
| "learning_rate": 3.8733541468711824e-05, | |
| "loss": 0.1955, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.1370943179126196, | |
| "eval_loss": 0.06311442703008652, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 104.2656, | |
| "eval_samples_per_second": 4.795, | |
| "eval_steps_per_second": 1.199, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.1398016853362212, | |
| "grad_norm": 3.631730318069458, | |
| "learning_rate": 3.870639337586535e-05, | |
| "loss": 0.1025, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 1.1425090527598227, | |
| "grad_norm": 0.006099845748394728, | |
| "learning_rate": 3.867924528301887e-05, | |
| "loss": 0.0781, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 1.145216420183424, | |
| "grad_norm": 0.0006692282040603459, | |
| "learning_rate": 3.8652097190172396e-05, | |
| "loss": 0.076, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 1.1479237876070256, | |
| "grad_norm": 0.023256322368979454, | |
| "learning_rate": 3.862494909732591e-05, | |
| "loss": 0.1354, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 1.1506311550306272, | |
| "grad_norm": 28.583494186401367, | |
| "learning_rate": 3.8597801004479437e-05, | |
| "loss": 0.1174, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.1506311550306272, | |
| "eval_loss": 0.07084568589925766, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 105.2689, | |
| "eval_samples_per_second": 4.75, | |
| "eval_steps_per_second": 1.187, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.1533385224542285, | |
| "grad_norm": 10.414568901062012, | |
| "learning_rate": 3.857065291163296e-05, | |
| "loss": 0.1554, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 1.15604588987783, | |
| "grad_norm": 8.933186531066895, | |
| "learning_rate": 3.854350481878648e-05, | |
| "loss": 0.0685, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 1.1587532573014316, | |
| "grad_norm": 2.6918463706970215, | |
| "learning_rate": 3.851635672594001e-05, | |
| "loss": 0.1234, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 1.161460624725033, | |
| "grad_norm": 1.297598123550415, | |
| "learning_rate": 3.8489208633093525e-05, | |
| "loss": 0.0685, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 1.1641679921486345, | |
| "grad_norm": 0.10164650529623032, | |
| "learning_rate": 3.846206054024705e-05, | |
| "loss": 0.0544, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.1641679921486345, | |
| "eval_loss": 0.07730449736118317, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 104.8653, | |
| "eval_samples_per_second": 4.768, | |
| "eval_steps_per_second": 1.192, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.1668753595722359, | |
| "grad_norm": 7.305291652679443, | |
| "learning_rate": 3.843491244740057e-05, | |
| "loss": 0.0794, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 1.1695827269958374, | |
| "grad_norm": 0.07961925864219666, | |
| "learning_rate": 3.840776435455409e-05, | |
| "loss": 0.044, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 1.172290094419439, | |
| "grad_norm": 0.023566043004393578, | |
| "learning_rate": 3.838061626170762e-05, | |
| "loss": 0.1417, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 1.1749974618430403, | |
| "grad_norm": 0.15135520696640015, | |
| "learning_rate": 3.835346816886114e-05, | |
| "loss": 0.1456, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 1.1777048292666419, | |
| "grad_norm": 0.07481474429368973, | |
| "learning_rate": 3.832632007601466e-05, | |
| "loss": 0.1329, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.1777048292666419, | |
| "eval_loss": 0.07221387326717377, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 104.6128, | |
| "eval_samples_per_second": 4.78, | |
| "eval_steps_per_second": 1.195, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.1804121966902432, | |
| "grad_norm": 1.943707823753357, | |
| "learning_rate": 3.8299171983168186e-05, | |
| "loss": 0.1643, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 1.1831195641138448, | |
| "grad_norm": 0.20479469001293182, | |
| "learning_rate": 3.827202389032171e-05, | |
| "loss": 0.1327, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 1.1858269315374463, | |
| "grad_norm": 1.116450548171997, | |
| "learning_rate": 3.824487579747523e-05, | |
| "loss": 0.1332, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 1.1885342989610477, | |
| "grad_norm": 1.1040481328964233, | |
| "learning_rate": 3.821772770462875e-05, | |
| "loss": 0.1719, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 1.1912416663846492, | |
| "grad_norm": 0.03437604010105133, | |
| "learning_rate": 3.8190579611782274e-05, | |
| "loss": 0.0317, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 1.1912416663846492, | |
| "eval_loss": 0.09452595561742783, | |
| "eval_model_preparation_time": 0.0086, | |
| "eval_runtime": 104.2972, | |
| "eval_samples_per_second": 4.794, | |
| "eval_steps_per_second": 1.198, | |
| "step": 8800 | |
| } | |
| ], | |
| "logging_steps": 20, | |
| "max_steps": 36935, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.119610830408488e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |