| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.039255671679837, | |
| "eval_steps": 500, | |
| "global_step": 8000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02549069589599796, | |
| "grad_norm": 3183.12646484375, | |
| "learning_rate": 2.5e-06, | |
| "loss": 227.1212, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.05098139179199592, | |
| "grad_norm": 363.3997802734375, | |
| "learning_rate": 5e-06, | |
| "loss": 88.4222, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.07647208768799388, | |
| "grad_norm": 263.5315246582031, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 67.8251, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.10196278358399184, | |
| "grad_norm": 328.4805908203125, | |
| "learning_rate": 1e-05, | |
| "loss": 63.1532, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.1274534794799898, | |
| "grad_norm": 496.2308044433594, | |
| "learning_rate": 9.995728791936505e-06, | |
| "loss": 60.2136, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.15294417537598776, | |
| "grad_norm": 405.20025634765625, | |
| "learning_rate": 9.98292246503335e-06, | |
| "loss": 59.0975, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.17843487127198573, | |
| "grad_norm": 451.58197021484375, | |
| "learning_rate": 9.961602898685225e-06, | |
| "loss": 55.2574, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.20392556716798368, | |
| "grad_norm": 127.22090148925781, | |
| "learning_rate": 9.931806517013612e-06, | |
| "loss": 56.7531, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.22941626306398163, | |
| "grad_norm": 431.8385925292969, | |
| "learning_rate": 9.893584226636773e-06, | |
| "loss": 57.5419, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.2549069589599796, | |
| "grad_norm": 249.0337677001953, | |
| "learning_rate": 9.847001329696653e-06, | |
| "loss": 57.2545, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.28039765485597756, | |
| "grad_norm": 726.9508056640625, | |
| "learning_rate": 9.792137412291265e-06, | |
| "loss": 56.1048, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.3058883507519755, | |
| "grad_norm": 811.3701171875, | |
| "learning_rate": 9.729086208503174e-06, | |
| "loss": 53.1675, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.33137904664797346, | |
| "grad_norm": 754.3577270507812, | |
| "learning_rate": 9.657955440256396e-06, | |
| "loss": 54.9404, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.35686974254397147, | |
| "grad_norm": 267.7255554199219, | |
| "learning_rate": 9.578866633275289e-06, | |
| "loss": 55.7077, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.3823604384399694, | |
| "grad_norm": 244.13253784179688, | |
| "learning_rate": 9.491954909459895e-06, | |
| "loss": 56.8997, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.40785113433596737, | |
| "grad_norm": 302.5712890625, | |
| "learning_rate": 9.397368756032445e-06, | |
| "loss": 50.8917, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.4333418302319653, | |
| "grad_norm": 660.6898193359375, | |
| "learning_rate": 9.295269771849426e-06, | |
| "loss": 49.8636, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.45883252612796327, | |
| "grad_norm": 235.12942504882812, | |
| "learning_rate": 9.185832391312644e-06, | |
| "loss": 53.0454, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.4843232220239613, | |
| "grad_norm": 231.6936492919922, | |
| "learning_rate": 9.069243586350976e-06, | |
| "loss": 56.1398, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.5098139179199592, | |
| "grad_norm": 460.98486328125, | |
| "learning_rate": 8.94570254698197e-06, | |
| "loss": 52.8867, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.5353046138159572, | |
| "grad_norm": 567.8645629882812, | |
| "learning_rate": 8.815420340999034e-06, | |
| "loss": 56.6194, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.5607953097119551, | |
| "grad_norm": 158.99224853515625, | |
| "learning_rate": 8.67861955336566e-06, | |
| "loss": 50.311, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.5862860056079531, | |
| "grad_norm": 463.69635009765625, | |
| "learning_rate": 8.535533905932739e-06, | |
| "loss": 48.126, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.611776701503951, | |
| "grad_norm": 346.60333251953125, | |
| "learning_rate": 8.386407858128707e-06, | |
| "loss": 51.602, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.637267397399949, | |
| "grad_norm": 430.1127624511719, | |
| "learning_rate": 8.231496189304704e-06, | |
| "loss": 50.3868, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.6627580932959469, | |
| "grad_norm": 430.7168884277344, | |
| "learning_rate": 8.071063563448341e-06, | |
| "loss": 49.5458, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.688248789191945, | |
| "grad_norm": 409.49114990234375, | |
| "learning_rate": 7.905384077009693e-06, | |
| "loss": 51.5311, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.7137394850879429, | |
| "grad_norm": 443.4747619628906, | |
| "learning_rate": 7.734740790612137e-06, | |
| "loss": 50.3823, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.7392301809839409, | |
| "grad_norm": 459.6910705566406, | |
| "learning_rate": 7.559425245448006e-06, | |
| "loss": 46.766, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.7647208768799388, | |
| "grad_norm": 579.1948852539062, | |
| "learning_rate": 7.379736965185369e-06, | |
| "loss": 51.7392, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.7902115727759368, | |
| "grad_norm": 329.1972961425781, | |
| "learning_rate": 7.195982944236853e-06, | |
| "loss": 51.0259, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.8157022686719347, | |
| "grad_norm": 488.2768859863281, | |
| "learning_rate": 7.008477123264849e-06, | |
| "loss": 53.4051, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.8411929645679327, | |
| "grad_norm": 463.906494140625, | |
| "learning_rate": 6.817539852819149e-06, | |
| "loss": 43.2941, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.8666836604639306, | |
| "grad_norm": 263.90185546875, | |
| "learning_rate": 6.6234973460234184e-06, | |
| "loss": 50.0815, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.8921743563599286, | |
| "grad_norm": 186.7496337890625, | |
| "learning_rate": 6.426681121245527e-06, | |
| "loss": 49.7881, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.9176650522559265, | |
| "grad_norm": 398.3317565917969, | |
| "learning_rate": 6.227427435703997e-06, | |
| "loss": 50.899, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.9431557481519246, | |
| "grad_norm": 238.25408935546875, | |
| "learning_rate": 6.026076710978172e-06, | |
| "loss": 44.1347, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.9686464440479226, | |
| "grad_norm": 567.3057861328125, | |
| "learning_rate": 5.82297295140367e-06, | |
| "loss": 50.0442, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.9941371399439205, | |
| "grad_norm": 226.54409790039062, | |
| "learning_rate": 5.61846315634674e-06, | |
| "loss": 41.8183, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 2458.4501953125, | |
| "eval_runtime": 28.4173, | |
| "eval_samples_per_second": 122.707, | |
| "eval_steps_per_second": 15.343, | |
| "step": 3923 | |
| }, | |
| { | |
| "epoch": 1.0196278358399185, | |
| "grad_norm": 698.6807861328125, | |
| "learning_rate": 5.412896727361663e-06, | |
| "loss": 44.033, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.0451185317359164, | |
| "grad_norm": 525.1095581054688, | |
| "learning_rate": 5.206624871244066e-06, | |
| "loss": 44.6573, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.0706092276319144, | |
| "grad_norm": 500.8614501953125, | |
| "learning_rate": 5e-06, | |
| "loss": 42.8117, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.0960999235279123, | |
| "grad_norm": 487.75689697265625, | |
| "learning_rate": 4.793375128755934e-06, | |
| "loss": 43.7467, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.1215906194239103, | |
| "grad_norm": 385.0135192871094, | |
| "learning_rate": 4.587103272638339e-06, | |
| "loss": 43.3253, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.1470813153199082, | |
| "grad_norm": 796.0384521484375, | |
| "learning_rate": 4.381536843653262e-06, | |
| "loss": 44.0489, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.1725720112159062, | |
| "grad_norm": 428.2261962890625, | |
| "learning_rate": 4.17702704859633e-06, | |
| "loss": 48.0535, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.198062707111904, | |
| "grad_norm": 385.2317199707031, | |
| "learning_rate": 3.973923289021829e-06, | |
| "loss": 43.4284, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.223553403007902, | |
| "grad_norm": 159.48098754882812, | |
| "learning_rate": 3.7725725642960047e-06, | |
| "loss": 40.4862, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.2490440989039, | |
| "grad_norm": 200.54164123535156, | |
| "learning_rate": 3.573318878754475e-06, | |
| "loss": 44.7052, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.274534794799898, | |
| "grad_norm": 301.8349304199219, | |
| "learning_rate": 3.3765026539765832e-06, | |
| "loss": 41.2848, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.3000254906958961, | |
| "grad_norm": 335.5228271484375, | |
| "learning_rate": 3.1824601471808504e-06, | |
| "loss": 43.5137, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.3255161865918939, | |
| "grad_norm": 94.21492767333984, | |
| "learning_rate": 2.991522876735154e-06, | |
| "loss": 41.9114, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.351006882487892, | |
| "grad_norm": 352.31048583984375, | |
| "learning_rate": 2.804017055763149e-06, | |
| "loss": 41.0477, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.3764975783838898, | |
| "grad_norm": 228.1314697265625, | |
| "learning_rate": 2.6202630348146323e-06, | |
| "loss": 38.6621, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.401988274279888, | |
| "grad_norm": 170.14947509765625, | |
| "learning_rate": 2.4405747545519966e-06, | |
| "loss": 40.0131, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.4274789701758859, | |
| "grad_norm": 166.94281005859375, | |
| "learning_rate": 2.265259209387867e-06, | |
| "loss": 42.6173, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.4529696660718838, | |
| "grad_norm": 255.24964904785156, | |
| "learning_rate": 2.094615922990309e-06, | |
| "loss": 41.4291, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.4784603619678818, | |
| "grad_norm": 296.5071716308594, | |
| "learning_rate": 1.928936436551661e-06, | |
| "loss": 39.6349, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.5039510578638797, | |
| "grad_norm": 227.43650817871094, | |
| "learning_rate": 1.7685038106952952e-06, | |
| "loss": 40.9158, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.5294417537598777, | |
| "grad_norm": 280.7147216796875, | |
| "learning_rate": 1.6135921418712959e-06, | |
| "loss": 45.4827, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.5549324496558756, | |
| "grad_norm": 166.27865600585938, | |
| "learning_rate": 1.4644660940672628e-06, | |
| "loss": 40.6627, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.5804231455518736, | |
| "grad_norm": 389.299560546875, | |
| "learning_rate": 1.321380446634342e-06, | |
| "loss": 38.7581, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.6059138414478715, | |
| "grad_norm": 170.0509033203125, | |
| "learning_rate": 1.1845796590009684e-06, | |
| "loss": 40.3946, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.6314045373438695, | |
| "grad_norm": 287.59423828125, | |
| "learning_rate": 1.0542974530180327e-06, | |
| "loss": 43.811, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 1.6568952332398674, | |
| "grad_norm": 150.49293518066406, | |
| "learning_rate": 9.307564136490255e-07, | |
| "loss": 40.4753, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.6823859291358654, | |
| "grad_norm": 264.41253662109375, | |
| "learning_rate": 8.141676086873574e-07, | |
| "loss": 43.6258, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.7078766250318633, | |
| "grad_norm": 157.13479614257812, | |
| "learning_rate": 7.047302281505735e-07, | |
| "loss": 43.768, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 1.7333673209278613, | |
| "grad_norm": 206.51609802246094, | |
| "learning_rate": 6.026312439675553e-07, | |
| "loss": 39.8553, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 1.7588580168238592, | |
| "grad_norm": 415.54351806640625, | |
| "learning_rate": 5.080450905401057e-07, | |
| "loss": 37.9228, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.7843487127198574, | |
| "grad_norm": 189.60276794433594, | |
| "learning_rate": 4.211333667247125e-07, | |
| "loss": 41.446, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.8098394086158551, | |
| "grad_norm": 612.336181640625, | |
| "learning_rate": 3.420445597436056e-07, | |
| "loss": 39.8801, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 1.8353301045118533, | |
| "grad_norm": 338.9154052734375, | |
| "learning_rate": 2.7091379149682683e-07, | |
| "loss": 41.9962, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.860820800407851, | |
| "grad_norm": 207.06942749023438, | |
| "learning_rate": 2.0786258770873647e-07, | |
| "loss": 40.2771, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 1.8863114963038492, | |
| "grad_norm": 585.3238525390625, | |
| "learning_rate": 1.5299867030334815e-07, | |
| "loss": 37.347, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.911802192199847, | |
| "grad_norm": 224.99607849121094, | |
| "learning_rate": 1.0641577336322761e-07, | |
| "loss": 45.2395, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.937292888095845, | |
| "grad_norm": 304.6292724609375, | |
| "learning_rate": 6.819348298638839e-08, | |
| "loss": 51.0356, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.9627835839918428, | |
| "grad_norm": 1007.5321044921875, | |
| "learning_rate": 3.839710131477492e-08, | |
| "loss": 44.2402, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.988274279887841, | |
| "grad_norm": 436.2712097167969, | |
| "learning_rate": 1.7077534966650767e-08, | |
| "loss": 41.5961, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 2133.657470703125, | |
| "eval_runtime": 28.5718, | |
| "eval_samples_per_second": 122.043, | |
| "eval_steps_per_second": 15.26, | |
| "step": 7846 | |
| }, | |
| { | |
| "epoch": 2.0137649757838387, | |
| "grad_norm": 354.5096740722656, | |
| "learning_rate": 4.2712080634949024e-09, | |
| "loss": 41.8052, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 2.039255671679837, | |
| "grad_norm": 220.47647094726562, | |
| "learning_rate": 0.0, | |
| "loss": 39.9241, | |
| "step": 8000 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 8000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 2000, | |
| "total_flos": 0.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |