| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9987515605493134, | |
| "eval_steps": 500, | |
| "global_step": 400, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0024968789013732834, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5487, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.004993757802746567, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5726, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.00749063670411985, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.4761, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.009987515605493134, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5582, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.012484394506866416, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5577, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0149812734082397, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5281, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.017478152309612985, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5385, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.019975031210986267, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.4239, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.02247191011235955, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.558, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.024968789013732832, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.4404, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02746566791510612, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5149, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.0299625468164794, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5407, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.03245942571785269, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5754, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.03495630461922597, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5222, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03745318352059925, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5261, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.039950062421972535, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5199, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.04244694132334582, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5327, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.0449438202247191, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.4925, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04744069912609238, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5452, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.049937578027465665, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.4986, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.052434456928838954, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.5148, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.05493133583021224, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.6131, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.05742821473158552, | |
| "grad_norm": NaN, | |
| "learning_rate": 5e-05, | |
| "loss": 2.534, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.0599250936329588, | |
| "grad_norm": 3.498404026031494, | |
| "learning_rate": 4.9875000000000006e-05, | |
| "loss": 2.5259, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.062421972534332085, | |
| "grad_norm": 3.7281525135040283, | |
| "learning_rate": 4.975e-05, | |
| "loss": 2.5641, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.06491885143570537, | |
| "grad_norm": 3.4878628253936768, | |
| "learning_rate": 4.962500000000001e-05, | |
| "loss": 2.5001, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.06741573033707865, | |
| "grad_norm": 3.498518228530884, | |
| "learning_rate": 4.9500000000000004e-05, | |
| "loss": 2.4923, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.06991260923845194, | |
| "grad_norm": 3.5697684288024902, | |
| "learning_rate": 4.937500000000001e-05, | |
| "loss": 2.4743, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.07240948813982521, | |
| "grad_norm": 3.58793044090271, | |
| "learning_rate": 4.9250000000000004e-05, | |
| "loss": 2.4403, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.0749063670411985, | |
| "grad_norm": 3.395378828048706, | |
| "learning_rate": 4.9125e-05, | |
| "loss": 2.4839, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07740324594257178, | |
| "grad_norm": 3.4821157455444336, | |
| "learning_rate": 4.9e-05, | |
| "loss": 2.4865, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.07990012484394507, | |
| "grad_norm": 3.5381717681884766, | |
| "learning_rate": 4.8875e-05, | |
| "loss": 2.3919, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.08239700374531835, | |
| "grad_norm": 3.7879738807678223, | |
| "learning_rate": 4.875e-05, | |
| "loss": 2.5253, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.08489388264669163, | |
| "grad_norm": 3.9536428451538086, | |
| "learning_rate": 4.8625e-05, | |
| "loss": 2.557, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.08739076154806492, | |
| "grad_norm": 3.5281002521514893, | |
| "learning_rate": 4.85e-05, | |
| "loss": 2.419, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.0898876404494382, | |
| "grad_norm": 3.8065526485443115, | |
| "learning_rate": 4.8375000000000004e-05, | |
| "loss": 2.4839, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.09238451935081149, | |
| "grad_norm": 3.48492431640625, | |
| "learning_rate": 4.825e-05, | |
| "loss": 2.3443, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.09488139825218476, | |
| "grad_norm": 3.5769503116607666, | |
| "learning_rate": 4.8125000000000004e-05, | |
| "loss": 2.5158, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.09737827715355805, | |
| "grad_norm": 3.5777034759521484, | |
| "learning_rate": 4.8e-05, | |
| "loss": 2.403, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.09987515605493133, | |
| "grad_norm": 3.471339225769043, | |
| "learning_rate": 4.7875000000000005e-05, | |
| "loss": 2.3998, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.10237203495630462, | |
| "grad_norm": 3.6594278812408447, | |
| "learning_rate": 4.775e-05, | |
| "loss": 2.4084, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.10486891385767791, | |
| "grad_norm": 3.5339417457580566, | |
| "learning_rate": 4.7625000000000006e-05, | |
| "loss": 2.4, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.10736579275905118, | |
| "grad_norm": 3.4090983867645264, | |
| "learning_rate": 4.75e-05, | |
| "loss": 2.367, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.10986267166042447, | |
| "grad_norm": 3.769968032836914, | |
| "learning_rate": 4.7375e-05, | |
| "loss": 2.3624, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.11235955056179775, | |
| "grad_norm": 3.788815975189209, | |
| "learning_rate": 4.7249999999999997e-05, | |
| "loss": 2.4006, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.11485642946317104, | |
| "grad_norm": 3.7179384231567383, | |
| "learning_rate": 4.7125e-05, | |
| "loss": 2.376, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.11735330836454431, | |
| "grad_norm": 3.4289679527282715, | |
| "learning_rate": 4.7e-05, | |
| "loss": 2.3354, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.1198501872659176, | |
| "grad_norm": 3.5412650108337402, | |
| "learning_rate": 4.6875e-05, | |
| "loss": 2.3317, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.12234706616729088, | |
| "grad_norm": 3.5559499263763428, | |
| "learning_rate": 4.6750000000000005e-05, | |
| "loss": 2.3663, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.12484394506866417, | |
| "grad_norm": 3.6865787506103516, | |
| "learning_rate": 4.6625e-05, | |
| "loss": 2.4062, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.12734082397003746, | |
| "grad_norm": 3.417897939682007, | |
| "learning_rate": 4.6500000000000005e-05, | |
| "loss": 2.3314, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.12983770287141075, | |
| "grad_norm": 3.5215134620666504, | |
| "learning_rate": 4.6375e-05, | |
| "loss": 2.367, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.132334581772784, | |
| "grad_norm": 3.478131055831909, | |
| "learning_rate": 4.6250000000000006e-05, | |
| "loss": 2.285, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.1348314606741573, | |
| "grad_norm": 3.8263564109802246, | |
| "learning_rate": 4.6125e-05, | |
| "loss": 2.402, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.1373283395755306, | |
| "grad_norm": 3.4977402687072754, | |
| "learning_rate": 4.600000000000001e-05, | |
| "loss": 2.317, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.13982521847690388, | |
| "grad_norm": 3.5797126293182373, | |
| "learning_rate": 4.5875000000000004e-05, | |
| "loss": 2.3097, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.14232209737827714, | |
| "grad_norm": 3.6681716442108154, | |
| "learning_rate": 4.575e-05, | |
| "loss": 2.3766, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.14481897627965043, | |
| "grad_norm": 3.6064581871032715, | |
| "learning_rate": 4.5625e-05, | |
| "loss": 2.2548, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.14731585518102372, | |
| "grad_norm": 3.680893659591675, | |
| "learning_rate": 4.55e-05, | |
| "loss": 2.4249, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.149812734082397, | |
| "grad_norm": 3.5588841438293457, | |
| "learning_rate": 4.5375e-05, | |
| "loss": 2.3109, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1523096129837703, | |
| "grad_norm": 3.4901130199432373, | |
| "learning_rate": 4.525e-05, | |
| "loss": 2.2421, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.15480649188514356, | |
| "grad_norm": 3.5400235652923584, | |
| "learning_rate": 4.5125e-05, | |
| "loss": 2.3226, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.15730337078651685, | |
| "grad_norm": 3.578420877456665, | |
| "learning_rate": 4.5e-05, | |
| "loss": 2.2904, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.15980024968789014, | |
| "grad_norm": 3.583136558532715, | |
| "learning_rate": 4.4875e-05, | |
| "loss": 2.3255, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.16229712858926343, | |
| "grad_norm": 3.35066819190979, | |
| "learning_rate": 4.4750000000000004e-05, | |
| "loss": 2.2183, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.1647940074906367, | |
| "grad_norm": 3.4735934734344482, | |
| "learning_rate": 4.4625e-05, | |
| "loss": 2.2754, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.16729088639200998, | |
| "grad_norm": 3.3307180404663086, | |
| "learning_rate": 4.4500000000000004e-05, | |
| "loss": 2.2964, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.16978776529338327, | |
| "grad_norm": 3.516148567199707, | |
| "learning_rate": 4.4375e-05, | |
| "loss": 2.3107, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.17228464419475656, | |
| "grad_norm": 3.4239909648895264, | |
| "learning_rate": 4.4250000000000005e-05, | |
| "loss": 2.2945, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.17478152309612985, | |
| "grad_norm": 3.507624626159668, | |
| "learning_rate": 4.4125e-05, | |
| "loss": 2.2859, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1772784019975031, | |
| "grad_norm": 3.6821298599243164, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 2.3248, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.1797752808988764, | |
| "grad_norm": 3.3753108978271484, | |
| "learning_rate": 4.3875e-05, | |
| "loss": 2.2001, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.1822721598002497, | |
| "grad_norm": 3.4027931690216064, | |
| "learning_rate": 4.375e-05, | |
| "loss": 2.2668, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.18476903870162298, | |
| "grad_norm": 3.524763584136963, | |
| "learning_rate": 4.3625e-05, | |
| "loss": 2.2686, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.18726591760299627, | |
| "grad_norm": 3.5769248008728027, | |
| "learning_rate": 4.35e-05, | |
| "loss": 2.2757, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.18976279650436953, | |
| "grad_norm": 3.5016684532165527, | |
| "learning_rate": 4.3375000000000004e-05, | |
| "loss": 2.2218, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.19225967540574282, | |
| "grad_norm": 3.45440411567688, | |
| "learning_rate": 4.325e-05, | |
| "loss": 2.2405, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.1947565543071161, | |
| "grad_norm": 3.5838842391967773, | |
| "learning_rate": 4.3125000000000005e-05, | |
| "loss": 2.2426, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.1972534332084894, | |
| "grad_norm": 3.716230869293213, | |
| "learning_rate": 4.3e-05, | |
| "loss": 2.2819, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.19975031210986266, | |
| "grad_norm": 3.258978843688965, | |
| "learning_rate": 4.2875000000000005e-05, | |
| "loss": 2.2049, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.20224719101123595, | |
| "grad_norm": 3.575076103210449, | |
| "learning_rate": 4.275e-05, | |
| "loss": 2.2412, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.20474406991260924, | |
| "grad_norm": 3.5371599197387695, | |
| "learning_rate": 4.2625000000000006e-05, | |
| "loss": 2.2206, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.20724094881398253, | |
| "grad_norm": 3.6981263160705566, | |
| "learning_rate": 4.25e-05, | |
| "loss": 2.266, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.20973782771535582, | |
| "grad_norm": 3.451237678527832, | |
| "learning_rate": 4.237500000000001e-05, | |
| "loss": 2.2564, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.21223470661672908, | |
| "grad_norm": 3.4362707138061523, | |
| "learning_rate": 4.2250000000000004e-05, | |
| "loss": 2.1987, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.21473158551810237, | |
| "grad_norm": 3.45649790763855, | |
| "learning_rate": 4.2125e-05, | |
| "loss": 2.2001, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.21722846441947566, | |
| "grad_norm": 3.476231575012207, | |
| "learning_rate": 4.2e-05, | |
| "loss": 2.1378, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.21972534332084895, | |
| "grad_norm": 3.5050230026245117, | |
| "learning_rate": 4.1875e-05, | |
| "loss": 2.1478, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.2222222222222222, | |
| "grad_norm": 3.455080509185791, | |
| "learning_rate": 4.175e-05, | |
| "loss": 2.2395, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.2247191011235955, | |
| "grad_norm": 3.500718832015991, | |
| "learning_rate": 4.1625e-05, | |
| "loss": 2.1794, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2272159800249688, | |
| "grad_norm": 3.53946852684021, | |
| "learning_rate": 4.15e-05, | |
| "loss": 2.2581, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.22971285892634208, | |
| "grad_norm": 3.506479024887085, | |
| "learning_rate": 4.1375e-05, | |
| "loss": 2.2028, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.23220973782771537, | |
| "grad_norm": 3.4707953929901123, | |
| "learning_rate": 4.125e-05, | |
| "loss": 2.1435, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.23470661672908863, | |
| "grad_norm": 3.4014856815338135, | |
| "learning_rate": 4.1125000000000004e-05, | |
| "loss": 2.2146, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.23720349563046192, | |
| "grad_norm": 3.42364764213562, | |
| "learning_rate": 4.1e-05, | |
| "loss": 2.1824, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.2397003745318352, | |
| "grad_norm": 3.480703115463257, | |
| "learning_rate": 4.0875000000000004e-05, | |
| "loss": 2.2208, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.2421972534332085, | |
| "grad_norm": 3.6021103858947754, | |
| "learning_rate": 4.075e-05, | |
| "loss": 2.1939, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.24469413233458176, | |
| "grad_norm": 3.673717498779297, | |
| "learning_rate": 4.0625000000000005e-05, | |
| "loss": 2.234, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.24719101123595505, | |
| "grad_norm": 3.511613368988037, | |
| "learning_rate": 4.05e-05, | |
| "loss": 2.2176, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.24968789013732834, | |
| "grad_norm": 3.5998849868774414, | |
| "learning_rate": 4.0375e-05, | |
| "loss": 2.2259, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.25218476903870163, | |
| "grad_norm": 3.4577770233154297, | |
| "learning_rate": 4.025e-05, | |
| "loss": 2.2238, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.2546816479400749, | |
| "grad_norm": 3.612593173980713, | |
| "learning_rate": 4.0125e-05, | |
| "loss": 2.161, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.2571785268414482, | |
| "grad_norm": 3.477313280105591, | |
| "learning_rate": 4e-05, | |
| "loss": 2.1839, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.2596754057428215, | |
| "grad_norm": 3.553819417953491, | |
| "learning_rate": 3.9875e-05, | |
| "loss": 2.2249, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.26217228464419473, | |
| "grad_norm": 3.412799596786499, | |
| "learning_rate": 3.9750000000000004e-05, | |
| "loss": 2.1834, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.264669163545568, | |
| "grad_norm": 3.4956414699554443, | |
| "learning_rate": 3.9625e-05, | |
| "loss": 2.1456, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.2671660424469413, | |
| "grad_norm": 3.6447088718414307, | |
| "learning_rate": 3.9500000000000005e-05, | |
| "loss": 2.166, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.2696629213483146, | |
| "grad_norm": 3.6700491905212402, | |
| "learning_rate": 3.9375e-05, | |
| "loss": 2.2626, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.2721598002496879, | |
| "grad_norm": 3.570895195007324, | |
| "learning_rate": 3.9250000000000005e-05, | |
| "loss": 2.1324, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.2746566791510612, | |
| "grad_norm": 3.456331968307495, | |
| "learning_rate": 3.9125e-05, | |
| "loss": 2.1244, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.27715355805243447, | |
| "grad_norm": 3.4046988487243652, | |
| "learning_rate": 3.9000000000000006e-05, | |
| "loss": 2.1525, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.27965043695380776, | |
| "grad_norm": 3.423374652862549, | |
| "learning_rate": 3.8875e-05, | |
| "loss": 2.1114, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.28214731585518105, | |
| "grad_norm": 3.3426084518432617, | |
| "learning_rate": 3.875e-05, | |
| "loss": 2.1863, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2846441947565543, | |
| "grad_norm": 3.2898874282836914, | |
| "learning_rate": 3.8625e-05, | |
| "loss": 2.2121, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.28714107365792757, | |
| "grad_norm": 3.395939826965332, | |
| "learning_rate": 3.85e-05, | |
| "loss": 2.1401, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.28963795255930086, | |
| "grad_norm": 3.3747150897979736, | |
| "learning_rate": 3.8375e-05, | |
| "loss": 2.0959, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.29213483146067415, | |
| "grad_norm": 3.5609467029571533, | |
| "learning_rate": 3.825e-05, | |
| "loss": 2.1268, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.29463171036204744, | |
| "grad_norm": 3.6404829025268555, | |
| "learning_rate": 3.8125e-05, | |
| "loss": 2.184, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.29712858926342073, | |
| "grad_norm": 3.50461745262146, | |
| "learning_rate": 3.8e-05, | |
| "loss": 2.168, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.299625468164794, | |
| "grad_norm": 3.4648733139038086, | |
| "learning_rate": 3.7875e-05, | |
| "loss": 2.1526, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3021223470661673, | |
| "grad_norm": 3.696610450744629, | |
| "learning_rate": 3.775e-05, | |
| "loss": 2.204, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.3046192259675406, | |
| "grad_norm": 3.4781978130340576, | |
| "learning_rate": 3.7625e-05, | |
| "loss": 2.1915, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.30711610486891383, | |
| "grad_norm": 3.490053176879883, | |
| "learning_rate": 3.7500000000000003e-05, | |
| "loss": 2.1437, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.3096129837702871, | |
| "grad_norm": 3.452174186706543, | |
| "learning_rate": 3.737500000000001e-05, | |
| "loss": 2.1535, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.3121098626716604, | |
| "grad_norm": 3.5188889503479004, | |
| "learning_rate": 3.7250000000000004e-05, | |
| "loss": 2.1696, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.3146067415730337, | |
| "grad_norm": 3.452965021133423, | |
| "learning_rate": 3.7125e-05, | |
| "loss": 2.1077, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.317103620474407, | |
| "grad_norm": 3.4781739711761475, | |
| "learning_rate": 3.7e-05, | |
| "loss": 2.1442, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.3196004993757803, | |
| "grad_norm": 3.455982208251953, | |
| "learning_rate": 3.6875e-05, | |
| "loss": 2.2422, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.32209737827715357, | |
| "grad_norm": 3.735812187194824, | |
| "learning_rate": 3.675e-05, | |
| "loss": 2.211, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.32459425717852686, | |
| "grad_norm": 3.7122321128845215, | |
| "learning_rate": 3.6625e-05, | |
| "loss": 2.1534, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.32709113607990015, | |
| "grad_norm": 3.407907724380493, | |
| "learning_rate": 3.65e-05, | |
| "loss": 2.135, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.3295880149812734, | |
| "grad_norm": 3.610645055770874, | |
| "learning_rate": 3.6375e-05, | |
| "loss": 2.1437, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.33208489388264667, | |
| "grad_norm": 3.316847085952759, | |
| "learning_rate": 3.625e-05, | |
| "loss": 2.1594, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.33458177278401996, | |
| "grad_norm": 3.504835367202759, | |
| "learning_rate": 3.6125000000000004e-05, | |
| "loss": 2.1369, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.33707865168539325, | |
| "grad_norm": 3.4886667728424072, | |
| "learning_rate": 3.6e-05, | |
| "loss": 2.1854, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.33957553058676654, | |
| "grad_norm": 3.4116599559783936, | |
| "learning_rate": 3.5875000000000005e-05, | |
| "loss": 2.0741, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.34207240948813983, | |
| "grad_norm": 3.4653735160827637, | |
| "learning_rate": 3.575e-05, | |
| "loss": 2.1779, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.3445692883895131, | |
| "grad_norm": 3.496469497680664, | |
| "learning_rate": 3.5625000000000005e-05, | |
| "loss": 2.1337, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.3470661672908864, | |
| "grad_norm": 3.4037115573883057, | |
| "learning_rate": 3.55e-05, | |
| "loss": 2.1346, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.3495630461922597, | |
| "grad_norm": 3.6134746074676514, | |
| "learning_rate": 3.5375e-05, | |
| "loss": 2.212, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.352059925093633, | |
| "grad_norm": 3.444700002670288, | |
| "learning_rate": 3.525e-05, | |
| "loss": 2.1676, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.3545568039950062, | |
| "grad_norm": 3.404010534286499, | |
| "learning_rate": 3.5125e-05, | |
| "loss": 2.0873, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.3570536828963795, | |
| "grad_norm": 3.3055148124694824, | |
| "learning_rate": 3.5e-05, | |
| "loss": 2.1762, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.3595505617977528, | |
| "grad_norm": 3.5183520317077637, | |
| "learning_rate": 3.4875e-05, | |
| "loss": 2.0908, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.3620474406991261, | |
| "grad_norm": 3.661907911300659, | |
| "learning_rate": 3.475e-05, | |
| "loss": 2.0899, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.3645443196004994, | |
| "grad_norm": 3.574948787689209, | |
| "learning_rate": 3.4625e-05, | |
| "loss": 2.1531, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.36704119850187267, | |
| "grad_norm": 3.4745309352874756, | |
| "learning_rate": 3.45e-05, | |
| "loss": 2.1238, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.36953807740324596, | |
| "grad_norm": 3.6377320289611816, | |
| "learning_rate": 3.4375e-05, | |
| "loss": 2.0556, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.37203495630461925, | |
| "grad_norm": 3.3080575466156006, | |
| "learning_rate": 3.4250000000000006e-05, | |
| "loss": 2.1218, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.37453183520599254, | |
| "grad_norm": 3.2884228229522705, | |
| "learning_rate": 3.4125e-05, | |
| "loss": 2.0843, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.37702871410736577, | |
| "grad_norm": 3.4235734939575195, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 2.1042, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.37952559300873906, | |
| "grad_norm": 3.284189462661743, | |
| "learning_rate": 3.3875000000000003e-05, | |
| "loss": 2.0937, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.38202247191011235, | |
| "grad_norm": 3.311265468597412, | |
| "learning_rate": 3.375000000000001e-05, | |
| "loss": 2.0621, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.38451935081148564, | |
| "grad_norm": 3.5135247707366943, | |
| "learning_rate": 3.3625000000000004e-05, | |
| "loss": 2.1317, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.38701622971285893, | |
| "grad_norm": 3.428800582885742, | |
| "learning_rate": 3.35e-05, | |
| "loss": 2.1447, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.3895131086142322, | |
| "grad_norm": 3.4077048301696777, | |
| "learning_rate": 3.3375e-05, | |
| "loss": 2.1757, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.3920099875156055, | |
| "grad_norm": 3.572735071182251, | |
| "learning_rate": 3.325e-05, | |
| "loss": 2.0373, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.3945068664169788, | |
| "grad_norm": 3.465317487716675, | |
| "learning_rate": 3.3125e-05, | |
| "loss": 2.0783, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.3970037453183521, | |
| "grad_norm": 3.5976178646087646, | |
| "learning_rate": 3.3e-05, | |
| "loss": 2.2142, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.3995006242197253, | |
| "grad_norm": 3.576270341873169, | |
| "learning_rate": 3.2875e-05, | |
| "loss": 2.134, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.4019975031210986, | |
| "grad_norm": 3.536389112472534, | |
| "learning_rate": 3.275e-05, | |
| "loss": 2.1068, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.4044943820224719, | |
| "grad_norm": 3.45173716545105, | |
| "learning_rate": 3.2625e-05, | |
| "loss": 2.0449, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.4069912609238452, | |
| "grad_norm": 3.498258352279663, | |
| "learning_rate": 3.2500000000000004e-05, | |
| "loss": 2.1712, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.4094881398252185, | |
| "grad_norm": 3.2850704193115234, | |
| "learning_rate": 3.2375e-05, | |
| "loss": 2.0633, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.41198501872659177, | |
| "grad_norm": 3.3497467041015625, | |
| "learning_rate": 3.2250000000000005e-05, | |
| "loss": 2.1995, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.41448189762796506, | |
| "grad_norm": 3.4341518878936768, | |
| "learning_rate": 3.2125e-05, | |
| "loss": 2.1297, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.41697877652933835, | |
| "grad_norm": 3.3831026554107666, | |
| "learning_rate": 3.2000000000000005e-05, | |
| "loss": 2.198, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.41947565543071164, | |
| "grad_norm": 3.450352191925049, | |
| "learning_rate": 3.1875e-05, | |
| "loss": 2.092, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.42197253433208487, | |
| "grad_norm": 3.3830325603485107, | |
| "learning_rate": 3.175e-05, | |
| "loss": 2.1015, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.42446941323345816, | |
| "grad_norm": 3.548405885696411, | |
| "learning_rate": 3.1624999999999996e-05, | |
| "loss": 2.0894, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.42696629213483145, | |
| "grad_norm": 3.675614595413208, | |
| "learning_rate": 3.15e-05, | |
| "loss": 2.1759, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.42946317103620474, | |
| "grad_norm": 3.520315408706665, | |
| "learning_rate": 3.1375e-05, | |
| "loss": 2.1231, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.43196004993757803, | |
| "grad_norm": 3.604353666305542, | |
| "learning_rate": 3.125e-05, | |
| "loss": 2.048, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.4344569288389513, | |
| "grad_norm": 3.609004259109497, | |
| "learning_rate": 3.1125000000000004e-05, | |
| "loss": 2.1223, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.4369538077403246, | |
| "grad_norm": 3.4398860931396484, | |
| "learning_rate": 3.1e-05, | |
| "loss": 2.0657, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.4394506866416979, | |
| "grad_norm": 3.4411044120788574, | |
| "learning_rate": 3.0875000000000005e-05, | |
| "loss": 2.0593, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.4419475655430712, | |
| "grad_norm": 3.5043869018554688, | |
| "learning_rate": 3.075e-05, | |
| "loss": 2.157, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.4444444444444444, | |
| "grad_norm": 3.3451061248779297, | |
| "learning_rate": 3.0625000000000006e-05, | |
| "loss": 1.996, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.4469413233458177, | |
| "grad_norm": 3.6858675479888916, | |
| "learning_rate": 3.05e-05, | |
| "loss": 2.1263, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.449438202247191, | |
| "grad_norm": 3.5967984199523926, | |
| "learning_rate": 3.0375000000000003e-05, | |
| "loss": 2.0165, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4519350811485643, | |
| "grad_norm": 3.5284011363983154, | |
| "learning_rate": 3.025e-05, | |
| "loss": 2.1286, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.4544319600499376, | |
| "grad_norm": 3.548715353012085, | |
| "learning_rate": 3.0125000000000004e-05, | |
| "loss": 2.0539, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.45692883895131087, | |
| "grad_norm": 3.52622127532959, | |
| "learning_rate": 3e-05, | |
| "loss": 2.1254, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.45942571785268416, | |
| "grad_norm": 3.3532912731170654, | |
| "learning_rate": 2.9875000000000004e-05, | |
| "loss": 2.1257, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.46192259675405745, | |
| "grad_norm": 3.4587035179138184, | |
| "learning_rate": 2.975e-05, | |
| "loss": 2.1171, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.46441947565543074, | |
| "grad_norm": 3.5046117305755615, | |
| "learning_rate": 2.9625000000000002e-05, | |
| "loss": 2.1057, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.46691635455680397, | |
| "grad_norm": 3.33771014213562, | |
| "learning_rate": 2.95e-05, | |
| "loss": 2.0668, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.46941323345817726, | |
| "grad_norm": 3.402855157852173, | |
| "learning_rate": 2.9375000000000003e-05, | |
| "loss": 2.087, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.47191011235955055, | |
| "grad_norm": 3.417536735534668, | |
| "learning_rate": 2.925e-05, | |
| "loss": 1.9994, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.47440699126092384, | |
| "grad_norm": 3.627957582473755, | |
| "learning_rate": 2.9125000000000003e-05, | |
| "loss": 2.1349, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4769038701622971, | |
| "grad_norm": 3.5531859397888184, | |
| "learning_rate": 2.9e-05, | |
| "loss": 2.1836, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.4794007490636704, | |
| "grad_norm": 3.5431745052337646, | |
| "learning_rate": 2.8875e-05, | |
| "loss": 2.0185, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.4818976279650437, | |
| "grad_norm": 3.592146158218384, | |
| "learning_rate": 2.8749999999999997e-05, | |
| "loss": 2.0973, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.484394506866417, | |
| "grad_norm": 3.4691829681396484, | |
| "learning_rate": 2.8625e-05, | |
| "loss": 2.1199, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.4868913857677903, | |
| "grad_norm": 3.599349021911621, | |
| "learning_rate": 2.8499999999999998e-05, | |
| "loss": 2.0849, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.4893882646691635, | |
| "grad_norm": 3.5513803958892822, | |
| "learning_rate": 2.8375000000000002e-05, | |
| "loss": 2.0859, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.4918851435705368, | |
| "grad_norm": 3.4182136058807373, | |
| "learning_rate": 2.825e-05, | |
| "loss": 2.0096, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.4943820224719101, | |
| "grad_norm": 3.347532272338867, | |
| "learning_rate": 2.8125000000000003e-05, | |
| "loss": 2.1039, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.4968789013732834, | |
| "grad_norm": 3.5202083587646484, | |
| "learning_rate": 2.8000000000000003e-05, | |
| "loss": 2.0436, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.4993757802746567, | |
| "grad_norm": 3.5688531398773193, | |
| "learning_rate": 2.7875e-05, | |
| "loss": 2.0736, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.50187265917603, | |
| "grad_norm": 3.531306743621826, | |
| "learning_rate": 2.7750000000000004e-05, | |
| "loss": 2.0244, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.5043695380774033, | |
| "grad_norm": 3.595841884613037, | |
| "learning_rate": 2.7625e-05, | |
| "loss": 2.0808, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.5068664169787765, | |
| "grad_norm": 3.610381603240967, | |
| "learning_rate": 2.7500000000000004e-05, | |
| "loss": 2.0863, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.5093632958801498, | |
| "grad_norm": 3.2456705570220947, | |
| "learning_rate": 2.7375e-05, | |
| "loss": 2.1102, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.5118601747815231, | |
| "grad_norm": 3.5384883880615234, | |
| "learning_rate": 2.725e-05, | |
| "loss": 2.0491, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.5143570536828964, | |
| "grad_norm": 3.4325191974639893, | |
| "learning_rate": 2.7125000000000002e-05, | |
| "loss": 2.0562, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.5168539325842697, | |
| "grad_norm": 3.428877115249634, | |
| "learning_rate": 2.7000000000000002e-05, | |
| "loss": 2.0329, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.519350811485643, | |
| "grad_norm": 3.451172351837158, | |
| "learning_rate": 2.6875e-05, | |
| "loss": 2.1357, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.5218476903870163, | |
| "grad_norm": 3.5145444869995117, | |
| "learning_rate": 2.6750000000000003e-05, | |
| "loss": 2.133, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.5243445692883895, | |
| "grad_norm": 3.452249050140381, | |
| "learning_rate": 2.6625e-05, | |
| "loss": 2.033, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.5268414481897628, | |
| "grad_norm": 3.493157148361206, | |
| "learning_rate": 2.6500000000000004e-05, | |
| "loss": 2.0387, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.529338327091136, | |
| "grad_norm": 3.4883360862731934, | |
| "learning_rate": 2.6375e-05, | |
| "loss": 2.0182, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.5318352059925093, | |
| "grad_norm": 3.377523183822632, | |
| "learning_rate": 2.625e-05, | |
| "loss": 2.0963, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.5343320848938826, | |
| "grad_norm": 3.4430289268493652, | |
| "learning_rate": 2.6124999999999998e-05, | |
| "loss": 2.0959, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.5368289637952559, | |
| "grad_norm": 3.380902051925659, | |
| "learning_rate": 2.6000000000000002e-05, | |
| "loss": 2.0883, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.5393258426966292, | |
| "grad_norm": 3.437469005584717, | |
| "learning_rate": 2.5875e-05, | |
| "loss": 2.0548, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.5418227215980025, | |
| "grad_norm": 3.628032684326172, | |
| "learning_rate": 2.5750000000000002e-05, | |
| "loss": 2.0784, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.5443196004993758, | |
| "grad_norm": 3.2870864868164062, | |
| "learning_rate": 2.5625e-05, | |
| "loss": 2.0463, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.5468164794007491, | |
| "grad_norm": 3.5962319374084473, | |
| "learning_rate": 2.5500000000000003e-05, | |
| "loss": 2.0625, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.5493133583021224, | |
| "grad_norm": 3.3193917274475098, | |
| "learning_rate": 2.5375e-05, | |
| "loss": 2.0249, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5518102372034956, | |
| "grad_norm": 3.4375598430633545, | |
| "learning_rate": 2.525e-05, | |
| "loss": 2.0802, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.5543071161048689, | |
| "grad_norm": 3.500824451446533, | |
| "learning_rate": 2.5124999999999997e-05, | |
| "loss": 2.0309, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.5568039950062422, | |
| "grad_norm": 3.5241281986236572, | |
| "learning_rate": 2.5e-05, | |
| "loss": 2.1013, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.5593008739076155, | |
| "grad_norm": 3.4705541133880615, | |
| "learning_rate": 2.4875e-05, | |
| "loss": 2.0883, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.5617977528089888, | |
| "grad_norm": 3.4071896076202393, | |
| "learning_rate": 2.4750000000000002e-05, | |
| "loss": 2.0965, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.5642946317103621, | |
| "grad_norm": 3.315619707107544, | |
| "learning_rate": 2.4625000000000002e-05, | |
| "loss": 2.0697, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.5667915106117354, | |
| "grad_norm": 3.280471086502075, | |
| "learning_rate": 2.45e-05, | |
| "loss": 2.1179, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.5692883895131086, | |
| "grad_norm": 3.7654521465301514, | |
| "learning_rate": 2.4375e-05, | |
| "loss": 2.0855, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.5717852684144819, | |
| "grad_norm": 3.7220211029052734, | |
| "learning_rate": 2.425e-05, | |
| "loss": 2.0319, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.5742821473158551, | |
| "grad_norm": 3.419952154159546, | |
| "learning_rate": 2.4125e-05, | |
| "loss": 2.0362, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5767790262172284, | |
| "grad_norm": 3.46474552154541, | |
| "learning_rate": 2.4e-05, | |
| "loss": 2.0325, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.5792759051186017, | |
| "grad_norm": 3.272773027420044, | |
| "learning_rate": 2.3875e-05, | |
| "loss": 2.0735, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.581772784019975, | |
| "grad_norm": 3.659705877304077, | |
| "learning_rate": 2.375e-05, | |
| "loss": 2.1283, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.5842696629213483, | |
| "grad_norm": 3.303420305252075, | |
| "learning_rate": 2.3624999999999998e-05, | |
| "loss": 2.1808, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.5867665418227216, | |
| "grad_norm": 3.622915744781494, | |
| "learning_rate": 2.35e-05, | |
| "loss": 2.0583, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.5892634207240949, | |
| "grad_norm": 3.3635072708129883, | |
| "learning_rate": 2.3375000000000002e-05, | |
| "loss": 2.0519, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.5917602996254682, | |
| "grad_norm": 3.5549280643463135, | |
| "learning_rate": 2.3250000000000003e-05, | |
| "loss": 2.0682, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.5942571785268415, | |
| "grad_norm": 3.578655958175659, | |
| "learning_rate": 2.3125000000000003e-05, | |
| "loss": 2.077, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.5967540574282147, | |
| "grad_norm": 3.611335277557373, | |
| "learning_rate": 2.3000000000000003e-05, | |
| "loss": 1.9911, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.599250936329588, | |
| "grad_norm": 3.3821637630462646, | |
| "learning_rate": 2.2875e-05, | |
| "loss": 2.0082, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.6017478152309613, | |
| "grad_norm": 3.4852335453033447, | |
| "learning_rate": 2.275e-05, | |
| "loss": 2.0278, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.6042446941323346, | |
| "grad_norm": 3.525944948196411, | |
| "learning_rate": 2.2625e-05, | |
| "loss": 2.0914, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.6067415730337079, | |
| "grad_norm": 3.3227696418762207, | |
| "learning_rate": 2.25e-05, | |
| "loss": 2.107, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.6092384519350812, | |
| "grad_norm": 3.669677257537842, | |
| "learning_rate": 2.2375000000000002e-05, | |
| "loss": 1.9829, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.6117353308364545, | |
| "grad_norm": 3.625420570373535, | |
| "learning_rate": 2.2250000000000002e-05, | |
| "loss": 2.0123, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.6142322097378277, | |
| "grad_norm": 3.4239187240600586, | |
| "learning_rate": 2.2125000000000002e-05, | |
| "loss": 2.0562, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.616729088639201, | |
| "grad_norm": 3.5444185733795166, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 2.0755, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.6192259675405742, | |
| "grad_norm": 3.500282049179077, | |
| "learning_rate": 2.1875e-05, | |
| "loss": 2.0275, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.6217228464419475, | |
| "grad_norm": 3.4052162170410156, | |
| "learning_rate": 2.175e-05, | |
| "loss": 2.0654, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.6242197253433208, | |
| "grad_norm": 3.5831384658813477, | |
| "learning_rate": 2.1625e-05, | |
| "loss": 1.994, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6267166042446941, | |
| "grad_norm": 3.322357654571533, | |
| "learning_rate": 2.15e-05, | |
| "loss": 2.0476, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.6292134831460674, | |
| "grad_norm": 3.543062448501587, | |
| "learning_rate": 2.1375e-05, | |
| "loss": 2.0585, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.6317103620474407, | |
| "grad_norm": 3.521042823791504, | |
| "learning_rate": 2.125e-05, | |
| "loss": 2.0239, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.634207240948814, | |
| "grad_norm": 3.395559549331665, | |
| "learning_rate": 2.1125000000000002e-05, | |
| "loss": 2.0957, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.6367041198501873, | |
| "grad_norm": 3.567340135574341, | |
| "learning_rate": 2.1e-05, | |
| "loss": 2.0649, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.6392009987515606, | |
| "grad_norm": 3.397399663925171, | |
| "learning_rate": 2.0875e-05, | |
| "loss": 2.1739, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.6416978776529338, | |
| "grad_norm": 3.4432034492492676, | |
| "learning_rate": 2.075e-05, | |
| "loss": 2.0686, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.6441947565543071, | |
| "grad_norm": 3.4990758895874023, | |
| "learning_rate": 2.0625e-05, | |
| "loss": 2.1194, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.6466916354556804, | |
| "grad_norm": 3.497478723526001, | |
| "learning_rate": 2.05e-05, | |
| "loss": 2.0429, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.6491885143570537, | |
| "grad_norm": 3.4457273483276367, | |
| "learning_rate": 2.0375e-05, | |
| "loss": 2.1038, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.651685393258427, | |
| "grad_norm": 3.3803672790527344, | |
| "learning_rate": 2.025e-05, | |
| "loss": 2.1119, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.6541822721598003, | |
| "grad_norm": 3.474200963973999, | |
| "learning_rate": 2.0125e-05, | |
| "loss": 2.0346, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.6566791510611736, | |
| "grad_norm": 3.4010913372039795, | |
| "learning_rate": 2e-05, | |
| "loss": 2.0361, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.6591760299625468, | |
| "grad_norm": 3.391678810119629, | |
| "learning_rate": 1.9875000000000002e-05, | |
| "loss": 2.0216, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.66167290886392, | |
| "grad_norm": 3.484393835067749, | |
| "learning_rate": 1.9750000000000002e-05, | |
| "loss": 2.1634, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.6641697877652933, | |
| "grad_norm": 3.491408586502075, | |
| "learning_rate": 1.9625000000000003e-05, | |
| "loss": 2.0471, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.6666666666666666, | |
| "grad_norm": 3.562328815460205, | |
| "learning_rate": 1.9500000000000003e-05, | |
| "loss": 2.0928, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.6691635455680399, | |
| "grad_norm": 3.6033616065979004, | |
| "learning_rate": 1.9375e-05, | |
| "loss": 2.0641, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.6716604244694132, | |
| "grad_norm": 3.441002130508423, | |
| "learning_rate": 1.925e-05, | |
| "loss": 2.0484, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.6741573033707865, | |
| "grad_norm": 3.4448623657226562, | |
| "learning_rate": 1.9125e-05, | |
| "loss": 2.0558, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6766541822721598, | |
| "grad_norm": 3.481809377670288, | |
| "learning_rate": 1.9e-05, | |
| "loss": 2.0449, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.6791510611735331, | |
| "grad_norm": 3.2882351875305176, | |
| "learning_rate": 1.8875e-05, | |
| "loss": 1.9441, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.6816479400749064, | |
| "grad_norm": 3.5338022708892822, | |
| "learning_rate": 1.8750000000000002e-05, | |
| "loss": 2.1015, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.6841448189762797, | |
| "grad_norm": 3.349963665008545, | |
| "learning_rate": 1.8625000000000002e-05, | |
| "loss": 2.0868, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.686641697877653, | |
| "grad_norm": 3.562603712081909, | |
| "learning_rate": 1.85e-05, | |
| "loss": 2.0308, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.6891385767790262, | |
| "grad_norm": 3.484811782836914, | |
| "learning_rate": 1.8375e-05, | |
| "loss": 2.0972, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.6916354556803995, | |
| "grad_norm": 3.4616386890411377, | |
| "learning_rate": 1.825e-05, | |
| "loss": 2.0084, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.6941323345817728, | |
| "grad_norm": 3.5441646575927734, | |
| "learning_rate": 1.8125e-05, | |
| "loss": 2.0395, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.6966292134831461, | |
| "grad_norm": 3.5133986473083496, | |
| "learning_rate": 1.8e-05, | |
| "loss": 2.0973, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.6991260923845194, | |
| "grad_norm": 3.6960537433624268, | |
| "learning_rate": 1.7875e-05, | |
| "loss": 1.9632, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.7016229712858927, | |
| "grad_norm": 3.499337911605835, | |
| "learning_rate": 1.775e-05, | |
| "loss": 2.0925, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.704119850187266, | |
| "grad_norm": 3.451720952987671, | |
| "learning_rate": 1.7625e-05, | |
| "loss": 2.0816, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.7066167290886392, | |
| "grad_norm": 3.540565252304077, | |
| "learning_rate": 1.75e-05, | |
| "loss": 2.0853, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.7091136079900124, | |
| "grad_norm": 3.5862390995025635, | |
| "learning_rate": 1.7375e-05, | |
| "loss": 1.9933, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.7116104868913857, | |
| "grad_norm": 3.436612844467163, | |
| "learning_rate": 1.725e-05, | |
| "loss": 2.0519, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.714107365792759, | |
| "grad_norm": 3.5170860290527344, | |
| "learning_rate": 1.7125000000000003e-05, | |
| "loss": 2.0087, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.7166042446941323, | |
| "grad_norm": 3.613799571990967, | |
| "learning_rate": 1.7000000000000003e-05, | |
| "loss": 2.0785, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.7191011235955056, | |
| "grad_norm": 3.4286880493164062, | |
| "learning_rate": 1.6875000000000004e-05, | |
| "loss": 2.0805, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.7215980024968789, | |
| "grad_norm": 3.4460251331329346, | |
| "learning_rate": 1.675e-05, | |
| "loss": 2.0529, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.7240948813982522, | |
| "grad_norm": 3.5601372718811035, | |
| "learning_rate": 1.6625e-05, | |
| "loss": 2.0795, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7265917602996255, | |
| "grad_norm": 3.3969712257385254, | |
| "learning_rate": 1.65e-05, | |
| "loss": 2.0261, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.7290886392009988, | |
| "grad_norm": 3.285691738128662, | |
| "learning_rate": 1.6375e-05, | |
| "loss": 2.0902, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.731585518102372, | |
| "grad_norm": 3.482919454574585, | |
| "learning_rate": 1.6250000000000002e-05, | |
| "loss": 2.0239, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.7340823970037453, | |
| "grad_norm": 3.489551544189453, | |
| "learning_rate": 1.6125000000000002e-05, | |
| "loss": 2.0235, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.7365792759051186, | |
| "grad_norm": 3.461838960647583, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 2.0562, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.7390761548064919, | |
| "grad_norm": 3.3764636516571045, | |
| "learning_rate": 1.5875e-05, | |
| "loss": 2.0649, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.7415730337078652, | |
| "grad_norm": 3.4460861682891846, | |
| "learning_rate": 1.575e-05, | |
| "loss": 2.0674, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.7440699126092385, | |
| "grad_norm": 3.5665292739868164, | |
| "learning_rate": 1.5625e-05, | |
| "loss": 2.0219, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.7465667915106118, | |
| "grad_norm": 3.505133867263794, | |
| "learning_rate": 1.55e-05, | |
| "loss": 2.0278, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.7490636704119851, | |
| "grad_norm": 3.615962266921997, | |
| "learning_rate": 1.5375e-05, | |
| "loss": 2.0006, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7515605493133583, | |
| "grad_norm": 3.462989330291748, | |
| "learning_rate": 1.525e-05, | |
| "loss": 2.0423, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.7540574282147315, | |
| "grad_norm": 3.351733446121216, | |
| "learning_rate": 1.5125e-05, | |
| "loss": 2.0353, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.7565543071161048, | |
| "grad_norm": 3.3703553676605225, | |
| "learning_rate": 1.5e-05, | |
| "loss": 2.0352, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.7590511860174781, | |
| "grad_norm": 3.578253984451294, | |
| "learning_rate": 1.4875e-05, | |
| "loss": 2.027, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.7615480649188514, | |
| "grad_norm": 3.304739236831665, | |
| "learning_rate": 1.475e-05, | |
| "loss": 2.0538, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.7640449438202247, | |
| "grad_norm": 3.5209975242614746, | |
| "learning_rate": 1.4625e-05, | |
| "loss": 1.9996, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.766541822721598, | |
| "grad_norm": 3.5905096530914307, | |
| "learning_rate": 1.45e-05, | |
| "loss": 2.0141, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.7690387016229713, | |
| "grad_norm": 3.383382558822632, | |
| "learning_rate": 1.4374999999999999e-05, | |
| "loss": 1.9648, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.7715355805243446, | |
| "grad_norm": 3.4596011638641357, | |
| "learning_rate": 1.4249999999999999e-05, | |
| "loss": 2.0336, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.7740324594257179, | |
| "grad_norm": 3.4215636253356934, | |
| "learning_rate": 1.4125e-05, | |
| "loss": 2.0386, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.7765293383270911, | |
| "grad_norm": 3.341522693634033, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 2.1078, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.7790262172284644, | |
| "grad_norm": 3.366393566131592, | |
| "learning_rate": 1.3875000000000002e-05, | |
| "loss": 2.079, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.7815230961298377, | |
| "grad_norm": 3.497201919555664, | |
| "learning_rate": 1.3750000000000002e-05, | |
| "loss": 2.0535, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.784019975031211, | |
| "grad_norm": 3.563197135925293, | |
| "learning_rate": 1.3625e-05, | |
| "loss": 2.0283, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.7865168539325843, | |
| "grad_norm": 3.4403908252716064, | |
| "learning_rate": 1.3500000000000001e-05, | |
| "loss": 2.0832, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.7890137328339576, | |
| "grad_norm": 3.4520885944366455, | |
| "learning_rate": 1.3375000000000002e-05, | |
| "loss": 2.0146, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.7915106117353309, | |
| "grad_norm": 3.5180275440216064, | |
| "learning_rate": 1.3250000000000002e-05, | |
| "loss": 2.004, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.7940074906367042, | |
| "grad_norm": 3.153761863708496, | |
| "learning_rate": 1.3125e-05, | |
| "loss": 2.0935, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.7965043695380774, | |
| "grad_norm": 3.468968629837036, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 2.066, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.7990012484394506, | |
| "grad_norm": 3.5003297328948975, | |
| "learning_rate": 1.2875000000000001e-05, | |
| "loss": 1.9856, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.8014981273408239, | |
| "grad_norm": 3.5374064445495605, | |
| "learning_rate": 1.2750000000000002e-05, | |
| "loss": 2.0671, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.8039950062421972, | |
| "grad_norm": 3.659308433532715, | |
| "learning_rate": 1.2625e-05, | |
| "loss": 1.9864, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.8064918851435705, | |
| "grad_norm": 3.5156655311584473, | |
| "learning_rate": 1.25e-05, | |
| "loss": 1.9394, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.8089887640449438, | |
| "grad_norm": 3.326158285140991, | |
| "learning_rate": 1.2375000000000001e-05, | |
| "loss": 1.961, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.8114856429463171, | |
| "grad_norm": 3.4294004440307617, | |
| "learning_rate": 1.225e-05, | |
| "loss": 2.0097, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.8139825218476904, | |
| "grad_norm": 3.4218974113464355, | |
| "learning_rate": 1.2125e-05, | |
| "loss": 2.0647, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.8164794007490637, | |
| "grad_norm": 3.438875198364258, | |
| "learning_rate": 1.2e-05, | |
| "loss": 2.0179, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.818976279650437, | |
| "grad_norm": 3.523358106613159, | |
| "learning_rate": 1.1875e-05, | |
| "loss": 2.0494, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.8214731585518102, | |
| "grad_norm": 3.506168842315674, | |
| "learning_rate": 1.175e-05, | |
| "loss": 2.1027, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.8239700374531835, | |
| "grad_norm": 3.548112154006958, | |
| "learning_rate": 1.1625000000000001e-05, | |
| "loss": 2.0389, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.8264669163545568, | |
| "grad_norm": 3.5158162117004395, | |
| "learning_rate": 1.1500000000000002e-05, | |
| "loss": 1.978, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.8289637952559301, | |
| "grad_norm": 3.4753963947296143, | |
| "learning_rate": 1.1375e-05, | |
| "loss": 2.0421, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.8314606741573034, | |
| "grad_norm": 3.5032706260681152, | |
| "learning_rate": 1.125e-05, | |
| "loss": 1.9888, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.8339575530586767, | |
| "grad_norm": 3.2991061210632324, | |
| "learning_rate": 1.1125000000000001e-05, | |
| "loss": 2.0151, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.83645443196005, | |
| "grad_norm": 3.425863027572632, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 2.0503, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.8389513108614233, | |
| "grad_norm": 3.469403028488159, | |
| "learning_rate": 1.0875e-05, | |
| "loss": 1.9425, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.8414481897627965, | |
| "grad_norm": 3.3070104122161865, | |
| "learning_rate": 1.075e-05, | |
| "loss": 2.0603, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.8439450686641697, | |
| "grad_norm": 3.418203115463257, | |
| "learning_rate": 1.0625e-05, | |
| "loss": 2.0018, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.846441947565543, | |
| "grad_norm": 3.5066540241241455, | |
| "learning_rate": 1.05e-05, | |
| "loss": 2.0122, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.8489388264669163, | |
| "grad_norm": 3.4425551891326904, | |
| "learning_rate": 1.0375e-05, | |
| "loss": 2.0225, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8514357053682896, | |
| "grad_norm": 3.343202829360962, | |
| "learning_rate": 1.025e-05, | |
| "loss": 2.0755, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.8539325842696629, | |
| "grad_norm": 3.28155255317688, | |
| "learning_rate": 1.0125e-05, | |
| "loss": 1.9847, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.8564294631710362, | |
| "grad_norm": 3.3546323776245117, | |
| "learning_rate": 1e-05, | |
| "loss": 1.991, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.8589263420724095, | |
| "grad_norm": 3.420783758163452, | |
| "learning_rate": 9.875000000000001e-06, | |
| "loss": 2.0471, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.8614232209737828, | |
| "grad_norm": 3.347703456878662, | |
| "learning_rate": 9.750000000000002e-06, | |
| "loss": 2.0444, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.8639200998751561, | |
| "grad_norm": 3.4276270866394043, | |
| "learning_rate": 9.625e-06, | |
| "loss": 2.0548, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.8664169787765293, | |
| "grad_norm": 3.3684396743774414, | |
| "learning_rate": 9.5e-06, | |
| "loss": 2.0664, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.8689138576779026, | |
| "grad_norm": 3.5308725833892822, | |
| "learning_rate": 9.375000000000001e-06, | |
| "loss": 2.0263, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.8714107365792759, | |
| "grad_norm": 3.5037009716033936, | |
| "learning_rate": 9.25e-06, | |
| "loss": 2.0629, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.8739076154806492, | |
| "grad_norm": 3.3732502460479736, | |
| "learning_rate": 9.125e-06, | |
| "loss": 2.0872, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.8764044943820225, | |
| "grad_norm": 3.379492998123169, | |
| "learning_rate": 9e-06, | |
| "loss": 2.0494, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.8789013732833958, | |
| "grad_norm": 3.356729030609131, | |
| "learning_rate": 8.875e-06, | |
| "loss": 2.0578, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.8813982521847691, | |
| "grad_norm": 3.505227565765381, | |
| "learning_rate": 8.75e-06, | |
| "loss": 2.0547, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.8838951310861424, | |
| "grad_norm": 3.325188159942627, | |
| "learning_rate": 8.625e-06, | |
| "loss": 2.0479, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.8863920099875156, | |
| "grad_norm": 3.220914363861084, | |
| "learning_rate": 8.500000000000002e-06, | |
| "loss": 2.1145, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.8888888888888888, | |
| "grad_norm": 3.2536563873291016, | |
| "learning_rate": 8.375e-06, | |
| "loss": 2.1226, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.8913857677902621, | |
| "grad_norm": 3.4424164295196533, | |
| "learning_rate": 8.25e-06, | |
| "loss": 1.9732, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.8938826466916354, | |
| "grad_norm": 3.210689067840576, | |
| "learning_rate": 8.125000000000001e-06, | |
| "loss": 2.0682, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.8963795255930087, | |
| "grad_norm": 3.6276254653930664, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.9996, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.898876404494382, | |
| "grad_norm": 3.284266471862793, | |
| "learning_rate": 7.875e-06, | |
| "loss": 1.9988, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.9013732833957553, | |
| "grad_norm": 3.3102824687957764, | |
| "learning_rate": 7.75e-06, | |
| "loss": 2.0986, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.9038701622971286, | |
| "grad_norm": 3.4718074798583984, | |
| "learning_rate": 7.625e-06, | |
| "loss": 2.0236, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.9063670411985019, | |
| "grad_norm": 3.356231689453125, | |
| "learning_rate": 7.5e-06, | |
| "loss": 1.9603, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.9088639200998752, | |
| "grad_norm": 3.488931655883789, | |
| "learning_rate": 7.375e-06, | |
| "loss": 2.0287, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.9113607990012484, | |
| "grad_norm": 3.370278835296631, | |
| "learning_rate": 7.25e-06, | |
| "loss": 2.055, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.9138576779026217, | |
| "grad_norm": 3.357985019683838, | |
| "learning_rate": 7.1249999999999995e-06, | |
| "loss": 2.0225, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.916354556803995, | |
| "grad_norm": 3.3029778003692627, | |
| "learning_rate": 7.000000000000001e-06, | |
| "loss": 1.9724, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.9188514357053683, | |
| "grad_norm": 3.6066527366638184, | |
| "learning_rate": 6.875000000000001e-06, | |
| "loss": 2.0172, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.9213483146067416, | |
| "grad_norm": 3.4643936157226562, | |
| "learning_rate": 6.750000000000001e-06, | |
| "loss": 2.0598, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.9238451935081149, | |
| "grad_norm": 3.2209830284118652, | |
| "learning_rate": 6.625000000000001e-06, | |
| "loss": 2.0395, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.9263420724094882, | |
| "grad_norm": 3.308652639389038, | |
| "learning_rate": 6.5000000000000004e-06, | |
| "loss": 2.0576, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.9288389513108615, | |
| "grad_norm": 3.378284215927124, | |
| "learning_rate": 6.375000000000001e-06, | |
| "loss": 2.1037, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.9313358302122348, | |
| "grad_norm": 3.408998489379883, | |
| "learning_rate": 6.25e-06, | |
| "loss": 2.0508, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.9338327091136079, | |
| "grad_norm": 3.4758336544036865, | |
| "learning_rate": 6.125e-06, | |
| "loss": 2.0642, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.9363295880149812, | |
| "grad_norm": 3.3644943237304688, | |
| "learning_rate": 6e-06, | |
| "loss": 2.104, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.9388264669163545, | |
| "grad_norm": 3.3621973991394043, | |
| "learning_rate": 5.875e-06, | |
| "loss": 2.05, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.9413233458177278, | |
| "grad_norm": 3.3936052322387695, | |
| "learning_rate": 5.750000000000001e-06, | |
| "loss": 1.9648, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.9438202247191011, | |
| "grad_norm": 3.28177809715271, | |
| "learning_rate": 5.625e-06, | |
| "loss": 2.0129, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.9463171036204744, | |
| "grad_norm": 3.1901121139526367, | |
| "learning_rate": 5.500000000000001e-06, | |
| "loss": 2.0564, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.9488139825218477, | |
| "grad_norm": 3.4150338172912598, | |
| "learning_rate": 5.375e-06, | |
| "loss": 2.0344, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.951310861423221, | |
| "grad_norm": 3.3532180786132812, | |
| "learning_rate": 5.25e-06, | |
| "loss": 2.012, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.9538077403245943, | |
| "grad_norm": 3.434931993484497, | |
| "learning_rate": 5.125e-06, | |
| "loss": 2.073, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.9563046192259675, | |
| "grad_norm": 3.34334135055542, | |
| "learning_rate": 5e-06, | |
| "loss": 2.0716, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.9588014981273408, | |
| "grad_norm": 3.3420920372009277, | |
| "learning_rate": 4.875000000000001e-06, | |
| "loss": 2.0691, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.9612983770287141, | |
| "grad_norm": 3.648798942565918, | |
| "learning_rate": 4.75e-06, | |
| "loss": 2.0198, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.9637952559300874, | |
| "grad_norm": 3.2924463748931885, | |
| "learning_rate": 4.625e-06, | |
| "loss": 2.015, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.9662921348314607, | |
| "grad_norm": 3.2124977111816406, | |
| "learning_rate": 4.5e-06, | |
| "loss": 2.1013, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.968789013732834, | |
| "grad_norm": 3.4171626567840576, | |
| "learning_rate": 4.375e-06, | |
| "loss": 2.0419, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.9712858926342073, | |
| "grad_norm": 3.4217112064361572, | |
| "learning_rate": 4.250000000000001e-06, | |
| "loss": 2.0808, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.9737827715355806, | |
| "grad_norm": 3.281397581100464, | |
| "learning_rate": 4.125e-06, | |
| "loss": 2.0552, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.9762796504369539, | |
| "grad_norm": 3.3025379180908203, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 2.0196, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.978776529338327, | |
| "grad_norm": 3.261284112930298, | |
| "learning_rate": 3.875e-06, | |
| "loss": 1.903, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.9812734082397003, | |
| "grad_norm": 3.3036253452301025, | |
| "learning_rate": 3.75e-06, | |
| "loss": 2.0455, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.9837702871410736, | |
| "grad_norm": 3.4235925674438477, | |
| "learning_rate": 3.625e-06, | |
| "loss": 2.0393, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.9862671660424469, | |
| "grad_norm": 3.347456455230713, | |
| "learning_rate": 3.5000000000000004e-06, | |
| "loss": 2.0227, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.9887640449438202, | |
| "grad_norm": 3.3116769790649414, | |
| "learning_rate": 3.3750000000000003e-06, | |
| "loss": 2.0891, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.9912609238451935, | |
| "grad_norm": 3.2966830730438232, | |
| "learning_rate": 3.2500000000000002e-06, | |
| "loss": 2.0784, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.9937578027465668, | |
| "grad_norm": 3.4097273349761963, | |
| "learning_rate": 3.125e-06, | |
| "loss": 2.0751, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.9962546816479401, | |
| "grad_norm": 3.3749406337738037, | |
| "learning_rate": 3e-06, | |
| "loss": 2.0916, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.9987515605493134, | |
| "grad_norm": 3.322603225708008, | |
| "learning_rate": 2.8750000000000004e-06, | |
| "loss": 2.0949, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.9987515605493134, | |
| "step": 400, | |
| "total_flos": 2.847507751108608e+17, | |
| "train_loss": 2.1486885741353037, | |
| "train_runtime": 92894.5901, | |
| "train_samples_per_second": 0.241, | |
| "train_steps_per_second": 0.004 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 400, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 400, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.847507751108608e+17, | |
| "train_batch_size": 14, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |