{ "best_global_step": 166236, "best_metric": 0.7059909142532652, "best_model_checkpoint": "modernbert-heritage-classification/checkpoint-166236", "epoch": 3.0, "eval_steps": 500, "global_step": 249354, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0060155441661252675, "grad_norm": 3.192460536956787, "learning_rate": 1.996005678673693e-05, "loss": 0.4511, "step": 500 }, { "epoch": 0.012031088332250535, "grad_norm": 2.3000173568725586, "learning_rate": 1.992003336621831e-05, "loss": 0.4198, "step": 1000 }, { "epoch": 0.018046632498375802, "grad_norm": 6.538833141326904, "learning_rate": 1.9879929738444142e-05, "loss": 0.4086, "step": 1500 }, { "epoch": 0.02406217666450107, "grad_norm": 0.18448229134082794, "learning_rate": 1.9839826110669975e-05, "loss": 0.3574, "step": 2000 }, { "epoch": 0.030077720830626337, "grad_norm": 5.782385349273682, "learning_rate": 1.9799722482895804e-05, "loss": 0.3798, "step": 2500 }, { "epoch": 0.036093264996751605, "grad_norm": 2.290853500366211, "learning_rate": 1.9759699062377184e-05, "loss": 0.38, "step": 3000 }, { "epoch": 0.042108809162876876, "grad_norm": 4.156062126159668, "learning_rate": 1.9719595434603017e-05, "loss": 0.3838, "step": 3500 }, { "epoch": 0.04812435332900214, "grad_norm": 2.1094439029693604, "learning_rate": 1.967949180682885e-05, "loss": 0.3625, "step": 4000 }, { "epoch": 0.05413989749512741, "grad_norm": 8.78498363494873, "learning_rate": 1.963938817905468e-05, "loss": 0.3834, "step": 4500 }, { "epoch": 0.060155441661252675, "grad_norm": 0.003325940575450659, "learning_rate": 1.959928455128051e-05, "loss": 0.3395, "step": 5000 }, { "epoch": 0.06617098582737795, "grad_norm": 0.030467336997389793, "learning_rate": 1.9559180923506344e-05, "loss": 0.3605, "step": 5500 }, { "epoch": 0.07218652999350321, "grad_norm": 0.7140659689903259, "learning_rate": 1.9519077295732173e-05, "loss": 0.352, "step": 6000 }, { "epoch": 0.07820207415962847, "grad_norm": 0.013019168749451637, "learning_rate": 1.9479053875213553e-05, "loss": 0.3533, "step": 6500 }, { "epoch": 0.08421761832575375, "grad_norm": 3.8174266815185547, "learning_rate": 1.9438950247439386e-05, "loss": 0.316, "step": 7000 }, { "epoch": 0.09023316249187902, "grad_norm": 7.174667835235596, "learning_rate": 1.939884661966522e-05, "loss": 0.3495, "step": 7500 }, { "epoch": 0.09624870665800428, "grad_norm": 1.6871285438537598, "learning_rate": 1.9358742991891048e-05, "loss": 0.3449, "step": 8000 }, { "epoch": 0.10226425082412956, "grad_norm": 0.04142903909087181, "learning_rate": 1.931863936411688e-05, "loss": 0.3212, "step": 8500 }, { "epoch": 0.10827979499025482, "grad_norm": 0.0040322355926036835, "learning_rate": 1.9278535736342713e-05, "loss": 0.3416, "step": 9000 }, { "epoch": 0.11429533915638009, "grad_norm": 7.649967193603516, "learning_rate": 1.9238512315824093e-05, "loss": 0.3615, "step": 9500 }, { "epoch": 0.12031088332250535, "grad_norm": 2.4331884384155273, "learning_rate": 1.9198408688049922e-05, "loss": 0.354, "step": 10000 }, { "epoch": 0.12632642748863063, "grad_norm": 8.657504081726074, "learning_rate": 1.9158305060275755e-05, "loss": 0.3413, "step": 10500 }, { "epoch": 0.1323419716547559, "grad_norm": 8.941267013549805, "learning_rate": 1.9118201432501588e-05, "loss": 0.3211, "step": 11000 }, { "epoch": 0.13835751582088116, "grad_norm": 16.12137794494629, "learning_rate": 1.9078178011982968e-05, "loss": 0.3323, "step": 11500 }, { "epoch": 0.14437305998700642, "grad_norm": 0.30122852325439453, "learning_rate": 1.9038074384208797e-05, "loss": 0.3163, "step": 12000 }, { "epoch": 0.15038860415313168, "grad_norm": 0.2980821132659912, "learning_rate": 1.899797075643463e-05, "loss": 0.3264, "step": 12500 }, { "epoch": 0.15640414831925695, "grad_norm": 0.006154663860797882, "learning_rate": 1.8957867128660462e-05, "loss": 0.3453, "step": 13000 }, { "epoch": 0.16241969248538224, "grad_norm": 0.8030881285667419, "learning_rate": 1.8917763500886292e-05, "loss": 0.3152, "step": 13500 }, { "epoch": 0.1684352366515075, "grad_norm": 17.813459396362305, "learning_rate": 1.8877659873112124e-05, "loss": 0.3189, "step": 14000 }, { "epoch": 0.17445078081763277, "grad_norm": 1.1400656700134277, "learning_rate": 1.8837556245337957e-05, "loss": 0.3434, "step": 14500 }, { "epoch": 0.18046632498375803, "grad_norm": 0.09848301857709885, "learning_rate": 1.8797452617563786e-05, "loss": 0.3087, "step": 15000 }, { "epoch": 0.1864818691498833, "grad_norm": 0.04860129952430725, "learning_rate": 1.875734898978962e-05, "loss": 0.3289, "step": 15500 }, { "epoch": 0.19249741331600856, "grad_norm": 0.012527902610599995, "learning_rate": 1.8717325569271e-05, "loss": 0.3324, "step": 16000 }, { "epoch": 0.19851295748213382, "grad_norm": 5.626002788543701, "learning_rate": 1.867730214875238e-05, "loss": 0.3351, "step": 16500 }, { "epoch": 0.20452850164825911, "grad_norm": 0.2671683430671692, "learning_rate": 1.8637198520978212e-05, "loss": 0.3467, "step": 17000 }, { "epoch": 0.21054404581438438, "grad_norm": 0.4044207036495209, "learning_rate": 1.859709489320404e-05, "loss": 0.3248, "step": 17500 }, { "epoch": 0.21655958998050964, "grad_norm": 0.19454579055309296, "learning_rate": 1.8556991265429874e-05, "loss": 0.339, "step": 18000 }, { "epoch": 0.2225751341466349, "grad_norm": 13.63963508605957, "learning_rate": 1.8516887637655703e-05, "loss": 0.3252, "step": 18500 }, { "epoch": 0.22859067831276017, "grad_norm": 3.8693294525146484, "learning_rate": 1.8476784009881536e-05, "loss": 0.3311, "step": 19000 }, { "epoch": 0.23460622247888543, "grad_norm": 2.644678831100464, "learning_rate": 1.8436760589362916e-05, "loss": 0.3378, "step": 19500 }, { "epoch": 0.2406217666450107, "grad_norm": 1.1504566669464111, "learning_rate": 1.8396656961588745e-05, "loss": 0.3162, "step": 20000 }, { "epoch": 0.24663731081113596, "grad_norm": 0.33002209663391113, "learning_rate": 1.8356553333814578e-05, "loss": 0.3112, "step": 20500 }, { "epoch": 0.25265285497726125, "grad_norm": 5.884070873260498, "learning_rate": 1.831644970604041e-05, "loss": 0.3375, "step": 21000 }, { "epoch": 0.2586683991433865, "grad_norm": 0.004158430732786655, "learning_rate": 1.827634607826624e-05, "loss": 0.2711, "step": 21500 }, { "epoch": 0.2646839433095118, "grad_norm": 6.601622581481934, "learning_rate": 1.8236242450492072e-05, "loss": 0.3031, "step": 22000 }, { "epoch": 0.27069948747563705, "grad_norm": 4.164157390594482, "learning_rate": 1.8196138822717905e-05, "loss": 0.3228, "step": 22500 }, { "epoch": 0.2767150316417623, "grad_norm": 34.43812942504883, "learning_rate": 1.8156035194943734e-05, "loss": 0.3061, "step": 23000 }, { "epoch": 0.2827305758078876, "grad_norm": 0.02143733762204647, "learning_rate": 1.8115931567169567e-05, "loss": 0.3163, "step": 23500 }, { "epoch": 0.28874611997401284, "grad_norm": 9.730279922485352, "learning_rate": 1.8075908146650947e-05, "loss": 0.2817, "step": 24000 }, { "epoch": 0.2947616641401381, "grad_norm": 9.340431213378906, "learning_rate": 1.803580451887678e-05, "loss": 0.3105, "step": 24500 }, { "epoch": 0.30077720830626337, "grad_norm": 9.22760009765625, "learning_rate": 1.799570089110261e-05, "loss": 0.309, "step": 25000 }, { "epoch": 0.30679275247238863, "grad_norm": 3.7310335636138916, "learning_rate": 1.795567747058399e-05, "loss": 0.3075, "step": 25500 }, { "epoch": 0.3128082966385139, "grad_norm": 4.968273639678955, "learning_rate": 1.7915573842809822e-05, "loss": 0.3056, "step": 26000 }, { "epoch": 0.3188238408046392, "grad_norm": 15.309103012084961, "learning_rate": 1.7875470215035654e-05, "loss": 0.3063, "step": 26500 }, { "epoch": 0.3248393849707645, "grad_norm": 0.028000958263874054, "learning_rate": 1.7835366587261484e-05, "loss": 0.3251, "step": 27000 }, { "epoch": 0.33085492913688974, "grad_norm": 11.808887481689453, "learning_rate": 1.7795262959487316e-05, "loss": 0.3335, "step": 27500 }, { "epoch": 0.336870473303015, "grad_norm": 6.1232123374938965, "learning_rate": 1.775515933171315e-05, "loss": 0.2928, "step": 28000 }, { "epoch": 0.34288601746914027, "grad_norm": 2.071223258972168, "learning_rate": 1.771505570393898e-05, "loss": 0.3228, "step": 28500 }, { "epoch": 0.34890156163526553, "grad_norm": 6.20293664932251, "learning_rate": 1.767495207616481e-05, "loss": 0.3115, "step": 29000 }, { "epoch": 0.3549171058013908, "grad_norm": 0.5319348573684692, "learning_rate": 1.7634848448390644e-05, "loss": 0.3424, "step": 29500 }, { "epoch": 0.36093264996751606, "grad_norm": 0.15102390944957733, "learning_rate": 1.7594825027872024e-05, "loss": 0.3198, "step": 30000 }, { "epoch": 0.3669481941336413, "grad_norm": 10.45068359375, "learning_rate": 1.7554721400097853e-05, "loss": 0.3083, "step": 30500 }, { "epoch": 0.3729637382997666, "grad_norm": 0.3780848979949951, "learning_rate": 1.7514697979579233e-05, "loss": 0.3079, "step": 31000 }, { "epoch": 0.37897928246589185, "grad_norm": 0.02213098295032978, "learning_rate": 1.7474674559060613e-05, "loss": 0.301, "step": 31500 }, { "epoch": 0.3849948266320171, "grad_norm": 2.598281145095825, "learning_rate": 1.7434570931286446e-05, "loss": 0.2878, "step": 32000 }, { "epoch": 0.3910103707981424, "grad_norm": 0.14519445598125458, "learning_rate": 1.739446730351228e-05, "loss": 0.3231, "step": 32500 }, { "epoch": 0.39702591496426765, "grad_norm": 0.04602254554629326, "learning_rate": 1.7354363675738108e-05, "loss": 0.2972, "step": 33000 }, { "epoch": 0.4030414591303929, "grad_norm": 18.00914192199707, "learning_rate": 1.731426004796394e-05, "loss": 0.323, "step": 33500 }, { "epoch": 0.40905700329651823, "grad_norm": 6.137303829193115, "learning_rate": 1.7274156420189773e-05, "loss": 0.3179, "step": 34000 }, { "epoch": 0.4150725474626435, "grad_norm": 3.145047187805176, "learning_rate": 1.7234052792415602e-05, "loss": 0.306, "step": 34500 }, { "epoch": 0.42108809162876876, "grad_norm": 4.087808609008789, "learning_rate": 1.7193949164641435e-05, "loss": 0.325, "step": 35000 }, { "epoch": 0.427103635794894, "grad_norm": 0.009332289919257164, "learning_rate": 1.7153845536867268e-05, "loss": 0.2979, "step": 35500 }, { "epoch": 0.4331191799610193, "grad_norm": 9.885307312011719, "learning_rate": 1.7113741909093097e-05, "loss": 0.3247, "step": 36000 }, { "epoch": 0.43913472412714455, "grad_norm": 0.0056028529070317745, "learning_rate": 1.707363828131893e-05, "loss": 0.2949, "step": 36500 }, { "epoch": 0.4451502682932698, "grad_norm": 8.444794654846191, "learning_rate": 1.7033534653544762e-05, "loss": 0.306, "step": 37000 }, { "epoch": 0.4511658124593951, "grad_norm": 2.0681827068328857, "learning_rate": 1.699343102577059e-05, "loss": 0.2792, "step": 37500 }, { "epoch": 0.45718135662552034, "grad_norm": 0.017593583092093468, "learning_rate": 1.695340760525197e-05, "loss": 0.3225, "step": 38000 }, { "epoch": 0.4631969007916456, "grad_norm": 3.3854148387908936, "learning_rate": 1.6913303977477804e-05, "loss": 0.2962, "step": 38500 }, { "epoch": 0.46921244495777087, "grad_norm": 0.4375062584877014, "learning_rate": 1.6873200349703637e-05, "loss": 0.3146, "step": 39000 }, { "epoch": 0.47522798912389613, "grad_norm": 0.24493420124053955, "learning_rate": 1.6833096721929466e-05, "loss": 0.2745, "step": 39500 }, { "epoch": 0.4812435332900214, "grad_norm": 5.30033540725708, "learning_rate": 1.6793153508666397e-05, "loss": 0.3242, "step": 40000 }, { "epoch": 0.48725907745614666, "grad_norm": 0.08991783857345581, "learning_rate": 1.6753049880892226e-05, "loss": 0.2889, "step": 40500 }, { "epoch": 0.4932746216222719, "grad_norm": 0.16233447194099426, "learning_rate": 1.671294625311806e-05, "loss": 0.3147, "step": 41000 }, { "epoch": 0.49929016578839724, "grad_norm": 21.49981117248535, "learning_rate": 1.6672842625343892e-05, "loss": 0.3097, "step": 41500 }, { "epoch": 0.5053057099545225, "grad_norm": 0.07683174312114716, "learning_rate": 1.663281920482527e-05, "loss": 0.3283, "step": 42000 }, { "epoch": 0.5113212541206478, "grad_norm": 5.969061374664307, "learning_rate": 1.65927155770511e-05, "loss": 0.2973, "step": 42500 }, { "epoch": 0.517336798286773, "grad_norm": 0.1165817528963089, "learning_rate": 1.6552611949276934e-05, "loss": 0.2865, "step": 43000 }, { "epoch": 0.5233523424528983, "grad_norm": 4.852444648742676, "learning_rate": 1.6512508321502766e-05, "loss": 0.2995, "step": 43500 }, { "epoch": 0.5293678866190236, "grad_norm": 0.3422677218914032, "learning_rate": 1.6472404693728596e-05, "loss": 0.3078, "step": 44000 }, { "epoch": 0.5353834307851488, "grad_norm": 0.02138070948421955, "learning_rate": 1.643230106595443e-05, "loss": 0.3041, "step": 44500 }, { "epoch": 0.5413989749512741, "grad_norm": 19.726158142089844, "learning_rate": 1.639219743818026e-05, "loss": 0.2747, "step": 45000 }, { "epoch": 0.5474145191173994, "grad_norm": 0.8494102358818054, "learning_rate": 1.635209381040609e-05, "loss": 0.3165, "step": 45500 }, { "epoch": 0.5534300632835246, "grad_norm": 69.28723907470703, "learning_rate": 1.6311990182631923e-05, "loss": 0.286, "step": 46000 }, { "epoch": 0.5594456074496499, "grad_norm": 5.112706661224365, "learning_rate": 1.6271886554857756e-05, "loss": 0.3103, "step": 46500 }, { "epoch": 0.5654611516157751, "grad_norm": 0.0373302698135376, "learning_rate": 1.6231782927083585e-05, "loss": 0.2966, "step": 47000 }, { "epoch": 0.5714766957819004, "grad_norm": 0.3029526472091675, "learning_rate": 1.6191679299309418e-05, "loss": 0.2785, "step": 47500 }, { "epoch": 0.5774922399480257, "grad_norm": 0.18373580276966095, "learning_rate": 1.6151655878790798e-05, "loss": 0.3328, "step": 48000 }, { "epoch": 0.5835077841141509, "grad_norm": 0.8737627267837524, "learning_rate": 1.611155225101663e-05, "loss": 0.2909, "step": 48500 }, { "epoch": 0.5895233282802762, "grad_norm": 2.7465710639953613, "learning_rate": 1.607144862324246e-05, "loss": 0.2991, "step": 49000 }, { "epoch": 0.5955388724464015, "grad_norm": 19.38477325439453, "learning_rate": 1.6031344995468292e-05, "loss": 0.2785, "step": 49500 }, { "epoch": 0.6015544166125267, "grad_norm": 0.013609528541564941, "learning_rate": 1.5991241367694125e-05, "loss": 0.3198, "step": 50000 }, { "epoch": 0.607569960778652, "grad_norm": 6.444379806518555, "learning_rate": 1.5951137739919954e-05, "loss": 0.2816, "step": 50500 }, { "epoch": 0.6135855049447773, "grad_norm": 6.260073661804199, "learning_rate": 1.5911114319401334e-05, "loss": 0.3193, "step": 51000 }, { "epoch": 0.6196010491109025, "grad_norm": 8.19072151184082, "learning_rate": 1.5871010691627167e-05, "loss": 0.3021, "step": 51500 }, { "epoch": 0.6256165932770278, "grad_norm": 5.27684211730957, "learning_rate": 1.5830907063853e-05, "loss": 0.2925, "step": 52000 }, { "epoch": 0.631632137443153, "grad_norm": 0.031686898320913315, "learning_rate": 1.579080343607883e-05, "loss": 0.3063, "step": 52500 }, { "epoch": 0.6376476816092784, "grad_norm": 0.15889908373355865, "learning_rate": 1.575078001556021e-05, "loss": 0.2893, "step": 53000 }, { "epoch": 0.6436632257754037, "grad_norm": 5.396053314208984, "learning_rate": 1.571067638778604e-05, "loss": 0.3009, "step": 53500 }, { "epoch": 0.649678769941529, "grad_norm": 0.15985211730003357, "learning_rate": 1.5670572760011874e-05, "loss": 0.2985, "step": 54000 }, { "epoch": 0.6556943141076542, "grad_norm": 2.195845127105713, "learning_rate": 1.5630469132237704e-05, "loss": 0.2886, "step": 54500 }, { "epoch": 0.6617098582737795, "grad_norm": 0.000236680411035195, "learning_rate": 1.5590365504463536e-05, "loss": 0.3041, "step": 55000 }, { "epoch": 0.6677254024399047, "grad_norm": 0.012887202203273773, "learning_rate": 1.555026187668937e-05, "loss": 0.2848, "step": 55500 }, { "epoch": 0.67374094660603, "grad_norm": 0.025558780878782272, "learning_rate": 1.5510158248915198e-05, "loss": 0.3037, "step": 56000 }, { "epoch": 0.6797564907721553, "grad_norm": 13.303611755371094, "learning_rate": 1.547013482839658e-05, "loss": 0.3052, "step": 56500 }, { "epoch": 0.6857720349382805, "grad_norm": 7.728808403015137, "learning_rate": 1.543003120062241e-05, "loss": 0.3076, "step": 57000 }, { "epoch": 0.6917875791044058, "grad_norm": 7.381441593170166, "learning_rate": 1.538992757284824e-05, "loss": 0.2838, "step": 57500 }, { "epoch": 0.6978031232705311, "grad_norm": 47.043766021728516, "learning_rate": 1.5349823945074073e-05, "loss": 0.2894, "step": 58000 }, { "epoch": 0.7038186674366563, "grad_norm": 1.3084138631820679, "learning_rate": 1.5309720317299902e-05, "loss": 0.3089, "step": 58500 }, { "epoch": 0.7098342116027816, "grad_norm": 0.28533193469047546, "learning_rate": 1.5269616689525735e-05, "loss": 0.3291, "step": 59000 }, { "epoch": 0.7158497557689069, "grad_norm": 0.006308581214398146, "learning_rate": 1.5229513061751566e-05, "loss": 0.2842, "step": 59500 }, { "epoch": 0.7218652999350321, "grad_norm": 0.18038663268089294, "learning_rate": 1.5189409433977399e-05, "loss": 0.2964, "step": 60000 }, { "epoch": 0.7278808441011574, "grad_norm": 9.797381401062012, "learning_rate": 1.514930580620323e-05, "loss": 0.3303, "step": 60500 }, { "epoch": 0.7338963882672827, "grad_norm": 0.27853924036026, "learning_rate": 1.510920217842906e-05, "loss": 0.2762, "step": 61000 }, { "epoch": 0.7399119324334079, "grad_norm": 0.13307029008865356, "learning_rate": 1.506917875791044e-05, "loss": 0.2798, "step": 61500 }, { "epoch": 0.7459274765995332, "grad_norm": 4.89689826965332, "learning_rate": 1.5029075130136273e-05, "loss": 0.3202, "step": 62000 }, { "epoch": 0.7519430207656584, "grad_norm": 14.662346839904785, "learning_rate": 1.4988971502362104e-05, "loss": 0.3051, "step": 62500 }, { "epoch": 0.7579585649317837, "grad_norm": 0.15676206350326538, "learning_rate": 1.4948867874587935e-05, "loss": 0.3065, "step": 63000 }, { "epoch": 0.763974109097909, "grad_norm": 10.646651268005371, "learning_rate": 1.4908764246813768e-05, "loss": 0.2741, "step": 63500 }, { "epoch": 0.7699896532640342, "grad_norm": 7.507605075836182, "learning_rate": 1.4868740826295148e-05, "loss": 0.3195, "step": 64000 }, { "epoch": 0.7760051974301595, "grad_norm": 0.24443697929382324, "learning_rate": 1.4828717405776526e-05, "loss": 0.2662, "step": 64500 }, { "epoch": 0.7820207415962848, "grad_norm": 7.655556678771973, "learning_rate": 1.4788613778002359e-05, "loss": 0.3168, "step": 65000 }, { "epoch": 0.78803628576241, "grad_norm": 7.646127223968506, "learning_rate": 1.474851015022819e-05, "loss": 0.2803, "step": 65500 }, { "epoch": 0.7940518299285353, "grad_norm": 0.0017321036430075765, "learning_rate": 1.4708406522454023e-05, "loss": 0.2779, "step": 66000 }, { "epoch": 0.8000673740946606, "grad_norm": 0.004854683298617601, "learning_rate": 1.4668302894679854e-05, "loss": 0.3048, "step": 66500 }, { "epoch": 0.8060829182607858, "grad_norm": 5.331120491027832, "learning_rate": 1.4628199266905685e-05, "loss": 0.3121, "step": 67000 }, { "epoch": 0.8120984624269111, "grad_norm": 4.585981369018555, "learning_rate": 1.4588095639131517e-05, "loss": 0.2968, "step": 67500 }, { "epoch": 0.8181140065930365, "grad_norm": 0.0019534584134817123, "learning_rate": 1.4548072218612896e-05, "loss": 0.2869, "step": 68000 }, { "epoch": 0.8241295507591617, "grad_norm": 0.03927936032414436, "learning_rate": 1.4507968590838728e-05, "loss": 0.2942, "step": 68500 }, { "epoch": 0.830145094925287, "grad_norm": 0.018528884276747704, "learning_rate": 1.446786496306456e-05, "loss": 0.2929, "step": 69000 }, { "epoch": 0.8361606390914122, "grad_norm": 2.7401936054229736, "learning_rate": 1.4427761335290392e-05, "loss": 0.272, "step": 69500 }, { "epoch": 0.8421761832575375, "grad_norm": 0.07152987271547318, "learning_rate": 1.4387657707516223e-05, "loss": 0.3162, "step": 70000 }, { "epoch": 0.8481917274236628, "grad_norm": 5.40221643447876, "learning_rate": 1.4347634286997603e-05, "loss": 0.2717, "step": 70500 }, { "epoch": 0.854207271589788, "grad_norm": 2.5662572383880615, "learning_rate": 1.4307530659223434e-05, "loss": 0.2913, "step": 71000 }, { "epoch": 0.8602228157559133, "grad_norm": 6.399110794067383, "learning_rate": 1.4267427031449267e-05, "loss": 0.279, "step": 71500 }, { "epoch": 0.8662383599220386, "grad_norm": 4.513882160186768, "learning_rate": 1.4227323403675098e-05, "loss": 0.3091, "step": 72000 }, { "epoch": 0.8722539040881638, "grad_norm": 2.024488925933838, "learning_rate": 1.4187219775900929e-05, "loss": 0.2831, "step": 72500 }, { "epoch": 0.8782694482542891, "grad_norm": 9.901945114135742, "learning_rate": 1.4147116148126761e-05, "loss": 0.2863, "step": 73000 }, { "epoch": 0.8842849924204144, "grad_norm": 1.1856168508529663, "learning_rate": 1.410709272760814e-05, "loss": 0.2754, "step": 73500 }, { "epoch": 0.8903005365865396, "grad_norm": 4.077794075012207, "learning_rate": 1.4066989099833972e-05, "loss": 0.298, "step": 74000 }, { "epoch": 0.8963160807526649, "grad_norm": 0.2611056864261627, "learning_rate": 1.4026885472059803e-05, "loss": 0.2985, "step": 74500 }, { "epoch": 0.9023316249187902, "grad_norm": 0.1290031224489212, "learning_rate": 1.3986781844285636e-05, "loss": 0.2798, "step": 75000 }, { "epoch": 0.9083471690849154, "grad_norm": 0.1582382321357727, "learning_rate": 1.3946678216511467e-05, "loss": 0.2908, "step": 75500 }, { "epoch": 0.9143627132510407, "grad_norm": 6.692793369293213, "learning_rate": 1.3906574588737298e-05, "loss": 0.2905, "step": 76000 }, { "epoch": 0.920378257417166, "grad_norm": 12.920578002929688, "learning_rate": 1.386647096096313e-05, "loss": 0.306, "step": 76500 }, { "epoch": 0.9263938015832912, "grad_norm": 0.14715702831745148, "learning_rate": 1.3826367333188961e-05, "loss": 0.2911, "step": 77000 }, { "epoch": 0.9324093457494165, "grad_norm": 7.250746726989746, "learning_rate": 1.3786343912670342e-05, "loss": 0.2911, "step": 77500 }, { "epoch": 0.9384248899155417, "grad_norm": 9.904165267944336, "learning_rate": 1.3746240284896173e-05, "loss": 0.2893, "step": 78000 }, { "epoch": 0.944440434081667, "grad_norm": 1.3197027444839478, "learning_rate": 1.3706136657122005e-05, "loss": 0.295, "step": 78500 }, { "epoch": 0.9504559782477923, "grad_norm": 0.5488787889480591, "learning_rate": 1.3666033029347836e-05, "loss": 0.2617, "step": 79000 }, { "epoch": 0.9564715224139175, "grad_norm": 0.007599683478474617, "learning_rate": 1.3625929401573667e-05, "loss": 0.2874, "step": 79500 }, { "epoch": 0.9624870665800428, "grad_norm": 28.689912796020508, "learning_rate": 1.35858257737995e-05, "loss": 0.2954, "step": 80000 }, { "epoch": 0.9685026107461681, "grad_norm": 1.1681454181671143, "learning_rate": 1.354572214602533e-05, "loss": 0.2893, "step": 80500 }, { "epoch": 0.9745181549122933, "grad_norm": 26.251056671142578, "learning_rate": 1.3505618518251162e-05, "loss": 0.2771, "step": 81000 }, { "epoch": 0.9805336990784186, "grad_norm": 19.04422950744629, "learning_rate": 1.3465514890476994e-05, "loss": 0.2556, "step": 81500 }, { "epoch": 0.9865492432445438, "grad_norm": 6.7204909324646, "learning_rate": 1.3425411262702825e-05, "loss": 0.2999, "step": 82000 }, { "epoch": 0.9925647874106691, "grad_norm": 0.3453606069087982, "learning_rate": 1.3385307634928656e-05, "loss": 0.2795, "step": 82500 }, { "epoch": 0.9985803315767945, "grad_norm": 2.7078826427459717, "learning_rate": 1.3345204007154489e-05, "loss": 0.2727, "step": 83000 }, { "epoch": 1.0, "eval_accuracy": 0.6738478320152713, "eval_f1": 0.69947127247092, "eval_loss": 0.287530779838562, "eval_roc_auc": 0.8124488682847013, "eval_runtime": 111.0668, "eval_samples_per_second": 66.032, "eval_steps_per_second": 66.032, "step": 83118 }, { "epoch": 1.0045958757429196, "grad_norm": 8.754840850830078, "learning_rate": 1.3305180586635869e-05, "loss": 0.2439, "step": 83500 }, { "epoch": 1.010611419909045, "grad_norm": 23.18345069885254, "learning_rate": 1.3265237373372797e-05, "loss": 0.2343, "step": 84000 }, { "epoch": 1.0166269640751702, "grad_norm": 0.2795548141002655, "learning_rate": 1.322513374559863e-05, "loss": 0.2591, "step": 84500 }, { "epoch": 1.0226425082412955, "grad_norm": 23.055418014526367, "learning_rate": 1.318503011782446e-05, "loss": 0.2403, "step": 85000 }, { "epoch": 1.0286580524074207, "grad_norm": 0.0004251549835316837, "learning_rate": 1.3144926490050291e-05, "loss": 0.2575, "step": 85500 }, { "epoch": 1.034673596573546, "grad_norm": 0.0705941691994667, "learning_rate": 1.3104822862276124e-05, "loss": 0.2298, "step": 86000 }, { "epoch": 1.0406891407396712, "grad_norm": 0.14872965216636658, "learning_rate": 1.3064719234501955e-05, "loss": 0.2556, "step": 86500 }, { "epoch": 1.0467046849057966, "grad_norm": 29.70933723449707, "learning_rate": 1.3024615606727786e-05, "loss": 0.246, "step": 87000 }, { "epoch": 1.0527202290719218, "grad_norm": 0.12189821898937225, "learning_rate": 1.2984511978953618e-05, "loss": 0.2486, "step": 87500 }, { "epoch": 1.0587357732380471, "grad_norm": 16.819997787475586, "learning_rate": 1.2944488558434999e-05, "loss": 0.2738, "step": 88000 }, { "epoch": 1.0647513174041723, "grad_norm": 0.504358172416687, "learning_rate": 1.290438493066083e-05, "loss": 0.2512, "step": 88500 }, { "epoch": 1.0707668615702977, "grad_norm": 33.048221588134766, "learning_rate": 1.286428130288666e-05, "loss": 0.2464, "step": 89000 }, { "epoch": 1.0767824057364228, "grad_norm": 10.764643669128418, "learning_rate": 1.2824177675112493e-05, "loss": 0.2633, "step": 89500 }, { "epoch": 1.0827979499025482, "grad_norm": 0.08640070259571075, "learning_rate": 1.2784074047338324e-05, "loss": 0.2406, "step": 90000 }, { "epoch": 1.0888134940686736, "grad_norm": 14.491482734680176, "learning_rate": 1.2744050626819704e-05, "loss": 0.2914, "step": 90500 }, { "epoch": 1.0948290382347987, "grad_norm": 6.458908557891846, "learning_rate": 1.2703946999045535e-05, "loss": 0.2768, "step": 91000 }, { "epoch": 1.100844582400924, "grad_norm": 3.36252498626709, "learning_rate": 1.2663923578526915e-05, "loss": 0.2275, "step": 91500 }, { "epoch": 1.1068601265670492, "grad_norm": 7.332001209259033, "learning_rate": 1.2623819950752746e-05, "loss": 0.2561, "step": 92000 }, { "epoch": 1.1128756707331746, "grad_norm": 0.10880118608474731, "learning_rate": 1.2583796530234126e-05, "loss": 0.26, "step": 92500 }, { "epoch": 1.1188912148992998, "grad_norm": 0.041975826025009155, "learning_rate": 1.2543692902459959e-05, "loss": 0.2588, "step": 93000 }, { "epoch": 1.1249067590654251, "grad_norm": 19.071786880493164, "learning_rate": 1.250358927468579e-05, "loss": 0.2537, "step": 93500 }, { "epoch": 1.1309223032315503, "grad_norm": 0.5620952248573303, "learning_rate": 1.2463485646911621e-05, "loss": 0.2694, "step": 94000 }, { "epoch": 1.1369378473976757, "grad_norm": 0.4685472548007965, "learning_rate": 1.2423382019137454e-05, "loss": 0.2255, "step": 94500 }, { "epoch": 1.1429533915638008, "grad_norm": 0.004374380223453045, "learning_rate": 1.2383278391363285e-05, "loss": 0.2979, "step": 95000 }, { "epoch": 1.1489689357299262, "grad_norm": 0.21592481434345245, "learning_rate": 1.2343174763589117e-05, "loss": 0.2425, "step": 95500 }, { "epoch": 1.1549844798960514, "grad_norm": 38.64509963989258, "learning_rate": 1.2303071135814946e-05, "loss": 0.2534, "step": 96000 }, { "epoch": 1.1610000240621767, "grad_norm": 0.1709357500076294, "learning_rate": 1.2262967508040777e-05, "loss": 0.2277, "step": 96500 }, { "epoch": 1.1670155682283019, "grad_norm": 0.02226692996919155, "learning_rate": 1.2222863880266608e-05, "loss": 0.248, "step": 97000 }, { "epoch": 1.1730311123944273, "grad_norm": 0.11720846593379974, "learning_rate": 1.2182760252492441e-05, "loss": 0.2818, "step": 97500 }, { "epoch": 1.1790466565605524, "grad_norm": 0.17743727564811707, "learning_rate": 1.2142656624718272e-05, "loss": 0.2625, "step": 98000 }, { "epoch": 1.1850622007266778, "grad_norm": 0.5575046539306641, "learning_rate": 1.2102633204199652e-05, "loss": 0.274, "step": 98500 }, { "epoch": 1.191077744892803, "grad_norm": 0.0813954770565033, "learning_rate": 1.2062529576425483e-05, "loss": 0.2764, "step": 99000 }, { "epoch": 1.1970932890589283, "grad_norm": 6.563291549682617, "learning_rate": 1.2022425948651316e-05, "loss": 0.2426, "step": 99500 }, { "epoch": 1.2031088332250535, "grad_norm": 4.957773126079701e-05, "learning_rate": 1.1982322320877147e-05, "loss": 0.2558, "step": 100000 }, { "epoch": 1.2091243773911788, "grad_norm": 0.08407856523990631, "learning_rate": 1.1942218693102978e-05, "loss": 0.244, "step": 100500 }, { "epoch": 1.215139921557304, "grad_norm": 0.04238196834921837, "learning_rate": 1.1902195272584358e-05, "loss": 0.2635, "step": 101000 }, { "epoch": 1.2211554657234294, "grad_norm": 0.10858649015426636, "learning_rate": 1.186209164481019e-05, "loss": 0.2536, "step": 101500 }, { "epoch": 1.2271710098895545, "grad_norm": 0.14419694244861603, "learning_rate": 1.1821988017036021e-05, "loss": 0.2803, "step": 102000 }, { "epoch": 1.23318655405568, "grad_norm": 0.1416100561618805, "learning_rate": 1.1781964596517402e-05, "loss": 0.2198, "step": 102500 }, { "epoch": 1.239202098221805, "grad_norm": 0.16191110014915466, "learning_rate": 1.1741860968743232e-05, "loss": 0.2687, "step": 103000 }, { "epoch": 1.2452176423879304, "grad_norm": 3.2424659729003906, "learning_rate": 1.1701757340969065e-05, "loss": 0.2531, "step": 103500 }, { "epoch": 1.2512331865540558, "grad_norm": 0.009034248068928719, "learning_rate": 1.1661653713194896e-05, "loss": 0.2144, "step": 104000 }, { "epoch": 1.257248730720181, "grad_norm": 4.68690299987793, "learning_rate": 1.1621550085420727e-05, "loss": 0.2536, "step": 104500 }, { "epoch": 1.263264274886306, "grad_norm": 0.04760544002056122, "learning_rate": 1.158144645764656e-05, "loss": 0.2394, "step": 105000 }, { "epoch": 1.2692798190524315, "grad_norm": 0.027170976623892784, "learning_rate": 1.154134282987239e-05, "loss": 0.2644, "step": 105500 }, { "epoch": 1.2752953632185569, "grad_norm": 29.608776092529297, "learning_rate": 1.1501239202098222e-05, "loss": 0.2456, "step": 106000 }, { "epoch": 1.281310907384682, "grad_norm": 3.261889934539795, "learning_rate": 1.1461135574324054e-05, "loss": 0.2182, "step": 106500 }, { "epoch": 1.2873264515508072, "grad_norm": 0.0033520832657814026, "learning_rate": 1.1421031946549885e-05, "loss": 0.2612, "step": 107000 }, { "epoch": 1.2933419957169325, "grad_norm": 0.0770372673869133, "learning_rate": 1.1380928318775716e-05, "loss": 0.2517, "step": 107500 }, { "epoch": 1.299357539883058, "grad_norm": 0.004508438985794783, "learning_rate": 1.1340904898257096e-05, "loss": 0.2552, "step": 108000 }, { "epoch": 1.305373084049183, "grad_norm": 0.05816527456045151, "learning_rate": 1.1300801270482929e-05, "loss": 0.2577, "step": 108500 }, { "epoch": 1.3113886282153084, "grad_norm": 8.914427757263184, "learning_rate": 1.126069764270876e-05, "loss": 0.2177, "step": 109000 }, { "epoch": 1.3174041723814336, "grad_norm": 0.0006245987024158239, "learning_rate": 1.1220594014934591e-05, "loss": 0.2321, "step": 109500 }, { "epoch": 1.323419716547559, "grad_norm": 10.717325210571289, "learning_rate": 1.1180490387160424e-05, "loss": 0.2532, "step": 110000 }, { "epoch": 1.3294352607136841, "grad_norm": 3.988328218460083, "learning_rate": 1.1140386759386255e-05, "loss": 0.2733, "step": 110500 }, { "epoch": 1.3354508048798095, "grad_norm": 9.221200942993164, "learning_rate": 1.1100363338867635e-05, "loss": 0.2611, "step": 111000 }, { "epoch": 1.3414663490459346, "grad_norm": 0.08797607570886612, "learning_rate": 1.1060259711093466e-05, "loss": 0.2608, "step": 111500 }, { "epoch": 1.34748189321206, "grad_norm": 0.018298862501978874, "learning_rate": 1.1020156083319298e-05, "loss": 0.2579, "step": 112000 }, { "epoch": 1.3534974373781852, "grad_norm": 0.033742956817150116, "learning_rate": 1.098005245554513e-05, "loss": 0.2431, "step": 112500 }, { "epoch": 1.3595129815443106, "grad_norm": 0.06899914145469666, "learning_rate": 1.093994882777096e-05, "loss": 0.253, "step": 113000 }, { "epoch": 1.3655285257104357, "grad_norm": 0.1469135880470276, "learning_rate": 1.0899845199996793e-05, "loss": 0.244, "step": 113500 }, { "epoch": 1.371544069876561, "grad_norm": 5.18328857421875, "learning_rate": 1.0859821779478173e-05, "loss": 0.2654, "step": 114000 }, { "epoch": 1.3775596140426862, "grad_norm": 24.231332778930664, "learning_rate": 1.0819718151704004e-05, "loss": 0.237, "step": 114500 }, { "epoch": 1.3835751582088116, "grad_norm": 0.2826213240623474, "learning_rate": 1.0779614523929835e-05, "loss": 0.2602, "step": 115000 }, { "epoch": 1.3895907023749368, "grad_norm": 0.09144386649131775, "learning_rate": 1.0739510896155668e-05, "loss": 0.249, "step": 115500 }, { "epoch": 1.3956062465410621, "grad_norm": 40.423797607421875, "learning_rate": 1.0699407268381499e-05, "loss": 0.2621, "step": 116000 }, { "epoch": 1.4016217907071873, "grad_norm": 0.01249714009463787, "learning_rate": 1.0659383847862879e-05, "loss": 0.2451, "step": 116500 }, { "epoch": 1.4076373348733127, "grad_norm": 0.14570143818855286, "learning_rate": 1.061928022008871e-05, "loss": 0.2481, "step": 117000 }, { "epoch": 1.413652879039438, "grad_norm": 8.342629432678223, "learning_rate": 1.0579176592314542e-05, "loss": 0.2813, "step": 117500 }, { "epoch": 1.4196684232055632, "grad_norm": 0.0001358153240289539, "learning_rate": 1.0539072964540373e-05, "loss": 0.2425, "step": 118000 }, { "epoch": 1.4256839673716883, "grad_norm": 0.2505891025066376, "learning_rate": 1.0498969336766204e-05, "loss": 0.243, "step": 118500 }, { "epoch": 1.4316995115378137, "grad_norm": 0.25133442878723145, "learning_rate": 1.0458865708992037e-05, "loss": 0.2313, "step": 119000 }, { "epoch": 1.437715055703939, "grad_norm": 0.00044232659274712205, "learning_rate": 1.0418762081217868e-05, "loss": 0.2426, "step": 119500 }, { "epoch": 1.4437305998700642, "grad_norm": 0.0011125517776235938, "learning_rate": 1.0378658453443699e-05, "loss": 0.2418, "step": 120000 }, { "epoch": 1.4497461440361894, "grad_norm": 7.526780605316162, "learning_rate": 1.0338635032925079e-05, "loss": 0.2797, "step": 120500 }, { "epoch": 1.4557616882023148, "grad_norm": 6.034137725830078, "learning_rate": 1.0298531405150912e-05, "loss": 0.2655, "step": 121000 }, { "epoch": 1.4617772323684401, "grad_norm": 0.021329816430807114, "learning_rate": 1.0258427777376743e-05, "loss": 0.2633, "step": 121500 }, { "epoch": 1.4677927765345653, "grad_norm": 0.013099109753966331, "learning_rate": 1.0218324149602574e-05, "loss": 0.2739, "step": 122000 }, { "epoch": 1.4738083207006905, "grad_norm": 71.02497863769531, "learning_rate": 1.0178300729083954e-05, "loss": 0.254, "step": 122500 }, { "epoch": 1.4798238648668158, "grad_norm": 15.551078796386719, "learning_rate": 1.0138197101309786e-05, "loss": 0.2602, "step": 123000 }, { "epoch": 1.4858394090329412, "grad_norm": 4.560909655992873e-05, "learning_rate": 1.0098093473535617e-05, "loss": 0.2387, "step": 123500 }, { "epoch": 1.4918549531990664, "grad_norm": 0.14415931701660156, "learning_rate": 1.0057989845761448e-05, "loss": 0.2823, "step": 124000 }, { "epoch": 1.4978704973651917, "grad_norm": 43.61802673339844, "learning_rate": 1.0017886217987281e-05, "loss": 0.2378, "step": 124500 }, { "epoch": 1.5038860415313169, "grad_norm": 0.20755548775196075, "learning_rate": 9.977782590213112e-06, "loss": 0.2433, "step": 125000 }, { "epoch": 1.5099015856974423, "grad_norm": 0.2520395815372467, "learning_rate": 9.937678962438943e-06, "loss": 0.2539, "step": 125500 }, { "epoch": 1.5159171298635674, "grad_norm": 0.3743610382080078, "learning_rate": 9.897575334664774e-06, "loss": 0.2789, "step": 126000 }, { "epoch": 1.5219326740296926, "grad_norm": 0.0024790179450064898, "learning_rate": 9.857551914146154e-06, "loss": 0.2258, "step": 126500 }, { "epoch": 1.527948218195818, "grad_norm": 21.557737350463867, "learning_rate": 9.817448286371985e-06, "loss": 0.2692, "step": 127000 }, { "epoch": 1.5339637623619433, "grad_norm": 0.00022073234140407294, "learning_rate": 9.777344658597818e-06, "loss": 0.2674, "step": 127500 }, { "epoch": 1.5399793065280685, "grad_norm": 0.0013155222404748201, "learning_rate": 9.737241030823649e-06, "loss": 0.242, "step": 128000 }, { "epoch": 1.5459948506941938, "grad_norm": 0.2000940591096878, "learning_rate": 9.69713740304948e-06, "loss": 0.2487, "step": 128500 }, { "epoch": 1.5520103948603192, "grad_norm": 0.0021518643479794264, "learning_rate": 9.657033775275312e-06, "loss": 0.2123, "step": 129000 }, { "epoch": 1.5580259390264444, "grad_norm": 0.24494774639606476, "learning_rate": 9.616930147501143e-06, "loss": 0.2975, "step": 129500 }, { "epoch": 1.5640414831925695, "grad_norm": 9.558631896972656, "learning_rate": 9.576906726982523e-06, "loss": 0.2485, "step": 130000 }, { "epoch": 1.570057027358695, "grad_norm": 6.8279128074646, "learning_rate": 9.536803099208354e-06, "loss": 0.2589, "step": 130500 }, { "epoch": 1.5760725715248203, "grad_norm": 0.018290262669324875, "learning_rate": 9.496699471434187e-06, "loss": 0.2605, "step": 131000 }, { "epoch": 1.5820881156909454, "grad_norm": 0.19554129242897034, "learning_rate": 9.456595843660018e-06, "loss": 0.2405, "step": 131500 }, { "epoch": 1.5881036598570706, "grad_norm": 26.12983512878418, "learning_rate": 9.416492215885849e-06, "loss": 0.2261, "step": 132000 }, { "epoch": 1.594119204023196, "grad_norm": 19.12807273864746, "learning_rate": 9.376388588111682e-06, "loss": 0.2454, "step": 132500 }, { "epoch": 1.6001347481893213, "grad_norm": 48.97768020629883, "learning_rate": 9.336284960337513e-06, "loss": 0.2488, "step": 133000 }, { "epoch": 1.6061502923554465, "grad_norm": 0.10775639116764069, "learning_rate": 9.296261539818893e-06, "loss": 0.2712, "step": 133500 }, { "epoch": 1.6121658365215716, "grad_norm": 18.01923942565918, "learning_rate": 9.256157912044724e-06, "loss": 0.2898, "step": 134000 }, { "epoch": 1.618181380687697, "grad_norm": 0.057373084127902985, "learning_rate": 9.216054284270556e-06, "loss": 0.2235, "step": 134500 }, { "epoch": 1.6241969248538224, "grad_norm": 0.0020849681459367275, "learning_rate": 9.175950656496387e-06, "loss": 0.2355, "step": 135000 }, { "epoch": 1.6302124690199475, "grad_norm": 0.15205715596675873, "learning_rate": 9.135847028722218e-06, "loss": 0.2791, "step": 135500 }, { "epoch": 1.6362280131860727, "grad_norm": 0.06727798283100128, "learning_rate": 9.095823608203598e-06, "loss": 0.2412, "step": 136000 }, { "epoch": 1.642243557352198, "grad_norm": 0.19422870874404907, "learning_rate": 9.055719980429431e-06, "loss": 0.2431, "step": 136500 }, { "epoch": 1.6482591015183234, "grad_norm": 0.06913313269615173, "learning_rate": 9.015616352655262e-06, "loss": 0.2687, "step": 137000 }, { "epoch": 1.6542746456844486, "grad_norm": 0.06762461364269257, "learning_rate": 8.975512724881093e-06, "loss": 0.2498, "step": 137500 }, { "epoch": 1.6602901898505737, "grad_norm": 0.1282985806465149, "learning_rate": 8.935409097106926e-06, "loss": 0.2503, "step": 138000 }, { "epoch": 1.6663057340166991, "grad_norm": 0.6339439153671265, "learning_rate": 8.895305469332756e-06, "loss": 0.2454, "step": 138500 }, { "epoch": 1.6723212781828245, "grad_norm": 0.15219150483608246, "learning_rate": 8.855201841558587e-06, "loss": 0.2845, "step": 139000 }, { "epoch": 1.6783368223489497, "grad_norm": 5.600707054138184, "learning_rate": 8.815178421039968e-06, "loss": 0.2459, "step": 139500 }, { "epoch": 1.6843523665150748, "grad_norm": 0.36313244700431824, "learning_rate": 8.775155000521348e-06, "loss": 0.2374, "step": 140000 }, { "epoch": 1.6903679106812002, "grad_norm": 4.401086807250977, "learning_rate": 8.73505137274718e-06, "loss": 0.2425, "step": 140500 }, { "epoch": 1.6963834548473256, "grad_norm": 0.00033720143255777657, "learning_rate": 8.694947744973011e-06, "loss": 0.2341, "step": 141000 }, { "epoch": 1.7023989990134507, "grad_norm": 0.0821937769651413, "learning_rate": 8.654844117198842e-06, "loss": 0.2326, "step": 141500 }, { "epoch": 1.7084145431795759, "grad_norm": 0.000721343734767288, "learning_rate": 8.614740489424675e-06, "loss": 0.2553, "step": 142000 }, { "epoch": 1.7144300873457012, "grad_norm": 0.3059915602207184, "learning_rate": 8.574636861650506e-06, "loss": 0.2743, "step": 142500 }, { "epoch": 1.7204456315118266, "grad_norm": 0.13650359213352203, "learning_rate": 8.534533233876337e-06, "loss": 0.255, "step": 143000 }, { "epoch": 1.7264611756779518, "grad_norm": 0.16923360526561737, "learning_rate": 8.49442960610217e-06, "loss": 0.2832, "step": 143500 }, { "epoch": 1.7324767198440771, "grad_norm": 0.0067064897157251835, "learning_rate": 8.454325978328e-06, "loss": 0.2425, "step": 144000 }, { "epoch": 1.7384922640102025, "grad_norm": 0.33390843868255615, "learning_rate": 8.41430255780938e-06, "loss": 0.2577, "step": 144500 }, { "epoch": 1.7445078081763277, "grad_norm": 0.11052338778972626, "learning_rate": 8.374198930035212e-06, "loss": 0.253, "step": 145000 }, { "epoch": 1.7505233523424528, "grad_norm": 0.06620516628026962, "learning_rate": 8.334095302261042e-06, "loss": 0.2555, "step": 145500 }, { "epoch": 1.7565388965085782, "grad_norm": 0.00559147447347641, "learning_rate": 8.293991674486873e-06, "loss": 0.2387, "step": 146000 }, { "epoch": 1.7625544406747036, "grad_norm": 0.1951608508825302, "learning_rate": 8.253888046712706e-06, "loss": 0.2249, "step": 146500 }, { "epoch": 1.7685699848408287, "grad_norm": 0.16532181203365326, "learning_rate": 8.213784418938537e-06, "loss": 0.2779, "step": 147000 }, { "epoch": 1.7745855290069539, "grad_norm": 0.024670429527759552, "learning_rate": 8.17368079116437e-06, "loss": 0.24, "step": 147500 }, { "epoch": 1.7806010731730793, "grad_norm": 0.7060034275054932, "learning_rate": 8.1335771633902e-06, "loss": 0.2646, "step": 148000 }, { "epoch": 1.7866166173392046, "grad_norm": 0.16653361916542053, "learning_rate": 8.093473535616032e-06, "loss": 0.2829, "step": 148500 }, { "epoch": 1.7926321615053298, "grad_norm": 9.015371322631836, "learning_rate": 8.053369907841864e-06, "loss": 0.2465, "step": 149000 }, { "epoch": 1.798647705671455, "grad_norm": 0.08503203839063644, "learning_rate": 8.013346487323243e-06, "loss": 0.2476, "step": 149500 }, { "epoch": 1.8046632498375803, "grad_norm": 0.31652989983558655, "learning_rate": 7.973242859549075e-06, "loss": 0.2807, "step": 150000 }, { "epoch": 1.8106787940037057, "grad_norm": 20.23206329345703, "learning_rate": 7.933139231774906e-06, "loss": 0.2393, "step": 150500 }, { "epoch": 1.8166943381698308, "grad_norm": 0.0023623614106327295, "learning_rate": 7.893035604000739e-06, "loss": 0.2606, "step": 151000 }, { "epoch": 1.822709882335956, "grad_norm": 0.0003463963221292943, "learning_rate": 7.85293197622657e-06, "loss": 0.2377, "step": 151500 }, { "epoch": 1.8287254265020814, "grad_norm": 22.177343368530273, "learning_rate": 7.81290855570795e-06, "loss": 0.225, "step": 152000 }, { "epoch": 1.8347409706682067, "grad_norm": 0.16396382451057434, "learning_rate": 7.772804927933781e-06, "loss": 0.2749, "step": 152500 }, { "epoch": 1.840756514834332, "grad_norm": 0.06666311621665955, "learning_rate": 7.732701300159614e-06, "loss": 0.2449, "step": 153000 }, { "epoch": 1.846772059000457, "grad_norm": 0.2808719575405121, "learning_rate": 7.692597672385445e-06, "loss": 0.2201, "step": 153500 }, { "epoch": 1.8527876031665824, "grad_norm": 0.02584654837846756, "learning_rate": 7.652574251866825e-06, "loss": 0.2555, "step": 154000 }, { "epoch": 1.8588031473327078, "grad_norm": 6.074542999267578, "learning_rate": 7.612470624092656e-06, "loss": 0.2652, "step": 154500 }, { "epoch": 1.864818691498833, "grad_norm": 0.007166345603764057, "learning_rate": 7.572447203574036e-06, "loss": 0.2777, "step": 155000 }, { "epoch": 1.870834235664958, "grad_norm": 2.7224321365356445, "learning_rate": 7.532423783055416e-06, "loss": 0.2614, "step": 155500 }, { "epoch": 1.8768497798310835, "grad_norm": 8.61175537109375, "learning_rate": 7.492320155281248e-06, "loss": 0.2716, "step": 156000 }, { "epoch": 1.8828653239972089, "grad_norm": 0.4761998951435089, "learning_rate": 7.452216527507079e-06, "loss": 0.2617, "step": 156500 }, { "epoch": 1.888880868163334, "grad_norm": 0.041520122438669205, "learning_rate": 7.4121128997329105e-06, "loss": 0.2233, "step": 157000 }, { "epoch": 1.8948964123294592, "grad_norm": 0.3298840820789337, "learning_rate": 7.372009271958742e-06, "loss": 0.2905, "step": 157500 }, { "epoch": 1.9009119564955848, "grad_norm": 0.14137232303619385, "learning_rate": 7.331905644184573e-06, "loss": 0.2489, "step": 158000 }, { "epoch": 1.90692750066171, "grad_norm": 0.017581721767783165, "learning_rate": 7.291802016410405e-06, "loss": 0.2531, "step": 158500 }, { "epoch": 1.912943044827835, "grad_norm": 0.10974815487861633, "learning_rate": 7.251698388636237e-06, "loss": 0.2387, "step": 159000 }, { "epoch": 1.9189585889939604, "grad_norm": 0.09865211695432663, "learning_rate": 7.211674968117617e-06, "loss": 0.2364, "step": 159500 }, { "epoch": 1.9249741331600858, "grad_norm": 0.004849567078053951, "learning_rate": 7.171571340343448e-06, "loss": 0.2751, "step": 160000 }, { "epoch": 1.930989677326211, "grad_norm": 0.017855796962976456, "learning_rate": 7.13146771256928e-06, "loss": 0.2422, "step": 160500 }, { "epoch": 1.9370052214923361, "grad_norm": 0.22124846279621124, "learning_rate": 7.091364084795112e-06, "loss": 0.2409, "step": 161000 }, { "epoch": 1.9430207656584615, "grad_norm": 0.0023982899729162455, "learning_rate": 7.051340664276492e-06, "loss": 0.2699, "step": 161500 }, { "epoch": 1.9490363098245869, "grad_norm": 8.00629997253418, "learning_rate": 7.011237036502323e-06, "loss": 0.2399, "step": 162000 }, { "epoch": 1.955051853990712, "grad_norm": 7.468501091003418, "learning_rate": 6.9711334087281545e-06, "loss": 0.2535, "step": 162500 }, { "epoch": 1.9610673981568372, "grad_norm": 0.4242821931838989, "learning_rate": 6.931029780953986e-06, "loss": 0.2506, "step": 163000 }, { "epoch": 1.9670829423229625, "grad_norm": 0.020416131243109703, "learning_rate": 6.890926153179817e-06, "loss": 0.2482, "step": 163500 }, { "epoch": 1.973098486489088, "grad_norm": 0.06718198210000992, "learning_rate": 6.850822525405649e-06, "loss": 0.2505, "step": 164000 }, { "epoch": 1.979114030655213, "grad_norm": 0.7040701508522034, "learning_rate": 6.81071889763148e-06, "loss": 0.2196, "step": 164500 }, { "epoch": 1.9851295748213382, "grad_norm": 0.1061306744813919, "learning_rate": 6.770615269857311e-06, "loss": 0.2699, "step": 165000 }, { "epoch": 1.9911451189874636, "grad_norm": 0.1626858115196228, "learning_rate": 6.730511642083143e-06, "loss": 0.2428, "step": 165500 }, { "epoch": 1.997160663153589, "grad_norm": 0.9182321429252625, "learning_rate": 6.690408014308975e-06, "loss": 0.2723, "step": 166000 }, { "epoch": 2.0, "eval_accuracy": 0.6772566130351786, "eval_f1": 0.7059909142532652, "eval_loss": 0.30793967843055725, "eval_roc_auc": 0.8147532042541586, "eval_runtime": 109.0185, "eval_samples_per_second": 67.273, "eval_steps_per_second": 67.273, "step": 166236 }, { "epoch": 2.003176207319714, "grad_norm": 0.0028744570445269346, "learning_rate": 6.650304386534806e-06, "loss": 0.2161, "step": 166500 }, { "epoch": 2.0091917514858393, "grad_norm": 0.10508907586336136, "learning_rate": 6.6102007587606375e-06, "loss": 0.1802, "step": 167000 }, { "epoch": 2.015207295651965, "grad_norm": 0.023608388379216194, "learning_rate": 6.570257545497566e-06, "loss": 0.1646, "step": 167500 }, { "epoch": 2.02122283981809, "grad_norm": 0.0004214489890728146, "learning_rate": 6.530153917723398e-06, "loss": 0.1425, "step": 168000 }, { "epoch": 2.027238383984215, "grad_norm": 0.15282946825027466, "learning_rate": 6.490050289949229e-06, "loss": 0.1767, "step": 168500 }, { "epoch": 2.0332539281503403, "grad_norm": 0.5594785213470459, "learning_rate": 6.4499466621750605e-06, "loss": 0.1431, "step": 169000 }, { "epoch": 2.039269472316466, "grad_norm": 0.006397734861820936, "learning_rate": 6.409843034400892e-06, "loss": 0.1433, "step": 169500 }, { "epoch": 2.045285016482591, "grad_norm": 0.02048441767692566, "learning_rate": 6.369739406626724e-06, "loss": 0.1751, "step": 170000 }, { "epoch": 2.0513005606487162, "grad_norm": 0.08668813109397888, "learning_rate": 6.329635778852555e-06, "loss": 0.1904, "step": 170500 }, { "epoch": 2.0573161048148414, "grad_norm": 1.3654146194458008, "learning_rate": 6.289612358333935e-06, "loss": 0.1477, "step": 171000 }, { "epoch": 2.063331648980967, "grad_norm": 0.10657373815774918, "learning_rate": 6.249508730559767e-06, "loss": 0.1651, "step": 171500 }, { "epoch": 2.069347193147092, "grad_norm": 0.0666826069355011, "learning_rate": 6.209405102785599e-06, "loss": 0.1896, "step": 172000 }, { "epoch": 2.0753627373132173, "grad_norm": 0.03260161355137825, "learning_rate": 6.16930147501143e-06, "loss": 0.1595, "step": 172500 }, { "epoch": 2.0813782814793425, "grad_norm": 11.851826667785645, "learning_rate": 6.129197847237262e-06, "loss": 0.1664, "step": 173000 }, { "epoch": 2.087393825645468, "grad_norm": 0.0001777495490387082, "learning_rate": 6.089094219463093e-06, "loss": 0.1719, "step": 173500 }, { "epoch": 2.093409369811593, "grad_norm": 0.001742048654705286, "learning_rate": 6.048990591688924e-06, "loss": 0.1336, "step": 174000 }, { "epoch": 2.0994249139777184, "grad_norm": 0.233998104929924, "learning_rate": 6.008886963914756e-06, "loss": 0.1659, "step": 174500 }, { "epoch": 2.1054404581438435, "grad_norm": 0.011589854955673218, "learning_rate": 5.968863543396136e-06, "loss": 0.1662, "step": 175000 }, { "epoch": 2.111456002309969, "grad_norm": 0.003179445629939437, "learning_rate": 5.928759915621968e-06, "loss": 0.201, "step": 175500 }, { "epoch": 2.1174715464760943, "grad_norm": 0.05272052437067032, "learning_rate": 5.888656287847799e-06, "loss": 0.1648, "step": 176000 }, { "epoch": 2.1234870906422194, "grad_norm": 17.65900421142578, "learning_rate": 5.848552660073631e-06, "loss": 0.1499, "step": 176500 }, { "epoch": 2.1295026348083446, "grad_norm": 0.017062697559595108, "learning_rate": 5.808449032299463e-06, "loss": 0.1812, "step": 177000 }, { "epoch": 2.13551817897447, "grad_norm": 0.062195923179388046, "learning_rate": 5.768505819036391e-06, "loss": 0.1745, "step": 177500 }, { "epoch": 2.1415337231405953, "grad_norm": 0.00937980879098177, "learning_rate": 5.728402191262222e-06, "loss": 0.138, "step": 178000 }, { "epoch": 2.1475492673067205, "grad_norm": 13.196198463439941, "learning_rate": 5.688298563488054e-06, "loss": 0.1717, "step": 178500 }, { "epoch": 2.1535648114728456, "grad_norm": 0.1440310925245285, "learning_rate": 5.648194935713886e-06, "loss": 0.1515, "step": 179000 }, { "epoch": 2.159580355638971, "grad_norm": 0.032506223767995834, "learning_rate": 5.608091307939717e-06, "loss": 0.1401, "step": 179500 }, { "epoch": 2.1655958998050964, "grad_norm": 6.104779458837584e-05, "learning_rate": 5.5679876801655484e-06, "loss": 0.1881, "step": 180000 }, { "epoch": 2.1716114439712215, "grad_norm": 0.04752574488520622, "learning_rate": 5.52788405239138e-06, "loss": 0.1603, "step": 180500 }, { "epoch": 2.177626988137347, "grad_norm": 0.005751811899244785, "learning_rate": 5.487780424617212e-06, "loss": 0.1983, "step": 181000 }, { "epoch": 2.1836425323034723, "grad_norm": 0.03546688333153725, "learning_rate": 5.447757004098591e-06, "loss": 0.1831, "step": 181500 }, { "epoch": 2.1896580764695974, "grad_norm": 0.21516333520412445, "learning_rate": 5.407653376324423e-06, "loss": 0.1708, "step": 182000 }, { "epoch": 2.1956736206357226, "grad_norm": 0.061349764466285706, "learning_rate": 5.367549748550255e-06, "loss": 0.1737, "step": 182500 }, { "epoch": 2.201689164801848, "grad_norm": 67.89110565185547, "learning_rate": 5.327446120776087e-06, "loss": 0.1465, "step": 183000 }, { "epoch": 2.2077047089679733, "grad_norm": 0.0042543611489236355, "learning_rate": 5.287342493001918e-06, "loss": 0.2034, "step": 183500 }, { "epoch": 2.2137202531340985, "grad_norm": 8.772153854370117, "learning_rate": 5.247238865227749e-06, "loss": 0.1354, "step": 184000 }, { "epoch": 2.2197357973002236, "grad_norm": 0.46139106154441833, "learning_rate": 5.207215444709128e-06, "loss": 0.1838, "step": 184500 }, { "epoch": 2.2257513414663492, "grad_norm": 0.007118485402315855, "learning_rate": 5.16711181693496e-06, "loss": 0.1702, "step": 185000 }, { "epoch": 2.2317668856324744, "grad_norm": 0.01587226428091526, "learning_rate": 5.127008189160792e-06, "loss": 0.1566, "step": 185500 }, { "epoch": 2.2377824297985995, "grad_norm": 0.006735761184245348, "learning_rate": 5.086904561386623e-06, "loss": 0.1644, "step": 186000 }, { "epoch": 2.2437979739647247, "grad_norm": 111.12641143798828, "learning_rate": 5.046881140868003e-06, "loss": 0.1578, "step": 186500 }, { "epoch": 2.2498135181308503, "grad_norm": 0.004494801629334688, "learning_rate": 5.0067775130938344e-06, "loss": 0.1625, "step": 187000 }, { "epoch": 2.2558290622969754, "grad_norm": 61.42302703857422, "learning_rate": 4.966673885319666e-06, "loss": 0.1725, "step": 187500 }, { "epoch": 2.2618446064631006, "grad_norm": 102.92018127441406, "learning_rate": 4.926570257545498e-06, "loss": 0.1503, "step": 188000 }, { "epoch": 2.2678601506292257, "grad_norm": 0.15618078410625458, "learning_rate": 4.886466629771329e-06, "loss": 0.1586, "step": 188500 }, { "epoch": 2.2738756947953513, "grad_norm": 0.1413143128156662, "learning_rate": 4.846363001997161e-06, "loss": 0.1985, "step": 189000 }, { "epoch": 2.2798912389614765, "grad_norm": 40.41890335083008, "learning_rate": 4.806259374222993e-06, "loss": 0.1302, "step": 189500 }, { "epoch": 2.2859067831276016, "grad_norm": 0.0008507597958669066, "learning_rate": 4.766235953704372e-06, "loss": 0.1409, "step": 190000 }, { "epoch": 2.291922327293727, "grad_norm": 2.4075923647615127e-05, "learning_rate": 4.726132325930204e-06, "loss": 0.1463, "step": 190500 }, { "epoch": 2.2979378714598524, "grad_norm": 0.05570465698838234, "learning_rate": 4.6860286981560356e-06, "loss": 0.1641, "step": 191000 }, { "epoch": 2.3039534156259776, "grad_norm": 177.34120178222656, "learning_rate": 4.645925070381867e-06, "loss": 0.1766, "step": 191500 }, { "epoch": 2.3099689597921027, "grad_norm": 25.356895446777344, "learning_rate": 4.605821442607698e-06, "loss": 0.1788, "step": 192000 }, { "epoch": 2.3159845039582283, "grad_norm": 26.419334411621094, "learning_rate": 4.56571781483353e-06, "loss": 0.1875, "step": 192500 }, { "epoch": 2.3220000481243535, "grad_norm": 0.05582762509584427, "learning_rate": 4.525614187059362e-06, "loss": 0.1357, "step": 193000 }, { "epoch": 2.3280155922904786, "grad_norm": 0.01142974104732275, "learning_rate": 4.485510559285193e-06, "loss": 0.148, "step": 193500 }, { "epoch": 2.3340311364566038, "grad_norm": 0.013201882131397724, "learning_rate": 4.445487138766573e-06, "loss": 0.1279, "step": 194000 }, { "epoch": 2.340046680622729, "grad_norm": 0.007936575450003147, "learning_rate": 4.405383510992405e-06, "loss": 0.1492, "step": 194500 }, { "epoch": 2.3460622247888545, "grad_norm": 0.030919533222913742, "learning_rate": 4.365360090473785e-06, "loss": 0.1298, "step": 195000 }, { "epoch": 2.3520777689549797, "grad_norm": 0.0013994334731251001, "learning_rate": 4.325256462699616e-06, "loss": 0.1555, "step": 195500 }, { "epoch": 2.358093313121105, "grad_norm": 0.004653325304389, "learning_rate": 4.285152834925448e-06, "loss": 0.1612, "step": 196000 }, { "epoch": 2.3641088572872304, "grad_norm": 0.034009434282779694, "learning_rate": 4.2450492071512796e-06, "loss": 0.1543, "step": 196500 }, { "epoch": 2.3701244014533556, "grad_norm": 0.6855669021606445, "learning_rate": 4.20502578663266e-06, "loss": 0.2072, "step": 197000 }, { "epoch": 2.3761399456194807, "grad_norm": 45.97907257080078, "learning_rate": 4.164922158858491e-06, "loss": 0.1637, "step": 197500 }, { "epoch": 2.382155489785606, "grad_norm": 0.36125409603118896, "learning_rate": 4.1248185310843216e-06, "loss": 0.142, "step": 198000 }, { "epoch": 2.388171033951731, "grad_norm": 0.013704453594982624, "learning_rate": 4.084714903310153e-06, "loss": 0.1702, "step": 198500 }, { "epoch": 2.3941865781178566, "grad_norm": 28.00567626953125, "learning_rate": 4.044611275535985e-06, "loss": 0.1641, "step": 199000 }, { "epoch": 2.4002021222839818, "grad_norm": 0.4930565655231476, "learning_rate": 4.0046680622729135e-06, "loss": 0.147, "step": 199500 }, { "epoch": 2.406217666450107, "grad_norm": 1.68142831325531, "learning_rate": 3.964564434498745e-06, "loss": 0.1543, "step": 200000 }, { "epoch": 2.4122332106162325, "grad_norm": 0.00046357992687262595, "learning_rate": 3.924460806724576e-06, "loss": 0.1383, "step": 200500 }, { "epoch": 2.4182487547823577, "grad_norm": 0.08767001330852509, "learning_rate": 3.884357178950408e-06, "loss": 0.1697, "step": 201000 }, { "epoch": 2.424264298948483, "grad_norm": 14.962833404541016, "learning_rate": 3.84425355117624e-06, "loss": 0.1919, "step": 201500 }, { "epoch": 2.430279843114608, "grad_norm": 0.0750858262181282, "learning_rate": 3.8041499234020714e-06, "loss": 0.1238, "step": 202000 }, { "epoch": 2.4362953872807336, "grad_norm": 0.011100132018327713, "learning_rate": 3.7640462956279028e-06, "loss": 0.1584, "step": 202500 }, { "epoch": 2.4423109314468587, "grad_norm": 0.0002675870491657406, "learning_rate": 3.723942667853734e-06, "loss": 0.1393, "step": 203000 }, { "epoch": 2.448326475612984, "grad_norm": 22.56096839904785, "learning_rate": 3.683839040079566e-06, "loss": 0.1284, "step": 203500 }, { "epoch": 2.454342019779109, "grad_norm": 0.0010695152450352907, "learning_rate": 3.643815619560946e-06, "loss": 0.1636, "step": 204000 }, { "epoch": 2.4603575639452346, "grad_norm": 0.05840376392006874, "learning_rate": 3.6037119917867774e-06, "loss": 0.1824, "step": 204500 }, { "epoch": 2.46637310811136, "grad_norm": 0.0005692685954272747, "learning_rate": 3.563608364012609e-06, "loss": 0.1189, "step": 205000 }, { "epoch": 2.472388652277485, "grad_norm": 12.676384925842285, "learning_rate": 3.5235047362384407e-06, "loss": 0.1589, "step": 205500 }, { "epoch": 2.47840419644361, "grad_norm": 0.024767233058810234, "learning_rate": 3.483401108464272e-06, "loss": 0.179, "step": 206000 }, { "epoch": 2.4844197406097357, "grad_norm": 0.014465268701314926, "learning_rate": 3.443297480690104e-06, "loss": 0.134, "step": 206500 }, { "epoch": 2.490435284775861, "grad_norm": 46.07574462890625, "learning_rate": 3.4031938529159353e-06, "loss": 0.1839, "step": 207000 }, { "epoch": 2.496450828941986, "grad_norm": 2.9257161617279053, "learning_rate": 3.3630902251417667e-06, "loss": 0.1522, "step": 207500 }, { "epoch": 2.5024663731081116, "grad_norm": 0.010684625245630741, "learning_rate": 3.3230668046231463e-06, "loss": 0.137, "step": 208000 }, { "epoch": 2.5084819172742367, "grad_norm": 0.023651426658034325, "learning_rate": 3.2829631768489777e-06, "loss": 0.1893, "step": 208500 }, { "epoch": 2.514497461440362, "grad_norm": 5.37974214553833, "learning_rate": 3.242939756330358e-06, "loss": 0.1688, "step": 209000 }, { "epoch": 2.520513005606487, "grad_norm": 0.21090950071811676, "learning_rate": 3.202836128556189e-06, "loss": 0.173, "step": 209500 }, { "epoch": 2.526528549772612, "grad_norm": 24.091960906982422, "learning_rate": 3.162732500782021e-06, "loss": 0.1614, "step": 210000 }, { "epoch": 2.532544093938738, "grad_norm": 0.032513245940208435, "learning_rate": 3.1226288730078524e-06, "loss": 0.1816, "step": 210500 }, { "epoch": 2.538559638104863, "grad_norm": 0.045939259231090546, "learning_rate": 3.082525245233684e-06, "loss": 0.1646, "step": 211000 }, { "epoch": 2.544575182270988, "grad_norm": 0.7573652863502502, "learning_rate": 3.0424216174595156e-06, "loss": 0.1224, "step": 211500 }, { "epoch": 2.5505907264371137, "grad_norm": 3.404059648513794, "learning_rate": 3.0023981969408957e-06, "loss": 0.1475, "step": 212000 }, { "epoch": 2.556606270603239, "grad_norm": 0.0026354603469371796, "learning_rate": 2.962294569166727e-06, "loss": 0.1591, "step": 212500 }, { "epoch": 2.562621814769364, "grad_norm": 21.002397537231445, "learning_rate": 2.9221909413925585e-06, "loss": 0.133, "step": 213000 }, { "epoch": 2.568637358935489, "grad_norm": 0.020220952108502388, "learning_rate": 2.8820873136183903e-06, "loss": 0.144, "step": 213500 }, { "epoch": 2.5746529031016143, "grad_norm": 0.02407023496925831, "learning_rate": 2.84206389309977e-06, "loss": 0.1522, "step": 214000 }, { "epoch": 2.58066844726774, "grad_norm": 0.05026541277766228, "learning_rate": 2.8019602653256018e-06, "loss": 0.1751, "step": 214500 }, { "epoch": 2.586683991433865, "grad_norm": 0.0013814900303259492, "learning_rate": 2.761856637551433e-06, "loss": 0.1606, "step": 215000 }, { "epoch": 2.5926995355999902, "grad_norm": 0.09587771445512772, "learning_rate": 2.721753009777265e-06, "loss": 0.1553, "step": 215500 }, { "epoch": 2.598715079766116, "grad_norm": 0.014226296916604042, "learning_rate": 2.6816493820030964e-06, "loss": 0.1732, "step": 216000 }, { "epoch": 2.604730623932241, "grad_norm": 0.020662061870098114, "learning_rate": 2.641545754228928e-06, "loss": 0.1851, "step": 216500 }, { "epoch": 2.610746168098366, "grad_norm": 0.020077265799045563, "learning_rate": 2.6015223337103074e-06, "loss": 0.1672, "step": 217000 }, { "epoch": 2.6167617122644913, "grad_norm": 17.917278289794922, "learning_rate": 2.561418705936139e-06, "loss": 0.1757, "step": 217500 }, { "epoch": 2.622777256430617, "grad_norm": 0.010099813342094421, "learning_rate": 2.5213150781619707e-06, "loss": 0.154, "step": 218000 }, { "epoch": 2.628792800596742, "grad_norm": 50.070430755615234, "learning_rate": 2.4812114503878025e-06, "loss": 0.1557, "step": 218500 }, { "epoch": 2.634808344762867, "grad_norm": 0.0430847629904747, "learning_rate": 2.441107822613634e-06, "loss": 0.1578, "step": 219000 }, { "epoch": 2.640823888928993, "grad_norm": 45.390384674072266, "learning_rate": 2.4010844020950135e-06, "loss": 0.1584, "step": 219500 }, { "epoch": 2.646839433095118, "grad_norm": 0.01655266433954239, "learning_rate": 2.361141188831942e-06, "loss": 0.1926, "step": 220000 }, { "epoch": 2.652854977261243, "grad_norm": 0.006697576493024826, "learning_rate": 2.3210375610577732e-06, "loss": 0.1469, "step": 220500 }, { "epoch": 2.6588705214273682, "grad_norm": 0.055724818259477615, "learning_rate": 2.280933933283605e-06, "loss": 0.1548, "step": 221000 }, { "epoch": 2.6648860655934934, "grad_norm": 0.09359436482191086, "learning_rate": 2.2408303055094365e-06, "loss": 0.1523, "step": 221500 }, { "epoch": 2.670901609759619, "grad_norm": 0.005888832733035088, "learning_rate": 2.2007266777352683e-06, "loss": 0.1896, "step": 222000 }, { "epoch": 2.676917153925744, "grad_norm": 0.018979087471961975, "learning_rate": 2.1606230499610997e-06, "loss": 0.123, "step": 222500 }, { "epoch": 2.6829326980918693, "grad_norm": 0.11329194158315659, "learning_rate": 2.1205194221869315e-06, "loss": 0.1556, "step": 223000 }, { "epoch": 2.688948242257995, "grad_norm": 0.018361905589699745, "learning_rate": 2.080415794412763e-06, "loss": 0.1621, "step": 223500 }, { "epoch": 2.69496378642412, "grad_norm": 0.00052923389011994, "learning_rate": 2.0403121666385943e-06, "loss": 0.1309, "step": 224000 }, { "epoch": 2.700979330590245, "grad_norm": 0.030753634870052338, "learning_rate": 2.0002085388644257e-06, "loss": 0.1401, "step": 224500 }, { "epoch": 2.7069948747563704, "grad_norm": 0.01107688806951046, "learning_rate": 1.9601049110902575e-06, "loss": 0.1444, "step": 225000 }, { "epoch": 2.7130104189224955, "grad_norm": 0.0049855210818350315, "learning_rate": 1.920001283316089e-06, "loss": 0.1398, "step": 225500 }, { "epoch": 2.719025963088621, "grad_norm": 52.67742156982422, "learning_rate": 1.8798976555419205e-06, "loss": 0.1743, "step": 226000 }, { "epoch": 2.7250415072547463, "grad_norm": 0.17802126705646515, "learning_rate": 1.839794027767752e-06, "loss": 0.1774, "step": 226500 }, { "epoch": 2.7310570514208714, "grad_norm": 0.03584326431155205, "learning_rate": 1.7996903999935835e-06, "loss": 0.1282, "step": 227000 }, { "epoch": 2.737072595586997, "grad_norm": 0.08680440485477448, "learning_rate": 1.7595867722194151e-06, "loss": 0.1281, "step": 227500 }, { "epoch": 2.743088139753122, "grad_norm": 0.2684645354747772, "learning_rate": 1.719563351700795e-06, "loss": 0.1414, "step": 228000 }, { "epoch": 2.7491036839192473, "grad_norm": 0.027800705283880234, "learning_rate": 1.679539931182175e-06, "loss": 0.1622, "step": 228500 }, { "epoch": 2.7551192280853725, "grad_norm": 11.01578140258789, "learning_rate": 1.6394363034080062e-06, "loss": 0.1396, "step": 229000 }, { "epoch": 2.7611347722514976, "grad_norm": 0.0035239350982010365, "learning_rate": 1.5993326756338379e-06, "loss": 0.1867, "step": 229500 }, { "epoch": 2.767150316417623, "grad_norm": 0.04345840960741043, "learning_rate": 1.5592290478596695e-06, "loss": 0.1404, "step": 230000 }, { "epoch": 2.7731658605837484, "grad_norm": 0.003262243466451764, "learning_rate": 1.5192056273410493e-06, "loss": 0.1605, "step": 230500 }, { "epoch": 2.7791814047498735, "grad_norm": 0.4090683162212372, "learning_rate": 1.479101999566881e-06, "loss": 0.1547, "step": 231000 }, { "epoch": 2.785196948915999, "grad_norm": 0.012451679445803165, "learning_rate": 1.4389983717927125e-06, "loss": 0.1147, "step": 231500 }, { "epoch": 2.7912124930821243, "grad_norm": 0.008698958903551102, "learning_rate": 1.3988947440185442e-06, "loss": 0.1869, "step": 232000 }, { "epoch": 2.7972280372482494, "grad_norm": 1.2418529987335205, "learning_rate": 1.3587911162443755e-06, "loss": 0.1428, "step": 232500 }, { "epoch": 2.8032435814143746, "grad_norm": 0.24010685086250305, "learning_rate": 1.3186874884702072e-06, "loss": 0.172, "step": 233000 }, { "epoch": 2.8092591255805, "grad_norm": 0.008175286464393139, "learning_rate": 1.2785838606960388e-06, "loss": 0.1932, "step": 233500 }, { "epoch": 2.8152746697466253, "grad_norm": 0.009441105648875237, "learning_rate": 1.2384802329218702e-06, "loss": 0.1202, "step": 234000 }, { "epoch": 2.8212902139127505, "grad_norm": 0.0015565038193017244, "learning_rate": 1.1984568124032502e-06, "loss": 0.1215, "step": 234500 }, { "epoch": 2.827305758078876, "grad_norm": 71.70852661132812, "learning_rate": 1.1583531846290816e-06, "loss": 0.1498, "step": 235000 }, { "epoch": 2.8333213022450012, "grad_norm": 4.693429470062256, "learning_rate": 1.1182495568549132e-06, "loss": 0.1463, "step": 235500 }, { "epoch": 2.8393368464111264, "grad_norm": 0.04238261282444, "learning_rate": 1.0781459290807446e-06, "loss": 0.1874, "step": 236000 }, { "epoch": 2.8453523905772515, "grad_norm": 0.091603122651577, "learning_rate": 1.0380423013065762e-06, "loss": 0.1495, "step": 236500 }, { "epoch": 2.8513679347433767, "grad_norm": 26.191734313964844, "learning_rate": 9.979386735324079e-07, "loss": 0.1246, "step": 237000 }, { "epoch": 2.8573834789095023, "grad_norm": 0.0009205593378283083, "learning_rate": 9.579152530137877e-07, "loss": 0.1417, "step": 237500 }, { "epoch": 2.8633990230756274, "grad_norm": 0.0035299675073474646, "learning_rate": 9.178116252396192e-07, "loss": 0.1665, "step": 238000 }, { "epoch": 2.8694145672417526, "grad_norm": 0.004722926300019026, "learning_rate": 8.777079974654508e-07, "loss": 0.144, "step": 238500 }, { "epoch": 2.875430111407878, "grad_norm": 0.007760610897094011, "learning_rate": 8.376043696912823e-07, "loss": 0.1325, "step": 239000 }, { "epoch": 2.8814456555740033, "grad_norm": 0.0008993456140160561, "learning_rate": 7.975007419171139e-07, "loss": 0.1572, "step": 239500 }, { "epoch": 2.8874611997401285, "grad_norm": 0.18450585007667542, "learning_rate": 7.574773213984937e-07, "loss": 0.1473, "step": 240000 }, { "epoch": 2.8934767439062536, "grad_norm": 0.0713871493935585, "learning_rate": 7.173736936243253e-07, "loss": 0.1216, "step": 240500 }, { "epoch": 2.899492288072379, "grad_norm": 0.0023451121523976326, "learning_rate": 6.773502731057053e-07, "loss": 0.1409, "step": 241000 }, { "epoch": 2.9055078322385044, "grad_norm": 0.014280580915510654, "learning_rate": 6.372466453315368e-07, "loss": 0.1946, "step": 241500 }, { "epoch": 2.9115233764046295, "grad_norm": 0.008721483871340752, "learning_rate": 5.971430175573683e-07, "loss": 0.1555, "step": 242000 }, { "epoch": 2.9175389205707547, "grad_norm": 0.039840217679739, "learning_rate": 5.570393897831999e-07, "loss": 0.1427, "step": 242500 }, { "epoch": 2.9235544647368803, "grad_norm": 0.002139493590220809, "learning_rate": 5.170159692645797e-07, "loss": 0.1436, "step": 243000 }, { "epoch": 2.9295700089030055, "grad_norm": 0.0007610457250848413, "learning_rate": 4.769123414904113e-07, "loss": 0.1393, "step": 243500 }, { "epoch": 2.9355855530691306, "grad_norm": 0.05628383159637451, "learning_rate": 4.368087137162428e-07, "loss": 0.1699, "step": 244000 }, { "epoch": 2.9416010972352558, "grad_norm": 0.02916884608566761, "learning_rate": 3.9670508594207435e-07, "loss": 0.1401, "step": 244500 }, { "epoch": 2.947616641401381, "grad_norm": 0.11827978491783142, "learning_rate": 3.5660145816790586e-07, "loss": 0.168, "step": 245000 }, { "epoch": 2.9536321855675065, "grad_norm": 0.0027151680551469326, "learning_rate": 3.165780376492858e-07, "loss": 0.1561, "step": 245500 }, { "epoch": 2.9596477297336317, "grad_norm": 0.040678609162569046, "learning_rate": 2.764744098751173e-07, "loss": 0.1305, "step": 246000 }, { "epoch": 2.965663273899757, "grad_norm": 7.518092632293701, "learning_rate": 2.3637078210094888e-07, "loss": 0.2053, "step": 246500 }, { "epoch": 2.9716788180658824, "grad_norm": 0.18074116110801697, "learning_rate": 1.962671543267804e-07, "loss": 0.1548, "step": 247000 }, { "epoch": 2.9776943622320076, "grad_norm": 7.7608208656311035, "learning_rate": 1.5616352655261197e-07, "loss": 0.1676, "step": 247500 }, { "epoch": 2.9837099063981327, "grad_norm": 0.00906237680464983, "learning_rate": 1.1605989877844351e-07, "loss": 0.1534, "step": 248000 }, { "epoch": 2.989725450564258, "grad_norm": 179.03021240234375, "learning_rate": 7.595627100427505e-08, "loss": 0.1176, "step": 248500 }, { "epoch": 2.9957409947303835, "grad_norm": 0.00344484462402761, "learning_rate": 3.58526432301066e-08, "loss": 0.1711, "step": 249000 }, { "epoch": 3.0, "eval_accuracy": 0.6820289064630488, "eval_f1": 0.7003498950314906, "eval_loss": 0.4521113932132721, "eval_roc_auc": 0.8145350422688847, "eval_runtime": 110.6722, "eval_samples_per_second": 66.268, "eval_steps_per_second": 66.268, "step": 249354 } ], "logging_steps": 500, "max_steps": 249354, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.6370896223238963e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }