diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,10498 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.956126096847579, + "eval_steps": 500, + "global_step": 30500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0025071034598027745, + "grad_norm": 0.09887482225894928, + "learning_rate": 3e-05, + "loss": 1.1694, + "step": 15 + }, + { + "epoch": 0.005014206919605549, + "grad_norm": 0.04257430136203766, + "learning_rate": 6e-05, + "loss": 1.0648, + "step": 30 + }, + { + "epoch": 0.007521310379408324, + "grad_norm": 0.025177787989377975, + "learning_rate": 9e-05, + "loss": 0.9901, + "step": 45 + }, + { + "epoch": 0.010028413839211098, + "grad_norm": 0.014405222609639168, + "learning_rate": 0.00012, + "loss": 0.9677, + "step": 60 + }, + { + "epoch": 0.012535517299013872, + "grad_norm": 0.012216474860906601, + "learning_rate": 0.00015000000000000001, + "loss": 0.9271, + "step": 75 + }, + { + "epoch": 0.015042620758816648, + "grad_norm": 0.01270276214927435, + "learning_rate": 0.00018, + "loss": 0.9087, + "step": 90 + }, + { + "epoch": 0.01754972421861942, + "grad_norm": 0.014931446872651577, + "learning_rate": 0.00019996645983565321, + "loss": 0.8911, + "step": 105 + }, + { + "epoch": 0.020056827678422196, + "grad_norm": 0.013434696942567825, + "learning_rate": 0.00019986583934261277, + "loss": 0.9009, + "step": 120 + }, + { + "epoch": 0.022563931138224972, + "grad_norm": 0.014926938340067863, + "learning_rate": 0.00019976521884957238, + "loss": 0.8827, + "step": 135 + }, + { + "epoch": 0.025071034598027744, + "grad_norm": 0.018700918182730675, + "learning_rate": 0.00019966459835653194, + "loss": 0.8815, + "step": 150 + }, + { + "epoch": 0.02757813805783052, + "grad_norm": 0.01647135801613331, + "learning_rate": 0.00019956397786349156, + "loss": 0.8762, + "step": 165 + }, + { + "epoch": 0.030085241517633296, + "grad_norm": 0.016942940652370453, + "learning_rate": 0.00019946335737045111, + "loss": 0.8648, + "step": 180 + }, + { + "epoch": 0.03259234497743607, + "grad_norm": 0.019231606274843216, + "learning_rate": 0.00019936273687741073, + "loss": 0.8567, + "step": 195 + }, + { + "epoch": 0.03509944843723884, + "grad_norm": 0.019707536324858665, + "learning_rate": 0.00019926211638437028, + "loss": 0.8568, + "step": 210 + }, + { + "epoch": 0.03760655189704162, + "grad_norm": 0.023225486278533936, + "learning_rate": 0.0001991614958913299, + "loss": 0.8507, + "step": 225 + }, + { + "epoch": 0.04011365535684439, + "grad_norm": 0.019466817378997803, + "learning_rate": 0.00019906087539828946, + "loss": 0.8559, + "step": 240 + }, + { + "epoch": 0.04262075881664717, + "grad_norm": 0.020322684198617935, + "learning_rate": 0.00019896025490524907, + "loss": 0.8411, + "step": 255 + }, + { + "epoch": 0.045127862276449944, + "grad_norm": 0.018744077533483505, + "learning_rate": 0.00019885963441220863, + "loss": 0.8587, + "step": 270 + }, + { + "epoch": 0.04763496573625271, + "grad_norm": 0.018990306183695793, + "learning_rate": 0.0001987590139191682, + "loss": 0.8434, + "step": 285 + }, + { + "epoch": 0.05014206919605549, + "grad_norm": 0.018513506278395653, + "learning_rate": 0.0001986583934261278, + "loss": 0.8573, + "step": 300 + }, + { + "epoch": 0.052649172655858265, + "grad_norm": 0.019561799243092537, + "learning_rate": 0.00019855777293308738, + "loss": 0.8459, + "step": 315 + }, + { + "epoch": 0.05515627611566104, + "grad_norm": 0.019172094762325287, + "learning_rate": 0.00019845715244004697, + "loss": 0.8653, + "step": 330 + }, + { + "epoch": 0.057663379575463816, + "grad_norm": 0.018002351745963097, + "learning_rate": 0.00019835653194700655, + "loss": 0.837, + "step": 345 + }, + { + "epoch": 0.06017048303526659, + "grad_norm": 0.01977609097957611, + "learning_rate": 0.00019825591145396614, + "loss": 0.8352, + "step": 360 + }, + { + "epoch": 0.06267758649506937, + "grad_norm": 0.019932597875595093, + "learning_rate": 0.0001981552909609257, + "loss": 0.8309, + "step": 375 + }, + { + "epoch": 0.06518468995487214, + "grad_norm": 0.01805214211344719, + "learning_rate": 0.0001980546704678853, + "loss": 0.8427, + "step": 390 + }, + { + "epoch": 0.0676917934146749, + "grad_norm": 0.018298575654625893, + "learning_rate": 0.00019795404997484487, + "loss": 0.8286, + "step": 405 + }, + { + "epoch": 0.07019889687447768, + "grad_norm": 0.01844840496778488, + "learning_rate": 0.00019785342948180448, + "loss": 0.8455, + "step": 420 + }, + { + "epoch": 0.07270600033428046, + "grad_norm": 0.018125606700778008, + "learning_rate": 0.00019775280898876404, + "loss": 0.8376, + "step": 435 + }, + { + "epoch": 0.07521310379408323, + "grad_norm": 0.01847078464925289, + "learning_rate": 0.00019765218849572365, + "loss": 0.8387, + "step": 450 + }, + { + "epoch": 0.07772020725388601, + "grad_norm": 0.0198803897947073, + "learning_rate": 0.0001975515680026832, + "loss": 0.8123, + "step": 465 + }, + { + "epoch": 0.08022731071368878, + "grad_norm": 0.018840806558728218, + "learning_rate": 0.00019745094750964282, + "loss": 0.8182, + "step": 480 + }, + { + "epoch": 0.08273441417349156, + "grad_norm": 0.02007896639406681, + "learning_rate": 0.00019735032701660238, + "loss": 0.8417, + "step": 495 + }, + { + "epoch": 0.08524151763329434, + "grad_norm": 0.02028510719537735, + "learning_rate": 0.000197249706523562, + "loss": 0.8359, + "step": 510 + }, + { + "epoch": 0.08774862109309711, + "grad_norm": 0.019926371052861214, + "learning_rate": 0.00019714908603052155, + "loss": 0.8253, + "step": 525 + }, + { + "epoch": 0.09025572455289989, + "grad_norm": 0.02000526711344719, + "learning_rate": 0.00019704846553748117, + "loss": 0.8136, + "step": 540 + }, + { + "epoch": 0.09276282801270266, + "grad_norm": 0.019949857145547867, + "learning_rate": 0.00019694784504444072, + "loss": 0.814, + "step": 555 + }, + { + "epoch": 0.09526993147250543, + "grad_norm": 0.018684815615415573, + "learning_rate": 0.0001968472245514003, + "loss": 0.8251, + "step": 570 + }, + { + "epoch": 0.0977770349323082, + "grad_norm": 0.01959558017551899, + "learning_rate": 0.0001967466040583599, + "loss": 0.8176, + "step": 585 + }, + { + "epoch": 0.10028413839211098, + "grad_norm": 0.02020624279975891, + "learning_rate": 0.00019664598356531948, + "loss": 0.8181, + "step": 600 + }, + { + "epoch": 0.10279124185191375, + "grad_norm": 0.019187506288290024, + "learning_rate": 0.00019654536307227907, + "loss": 0.8269, + "step": 615 + }, + { + "epoch": 0.10529834531171653, + "grad_norm": 0.019997362047433853, + "learning_rate": 0.00019644474257923865, + "loss": 0.8208, + "step": 630 + }, + { + "epoch": 0.1078054487715193, + "grad_norm": 0.02035447023808956, + "learning_rate": 0.00019634412208619824, + "loss": 0.822, + "step": 645 + }, + { + "epoch": 0.11031255223132208, + "grad_norm": 0.019842060282826424, + "learning_rate": 0.0001962435015931578, + "loss": 0.8123, + "step": 660 + }, + { + "epoch": 0.11281965569112486, + "grad_norm": 0.0202711783349514, + "learning_rate": 0.0001961428811001174, + "loss": 0.8137, + "step": 675 + }, + { + "epoch": 0.11532675915092763, + "grad_norm": 0.020544525235891342, + "learning_rate": 0.00019604226060707697, + "loss": 0.807, + "step": 690 + }, + { + "epoch": 0.11783386261073041, + "grad_norm": 0.02084393985569477, + "learning_rate": 0.00019594164011403658, + "loss": 0.8136, + "step": 705 + }, + { + "epoch": 0.12034096607053318, + "grad_norm": 0.020337115973234177, + "learning_rate": 0.00019584101962099614, + "loss": 0.8162, + "step": 720 + }, + { + "epoch": 0.12284806953033595, + "grad_norm": 0.019300837069749832, + "learning_rate": 0.00019574039912795575, + "loss": 0.8116, + "step": 735 + }, + { + "epoch": 0.12535517299013874, + "grad_norm": 0.020422646775841713, + "learning_rate": 0.0001956397786349153, + "loss": 0.8208, + "step": 750 + }, + { + "epoch": 0.1278622764499415, + "grad_norm": 0.019620511680841446, + "learning_rate": 0.00019553915814187492, + "loss": 0.7984, + "step": 765 + }, + { + "epoch": 0.1303693799097443, + "grad_norm": 0.018732598051428795, + "learning_rate": 0.00019543853764883448, + "loss": 0.8121, + "step": 780 + }, + { + "epoch": 0.13287648336954705, + "grad_norm": 0.020731749013066292, + "learning_rate": 0.0001953379171557941, + "loss": 0.8106, + "step": 795 + }, + { + "epoch": 0.1353835868293498, + "grad_norm": 0.019481362774968147, + "learning_rate": 0.00019523729666275365, + "loss": 0.8098, + "step": 810 + }, + { + "epoch": 0.1378906902891526, + "grad_norm": 0.021802278235554695, + "learning_rate": 0.00019513667616971326, + "loss": 0.7904, + "step": 825 + }, + { + "epoch": 0.14039779374895536, + "grad_norm": 0.02061532624065876, + "learning_rate": 0.00019503605567667282, + "loss": 0.8084, + "step": 840 + }, + { + "epoch": 0.14290489720875815, + "grad_norm": 0.01921633817255497, + "learning_rate": 0.00019493543518363243, + "loss": 0.8099, + "step": 855 + }, + { + "epoch": 0.14541200066856091, + "grad_norm": 0.020100874826312065, + "learning_rate": 0.000194834814690592, + "loss": 0.7984, + "step": 870 + }, + { + "epoch": 0.1479191041283637, + "grad_norm": 0.019538206979632378, + "learning_rate": 0.00019473419419755158, + "loss": 0.8003, + "step": 885 + }, + { + "epoch": 0.15042620758816647, + "grad_norm": 0.021109605208039284, + "learning_rate": 0.00019463357370451116, + "loss": 0.8027, + "step": 900 + }, + { + "epoch": 0.15293331104796926, + "grad_norm": 0.023268043994903564, + "learning_rate": 0.00019453295321147075, + "loss": 0.8137, + "step": 915 + }, + { + "epoch": 0.15544041450777202, + "grad_norm": 0.020138578489422798, + "learning_rate": 0.00019443233271843033, + "loss": 0.8015, + "step": 930 + }, + { + "epoch": 0.1579475179675748, + "grad_norm": 0.02234073542058468, + "learning_rate": 0.00019433171222538992, + "loss": 0.812, + "step": 945 + }, + { + "epoch": 0.16045462142737757, + "grad_norm": 0.02045338600873947, + "learning_rate": 0.0001942310917323495, + "loss": 0.7988, + "step": 960 + }, + { + "epoch": 0.16296172488718033, + "grad_norm": 0.020514754578471184, + "learning_rate": 0.00019413047123930906, + "loss": 0.7954, + "step": 975 + }, + { + "epoch": 0.16546882834698312, + "grad_norm": 0.020174162462353706, + "learning_rate": 0.00019402985074626867, + "loss": 0.7998, + "step": 990 + }, + { + "epoch": 0.16797593180678588, + "grad_norm": 0.020632125437259674, + "learning_rate": 0.00019392923025322823, + "loss": 0.8049, + "step": 1005 + }, + { + "epoch": 0.17048303526658867, + "grad_norm": 0.02562854066491127, + "learning_rate": 0.00019382860976018785, + "loss": 0.8058, + "step": 1020 + }, + { + "epoch": 0.17299013872639143, + "grad_norm": 0.019526248797774315, + "learning_rate": 0.0001937279892671474, + "loss": 0.7936, + "step": 1035 + }, + { + "epoch": 0.17549724218619422, + "grad_norm": 0.020038483664393425, + "learning_rate": 0.00019362736877410702, + "loss": 0.8057, + "step": 1050 + }, + { + "epoch": 0.178004345645997, + "grad_norm": 0.022498290985822678, + "learning_rate": 0.00019352674828106658, + "loss": 0.801, + "step": 1065 + }, + { + "epoch": 0.18051144910579978, + "grad_norm": 0.020256614312529564, + "learning_rate": 0.0001934261277880262, + "loss": 0.7984, + "step": 1080 + }, + { + "epoch": 0.18301855256560254, + "grad_norm": 0.021500416100025177, + "learning_rate": 0.00019332550729498575, + "loss": 0.8026, + "step": 1095 + }, + { + "epoch": 0.18552565602540533, + "grad_norm": 0.02118818834424019, + "learning_rate": 0.00019322488680194536, + "loss": 0.805, + "step": 1110 + }, + { + "epoch": 0.1880327594852081, + "grad_norm": 0.020503008738160133, + "learning_rate": 0.00019312426630890492, + "loss": 0.7991, + "step": 1125 + }, + { + "epoch": 0.19053986294501085, + "grad_norm": 0.02011336386203766, + "learning_rate": 0.00019302364581586453, + "loss": 0.7966, + "step": 1140 + }, + { + "epoch": 0.19304696640481364, + "grad_norm": 0.020229632034897804, + "learning_rate": 0.0001929230253228241, + "loss": 0.7925, + "step": 1155 + }, + { + "epoch": 0.1955540698646164, + "grad_norm": 0.021130822598934174, + "learning_rate": 0.00019282240482978367, + "loss": 0.7942, + "step": 1170 + }, + { + "epoch": 0.1980611733244192, + "grad_norm": 0.02094241976737976, + "learning_rate": 0.00019272178433674326, + "loss": 0.7857, + "step": 1185 + }, + { + "epoch": 0.20056827678422195, + "grad_norm": 0.01990380696952343, + "learning_rate": 0.00019262116384370284, + "loss": 0.7821, + "step": 1200 + }, + { + "epoch": 0.20307538024402474, + "grad_norm": 0.020579956471920013, + "learning_rate": 0.00019252054335066243, + "loss": 0.8054, + "step": 1215 + }, + { + "epoch": 0.2055824837038275, + "grad_norm": 0.02037345990538597, + "learning_rate": 0.00019241992285762202, + "loss": 0.789, + "step": 1230 + }, + { + "epoch": 0.2080895871636303, + "grad_norm": 0.02104773558676243, + "learning_rate": 0.0001923193023645816, + "loss": 0.7812, + "step": 1245 + }, + { + "epoch": 0.21059669062343306, + "grad_norm": 0.020640334114432335, + "learning_rate": 0.00019221868187154119, + "loss": 0.7976, + "step": 1260 + }, + { + "epoch": 0.21310379408323585, + "grad_norm": 0.02236183173954487, + "learning_rate": 0.00019211806137850077, + "loss": 0.7855, + "step": 1275 + }, + { + "epoch": 0.2156108975430386, + "grad_norm": 0.021454576402902603, + "learning_rate": 0.00019201744088546033, + "loss": 0.7921, + "step": 1290 + }, + { + "epoch": 0.21811800100284137, + "grad_norm": 0.0205401424318552, + "learning_rate": 0.00019191682039241994, + "loss": 0.7919, + "step": 1305 + }, + { + "epoch": 0.22062510446264416, + "grad_norm": 0.020427586510777473, + "learning_rate": 0.0001918161998993795, + "loss": 0.7842, + "step": 1320 + }, + { + "epoch": 0.22313220792244692, + "grad_norm": 0.020643971860408783, + "learning_rate": 0.0001917155794063391, + "loss": 0.8012, + "step": 1335 + }, + { + "epoch": 0.2256393113822497, + "grad_norm": 0.02139684371650219, + "learning_rate": 0.00019161495891329867, + "loss": 0.7853, + "step": 1350 + }, + { + "epoch": 0.22814641484205248, + "grad_norm": 0.020423240959644318, + "learning_rate": 0.00019151433842025828, + "loss": 0.7918, + "step": 1365 + }, + { + "epoch": 0.23065351830185526, + "grad_norm": 0.022509122267365456, + "learning_rate": 0.00019141371792721784, + "loss": 0.7867, + "step": 1380 + }, + { + "epoch": 0.23316062176165803, + "grad_norm": 0.021511022001504898, + "learning_rate": 0.00019131309743417746, + "loss": 0.7785, + "step": 1395 + }, + { + "epoch": 0.23566772522146082, + "grad_norm": 0.021113473922014236, + "learning_rate": 0.00019121247694113701, + "loss": 0.7835, + "step": 1410 + }, + { + "epoch": 0.23817482868126358, + "grad_norm": 0.02204412780702114, + "learning_rate": 0.00019111185644809663, + "loss": 0.7796, + "step": 1425 + }, + { + "epoch": 0.24068193214106637, + "grad_norm": 0.021546153351664543, + "learning_rate": 0.00019101123595505618, + "loss": 0.795, + "step": 1440 + }, + { + "epoch": 0.24318903560086913, + "grad_norm": 0.02130185067653656, + "learning_rate": 0.00019091061546201577, + "loss": 0.7743, + "step": 1455 + }, + { + "epoch": 0.2456961390606719, + "grad_norm": 0.020676780492067337, + "learning_rate": 0.00019080999496897536, + "loss": 0.7827, + "step": 1470 + }, + { + "epoch": 0.24820324252047468, + "grad_norm": 0.02096562273800373, + "learning_rate": 0.00019070937447593494, + "loss": 0.8017, + "step": 1485 + }, + { + "epoch": 0.25071034598027747, + "grad_norm": 0.021092170849442482, + "learning_rate": 0.00019060875398289453, + "loss": 0.7752, + "step": 1500 + }, + { + "epoch": 0.25321744944008023, + "grad_norm": 0.02107168734073639, + "learning_rate": 0.0001905081334898541, + "loss": 0.7819, + "step": 1515 + }, + { + "epoch": 0.255724552899883, + "grad_norm": 0.021778512746095657, + "learning_rate": 0.0001904075129968137, + "loss": 0.7911, + "step": 1530 + }, + { + "epoch": 0.25823165635968576, + "grad_norm": 0.020381765440106392, + "learning_rate": 0.00019030689250377328, + "loss": 0.787, + "step": 1545 + }, + { + "epoch": 0.2607387598194886, + "grad_norm": 0.02274371311068535, + "learning_rate": 0.00019020627201073287, + "loss": 0.7828, + "step": 1560 + }, + { + "epoch": 0.26324586327929134, + "grad_norm": 0.02126036398112774, + "learning_rate": 0.00019010565151769243, + "loss": 0.7798, + "step": 1575 + }, + { + "epoch": 0.2657529667390941, + "grad_norm": 0.02086903154850006, + "learning_rate": 0.00019000503102465204, + "loss": 0.7723, + "step": 1590 + }, + { + "epoch": 0.26826007019889686, + "grad_norm": 0.05506217107176781, + "learning_rate": 0.0001899044105316116, + "loss": 0.7798, + "step": 1605 + }, + { + "epoch": 0.2707671736586996, + "grad_norm": 0.02119087241590023, + "learning_rate": 0.0001898037900385712, + "loss": 0.7809, + "step": 1620 + }, + { + "epoch": 0.27327427711850244, + "grad_norm": 0.02315429411828518, + "learning_rate": 0.00018970316954553077, + "loss": 0.7881, + "step": 1635 + }, + { + "epoch": 0.2757813805783052, + "grad_norm": 0.021781641989946365, + "learning_rate": 0.00018960254905249038, + "loss": 0.7902, + "step": 1650 + }, + { + "epoch": 0.27828848403810796, + "grad_norm": 0.022906338796019554, + "learning_rate": 0.00018950192855944994, + "loss": 0.7766, + "step": 1665 + }, + { + "epoch": 0.2807955874979107, + "grad_norm": 0.021640203893184662, + "learning_rate": 0.00018940130806640955, + "loss": 0.777, + "step": 1680 + }, + { + "epoch": 0.28330269095771354, + "grad_norm": 0.02225816249847412, + "learning_rate": 0.0001893006875733691, + "loss": 0.7777, + "step": 1695 + }, + { + "epoch": 0.2858097944175163, + "grad_norm": 0.021424556151032448, + "learning_rate": 0.00018920006708032872, + "loss": 0.7609, + "step": 1710 + }, + { + "epoch": 0.28831689787731907, + "grad_norm": 0.02180912159383297, + "learning_rate": 0.00018909944658728828, + "loss": 0.7691, + "step": 1725 + }, + { + "epoch": 0.29082400133712183, + "grad_norm": 0.021193066611886024, + "learning_rate": 0.00018899882609424787, + "loss": 0.7636, + "step": 1740 + }, + { + "epoch": 0.29333110479692465, + "grad_norm": 0.021105512976646423, + "learning_rate": 0.00018889820560120745, + "loss": 0.7756, + "step": 1755 + }, + { + "epoch": 0.2958382082567274, + "grad_norm": 0.021696053445339203, + "learning_rate": 0.00018879758510816704, + "loss": 0.7814, + "step": 1770 + }, + { + "epoch": 0.29834531171653017, + "grad_norm": 0.021872224286198616, + "learning_rate": 0.00018869696461512662, + "loss": 0.7912, + "step": 1785 + }, + { + "epoch": 0.30085241517633293, + "grad_norm": 0.02113959938287735, + "learning_rate": 0.0001885963441220862, + "loss": 0.7775, + "step": 1800 + }, + { + "epoch": 0.3033595186361357, + "grad_norm": 0.020779291167855263, + "learning_rate": 0.0001884957236290458, + "loss": 0.7752, + "step": 1815 + }, + { + "epoch": 0.3058666220959385, + "grad_norm": 0.021366087719798088, + "learning_rate": 0.00018839510313600538, + "loss": 0.7748, + "step": 1830 + }, + { + "epoch": 0.3083737255557413, + "grad_norm": 0.02154374308884144, + "learning_rate": 0.00018829448264296497, + "loss": 0.7774, + "step": 1845 + }, + { + "epoch": 0.31088082901554404, + "grad_norm": 0.020630501210689545, + "learning_rate": 0.00018819386214992455, + "loss": 0.7832, + "step": 1860 + }, + { + "epoch": 0.3133879324753468, + "grad_norm": 0.022217195481061935, + "learning_rate": 0.00018809324165688414, + "loss": 0.7742, + "step": 1875 + }, + { + "epoch": 0.3158950359351496, + "grad_norm": 0.021622564643621445, + "learning_rate": 0.0001879926211638437, + "loss": 0.7782, + "step": 1890 + }, + { + "epoch": 0.3184021393949524, + "grad_norm": 0.02158367820084095, + "learning_rate": 0.0001878920006708033, + "loss": 0.7753, + "step": 1905 + }, + { + "epoch": 0.32090924285475514, + "grad_norm": 0.021993108093738556, + "learning_rate": 0.00018779138017776287, + "loss": 0.7801, + "step": 1920 + }, + { + "epoch": 0.3234163463145579, + "grad_norm": 0.02169063873589039, + "learning_rate": 0.00018769075968472248, + "loss": 0.7937, + "step": 1935 + }, + { + "epoch": 0.32592344977436066, + "grad_norm": 0.023950908333063126, + "learning_rate": 0.00018759013919168204, + "loss": 0.7668, + "step": 1950 + }, + { + "epoch": 0.3284305532341635, + "grad_norm": 0.02253536880016327, + "learning_rate": 0.00018748951869864165, + "loss": 0.7796, + "step": 1965 + }, + { + "epoch": 0.33093765669396624, + "grad_norm": 0.021693330258131027, + "learning_rate": 0.0001873888982056012, + "loss": 0.7634, + "step": 1980 + }, + { + "epoch": 0.333444760153769, + "grad_norm": 0.022510211914777756, + "learning_rate": 0.00018728827771256082, + "loss": 0.7664, + "step": 1995 + }, + { + "epoch": 0.33595186361357177, + "grad_norm": 0.021836843341588974, + "learning_rate": 0.00018718765721952038, + "loss": 0.7849, + "step": 2010 + }, + { + "epoch": 0.3384589670733746, + "grad_norm": 0.021421095356345177, + "learning_rate": 0.00018708703672647996, + "loss": 0.78, + "step": 2025 + }, + { + "epoch": 0.34096607053317735, + "grad_norm": 0.02277962490916252, + "learning_rate": 0.00018698641623343955, + "loss": 0.7935, + "step": 2040 + }, + { + "epoch": 0.3434731739929801, + "grad_norm": 0.022962411865592003, + "learning_rate": 0.00018688579574039913, + "loss": 0.7573, + "step": 2055 + }, + { + "epoch": 0.34598027745278287, + "grad_norm": 0.021092860028147697, + "learning_rate": 0.00018678517524735872, + "loss": 0.7725, + "step": 2070 + }, + { + "epoch": 0.3484873809125857, + "grad_norm": 0.0216389037668705, + "learning_rate": 0.0001866845547543183, + "loss": 0.7717, + "step": 2085 + }, + { + "epoch": 0.35099448437238845, + "grad_norm": 0.022193802520632744, + "learning_rate": 0.0001865839342612779, + "loss": 0.7671, + "step": 2100 + }, + { + "epoch": 0.3535015878321912, + "grad_norm": 0.021959876641631126, + "learning_rate": 0.00018648331376823748, + "loss": 0.7893, + "step": 2115 + }, + { + "epoch": 0.356008691291994, + "grad_norm": 0.022308630868792534, + "learning_rate": 0.00018638269327519706, + "loss": 0.7719, + "step": 2130 + }, + { + "epoch": 0.35851579475179673, + "grad_norm": 0.022814404219388962, + "learning_rate": 0.00018628207278215665, + "loss": 0.7687, + "step": 2145 + }, + { + "epoch": 0.36102289821159955, + "grad_norm": 0.021741073578596115, + "learning_rate": 0.00018618145228911623, + "loss": 0.7757, + "step": 2160 + }, + { + "epoch": 0.3635300016714023, + "grad_norm": 0.022700047120451927, + "learning_rate": 0.00018608083179607582, + "loss": 0.7724, + "step": 2175 + }, + { + "epoch": 0.3660371051312051, + "grad_norm": 0.023608332499861717, + "learning_rate": 0.0001859802113030354, + "loss": 0.7697, + "step": 2190 + }, + { + "epoch": 0.36854420859100784, + "grad_norm": 0.02221842296421528, + "learning_rate": 0.00018587959080999496, + "loss": 0.7746, + "step": 2205 + }, + { + "epoch": 0.37105131205081066, + "grad_norm": 0.022841554135084152, + "learning_rate": 0.00018577897031695457, + "loss": 0.7798, + "step": 2220 + }, + { + "epoch": 0.3735584155106134, + "grad_norm": 0.021496908739209175, + "learning_rate": 0.00018567834982391413, + "loss": 0.7608, + "step": 2235 + }, + { + "epoch": 0.3760655189704162, + "grad_norm": 0.022609667852520943, + "learning_rate": 0.00018557772933087375, + "loss": 0.7582, + "step": 2250 + }, + { + "epoch": 0.37857262243021894, + "grad_norm": 0.022388063371181488, + "learning_rate": 0.0001854771088378333, + "loss": 0.7736, + "step": 2265 + }, + { + "epoch": 0.3810797258900217, + "grad_norm": 0.021875958889722824, + "learning_rate": 0.00018537648834479292, + "loss": 0.764, + "step": 2280 + }, + { + "epoch": 0.3835868293498245, + "grad_norm": 0.023109521716833115, + "learning_rate": 0.00018527586785175247, + "loss": 0.7646, + "step": 2295 + }, + { + "epoch": 0.3860939328096273, + "grad_norm": 0.02191918157041073, + "learning_rate": 0.00018517524735871206, + "loss": 0.7688, + "step": 2310 + }, + { + "epoch": 0.38860103626943004, + "grad_norm": 0.022137146443128586, + "learning_rate": 0.00018507462686567165, + "loss": 0.7708, + "step": 2325 + }, + { + "epoch": 0.3911081397292328, + "grad_norm": 0.023074300959706306, + "learning_rate": 0.00018497400637263123, + "loss": 0.77, + "step": 2340 + }, + { + "epoch": 0.3936152431890356, + "grad_norm": 0.023129386827349663, + "learning_rate": 0.00018487338587959082, + "loss": 0.7645, + "step": 2355 + }, + { + "epoch": 0.3961223466488384, + "grad_norm": 0.022260216996073723, + "learning_rate": 0.0001847727653865504, + "loss": 0.7739, + "step": 2370 + }, + { + "epoch": 0.39862945010864115, + "grad_norm": 0.022797416895627975, + "learning_rate": 0.00018467214489351, + "loss": 0.7631, + "step": 2385 + }, + { + "epoch": 0.4011365535684439, + "grad_norm": 0.02237161435186863, + "learning_rate": 0.00018457152440046957, + "loss": 0.7708, + "step": 2400 + }, + { + "epoch": 0.40364365702824667, + "grad_norm": 0.023264579474925995, + "learning_rate": 0.00018447090390742916, + "loss": 0.7702, + "step": 2415 + }, + { + "epoch": 0.4061507604880495, + "grad_norm": 0.022827420383691788, + "learning_rate": 0.00018437028341438874, + "loss": 0.7696, + "step": 2430 + }, + { + "epoch": 0.40865786394785225, + "grad_norm": 0.022284789010882378, + "learning_rate": 0.00018426966292134833, + "loss": 0.7759, + "step": 2445 + }, + { + "epoch": 0.411164967407655, + "grad_norm": 0.0228969044983387, + "learning_rate": 0.00018416904242830791, + "loss": 0.7711, + "step": 2460 + }, + { + "epoch": 0.4136720708674578, + "grad_norm": 0.02203362248837948, + "learning_rate": 0.0001840684219352675, + "loss": 0.7557, + "step": 2475 + }, + { + "epoch": 0.4161791743272606, + "grad_norm": 0.022419359534978867, + "learning_rate": 0.00018396780144222706, + "loss": 0.7806, + "step": 2490 + }, + { + "epoch": 0.41868627778706335, + "grad_norm": 0.02259223349392414, + "learning_rate": 0.00018386718094918667, + "loss": 0.7616, + "step": 2505 + }, + { + "epoch": 0.4211933812468661, + "grad_norm": 0.023276396095752716, + "learning_rate": 0.00018376656045614623, + "loss": 0.7597, + "step": 2520 + }, + { + "epoch": 0.4237004847066689, + "grad_norm": 0.022209784016013145, + "learning_rate": 0.00018366593996310584, + "loss": 0.7554, + "step": 2535 + }, + { + "epoch": 0.4262075881664717, + "grad_norm": 0.022717982530593872, + "learning_rate": 0.0001835653194700654, + "loss": 0.7723, + "step": 2550 + }, + { + "epoch": 0.42871469162627446, + "grad_norm": 0.022251484915614128, + "learning_rate": 0.000183464698977025, + "loss": 0.7716, + "step": 2565 + }, + { + "epoch": 0.4312217950860772, + "grad_norm": 0.022392725571990013, + "learning_rate": 0.00018336407848398457, + "loss": 0.7654, + "step": 2580 + }, + { + "epoch": 0.43372889854588, + "grad_norm": 0.023053428158164024, + "learning_rate": 0.00018326345799094416, + "loss": 0.7579, + "step": 2595 + }, + { + "epoch": 0.43623600200568274, + "grad_norm": 0.02315950021147728, + "learning_rate": 0.00018316283749790374, + "loss": 0.7655, + "step": 2610 + }, + { + "epoch": 0.43874310546548556, + "grad_norm": 0.02267162874341011, + "learning_rate": 0.00018306221700486333, + "loss": 0.7503, + "step": 2625 + }, + { + "epoch": 0.4412502089252883, + "grad_norm": 0.022932684049010277, + "learning_rate": 0.0001829615965118229, + "loss": 0.7592, + "step": 2640 + }, + { + "epoch": 0.4437573123850911, + "grad_norm": 0.023299789056181908, + "learning_rate": 0.0001828609760187825, + "loss": 0.7805, + "step": 2655 + }, + { + "epoch": 0.44626441584489385, + "grad_norm": 0.022324666380882263, + "learning_rate": 0.00018276035552574208, + "loss": 0.7639, + "step": 2670 + }, + { + "epoch": 0.44877151930469666, + "grad_norm": 0.023942479863762856, + "learning_rate": 0.00018265973503270167, + "loss": 0.7506, + "step": 2685 + }, + { + "epoch": 0.4512786227644994, + "grad_norm": 0.022840656340122223, + "learning_rate": 0.00018255911453966126, + "loss": 0.7568, + "step": 2700 + }, + { + "epoch": 0.4537857262243022, + "grad_norm": 0.022889986634254456, + "learning_rate": 0.00018245849404662084, + "loss": 0.757, + "step": 2715 + }, + { + "epoch": 0.45629282968410495, + "grad_norm": 0.02276541106402874, + "learning_rate": 0.00018235787355358043, + "loss": 0.7702, + "step": 2730 + }, + { + "epoch": 0.4587999331439077, + "grad_norm": 0.022805610671639442, + "learning_rate": 0.00018225725306054, + "loss": 0.7522, + "step": 2745 + }, + { + "epoch": 0.46130703660371053, + "grad_norm": 0.02356228232383728, + "learning_rate": 0.0001821566325674996, + "loss": 0.7584, + "step": 2760 + }, + { + "epoch": 0.4638141400635133, + "grad_norm": 0.02339334785938263, + "learning_rate": 0.00018205601207445918, + "loss": 0.7551, + "step": 2775 + }, + { + "epoch": 0.46632124352331605, + "grad_norm": 0.022267676889896393, + "learning_rate": 0.00018195539158141877, + "loss": 0.7551, + "step": 2790 + }, + { + "epoch": 0.4688283469831188, + "grad_norm": 0.02209157682955265, + "learning_rate": 0.00018185477108837833, + "loss": 0.7555, + "step": 2805 + }, + { + "epoch": 0.47133545044292163, + "grad_norm": 0.023798322305083275, + "learning_rate": 0.00018175415059533794, + "loss": 0.7672, + "step": 2820 + }, + { + "epoch": 0.4738425539027244, + "grad_norm": 0.02182634547352791, + "learning_rate": 0.0001816535301022975, + "loss": 0.7523, + "step": 2835 + }, + { + "epoch": 0.47634965736252716, + "grad_norm": 0.02280135080218315, + "learning_rate": 0.0001815529096092571, + "loss": 0.7523, + "step": 2850 + }, + { + "epoch": 0.4788567608223299, + "grad_norm": 0.022913530468940735, + "learning_rate": 0.00018145228911621667, + "loss": 0.7664, + "step": 2865 + }, + { + "epoch": 0.48136386428213274, + "grad_norm": 0.022897444665431976, + "learning_rate": 0.00018135166862317625, + "loss": 0.7626, + "step": 2880 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 0.022669149562716484, + "learning_rate": 0.00018125104813013584, + "loss": 0.7666, + "step": 2895 + }, + { + "epoch": 0.48637807120173826, + "grad_norm": 0.022428149357438087, + "learning_rate": 0.00018115042763709542, + "loss": 0.7574, + "step": 2910 + }, + { + "epoch": 0.488885174661541, + "grad_norm": 0.02266399934887886, + "learning_rate": 0.000181049807144055, + "loss": 0.7713, + "step": 2925 + }, + { + "epoch": 0.4913922781213438, + "grad_norm": 0.022166673094034195, + "learning_rate": 0.0001809491866510146, + "loss": 0.7526, + "step": 2940 + }, + { + "epoch": 0.4938993815811466, + "grad_norm": 0.022409655153751373, + "learning_rate": 0.00018084856615797418, + "loss": 0.7378, + "step": 2955 + }, + { + "epoch": 0.49640648504094936, + "grad_norm": 0.02232409082353115, + "learning_rate": 0.00018074794566493377, + "loss": 0.7632, + "step": 2970 + }, + { + "epoch": 0.4989135885007521, + "grad_norm": 0.022124771028757095, + "learning_rate": 0.00018064732517189335, + "loss": 0.7493, + "step": 2985 + }, + { + "epoch": 0.5014206919605549, + "grad_norm": 0.023450786247849464, + "learning_rate": 0.00018054670467885294, + "loss": 0.751, + "step": 3000 + }, + { + "epoch": 0.5039277954203577, + "grad_norm": 0.023552143946290016, + "learning_rate": 0.00018044608418581252, + "loss": 0.7489, + "step": 3015 + }, + { + "epoch": 0.5064348988801605, + "grad_norm": 0.022822733968496323, + "learning_rate": 0.0001803454636927721, + "loss": 0.7464, + "step": 3030 + }, + { + "epoch": 0.5089420023399632, + "grad_norm": 0.02279839850962162, + "learning_rate": 0.0001802448431997317, + "loss": 0.7613, + "step": 3045 + }, + { + "epoch": 0.511449105799766, + "grad_norm": 0.023819871246814728, + "learning_rate": 0.00018014422270669128, + "loss": 0.7368, + "step": 3060 + }, + { + "epoch": 0.5139562092595688, + "grad_norm": 0.02348748780786991, + "learning_rate": 0.00018004360221365086, + "loss": 0.7485, + "step": 3075 + }, + { + "epoch": 0.5164633127193715, + "grad_norm": 0.02394930087029934, + "learning_rate": 0.00017994298172061045, + "loss": 0.7504, + "step": 3090 + }, + { + "epoch": 0.5189704161791743, + "grad_norm": 0.023166505619883537, + "learning_rate": 0.00017984236122757004, + "loss": 0.7526, + "step": 3105 + }, + { + "epoch": 0.5214775196389771, + "grad_norm": 0.023279821500182152, + "learning_rate": 0.0001797417407345296, + "loss": 0.7474, + "step": 3120 + }, + { + "epoch": 0.5239846230987799, + "grad_norm": 0.022907249629497528, + "learning_rate": 0.0001796411202414892, + "loss": 0.7555, + "step": 3135 + }, + { + "epoch": 0.5264917265585827, + "grad_norm": 0.023161666467785835, + "learning_rate": 0.00017954049974844877, + "loss": 0.7587, + "step": 3150 + }, + { + "epoch": 0.5289988300183854, + "grad_norm": 0.02453703060746193, + "learning_rate": 0.00017943987925540835, + "loss": 0.7623, + "step": 3165 + }, + { + "epoch": 0.5315059334781882, + "grad_norm": 0.02323891967535019, + "learning_rate": 0.00017933925876236794, + "loss": 0.7546, + "step": 3180 + }, + { + "epoch": 0.534013036937991, + "grad_norm": 0.022658636793494225, + "learning_rate": 0.00017923863826932752, + "loss": 0.7544, + "step": 3195 + }, + { + "epoch": 0.5365201403977937, + "grad_norm": 0.023256490007042885, + "learning_rate": 0.0001791380177762871, + "loss": 0.7631, + "step": 3210 + }, + { + "epoch": 0.5390272438575965, + "grad_norm": 0.02328312210738659, + "learning_rate": 0.0001790373972832467, + "loss": 0.7576, + "step": 3225 + }, + { + "epoch": 0.5415343473173992, + "grad_norm": 0.023502754047513008, + "learning_rate": 0.00017893677679020628, + "loss": 0.7558, + "step": 3240 + }, + { + "epoch": 0.5440414507772021, + "grad_norm": 0.02397795580327511, + "learning_rate": 0.00017883615629716586, + "loss": 0.7465, + "step": 3255 + }, + { + "epoch": 0.5465485542370049, + "grad_norm": 0.023231035098433495, + "learning_rate": 0.00017873553580412545, + "loss": 0.7596, + "step": 3270 + }, + { + "epoch": 0.5490556576968076, + "grad_norm": 0.023429760709404945, + "learning_rate": 0.00017863491531108503, + "loss": 0.7631, + "step": 3285 + }, + { + "epoch": 0.5515627611566104, + "grad_norm": 0.02327948808670044, + "learning_rate": 0.00017853429481804462, + "loss": 0.7493, + "step": 3300 + }, + { + "epoch": 0.5540698646164132, + "grad_norm": 0.023450564593076706, + "learning_rate": 0.0001784336743250042, + "loss": 0.7489, + "step": 3315 + }, + { + "epoch": 0.5565769680762159, + "grad_norm": 0.02356708236038685, + "learning_rate": 0.0001783330538319638, + "loss": 0.7541, + "step": 3330 + }, + { + "epoch": 0.5590840715360187, + "grad_norm": 0.024269040673971176, + "learning_rate": 0.00017823243333892338, + "loss": 0.7717, + "step": 3345 + }, + { + "epoch": 0.5615911749958215, + "grad_norm": 0.02358848787844181, + "learning_rate": 0.00017813181284588296, + "loss": 0.7553, + "step": 3360 + }, + { + "epoch": 0.5640982784556242, + "grad_norm": 0.02385580912232399, + "learning_rate": 0.00017803119235284255, + "loss": 0.7484, + "step": 3375 + }, + { + "epoch": 0.5666053819154271, + "grad_norm": 0.023820120841264725, + "learning_rate": 0.00017793057185980213, + "loss": 0.7529, + "step": 3390 + }, + { + "epoch": 0.5691124853752298, + "grad_norm": 0.023704256862401962, + "learning_rate": 0.00017782995136676172, + "loss": 0.763, + "step": 3405 + }, + { + "epoch": 0.5716195888350326, + "grad_norm": 0.02363293431699276, + "learning_rate": 0.0001777293308737213, + "loss": 0.7552, + "step": 3420 + }, + { + "epoch": 0.5741266922948354, + "grad_norm": 0.023471953347325325, + "learning_rate": 0.00017762871038068086, + "loss": 0.7516, + "step": 3435 + }, + { + "epoch": 0.5766337957546381, + "grad_norm": 0.023572325706481934, + "learning_rate": 0.00017752808988764045, + "loss": 0.7635, + "step": 3450 + }, + { + "epoch": 0.5791408992144409, + "grad_norm": 0.023114044219255447, + "learning_rate": 0.00017742746939460003, + "loss": 0.7376, + "step": 3465 + }, + { + "epoch": 0.5816480026742437, + "grad_norm": 0.022982290014624596, + "learning_rate": 0.00017732684890155962, + "loss": 0.7548, + "step": 3480 + }, + { + "epoch": 0.5841551061340464, + "grad_norm": 0.024818824604153633, + "learning_rate": 0.0001772262284085192, + "loss": 0.7555, + "step": 3495 + }, + { + "epoch": 0.5866622095938493, + "grad_norm": 0.024532759562134743, + "learning_rate": 0.0001771256079154788, + "loss": 0.7543, + "step": 3510 + }, + { + "epoch": 0.589169313053652, + "grad_norm": 0.023687878623604774, + "learning_rate": 0.00017702498742243837, + "loss": 0.7574, + "step": 3525 + }, + { + "epoch": 0.5916764165134548, + "grad_norm": 0.023244835436344147, + "learning_rate": 0.00017692436692939796, + "loss": 0.738, + "step": 3540 + }, + { + "epoch": 0.5941835199732576, + "grad_norm": 0.023271916434168816, + "learning_rate": 0.00017682374643635755, + "loss": 0.7472, + "step": 3555 + }, + { + "epoch": 0.5966906234330603, + "grad_norm": 0.023334383964538574, + "learning_rate": 0.00017672312594331713, + "loss": 0.7547, + "step": 3570 + }, + { + "epoch": 0.5991977268928631, + "grad_norm": 0.024023573845624924, + "learning_rate": 0.00017662250545027672, + "loss": 0.7516, + "step": 3585 + }, + { + "epoch": 0.6017048303526659, + "grad_norm": 0.023526392877101898, + "learning_rate": 0.0001765218849572363, + "loss": 0.7484, + "step": 3600 + }, + { + "epoch": 0.6042119338124686, + "grad_norm": 0.023420479148626328, + "learning_rate": 0.0001764212644641959, + "loss": 0.7368, + "step": 3615 + }, + { + "epoch": 0.6067190372722714, + "grad_norm": 0.024068370461463928, + "learning_rate": 0.00017632064397115547, + "loss": 0.7448, + "step": 3630 + }, + { + "epoch": 0.6092261407320743, + "grad_norm": 0.024318361654877663, + "learning_rate": 0.00017622002347811506, + "loss": 0.7544, + "step": 3645 + }, + { + "epoch": 0.611733244191877, + "grad_norm": 0.023683857172727585, + "learning_rate": 0.00017611940298507464, + "loss": 0.7583, + "step": 3660 + }, + { + "epoch": 0.6142403476516798, + "grad_norm": 0.023911328986287117, + "learning_rate": 0.00017601878249203423, + "loss": 0.7482, + "step": 3675 + }, + { + "epoch": 0.6167474511114825, + "grad_norm": 0.023844299837946892, + "learning_rate": 0.00017591816199899381, + "loss": 0.7406, + "step": 3690 + }, + { + "epoch": 0.6192545545712853, + "grad_norm": 0.023253358900547028, + "learning_rate": 0.0001758175415059534, + "loss": 0.7476, + "step": 3705 + }, + { + "epoch": 0.6217616580310881, + "grad_norm": 0.022935032844543457, + "learning_rate": 0.00017571692101291296, + "loss": 0.7563, + "step": 3720 + }, + { + "epoch": 0.6242687614908908, + "grad_norm": 0.02410741336643696, + "learning_rate": 0.00017561630051987254, + "loss": 0.7553, + "step": 3735 + }, + { + "epoch": 0.6267758649506936, + "grad_norm": 0.023733945563435555, + "learning_rate": 0.00017551568002683213, + "loss": 0.7395, + "step": 3750 + }, + { + "epoch": 0.6292829684104964, + "grad_norm": 0.024090424180030823, + "learning_rate": 0.00017541505953379171, + "loss": 0.7615, + "step": 3765 + }, + { + "epoch": 0.6317900718702992, + "grad_norm": 0.023794986307621002, + "learning_rate": 0.0001753144390407513, + "loss": 0.7527, + "step": 3780 + }, + { + "epoch": 0.634297175330102, + "grad_norm": 0.02363026887178421, + "learning_rate": 0.00017521381854771089, + "loss": 0.748, + "step": 3795 + }, + { + "epoch": 0.6368042787899048, + "grad_norm": 0.024967040866613388, + "learning_rate": 0.00017511319805467047, + "loss": 0.7501, + "step": 3810 + }, + { + "epoch": 0.6393113822497075, + "grad_norm": 0.02417265996336937, + "learning_rate": 0.00017501257756163006, + "loss": 0.7453, + "step": 3825 + }, + { + "epoch": 0.6418184857095103, + "grad_norm": 0.024464495480060577, + "learning_rate": 0.00017491195706858964, + "loss": 0.758, + "step": 3840 + }, + { + "epoch": 0.644325589169313, + "grad_norm": 0.023871179670095444, + "learning_rate": 0.00017481133657554923, + "loss": 0.7616, + "step": 3855 + }, + { + "epoch": 0.6468326926291158, + "grad_norm": 0.023780934512615204, + "learning_rate": 0.0001747107160825088, + "loss": 0.7453, + "step": 3870 + }, + { + "epoch": 0.6493397960889186, + "grad_norm": 0.02408822439610958, + "learning_rate": 0.0001746100955894684, + "loss": 0.7471, + "step": 3885 + }, + { + "epoch": 0.6518468995487213, + "grad_norm": 0.024668745696544647, + "learning_rate": 0.00017450947509642798, + "loss": 0.7333, + "step": 3900 + }, + { + "epoch": 0.6543540030085242, + "grad_norm": 0.023561371490359306, + "learning_rate": 0.00017440885460338757, + "loss": 0.7454, + "step": 3915 + }, + { + "epoch": 0.656861106468327, + "grad_norm": 0.02355646714568138, + "learning_rate": 0.00017430823411034716, + "loss": 0.7505, + "step": 3930 + }, + { + "epoch": 0.6593682099281297, + "grad_norm": 0.02338649332523346, + "learning_rate": 0.00017420761361730674, + "loss": 0.7615, + "step": 3945 + }, + { + "epoch": 0.6618753133879325, + "grad_norm": 0.024536214768886566, + "learning_rate": 0.00017410699312426633, + "loss": 0.7497, + "step": 3960 + }, + { + "epoch": 0.6643824168477352, + "grad_norm": 0.023618606850504875, + "learning_rate": 0.0001740063726312259, + "loss": 0.741, + "step": 3975 + }, + { + "epoch": 0.666889520307538, + "grad_norm": 0.023363051936030388, + "learning_rate": 0.0001739057521381855, + "loss": 0.7498, + "step": 3990 + }, + { + "epoch": 0.6693966237673408, + "grad_norm": 0.023151425644755363, + "learning_rate": 0.00017380513164514508, + "loss": 0.7436, + "step": 4005 + }, + { + "epoch": 0.6719037272271435, + "grad_norm": 0.024613911285996437, + "learning_rate": 0.00017370451115210464, + "loss": 0.7484, + "step": 4020 + }, + { + "epoch": 0.6744108306869463, + "grad_norm": 0.023703262209892273, + "learning_rate": 0.00017360389065906423, + "loss": 0.7401, + "step": 4035 + }, + { + "epoch": 0.6769179341467492, + "grad_norm": 0.02323344349861145, + "learning_rate": 0.0001735032701660238, + "loss": 0.7372, + "step": 4050 + }, + { + "epoch": 0.6794250376065519, + "grad_norm": 0.023779282346367836, + "learning_rate": 0.0001734026496729834, + "loss": 0.7474, + "step": 4065 + }, + { + "epoch": 0.6819321410663547, + "grad_norm": 0.024744119495153427, + "learning_rate": 0.00017330202917994298, + "loss": 0.7337, + "step": 4080 + }, + { + "epoch": 0.6844392445261575, + "grad_norm": 0.02366352453827858, + "learning_rate": 0.00017320140868690257, + "loss": 0.7755, + "step": 4095 + }, + { + "epoch": 0.6869463479859602, + "grad_norm": 0.02404959499835968, + "learning_rate": 0.00017310078819386215, + "loss": 0.7412, + "step": 4110 + }, + { + "epoch": 0.689453451445763, + "grad_norm": 0.024871889501810074, + "learning_rate": 0.00017300016770082174, + "loss": 0.7521, + "step": 4125 + }, + { + "epoch": 0.6919605549055657, + "grad_norm": 0.02386365458369255, + "learning_rate": 0.00017289954720778132, + "loss": 0.7431, + "step": 4140 + }, + { + "epoch": 0.6944676583653685, + "grad_norm": 0.025385569781064987, + "learning_rate": 0.0001727989267147409, + "loss": 0.73, + "step": 4155 + }, + { + "epoch": 0.6969747618251714, + "grad_norm": 0.024604368954896927, + "learning_rate": 0.0001726983062217005, + "loss": 0.7474, + "step": 4170 + }, + { + "epoch": 0.6994818652849741, + "grad_norm": 0.025954630225896835, + "learning_rate": 0.00017259768572866008, + "loss": 0.7473, + "step": 4185 + }, + { + "epoch": 0.7019889687447769, + "grad_norm": 0.02412698231637478, + "learning_rate": 0.00017249706523561967, + "loss": 0.7498, + "step": 4200 + }, + { + "epoch": 0.7044960722045797, + "grad_norm": 0.02433890663087368, + "learning_rate": 0.00017239644474257925, + "loss": 0.7474, + "step": 4215 + }, + { + "epoch": 0.7070031756643824, + "grad_norm": 0.02414149045944214, + "learning_rate": 0.00017229582424953884, + "loss": 0.7416, + "step": 4230 + }, + { + "epoch": 0.7095102791241852, + "grad_norm": 0.03919633850455284, + "learning_rate": 0.00017219520375649842, + "loss": 0.7502, + "step": 4245 + }, + { + "epoch": 0.712017382583988, + "grad_norm": 0.02408537268638611, + "learning_rate": 0.000172094583263458, + "loss": 0.7456, + "step": 4260 + }, + { + "epoch": 0.7145244860437907, + "grad_norm": 0.02505289390683174, + "learning_rate": 0.0001719939627704176, + "loss": 0.7412, + "step": 4275 + }, + { + "epoch": 0.7170315895035935, + "grad_norm": 0.02388434298336506, + "learning_rate": 0.00017189334227737718, + "loss": 0.739, + "step": 4290 + }, + { + "epoch": 0.7195386929633963, + "grad_norm": 0.02636132948100567, + "learning_rate": 0.00017179272178433674, + "loss": 0.7405, + "step": 4305 + }, + { + "epoch": 0.7220457964231991, + "grad_norm": 0.02557826228439808, + "learning_rate": 0.00017169210129129635, + "loss": 0.7424, + "step": 4320 + }, + { + "epoch": 0.7245528998830019, + "grad_norm": 0.02385845221579075, + "learning_rate": 0.0001715914807982559, + "loss": 0.7388, + "step": 4335 + }, + { + "epoch": 0.7270600033428046, + "grad_norm": 0.02581110969185829, + "learning_rate": 0.0001714908603052155, + "loss": 0.7414, + "step": 4350 + }, + { + "epoch": 0.7295671068026074, + "grad_norm": 0.025572916492819786, + "learning_rate": 0.00017139023981217508, + "loss": 0.7526, + "step": 4365 + }, + { + "epoch": 0.7320742102624102, + "grad_norm": 0.024279674515128136, + "learning_rate": 0.00017128961931913466, + "loss": 0.738, + "step": 4380 + }, + { + "epoch": 0.7345813137222129, + "grad_norm": 0.02414841763675213, + "learning_rate": 0.00017118899882609425, + "loss": 0.7387, + "step": 4395 + }, + { + "epoch": 0.7370884171820157, + "grad_norm": 0.024131467565894127, + "learning_rate": 0.00017108837833305384, + "loss": 0.743, + "step": 4410 + }, + { + "epoch": 0.7395955206418184, + "grad_norm": 0.024498678743839264, + "learning_rate": 0.00017098775784001342, + "loss": 0.7531, + "step": 4425 + }, + { + "epoch": 0.7421026241016213, + "grad_norm": 0.024572541937232018, + "learning_rate": 0.000170887137346973, + "loss": 0.7489, + "step": 4440 + }, + { + "epoch": 0.7446097275614241, + "grad_norm": 0.02463640458881855, + "learning_rate": 0.0001707865168539326, + "loss": 0.7379, + "step": 4455 + }, + { + "epoch": 0.7471168310212268, + "grad_norm": 0.024474984034895897, + "learning_rate": 0.00017068589636089218, + "loss": 0.7532, + "step": 4470 + }, + { + "epoch": 0.7496239344810296, + "grad_norm": 0.023911593481898308, + "learning_rate": 0.00017058527586785176, + "loss": 0.7346, + "step": 4485 + }, + { + "epoch": 0.7521310379408324, + "grad_norm": 0.024990247562527657, + "learning_rate": 0.00017048465537481135, + "loss": 0.727, + "step": 4500 + }, + { + "epoch": 0.7546381414006351, + "grad_norm": 0.024192336946725845, + "learning_rate": 0.00017038403488177093, + "loss": 0.7462, + "step": 4515 + }, + { + "epoch": 0.7571452448604379, + "grad_norm": 0.02413538470864296, + "learning_rate": 0.00017028341438873052, + "loss": 0.7364, + "step": 4530 + }, + { + "epoch": 0.7596523483202406, + "grad_norm": 0.02461206167936325, + "learning_rate": 0.0001701827938956901, + "loss": 0.7321, + "step": 4545 + }, + { + "epoch": 0.7621594517800434, + "grad_norm": 0.024669578298926353, + "learning_rate": 0.0001700821734026497, + "loss": 0.7503, + "step": 4560 + }, + { + "epoch": 0.7646665552398463, + "grad_norm": 0.02436312846839428, + "learning_rate": 0.00016998155290960928, + "loss": 0.7346, + "step": 4575 + }, + { + "epoch": 0.767173658699649, + "grad_norm": 0.025169432163238525, + "learning_rate": 0.00016988093241656883, + "loss": 0.7219, + "step": 4590 + }, + { + "epoch": 0.7696807621594518, + "grad_norm": 0.025311505421996117, + "learning_rate": 0.00016978031192352845, + "loss": 0.742, + "step": 4605 + }, + { + "epoch": 0.7721878656192546, + "grad_norm": 0.024896448478102684, + "learning_rate": 0.000169679691430488, + "loss": 0.746, + "step": 4620 + }, + { + "epoch": 0.7746949690790573, + "grad_norm": 0.025063227862119675, + "learning_rate": 0.0001695790709374476, + "loss": 0.7399, + "step": 4635 + }, + { + "epoch": 0.7772020725388601, + "grad_norm": 0.024744588881731033, + "learning_rate": 0.00016947845044440718, + "loss": 0.742, + "step": 4650 + }, + { + "epoch": 0.7797091759986629, + "grad_norm": 0.025170577690005302, + "learning_rate": 0.00016937782995136676, + "loss": 0.7288, + "step": 4665 + }, + { + "epoch": 0.7822162794584656, + "grad_norm": 0.024757632985711098, + "learning_rate": 0.00016927720945832635, + "loss": 0.7407, + "step": 4680 + }, + { + "epoch": 0.7847233829182684, + "grad_norm": 0.025282783433794975, + "learning_rate": 0.00016917658896528593, + "loss": 0.7294, + "step": 4695 + }, + { + "epoch": 0.7872304863780712, + "grad_norm": 0.025306588038802147, + "learning_rate": 0.00016907596847224552, + "loss": 0.7414, + "step": 4710 + }, + { + "epoch": 0.789737589837874, + "grad_norm": 0.024476177990436554, + "learning_rate": 0.0001689753479792051, + "loss": 0.7377, + "step": 4725 + }, + { + "epoch": 0.7922446932976768, + "grad_norm": 0.025107109919190407, + "learning_rate": 0.0001688747274861647, + "loss": 0.7378, + "step": 4740 + }, + { + "epoch": 0.7947517967574795, + "grad_norm": 0.024397587403655052, + "learning_rate": 0.00016877410699312427, + "loss": 0.7308, + "step": 4755 + }, + { + "epoch": 0.7972589002172823, + "grad_norm": 0.02418595738708973, + "learning_rate": 0.00016867348650008386, + "loss": 0.7437, + "step": 4770 + }, + { + "epoch": 0.7997660036770851, + "grad_norm": 0.025148652493953705, + "learning_rate": 0.00016857286600704345, + "loss": 0.7365, + "step": 4785 + }, + { + "epoch": 0.8022731071368878, + "grad_norm": 0.025669820606708527, + "learning_rate": 0.00016847224551400303, + "loss": 0.7392, + "step": 4800 + }, + { + "epoch": 0.8047802105966906, + "grad_norm": 0.02602335438132286, + "learning_rate": 0.00016837162502096262, + "loss": 0.7377, + "step": 4815 + }, + { + "epoch": 0.8072873140564933, + "grad_norm": 0.02492678537964821, + "learning_rate": 0.0001682710045279222, + "loss": 0.7435, + "step": 4830 + }, + { + "epoch": 0.8097944175162962, + "grad_norm": 0.02486814185976982, + "learning_rate": 0.0001681703840348818, + "loss": 0.7392, + "step": 4845 + }, + { + "epoch": 0.812301520976099, + "grad_norm": 0.026057204231619835, + "learning_rate": 0.00016806976354184137, + "loss": 0.7371, + "step": 4860 + }, + { + "epoch": 0.8148086244359017, + "grad_norm": 0.025231441482901573, + "learning_rate": 0.00016796914304880093, + "loss": 0.7429, + "step": 4875 + }, + { + "epoch": 0.8173157278957045, + "grad_norm": 0.025132806971669197, + "learning_rate": 0.00016786852255576054, + "loss": 0.7398, + "step": 4890 + }, + { + "epoch": 0.8198228313555073, + "grad_norm": 0.02506762556731701, + "learning_rate": 0.0001677679020627201, + "loss": 0.7463, + "step": 4905 + }, + { + "epoch": 0.82232993481531, + "grad_norm": 0.02398357354104519, + "learning_rate": 0.00016766728156967971, + "loss": 0.7309, + "step": 4920 + }, + { + "epoch": 0.8248370382751128, + "grad_norm": 0.025060344487428665, + "learning_rate": 0.00016756666107663927, + "loss": 0.7287, + "step": 4935 + }, + { + "epoch": 0.8273441417349156, + "grad_norm": 0.024265987798571587, + "learning_rate": 0.00016746604058359886, + "loss": 0.7439, + "step": 4950 + }, + { + "epoch": 0.8298512451947184, + "grad_norm": 0.025207631289958954, + "learning_rate": 0.00016736542009055844, + "loss": 0.7332, + "step": 4965 + }, + { + "epoch": 0.8323583486545212, + "grad_norm": 0.025070613250136375, + "learning_rate": 0.00016726479959751803, + "loss": 0.7356, + "step": 4980 + }, + { + "epoch": 0.834865452114324, + "grad_norm": 0.025521699339151382, + "learning_rate": 0.00016716417910447761, + "loss": 0.7345, + "step": 4995 + }, + { + "epoch": 0.8373725555741267, + "grad_norm": 0.025154948234558105, + "learning_rate": 0.0001670635586114372, + "loss": 0.7362, + "step": 5010 + }, + { + "epoch": 0.8398796590339295, + "grad_norm": 0.025558389723300934, + "learning_rate": 0.00016696293811839679, + "loss": 0.7294, + "step": 5025 + }, + { + "epoch": 0.8423867624937322, + "grad_norm": 0.026137180626392365, + "learning_rate": 0.00016686231762535637, + "loss": 0.7331, + "step": 5040 + }, + { + "epoch": 0.844893865953535, + "grad_norm": 0.024644847959280014, + "learning_rate": 0.00016676169713231596, + "loss": 0.7382, + "step": 5055 + }, + { + "epoch": 0.8474009694133378, + "grad_norm": 0.024775272235274315, + "learning_rate": 0.00016666107663927554, + "loss": 0.7242, + "step": 5070 + }, + { + "epoch": 0.8499080728731405, + "grad_norm": 0.025577571243047714, + "learning_rate": 0.00016656045614623513, + "loss": 0.7192, + "step": 5085 + }, + { + "epoch": 0.8524151763329434, + "grad_norm": 0.024751491844654083, + "learning_rate": 0.0001664598356531947, + "loss": 0.7219, + "step": 5100 + }, + { + "epoch": 0.8549222797927462, + "grad_norm": 0.025324271991848946, + "learning_rate": 0.0001663592151601543, + "loss": 0.7412, + "step": 5115 + }, + { + "epoch": 0.8574293832525489, + "grad_norm": 0.02564609982073307, + "learning_rate": 0.00016625859466711388, + "loss": 0.7366, + "step": 5130 + }, + { + "epoch": 0.8599364867123517, + "grad_norm": 0.02468453161418438, + "learning_rate": 0.00016615797417407347, + "loss": 0.7387, + "step": 5145 + }, + { + "epoch": 0.8624435901721544, + "grad_norm": 0.025196226313710213, + "learning_rate": 0.00016605735368103303, + "loss": 0.7299, + "step": 5160 + }, + { + "epoch": 0.8649506936319572, + "grad_norm": 0.02621576189994812, + "learning_rate": 0.00016595673318799264, + "loss": 0.7495, + "step": 5175 + }, + { + "epoch": 0.86745779709176, + "grad_norm": 0.025252273306250572, + "learning_rate": 0.0001658561126949522, + "loss": 0.7322, + "step": 5190 + }, + { + "epoch": 0.8699649005515627, + "grad_norm": 0.025535358116030693, + "learning_rate": 0.0001657554922019118, + "loss": 0.7281, + "step": 5205 + }, + { + "epoch": 0.8724720040113655, + "grad_norm": 0.024804269894957542, + "learning_rate": 0.00016565487170887137, + "loss": 0.7505, + "step": 5220 + }, + { + "epoch": 0.8749791074711684, + "grad_norm": 0.02469950169324875, + "learning_rate": 0.00016555425121583098, + "loss": 0.7265, + "step": 5235 + }, + { + "epoch": 0.8774862109309711, + "grad_norm": 0.02518155239522457, + "learning_rate": 0.00016545363072279054, + "loss": 0.7288, + "step": 5250 + }, + { + "epoch": 0.8799933143907739, + "grad_norm": 0.024804813787341118, + "learning_rate": 0.00016535301022975013, + "loss": 0.7382, + "step": 5265 + }, + { + "epoch": 0.8825004178505766, + "grad_norm": 0.024241533130407333, + "learning_rate": 0.0001652523897367097, + "loss": 0.7408, + "step": 5280 + }, + { + "epoch": 0.8850075213103794, + "grad_norm": 0.025099163874983788, + "learning_rate": 0.0001651517692436693, + "loss": 0.7324, + "step": 5295 + }, + { + "epoch": 0.8875146247701822, + "grad_norm": 0.025935839861631393, + "learning_rate": 0.00016505114875062888, + "loss": 0.7353, + "step": 5310 + }, + { + "epoch": 0.8900217282299849, + "grad_norm": 0.024958360940217972, + "learning_rate": 0.00016495052825758847, + "loss": 0.724, + "step": 5325 + }, + { + "epoch": 0.8925288316897877, + "grad_norm": 0.024382906034588814, + "learning_rate": 0.00016484990776454805, + "loss": 0.7372, + "step": 5340 + }, + { + "epoch": 0.8950359351495905, + "grad_norm": 0.02473212592303753, + "learning_rate": 0.00016474928727150764, + "loss": 0.7531, + "step": 5355 + }, + { + "epoch": 0.8975430386093933, + "grad_norm": 0.024407681077718735, + "learning_rate": 0.00016464866677846722, + "loss": 0.7489, + "step": 5370 + }, + { + "epoch": 0.9000501420691961, + "grad_norm": 0.02625984139740467, + "learning_rate": 0.0001645480462854268, + "loss": 0.7399, + "step": 5385 + }, + { + "epoch": 0.9025572455289989, + "grad_norm": 0.026630889624357224, + "learning_rate": 0.0001644474257923864, + "loss": 0.7322, + "step": 5400 + }, + { + "epoch": 0.9050643489888016, + "grad_norm": 0.025531059131026268, + "learning_rate": 0.00016434680529934598, + "loss": 0.7499, + "step": 5415 + }, + { + "epoch": 0.9075714524486044, + "grad_norm": 0.025482535362243652, + "learning_rate": 0.00016424618480630557, + "loss": 0.7287, + "step": 5430 + }, + { + "epoch": 0.9100785559084071, + "grad_norm": 0.026173191145062447, + "learning_rate": 0.00016414556431326512, + "loss": 0.7363, + "step": 5445 + }, + { + "epoch": 0.9125856593682099, + "grad_norm": 0.024878835305571556, + "learning_rate": 0.00016404494382022474, + "loss": 0.7592, + "step": 5460 + }, + { + "epoch": 0.9150927628280127, + "grad_norm": 0.025847023352980614, + "learning_rate": 0.0001639443233271843, + "loss": 0.7138, + "step": 5475 + }, + { + "epoch": 0.9175998662878154, + "grad_norm": 0.026683717966079712, + "learning_rate": 0.0001638437028341439, + "loss": 0.7266, + "step": 5490 + }, + { + "epoch": 0.9201069697476183, + "grad_norm": 0.02581162378191948, + "learning_rate": 0.00016374308234110347, + "loss": 0.7175, + "step": 5505 + }, + { + "epoch": 0.9226140732074211, + "grad_norm": 0.02513813227415085, + "learning_rate": 0.00016364246184806308, + "loss": 0.74, + "step": 5520 + }, + { + "epoch": 0.9251211766672238, + "grad_norm": 0.024819128215312958, + "learning_rate": 0.00016354184135502264, + "loss": 0.742, + "step": 5535 + }, + { + "epoch": 0.9276282801270266, + "grad_norm": 0.024832414463162422, + "learning_rate": 0.00016344122086198222, + "loss": 0.7465, + "step": 5550 + }, + { + "epoch": 0.9301353835868293, + "grad_norm": 0.02581876330077648, + "learning_rate": 0.0001633406003689418, + "loss": 0.7383, + "step": 5565 + }, + { + "epoch": 0.9326424870466321, + "grad_norm": 0.024939673021435738, + "learning_rate": 0.0001632399798759014, + "loss": 0.7479, + "step": 5580 + }, + { + "epoch": 0.9351495905064349, + "grad_norm": 0.025533072650432587, + "learning_rate": 0.00016313935938286098, + "loss": 0.7259, + "step": 5595 + }, + { + "epoch": 0.9376566939662376, + "grad_norm": 0.02547396905720234, + "learning_rate": 0.00016303873888982056, + "loss": 0.7258, + "step": 5610 + }, + { + "epoch": 0.9401637974260404, + "grad_norm": 0.025361550971865654, + "learning_rate": 0.00016293811839678015, + "loss": 0.7302, + "step": 5625 + }, + { + "epoch": 0.9426709008858433, + "grad_norm": 0.02566991187632084, + "learning_rate": 0.00016283749790373974, + "loss": 0.7319, + "step": 5640 + }, + { + "epoch": 0.945178004345646, + "grad_norm": 0.026383578777313232, + "learning_rate": 0.00016273687741069932, + "loss": 0.7523, + "step": 5655 + }, + { + "epoch": 0.9476851078054488, + "grad_norm": 0.025949161499738693, + "learning_rate": 0.0001626362569176589, + "loss": 0.7116, + "step": 5670 + }, + { + "epoch": 0.9501922112652516, + "grad_norm": 0.02509259060025215, + "learning_rate": 0.0001625356364246185, + "loss": 0.7259, + "step": 5685 + }, + { + "epoch": 0.9526993147250543, + "grad_norm": 0.025692781433463097, + "learning_rate": 0.00016243501593157808, + "loss": 0.7263, + "step": 5700 + }, + { + "epoch": 0.9552064181848571, + "grad_norm": 0.025238677859306335, + "learning_rate": 0.00016233439543853766, + "loss": 0.7315, + "step": 5715 + }, + { + "epoch": 0.9577135216446598, + "grad_norm": 0.025801653042435646, + "learning_rate": 0.00016223377494549722, + "loss": 0.7329, + "step": 5730 + }, + { + "epoch": 0.9602206251044626, + "grad_norm": 0.025331363081932068, + "learning_rate": 0.00016213315445245683, + "loss": 0.7245, + "step": 5745 + }, + { + "epoch": 0.9627277285642655, + "grad_norm": 0.025975272059440613, + "learning_rate": 0.0001620325339594164, + "loss": 0.7119, + "step": 5760 + }, + { + "epoch": 0.9652348320240682, + "grad_norm": 0.025318987667560577, + "learning_rate": 0.000161931913466376, + "loss": 0.7481, + "step": 5775 + }, + { + "epoch": 0.967741935483871, + "grad_norm": 0.02570466138422489, + "learning_rate": 0.00016183129297333556, + "loss": 0.7506, + "step": 5790 + }, + { + "epoch": 0.9702490389436738, + "grad_norm": 0.024957410991191864, + "learning_rate": 0.00016173067248029518, + "loss": 0.7277, + "step": 5805 + }, + { + "epoch": 0.9727561424034765, + "grad_norm": 0.026068007573485374, + "learning_rate": 0.00016163005198725473, + "loss": 0.7305, + "step": 5820 + }, + { + "epoch": 0.9752632458632793, + "grad_norm": 0.026244519278407097, + "learning_rate": 0.00016152943149421435, + "loss": 0.7184, + "step": 5835 + }, + { + "epoch": 0.977770349323082, + "grad_norm": 0.025324849411845207, + "learning_rate": 0.0001614288110011739, + "loss": 0.7264, + "step": 5850 + }, + { + "epoch": 0.9802774527828848, + "grad_norm": 0.025065554305911064, + "learning_rate": 0.0001613281905081335, + "loss": 0.7294, + "step": 5865 + }, + { + "epoch": 0.9827845562426876, + "grad_norm": 0.025444064289331436, + "learning_rate": 0.00016122757001509308, + "loss": 0.728, + "step": 5880 + }, + { + "epoch": 0.9852916597024904, + "grad_norm": 0.026068173348903656, + "learning_rate": 0.00016112694952205266, + "loss": 0.741, + "step": 5895 + }, + { + "epoch": 0.9877987631622932, + "grad_norm": 0.024954237043857574, + "learning_rate": 0.00016102632902901225, + "loss": 0.7375, + "step": 5910 + }, + { + "epoch": 0.990305866622096, + "grad_norm": 0.0247243270277977, + "learning_rate": 0.00016092570853597183, + "loss": 0.7375, + "step": 5925 + }, + { + "epoch": 0.9928129700818987, + "grad_norm": 0.025755500420928, + "learning_rate": 0.00016082508804293142, + "loss": 0.7368, + "step": 5940 + }, + { + "epoch": 0.9953200735417015, + "grad_norm": 0.026517482474446297, + "learning_rate": 0.000160724467549891, + "loss": 0.7203, + "step": 5955 + }, + { + "epoch": 0.9978271770015043, + "grad_norm": 0.025983400642871857, + "learning_rate": 0.0001606238470568506, + "loss": 0.7142, + "step": 5970 + }, + { + "epoch": 1.0003342804613071, + "grad_norm": 0.024920133873820305, + "learning_rate": 0.00016052322656381017, + "loss": 0.7166, + "step": 5985 + }, + { + "epoch": 1.0028413839211099, + "grad_norm": 0.02642948552966118, + "learning_rate": 0.00016042260607076976, + "loss": 0.7074, + "step": 6000 + }, + { + "epoch": 1.0053484873809126, + "grad_norm": 0.026105554774403572, + "learning_rate": 0.00016032198557772932, + "loss": 0.7139, + "step": 6015 + }, + { + "epoch": 1.0078555908407154, + "grad_norm": 0.0251301322132349, + "learning_rate": 0.00016022136508468893, + "loss": 0.7071, + "step": 6030 + }, + { + "epoch": 1.0103626943005182, + "grad_norm": 0.025702379643917084, + "learning_rate": 0.0001601207445916485, + "loss": 0.6975, + "step": 6045 + }, + { + "epoch": 1.012869797760321, + "grad_norm": 0.02600419521331787, + "learning_rate": 0.0001600201240986081, + "loss": 0.718, + "step": 6060 + }, + { + "epoch": 1.0153769012201237, + "grad_norm": 0.026151692494750023, + "learning_rate": 0.00015991950360556766, + "loss": 0.7069, + "step": 6075 + }, + { + "epoch": 1.0178840046799265, + "grad_norm": 0.025088109076023102, + "learning_rate": 0.00015981888311252727, + "loss": 0.708, + "step": 6090 + }, + { + "epoch": 1.0203911081397292, + "grad_norm": 0.026014501228928566, + "learning_rate": 0.00015971826261948683, + "loss": 0.7133, + "step": 6105 + }, + { + "epoch": 1.022898211599532, + "grad_norm": 0.02501731365919113, + "learning_rate": 0.00015961764212644644, + "loss": 0.7191, + "step": 6120 + }, + { + "epoch": 1.0254053150593347, + "grad_norm": 0.025291357189416885, + "learning_rate": 0.000159517021633406, + "loss": 0.7164, + "step": 6135 + }, + { + "epoch": 1.0279124185191375, + "grad_norm": 0.026282720267772675, + "learning_rate": 0.00015941640114036561, + "loss": 0.72, + "step": 6150 + }, + { + "epoch": 1.0304195219789403, + "grad_norm": 0.026496944949030876, + "learning_rate": 0.00015931578064732517, + "loss": 0.7212, + "step": 6165 + }, + { + "epoch": 1.032926625438743, + "grad_norm": 0.026129065081477165, + "learning_rate": 0.00015921516015428476, + "loss": 0.7121, + "step": 6180 + }, + { + "epoch": 1.0354337288985458, + "grad_norm": 0.027691906318068504, + "learning_rate": 0.00015911453966124434, + "loss": 0.7238, + "step": 6195 + }, + { + "epoch": 1.0379408323583486, + "grad_norm": 0.025952916592359543, + "learning_rate": 0.00015901391916820393, + "loss": 0.7141, + "step": 6210 + }, + { + "epoch": 1.0404479358181513, + "grad_norm": 0.0261197779327631, + "learning_rate": 0.00015891329867516351, + "loss": 0.7116, + "step": 6225 + }, + { + "epoch": 1.0429550392779543, + "grad_norm": 0.02602444589138031, + "learning_rate": 0.0001588126781821231, + "loss": 0.7212, + "step": 6240 + }, + { + "epoch": 1.045462142737757, + "grad_norm": 0.027341393753886223, + "learning_rate": 0.00015871205768908269, + "loss": 0.7065, + "step": 6255 + }, + { + "epoch": 1.0479692461975598, + "grad_norm": 0.026516225188970566, + "learning_rate": 0.00015861143719604227, + "loss": 0.7137, + "step": 6270 + }, + { + "epoch": 1.0504763496573626, + "grad_norm": 0.025233183056116104, + "learning_rate": 0.00015851081670300186, + "loss": 0.7286, + "step": 6285 + }, + { + "epoch": 1.0529834531171653, + "grad_norm": 0.025705158710479736, + "learning_rate": 0.00015841019620996141, + "loss": 0.7252, + "step": 6300 + }, + { + "epoch": 1.055490556576968, + "grad_norm": 0.025452638044953346, + "learning_rate": 0.00015830957571692103, + "loss": 0.7207, + "step": 6315 + }, + { + "epoch": 1.0579976600367709, + "grad_norm": 0.027089523151516914, + "learning_rate": 0.00015820895522388059, + "loss": 0.7123, + "step": 6330 + }, + { + "epoch": 1.0605047634965736, + "grad_norm": 0.02557321824133396, + "learning_rate": 0.0001581083347308402, + "loss": 0.7153, + "step": 6345 + }, + { + "epoch": 1.0630118669563764, + "grad_norm": 0.026297248899936676, + "learning_rate": 0.00015800771423779976, + "loss": 0.7183, + "step": 6360 + }, + { + "epoch": 1.0655189704161792, + "grad_norm": 0.026958812028169632, + "learning_rate": 0.00015790709374475937, + "loss": 0.7117, + "step": 6375 + }, + { + "epoch": 1.068026073875982, + "grad_norm": 0.026555512100458145, + "learning_rate": 0.00015780647325171893, + "loss": 0.7025, + "step": 6390 + }, + { + "epoch": 1.0705331773357847, + "grad_norm": 0.026713771745562553, + "learning_rate": 0.00015770585275867854, + "loss": 0.7158, + "step": 6405 + }, + { + "epoch": 1.0730402807955874, + "grad_norm": 0.02662680670619011, + "learning_rate": 0.0001576052322656381, + "loss": 0.7104, + "step": 6420 + }, + { + "epoch": 1.0755473842553902, + "grad_norm": 0.02612622268497944, + "learning_rate": 0.0001575046117725977, + "loss": 0.7186, + "step": 6435 + }, + { + "epoch": 1.078054487715193, + "grad_norm": 0.02652982994914055, + "learning_rate": 0.00015740399127955727, + "loss": 0.7097, + "step": 6450 + }, + { + "epoch": 1.0805615911749957, + "grad_norm": 0.026232892647385597, + "learning_rate": 0.00015730337078651685, + "loss": 0.7203, + "step": 6465 + }, + { + "epoch": 1.0830686946347985, + "grad_norm": 0.02632397972047329, + "learning_rate": 0.00015720275029347644, + "loss": 0.7209, + "step": 6480 + }, + { + "epoch": 1.0855757980946015, + "grad_norm": 0.02648136578500271, + "learning_rate": 0.00015710212980043603, + "loss": 0.7182, + "step": 6495 + }, + { + "epoch": 1.0880829015544042, + "grad_norm": 0.025636956095695496, + "learning_rate": 0.0001570015093073956, + "loss": 0.7077, + "step": 6510 + }, + { + "epoch": 1.090590005014207, + "grad_norm": 0.026664093136787415, + "learning_rate": 0.0001569008888143552, + "loss": 0.7216, + "step": 6525 + }, + { + "epoch": 1.0930971084740098, + "grad_norm": 0.02704274095594883, + "learning_rate": 0.00015680026832131478, + "loss": 0.714, + "step": 6540 + }, + { + "epoch": 1.0956042119338125, + "grad_norm": 0.026222985237836838, + "learning_rate": 0.00015669964782827437, + "loss": 0.7134, + "step": 6555 + }, + { + "epoch": 1.0981113153936153, + "grad_norm": 0.02940414845943451, + "learning_rate": 0.00015659902733523395, + "loss": 0.6986, + "step": 6570 + }, + { + "epoch": 1.100618418853418, + "grad_norm": 0.025812886655330658, + "learning_rate": 0.0001564984068421935, + "loss": 0.7163, + "step": 6585 + }, + { + "epoch": 1.1031255223132208, + "grad_norm": 0.026331394910812378, + "learning_rate": 0.00015639778634915312, + "loss": 0.7097, + "step": 6600 + }, + { + "epoch": 1.1056326257730236, + "grad_norm": 0.027025267481803894, + "learning_rate": 0.00015629716585611268, + "loss": 0.6983, + "step": 6615 + }, + { + "epoch": 1.1081397292328263, + "grad_norm": 0.02628287486732006, + "learning_rate": 0.0001561965453630723, + "loss": 0.7113, + "step": 6630 + }, + { + "epoch": 1.110646832692629, + "grad_norm": 0.0271297600120306, + "learning_rate": 0.00015609592487003185, + "loss": 0.7163, + "step": 6645 + }, + { + "epoch": 1.1131539361524319, + "grad_norm": 0.027640245854854584, + "learning_rate": 0.00015599530437699147, + "loss": 0.6974, + "step": 6660 + }, + { + "epoch": 1.1156610396122346, + "grad_norm": 0.026571575552225113, + "learning_rate": 0.00015589468388395102, + "loss": 0.7032, + "step": 6675 + }, + { + "epoch": 1.1181681430720374, + "grad_norm": 0.02639468014240265, + "learning_rate": 0.00015579406339091064, + "loss": 0.7139, + "step": 6690 + }, + { + "epoch": 1.1206752465318401, + "grad_norm": 0.026831267401576042, + "learning_rate": 0.0001556934428978702, + "loss": 0.7209, + "step": 6705 + }, + { + "epoch": 1.123182349991643, + "grad_norm": 0.028162870556116104, + "learning_rate": 0.0001555928224048298, + "loss": 0.7066, + "step": 6720 + }, + { + "epoch": 1.1256894534514457, + "grad_norm": 0.02714131958782673, + "learning_rate": 0.00015549220191178937, + "loss": 0.721, + "step": 6735 + }, + { + "epoch": 1.1281965569112486, + "grad_norm": 0.0281366016715765, + "learning_rate": 0.00015539158141874898, + "loss": 0.7024, + "step": 6750 + }, + { + "epoch": 1.1307036603710512, + "grad_norm": 0.027032790705561638, + "learning_rate": 0.00015529096092570854, + "loss": 0.7283, + "step": 6765 + }, + { + "epoch": 1.1332107638308542, + "grad_norm": 0.026658054441213608, + "learning_rate": 0.00015519034043266812, + "loss": 0.7111, + "step": 6780 + }, + { + "epoch": 1.135717867290657, + "grad_norm": 0.026945605874061584, + "learning_rate": 0.0001550897199396277, + "loss": 0.7298, + "step": 6795 + }, + { + "epoch": 1.1382249707504597, + "grad_norm": 0.02765739895403385, + "learning_rate": 0.0001549890994465873, + "loss": 0.7081, + "step": 6810 + }, + { + "epoch": 1.1407320742102625, + "grad_norm": 0.02612920291721821, + "learning_rate": 0.00015488847895354688, + "loss": 0.709, + "step": 6825 + }, + { + "epoch": 1.1432391776700652, + "grad_norm": 0.026704227551817894, + "learning_rate": 0.00015478785846050646, + "loss": 0.7088, + "step": 6840 + }, + { + "epoch": 1.145746281129868, + "grad_norm": 0.027153639122843742, + "learning_rate": 0.00015468723796746605, + "loss": 0.7166, + "step": 6855 + }, + { + "epoch": 1.1482533845896707, + "grad_norm": 0.02730732038617134, + "learning_rate": 0.0001545866174744256, + "loss": 0.7101, + "step": 6870 + }, + { + "epoch": 1.1507604880494735, + "grad_norm": 0.027596892789006233, + "learning_rate": 0.00015448599698138522, + "loss": 0.7122, + "step": 6885 + }, + { + "epoch": 1.1532675915092763, + "grad_norm": 0.02678474597632885, + "learning_rate": 0.00015438537648834478, + "loss": 0.7226, + "step": 6900 + }, + { + "epoch": 1.155774694969079, + "grad_norm": 0.02596975676715374, + "learning_rate": 0.0001542847559953044, + "loss": 0.7119, + "step": 6915 + }, + { + "epoch": 1.1582817984288818, + "grad_norm": 0.026990054175257683, + "learning_rate": 0.00015418413550226395, + "loss": 0.7186, + "step": 6930 + }, + { + "epoch": 1.1607889018886846, + "grad_norm": 0.026957310736179352, + "learning_rate": 0.00015408351500922356, + "loss": 0.7201, + "step": 6945 + }, + { + "epoch": 1.1632960053484873, + "grad_norm": 0.02676299959421158, + "learning_rate": 0.00015398289451618312, + "loss": 0.7116, + "step": 6960 + }, + { + "epoch": 1.16580310880829, + "grad_norm": 0.026614701375365257, + "learning_rate": 0.00015388227402314273, + "loss": 0.7181, + "step": 6975 + }, + { + "epoch": 1.1683102122680928, + "grad_norm": 0.02804492600262165, + "learning_rate": 0.0001537816535301023, + "loss": 0.7062, + "step": 6990 + }, + { + "epoch": 1.1708173157278958, + "grad_norm": 0.027462385594844818, + "learning_rate": 0.0001536810330370619, + "loss": 0.7274, + "step": 7005 + }, + { + "epoch": 1.1733244191876984, + "grad_norm": 0.026805778965353966, + "learning_rate": 0.00015358041254402146, + "loss": 0.7176, + "step": 7020 + }, + { + "epoch": 1.1758315226475013, + "grad_norm": 0.027235226705670357, + "learning_rate": 0.00015347979205098108, + "loss": 0.6944, + "step": 7035 + }, + { + "epoch": 1.178338626107304, + "grad_norm": 0.02651335299015045, + "learning_rate": 0.00015337917155794063, + "loss": 0.7084, + "step": 7050 + }, + { + "epoch": 1.1808457295671069, + "grad_norm": 0.027431068941950798, + "learning_rate": 0.00015327855106490025, + "loss": 0.7261, + "step": 7065 + }, + { + "epoch": 1.1833528330269096, + "grad_norm": 0.027069034054875374, + "learning_rate": 0.0001531779305718598, + "loss": 0.7269, + "step": 7080 + }, + { + "epoch": 1.1858599364867124, + "grad_norm": 0.026999959722161293, + "learning_rate": 0.0001530773100788194, + "loss": 0.713, + "step": 7095 + }, + { + "epoch": 1.1883670399465152, + "grad_norm": 0.027173152193427086, + "learning_rate": 0.00015297668958577898, + "loss": 0.7099, + "step": 7110 + }, + { + "epoch": 1.190874143406318, + "grad_norm": 0.026728777214884758, + "learning_rate": 0.00015287606909273856, + "loss": 0.7006, + "step": 7125 + }, + { + "epoch": 1.1933812468661207, + "grad_norm": 0.02722666971385479, + "learning_rate": 0.00015277544859969815, + "loss": 0.711, + "step": 7140 + }, + { + "epoch": 1.1958883503259234, + "grad_norm": 0.027167314663529396, + "learning_rate": 0.00015267482810665773, + "loss": 0.708, + "step": 7155 + }, + { + "epoch": 1.1983954537857262, + "grad_norm": 0.027100099250674248, + "learning_rate": 0.00015257420761361732, + "loss": 0.6944, + "step": 7170 + }, + { + "epoch": 1.200902557245529, + "grad_norm": 0.026492077857255936, + "learning_rate": 0.00015247358712057688, + "loss": 0.7122, + "step": 7185 + }, + { + "epoch": 1.2034096607053317, + "grad_norm": 0.027062034234404564, + "learning_rate": 0.0001523729666275365, + "loss": 0.7333, + "step": 7200 + }, + { + "epoch": 1.2059167641651345, + "grad_norm": 0.026957035064697266, + "learning_rate": 0.00015227234613449605, + "loss": 0.706, + "step": 7215 + }, + { + "epoch": 1.2084238676249373, + "grad_norm": 0.027580831199884415, + "learning_rate": 0.00015217172564145566, + "loss": 0.7113, + "step": 7230 + }, + { + "epoch": 1.21093097108474, + "grad_norm": 0.02672952227294445, + "learning_rate": 0.00015207110514841522, + "loss": 0.6934, + "step": 7245 + }, + { + "epoch": 1.2134380745445428, + "grad_norm": 0.026900822296738625, + "learning_rate": 0.00015197048465537483, + "loss": 0.7015, + "step": 7260 + }, + { + "epoch": 1.2159451780043455, + "grad_norm": 0.028098303824663162, + "learning_rate": 0.0001518698641623344, + "loss": 0.7143, + "step": 7275 + }, + { + "epoch": 1.2184522814641485, + "grad_norm": 0.026865461841225624, + "learning_rate": 0.000151769243669294, + "loss": 0.7253, + "step": 7290 + }, + { + "epoch": 1.2209593849239513, + "grad_norm": 0.02781241200864315, + "learning_rate": 0.00015166862317625356, + "loss": 0.7099, + "step": 7305 + }, + { + "epoch": 1.223466488383754, + "grad_norm": 0.027126578614115715, + "learning_rate": 0.00015156800268321317, + "loss": 0.6956, + "step": 7320 + }, + { + "epoch": 1.2259735918435568, + "grad_norm": 0.02705315686762333, + "learning_rate": 0.00015146738219017273, + "loss": 0.7037, + "step": 7335 + }, + { + "epoch": 1.2284806953033596, + "grad_norm": 0.027233878150582314, + "learning_rate": 0.00015136676169713234, + "loss": 0.7137, + "step": 7350 + }, + { + "epoch": 1.2309877987631623, + "grad_norm": 0.028538642451167107, + "learning_rate": 0.0001512661412040919, + "loss": 0.7123, + "step": 7365 + }, + { + "epoch": 1.233494902222965, + "grad_norm": 0.027490422129631042, + "learning_rate": 0.0001511655207110515, + "loss": 0.7208, + "step": 7380 + }, + { + "epoch": 1.2360020056827679, + "grad_norm": 0.02747008576989174, + "learning_rate": 0.00015106490021801107, + "loss": 0.7196, + "step": 7395 + }, + { + "epoch": 1.2385091091425706, + "grad_norm": 0.026851654052734375, + "learning_rate": 0.00015096427972497066, + "loss": 0.7149, + "step": 7410 + }, + { + "epoch": 1.2410162126023734, + "grad_norm": 0.02743196301162243, + "learning_rate": 0.00015086365923193024, + "loss": 0.7175, + "step": 7425 + }, + { + "epoch": 1.2435233160621761, + "grad_norm": 0.028329750522971153, + "learning_rate": 0.00015076303873888983, + "loss": 0.698, + "step": 7440 + }, + { + "epoch": 1.246030419521979, + "grad_norm": 0.026834193617105484, + "learning_rate": 0.00015066241824584941, + "loss": 0.7063, + "step": 7455 + }, + { + "epoch": 1.2485375229817817, + "grad_norm": 0.028689688071608543, + "learning_rate": 0.000150561797752809, + "loss": 0.704, + "step": 7470 + }, + { + "epoch": 1.2510446264415844, + "grad_norm": 0.02716403640806675, + "learning_rate": 0.00015046117725976859, + "loss": 0.7095, + "step": 7485 + }, + { + "epoch": 1.2535517299013872, + "grad_norm": 0.027952060103416443, + "learning_rate": 0.00015036055676672814, + "loss": 0.7341, + "step": 7500 + }, + { + "epoch": 1.2560588333611902, + "grad_norm": 0.028136277571320534, + "learning_rate": 0.00015025993627368776, + "loss": 0.6967, + "step": 7515 + }, + { + "epoch": 1.2585659368209927, + "grad_norm": 0.027513163164258003, + "learning_rate": 0.00015015931578064731, + "loss": 0.7053, + "step": 7530 + }, + { + "epoch": 1.2610730402807957, + "grad_norm": 0.027584819123148918, + "learning_rate": 0.00015005869528760693, + "loss": 0.7162, + "step": 7545 + }, + { + "epoch": 1.2635801437405982, + "grad_norm": 0.02737903967499733, + "learning_rate": 0.00014995807479456649, + "loss": 0.7211, + "step": 7560 + }, + { + "epoch": 1.2660872472004012, + "grad_norm": 0.028384409844875336, + "learning_rate": 0.0001498574543015261, + "loss": 0.7059, + "step": 7575 + }, + { + "epoch": 1.268594350660204, + "grad_norm": 0.027213079854846, + "learning_rate": 0.00014975683380848566, + "loss": 0.7064, + "step": 7590 + }, + { + "epoch": 1.2711014541200067, + "grad_norm": 0.02736794948577881, + "learning_rate": 0.00014965621331544527, + "loss": 0.712, + "step": 7605 + }, + { + "epoch": 1.2736085575798095, + "grad_norm": 0.026495933532714844, + "learning_rate": 0.00014955559282240483, + "loss": 0.7115, + "step": 7620 + }, + { + "epoch": 1.2761156610396123, + "grad_norm": 0.02718982845544815, + "learning_rate": 0.00014945497232936444, + "loss": 0.7039, + "step": 7635 + }, + { + "epoch": 1.278622764499415, + "grad_norm": 0.027888623997569084, + "learning_rate": 0.000149354351836324, + "loss": 0.6947, + "step": 7650 + }, + { + "epoch": 1.2811298679592178, + "grad_norm": 0.027887005358934402, + "learning_rate": 0.0001492537313432836, + "loss": 0.7092, + "step": 7665 + }, + { + "epoch": 1.2836369714190206, + "grad_norm": 0.02832951210439205, + "learning_rate": 0.00014915311085024317, + "loss": 0.7253, + "step": 7680 + }, + { + "epoch": 1.2861440748788233, + "grad_norm": 0.027755776420235634, + "learning_rate": 0.00014905249035720275, + "loss": 0.7051, + "step": 7695 + }, + { + "epoch": 1.288651178338626, + "grad_norm": 0.027755258604884148, + "learning_rate": 0.00014895186986416234, + "loss": 0.7131, + "step": 7710 + }, + { + "epoch": 1.2911582817984288, + "grad_norm": 0.027515331283211708, + "learning_rate": 0.00014885124937112193, + "loss": 0.6972, + "step": 7725 + }, + { + "epoch": 1.2936653852582316, + "grad_norm": 0.02867818996310234, + "learning_rate": 0.0001487506288780815, + "loss": 0.6909, + "step": 7740 + }, + { + "epoch": 1.2961724887180344, + "grad_norm": 0.027417359873652458, + "learning_rate": 0.0001486500083850411, + "loss": 0.7112, + "step": 7755 + }, + { + "epoch": 1.2986795921778371, + "grad_norm": 0.02725161798298359, + "learning_rate": 0.00014854938789200068, + "loss": 0.7172, + "step": 7770 + }, + { + "epoch": 1.30118669563764, + "grad_norm": 0.027100631967186928, + "learning_rate": 0.00014844876739896024, + "loss": 0.7079, + "step": 7785 + }, + { + "epoch": 1.3036937990974429, + "grad_norm": 0.026735814288258553, + "learning_rate": 0.00014834814690591985, + "loss": 0.7134, + "step": 7800 + }, + { + "epoch": 1.3062009025572454, + "grad_norm": 0.02827010303735733, + "learning_rate": 0.0001482475264128794, + "loss": 0.7077, + "step": 7815 + }, + { + "epoch": 1.3087080060170484, + "grad_norm": 0.02705741673707962, + "learning_rate": 0.00014814690591983902, + "loss": 0.7172, + "step": 7830 + }, + { + "epoch": 1.3112151094768512, + "grad_norm": 0.02796081081032753, + "learning_rate": 0.00014804628542679858, + "loss": 0.7232, + "step": 7845 + }, + { + "epoch": 1.313722212936654, + "grad_norm": 0.027841266244649887, + "learning_rate": 0.0001479456649337582, + "loss": 0.7113, + "step": 7860 + }, + { + "epoch": 1.3162293163964567, + "grad_norm": 0.030358731746673584, + "learning_rate": 0.00014784504444071775, + "loss": 0.7014, + "step": 7875 + }, + { + "epoch": 1.3187364198562594, + "grad_norm": 0.02849227376282215, + "learning_rate": 0.00014774442394767737, + "loss": 0.7233, + "step": 7890 + }, + { + "epoch": 1.3212435233160622, + "grad_norm": 0.02644391544163227, + "learning_rate": 0.00014764380345463692, + "loss": 0.729, + "step": 7905 + }, + { + "epoch": 1.323750626775865, + "grad_norm": 0.027298742905259132, + "learning_rate": 0.00014754318296159654, + "loss": 0.722, + "step": 7920 + }, + { + "epoch": 1.3262577302356677, + "grad_norm": 0.027199968695640564, + "learning_rate": 0.0001474425624685561, + "loss": 0.7041, + "step": 7935 + }, + { + "epoch": 1.3287648336954705, + "grad_norm": 0.027822501957416534, + "learning_rate": 0.0001473419419755157, + "loss": 0.7044, + "step": 7950 + }, + { + "epoch": 1.3312719371552733, + "grad_norm": 0.027914773672819138, + "learning_rate": 0.00014724132148247527, + "loss": 0.7074, + "step": 7965 + }, + { + "epoch": 1.333779040615076, + "grad_norm": 0.028190581128001213, + "learning_rate": 0.00014714070098943488, + "loss": 0.7014, + "step": 7980 + }, + { + "epoch": 1.3362861440748788, + "grad_norm": 0.027638264000415802, + "learning_rate": 0.00014704008049639444, + "loss": 0.6973, + "step": 7995 + }, + { + "epoch": 1.3387932475346815, + "grad_norm": 0.028353575617074966, + "learning_rate": 0.00014693946000335402, + "loss": 0.7191, + "step": 8010 + }, + { + "epoch": 1.3413003509944843, + "grad_norm": 0.027547866106033325, + "learning_rate": 0.0001468388395103136, + "loss": 0.7089, + "step": 8025 + }, + { + "epoch": 1.343807454454287, + "grad_norm": 0.02667342871427536, + "learning_rate": 0.0001467382190172732, + "loss": 0.7075, + "step": 8040 + }, + { + "epoch": 1.34631455791409, + "grad_norm": 0.028818530961871147, + "learning_rate": 0.00014663759852423278, + "loss": 0.7008, + "step": 8055 + }, + { + "epoch": 1.3488216613738926, + "grad_norm": 0.02606160379946232, + "learning_rate": 0.00014653697803119236, + "loss": 0.7207, + "step": 8070 + }, + { + "epoch": 1.3513287648336956, + "grad_norm": 0.028475910425186157, + "learning_rate": 0.00014643635753815195, + "loss": 0.6993, + "step": 8085 + }, + { + "epoch": 1.3538358682934981, + "grad_norm": 0.02790878899395466, + "learning_rate": 0.0001463357370451115, + "loss": 0.7058, + "step": 8100 + }, + { + "epoch": 1.356342971753301, + "grad_norm": 0.028986552730202675, + "learning_rate": 0.00014623511655207112, + "loss": 0.7056, + "step": 8115 + }, + { + "epoch": 1.3588500752131039, + "grad_norm": 0.02837732620537281, + "learning_rate": 0.00014613449605903068, + "loss": 0.708, + "step": 8130 + }, + { + "epoch": 1.3613571786729066, + "grad_norm": 0.027905132621526718, + "learning_rate": 0.0001460338755659903, + "loss": 0.7024, + "step": 8145 + }, + { + "epoch": 1.3638642821327094, + "grad_norm": 0.027892014011740685, + "learning_rate": 0.00014593325507294985, + "loss": 0.7011, + "step": 8160 + }, + { + "epoch": 1.3663713855925121, + "grad_norm": 0.02732338011264801, + "learning_rate": 0.00014583263457990946, + "loss": 0.6972, + "step": 8175 + }, + { + "epoch": 1.368878489052315, + "grad_norm": 0.028097622096538544, + "learning_rate": 0.00014573201408686902, + "loss": 0.6951, + "step": 8190 + }, + { + "epoch": 1.3713855925121177, + "grad_norm": 0.028016911819577217, + "learning_rate": 0.00014563139359382863, + "loss": 0.7156, + "step": 8205 + }, + { + "epoch": 1.3738926959719204, + "grad_norm": 0.02762255072593689, + "learning_rate": 0.0001455307731007882, + "loss": 0.7021, + "step": 8220 + }, + { + "epoch": 1.3763997994317232, + "grad_norm": 0.027654899284243584, + "learning_rate": 0.0001454301526077478, + "loss": 0.7074, + "step": 8235 + }, + { + "epoch": 1.378906902891526, + "grad_norm": 0.027378590777516365, + "learning_rate": 0.00014532953211470736, + "loss": 0.706, + "step": 8250 + }, + { + "epoch": 1.3814140063513287, + "grad_norm": 0.026956256479024887, + "learning_rate": 0.00014522891162166698, + "loss": 0.7183, + "step": 8265 + }, + { + "epoch": 1.3839211098111315, + "grad_norm": 0.027121366932988167, + "learning_rate": 0.00014512829112862653, + "loss": 0.7027, + "step": 8280 + }, + { + "epoch": 1.3864282132709342, + "grad_norm": 0.02765464223921299, + "learning_rate": 0.00014502767063558612, + "loss": 0.7132, + "step": 8295 + }, + { + "epoch": 1.3889353167307372, + "grad_norm": 0.02817637287080288, + "learning_rate": 0.0001449270501425457, + "loss": 0.6864, + "step": 8310 + }, + { + "epoch": 1.3914424201905398, + "grad_norm": 0.02854936383664608, + "learning_rate": 0.0001448264296495053, + "loss": 0.7248, + "step": 8325 + }, + { + "epoch": 1.3939495236503427, + "grad_norm": 0.028685523197054863, + "learning_rate": 0.00014472580915646488, + "loss": 0.7036, + "step": 8340 + }, + { + "epoch": 1.3964566271101453, + "grad_norm": 0.028023192659020424, + "learning_rate": 0.00014462518866342446, + "loss": 0.706, + "step": 8355 + }, + { + "epoch": 1.3989637305699483, + "grad_norm": 0.027805542573332787, + "learning_rate": 0.00014452456817038405, + "loss": 0.7126, + "step": 8370 + }, + { + "epoch": 1.401470834029751, + "grad_norm": 0.027909213677048683, + "learning_rate": 0.00014442394767734363, + "loss": 0.7133, + "step": 8385 + }, + { + "epoch": 1.4039779374895538, + "grad_norm": 0.02798452228307724, + "learning_rate": 0.00014432332718430322, + "loss": 0.7052, + "step": 8400 + }, + { + "epoch": 1.4064850409493566, + "grad_norm": 0.02735227160155773, + "learning_rate": 0.00014422270669126278, + "loss": 0.7096, + "step": 8415 + }, + { + "epoch": 1.4089921444091593, + "grad_norm": 0.027850987389683723, + "learning_rate": 0.0001441220861982224, + "loss": 0.718, + "step": 8430 + }, + { + "epoch": 1.411499247868962, + "grad_norm": 0.028347337618470192, + "learning_rate": 0.00014402146570518195, + "loss": 0.6989, + "step": 8445 + }, + { + "epoch": 1.4140063513287648, + "grad_norm": 0.028133846819400787, + "learning_rate": 0.00014392084521214156, + "loss": 0.7073, + "step": 8460 + }, + { + "epoch": 1.4165134547885676, + "grad_norm": 0.02889505960047245, + "learning_rate": 0.00014382022471910112, + "loss": 0.7157, + "step": 8475 + }, + { + "epoch": 1.4190205582483704, + "grad_norm": 0.02751564234495163, + "learning_rate": 0.00014371960422606073, + "loss": 0.7115, + "step": 8490 + }, + { + "epoch": 1.4215276617081731, + "grad_norm": 0.027201758697628975, + "learning_rate": 0.0001436189837330203, + "loss": 0.7121, + "step": 8505 + }, + { + "epoch": 1.424034765167976, + "grad_norm": 0.047122806310653687, + "learning_rate": 0.0001435183632399799, + "loss": 0.7103, + "step": 8520 + }, + { + "epoch": 1.4265418686277787, + "grad_norm": 0.028580831363797188, + "learning_rate": 0.00014341774274693946, + "loss": 0.6933, + "step": 8535 + }, + { + "epoch": 1.4290489720875814, + "grad_norm": 0.028754740953445435, + "learning_rate": 0.00014331712225389907, + "loss": 0.7155, + "step": 8550 + }, + { + "epoch": 1.4315560755473842, + "grad_norm": 0.028142735362052917, + "learning_rate": 0.00014321650176085863, + "loss": 0.7076, + "step": 8565 + }, + { + "epoch": 1.434063179007187, + "grad_norm": 0.02792290225625038, + "learning_rate": 0.00014311588126781822, + "loss": 0.7202, + "step": 8580 + }, + { + "epoch": 1.43657028246699, + "grad_norm": 0.027254393324255943, + "learning_rate": 0.0001430152607747778, + "loss": 0.7116, + "step": 8595 + }, + { + "epoch": 1.4390773859267925, + "grad_norm": 0.027158159762620926, + "learning_rate": 0.0001429146402817374, + "loss": 0.7034, + "step": 8610 + }, + { + "epoch": 1.4415844893865954, + "grad_norm": 0.028217531740665436, + "learning_rate": 0.00014281401978869697, + "loss": 0.7136, + "step": 8625 + }, + { + "epoch": 1.4440915928463982, + "grad_norm": 0.028678081929683685, + "learning_rate": 0.00014271339929565656, + "loss": 0.7053, + "step": 8640 + }, + { + "epoch": 1.446598696306201, + "grad_norm": 0.028371306136250496, + "learning_rate": 0.00014261277880261614, + "loss": 0.7115, + "step": 8655 + }, + { + "epoch": 1.4491057997660037, + "grad_norm": 0.027796892449259758, + "learning_rate": 0.00014251215830957573, + "loss": 0.7138, + "step": 8670 + }, + { + "epoch": 1.4516129032258065, + "grad_norm": 0.027524475008249283, + "learning_rate": 0.00014241153781653531, + "loss": 0.7048, + "step": 8685 + }, + { + "epoch": 1.4541200066856093, + "grad_norm": 0.02704106830060482, + "learning_rate": 0.00014231091732349487, + "loss": 0.6967, + "step": 8700 + }, + { + "epoch": 1.456627110145412, + "grad_norm": 0.028332151472568512, + "learning_rate": 0.00014221029683045448, + "loss": 0.7015, + "step": 8715 + }, + { + "epoch": 1.4591342136052148, + "grad_norm": 0.028455249965190887, + "learning_rate": 0.00014210967633741404, + "loss": 0.6973, + "step": 8730 + }, + { + "epoch": 1.4616413170650175, + "grad_norm": 0.028323177248239517, + "learning_rate": 0.00014200905584437366, + "loss": 0.7119, + "step": 8745 + }, + { + "epoch": 1.4641484205248203, + "grad_norm": 0.028827426955103874, + "learning_rate": 0.00014190843535133321, + "loss": 0.6973, + "step": 8760 + }, + { + "epoch": 1.466655523984623, + "grad_norm": 0.029024334624409676, + "learning_rate": 0.00014180781485829283, + "loss": 0.7109, + "step": 8775 + }, + { + "epoch": 1.4691626274444258, + "grad_norm": 0.02851213701069355, + "learning_rate": 0.00014170719436525239, + "loss": 0.7038, + "step": 8790 + }, + { + "epoch": 1.4716697309042286, + "grad_norm": 0.027595283463597298, + "learning_rate": 0.000141606573872212, + "loss": 0.7119, + "step": 8805 + }, + { + "epoch": 1.4741768343640314, + "grad_norm": 0.02817492000758648, + "learning_rate": 0.00014150595337917156, + "loss": 0.7007, + "step": 8820 + }, + { + "epoch": 1.4766839378238341, + "grad_norm": 0.028595896437764168, + "learning_rate": 0.00014140533288613117, + "loss": 0.7031, + "step": 8835 + }, + { + "epoch": 1.479191041283637, + "grad_norm": 0.028396232053637505, + "learning_rate": 0.00014130471239309073, + "loss": 0.6944, + "step": 8850 + }, + { + "epoch": 1.4816981447434396, + "grad_norm": 0.02777491882443428, + "learning_rate": 0.0001412040919000503, + "loss": 0.7069, + "step": 8865 + }, + { + "epoch": 1.4842052482032426, + "grad_norm": 0.02780229039490223, + "learning_rate": 0.0001411034714070099, + "loss": 0.6955, + "step": 8880 + }, + { + "epoch": 1.4867123516630452, + "grad_norm": 0.02779022604227066, + "learning_rate": 0.00014100285091396948, + "loss": 0.7101, + "step": 8895 + }, + { + "epoch": 1.4892194551228481, + "grad_norm": 0.029339686036109924, + "learning_rate": 0.00014090223042092907, + "loss": 0.7074, + "step": 8910 + }, + { + "epoch": 1.491726558582651, + "grad_norm": 0.0277661494910717, + "learning_rate": 0.00014080160992788865, + "loss": 0.7007, + "step": 8925 + }, + { + "epoch": 1.4942336620424537, + "grad_norm": 0.028384177014231682, + "learning_rate": 0.00014070098943484824, + "loss": 0.7006, + "step": 8940 + }, + { + "epoch": 1.4967407655022564, + "grad_norm": 0.027173573151230812, + "learning_rate": 0.00014060036894180783, + "loss": 0.7126, + "step": 8955 + }, + { + "epoch": 1.4992478689620592, + "grad_norm": 0.029250754043459892, + "learning_rate": 0.0001404997484487674, + "loss": 0.7053, + "step": 8970 + }, + { + "epoch": 1.501754972421862, + "grad_norm": 0.02840678207576275, + "learning_rate": 0.000140399127955727, + "loss": 0.691, + "step": 8985 + }, + { + "epoch": 1.5042620758816647, + "grad_norm": 0.029002662748098373, + "learning_rate": 0.00014029850746268658, + "loss": 0.6969, + "step": 9000 + }, + { + "epoch": 1.5067691793414675, + "grad_norm": 0.028643961995840073, + "learning_rate": 0.00014019788696964614, + "loss": 0.7145, + "step": 9015 + }, + { + "epoch": 1.5092762828012702, + "grad_norm": 0.027849212288856506, + "learning_rate": 0.00014009726647660575, + "loss": 0.7056, + "step": 9030 + }, + { + "epoch": 1.511783386261073, + "grad_norm": 0.02838641032576561, + "learning_rate": 0.0001399966459835653, + "loss": 0.7171, + "step": 9045 + }, + { + "epoch": 1.5142904897208758, + "grad_norm": 0.028329892084002495, + "learning_rate": 0.00013989602549052492, + "loss": 0.7021, + "step": 9060 + }, + { + "epoch": 1.5167975931806787, + "grad_norm": 0.0278428103774786, + "learning_rate": 0.00013979540499748448, + "loss": 0.7008, + "step": 9075 + }, + { + "epoch": 1.5193046966404813, + "grad_norm": 0.029085583984851837, + "learning_rate": 0.0001396947845044441, + "loss": 0.7014, + "step": 9090 + }, + { + "epoch": 1.5218118001002843, + "grad_norm": 0.028230642899870872, + "learning_rate": 0.00013959416401140365, + "loss": 0.7028, + "step": 9105 + }, + { + "epoch": 1.5243189035600868, + "grad_norm": 0.02829892747104168, + "learning_rate": 0.00013949354351836327, + "loss": 0.7052, + "step": 9120 + }, + { + "epoch": 1.5268260070198898, + "grad_norm": 0.02769339270889759, + "learning_rate": 0.00013939292302532282, + "loss": 0.705, + "step": 9135 + }, + { + "epoch": 1.5293331104796923, + "grad_norm": 0.02728847600519657, + "learning_rate": 0.0001392923025322824, + "loss": 0.7129, + "step": 9150 + }, + { + "epoch": 1.5318402139394953, + "grad_norm": 0.029400475323200226, + "learning_rate": 0.000139191682039242, + "loss": 0.7076, + "step": 9165 + }, + { + "epoch": 1.5343473173992979, + "grad_norm": 0.02829390950500965, + "learning_rate": 0.00013909106154620158, + "loss": 0.7032, + "step": 9180 + }, + { + "epoch": 1.5368544208591008, + "grad_norm": 0.028629042208194733, + "learning_rate": 0.00013899044105316117, + "loss": 0.6992, + "step": 9195 + }, + { + "epoch": 1.5393615243189036, + "grad_norm": 0.028124259784817696, + "learning_rate": 0.00013888982056012075, + "loss": 0.6928, + "step": 9210 + }, + { + "epoch": 1.5418686277787064, + "grad_norm": 0.027618682011961937, + "learning_rate": 0.00013878920006708034, + "loss": 0.6988, + "step": 9225 + }, + { + "epoch": 1.5443757312385091, + "grad_norm": 0.028371086344122887, + "learning_rate": 0.00013868857957403992, + "loss": 0.7068, + "step": 9240 + }, + { + "epoch": 1.546882834698312, + "grad_norm": 0.02925163321197033, + "learning_rate": 0.0001385879590809995, + "loss": 0.7044, + "step": 9255 + }, + { + "epoch": 1.5493899381581147, + "grad_norm": 0.027992991730570793, + "learning_rate": 0.0001384873385879591, + "loss": 0.7147, + "step": 9270 + }, + { + "epoch": 1.5518970416179174, + "grad_norm": 0.02831142581999302, + "learning_rate": 0.00013838671809491868, + "loss": 0.711, + "step": 9285 + }, + { + "epoch": 1.5544041450777202, + "grad_norm": 0.027344243600964546, + "learning_rate": 0.00013828609760187826, + "loss": 0.7043, + "step": 9300 + }, + { + "epoch": 1.556911248537523, + "grad_norm": 0.027959240600466728, + "learning_rate": 0.00013818547710883785, + "loss": 0.715, + "step": 9315 + }, + { + "epoch": 1.5594183519973257, + "grad_norm": 0.0285944901406765, + "learning_rate": 0.0001380848566157974, + "loss": 0.7104, + "step": 9330 + }, + { + "epoch": 1.5619254554571285, + "grad_norm": 0.02860502153635025, + "learning_rate": 0.00013798423612275702, + "loss": 0.7053, + "step": 9345 + }, + { + "epoch": 1.5644325589169314, + "grad_norm": 0.028087912127375603, + "learning_rate": 0.00013788361562971658, + "loss": 0.715, + "step": 9360 + }, + { + "epoch": 1.566939662376734, + "grad_norm": 0.028339073061943054, + "learning_rate": 0.0001377829951366762, + "loss": 0.7035, + "step": 9375 + }, + { + "epoch": 1.569446765836537, + "grad_norm": 0.027878131717443466, + "learning_rate": 0.00013768237464363575, + "loss": 0.7072, + "step": 9390 + }, + { + "epoch": 1.5719538692963395, + "grad_norm": 0.028305955231189728, + "learning_rate": 0.00013758175415059536, + "loss": 0.6994, + "step": 9405 + }, + { + "epoch": 1.5744609727561425, + "grad_norm": 0.028195269405841827, + "learning_rate": 0.00013748113365755492, + "loss": 0.7172, + "step": 9420 + }, + { + "epoch": 1.576968076215945, + "grad_norm": 0.028301289305090904, + "learning_rate": 0.00013738051316451453, + "loss": 0.6958, + "step": 9435 + }, + { + "epoch": 1.579475179675748, + "grad_norm": 0.029125042259693146, + "learning_rate": 0.0001372798926714741, + "loss": 0.7004, + "step": 9450 + }, + { + "epoch": 1.5819822831355508, + "grad_norm": 0.02798408642411232, + "learning_rate": 0.00013717927217843368, + "loss": 0.6995, + "step": 9465 + }, + { + "epoch": 1.5844893865953535, + "grad_norm": 0.029614899307489395, + "learning_rate": 0.00013707865168539326, + "loss": 0.7057, + "step": 9480 + }, + { + "epoch": 1.5869964900551563, + "grad_norm": 0.0279951523989439, + "learning_rate": 0.00013697803119235285, + "loss": 0.6949, + "step": 9495 + }, + { + "epoch": 1.589503593514959, + "grad_norm": 0.028490344062447548, + "learning_rate": 0.00013687741069931243, + "loss": 0.7003, + "step": 9510 + }, + { + "epoch": 1.5920106969747618, + "grad_norm": 0.028360631316900253, + "learning_rate": 0.00013677679020627202, + "loss": 0.7008, + "step": 9525 + }, + { + "epoch": 1.5945178004345646, + "grad_norm": 0.029337970539927483, + "learning_rate": 0.0001366761697132316, + "loss": 0.7143, + "step": 9540 + }, + { + "epoch": 1.5970249038943674, + "grad_norm": 0.02845313400030136, + "learning_rate": 0.0001365755492201912, + "loss": 0.7006, + "step": 9555 + }, + { + "epoch": 1.5995320073541701, + "grad_norm": 0.027560876682400703, + "learning_rate": 0.00013647492872715078, + "loss": 0.7102, + "step": 9570 + }, + { + "epoch": 1.6020391108139729, + "grad_norm": 0.028155362233519554, + "learning_rate": 0.00013637430823411036, + "loss": 0.6979, + "step": 9585 + }, + { + "epoch": 1.6045462142737756, + "grad_norm": 0.029344851151108742, + "learning_rate": 0.00013627368774106995, + "loss": 0.704, + "step": 9600 + }, + { + "epoch": 1.6070533177335786, + "grad_norm": 0.02839244157075882, + "learning_rate": 0.0001361730672480295, + "loss": 0.6977, + "step": 9615 + }, + { + "epoch": 1.6095604211933812, + "grad_norm": 0.027915630489587784, + "learning_rate": 0.00013607244675498912, + "loss": 0.7086, + "step": 9630 + }, + { + "epoch": 1.6120675246531841, + "grad_norm": 0.02826772816479206, + "learning_rate": 0.00013597182626194868, + "loss": 0.6951, + "step": 9645 + }, + { + "epoch": 1.6145746281129867, + "grad_norm": 0.02916094847023487, + "learning_rate": 0.0001358712057689083, + "loss": 0.7103, + "step": 9660 + }, + { + "epoch": 1.6170817315727897, + "grad_norm": 0.02921309880912304, + "learning_rate": 0.00013577058527586785, + "loss": 0.6987, + "step": 9675 + }, + { + "epoch": 1.6195888350325922, + "grad_norm": 0.028561830520629883, + "learning_rate": 0.00013566996478282746, + "loss": 0.7119, + "step": 9690 + }, + { + "epoch": 1.6220959384923952, + "grad_norm": 0.028445105999708176, + "learning_rate": 0.00013556934428978702, + "loss": 0.7022, + "step": 9705 + }, + { + "epoch": 1.6246030419521977, + "grad_norm": 0.029156696051359177, + "learning_rate": 0.00013546872379674663, + "loss": 0.6946, + "step": 9720 + }, + { + "epoch": 1.6271101454120007, + "grad_norm": 0.029195377603173256, + "learning_rate": 0.0001353681033037062, + "loss": 0.6919, + "step": 9735 + }, + { + "epoch": 1.6296172488718035, + "grad_norm": 0.028340883553028107, + "learning_rate": 0.00013526748281066577, + "loss": 0.6949, + "step": 9750 + }, + { + "epoch": 1.6321243523316062, + "grad_norm": 0.028798367828130722, + "learning_rate": 0.00013516686231762536, + "loss": 0.6939, + "step": 9765 + }, + { + "epoch": 1.634631455791409, + "grad_norm": 0.028108691796660423, + "learning_rate": 0.00013506624182458494, + "loss": 0.6877, + "step": 9780 + }, + { + "epoch": 1.6371385592512118, + "grad_norm": 0.029803916811943054, + "learning_rate": 0.00013496562133154453, + "loss": 0.7063, + "step": 9795 + }, + { + "epoch": 1.6396456627110145, + "grad_norm": 0.02933133766055107, + "learning_rate": 0.00013486500083850412, + "loss": 0.7105, + "step": 9810 + }, + { + "epoch": 1.6421527661708173, + "grad_norm": 0.02795150876045227, + "learning_rate": 0.0001347643803454637, + "loss": 0.7127, + "step": 9825 + }, + { + "epoch": 1.64465986963062, + "grad_norm": 0.028160467743873596, + "learning_rate": 0.00013466375985242329, + "loss": 0.6962, + "step": 9840 + }, + { + "epoch": 1.6471669730904228, + "grad_norm": 0.028696995228528976, + "learning_rate": 0.00013456313935938287, + "loss": 0.7104, + "step": 9855 + }, + { + "epoch": 1.6496740765502258, + "grad_norm": 0.028448186814785004, + "learning_rate": 0.00013446251886634246, + "loss": 0.7164, + "step": 9870 + }, + { + "epoch": 1.6521811800100283, + "grad_norm": 0.028285130858421326, + "learning_rate": 0.00013436189837330204, + "loss": 0.6969, + "step": 9885 + }, + { + "epoch": 1.6546882834698313, + "grad_norm": 0.02930794097483158, + "learning_rate": 0.00013426127788026163, + "loss": 0.6933, + "step": 9900 + }, + { + "epoch": 1.6571953869296339, + "grad_norm": 0.028923654928803444, + "learning_rate": 0.00013416065738722121, + "loss": 0.7023, + "step": 9915 + }, + { + "epoch": 1.6597024903894368, + "grad_norm": 0.029697788879275322, + "learning_rate": 0.00013406003689418077, + "loss": 0.7149, + "step": 9930 + }, + { + "epoch": 1.6622095938492394, + "grad_norm": 0.02780589461326599, + "learning_rate": 0.00013395941640114038, + "loss": 0.7018, + "step": 9945 + }, + { + "epoch": 1.6647166973090424, + "grad_norm": 0.028592998161911964, + "learning_rate": 0.00013385879590809994, + "loss": 0.6999, + "step": 9960 + }, + { + "epoch": 1.667223800768845, + "grad_norm": 0.028748946264386177, + "learning_rate": 0.00013375817541505956, + "loss": 0.7108, + "step": 9975 + }, + { + "epoch": 1.669730904228648, + "grad_norm": 0.02883664146065712, + "learning_rate": 0.00013365755492201911, + "loss": 0.7014, + "step": 9990 + }, + { + "epoch": 1.6722380076884507, + "grad_norm": 0.027728645130991936, + "learning_rate": 0.00013355693442897873, + "loss": 0.7098, + "step": 10005 + }, + { + "epoch": 1.6747451111482534, + "grad_norm": 0.028445927426218987, + "learning_rate": 0.00013345631393593828, + "loss": 0.7016, + "step": 10020 + }, + { + "epoch": 1.6772522146080562, + "grad_norm": 0.029764369130134583, + "learning_rate": 0.00013335569344289787, + "loss": 0.6983, + "step": 10035 + }, + { + "epoch": 1.679759318067859, + "grad_norm": 0.029188336804509163, + "learning_rate": 0.00013325507294985746, + "loss": 0.7158, + "step": 10050 + }, + { + "epoch": 1.6822664215276617, + "grad_norm": 0.028241556137800217, + "learning_rate": 0.00013315445245681704, + "loss": 0.6923, + "step": 10065 + }, + { + "epoch": 1.6847735249874645, + "grad_norm": 0.02920147404074669, + "learning_rate": 0.00013305383196377663, + "loss": 0.7157, + "step": 10080 + }, + { + "epoch": 1.6872806284472672, + "grad_norm": 0.027919236570596695, + "learning_rate": 0.0001329532114707362, + "loss": 0.6984, + "step": 10095 + }, + { + "epoch": 1.68978773190707, + "grad_norm": 0.0279484074562788, + "learning_rate": 0.0001328525909776958, + "loss": 0.688, + "step": 10110 + }, + { + "epoch": 1.6922948353668728, + "grad_norm": 0.02801922895014286, + "learning_rate": 0.00013275197048465538, + "loss": 0.7077, + "step": 10125 + }, + { + "epoch": 1.6948019388266755, + "grad_norm": 0.02875382825732231, + "learning_rate": 0.00013265134999161497, + "loss": 0.7097, + "step": 10140 + }, + { + "epoch": 1.6973090422864785, + "grad_norm": 0.028978591784834862, + "learning_rate": 0.00013255072949857455, + "loss": 0.6921, + "step": 10155 + }, + { + "epoch": 1.699816145746281, + "grad_norm": 0.028557538986206055, + "learning_rate": 0.00013245010900553414, + "loss": 0.7124, + "step": 10170 + }, + { + "epoch": 1.702323249206084, + "grad_norm": 0.02763993851840496, + "learning_rate": 0.00013234948851249372, + "loss": 0.6998, + "step": 10185 + }, + { + "epoch": 1.7048303526658866, + "grad_norm": 0.029599042609333992, + "learning_rate": 0.0001322488680194533, + "loss": 0.7116, + "step": 10200 + }, + { + "epoch": 1.7073374561256895, + "grad_norm": 0.028568753972649574, + "learning_rate": 0.0001321482475264129, + "loss": 0.6927, + "step": 10215 + }, + { + "epoch": 1.709844559585492, + "grad_norm": 0.028803616762161255, + "learning_rate": 0.00013204762703337248, + "loss": 0.7, + "step": 10230 + }, + { + "epoch": 1.712351663045295, + "grad_norm": 0.028020154684782028, + "learning_rate": 0.00013194700654033204, + "loss": 0.7024, + "step": 10245 + }, + { + "epoch": 1.7148587665050978, + "grad_norm": 0.029931314289569855, + "learning_rate": 0.00013184638604729165, + "loss": 0.6996, + "step": 10260 + }, + { + "epoch": 1.7173658699649006, + "grad_norm": 0.028297219425439835, + "learning_rate": 0.0001317457655542512, + "loss": 0.7055, + "step": 10275 + }, + { + "epoch": 1.7198729734247034, + "grad_norm": 0.02956199459731579, + "learning_rate": 0.00013164514506121082, + "loss": 0.6975, + "step": 10290 + }, + { + "epoch": 1.7223800768845061, + "grad_norm": 0.027763094753026962, + "learning_rate": 0.00013154452456817038, + "loss": 0.7072, + "step": 10305 + }, + { + "epoch": 1.7248871803443089, + "grad_norm": 0.027571503072977066, + "learning_rate": 0.00013144390407512997, + "loss": 0.7001, + "step": 10320 + }, + { + "epoch": 1.7273942838041116, + "grad_norm": 0.028334425762295723, + "learning_rate": 0.00013134328358208955, + "loss": 0.7059, + "step": 10335 + }, + { + "epoch": 1.7299013872639144, + "grad_norm": 0.027847876772284508, + "learning_rate": 0.00013124266308904914, + "loss": 0.6956, + "step": 10350 + }, + { + "epoch": 1.7324084907237172, + "grad_norm": 0.027983665466308594, + "learning_rate": 0.00013114204259600872, + "loss": 0.716, + "step": 10365 + }, + { + "epoch": 1.73491559418352, + "grad_norm": 0.028772972524166107, + "learning_rate": 0.0001310414221029683, + "loss": 0.7052, + "step": 10380 + }, + { + "epoch": 1.7374226976433227, + "grad_norm": 0.028679322451353073, + "learning_rate": 0.0001309408016099279, + "loss": 0.6948, + "step": 10395 + }, + { + "epoch": 1.7399298011031257, + "grad_norm": 0.02946317568421364, + "learning_rate": 0.00013084018111688748, + "loss": 0.7048, + "step": 10410 + }, + { + "epoch": 1.7424369045629282, + "grad_norm": 0.0287346001714468, + "learning_rate": 0.00013073956062384707, + "loss": 0.7047, + "step": 10425 + }, + { + "epoch": 1.7449440080227312, + "grad_norm": 0.02862308919429779, + "learning_rate": 0.00013063894013080665, + "loss": 0.6886, + "step": 10440 + }, + { + "epoch": 1.7474511114825337, + "grad_norm": 0.0288804080337286, + "learning_rate": 0.00013053831963776624, + "loss": 0.706, + "step": 10455 + }, + { + "epoch": 1.7499582149423367, + "grad_norm": 0.0278554018586874, + "learning_rate": 0.00013043769914472582, + "loss": 0.7169, + "step": 10470 + }, + { + "epoch": 1.7524653184021393, + "grad_norm": 0.02842450514435768, + "learning_rate": 0.0001303370786516854, + "loss": 0.7028, + "step": 10485 + }, + { + "epoch": 1.7549724218619422, + "grad_norm": 0.02780633233487606, + "learning_rate": 0.000130236458158645, + "loss": 0.6974, + "step": 10500 + }, + { + "epoch": 1.7574795253217448, + "grad_norm": 0.028826531022787094, + "learning_rate": 0.00013013583766560458, + "loss": 0.7024, + "step": 10515 + }, + { + "epoch": 1.7599866287815478, + "grad_norm": 0.028399532660841942, + "learning_rate": 0.00013003521717256414, + "loss": 0.7029, + "step": 10530 + }, + { + "epoch": 1.7624937322413505, + "grad_norm": 0.029726563021540642, + "learning_rate": 0.00012993459667952375, + "loss": 0.7033, + "step": 10545 + }, + { + "epoch": 1.7650008357011533, + "grad_norm": 0.028318284079432487, + "learning_rate": 0.0001298339761864833, + "loss": 0.7102, + "step": 10560 + }, + { + "epoch": 1.767507939160956, + "grad_norm": 0.02865464985370636, + "learning_rate": 0.00012973335569344292, + "loss": 0.7079, + "step": 10575 + }, + { + "epoch": 1.7700150426207588, + "grad_norm": 0.029711904004216194, + "learning_rate": 0.00012963273520040248, + "loss": 0.7003, + "step": 10590 + }, + { + "epoch": 1.7725221460805616, + "grad_norm": 0.02868981659412384, + "learning_rate": 0.00012953211470736206, + "loss": 0.7071, + "step": 10605 + }, + { + "epoch": 1.7750292495403643, + "grad_norm": 0.03023667074739933, + "learning_rate": 0.00012943149421432165, + "loss": 0.6988, + "step": 10620 + }, + { + "epoch": 1.777536353000167, + "grad_norm": 0.02855963073670864, + "learning_rate": 0.00012933087372128123, + "loss": 0.7001, + "step": 10635 + }, + { + "epoch": 1.7800434564599699, + "grad_norm": 0.02811777964234352, + "learning_rate": 0.00012923025322824082, + "loss": 0.6982, + "step": 10650 + }, + { + "epoch": 1.7825505599197728, + "grad_norm": 0.029220616444945335, + "learning_rate": 0.0001291296327352004, + "loss": 0.7123, + "step": 10665 + }, + { + "epoch": 1.7850576633795754, + "grad_norm": 0.02945820614695549, + "learning_rate": 0.00012902901224216, + "loss": 0.702, + "step": 10680 + }, + { + "epoch": 1.7875647668393784, + "grad_norm": 0.02915896289050579, + "learning_rate": 0.00012892839174911958, + "loss": 0.6996, + "step": 10695 + }, + { + "epoch": 1.790071870299181, + "grad_norm": 0.028102731332182884, + "learning_rate": 0.00012882777125607916, + "loss": 0.6931, + "step": 10710 + }, + { + "epoch": 1.792578973758984, + "grad_norm": 0.028598302975296974, + "learning_rate": 0.00012872715076303875, + "loss": 0.7049, + "step": 10725 + }, + { + "epoch": 1.7950860772187864, + "grad_norm": 0.02882864698767662, + "learning_rate": 0.00012862653026999833, + "loss": 0.6894, + "step": 10740 + }, + { + "epoch": 1.7975931806785894, + "grad_norm": 0.02864612452685833, + "learning_rate": 0.00012852590977695792, + "loss": 0.6959, + "step": 10755 + }, + { + "epoch": 1.800100284138392, + "grad_norm": 0.02791963331401348, + "learning_rate": 0.0001284252892839175, + "loss": 0.699, + "step": 10770 + }, + { + "epoch": 1.802607387598195, + "grad_norm": 0.029228495433926582, + "learning_rate": 0.0001283246687908771, + "loss": 0.7014, + "step": 10785 + }, + { + "epoch": 1.8051144910579977, + "grad_norm": 0.028694583103060722, + "learning_rate": 0.00012822404829783667, + "loss": 0.6981, + "step": 10800 + }, + { + "epoch": 1.8076215945178005, + "grad_norm": 0.028723234310746193, + "learning_rate": 0.00012812342780479626, + "loss": 0.6935, + "step": 10815 + }, + { + "epoch": 1.8101286979776032, + "grad_norm": 0.02791297808289528, + "learning_rate": 0.00012802280731175585, + "loss": 0.7023, + "step": 10830 + }, + { + "epoch": 1.812635801437406, + "grad_norm": 0.028474239632487297, + "learning_rate": 0.0001279221868187154, + "loss": 0.6966, + "step": 10845 + }, + { + "epoch": 1.8151429048972088, + "grad_norm": 0.028216082602739334, + "learning_rate": 0.00012782156632567502, + "loss": 0.7146, + "step": 10860 + }, + { + "epoch": 1.8176500083570115, + "grad_norm": 0.02868053875863552, + "learning_rate": 0.00012772094583263458, + "loss": 0.7018, + "step": 10875 + }, + { + "epoch": 1.8201571118168143, + "grad_norm": 0.029623722657561302, + "learning_rate": 0.00012762032533959416, + "loss": 0.7024, + "step": 10890 + }, + { + "epoch": 1.822664215276617, + "grad_norm": 0.029195398092269897, + "learning_rate": 0.00012751970484655375, + "loss": 0.7056, + "step": 10905 + }, + { + "epoch": 1.8251713187364198, + "grad_norm": 0.02803465723991394, + "learning_rate": 0.00012741908435351333, + "loss": 0.7013, + "step": 10920 + }, + { + "epoch": 1.8276784221962226, + "grad_norm": 0.02818216383457184, + "learning_rate": 0.00012731846386047292, + "loss": 0.7052, + "step": 10935 + }, + { + "epoch": 1.8301855256560255, + "grad_norm": 0.029034661129117012, + "learning_rate": 0.0001272178433674325, + "loss": 0.6918, + "step": 10950 + }, + { + "epoch": 1.832692629115828, + "grad_norm": 0.028653794899582863, + "learning_rate": 0.0001271172228743921, + "loss": 0.708, + "step": 10965 + }, + { + "epoch": 1.835199732575631, + "grad_norm": 0.02844145894050598, + "learning_rate": 0.00012701660238135167, + "loss": 0.7065, + "step": 10980 + }, + { + "epoch": 1.8377068360354336, + "grad_norm": 0.02880460023880005, + "learning_rate": 0.00012691598188831126, + "loss": 0.6931, + "step": 10995 + }, + { + "epoch": 1.8402139394952366, + "grad_norm": 0.02845979668200016, + "learning_rate": 0.00012681536139527084, + "loss": 0.6924, + "step": 11010 + }, + { + "epoch": 1.8427210429550391, + "grad_norm": 0.02805483527481556, + "learning_rate": 0.00012671474090223043, + "loss": 0.7064, + "step": 11025 + }, + { + "epoch": 1.8452281464148421, + "grad_norm": 0.029036138206720352, + "learning_rate": 0.00012661412040919002, + "loss": 0.7061, + "step": 11040 + }, + { + "epoch": 1.8477352498746449, + "grad_norm": 0.028865808621048927, + "learning_rate": 0.0001265134999161496, + "loss": 0.7088, + "step": 11055 + }, + { + "epoch": 1.8502423533344476, + "grad_norm": 0.028568295761942863, + "learning_rate": 0.00012641287942310919, + "loss": 0.7032, + "step": 11070 + }, + { + "epoch": 1.8527494567942504, + "grad_norm": 0.02971578575670719, + "learning_rate": 0.00012631225893006877, + "loss": 0.703, + "step": 11085 + }, + { + "epoch": 1.8552565602540532, + "grad_norm": 0.029128948226571083, + "learning_rate": 0.00012621163843702836, + "loss": 0.6964, + "step": 11100 + }, + { + "epoch": 1.857763663713856, + "grad_norm": 0.028951995074748993, + "learning_rate": 0.00012611101794398794, + "loss": 0.6998, + "step": 11115 + }, + { + "epoch": 1.8602707671736587, + "grad_norm": 0.029678482562303543, + "learning_rate": 0.00012601039745094753, + "loss": 0.6778, + "step": 11130 + }, + { + "epoch": 1.8627778706334615, + "grad_norm": 0.029598036780953407, + "learning_rate": 0.0001259097769579071, + "loss": 0.694, + "step": 11145 + }, + { + "epoch": 1.8652849740932642, + "grad_norm": 0.02879234589636326, + "learning_rate": 0.00012580915646486667, + "loss": 0.7085, + "step": 11160 + }, + { + "epoch": 1.867792077553067, + "grad_norm": 0.029246920719742775, + "learning_rate": 0.00012570853597182626, + "loss": 0.6932, + "step": 11175 + }, + { + "epoch": 1.8702991810128697, + "grad_norm": 0.030359363183379173, + "learning_rate": 0.00012560791547878584, + "loss": 0.7027, + "step": 11190 + }, + { + "epoch": 1.8728062844726727, + "grad_norm": 0.02991410344839096, + "learning_rate": 0.00012550729498574543, + "loss": 0.6842, + "step": 11205 + }, + { + "epoch": 1.8753133879324753, + "grad_norm": 0.028199173510074615, + "learning_rate": 0.00012540667449270501, + "loss": 0.6998, + "step": 11220 + }, + { + "epoch": 1.8778204913922782, + "grad_norm": 0.028087392449378967, + "learning_rate": 0.0001253060539996646, + "loss": 0.697, + "step": 11235 + }, + { + "epoch": 1.8803275948520808, + "grad_norm": 0.02853637933731079, + "learning_rate": 0.00012520543350662418, + "loss": 0.6874, + "step": 11250 + }, + { + "epoch": 1.8828346983118838, + "grad_norm": 0.028400765731930733, + "learning_rate": 0.00012510481301358377, + "loss": 0.6881, + "step": 11265 + }, + { + "epoch": 1.8853418017716863, + "grad_norm": 0.02928781695663929, + "learning_rate": 0.00012500419252054336, + "loss": 0.6951, + "step": 11280 + }, + { + "epoch": 1.8878489052314893, + "grad_norm": 0.028838330879807472, + "learning_rate": 0.00012490357202750294, + "loss": 0.6857, + "step": 11295 + }, + { + "epoch": 1.8903560086912918, + "grad_norm": 0.0293565783649683, + "learning_rate": 0.00012480295153446253, + "loss": 0.693, + "step": 11310 + }, + { + "epoch": 1.8928631121510948, + "grad_norm": 0.02845110557973385, + "learning_rate": 0.0001247023310414221, + "loss": 0.6999, + "step": 11325 + }, + { + "epoch": 1.8953702156108976, + "grad_norm": 0.029096076264977455, + "learning_rate": 0.0001246017105483817, + "loss": 0.6841, + "step": 11340 + }, + { + "epoch": 1.8978773190707003, + "grad_norm": 0.029120532795786858, + "learning_rate": 0.00012450109005534128, + "loss": 0.7009, + "step": 11355 + }, + { + "epoch": 1.900384422530503, + "grad_norm": 0.027919389307498932, + "learning_rate": 0.00012440046956230087, + "loss": 0.708, + "step": 11370 + }, + { + "epoch": 1.9028915259903059, + "grad_norm": 0.02887488156557083, + "learning_rate": 0.00012429984906926045, + "loss": 0.7048, + "step": 11385 + }, + { + "epoch": 1.9053986294501086, + "grad_norm": 0.028664030134677887, + "learning_rate": 0.00012419922857622004, + "loss": 0.7003, + "step": 11400 + }, + { + "epoch": 1.9079057329099114, + "grad_norm": 0.028661739081144333, + "learning_rate": 0.00012409860808317962, + "loss": 0.7009, + "step": 11415 + }, + { + "epoch": 1.9104128363697142, + "grad_norm": 0.02937045879662037, + "learning_rate": 0.0001239979875901392, + "loss": 0.6935, + "step": 11430 + }, + { + "epoch": 1.912919939829517, + "grad_norm": 0.030395416542887688, + "learning_rate": 0.0001238973670970988, + "loss": 0.6808, + "step": 11445 + }, + { + "epoch": 1.91542704328932, + "grad_norm": 0.030018294230103493, + "learning_rate": 0.00012379674660405835, + "loss": 0.6931, + "step": 11460 + }, + { + "epoch": 1.9179341467491224, + "grad_norm": 0.029583923518657684, + "learning_rate": 0.00012369612611101794, + "loss": 0.6844, + "step": 11475 + }, + { + "epoch": 1.9204412502089254, + "grad_norm": 0.028469126671552658, + "learning_rate": 0.00012359550561797752, + "loss": 0.7, + "step": 11490 + }, + { + "epoch": 1.922948353668728, + "grad_norm": 0.029069840908050537, + "learning_rate": 0.0001234948851249371, + "loss": 0.6769, + "step": 11505 + }, + { + "epoch": 1.925455457128531, + "grad_norm": 0.03039330244064331, + "learning_rate": 0.0001233942646318967, + "loss": 0.701, + "step": 11520 + }, + { + "epoch": 1.9279625605883335, + "grad_norm": 0.029704933986067772, + "learning_rate": 0.00012329364413885628, + "loss": 0.6955, + "step": 11535 + }, + { + "epoch": 1.9304696640481365, + "grad_norm": 0.02861003205180168, + "learning_rate": 0.00012319302364581587, + "loss": 0.7029, + "step": 11550 + }, + { + "epoch": 1.932976767507939, + "grad_norm": 0.028516478836536407, + "learning_rate": 0.00012309240315277545, + "loss": 0.685, + "step": 11565 + }, + { + "epoch": 1.935483870967742, + "grad_norm": 0.02939150668680668, + "learning_rate": 0.00012299178265973504, + "loss": 0.6862, + "step": 11580 + }, + { + "epoch": 1.9379909744275448, + "grad_norm": 0.029078399762511253, + "learning_rate": 0.00012289116216669462, + "loss": 0.6915, + "step": 11595 + }, + { + "epoch": 1.9404980778873475, + "grad_norm": 0.02967904321849346, + "learning_rate": 0.0001227905416736542, + "loss": 0.7078, + "step": 11610 + }, + { + "epoch": 1.9430051813471503, + "grad_norm": 0.02986898459494114, + "learning_rate": 0.0001226899211806138, + "loss": 0.6962, + "step": 11625 + }, + { + "epoch": 1.945512284806953, + "grad_norm": 0.029141373932361603, + "learning_rate": 0.00012258930068757338, + "loss": 0.6898, + "step": 11640 + }, + { + "epoch": 1.9480193882667558, + "grad_norm": 0.02856113389134407, + "learning_rate": 0.00012248868019453296, + "loss": 0.6932, + "step": 11655 + }, + { + "epoch": 1.9505264917265586, + "grad_norm": 0.02906043641269207, + "learning_rate": 0.00012238805970149255, + "loss": 0.6947, + "step": 11670 + }, + { + "epoch": 1.9530335951863613, + "grad_norm": 0.028559362515807152, + "learning_rate": 0.00012228743920845214, + "loss": 0.6959, + "step": 11685 + }, + { + "epoch": 1.955540698646164, + "grad_norm": 0.029632238671183586, + "learning_rate": 0.00012218681871541172, + "loss": 0.7044, + "step": 11700 + }, + { + "epoch": 1.958047802105967, + "grad_norm": 0.028845706954598427, + "learning_rate": 0.0001220861982223713, + "loss": 0.6845, + "step": 11715 + }, + { + "epoch": 1.9605549055657696, + "grad_norm": 0.029171636328101158, + "learning_rate": 0.00012198557772933088, + "loss": 0.7044, + "step": 11730 + }, + { + "epoch": 1.9630620090255726, + "grad_norm": 0.030526766553521156, + "learning_rate": 0.00012188495723629045, + "loss": 0.6881, + "step": 11745 + }, + { + "epoch": 1.9655691124853751, + "grad_norm": 0.029202323406934738, + "learning_rate": 0.00012178433674325005, + "loss": 0.6853, + "step": 11760 + }, + { + "epoch": 1.9680762159451781, + "grad_norm": 0.028741231188178062, + "learning_rate": 0.00012168371625020962, + "loss": 0.7085, + "step": 11775 + }, + { + "epoch": 1.9705833194049807, + "grad_norm": 0.029565809294581413, + "learning_rate": 0.00012158309575716922, + "loss": 0.6951, + "step": 11790 + }, + { + "epoch": 1.9730904228647836, + "grad_norm": 0.029546387493610382, + "learning_rate": 0.00012148247526412879, + "loss": 0.6961, + "step": 11805 + }, + { + "epoch": 1.9755975263245862, + "grad_norm": 0.029062774032354355, + "learning_rate": 0.00012138185477108839, + "loss": 0.6933, + "step": 11820 + }, + { + "epoch": 1.9781046297843892, + "grad_norm": 0.028955336660146713, + "learning_rate": 0.00012128123427804796, + "loss": 0.6898, + "step": 11835 + }, + { + "epoch": 1.980611733244192, + "grad_norm": 0.031218407675623894, + "learning_rate": 0.00012118061378500756, + "loss": 0.689, + "step": 11850 + }, + { + "epoch": 1.9831188367039947, + "grad_norm": 0.030403736978769302, + "learning_rate": 0.00012107999329196713, + "loss": 0.6981, + "step": 11865 + }, + { + "epoch": 1.9856259401637975, + "grad_norm": 0.030305424705147743, + "learning_rate": 0.00012097937279892673, + "loss": 0.6987, + "step": 11880 + }, + { + "epoch": 1.9881330436236002, + "grad_norm": 0.029590345919132233, + "learning_rate": 0.0001208787523058863, + "loss": 0.6984, + "step": 11895 + }, + { + "epoch": 1.990640147083403, + "grad_norm": 0.04296644404530525, + "learning_rate": 0.00012077813181284589, + "loss": 0.7018, + "step": 11910 + }, + { + "epoch": 1.9931472505432057, + "grad_norm": 0.029970306903123856, + "learning_rate": 0.00012067751131980548, + "loss": 0.6887, + "step": 11925 + }, + { + "epoch": 1.9956543540030085, + "grad_norm": 0.02884749509394169, + "learning_rate": 0.00012057689082676506, + "loss": 0.7004, + "step": 11940 + }, + { + "epoch": 1.9981614574628113, + "grad_norm": 0.030533695593476295, + "learning_rate": 0.00012047627033372463, + "loss": 0.6883, + "step": 11955 + }, + { + "epoch": 2.0006685609226142, + "grad_norm": 0.029126284644007683, + "learning_rate": 0.00012037564984068423, + "loss": 0.6984, + "step": 11970 + }, + { + "epoch": 2.003175664382417, + "grad_norm": 0.029292147606611252, + "learning_rate": 0.0001202750293476438, + "loss": 0.6894, + "step": 11985 + }, + { + "epoch": 2.0056827678422198, + "grad_norm": 0.029509389773011208, + "learning_rate": 0.0001201744088546034, + "loss": 0.6823, + "step": 12000 + }, + { + "epoch": 2.0081898713020223, + "grad_norm": 0.02902618609368801, + "learning_rate": 0.00012007378836156298, + "loss": 0.6763, + "step": 12015 + }, + { + "epoch": 2.0106969747618253, + "grad_norm": 0.028685985133051872, + "learning_rate": 0.00011997316786852255, + "loss": 0.6903, + "step": 12030 + }, + { + "epoch": 2.013204078221628, + "grad_norm": 0.029849760234355927, + "learning_rate": 0.00011987254737548215, + "loss": 0.6886, + "step": 12045 + }, + { + "epoch": 2.015711181681431, + "grad_norm": 0.030097436159849167, + "learning_rate": 0.00011977192688244172, + "loss": 0.6868, + "step": 12060 + }, + { + "epoch": 2.0182182851412334, + "grad_norm": 0.02963315322995186, + "learning_rate": 0.00011967130638940132, + "loss": 0.6856, + "step": 12075 + }, + { + "epoch": 2.0207253886010363, + "grad_norm": 0.030087383463978767, + "learning_rate": 0.00011957068589636089, + "loss": 0.6886, + "step": 12090 + }, + { + "epoch": 2.023232492060839, + "grad_norm": 0.029318705201148987, + "learning_rate": 0.00011947006540332049, + "loss": 0.6734, + "step": 12105 + }, + { + "epoch": 2.025739595520642, + "grad_norm": 0.029196394607424736, + "learning_rate": 0.00011936944491028006, + "loss": 0.674, + "step": 12120 + }, + { + "epoch": 2.0282466989804444, + "grad_norm": 0.029127739369869232, + "learning_rate": 0.00011926882441723966, + "loss": 0.6875, + "step": 12135 + }, + { + "epoch": 2.0307538024402474, + "grad_norm": 0.029445838183164597, + "learning_rate": 0.00011916820392419923, + "loss": 0.6869, + "step": 12150 + }, + { + "epoch": 2.03326090590005, + "grad_norm": 0.029497170820832253, + "learning_rate": 0.00011906758343115883, + "loss": 0.6717, + "step": 12165 + }, + { + "epoch": 2.035768009359853, + "grad_norm": 0.028793711215257645, + "learning_rate": 0.0001189669629381184, + "loss": 0.682, + "step": 12180 + }, + { + "epoch": 2.038275112819656, + "grad_norm": 0.029894977807998657, + "learning_rate": 0.00011886634244507799, + "loss": 0.6821, + "step": 12195 + }, + { + "epoch": 2.0407822162794584, + "grad_norm": 0.028813883662223816, + "learning_rate": 0.00011876572195203757, + "loss": 0.6678, + "step": 12210 + }, + { + "epoch": 2.0432893197392614, + "grad_norm": 0.029816757887601852, + "learning_rate": 0.00011866510145899716, + "loss": 0.693, + "step": 12225 + }, + { + "epoch": 2.045796423199064, + "grad_norm": 0.03083239123225212, + "learning_rate": 0.00011856448096595673, + "loss": 0.681, + "step": 12240 + }, + { + "epoch": 2.048303526658867, + "grad_norm": 0.029679182916879654, + "learning_rate": 0.00011846386047291633, + "loss": 0.6742, + "step": 12255 + }, + { + "epoch": 2.0508106301186695, + "grad_norm": 0.03096550703048706, + "learning_rate": 0.0001183632399798759, + "loss": 0.6836, + "step": 12270 + }, + { + "epoch": 2.0533177335784725, + "grad_norm": 0.030012456700205803, + "learning_rate": 0.0001182626194868355, + "loss": 0.6819, + "step": 12285 + }, + { + "epoch": 2.055824837038275, + "grad_norm": 0.029759397730231285, + "learning_rate": 0.00011816199899379507, + "loss": 0.6781, + "step": 12300 + }, + { + "epoch": 2.058331940498078, + "grad_norm": 0.030046438798308372, + "learning_rate": 0.00011806137850075464, + "loss": 0.6787, + "step": 12315 + }, + { + "epoch": 2.0608390439578805, + "grad_norm": 0.02959163673222065, + "learning_rate": 0.00011796075800771424, + "loss": 0.6828, + "step": 12330 + }, + { + "epoch": 2.0633461474176835, + "grad_norm": 0.02911483868956566, + "learning_rate": 0.00011786013751467382, + "loss": 0.682, + "step": 12345 + }, + { + "epoch": 2.065853250877486, + "grad_norm": 0.04046880826354027, + "learning_rate": 0.00011775951702163341, + "loss": 0.6852, + "step": 12360 + }, + { + "epoch": 2.068360354337289, + "grad_norm": 0.030412757769227028, + "learning_rate": 0.00011765889652859299, + "loss": 0.6783, + "step": 12375 + }, + { + "epoch": 2.0708674577970916, + "grad_norm": 0.029883218929171562, + "learning_rate": 0.00011755827603555259, + "loss": 0.6774, + "step": 12390 + }, + { + "epoch": 2.0733745612568946, + "grad_norm": 0.029417937621474266, + "learning_rate": 0.00011745765554251216, + "loss": 0.691, + "step": 12405 + }, + { + "epoch": 2.075881664716697, + "grad_norm": 0.03051302768290043, + "learning_rate": 0.00011735703504947176, + "loss": 0.6871, + "step": 12420 + }, + { + "epoch": 2.0783887681765, + "grad_norm": 0.030459176748991013, + "learning_rate": 0.00011725641455643133, + "loss": 0.6892, + "step": 12435 + }, + { + "epoch": 2.0808958716363026, + "grad_norm": 0.030476195737719536, + "learning_rate": 0.00011715579406339093, + "loss": 0.6875, + "step": 12450 + }, + { + "epoch": 2.0834029750961056, + "grad_norm": 0.02982410229742527, + "learning_rate": 0.0001170551735703505, + "loss": 0.6623, + "step": 12465 + }, + { + "epoch": 2.0859100785559086, + "grad_norm": 0.030465099960565567, + "learning_rate": 0.0001169545530773101, + "loss": 0.6841, + "step": 12480 + }, + { + "epoch": 2.088417182015711, + "grad_norm": 0.029227489605545998, + "learning_rate": 0.00011685393258426967, + "loss": 0.6777, + "step": 12495 + }, + { + "epoch": 2.090924285475514, + "grad_norm": 0.029344556853175163, + "learning_rate": 0.00011675331209122926, + "loss": 0.6823, + "step": 12510 + }, + { + "epoch": 2.0934313889353167, + "grad_norm": 0.030551349744200706, + "learning_rate": 0.00011665269159818884, + "loss": 0.6872, + "step": 12525 + }, + { + "epoch": 2.0959384923951196, + "grad_norm": 0.03063136897981167, + "learning_rate": 0.00011655207110514843, + "loss": 0.6767, + "step": 12540 + }, + { + "epoch": 2.098445595854922, + "grad_norm": 0.02986333705484867, + "learning_rate": 0.000116451450612108, + "loss": 0.6941, + "step": 12555 + }, + { + "epoch": 2.100952699314725, + "grad_norm": 0.030152348801493645, + "learning_rate": 0.0001163508301190676, + "loss": 0.6832, + "step": 12570 + }, + { + "epoch": 2.1034598027745277, + "grad_norm": 0.029383687302470207, + "learning_rate": 0.00011625020962602717, + "loss": 0.6676, + "step": 12585 + }, + { + "epoch": 2.1059669062343307, + "grad_norm": 0.03019135817885399, + "learning_rate": 0.00011614958913298674, + "loss": 0.6735, + "step": 12600 + }, + { + "epoch": 2.1084740096941332, + "grad_norm": 0.030429605394601822, + "learning_rate": 0.00011604896863994634, + "loss": 0.6837, + "step": 12615 + }, + { + "epoch": 2.110981113153936, + "grad_norm": 0.031370870769023895, + "learning_rate": 0.00011594834814690591, + "loss": 0.6753, + "step": 12630 + }, + { + "epoch": 2.1134882166137388, + "grad_norm": 0.030195990577340126, + "learning_rate": 0.00011584772765386551, + "loss": 0.669, + "step": 12645 + }, + { + "epoch": 2.1159953200735417, + "grad_norm": 0.03015013597905636, + "learning_rate": 0.00011574710716082508, + "loss": 0.6868, + "step": 12660 + }, + { + "epoch": 2.1185024235333443, + "grad_norm": 0.030749835073947906, + "learning_rate": 0.00011564648666778468, + "loss": 0.6664, + "step": 12675 + }, + { + "epoch": 2.1210095269931473, + "grad_norm": 0.03003542125225067, + "learning_rate": 0.00011554586617474425, + "loss": 0.6884, + "step": 12690 + }, + { + "epoch": 2.12351663045295, + "grad_norm": 0.02948312647640705, + "learning_rate": 0.00011544524568170385, + "loss": 0.686, + "step": 12705 + }, + { + "epoch": 2.126023733912753, + "grad_norm": 0.03116905875504017, + "learning_rate": 0.00011534462518866342, + "loss": 0.6917, + "step": 12720 + }, + { + "epoch": 2.1285308373725558, + "grad_norm": 0.03057217039167881, + "learning_rate": 0.00011524400469562302, + "loss": 0.6893, + "step": 12735 + }, + { + "epoch": 2.1310379408323583, + "grad_norm": 0.03055824153125286, + "learning_rate": 0.0001151433842025826, + "loss": 0.6749, + "step": 12750 + }, + { + "epoch": 2.1335450442921613, + "grad_norm": 0.030194489285349846, + "learning_rate": 0.0001150427637095422, + "loss": 0.6841, + "step": 12765 + }, + { + "epoch": 2.136052147751964, + "grad_norm": 0.030030904337763786, + "learning_rate": 0.00011494214321650177, + "loss": 0.6755, + "step": 12780 + }, + { + "epoch": 2.138559251211767, + "grad_norm": 0.030531438067555428, + "learning_rate": 0.00011484152272346137, + "loss": 0.6885, + "step": 12795 + }, + { + "epoch": 2.1410663546715694, + "grad_norm": 0.031014693900942802, + "learning_rate": 0.00011474090223042094, + "loss": 0.6872, + "step": 12810 + }, + { + "epoch": 2.1435734581313723, + "grad_norm": 0.03255138173699379, + "learning_rate": 0.00011464028173738052, + "loss": 0.6811, + "step": 12825 + }, + { + "epoch": 2.146080561591175, + "grad_norm": 0.02984030731022358, + "learning_rate": 0.00011453966124434011, + "loss": 0.684, + "step": 12840 + }, + { + "epoch": 2.148587665050978, + "grad_norm": 0.03038971871137619, + "learning_rate": 0.0001144390407512997, + "loss": 0.6901, + "step": 12855 + }, + { + "epoch": 2.1510947685107804, + "grad_norm": 0.03030613623559475, + "learning_rate": 0.00011433842025825927, + "loss": 0.6815, + "step": 12870 + }, + { + "epoch": 2.1536018719705834, + "grad_norm": 0.03107587993144989, + "learning_rate": 0.00011423779976521885, + "loss": 0.6773, + "step": 12885 + }, + { + "epoch": 2.156108975430386, + "grad_norm": 0.030311092734336853, + "learning_rate": 0.00011413717927217844, + "loss": 0.6859, + "step": 12900 + }, + { + "epoch": 2.158616078890189, + "grad_norm": 0.03004043735563755, + "learning_rate": 0.00011403655877913801, + "loss": 0.677, + "step": 12915 + }, + { + "epoch": 2.1611231823499915, + "grad_norm": 0.02978183701634407, + "learning_rate": 0.00011393593828609761, + "loss": 0.686, + "step": 12930 + }, + { + "epoch": 2.1636302858097944, + "grad_norm": 0.030549898743629456, + "learning_rate": 0.00011383531779305718, + "loss": 0.6875, + "step": 12945 + }, + { + "epoch": 2.166137389269597, + "grad_norm": 0.030601589009165764, + "learning_rate": 0.00011373469730001678, + "loss": 0.6817, + "step": 12960 + }, + { + "epoch": 2.1686444927294, + "grad_norm": 0.030580811202526093, + "learning_rate": 0.00011363407680697635, + "loss": 0.6887, + "step": 12975 + }, + { + "epoch": 2.171151596189203, + "grad_norm": 0.030157998204231262, + "learning_rate": 0.00011353345631393595, + "loss": 0.693, + "step": 12990 + }, + { + "epoch": 2.1736586996490055, + "grad_norm": 0.03086373209953308, + "learning_rate": 0.00011343283582089552, + "loss": 0.6962, + "step": 13005 + }, + { + "epoch": 2.1761658031088085, + "grad_norm": 0.02979792095720768, + "learning_rate": 0.00011333221532785512, + "loss": 0.6756, + "step": 13020 + }, + { + "epoch": 2.178672906568611, + "grad_norm": 0.03019995242357254, + "learning_rate": 0.00011323159483481469, + "loss": 0.6951, + "step": 13035 + }, + { + "epoch": 2.181180010028414, + "grad_norm": 0.030209194868803024, + "learning_rate": 0.00011313097434177429, + "loss": 0.6904, + "step": 13050 + }, + { + "epoch": 2.1836871134882165, + "grad_norm": 0.030985839664936066, + "learning_rate": 0.00011303035384873386, + "loss": 0.6808, + "step": 13065 + }, + { + "epoch": 2.1861942169480195, + "grad_norm": 0.03027096390724182, + "learning_rate": 0.00011292973335569346, + "loss": 0.6892, + "step": 13080 + }, + { + "epoch": 2.188701320407822, + "grad_norm": 0.03128921985626221, + "learning_rate": 0.00011282911286265303, + "loss": 0.6841, + "step": 13095 + }, + { + "epoch": 2.191208423867625, + "grad_norm": 0.030639823526144028, + "learning_rate": 0.00011272849236961262, + "loss": 0.6708, + "step": 13110 + }, + { + "epoch": 2.1937155273274276, + "grad_norm": 0.030816158279776573, + "learning_rate": 0.0001126278718765722, + "loss": 0.6792, + "step": 13125 + }, + { + "epoch": 2.1962226307872306, + "grad_norm": 0.03019116260111332, + "learning_rate": 0.00011252725138353179, + "loss": 0.6777, + "step": 13140 + }, + { + "epoch": 2.198729734247033, + "grad_norm": 0.030292050912976265, + "learning_rate": 0.00011242663089049136, + "loss": 0.6999, + "step": 13155 + }, + { + "epoch": 2.201236837706836, + "grad_norm": 0.029916753992438316, + "learning_rate": 0.00011232601039745095, + "loss": 0.6784, + "step": 13170 + }, + { + "epoch": 2.2037439411666386, + "grad_norm": 0.029692910611629486, + "learning_rate": 0.00011222538990441053, + "loss": 0.6921, + "step": 13185 + }, + { + "epoch": 2.2062510446264416, + "grad_norm": 0.030788224190473557, + "learning_rate": 0.0001121247694113701, + "loss": 0.6714, + "step": 13200 + }, + { + "epoch": 2.208758148086244, + "grad_norm": 0.031961727887392044, + "learning_rate": 0.0001120241489183297, + "loss": 0.6818, + "step": 13215 + }, + { + "epoch": 2.211265251546047, + "grad_norm": 0.030589012429118156, + "learning_rate": 0.00011192352842528928, + "loss": 0.6863, + "step": 13230 + }, + { + "epoch": 2.21377235500585, + "grad_norm": 0.03072304092347622, + "learning_rate": 0.00011182290793224888, + "loss": 0.6854, + "step": 13245 + }, + { + "epoch": 2.2162794584656527, + "grad_norm": 0.030577028170228004, + "learning_rate": 0.00011172228743920845, + "loss": 0.6781, + "step": 13260 + }, + { + "epoch": 2.2187865619254556, + "grad_norm": 0.030161473900079727, + "learning_rate": 0.00011162166694616805, + "loss": 0.6824, + "step": 13275 + }, + { + "epoch": 2.221293665385258, + "grad_norm": 0.030237851664423943, + "learning_rate": 0.00011152104645312762, + "loss": 0.6808, + "step": 13290 + }, + { + "epoch": 2.223800768845061, + "grad_norm": 0.030910607427358627, + "learning_rate": 0.00011142042596008722, + "loss": 0.6819, + "step": 13305 + }, + { + "epoch": 2.2263078723048637, + "grad_norm": 0.03041113168001175, + "learning_rate": 0.00011131980546704679, + "loss": 0.6784, + "step": 13320 + }, + { + "epoch": 2.2288149757646667, + "grad_norm": 0.0322742834687233, + "learning_rate": 0.00011121918497400639, + "loss": 0.6695, + "step": 13335 + }, + { + "epoch": 2.2313220792244692, + "grad_norm": 0.03125980496406555, + "learning_rate": 0.00011111856448096596, + "loss": 0.681, + "step": 13350 + }, + { + "epoch": 2.233829182684272, + "grad_norm": 0.030773991718888283, + "learning_rate": 0.00011101794398792556, + "loss": 0.6867, + "step": 13365 + }, + { + "epoch": 2.2363362861440748, + "grad_norm": 0.03200787305831909, + "learning_rate": 0.00011091732349488513, + "loss": 0.691, + "step": 13380 + }, + { + "epoch": 2.2388433896038777, + "grad_norm": 0.03116571344435215, + "learning_rate": 0.00011081670300184473, + "loss": 0.671, + "step": 13395 + }, + { + "epoch": 2.2413504930636803, + "grad_norm": 0.031088994815945625, + "learning_rate": 0.0001107160825088043, + "loss": 0.6726, + "step": 13410 + }, + { + "epoch": 2.2438575965234833, + "grad_norm": 0.03130762279033661, + "learning_rate": 0.00011061546201576389, + "loss": 0.6948, + "step": 13425 + }, + { + "epoch": 2.246364699983286, + "grad_norm": 0.03147103264927864, + "learning_rate": 0.00011051484152272347, + "loss": 0.6778, + "step": 13440 + }, + { + "epoch": 2.248871803443089, + "grad_norm": 0.02998683787882328, + "learning_rate": 0.00011041422102968304, + "loss": 0.6996, + "step": 13455 + }, + { + "epoch": 2.2513789069028913, + "grad_norm": 0.03249230980873108, + "learning_rate": 0.00011031360053664263, + "loss": 0.6949, + "step": 13470 + }, + { + "epoch": 2.2538860103626943, + "grad_norm": 0.030694512650370598, + "learning_rate": 0.00011021298004360222, + "loss": 0.6806, + "step": 13485 + }, + { + "epoch": 2.2563931138224973, + "grad_norm": 0.0317358560860157, + "learning_rate": 0.0001101123595505618, + "loss": 0.6844, + "step": 13500 + }, + { + "epoch": 2.2589002172823, + "grad_norm": 0.029508093371987343, + "learning_rate": 0.00011001173905752137, + "loss": 0.6723, + "step": 13515 + }, + { + "epoch": 2.2614073207421024, + "grad_norm": 0.03101976215839386, + "learning_rate": 0.00010991111856448097, + "loss": 0.6689, + "step": 13530 + }, + { + "epoch": 2.2639144242019054, + "grad_norm": 0.030808012932538986, + "learning_rate": 0.00010981049807144054, + "loss": 0.6701, + "step": 13545 + }, + { + "epoch": 2.2664215276617083, + "grad_norm": 0.03057938627898693, + "learning_rate": 0.00010970987757840014, + "loss": 0.684, + "step": 13560 + }, + { + "epoch": 2.268928631121511, + "grad_norm": 0.03127751499414444, + "learning_rate": 0.00010960925708535971, + "loss": 0.6776, + "step": 13575 + }, + { + "epoch": 2.271435734581314, + "grad_norm": 0.02989344857633114, + "learning_rate": 0.00010950863659231931, + "loss": 0.6889, + "step": 13590 + }, + { + "epoch": 2.2739428380411164, + "grad_norm": 0.03043249435722828, + "learning_rate": 0.00010940801609927889, + "loss": 0.6794, + "step": 13605 + }, + { + "epoch": 2.2764499415009194, + "grad_norm": 0.030408738180994987, + "learning_rate": 0.00010930739560623848, + "loss": 0.6815, + "step": 13620 + }, + { + "epoch": 2.278957044960722, + "grad_norm": 0.030735976994037628, + "learning_rate": 0.00010920677511319806, + "loss": 0.6809, + "step": 13635 + }, + { + "epoch": 2.281464148420525, + "grad_norm": 0.0312831737101078, + "learning_rate": 0.00010910615462015766, + "loss": 0.6788, + "step": 13650 + }, + { + "epoch": 2.2839712518803275, + "grad_norm": 0.030336899682879448, + "learning_rate": 0.00010900553412711723, + "loss": 0.6737, + "step": 13665 + }, + { + "epoch": 2.2864783553401304, + "grad_norm": 0.030938081443309784, + "learning_rate": 0.00010890491363407683, + "loss": 0.6803, + "step": 13680 + }, + { + "epoch": 2.288985458799933, + "grad_norm": 0.02994300052523613, + "learning_rate": 0.0001088042931410364, + "loss": 0.6714, + "step": 13695 + }, + { + "epoch": 2.291492562259736, + "grad_norm": 0.03124346025288105, + "learning_rate": 0.000108703672647996, + "loss": 0.678, + "step": 13710 + }, + { + "epoch": 2.2939996657195385, + "grad_norm": 0.030526146292686462, + "learning_rate": 0.00010860305215495557, + "loss": 0.6815, + "step": 13725 + }, + { + "epoch": 2.2965067691793415, + "grad_norm": 0.03184838965535164, + "learning_rate": 0.00010850243166191514, + "loss": 0.6768, + "step": 13740 + }, + { + "epoch": 2.2990138726391445, + "grad_norm": 0.03009560890495777, + "learning_rate": 0.00010840181116887474, + "loss": 0.6768, + "step": 13755 + }, + { + "epoch": 2.301520976098947, + "grad_norm": 0.029740184545516968, + "learning_rate": 0.00010830119067583431, + "loss": 0.681, + "step": 13770 + }, + { + "epoch": 2.3040280795587496, + "grad_norm": 0.030534571036696434, + "learning_rate": 0.0001082005701827939, + "loss": 0.6739, + "step": 13785 + }, + { + "epoch": 2.3065351830185525, + "grad_norm": 0.030200140550732613, + "learning_rate": 0.00010809994968975348, + "loss": 0.6695, + "step": 13800 + }, + { + "epoch": 2.3090422864783555, + "grad_norm": 0.031782638281583786, + "learning_rate": 0.00010799932919671307, + "loss": 0.6866, + "step": 13815 + }, + { + "epoch": 2.311549389938158, + "grad_norm": 0.03087507374584675, + "learning_rate": 0.00010789870870367264, + "loss": 0.6717, + "step": 13830 + }, + { + "epoch": 2.314056493397961, + "grad_norm": 0.030710799619555473, + "learning_rate": 0.00010779808821063224, + "loss": 0.6882, + "step": 13845 + }, + { + "epoch": 2.3165635968577636, + "grad_norm": 0.030561743304133415, + "learning_rate": 0.00010769746771759181, + "loss": 0.6814, + "step": 13860 + }, + { + "epoch": 2.3190707003175666, + "grad_norm": 0.030251817777752876, + "learning_rate": 0.00010759684722455141, + "loss": 0.6747, + "step": 13875 + }, + { + "epoch": 2.321577803777369, + "grad_norm": 0.030898461118340492, + "learning_rate": 0.00010749622673151098, + "loss": 0.6678, + "step": 13890 + }, + { + "epoch": 2.324084907237172, + "grad_norm": 0.031910572201013565, + "learning_rate": 0.00010739560623847058, + "loss": 0.6873, + "step": 13905 + }, + { + "epoch": 2.3265920106969746, + "grad_norm": 0.031096691265702248, + "learning_rate": 0.00010729498574543015, + "loss": 0.6761, + "step": 13920 + }, + { + "epoch": 2.3290991141567776, + "grad_norm": 0.030930999666452408, + "learning_rate": 0.00010719436525238975, + "loss": 0.6842, + "step": 13935 + }, + { + "epoch": 2.33160621761658, + "grad_norm": 0.030477695167064667, + "learning_rate": 0.00010709374475934932, + "loss": 0.6784, + "step": 13950 + }, + { + "epoch": 2.334113321076383, + "grad_norm": 0.03102184645831585, + "learning_rate": 0.00010699312426630892, + "loss": 0.679, + "step": 13965 + }, + { + "epoch": 2.3366204245361857, + "grad_norm": 0.02999734878540039, + "learning_rate": 0.0001068925037732685, + "loss": 0.6786, + "step": 13980 + }, + { + "epoch": 2.3391275279959887, + "grad_norm": 0.030323563143610954, + "learning_rate": 0.0001067918832802281, + "loss": 0.6825, + "step": 13995 + }, + { + "epoch": 2.3416346314557916, + "grad_norm": 0.030984263867139816, + "learning_rate": 0.00010669126278718767, + "loss": 0.6798, + "step": 14010 + }, + { + "epoch": 2.344141734915594, + "grad_norm": 0.03151758387684822, + "learning_rate": 0.00010659064229414724, + "loss": 0.6821, + "step": 14025 + }, + { + "epoch": 2.3466488383753967, + "grad_norm": 0.03008199669420719, + "learning_rate": 0.00010649002180110684, + "loss": 0.6778, + "step": 14040 + }, + { + "epoch": 2.3491559418351997, + "grad_norm": 0.030592739582061768, + "learning_rate": 0.00010638940130806641, + "loss": 0.6629, + "step": 14055 + }, + { + "epoch": 2.3516630452950027, + "grad_norm": 0.030223028734326363, + "learning_rate": 0.000106288780815026, + "loss": 0.6807, + "step": 14070 + }, + { + "epoch": 2.3541701487548052, + "grad_norm": 0.03019655868411064, + "learning_rate": 0.00010618816032198558, + "loss": 0.6875, + "step": 14085 + }, + { + "epoch": 2.356677252214608, + "grad_norm": 0.03179163858294487, + "learning_rate": 0.00010608753982894517, + "loss": 0.6744, + "step": 14100 + }, + { + "epoch": 2.3591843556744108, + "grad_norm": 0.030132126063108444, + "learning_rate": 0.00010598691933590474, + "loss": 0.6768, + "step": 14115 + }, + { + "epoch": 2.3616914591342137, + "grad_norm": 0.03125820681452751, + "learning_rate": 0.00010588629884286434, + "loss": 0.6734, + "step": 14130 + }, + { + "epoch": 2.3641985625940163, + "grad_norm": 0.03128393739461899, + "learning_rate": 0.00010578567834982391, + "loss": 0.6643, + "step": 14145 + }, + { + "epoch": 2.3667056660538193, + "grad_norm": 0.031101234257221222, + "learning_rate": 0.00010568505785678351, + "loss": 0.6937, + "step": 14160 + }, + { + "epoch": 2.369212769513622, + "grad_norm": 0.03127965331077576, + "learning_rate": 0.00010558443736374308, + "loss": 0.6837, + "step": 14175 + }, + { + "epoch": 2.371719872973425, + "grad_norm": 0.03142804279923439, + "learning_rate": 0.00010548381687070268, + "loss": 0.6643, + "step": 14190 + }, + { + "epoch": 2.3742269764332273, + "grad_norm": 0.03196566551923752, + "learning_rate": 0.00010538319637766225, + "loss": 0.6737, + "step": 14205 + }, + { + "epoch": 2.3767340798930303, + "grad_norm": 0.03105044923722744, + "learning_rate": 0.00010528257588462185, + "loss": 0.6733, + "step": 14220 + }, + { + "epoch": 2.379241183352833, + "grad_norm": 0.030758565291762352, + "learning_rate": 0.00010518195539158142, + "loss": 0.6959, + "step": 14235 + }, + { + "epoch": 2.381748286812636, + "grad_norm": 0.03046661615371704, + "learning_rate": 0.00010508133489854102, + "loss": 0.6958, + "step": 14250 + }, + { + "epoch": 2.384255390272439, + "grad_norm": 0.03125166893005371, + "learning_rate": 0.00010498071440550059, + "loss": 0.6763, + "step": 14265 + }, + { + "epoch": 2.3867624937322414, + "grad_norm": 0.031636305153369904, + "learning_rate": 0.00010488009391246019, + "loss": 0.6794, + "step": 14280 + }, + { + "epoch": 2.389269597192044, + "grad_norm": 0.030563022941350937, + "learning_rate": 0.00010477947341941976, + "loss": 0.6874, + "step": 14295 + }, + { + "epoch": 2.391776700651847, + "grad_norm": 0.03061690181493759, + "learning_rate": 0.00010467885292637933, + "loss": 0.6782, + "step": 14310 + }, + { + "epoch": 2.39428380411165, + "grad_norm": 0.0308393444865942, + "learning_rate": 0.00010457823243333893, + "loss": 0.6777, + "step": 14325 + }, + { + "epoch": 2.3967909075714524, + "grad_norm": 0.030834507197141647, + "learning_rate": 0.0001044776119402985, + "loss": 0.6854, + "step": 14340 + }, + { + "epoch": 2.3992980110312554, + "grad_norm": 0.031078575178980827, + "learning_rate": 0.0001043769914472581, + "loss": 0.6844, + "step": 14355 + }, + { + "epoch": 2.401805114491058, + "grad_norm": 0.030426884070038795, + "learning_rate": 0.00010427637095421768, + "loss": 0.6701, + "step": 14370 + }, + { + "epoch": 2.404312217950861, + "grad_norm": 0.03103550709784031, + "learning_rate": 0.00010417575046117726, + "loss": 0.685, + "step": 14385 + }, + { + "epoch": 2.4068193214106635, + "grad_norm": 0.030895834788680077, + "learning_rate": 0.00010407512996813685, + "loss": 0.6798, + "step": 14400 + }, + { + "epoch": 2.4093264248704664, + "grad_norm": 0.029942205175757408, + "learning_rate": 0.00010397450947509643, + "loss": 0.6848, + "step": 14415 + }, + { + "epoch": 2.411833528330269, + "grad_norm": 0.03145187348127365, + "learning_rate": 0.000103873888982056, + "loss": 0.6745, + "step": 14430 + }, + { + "epoch": 2.414340631790072, + "grad_norm": 0.03102920390665531, + "learning_rate": 0.0001037732684890156, + "loss": 0.6963, + "step": 14445 + }, + { + "epoch": 2.4168477352498745, + "grad_norm": 0.030479585751891136, + "learning_rate": 0.00010367264799597518, + "loss": 0.6779, + "step": 14460 + }, + { + "epoch": 2.4193548387096775, + "grad_norm": 0.0313333161175251, + "learning_rate": 0.00010357202750293477, + "loss": 0.6675, + "step": 14475 + }, + { + "epoch": 2.42186194216948, + "grad_norm": 0.031193213537335396, + "learning_rate": 0.00010347140700989435, + "loss": 0.6709, + "step": 14490 + }, + { + "epoch": 2.424369045629283, + "grad_norm": 0.031854551285505295, + "learning_rate": 0.00010337078651685395, + "loss": 0.6832, + "step": 14505 + }, + { + "epoch": 2.4268761490890856, + "grad_norm": 0.03131631389260292, + "learning_rate": 0.00010327016602381352, + "loss": 0.6831, + "step": 14520 + }, + { + "epoch": 2.4293832525488885, + "grad_norm": 0.030897963792085648, + "learning_rate": 0.00010316954553077312, + "loss": 0.6779, + "step": 14535 + }, + { + "epoch": 2.431890356008691, + "grad_norm": 0.030229298397898674, + "learning_rate": 0.00010306892503773269, + "loss": 0.6928, + "step": 14550 + }, + { + "epoch": 2.434397459468494, + "grad_norm": 0.03158511593937874, + "learning_rate": 0.00010296830454469229, + "loss": 0.6812, + "step": 14565 + }, + { + "epoch": 2.436904562928297, + "grad_norm": 0.03185586631298065, + "learning_rate": 0.00010286768405165186, + "loss": 0.6707, + "step": 14580 + }, + { + "epoch": 2.4394116663880996, + "grad_norm": 0.03139151632785797, + "learning_rate": 0.00010276706355861143, + "loss": 0.6814, + "step": 14595 + }, + { + "epoch": 2.4419187698479026, + "grad_norm": 0.03182042017579079, + "learning_rate": 0.00010266644306557103, + "loss": 0.6663, + "step": 14610 + }, + { + "epoch": 2.444425873307705, + "grad_norm": 0.030850499868392944, + "learning_rate": 0.0001025658225725306, + "loss": 0.6937, + "step": 14625 + }, + { + "epoch": 2.446932976767508, + "grad_norm": 0.032495591789484024, + "learning_rate": 0.0001024652020794902, + "loss": 0.6588, + "step": 14640 + }, + { + "epoch": 2.4494400802273106, + "grad_norm": 0.03162992000579834, + "learning_rate": 0.00010236458158644977, + "loss": 0.6848, + "step": 14655 + }, + { + "epoch": 2.4519471836871136, + "grad_norm": 0.031871598213911057, + "learning_rate": 0.00010226396109340937, + "loss": 0.6743, + "step": 14670 + }, + { + "epoch": 2.454454287146916, + "grad_norm": 0.031383831053972244, + "learning_rate": 0.00010216334060036894, + "loss": 0.6861, + "step": 14685 + }, + { + "epoch": 2.456961390606719, + "grad_norm": 0.03176445513963699, + "learning_rate": 0.00010206272010732853, + "loss": 0.6702, + "step": 14700 + }, + { + "epoch": 2.4594684940665217, + "grad_norm": 0.03109871782362461, + "learning_rate": 0.00010196209961428812, + "loss": 0.6776, + "step": 14715 + }, + { + "epoch": 2.4619755975263247, + "grad_norm": 0.031003376469016075, + "learning_rate": 0.0001018614791212477, + "loss": 0.688, + "step": 14730 + }, + { + "epoch": 2.464482700986127, + "grad_norm": 0.031020162627100945, + "learning_rate": 0.00010176085862820727, + "loss": 0.6713, + "step": 14745 + }, + { + "epoch": 2.46698980444593, + "grad_norm": 0.031086094677448273, + "learning_rate": 0.00010166023813516687, + "loss": 0.6769, + "step": 14760 + }, + { + "epoch": 2.4694969079057327, + "grad_norm": 0.03022875264286995, + "learning_rate": 0.00010155961764212644, + "loss": 0.6807, + "step": 14775 + }, + { + "epoch": 2.4720040113655357, + "grad_norm": 0.030896877869963646, + "learning_rate": 0.00010145899714908604, + "loss": 0.6927, + "step": 14790 + }, + { + "epoch": 2.4745111148253383, + "grad_norm": 0.031297486275434494, + "learning_rate": 0.00010135837665604561, + "loss": 0.6827, + "step": 14805 + }, + { + "epoch": 2.4770182182851412, + "grad_norm": 0.03127811476588249, + "learning_rate": 0.00010125775616300521, + "loss": 0.6962, + "step": 14820 + }, + { + "epoch": 2.479525321744944, + "grad_norm": 0.030049098655581474, + "learning_rate": 0.00010115713566996479, + "loss": 0.6666, + "step": 14835 + }, + { + "epoch": 2.4820324252047468, + "grad_norm": 0.031142529100179672, + "learning_rate": 0.00010105651517692438, + "loss": 0.6787, + "step": 14850 + }, + { + "epoch": 2.4845395286645497, + "grad_norm": 0.031707145273685455, + "learning_rate": 0.00010095589468388396, + "loss": 0.6741, + "step": 14865 + }, + { + "epoch": 2.4870466321243523, + "grad_norm": 0.03133350983262062, + "learning_rate": 0.00010085527419084353, + "loss": 0.6695, + "step": 14880 + }, + { + "epoch": 2.4895537355841553, + "grad_norm": 0.031642328947782516, + "learning_rate": 0.00010075465369780313, + "loss": 0.6853, + "step": 14895 + }, + { + "epoch": 2.492060839043958, + "grad_norm": 0.03161296248435974, + "learning_rate": 0.0001006540332047627, + "loss": 0.673, + "step": 14910 + }, + { + "epoch": 2.494567942503761, + "grad_norm": 0.03102605603635311, + "learning_rate": 0.0001005534127117223, + "loss": 0.6819, + "step": 14925 + }, + { + "epoch": 2.4970750459635633, + "grad_norm": 0.031027935445308685, + "learning_rate": 0.00010045279221868187, + "loss": 0.6835, + "step": 14940 + }, + { + "epoch": 2.4995821494233663, + "grad_norm": 0.031037239357829094, + "learning_rate": 0.00010035217172564147, + "loss": 0.6695, + "step": 14955 + }, + { + "epoch": 2.502089252883169, + "grad_norm": 0.030962081626057625, + "learning_rate": 0.00010025155123260104, + "loss": 0.6808, + "step": 14970 + }, + { + "epoch": 2.504596356342972, + "grad_norm": 0.030871711671352386, + "learning_rate": 0.00010015093073956063, + "loss": 0.6799, + "step": 14985 + }, + { + "epoch": 2.5071034598027744, + "grad_norm": 0.03209908306598663, + "learning_rate": 0.00010005031024652021, + "loss": 0.6785, + "step": 15000 + }, + { + "epoch": 2.5096105632625774, + "grad_norm": 0.031665463000535965, + "learning_rate": 9.99496897534798e-05, + "loss": 0.6871, + "step": 15015 + }, + { + "epoch": 2.5121176667223803, + "grad_norm": 0.031626634299755096, + "learning_rate": 9.984906926043938e-05, + "loss": 0.6706, + "step": 15030 + }, + { + "epoch": 2.514624770182183, + "grad_norm": 0.03143932297825813, + "learning_rate": 9.974844876739895e-05, + "loss": 0.6776, + "step": 15045 + }, + { + "epoch": 2.5171318736419854, + "grad_norm": 0.03138510882854462, + "learning_rate": 9.964782827435854e-05, + "loss": 0.6832, + "step": 15060 + }, + { + "epoch": 2.5196389771017884, + "grad_norm": 0.030731745064258575, + "learning_rate": 9.954720778131813e-05, + "loss": 0.6728, + "step": 15075 + }, + { + "epoch": 2.5221460805615914, + "grad_norm": 0.03058742918074131, + "learning_rate": 9.944658728827771e-05, + "loss": 0.6916, + "step": 15090 + }, + { + "epoch": 2.524653184021394, + "grad_norm": 0.030874596908688545, + "learning_rate": 9.93459667952373e-05, + "loss": 0.6692, + "step": 15105 + }, + { + "epoch": 2.5271602874811965, + "grad_norm": 0.03069966472685337, + "learning_rate": 9.924534630219688e-05, + "loss": 0.6836, + "step": 15120 + }, + { + "epoch": 2.5296673909409995, + "grad_norm": 0.031031129881739616, + "learning_rate": 9.914472580915647e-05, + "loss": 0.6806, + "step": 15135 + }, + { + "epoch": 2.5321744944008024, + "grad_norm": 0.03190414234995842, + "learning_rate": 9.904410531611605e-05, + "loss": 0.677, + "step": 15150 + }, + { + "epoch": 2.534681597860605, + "grad_norm": 0.03230069950222969, + "learning_rate": 9.894348482307564e-05, + "loss": 0.6761, + "step": 15165 + }, + { + "epoch": 2.537188701320408, + "grad_norm": 0.03053051233291626, + "learning_rate": 9.884286433003522e-05, + "loss": 0.6805, + "step": 15180 + }, + { + "epoch": 2.5396958047802105, + "grad_norm": 0.03064662776887417, + "learning_rate": 9.874224383699481e-05, + "loss": 0.6854, + "step": 15195 + }, + { + "epoch": 2.5422029082400135, + "grad_norm": 0.03142537549138069, + "learning_rate": 9.86416233439544e-05, + "loss": 0.6693, + "step": 15210 + }, + { + "epoch": 2.544710011699816, + "grad_norm": 0.031185530126094818, + "learning_rate": 9.854100285091398e-05, + "loss": 0.6626, + "step": 15225 + }, + { + "epoch": 2.547217115159619, + "grad_norm": 0.03198733925819397, + "learning_rate": 9.844038235787357e-05, + "loss": 0.6847, + "step": 15240 + }, + { + "epoch": 2.5497242186194216, + "grad_norm": 0.03293673321604729, + "learning_rate": 9.833976186483315e-05, + "loss": 0.6792, + "step": 15255 + }, + { + "epoch": 2.5522313220792245, + "grad_norm": 0.03125865384936333, + "learning_rate": 9.823914137179274e-05, + "loss": 0.6728, + "step": 15270 + }, + { + "epoch": 2.554738425539027, + "grad_norm": 0.0312894769012928, + "learning_rate": 9.813852087875232e-05, + "loss": 0.6748, + "step": 15285 + }, + { + "epoch": 2.55724552899883, + "grad_norm": 0.03170843422412872, + "learning_rate": 9.80379003857119e-05, + "loss": 0.6674, + "step": 15300 + }, + { + "epoch": 2.5597526324586326, + "grad_norm": 0.031321533024311066, + "learning_rate": 9.793727989267148e-05, + "loss": 0.6797, + "step": 15315 + }, + { + "epoch": 2.5622597359184356, + "grad_norm": 0.031243357807397842, + "learning_rate": 9.783665939963107e-05, + "loss": 0.6752, + "step": 15330 + }, + { + "epoch": 2.5647668393782386, + "grad_norm": 0.03241657465696335, + "learning_rate": 9.773603890659064e-05, + "loss": 0.6851, + "step": 15345 + }, + { + "epoch": 2.567273942838041, + "grad_norm": 0.032917000353336334, + "learning_rate": 9.763541841355022e-05, + "loss": 0.6858, + "step": 15360 + }, + { + "epoch": 2.5697810462978437, + "grad_norm": 0.03208984062075615, + "learning_rate": 9.753479792050981e-05, + "loss": 0.6684, + "step": 15375 + }, + { + "epoch": 2.5722881497576466, + "grad_norm": 0.03123905509710312, + "learning_rate": 9.74341774274694e-05, + "loss": 0.6789, + "step": 15390 + }, + { + "epoch": 2.5747952532174496, + "grad_norm": 0.030513722449541092, + "learning_rate": 9.733355693442898e-05, + "loss": 0.6875, + "step": 15405 + }, + { + "epoch": 2.577302356677252, + "grad_norm": 0.03204507753252983, + "learning_rate": 9.723293644138856e-05, + "loss": 0.6742, + "step": 15420 + }, + { + "epoch": 2.579809460137055, + "grad_norm": 0.031124508008360863, + "learning_rate": 9.713231594834815e-05, + "loss": 0.6839, + "step": 15435 + }, + { + "epoch": 2.5823165635968577, + "grad_norm": 0.03063870221376419, + "learning_rate": 9.703169545530774e-05, + "loss": 0.6736, + "step": 15450 + }, + { + "epoch": 2.5848236670566607, + "grad_norm": 0.030677396804094315, + "learning_rate": 9.693107496226732e-05, + "loss": 0.6844, + "step": 15465 + }, + { + "epoch": 2.587330770516463, + "grad_norm": 0.03137551248073578, + "learning_rate": 9.68304544692269e-05, + "loss": 0.6763, + "step": 15480 + }, + { + "epoch": 2.589837873976266, + "grad_norm": 0.030652204528450966, + "learning_rate": 9.672983397618649e-05, + "loss": 0.6676, + "step": 15495 + }, + { + "epoch": 2.5923449774360687, + "grad_norm": 0.03098338656127453, + "learning_rate": 9.662921348314608e-05, + "loss": 0.681, + "step": 15510 + }, + { + "epoch": 2.5948520808958717, + "grad_norm": 0.030911816284060478, + "learning_rate": 9.652859299010566e-05, + "loss": 0.675, + "step": 15525 + }, + { + "epoch": 2.5973591843556743, + "grad_norm": 0.03055042400956154, + "learning_rate": 9.642797249706525e-05, + "loss": 0.6789, + "step": 15540 + }, + { + "epoch": 2.5998662878154772, + "grad_norm": 0.03084755130112171, + "learning_rate": 9.632735200402483e-05, + "loss": 0.6728, + "step": 15555 + }, + { + "epoch": 2.60237339127528, + "grad_norm": 0.03066328726708889, + "learning_rate": 9.622673151098442e-05, + "loss": 0.693, + "step": 15570 + }, + { + "epoch": 2.6048804947350828, + "grad_norm": 0.03215918317437172, + "learning_rate": 9.612611101794399e-05, + "loss": 0.6832, + "step": 15585 + }, + { + "epoch": 2.6073875981948857, + "grad_norm": 0.03187975287437439, + "learning_rate": 9.602549052490358e-05, + "loss": 0.6799, + "step": 15600 + }, + { + "epoch": 2.6098947016546883, + "grad_norm": 0.03179864585399628, + "learning_rate": 9.592487003186316e-05, + "loss": 0.6865, + "step": 15615 + }, + { + "epoch": 2.612401805114491, + "grad_norm": 0.032180171459913254, + "learning_rate": 9.582424953882275e-05, + "loss": 0.664, + "step": 15630 + }, + { + "epoch": 2.614908908574294, + "grad_norm": 0.03252346068620682, + "learning_rate": 9.572362904578232e-05, + "loss": 0.6686, + "step": 15645 + }, + { + "epoch": 2.617416012034097, + "grad_norm": 0.03194168955087662, + "learning_rate": 9.56230085527419e-05, + "loss": 0.6711, + "step": 15660 + }, + { + "epoch": 2.6199231154938993, + "grad_norm": 0.03153575584292412, + "learning_rate": 9.552238805970149e-05, + "loss": 0.6787, + "step": 15675 + }, + { + "epoch": 2.6224302189537023, + "grad_norm": 0.03099830634891987, + "learning_rate": 9.542176756666108e-05, + "loss": 0.6638, + "step": 15690 + }, + { + "epoch": 2.624937322413505, + "grad_norm": 0.032073475420475006, + "learning_rate": 9.532114707362066e-05, + "loss": 0.6867, + "step": 15705 + }, + { + "epoch": 2.627444425873308, + "grad_norm": 0.03117840364575386, + "learning_rate": 9.522052658058025e-05, + "loss": 0.6751, + "step": 15720 + }, + { + "epoch": 2.6299515293331104, + "grad_norm": 0.031706538051366806, + "learning_rate": 9.511990608753983e-05, + "loss": 0.6751, + "step": 15735 + }, + { + "epoch": 2.6324586327929134, + "grad_norm": 0.0310919638723135, + "learning_rate": 9.501928559449942e-05, + "loss": 0.6818, + "step": 15750 + }, + { + "epoch": 2.634965736252716, + "grad_norm": 0.032505493611097336, + "learning_rate": 9.4918665101459e-05, + "loss": 0.6762, + "step": 15765 + }, + { + "epoch": 2.637472839712519, + "grad_norm": 0.03129402920603752, + "learning_rate": 9.481804460841859e-05, + "loss": 0.6774, + "step": 15780 + }, + { + "epoch": 2.6399799431723214, + "grad_norm": 0.030791781842708588, + "learning_rate": 9.471742411537817e-05, + "loss": 0.6791, + "step": 15795 + }, + { + "epoch": 2.6424870466321244, + "grad_norm": 0.030598165467381477, + "learning_rate": 9.461680362233776e-05, + "loss": 0.6751, + "step": 15810 + }, + { + "epoch": 2.644994150091927, + "grad_norm": 0.03157910704612732, + "learning_rate": 9.451618312929734e-05, + "loss": 0.6661, + "step": 15825 + }, + { + "epoch": 2.64750125355173, + "grad_norm": 0.031462252140045166, + "learning_rate": 9.441556263625693e-05, + "loss": 0.6783, + "step": 15840 + }, + { + "epoch": 2.650008357011533, + "grad_norm": 0.031676456332206726, + "learning_rate": 9.431494214321652e-05, + "loss": 0.6838, + "step": 15855 + }, + { + "epoch": 2.6525154604713355, + "grad_norm": 0.031083036214113235, + "learning_rate": 9.421432165017609e-05, + "loss": 0.678, + "step": 15870 + }, + { + "epoch": 2.655022563931138, + "grad_norm": 0.03105340152978897, + "learning_rate": 9.411370115713567e-05, + "loss": 0.6786, + "step": 15885 + }, + { + "epoch": 2.657529667390941, + "grad_norm": 0.03212074562907219, + "learning_rate": 9.401308066409526e-05, + "loss": 0.6724, + "step": 15900 + }, + { + "epoch": 2.660036770850744, + "grad_norm": 0.03203478455543518, + "learning_rate": 9.391246017105484e-05, + "loss": 0.6664, + "step": 15915 + }, + { + "epoch": 2.6625438743105465, + "grad_norm": 0.03217902034521103, + "learning_rate": 9.381183967801443e-05, + "loss": 0.6668, + "step": 15930 + }, + { + "epoch": 2.665050977770349, + "grad_norm": 0.032049164175987244, + "learning_rate": 9.371121918497402e-05, + "loss": 0.6633, + "step": 15945 + }, + { + "epoch": 2.667558081230152, + "grad_norm": 0.03231196105480194, + "learning_rate": 9.361059869193359e-05, + "loss": 0.6759, + "step": 15960 + }, + { + "epoch": 2.670065184689955, + "grad_norm": 0.03290446102619171, + "learning_rate": 9.350997819889317e-05, + "loss": 0.6686, + "step": 15975 + }, + { + "epoch": 2.6725722881497576, + "grad_norm": 0.03090088628232479, + "learning_rate": 9.340935770585276e-05, + "loss": 0.6841, + "step": 15990 + }, + { + "epoch": 2.6750793916095605, + "grad_norm": 0.031320635229349136, + "learning_rate": 9.330873721281234e-05, + "loss": 0.6697, + "step": 16005 + }, + { + "epoch": 2.677586495069363, + "grad_norm": 0.03119390271604061, + "learning_rate": 9.320811671977193e-05, + "loss": 0.679, + "step": 16020 + }, + { + "epoch": 2.680093598529166, + "grad_norm": 0.031817544251680374, + "learning_rate": 9.310749622673151e-05, + "loss": 0.68, + "step": 16035 + }, + { + "epoch": 2.6826007019889686, + "grad_norm": 0.030589740723371506, + "learning_rate": 9.30068757336911e-05, + "loss": 0.6881, + "step": 16050 + }, + { + "epoch": 2.6851078054487716, + "grad_norm": 0.031363166868686676, + "learning_rate": 9.290625524065069e-05, + "loss": 0.6755, + "step": 16065 + }, + { + "epoch": 2.687614908908574, + "grad_norm": 0.03159747272729874, + "learning_rate": 9.280563474761027e-05, + "loss": 0.682, + "step": 16080 + }, + { + "epoch": 2.690122012368377, + "grad_norm": 0.03237079828977585, + "learning_rate": 9.270501425456986e-05, + "loss": 0.6808, + "step": 16095 + }, + { + "epoch": 2.69262911582818, + "grad_norm": 0.031845077872276306, + "learning_rate": 9.260439376152944e-05, + "loss": 0.6764, + "step": 16110 + }, + { + "epoch": 2.6951362192879826, + "grad_norm": 0.031239351257681847, + "learning_rate": 9.250377326848903e-05, + "loss": 0.6649, + "step": 16125 + }, + { + "epoch": 2.697643322747785, + "grad_norm": 0.031146762892603874, + "learning_rate": 9.240315277544861e-05, + "loss": 0.6953, + "step": 16140 + }, + { + "epoch": 2.700150426207588, + "grad_norm": 0.0323052816092968, + "learning_rate": 9.230253228240818e-05, + "loss": 0.672, + "step": 16155 + }, + { + "epoch": 2.702657529667391, + "grad_norm": 0.031430117785930634, + "learning_rate": 9.220191178936777e-05, + "loss": 0.6796, + "step": 16170 + }, + { + "epoch": 2.7051646331271937, + "grad_norm": 0.03176365792751312, + "learning_rate": 9.210129129632736e-05, + "loss": 0.6716, + "step": 16185 + }, + { + "epoch": 2.7076717365869962, + "grad_norm": 0.031570978462696075, + "learning_rate": 9.200067080328694e-05, + "loss": 0.6779, + "step": 16200 + }, + { + "epoch": 2.710178840046799, + "grad_norm": 0.031726591289043427, + "learning_rate": 9.190005031024653e-05, + "loss": 0.6739, + "step": 16215 + }, + { + "epoch": 2.712685943506602, + "grad_norm": 0.03140697255730629, + "learning_rate": 9.179942981720611e-05, + "loss": 0.6795, + "step": 16230 + }, + { + "epoch": 2.7151930469664047, + "grad_norm": 0.03162944316864014, + "learning_rate": 9.16988093241657e-05, + "loss": 0.6772, + "step": 16245 + }, + { + "epoch": 2.7177001504262077, + "grad_norm": 0.03275005519390106, + "learning_rate": 9.159818883112527e-05, + "loss": 0.6726, + "step": 16260 + }, + { + "epoch": 2.7202072538860103, + "grad_norm": 0.0312725305557251, + "learning_rate": 9.149756833808485e-05, + "loss": 0.6795, + "step": 16275 + }, + { + "epoch": 2.7227143573458132, + "grad_norm": 0.03128618001937866, + "learning_rate": 9.139694784504444e-05, + "loss": 0.6867, + "step": 16290 + }, + { + "epoch": 2.725221460805616, + "grad_norm": 0.03184065595269203, + "learning_rate": 9.129632735200403e-05, + "loss": 0.6737, + "step": 16305 + }, + { + "epoch": 2.7277285642654188, + "grad_norm": 0.03144819289445877, + "learning_rate": 9.119570685896361e-05, + "loss": 0.6684, + "step": 16320 + }, + { + "epoch": 2.7302356677252213, + "grad_norm": 0.03168636932969093, + "learning_rate": 9.10950863659232e-05, + "loss": 0.6759, + "step": 16335 + }, + { + "epoch": 2.7327427711850243, + "grad_norm": 0.03160136938095093, + "learning_rate": 9.099446587288278e-05, + "loss": 0.67, + "step": 16350 + }, + { + "epoch": 2.7352498746448273, + "grad_norm": 0.032716233283281326, + "learning_rate": 9.089384537984237e-05, + "loss": 0.6775, + "step": 16365 + }, + { + "epoch": 2.73775697810463, + "grad_norm": 0.033191412687301636, + "learning_rate": 9.079322488680195e-05, + "loss": 0.6882, + "step": 16380 + }, + { + "epoch": 2.7402640815644324, + "grad_norm": 0.03207962587475777, + "learning_rate": 9.069260439376154e-05, + "loss": 0.6784, + "step": 16395 + }, + { + "epoch": 2.7427711850242353, + "grad_norm": 0.031515009701251984, + "learning_rate": 9.059198390072112e-05, + "loss": 0.6741, + "step": 16410 + }, + { + "epoch": 2.7452782884840383, + "grad_norm": 0.03187147155404091, + "learning_rate": 9.049136340768071e-05, + "loss": 0.6662, + "step": 16425 + }, + { + "epoch": 2.747785391943841, + "grad_norm": 0.03254789486527443, + "learning_rate": 9.039074291464028e-05, + "loss": 0.6879, + "step": 16440 + }, + { + "epoch": 2.7502924954036434, + "grad_norm": 0.03185366839170456, + "learning_rate": 9.029012242159987e-05, + "loss": 0.6781, + "step": 16455 + }, + { + "epoch": 2.7527995988634464, + "grad_norm": 0.03274752199649811, + "learning_rate": 9.018950192855945e-05, + "loss": 0.6779, + "step": 16470 + }, + { + "epoch": 2.7553067023232494, + "grad_norm": 0.030197665095329285, + "learning_rate": 9.008888143551904e-05, + "loss": 0.6732, + "step": 16485 + }, + { + "epoch": 2.757813805783052, + "grad_norm": 0.03070506826043129, + "learning_rate": 8.998826094247862e-05, + "loss": 0.6713, + "step": 16500 + }, + { + "epoch": 2.760320909242855, + "grad_norm": 0.03231901675462723, + "learning_rate": 8.988764044943821e-05, + "loss": 0.6743, + "step": 16515 + }, + { + "epoch": 2.7628280127026574, + "grad_norm": 0.031823549419641495, + "learning_rate": 8.97870199563978e-05, + "loss": 0.6738, + "step": 16530 + }, + { + "epoch": 2.7653351161624604, + "grad_norm": 0.03237045556306839, + "learning_rate": 8.968639946335738e-05, + "loss": 0.6844, + "step": 16545 + }, + { + "epoch": 2.767842219622263, + "grad_norm": 0.033365171402692795, + "learning_rate": 8.958577897031695e-05, + "loss": 0.6778, + "step": 16560 + }, + { + "epoch": 2.770349323082066, + "grad_norm": 0.03203690052032471, + "learning_rate": 8.948515847727654e-05, + "loss": 0.6781, + "step": 16575 + }, + { + "epoch": 2.7728564265418685, + "grad_norm": 0.0312359556555748, + "learning_rate": 8.938453798423612e-05, + "loss": 0.6863, + "step": 16590 + }, + { + "epoch": 2.7753635300016715, + "grad_norm": 0.03242425248026848, + "learning_rate": 8.928391749119571e-05, + "loss": 0.6754, + "step": 16605 + }, + { + "epoch": 2.7778706334614744, + "grad_norm": 0.03151217848062515, + "learning_rate": 8.91832969981553e-05, + "loss": 0.6799, + "step": 16620 + }, + { + "epoch": 2.780377736921277, + "grad_norm": 0.032228607684373856, + "learning_rate": 8.908267650511488e-05, + "loss": 0.6709, + "step": 16635 + }, + { + "epoch": 2.7828848403810795, + "grad_norm": 0.03263266757130623, + "learning_rate": 8.898205601207446e-05, + "loss": 0.6833, + "step": 16650 + }, + { + "epoch": 2.7853919438408825, + "grad_norm": 0.030848173424601555, + "learning_rate": 8.888143551903405e-05, + "loss": 0.6802, + "step": 16665 + }, + { + "epoch": 2.7878990473006855, + "grad_norm": 0.03234275057911873, + "learning_rate": 8.878081502599364e-05, + "loss": 0.6716, + "step": 16680 + }, + { + "epoch": 2.790406150760488, + "grad_norm": 0.03131961077451706, + "learning_rate": 8.868019453295322e-05, + "loss": 0.6813, + "step": 16695 + }, + { + "epoch": 2.7929132542202906, + "grad_norm": 0.03362729772925377, + "learning_rate": 8.85795740399128e-05, + "loss": 0.677, + "step": 16710 + }, + { + "epoch": 2.7954203576800936, + "grad_norm": 0.03228291869163513, + "learning_rate": 8.847895354687238e-05, + "loss": 0.6737, + "step": 16725 + }, + { + "epoch": 2.7979274611398965, + "grad_norm": 0.031786005944013596, + "learning_rate": 8.837833305383196e-05, + "loss": 0.6859, + "step": 16740 + }, + { + "epoch": 2.800434564599699, + "grad_norm": 0.03170496225357056, + "learning_rate": 8.827771256079155e-05, + "loss": 0.6756, + "step": 16755 + }, + { + "epoch": 2.802941668059502, + "grad_norm": 0.033506058156490326, + "learning_rate": 8.817709206775113e-05, + "loss": 0.6758, + "step": 16770 + }, + { + "epoch": 2.8054487715193046, + "grad_norm": 0.032467689365148544, + "learning_rate": 8.807647157471072e-05, + "loss": 0.6745, + "step": 16785 + }, + { + "epoch": 2.8079558749791076, + "grad_norm": 0.033489979803562164, + "learning_rate": 8.79758510816703e-05, + "loss": 0.6676, + "step": 16800 + }, + { + "epoch": 2.81046297843891, + "grad_norm": 0.032214514911174774, + "learning_rate": 8.787523058862989e-05, + "loss": 0.6675, + "step": 16815 + }, + { + "epoch": 2.812970081898713, + "grad_norm": 0.03187641128897667, + "learning_rate": 8.777461009558948e-05, + "loss": 0.6863, + "step": 16830 + }, + { + "epoch": 2.8154771853585157, + "grad_norm": 0.031782276928424835, + "learning_rate": 8.767398960254906e-05, + "loss": 0.6782, + "step": 16845 + }, + { + "epoch": 2.8179842888183186, + "grad_norm": 0.031185677275061607, + "learning_rate": 8.757336910950865e-05, + "loss": 0.6769, + "step": 16860 + }, + { + "epoch": 2.820491392278121, + "grad_norm": 0.03163639456033707, + "learning_rate": 8.747274861646822e-05, + "loss": 0.6744, + "step": 16875 + }, + { + "epoch": 2.822998495737924, + "grad_norm": 0.031712062656879425, + "learning_rate": 8.73721281234278e-05, + "loss": 0.6741, + "step": 16890 + }, + { + "epoch": 2.8255055991977267, + "grad_norm": 0.03253958374261856, + "learning_rate": 8.727150763038739e-05, + "loss": 0.6844, + "step": 16905 + }, + { + "epoch": 2.8280127026575297, + "grad_norm": 0.03280916064977646, + "learning_rate": 8.717088713734698e-05, + "loss": 0.6782, + "step": 16920 + }, + { + "epoch": 2.8305198061173327, + "grad_norm": 0.03310822695493698, + "learning_rate": 8.707026664430656e-05, + "loss": 0.6704, + "step": 16935 + }, + { + "epoch": 2.833026909577135, + "grad_norm": 0.031092172488570213, + "learning_rate": 8.696964615126615e-05, + "loss": 0.676, + "step": 16950 + }, + { + "epoch": 2.8355340130369378, + "grad_norm": 0.0315091609954834, + "learning_rate": 8.686902565822573e-05, + "loss": 0.6658, + "step": 16965 + }, + { + "epoch": 2.8380411164967407, + "grad_norm": 0.030993249267339706, + "learning_rate": 8.676840516518532e-05, + "loss": 0.6771, + "step": 16980 + }, + { + "epoch": 2.8405482199565437, + "grad_norm": 0.03143613040447235, + "learning_rate": 8.66677846721449e-05, + "loss": 0.6887, + "step": 16995 + }, + { + "epoch": 2.8430553234163463, + "grad_norm": 0.03253776207566261, + "learning_rate": 8.656716417910447e-05, + "loss": 0.6684, + "step": 17010 + }, + { + "epoch": 2.8455624268761492, + "grad_norm": 0.03285781666636467, + "learning_rate": 8.646654368606406e-05, + "loss": 0.6827, + "step": 17025 + }, + { + "epoch": 2.848069530335952, + "grad_norm": 0.03159667178988457, + "learning_rate": 8.636592319302365e-05, + "loss": 0.6703, + "step": 17040 + }, + { + "epoch": 2.8505766337957548, + "grad_norm": 0.0313105471432209, + "learning_rate": 8.626530269998323e-05, + "loss": 0.6754, + "step": 17055 + }, + { + "epoch": 2.8530837372555573, + "grad_norm": 0.0318642258644104, + "learning_rate": 8.616468220694282e-05, + "loss": 0.6697, + "step": 17070 + }, + { + "epoch": 2.8555908407153603, + "grad_norm": 0.03128768131136894, + "learning_rate": 8.60640617139024e-05, + "loss": 0.6681, + "step": 17085 + }, + { + "epoch": 2.858097944175163, + "grad_norm": 0.03178677707910538, + "learning_rate": 8.596344122086199e-05, + "loss": 0.6903, + "step": 17100 + }, + { + "epoch": 2.860605047634966, + "grad_norm": 0.03252077102661133, + "learning_rate": 8.586282072782157e-05, + "loss": 0.6779, + "step": 17115 + }, + { + "epoch": 2.8631121510947684, + "grad_norm": 0.032303210347890854, + "learning_rate": 8.576220023478116e-05, + "loss": 0.6683, + "step": 17130 + }, + { + "epoch": 2.8656192545545713, + "grad_norm": 0.031926706433296204, + "learning_rate": 8.566157974174074e-05, + "loss": 0.6769, + "step": 17145 + }, + { + "epoch": 2.868126358014374, + "grad_norm": 0.032100748270750046, + "learning_rate": 8.556095924870033e-05, + "loss": 0.6829, + "step": 17160 + }, + { + "epoch": 2.870633461474177, + "grad_norm": 0.032134201377630234, + "learning_rate": 8.54603387556599e-05, + "loss": 0.6838, + "step": 17175 + }, + { + "epoch": 2.87314056493398, + "grad_norm": 0.03254568204283714, + "learning_rate": 8.535971826261949e-05, + "loss": 0.6905, + "step": 17190 + }, + { + "epoch": 2.8756476683937824, + "grad_norm": 0.03155257925391197, + "learning_rate": 8.525909776957907e-05, + "loss": 0.6799, + "step": 17205 + }, + { + "epoch": 2.878154771853585, + "grad_norm": 0.03186199814081192, + "learning_rate": 8.515847727653866e-05, + "loss": 0.6889, + "step": 17220 + }, + { + "epoch": 2.880661875313388, + "grad_norm": 0.03130493685603142, + "learning_rate": 8.505785678349824e-05, + "loss": 0.6713, + "step": 17235 + }, + { + "epoch": 2.883168978773191, + "grad_norm": 0.032139863818883896, + "learning_rate": 8.495723629045783e-05, + "loss": 0.6695, + "step": 17250 + }, + { + "epoch": 2.8856760822329934, + "grad_norm": 0.03158256411552429, + "learning_rate": 8.485661579741741e-05, + "loss": 0.6696, + "step": 17265 + }, + { + "epoch": 2.8881831856927964, + "grad_norm": 0.03188103437423706, + "learning_rate": 8.4755995304377e-05, + "loss": 0.6631, + "step": 17280 + }, + { + "epoch": 2.890690289152599, + "grad_norm": 0.032000407576560974, + "learning_rate": 8.465537481133657e-05, + "loss": 0.6736, + "step": 17295 + }, + { + "epoch": 2.893197392612402, + "grad_norm": 0.0315023809671402, + "learning_rate": 8.455475431829616e-05, + "loss": 0.6765, + "step": 17310 + }, + { + "epoch": 2.8957044960722045, + "grad_norm": 0.031286027282476425, + "learning_rate": 8.445413382525574e-05, + "loss": 0.6718, + "step": 17325 + }, + { + "epoch": 2.8982115995320075, + "grad_norm": 0.031370267271995544, + "learning_rate": 8.435351333221533e-05, + "loss": 0.6742, + "step": 17340 + }, + { + "epoch": 2.90071870299181, + "grad_norm": 0.032615795731544495, + "learning_rate": 8.425289283917491e-05, + "loss": 0.6756, + "step": 17355 + }, + { + "epoch": 2.903225806451613, + "grad_norm": 0.03155631199479103, + "learning_rate": 8.41522723461345e-05, + "loss": 0.6872, + "step": 17370 + }, + { + "epoch": 2.9057329099114155, + "grad_norm": 0.03252144530415535, + "learning_rate": 8.405165185309408e-05, + "loss": 0.6744, + "step": 17385 + }, + { + "epoch": 2.9082400133712185, + "grad_norm": 0.034432001411914825, + "learning_rate": 8.395103136005367e-05, + "loss": 0.6824, + "step": 17400 + }, + { + "epoch": 2.910747116831021, + "grad_norm": 0.033232349902391434, + "learning_rate": 8.385041086701326e-05, + "loss": 0.6837, + "step": 17415 + }, + { + "epoch": 2.913254220290824, + "grad_norm": 0.03268026188015938, + "learning_rate": 8.374979037397284e-05, + "loss": 0.6858, + "step": 17430 + }, + { + "epoch": 2.915761323750627, + "grad_norm": 0.03303677961230278, + "learning_rate": 8.364916988093243e-05, + "loss": 0.6793, + "step": 17445 + }, + { + "epoch": 2.9182684272104296, + "grad_norm": 0.03122582472860813, + "learning_rate": 8.354854938789201e-05, + "loss": 0.6728, + "step": 17460 + }, + { + "epoch": 2.920775530670232, + "grad_norm": 0.03166414424777031, + "learning_rate": 8.34479288948516e-05, + "loss": 0.6747, + "step": 17475 + }, + { + "epoch": 2.923282634130035, + "grad_norm": 0.031359609216451645, + "learning_rate": 8.334730840181117e-05, + "loss": 0.6644, + "step": 17490 + }, + { + "epoch": 2.925789737589838, + "grad_norm": 0.032412488013505936, + "learning_rate": 8.324668790877075e-05, + "loss": 0.6654, + "step": 17505 + }, + { + "epoch": 2.9282968410496406, + "grad_norm": 0.030932830646634102, + "learning_rate": 8.314606741573034e-05, + "loss": 0.6783, + "step": 17520 + }, + { + "epoch": 2.930803944509443, + "grad_norm": 0.03237373009324074, + "learning_rate": 8.304544692268993e-05, + "loss": 0.6835, + "step": 17535 + }, + { + "epoch": 2.933311047969246, + "grad_norm": 0.030989525839686394, + "learning_rate": 8.294482642964951e-05, + "loss": 0.6774, + "step": 17550 + }, + { + "epoch": 2.935818151429049, + "grad_norm": 0.031636081635951996, + "learning_rate": 8.28442059366091e-05, + "loss": 0.6845, + "step": 17565 + }, + { + "epoch": 2.9383252548888517, + "grad_norm": 0.032494526356458664, + "learning_rate": 8.274358544356867e-05, + "loss": 0.6761, + "step": 17580 + }, + { + "epoch": 2.9408323583486546, + "grad_norm": 0.03214458376169205, + "learning_rate": 8.264296495052825e-05, + "loss": 0.6837, + "step": 17595 + }, + { + "epoch": 2.943339461808457, + "grad_norm": 0.03136483207345009, + "learning_rate": 8.254234445748784e-05, + "loss": 0.6642, + "step": 17610 + }, + { + "epoch": 2.94584656526826, + "grad_norm": 0.032171837985515594, + "learning_rate": 8.244172396444742e-05, + "loss": 0.6667, + "step": 17625 + }, + { + "epoch": 2.9483536687280627, + "grad_norm": 0.032955460250377655, + "learning_rate": 8.234110347140701e-05, + "loss": 0.6789, + "step": 17640 + }, + { + "epoch": 2.9508607721878657, + "grad_norm": 0.03368501737713814, + "learning_rate": 8.22404829783666e-05, + "loss": 0.6711, + "step": 17655 + }, + { + "epoch": 2.9533678756476682, + "grad_norm": 0.03181430697441101, + "learning_rate": 8.213986248532618e-05, + "loss": 0.6751, + "step": 17670 + }, + { + "epoch": 2.955874979107471, + "grad_norm": 0.03176905959844589, + "learning_rate": 8.203924199228577e-05, + "loss": 0.6713, + "step": 17685 + }, + { + "epoch": 2.958382082567274, + "grad_norm": 0.03155883401632309, + "learning_rate": 8.193862149924535e-05, + "loss": 0.6721, + "step": 17700 + }, + { + "epoch": 2.9608891860270767, + "grad_norm": 0.03286443278193474, + "learning_rate": 8.183800100620494e-05, + "loss": 0.6669, + "step": 17715 + }, + { + "epoch": 2.9633962894868793, + "grad_norm": 0.03203749656677246, + "learning_rate": 8.173738051316452e-05, + "loss": 0.6821, + "step": 17730 + }, + { + "epoch": 2.9659033929466823, + "grad_norm": 0.03260009363293648, + "learning_rate": 8.163676002012411e-05, + "loss": 0.6785, + "step": 17745 + }, + { + "epoch": 2.9684104964064852, + "grad_norm": 0.034267883747816086, + "learning_rate": 8.15361395270837e-05, + "loss": 0.6699, + "step": 17760 + }, + { + "epoch": 2.970917599866288, + "grad_norm": 0.03159933537244797, + "learning_rate": 8.143551903404328e-05, + "loss": 0.6736, + "step": 17775 + }, + { + "epoch": 2.9734247033260903, + "grad_norm": 0.03262174129486084, + "learning_rate": 8.133489854100285e-05, + "loss": 0.6775, + "step": 17790 + }, + { + "epoch": 2.9759318067858933, + "grad_norm": 0.03155225142836571, + "learning_rate": 8.123427804796244e-05, + "loss": 0.6767, + "step": 17805 + }, + { + "epoch": 2.9784389102456963, + "grad_norm": 0.03178201615810394, + "learning_rate": 8.113365755492202e-05, + "loss": 0.6741, + "step": 17820 + }, + { + "epoch": 2.980946013705499, + "grad_norm": 0.03577803075313568, + "learning_rate": 8.103303706188161e-05, + "loss": 0.6763, + "step": 17835 + }, + { + "epoch": 2.983453117165302, + "grad_norm": 0.03188909962773323, + "learning_rate": 8.093241656884119e-05, + "loss": 0.6685, + "step": 17850 + }, + { + "epoch": 2.9859602206251044, + "grad_norm": 0.03267040103673935, + "learning_rate": 8.083179607580078e-05, + "loss": 0.6716, + "step": 17865 + }, + { + "epoch": 2.9884673240849073, + "grad_norm": 0.03385018929839134, + "learning_rate": 8.073117558276035e-05, + "loss": 0.679, + "step": 17880 + }, + { + "epoch": 2.99097442754471, + "grad_norm": 0.030928703024983406, + "learning_rate": 8.063055508971994e-05, + "loss": 0.681, + "step": 17895 + }, + { + "epoch": 2.993481531004513, + "grad_norm": 0.032874446362257004, + "learning_rate": 8.052993459667952e-05, + "loss": 0.6637, + "step": 17910 + }, + { + "epoch": 2.9959886344643154, + "grad_norm": 0.033117108047008514, + "learning_rate": 8.04293141036391e-05, + "loss": 0.663, + "step": 17925 + }, + { + "epoch": 2.9984957379241184, + "grad_norm": 0.03306007385253906, + "learning_rate": 8.032869361059869e-05, + "loss": 0.6693, + "step": 17940 + }, + { + "epoch": 3.001002841383921, + "grad_norm": 0.0331704318523407, + "learning_rate": 8.022807311755828e-05, + "loss": 0.6718, + "step": 17955 + }, + { + "epoch": 3.003509944843724, + "grad_norm": 0.03156450018286705, + "learning_rate": 8.012745262451786e-05, + "loss": 0.6553, + "step": 17970 + }, + { + "epoch": 3.0060170483035265, + "grad_norm": 0.033207476139068604, + "learning_rate": 8.002683213147745e-05, + "loss": 0.6632, + "step": 17985 + }, + { + "epoch": 3.0085241517633294, + "grad_norm": 0.03195161744952202, + "learning_rate": 7.992621163843703e-05, + "loss": 0.6705, + "step": 18000 + }, + { + "epoch": 3.0110312552231324, + "grad_norm": 0.03295569121837616, + "learning_rate": 7.982559114539662e-05, + "loss": 0.6648, + "step": 18015 + }, + { + "epoch": 3.013538358682935, + "grad_norm": 0.031562697142362595, + "learning_rate": 7.97249706523562e-05, + "loss": 0.6576, + "step": 18030 + }, + { + "epoch": 3.016045462142738, + "grad_norm": 0.031858429312705994, + "learning_rate": 7.962435015931579e-05, + "loss": 0.6528, + "step": 18045 + }, + { + "epoch": 3.0185525656025405, + "grad_norm": 0.03421582654118538, + "learning_rate": 7.952372966627538e-05, + "loss": 0.6714, + "step": 18060 + }, + { + "epoch": 3.0210596690623435, + "grad_norm": 0.03287555277347565, + "learning_rate": 7.942310917323496e-05, + "loss": 0.6788, + "step": 18075 + }, + { + "epoch": 3.023566772522146, + "grad_norm": 0.03256657347083092, + "learning_rate": 7.932248868019453e-05, + "loss": 0.6485, + "step": 18090 + }, + { + "epoch": 3.026073875981949, + "grad_norm": 0.03333086147904396, + "learning_rate": 7.922186818715412e-05, + "loss": 0.6678, + "step": 18105 + }, + { + "epoch": 3.0285809794417515, + "grad_norm": 0.031958311796188354, + "learning_rate": 7.91212476941137e-05, + "loss": 0.666, + "step": 18120 + }, + { + "epoch": 3.0310880829015545, + "grad_norm": 0.033307287842035294, + "learning_rate": 7.902062720107329e-05, + "loss": 0.6738, + "step": 18135 + }, + { + "epoch": 3.033595186361357, + "grad_norm": 0.031850751489400864, + "learning_rate": 7.892000670803288e-05, + "loss": 0.6615, + "step": 18150 + }, + { + "epoch": 3.03610228982116, + "grad_norm": 0.031476084142923355, + "learning_rate": 7.881938621499245e-05, + "loss": 0.6597, + "step": 18165 + }, + { + "epoch": 3.0386093932809626, + "grad_norm": 0.032822057604789734, + "learning_rate": 7.871876572195203e-05, + "loss": 0.649, + "step": 18180 + }, + { + "epoch": 3.0411164967407656, + "grad_norm": 0.032040949910879135, + "learning_rate": 7.861814522891162e-05, + "loss": 0.6708, + "step": 18195 + }, + { + "epoch": 3.043623600200568, + "grad_norm": 0.03377379849553108, + "learning_rate": 7.85175247358712e-05, + "loss": 0.6705, + "step": 18210 + }, + { + "epoch": 3.046130703660371, + "grad_norm": 0.03224708139896393, + "learning_rate": 7.841690424283079e-05, + "loss": 0.6685, + "step": 18225 + }, + { + "epoch": 3.0486378071201736, + "grad_norm": 0.03286907821893692, + "learning_rate": 7.831628374979037e-05, + "loss": 0.6785, + "step": 18240 + }, + { + "epoch": 3.0511449105799766, + "grad_norm": 0.033475641161203384, + "learning_rate": 7.821566325674996e-05, + "loss": 0.6581, + "step": 18255 + }, + { + "epoch": 3.0536520140397796, + "grad_norm": 0.03226190432906151, + "learning_rate": 7.811504276370955e-05, + "loss": 0.6679, + "step": 18270 + }, + { + "epoch": 3.056159117499582, + "grad_norm": 0.03244561329483986, + "learning_rate": 7.801442227066913e-05, + "loss": 0.6616, + "step": 18285 + }, + { + "epoch": 3.058666220959385, + "grad_norm": 0.03348153084516525, + "learning_rate": 7.791380177762872e-05, + "loss": 0.667, + "step": 18300 + }, + { + "epoch": 3.0611733244191877, + "grad_norm": 0.0332234688103199, + "learning_rate": 7.78131812845883e-05, + "loss": 0.6632, + "step": 18315 + }, + { + "epoch": 3.0636804278789906, + "grad_norm": 0.0322151854634285, + "learning_rate": 7.771256079154789e-05, + "loss": 0.6657, + "step": 18330 + }, + { + "epoch": 3.066187531338793, + "grad_norm": 0.033220209181308746, + "learning_rate": 7.761194029850747e-05, + "loss": 0.6694, + "step": 18345 + }, + { + "epoch": 3.068694634798596, + "grad_norm": 0.03340331092476845, + "learning_rate": 7.751131980546706e-05, + "loss": 0.6557, + "step": 18360 + }, + { + "epoch": 3.0712017382583987, + "grad_norm": 0.03259367495775223, + "learning_rate": 7.741069931242664e-05, + "loss": 0.6595, + "step": 18375 + }, + { + "epoch": 3.0737088417182017, + "grad_norm": 0.03241262957453728, + "learning_rate": 7.731007881938623e-05, + "loss": 0.6579, + "step": 18390 + }, + { + "epoch": 3.0762159451780042, + "grad_norm": 0.0317726731300354, + "learning_rate": 7.72094583263458e-05, + "loss": 0.6595, + "step": 18405 + }, + { + "epoch": 3.078723048637807, + "grad_norm": 0.0336771234869957, + "learning_rate": 7.710883783330539e-05, + "loss": 0.6563, + "step": 18420 + }, + { + "epoch": 3.0812301520976098, + "grad_norm": 0.033014651387929916, + "learning_rate": 7.700821734026497e-05, + "loss": 0.6689, + "step": 18435 + }, + { + "epoch": 3.0837372555574127, + "grad_norm": 0.03290229290723801, + "learning_rate": 7.690759684722454e-05, + "loss": 0.6482, + "step": 18450 + }, + { + "epoch": 3.0862443590172153, + "grad_norm": 0.033278971910476685, + "learning_rate": 7.680697635418413e-05, + "loss": 0.6532, + "step": 18465 + }, + { + "epoch": 3.0887514624770183, + "grad_norm": 0.03247194364666939, + "learning_rate": 7.670635586114371e-05, + "loss": 0.6455, + "step": 18480 + }, + { + "epoch": 3.091258565936821, + "grad_norm": 0.03357178345322609, + "learning_rate": 7.66057353681033e-05, + "loss": 0.6735, + "step": 18495 + }, + { + "epoch": 3.093765669396624, + "grad_norm": 0.033263131976127625, + "learning_rate": 7.650511487506289e-05, + "loss": 0.661, + "step": 18510 + }, + { + "epoch": 3.0962727728564268, + "grad_norm": 0.03202125430107117, + "learning_rate": 7.640449438202247e-05, + "loss": 0.6573, + "step": 18525 + }, + { + "epoch": 3.0987798763162293, + "grad_norm": 0.033367644995450974, + "learning_rate": 7.630387388898206e-05, + "loss": 0.6536, + "step": 18540 + }, + { + "epoch": 3.1012869797760323, + "grad_norm": 0.03215374797582626, + "learning_rate": 7.620325339594164e-05, + "loss": 0.6677, + "step": 18555 + }, + { + "epoch": 3.103794083235835, + "grad_norm": 0.03222740814089775, + "learning_rate": 7.610263290290123e-05, + "loss": 0.6451, + "step": 18570 + }, + { + "epoch": 3.106301186695638, + "grad_norm": 0.032317865639925, + "learning_rate": 7.600201240986081e-05, + "loss": 0.6477, + "step": 18585 + }, + { + "epoch": 3.1088082901554404, + "grad_norm": 0.03358441963791847, + "learning_rate": 7.59013919168204e-05, + "loss": 0.6586, + "step": 18600 + }, + { + "epoch": 3.1113153936152433, + "grad_norm": 0.03255462273955345, + "learning_rate": 7.580077142377998e-05, + "loss": 0.6626, + "step": 18615 + }, + { + "epoch": 3.113822497075046, + "grad_norm": 0.032852429896593094, + "learning_rate": 7.570015093073957e-05, + "loss": 0.6663, + "step": 18630 + }, + { + "epoch": 3.116329600534849, + "grad_norm": 0.033199895173311234, + "learning_rate": 7.559953043769915e-05, + "loss": 0.6647, + "step": 18645 + }, + { + "epoch": 3.1188367039946514, + "grad_norm": 0.03370612487196922, + "learning_rate": 7.549890994465874e-05, + "loss": 0.664, + "step": 18660 + }, + { + "epoch": 3.1213438074544544, + "grad_norm": 0.03296181559562683, + "learning_rate": 7.539828945161833e-05, + "loss": 0.6509, + "step": 18675 + }, + { + "epoch": 3.123850910914257, + "grad_norm": 0.03318094462156296, + "learning_rate": 7.529766895857791e-05, + "loss": 0.6782, + "step": 18690 + }, + { + "epoch": 3.12635801437406, + "grad_norm": 0.03345433250069618, + "learning_rate": 7.519704846553748e-05, + "loss": 0.6615, + "step": 18705 + }, + { + "epoch": 3.1288651178338625, + "grad_norm": 0.03299123793840408, + "learning_rate": 7.509642797249707e-05, + "loss": 0.6567, + "step": 18720 + }, + { + "epoch": 3.1313722212936654, + "grad_norm": 0.033207185566425323, + "learning_rate": 7.499580747945665e-05, + "loss": 0.6657, + "step": 18735 + }, + { + "epoch": 3.133879324753468, + "grad_norm": 0.03186638280749321, + "learning_rate": 7.489518698641623e-05, + "loss": 0.6651, + "step": 18750 + }, + { + "epoch": 3.136386428213271, + "grad_norm": 0.03253698721528053, + "learning_rate": 7.479456649337581e-05, + "loss": 0.6419, + "step": 18765 + }, + { + "epoch": 3.138893531673074, + "grad_norm": 0.03357692062854767, + "learning_rate": 7.46939460003354e-05, + "loss": 0.6591, + "step": 18780 + }, + { + "epoch": 3.1414006351328765, + "grad_norm": 0.03315422683954239, + "learning_rate": 7.459332550729498e-05, + "loss": 0.6542, + "step": 18795 + }, + { + "epoch": 3.143907738592679, + "grad_norm": 0.03280475363135338, + "learning_rate": 7.449270501425457e-05, + "loss": 0.6488, + "step": 18810 + }, + { + "epoch": 3.146414842052482, + "grad_norm": 0.032748933881521225, + "learning_rate": 7.439208452121415e-05, + "loss": 0.6448, + "step": 18825 + }, + { + "epoch": 3.148921945512285, + "grad_norm": 0.03311785310506821, + "learning_rate": 7.429146402817374e-05, + "loss": 0.6798, + "step": 18840 + }, + { + "epoch": 3.1514290489720875, + "grad_norm": 0.034732621163129807, + "learning_rate": 7.419084353513332e-05, + "loss": 0.6582, + "step": 18855 + }, + { + "epoch": 3.1539361524318905, + "grad_norm": 0.03301689773797989, + "learning_rate": 7.409022304209291e-05, + "loss": 0.6545, + "step": 18870 + }, + { + "epoch": 3.156443255891693, + "grad_norm": 0.03281566500663757, + "learning_rate": 7.39896025490525e-05, + "loss": 0.6566, + "step": 18885 + }, + { + "epoch": 3.158950359351496, + "grad_norm": 0.034080591052770615, + "learning_rate": 7.388898205601208e-05, + "loss": 0.668, + "step": 18900 + }, + { + "epoch": 3.1614574628112986, + "grad_norm": 0.032824501395225525, + "learning_rate": 7.378836156297167e-05, + "loss": 0.6654, + "step": 18915 + }, + { + "epoch": 3.1639645662711016, + "grad_norm": 0.03173629939556122, + "learning_rate": 7.368774106993125e-05, + "loss": 0.6692, + "step": 18930 + }, + { + "epoch": 3.166471669730904, + "grad_norm": 0.03352899104356766, + "learning_rate": 7.358712057689084e-05, + "loss": 0.6511, + "step": 18945 + }, + { + "epoch": 3.168978773190707, + "grad_norm": 0.033694177865982056, + "learning_rate": 7.348650008385042e-05, + "loss": 0.6541, + "step": 18960 + }, + { + "epoch": 3.1714858766505096, + "grad_norm": 0.033731088042259216, + "learning_rate": 7.338587959081001e-05, + "loss": 0.6731, + "step": 18975 + }, + { + "epoch": 3.1739929801103126, + "grad_norm": 0.03405210003256798, + "learning_rate": 7.32852590977696e-05, + "loss": 0.6664, + "step": 18990 + }, + { + "epoch": 3.176500083570115, + "grad_norm": 0.03264220058917999, + "learning_rate": 7.318463860472917e-05, + "loss": 0.6638, + "step": 19005 + }, + { + "epoch": 3.179007187029918, + "grad_norm": 0.03331288322806358, + "learning_rate": 7.308401811168875e-05, + "loss": 0.6601, + "step": 19020 + }, + { + "epoch": 3.1815142904897207, + "grad_norm": 0.032871656119823456, + "learning_rate": 7.298339761864834e-05, + "loss": 0.6582, + "step": 19035 + }, + { + "epoch": 3.1840213939495237, + "grad_norm": 0.033308811485767365, + "learning_rate": 7.288277712560792e-05, + "loss": 0.6526, + "step": 19050 + }, + { + "epoch": 3.186528497409326, + "grad_norm": 0.034691642969846725, + "learning_rate": 7.27821566325675e-05, + "loss": 0.6562, + "step": 19065 + }, + { + "epoch": 3.189035600869129, + "grad_norm": 0.03305482491850853, + "learning_rate": 7.268153613952708e-05, + "loss": 0.6647, + "step": 19080 + }, + { + "epoch": 3.191542704328932, + "grad_norm": 0.03342653810977936, + "learning_rate": 7.258091564648666e-05, + "loss": 0.6547, + "step": 19095 + }, + { + "epoch": 3.1940498077887347, + "grad_norm": 0.032857514917850494, + "learning_rate": 7.248029515344625e-05, + "loss": 0.6675, + "step": 19110 + }, + { + "epoch": 3.1965569112485377, + "grad_norm": 0.033686548471450806, + "learning_rate": 7.237967466040584e-05, + "loss": 0.665, + "step": 19125 + }, + { + "epoch": 3.1990640147083402, + "grad_norm": 0.0347943976521492, + "learning_rate": 7.227905416736542e-05, + "loss": 0.6663, + "step": 19140 + }, + { + "epoch": 3.201571118168143, + "grad_norm": 0.0345291905105114, + "learning_rate": 7.2178433674325e-05, + "loss": 0.6712, + "step": 19155 + }, + { + "epoch": 3.2040782216279458, + "grad_norm": 0.03408714756369591, + "learning_rate": 7.207781318128459e-05, + "loss": 0.675, + "step": 19170 + }, + { + "epoch": 3.2065853250877487, + "grad_norm": 0.03278841823339462, + "learning_rate": 7.197719268824418e-05, + "loss": 0.6694, + "step": 19185 + }, + { + "epoch": 3.2090924285475513, + "grad_norm": 0.032465532422065735, + "learning_rate": 7.187657219520376e-05, + "loss": 0.6606, + "step": 19200 + }, + { + "epoch": 3.2115995320073543, + "grad_norm": 0.03292189538478851, + "learning_rate": 7.177595170216335e-05, + "loss": 0.6602, + "step": 19215 + }, + { + "epoch": 3.214106635467157, + "grad_norm": 0.03233910724520683, + "learning_rate": 7.167533120912293e-05, + "loss": 0.6736, + "step": 19230 + }, + { + "epoch": 3.21661373892696, + "grad_norm": 0.03283598646521568, + "learning_rate": 7.157471071608252e-05, + "loss": 0.6561, + "step": 19245 + }, + { + "epoch": 3.2191208423867623, + "grad_norm": 0.03332465514540672, + "learning_rate": 7.14740902230421e-05, + "loss": 0.6666, + "step": 19260 + }, + { + "epoch": 3.2216279458465653, + "grad_norm": 0.033038314431905746, + "learning_rate": 7.137346973000169e-05, + "loss": 0.6615, + "step": 19275 + }, + { + "epoch": 3.224135049306368, + "grad_norm": 0.033269450068473816, + "learning_rate": 7.127284923696128e-05, + "loss": 0.6429, + "step": 19290 + }, + { + "epoch": 3.226642152766171, + "grad_norm": 0.03383258357644081, + "learning_rate": 7.117222874392085e-05, + "loss": 0.6564, + "step": 19305 + }, + { + "epoch": 3.2291492562259734, + "grad_norm": 0.03422423452138901, + "learning_rate": 7.107160825088043e-05, + "loss": 0.6626, + "step": 19320 + }, + { + "epoch": 3.2316563596857764, + "grad_norm": 0.03328223526477814, + "learning_rate": 7.097098775784002e-05, + "loss": 0.6662, + "step": 19335 + }, + { + "epoch": 3.2341634631455793, + "grad_norm": 0.06695165485143661, + "learning_rate": 7.08703672647996e-05, + "loss": 0.6638, + "step": 19350 + }, + { + "epoch": 3.236670566605382, + "grad_norm": 0.034399405121803284, + "learning_rate": 7.076974677175918e-05, + "loss": 0.6696, + "step": 19365 + }, + { + "epoch": 3.239177670065185, + "grad_norm": 0.03373480588197708, + "learning_rate": 7.066912627871876e-05, + "loss": 0.6567, + "step": 19380 + }, + { + "epoch": 3.2416847735249874, + "grad_norm": 0.03468296676874161, + "learning_rate": 7.056850578567835e-05, + "loss": 0.6693, + "step": 19395 + }, + { + "epoch": 3.2441918769847904, + "grad_norm": 0.03372135013341904, + "learning_rate": 7.046788529263793e-05, + "loss": 0.6598, + "step": 19410 + }, + { + "epoch": 3.246698980444593, + "grad_norm": 0.033345624804496765, + "learning_rate": 7.036726479959752e-05, + "loss": 0.6618, + "step": 19425 + }, + { + "epoch": 3.249206083904396, + "grad_norm": 0.033639825880527496, + "learning_rate": 7.02666443065571e-05, + "loss": 0.675, + "step": 19440 + }, + { + "epoch": 3.2517131873641985, + "grad_norm": 0.032892145216464996, + "learning_rate": 7.016602381351669e-05, + "loss": 0.6725, + "step": 19455 + }, + { + "epoch": 3.2542202908240014, + "grad_norm": 0.03500252589583397, + "learning_rate": 7.006540332047627e-05, + "loss": 0.6648, + "step": 19470 + }, + { + "epoch": 3.256727394283804, + "grad_norm": 0.033187173306941986, + "learning_rate": 6.996478282743586e-05, + "loss": 0.6574, + "step": 19485 + }, + { + "epoch": 3.259234497743607, + "grad_norm": 0.033664412796497345, + "learning_rate": 6.986416233439545e-05, + "loss": 0.6555, + "step": 19500 + }, + { + "epoch": 3.2617416012034095, + "grad_norm": 0.031998805701732635, + "learning_rate": 6.976354184135503e-05, + "loss": 0.6621, + "step": 19515 + }, + { + "epoch": 3.2642487046632125, + "grad_norm": 0.03370664268732071, + "learning_rate": 6.966292134831462e-05, + "loss": 0.6575, + "step": 19530 + }, + { + "epoch": 3.266755808123015, + "grad_norm": 0.03247015178203583, + "learning_rate": 6.95623008552742e-05, + "loss": 0.6704, + "step": 19545 + }, + { + "epoch": 3.269262911582818, + "grad_norm": 0.03311055153608322, + "learning_rate": 6.946168036223379e-05, + "loss": 0.6616, + "step": 19560 + }, + { + "epoch": 3.2717700150426205, + "grad_norm": 0.0337141677737236, + "learning_rate": 6.936105986919337e-05, + "loss": 0.6755, + "step": 19575 + }, + { + "epoch": 3.2742771185024235, + "grad_norm": 0.03368309885263443, + "learning_rate": 6.926043937615294e-05, + "loss": 0.6641, + "step": 19590 + }, + { + "epoch": 3.2767842219622265, + "grad_norm": 0.035854946821928024, + "learning_rate": 6.915981888311253e-05, + "loss": 0.6417, + "step": 19605 + }, + { + "epoch": 3.279291325422029, + "grad_norm": 0.03423641249537468, + "learning_rate": 6.905919839007212e-05, + "loss": 0.6617, + "step": 19620 + }, + { + "epoch": 3.281798428881832, + "grad_norm": 0.033775344491004944, + "learning_rate": 6.89585778970317e-05, + "loss": 0.6636, + "step": 19635 + }, + { + "epoch": 3.2843055323416346, + "grad_norm": 0.03380570188164711, + "learning_rate": 6.885795740399129e-05, + "loss": 0.6803, + "step": 19650 + }, + { + "epoch": 3.2868126358014376, + "grad_norm": 0.03310622647404671, + "learning_rate": 6.875733691095086e-05, + "loss": 0.6565, + "step": 19665 + }, + { + "epoch": 3.28931973926124, + "grad_norm": 0.033229805529117584, + "learning_rate": 6.865671641791044e-05, + "loss": 0.6745, + "step": 19680 + }, + { + "epoch": 3.291826842721043, + "grad_norm": 0.032646749168634415, + "learning_rate": 6.855609592487003e-05, + "loss": 0.6621, + "step": 19695 + }, + { + "epoch": 3.2943339461808456, + "grad_norm": 0.03353268280625343, + "learning_rate": 6.845547543182961e-05, + "loss": 0.6591, + "step": 19710 + }, + { + "epoch": 3.2968410496406486, + "grad_norm": 0.03348655626177788, + "learning_rate": 6.83548549387892e-05, + "loss": 0.6595, + "step": 19725 + }, + { + "epoch": 3.299348153100451, + "grad_norm": 0.034753601998090744, + "learning_rate": 6.825423444574879e-05, + "loss": 0.6577, + "step": 19740 + }, + { + "epoch": 3.301855256560254, + "grad_norm": 0.03206339105963707, + "learning_rate": 6.815361395270837e-05, + "loss": 0.6532, + "step": 19755 + }, + { + "epoch": 3.3043623600200567, + "grad_norm": 0.03383897244930267, + "learning_rate": 6.805299345966796e-05, + "loss": 0.6582, + "step": 19770 + }, + { + "epoch": 3.3068694634798597, + "grad_norm": 0.03341525420546532, + "learning_rate": 6.795237296662754e-05, + "loss": 0.655, + "step": 19785 + }, + { + "epoch": 3.309376566939662, + "grad_norm": 0.033895961940288544, + "learning_rate": 6.785175247358713e-05, + "loss": 0.6635, + "step": 19800 + }, + { + "epoch": 3.311883670399465, + "grad_norm": 1.8668147325515747, + "learning_rate": 6.775113198054671e-05, + "loss": 0.6677, + "step": 19815 + }, + { + "epoch": 3.3143907738592677, + "grad_norm": 0.03255158290266991, + "learning_rate": 6.76505114875063e-05, + "loss": 0.6563, + "step": 19830 + }, + { + "epoch": 3.3168978773190707, + "grad_norm": 0.03239602968096733, + "learning_rate": 6.754989099446588e-05, + "loss": 0.6564, + "step": 19845 + }, + { + "epoch": 3.3194049807788737, + "grad_norm": 0.032781895250082016, + "learning_rate": 6.744927050142547e-05, + "loss": 0.6686, + "step": 19860 + }, + { + "epoch": 3.3219120842386762, + "grad_norm": 0.03344454988837242, + "learning_rate": 6.734865000838504e-05, + "loss": 0.6584, + "step": 19875 + }, + { + "epoch": 3.324419187698479, + "grad_norm": 0.0336369127035141, + "learning_rate": 6.724802951534463e-05, + "loss": 0.6561, + "step": 19890 + }, + { + "epoch": 3.3269262911582818, + "grad_norm": 0.0323907732963562, + "learning_rate": 6.714740902230421e-05, + "loss": 0.6587, + "step": 19905 + }, + { + "epoch": 3.3294333946180847, + "grad_norm": 0.03296393156051636, + "learning_rate": 6.70467885292638e-05, + "loss": 0.6666, + "step": 19920 + }, + { + "epoch": 3.3319404980778873, + "grad_norm": 0.032257601618766785, + "learning_rate": 6.694616803622338e-05, + "loss": 0.6488, + "step": 19935 + }, + { + "epoch": 3.3344476015376903, + "grad_norm": 0.03307221084833145, + "learning_rate": 6.684554754318297e-05, + "loss": 0.6546, + "step": 19950 + }, + { + "epoch": 3.336954704997493, + "grad_norm": 0.03298574313521385, + "learning_rate": 6.674492705014255e-05, + "loss": 0.6795, + "step": 19965 + }, + { + "epoch": 3.339461808457296, + "grad_norm": 0.0329146534204483, + "learning_rate": 6.664430655710213e-05, + "loss": 0.6571, + "step": 19980 + }, + { + "epoch": 3.3419689119170983, + "grad_norm": 0.03403447940945625, + "learning_rate": 6.654368606406171e-05, + "loss": 0.6778, + "step": 19995 + }, + { + "epoch": 3.3444760153769013, + "grad_norm": 0.03246279060840607, + "learning_rate": 6.64430655710213e-05, + "loss": 0.6612, + "step": 20010 + }, + { + "epoch": 3.346983118836704, + "grad_norm": 0.03411612659692764, + "learning_rate": 6.634244507798088e-05, + "loss": 0.6478, + "step": 20025 + }, + { + "epoch": 3.349490222296507, + "grad_norm": 0.033288851380348206, + "learning_rate": 6.624182458494047e-05, + "loss": 0.6551, + "step": 20040 + }, + { + "epoch": 3.3519973257563094, + "grad_norm": 0.03258313983678818, + "learning_rate": 6.614120409190005e-05, + "loss": 0.6761, + "step": 20055 + }, + { + "epoch": 3.3545044292161124, + "grad_norm": 0.033909596502780914, + "learning_rate": 6.604058359885964e-05, + "loss": 0.6707, + "step": 20070 + }, + { + "epoch": 3.357011532675915, + "grad_norm": 0.033043161034584045, + "learning_rate": 6.593996310581922e-05, + "loss": 0.6549, + "step": 20085 + }, + { + "epoch": 3.359518636135718, + "grad_norm": 0.03325843811035156, + "learning_rate": 6.583934261277881e-05, + "loss": 0.6655, + "step": 20100 + }, + { + "epoch": 3.362025739595521, + "grad_norm": 0.033593397587537766, + "learning_rate": 6.57387221197384e-05, + "loss": 0.659, + "step": 20115 + }, + { + "epoch": 3.3645328430553234, + "grad_norm": 0.032497063279151917, + "learning_rate": 6.563810162669798e-05, + "loss": 0.6728, + "step": 20130 + }, + { + "epoch": 3.367039946515126, + "grad_norm": 0.03400912135839462, + "learning_rate": 6.553748113365757e-05, + "loss": 0.6701, + "step": 20145 + }, + { + "epoch": 3.369547049974929, + "grad_norm": 0.033511847257614136, + "learning_rate": 6.543686064061714e-05, + "loss": 0.6522, + "step": 20160 + }, + { + "epoch": 3.372054153434732, + "grad_norm": 0.034270040690898895, + "learning_rate": 6.533624014757672e-05, + "loss": 0.6509, + "step": 20175 + }, + { + "epoch": 3.3745612568945345, + "grad_norm": 0.03471988067030907, + "learning_rate": 6.523561965453631e-05, + "loss": 0.657, + "step": 20190 + }, + { + "epoch": 3.3770683603543374, + "grad_norm": 0.033193349838256836, + "learning_rate": 6.51349991614959e-05, + "loss": 0.6602, + "step": 20205 + }, + { + "epoch": 3.37957546381414, + "grad_norm": 0.03341618552803993, + "learning_rate": 6.503437866845548e-05, + "loss": 0.6594, + "step": 20220 + }, + { + "epoch": 3.382082567273943, + "grad_norm": 0.03393018990755081, + "learning_rate": 6.493375817541507e-05, + "loss": 0.6692, + "step": 20235 + }, + { + "epoch": 3.3845896707337455, + "grad_norm": 0.032490186393260956, + "learning_rate": 6.483313768237465e-05, + "loss": 0.645, + "step": 20250 + }, + { + "epoch": 3.3870967741935485, + "grad_norm": 0.03293507918715477, + "learning_rate": 6.473251718933424e-05, + "loss": 0.6414, + "step": 20265 + }, + { + "epoch": 3.389603877653351, + "grad_norm": 0.033456623554229736, + "learning_rate": 6.463189669629381e-05, + "loss": 0.6547, + "step": 20280 + }, + { + "epoch": 3.392110981113154, + "grad_norm": 0.03404277190566063, + "learning_rate": 6.45312762032534e-05, + "loss": 0.6605, + "step": 20295 + }, + { + "epoch": 3.3946180845729566, + "grad_norm": 0.03427689149975777, + "learning_rate": 6.443065571021298e-05, + "loss": 0.6594, + "step": 20310 + }, + { + "epoch": 3.3971251880327595, + "grad_norm": 0.033457666635513306, + "learning_rate": 6.433003521717256e-05, + "loss": 0.6778, + "step": 20325 + }, + { + "epoch": 3.399632291492562, + "grad_norm": 0.034244317561388016, + "learning_rate": 6.422941472413215e-05, + "loss": 0.6742, + "step": 20340 + }, + { + "epoch": 3.402139394952365, + "grad_norm": 0.03270183503627777, + "learning_rate": 6.412879423109174e-05, + "loss": 0.6435, + "step": 20355 + }, + { + "epoch": 3.404646498412168, + "grad_norm": 0.034196462482213974, + "learning_rate": 6.402817373805132e-05, + "loss": 0.648, + "step": 20370 + }, + { + "epoch": 3.4071536018719706, + "grad_norm": 0.032824840396642685, + "learning_rate": 6.39275532450109e-05, + "loss": 0.6767, + "step": 20385 + }, + { + "epoch": 3.409660705331773, + "grad_norm": 0.03537704795598984, + "learning_rate": 6.382693275197049e-05, + "loss": 0.6721, + "step": 20400 + }, + { + "epoch": 3.412167808791576, + "grad_norm": 0.03542947396636009, + "learning_rate": 6.372631225893008e-05, + "loss": 0.6612, + "step": 20415 + }, + { + "epoch": 3.414674912251379, + "grad_norm": 0.033524878323078156, + "learning_rate": 6.362569176588966e-05, + "loss": 0.6709, + "step": 20430 + }, + { + "epoch": 3.4171820157111816, + "grad_norm": 0.03377537056803703, + "learning_rate": 6.352507127284923e-05, + "loss": 0.6711, + "step": 20445 + }, + { + "epoch": 3.4196891191709846, + "grad_norm": 0.033485304564237595, + "learning_rate": 6.342445077980882e-05, + "loss": 0.6622, + "step": 20460 + }, + { + "epoch": 3.422196222630787, + "grad_norm": 0.033119190484285355, + "learning_rate": 6.33238302867684e-05, + "loss": 0.66, + "step": 20475 + }, + { + "epoch": 3.42470332609059, + "grad_norm": 0.0347219854593277, + "learning_rate": 6.322320979372799e-05, + "loss": 0.6538, + "step": 20490 + }, + { + "epoch": 3.4272104295503927, + "grad_norm": 0.033375516533851624, + "learning_rate": 6.312258930068758e-05, + "loss": 0.6721, + "step": 20505 + }, + { + "epoch": 3.4297175330101957, + "grad_norm": 0.03268874064087868, + "learning_rate": 6.302196880764716e-05, + "loss": 0.6608, + "step": 20520 + }, + { + "epoch": 3.432224636469998, + "grad_norm": 0.03364133462309837, + "learning_rate": 6.292134831460675e-05, + "loss": 0.6673, + "step": 20535 + }, + { + "epoch": 3.434731739929801, + "grad_norm": 0.03426344320178032, + "learning_rate": 6.282072782156633e-05, + "loss": 0.6637, + "step": 20550 + }, + { + "epoch": 3.4372388433896037, + "grad_norm": 0.03355192393064499, + "learning_rate": 6.272010732852592e-05, + "loss": 0.6781, + "step": 20565 + }, + { + "epoch": 3.4397459468494067, + "grad_norm": 0.03330032154917717, + "learning_rate": 6.261948683548549e-05, + "loss": 0.6693, + "step": 20580 + }, + { + "epoch": 3.4422530503092093, + "grad_norm": 0.03315750136971474, + "learning_rate": 6.251886634244508e-05, + "loss": 0.66, + "step": 20595 + }, + { + "epoch": 3.4447601537690122, + "grad_norm": 0.033580485731363297, + "learning_rate": 6.241824584940466e-05, + "loss": 0.6749, + "step": 20610 + }, + { + "epoch": 3.447267257228815, + "grad_norm": 0.03357269614934921, + "learning_rate": 6.231762535636425e-05, + "loss": 0.6563, + "step": 20625 + }, + { + "epoch": 3.4497743606886178, + "grad_norm": 0.03285966068506241, + "learning_rate": 6.221700486332383e-05, + "loss": 0.6615, + "step": 20640 + }, + { + "epoch": 3.4522814641484203, + "grad_norm": 0.034152496606111526, + "learning_rate": 6.211638437028342e-05, + "loss": 0.666, + "step": 20655 + }, + { + "epoch": 3.4547885676082233, + "grad_norm": 0.03442816436290741, + "learning_rate": 6.2015763877243e-05, + "loss": 0.6637, + "step": 20670 + }, + { + "epoch": 3.4572956710680263, + "grad_norm": 0.03422121703624725, + "learning_rate": 6.191514338420259e-05, + "loss": 0.661, + "step": 20685 + }, + { + "epoch": 3.459802774527829, + "grad_norm": 0.035803407430648804, + "learning_rate": 6.181452289116217e-05, + "loss": 0.6659, + "step": 20700 + }, + { + "epoch": 3.462309877987632, + "grad_norm": 0.03371883183717728, + "learning_rate": 6.171390239812176e-05, + "loss": 0.6659, + "step": 20715 + }, + { + "epoch": 3.4648169814474343, + "grad_norm": 0.034014519304037094, + "learning_rate": 6.161328190508133e-05, + "loss": 0.662, + "step": 20730 + }, + { + "epoch": 3.4673240849072373, + "grad_norm": 0.032825078815221786, + "learning_rate": 6.151266141204092e-05, + "loss": 0.6645, + "step": 20745 + }, + { + "epoch": 3.46983118836704, + "grad_norm": 0.033502623438835144, + "learning_rate": 6.14120409190005e-05, + "loss": 0.6611, + "step": 20760 + }, + { + "epoch": 3.472338291826843, + "grad_norm": 0.03605775162577629, + "learning_rate": 6.131142042596009e-05, + "loss": 0.6781, + "step": 20775 + }, + { + "epoch": 3.4748453952866454, + "grad_norm": 0.03367803990840912, + "learning_rate": 6.121079993291967e-05, + "loss": 0.6547, + "step": 20790 + }, + { + "epoch": 3.4773524987464484, + "grad_norm": 0.034281060099601746, + "learning_rate": 6.111017943987926e-05, + "loss": 0.6607, + "step": 20805 + }, + { + "epoch": 3.479859602206251, + "grad_norm": 0.032581839710474014, + "learning_rate": 6.100955894683884e-05, + "loss": 0.6753, + "step": 20820 + }, + { + "epoch": 3.482366705666054, + "grad_norm": 0.03385984152555466, + "learning_rate": 6.090893845379842e-05, + "loss": 0.6582, + "step": 20835 + }, + { + "epoch": 3.4848738091258564, + "grad_norm": 0.03409432992339134, + "learning_rate": 6.080831796075801e-05, + "loss": 0.6541, + "step": 20850 + }, + { + "epoch": 3.4873809125856594, + "grad_norm": 0.032626084983348846, + "learning_rate": 6.0707697467717594e-05, + "loss": 0.6625, + "step": 20865 + }, + { + "epoch": 3.489888016045462, + "grad_norm": 0.034788578748703, + "learning_rate": 6.060707697467718e-05, + "loss": 0.6539, + "step": 20880 + }, + { + "epoch": 3.492395119505265, + "grad_norm": 0.034189485013484955, + "learning_rate": 6.0506456481636765e-05, + "loss": 0.6595, + "step": 20895 + }, + { + "epoch": 3.4949022229650675, + "grad_norm": 0.03473382443189621, + "learning_rate": 6.040583598859635e-05, + "loss": 0.6565, + "step": 20910 + }, + { + "epoch": 3.4974093264248705, + "grad_norm": 0.0336168147623539, + "learning_rate": 6.0305215495555936e-05, + "loss": 0.6584, + "step": 20925 + }, + { + "epoch": 3.4999164298846734, + "grad_norm": 0.0335552953183651, + "learning_rate": 6.020459500251552e-05, + "loss": 0.6557, + "step": 20940 + }, + { + "epoch": 3.502423533344476, + "grad_norm": 0.03281300142407417, + "learning_rate": 6.01039745094751e-05, + "loss": 0.668, + "step": 20955 + }, + { + "epoch": 3.5049306368042785, + "grad_norm": 0.03366611897945404, + "learning_rate": 6.0003354016434685e-05, + "loss": 0.6621, + "step": 20970 + }, + { + "epoch": 3.5074377402640815, + "grad_norm": 0.034198787063360214, + "learning_rate": 5.990273352339427e-05, + "loss": 0.6619, + "step": 20985 + }, + { + "epoch": 3.5099448437238845, + "grad_norm": 0.03419259935617447, + "learning_rate": 5.9802113030353856e-05, + "loss": 0.6737, + "step": 21000 + }, + { + "epoch": 3.5266588667892362, + "grad_norm": 0.03425095975399017, + "learning_rate": 5.913130974341775e-05, + "loss": 0.665, + "step": 21100 + }, + { + "epoch": 3.543372889854588, + "grad_norm": 0.033783555030822754, + "learning_rate": 5.846050645648163e-05, + "loss": 0.6577, + "step": 21200 + }, + { + "epoch": 3.5600869129199397, + "grad_norm": 0.03537527099251747, + "learning_rate": 5.778970316954553e-05, + "loss": 0.6635, + "step": 21300 + }, + { + "epoch": 3.5768009359852915, + "grad_norm": 0.03478403761982918, + "learning_rate": 5.7118899882609426e-05, + "loss": 0.6584, + "step": 21400 + }, + { + "epoch": 3.593514959050643, + "grad_norm": 0.034150175750255585, + "learning_rate": 5.644809659567332e-05, + "loss": 0.6596, + "step": 21500 + }, + { + "epoch": 3.6102289821159954, + "grad_norm": 0.034265898168087006, + "learning_rate": 5.5777293308737214e-05, + "loss": 0.6574, + "step": 21600 + }, + { + "epoch": 3.626943005181347, + "grad_norm": 0.03564199060201645, + "learning_rate": 5.510649002180111e-05, + "loss": 0.6549, + "step": 21700 + }, + { + "epoch": 3.643657028246699, + "grad_norm": 0.034864045679569244, + "learning_rate": 5.4435686734865e-05, + "loss": 0.6656, + "step": 21800 + }, + { + "epoch": 3.6603710513120506, + "grad_norm": 0.03352364897727966, + "learning_rate": 5.3764883447928896e-05, + "loss": 0.6622, + "step": 21900 + }, + { + "epoch": 3.677085074377403, + "grad_norm": 0.034029681235551834, + "learning_rate": 5.309408016099279e-05, + "loss": 0.6542, + "step": 22000 + }, + { + "epoch": 3.6937990974427546, + "grad_norm": 0.03352760896086693, + "learning_rate": 5.2423276874056684e-05, + "loss": 0.6664, + "step": 22100 + }, + { + "epoch": 3.7105131205081063, + "grad_norm": 0.03489440679550171, + "learning_rate": 5.175247358712058e-05, + "loss": 0.6629, + "step": 22200 + }, + { + "epoch": 3.727227143573458, + "grad_norm": 0.03430229052901268, + "learning_rate": 5.108167030018447e-05, + "loss": 0.6574, + "step": 22300 + }, + { + "epoch": 3.74394116663881, + "grad_norm": 0.033648181706666946, + "learning_rate": 5.0410867013248366e-05, + "loss": 0.6568, + "step": 22400 + }, + { + "epoch": 3.7606551897041616, + "grad_norm": 0.033668212592601776, + "learning_rate": 4.974006372631227e-05, + "loss": 0.6564, + "step": 22500 + }, + { + "epoch": 3.7773692127695138, + "grad_norm": 0.03772876039147377, + "learning_rate": 4.906926043937616e-05, + "loss": 0.6619, + "step": 22600 + }, + { + "epoch": 3.7940832358348655, + "grad_norm": 0.03430061787366867, + "learning_rate": 4.839845715244005e-05, + "loss": 0.6603, + "step": 22700 + }, + { + "epoch": 3.8107972589002173, + "grad_norm": 0.035929903388023376, + "learning_rate": 4.772765386550394e-05, + "loss": 0.6559, + "step": 22800 + }, + { + "epoch": 3.827511281965569, + "grad_norm": 0.03500952944159508, + "learning_rate": 4.7056850578567837e-05, + "loss": 0.6579, + "step": 22900 + }, + { + "epoch": 3.844225305030921, + "grad_norm": 0.03469489514827728, + "learning_rate": 4.638604729163173e-05, + "loss": 0.6558, + "step": 23000 + }, + { + "epoch": 3.860939328096273, + "grad_norm": 0.03354435786604881, + "learning_rate": 4.5715244004695625e-05, + "loss": 0.6623, + "step": 23100 + }, + { + "epoch": 3.8776533511616247, + "grad_norm": 0.03471764177083969, + "learning_rate": 4.504444071775952e-05, + "loss": 0.6648, + "step": 23200 + }, + { + "epoch": 3.8943673742269764, + "grad_norm": 0.03438182920217514, + "learning_rate": 4.437363743082341e-05, + "loss": 0.6606, + "step": 23300 + }, + { + "epoch": 3.911081397292328, + "grad_norm": 0.03417756408452988, + "learning_rate": 4.370283414388731e-05, + "loss": 0.6626, + "step": 23400 + }, + { + "epoch": 3.92779542035768, + "grad_norm": 0.03406790643930435, + "learning_rate": 4.30320308569512e-05, + "loss": 0.6609, + "step": 23500 + }, + { + "epoch": 3.9445094434230317, + "grad_norm": 0.035032719373703, + "learning_rate": 4.2361227570015095e-05, + "loss": 0.6579, + "step": 23600 + }, + { + "epoch": 3.961223466488384, + "grad_norm": 0.03397015482187271, + "learning_rate": 4.169042428307899e-05, + "loss": 0.6557, + "step": 23700 + }, + { + "epoch": 3.9779374895537356, + "grad_norm": 0.03334665298461914, + "learning_rate": 4.101962099614288e-05, + "loss": 0.6606, + "step": 23800 + }, + { + "epoch": 3.9946515126190874, + "grad_norm": 0.033896464854478836, + "learning_rate": 4.034881770920678e-05, + "loss": 0.6584, + "step": 23900 + }, + { + "epoch": 4.0113655356844395, + "grad_norm": 0.03485192731022835, + "learning_rate": 3.967801442227067e-05, + "loss": 0.6492, + "step": 24000 + }, + { + "epoch": 4.028079558749791, + "grad_norm": 0.0341389924287796, + "learning_rate": 3.9007211135334565e-05, + "loss": 0.6461, + "step": 24100 + }, + { + "epoch": 4.044793581815143, + "grad_norm": 0.03344714641571045, + "learning_rate": 3.833640784839846e-05, + "loss": 0.6475, + "step": 24200 + }, + { + "epoch": 4.061507604880495, + "grad_norm": 0.03415651619434357, + "learning_rate": 3.7665604561462354e-05, + "loss": 0.6533, + "step": 24300 + }, + { + "epoch": 4.0782216279458465, + "grad_norm": 0.03357802331447601, + "learning_rate": 3.699480127452625e-05, + "loss": 0.6513, + "step": 24400 + }, + { + "epoch": 4.094935651011198, + "grad_norm": 0.03374486416578293, + "learning_rate": 3.632399798759014e-05, + "loss": 0.6456, + "step": 24500 + }, + { + "epoch": 4.11164967407655, + "grad_norm": 0.03407549113035202, + "learning_rate": 3.5653194700654036e-05, + "loss": 0.651, + "step": 24600 + }, + { + "epoch": 4.128363697141902, + "grad_norm": 0.03402148187160492, + "learning_rate": 3.498239141371793e-05, + "loss": 0.6461, + "step": 24700 + }, + { + "epoch": 4.1450777202072535, + "grad_norm": 0.03708890080451965, + "learning_rate": 3.4311588126781824e-05, + "loss": 0.652, + "step": 24800 + }, + { + "epoch": 4.161791743272605, + "grad_norm": 0.034347113221883774, + "learning_rate": 3.364078483984572e-05, + "loss": 0.6509, + "step": 24900 + }, + { + "epoch": 4.178505766337958, + "grad_norm": 0.0340665765106678, + "learning_rate": 3.296998155290961e-05, + "loss": 0.6451, + "step": 25000 + }, + { + "epoch": 4.19521978940331, + "grad_norm": 0.035704102367162704, + "learning_rate": 3.2299178265973506e-05, + "loss": 0.6507, + "step": 25100 + }, + { + "epoch": 4.211933812468661, + "grad_norm": 0.03486304730176926, + "learning_rate": 3.16283749790374e-05, + "loss": 0.6547, + "step": 25200 + }, + { + "epoch": 4.228647835534013, + "grad_norm": 0.03282959759235382, + "learning_rate": 3.0957571692101294e-05, + "loss": 0.6523, + "step": 25300 + }, + { + "epoch": 4.245361858599365, + "grad_norm": 0.035729847848415375, + "learning_rate": 3.0286768405165188e-05, + "loss": 0.6503, + "step": 25400 + }, + { + "epoch": 4.262075881664717, + "grad_norm": 0.03634531795978546, + "learning_rate": 2.9615965118229082e-05, + "loss": 0.6493, + "step": 25500 + }, + { + "epoch": 4.278789904730068, + "grad_norm": 0.033321358263492584, + "learning_rate": 2.8945161831292976e-05, + "loss": 0.6492, + "step": 25600 + }, + { + "epoch": 4.29550392779542, + "grad_norm": 0.03541552275419235, + "learning_rate": 2.8274358544356867e-05, + "loss": 0.6466, + "step": 25700 + }, + { + "epoch": 4.312217950860772, + "grad_norm": 0.03528020158410072, + "learning_rate": 2.760355525742076e-05, + "loss": 0.6549, + "step": 25800 + }, + { + "epoch": 4.328931973926124, + "grad_norm": 0.03419233486056328, + "learning_rate": 2.6932751970484655e-05, + "loss": 0.6502, + "step": 25900 + }, + { + "epoch": 4.345645996991476, + "grad_norm": 0.03410422429442406, + "learning_rate": 2.626194868354855e-05, + "loss": 0.656, + "step": 26000 + }, + { + "epoch": 4.362360020056828, + "grad_norm": 0.033916935324668884, + "learning_rate": 2.5591145396612443e-05, + "loss": 0.643, + "step": 26100 + }, + { + "epoch": 4.37907404312218, + "grad_norm": 0.034409794956445694, + "learning_rate": 2.492034210967634e-05, + "loss": 0.6445, + "step": 26200 + }, + { + "epoch": 4.3957880661875315, + "grad_norm": 0.03435683995485306, + "learning_rate": 2.4249538822740235e-05, + "loss": 0.6493, + "step": 26300 + }, + { + "epoch": 4.412502089252883, + "grad_norm": 0.03483356907963753, + "learning_rate": 2.357873553580413e-05, + "loss": 0.649, + "step": 26400 + }, + { + "epoch": 4.429216112318235, + "grad_norm": 0.0342116504907608, + "learning_rate": 2.290793224886802e-05, + "loss": 0.6516, + "step": 26500 + }, + { + "epoch": 4.445930135383587, + "grad_norm": 0.035094503313302994, + "learning_rate": 2.2237128961931914e-05, + "loss": 0.6509, + "step": 26600 + }, + { + "epoch": 4.4626441584489385, + "grad_norm": 0.035515137016773224, + "learning_rate": 2.1566325674995808e-05, + "loss": 0.6491, + "step": 26700 + }, + { + "epoch": 4.47935818151429, + "grad_norm": 0.033778801560401917, + "learning_rate": 2.0895522388059702e-05, + "loss": 0.6516, + "step": 26800 + }, + { + "epoch": 4.496072204579642, + "grad_norm": 0.03408665210008621, + "learning_rate": 2.0224719101123596e-05, + "loss": 0.6482, + "step": 26900 + }, + { + "epoch": 4.512786227644995, + "grad_norm": 0.03422163799405098, + "learning_rate": 1.955391581418749e-05, + "loss": 0.6488, + "step": 27000 + }, + { + "epoch": 4.529500250710346, + "grad_norm": 0.03455764427781105, + "learning_rate": 1.8883112527251384e-05, + "loss": 0.6473, + "step": 27100 + }, + { + "epoch": 4.546214273775698, + "grad_norm": 0.033868152648210526, + "learning_rate": 1.8212309240315278e-05, + "loss": 0.6508, + "step": 27200 + }, + { + "epoch": 4.56292829684105, + "grad_norm": 0.03507550060749054, + "learning_rate": 1.7541505953379172e-05, + "loss": 0.6468, + "step": 27300 + }, + { + "epoch": 4.579642319906402, + "grad_norm": 0.03554074466228485, + "learning_rate": 1.6870702666443066e-05, + "loss": 0.6456, + "step": 27400 + }, + { + "epoch": 4.596356342971753, + "grad_norm": 0.03559200465679169, + "learning_rate": 1.619989937950696e-05, + "loss": 0.6544, + "step": 27500 + }, + { + "epoch": 4.613070366037105, + "grad_norm": 0.03546106070280075, + "learning_rate": 1.5529096092570854e-05, + "loss": 0.646, + "step": 27600 + }, + { + "epoch": 4.629784389102457, + "grad_norm": 0.03442246466875076, + "learning_rate": 1.4858292805634748e-05, + "loss": 0.6523, + "step": 27700 + }, + { + "epoch": 4.646498412167809, + "grad_norm": 0.03322317451238632, + "learning_rate": 1.4187489518698644e-05, + "loss": 0.6483, + "step": 27800 + }, + { + "epoch": 4.66321243523316, + "grad_norm": 0.0362270288169384, + "learning_rate": 1.3516686231762538e-05, + "loss": 0.649, + "step": 27900 + }, + { + "epoch": 4.679926458298512, + "grad_norm": 0.03510970249772072, + "learning_rate": 1.2845882944826429e-05, + "loss": 0.6461, + "step": 28000 + }, + { + "epoch": 4.696640481363865, + "grad_norm": 0.03399231657385826, + "learning_rate": 1.2175079657890325e-05, + "loss": 0.6491, + "step": 28100 + }, + { + "epoch": 4.713354504429216, + "grad_norm": 0.03436035290360451, + "learning_rate": 1.1504276370954219e-05, + "loss": 0.6499, + "step": 28200 + }, + { + "epoch": 4.730068527494568, + "grad_norm": 0.034751422703266144, + "learning_rate": 1.0833473084018113e-05, + "loss": 0.6476, + "step": 28300 + }, + { + "epoch": 4.74678255055992, + "grad_norm": 0.034067828208208084, + "learning_rate": 1.0162669797082005e-05, + "loss": 0.6463, + "step": 28400 + }, + { + "epoch": 4.763496573625272, + "grad_norm": 0.03397444635629654, + "learning_rate": 9.4918665101459e-06, + "loss": 0.6487, + "step": 28500 + }, + { + "epoch": 4.780210596690623, + "grad_norm": 0.03437269851565361, + "learning_rate": 8.821063223209793e-06, + "loss": 0.6477, + "step": 28600 + }, + { + "epoch": 4.796924619755975, + "grad_norm": 0.034697502851486206, + "learning_rate": 8.150259936273687e-06, + "loss": 0.6488, + "step": 28700 + }, + { + "epoch": 4.813638642821327, + "grad_norm": 0.03559542074799538, + "learning_rate": 7.479456649337582e-06, + "loss": 0.6492, + "step": 28800 + }, + { + "epoch": 4.830352665886679, + "grad_norm": 0.03439110890030861, + "learning_rate": 6.808653362401476e-06, + "loss": 0.6462, + "step": 28900 + }, + { + "epoch": 4.84706668895203, + "grad_norm": 0.03447462245821953, + "learning_rate": 6.13785007546537e-06, + "loss": 0.6499, + "step": 29000 + }, + { + "epoch": 4.863780712017382, + "grad_norm": 0.03418246656656265, + "learning_rate": 5.467046788529264e-06, + "loss": 0.6492, + "step": 29100 + }, + { + "epoch": 4.880494735082735, + "grad_norm": 0.03533853963017464, + "learning_rate": 4.796243501593159e-06, + "loss": 0.6513, + "step": 29200 + }, + { + "epoch": 4.8972087581480865, + "grad_norm": 0.03379116207361221, + "learning_rate": 4.125440214657052e-06, + "loss": 0.6491, + "step": 29300 + }, + { + "epoch": 4.913922781213438, + "grad_norm": 0.03501541167497635, + "learning_rate": 3.454636927720946e-06, + "loss": 0.6444, + "step": 29400 + }, + { + "epoch": 4.93063680427879, + "grad_norm": 0.03520382195711136, + "learning_rate": 2.78383364078484e-06, + "loss": 0.6499, + "step": 29500 + }, + { + "epoch": 4.809879753006175, + "grad_norm": 0.03432910144329071, + "learning_rate": 7.629605477665471e-06, + "loss": 0.6493, + "step": 29600 + }, + { + "epoch": 4.826129346766331, + "grad_norm": 0.03554558381438255, + "learning_rate": 6.9775024453863715e-06, + "loss": 0.6423, + "step": 29700 + }, + { + "epoch": 4.842378940526487, + "grad_norm": 0.03414788842201233, + "learning_rate": 6.3253994131072716e-06, + "loss": 0.6475, + "step": 29800 + }, + { + "epoch": 4.858628534286643, + "grad_norm": 0.03513456508517265, + "learning_rate": 5.673296380828172e-06, + "loss": 0.6474, + "step": 29900 + }, + { + "epoch": 4.874878128046799, + "grad_norm": 0.03595611825585365, + "learning_rate": 5.021193348549072e-06, + "loss": 0.6485, + "step": 30000 + }, + { + "epoch": 4.891127721806955, + "grad_norm": 0.035549987107515335, + "learning_rate": 4.369090316269971e-06, + "loss": 0.649, + "step": 30100 + }, + { + "epoch": 4.9073773155671105, + "grad_norm": 0.03475033864378929, + "learning_rate": 3.7169872839908704e-06, + "loss": 0.6468, + "step": 30200 + }, + { + "epoch": 4.923626909327266, + "grad_norm": 0.03467612341046333, + "learning_rate": 3.0648842517117705e-06, + "loss": 0.6404, + "step": 30300 + }, + { + "epoch": 4.939876503087422, + "grad_norm": 0.03498971089720726, + "learning_rate": 2.4127812194326705e-06, + "loss": 0.6544, + "step": 30400 + }, + { + "epoch": 4.956126096847579, + "grad_norm": 0.03452787175774574, + "learning_rate": 1.7606781871535704e-06, + "loss": 0.6491, + "step": 30500 + } + ], + "logging_steps": 100, + "max_steps": 30770, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9.120160458945331e+19, + "train_batch_size": 5, + "trial_name": null, + "trial_params": null +}