{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0000852151683, "eval_steps": 500, "global_step": 11736, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 8.521516829995739e-05, "grad_norm": 18.5, "learning_rate": 2.5000000000000002e-08, "loss": 1.21, "step": 1 }, { "epoch": 0.00017043033659991478, "grad_norm": 34.25, "learning_rate": 5.0000000000000004e-08, "loss": 1.5951, "step": 2 }, { "epoch": 0.00025564550489987217, "grad_norm": 32.25, "learning_rate": 7.500000000000001e-08, "loss": 1.4702, "step": 3 }, { "epoch": 0.00034086067319982955, "grad_norm": 27.5, "learning_rate": 1.0000000000000001e-07, "loss": 1.3454, "step": 4 }, { "epoch": 0.00042607584149978694, "grad_norm": 28.125, "learning_rate": 1.2500000000000002e-07, "loss": 1.16, "step": 5 }, { "epoch": 0.0005112910097997443, "grad_norm": 27.75, "learning_rate": 1.5000000000000002e-07, "loss": 1.6235, "step": 6 }, { "epoch": 0.0005965061780997018, "grad_norm": 27.0, "learning_rate": 1.7500000000000002e-07, "loss": 1.3127, "step": 7 }, { "epoch": 0.0006817213463996591, "grad_norm": 35.0, "learning_rate": 2.0000000000000002e-07, "loss": 1.4014, "step": 8 }, { "epoch": 0.0007669365146996165, "grad_norm": 19.375, "learning_rate": 2.2500000000000002e-07, "loss": 1.1322, "step": 9 }, { "epoch": 0.0008521516829995739, "grad_norm": 28.25, "learning_rate": 2.5000000000000004e-07, "loss": 1.2969, "step": 10 }, { "epoch": 0.0009373668512995313, "grad_norm": 43.0, "learning_rate": 2.75e-07, "loss": 1.4659, "step": 11 }, { "epoch": 0.0010225820195994887, "grad_norm": 14.875, "learning_rate": 3.0000000000000004e-07, "loss": 0.8019, "step": 12 }, { "epoch": 0.001107797187899446, "grad_norm": 45.75, "learning_rate": 3.25e-07, "loss": 1.4666, "step": 13 }, { "epoch": 0.0011930123561994035, "grad_norm": 52.0, "learning_rate": 3.5000000000000004e-07, "loss": 2.0725, "step": 14 }, { "epoch": 0.0012782275244993609, "grad_norm": 91.5, "learning_rate": 3.75e-07, "loss": 1.517, "step": 15 }, { "epoch": 0.0013634426927993182, "grad_norm": 53.5, "learning_rate": 4.0000000000000003e-07, "loss": 1.6968, "step": 16 }, { "epoch": 0.0014486578610992758, "grad_norm": 68.0, "learning_rate": 4.2500000000000006e-07, "loss": 1.3316, "step": 17 }, { "epoch": 0.001533873029399233, "grad_norm": 17.875, "learning_rate": 4.5000000000000003e-07, "loss": 1.1111, "step": 18 }, { "epoch": 0.0016190881976991904, "grad_norm": 36.75, "learning_rate": 4.7500000000000006e-07, "loss": 1.5615, "step": 19 }, { "epoch": 0.0017043033659991478, "grad_norm": 55.75, "learning_rate": 5.000000000000001e-07, "loss": 1.5168, "step": 20 }, { "epoch": 0.0017895185342991053, "grad_norm": 23.625, "learning_rate": 5.250000000000001e-07, "loss": 1.3197, "step": 21 }, { "epoch": 0.0018747337025990627, "grad_norm": 31.5, "learning_rate": 5.5e-07, "loss": 1.3818, "step": 22 }, { "epoch": 0.00195994887089902, "grad_norm": 24.75, "learning_rate": 5.750000000000001e-07, "loss": 1.4715, "step": 23 }, { "epoch": 0.0020451640391989773, "grad_norm": 26.25, "learning_rate": 6.000000000000001e-07, "loss": 1.2183, "step": 24 }, { "epoch": 0.0021303792074989347, "grad_norm": 52.5, "learning_rate": 6.25e-07, "loss": 1.5819, "step": 25 }, { "epoch": 0.002215594375798892, "grad_norm": 18.875, "learning_rate": 6.5e-07, "loss": 1.224, "step": 26 }, { "epoch": 0.0023008095440988498, "grad_norm": 47.75, "learning_rate": 6.750000000000001e-07, "loss": 1.0704, "step": 27 }, { "epoch": 0.002386024712398807, "grad_norm": 174.0, "learning_rate": 7.000000000000001e-07, "loss": 1.2073, "step": 28 }, { "epoch": 0.0024712398806987644, "grad_norm": 37.75, "learning_rate": 7.25e-07, "loss": 1.6614, "step": 29 }, { "epoch": 0.0025564550489987218, "grad_norm": 47.25, "learning_rate": 7.5e-07, "loss": 1.1309, "step": 30 }, { "epoch": 0.002641670217298679, "grad_norm": 24.625, "learning_rate": 7.750000000000001e-07, "loss": 1.4704, "step": 31 }, { "epoch": 0.0027268853855986364, "grad_norm": 53.25, "learning_rate": 8.000000000000001e-07, "loss": 0.93, "step": 32 }, { "epoch": 0.0028121005538985938, "grad_norm": 33.25, "learning_rate": 8.250000000000001e-07, "loss": 1.2446, "step": 33 }, { "epoch": 0.0028973157221985515, "grad_norm": 25.75, "learning_rate": 8.500000000000001e-07, "loss": 1.4553, "step": 34 }, { "epoch": 0.002982530890498509, "grad_norm": 17.0, "learning_rate": 8.75e-07, "loss": 1.0626, "step": 35 }, { "epoch": 0.003067746058798466, "grad_norm": 39.75, "learning_rate": 9.000000000000001e-07, "loss": 1.3777, "step": 36 }, { "epoch": 0.0031529612270984235, "grad_norm": 22.875, "learning_rate": 9.25e-07, "loss": 1.2808, "step": 37 }, { "epoch": 0.003238176395398381, "grad_norm": 15.125, "learning_rate": 9.500000000000001e-07, "loss": 1.0063, "step": 38 }, { "epoch": 0.003323391563698338, "grad_norm": 15.5625, "learning_rate": 9.750000000000002e-07, "loss": 1.2981, "step": 39 }, { "epoch": 0.0034086067319982955, "grad_norm": 15.75, "learning_rate": 1.0000000000000002e-06, "loss": 0.8237, "step": 40 }, { "epoch": 0.0034938219002982533, "grad_norm": 21.375, "learning_rate": 1.025e-06, "loss": 1.3627, "step": 41 }, { "epoch": 0.0035790370685982106, "grad_norm": 24.125, "learning_rate": 1.0500000000000001e-06, "loss": 1.4088, "step": 42 }, { "epoch": 0.003664252236898168, "grad_norm": 18.75, "learning_rate": 1.075e-06, "loss": 1.1094, "step": 43 }, { "epoch": 0.0037494674051981253, "grad_norm": 20.875, "learning_rate": 1.1e-06, "loss": 1.3761, "step": 44 }, { "epoch": 0.0038346825734980826, "grad_norm": 30.5, "learning_rate": 1.125e-06, "loss": 1.1231, "step": 45 }, { "epoch": 0.00391989774179804, "grad_norm": 18.0, "learning_rate": 1.1500000000000002e-06, "loss": 1.0932, "step": 46 }, { "epoch": 0.004005112910097998, "grad_norm": 34.0, "learning_rate": 1.175e-06, "loss": 1.5352, "step": 47 }, { "epoch": 0.004090328078397955, "grad_norm": 34.25, "learning_rate": 1.2000000000000002e-06, "loss": 1.4602, "step": 48 }, { "epoch": 0.004175543246697912, "grad_norm": 27.25, "learning_rate": 1.2250000000000001e-06, "loss": 1.0401, "step": 49 }, { "epoch": 0.004260758414997869, "grad_norm": 22.625, "learning_rate": 1.25e-06, "loss": 1.3274, "step": 50 }, { "epoch": 0.004345973583297827, "grad_norm": 27.0, "learning_rate": 1.275e-06, "loss": 1.5387, "step": 51 }, { "epoch": 0.004431188751597784, "grad_norm": 26.0, "learning_rate": 1.3e-06, "loss": 1.1358, "step": 52 }, { "epoch": 0.004516403919897742, "grad_norm": 27.375, "learning_rate": 1.3250000000000002e-06, "loss": 1.2844, "step": 53 }, { "epoch": 0.0046016190881976995, "grad_norm": 23.25, "learning_rate": 1.3500000000000002e-06, "loss": 1.1243, "step": 54 }, { "epoch": 0.004686834256497656, "grad_norm": 49.25, "learning_rate": 1.3750000000000002e-06, "loss": 1.3856, "step": 55 }, { "epoch": 0.004772049424797614, "grad_norm": 44.5, "learning_rate": 1.4000000000000001e-06, "loss": 1.4186, "step": 56 }, { "epoch": 0.004857264593097571, "grad_norm": 26.75, "learning_rate": 1.425e-06, "loss": 1.0897, "step": 57 }, { "epoch": 0.004942479761397529, "grad_norm": 17.0, "learning_rate": 1.45e-06, "loss": 1.0632, "step": 58 }, { "epoch": 0.005027694929697486, "grad_norm": 98.5, "learning_rate": 1.475e-06, "loss": 1.0083, "step": 59 }, { "epoch": 0.0051129100979974435, "grad_norm": 62.25, "learning_rate": 1.5e-06, "loss": 0.9524, "step": 60 }, { "epoch": 0.005198125266297401, "grad_norm": 74.0, "learning_rate": 1.525e-06, "loss": 1.3847, "step": 61 }, { "epoch": 0.005283340434597358, "grad_norm": 21.0, "learning_rate": 1.5500000000000002e-06, "loss": 0.7665, "step": 62 }, { "epoch": 0.005368555602897316, "grad_norm": 27.625, "learning_rate": 1.5750000000000002e-06, "loss": 0.9051, "step": 63 }, { "epoch": 0.005453770771197273, "grad_norm": 35.75, "learning_rate": 1.6000000000000001e-06, "loss": 1.1105, "step": 64 }, { "epoch": 0.005538985939497231, "grad_norm": 17.375, "learning_rate": 1.6250000000000001e-06, "loss": 0.9974, "step": 65 }, { "epoch": 0.0056242011077971875, "grad_norm": 13.1875, "learning_rate": 1.6500000000000003e-06, "loss": 0.9864, "step": 66 }, { "epoch": 0.005709416276097145, "grad_norm": 10.375, "learning_rate": 1.6750000000000003e-06, "loss": 0.5889, "step": 67 }, { "epoch": 0.005794631444397103, "grad_norm": 23.875, "learning_rate": 1.7000000000000002e-06, "loss": 1.0158, "step": 68 }, { "epoch": 0.00587984661269706, "grad_norm": 15.875, "learning_rate": 1.725e-06, "loss": 0.6812, "step": 69 }, { "epoch": 0.005965061780997018, "grad_norm": 35.75, "learning_rate": 1.75e-06, "loss": 1.4199, "step": 70 }, { "epoch": 0.006050276949296975, "grad_norm": 15.4375, "learning_rate": 1.7750000000000002e-06, "loss": 1.2153, "step": 71 }, { "epoch": 0.006135492117596932, "grad_norm": 13.75, "learning_rate": 1.8000000000000001e-06, "loss": 0.8568, "step": 72 }, { "epoch": 0.006220707285896889, "grad_norm": 15.8125, "learning_rate": 1.825e-06, "loss": 1.2163, "step": 73 }, { "epoch": 0.006305922454196847, "grad_norm": 22.5, "learning_rate": 1.85e-06, "loss": 1.0211, "step": 74 }, { "epoch": 0.006391137622496805, "grad_norm": 41.25, "learning_rate": 1.8750000000000003e-06, "loss": 1.7824, "step": 75 }, { "epoch": 0.006476352790796762, "grad_norm": 18.5, "learning_rate": 1.9000000000000002e-06, "loss": 0.9872, "step": 76 }, { "epoch": 0.0065615679590967195, "grad_norm": 14.75, "learning_rate": 1.925e-06, "loss": 1.1802, "step": 77 }, { "epoch": 0.006646783127396676, "grad_norm": 17.75, "learning_rate": 1.9500000000000004e-06, "loss": 1.0842, "step": 78 }, { "epoch": 0.006731998295696634, "grad_norm": 17.75, "learning_rate": 1.975e-06, "loss": 0.6949, "step": 79 }, { "epoch": 0.006817213463996591, "grad_norm": 15.6875, "learning_rate": 2.0000000000000003e-06, "loss": 1.0766, "step": 80 }, { "epoch": 0.006902428632296549, "grad_norm": 23.0, "learning_rate": 2.025e-06, "loss": 1.2883, "step": 81 }, { "epoch": 0.006987643800596507, "grad_norm": 14.0, "learning_rate": 2.05e-06, "loss": 0.9718, "step": 82 }, { "epoch": 0.0070728589688964635, "grad_norm": 13.1875, "learning_rate": 2.075e-06, "loss": 1.0055, "step": 83 }, { "epoch": 0.007158074137196421, "grad_norm": 24.25, "learning_rate": 2.1000000000000002e-06, "loss": 1.0122, "step": 84 }, { "epoch": 0.007243289305496378, "grad_norm": 13.375, "learning_rate": 2.125e-06, "loss": 0.6408, "step": 85 }, { "epoch": 0.007328504473796336, "grad_norm": 12.25, "learning_rate": 2.15e-06, "loss": 0.9325, "step": 86 }, { "epoch": 0.007413719642096293, "grad_norm": 16.75, "learning_rate": 2.1750000000000004e-06, "loss": 1.1798, "step": 87 }, { "epoch": 0.007498934810396251, "grad_norm": 14.875, "learning_rate": 2.2e-06, "loss": 0.9925, "step": 88 }, { "epoch": 0.0075841499786962075, "grad_norm": 16.25, "learning_rate": 2.2250000000000003e-06, "loss": 0.8689, "step": 89 }, { "epoch": 0.007669365146996165, "grad_norm": 19.25, "learning_rate": 2.25e-06, "loss": 1.1733, "step": 90 }, { "epoch": 0.007754580315296123, "grad_norm": 14.3125, "learning_rate": 2.2750000000000002e-06, "loss": 0.8938, "step": 91 }, { "epoch": 0.00783979548359608, "grad_norm": 15.25, "learning_rate": 2.3000000000000004e-06, "loss": 0.7673, "step": 92 }, { "epoch": 0.007925010651896037, "grad_norm": 17.5, "learning_rate": 2.325e-06, "loss": 0.8134, "step": 93 }, { "epoch": 0.008010225820195995, "grad_norm": 11.625, "learning_rate": 2.35e-06, "loss": 0.7227, "step": 94 }, { "epoch": 0.008095440988495952, "grad_norm": 20.375, "learning_rate": 2.375e-06, "loss": 0.8756, "step": 95 }, { "epoch": 0.00818065615679591, "grad_norm": 16.875, "learning_rate": 2.4000000000000003e-06, "loss": 1.0683, "step": 96 }, { "epoch": 0.008265871325095868, "grad_norm": 20.125, "learning_rate": 2.425e-06, "loss": 1.3241, "step": 97 }, { "epoch": 0.008351086493395825, "grad_norm": 13.375, "learning_rate": 2.4500000000000003e-06, "loss": 0.7591, "step": 98 }, { "epoch": 0.008436301661695782, "grad_norm": 12.9375, "learning_rate": 2.475e-06, "loss": 0.8464, "step": 99 }, { "epoch": 0.008521516829995739, "grad_norm": 20.75, "learning_rate": 2.5e-06, "loss": 1.0074, "step": 100 }, { "epoch": 0.008606731998295697, "grad_norm": 25.125, "learning_rate": 2.499999988705622e-06, "loss": 1.1472, "step": 101 }, { "epoch": 0.008691947166595654, "grad_norm": 21.375, "learning_rate": 2.4999999548224866e-06, "loss": 1.5212, "step": 102 }, { "epoch": 0.008777162334895611, "grad_norm": 13.25, "learning_rate": 2.499999898350595e-06, "loss": 0.6892, "step": 103 }, { "epoch": 0.008862377503195568, "grad_norm": 22.25, "learning_rate": 2.4999998192899483e-06, "loss": 1.219, "step": 104 }, { "epoch": 0.008947592671495527, "grad_norm": 19.875, "learning_rate": 2.4999997176405483e-06, "loss": 1.0446, "step": 105 }, { "epoch": 0.009032807839795483, "grad_norm": 12.5, "learning_rate": 2.4999995934023964e-06, "loss": 0.8379, "step": 106 }, { "epoch": 0.00911802300809544, "grad_norm": 19.25, "learning_rate": 2.4999994465754946e-06, "loss": 0.9465, "step": 107 }, { "epoch": 0.009203238176395399, "grad_norm": 19.5, "learning_rate": 2.4999992771598457e-06, "loss": 0.952, "step": 108 }, { "epoch": 0.009288453344695356, "grad_norm": 15.0, "learning_rate": 2.4999990851554527e-06, "loss": 0.8594, "step": 109 }, { "epoch": 0.009373668512995313, "grad_norm": 18.75, "learning_rate": 2.4999988705623203e-06, "loss": 0.9269, "step": 110 }, { "epoch": 0.00945888368129527, "grad_norm": 13.5, "learning_rate": 2.4999986333804506e-06, "loss": 0.7822, "step": 111 }, { "epoch": 0.009544098849595228, "grad_norm": 10.625, "learning_rate": 2.4999983736098486e-06, "loss": 0.5126, "step": 112 }, { "epoch": 0.009629314017895185, "grad_norm": 15.9375, "learning_rate": 2.499998091250519e-06, "loss": 0.9473, "step": 113 }, { "epoch": 0.009714529186195142, "grad_norm": 13.4375, "learning_rate": 2.499997786302467e-06, "loss": 0.7369, "step": 114 }, { "epoch": 0.0097997443544951, "grad_norm": 10.6875, "learning_rate": 2.499997458765698e-06, "loss": 0.5523, "step": 115 }, { "epoch": 0.009884959522795058, "grad_norm": 15.875, "learning_rate": 2.499997108640218e-06, "loss": 0.8407, "step": 116 }, { "epoch": 0.009970174691095015, "grad_norm": 18.875, "learning_rate": 2.4999967359260335e-06, "loss": 0.8085, "step": 117 }, { "epoch": 0.010055389859394971, "grad_norm": 15.3125, "learning_rate": 2.4999963406231508e-06, "loss": 0.923, "step": 118 }, { "epoch": 0.01014060502769493, "grad_norm": 15.4375, "learning_rate": 2.4999959227315773e-06, "loss": 1.1125, "step": 119 }, { "epoch": 0.010225820195994887, "grad_norm": 14.4375, "learning_rate": 2.4999954822513203e-06, "loss": 1.0749, "step": 120 }, { "epoch": 0.010311035364294844, "grad_norm": 13.4375, "learning_rate": 2.499995019182388e-06, "loss": 0.9686, "step": 121 }, { "epoch": 0.010396250532594803, "grad_norm": 21.125, "learning_rate": 2.4999945335247888e-06, "loss": 1.3533, "step": 122 }, { "epoch": 0.01048146570089476, "grad_norm": 14.625, "learning_rate": 2.4999940252785316e-06, "loss": 0.8111, "step": 123 }, { "epoch": 0.010566680869194716, "grad_norm": 13.0625, "learning_rate": 2.4999934944436256e-06, "loss": 0.6665, "step": 124 }, { "epoch": 0.010651896037494673, "grad_norm": 18.125, "learning_rate": 2.4999929410200797e-06, "loss": 0.7747, "step": 125 }, { "epoch": 0.010737111205794632, "grad_norm": 22.25, "learning_rate": 2.4999923650079044e-06, "loss": 1.0984, "step": 126 }, { "epoch": 0.010822326374094589, "grad_norm": 19.0, "learning_rate": 2.4999917664071105e-06, "loss": 1.0494, "step": 127 }, { "epoch": 0.010907541542394546, "grad_norm": 14.1875, "learning_rate": 2.499991145217708e-06, "loss": 0.9469, "step": 128 }, { "epoch": 0.010992756710694504, "grad_norm": 10.75, "learning_rate": 2.4999905014397087e-06, "loss": 0.5802, "step": 129 }, { "epoch": 0.011077971878994461, "grad_norm": 14.25, "learning_rate": 2.4999898350731242e-06, "loss": 0.9716, "step": 130 }, { "epoch": 0.011163187047294418, "grad_norm": 15.4375, "learning_rate": 2.4999891461179667e-06, "loss": 1.0489, "step": 131 }, { "epoch": 0.011248402215594375, "grad_norm": 21.0, "learning_rate": 2.4999884345742476e-06, "loss": 1.3434, "step": 132 }, { "epoch": 0.011333617383894334, "grad_norm": 14.8125, "learning_rate": 2.499987700441981e-06, "loss": 0.6333, "step": 133 }, { "epoch": 0.01141883255219429, "grad_norm": 15.375, "learning_rate": 2.4999869437211795e-06, "loss": 0.8497, "step": 134 }, { "epoch": 0.011504047720494247, "grad_norm": 22.875, "learning_rate": 2.499986164411857e-06, "loss": 1.2663, "step": 135 }, { "epoch": 0.011589262888794206, "grad_norm": 17.375, "learning_rate": 2.4999853625140277e-06, "loss": 1.0296, "step": 136 }, { "epoch": 0.011674478057094163, "grad_norm": 18.25, "learning_rate": 2.499984538027706e-06, "loss": 0.7126, "step": 137 }, { "epoch": 0.01175969322539412, "grad_norm": 25.375, "learning_rate": 2.4999836909529067e-06, "loss": 1.3332, "step": 138 }, { "epoch": 0.011844908393694077, "grad_norm": 12.8125, "learning_rate": 2.499982821289645e-06, "loss": 0.7095, "step": 139 }, { "epoch": 0.011930123561994035, "grad_norm": 15.3125, "learning_rate": 2.4999819290379367e-06, "loss": 1.0405, "step": 140 }, { "epoch": 0.012015338730293992, "grad_norm": 13.6875, "learning_rate": 2.499981014197798e-06, "loss": 0.537, "step": 141 }, { "epoch": 0.01210055389859395, "grad_norm": 12.125, "learning_rate": 2.4999800767692456e-06, "loss": 0.6729, "step": 142 }, { "epoch": 0.012185769066893908, "grad_norm": 20.125, "learning_rate": 2.4999791167522964e-06, "loss": 1.2783, "step": 143 }, { "epoch": 0.012270984235193865, "grad_norm": 16.125, "learning_rate": 2.4999781341469673e-06, "loss": 1.2955, "step": 144 }, { "epoch": 0.012356199403493822, "grad_norm": 11.1875, "learning_rate": 2.4999771289532765e-06, "loss": 0.6357, "step": 145 }, { "epoch": 0.012441414571793779, "grad_norm": 15.375, "learning_rate": 2.499976101171242e-06, "loss": 1.0036, "step": 146 }, { "epoch": 0.012526629740093737, "grad_norm": 24.125, "learning_rate": 2.4999750508008825e-06, "loss": 1.1779, "step": 147 }, { "epoch": 0.012611844908393694, "grad_norm": 13.0625, "learning_rate": 2.4999739778422164e-06, "loss": 0.8735, "step": 148 }, { "epoch": 0.012697060076693651, "grad_norm": 14.6875, "learning_rate": 2.499972882295264e-06, "loss": 0.7951, "step": 149 }, { "epoch": 0.01278227524499361, "grad_norm": 16.25, "learning_rate": 2.4999717641600446e-06, "loss": 0.8998, "step": 150 }, { "epoch": 0.012867490413293567, "grad_norm": 20.0, "learning_rate": 2.4999706234365786e-06, "loss": 1.4487, "step": 151 }, { "epoch": 0.012952705581593523, "grad_norm": 16.125, "learning_rate": 2.4999694601248865e-06, "loss": 1.0404, "step": 152 }, { "epoch": 0.01303792074989348, "grad_norm": 15.3125, "learning_rate": 2.4999682742249894e-06, "loss": 0.8277, "step": 153 }, { "epoch": 0.013123135918193439, "grad_norm": 18.875, "learning_rate": 2.499967065736908e-06, "loss": 0.9591, "step": 154 }, { "epoch": 0.013208351086493396, "grad_norm": 13.6875, "learning_rate": 2.4999658346606655e-06, "loss": 0.7346, "step": 155 }, { "epoch": 0.013293566254793353, "grad_norm": 19.25, "learning_rate": 2.499964580996283e-06, "loss": 0.8684, "step": 156 }, { "epoch": 0.013378781423093311, "grad_norm": 14.0625, "learning_rate": 2.499963304743784e-06, "loss": 0.8248, "step": 157 }, { "epoch": 0.013463996591393268, "grad_norm": 15.375, "learning_rate": 2.499962005903191e-06, "loss": 1.0007, "step": 158 }, { "epoch": 0.013549211759693225, "grad_norm": 12.9375, "learning_rate": 2.4999606844745272e-06, "loss": 0.8773, "step": 159 }, { "epoch": 0.013634426927993182, "grad_norm": 12.625, "learning_rate": 2.499959340457817e-06, "loss": 0.8744, "step": 160 }, { "epoch": 0.01371964209629314, "grad_norm": 20.25, "learning_rate": 2.499957973853085e-06, "loss": 1.0151, "step": 161 }, { "epoch": 0.013804857264593098, "grad_norm": 21.75, "learning_rate": 2.4999565846603556e-06, "loss": 1.0634, "step": 162 }, { "epoch": 0.013890072432893055, "grad_norm": 17.375, "learning_rate": 2.4999551728796533e-06, "loss": 0.9191, "step": 163 }, { "epoch": 0.013975287601193013, "grad_norm": 14.125, "learning_rate": 2.4999537385110047e-06, "loss": 0.7709, "step": 164 }, { "epoch": 0.01406050276949297, "grad_norm": 15.3125, "learning_rate": 2.4999522815544345e-06, "loss": 0.6242, "step": 165 }, { "epoch": 0.014145717937792927, "grad_norm": 12.375, "learning_rate": 2.49995080200997e-06, "loss": 0.6428, "step": 166 }, { "epoch": 0.014230933106092884, "grad_norm": 18.125, "learning_rate": 2.499949299877637e-06, "loss": 0.9351, "step": 167 }, { "epoch": 0.014316148274392843, "grad_norm": 27.75, "learning_rate": 2.499947775157464e-06, "loss": 1.3185, "step": 168 }, { "epoch": 0.0144013634426928, "grad_norm": 19.375, "learning_rate": 2.4999462278494776e-06, "loss": 0.847, "step": 169 }, { "epoch": 0.014486578610992756, "grad_norm": 16.25, "learning_rate": 2.4999446579537054e-06, "loss": 1.1577, "step": 170 }, { "epoch": 0.014571793779292713, "grad_norm": 10.25, "learning_rate": 2.499943065470177e-06, "loss": 0.4071, "step": 171 }, { "epoch": 0.014657008947592672, "grad_norm": 11.25, "learning_rate": 2.49994145039892e-06, "loss": 0.5795, "step": 172 }, { "epoch": 0.014742224115892629, "grad_norm": 14.875, "learning_rate": 2.499939812739964e-06, "loss": 0.7708, "step": 173 }, { "epoch": 0.014827439284192586, "grad_norm": 13.8125, "learning_rate": 2.499938152493339e-06, "loss": 0.9726, "step": 174 }, { "epoch": 0.014912654452492544, "grad_norm": 16.375, "learning_rate": 2.4999364696590747e-06, "loss": 0.9742, "step": 175 }, { "epoch": 0.014997869620792501, "grad_norm": 19.125, "learning_rate": 2.4999347642372017e-06, "loss": 0.7613, "step": 176 }, { "epoch": 0.015083084789092458, "grad_norm": 18.375, "learning_rate": 2.49993303622775e-06, "loss": 1.0638, "step": 177 }, { "epoch": 0.015168299957392415, "grad_norm": 16.25, "learning_rate": 2.499931285630752e-06, "loss": 0.7163, "step": 178 }, { "epoch": 0.015253515125692374, "grad_norm": 15.75, "learning_rate": 2.4999295124462384e-06, "loss": 0.958, "step": 179 }, { "epoch": 0.01533873029399233, "grad_norm": 13.9375, "learning_rate": 2.4999277166742418e-06, "loss": 0.8928, "step": 180 }, { "epoch": 0.015423945462292287, "grad_norm": 14.8125, "learning_rate": 2.4999258983147944e-06, "loss": 0.836, "step": 181 }, { "epoch": 0.015509160630592246, "grad_norm": 14.0, "learning_rate": 2.499924057367929e-06, "loss": 0.883, "step": 182 }, { "epoch": 0.015594375798892203, "grad_norm": 15.375, "learning_rate": 2.499922193833679e-06, "loss": 0.9203, "step": 183 }, { "epoch": 0.01567959096719216, "grad_norm": 10.375, "learning_rate": 2.4999203077120782e-06, "loss": 0.4823, "step": 184 }, { "epoch": 0.015764806135492117, "grad_norm": 20.125, "learning_rate": 2.4999183990031606e-06, "loss": 1.1846, "step": 185 }, { "epoch": 0.015850021303792074, "grad_norm": 14.125, "learning_rate": 2.4999164677069605e-06, "loss": 0.7984, "step": 186 }, { "epoch": 0.015935236472092034, "grad_norm": 15.375, "learning_rate": 2.499914513823513e-06, "loss": 0.7548, "step": 187 }, { "epoch": 0.01602045164039199, "grad_norm": 14.0, "learning_rate": 2.4999125373528532e-06, "loss": 0.8622, "step": 188 }, { "epoch": 0.016105666808691948, "grad_norm": 11.8125, "learning_rate": 2.499910538295017e-06, "loss": 0.4636, "step": 189 }, { "epoch": 0.016190881976991905, "grad_norm": 10.9375, "learning_rate": 2.4999085166500406e-06, "loss": 0.5639, "step": 190 }, { "epoch": 0.01627609714529186, "grad_norm": 13.75, "learning_rate": 2.4999064724179605e-06, "loss": 0.7793, "step": 191 }, { "epoch": 0.01636131231359182, "grad_norm": 13.4375, "learning_rate": 2.4999044055988135e-06, "loss": 0.8171, "step": 192 }, { "epoch": 0.016446527481891775, "grad_norm": 13.4375, "learning_rate": 2.499902316192637e-06, "loss": 0.6888, "step": 193 }, { "epoch": 0.016531742650191736, "grad_norm": 14.0625, "learning_rate": 2.4999002041994685e-06, "loss": 0.7249, "step": 194 }, { "epoch": 0.016616957818491693, "grad_norm": 13.1875, "learning_rate": 2.499898069619347e-06, "loss": 0.7304, "step": 195 }, { "epoch": 0.01670217298679165, "grad_norm": 12.6875, "learning_rate": 2.49989591245231e-06, "loss": 0.863, "step": 196 }, { "epoch": 0.016787388155091607, "grad_norm": 14.375, "learning_rate": 2.4998937326983974e-06, "loss": 0.7757, "step": 197 }, { "epoch": 0.016872603323391563, "grad_norm": 17.0, "learning_rate": 2.4998915303576478e-06, "loss": 0.8415, "step": 198 }, { "epoch": 0.01695781849169152, "grad_norm": 18.875, "learning_rate": 2.4998893054301014e-06, "loss": 1.3295, "step": 199 }, { "epoch": 0.017043033659991477, "grad_norm": 15.4375, "learning_rate": 2.4998870579157984e-06, "loss": 0.9835, "step": 200 }, { "epoch": 0.017128248828291438, "grad_norm": 21.75, "learning_rate": 2.4998847878147796e-06, "loss": 1.1561, "step": 201 }, { "epoch": 0.017213463996591395, "grad_norm": 10.4375, "learning_rate": 2.499882495127086e-06, "loss": 0.5615, "step": 202 }, { "epoch": 0.01729867916489135, "grad_norm": 14.3125, "learning_rate": 2.499880179852759e-06, "loss": 0.8957, "step": 203 }, { "epoch": 0.01738389433319131, "grad_norm": 17.25, "learning_rate": 2.4998778419918395e-06, "loss": 0.6773, "step": 204 }, { "epoch": 0.017469109501491265, "grad_norm": 12.25, "learning_rate": 2.4998754815443714e-06, "loss": 0.9909, "step": 205 }, { "epoch": 0.017554324669791222, "grad_norm": 13.1875, "learning_rate": 2.499873098510396e-06, "loss": 0.8327, "step": 206 }, { "epoch": 0.01763953983809118, "grad_norm": 15.5, "learning_rate": 2.4998706928899567e-06, "loss": 0.7118, "step": 207 }, { "epoch": 0.017724755006391136, "grad_norm": 13.0, "learning_rate": 2.4998682646830974e-06, "loss": 0.6673, "step": 208 }, { "epoch": 0.017809970174691096, "grad_norm": 19.75, "learning_rate": 2.499865813889862e-06, "loss": 0.6474, "step": 209 }, { "epoch": 0.017895185342991053, "grad_norm": 12.3125, "learning_rate": 2.499863340510294e-06, "loss": 0.6667, "step": 210 }, { "epoch": 0.01798040051129101, "grad_norm": 35.5, "learning_rate": 2.4998608445444387e-06, "loss": 0.7577, "step": 211 }, { "epoch": 0.018065615679590967, "grad_norm": 18.25, "learning_rate": 2.4998583259923413e-06, "loss": 0.5801, "step": 212 }, { "epoch": 0.018150830847890924, "grad_norm": 14.375, "learning_rate": 2.4998557848540473e-06, "loss": 0.6952, "step": 213 }, { "epoch": 0.01823604601619088, "grad_norm": 17.5, "learning_rate": 2.4998532211296024e-06, "loss": 0.9168, "step": 214 }, { "epoch": 0.018321261184490838, "grad_norm": 15.0, "learning_rate": 2.4998506348190523e-06, "loss": 0.7822, "step": 215 }, { "epoch": 0.018406476352790798, "grad_norm": 19.25, "learning_rate": 2.499848025922445e-06, "loss": 0.7619, "step": 216 }, { "epoch": 0.018491691521090755, "grad_norm": 16.25, "learning_rate": 2.4998453944398273e-06, "loss": 0.83, "step": 217 }, { "epoch": 0.018576906689390712, "grad_norm": 22.375, "learning_rate": 2.499842740371246e-06, "loss": 1.004, "step": 218 }, { "epoch": 0.01866212185769067, "grad_norm": 14.8125, "learning_rate": 2.4998400637167495e-06, "loss": 0.6864, "step": 219 }, { "epoch": 0.018747337025990626, "grad_norm": 13.25, "learning_rate": 2.4998373644763864e-06, "loss": 0.765, "step": 220 }, { "epoch": 0.018832552194290583, "grad_norm": 23.5, "learning_rate": 2.4998346426502052e-06, "loss": 1.162, "step": 221 }, { "epoch": 0.01891776736259054, "grad_norm": 12.375, "learning_rate": 2.4998318982382556e-06, "loss": 0.6497, "step": 222 }, { "epoch": 0.0190029825308905, "grad_norm": 13.0, "learning_rate": 2.4998291312405867e-06, "loss": 0.556, "step": 223 }, { "epoch": 0.019088197699190457, "grad_norm": 20.0, "learning_rate": 2.4998263416572484e-06, "loss": 0.8032, "step": 224 }, { "epoch": 0.019173412867490414, "grad_norm": 18.75, "learning_rate": 2.499823529488291e-06, "loss": 0.7872, "step": 225 }, { "epoch": 0.01925862803579037, "grad_norm": 16.75, "learning_rate": 2.499820694733766e-06, "loss": 0.7131, "step": 226 }, { "epoch": 0.019343843204090327, "grad_norm": 13.3125, "learning_rate": 2.499817837393724e-06, "loss": 0.7218, "step": 227 }, { "epoch": 0.019429058372390284, "grad_norm": 13.3125, "learning_rate": 2.4998149574682175e-06, "loss": 0.8036, "step": 228 }, { "epoch": 0.01951427354069024, "grad_norm": 14.9375, "learning_rate": 2.4998120549572972e-06, "loss": 0.9664, "step": 229 }, { "epoch": 0.0195994887089902, "grad_norm": 13.75, "learning_rate": 2.4998091298610168e-06, "loss": 0.8357, "step": 230 }, { "epoch": 0.01968470387729016, "grad_norm": 11.875, "learning_rate": 2.4998061821794282e-06, "loss": 0.7416, "step": 231 }, { "epoch": 0.019769919045590115, "grad_norm": 13.5, "learning_rate": 2.4998032119125858e-06, "loss": 0.5895, "step": 232 }, { "epoch": 0.019855134213890072, "grad_norm": 18.875, "learning_rate": 2.4998002190605423e-06, "loss": 0.8778, "step": 233 }, { "epoch": 0.01994034938219003, "grad_norm": 18.625, "learning_rate": 2.499797203623352e-06, "loss": 0.5428, "step": 234 }, { "epoch": 0.020025564550489986, "grad_norm": 16.75, "learning_rate": 2.4997941656010695e-06, "loss": 1.2933, "step": 235 }, { "epoch": 0.020110779718789943, "grad_norm": 16.25, "learning_rate": 2.4997911049937493e-06, "loss": 0.7574, "step": 236 }, { "epoch": 0.020195994887089903, "grad_norm": 18.75, "learning_rate": 2.4997880218014476e-06, "loss": 1.1304, "step": 237 }, { "epoch": 0.02028121005538986, "grad_norm": 14.4375, "learning_rate": 2.499784916024219e-06, "loss": 0.7651, "step": 238 }, { "epoch": 0.020366425223689817, "grad_norm": 16.625, "learning_rate": 2.4997817876621204e-06, "loss": 0.8226, "step": 239 }, { "epoch": 0.020451640391989774, "grad_norm": 14.5, "learning_rate": 2.4997786367152086e-06, "loss": 0.9215, "step": 240 }, { "epoch": 0.02053685556028973, "grad_norm": 17.25, "learning_rate": 2.4997754631835396e-06, "loss": 1.1801, "step": 241 }, { "epoch": 0.020622070728589688, "grad_norm": 13.5625, "learning_rate": 2.4997722670671715e-06, "loss": 0.8299, "step": 242 }, { "epoch": 0.020707285896889645, "grad_norm": 14.4375, "learning_rate": 2.4997690483661615e-06, "loss": 0.9563, "step": 243 }, { "epoch": 0.020792501065189605, "grad_norm": 11.4375, "learning_rate": 2.4997658070805684e-06, "loss": 0.5569, "step": 244 }, { "epoch": 0.020877716233489562, "grad_norm": 14.75, "learning_rate": 2.4997625432104504e-06, "loss": 0.9559, "step": 245 }, { "epoch": 0.02096293140178952, "grad_norm": 45.75, "learning_rate": 2.499759256755866e-06, "loss": 0.9006, "step": 246 }, { "epoch": 0.021048146570089476, "grad_norm": 18.875, "learning_rate": 2.499755947716876e-06, "loss": 1.1924, "step": 247 }, { "epoch": 0.021133361738389433, "grad_norm": 15.75, "learning_rate": 2.4997526160935385e-06, "loss": 0.8332, "step": 248 }, { "epoch": 0.02121857690668939, "grad_norm": 19.375, "learning_rate": 2.499749261885915e-06, "loss": 1.0387, "step": 249 }, { "epoch": 0.021303792074989347, "grad_norm": 18.625, "learning_rate": 2.4997458850940654e-06, "loss": 1.1405, "step": 250 }, { "epoch": 0.021389007243289307, "grad_norm": 10.3125, "learning_rate": 2.499742485718051e-06, "loss": 0.5173, "step": 251 }, { "epoch": 0.021474222411589264, "grad_norm": 13.625, "learning_rate": 2.4997390637579326e-06, "loss": 0.9637, "step": 252 }, { "epoch": 0.02155943757988922, "grad_norm": 21.75, "learning_rate": 2.4997356192137734e-06, "loss": 1.2523, "step": 253 }, { "epoch": 0.021644652748189178, "grad_norm": 17.875, "learning_rate": 2.4997321520856347e-06, "loss": 1.0835, "step": 254 }, { "epoch": 0.021729867916489135, "grad_norm": 14.6875, "learning_rate": 2.4997286623735788e-06, "loss": 0.8121, "step": 255 }, { "epoch": 0.02181508308478909, "grad_norm": 17.125, "learning_rate": 2.49972515007767e-06, "loss": 1.1225, "step": 256 }, { "epoch": 0.02190029825308905, "grad_norm": 9.75, "learning_rate": 2.499721615197971e-06, "loss": 0.3624, "step": 257 }, { "epoch": 0.02198551342138901, "grad_norm": 22.125, "learning_rate": 2.499718057734545e-06, "loss": 1.3517, "step": 258 }, { "epoch": 0.022070728589688966, "grad_norm": 23.625, "learning_rate": 2.4997144776874577e-06, "loss": 0.867, "step": 259 }, { "epoch": 0.022155943757988922, "grad_norm": 21.75, "learning_rate": 2.499710875056773e-06, "loss": 1.0509, "step": 260 }, { "epoch": 0.02224115892628888, "grad_norm": 39.75, "learning_rate": 2.4997072498425554e-06, "loss": 1.6689, "step": 261 }, { "epoch": 0.022326374094588836, "grad_norm": 36.0, "learning_rate": 2.4997036020448717e-06, "loss": 1.2132, "step": 262 }, { "epoch": 0.022411589262888793, "grad_norm": 16.25, "learning_rate": 2.4996999316637874e-06, "loss": 0.7928, "step": 263 }, { "epoch": 0.02249680443118875, "grad_norm": 12.9375, "learning_rate": 2.4996962386993685e-06, "loss": 0.8249, "step": 264 }, { "epoch": 0.02258201959948871, "grad_norm": 16.625, "learning_rate": 2.4996925231516816e-06, "loss": 0.744, "step": 265 }, { "epoch": 0.022667234767788667, "grad_norm": 11.25, "learning_rate": 2.4996887850207945e-06, "loss": 0.7313, "step": 266 }, { "epoch": 0.022752449936088624, "grad_norm": 9.0625, "learning_rate": 2.4996850243067744e-06, "loss": 0.3756, "step": 267 }, { "epoch": 0.02283766510438858, "grad_norm": 12.1875, "learning_rate": 2.4996812410096892e-06, "loss": 0.5983, "step": 268 }, { "epoch": 0.022922880272688538, "grad_norm": 11.875, "learning_rate": 2.499677435129607e-06, "loss": 0.4525, "step": 269 }, { "epoch": 0.023008095440988495, "grad_norm": 43.5, "learning_rate": 2.4996736066665973e-06, "loss": 1.4105, "step": 270 }, { "epoch": 0.023093310609288452, "grad_norm": 12.375, "learning_rate": 2.499669755620729e-06, "loss": 1.0042, "step": 271 }, { "epoch": 0.023178525777588412, "grad_norm": 14.3125, "learning_rate": 2.499665881992071e-06, "loss": 1.0983, "step": 272 }, { "epoch": 0.02326374094588837, "grad_norm": 17.875, "learning_rate": 2.4996619857806946e-06, "loss": 1.0489, "step": 273 }, { "epoch": 0.023348956114188326, "grad_norm": 14.25, "learning_rate": 2.499658066986669e-06, "loss": 0.9086, "step": 274 }, { "epoch": 0.023434171282488283, "grad_norm": 29.5, "learning_rate": 2.499654125610065e-06, "loss": 1.3386, "step": 275 }, { "epoch": 0.02351938645078824, "grad_norm": 16.75, "learning_rate": 2.4996501616509553e-06, "loss": 1.0422, "step": 276 }, { "epoch": 0.023604601619088197, "grad_norm": 13.375, "learning_rate": 2.4996461751094105e-06, "loss": 0.8694, "step": 277 }, { "epoch": 0.023689816787388154, "grad_norm": 16.125, "learning_rate": 2.4996421659855022e-06, "loss": 0.9185, "step": 278 }, { "epoch": 0.023775031955688114, "grad_norm": 16.25, "learning_rate": 2.4996381342793034e-06, "loss": 0.7571, "step": 279 }, { "epoch": 0.02386024712398807, "grad_norm": 16.75, "learning_rate": 2.4996340799908873e-06, "loss": 0.8722, "step": 280 }, { "epoch": 0.023945462292288028, "grad_norm": 12.0, "learning_rate": 2.4996300031203264e-06, "loss": 0.6721, "step": 281 }, { "epoch": 0.024030677460587985, "grad_norm": 21.25, "learning_rate": 2.499625903667695e-06, "loss": 0.7462, "step": 282 }, { "epoch": 0.02411589262888794, "grad_norm": 11.5, "learning_rate": 2.499621781633067e-06, "loss": 0.5317, "step": 283 }, { "epoch": 0.0242011077971879, "grad_norm": 13.9375, "learning_rate": 2.4996176370165164e-06, "loss": 0.7925, "step": 284 }, { "epoch": 0.024286322965487855, "grad_norm": 17.125, "learning_rate": 2.4996134698181184e-06, "loss": 1.1314, "step": 285 }, { "epoch": 0.024371538133787816, "grad_norm": 14.3125, "learning_rate": 2.499609280037949e-06, "loss": 0.8216, "step": 286 }, { "epoch": 0.024456753302087773, "grad_norm": 19.75, "learning_rate": 2.499605067676083e-06, "loss": 1.0857, "step": 287 }, { "epoch": 0.02454196847038773, "grad_norm": 12.0625, "learning_rate": 2.499600832732597e-06, "loss": 0.6202, "step": 288 }, { "epoch": 0.024627183638687686, "grad_norm": 18.375, "learning_rate": 2.4995965752075673e-06, "loss": 1.033, "step": 289 }, { "epoch": 0.024712398806987643, "grad_norm": 13.4375, "learning_rate": 2.499592295101071e-06, "loss": 0.648, "step": 290 }, { "epoch": 0.0247976139752876, "grad_norm": 18.75, "learning_rate": 2.499587992413185e-06, "loss": 1.016, "step": 291 }, { "epoch": 0.024882829143587557, "grad_norm": 14.8125, "learning_rate": 2.499583667143988e-06, "loss": 0.5957, "step": 292 }, { "epoch": 0.024968044311887518, "grad_norm": 21.875, "learning_rate": 2.499579319293557e-06, "loss": 0.9271, "step": 293 }, { "epoch": 0.025053259480187474, "grad_norm": 13.5625, "learning_rate": 2.4995749488619716e-06, "loss": 0.7341, "step": 294 }, { "epoch": 0.02513847464848743, "grad_norm": 14.375, "learning_rate": 2.49957055584931e-06, "loss": 0.8839, "step": 295 }, { "epoch": 0.025223689816787388, "grad_norm": 22.25, "learning_rate": 2.499566140255652e-06, "loss": 1.6852, "step": 296 }, { "epoch": 0.025308904985087345, "grad_norm": 26.5, "learning_rate": 2.4995617020810775e-06, "loss": 0.7103, "step": 297 }, { "epoch": 0.025394120153387302, "grad_norm": 16.125, "learning_rate": 2.4995572413256663e-06, "loss": 0.8754, "step": 298 }, { "epoch": 0.02547933532168726, "grad_norm": 17.125, "learning_rate": 2.4995527579894994e-06, "loss": 0.9294, "step": 299 }, { "epoch": 0.02556455048998722, "grad_norm": 14.0625, "learning_rate": 2.4995482520726573e-06, "loss": 0.6675, "step": 300 }, { "epoch": 0.025649765658287176, "grad_norm": 18.375, "learning_rate": 2.499543723575222e-06, "loss": 0.9184, "step": 301 }, { "epoch": 0.025734980826587133, "grad_norm": 21.125, "learning_rate": 2.499539172497275e-06, "loss": 0.9095, "step": 302 }, { "epoch": 0.02582019599488709, "grad_norm": 13.625, "learning_rate": 2.499534598838899e-06, "loss": 0.9908, "step": 303 }, { "epoch": 0.025905411163187047, "grad_norm": 17.5, "learning_rate": 2.4995300026001764e-06, "loss": 1.0894, "step": 304 }, { "epoch": 0.025990626331487004, "grad_norm": 12.125, "learning_rate": 2.4995253837811895e-06, "loss": 0.7298, "step": 305 }, { "epoch": 0.02607584149978696, "grad_norm": 16.5, "learning_rate": 2.499520742382023e-06, "loss": 0.8227, "step": 306 }, { "epoch": 0.02616105666808692, "grad_norm": 22.375, "learning_rate": 2.4995160784027596e-06, "loss": 1.1159, "step": 307 }, { "epoch": 0.026246271836386878, "grad_norm": 12.75, "learning_rate": 2.499511391843484e-06, "loss": 0.5054, "step": 308 }, { "epoch": 0.026331487004686835, "grad_norm": 13.6875, "learning_rate": 2.4995066827042823e-06, "loss": 0.8779, "step": 309 }, { "epoch": 0.026416702172986792, "grad_norm": 11.0625, "learning_rate": 2.4995019509852374e-06, "loss": 0.5208, "step": 310 }, { "epoch": 0.02650191734128675, "grad_norm": 15.75, "learning_rate": 2.4994971966864363e-06, "loss": 0.9356, "step": 311 }, { "epoch": 0.026587132509586706, "grad_norm": 21.125, "learning_rate": 2.499492419807964e-06, "loss": 0.8451, "step": 312 }, { "epoch": 0.026672347677886662, "grad_norm": 15.375, "learning_rate": 2.499487620349907e-06, "loss": 0.9195, "step": 313 }, { "epoch": 0.026757562846186623, "grad_norm": 15.5625, "learning_rate": 2.4994827983123527e-06, "loss": 0.8463, "step": 314 }, { "epoch": 0.02684277801448658, "grad_norm": 15.9375, "learning_rate": 2.4994779536953877e-06, "loss": 0.8846, "step": 315 }, { "epoch": 0.026927993182786537, "grad_norm": 23.625, "learning_rate": 2.4994730864990997e-06, "loss": 1.1352, "step": 316 }, { "epoch": 0.027013208351086494, "grad_norm": 27.5, "learning_rate": 2.4994681967235764e-06, "loss": 1.2751, "step": 317 }, { "epoch": 0.02709842351938645, "grad_norm": 13.25, "learning_rate": 2.4994632843689063e-06, "loss": 0.6847, "step": 318 }, { "epoch": 0.027183638687686407, "grad_norm": 29.25, "learning_rate": 2.4994583494351784e-06, "loss": 1.1252, "step": 319 }, { "epoch": 0.027268853855986364, "grad_norm": 24.625, "learning_rate": 2.4994533919224815e-06, "loss": 0.9692, "step": 320 }, { "epoch": 0.027354069024286325, "grad_norm": 17.25, "learning_rate": 2.4994484118309056e-06, "loss": 0.6664, "step": 321 }, { "epoch": 0.02743928419258628, "grad_norm": 20.375, "learning_rate": 2.4994434091605402e-06, "loss": 0.9731, "step": 322 }, { "epoch": 0.02752449936088624, "grad_norm": 20.0, "learning_rate": 2.4994383839114762e-06, "loss": 0.8914, "step": 323 }, { "epoch": 0.027609714529186195, "grad_norm": 13.9375, "learning_rate": 2.499433336083804e-06, "loss": 0.7852, "step": 324 }, { "epoch": 0.027694929697486152, "grad_norm": 13.6875, "learning_rate": 2.4994282656776154e-06, "loss": 0.5759, "step": 325 }, { "epoch": 0.02778014486578611, "grad_norm": 21.5, "learning_rate": 2.4994231726930012e-06, "loss": 0.8376, "step": 326 }, { "epoch": 0.027865360034086066, "grad_norm": 21.125, "learning_rate": 2.4994180571300546e-06, "loss": 1.1522, "step": 327 }, { "epoch": 0.027950575202386026, "grad_norm": 17.75, "learning_rate": 2.4994129189888667e-06, "loss": 0.6931, "step": 328 }, { "epoch": 0.028035790370685983, "grad_norm": 15.5625, "learning_rate": 2.4994077582695307e-06, "loss": 0.9413, "step": 329 }, { "epoch": 0.02812100553898594, "grad_norm": 11.9375, "learning_rate": 2.4994025749721405e-06, "loss": 0.7517, "step": 330 }, { "epoch": 0.028206220707285897, "grad_norm": 15.0625, "learning_rate": 2.49939736909679e-06, "loss": 0.622, "step": 331 }, { "epoch": 0.028291435875585854, "grad_norm": 20.625, "learning_rate": 2.499392140643572e-06, "loss": 1.185, "step": 332 }, { "epoch": 0.02837665104388581, "grad_norm": 22.5, "learning_rate": 2.4993868896125818e-06, "loss": 1.1348, "step": 333 }, { "epoch": 0.028461866212185768, "grad_norm": 13.0625, "learning_rate": 2.4993816160039137e-06, "loss": 0.5254, "step": 334 }, { "epoch": 0.028547081380485728, "grad_norm": 15.6875, "learning_rate": 2.499376319817664e-06, "loss": 1.1253, "step": 335 }, { "epoch": 0.028632296548785685, "grad_norm": 17.625, "learning_rate": 2.499371001053928e-06, "loss": 1.0347, "step": 336 }, { "epoch": 0.028717511717085642, "grad_norm": 11.25, "learning_rate": 2.499365659712801e-06, "loss": 0.5701, "step": 337 }, { "epoch": 0.0288027268853856, "grad_norm": 24.875, "learning_rate": 2.4993602957943808e-06, "loss": 1.1916, "step": 338 }, { "epoch": 0.028887942053685556, "grad_norm": 12.0625, "learning_rate": 2.4993549092987633e-06, "loss": 0.6434, "step": 339 }, { "epoch": 0.028973157221985513, "grad_norm": 14.625, "learning_rate": 2.4993495002260463e-06, "loss": 0.8653, "step": 340 }, { "epoch": 0.02905837239028547, "grad_norm": 15.375, "learning_rate": 2.4993440685763277e-06, "loss": 0.5834, "step": 341 }, { "epoch": 0.029143587558585426, "grad_norm": 11.9375, "learning_rate": 2.4993386143497056e-06, "loss": 0.5358, "step": 342 }, { "epoch": 0.029228802726885387, "grad_norm": 19.75, "learning_rate": 2.499333137546278e-06, "loss": 0.9414, "step": 343 }, { "epoch": 0.029314017895185344, "grad_norm": 21.75, "learning_rate": 2.499327638166145e-06, "loss": 1.1704, "step": 344 }, { "epoch": 0.0293992330634853, "grad_norm": 22.875, "learning_rate": 2.499322116209405e-06, "loss": 1.2015, "step": 345 }, { "epoch": 0.029484448231785258, "grad_norm": 14.5, "learning_rate": 2.4993165716761576e-06, "loss": 0.8069, "step": 346 }, { "epoch": 0.029569663400085214, "grad_norm": 11.9375, "learning_rate": 2.4993110045665037e-06, "loss": 0.4925, "step": 347 }, { "epoch": 0.02965487856838517, "grad_norm": 17.0, "learning_rate": 2.499305414880544e-06, "loss": 0.8036, "step": 348 }, { "epoch": 0.029740093736685128, "grad_norm": 13.125, "learning_rate": 2.499299802618379e-06, "loss": 0.6162, "step": 349 }, { "epoch": 0.02982530890498509, "grad_norm": 16.5, "learning_rate": 2.49929416778011e-06, "loss": 0.7732, "step": 350 }, { "epoch": 0.029910524073285046, "grad_norm": 23.0, "learning_rate": 2.4992885103658394e-06, "loss": 1.1847, "step": 351 }, { "epoch": 0.029995739241585002, "grad_norm": 13.25, "learning_rate": 2.4992828303756694e-06, "loss": 0.6363, "step": 352 }, { "epoch": 0.03008095440988496, "grad_norm": 11.25, "learning_rate": 2.4992771278097017e-06, "loss": 0.486, "step": 353 }, { "epoch": 0.030166169578184916, "grad_norm": 14.125, "learning_rate": 2.499271402668041e-06, "loss": 1.2398, "step": 354 }, { "epoch": 0.030251384746484873, "grad_norm": 13.4375, "learning_rate": 2.499265654950789e-06, "loss": 0.8022, "step": 355 }, { "epoch": 0.03033659991478483, "grad_norm": 19.75, "learning_rate": 2.4992598846580507e-06, "loss": 0.9641, "step": 356 }, { "epoch": 0.03042181508308479, "grad_norm": 16.125, "learning_rate": 2.4992540917899298e-06, "loss": 0.7756, "step": 357 }, { "epoch": 0.030507030251384747, "grad_norm": 16.125, "learning_rate": 2.499248276346532e-06, "loss": 1.0406, "step": 358 }, { "epoch": 0.030592245419684704, "grad_norm": 14.0625, "learning_rate": 2.499242438327961e-06, "loss": 0.6911, "step": 359 }, { "epoch": 0.03067746058798466, "grad_norm": 18.125, "learning_rate": 2.499236577734323e-06, "loss": 0.5916, "step": 360 }, { "epoch": 0.030762675756284618, "grad_norm": 16.0, "learning_rate": 2.499230694565724e-06, "loss": 0.9698, "step": 361 }, { "epoch": 0.030847890924584575, "grad_norm": 15.5, "learning_rate": 2.49922478882227e-06, "loss": 1.1002, "step": 362 }, { "epoch": 0.030933106092884532, "grad_norm": 15.3125, "learning_rate": 2.499218860504068e-06, "loss": 0.8531, "step": 363 }, { "epoch": 0.031018321261184492, "grad_norm": 22.5, "learning_rate": 2.4992129096112248e-06, "loss": 1.0199, "step": 364 }, { "epoch": 0.03110353642948445, "grad_norm": 15.375, "learning_rate": 2.499206936143848e-06, "loss": 0.65, "step": 365 }, { "epoch": 0.031188751597784406, "grad_norm": 19.5, "learning_rate": 2.499200940102046e-06, "loss": 0.8134, "step": 366 }, { "epoch": 0.03127396676608436, "grad_norm": 14.0625, "learning_rate": 2.499194921485927e-06, "loss": 0.9027, "step": 367 }, { "epoch": 0.03135918193438432, "grad_norm": 13.0625, "learning_rate": 2.4991888802955993e-06, "loss": 0.5978, "step": 368 }, { "epoch": 0.03144439710268428, "grad_norm": 11.25, "learning_rate": 2.4991828165311723e-06, "loss": 0.5651, "step": 369 }, { "epoch": 0.031529612270984234, "grad_norm": 20.625, "learning_rate": 2.499176730192756e-06, "loss": 1.0047, "step": 370 }, { "epoch": 0.03161482743928419, "grad_norm": 19.25, "learning_rate": 2.4991706212804596e-06, "loss": 0.9238, "step": 371 }, { "epoch": 0.03170004260758415, "grad_norm": 12.5625, "learning_rate": 2.4991644897943945e-06, "loss": 0.6923, "step": 372 }, { "epoch": 0.031785257775884104, "grad_norm": 13.4375, "learning_rate": 2.4991583357346705e-06, "loss": 0.7106, "step": 373 }, { "epoch": 0.03187047294418407, "grad_norm": 11.625, "learning_rate": 2.4991521591013995e-06, "loss": 0.4967, "step": 374 }, { "epoch": 0.031955688112484025, "grad_norm": 15.8125, "learning_rate": 2.499145959894693e-06, "loss": 0.9125, "step": 375 }, { "epoch": 0.03204090328078398, "grad_norm": 17.0, "learning_rate": 2.499139738114662e-06, "loss": 0.9366, "step": 376 }, { "epoch": 0.03212611844908394, "grad_norm": 17.0, "learning_rate": 2.4991334937614206e-06, "loss": 1.1503, "step": 377 }, { "epoch": 0.032211333617383896, "grad_norm": 13.5625, "learning_rate": 2.4991272268350807e-06, "loss": 0.7203, "step": 378 }, { "epoch": 0.03229654878568385, "grad_norm": 19.125, "learning_rate": 2.4991209373357558e-06, "loss": 0.6034, "step": 379 }, { "epoch": 0.03238176395398381, "grad_norm": 11.375, "learning_rate": 2.4991146252635593e-06, "loss": 0.5002, "step": 380 }, { "epoch": 0.032466979122283766, "grad_norm": 32.25, "learning_rate": 2.4991082906186057e-06, "loss": 1.7323, "step": 381 }, { "epoch": 0.03255219429058372, "grad_norm": 25.25, "learning_rate": 2.499101933401009e-06, "loss": 1.4725, "step": 382 }, { "epoch": 0.03263740945888368, "grad_norm": 14.4375, "learning_rate": 2.4990955536108846e-06, "loss": 0.7112, "step": 383 }, { "epoch": 0.03272262462718364, "grad_norm": 20.75, "learning_rate": 2.499089151248347e-06, "loss": 0.9326, "step": 384 }, { "epoch": 0.032807839795483594, "grad_norm": 19.625, "learning_rate": 2.4990827263135125e-06, "loss": 1.218, "step": 385 }, { "epoch": 0.03289305496378355, "grad_norm": 14.0625, "learning_rate": 2.499076278806498e-06, "loss": 0.8987, "step": 386 }, { "epoch": 0.03297827013208351, "grad_norm": 13.375, "learning_rate": 2.499069808727418e-06, "loss": 0.5982, "step": 387 }, { "epoch": 0.03306348530038347, "grad_norm": 21.375, "learning_rate": 2.4990633160763904e-06, "loss": 1.0252, "step": 388 }, { "epoch": 0.03314870046868343, "grad_norm": 13.8125, "learning_rate": 2.499056800853533e-06, "loss": 0.869, "step": 389 }, { "epoch": 0.033233915636983385, "grad_norm": 20.75, "learning_rate": 2.4990502630589637e-06, "loss": 0.5836, "step": 390 }, { "epoch": 0.03331913080528334, "grad_norm": 17.625, "learning_rate": 2.4990437026927993e-06, "loss": 1.0526, "step": 391 }, { "epoch": 0.0334043459735833, "grad_norm": 11.5625, "learning_rate": 2.49903711975516e-06, "loss": 0.5121, "step": 392 }, { "epoch": 0.033489561141883256, "grad_norm": 17.125, "learning_rate": 2.4990305142461635e-06, "loss": 0.7733, "step": 393 }, { "epoch": 0.03357477631018321, "grad_norm": 12.875, "learning_rate": 2.4990238861659293e-06, "loss": 0.6252, "step": 394 }, { "epoch": 0.03365999147848317, "grad_norm": 13.5, "learning_rate": 2.499017235514578e-06, "loss": 0.7226, "step": 395 }, { "epoch": 0.03374520664678313, "grad_norm": 20.875, "learning_rate": 2.499010562292229e-06, "loss": 1.1063, "step": 396 }, { "epoch": 0.033830421815083084, "grad_norm": 11.125, "learning_rate": 2.4990038664990036e-06, "loss": 0.523, "step": 397 }, { "epoch": 0.03391563698338304, "grad_norm": 20.25, "learning_rate": 2.4989971481350217e-06, "loss": 1.1988, "step": 398 }, { "epoch": 0.034000852151683, "grad_norm": 12.875, "learning_rate": 2.4989904072004056e-06, "loss": 0.6412, "step": 399 }, { "epoch": 0.034086067319982954, "grad_norm": 21.5, "learning_rate": 2.498983643695277e-06, "loss": 1.3073, "step": 400 }, { "epoch": 0.03417128248828291, "grad_norm": 14.6875, "learning_rate": 2.4989768576197583e-06, "loss": 0.8451, "step": 401 }, { "epoch": 0.034256497656582875, "grad_norm": 13.875, "learning_rate": 2.4989700489739713e-06, "loss": 0.6877, "step": 402 }, { "epoch": 0.03434171282488283, "grad_norm": 10.75, "learning_rate": 2.49896321775804e-06, "loss": 0.4754, "step": 403 }, { "epoch": 0.03442692799318279, "grad_norm": 17.875, "learning_rate": 2.498956363972087e-06, "loss": 0.9302, "step": 404 }, { "epoch": 0.034512143161482746, "grad_norm": 31.25, "learning_rate": 2.4989494876162368e-06, "loss": 1.2954, "step": 405 }, { "epoch": 0.0345973583297827, "grad_norm": 12.125, "learning_rate": 2.4989425886906133e-06, "loss": 0.5317, "step": 406 }, { "epoch": 0.03468257349808266, "grad_norm": 12.3125, "learning_rate": 2.4989356671953417e-06, "loss": 0.4213, "step": 407 }, { "epoch": 0.03476778866638262, "grad_norm": 20.375, "learning_rate": 2.4989287231305464e-06, "loss": 0.9238, "step": 408 }, { "epoch": 0.034853003834682574, "grad_norm": 13.6875, "learning_rate": 2.498921756496353e-06, "loss": 0.5157, "step": 409 }, { "epoch": 0.03493821900298253, "grad_norm": 17.5, "learning_rate": 2.4989147672928877e-06, "loss": 1.048, "step": 410 }, { "epoch": 0.03502343417128249, "grad_norm": 18.625, "learning_rate": 2.4989077555202766e-06, "loss": 0.9328, "step": 411 }, { "epoch": 0.035108649339582444, "grad_norm": 17.75, "learning_rate": 2.498900721178647e-06, "loss": 0.8179, "step": 412 }, { "epoch": 0.0351938645078824, "grad_norm": 18.75, "learning_rate": 2.498893664268125e-06, "loss": 1.1035, "step": 413 }, { "epoch": 0.03527907967618236, "grad_norm": 13.25, "learning_rate": 2.4988865847888388e-06, "loss": 0.7153, "step": 414 }, { "epoch": 0.035364294844482315, "grad_norm": 18.375, "learning_rate": 2.4988794827409165e-06, "loss": 0.9215, "step": 415 }, { "epoch": 0.03544951001278227, "grad_norm": 13.0, "learning_rate": 2.4988723581244857e-06, "loss": 0.5237, "step": 416 }, { "epoch": 0.035534725181082236, "grad_norm": 13.125, "learning_rate": 2.4988652109396756e-06, "loss": 0.8331, "step": 417 }, { "epoch": 0.03561994034938219, "grad_norm": 16.0, "learning_rate": 2.498858041186615e-06, "loss": 1.0079, "step": 418 }, { "epoch": 0.03570515551768215, "grad_norm": 15.1875, "learning_rate": 2.4988508488654342e-06, "loss": 0.6557, "step": 419 }, { "epoch": 0.035790370685982106, "grad_norm": 16.625, "learning_rate": 2.4988436339762625e-06, "loss": 1.0191, "step": 420 }, { "epoch": 0.03587558585428206, "grad_norm": 13.1875, "learning_rate": 2.4988363965192308e-06, "loss": 0.5937, "step": 421 }, { "epoch": 0.03596080102258202, "grad_norm": 18.0, "learning_rate": 2.4988291364944694e-06, "loss": 1.2424, "step": 422 }, { "epoch": 0.03604601619088198, "grad_norm": 21.625, "learning_rate": 2.4988218539021097e-06, "loss": 0.9282, "step": 423 }, { "epoch": 0.036131231359181934, "grad_norm": 13.0625, "learning_rate": 2.4988145487422833e-06, "loss": 0.4837, "step": 424 }, { "epoch": 0.03621644652748189, "grad_norm": 16.5, "learning_rate": 2.4988072210151224e-06, "loss": 0.8311, "step": 425 }, { "epoch": 0.03630166169578185, "grad_norm": 12.6875, "learning_rate": 2.4987998707207592e-06, "loss": 0.9997, "step": 426 }, { "epoch": 0.036386876864081805, "grad_norm": 13.5625, "learning_rate": 2.4987924978593263e-06, "loss": 0.8165, "step": 427 }, { "epoch": 0.03647209203238176, "grad_norm": 12.0625, "learning_rate": 2.4987851024309577e-06, "loss": 0.5936, "step": 428 }, { "epoch": 0.03655730720068172, "grad_norm": 14.375, "learning_rate": 2.4987776844357862e-06, "loss": 0.7518, "step": 429 }, { "epoch": 0.036642522368981675, "grad_norm": 16.125, "learning_rate": 2.4987702438739464e-06, "loss": 0.8977, "step": 430 }, { "epoch": 0.03672773753728164, "grad_norm": 38.75, "learning_rate": 2.498762780745572e-06, "loss": 1.6056, "step": 431 }, { "epoch": 0.036812952705581596, "grad_norm": 20.625, "learning_rate": 2.4987552950507994e-06, "loss": 1.0696, "step": 432 }, { "epoch": 0.03689816787388155, "grad_norm": 17.75, "learning_rate": 2.498747786789762e-06, "loss": 1.1169, "step": 433 }, { "epoch": 0.03698338304218151, "grad_norm": 10.375, "learning_rate": 2.4987402559625965e-06, "loss": 0.4596, "step": 434 }, { "epoch": 0.03706859821048147, "grad_norm": 14.25, "learning_rate": 2.4987327025694392e-06, "loss": 0.9313, "step": 435 }, { "epoch": 0.037153813378781424, "grad_norm": 12.6875, "learning_rate": 2.4987251266104263e-06, "loss": 0.7626, "step": 436 }, { "epoch": 0.03723902854708138, "grad_norm": 13.8125, "learning_rate": 2.4987175280856947e-06, "loss": 0.5739, "step": 437 }, { "epoch": 0.03732424371538134, "grad_norm": 30.625, "learning_rate": 2.4987099069953815e-06, "loss": 0.6925, "step": 438 }, { "epoch": 0.037409458883681294, "grad_norm": 13.6875, "learning_rate": 2.498702263339625e-06, "loss": 0.3508, "step": 439 }, { "epoch": 0.03749467405198125, "grad_norm": 15.4375, "learning_rate": 2.4986945971185623e-06, "loss": 0.9168, "step": 440 }, { "epoch": 0.03757988922028121, "grad_norm": 27.125, "learning_rate": 2.4986869083323324e-06, "loss": 0.8307, "step": 441 }, { "epoch": 0.037665104388581165, "grad_norm": 13.375, "learning_rate": 2.498679196981075e-06, "loss": 0.7669, "step": 442 }, { "epoch": 0.03775031955688112, "grad_norm": 19.25, "learning_rate": 2.4986714630649287e-06, "loss": 1.4351, "step": 443 }, { "epoch": 0.03783553472518108, "grad_norm": 16.375, "learning_rate": 2.4986637065840336e-06, "loss": 1.2245, "step": 444 }, { "epoch": 0.03792074989348104, "grad_norm": 13.3125, "learning_rate": 2.4986559275385294e-06, "loss": 0.7966, "step": 445 }, { "epoch": 0.038005965061781, "grad_norm": 12.8125, "learning_rate": 2.4986481259285574e-06, "loss": 0.6735, "step": 446 }, { "epoch": 0.03809118023008096, "grad_norm": 13.1875, "learning_rate": 2.4986403017542575e-06, "loss": 0.8942, "step": 447 }, { "epoch": 0.03817639539838091, "grad_norm": 14.75, "learning_rate": 2.498632455015772e-06, "loss": 0.8208, "step": 448 }, { "epoch": 0.03826161056668087, "grad_norm": 14.0625, "learning_rate": 2.4986245857132426e-06, "loss": 0.6493, "step": 449 }, { "epoch": 0.03834682573498083, "grad_norm": 15.1875, "learning_rate": 2.4986166938468114e-06, "loss": 0.9837, "step": 450 }, { "epoch": 0.038432040903280784, "grad_norm": 16.0, "learning_rate": 2.4986087794166206e-06, "loss": 0.8642, "step": 451 }, { "epoch": 0.03851725607158074, "grad_norm": 15.375, "learning_rate": 2.4986008424228143e-06, "loss": 0.9541, "step": 452 }, { "epoch": 0.0386024712398807, "grad_norm": 13.125, "learning_rate": 2.4985928828655346e-06, "loss": 0.7419, "step": 453 }, { "epoch": 0.038687686408180655, "grad_norm": 17.375, "learning_rate": 2.498584900744926e-06, "loss": 0.8758, "step": 454 }, { "epoch": 0.03877290157648061, "grad_norm": 13.875, "learning_rate": 2.498576896061133e-06, "loss": 0.7497, "step": 455 }, { "epoch": 0.03885811674478057, "grad_norm": 15.5625, "learning_rate": 2.4985688688142996e-06, "loss": 0.9892, "step": 456 }, { "epoch": 0.038943331913080526, "grad_norm": 12.0625, "learning_rate": 2.4985608190045714e-06, "loss": 0.6137, "step": 457 }, { "epoch": 0.03902854708138048, "grad_norm": 10.625, "learning_rate": 2.4985527466320937e-06, "loss": 0.4528, "step": 458 }, { "epoch": 0.039113762249680446, "grad_norm": 10.6875, "learning_rate": 2.4985446516970117e-06, "loss": 0.5635, "step": 459 }, { "epoch": 0.0391989774179804, "grad_norm": 16.625, "learning_rate": 2.4985365341994734e-06, "loss": 0.9851, "step": 460 }, { "epoch": 0.03928419258628036, "grad_norm": 24.625, "learning_rate": 2.498528394139624e-06, "loss": 1.373, "step": 461 }, { "epoch": 0.03936940775458032, "grad_norm": 17.875, "learning_rate": 2.498520231517611e-06, "loss": 0.8514, "step": 462 }, { "epoch": 0.039454622922880274, "grad_norm": 15.5, "learning_rate": 2.4985120463335816e-06, "loss": 1.1372, "step": 463 }, { "epoch": 0.03953983809118023, "grad_norm": 11.1875, "learning_rate": 2.498503838587684e-06, "loss": 0.4535, "step": 464 }, { "epoch": 0.03962505325948019, "grad_norm": 17.875, "learning_rate": 2.4984956082800673e-06, "loss": 0.8392, "step": 465 }, { "epoch": 0.039710268427780145, "grad_norm": 18.75, "learning_rate": 2.498487355410879e-06, "loss": 1.1626, "step": 466 }, { "epoch": 0.0397954835960801, "grad_norm": 16.5, "learning_rate": 2.498479079980269e-06, "loss": 1.2061, "step": 467 }, { "epoch": 0.03988069876438006, "grad_norm": 11.75, "learning_rate": 2.4984707819883863e-06, "loss": 0.6979, "step": 468 }, { "epoch": 0.039965913932680015, "grad_norm": 14.9375, "learning_rate": 2.4984624614353813e-06, "loss": 0.9957, "step": 469 }, { "epoch": 0.04005112910097997, "grad_norm": 17.25, "learning_rate": 2.4984541183214038e-06, "loss": 0.6148, "step": 470 }, { "epoch": 0.04013634426927993, "grad_norm": 13.8125, "learning_rate": 2.4984457526466055e-06, "loss": 0.614, "step": 471 }, { "epoch": 0.040221559437579886, "grad_norm": 13.5, "learning_rate": 2.498437364411137e-06, "loss": 0.7254, "step": 472 }, { "epoch": 0.04030677460587985, "grad_norm": 13.0, "learning_rate": 2.4984289536151496e-06, "loss": 0.7426, "step": 473 }, { "epoch": 0.04039198977417981, "grad_norm": 14.3125, "learning_rate": 2.4984205202587957e-06, "loss": 0.6586, "step": 474 }, { "epoch": 0.040477204942479764, "grad_norm": 16.375, "learning_rate": 2.498412064342228e-06, "loss": 0.7571, "step": 475 }, { "epoch": 0.04056242011077972, "grad_norm": 14.125, "learning_rate": 2.4984035858655985e-06, "loss": 0.8438, "step": 476 }, { "epoch": 0.04064763527907968, "grad_norm": 14.5, "learning_rate": 2.498395084829061e-06, "loss": 0.7932, "step": 477 }, { "epoch": 0.040732850447379634, "grad_norm": 10.8125, "learning_rate": 2.498386561232769e-06, "loss": 0.5459, "step": 478 }, { "epoch": 0.04081806561567959, "grad_norm": 12.5625, "learning_rate": 2.4983780150768763e-06, "loss": 0.529, "step": 479 }, { "epoch": 0.04090328078397955, "grad_norm": 13.625, "learning_rate": 2.498369446361538e-06, "loss": 0.6605, "step": 480 }, { "epoch": 0.040988495952279505, "grad_norm": 10.125, "learning_rate": 2.4983608550869078e-06, "loss": 0.4761, "step": 481 }, { "epoch": 0.04107371112057946, "grad_norm": 11.375, "learning_rate": 2.498352241253142e-06, "loss": 0.607, "step": 482 }, { "epoch": 0.04115892628887942, "grad_norm": 21.75, "learning_rate": 2.498343604860396e-06, "loss": 0.7981, "step": 483 }, { "epoch": 0.041244141457179376, "grad_norm": 16.0, "learning_rate": 2.498334945908825e-06, "loss": 1.15, "step": 484 }, { "epoch": 0.04132935662547933, "grad_norm": 16.875, "learning_rate": 2.4983262643985873e-06, "loss": 1.2257, "step": 485 }, { "epoch": 0.04141457179377929, "grad_norm": 32.5, "learning_rate": 2.498317560329838e-06, "loss": 1.0988, "step": 486 }, { "epoch": 0.04149978696207925, "grad_norm": 33.75, "learning_rate": 2.4983088337027354e-06, "loss": 1.1718, "step": 487 }, { "epoch": 0.04158500213037921, "grad_norm": 14.0, "learning_rate": 2.4983000845174365e-06, "loss": 0.9357, "step": 488 }, { "epoch": 0.04167021729867917, "grad_norm": 20.0, "learning_rate": 2.4982913127741006e-06, "loss": 1.2299, "step": 489 }, { "epoch": 0.041755432466979124, "grad_norm": 11.8125, "learning_rate": 2.4982825184728847e-06, "loss": 0.6561, "step": 490 }, { "epoch": 0.04184064763527908, "grad_norm": 17.125, "learning_rate": 2.4982737016139486e-06, "loss": 1.0078, "step": 491 }, { "epoch": 0.04192586280357904, "grad_norm": 13.5625, "learning_rate": 2.498264862197452e-06, "loss": 0.7485, "step": 492 }, { "epoch": 0.042011077971878995, "grad_norm": 15.375, "learning_rate": 2.4982560002235535e-06, "loss": 1.1153, "step": 493 }, { "epoch": 0.04209629314017895, "grad_norm": 19.25, "learning_rate": 2.498247115692414e-06, "loss": 1.0774, "step": 494 }, { "epoch": 0.04218150830847891, "grad_norm": 14.5, "learning_rate": 2.498238208604194e-06, "loss": 0.6382, "step": 495 }, { "epoch": 0.042266723476778865, "grad_norm": 12.625, "learning_rate": 2.498229278959054e-06, "loss": 0.72, "step": 496 }, { "epoch": 0.04235193864507882, "grad_norm": 13.3125, "learning_rate": 2.4982203267571563e-06, "loss": 0.8994, "step": 497 }, { "epoch": 0.04243715381337878, "grad_norm": 15.0, "learning_rate": 2.498211351998662e-06, "loss": 0.7694, "step": 498 }, { "epoch": 0.042522368981678736, "grad_norm": 13.125, "learning_rate": 2.498202354683733e-06, "loss": 0.7086, "step": 499 }, { "epoch": 0.04260758414997869, "grad_norm": 18.25, "learning_rate": 2.4981933348125327e-06, "loss": 1.0981, "step": 500 }, { "epoch": 0.04269279931827866, "grad_norm": 15.1875, "learning_rate": 2.4981842923852232e-06, "loss": 0.716, "step": 501 }, { "epoch": 0.042778014486578614, "grad_norm": 15.5625, "learning_rate": 2.498175227401969e-06, "loss": 0.8802, "step": 502 }, { "epoch": 0.04286322965487857, "grad_norm": 13.1875, "learning_rate": 2.498166139862933e-06, "loss": 0.613, "step": 503 }, { "epoch": 0.04294844482317853, "grad_norm": 10.875, "learning_rate": 2.49815702976828e-06, "loss": 0.4915, "step": 504 }, { "epoch": 0.043033659991478485, "grad_norm": 14.125, "learning_rate": 2.498147897118174e-06, "loss": 0.9676, "step": 505 }, { "epoch": 0.04311887515977844, "grad_norm": 17.125, "learning_rate": 2.4981387419127806e-06, "loss": 0.8427, "step": 506 }, { "epoch": 0.0432040903280784, "grad_norm": 15.375, "learning_rate": 2.4981295641522647e-06, "loss": 1.1045, "step": 507 }, { "epoch": 0.043289305496378355, "grad_norm": 17.25, "learning_rate": 2.498120363836793e-06, "loss": 1.0553, "step": 508 }, { "epoch": 0.04337452066467831, "grad_norm": 17.0, "learning_rate": 2.498111140966531e-06, "loss": 1.1626, "step": 509 }, { "epoch": 0.04345973583297827, "grad_norm": 17.0, "learning_rate": 2.4981018955416454e-06, "loss": 0.7456, "step": 510 }, { "epoch": 0.043544951001278226, "grad_norm": 20.875, "learning_rate": 2.498092627562304e-06, "loss": 0.9689, "step": 511 }, { "epoch": 0.04363016616957818, "grad_norm": 17.625, "learning_rate": 2.498083337028673e-06, "loss": 0.8882, "step": 512 }, { "epoch": 0.04371538133787814, "grad_norm": 10.5, "learning_rate": 2.4980740239409216e-06, "loss": 0.5084, "step": 513 }, { "epoch": 0.0438005965061781, "grad_norm": 13.875, "learning_rate": 2.498064688299218e-06, "loss": 0.8218, "step": 514 }, { "epoch": 0.04388581167447806, "grad_norm": 12.4375, "learning_rate": 2.4980553301037296e-06, "loss": 0.4602, "step": 515 }, { "epoch": 0.04397102684277802, "grad_norm": 17.375, "learning_rate": 2.4980459493546266e-06, "loss": 1.0326, "step": 516 }, { "epoch": 0.044056242011077974, "grad_norm": 31.25, "learning_rate": 2.4980365460520784e-06, "loss": 1.1643, "step": 517 }, { "epoch": 0.04414145717937793, "grad_norm": 13.5625, "learning_rate": 2.4980271201962547e-06, "loss": 0.7769, "step": 518 }, { "epoch": 0.04422667234767789, "grad_norm": 20.375, "learning_rate": 2.498017671787326e-06, "loss": 1.2848, "step": 519 }, { "epoch": 0.044311887515977845, "grad_norm": 17.125, "learning_rate": 2.498008200825463e-06, "loss": 0.9698, "step": 520 }, { "epoch": 0.0443971026842778, "grad_norm": 12.625, "learning_rate": 2.497998707310837e-06, "loss": 0.4529, "step": 521 }, { "epoch": 0.04448231785257776, "grad_norm": 16.125, "learning_rate": 2.497989191243619e-06, "loss": 1.1708, "step": 522 }, { "epoch": 0.044567533020877716, "grad_norm": 15.375, "learning_rate": 2.4979796526239816e-06, "loss": 0.9436, "step": 523 }, { "epoch": 0.04465274818917767, "grad_norm": 17.125, "learning_rate": 2.497970091452097e-06, "loss": 0.7137, "step": 524 }, { "epoch": 0.04473796335747763, "grad_norm": 16.875, "learning_rate": 2.4979605077281375e-06, "loss": 0.9704, "step": 525 }, { "epoch": 0.044823178525777586, "grad_norm": 12.5625, "learning_rate": 2.497950901452277e-06, "loss": 0.4724, "step": 526 }, { "epoch": 0.04490839369407754, "grad_norm": 19.25, "learning_rate": 2.4979412726246894e-06, "loss": 1.0186, "step": 527 }, { "epoch": 0.0449936088623775, "grad_norm": 22.25, "learning_rate": 2.4979316212455473e-06, "loss": 1.0909, "step": 528 }, { "epoch": 0.045078824030677464, "grad_norm": 13.5625, "learning_rate": 2.4979219473150263e-06, "loss": 0.5592, "step": 529 }, { "epoch": 0.04516403919897742, "grad_norm": 11.5, "learning_rate": 2.4979122508333004e-06, "loss": 0.678, "step": 530 }, { "epoch": 0.04524925436727738, "grad_norm": 15.8125, "learning_rate": 2.497902531800546e-06, "loss": 1.113, "step": 531 }, { "epoch": 0.045334469535577335, "grad_norm": 10.0625, "learning_rate": 2.4978927902169376e-06, "loss": 0.477, "step": 532 }, { "epoch": 0.04541968470387729, "grad_norm": 13.0625, "learning_rate": 2.4978830260826512e-06, "loss": 1.0077, "step": 533 }, { "epoch": 0.04550489987217725, "grad_norm": 14.125, "learning_rate": 2.4978732393978644e-06, "loss": 1.0396, "step": 534 }, { "epoch": 0.045590115040477205, "grad_norm": 15.875, "learning_rate": 2.497863430162753e-06, "loss": 1.0904, "step": 535 }, { "epoch": 0.04567533020877716, "grad_norm": 12.5, "learning_rate": 2.4978535983774945e-06, "loss": 0.5543, "step": 536 }, { "epoch": 0.04576054537707712, "grad_norm": 13.8125, "learning_rate": 2.4978437440422664e-06, "loss": 0.5372, "step": 537 }, { "epoch": 0.045845760545377076, "grad_norm": 11.5, "learning_rate": 2.4978338671572476e-06, "loss": 0.6376, "step": 538 }, { "epoch": 0.04593097571367703, "grad_norm": 15.25, "learning_rate": 2.4978239677226155e-06, "loss": 0.9596, "step": 539 }, { "epoch": 0.04601619088197699, "grad_norm": 28.875, "learning_rate": 2.49781404573855e-06, "loss": 1.2914, "step": 540 }, { "epoch": 0.04610140605027695, "grad_norm": 16.0, "learning_rate": 2.49780410120523e-06, "loss": 0.8702, "step": 541 }, { "epoch": 0.046186621218576904, "grad_norm": 14.9375, "learning_rate": 2.4977941341228347e-06, "loss": 0.6548, "step": 542 }, { "epoch": 0.04627183638687686, "grad_norm": 13.6875, "learning_rate": 2.4977841444915448e-06, "loss": 0.8117, "step": 543 }, { "epoch": 0.046357051555176824, "grad_norm": 12.625, "learning_rate": 2.4977741323115406e-06, "loss": 0.4431, "step": 544 }, { "epoch": 0.04644226672347678, "grad_norm": 14.4375, "learning_rate": 2.497764097583003e-06, "loss": 1.0855, "step": 545 }, { "epoch": 0.04652748189177674, "grad_norm": 14.3125, "learning_rate": 2.497754040306114e-06, "loss": 0.892, "step": 546 }, { "epoch": 0.046612697060076695, "grad_norm": 12.375, "learning_rate": 2.497743960481054e-06, "loss": 0.6311, "step": 547 }, { "epoch": 0.04669791222837665, "grad_norm": 14.6875, "learning_rate": 2.497733858108007e-06, "loss": 0.9285, "step": 548 }, { "epoch": 0.04678312739667661, "grad_norm": 21.25, "learning_rate": 2.4977237331871535e-06, "loss": 0.9391, "step": 549 }, { "epoch": 0.046868342564976566, "grad_norm": 15.375, "learning_rate": 2.497713585718678e-06, "loss": 0.8202, "step": 550 }, { "epoch": 0.04695355773327652, "grad_norm": 15.9375, "learning_rate": 2.497703415702763e-06, "loss": 0.9054, "step": 551 }, { "epoch": 0.04703877290157648, "grad_norm": 13.5625, "learning_rate": 2.4976932231395926e-06, "loss": 0.5876, "step": 552 }, { "epoch": 0.04712398806987644, "grad_norm": 12.375, "learning_rate": 2.4976830080293514e-06, "loss": 0.5749, "step": 553 }, { "epoch": 0.04720920323817639, "grad_norm": 16.75, "learning_rate": 2.497672770372223e-06, "loss": 1.0638, "step": 554 }, { "epoch": 0.04729441840647635, "grad_norm": 19.0, "learning_rate": 2.497662510168393e-06, "loss": 1.0446, "step": 555 }, { "epoch": 0.04737963357477631, "grad_norm": 28.25, "learning_rate": 2.4976522274180475e-06, "loss": 1.5739, "step": 556 }, { "epoch": 0.047464848743076264, "grad_norm": 22.125, "learning_rate": 2.497641922121371e-06, "loss": 1.1316, "step": 557 }, { "epoch": 0.04755006391137623, "grad_norm": 14.75, "learning_rate": 2.497631594278551e-06, "loss": 1.0194, "step": 558 }, { "epoch": 0.047635279079676185, "grad_norm": 12.25, "learning_rate": 2.497621243889773e-06, "loss": 0.5275, "step": 559 }, { "epoch": 0.04772049424797614, "grad_norm": 14.1875, "learning_rate": 2.4976108709552247e-06, "loss": 0.5929, "step": 560 }, { "epoch": 0.0478057094162761, "grad_norm": 18.25, "learning_rate": 2.4976004754750933e-06, "loss": 0.9813, "step": 561 }, { "epoch": 0.047890924584576056, "grad_norm": 11.1875, "learning_rate": 2.4975900574495672e-06, "loss": 0.3754, "step": 562 }, { "epoch": 0.04797613975287601, "grad_norm": 14.5, "learning_rate": 2.4975796168788337e-06, "loss": 0.7562, "step": 563 }, { "epoch": 0.04806135492117597, "grad_norm": 18.0, "learning_rate": 2.497569153763082e-06, "loss": 1.0602, "step": 564 }, { "epoch": 0.048146570089475926, "grad_norm": 21.375, "learning_rate": 2.4975586681025016e-06, "loss": 0.6363, "step": 565 }, { "epoch": 0.04823178525777588, "grad_norm": 23.625, "learning_rate": 2.497548159897281e-06, "loss": 0.9576, "step": 566 }, { "epoch": 0.04831700042607584, "grad_norm": 14.3125, "learning_rate": 2.497537629147611e-06, "loss": 0.901, "step": 567 }, { "epoch": 0.0484022155943758, "grad_norm": 13.4375, "learning_rate": 2.497527075853681e-06, "loss": 0.7384, "step": 568 }, { "epoch": 0.048487430762675754, "grad_norm": 15.875, "learning_rate": 2.497516500015683e-06, "loss": 0.8411, "step": 569 }, { "epoch": 0.04857264593097571, "grad_norm": 16.625, "learning_rate": 2.497505901633807e-06, "loss": 0.9076, "step": 570 }, { "epoch": 0.04865786109927567, "grad_norm": 14.6875, "learning_rate": 2.4974952807082454e-06, "loss": 0.6311, "step": 571 }, { "epoch": 0.04874307626757563, "grad_norm": 15.25, "learning_rate": 2.497484637239189e-06, "loss": 0.5999, "step": 572 }, { "epoch": 0.04882829143587559, "grad_norm": 15.5, "learning_rate": 2.4974739712268307e-06, "loss": 0.6612, "step": 573 }, { "epoch": 0.048913506604175545, "grad_norm": 15.625, "learning_rate": 2.4974632826713636e-06, "loss": 0.6295, "step": 574 }, { "epoch": 0.0489987217724755, "grad_norm": 14.1875, "learning_rate": 2.4974525715729808e-06, "loss": 0.4852, "step": 575 }, { "epoch": 0.04908393694077546, "grad_norm": 13.3125, "learning_rate": 2.497441837931875e-06, "loss": 0.5265, "step": 576 }, { "epoch": 0.049169152109075416, "grad_norm": 17.25, "learning_rate": 2.4974310817482414e-06, "loss": 0.891, "step": 577 }, { "epoch": 0.04925436727737537, "grad_norm": 21.125, "learning_rate": 2.4974203030222732e-06, "loss": 1.0509, "step": 578 }, { "epoch": 0.04933958244567533, "grad_norm": 15.0625, "learning_rate": 2.497409501754166e-06, "loss": 0.8276, "step": 579 }, { "epoch": 0.04942479761397529, "grad_norm": 17.0, "learning_rate": 2.497398677944115e-06, "loss": 0.9166, "step": 580 }, { "epoch": 0.049510012782275244, "grad_norm": 17.75, "learning_rate": 2.4973878315923153e-06, "loss": 1.0019, "step": 581 }, { "epoch": 0.0495952279505752, "grad_norm": 15.9375, "learning_rate": 2.497376962698963e-06, "loss": 1.0199, "step": 582 }, { "epoch": 0.04968044311887516, "grad_norm": 14.625, "learning_rate": 2.497366071264255e-06, "loss": 1.015, "step": 583 }, { "epoch": 0.049765658287175114, "grad_norm": 14.25, "learning_rate": 2.4973551572883875e-06, "loss": 0.9148, "step": 584 }, { "epoch": 0.04985087345547507, "grad_norm": 15.4375, "learning_rate": 2.497344220771558e-06, "loss": 1.0273, "step": 585 }, { "epoch": 0.049936088623775035, "grad_norm": 23.125, "learning_rate": 2.497333261713964e-06, "loss": 1.0809, "step": 586 }, { "epoch": 0.05002130379207499, "grad_norm": 18.375, "learning_rate": 2.4973222801158044e-06, "loss": 1.0111, "step": 587 }, { "epoch": 0.05010651896037495, "grad_norm": 10.5, "learning_rate": 2.4973112759772763e-06, "loss": 0.5785, "step": 588 }, { "epoch": 0.050191734128674906, "grad_norm": 15.8125, "learning_rate": 2.497300249298579e-06, "loss": 0.9292, "step": 589 }, { "epoch": 0.05027694929697486, "grad_norm": 18.625, "learning_rate": 2.497289200079912e-06, "loss": 1.1965, "step": 590 }, { "epoch": 0.05036216446527482, "grad_norm": 12.4375, "learning_rate": 2.4972781283214753e-06, "loss": 0.7628, "step": 591 }, { "epoch": 0.050447379633574777, "grad_norm": 18.625, "learning_rate": 2.497267034023468e-06, "loss": 1.1404, "step": 592 }, { "epoch": 0.05053259480187473, "grad_norm": 23.0, "learning_rate": 2.4972559171860917e-06, "loss": 0.9371, "step": 593 }, { "epoch": 0.05061780997017469, "grad_norm": 20.875, "learning_rate": 2.4972447778095468e-06, "loss": 1.2122, "step": 594 }, { "epoch": 0.05070302513847465, "grad_norm": 20.625, "learning_rate": 2.4972336158940344e-06, "loss": 1.4592, "step": 595 }, { "epoch": 0.050788240306774604, "grad_norm": 17.25, "learning_rate": 2.4972224314397564e-06, "loss": 0.9688, "step": 596 }, { "epoch": 0.05087345547507456, "grad_norm": 13.6875, "learning_rate": 2.497211224446915e-06, "loss": 0.7272, "step": 597 }, { "epoch": 0.05095867064337452, "grad_norm": 21.375, "learning_rate": 2.4971999949157123e-06, "loss": 0.7135, "step": 598 }, { "epoch": 0.051043885811674475, "grad_norm": 21.0, "learning_rate": 2.4971887428463515e-06, "loss": 1.1397, "step": 599 }, { "epoch": 0.05112910097997444, "grad_norm": 15.4375, "learning_rate": 2.4971774682390365e-06, "loss": 0.8959, "step": 600 }, { "epoch": 0.051214316148274396, "grad_norm": 25.625, "learning_rate": 2.49716617109397e-06, "loss": 1.153, "step": 601 }, { "epoch": 0.05129953131657435, "grad_norm": 14.3125, "learning_rate": 2.497154851411357e-06, "loss": 0.4679, "step": 602 }, { "epoch": 0.05138474648487431, "grad_norm": 15.125, "learning_rate": 2.4971435091914014e-06, "loss": 0.9237, "step": 603 }, { "epoch": 0.051469961653174266, "grad_norm": 12.3125, "learning_rate": 2.4971321444343084e-06, "loss": 0.6365, "step": 604 }, { "epoch": 0.05155517682147422, "grad_norm": 17.75, "learning_rate": 2.497120757140284e-06, "loss": 1.0036, "step": 605 }, { "epoch": 0.05164039198977418, "grad_norm": 12.625, "learning_rate": 2.4971093473095333e-06, "loss": 0.5024, "step": 606 }, { "epoch": 0.05172560715807414, "grad_norm": 15.3125, "learning_rate": 2.4970979149422624e-06, "loss": 0.5912, "step": 607 }, { "epoch": 0.051810822326374094, "grad_norm": 40.5, "learning_rate": 2.497086460038678e-06, "loss": 0.3844, "step": 608 }, { "epoch": 0.05189603749467405, "grad_norm": 14.0625, "learning_rate": 2.4970749825989877e-06, "loss": 0.9913, "step": 609 }, { "epoch": 0.05198125266297401, "grad_norm": 12.1875, "learning_rate": 2.497063482623398e-06, "loss": 0.56, "step": 610 }, { "epoch": 0.052066467831273965, "grad_norm": 27.875, "learning_rate": 2.4970519601121175e-06, "loss": 0.9598, "step": 611 }, { "epoch": 0.05215168299957392, "grad_norm": 12.25, "learning_rate": 2.4970404150653534e-06, "loss": 0.6372, "step": 612 }, { "epoch": 0.05223689816787388, "grad_norm": 16.375, "learning_rate": 2.497028847483316e-06, "loss": 1.0319, "step": 613 }, { "epoch": 0.05232211333617384, "grad_norm": 17.625, "learning_rate": 2.4970172573662124e-06, "loss": 1.1488, "step": 614 }, { "epoch": 0.0524073285044738, "grad_norm": 17.0, "learning_rate": 2.4970056447142535e-06, "loss": 0.7887, "step": 615 }, { "epoch": 0.052492543672773756, "grad_norm": 25.5, "learning_rate": 2.4969940095276483e-06, "loss": 1.1172, "step": 616 }, { "epoch": 0.05257775884107371, "grad_norm": 17.75, "learning_rate": 2.496982351806608e-06, "loss": 0.7054, "step": 617 }, { "epoch": 0.05266297400937367, "grad_norm": 18.125, "learning_rate": 2.496970671551342e-06, "loss": 1.325, "step": 618 }, { "epoch": 0.05274818917767363, "grad_norm": 20.25, "learning_rate": 2.4969589687620626e-06, "loss": 0.9802, "step": 619 }, { "epoch": 0.052833404345973584, "grad_norm": 15.25, "learning_rate": 2.49694724343898e-06, "loss": 1.0539, "step": 620 }, { "epoch": 0.05291861951427354, "grad_norm": 17.125, "learning_rate": 2.496935495582308e-06, "loss": 0.5991, "step": 621 }, { "epoch": 0.0530038346825735, "grad_norm": 12.875, "learning_rate": 2.4969237251922567e-06, "loss": 0.5742, "step": 622 }, { "epoch": 0.053089049850873454, "grad_norm": 20.0, "learning_rate": 2.49691193226904e-06, "loss": 1.1053, "step": 623 }, { "epoch": 0.05317426501917341, "grad_norm": 17.375, "learning_rate": 2.496900116812871e-06, "loss": 0.9491, "step": 624 }, { "epoch": 0.05325948018747337, "grad_norm": 20.625, "learning_rate": 2.496888278823963e-06, "loss": 1.0348, "step": 625 }, { "epoch": 0.053344695355773325, "grad_norm": 17.25, "learning_rate": 2.4968764183025297e-06, "loss": 0.7143, "step": 626 }, { "epoch": 0.05342991052407328, "grad_norm": 14.1875, "learning_rate": 2.496864535248786e-06, "loss": 0.8898, "step": 627 }, { "epoch": 0.053515125692373246, "grad_norm": 15.8125, "learning_rate": 2.496852629662946e-06, "loss": 1.0079, "step": 628 }, { "epoch": 0.0536003408606732, "grad_norm": 17.375, "learning_rate": 2.4968407015452255e-06, "loss": 0.9257, "step": 629 }, { "epoch": 0.05368555602897316, "grad_norm": 14.0625, "learning_rate": 2.4968287508958397e-06, "loss": 0.8194, "step": 630 }, { "epoch": 0.053770771197273116, "grad_norm": 15.9375, "learning_rate": 2.4968167777150044e-06, "loss": 0.7869, "step": 631 }, { "epoch": 0.05385598636557307, "grad_norm": 13.5, "learning_rate": 2.4968047820029364e-06, "loss": 0.4579, "step": 632 }, { "epoch": 0.05394120153387303, "grad_norm": 21.125, "learning_rate": 2.496792763759852e-06, "loss": 0.8788, "step": 633 }, { "epoch": 0.05402641670217299, "grad_norm": 16.0, "learning_rate": 2.4967807229859685e-06, "loss": 1.048, "step": 634 }, { "epoch": 0.054111631870472944, "grad_norm": 13.5, "learning_rate": 2.496768659681504e-06, "loss": 0.7781, "step": 635 }, { "epoch": 0.0541968470387729, "grad_norm": 12.0625, "learning_rate": 2.496756573846676e-06, "loss": 0.7897, "step": 636 }, { "epoch": 0.05428206220707286, "grad_norm": 14.25, "learning_rate": 2.496744465481703e-06, "loss": 0.7913, "step": 637 }, { "epoch": 0.054367277375372815, "grad_norm": 11.0625, "learning_rate": 2.4967323345868036e-06, "loss": 0.3933, "step": 638 }, { "epoch": 0.05445249254367277, "grad_norm": 13.5, "learning_rate": 2.496720181162197e-06, "loss": 0.8064, "step": 639 }, { "epoch": 0.05453770771197273, "grad_norm": 14.875, "learning_rate": 2.496708005208104e-06, "loss": 0.5314, "step": 640 }, { "epoch": 0.054622922880272685, "grad_norm": 20.0, "learning_rate": 2.4966958067247426e-06, "loss": 0.864, "step": 641 }, { "epoch": 0.05470813804857265, "grad_norm": 13.25, "learning_rate": 2.496683585712335e-06, "loss": 0.8099, "step": 642 }, { "epoch": 0.054793353216872606, "grad_norm": 12.0625, "learning_rate": 2.4966713421711013e-06, "loss": 0.3473, "step": 643 }, { "epoch": 0.05487856838517256, "grad_norm": 15.375, "learning_rate": 2.4966590761012627e-06, "loss": 0.7167, "step": 644 }, { "epoch": 0.05496378355347252, "grad_norm": 12.5625, "learning_rate": 2.496646787503041e-06, "loss": 0.4618, "step": 645 }, { "epoch": 0.05504899872177248, "grad_norm": 14.875, "learning_rate": 2.4966344763766583e-06, "loss": 0.8458, "step": 646 }, { "epoch": 0.055134213890072434, "grad_norm": 15.5, "learning_rate": 2.496622142722337e-06, "loss": 1.1295, "step": 647 }, { "epoch": 0.05521942905837239, "grad_norm": 12.6875, "learning_rate": 2.4966097865402995e-06, "loss": 0.6643, "step": 648 }, { "epoch": 0.05530464422667235, "grad_norm": 13.6875, "learning_rate": 2.4965974078307703e-06, "loss": 0.6109, "step": 649 }, { "epoch": 0.055389859394972304, "grad_norm": 13.6875, "learning_rate": 2.496585006593972e-06, "loss": 0.6909, "step": 650 }, { "epoch": 0.05547507456327226, "grad_norm": 20.375, "learning_rate": 2.4965725828301294e-06, "loss": 0.9731, "step": 651 }, { "epoch": 0.05556028973157222, "grad_norm": 14.0625, "learning_rate": 2.4965601365394666e-06, "loss": 0.9019, "step": 652 }, { "epoch": 0.055645504899872175, "grad_norm": 19.75, "learning_rate": 2.496547667722208e-06, "loss": 0.9211, "step": 653 }, { "epoch": 0.05573072006817213, "grad_norm": 14.625, "learning_rate": 2.4965351763785804e-06, "loss": 0.7265, "step": 654 }, { "epoch": 0.05581593523647209, "grad_norm": 17.75, "learning_rate": 2.4965226625088084e-06, "loss": 0.6484, "step": 655 }, { "epoch": 0.05590115040477205, "grad_norm": 16.25, "learning_rate": 2.4965101261131185e-06, "loss": 1.1539, "step": 656 }, { "epoch": 0.05598636557307201, "grad_norm": 12.6875, "learning_rate": 2.496497567191737e-06, "loss": 0.7442, "step": 657 }, { "epoch": 0.05607158074137197, "grad_norm": 45.0, "learning_rate": 2.4964849857448907e-06, "loss": 0.923, "step": 658 }, { "epoch": 0.056156795909671924, "grad_norm": 13.9375, "learning_rate": 2.496472381772808e-06, "loss": 0.5966, "step": 659 }, { "epoch": 0.05624201107797188, "grad_norm": 13.375, "learning_rate": 2.4964597552757155e-06, "loss": 0.6612, "step": 660 }, { "epoch": 0.05632722624627184, "grad_norm": 19.75, "learning_rate": 2.496447106253842e-06, "loss": 1.2901, "step": 661 }, { "epoch": 0.056412441414571794, "grad_norm": 21.125, "learning_rate": 2.4964344347074158e-06, "loss": 0.8304, "step": 662 }, { "epoch": 0.05649765658287175, "grad_norm": 16.125, "learning_rate": 2.496421740636666e-06, "loss": 1.0131, "step": 663 }, { "epoch": 0.05658287175117171, "grad_norm": 17.75, "learning_rate": 2.4964090240418215e-06, "loss": 0.9323, "step": 664 }, { "epoch": 0.056668086919471665, "grad_norm": 15.25, "learning_rate": 2.4963962849231135e-06, "loss": 0.7873, "step": 665 }, { "epoch": 0.05675330208777162, "grad_norm": 14.8125, "learning_rate": 2.4963835232807702e-06, "loss": 0.7644, "step": 666 }, { "epoch": 0.05683851725607158, "grad_norm": 16.875, "learning_rate": 2.496370739115024e-06, "loss": 1.2728, "step": 667 }, { "epoch": 0.056923732424371536, "grad_norm": 14.125, "learning_rate": 2.4963579324261052e-06, "loss": 0.6924, "step": 668 }, { "epoch": 0.05700894759267149, "grad_norm": 18.125, "learning_rate": 2.496345103214245e-06, "loss": 0.9334, "step": 669 }, { "epoch": 0.057094162760971456, "grad_norm": 14.8125, "learning_rate": 2.4963322514796755e-06, "loss": 0.6154, "step": 670 }, { "epoch": 0.05717937792927141, "grad_norm": 15.125, "learning_rate": 2.4963193772226295e-06, "loss": 1.0111, "step": 671 }, { "epoch": 0.05726459309757137, "grad_norm": 13.8125, "learning_rate": 2.4963064804433382e-06, "loss": 0.755, "step": 672 }, { "epoch": 0.05734980826587133, "grad_norm": 17.25, "learning_rate": 2.4962935611420364e-06, "loss": 1.1122, "step": 673 }, { "epoch": 0.057435023434171284, "grad_norm": 17.625, "learning_rate": 2.496280619318956e-06, "loss": 0.6942, "step": 674 }, { "epoch": 0.05752023860247124, "grad_norm": 20.25, "learning_rate": 2.496267654974332e-06, "loss": 0.9186, "step": 675 }, { "epoch": 0.0576054537707712, "grad_norm": 14.875, "learning_rate": 2.4962546681083982e-06, "loss": 0.7839, "step": 676 }, { "epoch": 0.057690668939071155, "grad_norm": 14.375, "learning_rate": 2.496241658721389e-06, "loss": 0.769, "step": 677 }, { "epoch": 0.05777588410737111, "grad_norm": 14.5625, "learning_rate": 2.4962286268135406e-06, "loss": 0.8106, "step": 678 }, { "epoch": 0.05786109927567107, "grad_norm": 15.5, "learning_rate": 2.4962155723850867e-06, "loss": 0.7286, "step": 679 }, { "epoch": 0.057946314443971025, "grad_norm": 18.5, "learning_rate": 2.4962024954362653e-06, "loss": 0.6561, "step": 680 }, { "epoch": 0.05803152961227098, "grad_norm": 19.875, "learning_rate": 2.496189395967311e-06, "loss": 0.9052, "step": 681 }, { "epoch": 0.05811674478057094, "grad_norm": 17.125, "learning_rate": 2.4961762739784613e-06, "loss": 1.5005, "step": 682 }, { "epoch": 0.058201959948870896, "grad_norm": 11.6875, "learning_rate": 2.496163129469953e-06, "loss": 0.4962, "step": 683 }, { "epoch": 0.05828717511717085, "grad_norm": 16.125, "learning_rate": 2.496149962442024e-06, "loss": 0.8615, "step": 684 }, { "epoch": 0.05837239028547082, "grad_norm": 11.0625, "learning_rate": 2.496136772894912e-06, "loss": 0.6755, "step": 685 }, { "epoch": 0.058457605453770774, "grad_norm": 20.125, "learning_rate": 2.496123560828856e-06, "loss": 1.1005, "step": 686 }, { "epoch": 0.05854282062207073, "grad_norm": 15.0625, "learning_rate": 2.4961103262440934e-06, "loss": 1.0075, "step": 687 }, { "epoch": 0.05862803579037069, "grad_norm": 15.1875, "learning_rate": 2.4960970691408645e-06, "loss": 0.6836, "step": 688 }, { "epoch": 0.058713250958670644, "grad_norm": 11.0, "learning_rate": 2.4960837895194083e-06, "loss": 0.7389, "step": 689 }, { "epoch": 0.0587984661269706, "grad_norm": 13.6875, "learning_rate": 2.4960704873799653e-06, "loss": 0.7842, "step": 690 }, { "epoch": 0.05888368129527056, "grad_norm": 15.4375, "learning_rate": 2.4960571627227753e-06, "loss": 0.2997, "step": 691 }, { "epoch": 0.058968896463570515, "grad_norm": 12.625, "learning_rate": 2.49604381554808e-06, "loss": 0.5738, "step": 692 }, { "epoch": 0.05905411163187047, "grad_norm": 11.5, "learning_rate": 2.496030445856119e-06, "loss": 0.4964, "step": 693 }, { "epoch": 0.05913932680017043, "grad_norm": 23.375, "learning_rate": 2.4960170536471355e-06, "loss": 0.8448, "step": 694 }, { "epoch": 0.059224541968470386, "grad_norm": 17.25, "learning_rate": 2.496003638921371e-06, "loss": 0.7069, "step": 695 }, { "epoch": 0.05930975713677034, "grad_norm": 17.375, "learning_rate": 2.495990201679067e-06, "loss": 0.7296, "step": 696 }, { "epoch": 0.0593949723050703, "grad_norm": 14.625, "learning_rate": 2.495976741920468e-06, "loss": 1.0706, "step": 697 }, { "epoch": 0.059480187473370257, "grad_norm": 16.375, "learning_rate": 2.4959632596458153e-06, "loss": 0.6849, "step": 698 }, { "epoch": 0.05956540264167022, "grad_norm": 14.25, "learning_rate": 2.4959497548553546e-06, "loss": 0.9897, "step": 699 }, { "epoch": 0.05965061780997018, "grad_norm": 16.25, "learning_rate": 2.4959362275493287e-06, "loss": 1.1302, "step": 700 }, { "epoch": 0.059735832978270134, "grad_norm": 14.0625, "learning_rate": 2.495922677727982e-06, "loss": 0.8003, "step": 701 }, { "epoch": 0.05982104814657009, "grad_norm": 11.3125, "learning_rate": 2.49590910539156e-06, "loss": 0.4348, "step": 702 }, { "epoch": 0.05990626331487005, "grad_norm": 19.125, "learning_rate": 2.495895510540307e-06, "loss": 0.8835, "step": 703 }, { "epoch": 0.059991478483170005, "grad_norm": 20.25, "learning_rate": 2.49588189317447e-06, "loss": 1.1262, "step": 704 }, { "epoch": 0.06007669365146996, "grad_norm": 13.25, "learning_rate": 2.495868253294294e-06, "loss": 0.6083, "step": 705 }, { "epoch": 0.06016190881976992, "grad_norm": 18.25, "learning_rate": 2.495854590900026e-06, "loss": 1.07, "step": 706 }, { "epoch": 0.060247123988069876, "grad_norm": 18.75, "learning_rate": 2.4958409059919124e-06, "loss": 1.0939, "step": 707 }, { "epoch": 0.06033233915636983, "grad_norm": 14.1875, "learning_rate": 2.495827198570201e-06, "loss": 1.0319, "step": 708 }, { "epoch": 0.06041755432466979, "grad_norm": 16.875, "learning_rate": 2.4958134686351396e-06, "loss": 0.9677, "step": 709 }, { "epoch": 0.060502769492969746, "grad_norm": 9.5, "learning_rate": 2.495799716186976e-06, "loss": 0.2821, "step": 710 }, { "epoch": 0.0605879846612697, "grad_norm": 16.75, "learning_rate": 2.4957859412259584e-06, "loss": 0.4358, "step": 711 }, { "epoch": 0.06067319982956966, "grad_norm": 14.5625, "learning_rate": 2.4957721437523365e-06, "loss": 1.0668, "step": 712 }, { "epoch": 0.060758414997869624, "grad_norm": 15.5625, "learning_rate": 2.495758323766359e-06, "loss": 0.9974, "step": 713 }, { "epoch": 0.06084363016616958, "grad_norm": 17.375, "learning_rate": 2.495744481268276e-06, "loss": 1.0701, "step": 714 }, { "epoch": 0.06092884533446954, "grad_norm": 10.375, "learning_rate": 2.4957306162583376e-06, "loss": 0.3589, "step": 715 }, { "epoch": 0.061014060502769495, "grad_norm": 13.375, "learning_rate": 2.495716728736794e-06, "loss": 0.6527, "step": 716 }, { "epoch": 0.06109927567106945, "grad_norm": 15.1875, "learning_rate": 2.4957028187038966e-06, "loss": 0.8977, "step": 717 }, { "epoch": 0.06118449083936941, "grad_norm": 16.5, "learning_rate": 2.4956888861598968e-06, "loss": 0.5937, "step": 718 }, { "epoch": 0.061269706007669365, "grad_norm": 30.625, "learning_rate": 2.495674931105046e-06, "loss": 1.0829, "step": 719 }, { "epoch": 0.06135492117596932, "grad_norm": 17.875, "learning_rate": 2.4956609535395966e-06, "loss": 1.0039, "step": 720 }, { "epoch": 0.06144013634426928, "grad_norm": 18.0, "learning_rate": 2.495646953463801e-06, "loss": 1.0935, "step": 721 }, { "epoch": 0.061525351512569236, "grad_norm": 18.875, "learning_rate": 2.4956329308779125e-06, "loss": 1.0674, "step": 722 }, { "epoch": 0.06161056668086919, "grad_norm": 19.375, "learning_rate": 2.4956188857821846e-06, "loss": 0.6899, "step": 723 }, { "epoch": 0.06169578184916915, "grad_norm": 16.75, "learning_rate": 2.495604818176871e-06, "loss": 0.7459, "step": 724 }, { "epoch": 0.06178099701746911, "grad_norm": 13.125, "learning_rate": 2.4955907280622253e-06, "loss": 0.6505, "step": 725 }, { "epoch": 0.061866212185769064, "grad_norm": 20.125, "learning_rate": 2.4955766154385025e-06, "loss": 1.2594, "step": 726 }, { "epoch": 0.06195142735406903, "grad_norm": 15.8125, "learning_rate": 2.495562480305958e-06, "loss": 0.8341, "step": 727 }, { "epoch": 0.062036642522368984, "grad_norm": 16.25, "learning_rate": 2.495548322664847e-06, "loss": 0.582, "step": 728 }, { "epoch": 0.06212185769066894, "grad_norm": 17.625, "learning_rate": 2.495534142515425e-06, "loss": 0.8811, "step": 729 }, { "epoch": 0.0622070728589689, "grad_norm": 16.75, "learning_rate": 2.495519939857949e-06, "loss": 0.984, "step": 730 }, { "epoch": 0.062292288027268855, "grad_norm": 23.75, "learning_rate": 2.4955057146926754e-06, "loss": 1.0588, "step": 731 }, { "epoch": 0.06237750319556881, "grad_norm": 32.75, "learning_rate": 2.4954914670198605e-06, "loss": 1.0906, "step": 732 }, { "epoch": 0.06246271836386877, "grad_norm": 39.5, "learning_rate": 2.4954771968397624e-06, "loss": 1.5367, "step": 733 }, { "epoch": 0.06254793353216873, "grad_norm": 16.0, "learning_rate": 2.4954629041526393e-06, "loss": 0.6287, "step": 734 }, { "epoch": 0.06263314870046868, "grad_norm": 10.375, "learning_rate": 2.495448588958749e-06, "loss": 0.3451, "step": 735 }, { "epoch": 0.06271836386876864, "grad_norm": 13.6875, "learning_rate": 2.4954342512583503e-06, "loss": 0.789, "step": 736 }, { "epoch": 0.0628035790370686, "grad_norm": 13.8125, "learning_rate": 2.495419891051702e-06, "loss": 0.4972, "step": 737 }, { "epoch": 0.06288879420536855, "grad_norm": 17.125, "learning_rate": 2.4954055083390643e-06, "loss": 0.7407, "step": 738 }, { "epoch": 0.06297400937366851, "grad_norm": 18.125, "learning_rate": 2.4953911031206966e-06, "loss": 1.0121, "step": 739 }, { "epoch": 0.06305922454196847, "grad_norm": 17.375, "learning_rate": 2.495376675396859e-06, "loss": 0.5802, "step": 740 }, { "epoch": 0.06314443971026842, "grad_norm": 11.875, "learning_rate": 2.4953622251678133e-06, "loss": 0.5409, "step": 741 }, { "epoch": 0.06322965487856838, "grad_norm": 12.5625, "learning_rate": 2.4953477524338194e-06, "loss": 0.6038, "step": 742 }, { "epoch": 0.06331487004686834, "grad_norm": 13.0, "learning_rate": 2.495333257195139e-06, "loss": 0.5836, "step": 743 }, { "epoch": 0.0634000852151683, "grad_norm": 17.5, "learning_rate": 2.4953187394520348e-06, "loss": 0.9286, "step": 744 }, { "epoch": 0.06348530038346825, "grad_norm": 18.875, "learning_rate": 2.4953041992047687e-06, "loss": 0.8986, "step": 745 }, { "epoch": 0.06357051555176821, "grad_norm": 11.125, "learning_rate": 2.4952896364536032e-06, "loss": 0.4778, "step": 746 }, { "epoch": 0.06365573072006817, "grad_norm": 20.25, "learning_rate": 2.495275051198802e-06, "loss": 1.0113, "step": 747 }, { "epoch": 0.06374094588836814, "grad_norm": 13.875, "learning_rate": 2.4952604434406285e-06, "loss": 0.9292, "step": 748 }, { "epoch": 0.0638261610566681, "grad_norm": 11.9375, "learning_rate": 2.495245813179346e-06, "loss": 0.8457, "step": 749 }, { "epoch": 0.06391137622496805, "grad_norm": 15.75, "learning_rate": 2.49523116041522e-06, "loss": 0.9048, "step": 750 }, { "epoch": 0.06399659139326801, "grad_norm": 11.25, "learning_rate": 2.4952164851485143e-06, "loss": 0.5674, "step": 751 }, { "epoch": 0.06408180656156796, "grad_norm": 14.9375, "learning_rate": 2.495201787379495e-06, "loss": 0.9348, "step": 752 }, { "epoch": 0.06416702172986792, "grad_norm": 15.125, "learning_rate": 2.495187067108427e-06, "loss": 1.1715, "step": 753 }, { "epoch": 0.06425223689816788, "grad_norm": 14.5, "learning_rate": 2.495172324335577e-06, "loss": 0.5497, "step": 754 }, { "epoch": 0.06433745206646783, "grad_norm": 13.375, "learning_rate": 2.4951575590612106e-06, "loss": 0.503, "step": 755 }, { "epoch": 0.06442266723476779, "grad_norm": 15.375, "learning_rate": 2.4951427712855952e-06, "loss": 1.2421, "step": 756 }, { "epoch": 0.06450788240306775, "grad_norm": 13.25, "learning_rate": 2.4951279610089978e-06, "loss": 0.5308, "step": 757 }, { "epoch": 0.0645930975713677, "grad_norm": 19.625, "learning_rate": 2.4951131282316858e-06, "loss": 1.1092, "step": 758 }, { "epoch": 0.06467831273966766, "grad_norm": 13.3125, "learning_rate": 2.4950982729539274e-06, "loss": 0.8727, "step": 759 }, { "epoch": 0.06476352790796762, "grad_norm": 13.6875, "learning_rate": 2.495083395175992e-06, "loss": 0.8149, "step": 760 }, { "epoch": 0.06484874307626758, "grad_norm": 13.875, "learning_rate": 2.495068494898147e-06, "loss": 0.6574, "step": 761 }, { "epoch": 0.06493395824456753, "grad_norm": 16.0, "learning_rate": 2.495053572120662e-06, "loss": 0.8122, "step": 762 }, { "epoch": 0.06501917341286749, "grad_norm": 12.1875, "learning_rate": 2.4950386268438074e-06, "loss": 0.6436, "step": 763 }, { "epoch": 0.06510438858116745, "grad_norm": 10.4375, "learning_rate": 2.4950236590678526e-06, "loss": 0.3646, "step": 764 }, { "epoch": 0.0651896037494674, "grad_norm": 20.875, "learning_rate": 2.4950086687930684e-06, "loss": 0.7299, "step": 765 }, { "epoch": 0.06527481891776736, "grad_norm": 14.5, "learning_rate": 2.494993656019726e-06, "loss": 0.8499, "step": 766 }, { "epoch": 0.06536003408606732, "grad_norm": 18.375, "learning_rate": 2.494978620748096e-06, "loss": 1.2044, "step": 767 }, { "epoch": 0.06544524925436727, "grad_norm": 13.8125, "learning_rate": 2.49496356297845e-06, "loss": 0.4475, "step": 768 }, { "epoch": 0.06553046442266723, "grad_norm": 11.25, "learning_rate": 2.494948482711061e-06, "loss": 0.5462, "step": 769 }, { "epoch": 0.06561567959096719, "grad_norm": 15.125, "learning_rate": 2.494933379946201e-06, "loss": 0.9137, "step": 770 }, { "epoch": 0.06570089475926714, "grad_norm": 15.5, "learning_rate": 2.4949182546841426e-06, "loss": 1.0257, "step": 771 }, { "epoch": 0.0657861099275671, "grad_norm": 13.6875, "learning_rate": 2.49490310692516e-06, "loss": 0.9288, "step": 772 }, { "epoch": 0.06587132509586706, "grad_norm": 13.625, "learning_rate": 2.494887936669526e-06, "loss": 0.7343, "step": 773 }, { "epoch": 0.06595654026416702, "grad_norm": 19.0, "learning_rate": 2.4948727439175153e-06, "loss": 0.9281, "step": 774 }, { "epoch": 0.06604175543246697, "grad_norm": 19.0, "learning_rate": 2.4948575286694022e-06, "loss": 0.632, "step": 775 }, { "epoch": 0.06612697060076694, "grad_norm": 17.25, "learning_rate": 2.494842290925462e-06, "loss": 0.7674, "step": 776 }, { "epoch": 0.0662121857690669, "grad_norm": 11.625, "learning_rate": 2.4948270306859695e-06, "loss": 0.5834, "step": 777 }, { "epoch": 0.06629740093736686, "grad_norm": 16.5, "learning_rate": 2.494811747951201e-06, "loss": 0.7233, "step": 778 }, { "epoch": 0.06638261610566681, "grad_norm": 12.3125, "learning_rate": 2.4947964427214326e-06, "loss": 0.75, "step": 779 }, { "epoch": 0.06646783127396677, "grad_norm": 14.625, "learning_rate": 2.49478111499694e-06, "loss": 0.591, "step": 780 }, { "epoch": 0.06655304644226673, "grad_norm": 16.5, "learning_rate": 2.494765764778002e-06, "loss": 0.8904, "step": 781 }, { "epoch": 0.06663826161056668, "grad_norm": 16.75, "learning_rate": 2.494750392064894e-06, "loss": 0.9807, "step": 782 }, { "epoch": 0.06672347677886664, "grad_norm": 34.5, "learning_rate": 2.4947349968578956e-06, "loss": 0.9434, "step": 783 }, { "epoch": 0.0668086919471666, "grad_norm": 10.5625, "learning_rate": 2.494719579157283e-06, "loss": 0.5366, "step": 784 }, { "epoch": 0.06689390711546656, "grad_norm": 13.1875, "learning_rate": 2.494704138963337e-06, "loss": 0.648, "step": 785 }, { "epoch": 0.06697912228376651, "grad_norm": 14.9375, "learning_rate": 2.494688676276335e-06, "loss": 1.1552, "step": 786 }, { "epoch": 0.06706433745206647, "grad_norm": 14.4375, "learning_rate": 2.494673191096557e-06, "loss": 1.0506, "step": 787 }, { "epoch": 0.06714955262036643, "grad_norm": 11.0, "learning_rate": 2.494657683424283e-06, "loss": 0.4094, "step": 788 }, { "epoch": 0.06723476778866638, "grad_norm": 17.0, "learning_rate": 2.494642153259793e-06, "loss": 0.9257, "step": 789 }, { "epoch": 0.06731998295696634, "grad_norm": 24.375, "learning_rate": 2.494626600603368e-06, "loss": 0.6784, "step": 790 }, { "epoch": 0.0674051981252663, "grad_norm": 14.125, "learning_rate": 2.4946110254552886e-06, "loss": 0.6307, "step": 791 }, { "epoch": 0.06749041329356625, "grad_norm": 18.75, "learning_rate": 2.494595427815836e-06, "loss": 1.0362, "step": 792 }, { "epoch": 0.06757562846186621, "grad_norm": 13.0625, "learning_rate": 2.4945798076852928e-06, "loss": 0.6927, "step": 793 }, { "epoch": 0.06766084363016617, "grad_norm": 15.0, "learning_rate": 2.494564165063941e-06, "loss": 1.057, "step": 794 }, { "epoch": 0.06774605879846612, "grad_norm": 16.125, "learning_rate": 2.494548499952063e-06, "loss": 0.5934, "step": 795 }, { "epoch": 0.06783127396676608, "grad_norm": 13.75, "learning_rate": 2.4945328123499424e-06, "loss": 0.7026, "step": 796 }, { "epoch": 0.06791648913506604, "grad_norm": 15.4375, "learning_rate": 2.4945171022578622e-06, "loss": 0.9534, "step": 797 }, { "epoch": 0.068001704303366, "grad_norm": 12.0625, "learning_rate": 2.4945013696761066e-06, "loss": 0.6558, "step": 798 }, { "epoch": 0.06808691947166595, "grad_norm": 35.25, "learning_rate": 2.49448561460496e-06, "loss": 1.0942, "step": 799 }, { "epoch": 0.06817213463996591, "grad_norm": 18.625, "learning_rate": 2.4944698370447066e-06, "loss": 0.7719, "step": 800 }, { "epoch": 0.06825734980826587, "grad_norm": 11.0625, "learning_rate": 2.494454036995632e-06, "loss": 0.7603, "step": 801 }, { "epoch": 0.06834256497656582, "grad_norm": 13.125, "learning_rate": 2.4944382144580213e-06, "loss": 0.7476, "step": 802 }, { "epoch": 0.06842778014486578, "grad_norm": 13.1875, "learning_rate": 2.494422369432161e-06, "loss": 0.3921, "step": 803 }, { "epoch": 0.06851299531316575, "grad_norm": 15.75, "learning_rate": 2.494406501918337e-06, "loss": 1.0187, "step": 804 }, { "epoch": 0.06859821048146571, "grad_norm": 13.625, "learning_rate": 2.494390611916836e-06, "loss": 0.6273, "step": 805 }, { "epoch": 0.06868342564976566, "grad_norm": 13.5, "learning_rate": 2.4943746994279456e-06, "loss": 0.7789, "step": 806 }, { "epoch": 0.06876864081806562, "grad_norm": 23.75, "learning_rate": 2.4943587644519528e-06, "loss": 1.0404, "step": 807 }, { "epoch": 0.06885385598636558, "grad_norm": 9.25, "learning_rate": 2.494342806989146e-06, "loss": 0.3221, "step": 808 }, { "epoch": 0.06893907115466553, "grad_norm": 21.75, "learning_rate": 2.4943268270398135e-06, "loss": 1.2056, "step": 809 }, { "epoch": 0.06902428632296549, "grad_norm": 14.3125, "learning_rate": 2.494310824604244e-06, "loss": 0.5997, "step": 810 }, { "epoch": 0.06910950149126545, "grad_norm": 13.6875, "learning_rate": 2.494294799682726e-06, "loss": 0.7444, "step": 811 }, { "epoch": 0.0691947166595654, "grad_norm": 24.25, "learning_rate": 2.494278752275551e-06, "loss": 0.7059, "step": 812 }, { "epoch": 0.06927993182786536, "grad_norm": 13.6875, "learning_rate": 2.494262682383006e-06, "loss": 0.9566, "step": 813 }, { "epoch": 0.06936514699616532, "grad_norm": 10.0625, "learning_rate": 2.4942465900053846e-06, "loss": 0.5052, "step": 814 }, { "epoch": 0.06945036216446528, "grad_norm": 6.25, "learning_rate": 2.4942304751429757e-06, "loss": 0.1778, "step": 815 }, { "epoch": 0.06953557733276523, "grad_norm": 13.1875, "learning_rate": 2.4942143377960708e-06, "loss": 0.6495, "step": 816 }, { "epoch": 0.06962079250106519, "grad_norm": 9.5625, "learning_rate": 2.4941981779649614e-06, "loss": 0.3381, "step": 817 }, { "epoch": 0.06970600766936515, "grad_norm": 13.0625, "learning_rate": 2.4941819956499404e-06, "loss": 0.8703, "step": 818 }, { "epoch": 0.0697912228376651, "grad_norm": 13.6875, "learning_rate": 2.4941657908512994e-06, "loss": 0.6306, "step": 819 }, { "epoch": 0.06987643800596506, "grad_norm": 12.375, "learning_rate": 2.494149563569331e-06, "loss": 0.4808, "step": 820 }, { "epoch": 0.06996165317426502, "grad_norm": 15.5, "learning_rate": 2.4941333138043294e-06, "loss": 0.7736, "step": 821 }, { "epoch": 0.07004686834256497, "grad_norm": 13.6875, "learning_rate": 2.4941170415565873e-06, "loss": 0.6722, "step": 822 }, { "epoch": 0.07013208351086493, "grad_norm": 11.6875, "learning_rate": 2.4941007468264e-06, "loss": 0.5138, "step": 823 }, { "epoch": 0.07021729867916489, "grad_norm": 14.25, "learning_rate": 2.4940844296140605e-06, "loss": 0.6698, "step": 824 }, { "epoch": 0.07030251384746485, "grad_norm": 15.1875, "learning_rate": 2.4940680899198644e-06, "loss": 0.8847, "step": 825 }, { "epoch": 0.0703877290157648, "grad_norm": 14.625, "learning_rate": 2.4940517277441073e-06, "loss": 0.841, "step": 826 }, { "epoch": 0.07047294418406476, "grad_norm": 12.25, "learning_rate": 2.494035343087084e-06, "loss": 0.4975, "step": 827 }, { "epoch": 0.07055815935236472, "grad_norm": 14.25, "learning_rate": 2.4940189359490915e-06, "loss": 0.6772, "step": 828 }, { "epoch": 0.07064337452066467, "grad_norm": 14.0625, "learning_rate": 2.4940025063304256e-06, "loss": 0.7616, "step": 829 }, { "epoch": 0.07072858968896463, "grad_norm": 35.25, "learning_rate": 2.4939860542313833e-06, "loss": 1.4091, "step": 830 }, { "epoch": 0.07081380485726459, "grad_norm": 11.8125, "learning_rate": 2.4939695796522624e-06, "loss": 0.5019, "step": 831 }, { "epoch": 0.07089902002556454, "grad_norm": 15.4375, "learning_rate": 2.49395308259336e-06, "loss": 0.656, "step": 832 }, { "epoch": 0.07098423519386451, "grad_norm": 18.375, "learning_rate": 2.493936563054975e-06, "loss": 0.8033, "step": 833 }, { "epoch": 0.07106945036216447, "grad_norm": 16.5, "learning_rate": 2.493920021037405e-06, "loss": 1.0798, "step": 834 }, { "epoch": 0.07115466553046443, "grad_norm": 12.375, "learning_rate": 2.4939034565409493e-06, "loss": 0.5937, "step": 835 }, { "epoch": 0.07123988069876439, "grad_norm": 11.9375, "learning_rate": 2.493886869565908e-06, "loss": 0.661, "step": 836 }, { "epoch": 0.07132509586706434, "grad_norm": 27.25, "learning_rate": 2.4938702601125792e-06, "loss": 1.5583, "step": 837 }, { "epoch": 0.0714103110353643, "grad_norm": 12.3125, "learning_rate": 2.4938536281812645e-06, "loss": 0.5201, "step": 838 }, { "epoch": 0.07149552620366426, "grad_norm": 15.4375, "learning_rate": 2.493836973772264e-06, "loss": 1.0613, "step": 839 }, { "epoch": 0.07158074137196421, "grad_norm": 12.4375, "learning_rate": 2.4938202968858784e-06, "loss": 0.6031, "step": 840 }, { "epoch": 0.07166595654026417, "grad_norm": 14.6875, "learning_rate": 2.4938035975224092e-06, "loss": 0.6495, "step": 841 }, { "epoch": 0.07175117170856413, "grad_norm": 20.125, "learning_rate": 2.4937868756821585e-06, "loss": 0.77, "step": 842 }, { "epoch": 0.07183638687686408, "grad_norm": 11.875, "learning_rate": 2.493770131365428e-06, "loss": 0.5273, "step": 843 }, { "epoch": 0.07192160204516404, "grad_norm": 27.0, "learning_rate": 2.493753364572521e-06, "loss": 1.2015, "step": 844 }, { "epoch": 0.072006817213464, "grad_norm": 13.4375, "learning_rate": 2.4937365753037394e-06, "loss": 0.7516, "step": 845 }, { "epoch": 0.07209203238176395, "grad_norm": 15.3125, "learning_rate": 2.4937197635593875e-06, "loss": 1.0393, "step": 846 }, { "epoch": 0.07217724755006391, "grad_norm": 16.625, "learning_rate": 2.4937029293397685e-06, "loss": 1.0639, "step": 847 }, { "epoch": 0.07226246271836387, "grad_norm": 14.375, "learning_rate": 2.4936860726451874e-06, "loss": 0.6474, "step": 848 }, { "epoch": 0.07234767788666382, "grad_norm": 13.3125, "learning_rate": 2.493669193475948e-06, "loss": 0.8118, "step": 849 }, { "epoch": 0.07243289305496378, "grad_norm": 13.5625, "learning_rate": 2.493652291832356e-06, "loss": 0.8965, "step": 850 }, { "epoch": 0.07251810822326374, "grad_norm": 9.8125, "learning_rate": 2.4936353677147165e-06, "loss": 0.3174, "step": 851 }, { "epoch": 0.0726033233915637, "grad_norm": 14.3125, "learning_rate": 2.493618421123335e-06, "loss": 1.0894, "step": 852 }, { "epoch": 0.07268853855986365, "grad_norm": 13.0625, "learning_rate": 2.4936014520585185e-06, "loss": 0.7397, "step": 853 }, { "epoch": 0.07277375372816361, "grad_norm": 12.9375, "learning_rate": 2.493584460520573e-06, "loss": 0.711, "step": 854 }, { "epoch": 0.07285896889646357, "grad_norm": 13.8125, "learning_rate": 2.4935674465098054e-06, "loss": 0.7559, "step": 855 }, { "epoch": 0.07294418406476352, "grad_norm": 16.5, "learning_rate": 2.493550410026524e-06, "loss": 0.669, "step": 856 }, { "epoch": 0.07302939923306348, "grad_norm": 16.5, "learning_rate": 2.493533351071036e-06, "loss": 1.0556, "step": 857 }, { "epoch": 0.07311461440136344, "grad_norm": 12.1875, "learning_rate": 2.4935162696436497e-06, "loss": 0.7872, "step": 858 }, { "epoch": 0.0731998295696634, "grad_norm": 17.625, "learning_rate": 2.4934991657446743e-06, "loss": 0.8275, "step": 859 }, { "epoch": 0.07328504473796335, "grad_norm": 11.3125, "learning_rate": 2.4934820393744187e-06, "loss": 0.6826, "step": 860 }, { "epoch": 0.07337025990626332, "grad_norm": 13.4375, "learning_rate": 2.4934648905331923e-06, "loss": 0.8516, "step": 861 }, { "epoch": 0.07345547507456328, "grad_norm": 14.0625, "learning_rate": 2.493447719221304e-06, "loss": 0.9051, "step": 862 }, { "epoch": 0.07354069024286324, "grad_norm": 19.25, "learning_rate": 2.4934305254390657e-06, "loss": 0.9234, "step": 863 }, { "epoch": 0.07362590541116319, "grad_norm": 13.8125, "learning_rate": 2.4934133091867873e-06, "loss": 0.6714, "step": 864 }, { "epoch": 0.07371112057946315, "grad_norm": 13.75, "learning_rate": 2.49339607046478e-06, "loss": 0.6635, "step": 865 }, { "epoch": 0.0737963357477631, "grad_norm": 15.3125, "learning_rate": 2.4933788092733556e-06, "loss": 0.9683, "step": 866 }, { "epoch": 0.07388155091606306, "grad_norm": 17.375, "learning_rate": 2.4933615256128256e-06, "loss": 1.2642, "step": 867 }, { "epoch": 0.07396676608436302, "grad_norm": 15.125, "learning_rate": 2.493344219483503e-06, "loss": 1.0422, "step": 868 }, { "epoch": 0.07405198125266298, "grad_norm": 17.5, "learning_rate": 2.4933268908857e-06, "loss": 0.8265, "step": 869 }, { "epoch": 0.07413719642096293, "grad_norm": 13.25, "learning_rate": 2.493309539819729e-06, "loss": 0.8263, "step": 870 }, { "epoch": 0.07422241158926289, "grad_norm": 15.75, "learning_rate": 2.493292166285905e-06, "loss": 0.7975, "step": 871 }, { "epoch": 0.07430762675756285, "grad_norm": 12.375, "learning_rate": 2.4932747702845413e-06, "loss": 0.5741, "step": 872 }, { "epoch": 0.0743928419258628, "grad_norm": 17.0, "learning_rate": 2.4932573518159522e-06, "loss": 0.9232, "step": 873 }, { "epoch": 0.07447805709416276, "grad_norm": 12.5, "learning_rate": 2.493239910880453e-06, "loss": 0.5227, "step": 874 }, { "epoch": 0.07456327226246272, "grad_norm": 14.125, "learning_rate": 2.493222447478358e-06, "loss": 0.6978, "step": 875 }, { "epoch": 0.07464848743076268, "grad_norm": 15.875, "learning_rate": 2.493204961609983e-06, "loss": 1.008, "step": 876 }, { "epoch": 0.07473370259906263, "grad_norm": 19.25, "learning_rate": 2.4931874532756447e-06, "loss": 0.8061, "step": 877 }, { "epoch": 0.07481891776736259, "grad_norm": 12.9375, "learning_rate": 2.493169922475658e-06, "loss": 0.7792, "step": 878 }, { "epoch": 0.07490413293566255, "grad_norm": 17.75, "learning_rate": 2.493152369210342e-06, "loss": 1.2115, "step": 879 }, { "epoch": 0.0749893481039625, "grad_norm": 15.625, "learning_rate": 2.493134793480012e-06, "loss": 0.8823, "step": 880 }, { "epoch": 0.07507456327226246, "grad_norm": 24.875, "learning_rate": 2.493117195284986e-06, "loss": 1.1535, "step": 881 }, { "epoch": 0.07515977844056242, "grad_norm": 22.25, "learning_rate": 2.493099574625583e-06, "loss": 1.3755, "step": 882 }, { "epoch": 0.07524499360886237, "grad_norm": 13.75, "learning_rate": 2.49308193150212e-06, "loss": 0.935, "step": 883 }, { "epoch": 0.07533020877716233, "grad_norm": 15.8125, "learning_rate": 2.493064265914916e-06, "loss": 1.0063, "step": 884 }, { "epoch": 0.07541542394546229, "grad_norm": 12.0, "learning_rate": 2.4930465778642916e-06, "loss": 0.5326, "step": 885 }, { "epoch": 0.07550063911376224, "grad_norm": 12.75, "learning_rate": 2.4930288673505652e-06, "loss": 0.6159, "step": 886 }, { "epoch": 0.0755858542820622, "grad_norm": 21.5, "learning_rate": 2.4930111343740574e-06, "loss": 0.887, "step": 887 }, { "epoch": 0.07567106945036216, "grad_norm": 13.1875, "learning_rate": 2.4929933789350884e-06, "loss": 0.9039, "step": 888 }, { "epoch": 0.07575628461866213, "grad_norm": 18.625, "learning_rate": 2.492975601033979e-06, "loss": 0.9532, "step": 889 }, { "epoch": 0.07584149978696209, "grad_norm": 17.0, "learning_rate": 2.4929578006710508e-06, "loss": 0.6734, "step": 890 }, { "epoch": 0.07592671495526204, "grad_norm": 18.25, "learning_rate": 2.492939977846625e-06, "loss": 0.8263, "step": 891 }, { "epoch": 0.076011930123562, "grad_norm": 21.875, "learning_rate": 2.4929221325610243e-06, "loss": 1.0999, "step": 892 }, { "epoch": 0.07609714529186196, "grad_norm": 17.0, "learning_rate": 2.4929042648145705e-06, "loss": 0.7115, "step": 893 }, { "epoch": 0.07618236046016191, "grad_norm": 14.3125, "learning_rate": 2.4928863746075867e-06, "loss": 0.6361, "step": 894 }, { "epoch": 0.07626757562846187, "grad_norm": 13.75, "learning_rate": 2.4928684619403964e-06, "loss": 1.11, "step": 895 }, { "epoch": 0.07635279079676183, "grad_norm": 17.125, "learning_rate": 2.4928505268133234e-06, "loss": 0.8011, "step": 896 }, { "epoch": 0.07643800596506178, "grad_norm": 12.9375, "learning_rate": 2.4928325692266914e-06, "loss": 0.3879, "step": 897 }, { "epoch": 0.07652322113336174, "grad_norm": 18.75, "learning_rate": 2.492814589180825e-06, "loss": 1.1808, "step": 898 }, { "epoch": 0.0766084363016617, "grad_norm": 15.875, "learning_rate": 2.4927965866760494e-06, "loss": 1.0175, "step": 899 }, { "epoch": 0.07669365146996165, "grad_norm": 18.25, "learning_rate": 2.49277856171269e-06, "loss": 1.067, "step": 900 }, { "epoch": 0.07677886663826161, "grad_norm": 13.4375, "learning_rate": 2.4927605142910716e-06, "loss": 0.6143, "step": 901 }, { "epoch": 0.07686408180656157, "grad_norm": 14.25, "learning_rate": 2.4927424444115217e-06, "loss": 0.7555, "step": 902 }, { "epoch": 0.07694929697486153, "grad_norm": 18.375, "learning_rate": 2.492724352074366e-06, "loss": 0.511, "step": 903 }, { "epoch": 0.07703451214316148, "grad_norm": 12.25, "learning_rate": 2.492706237279931e-06, "loss": 0.3614, "step": 904 }, { "epoch": 0.07711972731146144, "grad_norm": 16.875, "learning_rate": 2.4926881000285453e-06, "loss": 0.9427, "step": 905 }, { "epoch": 0.0772049424797614, "grad_norm": 13.625, "learning_rate": 2.4926699403205356e-06, "loss": 0.8772, "step": 906 }, { "epoch": 0.07729015764806135, "grad_norm": 13.5, "learning_rate": 2.492651758156231e-06, "loss": 0.6097, "step": 907 }, { "epoch": 0.07737537281636131, "grad_norm": 15.3125, "learning_rate": 2.492633553535959e-06, "loss": 1.0343, "step": 908 }, { "epoch": 0.07746058798466127, "grad_norm": 25.375, "learning_rate": 2.4926153264600494e-06, "loss": 0.4935, "step": 909 }, { "epoch": 0.07754580315296122, "grad_norm": 12.5625, "learning_rate": 2.4925970769288314e-06, "loss": 0.5237, "step": 910 }, { "epoch": 0.07763101832126118, "grad_norm": 15.6875, "learning_rate": 2.492578804942635e-06, "loss": 0.8587, "step": 911 }, { "epoch": 0.07771623348956114, "grad_norm": 16.0, "learning_rate": 2.4925605105017893e-06, "loss": 0.9437, "step": 912 }, { "epoch": 0.0778014486578611, "grad_norm": 15.625, "learning_rate": 2.492542193606626e-06, "loss": 0.8953, "step": 913 }, { "epoch": 0.07788666382616105, "grad_norm": 20.875, "learning_rate": 2.4925238542574758e-06, "loss": 0.9435, "step": 914 }, { "epoch": 0.07797187899446101, "grad_norm": 17.0, "learning_rate": 2.4925054924546702e-06, "loss": 1.2316, "step": 915 }, { "epoch": 0.07805709416276096, "grad_norm": 23.875, "learning_rate": 2.492487108198541e-06, "loss": 1.1427, "step": 916 }, { "epoch": 0.07814230933106094, "grad_norm": 12.125, "learning_rate": 2.49246870148942e-06, "loss": 0.6263, "step": 917 }, { "epoch": 0.07822752449936089, "grad_norm": 10.5, "learning_rate": 2.4924502723276403e-06, "loss": 0.4651, "step": 918 }, { "epoch": 0.07831273966766085, "grad_norm": 13.25, "learning_rate": 2.492431820713535e-06, "loss": 0.3829, "step": 919 }, { "epoch": 0.0783979548359608, "grad_norm": 11.5, "learning_rate": 2.4924133466474372e-06, "loss": 0.4553, "step": 920 }, { "epoch": 0.07848317000426076, "grad_norm": 17.125, "learning_rate": 2.4923948501296805e-06, "loss": 0.7863, "step": 921 }, { "epoch": 0.07856838517256072, "grad_norm": 16.25, "learning_rate": 2.4923763311606003e-06, "loss": 0.8902, "step": 922 }, { "epoch": 0.07865360034086068, "grad_norm": 12.875, "learning_rate": 2.4923577897405297e-06, "loss": 0.8082, "step": 923 }, { "epoch": 0.07873881550916063, "grad_norm": 16.875, "learning_rate": 2.492339225869805e-06, "loss": 0.9792, "step": 924 }, { "epoch": 0.07882403067746059, "grad_norm": 16.125, "learning_rate": 2.4923206395487616e-06, "loss": 1.0222, "step": 925 }, { "epoch": 0.07890924584576055, "grad_norm": 14.8125, "learning_rate": 2.4923020307777344e-06, "loss": 0.6075, "step": 926 }, { "epoch": 0.0789944610140605, "grad_norm": 12.625, "learning_rate": 2.4922833995570604e-06, "loss": 0.7074, "step": 927 }, { "epoch": 0.07907967618236046, "grad_norm": 14.5625, "learning_rate": 2.492264745887076e-06, "loss": 0.783, "step": 928 }, { "epoch": 0.07916489135066042, "grad_norm": 16.875, "learning_rate": 2.492246069768119e-06, "loss": 0.7242, "step": 929 }, { "epoch": 0.07925010651896038, "grad_norm": 11.625, "learning_rate": 2.4922273712005256e-06, "loss": 0.4453, "step": 930 }, { "epoch": 0.07933532168726033, "grad_norm": 22.625, "learning_rate": 2.492208650184635e-06, "loss": 1.0361, "step": 931 }, { "epoch": 0.07942053685556029, "grad_norm": 14.875, "learning_rate": 2.492189906720785e-06, "loss": 0.9549, "step": 932 }, { "epoch": 0.07950575202386025, "grad_norm": 10.625, "learning_rate": 2.4921711408093137e-06, "loss": 0.4851, "step": 933 }, { "epoch": 0.0795909671921602, "grad_norm": 10.875, "learning_rate": 2.492152352450561e-06, "loss": 0.3842, "step": 934 }, { "epoch": 0.07967618236046016, "grad_norm": 18.5, "learning_rate": 2.4921335416448668e-06, "loss": 1.4054, "step": 935 }, { "epoch": 0.07976139752876012, "grad_norm": 13.5625, "learning_rate": 2.4921147083925695e-06, "loss": 0.7233, "step": 936 }, { "epoch": 0.07984661269706007, "grad_norm": 11.375, "learning_rate": 2.492095852694011e-06, "loss": 0.7253, "step": 937 }, { "epoch": 0.07993182786536003, "grad_norm": 18.875, "learning_rate": 2.492076974549531e-06, "loss": 0.893, "step": 938 }, { "epoch": 0.08001704303365999, "grad_norm": 18.75, "learning_rate": 2.492058073959472e-06, "loss": 0.863, "step": 939 }, { "epoch": 0.08010225820195994, "grad_norm": 21.25, "learning_rate": 2.492039150924174e-06, "loss": 1.0865, "step": 940 }, { "epoch": 0.0801874733702599, "grad_norm": 15.8125, "learning_rate": 2.492020205443979e-06, "loss": 1.0112, "step": 941 }, { "epoch": 0.08027268853855986, "grad_norm": 11.3125, "learning_rate": 2.4920012375192305e-06, "loss": 0.53, "step": 942 }, { "epoch": 0.08035790370685982, "grad_norm": 11.9375, "learning_rate": 2.491982247150271e-06, "loss": 0.679, "step": 943 }, { "epoch": 0.08044311887515977, "grad_norm": 16.375, "learning_rate": 2.491963234337443e-06, "loss": 0.9403, "step": 944 }, { "epoch": 0.08052833404345973, "grad_norm": 14.1875, "learning_rate": 2.4919441990810904e-06, "loss": 0.8987, "step": 945 }, { "epoch": 0.0806135492117597, "grad_norm": 14.8125, "learning_rate": 2.4919251413815577e-06, "loss": 0.9483, "step": 946 }, { "epoch": 0.08069876438005966, "grad_norm": 20.75, "learning_rate": 2.4919060612391882e-06, "loss": 1.0219, "step": 947 }, { "epoch": 0.08078397954835961, "grad_norm": 14.0625, "learning_rate": 2.4918869586543276e-06, "loss": 0.8508, "step": 948 }, { "epoch": 0.08086919471665957, "grad_norm": 13.5625, "learning_rate": 2.491867833627321e-06, "loss": 0.8654, "step": 949 }, { "epoch": 0.08095440988495953, "grad_norm": 11.375, "learning_rate": 2.491848686158514e-06, "loss": 0.5309, "step": 950 }, { "epoch": 0.08103962505325948, "grad_norm": 17.875, "learning_rate": 2.491829516248252e-06, "loss": 0.9581, "step": 951 }, { "epoch": 0.08112484022155944, "grad_norm": 15.3125, "learning_rate": 2.491810323896882e-06, "loss": 0.9936, "step": 952 }, { "epoch": 0.0812100553898594, "grad_norm": 16.5, "learning_rate": 2.4917911091047507e-06, "loss": 0.9531, "step": 953 }, { "epoch": 0.08129527055815935, "grad_norm": 9.875, "learning_rate": 2.4917718718722056e-06, "loss": 0.3583, "step": 954 }, { "epoch": 0.08138048572645931, "grad_norm": 11.75, "learning_rate": 2.491752612199594e-06, "loss": 0.6067, "step": 955 }, { "epoch": 0.08146570089475927, "grad_norm": 15.1875, "learning_rate": 2.491733330087264e-06, "loss": 0.9601, "step": 956 }, { "epoch": 0.08155091606305923, "grad_norm": 13.8125, "learning_rate": 2.491714025535564e-06, "loss": 0.8267, "step": 957 }, { "epoch": 0.08163613123135918, "grad_norm": 11.0, "learning_rate": 2.491694698544843e-06, "loss": 0.5244, "step": 958 }, { "epoch": 0.08172134639965914, "grad_norm": 12.375, "learning_rate": 2.49167534911545e-06, "loss": 0.5638, "step": 959 }, { "epoch": 0.0818065615679591, "grad_norm": 16.625, "learning_rate": 2.491655977247735e-06, "loss": 0.8481, "step": 960 }, { "epoch": 0.08189177673625905, "grad_norm": 21.125, "learning_rate": 2.491636582942048e-06, "loss": 0.7118, "step": 961 }, { "epoch": 0.08197699190455901, "grad_norm": 24.625, "learning_rate": 2.491617166198739e-06, "loss": 1.2096, "step": 962 }, { "epoch": 0.08206220707285897, "grad_norm": 19.25, "learning_rate": 2.4915977270181594e-06, "loss": 1.4455, "step": 963 }, { "epoch": 0.08214742224115892, "grad_norm": 19.75, "learning_rate": 2.491578265400661e-06, "loss": 1.0058, "step": 964 }, { "epoch": 0.08223263740945888, "grad_norm": 25.5, "learning_rate": 2.4915587813465943e-06, "loss": 0.9515, "step": 965 }, { "epoch": 0.08231785257775884, "grad_norm": 14.125, "learning_rate": 2.4915392748563115e-06, "loss": 0.8341, "step": 966 }, { "epoch": 0.0824030677460588, "grad_norm": 18.625, "learning_rate": 2.4915197459301664e-06, "loss": 0.7962, "step": 967 }, { "epoch": 0.08248828291435875, "grad_norm": 14.0625, "learning_rate": 2.4915001945685106e-06, "loss": 0.7048, "step": 968 }, { "epoch": 0.08257349808265871, "grad_norm": 14.9375, "learning_rate": 2.491480620771698e-06, "loss": 0.5696, "step": 969 }, { "epoch": 0.08265871325095867, "grad_norm": 14.25, "learning_rate": 2.4914610245400816e-06, "loss": 0.7926, "step": 970 }, { "epoch": 0.08274392841925862, "grad_norm": 10.8125, "learning_rate": 2.4914414058740166e-06, "loss": 0.5785, "step": 971 }, { "epoch": 0.08282914358755858, "grad_norm": 20.0, "learning_rate": 2.491421764773857e-06, "loss": 0.9236, "step": 972 }, { "epoch": 0.08291435875585854, "grad_norm": 16.625, "learning_rate": 2.491402101239958e-06, "loss": 0.998, "step": 973 }, { "epoch": 0.0829995739241585, "grad_norm": 11.0625, "learning_rate": 2.4913824152726746e-06, "loss": 0.3975, "step": 974 }, { "epoch": 0.08308478909245846, "grad_norm": 19.75, "learning_rate": 2.491362706872362e-06, "loss": 0.9719, "step": 975 }, { "epoch": 0.08317000426075842, "grad_norm": 13.5625, "learning_rate": 2.491342976039378e-06, "loss": 0.7411, "step": 976 }, { "epoch": 0.08325521942905838, "grad_norm": 15.4375, "learning_rate": 2.4913232227740775e-06, "loss": 0.7216, "step": 977 }, { "epoch": 0.08334043459735833, "grad_norm": 18.875, "learning_rate": 2.4913034470768184e-06, "loss": 1.3156, "step": 978 }, { "epoch": 0.08342564976565829, "grad_norm": 11.875, "learning_rate": 2.4912836489479575e-06, "loss": 0.4551, "step": 979 }, { "epoch": 0.08351086493395825, "grad_norm": 15.75, "learning_rate": 2.4912638283878533e-06, "loss": 1.2054, "step": 980 }, { "epoch": 0.0835960801022582, "grad_norm": 16.625, "learning_rate": 2.491243985396863e-06, "loss": 1.133, "step": 981 }, { "epoch": 0.08368129527055816, "grad_norm": 12.25, "learning_rate": 2.491224119975346e-06, "loss": 0.8804, "step": 982 }, { "epoch": 0.08376651043885812, "grad_norm": 17.75, "learning_rate": 2.4912042321236614e-06, "loss": 0.7901, "step": 983 }, { "epoch": 0.08385172560715808, "grad_norm": 13.125, "learning_rate": 2.4911843218421673e-06, "loss": 0.7225, "step": 984 }, { "epoch": 0.08393694077545803, "grad_norm": 12.1875, "learning_rate": 2.4911643891312247e-06, "loss": 0.3106, "step": 985 }, { "epoch": 0.08402215594375799, "grad_norm": 11.8125, "learning_rate": 2.491144433991194e-06, "loss": 1.1128, "step": 986 }, { "epoch": 0.08410737111205795, "grad_norm": 11.9375, "learning_rate": 2.491124456422435e-06, "loss": 0.7801, "step": 987 }, { "epoch": 0.0841925862803579, "grad_norm": 14.25, "learning_rate": 2.491104456425309e-06, "loss": 0.9965, "step": 988 }, { "epoch": 0.08427780144865786, "grad_norm": 11.0625, "learning_rate": 2.4910844340001773e-06, "loss": 0.5957, "step": 989 }, { "epoch": 0.08436301661695782, "grad_norm": 15.875, "learning_rate": 2.4910643891474024e-06, "loss": 0.7323, "step": 990 }, { "epoch": 0.08444823178525777, "grad_norm": 16.125, "learning_rate": 2.4910443218673453e-06, "loss": 0.9596, "step": 991 }, { "epoch": 0.08453344695355773, "grad_norm": 16.25, "learning_rate": 2.49102423216037e-06, "loss": 0.764, "step": 992 }, { "epoch": 0.08461866212185769, "grad_norm": 14.375, "learning_rate": 2.4910041200268383e-06, "loss": 0.7823, "step": 993 }, { "epoch": 0.08470387729015764, "grad_norm": 14.8125, "learning_rate": 2.4909839854671147e-06, "loss": 0.7277, "step": 994 }, { "epoch": 0.0847890924584576, "grad_norm": 11.5, "learning_rate": 2.4909638284815624e-06, "loss": 0.5851, "step": 995 }, { "epoch": 0.08487430762675756, "grad_norm": 19.875, "learning_rate": 2.4909436490705456e-06, "loss": 1.0849, "step": 996 }, { "epoch": 0.08495952279505752, "grad_norm": 20.125, "learning_rate": 2.4909234472344297e-06, "loss": 1.1363, "step": 997 }, { "epoch": 0.08504473796335747, "grad_norm": 30.375, "learning_rate": 2.4909032229735786e-06, "loss": 1.4282, "step": 998 }, { "epoch": 0.08512995313165743, "grad_norm": 11.1875, "learning_rate": 2.4908829762883586e-06, "loss": 0.7139, "step": 999 }, { "epoch": 0.08521516829995739, "grad_norm": 13.0625, "learning_rate": 2.4908627071791357e-06, "loss": 0.5121, "step": 1000 }, { "epoch": 0.08530038346825734, "grad_norm": 15.8125, "learning_rate": 2.4908424156462757e-06, "loss": 0.9741, "step": 1001 }, { "epoch": 0.08538559863655731, "grad_norm": 21.25, "learning_rate": 2.490822101690146e-06, "loss": 1.3311, "step": 1002 }, { "epoch": 0.08547081380485727, "grad_norm": 15.5, "learning_rate": 2.4908017653111125e-06, "loss": 0.6252, "step": 1003 }, { "epoch": 0.08555602897315723, "grad_norm": 15.875, "learning_rate": 2.4907814065095437e-06, "loss": 1.0719, "step": 1004 }, { "epoch": 0.08564124414145718, "grad_norm": 13.1875, "learning_rate": 2.490761025285807e-06, "loss": 0.6015, "step": 1005 }, { "epoch": 0.08572645930975714, "grad_norm": 15.3125, "learning_rate": 2.4907406216402714e-06, "loss": 0.9579, "step": 1006 }, { "epoch": 0.0858116744780571, "grad_norm": 15.0625, "learning_rate": 2.4907201955733047e-06, "loss": 0.9232, "step": 1007 }, { "epoch": 0.08589688964635706, "grad_norm": 12.6875, "learning_rate": 2.490699747085276e-06, "loss": 0.7117, "step": 1008 }, { "epoch": 0.08598210481465701, "grad_norm": 17.75, "learning_rate": 2.490679276176556e-06, "loss": 0.7037, "step": 1009 }, { "epoch": 0.08606731998295697, "grad_norm": 17.75, "learning_rate": 2.4906587828475136e-06, "loss": 0.9069, "step": 1010 }, { "epoch": 0.08615253515125693, "grad_norm": 15.625, "learning_rate": 2.490638267098519e-06, "loss": 0.8875, "step": 1011 }, { "epoch": 0.08623775031955688, "grad_norm": 15.375, "learning_rate": 2.4906177289299438e-06, "loss": 0.9406, "step": 1012 }, { "epoch": 0.08632296548785684, "grad_norm": 11.625, "learning_rate": 2.4905971683421588e-06, "loss": 0.6685, "step": 1013 }, { "epoch": 0.0864081806561568, "grad_norm": 14.375, "learning_rate": 2.490576585335535e-06, "loss": 0.8152, "step": 1014 }, { "epoch": 0.08649339582445675, "grad_norm": 12.0625, "learning_rate": 2.4905559799104457e-06, "loss": 0.5096, "step": 1015 }, { "epoch": 0.08657861099275671, "grad_norm": 12.5625, "learning_rate": 2.4905353520672617e-06, "loss": 0.8894, "step": 1016 }, { "epoch": 0.08666382616105667, "grad_norm": 11.375, "learning_rate": 2.4905147018063563e-06, "loss": 0.5815, "step": 1017 }, { "epoch": 0.08674904132935662, "grad_norm": 22.875, "learning_rate": 2.490494029128103e-06, "loss": 1.2529, "step": 1018 }, { "epoch": 0.08683425649765658, "grad_norm": 14.3125, "learning_rate": 2.490473334032875e-06, "loss": 0.8467, "step": 1019 }, { "epoch": 0.08691947166595654, "grad_norm": 12.125, "learning_rate": 2.490452616521047e-06, "loss": 0.7415, "step": 1020 }, { "epoch": 0.0870046868342565, "grad_norm": 20.5, "learning_rate": 2.4904318765929924e-06, "loss": 1.1804, "step": 1021 }, { "epoch": 0.08708990200255645, "grad_norm": 12.875, "learning_rate": 2.4904111142490868e-06, "loss": 0.4877, "step": 1022 }, { "epoch": 0.08717511717085641, "grad_norm": 19.75, "learning_rate": 2.4903903294897043e-06, "loss": 1.1736, "step": 1023 }, { "epoch": 0.08726033233915637, "grad_norm": 17.125, "learning_rate": 2.4903695223152217e-06, "loss": 1.029, "step": 1024 }, { "epoch": 0.08734554750745632, "grad_norm": 13.125, "learning_rate": 2.4903486927260145e-06, "loss": 0.6916, "step": 1025 }, { "epoch": 0.08743076267575628, "grad_norm": 17.875, "learning_rate": 2.4903278407224597e-06, "loss": 0.8215, "step": 1026 }, { "epoch": 0.08751597784405624, "grad_norm": 10.625, "learning_rate": 2.490306966304933e-06, "loss": 0.7558, "step": 1027 }, { "epoch": 0.0876011930123562, "grad_norm": 22.0, "learning_rate": 2.490286069473812e-06, "loss": 1.0058, "step": 1028 }, { "epoch": 0.08768640818065615, "grad_norm": 10.625, "learning_rate": 2.490265150229475e-06, "loss": 0.4715, "step": 1029 }, { "epoch": 0.08777162334895612, "grad_norm": 10.875, "learning_rate": 2.4902442085722992e-06, "loss": 0.4575, "step": 1030 }, { "epoch": 0.08785683851725608, "grad_norm": 13.25, "learning_rate": 2.4902232445026636e-06, "loss": 0.5331, "step": 1031 }, { "epoch": 0.08794205368555603, "grad_norm": 15.5625, "learning_rate": 2.4902022580209465e-06, "loss": 0.9019, "step": 1032 }, { "epoch": 0.08802726885385599, "grad_norm": 14.625, "learning_rate": 2.490181249127528e-06, "loss": 0.768, "step": 1033 }, { "epoch": 0.08811248402215595, "grad_norm": 14.5, "learning_rate": 2.490160217822787e-06, "loss": 0.8467, "step": 1034 }, { "epoch": 0.0881976991904559, "grad_norm": 12.125, "learning_rate": 2.490139164107104e-06, "loss": 0.5999, "step": 1035 }, { "epoch": 0.08828291435875586, "grad_norm": 14.0625, "learning_rate": 2.490118087980859e-06, "loss": 1.0654, "step": 1036 }, { "epoch": 0.08836812952705582, "grad_norm": 14.3125, "learning_rate": 2.4900969894444328e-06, "loss": 0.7843, "step": 1037 }, { "epoch": 0.08845334469535578, "grad_norm": 14.25, "learning_rate": 2.4900758684982072e-06, "loss": 0.6245, "step": 1038 }, { "epoch": 0.08853855986365573, "grad_norm": 16.125, "learning_rate": 2.4900547251425638e-06, "loss": 0.3501, "step": 1039 }, { "epoch": 0.08862377503195569, "grad_norm": 11.5625, "learning_rate": 2.4900335593778844e-06, "loss": 0.9506, "step": 1040 }, { "epoch": 0.08870899020025565, "grad_norm": 13.625, "learning_rate": 2.490012371204552e-06, "loss": 1.1008, "step": 1041 }, { "epoch": 0.0887942053685556, "grad_norm": 17.5, "learning_rate": 2.4899911606229487e-06, "loss": 0.8739, "step": 1042 }, { "epoch": 0.08887942053685556, "grad_norm": 15.0625, "learning_rate": 2.489969927633459e-06, "loss": 0.4483, "step": 1043 }, { "epoch": 0.08896463570515552, "grad_norm": 27.125, "learning_rate": 2.4899486722364645e-06, "loss": 1.1257, "step": 1044 }, { "epoch": 0.08904985087345547, "grad_norm": 17.75, "learning_rate": 2.4899273944323516e-06, "loss": 1.1577, "step": 1045 }, { "epoch": 0.08913506604175543, "grad_norm": 14.3125, "learning_rate": 2.4899060942215037e-06, "loss": 0.8145, "step": 1046 }, { "epoch": 0.08922028121005539, "grad_norm": 13.4375, "learning_rate": 2.4898847716043055e-06, "loss": 0.5171, "step": 1047 }, { "epoch": 0.08930549637835535, "grad_norm": 21.0, "learning_rate": 2.489863426581143e-06, "loss": 1.287, "step": 1048 }, { "epoch": 0.0893907115466553, "grad_norm": 18.125, "learning_rate": 2.4898420591524016e-06, "loss": 0.7502, "step": 1049 }, { "epoch": 0.08947592671495526, "grad_norm": 24.625, "learning_rate": 2.4898206693184672e-06, "loss": 1.3815, "step": 1050 }, { "epoch": 0.08956114188325522, "grad_norm": 13.875, "learning_rate": 2.489799257079727e-06, "loss": 0.6419, "step": 1051 }, { "epoch": 0.08964635705155517, "grad_norm": 19.25, "learning_rate": 2.4897778224365672e-06, "loss": 1.0908, "step": 1052 }, { "epoch": 0.08973157221985513, "grad_norm": 24.5, "learning_rate": 2.4897563653893754e-06, "loss": 1.2304, "step": 1053 }, { "epoch": 0.08981678738815509, "grad_norm": 19.5, "learning_rate": 2.4897348859385395e-06, "loss": 1.2056, "step": 1054 }, { "epoch": 0.08990200255645504, "grad_norm": 14.6875, "learning_rate": 2.4897133840844474e-06, "loss": 0.6569, "step": 1055 }, { "epoch": 0.089987217724755, "grad_norm": 13.875, "learning_rate": 2.4896918598274884e-06, "loss": 0.5882, "step": 1056 }, { "epoch": 0.09007243289305496, "grad_norm": 16.75, "learning_rate": 2.4896703131680503e-06, "loss": 1.1163, "step": 1057 }, { "epoch": 0.09015764806135493, "grad_norm": 15.375, "learning_rate": 2.489648744106523e-06, "loss": 0.7096, "step": 1058 }, { "epoch": 0.09024286322965488, "grad_norm": 14.625, "learning_rate": 2.4896271526432966e-06, "loss": 0.7612, "step": 1059 }, { "epoch": 0.09032807839795484, "grad_norm": 13.0625, "learning_rate": 2.489605538778761e-06, "loss": 0.6753, "step": 1060 }, { "epoch": 0.0904132935662548, "grad_norm": 12.875, "learning_rate": 2.4895839025133067e-06, "loss": 0.5948, "step": 1061 }, { "epoch": 0.09049850873455476, "grad_norm": 14.75, "learning_rate": 2.489562243847325e-06, "loss": 0.6922, "step": 1062 }, { "epoch": 0.09058372390285471, "grad_norm": 18.375, "learning_rate": 2.489540562781207e-06, "loss": 0.8443, "step": 1063 }, { "epoch": 0.09066893907115467, "grad_norm": 12.375, "learning_rate": 2.4895188593153444e-06, "loss": 0.6662, "step": 1064 }, { "epoch": 0.09075415423945463, "grad_norm": 16.375, "learning_rate": 2.4894971334501295e-06, "loss": 1.2845, "step": 1065 }, { "epoch": 0.09083936940775458, "grad_norm": 14.8125, "learning_rate": 2.4894753851859555e-06, "loss": 1.016, "step": 1066 }, { "epoch": 0.09092458457605454, "grad_norm": 15.0, "learning_rate": 2.489453614523214e-06, "loss": 0.7773, "step": 1067 }, { "epoch": 0.0910097997443545, "grad_norm": 13.8125, "learning_rate": 2.4894318214623004e-06, "loss": 0.7754, "step": 1068 }, { "epoch": 0.09109501491265445, "grad_norm": 20.5, "learning_rate": 2.4894100060036066e-06, "loss": 1.1329, "step": 1069 }, { "epoch": 0.09118023008095441, "grad_norm": 11.75, "learning_rate": 2.489388168147528e-06, "loss": 0.4827, "step": 1070 }, { "epoch": 0.09126544524925437, "grad_norm": 18.0, "learning_rate": 2.4893663078944593e-06, "loss": 0.8682, "step": 1071 }, { "epoch": 0.09135066041755432, "grad_norm": 19.75, "learning_rate": 2.4893444252447948e-06, "loss": 0.9551, "step": 1072 }, { "epoch": 0.09143587558585428, "grad_norm": 16.75, "learning_rate": 2.4893225201989295e-06, "loss": 0.7951, "step": 1073 }, { "epoch": 0.09152109075415424, "grad_norm": 14.8125, "learning_rate": 2.489300592757261e-06, "loss": 0.995, "step": 1074 }, { "epoch": 0.0916063059224542, "grad_norm": 20.875, "learning_rate": 2.4892786429201843e-06, "loss": 1.0128, "step": 1075 }, { "epoch": 0.09169152109075415, "grad_norm": 17.25, "learning_rate": 2.4892566706880963e-06, "loss": 0.8333, "step": 1076 }, { "epoch": 0.09177673625905411, "grad_norm": 13.4375, "learning_rate": 2.489234676061394e-06, "loss": 0.9616, "step": 1077 }, { "epoch": 0.09186195142735407, "grad_norm": 21.0, "learning_rate": 2.489212659040475e-06, "loss": 1.1042, "step": 1078 }, { "epoch": 0.09194716659565402, "grad_norm": 14.875, "learning_rate": 2.4891906196257377e-06, "loss": 0.821, "step": 1079 }, { "epoch": 0.09203238176395398, "grad_norm": 21.625, "learning_rate": 2.489168557817579e-06, "loss": 1.1489, "step": 1080 }, { "epoch": 0.09211759693225394, "grad_norm": 11.6875, "learning_rate": 2.489146473616399e-06, "loss": 0.5219, "step": 1081 }, { "epoch": 0.0922028121005539, "grad_norm": 16.0, "learning_rate": 2.4891243670225955e-06, "loss": 0.7303, "step": 1082 }, { "epoch": 0.09228802726885385, "grad_norm": 19.125, "learning_rate": 2.4891022380365693e-06, "loss": 0.7661, "step": 1083 }, { "epoch": 0.09237324243715381, "grad_norm": 17.625, "learning_rate": 2.4890800866587193e-06, "loss": 1.1133, "step": 1084 }, { "epoch": 0.09245845760545376, "grad_norm": 17.875, "learning_rate": 2.489057912889446e-06, "loss": 0.7884, "step": 1085 }, { "epoch": 0.09254367277375372, "grad_norm": 12.4375, "learning_rate": 2.4890357167291507e-06, "loss": 0.6475, "step": 1086 }, { "epoch": 0.09262888794205369, "grad_norm": 18.375, "learning_rate": 2.489013498178234e-06, "loss": 0.9722, "step": 1087 }, { "epoch": 0.09271410311035365, "grad_norm": 17.375, "learning_rate": 2.488991257237097e-06, "loss": 0.8523, "step": 1088 }, { "epoch": 0.0927993182786536, "grad_norm": 18.875, "learning_rate": 2.488968993906142e-06, "loss": 0.7299, "step": 1089 }, { "epoch": 0.09288453344695356, "grad_norm": 18.375, "learning_rate": 2.488946708185772e-06, "loss": 1.2717, "step": 1090 }, { "epoch": 0.09296974861525352, "grad_norm": 17.375, "learning_rate": 2.488924400076389e-06, "loss": 0.8368, "step": 1091 }, { "epoch": 0.09305496378355348, "grad_norm": 12.5625, "learning_rate": 2.4889020695783962e-06, "loss": 0.7039, "step": 1092 }, { "epoch": 0.09314017895185343, "grad_norm": 18.625, "learning_rate": 2.4888797166921967e-06, "loss": 1.0184, "step": 1093 }, { "epoch": 0.09322539412015339, "grad_norm": 15.125, "learning_rate": 2.4888573414181956e-06, "loss": 0.8016, "step": 1094 }, { "epoch": 0.09331060928845335, "grad_norm": 13.75, "learning_rate": 2.488834943756796e-06, "loss": 0.608, "step": 1095 }, { "epoch": 0.0933958244567533, "grad_norm": 19.875, "learning_rate": 2.4888125237084037e-06, "loss": 1.0198, "step": 1096 }, { "epoch": 0.09348103962505326, "grad_norm": 15.875, "learning_rate": 2.488790081273423e-06, "loss": 1.0101, "step": 1097 }, { "epoch": 0.09356625479335322, "grad_norm": 15.6875, "learning_rate": 2.48876761645226e-06, "loss": 0.9163, "step": 1098 }, { "epoch": 0.09365146996165317, "grad_norm": 15.6875, "learning_rate": 2.4887451292453208e-06, "loss": 1.0795, "step": 1099 }, { "epoch": 0.09373668512995313, "grad_norm": 19.125, "learning_rate": 2.488722619653011e-06, "loss": 0.7447, "step": 1100 }, { "epoch": 0.09382190029825309, "grad_norm": 10.25, "learning_rate": 2.4887000876757383e-06, "loss": 0.4926, "step": 1101 }, { "epoch": 0.09390711546655305, "grad_norm": 18.125, "learning_rate": 2.488677533313909e-06, "loss": 1.1527, "step": 1102 }, { "epoch": 0.093992330634853, "grad_norm": 16.375, "learning_rate": 2.4886549565679314e-06, "loss": 1.2142, "step": 1103 }, { "epoch": 0.09407754580315296, "grad_norm": 16.125, "learning_rate": 2.488632357438213e-06, "loss": 1.0449, "step": 1104 }, { "epoch": 0.09416276097145292, "grad_norm": 26.625, "learning_rate": 2.488609735925162e-06, "loss": 1.2415, "step": 1105 }, { "epoch": 0.09424797613975287, "grad_norm": 11.9375, "learning_rate": 2.4885870920291883e-06, "loss": 0.5278, "step": 1106 }, { "epoch": 0.09433319130805283, "grad_norm": 13.75, "learning_rate": 2.4885644257506997e-06, "loss": 0.5945, "step": 1107 }, { "epoch": 0.09441840647635279, "grad_norm": 18.25, "learning_rate": 2.4885417370901066e-06, "loss": 1.0857, "step": 1108 }, { "epoch": 0.09450362164465274, "grad_norm": 18.125, "learning_rate": 2.4885190260478194e-06, "loss": 1.0826, "step": 1109 }, { "epoch": 0.0945888368129527, "grad_norm": 12.0625, "learning_rate": 2.4884962926242473e-06, "loss": 0.5277, "step": 1110 }, { "epoch": 0.09467405198125266, "grad_norm": 13.4375, "learning_rate": 2.4884735368198022e-06, "loss": 0.8315, "step": 1111 }, { "epoch": 0.09475926714955261, "grad_norm": 12.8125, "learning_rate": 2.4884507586348943e-06, "loss": 0.7635, "step": 1112 }, { "epoch": 0.09484448231785257, "grad_norm": 17.75, "learning_rate": 2.4884279580699367e-06, "loss": 0.8582, "step": 1113 }, { "epoch": 0.09492969748615253, "grad_norm": 21.0, "learning_rate": 2.48840513512534e-06, "loss": 1.1656, "step": 1114 }, { "epoch": 0.0950149126544525, "grad_norm": 13.1875, "learning_rate": 2.4883822898015174e-06, "loss": 0.7417, "step": 1115 }, { "epoch": 0.09510012782275246, "grad_norm": 18.375, "learning_rate": 2.4883594220988815e-06, "loss": 1.1427, "step": 1116 }, { "epoch": 0.09518534299105241, "grad_norm": 19.5, "learning_rate": 2.488336532017845e-06, "loss": 0.9757, "step": 1117 }, { "epoch": 0.09527055815935237, "grad_norm": 21.75, "learning_rate": 2.488313619558823e-06, "loss": 1.0315, "step": 1118 }, { "epoch": 0.09535577332765233, "grad_norm": 12.8125, "learning_rate": 2.488290684722228e-06, "loss": 0.6196, "step": 1119 }, { "epoch": 0.09544098849595228, "grad_norm": 17.0, "learning_rate": 2.488267727508476e-06, "loss": 1.1974, "step": 1120 }, { "epoch": 0.09552620366425224, "grad_norm": 12.875, "learning_rate": 2.4882447479179804e-06, "loss": 0.638, "step": 1121 }, { "epoch": 0.0956114188325522, "grad_norm": 11.375, "learning_rate": 2.488221745951157e-06, "loss": 0.6605, "step": 1122 }, { "epoch": 0.09569663400085215, "grad_norm": 14.0625, "learning_rate": 2.4881987216084217e-06, "loss": 0.5806, "step": 1123 }, { "epoch": 0.09578184916915211, "grad_norm": 19.625, "learning_rate": 2.4881756748901903e-06, "loss": 1.0507, "step": 1124 }, { "epoch": 0.09586706433745207, "grad_norm": 13.6875, "learning_rate": 2.488152605796879e-06, "loss": 1.028, "step": 1125 }, { "epoch": 0.09595227950575203, "grad_norm": 14.1875, "learning_rate": 2.488129514328906e-06, "loss": 0.5205, "step": 1126 }, { "epoch": 0.09603749467405198, "grad_norm": 16.0, "learning_rate": 2.488106400486687e-06, "loss": 0.8333, "step": 1127 }, { "epoch": 0.09612270984235194, "grad_norm": 16.25, "learning_rate": 2.4880832642706404e-06, "loss": 0.7568, "step": 1128 }, { "epoch": 0.0962079250106519, "grad_norm": 26.125, "learning_rate": 2.4880601056811843e-06, "loss": 1.2304, "step": 1129 }, { "epoch": 0.09629314017895185, "grad_norm": 20.5, "learning_rate": 2.488036924718737e-06, "loss": 1.0342, "step": 1130 }, { "epoch": 0.09637835534725181, "grad_norm": 18.375, "learning_rate": 2.4880137213837176e-06, "loss": 1.3621, "step": 1131 }, { "epoch": 0.09646357051555177, "grad_norm": 27.75, "learning_rate": 2.487990495676545e-06, "loss": 1.5038, "step": 1132 }, { "epoch": 0.09654878568385172, "grad_norm": 14.8125, "learning_rate": 2.48796724759764e-06, "loss": 0.9412, "step": 1133 }, { "epoch": 0.09663400085215168, "grad_norm": 14.0625, "learning_rate": 2.487943977147421e-06, "loss": 0.467, "step": 1134 }, { "epoch": 0.09671921602045164, "grad_norm": 14.1875, "learning_rate": 2.4879206843263105e-06, "loss": 0.7452, "step": 1135 }, { "epoch": 0.0968044311887516, "grad_norm": 12.6875, "learning_rate": 2.487897369134728e-06, "loss": 0.6697, "step": 1136 }, { "epoch": 0.09688964635705155, "grad_norm": 15.1875, "learning_rate": 2.4878740315730947e-06, "loss": 1.3572, "step": 1137 }, { "epoch": 0.09697486152535151, "grad_norm": 15.0625, "learning_rate": 2.487850671641834e-06, "loss": 0.9798, "step": 1138 }, { "epoch": 0.09706007669365146, "grad_norm": 14.0, "learning_rate": 2.4878272893413656e-06, "loss": 0.5652, "step": 1139 }, { "epoch": 0.09714529186195142, "grad_norm": 14.625, "learning_rate": 2.487803884672114e-06, "loss": 0.912, "step": 1140 }, { "epoch": 0.09723050703025138, "grad_norm": 12.75, "learning_rate": 2.4877804576345017e-06, "loss": 0.3384, "step": 1141 }, { "epoch": 0.09731572219855134, "grad_norm": 19.5, "learning_rate": 2.4877570082289518e-06, "loss": 1.1014, "step": 1142 }, { "epoch": 0.0974009373668513, "grad_norm": 13.1875, "learning_rate": 2.487733536455888e-06, "loss": 0.5668, "step": 1143 }, { "epoch": 0.09748615253515126, "grad_norm": 11.625, "learning_rate": 2.487710042315734e-06, "loss": 0.6697, "step": 1144 }, { "epoch": 0.09757136770345122, "grad_norm": 12.25, "learning_rate": 2.4876865258089157e-06, "loss": 0.7408, "step": 1145 }, { "epoch": 0.09765658287175118, "grad_norm": 11.25, "learning_rate": 2.487662986935857e-06, "loss": 0.5894, "step": 1146 }, { "epoch": 0.09774179804005113, "grad_norm": 12.3125, "learning_rate": 2.487639425696984e-06, "loss": 0.4808, "step": 1147 }, { "epoch": 0.09782701320835109, "grad_norm": 15.6875, "learning_rate": 2.487615842092721e-06, "loss": 0.9066, "step": 1148 }, { "epoch": 0.09791222837665105, "grad_norm": 14.4375, "learning_rate": 2.487592236123496e-06, "loss": 0.5614, "step": 1149 }, { "epoch": 0.097997443544951, "grad_norm": 10.75, "learning_rate": 2.487568607789735e-06, "loss": 0.3832, "step": 1150 }, { "epoch": 0.09808265871325096, "grad_norm": 11.125, "learning_rate": 2.4875449570918644e-06, "loss": 0.3155, "step": 1151 }, { "epoch": 0.09816787388155092, "grad_norm": 18.25, "learning_rate": 2.4875212840303115e-06, "loss": 0.9819, "step": 1152 }, { "epoch": 0.09825308904985088, "grad_norm": 24.125, "learning_rate": 2.4874975886055054e-06, "loss": 1.5273, "step": 1153 }, { "epoch": 0.09833830421815083, "grad_norm": 15.75, "learning_rate": 2.487473870817873e-06, "loss": 0.7491, "step": 1154 }, { "epoch": 0.09842351938645079, "grad_norm": 14.9375, "learning_rate": 2.4874501306678436e-06, "loss": 0.9202, "step": 1155 }, { "epoch": 0.09850873455475075, "grad_norm": 17.5, "learning_rate": 2.4874263681558456e-06, "loss": 0.8345, "step": 1156 }, { "epoch": 0.0985939497230507, "grad_norm": 14.0625, "learning_rate": 2.4874025832823096e-06, "loss": 0.8706, "step": 1157 }, { "epoch": 0.09867916489135066, "grad_norm": 26.0, "learning_rate": 2.487378776047664e-06, "loss": 0.9641, "step": 1158 }, { "epoch": 0.09876438005965062, "grad_norm": 14.625, "learning_rate": 2.4873549464523392e-06, "loss": 0.3667, "step": 1159 }, { "epoch": 0.09884959522795057, "grad_norm": 13.8125, "learning_rate": 2.4873310944967673e-06, "loss": 0.8958, "step": 1160 }, { "epoch": 0.09893481039625053, "grad_norm": 22.625, "learning_rate": 2.487307220181378e-06, "loss": 0.8326, "step": 1161 }, { "epoch": 0.09902002556455049, "grad_norm": 17.125, "learning_rate": 2.4872833235066024e-06, "loss": 0.745, "step": 1162 }, { "epoch": 0.09910524073285044, "grad_norm": 13.125, "learning_rate": 2.487259404472874e-06, "loss": 0.6665, "step": 1163 }, { "epoch": 0.0991904559011504, "grad_norm": 12.4375, "learning_rate": 2.487235463080623e-06, "loss": 0.6353, "step": 1164 }, { "epoch": 0.09927567106945036, "grad_norm": 16.0, "learning_rate": 2.4872114993302835e-06, "loss": 1.1939, "step": 1165 }, { "epoch": 0.09936088623775031, "grad_norm": 15.3125, "learning_rate": 2.4871875132222877e-06, "loss": 1.0054, "step": 1166 }, { "epoch": 0.09944610140605027, "grad_norm": 12.1875, "learning_rate": 2.48716350475707e-06, "loss": 0.5284, "step": 1167 }, { "epoch": 0.09953131657435023, "grad_norm": 13.8125, "learning_rate": 2.4871394739350636e-06, "loss": 0.7509, "step": 1168 }, { "epoch": 0.09961653174265019, "grad_norm": 15.375, "learning_rate": 2.4871154207567027e-06, "loss": 0.8657, "step": 1169 }, { "epoch": 0.09970174691095014, "grad_norm": 14.1875, "learning_rate": 2.487091345222422e-06, "loss": 0.6932, "step": 1170 }, { "epoch": 0.09978696207925011, "grad_norm": 30.75, "learning_rate": 2.4870672473326567e-06, "loss": 0.9452, "step": 1171 }, { "epoch": 0.09987217724755007, "grad_norm": 12.625, "learning_rate": 2.4870431270878427e-06, "loss": 0.6746, "step": 1172 }, { "epoch": 0.09995739241585003, "grad_norm": 13.6875, "learning_rate": 2.487018984488415e-06, "loss": 0.677, "step": 1173 }, { "epoch": 0.10004260758414998, "grad_norm": 12.25, "learning_rate": 2.486994819534811e-06, "loss": 0.6327, "step": 1174 }, { "epoch": 0.10012782275244994, "grad_norm": 15.375, "learning_rate": 2.4869706322274663e-06, "loss": 0.8116, "step": 1175 }, { "epoch": 0.1002130379207499, "grad_norm": 16.0, "learning_rate": 2.486946422566818e-06, "loss": 0.6631, "step": 1176 }, { "epoch": 0.10029825308904985, "grad_norm": 10.9375, "learning_rate": 2.486922190553304e-06, "loss": 0.5065, "step": 1177 }, { "epoch": 0.10038346825734981, "grad_norm": 15.4375, "learning_rate": 2.486897936187363e-06, "loss": 1.0097, "step": 1178 }, { "epoch": 0.10046868342564977, "grad_norm": 17.75, "learning_rate": 2.486873659469432e-06, "loss": 0.7886, "step": 1179 }, { "epoch": 0.10055389859394973, "grad_norm": 14.625, "learning_rate": 2.4868493603999503e-06, "loss": 0.8535, "step": 1180 }, { "epoch": 0.10063911376224968, "grad_norm": 16.25, "learning_rate": 2.486825038979357e-06, "loss": 0.7982, "step": 1181 }, { "epoch": 0.10072432893054964, "grad_norm": 12.3125, "learning_rate": 2.486800695208091e-06, "loss": 0.5671, "step": 1182 }, { "epoch": 0.1008095440988496, "grad_norm": 16.125, "learning_rate": 2.486776329086593e-06, "loss": 0.8319, "step": 1183 }, { "epoch": 0.10089475926714955, "grad_norm": 20.625, "learning_rate": 2.4867519406153036e-06, "loss": 1.1765, "step": 1184 }, { "epoch": 0.10097997443544951, "grad_norm": 22.125, "learning_rate": 2.4867275297946627e-06, "loss": 1.0212, "step": 1185 }, { "epoch": 0.10106518960374947, "grad_norm": 14.0, "learning_rate": 2.4867030966251112e-06, "loss": 0.9917, "step": 1186 }, { "epoch": 0.10115040477204942, "grad_norm": 18.25, "learning_rate": 2.4866786411070914e-06, "loss": 0.8056, "step": 1187 }, { "epoch": 0.10123561994034938, "grad_norm": 14.1875, "learning_rate": 2.486654163241045e-06, "loss": 0.8442, "step": 1188 }, { "epoch": 0.10132083510864934, "grad_norm": 15.8125, "learning_rate": 2.4866296630274144e-06, "loss": 1.195, "step": 1189 }, { "epoch": 0.1014060502769493, "grad_norm": 16.125, "learning_rate": 2.4866051404666424e-06, "loss": 0.9384, "step": 1190 }, { "epoch": 0.10149126544524925, "grad_norm": 16.25, "learning_rate": 2.4865805955591723e-06, "loss": 0.6707, "step": 1191 }, { "epoch": 0.10157648061354921, "grad_norm": 13.6875, "learning_rate": 2.4865560283054466e-06, "loss": 0.805, "step": 1192 }, { "epoch": 0.10166169578184917, "grad_norm": 14.8125, "learning_rate": 2.48653143870591e-06, "loss": 0.4752, "step": 1193 }, { "epoch": 0.10174691095014912, "grad_norm": 19.25, "learning_rate": 2.4865068267610077e-06, "loss": 0.9507, "step": 1194 }, { "epoch": 0.10183212611844908, "grad_norm": 22.625, "learning_rate": 2.486482192471183e-06, "loss": 0.9932, "step": 1195 }, { "epoch": 0.10191734128674904, "grad_norm": 23.0, "learning_rate": 2.486457535836882e-06, "loss": 0.9625, "step": 1196 }, { "epoch": 0.10200255645504899, "grad_norm": 21.125, "learning_rate": 2.48643285685855e-06, "loss": 1.0072, "step": 1197 }, { "epoch": 0.10208777162334895, "grad_norm": 21.5, "learning_rate": 2.4864081555366327e-06, "loss": 1.1831, "step": 1198 }, { "epoch": 0.10217298679164892, "grad_norm": 12.625, "learning_rate": 2.4863834318715767e-06, "loss": 0.7271, "step": 1199 }, { "epoch": 0.10225820195994888, "grad_norm": 12.25, "learning_rate": 2.4863586858638287e-06, "loss": 0.5084, "step": 1200 }, { "epoch": 0.10234341712824883, "grad_norm": 16.125, "learning_rate": 2.4863339175138364e-06, "loss": 0.9377, "step": 1201 }, { "epoch": 0.10242863229654879, "grad_norm": 14.0, "learning_rate": 2.486309126822047e-06, "loss": 0.7694, "step": 1202 }, { "epoch": 0.10251384746484875, "grad_norm": 10.3125, "learning_rate": 2.4862843137889083e-06, "loss": 0.4483, "step": 1203 }, { "epoch": 0.1025990626331487, "grad_norm": 22.375, "learning_rate": 2.4862594784148687e-06, "loss": 1.1014, "step": 1204 }, { "epoch": 0.10268427780144866, "grad_norm": 11.6875, "learning_rate": 2.4862346207003775e-06, "loss": 0.4566, "step": 1205 }, { "epoch": 0.10276949296974862, "grad_norm": 15.9375, "learning_rate": 2.486209740645883e-06, "loss": 0.8798, "step": 1206 }, { "epoch": 0.10285470813804858, "grad_norm": 10.5625, "learning_rate": 2.486184838251836e-06, "loss": 0.4533, "step": 1207 }, { "epoch": 0.10293992330634853, "grad_norm": 12.375, "learning_rate": 2.4861599135186855e-06, "loss": 0.4927, "step": 1208 }, { "epoch": 0.10302513847464849, "grad_norm": 13.8125, "learning_rate": 2.4861349664468825e-06, "loss": 0.7832, "step": 1209 }, { "epoch": 0.10311035364294845, "grad_norm": 15.375, "learning_rate": 2.4861099970368772e-06, "loss": 0.8285, "step": 1210 }, { "epoch": 0.1031955688112484, "grad_norm": 13.125, "learning_rate": 2.486085005289122e-06, "loss": 0.5292, "step": 1211 }, { "epoch": 0.10328078397954836, "grad_norm": 12.5625, "learning_rate": 2.4860599912040672e-06, "loss": 0.7756, "step": 1212 }, { "epoch": 0.10336599914784832, "grad_norm": 12.8125, "learning_rate": 2.486034954782166e-06, "loss": 0.8251, "step": 1213 }, { "epoch": 0.10345121431614827, "grad_norm": 14.1875, "learning_rate": 2.4860098960238697e-06, "loss": 0.8704, "step": 1214 }, { "epoch": 0.10353642948444823, "grad_norm": 14.1875, "learning_rate": 2.4859848149296315e-06, "loss": 0.94, "step": 1215 }, { "epoch": 0.10362164465274819, "grad_norm": 13.6875, "learning_rate": 2.4859597114999054e-06, "loss": 0.6716, "step": 1216 }, { "epoch": 0.10370685982104814, "grad_norm": 29.375, "learning_rate": 2.4859345857351443e-06, "loss": 1.5182, "step": 1217 }, { "epoch": 0.1037920749893481, "grad_norm": 21.875, "learning_rate": 2.4859094376358024e-06, "loss": 1.1244, "step": 1218 }, { "epoch": 0.10387729015764806, "grad_norm": 13.9375, "learning_rate": 2.4858842672023337e-06, "loss": 0.8572, "step": 1219 }, { "epoch": 0.10396250532594802, "grad_norm": 13.8125, "learning_rate": 2.485859074435194e-06, "loss": 0.6713, "step": 1220 }, { "epoch": 0.10404772049424797, "grad_norm": 21.375, "learning_rate": 2.485833859334838e-06, "loss": 0.9288, "step": 1221 }, { "epoch": 0.10413293566254793, "grad_norm": 13.0, "learning_rate": 2.4858086219017214e-06, "loss": 0.5945, "step": 1222 }, { "epoch": 0.10421815083084789, "grad_norm": 17.5, "learning_rate": 2.4857833621363003e-06, "loss": 0.7457, "step": 1223 }, { "epoch": 0.10430336599914784, "grad_norm": 16.625, "learning_rate": 2.4857580800390312e-06, "loss": 0.7915, "step": 1224 }, { "epoch": 0.1043885811674478, "grad_norm": 14.5625, "learning_rate": 2.485732775610371e-06, "loss": 0.9525, "step": 1225 }, { "epoch": 0.10447379633574776, "grad_norm": 14.0, "learning_rate": 2.4857074488507765e-06, "loss": 0.4845, "step": 1226 }, { "epoch": 0.10455901150404771, "grad_norm": 16.75, "learning_rate": 2.485682099760706e-06, "loss": 0.889, "step": 1227 }, { "epoch": 0.10464422667234768, "grad_norm": 15.375, "learning_rate": 2.4856567283406177e-06, "loss": 0.7665, "step": 1228 }, { "epoch": 0.10472944184064764, "grad_norm": 13.875, "learning_rate": 2.4856313345909697e-06, "loss": 0.5197, "step": 1229 }, { "epoch": 0.1048146570089476, "grad_norm": 20.625, "learning_rate": 2.4856059185122208e-06, "loss": 1.014, "step": 1230 }, { "epoch": 0.10489987217724756, "grad_norm": 10.875, "learning_rate": 2.4855804801048304e-06, "loss": 0.389, "step": 1231 }, { "epoch": 0.10498508734554751, "grad_norm": 22.75, "learning_rate": 2.4855550193692585e-06, "loss": 1.3122, "step": 1232 }, { "epoch": 0.10507030251384747, "grad_norm": 12.8125, "learning_rate": 2.4855295363059646e-06, "loss": 0.7156, "step": 1233 }, { "epoch": 0.10515551768214743, "grad_norm": 13.6875, "learning_rate": 2.48550403091541e-06, "loss": 0.6026, "step": 1234 }, { "epoch": 0.10524073285044738, "grad_norm": 14.1875, "learning_rate": 2.4854785031980547e-06, "loss": 0.7546, "step": 1235 }, { "epoch": 0.10532594801874734, "grad_norm": 11.0625, "learning_rate": 2.485452953154361e-06, "loss": 0.4887, "step": 1236 }, { "epoch": 0.1054111631870473, "grad_norm": 13.1875, "learning_rate": 2.4854273807847893e-06, "loss": 0.7915, "step": 1237 }, { "epoch": 0.10549637835534725, "grad_norm": 16.5, "learning_rate": 2.485401786089803e-06, "loss": 1.1728, "step": 1238 }, { "epoch": 0.10558159352364721, "grad_norm": 20.625, "learning_rate": 2.4853761690698643e-06, "loss": 1.2652, "step": 1239 }, { "epoch": 0.10566680869194717, "grad_norm": 24.125, "learning_rate": 2.485350529725436e-06, "loss": 1.2301, "step": 1240 }, { "epoch": 0.10575202386024712, "grad_norm": 17.375, "learning_rate": 2.485324868056981e-06, "loss": 0.9542, "step": 1241 }, { "epoch": 0.10583723902854708, "grad_norm": 20.125, "learning_rate": 2.4852991840649638e-06, "loss": 0.576, "step": 1242 }, { "epoch": 0.10592245419684704, "grad_norm": 17.5, "learning_rate": 2.4852734777498478e-06, "loss": 0.972, "step": 1243 }, { "epoch": 0.106007669365147, "grad_norm": 12.3125, "learning_rate": 2.485247749112098e-06, "loss": 0.6378, "step": 1244 }, { "epoch": 0.10609288453344695, "grad_norm": 13.875, "learning_rate": 2.485221998152179e-06, "loss": 0.3143, "step": 1245 }, { "epoch": 0.10617809970174691, "grad_norm": 11.375, "learning_rate": 2.485196224870557e-06, "loss": 0.6087, "step": 1246 }, { "epoch": 0.10626331487004687, "grad_norm": 12.875, "learning_rate": 2.485170429267697e-06, "loss": 0.6459, "step": 1247 }, { "epoch": 0.10634853003834682, "grad_norm": 24.75, "learning_rate": 2.4851446113440647e-06, "loss": 0.9692, "step": 1248 }, { "epoch": 0.10643374520664678, "grad_norm": 15.125, "learning_rate": 2.485118771100128e-06, "loss": 0.7668, "step": 1249 }, { "epoch": 0.10651896037494674, "grad_norm": 16.0, "learning_rate": 2.4850929085363525e-06, "loss": 0.6654, "step": 1250 }, { "epoch": 0.1066041755432467, "grad_norm": 12.1875, "learning_rate": 2.4850670236532065e-06, "loss": 0.6495, "step": 1251 }, { "epoch": 0.10668939071154665, "grad_norm": 14.0625, "learning_rate": 2.4850411164511573e-06, "loss": 0.7962, "step": 1252 }, { "epoch": 0.1067746058798466, "grad_norm": 13.4375, "learning_rate": 2.4850151869306734e-06, "loss": 0.7298, "step": 1253 }, { "epoch": 0.10685982104814656, "grad_norm": 18.375, "learning_rate": 2.4849892350922228e-06, "loss": 0.8274, "step": 1254 }, { "epoch": 0.10694503621644652, "grad_norm": 22.625, "learning_rate": 2.4849632609362747e-06, "loss": 1.147, "step": 1255 }, { "epoch": 0.10703025138474649, "grad_norm": 17.125, "learning_rate": 2.4849372644632993e-06, "loss": 0.8296, "step": 1256 }, { "epoch": 0.10711546655304645, "grad_norm": 14.875, "learning_rate": 2.484911245673765e-06, "loss": 0.73, "step": 1257 }, { "epoch": 0.1072006817213464, "grad_norm": 12.125, "learning_rate": 2.4848852045681434e-06, "loss": 0.6973, "step": 1258 }, { "epoch": 0.10728589688964636, "grad_norm": 18.625, "learning_rate": 2.4848591411469035e-06, "loss": 1.1448, "step": 1259 }, { "epoch": 0.10737111205794632, "grad_norm": 14.0, "learning_rate": 2.484833055410518e-06, "loss": 0.7084, "step": 1260 }, { "epoch": 0.10745632722624628, "grad_norm": 14.125, "learning_rate": 2.4848069473594573e-06, "loss": 0.8765, "step": 1261 }, { "epoch": 0.10754154239454623, "grad_norm": 13.5625, "learning_rate": 2.484780816994193e-06, "loss": 0.5018, "step": 1262 }, { "epoch": 0.10762675756284619, "grad_norm": 10.0, "learning_rate": 2.484754664315198e-06, "loss": 0.4434, "step": 1263 }, { "epoch": 0.10771197273114615, "grad_norm": 13.875, "learning_rate": 2.4847284893229445e-06, "loss": 0.7946, "step": 1264 }, { "epoch": 0.1077971878994461, "grad_norm": 16.75, "learning_rate": 2.4847022920179054e-06, "loss": 0.9332, "step": 1265 }, { "epoch": 0.10788240306774606, "grad_norm": 20.75, "learning_rate": 2.4846760724005544e-06, "loss": 1.1173, "step": 1266 }, { "epoch": 0.10796761823604602, "grad_norm": 11.9375, "learning_rate": 2.484649830471365e-06, "loss": 0.4956, "step": 1267 }, { "epoch": 0.10805283340434597, "grad_norm": 12.75, "learning_rate": 2.4846235662308123e-06, "loss": 0.4356, "step": 1268 }, { "epoch": 0.10813804857264593, "grad_norm": 17.0, "learning_rate": 2.48459727967937e-06, "loss": 1.0364, "step": 1269 }, { "epoch": 0.10822326374094589, "grad_norm": 16.0, "learning_rate": 2.484570970817513e-06, "loss": 0.8467, "step": 1270 }, { "epoch": 0.10830847890924584, "grad_norm": 13.375, "learning_rate": 2.4845446396457173e-06, "loss": 0.591, "step": 1271 }, { "epoch": 0.1083936940775458, "grad_norm": 12.5625, "learning_rate": 2.4845182861644584e-06, "loss": 0.4932, "step": 1272 }, { "epoch": 0.10847890924584576, "grad_norm": 14.9375, "learning_rate": 2.4844919103742126e-06, "loss": 0.8037, "step": 1273 }, { "epoch": 0.10856412441414572, "grad_norm": 11.6875, "learning_rate": 2.4844655122754567e-06, "loss": 0.5426, "step": 1274 }, { "epoch": 0.10864933958244567, "grad_norm": 13.25, "learning_rate": 2.4844390918686678e-06, "loss": 0.7429, "step": 1275 }, { "epoch": 0.10873455475074563, "grad_norm": 33.25, "learning_rate": 2.4844126491543234e-06, "loss": 1.3516, "step": 1276 }, { "epoch": 0.10881976991904559, "grad_norm": 15.375, "learning_rate": 2.484386184132901e-06, "loss": 0.9762, "step": 1277 }, { "epoch": 0.10890498508734554, "grad_norm": 19.625, "learning_rate": 2.4843596968048784e-06, "loss": 1.1892, "step": 1278 }, { "epoch": 0.1089902002556455, "grad_norm": 12.4375, "learning_rate": 2.4843331871707355e-06, "loss": 0.6167, "step": 1279 }, { "epoch": 0.10907541542394546, "grad_norm": 13.5, "learning_rate": 2.48430665523095e-06, "loss": 0.6848, "step": 1280 }, { "epoch": 0.10916063059224541, "grad_norm": 17.125, "learning_rate": 2.484280100986003e-06, "loss": 1.2219, "step": 1281 }, { "epoch": 0.10924584576054537, "grad_norm": 12.0, "learning_rate": 2.484253524436373e-06, "loss": 0.5148, "step": 1282 }, { "epoch": 0.10933106092884533, "grad_norm": 20.5, "learning_rate": 2.48422692558254e-06, "loss": 1.3084, "step": 1283 }, { "epoch": 0.1094162760971453, "grad_norm": 14.5, "learning_rate": 2.484200304424986e-06, "loss": 0.6643, "step": 1284 }, { "epoch": 0.10950149126544526, "grad_norm": 14.4375, "learning_rate": 2.4841736609641916e-06, "loss": 1.0793, "step": 1285 }, { "epoch": 0.10958670643374521, "grad_norm": 15.75, "learning_rate": 2.484146995200638e-06, "loss": 1.0071, "step": 1286 }, { "epoch": 0.10967192160204517, "grad_norm": 15.0625, "learning_rate": 2.4841203071348067e-06, "loss": 0.6509, "step": 1287 }, { "epoch": 0.10975713677034513, "grad_norm": 15.3125, "learning_rate": 2.4840935967671807e-06, "loss": 0.6941, "step": 1288 }, { "epoch": 0.10984235193864508, "grad_norm": 11.375, "learning_rate": 2.484066864098243e-06, "loss": 0.6386, "step": 1289 }, { "epoch": 0.10992756710694504, "grad_norm": 10.875, "learning_rate": 2.4840401091284754e-06, "loss": 0.5552, "step": 1290 }, { "epoch": 0.110012782275245, "grad_norm": 10.25, "learning_rate": 2.4840133318583625e-06, "loss": 0.541, "step": 1291 }, { "epoch": 0.11009799744354495, "grad_norm": 17.25, "learning_rate": 2.4839865322883873e-06, "loss": 1.0938, "step": 1292 }, { "epoch": 0.11018321261184491, "grad_norm": 16.875, "learning_rate": 2.483959710419035e-06, "loss": 0.8905, "step": 1293 }, { "epoch": 0.11026842778014487, "grad_norm": 23.375, "learning_rate": 2.4839328662507896e-06, "loss": 0.9578, "step": 1294 }, { "epoch": 0.11035364294844482, "grad_norm": 18.75, "learning_rate": 2.483905999784137e-06, "loss": 0.7415, "step": 1295 }, { "epoch": 0.11043885811674478, "grad_norm": 12.4375, "learning_rate": 2.4838791110195616e-06, "loss": 0.6517, "step": 1296 }, { "epoch": 0.11052407328504474, "grad_norm": 17.375, "learning_rate": 2.48385219995755e-06, "loss": 0.7366, "step": 1297 }, { "epoch": 0.1106092884533447, "grad_norm": 17.875, "learning_rate": 2.4838252665985887e-06, "loss": 0.6285, "step": 1298 }, { "epoch": 0.11069450362164465, "grad_norm": 14.5, "learning_rate": 2.483798310943164e-06, "loss": 0.6958, "step": 1299 }, { "epoch": 0.11077971878994461, "grad_norm": 15.3125, "learning_rate": 2.4837713329917634e-06, "loss": 0.5696, "step": 1300 }, { "epoch": 0.11086493395824457, "grad_norm": 22.875, "learning_rate": 2.483744332744874e-06, "loss": 1.282, "step": 1301 }, { "epoch": 0.11095014912654452, "grad_norm": 20.25, "learning_rate": 2.4837173102029833e-06, "loss": 1.2556, "step": 1302 }, { "epoch": 0.11103536429484448, "grad_norm": 15.5625, "learning_rate": 2.483690265366581e-06, "loss": 0.9616, "step": 1303 }, { "epoch": 0.11112057946314444, "grad_norm": 13.8125, "learning_rate": 2.4836631982361546e-06, "loss": 0.8363, "step": 1304 }, { "epoch": 0.1112057946314444, "grad_norm": 14.0, "learning_rate": 2.483636108812194e-06, "loss": 0.5904, "step": 1305 }, { "epoch": 0.11129100979974435, "grad_norm": 20.25, "learning_rate": 2.483608997095188e-06, "loss": 1.1782, "step": 1306 }, { "epoch": 0.11137622496804431, "grad_norm": 13.875, "learning_rate": 2.483581863085627e-06, "loss": 0.5778, "step": 1307 }, { "epoch": 0.11146144013634426, "grad_norm": 18.5, "learning_rate": 2.4835547067840015e-06, "loss": 0.9187, "step": 1308 }, { "epoch": 0.11154665530464422, "grad_norm": 22.375, "learning_rate": 2.483527528190802e-06, "loss": 1.2057, "step": 1309 }, { "epoch": 0.11163187047294418, "grad_norm": 13.0, "learning_rate": 2.4835003273065197e-06, "loss": 0.8024, "step": 1310 }, { "epoch": 0.11171708564124413, "grad_norm": 11.1875, "learning_rate": 2.4834731041316457e-06, "loss": 0.5612, "step": 1311 }, { "epoch": 0.1118023008095441, "grad_norm": 19.5, "learning_rate": 2.4834458586666725e-06, "loss": 0.5, "step": 1312 }, { "epoch": 0.11188751597784406, "grad_norm": 11.8125, "learning_rate": 2.483418590912093e-06, "loss": 0.5764, "step": 1313 }, { "epoch": 0.11197273114614402, "grad_norm": 11.0, "learning_rate": 2.4833913008683985e-06, "loss": 0.5075, "step": 1314 }, { "epoch": 0.11205794631444398, "grad_norm": 10.5, "learning_rate": 2.483363988536083e-06, "loss": 0.3628, "step": 1315 }, { "epoch": 0.11214316148274393, "grad_norm": 19.375, "learning_rate": 2.48333665391564e-06, "loss": 0.8608, "step": 1316 }, { "epoch": 0.11222837665104389, "grad_norm": 17.0, "learning_rate": 2.4833092970075638e-06, "loss": 1.0525, "step": 1317 }, { "epoch": 0.11231359181934385, "grad_norm": 13.5, "learning_rate": 2.4832819178123477e-06, "loss": 0.9775, "step": 1318 }, { "epoch": 0.1123988069876438, "grad_norm": 15.3125, "learning_rate": 2.483254516330488e-06, "loss": 0.6232, "step": 1319 }, { "epoch": 0.11248402215594376, "grad_norm": 24.125, "learning_rate": 2.4832270925624784e-06, "loss": 0.9508, "step": 1320 }, { "epoch": 0.11256923732424372, "grad_norm": 22.0, "learning_rate": 2.4831996465088156e-06, "loss": 0.8841, "step": 1321 }, { "epoch": 0.11265445249254367, "grad_norm": 12.3125, "learning_rate": 2.483172178169995e-06, "loss": 0.6611, "step": 1322 }, { "epoch": 0.11273966766084363, "grad_norm": 15.625, "learning_rate": 2.4831446875465133e-06, "loss": 1.064, "step": 1323 }, { "epoch": 0.11282488282914359, "grad_norm": 20.75, "learning_rate": 2.483117174638867e-06, "loss": 0.9941, "step": 1324 }, { "epoch": 0.11291009799744355, "grad_norm": 17.0, "learning_rate": 2.483089639447553e-06, "loss": 0.9082, "step": 1325 }, { "epoch": 0.1129953131657435, "grad_norm": 20.75, "learning_rate": 2.4830620819730693e-06, "loss": 0.8576, "step": 1326 }, { "epoch": 0.11308052833404346, "grad_norm": 18.375, "learning_rate": 2.483034502215914e-06, "loss": 1.2188, "step": 1327 }, { "epoch": 0.11316574350234342, "grad_norm": 11.1875, "learning_rate": 2.483006900176585e-06, "loss": 0.4908, "step": 1328 }, { "epoch": 0.11325095867064337, "grad_norm": 13.25, "learning_rate": 2.4829792758555817e-06, "loss": 0.8232, "step": 1329 }, { "epoch": 0.11333617383894333, "grad_norm": 16.125, "learning_rate": 2.482951629253403e-06, "loss": 1.0177, "step": 1330 }, { "epoch": 0.11342138900724329, "grad_norm": 13.9375, "learning_rate": 2.4829239603705486e-06, "loss": 0.7121, "step": 1331 }, { "epoch": 0.11350660417554324, "grad_norm": 15.0625, "learning_rate": 2.4828962692075187e-06, "loss": 0.6933, "step": 1332 }, { "epoch": 0.1135918193438432, "grad_norm": 16.25, "learning_rate": 2.482868555764813e-06, "loss": 0.9299, "step": 1333 }, { "epoch": 0.11367703451214316, "grad_norm": 12.6875, "learning_rate": 2.482840820042933e-06, "loss": 0.5928, "step": 1334 }, { "epoch": 0.11376224968044311, "grad_norm": 10.875, "learning_rate": 2.4828130620423795e-06, "loss": 0.4289, "step": 1335 }, { "epoch": 0.11384746484874307, "grad_norm": 15.5, "learning_rate": 2.4827852817636545e-06, "loss": 0.7182, "step": 1336 }, { "epoch": 0.11393268001704303, "grad_norm": 11.0, "learning_rate": 2.4827574792072594e-06, "loss": 0.4529, "step": 1337 }, { "epoch": 0.11401789518534299, "grad_norm": 21.0, "learning_rate": 2.482729654373697e-06, "loss": 0.8668, "step": 1338 }, { "epoch": 0.11410311035364294, "grad_norm": 17.625, "learning_rate": 2.4827018072634707e-06, "loss": 1.015, "step": 1339 }, { "epoch": 0.11418832552194291, "grad_norm": 14.0625, "learning_rate": 2.4826739378770833e-06, "loss": 0.7146, "step": 1340 }, { "epoch": 0.11427354069024287, "grad_norm": 10.9375, "learning_rate": 2.4826460462150376e-06, "loss": 0.3754, "step": 1341 }, { "epoch": 0.11435875585854283, "grad_norm": 16.875, "learning_rate": 2.4826181322778385e-06, "loss": 0.8204, "step": 1342 }, { "epoch": 0.11444397102684278, "grad_norm": 15.5625, "learning_rate": 2.4825901960659904e-06, "loss": 0.947, "step": 1343 }, { "epoch": 0.11452918619514274, "grad_norm": 11.4375, "learning_rate": 2.482562237579998e-06, "loss": 0.415, "step": 1344 }, { "epoch": 0.1146144013634427, "grad_norm": 18.75, "learning_rate": 2.4825342568203664e-06, "loss": 0.9712, "step": 1345 }, { "epoch": 0.11469961653174265, "grad_norm": 15.75, "learning_rate": 2.4825062537876014e-06, "loss": 0.8323, "step": 1346 }, { "epoch": 0.11478483170004261, "grad_norm": 15.0625, "learning_rate": 2.4824782284822092e-06, "loss": 0.9422, "step": 1347 }, { "epoch": 0.11487004686834257, "grad_norm": 14.8125, "learning_rate": 2.4824501809046956e-06, "loss": 0.4773, "step": 1348 }, { "epoch": 0.11495526203664252, "grad_norm": 13.3125, "learning_rate": 2.4824221110555683e-06, "loss": 0.7092, "step": 1349 }, { "epoch": 0.11504047720494248, "grad_norm": 15.5625, "learning_rate": 2.482394018935334e-06, "loss": 0.8438, "step": 1350 }, { "epoch": 0.11512569237324244, "grad_norm": 14.3125, "learning_rate": 2.482365904544501e-06, "loss": 0.7924, "step": 1351 }, { "epoch": 0.1152109075415424, "grad_norm": 16.125, "learning_rate": 2.4823377678835763e-06, "loss": 0.9336, "step": 1352 }, { "epoch": 0.11529612270984235, "grad_norm": 18.0, "learning_rate": 2.4823096089530692e-06, "loss": 0.926, "step": 1353 }, { "epoch": 0.11538133787814231, "grad_norm": 12.125, "learning_rate": 2.482281427753488e-06, "loss": 0.6261, "step": 1354 }, { "epoch": 0.11546655304644227, "grad_norm": 19.75, "learning_rate": 2.482253224285342e-06, "loss": 0.4516, "step": 1355 }, { "epoch": 0.11555176821474222, "grad_norm": 10.9375, "learning_rate": 2.4822249985491416e-06, "loss": 0.5531, "step": 1356 }, { "epoch": 0.11563698338304218, "grad_norm": 13.625, "learning_rate": 2.4821967505453965e-06, "loss": 0.7452, "step": 1357 }, { "epoch": 0.11572219855134214, "grad_norm": 13.8125, "learning_rate": 2.4821684802746167e-06, "loss": 0.7824, "step": 1358 }, { "epoch": 0.1158074137196421, "grad_norm": 15.6875, "learning_rate": 2.4821401877373135e-06, "loss": 0.6703, "step": 1359 }, { "epoch": 0.11589262888794205, "grad_norm": 18.875, "learning_rate": 2.4821118729339985e-06, "loss": 0.873, "step": 1360 }, { "epoch": 0.11597784405624201, "grad_norm": 27.0, "learning_rate": 2.4820835358651823e-06, "loss": 0.9078, "step": 1361 }, { "epoch": 0.11606305922454196, "grad_norm": 13.5625, "learning_rate": 2.4820551765313784e-06, "loss": 0.6403, "step": 1362 }, { "epoch": 0.11614827439284192, "grad_norm": 18.0, "learning_rate": 2.482026794933098e-06, "loss": 1.0153, "step": 1363 }, { "epoch": 0.11623348956114188, "grad_norm": 17.5, "learning_rate": 2.4819983910708546e-06, "loss": 0.8461, "step": 1364 }, { "epoch": 0.11631870472944184, "grad_norm": 14.625, "learning_rate": 2.4819699649451613e-06, "loss": 0.9227, "step": 1365 }, { "epoch": 0.11640391989774179, "grad_norm": 15.5, "learning_rate": 2.4819415165565317e-06, "loss": 0.6959, "step": 1366 }, { "epoch": 0.11648913506604175, "grad_norm": 13.25, "learning_rate": 2.481913045905481e-06, "loss": 0.6191, "step": 1367 }, { "epoch": 0.1165743502343417, "grad_norm": 14.625, "learning_rate": 2.4818845529925216e-06, "loss": 0.3106, "step": 1368 }, { "epoch": 0.11665956540264168, "grad_norm": 24.25, "learning_rate": 2.4818560378181707e-06, "loss": 1.1964, "step": 1369 }, { "epoch": 0.11674478057094163, "grad_norm": 19.375, "learning_rate": 2.481827500382942e-06, "loss": 0.9775, "step": 1370 }, { "epoch": 0.11682999573924159, "grad_norm": 12.875, "learning_rate": 2.4817989406873512e-06, "loss": 0.7623, "step": 1371 }, { "epoch": 0.11691521090754155, "grad_norm": 28.375, "learning_rate": 2.481770358731915e-06, "loss": 0.5829, "step": 1372 }, { "epoch": 0.1170004260758415, "grad_norm": 11.4375, "learning_rate": 2.48174175451715e-06, "loss": 0.5868, "step": 1373 }, { "epoch": 0.11708564124414146, "grad_norm": 12.75, "learning_rate": 2.481713128043573e-06, "loss": 0.5655, "step": 1374 }, { "epoch": 0.11717085641244142, "grad_norm": 16.75, "learning_rate": 2.4816844793117014e-06, "loss": 0.9376, "step": 1375 }, { "epoch": 0.11725607158074138, "grad_norm": 14.25, "learning_rate": 2.481655808322052e-06, "loss": 0.5245, "step": 1376 }, { "epoch": 0.11734128674904133, "grad_norm": 13.9375, "learning_rate": 2.4816271150751444e-06, "loss": 0.5217, "step": 1377 }, { "epoch": 0.11742650191734129, "grad_norm": 14.75, "learning_rate": 2.481598399571496e-06, "loss": 0.8422, "step": 1378 }, { "epoch": 0.11751171708564125, "grad_norm": 15.625, "learning_rate": 2.4815696618116257e-06, "loss": 1.1816, "step": 1379 }, { "epoch": 0.1175969322539412, "grad_norm": 13.5625, "learning_rate": 2.4815409017960538e-06, "loss": 0.8437, "step": 1380 }, { "epoch": 0.11768214742224116, "grad_norm": 13.3125, "learning_rate": 2.481512119525299e-06, "loss": 0.8476, "step": 1381 }, { "epoch": 0.11776736259054112, "grad_norm": 13.5, "learning_rate": 2.481483314999882e-06, "loss": 0.8877, "step": 1382 }, { "epoch": 0.11785257775884107, "grad_norm": 12.875, "learning_rate": 2.481454488220323e-06, "loss": 0.7549, "step": 1383 }, { "epoch": 0.11793779292714103, "grad_norm": 19.375, "learning_rate": 2.4814256391871435e-06, "loss": 0.7129, "step": 1384 }, { "epoch": 0.11802300809544099, "grad_norm": 11.1875, "learning_rate": 2.481396767900864e-06, "loss": 0.5135, "step": 1385 }, { "epoch": 0.11810822326374094, "grad_norm": 14.625, "learning_rate": 2.481367874362006e-06, "loss": 0.7443, "step": 1386 }, { "epoch": 0.1181934384320409, "grad_norm": 9.1875, "learning_rate": 2.4813389585710933e-06, "loss": 0.3536, "step": 1387 }, { "epoch": 0.11827865360034086, "grad_norm": 14.75, "learning_rate": 2.481310020528647e-06, "loss": 0.6719, "step": 1388 }, { "epoch": 0.11836386876864081, "grad_norm": 10.9375, "learning_rate": 2.4812810602351907e-06, "loss": 0.5734, "step": 1389 }, { "epoch": 0.11844908393694077, "grad_norm": 19.375, "learning_rate": 2.481252077691247e-06, "loss": 1.0354, "step": 1390 }, { "epoch": 0.11853429910524073, "grad_norm": 16.75, "learning_rate": 2.4812230728973407e-06, "loss": 0.8676, "step": 1391 }, { "epoch": 0.11861951427354069, "grad_norm": 13.5, "learning_rate": 2.481194045853995e-06, "loss": 0.6957, "step": 1392 }, { "epoch": 0.11870472944184064, "grad_norm": 16.0, "learning_rate": 2.481164996561735e-06, "loss": 0.9095, "step": 1393 }, { "epoch": 0.1187899446101406, "grad_norm": 10.5625, "learning_rate": 2.4811359250210852e-06, "loss": 0.4006, "step": 1394 }, { "epoch": 0.11887515977844056, "grad_norm": 12.3125, "learning_rate": 2.4811068312325713e-06, "loss": 0.6876, "step": 1395 }, { "epoch": 0.11896037494674051, "grad_norm": 12.6875, "learning_rate": 2.4810777151967196e-06, "loss": 0.7951, "step": 1396 }, { "epoch": 0.11904559011504048, "grad_norm": 15.4375, "learning_rate": 2.4810485769140554e-06, "loss": 0.6443, "step": 1397 }, { "epoch": 0.11913080528334044, "grad_norm": 14.5625, "learning_rate": 2.481019416385105e-06, "loss": 0.9024, "step": 1398 }, { "epoch": 0.1192160204516404, "grad_norm": 13.375, "learning_rate": 2.480990233610396e-06, "loss": 0.6823, "step": 1399 }, { "epoch": 0.11930123561994035, "grad_norm": 24.125, "learning_rate": 2.4809610285904558e-06, "loss": 0.7423, "step": 1400 }, { "epoch": 0.11938645078824031, "grad_norm": 19.0, "learning_rate": 2.480931801325812e-06, "loss": 0.9985, "step": 1401 }, { "epoch": 0.11947166595654027, "grad_norm": 16.25, "learning_rate": 2.480902551816993e-06, "loss": 0.7579, "step": 1402 }, { "epoch": 0.11955688112484023, "grad_norm": 16.75, "learning_rate": 2.4808732800645264e-06, "loss": 1.2039, "step": 1403 }, { "epoch": 0.11964209629314018, "grad_norm": 15.4375, "learning_rate": 2.4808439860689427e-06, "loss": 0.6401, "step": 1404 }, { "epoch": 0.11972731146144014, "grad_norm": 25.125, "learning_rate": 2.48081466983077e-06, "loss": 0.9703, "step": 1405 }, { "epoch": 0.1198125266297401, "grad_norm": 15.5, "learning_rate": 2.4807853313505385e-06, "loss": 0.8871, "step": 1406 }, { "epoch": 0.11989774179804005, "grad_norm": 25.875, "learning_rate": 2.4807559706287786e-06, "loss": 0.7812, "step": 1407 }, { "epoch": 0.11998295696634001, "grad_norm": 13.8125, "learning_rate": 2.4807265876660205e-06, "loss": 0.8993, "step": 1408 }, { "epoch": 0.12006817213463997, "grad_norm": 22.375, "learning_rate": 2.4806971824627955e-06, "loss": 1.2011, "step": 1409 }, { "epoch": 0.12015338730293992, "grad_norm": 14.375, "learning_rate": 2.480667755019635e-06, "loss": 0.6522, "step": 1410 }, { "epoch": 0.12023860247123988, "grad_norm": 16.375, "learning_rate": 2.4806383053370704e-06, "loss": 0.703, "step": 1411 }, { "epoch": 0.12032381763953984, "grad_norm": 16.25, "learning_rate": 2.480608833415634e-06, "loss": 1.1941, "step": 1412 }, { "epoch": 0.1204090328078398, "grad_norm": 16.375, "learning_rate": 2.4805793392558587e-06, "loss": 0.6089, "step": 1413 }, { "epoch": 0.12049424797613975, "grad_norm": 14.25, "learning_rate": 2.4805498228582776e-06, "loss": 0.8614, "step": 1414 }, { "epoch": 0.12057946314443971, "grad_norm": 13.5, "learning_rate": 2.4805202842234233e-06, "loss": 0.7704, "step": 1415 }, { "epoch": 0.12066467831273966, "grad_norm": 12.875, "learning_rate": 2.4804907233518305e-06, "loss": 0.827, "step": 1416 }, { "epoch": 0.12074989348103962, "grad_norm": 12.5, "learning_rate": 2.4804611402440326e-06, "loss": 0.6887, "step": 1417 }, { "epoch": 0.12083510864933958, "grad_norm": 15.5, "learning_rate": 2.480431534900565e-06, "loss": 0.703, "step": 1418 }, { "epoch": 0.12092032381763954, "grad_norm": 15.5625, "learning_rate": 2.4804019073219617e-06, "loss": 0.7831, "step": 1419 }, { "epoch": 0.12100553898593949, "grad_norm": 16.375, "learning_rate": 2.480372257508759e-06, "loss": 0.946, "step": 1420 }, { "epoch": 0.12109075415423945, "grad_norm": 15.4375, "learning_rate": 2.480342585461492e-06, "loss": 0.7256, "step": 1421 }, { "epoch": 0.1211759693225394, "grad_norm": 31.75, "learning_rate": 2.480312891180698e-06, "loss": 1.0698, "step": 1422 }, { "epoch": 0.12126118449083936, "grad_norm": 17.625, "learning_rate": 2.480283174666912e-06, "loss": 1.0257, "step": 1423 }, { "epoch": 0.12134639965913932, "grad_norm": 12.375, "learning_rate": 2.4802534359206727e-06, "loss": 0.5837, "step": 1424 }, { "epoch": 0.12143161482743929, "grad_norm": 19.625, "learning_rate": 2.480223674942516e-06, "loss": 1.1336, "step": 1425 }, { "epoch": 0.12151682999573925, "grad_norm": 14.875, "learning_rate": 2.4801938917329808e-06, "loss": 0.6785, "step": 1426 }, { "epoch": 0.1216020451640392, "grad_norm": 17.375, "learning_rate": 2.4801640862926054e-06, "loss": 1.2337, "step": 1427 }, { "epoch": 0.12168726033233916, "grad_norm": 14.75, "learning_rate": 2.4801342586219273e-06, "loss": 0.8515, "step": 1428 }, { "epoch": 0.12177247550063912, "grad_norm": 12.5625, "learning_rate": 2.4801044087214867e-06, "loss": 0.6902, "step": 1429 }, { "epoch": 0.12185769066893908, "grad_norm": 17.0, "learning_rate": 2.4800745365918216e-06, "loss": 0.6936, "step": 1430 }, { "epoch": 0.12194290583723903, "grad_norm": 20.5, "learning_rate": 2.480044642233473e-06, "loss": 1.3433, "step": 1431 }, { "epoch": 0.12202812100553899, "grad_norm": 14.3125, "learning_rate": 2.480014725646981e-06, "loss": 0.793, "step": 1432 }, { "epoch": 0.12211333617383895, "grad_norm": 11.25, "learning_rate": 2.4799847868328863e-06, "loss": 0.622, "step": 1433 }, { "epoch": 0.1221985513421389, "grad_norm": 12.5, "learning_rate": 2.4799548257917292e-06, "loss": 0.741, "step": 1434 }, { "epoch": 0.12228376651043886, "grad_norm": 14.25, "learning_rate": 2.479924842524052e-06, "loss": 0.8168, "step": 1435 }, { "epoch": 0.12236898167873882, "grad_norm": 15.1875, "learning_rate": 2.4798948370303953e-06, "loss": 0.8108, "step": 1436 }, { "epoch": 0.12245419684703877, "grad_norm": 19.25, "learning_rate": 2.4798648093113027e-06, "loss": 0.8577, "step": 1437 }, { "epoch": 0.12253941201533873, "grad_norm": 11.125, "learning_rate": 2.479834759367316e-06, "loss": 0.4022, "step": 1438 }, { "epoch": 0.12262462718363869, "grad_norm": 13.8125, "learning_rate": 2.479804687198979e-06, "loss": 1.0363, "step": 1439 }, { "epoch": 0.12270984235193864, "grad_norm": 15.0, "learning_rate": 2.4797745928068346e-06, "loss": 0.9002, "step": 1440 }, { "epoch": 0.1227950575202386, "grad_norm": 13.6875, "learning_rate": 2.479744476191426e-06, "loss": 0.661, "step": 1441 }, { "epoch": 0.12288027268853856, "grad_norm": 18.0, "learning_rate": 2.479714337353299e-06, "loss": 0.9553, "step": 1442 }, { "epoch": 0.12296548785683852, "grad_norm": 15.1875, "learning_rate": 2.4796841762929967e-06, "loss": 0.9295, "step": 1443 }, { "epoch": 0.12305070302513847, "grad_norm": 20.0, "learning_rate": 2.4796539930110652e-06, "loss": 0.8901, "step": 1444 }, { "epoch": 0.12313591819343843, "grad_norm": 13.9375, "learning_rate": 2.479623787508049e-06, "loss": 0.6673, "step": 1445 }, { "epoch": 0.12322113336173839, "grad_norm": 17.25, "learning_rate": 2.479593559784495e-06, "loss": 1.2743, "step": 1446 }, { "epoch": 0.12330634853003834, "grad_norm": 14.875, "learning_rate": 2.479563309840949e-06, "loss": 0.8796, "step": 1447 }, { "epoch": 0.1233915636983383, "grad_norm": 13.9375, "learning_rate": 2.479533037677957e-06, "loss": 0.6922, "step": 1448 }, { "epoch": 0.12347677886663826, "grad_norm": 11.125, "learning_rate": 2.4795027432960673e-06, "loss": 0.5181, "step": 1449 }, { "epoch": 0.12356199403493821, "grad_norm": 12.3125, "learning_rate": 2.479472426695826e-06, "loss": 0.5184, "step": 1450 }, { "epoch": 0.12364720920323817, "grad_norm": 17.5, "learning_rate": 2.4794420878777818e-06, "loss": 1.2067, "step": 1451 }, { "epoch": 0.12373242437153813, "grad_norm": 17.625, "learning_rate": 2.4794117268424834e-06, "loss": 0.9037, "step": 1452 }, { "epoch": 0.1238176395398381, "grad_norm": 18.625, "learning_rate": 2.4793813435904782e-06, "loss": 1.0361, "step": 1453 }, { "epoch": 0.12390285470813805, "grad_norm": 13.75, "learning_rate": 2.479350938122316e-06, "loss": 0.8246, "step": 1454 }, { "epoch": 0.12398806987643801, "grad_norm": 15.1875, "learning_rate": 2.479320510438546e-06, "loss": 0.7256, "step": 1455 }, { "epoch": 0.12407328504473797, "grad_norm": 24.5, "learning_rate": 2.4792900605397185e-06, "loss": 1.6031, "step": 1456 }, { "epoch": 0.12415850021303793, "grad_norm": 17.75, "learning_rate": 2.4792595884263833e-06, "loss": 1.1003, "step": 1457 }, { "epoch": 0.12424371538133788, "grad_norm": 16.875, "learning_rate": 2.479229094099091e-06, "loss": 0.7001, "step": 1458 }, { "epoch": 0.12432893054963784, "grad_norm": 20.0, "learning_rate": 2.479198577558393e-06, "loss": 0.9321, "step": 1459 }, { "epoch": 0.1244141457179378, "grad_norm": 16.125, "learning_rate": 2.479168038804841e-06, "loss": 0.9677, "step": 1460 }, { "epoch": 0.12449936088623775, "grad_norm": 13.875, "learning_rate": 2.4791374778389866e-06, "loss": 0.7319, "step": 1461 }, { "epoch": 0.12458457605453771, "grad_norm": 14.875, "learning_rate": 2.4791068946613815e-06, "loss": 0.8257, "step": 1462 }, { "epoch": 0.12466979122283767, "grad_norm": 13.375, "learning_rate": 2.479076289272579e-06, "loss": 0.7955, "step": 1463 }, { "epoch": 0.12475500639113762, "grad_norm": 15.1875, "learning_rate": 2.479045661673132e-06, "loss": 0.4965, "step": 1464 }, { "epoch": 0.12484022155943758, "grad_norm": 16.5, "learning_rate": 2.479015011863594e-06, "loss": 0.6017, "step": 1465 }, { "epoch": 0.12492543672773754, "grad_norm": 15.875, "learning_rate": 2.478984339844519e-06, "loss": 0.7101, "step": 1466 }, { "epoch": 0.1250106518960375, "grad_norm": 13.625, "learning_rate": 2.478953645616461e-06, "loss": 0.9422, "step": 1467 }, { "epoch": 0.12509586706433745, "grad_norm": 14.3125, "learning_rate": 2.4789229291799748e-06, "loss": 0.8229, "step": 1468 }, { "epoch": 0.1251810822326374, "grad_norm": 10.1875, "learning_rate": 2.4788921905356157e-06, "loss": 0.3218, "step": 1469 }, { "epoch": 0.12526629740093737, "grad_norm": 20.125, "learning_rate": 2.4788614296839387e-06, "loss": 0.7974, "step": 1470 }, { "epoch": 0.12535151256923732, "grad_norm": 13.375, "learning_rate": 2.4788306466255e-06, "loss": 0.5381, "step": 1471 }, { "epoch": 0.12543672773753728, "grad_norm": 13.0, "learning_rate": 2.4787998413608556e-06, "loss": 0.6639, "step": 1472 }, { "epoch": 0.12552194290583724, "grad_norm": 23.75, "learning_rate": 2.478769013890563e-06, "loss": 1.1722, "step": 1473 }, { "epoch": 0.1256071580741372, "grad_norm": 14.0, "learning_rate": 2.478738164215178e-06, "loss": 0.7418, "step": 1474 }, { "epoch": 0.12569237324243715, "grad_norm": 15.3125, "learning_rate": 2.4787072923352597e-06, "loss": 1.033, "step": 1475 }, { "epoch": 0.1257775884107371, "grad_norm": 14.25, "learning_rate": 2.478676398251364e-06, "loss": 0.6169, "step": 1476 }, { "epoch": 0.12586280357903706, "grad_norm": 16.125, "learning_rate": 2.478645481964051e-06, "loss": 1.163, "step": 1477 }, { "epoch": 0.12594801874733702, "grad_norm": 25.375, "learning_rate": 2.4786145434738783e-06, "loss": 1.371, "step": 1478 }, { "epoch": 0.12603323391563698, "grad_norm": 15.125, "learning_rate": 2.4785835827814055e-06, "loss": 0.672, "step": 1479 }, { "epoch": 0.12611844908393693, "grad_norm": 14.0, "learning_rate": 2.478552599887192e-06, "loss": 0.8676, "step": 1480 }, { "epoch": 0.1262036642522369, "grad_norm": 22.25, "learning_rate": 2.4785215947917973e-06, "loss": 1.2902, "step": 1481 }, { "epoch": 0.12628887942053685, "grad_norm": 16.125, "learning_rate": 2.478490567495782e-06, "loss": 0.8189, "step": 1482 }, { "epoch": 0.1263740945888368, "grad_norm": 18.25, "learning_rate": 2.4784595179997073e-06, "loss": 0.9453, "step": 1483 }, { "epoch": 0.12645930975713676, "grad_norm": 11.0625, "learning_rate": 2.4784284463041334e-06, "loss": 0.6488, "step": 1484 }, { "epoch": 0.12654452492543672, "grad_norm": 12.0625, "learning_rate": 2.4783973524096226e-06, "loss": 0.5557, "step": 1485 }, { "epoch": 0.12662974009373668, "grad_norm": 14.5, "learning_rate": 2.478366236316736e-06, "loss": 0.7248, "step": 1486 }, { "epoch": 0.12671495526203663, "grad_norm": 20.75, "learning_rate": 2.4783350980260366e-06, "loss": 0.633, "step": 1487 }, { "epoch": 0.1268001704303366, "grad_norm": 13.375, "learning_rate": 2.4783039375380864e-06, "loss": 0.5183, "step": 1488 }, { "epoch": 0.12688538559863655, "grad_norm": 14.5625, "learning_rate": 2.478272754853449e-06, "loss": 0.8545, "step": 1489 }, { "epoch": 0.1269706007669365, "grad_norm": 14.5625, "learning_rate": 2.4782415499726883e-06, "loss": 0.6475, "step": 1490 }, { "epoch": 0.12705581593523646, "grad_norm": 9.6875, "learning_rate": 2.478210322896367e-06, "loss": 0.3222, "step": 1491 }, { "epoch": 0.12714103110353642, "grad_norm": 23.0, "learning_rate": 2.4781790736250507e-06, "loss": 1.0186, "step": 1492 }, { "epoch": 0.12722624627183637, "grad_norm": 15.75, "learning_rate": 2.478147802159303e-06, "loss": 0.5955, "step": 1493 }, { "epoch": 0.12731146144013633, "grad_norm": 12.375, "learning_rate": 2.4781165084996895e-06, "loss": 0.6227, "step": 1494 }, { "epoch": 0.1273966766084363, "grad_norm": 27.125, "learning_rate": 2.478085192646776e-06, "loss": 1.0827, "step": 1495 }, { "epoch": 0.12748189177673627, "grad_norm": 12.875, "learning_rate": 2.478053854601128e-06, "loss": 0.492, "step": 1496 }, { "epoch": 0.12756710694503623, "grad_norm": 14.625, "learning_rate": 2.478022494363312e-06, "loss": 0.5216, "step": 1497 }, { "epoch": 0.1276523221133362, "grad_norm": 15.375, "learning_rate": 2.4779911119338948e-06, "loss": 0.7451, "step": 1498 }, { "epoch": 0.12773753728163614, "grad_norm": 12.125, "learning_rate": 2.477959707313443e-06, "loss": 0.6604, "step": 1499 }, { "epoch": 0.1278227524499361, "grad_norm": 14.875, "learning_rate": 2.477928280502525e-06, "loss": 0.8966, "step": 1500 }, { "epoch": 0.12790796761823606, "grad_norm": 19.5, "learning_rate": 2.477896831501708e-06, "loss": 1.037, "step": 1501 }, { "epoch": 0.12799318278653601, "grad_norm": 12.875, "learning_rate": 2.47786536031156e-06, "loss": 0.5206, "step": 1502 }, { "epoch": 0.12807839795483597, "grad_norm": 17.0, "learning_rate": 2.4778338669326507e-06, "loss": 0.9273, "step": 1503 }, { "epoch": 0.12816361312313593, "grad_norm": 14.4375, "learning_rate": 2.477802351365549e-06, "loss": 0.7273, "step": 1504 }, { "epoch": 0.12824882829143588, "grad_norm": 15.125, "learning_rate": 2.4777708136108235e-06, "loss": 1.0315, "step": 1505 }, { "epoch": 0.12833404345973584, "grad_norm": 16.125, "learning_rate": 2.4777392536690455e-06, "loss": 0.8453, "step": 1506 }, { "epoch": 0.1284192586280358, "grad_norm": 15.25, "learning_rate": 2.477707671540784e-06, "loss": 1.0709, "step": 1507 }, { "epoch": 0.12850447379633576, "grad_norm": 14.125, "learning_rate": 2.4776760672266102e-06, "loss": 0.4945, "step": 1508 }, { "epoch": 0.1285896889646357, "grad_norm": 12.3125, "learning_rate": 2.477644440727096e-06, "loss": 0.7331, "step": 1509 }, { "epoch": 0.12867490413293567, "grad_norm": 13.125, "learning_rate": 2.477612792042812e-06, "loss": 0.772, "step": 1510 }, { "epoch": 0.12876011930123563, "grad_norm": 13.9375, "learning_rate": 2.47758112117433e-06, "loss": 1.0932, "step": 1511 }, { "epoch": 0.12884533446953558, "grad_norm": 15.0625, "learning_rate": 2.4775494281222236e-06, "loss": 0.5275, "step": 1512 }, { "epoch": 0.12893054963783554, "grad_norm": 16.75, "learning_rate": 2.4775177128870636e-06, "loss": 1.2726, "step": 1513 }, { "epoch": 0.1290157648061355, "grad_norm": 14.5, "learning_rate": 2.477485975469425e-06, "loss": 1.0652, "step": 1514 }, { "epoch": 0.12910097997443545, "grad_norm": 14.8125, "learning_rate": 2.47745421586988e-06, "loss": 0.765, "step": 1515 }, { "epoch": 0.1291861951427354, "grad_norm": 16.125, "learning_rate": 2.477422434089003e-06, "loss": 0.6144, "step": 1516 }, { "epoch": 0.12927141031103537, "grad_norm": 12.125, "learning_rate": 2.4773906301273686e-06, "loss": 0.3885, "step": 1517 }, { "epoch": 0.12935662547933532, "grad_norm": 14.375, "learning_rate": 2.477358803985551e-06, "loss": 0.2456, "step": 1518 }, { "epoch": 0.12944184064763528, "grad_norm": 11.4375, "learning_rate": 2.477326955664126e-06, "loss": 0.4207, "step": 1519 }, { "epoch": 0.12952705581593524, "grad_norm": 12.6875, "learning_rate": 2.4772950851636686e-06, "loss": 0.7109, "step": 1520 }, { "epoch": 0.1296122709842352, "grad_norm": 13.9375, "learning_rate": 2.4772631924847547e-06, "loss": 0.747, "step": 1521 }, { "epoch": 0.12969748615253515, "grad_norm": 18.5, "learning_rate": 2.477231277627961e-06, "loss": 0.8909, "step": 1522 }, { "epoch": 0.1297827013208351, "grad_norm": 13.75, "learning_rate": 2.4771993405938636e-06, "loss": 0.7773, "step": 1523 }, { "epoch": 0.12986791648913507, "grad_norm": 18.25, "learning_rate": 2.4771673813830405e-06, "loss": 0.8698, "step": 1524 }, { "epoch": 0.12995313165743502, "grad_norm": 18.0, "learning_rate": 2.477135399996069e-06, "loss": 0.9722, "step": 1525 }, { "epoch": 0.13003834682573498, "grad_norm": 15.0625, "learning_rate": 2.4771033964335263e-06, "loss": 0.7828, "step": 1526 }, { "epoch": 0.13012356199403494, "grad_norm": 18.875, "learning_rate": 2.4770713706959914e-06, "loss": 0.7726, "step": 1527 }, { "epoch": 0.1302087771623349, "grad_norm": 17.625, "learning_rate": 2.4770393227840432e-06, "loss": 0.7019, "step": 1528 }, { "epoch": 0.13029399233063485, "grad_norm": 16.375, "learning_rate": 2.4770072526982607e-06, "loss": 0.9926, "step": 1529 }, { "epoch": 0.1303792074989348, "grad_norm": 17.75, "learning_rate": 2.476975160439223e-06, "loss": 0.8013, "step": 1530 }, { "epoch": 0.13046442266723476, "grad_norm": 23.125, "learning_rate": 2.47694304600751e-06, "loss": 1.0229, "step": 1531 }, { "epoch": 0.13054963783553472, "grad_norm": 15.125, "learning_rate": 2.4769109094037027e-06, "loss": 0.69, "step": 1532 }, { "epoch": 0.13063485300383468, "grad_norm": 13.625, "learning_rate": 2.4768787506283813e-06, "loss": 0.808, "step": 1533 }, { "epoch": 0.13072006817213463, "grad_norm": 13.25, "learning_rate": 2.4768465696821275e-06, "loss": 0.7263, "step": 1534 }, { "epoch": 0.1308052833404346, "grad_norm": 15.4375, "learning_rate": 2.4768143665655224e-06, "loss": 0.5559, "step": 1535 }, { "epoch": 0.13089049850873455, "grad_norm": 11.3125, "learning_rate": 2.4767821412791482e-06, "loss": 0.5647, "step": 1536 }, { "epoch": 0.1309757136770345, "grad_norm": 15.75, "learning_rate": 2.4767498938235866e-06, "loss": 1.0662, "step": 1537 }, { "epoch": 0.13106092884533446, "grad_norm": 19.875, "learning_rate": 2.476717624199421e-06, "loss": 0.9566, "step": 1538 }, { "epoch": 0.13114614401363442, "grad_norm": 11.125, "learning_rate": 2.476685332407235e-06, "loss": 0.4021, "step": 1539 }, { "epoch": 0.13123135918193438, "grad_norm": 17.625, "learning_rate": 2.4766530184476107e-06, "loss": 1.0331, "step": 1540 }, { "epoch": 0.13131657435023433, "grad_norm": 14.0625, "learning_rate": 2.4766206823211335e-06, "loss": 1.2049, "step": 1541 }, { "epoch": 0.1314017895185343, "grad_norm": 13.4375, "learning_rate": 2.4765883240283868e-06, "loss": 0.5505, "step": 1542 }, { "epoch": 0.13148700468683425, "grad_norm": 12.75, "learning_rate": 2.4765559435699554e-06, "loss": 0.5995, "step": 1543 }, { "epoch": 0.1315722198551342, "grad_norm": 13.3125, "learning_rate": 2.476523540946425e-06, "loss": 0.8021, "step": 1544 }, { "epoch": 0.13165743502343416, "grad_norm": 12.125, "learning_rate": 2.4764911161583806e-06, "loss": 0.5415, "step": 1545 }, { "epoch": 0.13174265019173412, "grad_norm": 17.375, "learning_rate": 2.4764586692064086e-06, "loss": 0.9332, "step": 1546 }, { "epoch": 0.13182786536003407, "grad_norm": 15.875, "learning_rate": 2.4764262000910953e-06, "loss": 0.5249, "step": 1547 }, { "epoch": 0.13191308052833403, "grad_norm": 12.0, "learning_rate": 2.476393708813027e-06, "loss": 0.448, "step": 1548 }, { "epoch": 0.131998295696634, "grad_norm": 13.75, "learning_rate": 2.476361195372791e-06, "loss": 0.4997, "step": 1549 }, { "epoch": 0.13208351086493395, "grad_norm": 20.125, "learning_rate": 2.4763286597709755e-06, "loss": 1.3269, "step": 1550 }, { "epoch": 0.1321687260332339, "grad_norm": 13.25, "learning_rate": 2.476296102008168e-06, "loss": 0.6287, "step": 1551 }, { "epoch": 0.1322539412015339, "grad_norm": 11.125, "learning_rate": 2.476263522084956e-06, "loss": 0.4037, "step": 1552 }, { "epoch": 0.13233915636983384, "grad_norm": 18.375, "learning_rate": 2.4762309200019296e-06, "loss": 0.5243, "step": 1553 }, { "epoch": 0.1324243715381338, "grad_norm": 19.5, "learning_rate": 2.4761982957596776e-06, "loss": 0.793, "step": 1554 }, { "epoch": 0.13250958670643376, "grad_norm": 12.75, "learning_rate": 2.476165649358789e-06, "loss": 0.655, "step": 1555 }, { "epoch": 0.13259480187473371, "grad_norm": 11.375, "learning_rate": 2.4761329807998546e-06, "loss": 0.6106, "step": 1556 }, { "epoch": 0.13268001704303367, "grad_norm": 22.5, "learning_rate": 2.4761002900834635e-06, "loss": 0.905, "step": 1557 }, { "epoch": 0.13276523221133363, "grad_norm": 22.125, "learning_rate": 2.4760675772102077e-06, "loss": 0.7622, "step": 1558 }, { "epoch": 0.13285044737963358, "grad_norm": 11.6875, "learning_rate": 2.476034842180678e-06, "loss": 0.3594, "step": 1559 }, { "epoch": 0.13293566254793354, "grad_norm": 12.375, "learning_rate": 2.4760020849954653e-06, "loss": 0.7484, "step": 1560 }, { "epoch": 0.1330208777162335, "grad_norm": 13.0, "learning_rate": 2.4759693056551626e-06, "loss": 0.5805, "step": 1561 }, { "epoch": 0.13310609288453346, "grad_norm": 12.375, "learning_rate": 2.4759365041603617e-06, "loss": 0.5127, "step": 1562 }, { "epoch": 0.1331913080528334, "grad_norm": 15.6875, "learning_rate": 2.475903680511655e-06, "loss": 0.9681, "step": 1563 }, { "epoch": 0.13327652322113337, "grad_norm": 20.125, "learning_rate": 2.4758708347096363e-06, "loss": 1.1556, "step": 1564 }, { "epoch": 0.13336173838943333, "grad_norm": 14.625, "learning_rate": 2.475837966754899e-06, "loss": 0.7612, "step": 1565 }, { "epoch": 0.13344695355773328, "grad_norm": 17.75, "learning_rate": 2.475805076648037e-06, "loss": 1.1628, "step": 1566 }, { "epoch": 0.13353216872603324, "grad_norm": 17.0, "learning_rate": 2.4757721643896444e-06, "loss": 0.9244, "step": 1567 }, { "epoch": 0.1336173838943332, "grad_norm": 13.8125, "learning_rate": 2.4757392299803163e-06, "loss": 0.6452, "step": 1568 }, { "epoch": 0.13370259906263315, "grad_norm": 12.5625, "learning_rate": 2.4757062734206476e-06, "loss": 0.61, "step": 1569 }, { "epoch": 0.1337878142309331, "grad_norm": 16.75, "learning_rate": 2.4756732947112343e-06, "loss": 1.1258, "step": 1570 }, { "epoch": 0.13387302939923307, "grad_norm": 18.125, "learning_rate": 2.475640293852672e-06, "loss": 0.9646, "step": 1571 }, { "epoch": 0.13395824456753302, "grad_norm": 13.9375, "learning_rate": 2.4756072708455565e-06, "loss": 0.7332, "step": 1572 }, { "epoch": 0.13404345973583298, "grad_norm": 19.125, "learning_rate": 2.4755742256904856e-06, "loss": 1.1055, "step": 1573 }, { "epoch": 0.13412867490413294, "grad_norm": 13.9375, "learning_rate": 2.475541158388056e-06, "loss": 0.9313, "step": 1574 }, { "epoch": 0.1342138900724329, "grad_norm": 15.0, "learning_rate": 2.4755080689388656e-06, "loss": 1.0465, "step": 1575 }, { "epoch": 0.13429910524073285, "grad_norm": 14.4375, "learning_rate": 2.475474957343512e-06, "loss": 0.8469, "step": 1576 }, { "epoch": 0.1343843204090328, "grad_norm": 14.875, "learning_rate": 2.4754418236025935e-06, "loss": 0.6181, "step": 1577 }, { "epoch": 0.13446953557733277, "grad_norm": 15.4375, "learning_rate": 2.4754086677167087e-06, "loss": 0.7345, "step": 1578 }, { "epoch": 0.13455475074563272, "grad_norm": 17.75, "learning_rate": 2.4753754896864575e-06, "loss": 0.9474, "step": 1579 }, { "epoch": 0.13463996591393268, "grad_norm": 14.125, "learning_rate": 2.4753422895124387e-06, "loss": 0.7746, "step": 1580 }, { "epoch": 0.13472518108223264, "grad_norm": 13.4375, "learning_rate": 2.4753090671952523e-06, "loss": 0.6436, "step": 1581 }, { "epoch": 0.1348103962505326, "grad_norm": 10.6875, "learning_rate": 2.475275822735499e-06, "loss": 0.4949, "step": 1582 }, { "epoch": 0.13489561141883255, "grad_norm": 23.75, "learning_rate": 2.47524255613378e-06, "loss": 0.9854, "step": 1583 }, { "epoch": 0.1349808265871325, "grad_norm": 17.25, "learning_rate": 2.4752092673906955e-06, "loss": 1.0981, "step": 1584 }, { "epoch": 0.13506604175543246, "grad_norm": 13.3125, "learning_rate": 2.4751759565068476e-06, "loss": 0.5711, "step": 1585 }, { "epoch": 0.13515125692373242, "grad_norm": 11.9375, "learning_rate": 2.475142623482838e-06, "loss": 0.7141, "step": 1586 }, { "epoch": 0.13523647209203238, "grad_norm": 15.125, "learning_rate": 2.4751092683192696e-06, "loss": 0.6999, "step": 1587 }, { "epoch": 0.13532168726033234, "grad_norm": 11.3125, "learning_rate": 2.4750758910167443e-06, "loss": 0.4597, "step": 1588 }, { "epoch": 0.1354069024286323, "grad_norm": 34.25, "learning_rate": 2.475042491575866e-06, "loss": 1.183, "step": 1589 }, { "epoch": 0.13549211759693225, "grad_norm": 12.0, "learning_rate": 2.4750090699972383e-06, "loss": 0.673, "step": 1590 }, { "epoch": 0.1355773327652322, "grad_norm": 12.6875, "learning_rate": 2.4749756262814643e-06, "loss": 0.5467, "step": 1591 }, { "epoch": 0.13566254793353216, "grad_norm": 13.8125, "learning_rate": 2.474942160429149e-06, "loss": 0.5855, "step": 1592 }, { "epoch": 0.13574776310183212, "grad_norm": 15.0, "learning_rate": 2.4749086724408975e-06, "loss": 0.7791, "step": 1593 }, { "epoch": 0.13583297827013208, "grad_norm": 15.5, "learning_rate": 2.4748751623173143e-06, "loss": 0.8921, "step": 1594 }, { "epoch": 0.13591819343843203, "grad_norm": 15.25, "learning_rate": 2.474841630059005e-06, "loss": 0.661, "step": 1595 }, { "epoch": 0.136003408606732, "grad_norm": 13.875, "learning_rate": 2.474808075666576e-06, "loss": 0.8196, "step": 1596 }, { "epoch": 0.13608862377503195, "grad_norm": 14.8125, "learning_rate": 2.4747744991406337e-06, "loss": 0.9434, "step": 1597 }, { "epoch": 0.1361738389433319, "grad_norm": 13.0, "learning_rate": 2.4747409004817842e-06, "loss": 0.6281, "step": 1598 }, { "epoch": 0.13625905411163186, "grad_norm": 14.25, "learning_rate": 2.474707279690635e-06, "loss": 0.7573, "step": 1599 }, { "epoch": 0.13634426927993182, "grad_norm": 23.0, "learning_rate": 2.474673636767794e-06, "loss": 1.0137, "step": 1600 }, { "epoch": 0.13642948444823177, "grad_norm": 14.0625, "learning_rate": 2.474639971713869e-06, "loss": 0.6528, "step": 1601 }, { "epoch": 0.13651469961653173, "grad_norm": 15.625, "learning_rate": 2.474606284529468e-06, "loss": 1.0082, "step": 1602 }, { "epoch": 0.1365999147848317, "grad_norm": 13.0, "learning_rate": 2.4745725752152e-06, "loss": 0.6146, "step": 1603 }, { "epoch": 0.13668512995313165, "grad_norm": 14.6875, "learning_rate": 2.474538843771674e-06, "loss": 1.1543, "step": 1604 }, { "epoch": 0.1367703451214316, "grad_norm": 18.75, "learning_rate": 2.4745050901994998e-06, "loss": 0.6986, "step": 1605 }, { "epoch": 0.13685556028973156, "grad_norm": 13.0, "learning_rate": 2.4744713144992875e-06, "loss": 0.6663, "step": 1606 }, { "epoch": 0.13694077545803152, "grad_norm": 14.375, "learning_rate": 2.4744375166716473e-06, "loss": 0.9152, "step": 1607 }, { "epoch": 0.1370259906263315, "grad_norm": 16.125, "learning_rate": 2.47440369671719e-06, "loss": 0.894, "step": 1608 }, { "epoch": 0.13711120579463146, "grad_norm": 13.5, "learning_rate": 2.474369854636526e-06, "loss": 0.9786, "step": 1609 }, { "epoch": 0.13719642096293141, "grad_norm": 16.375, "learning_rate": 2.474335990430268e-06, "loss": 0.4418, "step": 1610 }, { "epoch": 0.13728163613123137, "grad_norm": 10.875, "learning_rate": 2.4743021040990272e-06, "loss": 0.5231, "step": 1611 }, { "epoch": 0.13736685129953133, "grad_norm": 11.625, "learning_rate": 2.4742681956434165e-06, "loss": 0.4635, "step": 1612 }, { "epoch": 0.13745206646783129, "grad_norm": 14.6875, "learning_rate": 2.4742342650640485e-06, "loss": 0.649, "step": 1613 }, { "epoch": 0.13753728163613124, "grad_norm": 19.5, "learning_rate": 2.474200312361536e-06, "loss": 0.8201, "step": 1614 }, { "epoch": 0.1376224968044312, "grad_norm": 15.625, "learning_rate": 2.4741663375364934e-06, "loss": 0.7977, "step": 1615 }, { "epoch": 0.13770771197273116, "grad_norm": 16.0, "learning_rate": 2.4741323405895334e-06, "loss": 0.9941, "step": 1616 }, { "epoch": 0.1377929271410311, "grad_norm": 21.5, "learning_rate": 2.474098321521271e-06, "loss": 0.8637, "step": 1617 }, { "epoch": 0.13787814230933107, "grad_norm": 9.8125, "learning_rate": 2.4740642803323217e-06, "loss": 0.3601, "step": 1618 }, { "epoch": 0.13796335747763103, "grad_norm": 12.875, "learning_rate": 2.4740302170232995e-06, "loss": 0.6797, "step": 1619 }, { "epoch": 0.13804857264593098, "grad_norm": 14.125, "learning_rate": 2.4739961315948204e-06, "loss": 0.8079, "step": 1620 }, { "epoch": 0.13813378781423094, "grad_norm": 18.125, "learning_rate": 2.4739620240475006e-06, "loss": 0.9553, "step": 1621 }, { "epoch": 0.1382190029825309, "grad_norm": 20.875, "learning_rate": 2.4739278943819563e-06, "loss": 1.2068, "step": 1622 }, { "epoch": 0.13830421815083085, "grad_norm": 16.0, "learning_rate": 2.4738937425988037e-06, "loss": 0.7125, "step": 1623 }, { "epoch": 0.1383894333191308, "grad_norm": 12.0625, "learning_rate": 2.4738595686986603e-06, "loss": 0.7978, "step": 1624 }, { "epoch": 0.13847464848743077, "grad_norm": 20.875, "learning_rate": 2.473825372682145e-06, "loss": 1.0319, "step": 1625 }, { "epoch": 0.13855986365573073, "grad_norm": 13.875, "learning_rate": 2.4737911545498733e-06, "loss": 0.7193, "step": 1626 }, { "epoch": 0.13864507882403068, "grad_norm": 12.5, "learning_rate": 2.4737569143024654e-06, "loss": 0.5906, "step": 1627 }, { "epoch": 0.13873029399233064, "grad_norm": 19.0, "learning_rate": 2.4737226519405393e-06, "loss": 0.8338, "step": 1628 }, { "epoch": 0.1388155091606306, "grad_norm": 18.5, "learning_rate": 2.473688367464715e-06, "loss": 1.0743, "step": 1629 }, { "epoch": 0.13890072432893055, "grad_norm": 20.375, "learning_rate": 2.4736540608756105e-06, "loss": 0.6711, "step": 1630 }, { "epoch": 0.1389859394972305, "grad_norm": 27.125, "learning_rate": 2.473619732173847e-06, "loss": 1.1368, "step": 1631 }, { "epoch": 0.13907115466553047, "grad_norm": 18.125, "learning_rate": 2.473585381360044e-06, "loss": 1.0869, "step": 1632 }, { "epoch": 0.13915636983383042, "grad_norm": 17.5, "learning_rate": 2.4735510084348236e-06, "loss": 0.8473, "step": 1633 }, { "epoch": 0.13924158500213038, "grad_norm": 12.4375, "learning_rate": 2.4735166133988057e-06, "loss": 0.5834, "step": 1634 }, { "epoch": 0.13932680017043034, "grad_norm": 17.125, "learning_rate": 2.4734821962526122e-06, "loss": 0.8453, "step": 1635 }, { "epoch": 0.1394120153387303, "grad_norm": 22.875, "learning_rate": 2.473447756996865e-06, "loss": 0.7648, "step": 1636 }, { "epoch": 0.13949723050703025, "grad_norm": 11.6875, "learning_rate": 2.4734132956321872e-06, "loss": 0.4743, "step": 1637 }, { "epoch": 0.1395824456753302, "grad_norm": 15.5625, "learning_rate": 2.4733788121592004e-06, "loss": 1.1727, "step": 1638 }, { "epoch": 0.13966766084363016, "grad_norm": 16.0, "learning_rate": 2.4733443065785283e-06, "loss": 1.2217, "step": 1639 }, { "epoch": 0.13975287601193012, "grad_norm": 16.25, "learning_rate": 2.4733097788907947e-06, "loss": 0.8836, "step": 1640 }, { "epoch": 0.13983809118023008, "grad_norm": 20.125, "learning_rate": 2.4732752290966227e-06, "loss": 1.039, "step": 1641 }, { "epoch": 0.13992330634853004, "grad_norm": 20.125, "learning_rate": 2.4732406571966373e-06, "loss": 0.864, "step": 1642 }, { "epoch": 0.14000852151683, "grad_norm": 18.25, "learning_rate": 2.4732060631914633e-06, "loss": 0.8466, "step": 1643 }, { "epoch": 0.14009373668512995, "grad_norm": 17.5, "learning_rate": 2.4731714470817257e-06, "loss": 0.9993, "step": 1644 }, { "epoch": 0.1401789518534299, "grad_norm": 19.625, "learning_rate": 2.47313680886805e-06, "loss": 0.9561, "step": 1645 }, { "epoch": 0.14026416702172986, "grad_norm": 18.25, "learning_rate": 2.4731021485510625e-06, "loss": 0.8679, "step": 1646 }, { "epoch": 0.14034938219002982, "grad_norm": 11.375, "learning_rate": 2.4730674661313892e-06, "loss": 0.5669, "step": 1647 }, { "epoch": 0.14043459735832978, "grad_norm": 16.5, "learning_rate": 2.4730327616096566e-06, "loss": 1.0393, "step": 1648 }, { "epoch": 0.14051981252662973, "grad_norm": 19.5, "learning_rate": 2.472998034986493e-06, "loss": 0.9216, "step": 1649 }, { "epoch": 0.1406050276949297, "grad_norm": 16.5, "learning_rate": 2.472963286262524e-06, "loss": 0.9496, "step": 1650 }, { "epoch": 0.14069024286322965, "grad_norm": 15.5, "learning_rate": 2.4729285154383792e-06, "loss": 0.8578, "step": 1651 }, { "epoch": 0.1407754580315296, "grad_norm": 11.875, "learning_rate": 2.472893722514686e-06, "loss": 0.4865, "step": 1652 }, { "epoch": 0.14086067319982956, "grad_norm": 16.375, "learning_rate": 2.472858907492074e-06, "loss": 0.9434, "step": 1653 }, { "epoch": 0.14094588836812952, "grad_norm": 11.75, "learning_rate": 2.4728240703711712e-06, "loss": 0.5191, "step": 1654 }, { "epoch": 0.14103110353642948, "grad_norm": 15.125, "learning_rate": 2.4727892111526084e-06, "loss": 0.8929, "step": 1655 }, { "epoch": 0.14111631870472943, "grad_norm": 11.0625, "learning_rate": 2.4727543298370148e-06, "loss": 0.4518, "step": 1656 }, { "epoch": 0.1412015338730294, "grad_norm": 12.4375, "learning_rate": 2.472719426425021e-06, "loss": 0.6647, "step": 1657 }, { "epoch": 0.14128674904132935, "grad_norm": 17.0, "learning_rate": 2.4726845009172572e-06, "loss": 0.847, "step": 1658 }, { "epoch": 0.1413719642096293, "grad_norm": 15.4375, "learning_rate": 2.472649553314355e-06, "loss": 1.0141, "step": 1659 }, { "epoch": 0.14145717937792926, "grad_norm": 19.25, "learning_rate": 2.472614583616946e-06, "loss": 0.6686, "step": 1660 }, { "epoch": 0.14154239454622922, "grad_norm": 12.5625, "learning_rate": 2.4725795918256628e-06, "loss": 0.7507, "step": 1661 }, { "epoch": 0.14162760971452917, "grad_norm": 17.0, "learning_rate": 2.472544577941136e-06, "loss": 0.6057, "step": 1662 }, { "epoch": 0.14171282488282913, "grad_norm": 12.4375, "learning_rate": 2.472509541964e-06, "loss": 0.5135, "step": 1663 }, { "epoch": 0.1417980400511291, "grad_norm": 16.25, "learning_rate": 2.472474483894887e-06, "loss": 0.7836, "step": 1664 }, { "epoch": 0.14188325521942907, "grad_norm": 12.375, "learning_rate": 2.4724394037344302e-06, "loss": 0.4222, "step": 1665 }, { "epoch": 0.14196847038772903, "grad_norm": 11.125, "learning_rate": 2.472404301483265e-06, "loss": 0.4369, "step": 1666 }, { "epoch": 0.14205368555602899, "grad_norm": 14.3125, "learning_rate": 2.4723691771420243e-06, "loss": 0.8266, "step": 1667 }, { "epoch": 0.14213890072432894, "grad_norm": 18.5, "learning_rate": 2.472334030711344e-06, "loss": 0.9728, "step": 1668 }, { "epoch": 0.1422241158926289, "grad_norm": 13.8125, "learning_rate": 2.4722988621918575e-06, "loss": 0.6679, "step": 1669 }, { "epoch": 0.14230933106092886, "grad_norm": 42.25, "learning_rate": 2.472263671584202e-06, "loss": 1.1168, "step": 1670 }, { "epoch": 0.1423945462292288, "grad_norm": 14.5, "learning_rate": 2.4722284588890135e-06, "loss": 0.603, "step": 1671 }, { "epoch": 0.14247976139752877, "grad_norm": 14.0, "learning_rate": 2.4721932241069267e-06, "loss": 0.8431, "step": 1672 }, { "epoch": 0.14256497656582873, "grad_norm": 18.25, "learning_rate": 2.47215796723858e-06, "loss": 1.4129, "step": 1673 }, { "epoch": 0.14265019173412868, "grad_norm": 19.375, "learning_rate": 2.4721226882846096e-06, "loss": 1.0845, "step": 1674 }, { "epoch": 0.14273540690242864, "grad_norm": 39.0, "learning_rate": 2.4720873872456532e-06, "loss": 1.4384, "step": 1675 }, { "epoch": 0.1428206220707286, "grad_norm": 16.0, "learning_rate": 2.472052064122349e-06, "loss": 0.4952, "step": 1676 }, { "epoch": 0.14290583723902855, "grad_norm": 10.5, "learning_rate": 2.4720167189153352e-06, "loss": 0.4848, "step": 1677 }, { "epoch": 0.1429910524073285, "grad_norm": 16.25, "learning_rate": 2.47198135162525e-06, "loss": 0.5999, "step": 1678 }, { "epoch": 0.14307626757562847, "grad_norm": 18.875, "learning_rate": 2.4719459622527335e-06, "loss": 1.2016, "step": 1679 }, { "epoch": 0.14316148274392843, "grad_norm": 18.625, "learning_rate": 2.4719105507984246e-06, "loss": 0.6198, "step": 1680 }, { "epoch": 0.14324669791222838, "grad_norm": 12.875, "learning_rate": 2.4718751172629633e-06, "loss": 0.6453, "step": 1681 }, { "epoch": 0.14333191308052834, "grad_norm": 12.1875, "learning_rate": 2.4718396616469896e-06, "loss": 0.7325, "step": 1682 }, { "epoch": 0.1434171282488283, "grad_norm": 13.8125, "learning_rate": 2.471804183951145e-06, "loss": 0.6485, "step": 1683 }, { "epoch": 0.14350234341712825, "grad_norm": 13.0625, "learning_rate": 2.4717686841760703e-06, "loss": 0.7609, "step": 1684 }, { "epoch": 0.1435875585854282, "grad_norm": 16.0, "learning_rate": 2.4717331623224064e-06, "loss": 0.4967, "step": 1685 }, { "epoch": 0.14367277375372817, "grad_norm": 12.6875, "learning_rate": 2.471697618390796e-06, "loss": 0.6717, "step": 1686 }, { "epoch": 0.14375798892202812, "grad_norm": 12.1875, "learning_rate": 2.4716620523818805e-06, "loss": 0.4425, "step": 1687 }, { "epoch": 0.14384320409032808, "grad_norm": 16.625, "learning_rate": 2.471626464296304e-06, "loss": 0.8764, "step": 1688 }, { "epoch": 0.14392841925862804, "grad_norm": 18.0, "learning_rate": 2.471590854134709e-06, "loss": 1.065, "step": 1689 }, { "epoch": 0.144013634426928, "grad_norm": 15.0, "learning_rate": 2.4715552218977383e-06, "loss": 0.647, "step": 1690 }, { "epoch": 0.14409884959522795, "grad_norm": 12.25, "learning_rate": 2.4715195675860366e-06, "loss": 0.7388, "step": 1691 }, { "epoch": 0.1441840647635279, "grad_norm": 17.625, "learning_rate": 2.471483891200248e-06, "loss": 0.9272, "step": 1692 }, { "epoch": 0.14426927993182787, "grad_norm": 13.0, "learning_rate": 2.471448192741017e-06, "loss": 0.6184, "step": 1693 }, { "epoch": 0.14435449510012782, "grad_norm": 13.4375, "learning_rate": 2.471412472208989e-06, "loss": 0.7274, "step": 1694 }, { "epoch": 0.14443971026842778, "grad_norm": 16.0, "learning_rate": 2.47137672960481e-06, "loss": 0.5611, "step": 1695 }, { "epoch": 0.14452492543672774, "grad_norm": 16.75, "learning_rate": 2.4713409649291243e-06, "loss": 1.1467, "step": 1696 }, { "epoch": 0.1446101406050277, "grad_norm": 15.625, "learning_rate": 2.47130517818258e-06, "loss": 0.7028, "step": 1697 }, { "epoch": 0.14469535577332765, "grad_norm": 18.0, "learning_rate": 2.4712693693658222e-06, "loss": 0.9602, "step": 1698 }, { "epoch": 0.1447805709416276, "grad_norm": 14.625, "learning_rate": 2.4712335384794995e-06, "loss": 0.6669, "step": 1699 }, { "epoch": 0.14486578610992756, "grad_norm": 16.375, "learning_rate": 2.471197685524258e-06, "loss": 0.9796, "step": 1700 }, { "epoch": 0.14495100127822752, "grad_norm": 18.875, "learning_rate": 2.471161810500747e-06, "loss": 1.2473, "step": 1701 }, { "epoch": 0.14503621644652748, "grad_norm": 17.0, "learning_rate": 2.471125913409614e-06, "loss": 0.9406, "step": 1702 }, { "epoch": 0.14512143161482743, "grad_norm": 9.9375, "learning_rate": 2.4710899942515077e-06, "loss": 0.2845, "step": 1703 }, { "epoch": 0.1452066467831274, "grad_norm": 16.875, "learning_rate": 2.471054053027077e-06, "loss": 0.9779, "step": 1704 }, { "epoch": 0.14529186195142735, "grad_norm": 12.875, "learning_rate": 2.4710180897369725e-06, "loss": 0.6043, "step": 1705 }, { "epoch": 0.1453770771197273, "grad_norm": 15.8125, "learning_rate": 2.470982104381843e-06, "loss": 1.1665, "step": 1706 }, { "epoch": 0.14546229228802726, "grad_norm": 16.25, "learning_rate": 2.4709460969623385e-06, "loss": 0.7901, "step": 1707 }, { "epoch": 0.14554750745632722, "grad_norm": 19.125, "learning_rate": 2.4709100674791108e-06, "loss": 1.1028, "step": 1708 }, { "epoch": 0.14563272262462718, "grad_norm": 11.75, "learning_rate": 2.4708740159328105e-06, "loss": 0.591, "step": 1709 }, { "epoch": 0.14571793779292713, "grad_norm": 17.5, "learning_rate": 2.4708379423240887e-06, "loss": 0.8908, "step": 1710 }, { "epoch": 0.1458031529612271, "grad_norm": 16.625, "learning_rate": 2.4708018466535976e-06, "loss": 0.9323, "step": 1711 }, { "epoch": 0.14588836812952705, "grad_norm": 13.625, "learning_rate": 2.4707657289219898e-06, "loss": 0.6885, "step": 1712 }, { "epoch": 0.145973583297827, "grad_norm": 19.0, "learning_rate": 2.4707295891299175e-06, "loss": 0.7523, "step": 1713 }, { "epoch": 0.14605879846612696, "grad_norm": 14.5, "learning_rate": 2.4706934272780342e-06, "loss": 0.7756, "step": 1714 }, { "epoch": 0.14614401363442692, "grad_norm": 13.75, "learning_rate": 2.470657243366993e-06, "loss": 0.759, "step": 1715 }, { "epoch": 0.14622922880272687, "grad_norm": 13.75, "learning_rate": 2.4706210373974483e-06, "loss": 0.834, "step": 1716 }, { "epoch": 0.14631444397102683, "grad_norm": 14.75, "learning_rate": 2.4705848093700534e-06, "loss": 0.9064, "step": 1717 }, { "epoch": 0.1463996591393268, "grad_norm": 17.125, "learning_rate": 2.470548559285464e-06, "loss": 1.0722, "step": 1718 }, { "epoch": 0.14648487430762674, "grad_norm": 16.25, "learning_rate": 2.470512287144335e-06, "loss": 0.8775, "step": 1719 }, { "epoch": 0.1465700894759267, "grad_norm": 33.0, "learning_rate": 2.4704759929473213e-06, "loss": 1.0421, "step": 1720 }, { "epoch": 0.14665530464422669, "grad_norm": 11.125, "learning_rate": 2.470439676695079e-06, "loss": 0.4325, "step": 1721 }, { "epoch": 0.14674051981252664, "grad_norm": 13.625, "learning_rate": 2.470403338388265e-06, "loss": 0.5844, "step": 1722 }, { "epoch": 0.1468257349808266, "grad_norm": 11.1875, "learning_rate": 2.4703669780275353e-06, "loss": 0.4866, "step": 1723 }, { "epoch": 0.14691095014912656, "grad_norm": 19.75, "learning_rate": 2.4703305956135466e-06, "loss": 0.9827, "step": 1724 }, { "epoch": 0.1469961653174265, "grad_norm": 18.5, "learning_rate": 2.4702941911469573e-06, "loss": 1.1083, "step": 1725 }, { "epoch": 0.14708138048572647, "grad_norm": 17.625, "learning_rate": 2.470257764628425e-06, "loss": 1.0822, "step": 1726 }, { "epoch": 0.14716659565402643, "grad_norm": 16.125, "learning_rate": 2.470221316058608e-06, "loss": 0.3616, "step": 1727 }, { "epoch": 0.14725181082232638, "grad_norm": 11.8125, "learning_rate": 2.4701848454381643e-06, "loss": 0.4721, "step": 1728 }, { "epoch": 0.14733702599062634, "grad_norm": 14.5, "learning_rate": 2.4701483527677536e-06, "loss": 0.7631, "step": 1729 }, { "epoch": 0.1474222411589263, "grad_norm": 15.0, "learning_rate": 2.470111838048035e-06, "loss": 0.7506, "step": 1730 }, { "epoch": 0.14750745632722626, "grad_norm": 21.25, "learning_rate": 2.470075301279669e-06, "loss": 0.8739, "step": 1731 }, { "epoch": 0.1475926714955262, "grad_norm": 15.625, "learning_rate": 2.4700387424633154e-06, "loss": 0.8766, "step": 1732 }, { "epoch": 0.14767788666382617, "grad_norm": 17.25, "learning_rate": 2.4700021615996343e-06, "loss": 1.1099, "step": 1733 }, { "epoch": 0.14776310183212613, "grad_norm": 30.25, "learning_rate": 2.4699655586892877e-06, "loss": 1.156, "step": 1734 }, { "epoch": 0.14784831700042608, "grad_norm": 24.625, "learning_rate": 2.4699289337329373e-06, "loss": 0.9198, "step": 1735 }, { "epoch": 0.14793353216872604, "grad_norm": 11.1875, "learning_rate": 2.4698922867312436e-06, "loss": 0.342, "step": 1736 }, { "epoch": 0.148018747337026, "grad_norm": 12.6875, "learning_rate": 2.46985561768487e-06, "loss": 0.7033, "step": 1737 }, { "epoch": 0.14810396250532595, "grad_norm": 20.625, "learning_rate": 2.469818926594478e-06, "loss": 1.0989, "step": 1738 }, { "epoch": 0.1481891776736259, "grad_norm": 16.375, "learning_rate": 2.469782213460732e-06, "loss": 0.9749, "step": 1739 }, { "epoch": 0.14827439284192587, "grad_norm": 9.75, "learning_rate": 2.4697454782842946e-06, "loss": 0.4032, "step": 1740 }, { "epoch": 0.14835960801022582, "grad_norm": 12.0625, "learning_rate": 2.4697087210658297e-06, "loss": 0.4214, "step": 1741 }, { "epoch": 0.14844482317852578, "grad_norm": 11.5625, "learning_rate": 2.4696719418060023e-06, "loss": 0.5647, "step": 1742 }, { "epoch": 0.14853003834682574, "grad_norm": 14.1875, "learning_rate": 2.4696351405054763e-06, "loss": 0.9841, "step": 1743 }, { "epoch": 0.1486152535151257, "grad_norm": 15.1875, "learning_rate": 2.4695983171649162e-06, "loss": 0.804, "step": 1744 }, { "epoch": 0.14870046868342565, "grad_norm": 13.8125, "learning_rate": 2.4695614717849888e-06, "loss": 0.9698, "step": 1745 }, { "epoch": 0.1487856838517256, "grad_norm": 16.875, "learning_rate": 2.469524604366359e-06, "loss": 0.7278, "step": 1746 }, { "epoch": 0.14887089902002557, "grad_norm": 12.6875, "learning_rate": 2.4694877149096934e-06, "loss": 0.5082, "step": 1747 }, { "epoch": 0.14895611418832552, "grad_norm": 18.625, "learning_rate": 2.4694508034156583e-06, "loss": 0.9462, "step": 1748 }, { "epoch": 0.14904132935662548, "grad_norm": 11.75, "learning_rate": 2.4694138698849206e-06, "loss": 0.4739, "step": 1749 }, { "epoch": 0.14912654452492544, "grad_norm": 13.8125, "learning_rate": 2.469376914318149e-06, "loss": 0.5964, "step": 1750 }, { "epoch": 0.1492117596932254, "grad_norm": 20.25, "learning_rate": 2.4693399367160095e-06, "loss": 0.9735, "step": 1751 }, { "epoch": 0.14929697486152535, "grad_norm": 16.25, "learning_rate": 2.4693029370791715e-06, "loss": 1.0902, "step": 1752 }, { "epoch": 0.1493821900298253, "grad_norm": 11.8125, "learning_rate": 2.4692659154083033e-06, "loss": 0.72, "step": 1753 }, { "epoch": 0.14946740519812526, "grad_norm": 13.0, "learning_rate": 2.469228871704074e-06, "loss": 0.6453, "step": 1754 }, { "epoch": 0.14955262036642522, "grad_norm": 11.5625, "learning_rate": 2.469191805967153e-06, "loss": 0.5758, "step": 1755 }, { "epoch": 0.14963783553472518, "grad_norm": 12.1875, "learning_rate": 2.4691547181982102e-06, "loss": 0.5256, "step": 1756 }, { "epoch": 0.14972305070302513, "grad_norm": 16.75, "learning_rate": 2.4691176083979155e-06, "loss": 0.9599, "step": 1757 }, { "epoch": 0.1498082658713251, "grad_norm": 19.25, "learning_rate": 2.4690804765669396e-06, "loss": 0.9446, "step": 1758 }, { "epoch": 0.14989348103962505, "grad_norm": 19.0, "learning_rate": 2.469043322705954e-06, "loss": 1.1701, "step": 1759 }, { "epoch": 0.149978696207925, "grad_norm": 13.9375, "learning_rate": 2.4690061468156293e-06, "loss": 0.7094, "step": 1760 }, { "epoch": 0.15006391137622496, "grad_norm": 10.0625, "learning_rate": 2.4689689488966377e-06, "loss": 0.292, "step": 1761 }, { "epoch": 0.15014912654452492, "grad_norm": 11.8125, "learning_rate": 2.468931728949652e-06, "loss": 0.5294, "step": 1762 }, { "epoch": 0.15023434171282488, "grad_norm": 14.6875, "learning_rate": 2.4688944869753435e-06, "loss": 0.9632, "step": 1763 }, { "epoch": 0.15031955688112483, "grad_norm": 12.1875, "learning_rate": 2.4688572229743863e-06, "loss": 0.3175, "step": 1764 }, { "epoch": 0.1504047720494248, "grad_norm": 30.125, "learning_rate": 2.4688199369474532e-06, "loss": 1.283, "step": 1765 }, { "epoch": 0.15048998721772475, "grad_norm": 15.375, "learning_rate": 2.4687826288952186e-06, "loss": 0.9624, "step": 1766 }, { "epoch": 0.1505752023860247, "grad_norm": 16.5, "learning_rate": 2.468745298818356e-06, "loss": 0.8091, "step": 1767 }, { "epoch": 0.15066041755432466, "grad_norm": 15.25, "learning_rate": 2.4687079467175402e-06, "loss": 0.6241, "step": 1768 }, { "epoch": 0.15074563272262462, "grad_norm": 13.0625, "learning_rate": 2.4686705725934467e-06, "loss": 0.6405, "step": 1769 }, { "epoch": 0.15083084789092457, "grad_norm": 15.0, "learning_rate": 2.46863317644675e-06, "loss": 0.8592, "step": 1770 }, { "epoch": 0.15091606305922453, "grad_norm": 19.125, "learning_rate": 2.468595758278127e-06, "loss": 0.775, "step": 1771 }, { "epoch": 0.1510012782275245, "grad_norm": 10.6875, "learning_rate": 2.468558318088253e-06, "loss": 0.4237, "step": 1772 }, { "epoch": 0.15108649339582444, "grad_norm": 26.375, "learning_rate": 2.4685208558778045e-06, "loss": 0.6044, "step": 1773 }, { "epoch": 0.1511717085641244, "grad_norm": 22.125, "learning_rate": 2.4684833716474594e-06, "loss": 1.1508, "step": 1774 }, { "epoch": 0.15125692373242436, "grad_norm": 13.5625, "learning_rate": 2.468445865397894e-06, "loss": 0.6571, "step": 1775 }, { "epoch": 0.15134213890072432, "grad_norm": 11.75, "learning_rate": 2.4684083371297866e-06, "loss": 0.4636, "step": 1776 }, { "epoch": 0.15142735406902427, "grad_norm": 14.0625, "learning_rate": 2.4683707868438156e-06, "loss": 0.8693, "step": 1777 }, { "epoch": 0.15151256923732426, "grad_norm": 13.4375, "learning_rate": 2.468333214540659e-06, "loss": 0.7335, "step": 1778 }, { "epoch": 0.15159778440562421, "grad_norm": 20.125, "learning_rate": 2.4682956202209963e-06, "loss": 0.9285, "step": 1779 }, { "epoch": 0.15168299957392417, "grad_norm": 16.125, "learning_rate": 2.468258003885507e-06, "loss": 1.2837, "step": 1780 }, { "epoch": 0.15176821474222413, "grad_norm": 37.0, "learning_rate": 2.46822036553487e-06, "loss": 1.6967, "step": 1781 }, { "epoch": 0.15185342991052408, "grad_norm": 14.75, "learning_rate": 2.468182705169766e-06, "loss": 1.1296, "step": 1782 }, { "epoch": 0.15193864507882404, "grad_norm": 15.0, "learning_rate": 2.468145022790876e-06, "loss": 0.8443, "step": 1783 }, { "epoch": 0.152023860247124, "grad_norm": 18.375, "learning_rate": 2.46810731839888e-06, "loss": 1.1151, "step": 1784 }, { "epoch": 0.15210907541542396, "grad_norm": 13.5625, "learning_rate": 2.46806959199446e-06, "loss": 0.3695, "step": 1785 }, { "epoch": 0.1521942905837239, "grad_norm": 18.625, "learning_rate": 2.4680318435782977e-06, "loss": 0.9289, "step": 1786 }, { "epoch": 0.15227950575202387, "grad_norm": 13.6875, "learning_rate": 2.467994073151075e-06, "loss": 0.7206, "step": 1787 }, { "epoch": 0.15236472092032383, "grad_norm": 13.375, "learning_rate": 2.4679562807134746e-06, "loss": 0.5083, "step": 1788 }, { "epoch": 0.15244993608862378, "grad_norm": 19.75, "learning_rate": 2.4679184662661797e-06, "loss": 1.1709, "step": 1789 }, { "epoch": 0.15253515125692374, "grad_norm": 19.625, "learning_rate": 2.4678806298098732e-06, "loss": 1.1609, "step": 1790 }, { "epoch": 0.1526203664252237, "grad_norm": 11.875, "learning_rate": 2.467842771345239e-06, "loss": 0.7869, "step": 1791 }, { "epoch": 0.15270558159352365, "grad_norm": 12.3125, "learning_rate": 2.4678048908729615e-06, "loss": 0.7349, "step": 1792 }, { "epoch": 0.1527907967618236, "grad_norm": 15.5625, "learning_rate": 2.467766988393725e-06, "loss": 0.8078, "step": 1793 }, { "epoch": 0.15287601193012357, "grad_norm": 16.875, "learning_rate": 2.4677290639082143e-06, "loss": 0.7204, "step": 1794 }, { "epoch": 0.15296122709842352, "grad_norm": 13.125, "learning_rate": 2.467691117417115e-06, "loss": 0.6347, "step": 1795 }, { "epoch": 0.15304644226672348, "grad_norm": 15.9375, "learning_rate": 2.4676531489211127e-06, "loss": 0.9882, "step": 1796 }, { "epoch": 0.15313165743502344, "grad_norm": 13.75, "learning_rate": 2.4676151584208937e-06, "loss": 0.638, "step": 1797 }, { "epoch": 0.1532168726033234, "grad_norm": 17.25, "learning_rate": 2.4675771459171443e-06, "loss": 0.7384, "step": 1798 }, { "epoch": 0.15330208777162335, "grad_norm": 11.75, "learning_rate": 2.467539111410551e-06, "loss": 0.4922, "step": 1799 }, { "epoch": 0.1533873029399233, "grad_norm": 19.75, "learning_rate": 2.4675010549018024e-06, "loss": 1.1412, "step": 1800 }, { "epoch": 0.15347251810822327, "grad_norm": 11.375, "learning_rate": 2.467462976391585e-06, "loss": 0.4821, "step": 1801 }, { "epoch": 0.15355773327652322, "grad_norm": 22.625, "learning_rate": 2.4674248758805873e-06, "loss": 1.1391, "step": 1802 }, { "epoch": 0.15364294844482318, "grad_norm": 15.25, "learning_rate": 2.4673867533694982e-06, "loss": 1.0296, "step": 1803 }, { "epoch": 0.15372816361312314, "grad_norm": 15.6875, "learning_rate": 2.4673486088590064e-06, "loss": 0.8798, "step": 1804 }, { "epoch": 0.1538133787814231, "grad_norm": 15.9375, "learning_rate": 2.4673104423498005e-06, "loss": 1.0988, "step": 1805 }, { "epoch": 0.15389859394972305, "grad_norm": 16.625, "learning_rate": 2.4672722538425713e-06, "loss": 1.1313, "step": 1806 }, { "epoch": 0.153983809118023, "grad_norm": 13.75, "learning_rate": 2.4672340433380083e-06, "loss": 0.7441, "step": 1807 }, { "epoch": 0.15406902428632296, "grad_norm": 20.375, "learning_rate": 2.467195810836802e-06, "loss": 0.914, "step": 1808 }, { "epoch": 0.15415423945462292, "grad_norm": 12.0625, "learning_rate": 2.4671575563396437e-06, "loss": 0.3241, "step": 1809 }, { "epoch": 0.15423945462292288, "grad_norm": 21.875, "learning_rate": 2.467119279847224e-06, "loss": 1.3322, "step": 1810 }, { "epoch": 0.15432466979122283, "grad_norm": 9.8125, "learning_rate": 2.4670809813602355e-06, "loss": 0.3766, "step": 1811 }, { "epoch": 0.1544098849595228, "grad_norm": 14.9375, "learning_rate": 2.467042660879369e-06, "loss": 0.6851, "step": 1812 }, { "epoch": 0.15449510012782275, "grad_norm": 14.375, "learning_rate": 2.4670043184053184e-06, "loss": 0.9174, "step": 1813 }, { "epoch": 0.1545803152961227, "grad_norm": 18.875, "learning_rate": 2.466965953938776e-06, "loss": 1.0705, "step": 1814 }, { "epoch": 0.15466553046442266, "grad_norm": 20.25, "learning_rate": 2.4669275674804347e-06, "loss": 1.5079, "step": 1815 }, { "epoch": 0.15475074563272262, "grad_norm": 19.375, "learning_rate": 2.466889159030989e-06, "loss": 0.9318, "step": 1816 }, { "epoch": 0.15483596080102258, "grad_norm": 18.625, "learning_rate": 2.466850728591132e-06, "loss": 0.8093, "step": 1817 }, { "epoch": 0.15492117596932253, "grad_norm": 12.125, "learning_rate": 2.466812276161559e-06, "loss": 0.4672, "step": 1818 }, { "epoch": 0.1550063911376225, "grad_norm": 14.3125, "learning_rate": 2.4667738017429648e-06, "loss": 0.6757, "step": 1819 }, { "epoch": 0.15509160630592245, "grad_norm": 11.375, "learning_rate": 2.466735305336044e-06, "loss": 0.5264, "step": 1820 }, { "epoch": 0.1551768214742224, "grad_norm": 14.3125, "learning_rate": 2.4666967869414932e-06, "loss": 0.7891, "step": 1821 }, { "epoch": 0.15526203664252236, "grad_norm": 21.125, "learning_rate": 2.4666582465600073e-06, "loss": 0.8732, "step": 1822 }, { "epoch": 0.15534725181082232, "grad_norm": 12.1875, "learning_rate": 2.466619684192284e-06, "loss": 0.5744, "step": 1823 }, { "epoch": 0.15543246697912227, "grad_norm": 11.625, "learning_rate": 2.4665810998390196e-06, "loss": 0.5755, "step": 1824 }, { "epoch": 0.15551768214742223, "grad_norm": 13.8125, "learning_rate": 2.4665424935009106e-06, "loss": 0.746, "step": 1825 }, { "epoch": 0.1556028973157222, "grad_norm": 15.125, "learning_rate": 2.466503865178656e-06, "loss": 0.7084, "step": 1826 }, { "epoch": 0.15568811248402215, "grad_norm": 16.25, "learning_rate": 2.4664652148729533e-06, "loss": 0.6737, "step": 1827 }, { "epoch": 0.1557733276523221, "grad_norm": 12.0625, "learning_rate": 2.466426542584501e-06, "loss": 0.6415, "step": 1828 }, { "epoch": 0.15585854282062206, "grad_norm": 12.125, "learning_rate": 2.466387848313998e-06, "loss": 0.8392, "step": 1829 }, { "epoch": 0.15594375798892202, "grad_norm": 14.8125, "learning_rate": 2.466349132062143e-06, "loss": 0.7633, "step": 1830 }, { "epoch": 0.15602897315722197, "grad_norm": 22.875, "learning_rate": 2.466310393829636e-06, "loss": 1.0498, "step": 1831 }, { "epoch": 0.15611418832552193, "grad_norm": 12.0625, "learning_rate": 2.4662716336171778e-06, "loss": 0.4587, "step": 1832 }, { "epoch": 0.1561994034938219, "grad_norm": 15.6875, "learning_rate": 2.4662328514254672e-06, "loss": 0.6356, "step": 1833 }, { "epoch": 0.15628461866212187, "grad_norm": 17.5, "learning_rate": 2.4661940472552067e-06, "loss": 1.142, "step": 1834 }, { "epoch": 0.15636983383042183, "grad_norm": 13.25, "learning_rate": 2.466155221107096e-06, "loss": 0.7789, "step": 1835 }, { "epoch": 0.15645504899872179, "grad_norm": 16.125, "learning_rate": 2.4661163729818383e-06, "loss": 0.9163, "step": 1836 }, { "epoch": 0.15654026416702174, "grad_norm": 15.1875, "learning_rate": 2.4660775028801345e-06, "loss": 0.5911, "step": 1837 }, { "epoch": 0.1566254793353217, "grad_norm": 14.5, "learning_rate": 2.4660386108026872e-06, "loss": 0.8785, "step": 1838 }, { "epoch": 0.15671069450362166, "grad_norm": 23.75, "learning_rate": 2.4659996967501997e-06, "loss": 0.7775, "step": 1839 }, { "epoch": 0.1567959096719216, "grad_norm": 12.0, "learning_rate": 2.4659607607233748e-06, "loss": 0.4976, "step": 1840 }, { "epoch": 0.15688112484022157, "grad_norm": 19.5, "learning_rate": 2.465921802722916e-06, "loss": 0.9605, "step": 1841 }, { "epoch": 0.15696634000852153, "grad_norm": 16.125, "learning_rate": 2.465882822749528e-06, "loss": 0.8789, "step": 1842 }, { "epoch": 0.15705155517682148, "grad_norm": 15.4375, "learning_rate": 2.465843820803914e-06, "loss": 1.0392, "step": 1843 }, { "epoch": 0.15713677034512144, "grad_norm": 13.75, "learning_rate": 2.46580479688678e-06, "loss": 0.5005, "step": 1844 }, { "epoch": 0.1572219855134214, "grad_norm": 20.25, "learning_rate": 2.465765750998831e-06, "loss": 1.0982, "step": 1845 }, { "epoch": 0.15730720068172135, "grad_norm": 13.125, "learning_rate": 2.4657266831407717e-06, "loss": 0.7649, "step": 1846 }, { "epoch": 0.1573924158500213, "grad_norm": 30.625, "learning_rate": 2.465687593313309e-06, "loss": 0.9712, "step": 1847 }, { "epoch": 0.15747763101832127, "grad_norm": 11.9375, "learning_rate": 2.4656484815171494e-06, "loss": 0.7227, "step": 1848 }, { "epoch": 0.15756284618662122, "grad_norm": 13.625, "learning_rate": 2.4656093477529986e-06, "loss": 0.7647, "step": 1849 }, { "epoch": 0.15764806135492118, "grad_norm": 13.6875, "learning_rate": 2.465570192021565e-06, "loss": 0.6293, "step": 1850 }, { "epoch": 0.15773327652322114, "grad_norm": 17.75, "learning_rate": 2.4655310143235556e-06, "loss": 0.6958, "step": 1851 }, { "epoch": 0.1578184916915211, "grad_norm": 27.0, "learning_rate": 2.4654918146596783e-06, "loss": 0.9169, "step": 1852 }, { "epoch": 0.15790370685982105, "grad_norm": 19.125, "learning_rate": 2.465452593030642e-06, "loss": 0.9063, "step": 1853 }, { "epoch": 0.157988922028121, "grad_norm": 13.875, "learning_rate": 2.4654133494371547e-06, "loss": 0.5622, "step": 1854 }, { "epoch": 0.15807413719642097, "grad_norm": 9.6875, "learning_rate": 2.465374083879926e-06, "loss": 0.3138, "step": 1855 }, { "epoch": 0.15815935236472092, "grad_norm": 12.0625, "learning_rate": 2.465334796359666e-06, "loss": 0.5773, "step": 1856 }, { "epoch": 0.15824456753302088, "grad_norm": 21.125, "learning_rate": 2.4652954868770834e-06, "loss": 0.8814, "step": 1857 }, { "epoch": 0.15832978270132084, "grad_norm": 31.0, "learning_rate": 2.4652561554328895e-06, "loss": 0.8206, "step": 1858 }, { "epoch": 0.1584149978696208, "grad_norm": 12.5, "learning_rate": 2.465216802027795e-06, "loss": 0.6012, "step": 1859 }, { "epoch": 0.15850021303792075, "grad_norm": 13.75, "learning_rate": 2.465177426662511e-06, "loss": 0.7477, "step": 1860 }, { "epoch": 0.1585854282062207, "grad_norm": 15.5, "learning_rate": 2.4651380293377486e-06, "loss": 1.0542, "step": 1861 }, { "epoch": 0.15867064337452066, "grad_norm": 15.625, "learning_rate": 2.46509861005422e-06, "loss": 0.745, "step": 1862 }, { "epoch": 0.15875585854282062, "grad_norm": 16.375, "learning_rate": 2.4650591688126384e-06, "loss": 0.7779, "step": 1863 }, { "epoch": 0.15884107371112058, "grad_norm": 19.0, "learning_rate": 2.465019705613715e-06, "loss": 0.9052, "step": 1864 }, { "epoch": 0.15892628887942054, "grad_norm": 15.1875, "learning_rate": 2.4649802204581643e-06, "loss": 0.7703, "step": 1865 }, { "epoch": 0.1590115040477205, "grad_norm": 13.25, "learning_rate": 2.4649407133466987e-06, "loss": 0.639, "step": 1866 }, { "epoch": 0.15909671921602045, "grad_norm": 17.125, "learning_rate": 2.464901184280033e-06, "loss": 0.809, "step": 1867 }, { "epoch": 0.1591819343843204, "grad_norm": 17.375, "learning_rate": 2.4648616332588807e-06, "loss": 0.6356, "step": 1868 }, { "epoch": 0.15926714955262036, "grad_norm": 16.625, "learning_rate": 2.4648220602839577e-06, "loss": 0.9886, "step": 1869 }, { "epoch": 0.15935236472092032, "grad_norm": 16.125, "learning_rate": 2.4647824653559786e-06, "loss": 1.0655, "step": 1870 }, { "epoch": 0.15943757988922028, "grad_norm": 14.5625, "learning_rate": 2.4647428484756584e-06, "loss": 0.9297, "step": 1871 }, { "epoch": 0.15952279505752023, "grad_norm": 14.25, "learning_rate": 2.4647032096437133e-06, "loss": 1.1566, "step": 1872 }, { "epoch": 0.1596080102258202, "grad_norm": 14.875, "learning_rate": 2.46466354886086e-06, "loss": 0.9909, "step": 1873 }, { "epoch": 0.15969322539412015, "grad_norm": 18.875, "learning_rate": 2.4646238661278148e-06, "loss": 0.9221, "step": 1874 }, { "epoch": 0.1597784405624201, "grad_norm": 18.875, "learning_rate": 2.4645841614452955e-06, "loss": 0.9497, "step": 1875 }, { "epoch": 0.15986365573072006, "grad_norm": 8.8125, "learning_rate": 2.464544434814018e-06, "loss": 0.3158, "step": 1876 }, { "epoch": 0.15994887089902002, "grad_norm": 22.625, "learning_rate": 2.4645046862347024e-06, "loss": 1.2811, "step": 1877 }, { "epoch": 0.16003408606731997, "grad_norm": 24.375, "learning_rate": 2.4644649157080647e-06, "loss": 0.6415, "step": 1878 }, { "epoch": 0.16011930123561993, "grad_norm": 16.125, "learning_rate": 2.4644251232348253e-06, "loss": 1.0919, "step": 1879 }, { "epoch": 0.1602045164039199, "grad_norm": 16.0, "learning_rate": 2.4643853088157026e-06, "loss": 0.6815, "step": 1880 }, { "epoch": 0.16028973157221985, "grad_norm": 13.125, "learning_rate": 2.4643454724514165e-06, "loss": 0.766, "step": 1881 }, { "epoch": 0.1603749467405198, "grad_norm": 15.375, "learning_rate": 2.464305614142686e-06, "loss": 0.843, "step": 1882 }, { "epoch": 0.16046016190881976, "grad_norm": 14.375, "learning_rate": 2.4642657338902324e-06, "loss": 0.5713, "step": 1883 }, { "epoch": 0.16054537707711972, "grad_norm": 16.625, "learning_rate": 2.464225831694776e-06, "loss": 0.2904, "step": 1884 }, { "epoch": 0.16063059224541967, "grad_norm": 35.5, "learning_rate": 2.464185907557038e-06, "loss": 0.6779, "step": 1885 }, { "epoch": 0.16071580741371963, "grad_norm": 10.625, "learning_rate": 2.4641459614777393e-06, "loss": 0.4465, "step": 1886 }, { "epoch": 0.1608010225820196, "grad_norm": 12.875, "learning_rate": 2.4641059934576022e-06, "loss": 0.3821, "step": 1887 }, { "epoch": 0.16088623775031954, "grad_norm": 17.625, "learning_rate": 2.4640660034973492e-06, "loss": 0.8156, "step": 1888 }, { "epoch": 0.1609714529186195, "grad_norm": 25.0, "learning_rate": 2.4640259915977024e-06, "loss": 0.854, "step": 1889 }, { "epoch": 0.16105666808691946, "grad_norm": 10.75, "learning_rate": 2.463985957759385e-06, "loss": 0.3041, "step": 1890 }, { "epoch": 0.16114188325521944, "grad_norm": 15.25, "learning_rate": 2.463945901983121e-06, "loss": 0.8962, "step": 1891 }, { "epoch": 0.1612270984235194, "grad_norm": 10.3125, "learning_rate": 2.4639058242696337e-06, "loss": 0.3486, "step": 1892 }, { "epoch": 0.16131231359181936, "grad_norm": 11.8125, "learning_rate": 2.4638657246196474e-06, "loss": 0.5969, "step": 1893 }, { "epoch": 0.1613975287601193, "grad_norm": 34.5, "learning_rate": 2.463825603033887e-06, "loss": 1.2165, "step": 1894 }, { "epoch": 0.16148274392841927, "grad_norm": 12.9375, "learning_rate": 2.4637854595130774e-06, "loss": 0.6792, "step": 1895 }, { "epoch": 0.16156795909671923, "grad_norm": 17.625, "learning_rate": 2.4637452940579434e-06, "loss": 0.7919, "step": 1896 }, { "epoch": 0.16165317426501918, "grad_norm": 11.6875, "learning_rate": 2.463705106669212e-06, "loss": 0.5261, "step": 1897 }, { "epoch": 0.16173838943331914, "grad_norm": 30.25, "learning_rate": 2.4636648973476086e-06, "loss": 1.0703, "step": 1898 }, { "epoch": 0.1618236046016191, "grad_norm": 16.75, "learning_rate": 2.46362466609386e-06, "loss": 0.9846, "step": 1899 }, { "epoch": 0.16190881976991905, "grad_norm": 16.5, "learning_rate": 2.4635844129086934e-06, "loss": 0.5692, "step": 1900 }, { "epoch": 0.161994034938219, "grad_norm": 17.75, "learning_rate": 2.463544137792836e-06, "loss": 0.9089, "step": 1901 }, { "epoch": 0.16207925010651897, "grad_norm": 21.5, "learning_rate": 2.463503840747016e-06, "loss": 1.0333, "step": 1902 }, { "epoch": 0.16216446527481893, "grad_norm": 10.25, "learning_rate": 2.463463521771961e-06, "loss": 0.3731, "step": 1903 }, { "epoch": 0.16224968044311888, "grad_norm": 13.375, "learning_rate": 2.4634231808684e-06, "loss": 0.5566, "step": 1904 }, { "epoch": 0.16233489561141884, "grad_norm": 12.0625, "learning_rate": 2.463382818037062e-06, "loss": 0.5679, "step": 1905 }, { "epoch": 0.1624201107797188, "grad_norm": 13.5, "learning_rate": 2.4633424332786767e-06, "loss": 0.8186, "step": 1906 }, { "epoch": 0.16250532594801875, "grad_norm": 15.625, "learning_rate": 2.463302026593973e-06, "loss": 0.7229, "step": 1907 }, { "epoch": 0.1625905411163187, "grad_norm": 14.5625, "learning_rate": 2.463261597983682e-06, "loss": 0.8144, "step": 1908 }, { "epoch": 0.16267575628461867, "grad_norm": 16.25, "learning_rate": 2.4632211474485333e-06, "loss": 0.9194, "step": 1909 }, { "epoch": 0.16276097145291862, "grad_norm": 14.1875, "learning_rate": 2.463180674989259e-06, "loss": 0.5902, "step": 1910 }, { "epoch": 0.16284618662121858, "grad_norm": 11.1875, "learning_rate": 2.4631401806065898e-06, "loss": 0.4058, "step": 1911 }, { "epoch": 0.16293140178951854, "grad_norm": 15.1875, "learning_rate": 2.4630996643012577e-06, "loss": 1.0518, "step": 1912 }, { "epoch": 0.1630166169578185, "grad_norm": 16.125, "learning_rate": 2.4630591260739946e-06, "loss": 0.8537, "step": 1913 }, { "epoch": 0.16310183212611845, "grad_norm": 22.875, "learning_rate": 2.4630185659255333e-06, "loss": 1.2322, "step": 1914 }, { "epoch": 0.1631870472944184, "grad_norm": 13.4375, "learning_rate": 2.462977983856607e-06, "loss": 0.5618, "step": 1915 }, { "epoch": 0.16327226246271836, "grad_norm": 12.0, "learning_rate": 2.4629373798679487e-06, "loss": 0.6039, "step": 1916 }, { "epoch": 0.16335747763101832, "grad_norm": 13.375, "learning_rate": 2.462896753960292e-06, "loss": 0.7162, "step": 1917 }, { "epoch": 0.16344269279931828, "grad_norm": 15.9375, "learning_rate": 2.4628561061343716e-06, "loss": 1.0867, "step": 1918 }, { "epoch": 0.16352790796761824, "grad_norm": 16.25, "learning_rate": 2.462815436390922e-06, "loss": 0.9963, "step": 1919 }, { "epoch": 0.1636131231359182, "grad_norm": 13.1875, "learning_rate": 2.462774744730677e-06, "loss": 0.6273, "step": 1920 }, { "epoch": 0.16369833830421815, "grad_norm": 13.3125, "learning_rate": 2.4627340311543735e-06, "loss": 0.7721, "step": 1921 }, { "epoch": 0.1637835534725181, "grad_norm": 18.5, "learning_rate": 2.462693295662746e-06, "loss": 0.6745, "step": 1922 }, { "epoch": 0.16386876864081806, "grad_norm": 11.5625, "learning_rate": 2.462652538256532e-06, "loss": 0.4459, "step": 1923 }, { "epoch": 0.16395398380911802, "grad_norm": 12.4375, "learning_rate": 2.462611758936467e-06, "loss": 0.6698, "step": 1924 }, { "epoch": 0.16403919897741798, "grad_norm": 13.5625, "learning_rate": 2.4625709577032876e-06, "loss": 0.9035, "step": 1925 }, { "epoch": 0.16412441414571793, "grad_norm": 14.625, "learning_rate": 2.4625301345577318e-06, "loss": 0.8877, "step": 1926 }, { "epoch": 0.1642096293140179, "grad_norm": 16.875, "learning_rate": 2.4624892895005377e-06, "loss": 0.8654, "step": 1927 }, { "epoch": 0.16429484448231785, "grad_norm": 12.875, "learning_rate": 2.4624484225324424e-06, "loss": 0.8497, "step": 1928 }, { "epoch": 0.1643800596506178, "grad_norm": 17.875, "learning_rate": 2.4624075336541848e-06, "loss": 0.8131, "step": 1929 }, { "epoch": 0.16446527481891776, "grad_norm": 16.75, "learning_rate": 2.4623666228665043e-06, "loss": 0.8727, "step": 1930 }, { "epoch": 0.16455048998721772, "grad_norm": 14.875, "learning_rate": 2.4623256901701396e-06, "loss": 0.8846, "step": 1931 }, { "epoch": 0.16463570515551768, "grad_norm": 12.0, "learning_rate": 2.4622847355658306e-06, "loss": 0.7339, "step": 1932 }, { "epoch": 0.16472092032381763, "grad_norm": 17.5, "learning_rate": 2.4622437590543176e-06, "loss": 0.9891, "step": 1933 }, { "epoch": 0.1648061354921176, "grad_norm": 14.125, "learning_rate": 2.46220276063634e-06, "loss": 0.5347, "step": 1934 }, { "epoch": 0.16489135066041755, "grad_norm": 15.9375, "learning_rate": 2.4621617403126407e-06, "loss": 1.1538, "step": 1935 }, { "epoch": 0.1649765658287175, "grad_norm": 20.625, "learning_rate": 2.462120698083959e-06, "loss": 1.0269, "step": 1936 }, { "epoch": 0.16506178099701746, "grad_norm": 12.9375, "learning_rate": 2.4620796339510382e-06, "loss": 0.7374, "step": 1937 }, { "epoch": 0.16514699616531742, "grad_norm": 11.9375, "learning_rate": 2.462038547914619e-06, "loss": 0.4763, "step": 1938 }, { "epoch": 0.16523221133361737, "grad_norm": 24.375, "learning_rate": 2.4619974399754444e-06, "loss": 1.3747, "step": 1939 }, { "epoch": 0.16531742650191733, "grad_norm": 16.125, "learning_rate": 2.4619563101342577e-06, "loss": 1.0067, "step": 1940 }, { "epoch": 0.1654026416702173, "grad_norm": 15.25, "learning_rate": 2.4619151583918013e-06, "loss": 1.0633, "step": 1941 }, { "epoch": 0.16548785683851724, "grad_norm": 13.0, "learning_rate": 2.4618739847488196e-06, "loss": 0.5575, "step": 1942 }, { "epoch": 0.1655730720068172, "grad_norm": 14.25, "learning_rate": 2.461832789206056e-06, "loss": 0.5893, "step": 1943 }, { "epoch": 0.16565828717511716, "grad_norm": 12.9375, "learning_rate": 2.461791571764256e-06, "loss": 0.7454, "step": 1944 }, { "epoch": 0.16574350234341712, "grad_norm": 20.5, "learning_rate": 2.4617503324241633e-06, "loss": 0.9566, "step": 1945 }, { "epoch": 0.16582871751171707, "grad_norm": 12.5625, "learning_rate": 2.4617090711865235e-06, "loss": 0.7483, "step": 1946 }, { "epoch": 0.16591393268001706, "grad_norm": 22.625, "learning_rate": 2.4616677880520825e-06, "loss": 0.6018, "step": 1947 }, { "epoch": 0.165999147848317, "grad_norm": 13.5625, "learning_rate": 2.461626483021586e-06, "loss": 0.7312, "step": 1948 }, { "epoch": 0.16608436301661697, "grad_norm": 17.5, "learning_rate": 2.461585156095781e-06, "loss": 0.7235, "step": 1949 }, { "epoch": 0.16616957818491693, "grad_norm": 29.125, "learning_rate": 2.4615438072754134e-06, "loss": 1.2423, "step": 1950 }, { "epoch": 0.16625479335321688, "grad_norm": 14.6875, "learning_rate": 2.461502436561231e-06, "loss": 0.7071, "step": 1951 }, { "epoch": 0.16634000852151684, "grad_norm": 28.875, "learning_rate": 2.4614610439539813e-06, "loss": 0.7138, "step": 1952 }, { "epoch": 0.1664252236898168, "grad_norm": 12.375, "learning_rate": 2.4614196294544122e-06, "loss": 0.4549, "step": 1953 }, { "epoch": 0.16651043885811675, "grad_norm": 15.9375, "learning_rate": 2.4613781930632723e-06, "loss": 0.7831, "step": 1954 }, { "epoch": 0.1665956540264167, "grad_norm": 22.125, "learning_rate": 2.4613367347813105e-06, "loss": 1.1, "step": 1955 }, { "epoch": 0.16668086919471667, "grad_norm": 15.3125, "learning_rate": 2.4612952546092763e-06, "loss": 0.8827, "step": 1956 }, { "epoch": 0.16676608436301663, "grad_norm": 13.9375, "learning_rate": 2.461253752547918e-06, "loss": 0.6588, "step": 1957 }, { "epoch": 0.16685129953131658, "grad_norm": 16.625, "learning_rate": 2.4612122285979863e-06, "loss": 1.0942, "step": 1958 }, { "epoch": 0.16693651469961654, "grad_norm": 24.75, "learning_rate": 2.4611706827602324e-06, "loss": 1.3617, "step": 1959 }, { "epoch": 0.1670217298679165, "grad_norm": 14.1875, "learning_rate": 2.4611291150354054e-06, "loss": 0.5499, "step": 1960 }, { "epoch": 0.16710694503621645, "grad_norm": 19.0, "learning_rate": 2.461087525424258e-06, "loss": 0.4473, "step": 1961 }, { "epoch": 0.1671921602045164, "grad_norm": 16.875, "learning_rate": 2.461045913927541e-06, "loss": 0.92, "step": 1962 }, { "epoch": 0.16727737537281637, "grad_norm": 20.75, "learning_rate": 2.461004280546007e-06, "loss": 0.6915, "step": 1963 }, { "epoch": 0.16736259054111632, "grad_norm": 13.9375, "learning_rate": 2.4609626252804074e-06, "loss": 0.6387, "step": 1964 }, { "epoch": 0.16744780570941628, "grad_norm": 13.75, "learning_rate": 2.4609209481314957e-06, "loss": 0.7349, "step": 1965 }, { "epoch": 0.16753302087771624, "grad_norm": 17.0, "learning_rate": 2.4608792491000245e-06, "loss": 0.7464, "step": 1966 }, { "epoch": 0.1676182360460162, "grad_norm": 14.6875, "learning_rate": 2.4608375281867475e-06, "loss": 0.6818, "step": 1967 }, { "epoch": 0.16770345121431615, "grad_norm": 12.25, "learning_rate": 2.460795785392419e-06, "loss": 0.6621, "step": 1968 }, { "epoch": 0.1677886663826161, "grad_norm": 18.25, "learning_rate": 2.4607540207177934e-06, "loss": 0.9225, "step": 1969 }, { "epoch": 0.16787388155091607, "grad_norm": 14.3125, "learning_rate": 2.460712234163625e-06, "loss": 0.6546, "step": 1970 }, { "epoch": 0.16795909671921602, "grad_norm": 16.375, "learning_rate": 2.460670425730669e-06, "loss": 0.8066, "step": 1971 }, { "epoch": 0.16804431188751598, "grad_norm": 10.75, "learning_rate": 2.4606285954196805e-06, "loss": 0.5965, "step": 1972 }, { "epoch": 0.16812952705581594, "grad_norm": 27.125, "learning_rate": 2.4605867432314167e-06, "loss": 0.9837, "step": 1973 }, { "epoch": 0.1682147422241159, "grad_norm": 16.125, "learning_rate": 2.460544869166633e-06, "loss": 0.9392, "step": 1974 }, { "epoch": 0.16829995739241585, "grad_norm": 15.125, "learning_rate": 2.460502973226086e-06, "loss": 0.9626, "step": 1975 }, { "epoch": 0.1683851725607158, "grad_norm": 25.0, "learning_rate": 2.4604610554105326e-06, "loss": 1.2058, "step": 1976 }, { "epoch": 0.16847038772901576, "grad_norm": 17.875, "learning_rate": 2.460419115720731e-06, "loss": 1.023, "step": 1977 }, { "epoch": 0.16855560289731572, "grad_norm": 14.25, "learning_rate": 2.460377154157439e-06, "loss": 0.6717, "step": 1978 }, { "epoch": 0.16864081806561568, "grad_norm": 13.0625, "learning_rate": 2.4603351707214144e-06, "loss": 0.6409, "step": 1979 }, { "epoch": 0.16872603323391563, "grad_norm": 15.125, "learning_rate": 2.4602931654134165e-06, "loss": 0.959, "step": 1980 }, { "epoch": 0.1688112484022156, "grad_norm": 18.25, "learning_rate": 2.460251138234204e-06, "loss": 1.0022, "step": 1981 }, { "epoch": 0.16889646357051555, "grad_norm": 14.3125, "learning_rate": 2.4602090891845363e-06, "loss": 0.6976, "step": 1982 }, { "epoch": 0.1689816787388155, "grad_norm": 13.0, "learning_rate": 2.4601670182651736e-06, "loss": 0.5451, "step": 1983 }, { "epoch": 0.16906689390711546, "grad_norm": 16.875, "learning_rate": 2.4601249254768762e-06, "loss": 1.0195, "step": 1984 }, { "epoch": 0.16915210907541542, "grad_norm": 19.0, "learning_rate": 2.460082810820404e-06, "loss": 0.8803, "step": 1985 }, { "epoch": 0.16923732424371538, "grad_norm": 18.875, "learning_rate": 2.460040674296519e-06, "loss": 1.1858, "step": 1986 }, { "epoch": 0.16932253941201533, "grad_norm": 18.375, "learning_rate": 2.4599985159059817e-06, "loss": 1.0269, "step": 1987 }, { "epoch": 0.1694077545803153, "grad_norm": 11.5, "learning_rate": 2.4599563356495547e-06, "loss": 0.4623, "step": 1988 }, { "epoch": 0.16949296974861525, "grad_norm": 12.875, "learning_rate": 2.459914133528e-06, "loss": 0.5695, "step": 1989 }, { "epoch": 0.1695781849169152, "grad_norm": 13.6875, "learning_rate": 2.4598719095420803e-06, "loss": 0.7257, "step": 1990 }, { "epoch": 0.16966340008521516, "grad_norm": 18.875, "learning_rate": 2.4598296636925586e-06, "loss": 0.8458, "step": 1991 }, { "epoch": 0.16974861525351512, "grad_norm": 15.0625, "learning_rate": 2.4597873959801982e-06, "loss": 0.6678, "step": 1992 }, { "epoch": 0.16983383042181507, "grad_norm": 13.6875, "learning_rate": 2.459745106405763e-06, "loss": 0.4443, "step": 1993 }, { "epoch": 0.16991904559011503, "grad_norm": 14.25, "learning_rate": 2.4597027949700176e-06, "loss": 0.6835, "step": 1994 }, { "epoch": 0.170004260758415, "grad_norm": 17.5, "learning_rate": 2.4596604616737257e-06, "loss": 0.7451, "step": 1995 }, { "epoch": 0.17008947592671494, "grad_norm": 15.4375, "learning_rate": 2.459618106517653e-06, "loss": 0.6533, "step": 1996 }, { "epoch": 0.1701746910950149, "grad_norm": 16.875, "learning_rate": 2.459575729502565e-06, "loss": 1.0351, "step": 1997 }, { "epoch": 0.17025990626331486, "grad_norm": 21.875, "learning_rate": 2.4595333306292272e-06, "loss": 0.3285, "step": 1998 }, { "epoch": 0.17034512143161482, "grad_norm": 12.5, "learning_rate": 2.4594909098984058e-06, "loss": 0.6, "step": 1999 }, { "epoch": 0.17043033659991477, "grad_norm": 16.0, "learning_rate": 2.459448467310867e-06, "loss": 1.0138, "step": 2000 }, { "epoch": 0.17051555176821473, "grad_norm": 11.375, "learning_rate": 2.4594060028673788e-06, "loss": 0.5304, "step": 2001 }, { "epoch": 0.1706007669365147, "grad_norm": 18.75, "learning_rate": 2.4593635165687076e-06, "loss": 0.7781, "step": 2002 }, { "epoch": 0.17068598210481467, "grad_norm": 22.625, "learning_rate": 2.459321008415622e-06, "loss": 1.0972, "step": 2003 }, { "epoch": 0.17077119727311463, "grad_norm": 12.125, "learning_rate": 2.459278478408889e-06, "loss": 0.4955, "step": 2004 }, { "epoch": 0.17085641244141458, "grad_norm": 11.25, "learning_rate": 2.4592359265492786e-06, "loss": 0.472, "step": 2005 }, { "epoch": 0.17094162760971454, "grad_norm": 12.5625, "learning_rate": 2.4591933528375586e-06, "loss": 0.475, "step": 2006 }, { "epoch": 0.1710268427780145, "grad_norm": 14.4375, "learning_rate": 2.459150757274499e-06, "loss": 0.9155, "step": 2007 }, { "epoch": 0.17111205794631446, "grad_norm": 19.5, "learning_rate": 2.459108139860869e-06, "loss": 1.1182, "step": 2008 }, { "epoch": 0.1711972731146144, "grad_norm": 13.0625, "learning_rate": 2.459065500597439e-06, "loss": 0.619, "step": 2009 }, { "epoch": 0.17128248828291437, "grad_norm": 15.6875, "learning_rate": 2.45902283948498e-06, "loss": 0.9494, "step": 2010 }, { "epoch": 0.17136770345121433, "grad_norm": 12.375, "learning_rate": 2.4589801565242627e-06, "loss": 0.5641, "step": 2011 }, { "epoch": 0.17145291861951428, "grad_norm": 12.75, "learning_rate": 2.458937451716058e-06, "loss": 0.6059, "step": 2012 }, { "epoch": 0.17153813378781424, "grad_norm": 14.375, "learning_rate": 2.4588947250611377e-06, "loss": 0.8325, "step": 2013 }, { "epoch": 0.1716233489561142, "grad_norm": 12.25, "learning_rate": 2.4588519765602743e-06, "loss": 0.6977, "step": 2014 }, { "epoch": 0.17170856412441415, "grad_norm": 15.625, "learning_rate": 2.4588092062142394e-06, "loss": 0.6614, "step": 2015 }, { "epoch": 0.1717937792927141, "grad_norm": 18.125, "learning_rate": 2.4587664140238076e-06, "loss": 1.1969, "step": 2016 }, { "epoch": 0.17187899446101407, "grad_norm": 22.125, "learning_rate": 2.458723599989751e-06, "loss": 0.9492, "step": 2017 }, { "epoch": 0.17196420962931402, "grad_norm": 13.6875, "learning_rate": 2.458680764112843e-06, "loss": 0.6721, "step": 2018 }, { "epoch": 0.17204942479761398, "grad_norm": 18.5, "learning_rate": 2.458637906393858e-06, "loss": 0.8085, "step": 2019 }, { "epoch": 0.17213463996591394, "grad_norm": 11.0, "learning_rate": 2.458595026833571e-06, "loss": 0.4131, "step": 2020 }, { "epoch": 0.1722198551342139, "grad_norm": 18.625, "learning_rate": 2.4585521254327567e-06, "loss": 0.8082, "step": 2021 }, { "epoch": 0.17230507030251385, "grad_norm": 12.9375, "learning_rate": 2.45850920219219e-06, "loss": 0.8042, "step": 2022 }, { "epoch": 0.1723902854708138, "grad_norm": 13.4375, "learning_rate": 2.458466257112647e-06, "loss": 0.7371, "step": 2023 }, { "epoch": 0.17247550063911377, "grad_norm": 10.1875, "learning_rate": 2.4584232901949036e-06, "loss": 0.4148, "step": 2024 }, { "epoch": 0.17256071580741372, "grad_norm": 13.5625, "learning_rate": 2.4583803014397355e-06, "loss": 0.7513, "step": 2025 }, { "epoch": 0.17264593097571368, "grad_norm": 8.125, "learning_rate": 2.458337290847921e-06, "loss": 0.2694, "step": 2026 }, { "epoch": 0.17273114614401364, "grad_norm": 12.6875, "learning_rate": 2.4582942584202363e-06, "loss": 1.0149, "step": 2027 }, { "epoch": 0.1728163613123136, "grad_norm": 17.5, "learning_rate": 2.458251204157459e-06, "loss": 0.9005, "step": 2028 }, { "epoch": 0.17290157648061355, "grad_norm": 12.5625, "learning_rate": 2.458208128060368e-06, "loss": 0.6792, "step": 2029 }, { "epoch": 0.1729867916489135, "grad_norm": 16.375, "learning_rate": 2.458165030129741e-06, "loss": 0.6161, "step": 2030 }, { "epoch": 0.17307200681721346, "grad_norm": 15.1875, "learning_rate": 2.458121910366357e-06, "loss": 0.4292, "step": 2031 }, { "epoch": 0.17315722198551342, "grad_norm": 18.125, "learning_rate": 2.4580787687709947e-06, "loss": 1.0835, "step": 2032 }, { "epoch": 0.17324243715381338, "grad_norm": 12.875, "learning_rate": 2.458035605344435e-06, "loss": 0.7919, "step": 2033 }, { "epoch": 0.17332765232211333, "grad_norm": 15.5, "learning_rate": 2.4579924200874566e-06, "loss": 0.7203, "step": 2034 }, { "epoch": 0.1734128674904133, "grad_norm": 22.0, "learning_rate": 2.4579492130008406e-06, "loss": 1.0555, "step": 2035 }, { "epoch": 0.17349808265871325, "grad_norm": 16.25, "learning_rate": 2.4579059840853677e-06, "loss": 0.8128, "step": 2036 }, { "epoch": 0.1735832978270132, "grad_norm": 31.375, "learning_rate": 2.457862733341819e-06, "loss": 0.9135, "step": 2037 }, { "epoch": 0.17366851299531316, "grad_norm": 14.125, "learning_rate": 2.457819460770976e-06, "loss": 0.713, "step": 2038 }, { "epoch": 0.17375372816361312, "grad_norm": 13.0625, "learning_rate": 2.457776166373621e-06, "loss": 0.6362, "step": 2039 }, { "epoch": 0.17383894333191308, "grad_norm": 16.5, "learning_rate": 2.457732850150536e-06, "loss": 1.0217, "step": 2040 }, { "epoch": 0.17392415850021303, "grad_norm": 17.625, "learning_rate": 2.457689512102504e-06, "loss": 0.6053, "step": 2041 }, { "epoch": 0.174009373668513, "grad_norm": 14.75, "learning_rate": 2.457646152230308e-06, "loss": 0.8687, "step": 2042 }, { "epoch": 0.17409458883681295, "grad_norm": 12.25, "learning_rate": 2.4576027705347316e-06, "loss": 0.5309, "step": 2043 }, { "epoch": 0.1741798040051129, "grad_norm": 21.75, "learning_rate": 2.457559367016559e-06, "loss": 0.903, "step": 2044 }, { "epoch": 0.17426501917341286, "grad_norm": 15.0, "learning_rate": 2.4575159416765742e-06, "loss": 0.6945, "step": 2045 }, { "epoch": 0.17435023434171282, "grad_norm": 17.5, "learning_rate": 2.4574724945155624e-06, "loss": 0.9111, "step": 2046 }, { "epoch": 0.17443544951001277, "grad_norm": 14.875, "learning_rate": 2.457429025534308e-06, "loss": 0.8245, "step": 2047 }, { "epoch": 0.17452066467831273, "grad_norm": 25.125, "learning_rate": 2.4573855347335972e-06, "loss": 1.5115, "step": 2048 }, { "epoch": 0.1746058798466127, "grad_norm": 15.9375, "learning_rate": 2.4573420221142155e-06, "loss": 0.8505, "step": 2049 }, { "epoch": 0.17469109501491265, "grad_norm": 17.0, "learning_rate": 2.4572984876769495e-06, "loss": 0.8712, "step": 2050 }, { "epoch": 0.1747763101832126, "grad_norm": 13.9375, "learning_rate": 2.457254931422586e-06, "loss": 0.6953, "step": 2051 }, { "epoch": 0.17486152535151256, "grad_norm": 15.875, "learning_rate": 2.4572113533519116e-06, "loss": 0.5377, "step": 2052 }, { "epoch": 0.17494674051981252, "grad_norm": 13.1875, "learning_rate": 2.457167753465714e-06, "loss": 0.7744, "step": 2053 }, { "epoch": 0.17503195568811247, "grad_norm": 21.5, "learning_rate": 2.457124131764781e-06, "loss": 0.846, "step": 2054 }, { "epoch": 0.17511717085641243, "grad_norm": 14.875, "learning_rate": 2.4570804882499015e-06, "loss": 0.9177, "step": 2055 }, { "epoch": 0.1752023860247124, "grad_norm": 12.125, "learning_rate": 2.457036822921864e-06, "loss": 0.802, "step": 2056 }, { "epoch": 0.17528760119301234, "grad_norm": 12.6875, "learning_rate": 2.456993135781457e-06, "loss": 0.5264, "step": 2057 }, { "epoch": 0.1753728163613123, "grad_norm": 12.25, "learning_rate": 2.4569494268294702e-06, "loss": 0.5958, "step": 2058 }, { "epoch": 0.17545803152961226, "grad_norm": 19.5, "learning_rate": 2.4569056960666937e-06, "loss": 1.366, "step": 2059 }, { "epoch": 0.17554324669791224, "grad_norm": 12.3125, "learning_rate": 2.456861943493918e-06, "loss": 0.7018, "step": 2060 }, { "epoch": 0.1756284618662122, "grad_norm": 11.8125, "learning_rate": 2.456818169111933e-06, "loss": 0.7075, "step": 2061 }, { "epoch": 0.17571367703451216, "grad_norm": 16.25, "learning_rate": 2.45677437292153e-06, "loss": 0.8157, "step": 2062 }, { "epoch": 0.1757988922028121, "grad_norm": 18.75, "learning_rate": 2.4567305549235008e-06, "loss": 1.0475, "step": 2063 }, { "epoch": 0.17588410737111207, "grad_norm": 10.5625, "learning_rate": 2.456686715118637e-06, "loss": 0.327, "step": 2064 }, { "epoch": 0.17596932253941203, "grad_norm": 19.0, "learning_rate": 2.456642853507731e-06, "loss": 1.1008, "step": 2065 }, { "epoch": 0.17605453770771198, "grad_norm": 14.5, "learning_rate": 2.4565989700915746e-06, "loss": 1.0194, "step": 2066 }, { "epoch": 0.17613975287601194, "grad_norm": 26.0, "learning_rate": 2.456555064870962e-06, "loss": 0.8765, "step": 2067 }, { "epoch": 0.1762249680443119, "grad_norm": 13.5625, "learning_rate": 2.456511137846686e-06, "loss": 0.9368, "step": 2068 }, { "epoch": 0.17631018321261185, "grad_norm": 16.5, "learning_rate": 2.4564671890195406e-06, "loss": 0.9802, "step": 2069 }, { "epoch": 0.1763953983809118, "grad_norm": 17.125, "learning_rate": 2.4564232183903196e-06, "loss": 1.0024, "step": 2070 }, { "epoch": 0.17648061354921177, "grad_norm": 15.125, "learning_rate": 2.4563792259598186e-06, "loss": 0.9557, "step": 2071 }, { "epoch": 0.17656582871751172, "grad_norm": 17.25, "learning_rate": 2.4563352117288312e-06, "loss": 0.5583, "step": 2072 }, { "epoch": 0.17665104388581168, "grad_norm": 14.75, "learning_rate": 2.4562911756981537e-06, "loss": 1.0684, "step": 2073 }, { "epoch": 0.17673625905411164, "grad_norm": 13.625, "learning_rate": 2.4562471178685815e-06, "loss": 0.9078, "step": 2074 }, { "epoch": 0.1768214742224116, "grad_norm": 14.9375, "learning_rate": 2.456203038240911e-06, "loss": 0.9359, "step": 2075 }, { "epoch": 0.17690668939071155, "grad_norm": 19.75, "learning_rate": 2.4561589368159388e-06, "loss": 1.1046, "step": 2076 }, { "epoch": 0.1769919045590115, "grad_norm": 15.9375, "learning_rate": 2.4561148135944616e-06, "loss": 0.8891, "step": 2077 }, { "epoch": 0.17707711972731147, "grad_norm": 28.25, "learning_rate": 2.4560706685772767e-06, "loss": 0.8514, "step": 2078 }, { "epoch": 0.17716233489561142, "grad_norm": 28.0, "learning_rate": 2.4560265017651826e-06, "loss": 0.5024, "step": 2079 }, { "epoch": 0.17724755006391138, "grad_norm": 18.5, "learning_rate": 2.455982313158976e-06, "loss": 1.2425, "step": 2080 }, { "epoch": 0.17733276523221134, "grad_norm": 13.4375, "learning_rate": 2.455938102759457e-06, "loss": 0.7255, "step": 2081 }, { "epoch": 0.1774179804005113, "grad_norm": 14.625, "learning_rate": 2.4558938705674236e-06, "loss": 1.0185, "step": 2082 }, { "epoch": 0.17750319556881125, "grad_norm": 16.25, "learning_rate": 2.455849616583675e-06, "loss": 0.8518, "step": 2083 }, { "epoch": 0.1775884107371112, "grad_norm": 11.25, "learning_rate": 2.4558053408090115e-06, "loss": 0.9188, "step": 2084 }, { "epoch": 0.17767362590541116, "grad_norm": 12.75, "learning_rate": 2.455761043244233e-06, "loss": 0.8994, "step": 2085 }, { "epoch": 0.17775884107371112, "grad_norm": 13.25, "learning_rate": 2.4557167238901404e-06, "loss": 0.6106, "step": 2086 }, { "epoch": 0.17784405624201108, "grad_norm": 20.875, "learning_rate": 2.4556723827475333e-06, "loss": 0.9989, "step": 2087 }, { "epoch": 0.17792927141031104, "grad_norm": 14.625, "learning_rate": 2.4556280198172144e-06, "loss": 0.6338, "step": 2088 }, { "epoch": 0.178014486578611, "grad_norm": 15.75, "learning_rate": 2.455583635099985e-06, "loss": 0.8638, "step": 2089 }, { "epoch": 0.17809970174691095, "grad_norm": 15.4375, "learning_rate": 2.4555392285966467e-06, "loss": 0.8543, "step": 2090 }, { "epoch": 0.1781849169152109, "grad_norm": 10.0625, "learning_rate": 2.455494800308002e-06, "loss": 0.3567, "step": 2091 }, { "epoch": 0.17827013208351086, "grad_norm": 14.25, "learning_rate": 2.4554503502348544e-06, "loss": 0.5069, "step": 2092 }, { "epoch": 0.17835534725181082, "grad_norm": 16.875, "learning_rate": 2.4554058783780064e-06, "loss": 0.6473, "step": 2093 }, { "epoch": 0.17844056242011078, "grad_norm": 14.1875, "learning_rate": 2.455361384738263e-06, "loss": 0.735, "step": 2094 }, { "epoch": 0.17852577758841073, "grad_norm": 23.625, "learning_rate": 2.4553168693164262e-06, "loss": 1.4259, "step": 2095 }, { "epoch": 0.1786109927567107, "grad_norm": 12.625, "learning_rate": 2.455272332113302e-06, "loss": 0.6685, "step": 2096 }, { "epoch": 0.17869620792501065, "grad_norm": 15.5, "learning_rate": 2.4552277731296952e-06, "loss": 0.7064, "step": 2097 }, { "epoch": 0.1787814230933106, "grad_norm": 18.375, "learning_rate": 2.4551831923664098e-06, "loss": 0.8756, "step": 2098 }, { "epoch": 0.17886663826161056, "grad_norm": 19.0, "learning_rate": 2.4551385898242525e-06, "loss": 1.028, "step": 2099 }, { "epoch": 0.17895185342991052, "grad_norm": 15.9375, "learning_rate": 2.455093965504029e-06, "loss": 0.7538, "step": 2100 }, { "epoch": 0.17903706859821047, "grad_norm": 18.375, "learning_rate": 2.4550493194065462e-06, "loss": 0.9061, "step": 2101 }, { "epoch": 0.17912228376651043, "grad_norm": 19.375, "learning_rate": 2.45500465153261e-06, "loss": 0.9957, "step": 2102 }, { "epoch": 0.1792074989348104, "grad_norm": 16.375, "learning_rate": 2.454959961883028e-06, "loss": 0.7387, "step": 2103 }, { "epoch": 0.17929271410311035, "grad_norm": 12.875, "learning_rate": 2.454915250458608e-06, "loss": 0.6402, "step": 2104 }, { "epoch": 0.1793779292714103, "grad_norm": 11.5, "learning_rate": 2.4548705172601574e-06, "loss": 0.4893, "step": 2105 }, { "epoch": 0.17946314443971026, "grad_norm": 13.0, "learning_rate": 2.4548257622884853e-06, "loss": 0.6715, "step": 2106 }, { "epoch": 0.17954835960801022, "grad_norm": 13.5625, "learning_rate": 2.4547809855444e-06, "loss": 0.9421, "step": 2107 }, { "epoch": 0.17963357477631017, "grad_norm": 13.875, "learning_rate": 2.454736187028711e-06, "loss": 0.6871, "step": 2108 }, { "epoch": 0.17971878994461013, "grad_norm": 13.125, "learning_rate": 2.4546913667422272e-06, "loss": 0.6145, "step": 2109 }, { "epoch": 0.1798040051129101, "grad_norm": 15.1875, "learning_rate": 2.4546465246857595e-06, "loss": 0.6275, "step": 2110 }, { "epoch": 0.17988922028121004, "grad_norm": 17.375, "learning_rate": 2.4546016608601173e-06, "loss": 0.9303, "step": 2111 }, { "epoch": 0.17997443544951, "grad_norm": 13.125, "learning_rate": 2.454556775266112e-06, "loss": 0.6837, "step": 2112 }, { "epoch": 0.18005965061780996, "grad_norm": 17.25, "learning_rate": 2.454511867904555e-06, "loss": 0.9435, "step": 2113 }, { "epoch": 0.18014486578610991, "grad_norm": 15.5, "learning_rate": 2.4544669387762567e-06, "loss": 0.7194, "step": 2114 }, { "epoch": 0.18023008095440987, "grad_norm": 16.75, "learning_rate": 2.45442198788203e-06, "loss": 1.0713, "step": 2115 }, { "epoch": 0.18031529612270986, "grad_norm": 24.375, "learning_rate": 2.4543770152226857e-06, "loss": 1.5671, "step": 2116 }, { "epoch": 0.1804005112910098, "grad_norm": 21.625, "learning_rate": 2.454332020799039e-06, "loss": 0.9992, "step": 2117 }, { "epoch": 0.18048572645930977, "grad_norm": 13.5, "learning_rate": 2.454287004611901e-06, "loss": 0.7161, "step": 2118 }, { "epoch": 0.18057094162760973, "grad_norm": 15.4375, "learning_rate": 2.454241966662086e-06, "loss": 0.756, "step": 2119 }, { "epoch": 0.18065615679590968, "grad_norm": 12.625, "learning_rate": 2.4541969069504075e-06, "loss": 0.7193, "step": 2120 }, { "epoch": 0.18074137196420964, "grad_norm": 11.4375, "learning_rate": 2.45415182547768e-06, "loss": 0.5944, "step": 2121 }, { "epoch": 0.1808265871325096, "grad_norm": 18.75, "learning_rate": 2.4541067222447184e-06, "loss": 1.2357, "step": 2122 }, { "epoch": 0.18091180230080955, "grad_norm": 14.5, "learning_rate": 2.4540615972523373e-06, "loss": 0.3164, "step": 2123 }, { "epoch": 0.1809970174691095, "grad_norm": 13.25, "learning_rate": 2.4540164505013523e-06, "loss": 0.6552, "step": 2124 }, { "epoch": 0.18108223263740947, "grad_norm": 19.75, "learning_rate": 2.45397128199258e-06, "loss": 1.3354, "step": 2125 }, { "epoch": 0.18116744780570943, "grad_norm": 12.625, "learning_rate": 2.4539260917268354e-06, "loss": 0.708, "step": 2126 }, { "epoch": 0.18125266297400938, "grad_norm": 15.3125, "learning_rate": 2.4538808797049357e-06, "loss": 0.7667, "step": 2127 }, { "epoch": 0.18133787814230934, "grad_norm": 16.0, "learning_rate": 2.4538356459276977e-06, "loss": 0.9075, "step": 2128 }, { "epoch": 0.1814230933106093, "grad_norm": 13.6875, "learning_rate": 2.4537903903959394e-06, "loss": 0.7951, "step": 2129 }, { "epoch": 0.18150830847890925, "grad_norm": 14.125, "learning_rate": 2.4537451131104776e-06, "loss": 0.8273, "step": 2130 }, { "epoch": 0.1815935236472092, "grad_norm": 14.75, "learning_rate": 2.453699814072132e-06, "loss": 0.7433, "step": 2131 }, { "epoch": 0.18167873881550917, "grad_norm": 63.5, "learning_rate": 2.45365449328172e-06, "loss": 1.5492, "step": 2132 }, { "epoch": 0.18176395398380912, "grad_norm": 16.0, "learning_rate": 2.4536091507400608e-06, "loss": 0.9812, "step": 2133 }, { "epoch": 0.18184916915210908, "grad_norm": 14.125, "learning_rate": 2.453563786447974e-06, "loss": 0.7726, "step": 2134 }, { "epoch": 0.18193438432040904, "grad_norm": 15.5, "learning_rate": 2.4535184004062795e-06, "loss": 1.0586, "step": 2135 }, { "epoch": 0.182019599488709, "grad_norm": 14.8125, "learning_rate": 2.453472992615797e-06, "loss": 0.5624, "step": 2136 }, { "epoch": 0.18210481465700895, "grad_norm": 13.0, "learning_rate": 2.4534275630773475e-06, "loss": 0.5086, "step": 2137 }, { "epoch": 0.1821900298253089, "grad_norm": 12.75, "learning_rate": 2.453382111791752e-06, "loss": 0.5329, "step": 2138 }, { "epoch": 0.18227524499360886, "grad_norm": 13.875, "learning_rate": 2.453336638759831e-06, "loss": 0.6857, "step": 2139 }, { "epoch": 0.18236046016190882, "grad_norm": 13.9375, "learning_rate": 2.453291143982407e-06, "loss": 0.8337, "step": 2140 }, { "epoch": 0.18244567533020878, "grad_norm": 21.5, "learning_rate": 2.4532456274603024e-06, "loss": 1.211, "step": 2141 }, { "epoch": 0.18253089049850874, "grad_norm": 14.5625, "learning_rate": 2.4532000891943393e-06, "loss": 0.8509, "step": 2142 }, { "epoch": 0.1826161056668087, "grad_norm": 17.25, "learning_rate": 2.4531545291853407e-06, "loss": 0.831, "step": 2143 }, { "epoch": 0.18270132083510865, "grad_norm": 12.0625, "learning_rate": 2.4531089474341296e-06, "loss": 0.7519, "step": 2144 }, { "epoch": 0.1827865360034086, "grad_norm": 9.5, "learning_rate": 2.45306334394153e-06, "loss": 0.2772, "step": 2145 }, { "epoch": 0.18287175117170856, "grad_norm": 15.75, "learning_rate": 2.4530177187083666e-06, "loss": 0.5885, "step": 2146 }, { "epoch": 0.18295696634000852, "grad_norm": 15.9375, "learning_rate": 2.4529720717354627e-06, "loss": 1.0621, "step": 2147 }, { "epoch": 0.18304218150830848, "grad_norm": 17.375, "learning_rate": 2.4529264030236445e-06, "loss": 0.9031, "step": 2148 }, { "epoch": 0.18312739667660843, "grad_norm": 16.25, "learning_rate": 2.452880712573736e-06, "loss": 0.8009, "step": 2149 }, { "epoch": 0.1832126118449084, "grad_norm": 13.5625, "learning_rate": 2.452835000386563e-06, "loss": 0.7724, "step": 2150 }, { "epoch": 0.18329782701320835, "grad_norm": 21.75, "learning_rate": 2.4527892664629525e-06, "loss": 0.8147, "step": 2151 }, { "epoch": 0.1833830421815083, "grad_norm": 24.75, "learning_rate": 2.452743510803731e-06, "loss": 0.7664, "step": 2152 }, { "epoch": 0.18346825734980826, "grad_norm": 20.75, "learning_rate": 2.452697733409724e-06, "loss": 1.2781, "step": 2153 }, { "epoch": 0.18355347251810822, "grad_norm": 20.625, "learning_rate": 2.4526519342817593e-06, "loss": 1.1239, "step": 2154 }, { "epoch": 0.18363868768640818, "grad_norm": 14.1875, "learning_rate": 2.452606113420666e-06, "loss": 0.7286, "step": 2155 }, { "epoch": 0.18372390285470813, "grad_norm": 19.75, "learning_rate": 2.45256027082727e-06, "loss": 0.8668, "step": 2156 }, { "epoch": 0.1838091180230081, "grad_norm": 17.5, "learning_rate": 2.452514406502401e-06, "loss": 0.8513, "step": 2157 }, { "epoch": 0.18389433319130805, "grad_norm": 13.8125, "learning_rate": 2.452468520446887e-06, "loss": 0.9288, "step": 2158 }, { "epoch": 0.183979548359608, "grad_norm": 28.125, "learning_rate": 2.452422612661558e-06, "loss": 1.1164, "step": 2159 }, { "epoch": 0.18406476352790796, "grad_norm": 14.9375, "learning_rate": 2.452376683147243e-06, "loss": 0.846, "step": 2160 }, { "epoch": 0.18414997869620792, "grad_norm": 13.5625, "learning_rate": 2.4523307319047726e-06, "loss": 0.562, "step": 2161 }, { "epoch": 0.18423519386450787, "grad_norm": 10.4375, "learning_rate": 2.4522847589349768e-06, "loss": 0.561, "step": 2162 }, { "epoch": 0.18432040903280783, "grad_norm": 14.3125, "learning_rate": 2.4522387642386863e-06, "loss": 0.7294, "step": 2163 }, { "epoch": 0.1844056242011078, "grad_norm": 22.375, "learning_rate": 2.4521927478167325e-06, "loss": 1.2182, "step": 2164 }, { "epoch": 0.18449083936940774, "grad_norm": 12.125, "learning_rate": 2.4521467096699464e-06, "loss": 0.5937, "step": 2165 }, { "epoch": 0.1845760545377077, "grad_norm": 13.625, "learning_rate": 2.452100649799161e-06, "loss": 0.7398, "step": 2166 }, { "epoch": 0.18466126970600766, "grad_norm": 30.75, "learning_rate": 2.452054568205208e-06, "loss": 0.9536, "step": 2167 }, { "epoch": 0.18474648487430761, "grad_norm": 11.25, "learning_rate": 2.4520084648889198e-06, "loss": 0.4099, "step": 2168 }, { "epoch": 0.18483170004260757, "grad_norm": 13.0, "learning_rate": 2.45196233985113e-06, "loss": 0.5028, "step": 2169 }, { "epoch": 0.18491691521090753, "grad_norm": 14.75, "learning_rate": 2.451916193092672e-06, "loss": 0.9825, "step": 2170 }, { "epoch": 0.18500213037920749, "grad_norm": 11.625, "learning_rate": 2.45187002461438e-06, "loss": 0.7238, "step": 2171 }, { "epoch": 0.18508734554750744, "grad_norm": 13.0, "learning_rate": 2.451823834417088e-06, "loss": 0.6578, "step": 2172 }, { "epoch": 0.18517256071580743, "grad_norm": 11.5625, "learning_rate": 2.451777622501631e-06, "loss": 0.5017, "step": 2173 }, { "epoch": 0.18525777588410738, "grad_norm": 11.625, "learning_rate": 2.4517313888688437e-06, "loss": 0.5247, "step": 2174 }, { "epoch": 0.18534299105240734, "grad_norm": 20.5, "learning_rate": 2.451685133519562e-06, "loss": 1.0488, "step": 2175 }, { "epoch": 0.1854282062207073, "grad_norm": 15.625, "learning_rate": 2.451638856454621e-06, "loss": 0.972, "step": 2176 }, { "epoch": 0.18551342138900725, "grad_norm": 14.875, "learning_rate": 2.4515925576748578e-06, "loss": 0.6616, "step": 2177 }, { "epoch": 0.1855986365573072, "grad_norm": 13.25, "learning_rate": 2.451546237181109e-06, "loss": 0.6429, "step": 2178 }, { "epoch": 0.18568385172560717, "grad_norm": 20.125, "learning_rate": 2.4514998949742115e-06, "loss": 1.1132, "step": 2179 }, { "epoch": 0.18576906689390713, "grad_norm": 17.25, "learning_rate": 2.4514535310550023e-06, "loss": 1.0648, "step": 2180 }, { "epoch": 0.18585428206220708, "grad_norm": 13.5, "learning_rate": 2.45140714542432e-06, "loss": 1.0407, "step": 2181 }, { "epoch": 0.18593949723050704, "grad_norm": 20.5, "learning_rate": 2.451360738083002e-06, "loss": 0.9394, "step": 2182 }, { "epoch": 0.186024712398807, "grad_norm": 19.875, "learning_rate": 2.4513143090318877e-06, "loss": 1.2023, "step": 2183 }, { "epoch": 0.18610992756710695, "grad_norm": 15.6875, "learning_rate": 2.4512678582718156e-06, "loss": 0.638, "step": 2184 }, { "epoch": 0.1861951427354069, "grad_norm": 14.5625, "learning_rate": 2.4512213858036258e-06, "loss": 0.9932, "step": 2185 }, { "epoch": 0.18628035790370687, "grad_norm": 16.875, "learning_rate": 2.4511748916281575e-06, "loss": 0.5861, "step": 2186 }, { "epoch": 0.18636557307200682, "grad_norm": 12.625, "learning_rate": 2.4511283757462507e-06, "loss": 0.57, "step": 2187 }, { "epoch": 0.18645078824030678, "grad_norm": 14.875, "learning_rate": 2.4510818381587464e-06, "loss": 0.8754, "step": 2188 }, { "epoch": 0.18653600340860674, "grad_norm": 16.5, "learning_rate": 2.451035278866486e-06, "loss": 0.8999, "step": 2189 }, { "epoch": 0.1866212185769067, "grad_norm": 18.25, "learning_rate": 2.4509886978703097e-06, "loss": 0.9735, "step": 2190 }, { "epoch": 0.18670643374520665, "grad_norm": 13.3125, "learning_rate": 2.4509420951710605e-06, "loss": 0.6872, "step": 2191 }, { "epoch": 0.1867916489135066, "grad_norm": 9.625, "learning_rate": 2.45089547076958e-06, "loss": 0.3071, "step": 2192 }, { "epoch": 0.18687686408180657, "grad_norm": 14.125, "learning_rate": 2.45084882466671e-06, "loss": 0.5016, "step": 2193 }, { "epoch": 0.18696207925010652, "grad_norm": 15.1875, "learning_rate": 2.4508021568632952e-06, "loss": 0.5695, "step": 2194 }, { "epoch": 0.18704729441840648, "grad_norm": 17.0, "learning_rate": 2.450755467360177e-06, "loss": 0.7118, "step": 2195 }, { "epoch": 0.18713250958670644, "grad_norm": 12.625, "learning_rate": 2.4507087561582004e-06, "loss": 0.5187, "step": 2196 }, { "epoch": 0.1872177247550064, "grad_norm": 19.75, "learning_rate": 2.4506620232582096e-06, "loss": 0.855, "step": 2197 }, { "epoch": 0.18730293992330635, "grad_norm": 21.875, "learning_rate": 2.450615268661048e-06, "loss": 0.8528, "step": 2198 }, { "epoch": 0.1873881550916063, "grad_norm": 20.25, "learning_rate": 2.4505684923675614e-06, "loss": 1.024, "step": 2199 }, { "epoch": 0.18747337025990626, "grad_norm": 14.6875, "learning_rate": 2.450521694378595e-06, "loss": 0.6702, "step": 2200 }, { "epoch": 0.18755858542820622, "grad_norm": 13.9375, "learning_rate": 2.4504748746949946e-06, "loss": 0.6394, "step": 2201 }, { "epoch": 0.18764380059650618, "grad_norm": 14.75, "learning_rate": 2.4504280333176054e-06, "loss": 0.8527, "step": 2202 }, { "epoch": 0.18772901576480613, "grad_norm": 13.1875, "learning_rate": 2.450381170247275e-06, "loss": 0.6188, "step": 2203 }, { "epoch": 0.1878142309331061, "grad_norm": 18.125, "learning_rate": 2.45033428548485e-06, "loss": 0.4409, "step": 2204 }, { "epoch": 0.18789944610140605, "grad_norm": 17.375, "learning_rate": 2.450287379031177e-06, "loss": 1.0834, "step": 2205 }, { "epoch": 0.187984661269706, "grad_norm": 15.0, "learning_rate": 2.4502404508871042e-06, "loss": 0.7946, "step": 2206 }, { "epoch": 0.18806987643800596, "grad_norm": 14.0625, "learning_rate": 2.4501935010534794e-06, "loss": 0.8715, "step": 2207 }, { "epoch": 0.18815509160630592, "grad_norm": 12.25, "learning_rate": 2.4501465295311512e-06, "loss": 0.5447, "step": 2208 }, { "epoch": 0.18824030677460588, "grad_norm": 19.75, "learning_rate": 2.4500995363209684e-06, "loss": 1.1739, "step": 2209 }, { "epoch": 0.18832552194290583, "grad_norm": 12.75, "learning_rate": 2.4500525214237804e-06, "loss": 0.6727, "step": 2210 }, { "epoch": 0.1884107371112058, "grad_norm": 13.9375, "learning_rate": 2.4500054848404365e-06, "loss": 0.6016, "step": 2211 }, { "epoch": 0.18849595227950575, "grad_norm": 14.875, "learning_rate": 2.4499584265717866e-06, "loss": 0.8378, "step": 2212 }, { "epoch": 0.1885811674478057, "grad_norm": 17.625, "learning_rate": 2.4499113466186812e-06, "loss": 0.6287, "step": 2213 }, { "epoch": 0.18866638261610566, "grad_norm": 16.75, "learning_rate": 2.4498642449819716e-06, "loss": 0.9848, "step": 2214 }, { "epoch": 0.18875159778440562, "grad_norm": 15.875, "learning_rate": 2.4498171216625085e-06, "loss": 1.0045, "step": 2215 }, { "epoch": 0.18883681295270557, "grad_norm": 16.25, "learning_rate": 2.449769976661143e-06, "loss": 0.9238, "step": 2216 }, { "epoch": 0.18892202812100553, "grad_norm": 15.75, "learning_rate": 2.449722809978728e-06, "loss": 1.0048, "step": 2217 }, { "epoch": 0.1890072432893055, "grad_norm": 13.6875, "learning_rate": 2.449675621616115e-06, "loss": 0.5783, "step": 2218 }, { "epoch": 0.18909245845760544, "grad_norm": 12.5, "learning_rate": 2.4496284115741575e-06, "loss": 0.3754, "step": 2219 }, { "epoch": 0.1891776736259054, "grad_norm": 14.1875, "learning_rate": 2.4495811798537082e-06, "loss": 0.916, "step": 2220 }, { "epoch": 0.18926288879420536, "grad_norm": 17.0, "learning_rate": 2.4495339264556208e-06, "loss": 0.7974, "step": 2221 }, { "epoch": 0.18934810396250532, "grad_norm": 13.6875, "learning_rate": 2.449486651380749e-06, "loss": 0.761, "step": 2222 }, { "epoch": 0.18943331913080527, "grad_norm": 15.3125, "learning_rate": 2.4494393546299473e-06, "loss": 1.0153, "step": 2223 }, { "epoch": 0.18951853429910523, "grad_norm": 14.0, "learning_rate": 2.4493920362040703e-06, "loss": 0.6706, "step": 2224 }, { "epoch": 0.18960374946740519, "grad_norm": 13.1875, "learning_rate": 2.449344696103973e-06, "loss": 0.7151, "step": 2225 }, { "epoch": 0.18968896463570514, "grad_norm": 12.75, "learning_rate": 2.449297334330511e-06, "loss": 0.4276, "step": 2226 }, { "epoch": 0.1897741798040051, "grad_norm": 15.4375, "learning_rate": 2.4492499508845403e-06, "loss": 0.7543, "step": 2227 }, { "epoch": 0.18985939497230506, "grad_norm": 11.9375, "learning_rate": 2.449202545766917e-06, "loss": 0.6175, "step": 2228 }, { "epoch": 0.18994461014060504, "grad_norm": 13.5, "learning_rate": 2.4491551189784978e-06, "loss": 0.7408, "step": 2229 }, { "epoch": 0.190029825308905, "grad_norm": 13.875, "learning_rate": 2.44910767052014e-06, "loss": 0.7936, "step": 2230 }, { "epoch": 0.19011504047720496, "grad_norm": 17.375, "learning_rate": 2.4490602003927006e-06, "loss": 0.5238, "step": 2231 }, { "epoch": 0.1902002556455049, "grad_norm": 13.6875, "learning_rate": 2.4490127085970376e-06, "loss": 0.8359, "step": 2232 }, { "epoch": 0.19028547081380487, "grad_norm": 11.5625, "learning_rate": 2.4489651951340093e-06, "loss": 0.5499, "step": 2233 }, { "epoch": 0.19037068598210483, "grad_norm": 24.875, "learning_rate": 2.448917660004474e-06, "loss": 1.2219, "step": 2234 }, { "epoch": 0.19045590115040478, "grad_norm": 14.4375, "learning_rate": 2.4488701032092917e-06, "loss": 0.8399, "step": 2235 }, { "epoch": 0.19054111631870474, "grad_norm": 19.125, "learning_rate": 2.44882252474932e-06, "loss": 0.8813, "step": 2236 }, { "epoch": 0.1906263314870047, "grad_norm": 11.6875, "learning_rate": 2.4487749246254206e-06, "loss": 0.4664, "step": 2237 }, { "epoch": 0.19071154665530465, "grad_norm": 14.375, "learning_rate": 2.448727302838453e-06, "loss": 0.8084, "step": 2238 }, { "epoch": 0.1907967618236046, "grad_norm": 15.5, "learning_rate": 2.448679659389277e-06, "loss": 0.3714, "step": 2239 }, { "epoch": 0.19088197699190457, "grad_norm": 13.1875, "learning_rate": 2.4486319942787547e-06, "loss": 0.6104, "step": 2240 }, { "epoch": 0.19096719216020452, "grad_norm": 21.375, "learning_rate": 2.4485843075077464e-06, "loss": 0.9878, "step": 2241 }, { "epoch": 0.19105240732850448, "grad_norm": 18.5, "learning_rate": 2.4485365990771145e-06, "loss": 1.0321, "step": 2242 }, { "epoch": 0.19113762249680444, "grad_norm": 15.0625, "learning_rate": 2.448488868987721e-06, "loss": 0.8145, "step": 2243 }, { "epoch": 0.1912228376651044, "grad_norm": 18.5, "learning_rate": 2.448441117240429e-06, "loss": 1.2219, "step": 2244 }, { "epoch": 0.19130805283340435, "grad_norm": 15.0, "learning_rate": 2.4483933438361005e-06, "loss": 0.6343, "step": 2245 }, { "epoch": 0.1913932680017043, "grad_norm": 15.5625, "learning_rate": 2.4483455487755992e-06, "loss": 0.7822, "step": 2246 }, { "epoch": 0.19147848317000427, "grad_norm": 15.875, "learning_rate": 2.448297732059789e-06, "loss": 0.8574, "step": 2247 }, { "epoch": 0.19156369833830422, "grad_norm": 15.25, "learning_rate": 2.4482498936895337e-06, "loss": 1.0431, "step": 2248 }, { "epoch": 0.19164891350660418, "grad_norm": 15.9375, "learning_rate": 2.4482020336656978e-06, "loss": 0.7653, "step": 2249 }, { "epoch": 0.19173412867490414, "grad_norm": 16.625, "learning_rate": 2.4481541519891464e-06, "loss": 0.869, "step": 2250 }, { "epoch": 0.1918193438432041, "grad_norm": 14.125, "learning_rate": 2.4481062486607447e-06, "loss": 0.7833, "step": 2251 }, { "epoch": 0.19190455901150405, "grad_norm": 16.875, "learning_rate": 2.448058323681358e-06, "loss": 0.9384, "step": 2252 }, { "epoch": 0.191989774179804, "grad_norm": 25.0, "learning_rate": 2.4480103770518528e-06, "loss": 0.7465, "step": 2253 }, { "epoch": 0.19207498934810396, "grad_norm": 16.0, "learning_rate": 2.447962408773095e-06, "loss": 0.449, "step": 2254 }, { "epoch": 0.19216020451640392, "grad_norm": 9.125, "learning_rate": 2.4479144188459524e-06, "loss": 0.3205, "step": 2255 }, { "epoch": 0.19224541968470388, "grad_norm": 14.6875, "learning_rate": 2.4478664072712916e-06, "loss": 0.8656, "step": 2256 }, { "epoch": 0.19233063485300383, "grad_norm": 14.9375, "learning_rate": 2.4478183740499805e-06, "loss": 0.9468, "step": 2257 }, { "epoch": 0.1924158500213038, "grad_norm": 12.375, "learning_rate": 2.447770319182886e-06, "loss": 0.5751, "step": 2258 }, { "epoch": 0.19250106518960375, "grad_norm": 26.125, "learning_rate": 2.4477222426708783e-06, "loss": 1.1603, "step": 2259 }, { "epoch": 0.1925862803579037, "grad_norm": 11.125, "learning_rate": 2.447674144514825e-06, "loss": 0.4086, "step": 2260 }, { "epoch": 0.19267149552620366, "grad_norm": 16.125, "learning_rate": 2.4476260247155955e-06, "loss": 0.9462, "step": 2261 }, { "epoch": 0.19275671069450362, "grad_norm": 23.625, "learning_rate": 2.447577883274059e-06, "loss": 0.9794, "step": 2262 }, { "epoch": 0.19284192586280358, "grad_norm": 13.75, "learning_rate": 2.4475297201910865e-06, "loss": 0.6012, "step": 2263 }, { "epoch": 0.19292714103110353, "grad_norm": 12.9375, "learning_rate": 2.4474815354675473e-06, "loss": 0.7466, "step": 2264 }, { "epoch": 0.1930123561994035, "grad_norm": 30.375, "learning_rate": 2.4474333291043127e-06, "loss": 1.0961, "step": 2265 }, { "epoch": 0.19309757136770345, "grad_norm": 19.625, "learning_rate": 2.4473851011022544e-06, "loss": 0.5913, "step": 2266 }, { "epoch": 0.1931827865360034, "grad_norm": 32.25, "learning_rate": 2.4473368514622426e-06, "loss": 1.0623, "step": 2267 }, { "epoch": 0.19326800170430336, "grad_norm": 16.375, "learning_rate": 2.44728858018515e-06, "loss": 0.8422, "step": 2268 }, { "epoch": 0.19335321687260332, "grad_norm": 13.375, "learning_rate": 2.4472402872718486e-06, "loss": 0.5212, "step": 2269 }, { "epoch": 0.19343843204090327, "grad_norm": 15.6875, "learning_rate": 2.4471919727232113e-06, "loss": 0.7781, "step": 2270 }, { "epoch": 0.19352364720920323, "grad_norm": 22.625, "learning_rate": 2.447143636540111e-06, "loss": 0.7535, "step": 2271 }, { "epoch": 0.1936088623775032, "grad_norm": 24.25, "learning_rate": 2.447095278723422e-06, "loss": 1.0257, "step": 2272 }, { "epoch": 0.19369407754580314, "grad_norm": 12.375, "learning_rate": 2.4470468992740173e-06, "loss": 0.6115, "step": 2273 }, { "epoch": 0.1937792927141031, "grad_norm": 15.875, "learning_rate": 2.4469984981927714e-06, "loss": 0.7066, "step": 2274 }, { "epoch": 0.19386450788240306, "grad_norm": 13.0625, "learning_rate": 2.446950075480559e-06, "loss": 0.5466, "step": 2275 }, { "epoch": 0.19394972305070302, "grad_norm": 14.4375, "learning_rate": 2.4469016311382547e-06, "loss": 0.7272, "step": 2276 }, { "epoch": 0.19403493821900297, "grad_norm": 18.75, "learning_rate": 2.4468531651667347e-06, "loss": 1.1441, "step": 2277 }, { "epoch": 0.19412015338730293, "grad_norm": 16.5, "learning_rate": 2.4468046775668746e-06, "loss": 0.8277, "step": 2278 }, { "epoch": 0.1942053685556029, "grad_norm": 18.875, "learning_rate": 2.4467561683395504e-06, "loss": 0.9888, "step": 2279 }, { "epoch": 0.19429058372390284, "grad_norm": 13.0625, "learning_rate": 2.446707637485639e-06, "loss": 0.6971, "step": 2280 }, { "epoch": 0.1943757988922028, "grad_norm": 14.625, "learning_rate": 2.446659085006017e-06, "loss": 0.6163, "step": 2281 }, { "epoch": 0.19446101406050276, "grad_norm": 10.3125, "learning_rate": 2.4466105109015618e-06, "loss": 0.754, "step": 2282 }, { "epoch": 0.19454622922880271, "grad_norm": 11.25, "learning_rate": 2.4465619151731513e-06, "loss": 0.5766, "step": 2283 }, { "epoch": 0.19463144439710267, "grad_norm": 14.0625, "learning_rate": 2.446513297821664e-06, "loss": 0.8218, "step": 2284 }, { "epoch": 0.19471665956540266, "grad_norm": 14.3125, "learning_rate": 2.4464646588479784e-06, "loss": 0.8447, "step": 2285 }, { "epoch": 0.1948018747337026, "grad_norm": 15.5, "learning_rate": 2.446415998252973e-06, "loss": 0.9637, "step": 2286 }, { "epoch": 0.19488708990200257, "grad_norm": 17.125, "learning_rate": 2.4463673160375274e-06, "loss": 0.4817, "step": 2287 }, { "epoch": 0.19497230507030253, "grad_norm": 15.625, "learning_rate": 2.4463186122025217e-06, "loss": 0.6415, "step": 2288 }, { "epoch": 0.19505752023860248, "grad_norm": 11.6875, "learning_rate": 2.4462698867488356e-06, "loss": 0.6263, "step": 2289 }, { "epoch": 0.19514273540690244, "grad_norm": 12.8125, "learning_rate": 2.4462211396773495e-06, "loss": 0.9363, "step": 2290 }, { "epoch": 0.1952279505752024, "grad_norm": 16.25, "learning_rate": 2.4461723709889448e-06, "loss": 0.7639, "step": 2291 }, { "epoch": 0.19531316574350235, "grad_norm": 12.0, "learning_rate": 2.446123580684502e-06, "loss": 0.6631, "step": 2292 }, { "epoch": 0.1953983809118023, "grad_norm": 14.625, "learning_rate": 2.4460747687649035e-06, "loss": 0.9672, "step": 2293 }, { "epoch": 0.19548359608010227, "grad_norm": 15.8125, "learning_rate": 2.4460259352310313e-06, "loss": 0.8038, "step": 2294 }, { "epoch": 0.19556881124840222, "grad_norm": 13.0, "learning_rate": 2.445977080083768e-06, "loss": 0.4675, "step": 2295 }, { "epoch": 0.19565402641670218, "grad_norm": 10.4375, "learning_rate": 2.4459282033239957e-06, "loss": 0.3959, "step": 2296 }, { "epoch": 0.19573924158500214, "grad_norm": 15.6875, "learning_rate": 2.4458793049525987e-06, "loss": 0.9084, "step": 2297 }, { "epoch": 0.1958244567533021, "grad_norm": 16.75, "learning_rate": 2.4458303849704597e-06, "loss": 0.7232, "step": 2298 }, { "epoch": 0.19590967192160205, "grad_norm": 13.875, "learning_rate": 2.4457814433784633e-06, "loss": 0.722, "step": 2299 }, { "epoch": 0.195994887089902, "grad_norm": 16.125, "learning_rate": 2.4457324801774934e-06, "loss": 1.0233, "step": 2300 }, { "epoch": 0.19608010225820197, "grad_norm": 15.25, "learning_rate": 2.4456834953684357e-06, "loss": 0.5372, "step": 2301 }, { "epoch": 0.19616531742650192, "grad_norm": 13.375, "learning_rate": 2.445634488952175e-06, "loss": 0.7053, "step": 2302 }, { "epoch": 0.19625053259480188, "grad_norm": 22.875, "learning_rate": 2.4455854609295963e-06, "loss": 1.2031, "step": 2303 }, { "epoch": 0.19633574776310184, "grad_norm": 11.875, "learning_rate": 2.4455364113015863e-06, "loss": 0.6793, "step": 2304 }, { "epoch": 0.1964209629314018, "grad_norm": 16.5, "learning_rate": 2.445487340069031e-06, "loss": 1.0749, "step": 2305 }, { "epoch": 0.19650617809970175, "grad_norm": 13.0625, "learning_rate": 2.4454382472328174e-06, "loss": 0.8275, "step": 2306 }, { "epoch": 0.1965913932680017, "grad_norm": 13.1875, "learning_rate": 2.4453891327938324e-06, "loss": 0.8679, "step": 2307 }, { "epoch": 0.19667660843630166, "grad_norm": 21.625, "learning_rate": 2.445339996752964e-06, "loss": 1.1166, "step": 2308 }, { "epoch": 0.19676182360460162, "grad_norm": 12.25, "learning_rate": 2.4452908391111e-06, "loss": 0.4847, "step": 2309 }, { "epoch": 0.19684703877290158, "grad_norm": 20.375, "learning_rate": 2.445241659869128e-06, "loss": 1.0299, "step": 2310 }, { "epoch": 0.19693225394120153, "grad_norm": 12.5, "learning_rate": 2.4451924590279374e-06, "loss": 0.5539, "step": 2311 }, { "epoch": 0.1970174691095015, "grad_norm": 20.5, "learning_rate": 2.4451432365884175e-06, "loss": 0.8347, "step": 2312 }, { "epoch": 0.19710268427780145, "grad_norm": 16.25, "learning_rate": 2.4450939925514577e-06, "loss": 0.7147, "step": 2313 }, { "epoch": 0.1971878994461014, "grad_norm": 14.25, "learning_rate": 2.445044726917947e-06, "loss": 0.9938, "step": 2314 }, { "epoch": 0.19727311461440136, "grad_norm": 16.375, "learning_rate": 2.444995439688777e-06, "loss": 0.9131, "step": 2315 }, { "epoch": 0.19735832978270132, "grad_norm": 19.5, "learning_rate": 2.4449461308648375e-06, "loss": 0.9354, "step": 2316 }, { "epoch": 0.19744354495100128, "grad_norm": 12.875, "learning_rate": 2.44489680044702e-06, "loss": 0.6274, "step": 2317 }, { "epoch": 0.19752876011930123, "grad_norm": 13.0, "learning_rate": 2.4448474484362157e-06, "loss": 0.4424, "step": 2318 }, { "epoch": 0.1976139752876012, "grad_norm": 17.0, "learning_rate": 2.444798074833316e-06, "loss": 0.7284, "step": 2319 }, { "epoch": 0.19769919045590115, "grad_norm": 20.75, "learning_rate": 2.444748679639214e-06, "loss": 1.13, "step": 2320 }, { "epoch": 0.1977844056242011, "grad_norm": 15.375, "learning_rate": 2.444699262854802e-06, "loss": 1.0087, "step": 2321 }, { "epoch": 0.19786962079250106, "grad_norm": 13.5, "learning_rate": 2.444649824480973e-06, "loss": 0.7771, "step": 2322 }, { "epoch": 0.19795483596080102, "grad_norm": 22.875, "learning_rate": 2.4446003645186204e-06, "loss": 1.2427, "step": 2323 }, { "epoch": 0.19804005112910097, "grad_norm": 16.25, "learning_rate": 2.4445508829686374e-06, "loss": 0.9601, "step": 2324 }, { "epoch": 0.19812526629740093, "grad_norm": 13.1875, "learning_rate": 2.4445013798319194e-06, "loss": 0.8291, "step": 2325 }, { "epoch": 0.1982104814657009, "grad_norm": 19.0, "learning_rate": 2.44445185510936e-06, "loss": 1.0551, "step": 2326 }, { "epoch": 0.19829569663400085, "grad_norm": 13.8125, "learning_rate": 2.444402308801855e-06, "loss": 0.5958, "step": 2327 }, { "epoch": 0.1983809118023008, "grad_norm": 16.625, "learning_rate": 2.4443527409102986e-06, "loss": 1.0035, "step": 2328 }, { "epoch": 0.19846612697060076, "grad_norm": 12.3125, "learning_rate": 2.4443031514355873e-06, "loss": 0.6038, "step": 2329 }, { "epoch": 0.19855134213890072, "grad_norm": 15.625, "learning_rate": 2.444253540378617e-06, "loss": 1.0203, "step": 2330 }, { "epoch": 0.19863655730720067, "grad_norm": 20.0, "learning_rate": 2.4442039077402845e-06, "loss": 1.0133, "step": 2331 }, { "epoch": 0.19872177247550063, "grad_norm": 17.125, "learning_rate": 2.444154253521486e-06, "loss": 0.9023, "step": 2332 }, { "epoch": 0.1988069876438006, "grad_norm": 15.3125, "learning_rate": 2.4441045777231197e-06, "loss": 1.0655, "step": 2333 }, { "epoch": 0.19889220281210054, "grad_norm": 21.75, "learning_rate": 2.4440548803460835e-06, "loss": 1.1842, "step": 2334 }, { "epoch": 0.1989774179804005, "grad_norm": 18.5, "learning_rate": 2.444005161391274e-06, "loss": 1.0103, "step": 2335 }, { "epoch": 0.19906263314870046, "grad_norm": 21.75, "learning_rate": 2.443955420859591e-06, "loss": 0.8935, "step": 2336 }, { "epoch": 0.19914784831700041, "grad_norm": 17.75, "learning_rate": 2.4439056587519327e-06, "loss": 1.0824, "step": 2337 }, { "epoch": 0.19923306348530037, "grad_norm": 14.0, "learning_rate": 2.443855875069199e-06, "loss": 0.8669, "step": 2338 }, { "epoch": 0.19931827865360033, "grad_norm": 14.375, "learning_rate": 2.443806069812289e-06, "loss": 0.7385, "step": 2339 }, { "epoch": 0.19940349382190029, "grad_norm": 13.5, "learning_rate": 2.4437562429821026e-06, "loss": 0.7603, "step": 2340 }, { "epoch": 0.19948870899020024, "grad_norm": 12.5, "learning_rate": 2.4437063945795405e-06, "loss": 0.4115, "step": 2341 }, { "epoch": 0.19957392415850023, "grad_norm": 21.875, "learning_rate": 2.4436565246055035e-06, "loss": 1.0797, "step": 2342 }, { "epoch": 0.19965913932680018, "grad_norm": 13.75, "learning_rate": 2.443606633060893e-06, "loss": 0.9298, "step": 2343 }, { "epoch": 0.19974435449510014, "grad_norm": 11.4375, "learning_rate": 2.44355671994661e-06, "loss": 0.4778, "step": 2344 }, { "epoch": 0.1998295696634001, "grad_norm": 18.625, "learning_rate": 2.443506785263557e-06, "loss": 0.786, "step": 2345 }, { "epoch": 0.19991478483170005, "grad_norm": 15.625, "learning_rate": 2.4434568290126364e-06, "loss": 0.7459, "step": 2346 }, { "epoch": 0.2, "grad_norm": 20.5, "learning_rate": 2.443406851194751e-06, "loss": 0.77, "step": 2347 }, { "epoch": 0.20008521516829997, "grad_norm": 19.125, "learning_rate": 2.443356851810803e-06, "loss": 0.9424, "step": 2348 }, { "epoch": 0.20017043033659992, "grad_norm": 14.25, "learning_rate": 2.4433068308616976e-06, "loss": 0.5636, "step": 2349 }, { "epoch": 0.20025564550489988, "grad_norm": 15.25, "learning_rate": 2.443256788348337e-06, "loss": 0.6556, "step": 2350 }, { "epoch": 0.20034086067319984, "grad_norm": 16.375, "learning_rate": 2.443206724271627e-06, "loss": 1.2988, "step": 2351 }, { "epoch": 0.2004260758414998, "grad_norm": 14.4375, "learning_rate": 2.443156638632471e-06, "loss": 0.6152, "step": 2352 }, { "epoch": 0.20051129100979975, "grad_norm": 14.5, "learning_rate": 2.4431065314317747e-06, "loss": 0.6758, "step": 2353 }, { "epoch": 0.2005965061780997, "grad_norm": 17.125, "learning_rate": 2.4430564026704444e-06, "loss": 1.0294, "step": 2354 }, { "epoch": 0.20068172134639967, "grad_norm": 13.4375, "learning_rate": 2.4430062523493845e-06, "loss": 0.7881, "step": 2355 }, { "epoch": 0.20076693651469962, "grad_norm": 12.5, "learning_rate": 2.442956080469502e-06, "loss": 0.5759, "step": 2356 }, { "epoch": 0.20085215168299958, "grad_norm": 13.875, "learning_rate": 2.442905887031704e-06, "loss": 0.7606, "step": 2357 }, { "epoch": 0.20093736685129954, "grad_norm": 12.75, "learning_rate": 2.442855672036897e-06, "loss": 0.6191, "step": 2358 }, { "epoch": 0.2010225820195995, "grad_norm": 16.375, "learning_rate": 2.4428054354859878e-06, "loss": 0.7308, "step": 2359 }, { "epoch": 0.20110779718789945, "grad_norm": 17.25, "learning_rate": 2.442755177379885e-06, "loss": 0.999, "step": 2360 }, { "epoch": 0.2011930123561994, "grad_norm": 25.875, "learning_rate": 2.4427048977194975e-06, "loss": 1.0313, "step": 2361 }, { "epoch": 0.20127822752449936, "grad_norm": 17.125, "learning_rate": 2.442654596505733e-06, "loss": 1.0347, "step": 2362 }, { "epoch": 0.20136344269279932, "grad_norm": 12.125, "learning_rate": 2.4426042737395e-06, "loss": 0.6473, "step": 2363 }, { "epoch": 0.20144865786109928, "grad_norm": 17.0, "learning_rate": 2.4425539294217087e-06, "loss": 0.6536, "step": 2364 }, { "epoch": 0.20153387302939924, "grad_norm": 22.5, "learning_rate": 2.442503563553269e-06, "loss": 1.0198, "step": 2365 }, { "epoch": 0.2016190881976992, "grad_norm": 17.0, "learning_rate": 2.4424531761350906e-06, "loss": 0.5529, "step": 2366 }, { "epoch": 0.20170430336599915, "grad_norm": 13.75, "learning_rate": 2.4424027671680845e-06, "loss": 0.4424, "step": 2367 }, { "epoch": 0.2017895185342991, "grad_norm": 21.625, "learning_rate": 2.442352336653161e-06, "loss": 1.1686, "step": 2368 }, { "epoch": 0.20187473370259906, "grad_norm": 9.75, "learning_rate": 2.4423018845912315e-06, "loss": 0.3541, "step": 2369 }, { "epoch": 0.20195994887089902, "grad_norm": 36.0, "learning_rate": 2.4422514109832083e-06, "loss": 0.951, "step": 2370 }, { "epoch": 0.20204516403919898, "grad_norm": 11.5625, "learning_rate": 2.4422009158300035e-06, "loss": 0.4967, "step": 2371 }, { "epoch": 0.20213037920749893, "grad_norm": 14.25, "learning_rate": 2.442150399132529e-06, "loss": 0.7606, "step": 2372 }, { "epoch": 0.2022155943757989, "grad_norm": 13.25, "learning_rate": 2.442099860891698e-06, "loss": 0.504, "step": 2373 }, { "epoch": 0.20230080954409885, "grad_norm": 10.25, "learning_rate": 2.442049301108424e-06, "loss": 0.3663, "step": 2374 }, { "epoch": 0.2023860247123988, "grad_norm": 9.875, "learning_rate": 2.4419987197836204e-06, "loss": 0.4461, "step": 2375 }, { "epoch": 0.20247123988069876, "grad_norm": 17.375, "learning_rate": 2.4419481169182006e-06, "loss": 0.7311, "step": 2376 }, { "epoch": 0.20255645504899872, "grad_norm": 19.125, "learning_rate": 2.4418974925130805e-06, "loss": 1.0781, "step": 2377 }, { "epoch": 0.20264167021729868, "grad_norm": 16.125, "learning_rate": 2.4418468465691743e-06, "loss": 0.7007, "step": 2378 }, { "epoch": 0.20272688538559863, "grad_norm": 20.75, "learning_rate": 2.4417961790873963e-06, "loss": 0.8812, "step": 2379 }, { "epoch": 0.2028121005538986, "grad_norm": 11.875, "learning_rate": 2.4417454900686634e-06, "loss": 0.5282, "step": 2380 }, { "epoch": 0.20289731572219855, "grad_norm": 12.6875, "learning_rate": 2.441694779513891e-06, "loss": 0.6174, "step": 2381 }, { "epoch": 0.2029825308904985, "grad_norm": 30.0, "learning_rate": 2.4416440474239953e-06, "loss": 1.0134, "step": 2382 }, { "epoch": 0.20306774605879846, "grad_norm": 17.25, "learning_rate": 2.4415932937998936e-06, "loss": 0.6321, "step": 2383 }, { "epoch": 0.20315296122709842, "grad_norm": 10.625, "learning_rate": 2.441542518642503e-06, "loss": 0.4032, "step": 2384 }, { "epoch": 0.20323817639539837, "grad_norm": 15.25, "learning_rate": 2.441491721952741e-06, "loss": 0.8512, "step": 2385 }, { "epoch": 0.20332339156369833, "grad_norm": 18.75, "learning_rate": 2.4414409037315255e-06, "loss": 0.8377, "step": 2386 }, { "epoch": 0.2034086067319983, "grad_norm": 13.1875, "learning_rate": 2.4413900639797745e-06, "loss": 0.7815, "step": 2387 }, { "epoch": 0.20349382190029824, "grad_norm": 12.5, "learning_rate": 2.4413392026984066e-06, "loss": 0.7139, "step": 2388 }, { "epoch": 0.2035790370685982, "grad_norm": 14.75, "learning_rate": 2.441288319888342e-06, "loss": 0.5836, "step": 2389 }, { "epoch": 0.20366425223689816, "grad_norm": 10.0, "learning_rate": 2.4412374155504993e-06, "loss": 0.3388, "step": 2390 }, { "epoch": 0.20374946740519811, "grad_norm": 19.875, "learning_rate": 2.4411864896857986e-06, "loss": 1.0113, "step": 2391 }, { "epoch": 0.20383468257349807, "grad_norm": 15.5, "learning_rate": 2.4411355422951607e-06, "loss": 0.9502, "step": 2392 }, { "epoch": 0.20391989774179803, "grad_norm": 34.5, "learning_rate": 2.4410845733795047e-06, "loss": 0.6113, "step": 2393 }, { "epoch": 0.20400511291009799, "grad_norm": 13.4375, "learning_rate": 2.4410335829397535e-06, "loss": 0.8717, "step": 2394 }, { "epoch": 0.20409032807839794, "grad_norm": 13.5, "learning_rate": 2.4409825709768277e-06, "loss": 0.9236, "step": 2395 }, { "epoch": 0.2041755432466979, "grad_norm": 23.75, "learning_rate": 2.440931537491649e-06, "loss": 0.8897, "step": 2396 }, { "epoch": 0.20426075841499786, "grad_norm": 11.875, "learning_rate": 2.44088048248514e-06, "loss": 0.5197, "step": 2397 }, { "epoch": 0.20434597358329784, "grad_norm": 17.25, "learning_rate": 2.440829405958223e-06, "loss": 1.1689, "step": 2398 }, { "epoch": 0.2044311887515978, "grad_norm": 17.25, "learning_rate": 2.4407783079118214e-06, "loss": 1.0093, "step": 2399 }, { "epoch": 0.20451640391989775, "grad_norm": 12.75, "learning_rate": 2.440727188346858e-06, "loss": 0.4181, "step": 2400 }, { "epoch": 0.2046016190881977, "grad_norm": 11.125, "learning_rate": 2.4406760472642576e-06, "loss": 0.6745, "step": 2401 }, { "epoch": 0.20468683425649767, "grad_norm": 14.5625, "learning_rate": 2.440624884664943e-06, "loss": 0.7777, "step": 2402 }, { "epoch": 0.20477204942479763, "grad_norm": 12.875, "learning_rate": 2.44057370054984e-06, "loss": 0.6037, "step": 2403 }, { "epoch": 0.20485726459309758, "grad_norm": 13.5, "learning_rate": 2.4405224949198725e-06, "loss": 0.7889, "step": 2404 }, { "epoch": 0.20494247976139754, "grad_norm": 18.625, "learning_rate": 2.4404712677759667e-06, "loss": 0.6122, "step": 2405 }, { "epoch": 0.2050276949296975, "grad_norm": 21.5, "learning_rate": 2.440420019119048e-06, "loss": 0.9379, "step": 2406 }, { "epoch": 0.20511291009799745, "grad_norm": 19.875, "learning_rate": 2.440368748950042e-06, "loss": 1.0805, "step": 2407 }, { "epoch": 0.2051981252662974, "grad_norm": 12.75, "learning_rate": 2.4403174572698764e-06, "loss": 0.7804, "step": 2408 }, { "epoch": 0.20528334043459737, "grad_norm": 13.8125, "learning_rate": 2.440266144079477e-06, "loss": 0.8601, "step": 2409 }, { "epoch": 0.20536855560289732, "grad_norm": 11.6875, "learning_rate": 2.4402148093797717e-06, "loss": 0.5067, "step": 2410 }, { "epoch": 0.20545377077119728, "grad_norm": 19.125, "learning_rate": 2.440163453171688e-06, "loss": 1.0207, "step": 2411 }, { "epoch": 0.20553898593949724, "grad_norm": 19.0, "learning_rate": 2.4401120754561535e-06, "loss": 0.9645, "step": 2412 }, { "epoch": 0.2056242011077972, "grad_norm": 11.625, "learning_rate": 2.4400606762340973e-06, "loss": 0.3906, "step": 2413 }, { "epoch": 0.20570941627609715, "grad_norm": 17.375, "learning_rate": 2.4400092555064475e-06, "loss": 0.7098, "step": 2414 }, { "epoch": 0.2057946314443971, "grad_norm": 11.0625, "learning_rate": 2.4399578132741343e-06, "loss": 0.6405, "step": 2415 }, { "epoch": 0.20587984661269706, "grad_norm": 14.0625, "learning_rate": 2.439906349538087e-06, "loss": 0.6483, "step": 2416 }, { "epoch": 0.20596506178099702, "grad_norm": 16.625, "learning_rate": 2.439854864299235e-06, "loss": 0.6714, "step": 2417 }, { "epoch": 0.20605027694929698, "grad_norm": 17.625, "learning_rate": 2.4398033575585095e-06, "loss": 0.7971, "step": 2418 }, { "epoch": 0.20613549211759694, "grad_norm": 13.375, "learning_rate": 2.4397518293168403e-06, "loss": 0.5174, "step": 2419 }, { "epoch": 0.2062207072858969, "grad_norm": 17.0, "learning_rate": 2.4397002795751595e-06, "loss": 0.9742, "step": 2420 }, { "epoch": 0.20630592245419685, "grad_norm": 20.375, "learning_rate": 2.4396487083343984e-06, "loss": 0.95, "step": 2421 }, { "epoch": 0.2063911376224968, "grad_norm": 11.9375, "learning_rate": 2.439597115595489e-06, "loss": 0.4189, "step": 2422 }, { "epoch": 0.20647635279079676, "grad_norm": 10.1875, "learning_rate": 2.4395455013593634e-06, "loss": 0.4058, "step": 2423 }, { "epoch": 0.20656156795909672, "grad_norm": 10.25, "learning_rate": 2.439493865626954e-06, "loss": 0.3387, "step": 2424 }, { "epoch": 0.20664678312739668, "grad_norm": 37.25, "learning_rate": 2.439442208399195e-06, "loss": 0.9265, "step": 2425 }, { "epoch": 0.20673199829569663, "grad_norm": 23.625, "learning_rate": 2.4393905296770183e-06, "loss": 0.9737, "step": 2426 }, { "epoch": 0.2068172134639966, "grad_norm": 13.625, "learning_rate": 2.4393388294613596e-06, "loss": 0.8829, "step": 2427 }, { "epoch": 0.20690242863229655, "grad_norm": 14.5, "learning_rate": 2.4392871077531516e-06, "loss": 0.6078, "step": 2428 }, { "epoch": 0.2069876438005965, "grad_norm": 12.1875, "learning_rate": 2.43923536455333e-06, "loss": 0.4589, "step": 2429 }, { "epoch": 0.20707285896889646, "grad_norm": 16.375, "learning_rate": 2.4391835998628295e-06, "loss": 1.1341, "step": 2430 }, { "epoch": 0.20715807413719642, "grad_norm": 16.875, "learning_rate": 2.4391318136825855e-06, "loss": 0.7221, "step": 2431 }, { "epoch": 0.20724328930549638, "grad_norm": 17.25, "learning_rate": 2.439080006013534e-06, "loss": 0.7102, "step": 2432 }, { "epoch": 0.20732850447379633, "grad_norm": 15.5625, "learning_rate": 2.4390281768566108e-06, "loss": 0.7828, "step": 2433 }, { "epoch": 0.2074137196420963, "grad_norm": 13.0625, "learning_rate": 2.4389763262127532e-06, "loss": 0.7476, "step": 2434 }, { "epoch": 0.20749893481039625, "grad_norm": 11.9375, "learning_rate": 2.438924454082897e-06, "loss": 0.5536, "step": 2435 }, { "epoch": 0.2075841499786962, "grad_norm": 14.125, "learning_rate": 2.4388725604679807e-06, "loss": 0.7197, "step": 2436 }, { "epoch": 0.20766936514699616, "grad_norm": 14.3125, "learning_rate": 2.4388206453689422e-06, "loss": 0.7193, "step": 2437 }, { "epoch": 0.20775458031529612, "grad_norm": 14.9375, "learning_rate": 2.438768708786719e-06, "loss": 0.9624, "step": 2438 }, { "epoch": 0.20783979548359607, "grad_norm": 10.3125, "learning_rate": 2.4387167507222494e-06, "loss": 0.419, "step": 2439 }, { "epoch": 0.20792501065189603, "grad_norm": 16.5, "learning_rate": 2.4386647711764732e-06, "loss": 0.9562, "step": 2440 }, { "epoch": 0.208010225820196, "grad_norm": 15.375, "learning_rate": 2.438612770150329e-06, "loss": 0.829, "step": 2441 }, { "epoch": 0.20809544098849594, "grad_norm": 12.5625, "learning_rate": 2.438560747644757e-06, "loss": 0.5211, "step": 2442 }, { "epoch": 0.2081806561567959, "grad_norm": 19.375, "learning_rate": 2.438508703660697e-06, "loss": 1.0678, "step": 2443 }, { "epoch": 0.20826587132509586, "grad_norm": 11.6875, "learning_rate": 2.4384566381990897e-06, "loss": 0.4248, "step": 2444 }, { "epoch": 0.20835108649339582, "grad_norm": 17.75, "learning_rate": 2.438404551260876e-06, "loss": 0.7505, "step": 2445 }, { "epoch": 0.20843630166169577, "grad_norm": 17.5, "learning_rate": 2.438352442846997e-06, "loss": 1.0375, "step": 2446 }, { "epoch": 0.20852151682999573, "grad_norm": 16.5, "learning_rate": 2.438300312958394e-06, "loss": 0.8947, "step": 2447 }, { "epoch": 0.20860673199829569, "grad_norm": 11.0, "learning_rate": 2.4382481615960096e-06, "loss": 0.498, "step": 2448 }, { "epoch": 0.20869194716659564, "grad_norm": 18.125, "learning_rate": 2.438195988760786e-06, "loss": 0.7602, "step": 2449 }, { "epoch": 0.2087771623348956, "grad_norm": 12.5, "learning_rate": 2.438143794453666e-06, "loss": 0.5734, "step": 2450 }, { "epoch": 0.20886237750319556, "grad_norm": 17.75, "learning_rate": 2.438091578675593e-06, "loss": 1.0439, "step": 2451 }, { "epoch": 0.2089475926714955, "grad_norm": 17.25, "learning_rate": 2.4380393414275107e-06, "loss": 1.0062, "step": 2452 }, { "epoch": 0.20903280783979547, "grad_norm": 15.9375, "learning_rate": 2.4379870827103626e-06, "loss": 0.8711, "step": 2453 }, { "epoch": 0.20911802300809543, "grad_norm": 11.375, "learning_rate": 2.4379348025250933e-06, "loss": 0.5635, "step": 2454 }, { "epoch": 0.2092032381763954, "grad_norm": 14.5625, "learning_rate": 2.4378825008726473e-06, "loss": 0.8082, "step": 2455 }, { "epoch": 0.20928845334469537, "grad_norm": 14.3125, "learning_rate": 2.43783017775397e-06, "loss": 0.7393, "step": 2456 }, { "epoch": 0.20937366851299533, "grad_norm": 16.25, "learning_rate": 2.4377778331700077e-06, "loss": 0.7298, "step": 2457 }, { "epoch": 0.20945888368129528, "grad_norm": 24.375, "learning_rate": 2.437725467121705e-06, "loss": 1.121, "step": 2458 }, { "epoch": 0.20954409884959524, "grad_norm": 14.5, "learning_rate": 2.4376730796100084e-06, "loss": 0.6564, "step": 2459 }, { "epoch": 0.2096293140178952, "grad_norm": 11.4375, "learning_rate": 2.437620670635865e-06, "loss": 0.6601, "step": 2460 }, { "epoch": 0.20971452918619515, "grad_norm": 14.6875, "learning_rate": 2.4375682402002227e-06, "loss": 0.8472, "step": 2461 }, { "epoch": 0.2097997443544951, "grad_norm": 11.8125, "learning_rate": 2.4375157883040273e-06, "loss": 0.5964, "step": 2462 }, { "epoch": 0.20988495952279507, "grad_norm": 20.625, "learning_rate": 2.437463314948228e-06, "loss": 1.0367, "step": 2463 }, { "epoch": 0.20997017469109502, "grad_norm": 20.25, "learning_rate": 2.437410820133772e-06, "loss": 1.0741, "step": 2464 }, { "epoch": 0.21005538985939498, "grad_norm": 21.75, "learning_rate": 2.437358303861609e-06, "loss": 1.0369, "step": 2465 }, { "epoch": 0.21014060502769494, "grad_norm": 17.875, "learning_rate": 2.437305766132687e-06, "loss": 0.7807, "step": 2466 }, { "epoch": 0.2102258201959949, "grad_norm": 12.875, "learning_rate": 2.437253206947956e-06, "loss": 0.7198, "step": 2467 }, { "epoch": 0.21031103536429485, "grad_norm": 17.75, "learning_rate": 2.437200626308366e-06, "loss": 0.6326, "step": 2468 }, { "epoch": 0.2103962505325948, "grad_norm": 11.5, "learning_rate": 2.437148024214866e-06, "loss": 0.5773, "step": 2469 }, { "epoch": 0.21048146570089477, "grad_norm": 14.8125, "learning_rate": 2.4370954006684083e-06, "loss": 0.8714, "step": 2470 }, { "epoch": 0.21056668086919472, "grad_norm": 16.375, "learning_rate": 2.437042755669943e-06, "loss": 0.79, "step": 2471 }, { "epoch": 0.21065189603749468, "grad_norm": 14.0, "learning_rate": 2.4369900892204205e-06, "loss": 0.7128, "step": 2472 }, { "epoch": 0.21073711120579464, "grad_norm": 25.75, "learning_rate": 2.4369374013207946e-06, "loss": 0.9695, "step": 2473 }, { "epoch": 0.2108223263740946, "grad_norm": 13.0, "learning_rate": 2.4368846919720156e-06, "loss": 0.6582, "step": 2474 }, { "epoch": 0.21090754154239455, "grad_norm": 11.625, "learning_rate": 2.4368319611750367e-06, "loss": 0.5139, "step": 2475 }, { "epoch": 0.2109927567106945, "grad_norm": 13.5625, "learning_rate": 2.4367792089308114e-06, "loss": 0.8152, "step": 2476 }, { "epoch": 0.21107797187899446, "grad_norm": 16.0, "learning_rate": 2.436726435240292e-06, "loss": 0.6635, "step": 2477 }, { "epoch": 0.21116318704729442, "grad_norm": 15.3125, "learning_rate": 2.436673640104433e-06, "loss": 0.5012, "step": 2478 }, { "epoch": 0.21124840221559438, "grad_norm": 18.375, "learning_rate": 2.436620823524187e-06, "loss": 0.9702, "step": 2479 }, { "epoch": 0.21133361738389433, "grad_norm": 13.0625, "learning_rate": 2.4365679855005104e-06, "loss": 0.5865, "step": 2480 }, { "epoch": 0.2114188325521943, "grad_norm": 14.125, "learning_rate": 2.436515126034357e-06, "loss": 0.8313, "step": 2481 }, { "epoch": 0.21150404772049425, "grad_norm": 14.75, "learning_rate": 2.436462245126682e-06, "loss": 0.4818, "step": 2482 }, { "epoch": 0.2115892628887942, "grad_norm": 11.1875, "learning_rate": 2.436409342778441e-06, "loss": 0.8752, "step": 2483 }, { "epoch": 0.21167447805709416, "grad_norm": 11.25, "learning_rate": 2.4363564189905904e-06, "loss": 0.3124, "step": 2484 }, { "epoch": 0.21175969322539412, "grad_norm": 17.75, "learning_rate": 2.436303473764086e-06, "loss": 0.8358, "step": 2485 }, { "epoch": 0.21184490839369408, "grad_norm": 9.75, "learning_rate": 2.4362505070998847e-06, "loss": 0.2835, "step": 2486 }, { "epoch": 0.21193012356199403, "grad_norm": 15.4375, "learning_rate": 2.4361975189989447e-06, "loss": 0.9199, "step": 2487 }, { "epoch": 0.212015338730294, "grad_norm": 18.125, "learning_rate": 2.4361445094622223e-06, "loss": 0.7091, "step": 2488 }, { "epoch": 0.21210055389859395, "grad_norm": 13.9375, "learning_rate": 2.4360914784906757e-06, "loss": 0.8237, "step": 2489 }, { "epoch": 0.2121857690668939, "grad_norm": 15.1875, "learning_rate": 2.436038426085264e-06, "loss": 0.9451, "step": 2490 }, { "epoch": 0.21227098423519386, "grad_norm": 12.375, "learning_rate": 2.4359853522469447e-06, "loss": 0.7429, "step": 2491 }, { "epoch": 0.21235619940349382, "grad_norm": 16.125, "learning_rate": 2.4359322569766776e-06, "loss": 0.9901, "step": 2492 }, { "epoch": 0.21244141457179377, "grad_norm": 16.5, "learning_rate": 2.4358791402754216e-06, "loss": 0.5445, "step": 2493 }, { "epoch": 0.21252662974009373, "grad_norm": 11.75, "learning_rate": 2.4358260021441376e-06, "loss": 0.662, "step": 2494 }, { "epoch": 0.2126118449083937, "grad_norm": 13.75, "learning_rate": 2.4357728425837853e-06, "loss": 0.4285, "step": 2495 }, { "epoch": 0.21269706007669364, "grad_norm": 13.875, "learning_rate": 2.4357196615953256e-06, "loss": 0.8422, "step": 2496 }, { "epoch": 0.2127822752449936, "grad_norm": 16.25, "learning_rate": 2.4356664591797186e-06, "loss": 0.5245, "step": 2497 }, { "epoch": 0.21286749041329356, "grad_norm": 13.5625, "learning_rate": 2.435613235337927e-06, "loss": 1.0852, "step": 2498 }, { "epoch": 0.21295270558159352, "grad_norm": 16.875, "learning_rate": 2.4355599900709114e-06, "loss": 0.3149, "step": 2499 }, { "epoch": 0.21303792074989347, "grad_norm": 18.75, "learning_rate": 2.4355067233796352e-06, "loss": 1.0901, "step": 2500 }, { "epoch": 0.21312313591819343, "grad_norm": 14.5, "learning_rate": 2.43545343526506e-06, "loss": 0.8986, "step": 2501 }, { "epoch": 0.2132083510864934, "grad_norm": 18.125, "learning_rate": 2.4354001257281497e-06, "loss": 0.8688, "step": 2502 }, { "epoch": 0.21329356625479334, "grad_norm": 19.375, "learning_rate": 2.4353467947698665e-06, "loss": 1.086, "step": 2503 }, { "epoch": 0.2133787814230933, "grad_norm": 24.375, "learning_rate": 2.435293442391175e-06, "loss": 1.2636, "step": 2504 }, { "epoch": 0.21346399659139326, "grad_norm": 13.9375, "learning_rate": 2.4352400685930388e-06, "loss": 0.5842, "step": 2505 }, { "epoch": 0.2135492117596932, "grad_norm": 11.125, "learning_rate": 2.435186673376423e-06, "loss": 0.5385, "step": 2506 }, { "epoch": 0.21363442692799317, "grad_norm": 17.5, "learning_rate": 2.4351332567422924e-06, "loss": 0.6024, "step": 2507 }, { "epoch": 0.21371964209629313, "grad_norm": 12.5, "learning_rate": 2.4350798186916118e-06, "loss": 0.6682, "step": 2508 }, { "epoch": 0.21380485726459308, "grad_norm": 13.8125, "learning_rate": 2.4350263592253476e-06, "loss": 1.0078, "step": 2509 }, { "epoch": 0.21389007243289304, "grad_norm": 15.1875, "learning_rate": 2.4349728783444646e-06, "loss": 0.6571, "step": 2510 }, { "epoch": 0.21397528760119303, "grad_norm": 11.9375, "learning_rate": 2.434919376049931e-06, "loss": 0.5937, "step": 2511 }, { "epoch": 0.21406050276949298, "grad_norm": 13.5625, "learning_rate": 2.4348658523427125e-06, "loss": 0.7072, "step": 2512 }, { "epoch": 0.21414571793779294, "grad_norm": 18.75, "learning_rate": 2.4348123072237766e-06, "loss": 0.8007, "step": 2513 }, { "epoch": 0.2142309331060929, "grad_norm": 15.8125, "learning_rate": 2.4347587406940908e-06, "loss": 0.8709, "step": 2514 }, { "epoch": 0.21431614827439285, "grad_norm": 13.125, "learning_rate": 2.4347051527546227e-06, "loss": 0.9103, "step": 2515 }, { "epoch": 0.2144013634426928, "grad_norm": 11.625, "learning_rate": 2.434651543406342e-06, "loss": 0.6602, "step": 2516 }, { "epoch": 0.21448657861099277, "grad_norm": 15.625, "learning_rate": 2.4345979126502164e-06, "loss": 0.7888, "step": 2517 }, { "epoch": 0.21457179377929272, "grad_norm": 13.125, "learning_rate": 2.434544260487215e-06, "loss": 0.6301, "step": 2518 }, { "epoch": 0.21465700894759268, "grad_norm": 17.25, "learning_rate": 2.434490586918308e-06, "loss": 1.1467, "step": 2519 }, { "epoch": 0.21474222411589264, "grad_norm": 20.0, "learning_rate": 2.4344368919444646e-06, "loss": 0.903, "step": 2520 }, { "epoch": 0.2148274392841926, "grad_norm": 13.625, "learning_rate": 2.4343831755666557e-06, "loss": 0.7378, "step": 2521 }, { "epoch": 0.21491265445249255, "grad_norm": 13.6875, "learning_rate": 2.434329437785852e-06, "loss": 0.7339, "step": 2522 }, { "epoch": 0.2149978696207925, "grad_norm": 20.875, "learning_rate": 2.4342756786030246e-06, "loss": 1.2273, "step": 2523 }, { "epoch": 0.21508308478909247, "grad_norm": 10.0625, "learning_rate": 2.434221898019144e-06, "loss": 0.4503, "step": 2524 }, { "epoch": 0.21516829995739242, "grad_norm": 13.9375, "learning_rate": 2.434168096035184e-06, "loss": 0.5745, "step": 2525 }, { "epoch": 0.21525351512569238, "grad_norm": 15.0625, "learning_rate": 2.434114272652115e-06, "loss": 1.0181, "step": 2526 }, { "epoch": 0.21533873029399234, "grad_norm": 14.1875, "learning_rate": 2.43406042787091e-06, "loss": 0.8084, "step": 2527 }, { "epoch": 0.2154239454622923, "grad_norm": 14.625, "learning_rate": 2.4340065616925433e-06, "loss": 0.8192, "step": 2528 }, { "epoch": 0.21550916063059225, "grad_norm": 12.875, "learning_rate": 2.433952674117987e-06, "loss": 0.7902, "step": 2529 }, { "epoch": 0.2155943757988922, "grad_norm": 15.3125, "learning_rate": 2.4338987651482153e-06, "loss": 0.904, "step": 2530 }, { "epoch": 0.21567959096719216, "grad_norm": 18.125, "learning_rate": 2.4338448347842027e-06, "loss": 0.8997, "step": 2531 }, { "epoch": 0.21576480613549212, "grad_norm": 15.5625, "learning_rate": 2.433790883026923e-06, "loss": 0.7253, "step": 2532 }, { "epoch": 0.21585002130379208, "grad_norm": 15.1875, "learning_rate": 2.433736909877352e-06, "loss": 0.9365, "step": 2533 }, { "epoch": 0.21593523647209203, "grad_norm": 16.75, "learning_rate": 2.433682915336465e-06, "loss": 0.9074, "step": 2534 }, { "epoch": 0.216020451640392, "grad_norm": 10.875, "learning_rate": 2.4336288994052364e-06, "loss": 0.359, "step": 2535 }, { "epoch": 0.21610566680869195, "grad_norm": 14.1875, "learning_rate": 2.4335748620846444e-06, "loss": 0.6362, "step": 2536 }, { "epoch": 0.2161908819769919, "grad_norm": 11.5625, "learning_rate": 2.433520803375664e-06, "loss": 0.418, "step": 2537 }, { "epoch": 0.21627609714529186, "grad_norm": 14.0, "learning_rate": 2.4334667232792724e-06, "loss": 0.7132, "step": 2538 }, { "epoch": 0.21636131231359182, "grad_norm": 13.6875, "learning_rate": 2.4334126217964475e-06, "loss": 0.5964, "step": 2539 }, { "epoch": 0.21644652748189178, "grad_norm": 26.75, "learning_rate": 2.4333584989281663e-06, "loss": 0.6838, "step": 2540 }, { "epoch": 0.21653174265019173, "grad_norm": 18.875, "learning_rate": 2.433304354675407e-06, "loss": 0.9192, "step": 2541 }, { "epoch": 0.2166169578184917, "grad_norm": 19.375, "learning_rate": 2.433250189039148e-06, "loss": 0.9965, "step": 2542 }, { "epoch": 0.21670217298679165, "grad_norm": 13.625, "learning_rate": 2.4331960020203686e-06, "loss": 0.4586, "step": 2543 }, { "epoch": 0.2167873881550916, "grad_norm": 21.625, "learning_rate": 2.4331417936200475e-06, "loss": 0.6214, "step": 2544 }, { "epoch": 0.21687260332339156, "grad_norm": 13.375, "learning_rate": 2.4330875638391644e-06, "loss": 0.5138, "step": 2545 }, { "epoch": 0.21695781849169152, "grad_norm": 11.0625, "learning_rate": 2.433033312678699e-06, "loss": 0.6062, "step": 2546 }, { "epoch": 0.21704303365999147, "grad_norm": 28.125, "learning_rate": 2.4329790401396326e-06, "loss": 0.8695, "step": 2547 }, { "epoch": 0.21712824882829143, "grad_norm": 38.75, "learning_rate": 2.432924746222945e-06, "loss": 1.2791, "step": 2548 }, { "epoch": 0.2172134639965914, "grad_norm": 16.25, "learning_rate": 2.432870430929618e-06, "loss": 0.8814, "step": 2549 }, { "epoch": 0.21729867916489135, "grad_norm": 15.375, "learning_rate": 2.4328160942606324e-06, "loss": 0.7691, "step": 2550 }, { "epoch": 0.2173838943331913, "grad_norm": 28.0, "learning_rate": 2.432761736216971e-06, "loss": 0.7908, "step": 2551 }, { "epoch": 0.21746910950149126, "grad_norm": 21.375, "learning_rate": 2.4327073567996152e-06, "loss": 0.9392, "step": 2552 }, { "epoch": 0.21755432466979122, "grad_norm": 11.9375, "learning_rate": 2.4326529560095486e-06, "loss": 0.6047, "step": 2553 }, { "epoch": 0.21763953983809117, "grad_norm": 11.8125, "learning_rate": 2.4325985338477537e-06, "loss": 0.4832, "step": 2554 }, { "epoch": 0.21772475500639113, "grad_norm": 11.9375, "learning_rate": 2.432544090315214e-06, "loss": 0.4691, "step": 2555 }, { "epoch": 0.2178099701746911, "grad_norm": 20.125, "learning_rate": 2.4324896254129137e-06, "loss": 0.727, "step": 2556 }, { "epoch": 0.21789518534299104, "grad_norm": 16.125, "learning_rate": 2.4324351391418364e-06, "loss": 0.7773, "step": 2557 }, { "epoch": 0.217980400511291, "grad_norm": 18.5, "learning_rate": 2.4323806315029673e-06, "loss": 1.0957, "step": 2558 }, { "epoch": 0.21806561567959096, "grad_norm": 19.125, "learning_rate": 2.432326102497291e-06, "loss": 1.3076, "step": 2559 }, { "epoch": 0.21815083084789091, "grad_norm": 12.4375, "learning_rate": 2.4322715521257934e-06, "loss": 0.6208, "step": 2560 }, { "epoch": 0.21823604601619087, "grad_norm": 16.0, "learning_rate": 2.43221698038946e-06, "loss": 0.8164, "step": 2561 }, { "epoch": 0.21832126118449083, "grad_norm": 11.125, "learning_rate": 2.432162387289277e-06, "loss": 0.4064, "step": 2562 }, { "epoch": 0.21840647635279078, "grad_norm": 12.0625, "learning_rate": 2.432107772826231e-06, "loss": 0.3819, "step": 2563 }, { "epoch": 0.21849169152109074, "grad_norm": 16.875, "learning_rate": 2.4320531370013084e-06, "loss": 0.904, "step": 2564 }, { "epoch": 0.2185769066893907, "grad_norm": 14.375, "learning_rate": 2.4319984798154973e-06, "loss": 0.8427, "step": 2565 }, { "epoch": 0.21866212185769066, "grad_norm": 13.9375, "learning_rate": 2.431943801269785e-06, "loss": 0.6886, "step": 2566 }, { "epoch": 0.2187473370259906, "grad_norm": 17.0, "learning_rate": 2.4318891013651594e-06, "loss": 0.8962, "step": 2567 }, { "epoch": 0.2188325521942906, "grad_norm": 9.3125, "learning_rate": 2.4318343801026095e-06, "loss": 0.3265, "step": 2568 }, { "epoch": 0.21891776736259055, "grad_norm": 18.625, "learning_rate": 2.4317796374831238e-06, "loss": 0.957, "step": 2569 }, { "epoch": 0.2190029825308905, "grad_norm": 13.9375, "learning_rate": 2.431724873507692e-06, "loss": 0.8619, "step": 2570 }, { "epoch": 0.21908819769919047, "grad_norm": 11.5625, "learning_rate": 2.431670088177303e-06, "loss": 0.507, "step": 2571 }, { "epoch": 0.21917341286749042, "grad_norm": 17.25, "learning_rate": 2.4316152814929478e-06, "loss": 0.7112, "step": 2572 }, { "epoch": 0.21925862803579038, "grad_norm": 13.125, "learning_rate": 2.431560453455616e-06, "loss": 0.4879, "step": 2573 }, { "epoch": 0.21934384320409034, "grad_norm": 13.6875, "learning_rate": 2.431505604066298e-06, "loss": 0.8353, "step": 2574 }, { "epoch": 0.2194290583723903, "grad_norm": 16.625, "learning_rate": 2.4314507333259864e-06, "loss": 0.9581, "step": 2575 }, { "epoch": 0.21951427354069025, "grad_norm": 17.5, "learning_rate": 2.431395841235672e-06, "loss": 0.9309, "step": 2576 }, { "epoch": 0.2195994887089902, "grad_norm": 19.75, "learning_rate": 2.4313409277963463e-06, "loss": 0.7396, "step": 2577 }, { "epoch": 0.21968470387729017, "grad_norm": 15.0, "learning_rate": 2.4312859930090025e-06, "loss": 0.8397, "step": 2578 }, { "epoch": 0.21976991904559012, "grad_norm": 16.25, "learning_rate": 2.4312310368746327e-06, "loss": 0.9028, "step": 2579 }, { "epoch": 0.21985513421389008, "grad_norm": 13.625, "learning_rate": 2.431176059394231e-06, "loss": 0.8091, "step": 2580 }, { "epoch": 0.21994034938219004, "grad_norm": 18.0, "learning_rate": 2.431121060568789e-06, "loss": 0.96, "step": 2581 }, { "epoch": 0.22002556455049, "grad_norm": 15.8125, "learning_rate": 2.4310660403993023e-06, "loss": 0.7458, "step": 2582 }, { "epoch": 0.22011077971878995, "grad_norm": 11.5625, "learning_rate": 2.4310109988867647e-06, "loss": 0.5909, "step": 2583 }, { "epoch": 0.2201959948870899, "grad_norm": 13.375, "learning_rate": 2.4309559360321707e-06, "loss": 0.7709, "step": 2584 }, { "epoch": 0.22028121005538986, "grad_norm": 17.0, "learning_rate": 2.430900851836515e-06, "loss": 0.7266, "step": 2585 }, { "epoch": 0.22036642522368982, "grad_norm": 14.125, "learning_rate": 2.430845746300794e-06, "loss": 0.3861, "step": 2586 }, { "epoch": 0.22045164039198978, "grad_norm": 17.375, "learning_rate": 2.430790619426002e-06, "loss": 1.3531, "step": 2587 }, { "epoch": 0.22053685556028974, "grad_norm": 14.0625, "learning_rate": 2.430735471213137e-06, "loss": 0.7736, "step": 2588 }, { "epoch": 0.2206220707285897, "grad_norm": 11.625, "learning_rate": 2.4306803016631945e-06, "loss": 0.5715, "step": 2589 }, { "epoch": 0.22070728589688965, "grad_norm": 18.75, "learning_rate": 2.4306251107771717e-06, "loss": 0.9396, "step": 2590 }, { "epoch": 0.2207925010651896, "grad_norm": 13.8125, "learning_rate": 2.430569898556066e-06, "loss": 0.821, "step": 2591 }, { "epoch": 0.22087771623348956, "grad_norm": 16.25, "learning_rate": 2.430514665000875e-06, "loss": 0.9442, "step": 2592 }, { "epoch": 0.22096293140178952, "grad_norm": 14.5, "learning_rate": 2.4304594101125967e-06, "loss": 1.0966, "step": 2593 }, { "epoch": 0.22104814657008948, "grad_norm": 13.125, "learning_rate": 2.4304041338922303e-06, "loss": 0.6438, "step": 2594 }, { "epoch": 0.22113336173838943, "grad_norm": 21.25, "learning_rate": 2.4303488363407736e-06, "loss": 1.0892, "step": 2595 }, { "epoch": 0.2212185769066894, "grad_norm": 22.5, "learning_rate": 2.430293517459227e-06, "loss": 0.9584, "step": 2596 }, { "epoch": 0.22130379207498935, "grad_norm": 14.75, "learning_rate": 2.4302381772485893e-06, "loss": 1.055, "step": 2597 }, { "epoch": 0.2213890072432893, "grad_norm": 15.25, "learning_rate": 2.430182815709861e-06, "loss": 1.1028, "step": 2598 }, { "epoch": 0.22147422241158926, "grad_norm": 13.4375, "learning_rate": 2.4301274328440427e-06, "loss": 0.7913, "step": 2599 }, { "epoch": 0.22155943757988922, "grad_norm": 14.25, "learning_rate": 2.430072028652135e-06, "loss": 0.5136, "step": 2600 }, { "epoch": 0.22164465274818917, "grad_norm": 15.4375, "learning_rate": 2.4300166031351384e-06, "loss": 0.7976, "step": 2601 }, { "epoch": 0.22172986791648913, "grad_norm": 15.0, "learning_rate": 2.429961156294056e-06, "loss": 0.7175, "step": 2602 }, { "epoch": 0.2218150830847891, "grad_norm": 23.125, "learning_rate": 2.429905688129889e-06, "loss": 1.0837, "step": 2603 }, { "epoch": 0.22190029825308905, "grad_norm": 23.25, "learning_rate": 2.429850198643639e-06, "loss": 0.8739, "step": 2604 }, { "epoch": 0.221985513421389, "grad_norm": 15.9375, "learning_rate": 2.42979468783631e-06, "loss": 0.6373, "step": 2605 }, { "epoch": 0.22207072858968896, "grad_norm": 15.4375, "learning_rate": 2.4297391557089044e-06, "loss": 0.7459, "step": 2606 }, { "epoch": 0.22215594375798892, "grad_norm": 16.125, "learning_rate": 2.429683602262426e-06, "loss": 1.026, "step": 2607 }, { "epoch": 0.22224115892628887, "grad_norm": 14.5625, "learning_rate": 2.429628027497879e-06, "loss": 0.6947, "step": 2608 }, { "epoch": 0.22232637409458883, "grad_norm": 11.5625, "learning_rate": 2.4295724314162668e-06, "loss": 0.637, "step": 2609 }, { "epoch": 0.2224115892628888, "grad_norm": 16.125, "learning_rate": 2.4295168140185945e-06, "loss": 1.2364, "step": 2610 }, { "epoch": 0.22249680443118874, "grad_norm": 14.3125, "learning_rate": 2.429461175305868e-06, "loss": 0.5258, "step": 2611 }, { "epoch": 0.2225820195994887, "grad_norm": 17.75, "learning_rate": 2.429405515279091e-06, "loss": 0.9123, "step": 2612 }, { "epoch": 0.22266723476778866, "grad_norm": 18.625, "learning_rate": 2.4293498339392715e-06, "loss": 0.5855, "step": 2613 }, { "epoch": 0.22275244993608861, "grad_norm": 19.625, "learning_rate": 2.4292941312874137e-06, "loss": 0.913, "step": 2614 }, { "epoch": 0.22283766510438857, "grad_norm": 12.875, "learning_rate": 2.4292384073245255e-06, "loss": 0.5384, "step": 2615 }, { "epoch": 0.22292288027268853, "grad_norm": 16.0, "learning_rate": 2.4291826620516135e-06, "loss": 0.9442, "step": 2616 }, { "epoch": 0.22300809544098849, "grad_norm": 19.75, "learning_rate": 2.429126895469685e-06, "loss": 0.6352, "step": 2617 }, { "epoch": 0.22309331060928844, "grad_norm": 11.375, "learning_rate": 2.4290711075797475e-06, "loss": 0.6309, "step": 2618 }, { "epoch": 0.2231785257775884, "grad_norm": 14.75, "learning_rate": 2.4290152983828095e-06, "loss": 0.4835, "step": 2619 }, { "epoch": 0.22326374094588836, "grad_norm": 15.375, "learning_rate": 2.4289594678798796e-06, "loss": 0.6368, "step": 2620 }, { "epoch": 0.2233489561141883, "grad_norm": 15.5625, "learning_rate": 2.4289036160719666e-06, "loss": 0.9495, "step": 2621 }, { "epoch": 0.22343417128248827, "grad_norm": 18.625, "learning_rate": 2.4288477429600798e-06, "loss": 0.4465, "step": 2622 }, { "epoch": 0.22351938645078823, "grad_norm": 11.0625, "learning_rate": 2.4287918485452288e-06, "loss": 0.4655, "step": 2623 }, { "epoch": 0.2236046016190882, "grad_norm": 16.125, "learning_rate": 2.4287359328284237e-06, "loss": 0.9579, "step": 2624 }, { "epoch": 0.22368981678738817, "grad_norm": 19.0, "learning_rate": 2.428679995810675e-06, "loss": 0.717, "step": 2625 }, { "epoch": 0.22377503195568813, "grad_norm": 23.125, "learning_rate": 2.4286240374929937e-06, "loss": 1.1671, "step": 2626 }, { "epoch": 0.22386024712398808, "grad_norm": 16.0, "learning_rate": 2.4285680578763907e-06, "loss": 0.914, "step": 2627 }, { "epoch": 0.22394546229228804, "grad_norm": 25.25, "learning_rate": 2.4285120569618777e-06, "loss": 1.0151, "step": 2628 }, { "epoch": 0.224030677460588, "grad_norm": 64.5, "learning_rate": 2.4284560347504666e-06, "loss": 0.6493, "step": 2629 }, { "epoch": 0.22411589262888795, "grad_norm": 17.375, "learning_rate": 2.4283999912431704e-06, "loss": 0.8151, "step": 2630 }, { "epoch": 0.2242011077971879, "grad_norm": 15.9375, "learning_rate": 2.428343926441001e-06, "loss": 0.7265, "step": 2631 }, { "epoch": 0.22428632296548787, "grad_norm": 19.0, "learning_rate": 2.4282878403449724e-06, "loss": 0.7539, "step": 2632 }, { "epoch": 0.22437153813378782, "grad_norm": 15.75, "learning_rate": 2.428231732956097e-06, "loss": 0.8051, "step": 2633 }, { "epoch": 0.22445675330208778, "grad_norm": 15.3125, "learning_rate": 2.4281756042753894e-06, "loss": 1.016, "step": 2634 }, { "epoch": 0.22454196847038774, "grad_norm": 12.6875, "learning_rate": 2.4281194543038643e-06, "loss": 0.4299, "step": 2635 }, { "epoch": 0.2246271836386877, "grad_norm": 14.25, "learning_rate": 2.428063283042536e-06, "loss": 0.762, "step": 2636 }, { "epoch": 0.22471239880698765, "grad_norm": 17.125, "learning_rate": 2.428007090492419e-06, "loss": 0.9848, "step": 2637 }, { "epoch": 0.2247976139752876, "grad_norm": 12.6875, "learning_rate": 2.4279508766545303e-06, "loss": 0.5421, "step": 2638 }, { "epoch": 0.22488282914358756, "grad_norm": 13.8125, "learning_rate": 2.427894641529884e-06, "loss": 0.6195, "step": 2639 }, { "epoch": 0.22496804431188752, "grad_norm": 12.75, "learning_rate": 2.427838385119497e-06, "loss": 0.6019, "step": 2640 }, { "epoch": 0.22505325948018748, "grad_norm": 13.0, "learning_rate": 2.4277821074243862e-06, "loss": 0.4765, "step": 2641 }, { "epoch": 0.22513847464848744, "grad_norm": 16.625, "learning_rate": 2.4277258084455682e-06, "loss": 0.8688, "step": 2642 }, { "epoch": 0.2252236898167874, "grad_norm": 13.8125, "learning_rate": 2.4276694881840606e-06, "loss": 0.6365, "step": 2643 }, { "epoch": 0.22530890498508735, "grad_norm": 20.0, "learning_rate": 2.4276131466408815e-06, "loss": 1.0313, "step": 2644 }, { "epoch": 0.2253941201533873, "grad_norm": 14.4375, "learning_rate": 2.4275567838170483e-06, "loss": 0.6653, "step": 2645 }, { "epoch": 0.22547933532168726, "grad_norm": 11.9375, "learning_rate": 2.42750039971358e-06, "loss": 0.5904, "step": 2646 }, { "epoch": 0.22556455048998722, "grad_norm": 14.0, "learning_rate": 2.427443994331495e-06, "loss": 0.6693, "step": 2647 }, { "epoch": 0.22564976565828718, "grad_norm": 13.875, "learning_rate": 2.4273875676718133e-06, "loss": 0.5469, "step": 2648 }, { "epoch": 0.22573498082658713, "grad_norm": 16.375, "learning_rate": 2.427331119735554e-06, "loss": 1.003, "step": 2649 }, { "epoch": 0.2258201959948871, "grad_norm": 12.6875, "learning_rate": 2.427274650523738e-06, "loss": 0.4572, "step": 2650 }, { "epoch": 0.22590541116318705, "grad_norm": 19.625, "learning_rate": 2.427218160037385e-06, "loss": 0.9576, "step": 2651 }, { "epoch": 0.225990626331487, "grad_norm": 13.6875, "learning_rate": 2.4271616482775164e-06, "loss": 0.6773, "step": 2652 }, { "epoch": 0.22607584149978696, "grad_norm": 13.5, "learning_rate": 2.4271051152451526e-06, "loss": 0.5406, "step": 2653 }, { "epoch": 0.22616105666808692, "grad_norm": 14.25, "learning_rate": 2.427048560941316e-06, "loss": 0.622, "step": 2654 }, { "epoch": 0.22624627183638688, "grad_norm": 17.875, "learning_rate": 2.426991985367028e-06, "loss": 1.1171, "step": 2655 }, { "epoch": 0.22633148700468683, "grad_norm": 16.375, "learning_rate": 2.4269353885233114e-06, "loss": 1.0306, "step": 2656 }, { "epoch": 0.2264167021729868, "grad_norm": 10.375, "learning_rate": 2.426878770411189e-06, "loss": 0.4994, "step": 2657 }, { "epoch": 0.22650191734128675, "grad_norm": 10.9375, "learning_rate": 2.426822131031684e-06, "loss": 0.4089, "step": 2658 }, { "epoch": 0.2265871325095867, "grad_norm": 16.625, "learning_rate": 2.426765470385819e-06, "loss": 0.9012, "step": 2659 }, { "epoch": 0.22667234767788666, "grad_norm": 21.375, "learning_rate": 2.426708788474619e-06, "loss": 1.65, "step": 2660 }, { "epoch": 0.22675756284618662, "grad_norm": 19.5, "learning_rate": 2.426652085299108e-06, "loss": 1.0165, "step": 2661 }, { "epoch": 0.22684277801448657, "grad_norm": 13.1875, "learning_rate": 2.4265953608603105e-06, "loss": 0.6429, "step": 2662 }, { "epoch": 0.22692799318278653, "grad_norm": 12.4375, "learning_rate": 2.426538615159252e-06, "loss": 0.5751, "step": 2663 }, { "epoch": 0.2270132083510865, "grad_norm": 16.0, "learning_rate": 2.426481848196957e-06, "loss": 1.1199, "step": 2664 }, { "epoch": 0.22709842351938644, "grad_norm": 15.0625, "learning_rate": 2.4264250599744522e-06, "loss": 0.5531, "step": 2665 }, { "epoch": 0.2271836386876864, "grad_norm": 13.8125, "learning_rate": 2.4263682504927635e-06, "loss": 0.8152, "step": 2666 }, { "epoch": 0.22726885385598636, "grad_norm": 43.75, "learning_rate": 2.426311419752918e-06, "loss": 1.0729, "step": 2667 }, { "epoch": 0.22735406902428631, "grad_norm": 14.9375, "learning_rate": 2.4262545677559418e-06, "loss": 0.9786, "step": 2668 }, { "epoch": 0.22743928419258627, "grad_norm": 14.875, "learning_rate": 2.426197694502863e-06, "loss": 0.7922, "step": 2669 }, { "epoch": 0.22752449936088623, "grad_norm": 10.6875, "learning_rate": 2.426140799994709e-06, "loss": 0.4565, "step": 2670 }, { "epoch": 0.22760971452918619, "grad_norm": 15.1875, "learning_rate": 2.426083884232508e-06, "loss": 0.7747, "step": 2671 }, { "epoch": 0.22769492969748614, "grad_norm": 17.375, "learning_rate": 2.4260269472172887e-06, "loss": 0.5734, "step": 2672 }, { "epoch": 0.2277801448657861, "grad_norm": 14.0, "learning_rate": 2.4259699889500797e-06, "loss": 0.9756, "step": 2673 }, { "epoch": 0.22786536003408606, "grad_norm": 13.8125, "learning_rate": 2.4259130094319106e-06, "loss": 0.6701, "step": 2674 }, { "epoch": 0.227950575202386, "grad_norm": 13.625, "learning_rate": 2.4258560086638113e-06, "loss": 0.7146, "step": 2675 }, { "epoch": 0.22803579037068597, "grad_norm": 18.75, "learning_rate": 2.4257989866468105e-06, "loss": 1.1415, "step": 2676 }, { "epoch": 0.22812100553898593, "grad_norm": 16.625, "learning_rate": 2.425741943381941e-06, "loss": 0.8519, "step": 2677 }, { "epoch": 0.22820622070728588, "grad_norm": 11.375, "learning_rate": 2.4256848788702314e-06, "loss": 0.7111, "step": 2678 }, { "epoch": 0.22829143587558584, "grad_norm": 14.6875, "learning_rate": 2.425627793112714e-06, "loss": 0.7178, "step": 2679 }, { "epoch": 0.22837665104388583, "grad_norm": 15.8125, "learning_rate": 2.42557068611042e-06, "loss": 0.8471, "step": 2680 }, { "epoch": 0.22846186621218578, "grad_norm": 13.625, "learning_rate": 2.4255135578643817e-06, "loss": 0.7265, "step": 2681 }, { "epoch": 0.22854708138048574, "grad_norm": 13.4375, "learning_rate": 2.4254564083756318e-06, "loss": 0.5264, "step": 2682 }, { "epoch": 0.2286322965487857, "grad_norm": 15.4375, "learning_rate": 2.4253992376452017e-06, "loss": 0.9509, "step": 2683 }, { "epoch": 0.22871751171708565, "grad_norm": 15.5625, "learning_rate": 2.4253420456741263e-06, "loss": 0.9853, "step": 2684 }, { "epoch": 0.2288027268853856, "grad_norm": 23.375, "learning_rate": 2.4252848324634375e-06, "loss": 1.2609, "step": 2685 }, { "epoch": 0.22888794205368557, "grad_norm": 13.0625, "learning_rate": 2.4252275980141705e-06, "loss": 0.7184, "step": 2686 }, { "epoch": 0.22897315722198552, "grad_norm": 15.5625, "learning_rate": 2.4251703423273586e-06, "loss": 0.8876, "step": 2687 }, { "epoch": 0.22905837239028548, "grad_norm": 12.75, "learning_rate": 2.4251130654040374e-06, "loss": 0.7087, "step": 2688 }, { "epoch": 0.22914358755858544, "grad_norm": 17.125, "learning_rate": 2.4250557672452406e-06, "loss": 0.8459, "step": 2689 }, { "epoch": 0.2292288027268854, "grad_norm": 15.0625, "learning_rate": 2.424998447852005e-06, "loss": 0.9062, "step": 2690 }, { "epoch": 0.22931401789518535, "grad_norm": 13.8125, "learning_rate": 2.424941107225366e-06, "loss": 0.6738, "step": 2691 }, { "epoch": 0.2293992330634853, "grad_norm": 16.125, "learning_rate": 2.42488374536636e-06, "loss": 0.8291, "step": 2692 }, { "epoch": 0.22948444823178527, "grad_norm": 14.9375, "learning_rate": 2.424826362276023e-06, "loss": 0.5306, "step": 2693 }, { "epoch": 0.22956966340008522, "grad_norm": 17.0, "learning_rate": 2.4247689579553918e-06, "loss": 0.8042, "step": 2694 }, { "epoch": 0.22965487856838518, "grad_norm": 14.0625, "learning_rate": 2.4247115324055045e-06, "loss": 1.0936, "step": 2695 }, { "epoch": 0.22974009373668514, "grad_norm": 11.5, "learning_rate": 2.424654085627398e-06, "loss": 0.5308, "step": 2696 }, { "epoch": 0.2298253089049851, "grad_norm": 11.875, "learning_rate": 2.4245966176221118e-06, "loss": 0.4608, "step": 2697 }, { "epoch": 0.22991052407328505, "grad_norm": 14.875, "learning_rate": 2.4245391283906837e-06, "loss": 0.6074, "step": 2698 }, { "epoch": 0.229995739241585, "grad_norm": 15.875, "learning_rate": 2.424481617934152e-06, "loss": 0.9357, "step": 2699 }, { "epoch": 0.23008095440988496, "grad_norm": 14.375, "learning_rate": 2.4244240862535563e-06, "loss": 0.9879, "step": 2700 }, { "epoch": 0.23016616957818492, "grad_norm": 19.125, "learning_rate": 2.4243665333499365e-06, "loss": 1.1863, "step": 2701 }, { "epoch": 0.23025138474648488, "grad_norm": 15.3125, "learning_rate": 2.424308959224333e-06, "loss": 0.8961, "step": 2702 }, { "epoch": 0.23033659991478483, "grad_norm": 12.375, "learning_rate": 2.4242513638777854e-06, "loss": 0.5673, "step": 2703 }, { "epoch": 0.2304218150830848, "grad_norm": 10.9375, "learning_rate": 2.424193747311335e-06, "loss": 0.4202, "step": 2704 }, { "epoch": 0.23050703025138475, "grad_norm": 17.375, "learning_rate": 2.4241361095260225e-06, "loss": 0.6607, "step": 2705 }, { "epoch": 0.2305922454196847, "grad_norm": 13.625, "learning_rate": 2.4240784505228897e-06, "loss": 0.6322, "step": 2706 }, { "epoch": 0.23067746058798466, "grad_norm": 12.5625, "learning_rate": 2.4240207703029788e-06, "loss": 0.6797, "step": 2707 }, { "epoch": 0.23076267575628462, "grad_norm": 12.6875, "learning_rate": 2.4239630688673324e-06, "loss": 0.4156, "step": 2708 }, { "epoch": 0.23084789092458458, "grad_norm": 17.375, "learning_rate": 2.4239053462169924e-06, "loss": 1.1465, "step": 2709 }, { "epoch": 0.23093310609288453, "grad_norm": 12.3125, "learning_rate": 2.4238476023530028e-06, "loss": 0.4007, "step": 2710 }, { "epoch": 0.2310183212611845, "grad_norm": 10.625, "learning_rate": 2.423789837276406e-06, "loss": 0.3558, "step": 2711 }, { "epoch": 0.23110353642948445, "grad_norm": 11.6875, "learning_rate": 2.423732050988247e-06, "loss": 0.5028, "step": 2712 }, { "epoch": 0.2311887515977844, "grad_norm": 13.25, "learning_rate": 2.42367424348957e-06, "loss": 0.7019, "step": 2713 }, { "epoch": 0.23127396676608436, "grad_norm": 12.5625, "learning_rate": 2.4236164147814184e-06, "loss": 0.47, "step": 2714 }, { "epoch": 0.23135918193438432, "grad_norm": 20.0, "learning_rate": 2.423558564864838e-06, "loss": 0.9292, "step": 2715 }, { "epoch": 0.23144439710268427, "grad_norm": 16.75, "learning_rate": 2.4235006937408743e-06, "loss": 1.0831, "step": 2716 }, { "epoch": 0.23152961227098423, "grad_norm": 15.3125, "learning_rate": 2.4234428014105735e-06, "loss": 0.7353, "step": 2717 }, { "epoch": 0.2316148274392842, "grad_norm": 21.25, "learning_rate": 2.423384887874981e-06, "loss": 1.2695, "step": 2718 }, { "epoch": 0.23170004260758414, "grad_norm": 13.125, "learning_rate": 2.4233269531351437e-06, "loss": 0.7943, "step": 2719 }, { "epoch": 0.2317852577758841, "grad_norm": 12.1875, "learning_rate": 2.423268997192108e-06, "loss": 0.6412, "step": 2720 }, { "epoch": 0.23187047294418406, "grad_norm": 14.5, "learning_rate": 2.4232110200469223e-06, "loss": 0.6674, "step": 2721 }, { "epoch": 0.23195568811248402, "grad_norm": 16.0, "learning_rate": 2.423153021700634e-06, "loss": 0.7757, "step": 2722 }, { "epoch": 0.23204090328078397, "grad_norm": 13.875, "learning_rate": 2.42309500215429e-06, "loss": 1.1354, "step": 2723 }, { "epoch": 0.23212611844908393, "grad_norm": 18.875, "learning_rate": 2.4230369614089404e-06, "loss": 0.9389, "step": 2724 }, { "epoch": 0.23221133361738389, "grad_norm": 15.375, "learning_rate": 2.422978899465633e-06, "loss": 0.7042, "step": 2725 }, { "epoch": 0.23229654878568384, "grad_norm": 11.8125, "learning_rate": 2.4229208163254176e-06, "loss": 0.5903, "step": 2726 }, { "epoch": 0.2323817639539838, "grad_norm": 16.25, "learning_rate": 2.4228627119893434e-06, "loss": 0.7102, "step": 2727 }, { "epoch": 0.23246697912228376, "grad_norm": 15.6875, "learning_rate": 2.42280458645846e-06, "loss": 0.9055, "step": 2728 }, { "epoch": 0.2325521942905837, "grad_norm": 15.4375, "learning_rate": 2.4227464397338195e-06, "loss": 0.8032, "step": 2729 }, { "epoch": 0.23263740945888367, "grad_norm": 22.125, "learning_rate": 2.4226882718164708e-06, "loss": 1.1022, "step": 2730 }, { "epoch": 0.23272262462718363, "grad_norm": 19.0, "learning_rate": 2.422630082707466e-06, "loss": 0.8255, "step": 2731 }, { "epoch": 0.23280783979548358, "grad_norm": 21.0, "learning_rate": 2.4225718724078565e-06, "loss": 0.7442, "step": 2732 }, { "epoch": 0.23289305496378354, "grad_norm": 18.375, "learning_rate": 2.422513640918694e-06, "loss": 0.688, "step": 2733 }, { "epoch": 0.2329782701320835, "grad_norm": 24.5, "learning_rate": 2.4224553882410313e-06, "loss": 0.693, "step": 2734 }, { "epoch": 0.23306348530038346, "grad_norm": 13.9375, "learning_rate": 2.42239711437592e-06, "loss": 1.0697, "step": 2735 }, { "epoch": 0.2331487004686834, "grad_norm": 18.125, "learning_rate": 2.4223388193244144e-06, "loss": 0.9238, "step": 2736 }, { "epoch": 0.2332339156369834, "grad_norm": 17.0, "learning_rate": 2.4222805030875677e-06, "loss": 0.892, "step": 2737 }, { "epoch": 0.23331913080528335, "grad_norm": 12.125, "learning_rate": 2.4222221656664334e-06, "loss": 0.6879, "step": 2738 }, { "epoch": 0.2334043459735833, "grad_norm": 13.9375, "learning_rate": 2.4221638070620655e-06, "loss": 0.6441, "step": 2739 }, { "epoch": 0.23348956114188327, "grad_norm": 16.375, "learning_rate": 2.422105427275519e-06, "loss": 0.6626, "step": 2740 }, { "epoch": 0.23357477631018322, "grad_norm": 10.1875, "learning_rate": 2.422047026307849e-06, "loss": 0.4695, "step": 2741 }, { "epoch": 0.23365999147848318, "grad_norm": 11.0, "learning_rate": 2.4219886041601104e-06, "loss": 0.3842, "step": 2742 }, { "epoch": 0.23374520664678314, "grad_norm": 11.4375, "learning_rate": 2.4219301608333595e-06, "loss": 0.4441, "step": 2743 }, { "epoch": 0.2338304218150831, "grad_norm": 14.0, "learning_rate": 2.4218716963286516e-06, "loss": 0.5696, "step": 2744 }, { "epoch": 0.23391563698338305, "grad_norm": 13.375, "learning_rate": 2.4218132106470443e-06, "loss": 0.6914, "step": 2745 }, { "epoch": 0.234000852151683, "grad_norm": 12.375, "learning_rate": 2.4217547037895935e-06, "loss": 0.5383, "step": 2746 }, { "epoch": 0.23408606731998297, "grad_norm": 11.9375, "learning_rate": 2.421696175757357e-06, "loss": 0.6258, "step": 2747 }, { "epoch": 0.23417128248828292, "grad_norm": 22.5, "learning_rate": 2.4216376265513928e-06, "loss": 1.0195, "step": 2748 }, { "epoch": 0.23425649765658288, "grad_norm": 17.0, "learning_rate": 2.421579056172758e-06, "loss": 0.9471, "step": 2749 }, { "epoch": 0.23434171282488284, "grad_norm": 35.75, "learning_rate": 2.4215204646225118e-06, "loss": 0.9019, "step": 2750 }, { "epoch": 0.2344269279931828, "grad_norm": 13.625, "learning_rate": 2.4214618519017125e-06, "loss": 0.5979, "step": 2751 }, { "epoch": 0.23451214316148275, "grad_norm": 20.5, "learning_rate": 2.42140321801142e-06, "loss": 1.1552, "step": 2752 }, { "epoch": 0.2345973583297827, "grad_norm": 14.0, "learning_rate": 2.421344562952693e-06, "loss": 0.5761, "step": 2753 }, { "epoch": 0.23468257349808266, "grad_norm": 13.5, "learning_rate": 2.421285886726592e-06, "loss": 0.7395, "step": 2754 }, { "epoch": 0.23476778866638262, "grad_norm": 19.125, "learning_rate": 2.421227189334177e-06, "loss": 0.9516, "step": 2755 }, { "epoch": 0.23485300383468258, "grad_norm": 15.0, "learning_rate": 2.4211684707765094e-06, "loss": 1.0189, "step": 2756 }, { "epoch": 0.23493821900298253, "grad_norm": 17.125, "learning_rate": 2.4211097310546495e-06, "loss": 1.0226, "step": 2757 }, { "epoch": 0.2350234341712825, "grad_norm": 26.625, "learning_rate": 2.4210509701696592e-06, "loss": 0.9777, "step": 2758 }, { "epoch": 0.23510864933958245, "grad_norm": 29.125, "learning_rate": 2.4209921881226e-06, "loss": 1.006, "step": 2759 }, { "epoch": 0.2351938645078824, "grad_norm": 14.125, "learning_rate": 2.420933384914535e-06, "loss": 0.833, "step": 2760 }, { "epoch": 0.23527907967618236, "grad_norm": 14.1875, "learning_rate": 2.420874560546526e-06, "loss": 0.9195, "step": 2761 }, { "epoch": 0.23536429484448232, "grad_norm": 13.375, "learning_rate": 2.420815715019636e-06, "loss": 0.84, "step": 2762 }, { "epoch": 0.23544951001278228, "grad_norm": 18.0, "learning_rate": 2.420756848334929e-06, "loss": 0.5436, "step": 2763 }, { "epoch": 0.23553472518108223, "grad_norm": 10.625, "learning_rate": 2.420697960493469e-06, "loss": 0.5073, "step": 2764 }, { "epoch": 0.2356199403493822, "grad_norm": 15.625, "learning_rate": 2.4206390514963186e-06, "loss": 0.5162, "step": 2765 }, { "epoch": 0.23570515551768215, "grad_norm": 20.75, "learning_rate": 2.420580121344544e-06, "loss": 0.9209, "step": 2766 }, { "epoch": 0.2357903706859821, "grad_norm": 12.25, "learning_rate": 2.420521170039209e-06, "loss": 0.5864, "step": 2767 }, { "epoch": 0.23587558585428206, "grad_norm": 15.4375, "learning_rate": 2.42046219758138e-06, "loss": 0.9031, "step": 2768 }, { "epoch": 0.23596080102258202, "grad_norm": 13.0, "learning_rate": 2.420403203972122e-06, "loss": 0.6023, "step": 2769 }, { "epoch": 0.23604601619088197, "grad_norm": 14.5, "learning_rate": 2.4203441892125007e-06, "loss": 0.7756, "step": 2770 }, { "epoch": 0.23613123135918193, "grad_norm": 16.375, "learning_rate": 2.420285153303583e-06, "loss": 0.83, "step": 2771 }, { "epoch": 0.2362164465274819, "grad_norm": 16.375, "learning_rate": 2.4202260962464365e-06, "loss": 0.6134, "step": 2772 }, { "epoch": 0.23630166169578184, "grad_norm": 11.6875, "learning_rate": 2.4201670180421273e-06, "loss": 0.5596, "step": 2773 }, { "epoch": 0.2363868768640818, "grad_norm": 16.0, "learning_rate": 2.420107918691723e-06, "loss": 1.1389, "step": 2774 }, { "epoch": 0.23647209203238176, "grad_norm": 17.625, "learning_rate": 2.4200487981962926e-06, "loss": 1.1048, "step": 2775 }, { "epoch": 0.23655730720068172, "grad_norm": 16.625, "learning_rate": 2.4199896565569037e-06, "loss": 0.7548, "step": 2776 }, { "epoch": 0.23664252236898167, "grad_norm": 14.625, "learning_rate": 2.419930493774625e-06, "loss": 0.6183, "step": 2777 }, { "epoch": 0.23672773753728163, "grad_norm": 14.1875, "learning_rate": 2.419871309850526e-06, "loss": 1.0604, "step": 2778 }, { "epoch": 0.2368129527055816, "grad_norm": 19.625, "learning_rate": 2.419812104785676e-06, "loss": 1.0181, "step": 2779 }, { "epoch": 0.23689816787388154, "grad_norm": 17.375, "learning_rate": 2.4197528785811447e-06, "loss": 0.9889, "step": 2780 }, { "epoch": 0.2369833830421815, "grad_norm": 15.5625, "learning_rate": 2.4196936312380025e-06, "loss": 0.946, "step": 2781 }, { "epoch": 0.23706859821048146, "grad_norm": 11.625, "learning_rate": 2.4196343627573203e-06, "loss": 0.7725, "step": 2782 }, { "epoch": 0.23715381337878141, "grad_norm": 19.0, "learning_rate": 2.4195750731401698e-06, "loss": 0.9169, "step": 2783 }, { "epoch": 0.23723902854708137, "grad_norm": 19.625, "learning_rate": 2.4195157623876206e-06, "loss": 0.891, "step": 2784 }, { "epoch": 0.23732424371538133, "grad_norm": 17.0, "learning_rate": 2.419456430500746e-06, "loss": 1.0242, "step": 2785 }, { "epoch": 0.23740945888368128, "grad_norm": 12.375, "learning_rate": 2.4193970774806175e-06, "loss": 0.7015, "step": 2786 }, { "epoch": 0.23749467405198124, "grad_norm": 14.375, "learning_rate": 2.4193377033283082e-06, "loss": 0.8451, "step": 2787 }, { "epoch": 0.2375798892202812, "grad_norm": 15.5625, "learning_rate": 2.4192783080448906e-06, "loss": 1.0183, "step": 2788 }, { "epoch": 0.23766510438858116, "grad_norm": 11.6875, "learning_rate": 2.419218891631438e-06, "loss": 0.5921, "step": 2789 }, { "epoch": 0.2377503195568811, "grad_norm": 19.5, "learning_rate": 2.4191594540890246e-06, "loss": 0.7494, "step": 2790 }, { "epoch": 0.23783553472518107, "grad_norm": 21.625, "learning_rate": 2.419099995418724e-06, "loss": 0.9608, "step": 2791 }, { "epoch": 0.23792074989348103, "grad_norm": 12.5, "learning_rate": 2.4190405156216105e-06, "loss": 0.5553, "step": 2792 }, { "epoch": 0.238005965061781, "grad_norm": 24.0, "learning_rate": 2.41898101469876e-06, "loss": 0.9755, "step": 2793 }, { "epoch": 0.23809118023008097, "grad_norm": 17.75, "learning_rate": 2.4189214926512467e-06, "loss": 0.8864, "step": 2794 }, { "epoch": 0.23817639539838092, "grad_norm": 16.75, "learning_rate": 2.4188619494801463e-06, "loss": 0.8475, "step": 2795 }, { "epoch": 0.23826161056668088, "grad_norm": 19.5, "learning_rate": 2.4188023851865354e-06, "loss": 0.9995, "step": 2796 }, { "epoch": 0.23834682573498084, "grad_norm": 14.3125, "learning_rate": 2.41874279977149e-06, "loss": 0.612, "step": 2797 }, { "epoch": 0.2384320409032808, "grad_norm": 13.3125, "learning_rate": 2.4186831932360873e-06, "loss": 0.7524, "step": 2798 }, { "epoch": 0.23851725607158075, "grad_norm": 19.75, "learning_rate": 2.4186235655814037e-06, "loss": 1.0977, "step": 2799 }, { "epoch": 0.2386024712398807, "grad_norm": 23.25, "learning_rate": 2.4185639168085175e-06, "loss": 1.1972, "step": 2800 }, { "epoch": 0.23868768640818067, "grad_norm": 18.625, "learning_rate": 2.418504246918506e-06, "loss": 0.735, "step": 2801 }, { "epoch": 0.23877290157648062, "grad_norm": 16.75, "learning_rate": 2.4184445559124476e-06, "loss": 1.0113, "step": 2802 }, { "epoch": 0.23885811674478058, "grad_norm": 13.4375, "learning_rate": 2.4183848437914216e-06, "loss": 0.6332, "step": 2803 }, { "epoch": 0.23894333191308054, "grad_norm": 16.375, "learning_rate": 2.4183251105565058e-06, "loss": 1.0582, "step": 2804 }, { "epoch": 0.2390285470813805, "grad_norm": 12.75, "learning_rate": 2.418265356208781e-06, "loss": 0.8828, "step": 2805 }, { "epoch": 0.23911376224968045, "grad_norm": 12.5, "learning_rate": 2.4182055807493263e-06, "loss": 0.4269, "step": 2806 }, { "epoch": 0.2391989774179804, "grad_norm": 20.875, "learning_rate": 2.418145784179222e-06, "loss": 1.118, "step": 2807 }, { "epoch": 0.23928419258628036, "grad_norm": 13.5625, "learning_rate": 2.4180859664995486e-06, "loss": 0.8421, "step": 2808 }, { "epoch": 0.23936940775458032, "grad_norm": 13.4375, "learning_rate": 2.4180261277113874e-06, "loss": 0.633, "step": 2809 }, { "epoch": 0.23945462292288028, "grad_norm": 14.0, "learning_rate": 2.4179662678158195e-06, "loss": 0.7742, "step": 2810 }, { "epoch": 0.23953983809118023, "grad_norm": 16.75, "learning_rate": 2.4179063868139265e-06, "loss": 1.0349, "step": 2811 }, { "epoch": 0.2396250532594802, "grad_norm": 11.0, "learning_rate": 2.417846484706791e-06, "loss": 0.435, "step": 2812 }, { "epoch": 0.23971026842778015, "grad_norm": 27.5, "learning_rate": 2.417786561495495e-06, "loss": 1.2232, "step": 2813 }, { "epoch": 0.2397954835960801, "grad_norm": 30.875, "learning_rate": 2.417726617181121e-06, "loss": 0.952, "step": 2814 }, { "epoch": 0.23988069876438006, "grad_norm": 15.5, "learning_rate": 2.4176666517647534e-06, "loss": 0.7361, "step": 2815 }, { "epoch": 0.23996591393268002, "grad_norm": 14.375, "learning_rate": 2.417606665247475e-06, "loss": 0.7363, "step": 2816 }, { "epoch": 0.24005112910097998, "grad_norm": 14.0, "learning_rate": 2.4175466576303703e-06, "loss": 0.7376, "step": 2817 }, { "epoch": 0.24013634426927993, "grad_norm": 18.875, "learning_rate": 2.417486628914523e-06, "loss": 1.02, "step": 2818 }, { "epoch": 0.2402215594375799, "grad_norm": 21.625, "learning_rate": 2.4174265791010187e-06, "loss": 0.9254, "step": 2819 }, { "epoch": 0.24030677460587985, "grad_norm": 19.5, "learning_rate": 2.417366508190942e-06, "loss": 0.6651, "step": 2820 }, { "epoch": 0.2403919897741798, "grad_norm": 12.25, "learning_rate": 2.4173064161853788e-06, "loss": 0.7512, "step": 2821 }, { "epoch": 0.24047720494247976, "grad_norm": 15.0, "learning_rate": 2.4172463030854146e-06, "loss": 0.5555, "step": 2822 }, { "epoch": 0.24056242011077972, "grad_norm": 13.75, "learning_rate": 2.417186168892136e-06, "loss": 0.715, "step": 2823 }, { "epoch": 0.24064763527907967, "grad_norm": 20.25, "learning_rate": 2.41712601360663e-06, "loss": 1.1555, "step": 2824 }, { "epoch": 0.24073285044737963, "grad_norm": 12.25, "learning_rate": 2.4170658372299827e-06, "loss": 0.3854, "step": 2825 }, { "epoch": 0.2408180656156796, "grad_norm": 10.5625, "learning_rate": 2.4170056397632823e-06, "loss": 0.484, "step": 2826 }, { "epoch": 0.24090328078397955, "grad_norm": 15.375, "learning_rate": 2.4169454212076167e-06, "loss": 1.2465, "step": 2827 }, { "epoch": 0.2409884959522795, "grad_norm": 14.5, "learning_rate": 2.416885181564074e-06, "loss": 0.839, "step": 2828 }, { "epoch": 0.24107371112057946, "grad_norm": 15.375, "learning_rate": 2.4168249208337422e-06, "loss": 1.0157, "step": 2829 }, { "epoch": 0.24115892628887942, "grad_norm": 21.875, "learning_rate": 2.4167646390177115e-06, "loss": 1.0009, "step": 2830 }, { "epoch": 0.24124414145717937, "grad_norm": 23.125, "learning_rate": 2.41670433611707e-06, "loss": 0.9316, "step": 2831 }, { "epoch": 0.24132935662547933, "grad_norm": 12.25, "learning_rate": 2.416644012132908e-06, "loss": 0.5364, "step": 2832 }, { "epoch": 0.2414145717937793, "grad_norm": 13.0, "learning_rate": 2.416583667066315e-06, "loss": 0.6709, "step": 2833 }, { "epoch": 0.24149978696207924, "grad_norm": 14.875, "learning_rate": 2.4165233009183827e-06, "loss": 0.9483, "step": 2834 }, { "epoch": 0.2415850021303792, "grad_norm": 10.5625, "learning_rate": 2.416462913690201e-06, "loss": 0.9666, "step": 2835 }, { "epoch": 0.24167021729867916, "grad_norm": 12.0, "learning_rate": 2.416402505382862e-06, "loss": 0.5579, "step": 2836 }, { "epoch": 0.24175543246697911, "grad_norm": 14.0625, "learning_rate": 2.416342075997456e-06, "loss": 0.734, "step": 2837 }, { "epoch": 0.24184064763527907, "grad_norm": 15.5, "learning_rate": 2.4162816255350764e-06, "loss": 0.6405, "step": 2838 }, { "epoch": 0.24192586280357903, "grad_norm": 16.125, "learning_rate": 2.416221153996815e-06, "loss": 0.6583, "step": 2839 }, { "epoch": 0.24201107797187899, "grad_norm": 14.4375, "learning_rate": 2.4161606613837645e-06, "loss": 0.8384, "step": 2840 }, { "epoch": 0.24209629314017894, "grad_norm": 15.75, "learning_rate": 2.4161001476970186e-06, "loss": 0.7441, "step": 2841 }, { "epoch": 0.2421815083084789, "grad_norm": 18.25, "learning_rate": 2.41603961293767e-06, "loss": 1.0302, "step": 2842 }, { "epoch": 0.24226672347677886, "grad_norm": 14.4375, "learning_rate": 2.415979057106813e-06, "loss": 0.6088, "step": 2843 }, { "epoch": 0.2423519386450788, "grad_norm": 18.625, "learning_rate": 2.415918480205542e-06, "loss": 0.98, "step": 2844 }, { "epoch": 0.24243715381337877, "grad_norm": 20.625, "learning_rate": 2.4158578822349516e-06, "loss": 1.1366, "step": 2845 }, { "epoch": 0.24252236898167873, "grad_norm": 13.8125, "learning_rate": 2.4157972631961374e-06, "loss": 0.7558, "step": 2846 }, { "epoch": 0.24260758414997868, "grad_norm": 26.875, "learning_rate": 2.415736623090194e-06, "loss": 0.5989, "step": 2847 }, { "epoch": 0.24269279931827864, "grad_norm": 16.125, "learning_rate": 2.4156759619182175e-06, "loss": 0.7427, "step": 2848 }, { "epoch": 0.2427780144865786, "grad_norm": 12.0625, "learning_rate": 2.4156152796813044e-06, "loss": 0.5771, "step": 2849 }, { "epoch": 0.24286322965487858, "grad_norm": 15.5625, "learning_rate": 2.415554576380551e-06, "loss": 0.6352, "step": 2850 }, { "epoch": 0.24294844482317854, "grad_norm": 11.0, "learning_rate": 2.415493852017054e-06, "loss": 0.5072, "step": 2851 }, { "epoch": 0.2430336599914785, "grad_norm": 13.5625, "learning_rate": 2.415433106591911e-06, "loss": 0.7335, "step": 2852 }, { "epoch": 0.24311887515977845, "grad_norm": 12.4375, "learning_rate": 2.4153723401062207e-06, "loss": 0.4162, "step": 2853 }, { "epoch": 0.2432040903280784, "grad_norm": 16.125, "learning_rate": 2.4153115525610797e-06, "loss": 0.6565, "step": 2854 }, { "epoch": 0.24328930549637837, "grad_norm": 14.25, "learning_rate": 2.4152507439575874e-06, "loss": 0.5961, "step": 2855 }, { "epoch": 0.24337452066467832, "grad_norm": 14.3125, "learning_rate": 2.415189914296842e-06, "loss": 0.6748, "step": 2856 }, { "epoch": 0.24345973583297828, "grad_norm": 13.4375, "learning_rate": 2.4151290635799436e-06, "loss": 0.6063, "step": 2857 }, { "epoch": 0.24354495100127824, "grad_norm": 15.375, "learning_rate": 2.4150681918079914e-06, "loss": 0.9039, "step": 2858 }, { "epoch": 0.2436301661695782, "grad_norm": 18.875, "learning_rate": 2.4150072989820853e-06, "loss": 0.815, "step": 2859 }, { "epoch": 0.24371538133787815, "grad_norm": 79.0, "learning_rate": 2.4149463851033257e-06, "loss": 0.7938, "step": 2860 }, { "epoch": 0.2438005965061781, "grad_norm": 14.5, "learning_rate": 2.4148854501728137e-06, "loss": 0.8861, "step": 2861 }, { "epoch": 0.24388581167447806, "grad_norm": 13.5625, "learning_rate": 2.41482449419165e-06, "loss": 0.7299, "step": 2862 }, { "epoch": 0.24397102684277802, "grad_norm": 14.875, "learning_rate": 2.414763517160936e-06, "loss": 0.8017, "step": 2863 }, { "epoch": 0.24405624201107798, "grad_norm": 12.9375, "learning_rate": 2.4147025190817745e-06, "loss": 0.4447, "step": 2864 }, { "epoch": 0.24414145717937794, "grad_norm": 14.3125, "learning_rate": 2.4146414999552677e-06, "loss": 0.7857, "step": 2865 }, { "epoch": 0.2442266723476779, "grad_norm": 18.625, "learning_rate": 2.414580459782517e-06, "loss": 0.5316, "step": 2866 }, { "epoch": 0.24431188751597785, "grad_norm": 45.0, "learning_rate": 2.414519398564627e-06, "loss": 0.7077, "step": 2867 }, { "epoch": 0.2443971026842778, "grad_norm": 14.4375, "learning_rate": 2.4144583163027e-06, "loss": 0.5152, "step": 2868 }, { "epoch": 0.24448231785257776, "grad_norm": 13.8125, "learning_rate": 2.4143972129978406e-06, "loss": 0.6902, "step": 2869 }, { "epoch": 0.24456753302087772, "grad_norm": 13.25, "learning_rate": 2.4143360886511523e-06, "loss": 0.8894, "step": 2870 }, { "epoch": 0.24465274818917768, "grad_norm": 16.625, "learning_rate": 2.4142749432637404e-06, "loss": 1.2033, "step": 2871 }, { "epoch": 0.24473796335747763, "grad_norm": 19.75, "learning_rate": 2.414213776836709e-06, "loss": 0.9076, "step": 2872 }, { "epoch": 0.2448231785257776, "grad_norm": 17.625, "learning_rate": 2.4141525893711643e-06, "loss": 0.72, "step": 2873 }, { "epoch": 0.24490839369407755, "grad_norm": 34.75, "learning_rate": 2.4140913808682117e-06, "loss": 1.1777, "step": 2874 }, { "epoch": 0.2449936088623775, "grad_norm": 12.125, "learning_rate": 2.414030151328957e-06, "loss": 0.623, "step": 2875 }, { "epoch": 0.24507882403067746, "grad_norm": 15.25, "learning_rate": 2.4139689007545072e-06, "loss": 0.9947, "step": 2876 }, { "epoch": 0.24516403919897742, "grad_norm": 19.125, "learning_rate": 2.413907629145969e-06, "loss": 1.0784, "step": 2877 }, { "epoch": 0.24524925436727738, "grad_norm": 14.9375, "learning_rate": 2.4138463365044492e-06, "loss": 0.9475, "step": 2878 }, { "epoch": 0.24533446953557733, "grad_norm": 16.5, "learning_rate": 2.413785022831056e-06, "loss": 0.9571, "step": 2879 }, { "epoch": 0.2454196847038773, "grad_norm": 12.8125, "learning_rate": 2.413723688126897e-06, "loss": 0.6369, "step": 2880 }, { "epoch": 0.24550489987217725, "grad_norm": 13.25, "learning_rate": 2.413662332393081e-06, "loss": 0.6627, "step": 2881 }, { "epoch": 0.2455901150404772, "grad_norm": 12.875, "learning_rate": 2.4136009556307162e-06, "loss": 0.6959, "step": 2882 }, { "epoch": 0.24567533020877716, "grad_norm": 13.5, "learning_rate": 2.413539557840912e-06, "loss": 0.6722, "step": 2883 }, { "epoch": 0.24576054537707712, "grad_norm": 13.8125, "learning_rate": 2.4134781390247785e-06, "loss": 0.5039, "step": 2884 }, { "epoch": 0.24584576054537707, "grad_norm": 16.5, "learning_rate": 2.4134166991834247e-06, "loss": 0.5201, "step": 2885 }, { "epoch": 0.24593097571367703, "grad_norm": 16.75, "learning_rate": 2.413355238317961e-06, "loss": 0.7207, "step": 2886 }, { "epoch": 0.246016190881977, "grad_norm": 15.75, "learning_rate": 2.413293756429499e-06, "loss": 1.0481, "step": 2887 }, { "epoch": 0.24610140605027694, "grad_norm": 13.8125, "learning_rate": 2.4132322535191482e-06, "loss": 0.7666, "step": 2888 }, { "epoch": 0.2461866212185769, "grad_norm": 17.0, "learning_rate": 2.4131707295880213e-06, "loss": 0.8501, "step": 2889 }, { "epoch": 0.24627183638687686, "grad_norm": 14.5625, "learning_rate": 2.4131091846372293e-06, "loss": 0.8717, "step": 2890 }, { "epoch": 0.24635705155517681, "grad_norm": 13.0625, "learning_rate": 2.4130476186678852e-06, "loss": 0.6441, "step": 2891 }, { "epoch": 0.24644226672347677, "grad_norm": 15.5, "learning_rate": 2.412986031681101e-06, "loss": 0.8361, "step": 2892 }, { "epoch": 0.24652748189177673, "grad_norm": 17.375, "learning_rate": 2.4129244236779897e-06, "loss": 0.6939, "step": 2893 }, { "epoch": 0.24661269706007669, "grad_norm": 19.25, "learning_rate": 2.4128627946596648e-06, "loss": 0.8974, "step": 2894 }, { "epoch": 0.24669791222837664, "grad_norm": 13.125, "learning_rate": 2.4128011446272397e-06, "loss": 0.571, "step": 2895 }, { "epoch": 0.2467831273966766, "grad_norm": 16.0, "learning_rate": 2.4127394735818283e-06, "loss": 1.073, "step": 2896 }, { "epoch": 0.24686834256497656, "grad_norm": 18.625, "learning_rate": 2.412677781524546e-06, "loss": 0.8276, "step": 2897 }, { "epoch": 0.2469535577332765, "grad_norm": 13.875, "learning_rate": 2.4126160684565065e-06, "loss": 0.9196, "step": 2898 }, { "epoch": 0.24703877290157647, "grad_norm": 21.375, "learning_rate": 2.412554334378826e-06, "loss": 0.8342, "step": 2899 }, { "epoch": 0.24712398806987643, "grad_norm": 11.5625, "learning_rate": 2.4124925792926195e-06, "loss": 0.4633, "step": 2900 }, { "epoch": 0.24720920323817638, "grad_norm": 15.5625, "learning_rate": 2.4124308031990028e-06, "loss": 0.8989, "step": 2901 }, { "epoch": 0.24729441840647634, "grad_norm": 11.0, "learning_rate": 2.4123690060990927e-06, "loss": 0.3603, "step": 2902 }, { "epoch": 0.2473796335747763, "grad_norm": 13.0, "learning_rate": 2.412307187994006e-06, "loss": 0.5763, "step": 2903 }, { "epoch": 0.24746484874307625, "grad_norm": 16.375, "learning_rate": 2.4122453488848595e-06, "loss": 0.6921, "step": 2904 }, { "epoch": 0.2475500639113762, "grad_norm": 33.25, "learning_rate": 2.412183488772771e-06, "loss": 0.9493, "step": 2905 }, { "epoch": 0.2476352790796762, "grad_norm": 14.0, "learning_rate": 2.4121216076588577e-06, "loss": 0.7555, "step": 2906 }, { "epoch": 0.24772049424797615, "grad_norm": 17.5, "learning_rate": 2.412059705544239e-06, "loss": 0.5751, "step": 2907 }, { "epoch": 0.2478057094162761, "grad_norm": 12.4375, "learning_rate": 2.4119977824300324e-06, "loss": 0.5633, "step": 2908 }, { "epoch": 0.24789092458457607, "grad_norm": 17.875, "learning_rate": 2.4119358383173573e-06, "loss": 0.7717, "step": 2909 }, { "epoch": 0.24797613975287602, "grad_norm": 14.0625, "learning_rate": 2.411873873207334e-06, "loss": 0.8224, "step": 2910 }, { "epoch": 0.24806135492117598, "grad_norm": 15.1875, "learning_rate": 2.41181188710108e-06, "loss": 0.7632, "step": 2911 }, { "epoch": 0.24814657008947594, "grad_norm": 12.4375, "learning_rate": 2.411749879999718e-06, "loss": 0.6182, "step": 2912 }, { "epoch": 0.2482317852577759, "grad_norm": 11.6875, "learning_rate": 2.4116878519043676e-06, "loss": 0.529, "step": 2913 }, { "epoch": 0.24831700042607585, "grad_norm": 14.0, "learning_rate": 2.411625802816149e-06, "loss": 0.6263, "step": 2914 }, { "epoch": 0.2484022155943758, "grad_norm": 15.8125, "learning_rate": 2.4115637327361844e-06, "loss": 0.9153, "step": 2915 }, { "epoch": 0.24848743076267577, "grad_norm": 13.5, "learning_rate": 2.4115016416655947e-06, "loss": 0.8225, "step": 2916 }, { "epoch": 0.24857264593097572, "grad_norm": 13.5, "learning_rate": 2.411439529605503e-06, "loss": 0.8688, "step": 2917 }, { "epoch": 0.24865786109927568, "grad_norm": 12.125, "learning_rate": 2.41137739655703e-06, "loss": 0.5164, "step": 2918 }, { "epoch": 0.24874307626757564, "grad_norm": 12.75, "learning_rate": 2.4113152425213006e-06, "loss": 0.6873, "step": 2919 }, { "epoch": 0.2488282914358756, "grad_norm": 12.375, "learning_rate": 2.4112530674994367e-06, "loss": 0.4927, "step": 2920 }, { "epoch": 0.24891350660417555, "grad_norm": 14.75, "learning_rate": 2.411190871492562e-06, "loss": 0.4649, "step": 2921 }, { "epoch": 0.2489987217724755, "grad_norm": 14.0625, "learning_rate": 2.4111286545018006e-06, "loss": 0.8986, "step": 2922 }, { "epoch": 0.24908393694077546, "grad_norm": 9.8125, "learning_rate": 2.411066416528277e-06, "loss": 0.321, "step": 2923 }, { "epoch": 0.24916915210907542, "grad_norm": 16.0, "learning_rate": 2.4110041575731156e-06, "loss": 1.0307, "step": 2924 }, { "epoch": 0.24925436727737538, "grad_norm": 15.4375, "learning_rate": 2.4109418776374417e-06, "loss": 0.7543, "step": 2925 }, { "epoch": 0.24933958244567533, "grad_norm": 31.0, "learning_rate": 2.4108795767223803e-06, "loss": 1.2791, "step": 2926 }, { "epoch": 0.2494247976139753, "grad_norm": 13.3125, "learning_rate": 2.4108172548290576e-06, "loss": 0.8779, "step": 2927 }, { "epoch": 0.24951001278227525, "grad_norm": 20.625, "learning_rate": 2.4107549119586e-06, "loss": 1.0068, "step": 2928 }, { "epoch": 0.2495952279505752, "grad_norm": 13.6875, "learning_rate": 2.410692548112134e-06, "loss": 0.7455, "step": 2929 }, { "epoch": 0.24968044311887516, "grad_norm": 13.9375, "learning_rate": 2.4106301632907865e-06, "loss": 0.8551, "step": 2930 }, { "epoch": 0.24976565828717512, "grad_norm": 22.125, "learning_rate": 2.410567757495685e-06, "loss": 1.1662, "step": 2931 }, { "epoch": 0.24985087345547508, "grad_norm": 17.625, "learning_rate": 2.4105053307279567e-06, "loss": 1.0824, "step": 2932 }, { "epoch": 0.24993608862377503, "grad_norm": 9.6875, "learning_rate": 2.41044288298873e-06, "loss": 0.367, "step": 2933 }, { "epoch": 0.250021303792075, "grad_norm": 11.8125, "learning_rate": 2.410380414279134e-06, "loss": 0.5967, "step": 2934 }, { "epoch": 0.25010651896037495, "grad_norm": 15.0625, "learning_rate": 2.4103179246002967e-06, "loss": 0.6324, "step": 2935 }, { "epoch": 0.2501917341286749, "grad_norm": 15.375, "learning_rate": 2.410255413953348e-06, "loss": 0.8187, "step": 2936 }, { "epoch": 0.25027694929697486, "grad_norm": 21.0, "learning_rate": 2.4101928823394167e-06, "loss": 0.6911, "step": 2937 }, { "epoch": 0.2503621644652748, "grad_norm": 19.625, "learning_rate": 2.410130329759634e-06, "loss": 1.0849, "step": 2938 }, { "epoch": 0.2504473796335748, "grad_norm": 21.875, "learning_rate": 2.4100677562151294e-06, "loss": 1.3756, "step": 2939 }, { "epoch": 0.25053259480187473, "grad_norm": 17.25, "learning_rate": 2.4100051617070334e-06, "loss": 1.3955, "step": 2940 }, { "epoch": 0.2506178099701747, "grad_norm": 38.0, "learning_rate": 2.4099425462364784e-06, "loss": 1.5249, "step": 2941 }, { "epoch": 0.25070302513847464, "grad_norm": 33.25, "learning_rate": 2.4098799098045948e-06, "loss": 1.2549, "step": 2942 }, { "epoch": 0.2507882403067746, "grad_norm": 12.0, "learning_rate": 2.4098172524125153e-06, "loss": 0.566, "step": 2943 }, { "epoch": 0.25087345547507456, "grad_norm": 13.625, "learning_rate": 2.409754574061371e-06, "loss": 0.7195, "step": 2944 }, { "epoch": 0.2509586706433745, "grad_norm": 11.5625, "learning_rate": 2.4096918747522956e-06, "loss": 0.5083, "step": 2945 }, { "epoch": 0.25104388581167447, "grad_norm": 11.125, "learning_rate": 2.409629154486422e-06, "loss": 0.4639, "step": 2946 }, { "epoch": 0.25112910097997443, "grad_norm": 13.75, "learning_rate": 2.409566413264884e-06, "loss": 0.6159, "step": 2947 }, { "epoch": 0.2512143161482744, "grad_norm": 18.125, "learning_rate": 2.409503651088814e-06, "loss": 0.7635, "step": 2948 }, { "epoch": 0.25129953131657434, "grad_norm": 14.75, "learning_rate": 2.4094408679593477e-06, "loss": 0.8745, "step": 2949 }, { "epoch": 0.2513847464848743, "grad_norm": 13.0625, "learning_rate": 2.4093780638776188e-06, "loss": 1.0396, "step": 2950 }, { "epoch": 0.25146996165317426, "grad_norm": 14.6875, "learning_rate": 2.409315238844762e-06, "loss": 0.5989, "step": 2951 }, { "epoch": 0.2515551768214742, "grad_norm": 14.75, "learning_rate": 2.4092523928619137e-06, "loss": 1.1529, "step": 2952 }, { "epoch": 0.25164039198977417, "grad_norm": 18.25, "learning_rate": 2.4091895259302084e-06, "loss": 1.2555, "step": 2953 }, { "epoch": 0.2517256071580741, "grad_norm": 13.9375, "learning_rate": 2.409126638050783e-06, "loss": 0.6195, "step": 2954 }, { "epoch": 0.2518108223263741, "grad_norm": 16.5, "learning_rate": 2.4090637292247737e-06, "loss": 0.9051, "step": 2955 }, { "epoch": 0.25189603749467404, "grad_norm": 13.5625, "learning_rate": 2.409000799453317e-06, "loss": 0.8554, "step": 2956 }, { "epoch": 0.251981252662974, "grad_norm": 13.6875, "learning_rate": 2.408937848737551e-06, "loss": 0.9069, "step": 2957 }, { "epoch": 0.25206646783127395, "grad_norm": 16.375, "learning_rate": 2.408874877078612e-06, "loss": 1.0019, "step": 2958 }, { "epoch": 0.2521516829995739, "grad_norm": 12.5625, "learning_rate": 2.408811884477639e-06, "loss": 0.7156, "step": 2959 }, { "epoch": 0.25223689816787387, "grad_norm": 18.25, "learning_rate": 2.4087488709357698e-06, "loss": 0.95, "step": 2960 }, { "epoch": 0.2523221133361738, "grad_norm": 19.0, "learning_rate": 2.4086858364541437e-06, "loss": 1.1111, "step": 2961 }, { "epoch": 0.2524073285044738, "grad_norm": 13.9375, "learning_rate": 2.4086227810338987e-06, "loss": 0.7465, "step": 2962 }, { "epoch": 0.25249254367277374, "grad_norm": 12.25, "learning_rate": 2.4085597046761754e-06, "loss": 0.5839, "step": 2963 }, { "epoch": 0.2525777588410737, "grad_norm": 17.5, "learning_rate": 2.408496607382113e-06, "loss": 0.939, "step": 2964 }, { "epoch": 0.25266297400937365, "grad_norm": 17.0, "learning_rate": 2.408433489152852e-06, "loss": 1.0018, "step": 2965 }, { "epoch": 0.2527481891776736, "grad_norm": 17.625, "learning_rate": 2.4083703499895335e-06, "loss": 1.0089, "step": 2966 }, { "epoch": 0.25283340434597357, "grad_norm": 14.375, "learning_rate": 2.4083071898932975e-06, "loss": 1.0389, "step": 2967 }, { "epoch": 0.2529186195142735, "grad_norm": 15.625, "learning_rate": 2.4082440088652857e-06, "loss": 0.7878, "step": 2968 }, { "epoch": 0.2530038346825735, "grad_norm": 14.0625, "learning_rate": 2.4081808069066396e-06, "loss": 1.1512, "step": 2969 }, { "epoch": 0.25308904985087344, "grad_norm": 17.5, "learning_rate": 2.4081175840185023e-06, "loss": 0.9583, "step": 2970 }, { "epoch": 0.2531742650191734, "grad_norm": 16.625, "learning_rate": 2.4080543402020155e-06, "loss": 0.4247, "step": 2971 }, { "epoch": 0.25325948018747335, "grad_norm": 11.125, "learning_rate": 2.4079910754583223e-06, "loss": 0.4451, "step": 2972 }, { "epoch": 0.2533446953557733, "grad_norm": 17.625, "learning_rate": 2.4079277897885662e-06, "loss": 0.6546, "step": 2973 }, { "epoch": 0.25342991052407327, "grad_norm": 12.625, "learning_rate": 2.40786448319389e-06, "loss": 0.5652, "step": 2974 }, { "epoch": 0.2535151256923732, "grad_norm": 19.25, "learning_rate": 2.4078011556754386e-06, "loss": 0.7326, "step": 2975 }, { "epoch": 0.2536003408606732, "grad_norm": 13.75, "learning_rate": 2.407737807234356e-06, "loss": 0.5198, "step": 2976 }, { "epoch": 0.25368555602897314, "grad_norm": 14.9375, "learning_rate": 2.407674437871787e-06, "loss": 1.0169, "step": 2977 }, { "epoch": 0.2537707711972731, "grad_norm": 11.1875, "learning_rate": 2.4076110475888763e-06, "loss": 0.3128, "step": 2978 }, { "epoch": 0.25385598636557305, "grad_norm": 15.8125, "learning_rate": 2.4075476363867707e-06, "loss": 0.8734, "step": 2979 }, { "epoch": 0.253941201533873, "grad_norm": 14.1875, "learning_rate": 2.4074842042666147e-06, "loss": 0.8517, "step": 2980 }, { "epoch": 0.25402641670217296, "grad_norm": 16.375, "learning_rate": 2.4074207512295557e-06, "loss": 0.8789, "step": 2981 }, { "epoch": 0.2541116318704729, "grad_norm": 14.0625, "learning_rate": 2.4073572772767396e-06, "loss": 0.604, "step": 2982 }, { "epoch": 0.2541968470387729, "grad_norm": 33.75, "learning_rate": 2.4072937824093138e-06, "loss": 0.9227, "step": 2983 }, { "epoch": 0.25428206220707283, "grad_norm": 18.625, "learning_rate": 2.4072302666284255e-06, "loss": 0.8889, "step": 2984 }, { "epoch": 0.2543672773753728, "grad_norm": 16.875, "learning_rate": 2.4071667299352225e-06, "loss": 0.7892, "step": 2985 }, { "epoch": 0.25445249254367275, "grad_norm": 16.125, "learning_rate": 2.407103172330853e-06, "loss": 0.647, "step": 2986 }, { "epoch": 0.2545377077119727, "grad_norm": 19.625, "learning_rate": 2.407039593816466e-06, "loss": 1.0276, "step": 2987 }, { "epoch": 0.25462292288027266, "grad_norm": 18.125, "learning_rate": 2.4069759943932096e-06, "loss": 0.5737, "step": 2988 }, { "epoch": 0.2547081380485726, "grad_norm": 11.5, "learning_rate": 2.406912374062234e-06, "loss": 0.8007, "step": 2989 }, { "epoch": 0.2547933532168726, "grad_norm": 11.625, "learning_rate": 2.406848732824688e-06, "loss": 0.4117, "step": 2990 }, { "epoch": 0.2548785683851726, "grad_norm": 14.25, "learning_rate": 2.4067850706817224e-06, "loss": 0.6301, "step": 2991 }, { "epoch": 0.25496378355347254, "grad_norm": 15.1875, "learning_rate": 2.4067213876344873e-06, "loss": 0.7766, "step": 2992 }, { "epoch": 0.2550489987217725, "grad_norm": 16.75, "learning_rate": 2.4066576836841334e-06, "loss": 0.9498, "step": 2993 }, { "epoch": 0.25513421389007246, "grad_norm": 17.75, "learning_rate": 2.406593958831812e-06, "loss": 0.9111, "step": 2994 }, { "epoch": 0.2552194290583724, "grad_norm": 20.625, "learning_rate": 2.406530213078675e-06, "loss": 0.7957, "step": 2995 }, { "epoch": 0.2553046442266724, "grad_norm": 14.4375, "learning_rate": 2.406466446425874e-06, "loss": 0.618, "step": 2996 }, { "epoch": 0.25538985939497233, "grad_norm": 14.1875, "learning_rate": 2.4064026588745616e-06, "loss": 0.6432, "step": 2997 }, { "epoch": 0.2554750745632723, "grad_norm": 13.4375, "learning_rate": 2.4063388504258897e-06, "loss": 0.7925, "step": 2998 }, { "epoch": 0.25556028973157224, "grad_norm": 18.125, "learning_rate": 2.4062750210810124e-06, "loss": 1.2936, "step": 2999 }, { "epoch": 0.2556455048998722, "grad_norm": 14.875, "learning_rate": 2.4062111708410825e-06, "loss": 0.8411, "step": 3000 }, { "epoch": 0.25573072006817216, "grad_norm": 10.875, "learning_rate": 2.406147299707254e-06, "loss": 0.4682, "step": 3001 }, { "epoch": 0.2558159352364721, "grad_norm": 11.8125, "learning_rate": 2.406083407680682e-06, "loss": 0.6219, "step": 3002 }, { "epoch": 0.25590115040477207, "grad_norm": 17.0, "learning_rate": 2.4060194947625195e-06, "loss": 0.9872, "step": 3003 }, { "epoch": 0.25598636557307203, "grad_norm": 13.5625, "learning_rate": 2.4059555609539226e-06, "loss": 0.6533, "step": 3004 }, { "epoch": 0.256071580741372, "grad_norm": 17.875, "learning_rate": 2.405891606256046e-06, "loss": 1.0268, "step": 3005 }, { "epoch": 0.25615679590967194, "grad_norm": 16.625, "learning_rate": 2.4058276306700462e-06, "loss": 0.8031, "step": 3006 }, { "epoch": 0.2562420110779719, "grad_norm": 19.625, "learning_rate": 2.4057636341970786e-06, "loss": 0.8053, "step": 3007 }, { "epoch": 0.25632722624627186, "grad_norm": 71.5, "learning_rate": 2.4056996168382996e-06, "loss": 0.8514, "step": 3008 }, { "epoch": 0.2564124414145718, "grad_norm": 17.75, "learning_rate": 2.405635578594867e-06, "loss": 1.1079, "step": 3009 }, { "epoch": 0.25649765658287177, "grad_norm": 15.0, "learning_rate": 2.405571519467937e-06, "loss": 0.8001, "step": 3010 }, { "epoch": 0.2565828717511717, "grad_norm": 16.75, "learning_rate": 2.4055074394586676e-06, "loss": 0.8068, "step": 3011 }, { "epoch": 0.2566680869194717, "grad_norm": 12.0625, "learning_rate": 2.405443338568217e-06, "loss": 0.5894, "step": 3012 }, { "epoch": 0.25675330208777164, "grad_norm": 16.5, "learning_rate": 2.405379216797743e-06, "loss": 0.7341, "step": 3013 }, { "epoch": 0.2568385172560716, "grad_norm": 14.5, "learning_rate": 2.4053150741484056e-06, "loss": 0.6545, "step": 3014 }, { "epoch": 0.25692373242437155, "grad_norm": 14.5625, "learning_rate": 2.4052509106213623e-06, "loss": 0.7748, "step": 3015 }, { "epoch": 0.2570089475926715, "grad_norm": 28.125, "learning_rate": 2.4051867262177736e-06, "loss": 1.1048, "step": 3016 }, { "epoch": 0.25709416276097147, "grad_norm": 11.3125, "learning_rate": 2.4051225209387992e-06, "loss": 0.6033, "step": 3017 }, { "epoch": 0.2571793779292714, "grad_norm": 23.375, "learning_rate": 2.4050582947855987e-06, "loss": 1.2305, "step": 3018 }, { "epoch": 0.2572645930975714, "grad_norm": 22.75, "learning_rate": 2.404994047759334e-06, "loss": 0.9808, "step": 3019 }, { "epoch": 0.25734980826587134, "grad_norm": 14.4375, "learning_rate": 2.4049297798611654e-06, "loss": 0.8482, "step": 3020 }, { "epoch": 0.2574350234341713, "grad_norm": 14.8125, "learning_rate": 2.4048654910922537e-06, "loss": 0.5807, "step": 3021 }, { "epoch": 0.25752023860247125, "grad_norm": 12.0625, "learning_rate": 2.404801181453762e-06, "loss": 0.7013, "step": 3022 }, { "epoch": 0.2576054537707712, "grad_norm": 17.375, "learning_rate": 2.404736850946851e-06, "loss": 0.7474, "step": 3023 }, { "epoch": 0.25769066893907117, "grad_norm": 15.125, "learning_rate": 2.4046724995726844e-06, "loss": 0.6789, "step": 3024 }, { "epoch": 0.2577758841073711, "grad_norm": 10.875, "learning_rate": 2.4046081273324246e-06, "loss": 0.394, "step": 3025 }, { "epoch": 0.2578610992756711, "grad_norm": 13.0, "learning_rate": 2.404543734227235e-06, "loss": 0.8614, "step": 3026 }, { "epoch": 0.25794631444397104, "grad_norm": 13.125, "learning_rate": 2.404479320258279e-06, "loss": 0.4764, "step": 3027 }, { "epoch": 0.258031529612271, "grad_norm": 13.375, "learning_rate": 2.4044148854267203e-06, "loss": 0.7893, "step": 3028 }, { "epoch": 0.25811674478057095, "grad_norm": 14.625, "learning_rate": 2.4043504297337244e-06, "loss": 0.4692, "step": 3029 }, { "epoch": 0.2582019599488709, "grad_norm": 14.8125, "learning_rate": 2.404285953180455e-06, "loss": 0.8938, "step": 3030 }, { "epoch": 0.25828717511717086, "grad_norm": 12.5, "learning_rate": 2.4042214557680783e-06, "loss": 0.6956, "step": 3031 }, { "epoch": 0.2583723902854708, "grad_norm": 21.5, "learning_rate": 2.4041569374977587e-06, "loss": 1.0409, "step": 3032 }, { "epoch": 0.2584576054537708, "grad_norm": 18.25, "learning_rate": 2.4040923983706626e-06, "loss": 1.004, "step": 3033 }, { "epoch": 0.25854282062207073, "grad_norm": 19.375, "learning_rate": 2.404027838387957e-06, "loss": 1.1083, "step": 3034 }, { "epoch": 0.2586280357903707, "grad_norm": 17.125, "learning_rate": 2.403963257550807e-06, "loss": 0.7026, "step": 3035 }, { "epoch": 0.25871325095867065, "grad_norm": 10.9375, "learning_rate": 2.403898655860381e-06, "loss": 0.4245, "step": 3036 }, { "epoch": 0.2587984661269706, "grad_norm": 15.6875, "learning_rate": 2.403834033317846e-06, "loss": 0.6608, "step": 3037 }, { "epoch": 0.25888368129527056, "grad_norm": 12.9375, "learning_rate": 2.4037693899243696e-06, "loss": 0.6097, "step": 3038 }, { "epoch": 0.2589688964635705, "grad_norm": 21.0, "learning_rate": 2.4037047256811204e-06, "loss": 1.1294, "step": 3039 }, { "epoch": 0.2590541116318705, "grad_norm": 12.8125, "learning_rate": 2.403640040589266e-06, "loss": 0.6969, "step": 3040 }, { "epoch": 0.25913932680017043, "grad_norm": 12.5625, "learning_rate": 2.403575334649977e-06, "loss": 0.7207, "step": 3041 }, { "epoch": 0.2592245419684704, "grad_norm": 14.625, "learning_rate": 2.403510607864421e-06, "loss": 0.8406, "step": 3042 }, { "epoch": 0.25930975713677035, "grad_norm": 18.75, "learning_rate": 2.4034458602337684e-06, "loss": 1.0368, "step": 3043 }, { "epoch": 0.2593949723050703, "grad_norm": 18.75, "learning_rate": 2.403381091759189e-06, "loss": 1.1469, "step": 3044 }, { "epoch": 0.25948018747337026, "grad_norm": 15.5625, "learning_rate": 2.4033163024418543e-06, "loss": 1.0348, "step": 3045 }, { "epoch": 0.2595654026416702, "grad_norm": 13.8125, "learning_rate": 2.4032514922829337e-06, "loss": 0.8978, "step": 3046 }, { "epoch": 0.2596506178099702, "grad_norm": 14.5625, "learning_rate": 2.4031866612835993e-06, "loss": 0.1999, "step": 3047 }, { "epoch": 0.25973583297827013, "grad_norm": 11.125, "learning_rate": 2.4031218094450215e-06, "loss": 0.4699, "step": 3048 }, { "epoch": 0.2598210481465701, "grad_norm": 20.125, "learning_rate": 2.4030569367683736e-06, "loss": 0.6024, "step": 3049 }, { "epoch": 0.25990626331487005, "grad_norm": 33.75, "learning_rate": 2.4029920432548275e-06, "loss": 1.1232, "step": 3050 }, { "epoch": 0.25999147848317, "grad_norm": 14.1875, "learning_rate": 2.4029271289055558e-06, "loss": 0.6519, "step": 3051 }, { "epoch": 0.26007669365146996, "grad_norm": 14.25, "learning_rate": 2.4028621937217312e-06, "loss": 0.5323, "step": 3052 }, { "epoch": 0.2601619088197699, "grad_norm": 12.8125, "learning_rate": 2.4027972377045274e-06, "loss": 0.5999, "step": 3053 }, { "epoch": 0.2602471239880699, "grad_norm": 16.125, "learning_rate": 2.4027322608551183e-06, "loss": 0.6853, "step": 3054 }, { "epoch": 0.26033233915636983, "grad_norm": 14.5625, "learning_rate": 2.4026672631746783e-06, "loss": 0.8008, "step": 3055 }, { "epoch": 0.2604175543246698, "grad_norm": 17.375, "learning_rate": 2.402602244664382e-06, "loss": 0.6959, "step": 3056 }, { "epoch": 0.26050276949296974, "grad_norm": 14.5625, "learning_rate": 2.402537205325404e-06, "loss": 0.8843, "step": 3057 }, { "epoch": 0.2605879846612697, "grad_norm": 11.8125, "learning_rate": 2.4024721451589194e-06, "loss": 0.4014, "step": 3058 }, { "epoch": 0.26067319982956966, "grad_norm": 15.25, "learning_rate": 2.4024070641661043e-06, "loss": 0.9268, "step": 3059 }, { "epoch": 0.2607584149978696, "grad_norm": 13.125, "learning_rate": 2.4023419623481346e-06, "loss": 0.3544, "step": 3060 }, { "epoch": 0.26084363016616957, "grad_norm": 14.6875, "learning_rate": 2.4022768397061873e-06, "loss": 0.7147, "step": 3061 }, { "epoch": 0.26092884533446953, "grad_norm": 17.875, "learning_rate": 2.4022116962414384e-06, "loss": 0.7709, "step": 3062 }, { "epoch": 0.2610140605027695, "grad_norm": 13.375, "learning_rate": 2.402146531955066e-06, "loss": 0.3639, "step": 3063 }, { "epoch": 0.26109927567106944, "grad_norm": 17.5, "learning_rate": 2.402081346848247e-06, "loss": 0.7288, "step": 3064 }, { "epoch": 0.2611844908393694, "grad_norm": 26.5, "learning_rate": 2.4020161409221594e-06, "loss": 1.3285, "step": 3065 }, { "epoch": 0.26126970600766936, "grad_norm": 20.75, "learning_rate": 2.4019509141779816e-06, "loss": 1.0347, "step": 3066 }, { "epoch": 0.2613549211759693, "grad_norm": 13.8125, "learning_rate": 2.4018856666168925e-06, "loss": 0.8393, "step": 3067 }, { "epoch": 0.26144013634426927, "grad_norm": 16.125, "learning_rate": 2.4018203982400713e-06, "loss": 0.7065, "step": 3068 }, { "epoch": 0.2615253515125692, "grad_norm": 14.0625, "learning_rate": 2.401755109048697e-06, "loss": 0.6089, "step": 3069 }, { "epoch": 0.2616105666808692, "grad_norm": 14.4375, "learning_rate": 2.4016897990439498e-06, "loss": 0.587, "step": 3070 }, { "epoch": 0.26169578184916914, "grad_norm": 16.25, "learning_rate": 2.4016244682270097e-06, "loss": 0.8635, "step": 3071 }, { "epoch": 0.2617809970174691, "grad_norm": 30.0, "learning_rate": 2.4015591165990576e-06, "loss": 0.9933, "step": 3072 }, { "epoch": 0.26186621218576905, "grad_norm": 12.6875, "learning_rate": 2.401493744161274e-06, "loss": 0.4795, "step": 3073 }, { "epoch": 0.261951427354069, "grad_norm": 13.5, "learning_rate": 2.4014283509148406e-06, "loss": 0.5747, "step": 3074 }, { "epoch": 0.26203664252236897, "grad_norm": 15.75, "learning_rate": 2.4013629368609393e-06, "loss": 0.5161, "step": 3075 }, { "epoch": 0.2621218576906689, "grad_norm": 16.375, "learning_rate": 2.4012975020007516e-06, "loss": 0.6793, "step": 3076 }, { "epoch": 0.2622070728589689, "grad_norm": 12.0, "learning_rate": 2.4012320463354606e-06, "loss": 0.4181, "step": 3077 }, { "epoch": 0.26229228802726884, "grad_norm": 18.0, "learning_rate": 2.401166569866249e-06, "loss": 0.6657, "step": 3078 }, { "epoch": 0.2623775031955688, "grad_norm": 26.25, "learning_rate": 2.4011010725942998e-06, "loss": 1.0912, "step": 3079 }, { "epoch": 0.26246271836386875, "grad_norm": 19.25, "learning_rate": 2.4010355545207965e-06, "loss": 0.8171, "step": 3080 }, { "epoch": 0.2625479335321687, "grad_norm": 15.0625, "learning_rate": 2.4009700156469233e-06, "loss": 0.6955, "step": 3081 }, { "epoch": 0.26263314870046867, "grad_norm": 15.5, "learning_rate": 2.4009044559738646e-06, "loss": 0.688, "step": 3082 }, { "epoch": 0.2627183638687686, "grad_norm": 13.0, "learning_rate": 2.400838875502805e-06, "loss": 0.6327, "step": 3083 }, { "epoch": 0.2628035790370686, "grad_norm": 16.375, "learning_rate": 2.40077327423493e-06, "loss": 1.2231, "step": 3084 }, { "epoch": 0.26288879420536854, "grad_norm": 16.75, "learning_rate": 2.400707652171424e-06, "loss": 1.0987, "step": 3085 }, { "epoch": 0.2629740093736685, "grad_norm": 15.6875, "learning_rate": 2.4006420093134744e-06, "loss": 0.5817, "step": 3086 }, { "epoch": 0.26305922454196845, "grad_norm": 17.625, "learning_rate": 2.4005763456622663e-06, "loss": 0.8865, "step": 3087 }, { "epoch": 0.2631444397102684, "grad_norm": 12.75, "learning_rate": 2.4005106612189866e-06, "loss": 0.5989, "step": 3088 }, { "epoch": 0.26322965487856836, "grad_norm": 16.5, "learning_rate": 2.4004449559848224e-06, "loss": 1.0143, "step": 3089 }, { "epoch": 0.2633148700468683, "grad_norm": 11.9375, "learning_rate": 2.400379229960961e-06, "loss": 0.4729, "step": 3090 }, { "epoch": 0.2634000852151683, "grad_norm": 17.0, "learning_rate": 2.40031348314859e-06, "loss": 0.7569, "step": 3091 }, { "epoch": 0.26348530038346824, "grad_norm": 24.75, "learning_rate": 2.4002477155488976e-06, "loss": 0.9389, "step": 3092 }, { "epoch": 0.2635705155517682, "grad_norm": 13.0625, "learning_rate": 2.4001819271630726e-06, "loss": 0.7724, "step": 3093 }, { "epoch": 0.26365573072006815, "grad_norm": 12.125, "learning_rate": 2.4001161179923034e-06, "loss": 0.6507, "step": 3094 }, { "epoch": 0.2637409458883681, "grad_norm": 11.3125, "learning_rate": 2.400050288037779e-06, "loss": 0.506, "step": 3095 }, { "epoch": 0.26382616105666806, "grad_norm": 13.6875, "learning_rate": 2.3999844373006904e-06, "loss": 0.5859, "step": 3096 }, { "epoch": 0.263911376224968, "grad_norm": 12.9375, "learning_rate": 2.3999185657822256e-06, "loss": 0.6689, "step": 3097 }, { "epoch": 0.263996591393268, "grad_norm": 12.25, "learning_rate": 2.3998526734835765e-06, "loss": 0.7123, "step": 3098 }, { "epoch": 0.26408180656156793, "grad_norm": 16.875, "learning_rate": 2.3997867604059335e-06, "loss": 0.7335, "step": 3099 }, { "epoch": 0.2641670217298679, "grad_norm": 17.25, "learning_rate": 2.399720826550487e-06, "loss": 0.9156, "step": 3100 }, { "epoch": 0.26425223689816785, "grad_norm": 17.75, "learning_rate": 2.3996548719184293e-06, "loss": 0.7847, "step": 3101 }, { "epoch": 0.2643374520664678, "grad_norm": 14.4375, "learning_rate": 2.399588896510952e-06, "loss": 0.7428, "step": 3102 }, { "epoch": 0.26442266723476776, "grad_norm": 12.1875, "learning_rate": 2.3995229003292473e-06, "loss": 0.6125, "step": 3103 }, { "epoch": 0.2645078824030678, "grad_norm": 12.75, "learning_rate": 2.3994568833745085e-06, "loss": 0.5944, "step": 3104 }, { "epoch": 0.26459309757136773, "grad_norm": 15.25, "learning_rate": 2.399390845647927e-06, "loss": 0.7549, "step": 3105 }, { "epoch": 0.2646783127396677, "grad_norm": 36.25, "learning_rate": 2.3993247871506976e-06, "loss": 1.4105, "step": 3106 }, { "epoch": 0.26476352790796764, "grad_norm": 13.0, "learning_rate": 2.3992587078840134e-06, "loss": 0.7736, "step": 3107 }, { "epoch": 0.2648487430762676, "grad_norm": 17.625, "learning_rate": 2.3991926078490686e-06, "loss": 0.9964, "step": 3108 }, { "epoch": 0.26493395824456756, "grad_norm": 11.3125, "learning_rate": 2.3991264870470583e-06, "loss": 0.5037, "step": 3109 }, { "epoch": 0.2650191734128675, "grad_norm": 15.375, "learning_rate": 2.399060345479176e-06, "loss": 0.4565, "step": 3110 }, { "epoch": 0.26510438858116747, "grad_norm": 10.3125, "learning_rate": 2.3989941831466184e-06, "loss": 0.4318, "step": 3111 }, { "epoch": 0.26518960374946743, "grad_norm": 10.5625, "learning_rate": 2.3989280000505805e-06, "loss": 0.3038, "step": 3112 }, { "epoch": 0.2652748189177674, "grad_norm": 11.8125, "learning_rate": 2.3988617961922578e-06, "loss": 0.5742, "step": 3113 }, { "epoch": 0.26536003408606734, "grad_norm": 14.0, "learning_rate": 2.398795571572848e-06, "loss": 0.8746, "step": 3114 }, { "epoch": 0.2654452492543673, "grad_norm": 17.125, "learning_rate": 2.398729326193546e-06, "loss": 1.365, "step": 3115 }, { "epoch": 0.26553046442266726, "grad_norm": 22.375, "learning_rate": 2.398663060055551e-06, "loss": 1.0579, "step": 3116 }, { "epoch": 0.2656156795909672, "grad_norm": 32.0, "learning_rate": 2.3985967731600586e-06, "loss": 0.9658, "step": 3117 }, { "epoch": 0.26570089475926717, "grad_norm": 21.5, "learning_rate": 2.398530465508268e-06, "loss": 0.747, "step": 3118 }, { "epoch": 0.2657861099275671, "grad_norm": 17.625, "learning_rate": 2.3984641371013765e-06, "loss": 0.9848, "step": 3119 }, { "epoch": 0.2658713250958671, "grad_norm": 20.875, "learning_rate": 2.3983977879405834e-06, "loss": 0.879, "step": 3120 }, { "epoch": 0.26595654026416704, "grad_norm": 13.0, "learning_rate": 2.398331418027088e-06, "loss": 0.375, "step": 3121 }, { "epoch": 0.266041755432467, "grad_norm": 11.3125, "learning_rate": 2.3982650273620882e-06, "loss": 0.5694, "step": 3122 }, { "epoch": 0.26612697060076695, "grad_norm": 14.3125, "learning_rate": 2.398198615946785e-06, "loss": 0.6548, "step": 3123 }, { "epoch": 0.2662121857690669, "grad_norm": 16.25, "learning_rate": 2.3981321837823784e-06, "loss": 0.7613, "step": 3124 }, { "epoch": 0.26629740093736687, "grad_norm": 22.625, "learning_rate": 2.398065730870069e-06, "loss": 0.8212, "step": 3125 }, { "epoch": 0.2663826161056668, "grad_norm": 21.75, "learning_rate": 2.397999257211057e-06, "loss": 0.922, "step": 3126 }, { "epoch": 0.2664678312739668, "grad_norm": 12.1875, "learning_rate": 2.397932762806544e-06, "loss": 0.5572, "step": 3127 }, { "epoch": 0.26655304644226674, "grad_norm": 24.75, "learning_rate": 2.3978662476577313e-06, "loss": 0.9428, "step": 3128 }, { "epoch": 0.2666382616105667, "grad_norm": 15.5625, "learning_rate": 2.397799711765822e-06, "loss": 0.9027, "step": 3129 }, { "epoch": 0.26672347677886665, "grad_norm": 14.8125, "learning_rate": 2.397733155132017e-06, "loss": 0.6059, "step": 3130 }, { "epoch": 0.2668086919471666, "grad_norm": 13.8125, "learning_rate": 2.3976665777575203e-06, "loss": 0.9804, "step": 3131 }, { "epoch": 0.26689390711546657, "grad_norm": 10.75, "learning_rate": 2.3975999796435343e-06, "loss": 0.3274, "step": 3132 }, { "epoch": 0.2669791222837665, "grad_norm": 12.625, "learning_rate": 2.397533360791263e-06, "loss": 0.5653, "step": 3133 }, { "epoch": 0.2670643374520665, "grad_norm": 15.8125, "learning_rate": 2.397466721201909e-06, "loss": 0.5945, "step": 3134 }, { "epoch": 0.26714955262036644, "grad_norm": 10.5, "learning_rate": 2.3974000608766777e-06, "loss": 0.5088, "step": 3135 }, { "epoch": 0.2672347677886664, "grad_norm": 12.8125, "learning_rate": 2.3973333798167744e-06, "loss": 0.5068, "step": 3136 }, { "epoch": 0.26731998295696635, "grad_norm": 14.625, "learning_rate": 2.3972666780234026e-06, "loss": 0.905, "step": 3137 }, { "epoch": 0.2674051981252663, "grad_norm": 13.125, "learning_rate": 2.3971999554977683e-06, "loss": 0.713, "step": 3138 }, { "epoch": 0.26749041329356626, "grad_norm": 11.4375, "learning_rate": 2.397133212241077e-06, "loss": 0.3716, "step": 3139 }, { "epoch": 0.2675756284618662, "grad_norm": 21.375, "learning_rate": 2.3970664482545352e-06, "loss": 0.9597, "step": 3140 }, { "epoch": 0.2676608436301662, "grad_norm": 11.25, "learning_rate": 2.3969996635393493e-06, "loss": 0.4993, "step": 3141 }, { "epoch": 0.26774605879846614, "grad_norm": 13.125, "learning_rate": 2.396932858096726e-06, "loss": 0.7593, "step": 3142 }, { "epoch": 0.2678312739667661, "grad_norm": 15.875, "learning_rate": 2.3968660319278725e-06, "loss": 0.7841, "step": 3143 }, { "epoch": 0.26791648913506605, "grad_norm": 16.25, "learning_rate": 2.3967991850339965e-06, "loss": 0.7095, "step": 3144 }, { "epoch": 0.268001704303366, "grad_norm": 17.625, "learning_rate": 2.3967323174163066e-06, "loss": 1.0058, "step": 3145 }, { "epoch": 0.26808691947166596, "grad_norm": 10.8125, "learning_rate": 2.39666542907601e-06, "loss": 0.3677, "step": 3146 }, { "epoch": 0.2681721346399659, "grad_norm": 13.5, "learning_rate": 2.3965985200143163e-06, "loss": 0.4982, "step": 3147 }, { "epoch": 0.2682573498082659, "grad_norm": 22.75, "learning_rate": 2.396531590232434e-06, "loss": 0.8559, "step": 3148 }, { "epoch": 0.26834256497656583, "grad_norm": 17.375, "learning_rate": 2.3964646397315735e-06, "loss": 1.0388, "step": 3149 }, { "epoch": 0.2684277801448658, "grad_norm": 13.0625, "learning_rate": 2.3963976685129437e-06, "loss": 0.6354, "step": 3150 }, { "epoch": 0.26851299531316575, "grad_norm": 13.625, "learning_rate": 2.3963306765777554e-06, "loss": 0.4669, "step": 3151 }, { "epoch": 0.2685982104814657, "grad_norm": 15.3125, "learning_rate": 2.3962636639272193e-06, "loss": 0.9461, "step": 3152 }, { "epoch": 0.26868342564976566, "grad_norm": 13.9375, "learning_rate": 2.396196630562546e-06, "loss": 0.6262, "step": 3153 }, { "epoch": 0.2687686408180656, "grad_norm": 10.9375, "learning_rate": 2.3961295764849462e-06, "loss": 0.374, "step": 3154 }, { "epoch": 0.2688538559863656, "grad_norm": 14.5, "learning_rate": 2.3960625016956335e-06, "loss": 0.9857, "step": 3155 }, { "epoch": 0.26893907115466553, "grad_norm": 11.375, "learning_rate": 2.3959954061958184e-06, "loss": 0.5476, "step": 3156 }, { "epoch": 0.2690242863229655, "grad_norm": 11.75, "learning_rate": 2.395928289986714e-06, "loss": 0.741, "step": 3157 }, { "epoch": 0.26910950149126545, "grad_norm": 13.8125, "learning_rate": 2.395861153069533e-06, "loss": 0.7711, "step": 3158 }, { "epoch": 0.2691947166595654, "grad_norm": 11.375, "learning_rate": 2.395793995445489e-06, "loss": 0.6454, "step": 3159 }, { "epoch": 0.26927993182786536, "grad_norm": 13.5625, "learning_rate": 2.395726817115795e-06, "loss": 0.6697, "step": 3160 }, { "epoch": 0.2693651469961653, "grad_norm": 15.3125, "learning_rate": 2.3956596180816655e-06, "loss": 0.7545, "step": 3161 }, { "epoch": 0.2694503621644653, "grad_norm": 12.625, "learning_rate": 2.395592398344315e-06, "loss": 0.6582, "step": 3162 }, { "epoch": 0.26953557733276523, "grad_norm": 12.8125, "learning_rate": 2.395525157904957e-06, "loss": 0.6127, "step": 3163 }, { "epoch": 0.2696207925010652, "grad_norm": 16.75, "learning_rate": 2.395457896764808e-06, "loss": 0.9324, "step": 3164 }, { "epoch": 0.26970600766936514, "grad_norm": 12.1875, "learning_rate": 2.395390614925083e-06, "loss": 0.546, "step": 3165 }, { "epoch": 0.2697912228376651, "grad_norm": 12.125, "learning_rate": 2.3953233123869975e-06, "loss": 0.648, "step": 3166 }, { "epoch": 0.26987643800596506, "grad_norm": 10.6875, "learning_rate": 2.3952559891517685e-06, "loss": 0.4571, "step": 3167 }, { "epoch": 0.269961653174265, "grad_norm": 11.9375, "learning_rate": 2.395188645220612e-06, "loss": 0.3664, "step": 3168 }, { "epoch": 0.27004686834256497, "grad_norm": 22.125, "learning_rate": 2.395121280594745e-06, "loss": 1.4256, "step": 3169 }, { "epoch": 0.27013208351086493, "grad_norm": 24.125, "learning_rate": 2.3950538952753847e-06, "loss": 1.0846, "step": 3170 }, { "epoch": 0.2702172986791649, "grad_norm": 17.125, "learning_rate": 2.394986489263749e-06, "loss": 1.0623, "step": 3171 }, { "epoch": 0.27030251384746484, "grad_norm": 13.9375, "learning_rate": 2.394919062561057e-06, "loss": 0.7516, "step": 3172 }, { "epoch": 0.2703877290157648, "grad_norm": 14.4375, "learning_rate": 2.3948516151685253e-06, "loss": 0.8134, "step": 3173 }, { "epoch": 0.27047294418406476, "grad_norm": 16.0, "learning_rate": 2.394784147087374e-06, "loss": 0.8237, "step": 3174 }, { "epoch": 0.2705581593523647, "grad_norm": 13.375, "learning_rate": 2.394716658318822e-06, "loss": 0.8622, "step": 3175 }, { "epoch": 0.27064337452066467, "grad_norm": 12.0625, "learning_rate": 2.3946491488640887e-06, "loss": 0.5916, "step": 3176 }, { "epoch": 0.2707285896889646, "grad_norm": 14.625, "learning_rate": 2.394581618724395e-06, "loss": 0.5921, "step": 3177 }, { "epoch": 0.2708138048572646, "grad_norm": 17.625, "learning_rate": 2.3945140679009593e-06, "loss": 0.7363, "step": 3178 }, { "epoch": 0.27089902002556454, "grad_norm": 19.375, "learning_rate": 2.394446496395004e-06, "loss": 0.9161, "step": 3179 }, { "epoch": 0.2709842351938645, "grad_norm": 32.0, "learning_rate": 2.39437890420775e-06, "loss": 1.0106, "step": 3180 }, { "epoch": 0.27106945036216445, "grad_norm": 21.25, "learning_rate": 2.394311291340418e-06, "loss": 0.6123, "step": 3181 }, { "epoch": 0.2711546655304644, "grad_norm": 14.9375, "learning_rate": 2.3942436577942308e-06, "loss": 0.6568, "step": 3182 }, { "epoch": 0.27123988069876437, "grad_norm": 20.375, "learning_rate": 2.39417600357041e-06, "loss": 0.804, "step": 3183 }, { "epoch": 0.2713250958670643, "grad_norm": 12.6875, "learning_rate": 2.3941083286701778e-06, "loss": 0.4972, "step": 3184 }, { "epoch": 0.2714103110353643, "grad_norm": 20.625, "learning_rate": 2.3940406330947578e-06, "loss": 0.8979, "step": 3185 }, { "epoch": 0.27149552620366424, "grad_norm": 8.8125, "learning_rate": 2.393972916845373e-06, "loss": 0.5903, "step": 3186 }, { "epoch": 0.2715807413719642, "grad_norm": 11.0625, "learning_rate": 2.3939051799232476e-06, "loss": 0.4858, "step": 3187 }, { "epoch": 0.27166595654026415, "grad_norm": 22.625, "learning_rate": 2.393837422329605e-06, "loss": 0.8938, "step": 3188 }, { "epoch": 0.2717511717085641, "grad_norm": 31.0, "learning_rate": 2.3937696440656703e-06, "loss": 1.5027, "step": 3189 }, { "epoch": 0.27183638687686407, "grad_norm": 11.0625, "learning_rate": 2.3937018451326677e-06, "loss": 0.4057, "step": 3190 }, { "epoch": 0.271921602045164, "grad_norm": 15.6875, "learning_rate": 2.3936340255318223e-06, "loss": 0.8049, "step": 3191 }, { "epoch": 0.272006817213464, "grad_norm": 10.75, "learning_rate": 2.3935661852643606e-06, "loss": 0.4454, "step": 3192 }, { "epoch": 0.27209203238176394, "grad_norm": 15.3125, "learning_rate": 2.393498324331508e-06, "loss": 0.6401, "step": 3193 }, { "epoch": 0.2721772475500639, "grad_norm": 23.0, "learning_rate": 2.3934304427344906e-06, "loss": 1.0184, "step": 3194 }, { "epoch": 0.27226246271836385, "grad_norm": 10.875, "learning_rate": 2.3933625404745357e-06, "loss": 0.4099, "step": 3195 }, { "epoch": 0.2723476778866638, "grad_norm": 20.875, "learning_rate": 2.3932946175528693e-06, "loss": 1.134, "step": 3196 }, { "epoch": 0.27243289305496377, "grad_norm": 29.125, "learning_rate": 2.39322667397072e-06, "loss": 0.8978, "step": 3197 }, { "epoch": 0.2725181082232637, "grad_norm": 15.4375, "learning_rate": 2.393158709729315e-06, "loss": 1.0133, "step": 3198 }, { "epoch": 0.2726033233915637, "grad_norm": 18.75, "learning_rate": 2.3930907248298825e-06, "loss": 1.0383, "step": 3199 }, { "epoch": 0.27268853855986364, "grad_norm": 14.625, "learning_rate": 2.393022719273651e-06, "loss": 0.4778, "step": 3200 }, { "epoch": 0.2727737537281636, "grad_norm": 16.25, "learning_rate": 2.3929546930618496e-06, "loss": 0.9558, "step": 3201 }, { "epoch": 0.27285896889646355, "grad_norm": 17.75, "learning_rate": 2.392886646195708e-06, "loss": 0.7627, "step": 3202 }, { "epoch": 0.2729441840647635, "grad_norm": 19.375, "learning_rate": 2.392818578676455e-06, "loss": 1.3218, "step": 3203 }, { "epoch": 0.27302939923306346, "grad_norm": 13.5, "learning_rate": 2.392750490505321e-06, "loss": 0.7897, "step": 3204 }, { "epoch": 0.2731146144013634, "grad_norm": 13.625, "learning_rate": 2.392682381683537e-06, "loss": 0.7711, "step": 3205 }, { "epoch": 0.2731998295696634, "grad_norm": 13.125, "learning_rate": 2.3926142522123327e-06, "loss": 0.6208, "step": 3206 }, { "epoch": 0.27328504473796333, "grad_norm": 19.375, "learning_rate": 2.3925461020929404e-06, "loss": 0.5715, "step": 3207 }, { "epoch": 0.2733702599062633, "grad_norm": 13.5625, "learning_rate": 2.3924779313265906e-06, "loss": 0.5413, "step": 3208 }, { "epoch": 0.27345547507456325, "grad_norm": 15.1875, "learning_rate": 2.3924097399145165e-06, "loss": 0.5909, "step": 3209 }, { "epoch": 0.2735406902428632, "grad_norm": 14.5, "learning_rate": 2.3923415278579486e-06, "loss": 0.4819, "step": 3210 }, { "epoch": 0.27362590541116316, "grad_norm": 18.875, "learning_rate": 2.3922732951581213e-06, "loss": 1.0415, "step": 3211 }, { "epoch": 0.2737111205794631, "grad_norm": 13.75, "learning_rate": 2.3922050418162667e-06, "loss": 0.741, "step": 3212 }, { "epoch": 0.2737963357477631, "grad_norm": 10.875, "learning_rate": 2.392136767833618e-06, "loss": 0.381, "step": 3213 }, { "epoch": 0.27388155091606303, "grad_norm": 14.6875, "learning_rate": 2.3920684732114103e-06, "loss": 0.9179, "step": 3214 }, { "epoch": 0.273966766084363, "grad_norm": 13.125, "learning_rate": 2.3920001579508764e-06, "loss": 0.5845, "step": 3215 }, { "epoch": 0.274051981252663, "grad_norm": 14.0625, "learning_rate": 2.391931822053251e-06, "loss": 0.6268, "step": 3216 }, { "epoch": 0.27413719642096296, "grad_norm": 21.375, "learning_rate": 2.3918634655197698e-06, "loss": 1.2118, "step": 3217 }, { "epoch": 0.2742224115892629, "grad_norm": 14.5625, "learning_rate": 2.3917950883516674e-06, "loss": 0.8733, "step": 3218 }, { "epoch": 0.2743076267575629, "grad_norm": 27.5, "learning_rate": 2.3917266905501796e-06, "loss": 1.2663, "step": 3219 }, { "epoch": 0.27439284192586283, "grad_norm": 13.75, "learning_rate": 2.391658272116542e-06, "loss": 0.7653, "step": 3220 }, { "epoch": 0.2744780570941628, "grad_norm": 15.0625, "learning_rate": 2.3915898330519917e-06, "loss": 0.7122, "step": 3221 }, { "epoch": 0.27456327226246274, "grad_norm": 15.6875, "learning_rate": 2.3915213733577654e-06, "loss": 0.9794, "step": 3222 }, { "epoch": 0.2746484874307627, "grad_norm": 13.6875, "learning_rate": 2.3914528930350998e-06, "loss": 0.7808, "step": 3223 }, { "epoch": 0.27473370259906266, "grad_norm": 19.5, "learning_rate": 2.3913843920852325e-06, "loss": 0.9757, "step": 3224 }, { "epoch": 0.2748189177673626, "grad_norm": 18.125, "learning_rate": 2.3913158705094014e-06, "loss": 0.7532, "step": 3225 }, { "epoch": 0.27490413293566257, "grad_norm": 9.9375, "learning_rate": 2.391247328308845e-06, "loss": 0.4005, "step": 3226 }, { "epoch": 0.2749893481039625, "grad_norm": 22.375, "learning_rate": 2.391178765484802e-06, "loss": 1.2088, "step": 3227 }, { "epoch": 0.2750745632722625, "grad_norm": 21.125, "learning_rate": 2.391110182038511e-06, "loss": 0.7223, "step": 3228 }, { "epoch": 0.27515977844056244, "grad_norm": 16.375, "learning_rate": 2.391041577971211e-06, "loss": 1.0136, "step": 3229 }, { "epoch": 0.2752449936088624, "grad_norm": 15.75, "learning_rate": 2.390972953284143e-06, "loss": 0.9312, "step": 3230 }, { "epoch": 0.27533020877716236, "grad_norm": 15.5, "learning_rate": 2.390904307978546e-06, "loss": 1.1171, "step": 3231 }, { "epoch": 0.2754154239454623, "grad_norm": 15.9375, "learning_rate": 2.390835642055661e-06, "loss": 0.7454, "step": 3232 }, { "epoch": 0.27550063911376227, "grad_norm": 16.875, "learning_rate": 2.3907669555167288e-06, "loss": 0.8645, "step": 3233 }, { "epoch": 0.2755858542820622, "grad_norm": 13.8125, "learning_rate": 2.3906982483629903e-06, "loss": 0.5586, "step": 3234 }, { "epoch": 0.2756710694503622, "grad_norm": 13.875, "learning_rate": 2.3906295205956877e-06, "loss": 1.0113, "step": 3235 }, { "epoch": 0.27575628461866214, "grad_norm": 12.25, "learning_rate": 2.3905607722160625e-06, "loss": 0.6721, "step": 3236 }, { "epoch": 0.2758414997869621, "grad_norm": 17.125, "learning_rate": 2.3904920032253575e-06, "loss": 1.0627, "step": 3237 }, { "epoch": 0.27592671495526205, "grad_norm": 13.5625, "learning_rate": 2.390423213624815e-06, "loss": 0.7827, "step": 3238 }, { "epoch": 0.276011930123562, "grad_norm": 10.0, "learning_rate": 2.390354403415678e-06, "loss": 0.3368, "step": 3239 }, { "epoch": 0.27609714529186197, "grad_norm": 15.125, "learning_rate": 2.3902855725991907e-06, "loss": 0.5098, "step": 3240 }, { "epoch": 0.2761823604601619, "grad_norm": 14.375, "learning_rate": 2.3902167211765958e-06, "loss": 0.649, "step": 3241 }, { "epoch": 0.2762675756284619, "grad_norm": 15.625, "learning_rate": 2.390147849149139e-06, "loss": 0.6001, "step": 3242 }, { "epoch": 0.27635279079676184, "grad_norm": 14.375, "learning_rate": 2.3900789565180632e-06, "loss": 0.8388, "step": 3243 }, { "epoch": 0.2764380059650618, "grad_norm": 13.1875, "learning_rate": 2.390010043284615e-06, "loss": 0.587, "step": 3244 }, { "epoch": 0.27652322113336175, "grad_norm": 15.3125, "learning_rate": 2.3899411094500385e-06, "loss": 0.5722, "step": 3245 }, { "epoch": 0.2766084363016617, "grad_norm": 15.6875, "learning_rate": 2.38987215501558e-06, "loss": 0.9183, "step": 3246 }, { "epoch": 0.27669365146996167, "grad_norm": 13.0, "learning_rate": 2.389803179982485e-06, "loss": 0.6159, "step": 3247 }, { "epoch": 0.2767788666382616, "grad_norm": 16.5, "learning_rate": 2.389734184352001e-06, "loss": 0.7564, "step": 3248 }, { "epoch": 0.2768640818065616, "grad_norm": 12.9375, "learning_rate": 2.389665168125374e-06, "loss": 0.7234, "step": 3249 }, { "epoch": 0.27694929697486154, "grad_norm": 11.4375, "learning_rate": 2.3895961313038515e-06, "loss": 0.4677, "step": 3250 }, { "epoch": 0.2770345121431615, "grad_norm": 16.625, "learning_rate": 2.3895270738886807e-06, "loss": 1.1617, "step": 3251 }, { "epoch": 0.27711972731146145, "grad_norm": 17.625, "learning_rate": 2.38945799588111e-06, "loss": 0.8164, "step": 3252 }, { "epoch": 0.2772049424797614, "grad_norm": 13.4375, "learning_rate": 2.389388897282387e-06, "loss": 0.5967, "step": 3253 }, { "epoch": 0.27729015764806136, "grad_norm": 14.5625, "learning_rate": 2.389319778093761e-06, "loss": 0.8305, "step": 3254 }, { "epoch": 0.2773753728163613, "grad_norm": 19.375, "learning_rate": 2.3892506383164816e-06, "loss": 0.7634, "step": 3255 }, { "epoch": 0.2774605879846613, "grad_norm": 23.25, "learning_rate": 2.389181477951797e-06, "loss": 1.2268, "step": 3256 }, { "epoch": 0.27754580315296123, "grad_norm": 23.75, "learning_rate": 2.3891122970009577e-06, "loss": 1.534, "step": 3257 }, { "epoch": 0.2776310183212612, "grad_norm": 15.75, "learning_rate": 2.389043095465214e-06, "loss": 0.8984, "step": 3258 }, { "epoch": 0.27771623348956115, "grad_norm": 27.125, "learning_rate": 2.3889738733458155e-06, "loss": 0.8182, "step": 3259 }, { "epoch": 0.2778014486578611, "grad_norm": 14.0, "learning_rate": 2.3889046306440146e-06, "loss": 0.8357, "step": 3260 }, { "epoch": 0.27788666382616106, "grad_norm": 14.1875, "learning_rate": 2.3888353673610613e-06, "loss": 0.8822, "step": 3261 }, { "epoch": 0.277971878994461, "grad_norm": 14.0, "learning_rate": 2.3887660834982076e-06, "loss": 0.6939, "step": 3262 }, { "epoch": 0.278057094162761, "grad_norm": 11.375, "learning_rate": 2.388696779056706e-06, "loss": 0.6559, "step": 3263 }, { "epoch": 0.27814230933106093, "grad_norm": 18.375, "learning_rate": 2.3886274540378083e-06, "loss": 0.5942, "step": 3264 }, { "epoch": 0.2782275244993609, "grad_norm": 13.4375, "learning_rate": 2.388558108442768e-06, "loss": 0.6218, "step": 3265 }, { "epoch": 0.27831273966766085, "grad_norm": 13.125, "learning_rate": 2.388488742272837e-06, "loss": 0.5284, "step": 3266 }, { "epoch": 0.2783979548359608, "grad_norm": 11.25, "learning_rate": 2.38841935552927e-06, "loss": 0.6733, "step": 3267 }, { "epoch": 0.27848317000426076, "grad_norm": 15.5, "learning_rate": 2.3883499482133206e-06, "loss": 0.9241, "step": 3268 }, { "epoch": 0.2785683851725607, "grad_norm": 14.375, "learning_rate": 2.3882805203262425e-06, "loss": 0.7892, "step": 3269 }, { "epoch": 0.2786536003408607, "grad_norm": 13.125, "learning_rate": 2.388211071869291e-06, "loss": 0.7357, "step": 3270 }, { "epoch": 0.27873881550916063, "grad_norm": 15.5, "learning_rate": 2.388141602843721e-06, "loss": 0.8892, "step": 3271 }, { "epoch": 0.2788240306774606, "grad_norm": 12.5, "learning_rate": 2.388072113250788e-06, "loss": 0.6221, "step": 3272 }, { "epoch": 0.27890924584576055, "grad_norm": 12.6875, "learning_rate": 2.388002603091747e-06, "loss": 0.6881, "step": 3273 }, { "epoch": 0.2789944610140605, "grad_norm": 15.4375, "learning_rate": 2.3879330723678547e-06, "loss": 0.9123, "step": 3274 }, { "epoch": 0.27907967618236046, "grad_norm": 13.5625, "learning_rate": 2.3878635210803675e-06, "loss": 0.5552, "step": 3275 }, { "epoch": 0.2791648913506604, "grad_norm": 13.8125, "learning_rate": 2.3877939492305423e-06, "loss": 0.7999, "step": 3276 }, { "epoch": 0.2792501065189604, "grad_norm": 14.25, "learning_rate": 2.3877243568196364e-06, "loss": 0.5214, "step": 3277 }, { "epoch": 0.27933532168726033, "grad_norm": 17.25, "learning_rate": 2.3876547438489073e-06, "loss": 0.9477, "step": 3278 }, { "epoch": 0.2794205368555603, "grad_norm": 11.6875, "learning_rate": 2.387585110319613e-06, "loss": 0.5248, "step": 3279 }, { "epoch": 0.27950575202386024, "grad_norm": 14.75, "learning_rate": 2.3875154562330117e-06, "loss": 0.5613, "step": 3280 }, { "epoch": 0.2795909671921602, "grad_norm": 15.625, "learning_rate": 2.3874457815903622e-06, "loss": 0.664, "step": 3281 }, { "epoch": 0.27967618236046016, "grad_norm": 10.875, "learning_rate": 2.387376086392924e-06, "loss": 0.3871, "step": 3282 }, { "epoch": 0.2797613975287601, "grad_norm": 24.375, "learning_rate": 2.3873063706419557e-06, "loss": 1.2625, "step": 3283 }, { "epoch": 0.27984661269706007, "grad_norm": 16.25, "learning_rate": 2.3872366343387178e-06, "loss": 1.1833, "step": 3284 }, { "epoch": 0.27993182786536003, "grad_norm": 12.3125, "learning_rate": 2.3871668774844705e-06, "loss": 0.769, "step": 3285 }, { "epoch": 0.28001704303366, "grad_norm": 12.0, "learning_rate": 2.3870971000804745e-06, "loss": 0.5318, "step": 3286 }, { "epoch": 0.28010225820195994, "grad_norm": 14.5, "learning_rate": 2.3870273021279896e-06, "loss": 0.8197, "step": 3287 }, { "epoch": 0.2801874733702599, "grad_norm": 15.4375, "learning_rate": 2.3869574836282784e-06, "loss": 0.8311, "step": 3288 }, { "epoch": 0.28027268853855986, "grad_norm": 20.25, "learning_rate": 2.3868876445826024e-06, "loss": 1.1167, "step": 3289 }, { "epoch": 0.2803579037068598, "grad_norm": 15.3125, "learning_rate": 2.386817784992223e-06, "loss": 0.6433, "step": 3290 }, { "epoch": 0.28044311887515977, "grad_norm": 23.25, "learning_rate": 2.3867479048584036e-06, "loss": 1.0603, "step": 3291 }, { "epoch": 0.2805283340434597, "grad_norm": 13.8125, "learning_rate": 2.386678004182406e-06, "loss": 0.3652, "step": 3292 }, { "epoch": 0.2806135492117597, "grad_norm": 13.1875, "learning_rate": 2.3866080829654944e-06, "loss": 0.5017, "step": 3293 }, { "epoch": 0.28069876438005964, "grad_norm": 19.125, "learning_rate": 2.386538141208931e-06, "loss": 0.6801, "step": 3294 }, { "epoch": 0.2807839795483596, "grad_norm": 13.0, "learning_rate": 2.386468178913981e-06, "loss": 0.7981, "step": 3295 }, { "epoch": 0.28086919471665955, "grad_norm": 14.375, "learning_rate": 2.3863981960819085e-06, "loss": 0.7691, "step": 3296 }, { "epoch": 0.2809544098849595, "grad_norm": 13.625, "learning_rate": 2.3863281927139773e-06, "loss": 0.8362, "step": 3297 }, { "epoch": 0.28103962505325947, "grad_norm": 18.5, "learning_rate": 2.3862581688114535e-06, "loss": 0.592, "step": 3298 }, { "epoch": 0.2811248402215594, "grad_norm": 18.875, "learning_rate": 2.3861881243756018e-06, "loss": 0.9038, "step": 3299 }, { "epoch": 0.2812100553898594, "grad_norm": 11.9375, "learning_rate": 2.386118059407688e-06, "loss": 0.5841, "step": 3300 }, { "epoch": 0.28129527055815934, "grad_norm": 15.9375, "learning_rate": 2.386047973908978e-06, "loss": 0.7435, "step": 3301 }, { "epoch": 0.2813804857264593, "grad_norm": 13.0625, "learning_rate": 2.3859778678807393e-06, "loss": 0.7852, "step": 3302 }, { "epoch": 0.28146570089475925, "grad_norm": 16.5, "learning_rate": 2.385907741324238e-06, "loss": 0.7224, "step": 3303 }, { "epoch": 0.2815509160630592, "grad_norm": 16.25, "learning_rate": 2.385837594240742e-06, "loss": 0.8005, "step": 3304 }, { "epoch": 0.28163613123135917, "grad_norm": 13.75, "learning_rate": 2.385767426631518e-06, "loss": 0.884, "step": 3305 }, { "epoch": 0.2817213463996591, "grad_norm": 12.6875, "learning_rate": 2.385697238497835e-06, "loss": 0.5361, "step": 3306 }, { "epoch": 0.2818065615679591, "grad_norm": 13.8125, "learning_rate": 2.3856270298409603e-06, "loss": 0.6809, "step": 3307 }, { "epoch": 0.28189177673625904, "grad_norm": 12.6875, "learning_rate": 2.3855568006621634e-06, "loss": 0.4543, "step": 3308 }, { "epoch": 0.281976991904559, "grad_norm": 20.875, "learning_rate": 2.3854865509627135e-06, "loss": 1.0276, "step": 3309 }, { "epoch": 0.28206220707285895, "grad_norm": 12.4375, "learning_rate": 2.38541628074388e-06, "loss": 0.5687, "step": 3310 }, { "epoch": 0.2821474222411589, "grad_norm": 11.3125, "learning_rate": 2.3853459900069317e-06, "loss": 0.6423, "step": 3311 }, { "epoch": 0.28223263740945886, "grad_norm": 14.9375, "learning_rate": 2.38527567875314e-06, "loss": 0.7432, "step": 3312 }, { "epoch": 0.2823178525777588, "grad_norm": 11.25, "learning_rate": 2.3852053469837756e-06, "loss": 0.6378, "step": 3313 }, { "epoch": 0.2824030677460588, "grad_norm": 25.625, "learning_rate": 2.3851349947001087e-06, "loss": 1.0169, "step": 3314 }, { "epoch": 0.28248828291435873, "grad_norm": 11.0625, "learning_rate": 2.3850646219034107e-06, "loss": 0.5053, "step": 3315 }, { "epoch": 0.2825734980826587, "grad_norm": 21.5, "learning_rate": 2.384994228594954e-06, "loss": 0.5648, "step": 3316 }, { "epoch": 0.28265871325095865, "grad_norm": 23.5, "learning_rate": 2.38492381477601e-06, "loss": 1.022, "step": 3317 }, { "epoch": 0.2827439284192586, "grad_norm": 13.875, "learning_rate": 2.384853380447852e-06, "loss": 0.677, "step": 3318 }, { "epoch": 0.28282914358755856, "grad_norm": 11.8125, "learning_rate": 2.3847829256117516e-06, "loss": 0.489, "step": 3319 }, { "epoch": 0.2829143587558585, "grad_norm": 13.625, "learning_rate": 2.384712450268983e-06, "loss": 0.7879, "step": 3320 }, { "epoch": 0.2829995739241585, "grad_norm": 18.25, "learning_rate": 2.384641954420819e-06, "loss": 1.1059, "step": 3321 }, { "epoch": 0.28308478909245843, "grad_norm": 11.5625, "learning_rate": 2.3845714380685342e-06, "loss": 0.2821, "step": 3322 }, { "epoch": 0.2831700042607584, "grad_norm": 14.0625, "learning_rate": 2.3845009012134027e-06, "loss": 0.4754, "step": 3323 }, { "epoch": 0.28325521942905835, "grad_norm": 32.5, "learning_rate": 2.384430343856699e-06, "loss": 1.4646, "step": 3324 }, { "epoch": 0.2833404345973583, "grad_norm": 13.25, "learning_rate": 2.384359765999698e-06, "loss": 0.7753, "step": 3325 }, { "epoch": 0.28342564976565826, "grad_norm": 17.5, "learning_rate": 2.3842891676436756e-06, "loss": 0.2746, "step": 3326 }, { "epoch": 0.2835108649339582, "grad_norm": 11.625, "learning_rate": 2.3842185487899073e-06, "loss": 0.4905, "step": 3327 }, { "epoch": 0.2835960801022582, "grad_norm": 15.375, "learning_rate": 2.384147909439669e-06, "loss": 0.6193, "step": 3328 }, { "epoch": 0.2836812952705582, "grad_norm": 12.8125, "learning_rate": 2.3840772495942378e-06, "loss": 0.7505, "step": 3329 }, { "epoch": 0.28376651043885814, "grad_norm": 20.375, "learning_rate": 2.3840065692548904e-06, "loss": 0.8517, "step": 3330 }, { "epoch": 0.2838517256071581, "grad_norm": 22.5, "learning_rate": 2.3839358684229035e-06, "loss": 0.904, "step": 3331 }, { "epoch": 0.28393694077545806, "grad_norm": 14.0, "learning_rate": 2.3838651470995556e-06, "loss": 0.5554, "step": 3332 }, { "epoch": 0.284022155943758, "grad_norm": 14.625, "learning_rate": 2.383794405286124e-06, "loss": 0.6426, "step": 3333 }, { "epoch": 0.28410737111205797, "grad_norm": 27.125, "learning_rate": 2.3837236429838878e-06, "loss": 0.8508, "step": 3334 }, { "epoch": 0.28419258628035793, "grad_norm": 23.5, "learning_rate": 2.383652860194125e-06, "loss": 0.3894, "step": 3335 }, { "epoch": 0.2842778014486579, "grad_norm": 13.0, "learning_rate": 2.383582056918115e-06, "loss": 0.5137, "step": 3336 }, { "epoch": 0.28436301661695784, "grad_norm": 16.5, "learning_rate": 2.3835112331571374e-06, "loss": 0.7774, "step": 3337 }, { "epoch": 0.2844482317852578, "grad_norm": 23.625, "learning_rate": 2.383440388912472e-06, "loss": 1.0577, "step": 3338 }, { "epoch": 0.28453344695355776, "grad_norm": 22.0, "learning_rate": 2.3833695241853993e-06, "loss": 1.1389, "step": 3339 }, { "epoch": 0.2846186621218577, "grad_norm": 14.875, "learning_rate": 2.383298638977199e-06, "loss": 1.037, "step": 3340 }, { "epoch": 0.28470387729015767, "grad_norm": 14.4375, "learning_rate": 2.3832277332891534e-06, "loss": 0.8874, "step": 3341 }, { "epoch": 0.2847890924584576, "grad_norm": 13.5, "learning_rate": 2.3831568071225424e-06, "loss": 0.7628, "step": 3342 }, { "epoch": 0.2848743076267576, "grad_norm": 27.25, "learning_rate": 2.383085860478649e-06, "loss": 0.9163, "step": 3343 }, { "epoch": 0.28495952279505754, "grad_norm": 13.1875, "learning_rate": 2.3830148933587545e-06, "loss": 0.469, "step": 3344 }, { "epoch": 0.2850447379633575, "grad_norm": 13.875, "learning_rate": 2.3829439057641415e-06, "loss": 0.4983, "step": 3345 }, { "epoch": 0.28512995313165745, "grad_norm": 16.25, "learning_rate": 2.382872897696093e-06, "loss": 0.5694, "step": 3346 }, { "epoch": 0.2852151682999574, "grad_norm": 12.875, "learning_rate": 2.3828018691558915e-06, "loss": 0.589, "step": 3347 }, { "epoch": 0.28530038346825737, "grad_norm": 15.875, "learning_rate": 2.3827308201448215e-06, "loss": 1.1654, "step": 3348 }, { "epoch": 0.2853855986365573, "grad_norm": 12.75, "learning_rate": 2.382659750664166e-06, "loss": 0.5626, "step": 3349 }, { "epoch": 0.2854708138048573, "grad_norm": 15.0, "learning_rate": 2.382588660715211e-06, "loss": 0.9217, "step": 3350 }, { "epoch": 0.28555602897315724, "grad_norm": 20.875, "learning_rate": 2.3825175502992394e-06, "loss": 0.9546, "step": 3351 }, { "epoch": 0.2856412441414572, "grad_norm": 11.5625, "learning_rate": 2.3824464194175366e-06, "loss": 0.3705, "step": 3352 }, { "epoch": 0.28572645930975715, "grad_norm": 13.1875, "learning_rate": 2.3823752680713887e-06, "loss": 0.4902, "step": 3353 }, { "epoch": 0.2858116744780571, "grad_norm": 16.75, "learning_rate": 2.3823040962620805e-06, "loss": 1.0502, "step": 3354 }, { "epoch": 0.28589688964635707, "grad_norm": 13.625, "learning_rate": 2.382232903990899e-06, "loss": 0.547, "step": 3355 }, { "epoch": 0.285982104814657, "grad_norm": 19.0, "learning_rate": 2.38216169125913e-06, "loss": 1.1682, "step": 3356 }, { "epoch": 0.286067319982957, "grad_norm": 13.9375, "learning_rate": 2.382090458068061e-06, "loss": 0.8007, "step": 3357 }, { "epoch": 0.28615253515125694, "grad_norm": 16.0, "learning_rate": 2.3820192044189794e-06, "loss": 0.952, "step": 3358 }, { "epoch": 0.2862377503195569, "grad_norm": 15.1875, "learning_rate": 2.3819479303131722e-06, "loss": 0.6862, "step": 3359 }, { "epoch": 0.28632296548785685, "grad_norm": 30.875, "learning_rate": 2.3818766357519273e-06, "loss": 0.8938, "step": 3360 }, { "epoch": 0.2864081806561568, "grad_norm": 17.5, "learning_rate": 2.3818053207365337e-06, "loss": 0.5579, "step": 3361 }, { "epoch": 0.28649339582445676, "grad_norm": 11.6875, "learning_rate": 2.38173398526828e-06, "loss": 0.4603, "step": 3362 }, { "epoch": 0.2865786109927567, "grad_norm": 15.4375, "learning_rate": 2.381662629348455e-06, "loss": 1.006, "step": 3363 }, { "epoch": 0.2866638261610567, "grad_norm": 14.5625, "learning_rate": 2.381591252978348e-06, "loss": 0.8126, "step": 3364 }, { "epoch": 0.28674904132935664, "grad_norm": 16.875, "learning_rate": 2.381519856159249e-06, "loss": 0.8469, "step": 3365 }, { "epoch": 0.2868342564976566, "grad_norm": 12.6875, "learning_rate": 2.3814484388924487e-06, "loss": 0.4767, "step": 3366 }, { "epoch": 0.28691947166595655, "grad_norm": 18.125, "learning_rate": 2.3813770011792373e-06, "loss": 0.7269, "step": 3367 }, { "epoch": 0.2870046868342565, "grad_norm": 14.6875, "learning_rate": 2.3813055430209063e-06, "loss": 0.5826, "step": 3368 }, { "epoch": 0.28708990200255646, "grad_norm": 21.875, "learning_rate": 2.381234064418746e-06, "loss": 1.1915, "step": 3369 }, { "epoch": 0.2871751171708564, "grad_norm": 11.5625, "learning_rate": 2.3811625653740487e-06, "loss": 0.4676, "step": 3370 }, { "epoch": 0.2872603323391564, "grad_norm": 14.375, "learning_rate": 2.3810910458881064e-06, "loss": 0.6785, "step": 3371 }, { "epoch": 0.28734554750745633, "grad_norm": 15.3125, "learning_rate": 2.3810195059622117e-06, "loss": 0.6019, "step": 3372 }, { "epoch": 0.2874307626757563, "grad_norm": 13.9375, "learning_rate": 2.380947945597657e-06, "loss": 0.7969, "step": 3373 }, { "epoch": 0.28751597784405625, "grad_norm": 12.3125, "learning_rate": 2.3808763647957355e-06, "loss": 0.4861, "step": 3374 }, { "epoch": 0.2876011930123562, "grad_norm": 14.125, "learning_rate": 2.380804763557741e-06, "loss": 0.87, "step": 3375 }, { "epoch": 0.28768640818065616, "grad_norm": 11.5625, "learning_rate": 2.3807331418849675e-06, "loss": 0.4479, "step": 3376 }, { "epoch": 0.2877716233489561, "grad_norm": 23.625, "learning_rate": 2.3806614997787084e-06, "loss": 0.8922, "step": 3377 }, { "epoch": 0.2878568385172561, "grad_norm": 20.625, "learning_rate": 2.3805898372402605e-06, "loss": 0.7641, "step": 3378 }, { "epoch": 0.28794205368555603, "grad_norm": 12.8125, "learning_rate": 2.380518154270916e-06, "loss": 0.5588, "step": 3379 }, { "epoch": 0.288027268853856, "grad_norm": 13.375, "learning_rate": 2.380446450871972e-06, "loss": 0.6506, "step": 3380 }, { "epoch": 0.28811248402215595, "grad_norm": 12.1875, "learning_rate": 2.380374727044724e-06, "loss": 0.6244, "step": 3381 }, { "epoch": 0.2881976991904559, "grad_norm": 14.3125, "learning_rate": 2.380302982790468e-06, "loss": 1.1131, "step": 3382 }, { "epoch": 0.28828291435875586, "grad_norm": 18.0, "learning_rate": 2.3802312181105e-06, "loss": 1.1048, "step": 3383 }, { "epoch": 0.2883681295270558, "grad_norm": 19.25, "learning_rate": 2.3801594330061183e-06, "loss": 0.6149, "step": 3384 }, { "epoch": 0.2884533446953558, "grad_norm": 28.375, "learning_rate": 2.3800876274786185e-06, "loss": 0.8047, "step": 3385 }, { "epoch": 0.28853855986365573, "grad_norm": 8.6875, "learning_rate": 2.380015801529299e-06, "loss": 0.4653, "step": 3386 }, { "epoch": 0.2886237750319557, "grad_norm": 15.3125, "learning_rate": 2.379943955159458e-06, "loss": 0.7847, "step": 3387 }, { "epoch": 0.28870899020025564, "grad_norm": 13.5625, "learning_rate": 2.379872088370393e-06, "loss": 0.5854, "step": 3388 }, { "epoch": 0.2887942053685556, "grad_norm": 15.875, "learning_rate": 2.3798002011634035e-06, "loss": 0.7127, "step": 3389 }, { "epoch": 0.28887942053685556, "grad_norm": 24.25, "learning_rate": 2.379728293539788e-06, "loss": 1.0345, "step": 3390 }, { "epoch": 0.2889646357051555, "grad_norm": 11.875, "learning_rate": 2.3796563655008466e-06, "loss": 0.4436, "step": 3391 }, { "epoch": 0.28904985087345547, "grad_norm": 23.75, "learning_rate": 2.3795844170478783e-06, "loss": 1.2227, "step": 3392 }, { "epoch": 0.28913506604175543, "grad_norm": 17.375, "learning_rate": 2.379512448182184e-06, "loss": 0.9114, "step": 3393 }, { "epoch": 0.2892202812100554, "grad_norm": 14.4375, "learning_rate": 2.3794404589050636e-06, "loss": 0.8593, "step": 3394 }, { "epoch": 0.28930549637835534, "grad_norm": 14.75, "learning_rate": 2.3793684492178186e-06, "loss": 0.6833, "step": 3395 }, { "epoch": 0.2893907115466553, "grad_norm": 13.875, "learning_rate": 2.37929641912175e-06, "loss": 0.9602, "step": 3396 }, { "epoch": 0.28947592671495526, "grad_norm": 16.75, "learning_rate": 2.37922436861816e-06, "loss": 0.7795, "step": 3397 }, { "epoch": 0.2895611418832552, "grad_norm": 15.4375, "learning_rate": 2.379152297708349e-06, "loss": 0.998, "step": 3398 }, { "epoch": 0.28964635705155517, "grad_norm": 13.125, "learning_rate": 2.3790802063936215e-06, "loss": 0.496, "step": 3399 }, { "epoch": 0.2897315722198551, "grad_norm": 10.4375, "learning_rate": 2.379008094675279e-06, "loss": 0.448, "step": 3400 }, { "epoch": 0.2898167873881551, "grad_norm": 12.3125, "learning_rate": 2.378935962554625e-06, "loss": 0.5801, "step": 3401 }, { "epoch": 0.28990200255645504, "grad_norm": 15.375, "learning_rate": 2.378863810032963e-06, "loss": 0.763, "step": 3402 }, { "epoch": 0.289987217724755, "grad_norm": 16.5, "learning_rate": 2.3787916371115966e-06, "loss": 1.0671, "step": 3403 }, { "epoch": 0.29007243289305495, "grad_norm": 14.1875, "learning_rate": 2.3787194437918302e-06, "loss": 0.6688, "step": 3404 }, { "epoch": 0.2901576480613549, "grad_norm": 14.1875, "learning_rate": 2.3786472300749688e-06, "loss": 0.5871, "step": 3405 }, { "epoch": 0.29024286322965487, "grad_norm": 12.3125, "learning_rate": 2.3785749959623167e-06, "loss": 0.5202, "step": 3406 }, { "epoch": 0.2903280783979548, "grad_norm": 21.625, "learning_rate": 2.3785027414551793e-06, "loss": 1.0641, "step": 3407 }, { "epoch": 0.2904132935662548, "grad_norm": 11.0625, "learning_rate": 2.378430466554863e-06, "loss": 0.3786, "step": 3408 }, { "epoch": 0.29049850873455474, "grad_norm": 13.0, "learning_rate": 2.3783581712626733e-06, "loss": 0.6681, "step": 3409 }, { "epoch": 0.2905837239028547, "grad_norm": 14.8125, "learning_rate": 2.3782858555799167e-06, "loss": 0.834, "step": 3410 }, { "epoch": 0.29066893907115465, "grad_norm": 14.4375, "learning_rate": 2.3782135195079004e-06, "loss": 0.6867, "step": 3411 }, { "epoch": 0.2907541542394546, "grad_norm": 14.0, "learning_rate": 2.3781411630479307e-06, "loss": 0.8765, "step": 3412 }, { "epoch": 0.29083936940775457, "grad_norm": 14.5, "learning_rate": 2.3780687862013162e-06, "loss": 0.7197, "step": 3413 }, { "epoch": 0.2909245845760545, "grad_norm": 16.0, "learning_rate": 2.3779963889693644e-06, "loss": 0.7068, "step": 3414 }, { "epoch": 0.2910097997443545, "grad_norm": 13.9375, "learning_rate": 2.377923971353384e-06, "loss": 0.619, "step": 3415 }, { "epoch": 0.29109501491265444, "grad_norm": 13.375, "learning_rate": 2.377851533354682e-06, "loss": 0.4641, "step": 3416 }, { "epoch": 0.2911802300809544, "grad_norm": 13.1875, "learning_rate": 2.3777790749745695e-06, "loss": 0.5483, "step": 3417 }, { "epoch": 0.29126544524925435, "grad_norm": 12.125, "learning_rate": 2.3777065962143547e-06, "loss": 0.617, "step": 3418 }, { "epoch": 0.2913506604175543, "grad_norm": 15.125, "learning_rate": 2.377634097075348e-06, "loss": 0.8582, "step": 3419 }, { "epoch": 0.29143587558585426, "grad_norm": 13.125, "learning_rate": 2.3775615775588594e-06, "loss": 0.4053, "step": 3420 }, { "epoch": 0.2915210907541542, "grad_norm": 20.25, "learning_rate": 2.377489037666199e-06, "loss": 1.0821, "step": 3421 }, { "epoch": 0.2916063059224542, "grad_norm": 19.0, "learning_rate": 2.3774164773986775e-06, "loss": 1.035, "step": 3422 }, { "epoch": 0.29169152109075414, "grad_norm": 13.625, "learning_rate": 2.3773438967576075e-06, "loss": 0.4144, "step": 3423 }, { "epoch": 0.2917767362590541, "grad_norm": 13.375, "learning_rate": 2.377271295744299e-06, "loss": 0.6282, "step": 3424 }, { "epoch": 0.29186195142735405, "grad_norm": 20.25, "learning_rate": 2.3771986743600643e-06, "loss": 1.0134, "step": 3425 }, { "epoch": 0.291947166595654, "grad_norm": 14.5, "learning_rate": 2.377126032606217e-06, "loss": 0.753, "step": 3426 }, { "epoch": 0.29203238176395396, "grad_norm": 21.875, "learning_rate": 2.377053370484068e-06, "loss": 1.1791, "step": 3427 }, { "epoch": 0.2921175969322539, "grad_norm": 19.75, "learning_rate": 2.376980687994932e-06, "loss": 0.9957, "step": 3428 }, { "epoch": 0.2922028121005539, "grad_norm": 16.5, "learning_rate": 2.376907985140121e-06, "loss": 1.0694, "step": 3429 }, { "epoch": 0.29228802726885383, "grad_norm": 16.25, "learning_rate": 2.37683526192095e-06, "loss": 1.0766, "step": 3430 }, { "epoch": 0.2923732424371538, "grad_norm": 14.0, "learning_rate": 2.3767625183387323e-06, "loss": 0.79, "step": 3431 }, { "epoch": 0.29245845760545375, "grad_norm": 13.5625, "learning_rate": 2.376689754394783e-06, "loss": 0.7916, "step": 3432 }, { "epoch": 0.2925436727737537, "grad_norm": 15.875, "learning_rate": 2.3766169700904176e-06, "loss": 0.888, "step": 3433 }, { "epoch": 0.29262888794205366, "grad_norm": 13.8125, "learning_rate": 2.3765441654269495e-06, "loss": 0.8489, "step": 3434 }, { "epoch": 0.2927141031103536, "grad_norm": 15.9375, "learning_rate": 2.376471340405696e-06, "loss": 0.7272, "step": 3435 }, { "epoch": 0.2927993182786536, "grad_norm": 21.0, "learning_rate": 2.376398495027973e-06, "loss": 0.9728, "step": 3436 }, { "epoch": 0.29288453344695353, "grad_norm": 11.9375, "learning_rate": 2.376325629295096e-06, "loss": 0.3709, "step": 3437 }, { "epoch": 0.2929697486152535, "grad_norm": 16.625, "learning_rate": 2.376252743208382e-06, "loss": 0.8203, "step": 3438 }, { "epoch": 0.29305496378355345, "grad_norm": 14.0, "learning_rate": 2.376179836769149e-06, "loss": 0.8175, "step": 3439 }, { "epoch": 0.2931401789518534, "grad_norm": 13.8125, "learning_rate": 2.376106909978714e-06, "loss": 0.7277, "step": 3440 }, { "epoch": 0.29322539412015336, "grad_norm": 16.125, "learning_rate": 2.3760339628383944e-06, "loss": 0.7932, "step": 3441 }, { "epoch": 0.29331060928845337, "grad_norm": 19.375, "learning_rate": 2.375960995349509e-06, "loss": 0.3535, "step": 3442 }, { "epoch": 0.29339582445675333, "grad_norm": 14.4375, "learning_rate": 2.3758880075133754e-06, "loss": 0.7799, "step": 3443 }, { "epoch": 0.2934810396250533, "grad_norm": 13.375, "learning_rate": 2.3758149993313145e-06, "loss": 0.6711, "step": 3444 }, { "epoch": 0.29356625479335324, "grad_norm": 12.9375, "learning_rate": 2.3757419708046436e-06, "loss": 0.5204, "step": 3445 }, { "epoch": 0.2936514699616532, "grad_norm": 18.375, "learning_rate": 2.3756689219346833e-06, "loss": 0.8879, "step": 3446 }, { "epoch": 0.29373668512995316, "grad_norm": 13.5, "learning_rate": 2.375595852722754e-06, "loss": 0.7165, "step": 3447 }, { "epoch": 0.2938219002982531, "grad_norm": 11.125, "learning_rate": 2.375522763170176e-06, "loss": 0.4395, "step": 3448 }, { "epoch": 0.29390711546655307, "grad_norm": 15.3125, "learning_rate": 2.3754496532782696e-06, "loss": 0.9505, "step": 3449 }, { "epoch": 0.293992330634853, "grad_norm": 12.6875, "learning_rate": 2.375376523048356e-06, "loss": 0.635, "step": 3450 }, { "epoch": 0.294077545803153, "grad_norm": 13.5, "learning_rate": 2.375303372481757e-06, "loss": 0.6056, "step": 3451 }, { "epoch": 0.29416276097145294, "grad_norm": 15.75, "learning_rate": 2.3752302015797944e-06, "loss": 0.7424, "step": 3452 }, { "epoch": 0.2942479761397529, "grad_norm": 16.375, "learning_rate": 2.375157010343791e-06, "loss": 1.001, "step": 3453 }, { "epoch": 0.29433319130805286, "grad_norm": 13.375, "learning_rate": 2.375083798775069e-06, "loss": 0.3734, "step": 3454 }, { "epoch": 0.2944184064763528, "grad_norm": 17.125, "learning_rate": 2.3750105668749513e-06, "loss": 0.956, "step": 3455 }, { "epoch": 0.29450362164465277, "grad_norm": 12.5625, "learning_rate": 2.374937314644761e-06, "loss": 0.4849, "step": 3456 }, { "epoch": 0.2945888368129527, "grad_norm": 13.125, "learning_rate": 2.3748640420858228e-06, "loss": 0.5476, "step": 3457 }, { "epoch": 0.2946740519812527, "grad_norm": 18.5, "learning_rate": 2.3747907491994597e-06, "loss": 0.8018, "step": 3458 }, { "epoch": 0.29475926714955264, "grad_norm": 13.5625, "learning_rate": 2.374717435986997e-06, "loss": 0.6866, "step": 3459 }, { "epoch": 0.2948444823178526, "grad_norm": 11.5, "learning_rate": 2.3746441024497586e-06, "loss": 0.5413, "step": 3460 }, { "epoch": 0.29492969748615255, "grad_norm": 14.4375, "learning_rate": 2.374570748589071e-06, "loss": 0.6204, "step": 3461 }, { "epoch": 0.2950149126544525, "grad_norm": 12.6875, "learning_rate": 2.374497374406259e-06, "loss": 0.2954, "step": 3462 }, { "epoch": 0.29510012782275247, "grad_norm": 13.4375, "learning_rate": 2.3744239799026486e-06, "loss": 0.8429, "step": 3463 }, { "epoch": 0.2951853429910524, "grad_norm": 15.125, "learning_rate": 2.3743505650795663e-06, "loss": 0.8303, "step": 3464 }, { "epoch": 0.2952705581593524, "grad_norm": 17.75, "learning_rate": 2.3742771299383387e-06, "loss": 1.0318, "step": 3465 }, { "epoch": 0.29535577332765234, "grad_norm": 13.6875, "learning_rate": 2.3742036744802927e-06, "loss": 0.5929, "step": 3466 }, { "epoch": 0.2954409884959523, "grad_norm": 17.25, "learning_rate": 2.3741301987067557e-06, "loss": 0.8577, "step": 3467 }, { "epoch": 0.29552620366425225, "grad_norm": 17.625, "learning_rate": 2.374056702619056e-06, "loss": 1.2268, "step": 3468 }, { "epoch": 0.2956114188325522, "grad_norm": 14.3125, "learning_rate": 2.373983186218521e-06, "loss": 1.0293, "step": 3469 }, { "epoch": 0.29569663400085217, "grad_norm": 13.6875, "learning_rate": 2.37390964950648e-06, "loss": 0.6229, "step": 3470 }, { "epoch": 0.2957818491691521, "grad_norm": 12.5625, "learning_rate": 2.373836092484261e-06, "loss": 0.5964, "step": 3471 }, { "epoch": 0.2958670643374521, "grad_norm": 11.375, "learning_rate": 2.3737625151531937e-06, "loss": 0.4489, "step": 3472 }, { "epoch": 0.29595227950575204, "grad_norm": 14.1875, "learning_rate": 2.373688917514608e-06, "loss": 0.6087, "step": 3473 }, { "epoch": 0.296037494674052, "grad_norm": 14.6875, "learning_rate": 2.3736152995698335e-06, "loss": 0.8158, "step": 3474 }, { "epoch": 0.29612270984235195, "grad_norm": 12.125, "learning_rate": 2.3735416613202005e-06, "loss": 0.8275, "step": 3475 }, { "epoch": 0.2962079250106519, "grad_norm": 13.3125, "learning_rate": 2.37346800276704e-06, "loss": 0.6294, "step": 3476 }, { "epoch": 0.29629314017895186, "grad_norm": 15.0625, "learning_rate": 2.373394323911683e-06, "loss": 0.8348, "step": 3477 }, { "epoch": 0.2963783553472518, "grad_norm": 14.375, "learning_rate": 2.3733206247554607e-06, "loss": 0.8657, "step": 3478 }, { "epoch": 0.2964635705155518, "grad_norm": 13.8125, "learning_rate": 2.3732469052997055e-06, "loss": 1.0255, "step": 3479 }, { "epoch": 0.29654878568385173, "grad_norm": 17.125, "learning_rate": 2.3731731655457492e-06, "loss": 0.646, "step": 3480 }, { "epoch": 0.2966340008521517, "grad_norm": 20.875, "learning_rate": 2.373099405494924e-06, "loss": 1.4531, "step": 3481 }, { "epoch": 0.29671921602045165, "grad_norm": 23.25, "learning_rate": 2.3730256251485633e-06, "loss": 1.1542, "step": 3482 }, { "epoch": 0.2968044311887516, "grad_norm": 19.0, "learning_rate": 2.372951824508e-06, "loss": 1.0287, "step": 3483 }, { "epoch": 0.29688964635705156, "grad_norm": 15.875, "learning_rate": 2.3728780035745684e-06, "loss": 0.9181, "step": 3484 }, { "epoch": 0.2969748615253515, "grad_norm": 13.5625, "learning_rate": 2.3728041623496023e-06, "loss": 0.7606, "step": 3485 }, { "epoch": 0.2970600766936515, "grad_norm": 19.0, "learning_rate": 2.372730300834436e-06, "loss": 0.9599, "step": 3486 }, { "epoch": 0.29714529186195143, "grad_norm": 13.0625, "learning_rate": 2.3726564190304036e-06, "loss": 0.6571, "step": 3487 }, { "epoch": 0.2972305070302514, "grad_norm": 17.125, "learning_rate": 2.372582516938841e-06, "loss": 0.8738, "step": 3488 }, { "epoch": 0.29731572219855135, "grad_norm": 12.25, "learning_rate": 2.3725085945610833e-06, "loss": 0.5681, "step": 3489 }, { "epoch": 0.2974009373668513, "grad_norm": 16.75, "learning_rate": 2.372434651898467e-06, "loss": 0.9971, "step": 3490 }, { "epoch": 0.29748615253515126, "grad_norm": 11.0, "learning_rate": 2.3723606889523277e-06, "loss": 0.5215, "step": 3491 }, { "epoch": 0.2975713677034512, "grad_norm": 11.75, "learning_rate": 2.372286705724002e-06, "loss": 0.428, "step": 3492 }, { "epoch": 0.2976565828717512, "grad_norm": 16.625, "learning_rate": 2.3722127022148267e-06, "loss": 0.9708, "step": 3493 }, { "epoch": 0.29774179804005113, "grad_norm": 21.0, "learning_rate": 2.3721386784261398e-06, "loss": 0.6979, "step": 3494 }, { "epoch": 0.2978270132083511, "grad_norm": 11.75, "learning_rate": 2.3720646343592783e-06, "loss": 0.5514, "step": 3495 }, { "epoch": 0.29791222837665104, "grad_norm": 15.9375, "learning_rate": 2.371990570015581e-06, "loss": 1.0081, "step": 3496 }, { "epoch": 0.297997443544951, "grad_norm": 14.6875, "learning_rate": 2.3719164853963857e-06, "loss": 0.5804, "step": 3497 }, { "epoch": 0.29808265871325096, "grad_norm": 21.375, "learning_rate": 2.371842380503031e-06, "loss": 0.7738, "step": 3498 }, { "epoch": 0.2981678738815509, "grad_norm": 15.75, "learning_rate": 2.3717682553368566e-06, "loss": 0.4976, "step": 3499 }, { "epoch": 0.2982530890498509, "grad_norm": 14.5625, "learning_rate": 2.371694109899202e-06, "loss": 0.9709, "step": 3500 }, { "epoch": 0.29833830421815083, "grad_norm": 14.375, "learning_rate": 2.3716199441914064e-06, "loss": 0.9336, "step": 3501 }, { "epoch": 0.2984235193864508, "grad_norm": 15.125, "learning_rate": 2.371545758214811e-06, "loss": 0.7563, "step": 3502 }, { "epoch": 0.29850873455475074, "grad_norm": 13.75, "learning_rate": 2.3714715519707556e-06, "loss": 0.4922, "step": 3503 }, { "epoch": 0.2985939497230507, "grad_norm": 11.5625, "learning_rate": 2.3713973254605816e-06, "loss": 0.3999, "step": 3504 }, { "epoch": 0.29867916489135066, "grad_norm": 15.0, "learning_rate": 2.3713230786856305e-06, "loss": 0.845, "step": 3505 }, { "epoch": 0.2987643800596506, "grad_norm": 13.75, "learning_rate": 2.3712488116472437e-06, "loss": 0.7819, "step": 3506 }, { "epoch": 0.29884959522795057, "grad_norm": 17.875, "learning_rate": 2.371174524346763e-06, "loss": 0.9408, "step": 3507 }, { "epoch": 0.2989348103962505, "grad_norm": 17.875, "learning_rate": 2.371100216785532e-06, "loss": 1.0446, "step": 3508 }, { "epoch": 0.2990200255645505, "grad_norm": 12.0625, "learning_rate": 2.371025888964892e-06, "loss": 0.6994, "step": 3509 }, { "epoch": 0.29910524073285044, "grad_norm": 20.75, "learning_rate": 2.3709515408861873e-06, "loss": 1.1759, "step": 3510 }, { "epoch": 0.2991904559011504, "grad_norm": 13.6875, "learning_rate": 2.3708771725507605e-06, "loss": 0.7445, "step": 3511 }, { "epoch": 0.29927567106945036, "grad_norm": 17.75, "learning_rate": 2.3708027839599567e-06, "loss": 0.8531, "step": 3512 }, { "epoch": 0.2993608862377503, "grad_norm": 12.6875, "learning_rate": 2.3707283751151195e-06, "loss": 0.45, "step": 3513 }, { "epoch": 0.29944610140605027, "grad_norm": 9.75, "learning_rate": 2.3706539460175933e-06, "loss": 0.373, "step": 3514 }, { "epoch": 0.2995313165743502, "grad_norm": 11.1875, "learning_rate": 2.3705794966687236e-06, "loss": 0.4872, "step": 3515 }, { "epoch": 0.2996165317426502, "grad_norm": 15.625, "learning_rate": 2.3705050270698554e-06, "loss": 0.8585, "step": 3516 }, { "epoch": 0.29970174691095014, "grad_norm": 12.625, "learning_rate": 2.3704305372223346e-06, "loss": 0.5092, "step": 3517 }, { "epoch": 0.2997869620792501, "grad_norm": 18.625, "learning_rate": 2.3703560271275073e-06, "loss": 1.1137, "step": 3518 }, { "epoch": 0.29987217724755005, "grad_norm": 13.3125, "learning_rate": 2.37028149678672e-06, "loss": 0.4338, "step": 3519 }, { "epoch": 0.29995739241585, "grad_norm": 12.125, "learning_rate": 2.3702069462013196e-06, "loss": 0.4063, "step": 3520 }, { "epoch": 0.30004260758414997, "grad_norm": 17.875, "learning_rate": 2.3701323753726536e-06, "loss": 0.8559, "step": 3521 }, { "epoch": 0.3001278227524499, "grad_norm": 19.0, "learning_rate": 2.3700577843020686e-06, "loss": 1.0492, "step": 3522 }, { "epoch": 0.3002130379207499, "grad_norm": 14.9375, "learning_rate": 2.3699831729909135e-06, "loss": 0.8995, "step": 3523 }, { "epoch": 0.30029825308904984, "grad_norm": 11.125, "learning_rate": 2.369908541440536e-06, "loss": 0.5174, "step": 3524 }, { "epoch": 0.3003834682573498, "grad_norm": 11.625, "learning_rate": 2.369833889652285e-06, "loss": 0.4916, "step": 3525 }, { "epoch": 0.30046868342564975, "grad_norm": 14.1875, "learning_rate": 2.36975921762751e-06, "loss": 0.8715, "step": 3526 }, { "epoch": 0.3005538985939497, "grad_norm": 9.6875, "learning_rate": 2.369684525367559e-06, "loss": 0.441, "step": 3527 }, { "epoch": 0.30063911376224967, "grad_norm": 16.125, "learning_rate": 2.3696098128737833e-06, "loss": 0.4105, "step": 3528 }, { "epoch": 0.3007243289305496, "grad_norm": 13.125, "learning_rate": 2.3695350801475325e-06, "loss": 0.6402, "step": 3529 }, { "epoch": 0.3008095440988496, "grad_norm": 14.875, "learning_rate": 2.369460327190157e-06, "loss": 0.8228, "step": 3530 }, { "epoch": 0.30089475926714954, "grad_norm": 17.0, "learning_rate": 2.3693855540030074e-06, "loss": 0.6175, "step": 3531 }, { "epoch": 0.3009799744354495, "grad_norm": 46.5, "learning_rate": 2.3693107605874355e-06, "loss": 0.6063, "step": 3532 }, { "epoch": 0.30106518960374945, "grad_norm": 21.125, "learning_rate": 2.3692359469447923e-06, "loss": 1.1743, "step": 3533 }, { "epoch": 0.3011504047720494, "grad_norm": 14.1875, "learning_rate": 2.3691611130764303e-06, "loss": 0.9115, "step": 3534 }, { "epoch": 0.30123561994034936, "grad_norm": 12.875, "learning_rate": 2.3690862589837015e-06, "loss": 0.5827, "step": 3535 }, { "epoch": 0.3013208351086493, "grad_norm": 17.125, "learning_rate": 2.3690113846679586e-06, "loss": 0.9201, "step": 3536 }, { "epoch": 0.3014060502769493, "grad_norm": 21.125, "learning_rate": 2.368936490130555e-06, "loss": 1.3266, "step": 3537 }, { "epoch": 0.30149126544524923, "grad_norm": 17.75, "learning_rate": 2.3688615753728435e-06, "loss": 0.9453, "step": 3538 }, { "epoch": 0.3015764806135492, "grad_norm": 12.3125, "learning_rate": 2.3687866403961784e-06, "loss": 0.6202, "step": 3539 }, { "epoch": 0.30166169578184915, "grad_norm": 13.25, "learning_rate": 2.368711685201914e-06, "loss": 0.553, "step": 3540 }, { "epoch": 0.3017469109501491, "grad_norm": 16.625, "learning_rate": 2.3686367097914038e-06, "loss": 0.5821, "step": 3541 }, { "epoch": 0.30183212611844906, "grad_norm": 13.8125, "learning_rate": 2.368561714166004e-06, "loss": 0.5321, "step": 3542 }, { "epoch": 0.301917341286749, "grad_norm": 12.8125, "learning_rate": 2.3684866983270687e-06, "loss": 0.4729, "step": 3543 }, { "epoch": 0.302002556455049, "grad_norm": 13.5, "learning_rate": 2.3684116622759547e-06, "loss": 0.6648, "step": 3544 }, { "epoch": 0.30208777162334893, "grad_norm": 21.125, "learning_rate": 2.3683366060140166e-06, "loss": 0.8463, "step": 3545 }, { "epoch": 0.3021729867916489, "grad_norm": 14.25, "learning_rate": 2.368261529542612e-06, "loss": 0.7165, "step": 3546 }, { "epoch": 0.30225820195994885, "grad_norm": 14.1875, "learning_rate": 2.368186432863097e-06, "loss": 0.597, "step": 3547 }, { "epoch": 0.3023434171282488, "grad_norm": 15.1875, "learning_rate": 2.3681113159768284e-06, "loss": 0.7239, "step": 3548 }, { "epoch": 0.30242863229654876, "grad_norm": 27.875, "learning_rate": 2.368036178885164e-06, "loss": 1.1407, "step": 3549 }, { "epoch": 0.3025138474648487, "grad_norm": 12.1875, "learning_rate": 2.367961021589462e-06, "loss": 0.4478, "step": 3550 }, { "epoch": 0.3025990626331487, "grad_norm": 17.875, "learning_rate": 2.3678858440910795e-06, "loss": 1.024, "step": 3551 }, { "epoch": 0.30268427780144863, "grad_norm": 11.4375, "learning_rate": 2.3678106463913765e-06, "loss": 0.3483, "step": 3552 }, { "epoch": 0.3027694929697486, "grad_norm": 14.75, "learning_rate": 2.3677354284917102e-06, "loss": 0.7777, "step": 3553 }, { "epoch": 0.30285470813804855, "grad_norm": 16.125, "learning_rate": 2.3676601903934417e-06, "loss": 1.0319, "step": 3554 }, { "epoch": 0.30293992330634856, "grad_norm": 13.4375, "learning_rate": 2.367584932097929e-06, "loss": 0.8394, "step": 3555 }, { "epoch": 0.3030251384746485, "grad_norm": 16.625, "learning_rate": 2.3675096536065327e-06, "loss": 0.7785, "step": 3556 }, { "epoch": 0.30311035364294847, "grad_norm": 13.0, "learning_rate": 2.3674343549206137e-06, "loss": 0.7054, "step": 3557 }, { "epoch": 0.30319556881124843, "grad_norm": 12.625, "learning_rate": 2.3673590360415318e-06, "loss": 0.5428, "step": 3558 }, { "epoch": 0.3032807839795484, "grad_norm": 18.125, "learning_rate": 2.3672836969706484e-06, "loss": 1.0269, "step": 3559 }, { "epoch": 0.30336599914784834, "grad_norm": 13.5625, "learning_rate": 2.3672083377093253e-06, "loss": 0.7792, "step": 3560 }, { "epoch": 0.3034512143161483, "grad_norm": 16.25, "learning_rate": 2.367132958258924e-06, "loss": 0.8149, "step": 3561 }, { "epoch": 0.30353642948444826, "grad_norm": 15.8125, "learning_rate": 2.367057558620807e-06, "loss": 1.015, "step": 3562 }, { "epoch": 0.3036216446527482, "grad_norm": 14.625, "learning_rate": 2.366982138796337e-06, "loss": 0.9353, "step": 3563 }, { "epoch": 0.30370685982104817, "grad_norm": 21.125, "learning_rate": 2.366906698786876e-06, "loss": 0.5762, "step": 3564 }, { "epoch": 0.3037920749893481, "grad_norm": 14.1875, "learning_rate": 2.366831238593788e-06, "loss": 1.0013, "step": 3565 }, { "epoch": 0.3038772901576481, "grad_norm": 11.0625, "learning_rate": 2.366755758218436e-06, "loss": 0.3632, "step": 3566 }, { "epoch": 0.30396250532594804, "grad_norm": 15.875, "learning_rate": 2.366680257662185e-06, "loss": 0.9475, "step": 3567 }, { "epoch": 0.304047720494248, "grad_norm": 13.0, "learning_rate": 2.366604736926399e-06, "loss": 0.5595, "step": 3568 }, { "epoch": 0.30413293566254795, "grad_norm": 12.125, "learning_rate": 2.3665291960124422e-06, "loss": 0.5741, "step": 3569 }, { "epoch": 0.3042181508308479, "grad_norm": 12.1875, "learning_rate": 2.3664536349216797e-06, "loss": 0.4291, "step": 3570 }, { "epoch": 0.30430336599914787, "grad_norm": 13.6875, "learning_rate": 2.366378053655478e-06, "loss": 0.7145, "step": 3571 }, { "epoch": 0.3043885811674478, "grad_norm": 15.8125, "learning_rate": 2.3663024522152015e-06, "loss": 0.8785, "step": 3572 }, { "epoch": 0.3044737963357478, "grad_norm": 10.125, "learning_rate": 2.366226830602218e-06, "loss": 0.3718, "step": 3573 }, { "epoch": 0.30455901150404774, "grad_norm": 15.6875, "learning_rate": 2.3661511888178928e-06, "loss": 0.5046, "step": 3574 }, { "epoch": 0.3046442266723477, "grad_norm": 21.875, "learning_rate": 2.3660755268635933e-06, "loss": 0.8949, "step": 3575 }, { "epoch": 0.30472944184064765, "grad_norm": 17.875, "learning_rate": 2.365999844740687e-06, "loss": 1.0063, "step": 3576 }, { "epoch": 0.3048146570089476, "grad_norm": 14.0, "learning_rate": 2.365924142450541e-06, "loss": 0.7833, "step": 3577 }, { "epoch": 0.30489987217724757, "grad_norm": 14.0, "learning_rate": 2.3658484199945237e-06, "loss": 0.6318, "step": 3578 }, { "epoch": 0.3049850873455475, "grad_norm": 16.625, "learning_rate": 2.365772677374003e-06, "loss": 0.8016, "step": 3579 }, { "epoch": 0.3050703025138475, "grad_norm": 17.125, "learning_rate": 2.3656969145903486e-06, "loss": 0.9323, "step": 3580 }, { "epoch": 0.30515551768214744, "grad_norm": 16.875, "learning_rate": 2.3656211316449287e-06, "loss": 1.0339, "step": 3581 }, { "epoch": 0.3052407328504474, "grad_norm": 16.75, "learning_rate": 2.3655453285391133e-06, "loss": 0.6346, "step": 3582 }, { "epoch": 0.30532594801874735, "grad_norm": 21.125, "learning_rate": 2.3654695052742717e-06, "loss": 0.9334, "step": 3583 }, { "epoch": 0.3054111631870473, "grad_norm": 17.0, "learning_rate": 2.3653936618517747e-06, "loss": 0.9575, "step": 3584 }, { "epoch": 0.30549637835534726, "grad_norm": 13.625, "learning_rate": 2.365317798272993e-06, "loss": 0.615, "step": 3585 }, { "epoch": 0.3055815935236472, "grad_norm": 12.75, "learning_rate": 2.365241914539296e-06, "loss": 0.6607, "step": 3586 }, { "epoch": 0.3056668086919472, "grad_norm": 15.375, "learning_rate": 2.3651660106520574e-06, "loss": 0.6869, "step": 3587 }, { "epoch": 0.30575202386024714, "grad_norm": 15.75, "learning_rate": 2.365090086612647e-06, "loss": 0.7421, "step": 3588 }, { "epoch": 0.3058372390285471, "grad_norm": 12.6875, "learning_rate": 2.365014142422438e-06, "loss": 0.6018, "step": 3589 }, { "epoch": 0.30592245419684705, "grad_norm": 15.75, "learning_rate": 2.364938178082801e-06, "loss": 0.9748, "step": 3590 }, { "epoch": 0.306007669365147, "grad_norm": 16.75, "learning_rate": 2.364862193595111e-06, "loss": 0.93, "step": 3591 }, { "epoch": 0.30609288453344696, "grad_norm": 12.5, "learning_rate": 2.36478618896074e-06, "loss": 0.6869, "step": 3592 }, { "epoch": 0.3061780997017469, "grad_norm": 15.8125, "learning_rate": 2.364710164181061e-06, "loss": 0.691, "step": 3593 }, { "epoch": 0.3062633148700469, "grad_norm": 25.75, "learning_rate": 2.364634119257449e-06, "loss": 1.1043, "step": 3594 }, { "epoch": 0.30634853003834683, "grad_norm": 14.9375, "learning_rate": 2.3645580541912773e-06, "loss": 0.6804, "step": 3595 }, { "epoch": 0.3064337452066468, "grad_norm": 16.875, "learning_rate": 2.3644819689839205e-06, "loss": 0.9131, "step": 3596 }, { "epoch": 0.30651896037494675, "grad_norm": 11.375, "learning_rate": 2.3644058636367544e-06, "loss": 0.5796, "step": 3597 }, { "epoch": 0.3066041755432467, "grad_norm": 11.5, "learning_rate": 2.364329738151153e-06, "loss": 0.5598, "step": 3598 }, { "epoch": 0.30668939071154666, "grad_norm": 13.1875, "learning_rate": 2.3642535925284937e-06, "loss": 0.78, "step": 3599 }, { "epoch": 0.3067746058798466, "grad_norm": 10.125, "learning_rate": 2.3641774267701505e-06, "loss": 0.4551, "step": 3600 }, { "epoch": 0.3068598210481466, "grad_norm": 12.375, "learning_rate": 2.3641012408775018e-06, "loss": 0.4729, "step": 3601 }, { "epoch": 0.30694503621644653, "grad_norm": 15.125, "learning_rate": 2.364025034851923e-06, "loss": 0.8796, "step": 3602 }, { "epoch": 0.3070302513847465, "grad_norm": 17.625, "learning_rate": 2.363948808694791e-06, "loss": 0.7959, "step": 3603 }, { "epoch": 0.30711546655304645, "grad_norm": 10.875, "learning_rate": 2.3638725624074844e-06, "loss": 0.9934, "step": 3604 }, { "epoch": 0.3072006817213464, "grad_norm": 21.875, "learning_rate": 2.36379629599138e-06, "loss": 0.9451, "step": 3605 }, { "epoch": 0.30728589688964636, "grad_norm": 20.875, "learning_rate": 2.363720009447857e-06, "loss": 1.09, "step": 3606 }, { "epoch": 0.3073711120579463, "grad_norm": 12.5, "learning_rate": 2.3636437027782937e-06, "loss": 0.6467, "step": 3607 }, { "epoch": 0.3074563272262463, "grad_norm": 12.4375, "learning_rate": 2.3635673759840685e-06, "loss": 0.6728, "step": 3608 }, { "epoch": 0.30754154239454623, "grad_norm": 23.5, "learning_rate": 2.3634910290665614e-06, "loss": 1.1796, "step": 3609 }, { "epoch": 0.3076267575628462, "grad_norm": 15.9375, "learning_rate": 2.363414662027151e-06, "loss": 1.0791, "step": 3610 }, { "epoch": 0.30771197273114614, "grad_norm": 12.125, "learning_rate": 2.3633382748672186e-06, "loss": 0.5334, "step": 3611 }, { "epoch": 0.3077971878994461, "grad_norm": 13.0625, "learning_rate": 2.363261867588144e-06, "loss": 0.7565, "step": 3612 }, { "epoch": 0.30788240306774606, "grad_norm": 15.375, "learning_rate": 2.3631854401913077e-06, "loss": 0.8815, "step": 3613 }, { "epoch": 0.307967618236046, "grad_norm": 17.25, "learning_rate": 2.3631089926780913e-06, "loss": 1.0213, "step": 3614 }, { "epoch": 0.30805283340434597, "grad_norm": 18.125, "learning_rate": 2.363032525049876e-06, "loss": 0.7114, "step": 3615 }, { "epoch": 0.30813804857264593, "grad_norm": 17.25, "learning_rate": 2.3629560373080436e-06, "loss": 0.8506, "step": 3616 }, { "epoch": 0.3082232637409459, "grad_norm": 29.625, "learning_rate": 2.3628795294539767e-06, "loss": 1.0163, "step": 3617 }, { "epoch": 0.30830847890924584, "grad_norm": 11.1875, "learning_rate": 2.3628030014890576e-06, "loss": 0.3581, "step": 3618 }, { "epoch": 0.3083936940775458, "grad_norm": 14.8125, "learning_rate": 2.362726453414669e-06, "loss": 0.5291, "step": 3619 }, { "epoch": 0.30847890924584576, "grad_norm": 40.25, "learning_rate": 2.3626498852321947e-06, "loss": 0.9466, "step": 3620 }, { "epoch": 0.3085641244141457, "grad_norm": 12.0, "learning_rate": 2.362573296943018e-06, "loss": 0.5573, "step": 3621 }, { "epoch": 0.30864933958244567, "grad_norm": 11.4375, "learning_rate": 2.362496688548523e-06, "loss": 0.4074, "step": 3622 }, { "epoch": 0.3087345547507456, "grad_norm": 15.25, "learning_rate": 2.3624200600500943e-06, "loss": 1.2185, "step": 3623 }, { "epoch": 0.3088197699190456, "grad_norm": 14.625, "learning_rate": 2.3623434114491166e-06, "loss": 0.8924, "step": 3624 }, { "epoch": 0.30890498508734554, "grad_norm": 14.5625, "learning_rate": 2.3622667427469743e-06, "loss": 0.7894, "step": 3625 }, { "epoch": 0.3089902002556455, "grad_norm": 16.875, "learning_rate": 2.3621900539450544e-06, "loss": 1.0706, "step": 3626 }, { "epoch": 0.30907541542394545, "grad_norm": 23.875, "learning_rate": 2.362113345044741e-06, "loss": 1.1409, "step": 3627 }, { "epoch": 0.3091606305922454, "grad_norm": 12.125, "learning_rate": 2.362036616047422e-06, "loss": 0.7046, "step": 3628 }, { "epoch": 0.30924584576054537, "grad_norm": 13.25, "learning_rate": 2.3619598669544824e-06, "loss": 0.6346, "step": 3629 }, { "epoch": 0.3093310609288453, "grad_norm": 13.5625, "learning_rate": 2.3618830977673096e-06, "loss": 0.5055, "step": 3630 }, { "epoch": 0.3094162760971453, "grad_norm": 19.0, "learning_rate": 2.3618063084872917e-06, "loss": 0.9511, "step": 3631 }, { "epoch": 0.30950149126544524, "grad_norm": 13.9375, "learning_rate": 2.361729499115816e-06, "loss": 0.8389, "step": 3632 }, { "epoch": 0.3095867064337452, "grad_norm": 15.875, "learning_rate": 2.3616526696542698e-06, "loss": 0.8227, "step": 3633 }, { "epoch": 0.30967192160204515, "grad_norm": 15.9375, "learning_rate": 2.361575820104042e-06, "loss": 0.9879, "step": 3634 }, { "epoch": 0.3097571367703451, "grad_norm": 17.5, "learning_rate": 2.3614989504665216e-06, "loss": 0.9241, "step": 3635 }, { "epoch": 0.30984235193864507, "grad_norm": 14.625, "learning_rate": 2.3614220607430973e-06, "loss": 0.8256, "step": 3636 }, { "epoch": 0.309927567106945, "grad_norm": 12.4375, "learning_rate": 2.361345150935159e-06, "loss": 0.5973, "step": 3637 }, { "epoch": 0.310012782275245, "grad_norm": 12.75, "learning_rate": 2.361268221044096e-06, "loss": 0.5119, "step": 3638 }, { "epoch": 0.31009799744354494, "grad_norm": 18.375, "learning_rate": 2.3611912710712987e-06, "loss": 0.7814, "step": 3639 }, { "epoch": 0.3101832126118449, "grad_norm": 10.5, "learning_rate": 2.361114301018158e-06, "loss": 0.4291, "step": 3640 }, { "epoch": 0.31026842778014485, "grad_norm": 17.875, "learning_rate": 2.3610373108860646e-06, "loss": 0.7523, "step": 3641 }, { "epoch": 0.3103536429484448, "grad_norm": 15.3125, "learning_rate": 2.36096030067641e-06, "loss": 0.7863, "step": 3642 }, { "epoch": 0.31043885811674476, "grad_norm": 14.8125, "learning_rate": 2.3608832703905853e-06, "loss": 0.7521, "step": 3643 }, { "epoch": 0.3105240732850447, "grad_norm": 14.0625, "learning_rate": 2.3608062200299825e-06, "loss": 1.0106, "step": 3644 }, { "epoch": 0.3106092884533447, "grad_norm": 16.375, "learning_rate": 2.3607291495959947e-06, "loss": 0.7794, "step": 3645 }, { "epoch": 0.31069450362164464, "grad_norm": 13.3125, "learning_rate": 2.360652059090014e-06, "loss": 0.7421, "step": 3646 }, { "epoch": 0.3107797187899446, "grad_norm": 15.0, "learning_rate": 2.360574948513434e-06, "loss": 0.78, "step": 3647 }, { "epoch": 0.31086493395824455, "grad_norm": 12.6875, "learning_rate": 2.360497817867648e-06, "loss": 0.5243, "step": 3648 }, { "epoch": 0.3109501491265445, "grad_norm": 12.125, "learning_rate": 2.3604206671540495e-06, "loss": 0.5473, "step": 3649 }, { "epoch": 0.31103536429484446, "grad_norm": 14.25, "learning_rate": 2.360343496374033e-06, "loss": 0.9088, "step": 3650 }, { "epoch": 0.3111205794631444, "grad_norm": 17.25, "learning_rate": 2.360266305528993e-06, "loss": 0.6185, "step": 3651 }, { "epoch": 0.3112057946314444, "grad_norm": 13.375, "learning_rate": 2.3601890946203245e-06, "loss": 0.641, "step": 3652 }, { "epoch": 0.31129100979974433, "grad_norm": 18.25, "learning_rate": 2.3601118636494223e-06, "loss": 0.6078, "step": 3653 }, { "epoch": 0.3113762249680443, "grad_norm": 16.625, "learning_rate": 2.3600346126176827e-06, "loss": 0.7837, "step": 3654 }, { "epoch": 0.31146144013634425, "grad_norm": 10.8125, "learning_rate": 2.3599573415265013e-06, "loss": 0.4894, "step": 3655 }, { "epoch": 0.3115466553046442, "grad_norm": 14.3125, "learning_rate": 2.3598800503772747e-06, "loss": 0.8935, "step": 3656 }, { "epoch": 0.31163187047294416, "grad_norm": 11.3125, "learning_rate": 2.3598027391713994e-06, "loss": 0.5816, "step": 3657 }, { "epoch": 0.3117170856412441, "grad_norm": 18.375, "learning_rate": 2.359725407910273e-06, "loss": 0.977, "step": 3658 }, { "epoch": 0.3118023008095441, "grad_norm": 14.125, "learning_rate": 2.3596480565952923e-06, "loss": 0.5791, "step": 3659 }, { "epoch": 0.31188751597784403, "grad_norm": 14.625, "learning_rate": 2.3595706852278554e-06, "loss": 0.8229, "step": 3660 }, { "epoch": 0.311972731146144, "grad_norm": 17.0, "learning_rate": 2.35949329380936e-06, "loss": 0.7805, "step": 3661 }, { "epoch": 0.31205794631444395, "grad_norm": 12.5625, "learning_rate": 2.3594158823412057e-06, "loss": 0.5918, "step": 3662 }, { "epoch": 0.3121431614827439, "grad_norm": 11.5625, "learning_rate": 2.3593384508247906e-06, "loss": 0.552, "step": 3663 }, { "epoch": 0.31222837665104386, "grad_norm": 17.75, "learning_rate": 2.359260999261514e-06, "loss": 1.0187, "step": 3664 }, { "epoch": 0.3123135918193438, "grad_norm": 22.75, "learning_rate": 2.359183527652776e-06, "loss": 0.9452, "step": 3665 }, { "epoch": 0.3123988069876438, "grad_norm": 17.125, "learning_rate": 2.359106035999976e-06, "loss": 0.9554, "step": 3666 }, { "epoch": 0.31248402215594373, "grad_norm": 13.8125, "learning_rate": 2.359028524304515e-06, "loss": 0.4401, "step": 3667 }, { "epoch": 0.31256923732424374, "grad_norm": 19.125, "learning_rate": 2.358950992567793e-06, "loss": 1.0082, "step": 3668 }, { "epoch": 0.3126544524925437, "grad_norm": 15.4375, "learning_rate": 2.3588734407912116e-06, "loss": 0.6147, "step": 3669 }, { "epoch": 0.31273966766084366, "grad_norm": 15.5625, "learning_rate": 2.358795868976172e-06, "loss": 0.7906, "step": 3670 }, { "epoch": 0.3128248828291436, "grad_norm": 16.5, "learning_rate": 2.3587182771240764e-06, "loss": 0.7904, "step": 3671 }, { "epoch": 0.31291009799744357, "grad_norm": 17.0, "learning_rate": 2.358640665236326e-06, "loss": 0.8578, "step": 3672 }, { "epoch": 0.3129953131657435, "grad_norm": 19.0, "learning_rate": 2.3585630333143244e-06, "loss": 1.0, "step": 3673 }, { "epoch": 0.3130805283340435, "grad_norm": 12.375, "learning_rate": 2.358485381359474e-06, "loss": 0.4281, "step": 3674 }, { "epoch": 0.31316574350234344, "grad_norm": 14.375, "learning_rate": 2.3584077093731782e-06, "loss": 0.855, "step": 3675 }, { "epoch": 0.3132509586706434, "grad_norm": 12.6875, "learning_rate": 2.35833001735684e-06, "loss": 0.5559, "step": 3676 }, { "epoch": 0.31333617383894335, "grad_norm": 18.875, "learning_rate": 2.3582523053118644e-06, "loss": 0.8259, "step": 3677 }, { "epoch": 0.3134213890072433, "grad_norm": 30.0, "learning_rate": 2.358174573239655e-06, "loss": 1.0447, "step": 3678 }, { "epoch": 0.31350660417554327, "grad_norm": 11.5, "learning_rate": 2.358096821141617e-06, "loss": 0.349, "step": 3679 }, { "epoch": 0.3135918193438432, "grad_norm": 12.25, "learning_rate": 2.3580190490191545e-06, "loss": 0.4728, "step": 3680 }, { "epoch": 0.3136770345121432, "grad_norm": 14.5625, "learning_rate": 2.3579412568736743e-06, "loss": 1.0822, "step": 3681 }, { "epoch": 0.31376224968044314, "grad_norm": 15.4375, "learning_rate": 2.3578634447065808e-06, "loss": 0.6368, "step": 3682 }, { "epoch": 0.3138474648487431, "grad_norm": 14.5, "learning_rate": 2.357785612519281e-06, "loss": 0.5831, "step": 3683 }, { "epoch": 0.31393268001704305, "grad_norm": 18.125, "learning_rate": 2.3577077603131816e-06, "loss": 1.168, "step": 3684 }, { "epoch": 0.314017895185343, "grad_norm": 15.3125, "learning_rate": 2.3576298880896886e-06, "loss": 0.9287, "step": 3685 }, { "epoch": 0.31410311035364297, "grad_norm": 21.75, "learning_rate": 2.3575519958502103e-06, "loss": 0.7875, "step": 3686 }, { "epoch": 0.3141883255219429, "grad_norm": 20.625, "learning_rate": 2.357474083596153e-06, "loss": 0.7885, "step": 3687 }, { "epoch": 0.3142735406902429, "grad_norm": 19.5, "learning_rate": 2.357396151328926e-06, "loss": 0.9576, "step": 3688 }, { "epoch": 0.31435875585854284, "grad_norm": 14.5625, "learning_rate": 2.3573181990499367e-06, "loss": 0.5863, "step": 3689 }, { "epoch": 0.3144439710268428, "grad_norm": 21.75, "learning_rate": 2.357240226760594e-06, "loss": 0.7171, "step": 3690 }, { "epoch": 0.31452918619514275, "grad_norm": 11.625, "learning_rate": 2.357162234462307e-06, "loss": 0.6142, "step": 3691 }, { "epoch": 0.3146144013634427, "grad_norm": 12.5625, "learning_rate": 2.357084222156485e-06, "loss": 0.6429, "step": 3692 }, { "epoch": 0.31469961653174267, "grad_norm": 16.875, "learning_rate": 2.357006189844538e-06, "loss": 1.093, "step": 3693 }, { "epoch": 0.3147848317000426, "grad_norm": 15.0625, "learning_rate": 2.356928137527876e-06, "loss": 0.5054, "step": 3694 }, { "epoch": 0.3148700468683426, "grad_norm": 12.8125, "learning_rate": 2.3568500652079096e-06, "loss": 0.6249, "step": 3695 }, { "epoch": 0.31495526203664254, "grad_norm": 14.8125, "learning_rate": 2.356771972886049e-06, "loss": 0.7313, "step": 3696 }, { "epoch": 0.3150404772049425, "grad_norm": 14.375, "learning_rate": 2.3566938605637063e-06, "loss": 0.7174, "step": 3697 }, { "epoch": 0.31512569237324245, "grad_norm": 12.375, "learning_rate": 2.3566157282422924e-06, "loss": 0.4473, "step": 3698 }, { "epoch": 0.3152109075415424, "grad_norm": 16.375, "learning_rate": 2.35653757592322e-06, "loss": 1.2287, "step": 3699 }, { "epoch": 0.31529612270984236, "grad_norm": 11.9375, "learning_rate": 2.3564594036079003e-06, "loss": 0.5213, "step": 3700 }, { "epoch": 0.3153813378781423, "grad_norm": 15.125, "learning_rate": 2.356381211297747e-06, "loss": 0.6685, "step": 3701 }, { "epoch": 0.3154665530464423, "grad_norm": 18.25, "learning_rate": 2.356302998994172e-06, "loss": 1.0843, "step": 3702 }, { "epoch": 0.31555176821474223, "grad_norm": 12.0625, "learning_rate": 2.3562247666985903e-06, "loss": 0.5893, "step": 3703 }, { "epoch": 0.3156369833830422, "grad_norm": 18.25, "learning_rate": 2.356146514412414e-06, "loss": 0.9412, "step": 3704 }, { "epoch": 0.31572219855134215, "grad_norm": 11.9375, "learning_rate": 2.356068242137058e-06, "loss": 0.4863, "step": 3705 }, { "epoch": 0.3158074137196421, "grad_norm": 13.75, "learning_rate": 2.3559899498739367e-06, "loss": 1.0099, "step": 3706 }, { "epoch": 0.31589262888794206, "grad_norm": 19.875, "learning_rate": 2.355911637624465e-06, "loss": 0.7066, "step": 3707 }, { "epoch": 0.315977844056242, "grad_norm": 14.3125, "learning_rate": 2.3558333053900574e-06, "loss": 0.6905, "step": 3708 }, { "epoch": 0.316063059224542, "grad_norm": 12.25, "learning_rate": 2.3557549531721306e-06, "loss": 0.6421, "step": 3709 }, { "epoch": 0.31614827439284193, "grad_norm": 12.125, "learning_rate": 2.3556765809720994e-06, "loss": 0.6885, "step": 3710 }, { "epoch": 0.3162334895611419, "grad_norm": 15.0, "learning_rate": 2.355598188791381e-06, "loss": 0.7999, "step": 3711 }, { "epoch": 0.31631870472944185, "grad_norm": 28.875, "learning_rate": 2.355519776631391e-06, "loss": 1.2331, "step": 3712 }, { "epoch": 0.3164039198977418, "grad_norm": 9.9375, "learning_rate": 2.3554413444935474e-06, "loss": 0.4101, "step": 3713 }, { "epoch": 0.31648913506604176, "grad_norm": 12.25, "learning_rate": 2.3553628923792666e-06, "loss": 0.5618, "step": 3714 }, { "epoch": 0.3165743502343417, "grad_norm": 19.25, "learning_rate": 2.3552844202899673e-06, "loss": 0.9511, "step": 3715 }, { "epoch": 0.3166595654026417, "grad_norm": 29.375, "learning_rate": 2.355205928227067e-06, "loss": 1.1118, "step": 3716 }, { "epoch": 0.31674478057094163, "grad_norm": 14.375, "learning_rate": 2.355127416191984e-06, "loss": 0.8699, "step": 3717 }, { "epoch": 0.3168299957392416, "grad_norm": 13.625, "learning_rate": 2.3550488841861374e-06, "loss": 0.5536, "step": 3718 }, { "epoch": 0.31691521090754154, "grad_norm": 14.9375, "learning_rate": 2.354970332210946e-06, "loss": 0.7641, "step": 3719 }, { "epoch": 0.3170004260758415, "grad_norm": 15.75, "learning_rate": 2.35489176026783e-06, "loss": 0.5738, "step": 3720 }, { "epoch": 0.31708564124414146, "grad_norm": 11.8125, "learning_rate": 2.3548131683582088e-06, "loss": 0.497, "step": 3721 }, { "epoch": 0.3171708564124414, "grad_norm": 14.8125, "learning_rate": 2.354734556483503e-06, "loss": 1.0743, "step": 3722 }, { "epoch": 0.3172560715807414, "grad_norm": 11.9375, "learning_rate": 2.354655924645132e-06, "loss": 0.7099, "step": 3723 }, { "epoch": 0.31734128674904133, "grad_norm": 12.3125, "learning_rate": 2.354577272844518e-06, "loss": 0.6531, "step": 3724 }, { "epoch": 0.3174265019173413, "grad_norm": 19.25, "learning_rate": 2.354498601083082e-06, "loss": 0.8511, "step": 3725 }, { "epoch": 0.31751171708564124, "grad_norm": 22.5, "learning_rate": 2.3544199093622455e-06, "loss": 1.1105, "step": 3726 }, { "epoch": 0.3175969322539412, "grad_norm": 14.125, "learning_rate": 2.3543411976834305e-06, "loss": 0.6921, "step": 3727 }, { "epoch": 0.31768214742224116, "grad_norm": 13.125, "learning_rate": 2.3542624660480603e-06, "loss": 0.4625, "step": 3728 }, { "epoch": 0.3177673625905411, "grad_norm": 14.75, "learning_rate": 2.3541837144575565e-06, "loss": 1.1254, "step": 3729 }, { "epoch": 0.31785257775884107, "grad_norm": 17.375, "learning_rate": 2.354104942913342e-06, "loss": 1.0532, "step": 3730 }, { "epoch": 0.317937792927141, "grad_norm": 17.625, "learning_rate": 2.3540261514168418e-06, "loss": 0.6351, "step": 3731 }, { "epoch": 0.318023008095441, "grad_norm": 252.0, "learning_rate": 2.3539473399694782e-06, "loss": 1.0827, "step": 3732 }, { "epoch": 0.31810822326374094, "grad_norm": 10.5625, "learning_rate": 2.3538685085726766e-06, "loss": 0.4128, "step": 3733 }, { "epoch": 0.3181934384320409, "grad_norm": 20.125, "learning_rate": 2.353789657227861e-06, "loss": 0.355, "step": 3734 }, { "epoch": 0.31827865360034086, "grad_norm": 9.9375, "learning_rate": 2.353710785936456e-06, "loss": 0.444, "step": 3735 }, { "epoch": 0.3183638687686408, "grad_norm": 17.0, "learning_rate": 2.3536318946998877e-06, "loss": 0.5497, "step": 3736 }, { "epoch": 0.31844908393694077, "grad_norm": 13.0, "learning_rate": 2.353552983519581e-06, "loss": 0.6666, "step": 3737 }, { "epoch": 0.3185342991052407, "grad_norm": 11.6875, "learning_rate": 2.353474052396962e-06, "loss": 0.4177, "step": 3738 }, { "epoch": 0.3186195142735407, "grad_norm": 16.875, "learning_rate": 2.3533951013334575e-06, "loss": 0.6591, "step": 3739 }, { "epoch": 0.31870472944184064, "grad_norm": 14.125, "learning_rate": 2.353316130330494e-06, "loss": 0.8239, "step": 3740 }, { "epoch": 0.3187899446101406, "grad_norm": 11.4375, "learning_rate": 2.3532371393894983e-06, "loss": 0.3792, "step": 3741 }, { "epoch": 0.31887515977844055, "grad_norm": 14.5, "learning_rate": 2.3531581285118983e-06, "loss": 0.7582, "step": 3742 }, { "epoch": 0.3189603749467405, "grad_norm": 18.75, "learning_rate": 2.3530790976991214e-06, "loss": 1.1234, "step": 3743 }, { "epoch": 0.31904559011504047, "grad_norm": 13.8125, "learning_rate": 2.353000046952596e-06, "loss": 0.7819, "step": 3744 }, { "epoch": 0.3191308052833404, "grad_norm": 12.9375, "learning_rate": 2.352920976273751e-06, "loss": 0.6734, "step": 3745 }, { "epoch": 0.3192160204516404, "grad_norm": 16.125, "learning_rate": 2.3528418856640145e-06, "loss": 0.542, "step": 3746 }, { "epoch": 0.31930123561994034, "grad_norm": 17.0, "learning_rate": 2.352762775124816e-06, "loss": 1.0065, "step": 3747 }, { "epoch": 0.3193864507882403, "grad_norm": 34.5, "learning_rate": 2.352683644657586e-06, "loss": 0.9629, "step": 3748 }, { "epoch": 0.31947166595654025, "grad_norm": 12.25, "learning_rate": 2.3526044942637526e-06, "loss": 0.7039, "step": 3749 }, { "epoch": 0.3195568811248402, "grad_norm": 13.8125, "learning_rate": 2.352525323944748e-06, "loss": 0.7703, "step": 3750 }, { "epoch": 0.31964209629314017, "grad_norm": 15.5625, "learning_rate": 2.352446133702002e-06, "loss": 0.7529, "step": 3751 }, { "epoch": 0.3197273114614401, "grad_norm": 13.4375, "learning_rate": 2.3523669235369457e-06, "loss": 0.4563, "step": 3752 }, { "epoch": 0.3198125266297401, "grad_norm": 12.3125, "learning_rate": 2.3522876934510104e-06, "loss": 0.5116, "step": 3753 }, { "epoch": 0.31989774179804004, "grad_norm": 15.25, "learning_rate": 2.352208443445628e-06, "loss": 0.9719, "step": 3754 }, { "epoch": 0.31998295696634, "grad_norm": 15.0, "learning_rate": 2.352129173522231e-06, "loss": 0.6953, "step": 3755 }, { "epoch": 0.32006817213463995, "grad_norm": 9.5625, "learning_rate": 2.3520498836822515e-06, "loss": 0.288, "step": 3756 }, { "epoch": 0.3201533873029399, "grad_norm": 12.5625, "learning_rate": 2.3519705739271224e-06, "loss": 0.5002, "step": 3757 }, { "epoch": 0.32023860247123986, "grad_norm": 14.625, "learning_rate": 2.351891244258277e-06, "loss": 0.7877, "step": 3758 }, { "epoch": 0.3203238176395398, "grad_norm": 9.5625, "learning_rate": 2.3518118946771484e-06, "loss": 0.3697, "step": 3759 }, { "epoch": 0.3204090328078398, "grad_norm": 15.25, "learning_rate": 2.3517325251851716e-06, "loss": 0.6415, "step": 3760 }, { "epoch": 0.32049424797613973, "grad_norm": 17.625, "learning_rate": 2.3516531357837793e-06, "loss": 0.9096, "step": 3761 }, { "epoch": 0.3205794631444397, "grad_norm": 15.375, "learning_rate": 2.3515737264744077e-06, "loss": 1.1126, "step": 3762 }, { "epoch": 0.32066467831273965, "grad_norm": 12.5625, "learning_rate": 2.351494297258491e-06, "loss": 0.4516, "step": 3763 }, { "epoch": 0.3207498934810396, "grad_norm": 17.75, "learning_rate": 2.3514148481374643e-06, "loss": 0.9108, "step": 3764 }, { "epoch": 0.32083510864933956, "grad_norm": 14.9375, "learning_rate": 2.351335379112764e-06, "loss": 0.8066, "step": 3765 }, { "epoch": 0.3209203238176395, "grad_norm": 13.9375, "learning_rate": 2.351255890185826e-06, "loss": 0.6022, "step": 3766 }, { "epoch": 0.3210055389859395, "grad_norm": 15.75, "learning_rate": 2.351176381358087e-06, "loss": 1.2152, "step": 3767 }, { "epoch": 0.32109075415423943, "grad_norm": 14.0625, "learning_rate": 2.351096852630983e-06, "loss": 0.8962, "step": 3768 }, { "epoch": 0.3211759693225394, "grad_norm": 15.125, "learning_rate": 2.351017304005952e-06, "loss": 0.7988, "step": 3769 }, { "epoch": 0.32126118449083935, "grad_norm": 11.625, "learning_rate": 2.350937735484431e-06, "loss": 0.3469, "step": 3770 }, { "epoch": 0.3213463996591393, "grad_norm": 17.375, "learning_rate": 2.3508581470678576e-06, "loss": 1.0225, "step": 3771 }, { "epoch": 0.32143161482743926, "grad_norm": 16.625, "learning_rate": 2.350778538757671e-06, "loss": 0.7168, "step": 3772 }, { "epoch": 0.3215168299957392, "grad_norm": 17.625, "learning_rate": 2.350698910555309e-06, "loss": 0.7265, "step": 3773 }, { "epoch": 0.3216020451640392, "grad_norm": 19.625, "learning_rate": 2.350619262462211e-06, "loss": 0.8753, "step": 3774 }, { "epoch": 0.32168726033233913, "grad_norm": 15.4375, "learning_rate": 2.350539594479816e-06, "loss": 0.6382, "step": 3775 }, { "epoch": 0.3217724755006391, "grad_norm": 12.25, "learning_rate": 2.350459906609564e-06, "loss": 0.5823, "step": 3776 }, { "epoch": 0.32185769066893904, "grad_norm": 16.875, "learning_rate": 2.350380198852895e-06, "loss": 0.6452, "step": 3777 }, { "epoch": 0.321942905837239, "grad_norm": 22.375, "learning_rate": 2.3503004712112495e-06, "loss": 0.9861, "step": 3778 }, { "epoch": 0.32202812100553896, "grad_norm": 29.375, "learning_rate": 2.3502207236860673e-06, "loss": 1.3897, "step": 3779 }, { "epoch": 0.3221133361738389, "grad_norm": 11.5625, "learning_rate": 2.350140956278791e-06, "loss": 0.4045, "step": 3780 }, { "epoch": 0.32219855134213893, "grad_norm": 12.5, "learning_rate": 2.3500611689908607e-06, "loss": 0.6438, "step": 3781 }, { "epoch": 0.3222837665104389, "grad_norm": 14.5625, "learning_rate": 2.349981361823719e-06, "loss": 0.8372, "step": 3782 }, { "epoch": 0.32236898167873884, "grad_norm": 11.0, "learning_rate": 2.3499015347788083e-06, "loss": 0.6228, "step": 3783 }, { "epoch": 0.3224541968470388, "grad_norm": 10.5625, "learning_rate": 2.349821687857571e-06, "loss": 0.4362, "step": 3784 }, { "epoch": 0.32253941201533876, "grad_norm": 12.375, "learning_rate": 2.3497418210614494e-06, "loss": 0.6263, "step": 3785 }, { "epoch": 0.3226246271836387, "grad_norm": 14.0625, "learning_rate": 2.3496619343918873e-06, "loss": 0.5625, "step": 3786 }, { "epoch": 0.32270984235193867, "grad_norm": 14.75, "learning_rate": 2.3495820278503284e-06, "loss": 0.7221, "step": 3787 }, { "epoch": 0.3227950575202386, "grad_norm": 16.0, "learning_rate": 2.3495021014382163e-06, "loss": 0.8427, "step": 3788 }, { "epoch": 0.3228802726885386, "grad_norm": 19.0, "learning_rate": 2.3494221551569958e-06, "loss": 1.0603, "step": 3789 }, { "epoch": 0.32296548785683854, "grad_norm": 12.8125, "learning_rate": 2.3493421890081113e-06, "loss": 0.6749, "step": 3790 }, { "epoch": 0.3230507030251385, "grad_norm": 13.4375, "learning_rate": 2.3492622029930085e-06, "loss": 0.5493, "step": 3791 }, { "epoch": 0.32313591819343845, "grad_norm": 12.25, "learning_rate": 2.349182197113132e-06, "loss": 0.6549, "step": 3792 }, { "epoch": 0.3232211333617384, "grad_norm": 17.125, "learning_rate": 2.3491021713699273e-06, "loss": 1.0181, "step": 3793 }, { "epoch": 0.32330634853003837, "grad_norm": 19.5, "learning_rate": 2.3490221257648415e-06, "loss": 0.5747, "step": 3794 }, { "epoch": 0.3233915636983383, "grad_norm": 27.375, "learning_rate": 2.348942060299321e-06, "loss": 0.975, "step": 3795 }, { "epoch": 0.3234767788666383, "grad_norm": 11.5, "learning_rate": 2.348861974974812e-06, "loss": 0.4122, "step": 3796 }, { "epoch": 0.32356199403493824, "grad_norm": 15.5625, "learning_rate": 2.3487818697927623e-06, "loss": 1.05, "step": 3797 }, { "epoch": 0.3236472092032382, "grad_norm": 13.3125, "learning_rate": 2.3487017447546196e-06, "loss": 0.8526, "step": 3798 }, { "epoch": 0.32373242437153815, "grad_norm": 17.625, "learning_rate": 2.348621599861831e-06, "loss": 1.1766, "step": 3799 }, { "epoch": 0.3238176395398381, "grad_norm": 15.875, "learning_rate": 2.3485414351158454e-06, "loss": 0.5928, "step": 3800 }, { "epoch": 0.32390285470813807, "grad_norm": 14.5625, "learning_rate": 2.3484612505181117e-06, "loss": 0.888, "step": 3801 }, { "epoch": 0.323988069876438, "grad_norm": 14.0, "learning_rate": 2.3483810460700783e-06, "loss": 0.5663, "step": 3802 }, { "epoch": 0.324073285044738, "grad_norm": 13.1875, "learning_rate": 2.3483008217731953e-06, "loss": 0.4324, "step": 3803 }, { "epoch": 0.32415850021303794, "grad_norm": 12.25, "learning_rate": 2.348220577628912e-06, "loss": 0.5498, "step": 3804 }, { "epoch": 0.3242437153813379, "grad_norm": 28.25, "learning_rate": 2.348140313638678e-06, "loss": 1.1247, "step": 3805 }, { "epoch": 0.32432893054963785, "grad_norm": 17.875, "learning_rate": 2.3480600298039446e-06, "loss": 0.9968, "step": 3806 }, { "epoch": 0.3244141457179378, "grad_norm": 14.375, "learning_rate": 2.3479797261261623e-06, "loss": 0.5682, "step": 3807 }, { "epoch": 0.32449936088623776, "grad_norm": 19.5, "learning_rate": 2.3478994026067824e-06, "loss": 0.9215, "step": 3808 }, { "epoch": 0.3245845760545377, "grad_norm": 15.3125, "learning_rate": 2.347819059247256e-06, "loss": 0.9493, "step": 3809 }, { "epoch": 0.3246697912228377, "grad_norm": 13.6875, "learning_rate": 2.3477386960490358e-06, "loss": 0.4373, "step": 3810 }, { "epoch": 0.32475500639113764, "grad_norm": 20.375, "learning_rate": 2.3476583130135727e-06, "loss": 0.9943, "step": 3811 }, { "epoch": 0.3248402215594376, "grad_norm": 13.4375, "learning_rate": 2.3475779101423205e-06, "loss": 0.6435, "step": 3812 }, { "epoch": 0.32492543672773755, "grad_norm": 13.125, "learning_rate": 2.3474974874367317e-06, "loss": 0.7465, "step": 3813 }, { "epoch": 0.3250106518960375, "grad_norm": 14.0, "learning_rate": 2.3474170448982596e-06, "loss": 0.8447, "step": 3814 }, { "epoch": 0.32509586706433746, "grad_norm": 16.125, "learning_rate": 2.3473365825283583e-06, "loss": 1.0348, "step": 3815 }, { "epoch": 0.3251810822326374, "grad_norm": 11.5625, "learning_rate": 2.347256100328481e-06, "loss": 0.4889, "step": 3816 }, { "epoch": 0.3252662974009374, "grad_norm": 12.1875, "learning_rate": 2.3471755983000825e-06, "loss": 0.6483, "step": 3817 }, { "epoch": 0.32535151256923733, "grad_norm": 15.8125, "learning_rate": 2.347095076444618e-06, "loss": 0.5665, "step": 3818 }, { "epoch": 0.3254367277375373, "grad_norm": 16.75, "learning_rate": 2.3470145347635427e-06, "loss": 0.8475, "step": 3819 }, { "epoch": 0.32552194290583725, "grad_norm": 13.8125, "learning_rate": 2.346933973258311e-06, "loss": 0.908, "step": 3820 }, { "epoch": 0.3256071580741372, "grad_norm": 14.6875, "learning_rate": 2.3468533919303797e-06, "loss": 0.6939, "step": 3821 }, { "epoch": 0.32569237324243716, "grad_norm": 16.25, "learning_rate": 2.3467727907812044e-06, "loss": 1.059, "step": 3822 }, { "epoch": 0.3257775884107371, "grad_norm": 12.875, "learning_rate": 2.346692169812242e-06, "loss": 0.7456, "step": 3823 }, { "epoch": 0.3258628035790371, "grad_norm": 13.25, "learning_rate": 2.3466115290249493e-06, "loss": 0.6671, "step": 3824 }, { "epoch": 0.32594801874733703, "grad_norm": 14.0625, "learning_rate": 2.346530868420783e-06, "loss": 0.7496, "step": 3825 }, { "epoch": 0.326033233915637, "grad_norm": 14.75, "learning_rate": 2.346450188001202e-06, "loss": 0.7022, "step": 3826 }, { "epoch": 0.32611844908393695, "grad_norm": 16.75, "learning_rate": 2.346369487767663e-06, "loss": 0.7458, "step": 3827 }, { "epoch": 0.3262036642522369, "grad_norm": 18.0, "learning_rate": 2.3462887677216254e-06, "loss": 0.9712, "step": 3828 }, { "epoch": 0.32628887942053686, "grad_norm": 13.625, "learning_rate": 2.346208027864547e-06, "loss": 0.667, "step": 3829 }, { "epoch": 0.3263740945888368, "grad_norm": 15.6875, "learning_rate": 2.3461272681978873e-06, "loss": 0.9095, "step": 3830 }, { "epoch": 0.3264593097571368, "grad_norm": 14.6875, "learning_rate": 2.346046488723106e-06, "loss": 0.7896, "step": 3831 }, { "epoch": 0.32654452492543673, "grad_norm": 12.5, "learning_rate": 2.3459656894416616e-06, "loss": 0.5909, "step": 3832 }, { "epoch": 0.3266297400937367, "grad_norm": 15.75, "learning_rate": 2.345884870355016e-06, "loss": 0.6627, "step": 3833 }, { "epoch": 0.32671495526203664, "grad_norm": 11.9375, "learning_rate": 2.345804031464628e-06, "loss": 0.4878, "step": 3834 }, { "epoch": 0.3268001704303366, "grad_norm": 17.75, "learning_rate": 2.34572317277196e-06, "loss": 1.0159, "step": 3835 }, { "epoch": 0.32688538559863656, "grad_norm": 18.125, "learning_rate": 2.3456422942784718e-06, "loss": 0.7609, "step": 3836 }, { "epoch": 0.3269706007669365, "grad_norm": 10.6875, "learning_rate": 2.345561395985626e-06, "loss": 0.5056, "step": 3837 }, { "epoch": 0.32705581593523647, "grad_norm": 11.4375, "learning_rate": 2.345480477894884e-06, "loss": 0.5038, "step": 3838 }, { "epoch": 0.32714103110353643, "grad_norm": 10.25, "learning_rate": 2.3453995400077085e-06, "loss": 0.4173, "step": 3839 }, { "epoch": 0.3272262462718364, "grad_norm": 27.25, "learning_rate": 2.345318582325561e-06, "loss": 1.0164, "step": 3840 }, { "epoch": 0.32731146144013634, "grad_norm": 12.125, "learning_rate": 2.345237604849906e-06, "loss": 0.5116, "step": 3841 }, { "epoch": 0.3273966766084363, "grad_norm": 11.375, "learning_rate": 2.345156607582206e-06, "loss": 0.3717, "step": 3842 }, { "epoch": 0.32748189177673626, "grad_norm": 13.0625, "learning_rate": 2.3450755905239246e-06, "loss": 0.555, "step": 3843 }, { "epoch": 0.3275671069450362, "grad_norm": 13.625, "learning_rate": 2.344994553676526e-06, "loss": 0.5815, "step": 3844 }, { "epoch": 0.32765232211333617, "grad_norm": 13.3125, "learning_rate": 2.344913497041475e-06, "loss": 0.5783, "step": 3845 }, { "epoch": 0.3277375372816361, "grad_norm": 12.5, "learning_rate": 2.344832420620236e-06, "loss": 0.6687, "step": 3846 }, { "epoch": 0.3278227524499361, "grad_norm": 19.5, "learning_rate": 2.3447513244142745e-06, "loss": 0.8924, "step": 3847 }, { "epoch": 0.32790796761823604, "grad_norm": 14.0625, "learning_rate": 2.3446702084250552e-06, "loss": 0.9233, "step": 3848 }, { "epoch": 0.327993182786536, "grad_norm": 19.125, "learning_rate": 2.344589072654045e-06, "loss": 0.9593, "step": 3849 }, { "epoch": 0.32807839795483595, "grad_norm": 12.1875, "learning_rate": 2.344507917102709e-06, "loss": 0.5729, "step": 3850 }, { "epoch": 0.3281636131231359, "grad_norm": 17.25, "learning_rate": 2.3444267417725147e-06, "loss": 1.1154, "step": 3851 }, { "epoch": 0.32824882829143587, "grad_norm": 20.75, "learning_rate": 2.3443455466649285e-06, "loss": 0.7493, "step": 3852 }, { "epoch": 0.3283340434597358, "grad_norm": 25.875, "learning_rate": 2.344264331781418e-06, "loss": 0.8772, "step": 3853 }, { "epoch": 0.3284192586280358, "grad_norm": 29.125, "learning_rate": 2.3441830971234503e-06, "loss": 1.1599, "step": 3854 }, { "epoch": 0.32850447379633574, "grad_norm": 12.875, "learning_rate": 2.3441018426924937e-06, "loss": 0.5125, "step": 3855 }, { "epoch": 0.3285896889646357, "grad_norm": 19.375, "learning_rate": 2.344020568490017e-06, "loss": 1.3051, "step": 3856 }, { "epoch": 0.32867490413293565, "grad_norm": 18.5, "learning_rate": 2.3439392745174885e-06, "loss": 0.6417, "step": 3857 }, { "epoch": 0.3287601193012356, "grad_norm": 17.125, "learning_rate": 2.3438579607763774e-06, "loss": 1.1153, "step": 3858 }, { "epoch": 0.32884533446953557, "grad_norm": 19.25, "learning_rate": 2.3437766272681523e-06, "loss": 1.2917, "step": 3859 }, { "epoch": 0.3289305496378355, "grad_norm": 14.1875, "learning_rate": 2.3436952739942843e-06, "loss": 0.7366, "step": 3860 }, { "epoch": 0.3290157648061355, "grad_norm": 17.0, "learning_rate": 2.3436139009562425e-06, "loss": 0.9502, "step": 3861 }, { "epoch": 0.32910097997443544, "grad_norm": 19.25, "learning_rate": 2.343532508155498e-06, "loss": 0.8141, "step": 3862 }, { "epoch": 0.3291861951427354, "grad_norm": 12.4375, "learning_rate": 2.3434510955935217e-06, "loss": 0.5179, "step": 3863 }, { "epoch": 0.32927141031103535, "grad_norm": 11.0625, "learning_rate": 2.343369663271784e-06, "loss": 0.5175, "step": 3864 }, { "epoch": 0.3293566254793353, "grad_norm": 21.5, "learning_rate": 2.3432882111917574e-06, "loss": 0.9953, "step": 3865 }, { "epoch": 0.32944184064763526, "grad_norm": 12.6875, "learning_rate": 2.3432067393549134e-06, "loss": 0.5388, "step": 3866 }, { "epoch": 0.3295270558159352, "grad_norm": 14.3125, "learning_rate": 2.3431252477627242e-06, "loss": 0.7496, "step": 3867 }, { "epoch": 0.3296122709842352, "grad_norm": 13.0, "learning_rate": 2.3430437364166624e-06, "loss": 0.4234, "step": 3868 }, { "epoch": 0.32969748615253514, "grad_norm": 12.5625, "learning_rate": 2.3429622053182018e-06, "loss": 0.6817, "step": 3869 }, { "epoch": 0.3297827013208351, "grad_norm": 17.0, "learning_rate": 2.3428806544688145e-06, "loss": 0.8044, "step": 3870 }, { "epoch": 0.32986791648913505, "grad_norm": 15.0625, "learning_rate": 2.342799083869975e-06, "loss": 0.8683, "step": 3871 }, { "epoch": 0.329953131657435, "grad_norm": 15.625, "learning_rate": 2.342717493523157e-06, "loss": 1.0609, "step": 3872 }, { "epoch": 0.33003834682573496, "grad_norm": 14.375, "learning_rate": 2.342635883429835e-06, "loss": 0.5665, "step": 3873 }, { "epoch": 0.3301235619940349, "grad_norm": 11.5625, "learning_rate": 2.342554253591484e-06, "loss": 0.5236, "step": 3874 }, { "epoch": 0.3302087771623349, "grad_norm": 15.75, "learning_rate": 2.3424726040095793e-06, "loss": 0.7493, "step": 3875 }, { "epoch": 0.33029399233063483, "grad_norm": 24.625, "learning_rate": 2.342390934685596e-06, "loss": 0.9125, "step": 3876 }, { "epoch": 0.3303792074989348, "grad_norm": 10.5625, "learning_rate": 2.34230924562101e-06, "loss": 0.368, "step": 3877 }, { "epoch": 0.33046442266723475, "grad_norm": 11.9375, "learning_rate": 2.342227536817297e-06, "loss": 0.5587, "step": 3878 }, { "epoch": 0.3305496378355347, "grad_norm": 15.1875, "learning_rate": 2.3421458082759347e-06, "loss": 0.4354, "step": 3879 }, { "epoch": 0.33063485300383466, "grad_norm": 18.125, "learning_rate": 2.342064059998399e-06, "loss": 0.7428, "step": 3880 }, { "epoch": 0.3307200681721346, "grad_norm": 13.875, "learning_rate": 2.341982291986168e-06, "loss": 0.6956, "step": 3881 }, { "epoch": 0.3308052833404346, "grad_norm": 16.75, "learning_rate": 2.3419005042407184e-06, "loss": 1.0907, "step": 3882 }, { "epoch": 0.33089049850873453, "grad_norm": 14.1875, "learning_rate": 2.341818696763529e-06, "loss": 0.8007, "step": 3883 }, { "epoch": 0.3309757136770345, "grad_norm": 19.0, "learning_rate": 2.3417368695560777e-06, "loss": 1.1786, "step": 3884 }, { "epoch": 0.33106092884533445, "grad_norm": 16.875, "learning_rate": 2.3416550226198435e-06, "loss": 0.6097, "step": 3885 }, { "epoch": 0.3311461440136344, "grad_norm": 25.375, "learning_rate": 2.341573155956305e-06, "loss": 0.7952, "step": 3886 }, { "epoch": 0.33123135918193436, "grad_norm": 13.3125, "learning_rate": 2.3414912695669424e-06, "loss": 0.6954, "step": 3887 }, { "epoch": 0.3313165743502343, "grad_norm": 12.9375, "learning_rate": 2.3414093634532345e-06, "loss": 0.619, "step": 3888 }, { "epoch": 0.3314017895185343, "grad_norm": 19.875, "learning_rate": 2.3413274376166623e-06, "loss": 0.7765, "step": 3889 }, { "epoch": 0.33148700468683423, "grad_norm": 16.25, "learning_rate": 2.3412454920587057e-06, "loss": 1.0166, "step": 3890 }, { "epoch": 0.3315722198551342, "grad_norm": 15.0625, "learning_rate": 2.341163526780846e-06, "loss": 0.9749, "step": 3891 }, { "epoch": 0.33165743502343414, "grad_norm": 13.6875, "learning_rate": 2.341081541784563e-06, "loss": 0.8066, "step": 3892 }, { "epoch": 0.33174265019173416, "grad_norm": 15.8125, "learning_rate": 2.3409995370713405e-06, "loss": 0.5688, "step": 3893 }, { "epoch": 0.3318278653600341, "grad_norm": 12.375, "learning_rate": 2.340917512642659e-06, "loss": 0.8135, "step": 3894 }, { "epoch": 0.33191308052833407, "grad_norm": 13.5625, "learning_rate": 2.340835468500001e-06, "loss": 0.6521, "step": 3895 }, { "epoch": 0.331998295696634, "grad_norm": 17.75, "learning_rate": 2.340753404644849e-06, "loss": 0.8416, "step": 3896 }, { "epoch": 0.332083510864934, "grad_norm": 10.375, "learning_rate": 2.3406713210786867e-06, "loss": 0.3349, "step": 3897 }, { "epoch": 0.33216872603323394, "grad_norm": 17.375, "learning_rate": 2.3405892178029964e-06, "loss": 1.0132, "step": 3898 }, { "epoch": 0.3322539412015339, "grad_norm": 18.625, "learning_rate": 2.3405070948192625e-06, "loss": 1.0266, "step": 3899 }, { "epoch": 0.33233915636983385, "grad_norm": 15.875, "learning_rate": 2.340424952128969e-06, "loss": 1.0015, "step": 3900 }, { "epoch": 0.3324243715381338, "grad_norm": 16.25, "learning_rate": 2.3403427897335995e-06, "loss": 0.7218, "step": 3901 }, { "epoch": 0.33250958670643377, "grad_norm": 16.75, "learning_rate": 2.34026060763464e-06, "loss": 0.7724, "step": 3902 }, { "epoch": 0.3325948018747337, "grad_norm": 16.75, "learning_rate": 2.3401784058335744e-06, "loss": 0.7607, "step": 3903 }, { "epoch": 0.3326800170430337, "grad_norm": 12.375, "learning_rate": 2.340096184331889e-06, "loss": 0.7528, "step": 3904 }, { "epoch": 0.33276523221133364, "grad_norm": 21.0, "learning_rate": 2.3400139431310694e-06, "loss": 0.6997, "step": 3905 }, { "epoch": 0.3328504473796336, "grad_norm": 11.25, "learning_rate": 2.3399316822326018e-06, "loss": 0.6647, "step": 3906 }, { "epoch": 0.33293566254793355, "grad_norm": 13.3125, "learning_rate": 2.3398494016379725e-06, "loss": 0.4979, "step": 3907 }, { "epoch": 0.3330208777162335, "grad_norm": 18.125, "learning_rate": 2.339767101348669e-06, "loss": 0.8077, "step": 3908 }, { "epoch": 0.33310609288453347, "grad_norm": 19.5, "learning_rate": 2.339684781366178e-06, "loss": 0.9445, "step": 3909 }, { "epoch": 0.3331913080528334, "grad_norm": 14.3125, "learning_rate": 2.3396024416919864e-06, "loss": 0.8323, "step": 3910 }, { "epoch": 0.3332765232211334, "grad_norm": 17.625, "learning_rate": 2.339520082327584e-06, "loss": 0.9099, "step": 3911 }, { "epoch": 0.33336173838943334, "grad_norm": 17.25, "learning_rate": 2.339437703274458e-06, "loss": 0.6971, "step": 3912 }, { "epoch": 0.3334469535577333, "grad_norm": 16.25, "learning_rate": 2.3393553045340967e-06, "loss": 0.9556, "step": 3913 }, { "epoch": 0.33353216872603325, "grad_norm": 12.0, "learning_rate": 2.33927288610799e-06, "loss": 0.5052, "step": 3914 }, { "epoch": 0.3336173838943332, "grad_norm": 14.4375, "learning_rate": 2.3391904479976267e-06, "loss": 0.8101, "step": 3915 }, { "epoch": 0.33370259906263317, "grad_norm": 13.5, "learning_rate": 2.3391079902044968e-06, "loss": 0.6602, "step": 3916 }, { "epoch": 0.3337878142309331, "grad_norm": 17.125, "learning_rate": 2.3390255127300905e-06, "loss": 0.96, "step": 3917 }, { "epoch": 0.3338730293992331, "grad_norm": 14.5, "learning_rate": 2.338943015575898e-06, "loss": 0.9262, "step": 3918 }, { "epoch": 0.33395824456753304, "grad_norm": 14.4375, "learning_rate": 2.3388604987434097e-06, "loss": 0.6437, "step": 3919 }, { "epoch": 0.334043459735833, "grad_norm": 13.0, "learning_rate": 2.338777962234118e-06, "loss": 0.8867, "step": 3920 }, { "epoch": 0.33412867490413295, "grad_norm": 12.625, "learning_rate": 2.338695406049513e-06, "loss": 0.4617, "step": 3921 }, { "epoch": 0.3342138900724329, "grad_norm": 12.9375, "learning_rate": 2.338612830191088e-06, "loss": 0.5451, "step": 3922 }, { "epoch": 0.33429910524073286, "grad_norm": 13.625, "learning_rate": 2.338530234660334e-06, "loss": 0.2578, "step": 3923 }, { "epoch": 0.3343843204090328, "grad_norm": 28.25, "learning_rate": 2.338447619458744e-06, "loss": 1.023, "step": 3924 }, { "epoch": 0.3344695355773328, "grad_norm": 22.5, "learning_rate": 2.3383649845878116e-06, "loss": 0.9503, "step": 3925 }, { "epoch": 0.33455475074563273, "grad_norm": 25.0, "learning_rate": 2.3382823300490286e-06, "loss": 1.301, "step": 3926 }, { "epoch": 0.3346399659139327, "grad_norm": 13.9375, "learning_rate": 2.33819965584389e-06, "loss": 0.5245, "step": 3927 }, { "epoch": 0.33472518108223265, "grad_norm": 24.375, "learning_rate": 2.33811696197389e-06, "loss": 0.4129, "step": 3928 }, { "epoch": 0.3348103962505326, "grad_norm": 13.0625, "learning_rate": 2.338034248440521e-06, "loss": 0.4974, "step": 3929 }, { "epoch": 0.33489561141883256, "grad_norm": 15.75, "learning_rate": 2.33795151524528e-06, "loss": 0.7743, "step": 3930 }, { "epoch": 0.3349808265871325, "grad_norm": 21.25, "learning_rate": 2.3378687623896606e-06, "loss": 0.9439, "step": 3931 }, { "epoch": 0.3350660417554325, "grad_norm": 11.75, "learning_rate": 2.337785989875159e-06, "loss": 0.4533, "step": 3932 }, { "epoch": 0.33515125692373243, "grad_norm": 15.5, "learning_rate": 2.3377031977032704e-06, "loss": 0.994, "step": 3933 }, { "epoch": 0.3352364720920324, "grad_norm": 12.25, "learning_rate": 2.3376203858754916e-06, "loss": 0.507, "step": 3934 }, { "epoch": 0.33532168726033235, "grad_norm": 17.875, "learning_rate": 2.3375375543933186e-06, "loss": 0.4147, "step": 3935 }, { "epoch": 0.3354069024286323, "grad_norm": 26.875, "learning_rate": 2.3374547032582482e-06, "loss": 1.3156, "step": 3936 }, { "epoch": 0.33549211759693226, "grad_norm": 12.125, "learning_rate": 2.337371832471778e-06, "loss": 0.6413, "step": 3937 }, { "epoch": 0.3355773327652322, "grad_norm": 13.1875, "learning_rate": 2.3372889420354055e-06, "loss": 0.6281, "step": 3938 }, { "epoch": 0.3356625479335322, "grad_norm": 17.25, "learning_rate": 2.3372060319506283e-06, "loss": 0.8928, "step": 3939 }, { "epoch": 0.33574776310183213, "grad_norm": 12.625, "learning_rate": 2.3371231022189447e-06, "loss": 0.7688, "step": 3940 }, { "epoch": 0.3358329782701321, "grad_norm": 14.6875, "learning_rate": 2.3370401528418535e-06, "loss": 0.7262, "step": 3941 }, { "epoch": 0.33591819343843204, "grad_norm": 17.125, "learning_rate": 2.3369571838208537e-06, "loss": 0.5853, "step": 3942 }, { "epoch": 0.336003408606732, "grad_norm": 16.0, "learning_rate": 2.336874195157445e-06, "loss": 0.5609, "step": 3943 }, { "epoch": 0.33608862377503196, "grad_norm": 10.4375, "learning_rate": 2.336791186853126e-06, "loss": 0.3731, "step": 3944 }, { "epoch": 0.3361738389433319, "grad_norm": 12.75, "learning_rate": 2.3367081589093976e-06, "loss": 0.8264, "step": 3945 }, { "epoch": 0.33625905411163187, "grad_norm": 10.8125, "learning_rate": 2.3366251113277597e-06, "loss": 0.4323, "step": 3946 }, { "epoch": 0.33634426927993183, "grad_norm": 16.0, "learning_rate": 2.336542044109714e-06, "loss": 0.5537, "step": 3947 }, { "epoch": 0.3364294844482318, "grad_norm": 20.125, "learning_rate": 2.3364589572567608e-06, "loss": 1.0855, "step": 3948 }, { "epoch": 0.33651469961653174, "grad_norm": 11.875, "learning_rate": 2.3363758507704015e-06, "loss": 0.679, "step": 3949 }, { "epoch": 0.3365999147848317, "grad_norm": 18.0, "learning_rate": 2.336292724652138e-06, "loss": 0.8954, "step": 3950 }, { "epoch": 0.33668512995313166, "grad_norm": 15.0, "learning_rate": 2.336209578903473e-06, "loss": 0.9358, "step": 3951 }, { "epoch": 0.3367703451214316, "grad_norm": 10.25, "learning_rate": 2.3361264135259083e-06, "loss": 0.3877, "step": 3952 }, { "epoch": 0.33685556028973157, "grad_norm": 16.25, "learning_rate": 2.336043228520947e-06, "loss": 0.6617, "step": 3953 }, { "epoch": 0.3369407754580315, "grad_norm": 17.75, "learning_rate": 2.3359600238900932e-06, "loss": 1.2946, "step": 3954 }, { "epoch": 0.3370259906263315, "grad_norm": 12.5625, "learning_rate": 2.3358767996348492e-06, "loss": 0.2643, "step": 3955 }, { "epoch": 0.33711120579463144, "grad_norm": 14.6875, "learning_rate": 2.3357935557567195e-06, "loss": 0.8753, "step": 3956 }, { "epoch": 0.3371964209629314, "grad_norm": 20.875, "learning_rate": 2.3357102922572083e-06, "loss": 0.9863, "step": 3957 }, { "epoch": 0.33728163613123135, "grad_norm": 11.3125, "learning_rate": 2.335627009137821e-06, "loss": 0.4275, "step": 3958 }, { "epoch": 0.3373668512995313, "grad_norm": 15.125, "learning_rate": 2.335543706400061e-06, "loss": 0.7731, "step": 3959 }, { "epoch": 0.33745206646783127, "grad_norm": 12.8125, "learning_rate": 2.3354603840454355e-06, "loss": 0.6382, "step": 3960 }, { "epoch": 0.3375372816361312, "grad_norm": 20.25, "learning_rate": 2.335377042075449e-06, "loss": 1.0024, "step": 3961 }, { "epoch": 0.3376224968044312, "grad_norm": 12.875, "learning_rate": 2.3352936804916076e-06, "loss": 0.6543, "step": 3962 }, { "epoch": 0.33770771197273114, "grad_norm": 14.625, "learning_rate": 2.3352102992954183e-06, "loss": 0.9571, "step": 3963 }, { "epoch": 0.3377929271410311, "grad_norm": 16.125, "learning_rate": 2.335126898488388e-06, "loss": 1.1574, "step": 3964 }, { "epoch": 0.33787814230933105, "grad_norm": 17.875, "learning_rate": 2.335043478072023e-06, "loss": 0.7375, "step": 3965 }, { "epoch": 0.337963357477631, "grad_norm": 12.6875, "learning_rate": 2.3349600380478317e-06, "loss": 0.486, "step": 3966 }, { "epoch": 0.33804857264593097, "grad_norm": 11.4375, "learning_rate": 2.334876578417321e-06, "loss": 0.3549, "step": 3967 }, { "epoch": 0.3381337878142309, "grad_norm": 17.5, "learning_rate": 2.3347930991819996e-06, "loss": 0.7313, "step": 3968 }, { "epoch": 0.3382190029825309, "grad_norm": 16.125, "learning_rate": 2.334709600343376e-06, "loss": 0.911, "step": 3969 }, { "epoch": 0.33830421815083084, "grad_norm": 16.75, "learning_rate": 2.3346260819029597e-06, "loss": 0.797, "step": 3970 }, { "epoch": 0.3383894333191308, "grad_norm": 8.6875, "learning_rate": 2.3345425438622593e-06, "loss": 0.2527, "step": 3971 }, { "epoch": 0.33847464848743075, "grad_norm": 13.0625, "learning_rate": 2.3344589862227844e-06, "loss": 0.5059, "step": 3972 }, { "epoch": 0.3385598636557307, "grad_norm": 13.625, "learning_rate": 2.334375408986045e-06, "loss": 0.8461, "step": 3973 }, { "epoch": 0.33864507882403067, "grad_norm": 25.5, "learning_rate": 2.3342918121535516e-06, "loss": 1.3825, "step": 3974 }, { "epoch": 0.3387302939923306, "grad_norm": 15.4375, "learning_rate": 2.3342081957268146e-06, "loss": 0.9739, "step": 3975 }, { "epoch": 0.3388155091606306, "grad_norm": 22.125, "learning_rate": 2.334124559707346e-06, "loss": 0.7812, "step": 3976 }, { "epoch": 0.33890072432893054, "grad_norm": 14.1875, "learning_rate": 2.3340409040966555e-06, "loss": 0.8652, "step": 3977 }, { "epoch": 0.3389859394972305, "grad_norm": 30.875, "learning_rate": 2.3339572288962564e-06, "loss": 1.031, "step": 3978 }, { "epoch": 0.33907115466553045, "grad_norm": 12.375, "learning_rate": 2.3338735341076597e-06, "loss": 0.7422, "step": 3979 }, { "epoch": 0.3391563698338304, "grad_norm": 18.75, "learning_rate": 2.333789819732379e-06, "loss": 0.711, "step": 3980 }, { "epoch": 0.33924158500213036, "grad_norm": 24.25, "learning_rate": 2.333706085771926e-06, "loss": 1.0343, "step": 3981 }, { "epoch": 0.3393268001704303, "grad_norm": 11.4375, "learning_rate": 2.333622332227814e-06, "loss": 0.6232, "step": 3982 }, { "epoch": 0.3394120153387303, "grad_norm": 18.25, "learning_rate": 2.333538559101557e-06, "loss": 0.7608, "step": 3983 }, { "epoch": 0.33949723050703023, "grad_norm": 20.0, "learning_rate": 2.333454766394669e-06, "loss": 0.8179, "step": 3984 }, { "epoch": 0.3395824456753302, "grad_norm": 21.875, "learning_rate": 2.3333709541086636e-06, "loss": 1.131, "step": 3985 }, { "epoch": 0.33966766084363015, "grad_norm": 18.25, "learning_rate": 2.3332871222450554e-06, "loss": 0.9852, "step": 3986 }, { "epoch": 0.3397528760119301, "grad_norm": 19.75, "learning_rate": 2.33320327080536e-06, "loss": 0.8217, "step": 3987 }, { "epoch": 0.33983809118023006, "grad_norm": 14.375, "learning_rate": 2.333119399791092e-06, "loss": 0.4423, "step": 3988 }, { "epoch": 0.33992330634853, "grad_norm": 14.125, "learning_rate": 2.3330355092037672e-06, "loss": 0.6775, "step": 3989 }, { "epoch": 0.34000852151683, "grad_norm": 18.25, "learning_rate": 2.332951599044902e-06, "loss": 0.6429, "step": 3990 }, { "epoch": 0.34009373668512993, "grad_norm": 13.625, "learning_rate": 2.3328676693160124e-06, "loss": 0.7431, "step": 3991 }, { "epoch": 0.3401789518534299, "grad_norm": 15.75, "learning_rate": 2.332783720018615e-06, "loss": 0.6359, "step": 3992 }, { "epoch": 0.34026416702172985, "grad_norm": 15.3125, "learning_rate": 2.3326997511542267e-06, "loss": 1.0031, "step": 3993 }, { "epoch": 0.3403493821900298, "grad_norm": 14.1875, "learning_rate": 2.3326157627243653e-06, "loss": 0.5978, "step": 3994 }, { "epoch": 0.34043459735832976, "grad_norm": 12.9375, "learning_rate": 2.332531754730549e-06, "loss": 0.5315, "step": 3995 }, { "epoch": 0.3405198125266297, "grad_norm": 13.875, "learning_rate": 2.3324477271742943e-06, "loss": 0.6428, "step": 3996 }, { "epoch": 0.3406050276949297, "grad_norm": 17.125, "learning_rate": 2.3323636800571212e-06, "loss": 0.6487, "step": 3997 }, { "epoch": 0.34069024286322963, "grad_norm": 12.8125, "learning_rate": 2.332279613380548e-06, "loss": 0.8349, "step": 3998 }, { "epoch": 0.3407754580315296, "grad_norm": 15.8125, "learning_rate": 2.332195527146094e-06, "loss": 0.641, "step": 3999 }, { "epoch": 0.34086067319982954, "grad_norm": 13.4375, "learning_rate": 2.332111421355278e-06, "loss": 1.024, "step": 4000 }, { "epoch": 0.3409458883681295, "grad_norm": 19.625, "learning_rate": 2.3320272960096212e-06, "loss": 1.2106, "step": 4001 }, { "epoch": 0.34103110353642946, "grad_norm": 16.875, "learning_rate": 2.331943151110643e-06, "loss": 1.1761, "step": 4002 }, { "epoch": 0.3411163187047294, "grad_norm": 17.875, "learning_rate": 2.331858986659864e-06, "loss": 0.8235, "step": 4003 }, { "epoch": 0.3412015338730294, "grad_norm": 11.9375, "learning_rate": 2.331774802658805e-06, "loss": 0.4329, "step": 4004 }, { "epoch": 0.34128674904132933, "grad_norm": 13.625, "learning_rate": 2.3316905991089877e-06, "loss": 0.5224, "step": 4005 }, { "epoch": 0.34137196420962934, "grad_norm": 21.0, "learning_rate": 2.331606376011933e-06, "loss": 1.1005, "step": 4006 }, { "epoch": 0.3414571793779293, "grad_norm": 15.25, "learning_rate": 2.331522133369164e-06, "loss": 1.0837, "step": 4007 }, { "epoch": 0.34154239454622926, "grad_norm": 15.125, "learning_rate": 2.3314378711822024e-06, "loss": 0.9544, "step": 4008 }, { "epoch": 0.3416276097145292, "grad_norm": 16.0, "learning_rate": 2.3313535894525713e-06, "loss": 0.815, "step": 4009 }, { "epoch": 0.34171282488282917, "grad_norm": 13.0625, "learning_rate": 2.331269288181793e-06, "loss": 0.4748, "step": 4010 }, { "epoch": 0.3417980400511291, "grad_norm": 18.75, "learning_rate": 2.3311849673713913e-06, "loss": 1.3333, "step": 4011 }, { "epoch": 0.3418832552194291, "grad_norm": 14.0, "learning_rate": 2.3311006270228905e-06, "loss": 0.4996, "step": 4012 }, { "epoch": 0.34196847038772904, "grad_norm": 12.4375, "learning_rate": 2.331016267137814e-06, "loss": 0.5874, "step": 4013 }, { "epoch": 0.342053685556029, "grad_norm": 16.125, "learning_rate": 2.3309318877176865e-06, "loss": 0.9413, "step": 4014 }, { "epoch": 0.34213890072432895, "grad_norm": 25.375, "learning_rate": 2.3308474887640327e-06, "loss": 1.0336, "step": 4015 }, { "epoch": 0.3422241158926289, "grad_norm": 12.1875, "learning_rate": 2.330763070278378e-06, "loss": 0.5361, "step": 4016 }, { "epoch": 0.34230933106092887, "grad_norm": 17.375, "learning_rate": 2.330678632262248e-06, "loss": 0.3863, "step": 4017 }, { "epoch": 0.3423945462292288, "grad_norm": 12.0625, "learning_rate": 2.3305941747171684e-06, "loss": 0.7383, "step": 4018 }, { "epoch": 0.3424797613975288, "grad_norm": 18.0, "learning_rate": 2.3305096976446652e-06, "loss": 0.9696, "step": 4019 }, { "epoch": 0.34256497656582874, "grad_norm": 12.5, "learning_rate": 2.3304252010462654e-06, "loss": 0.5396, "step": 4020 }, { "epoch": 0.3426501917341287, "grad_norm": 14.3125, "learning_rate": 2.3303406849234956e-06, "loss": 0.6174, "step": 4021 }, { "epoch": 0.34273540690242865, "grad_norm": 20.875, "learning_rate": 2.330256149277883e-06, "loss": 1.2139, "step": 4022 }, { "epoch": 0.3428206220707286, "grad_norm": 20.25, "learning_rate": 2.330171594110956e-06, "loss": 0.904, "step": 4023 }, { "epoch": 0.34290583723902857, "grad_norm": 33.25, "learning_rate": 2.330087019424242e-06, "loss": 0.9877, "step": 4024 }, { "epoch": 0.3429910524073285, "grad_norm": 14.625, "learning_rate": 2.3300024252192692e-06, "loss": 0.6767, "step": 4025 }, { "epoch": 0.3430762675756285, "grad_norm": 14.4375, "learning_rate": 2.329917811497567e-06, "loss": 0.6389, "step": 4026 }, { "epoch": 0.34316148274392844, "grad_norm": 18.375, "learning_rate": 2.3298331782606638e-06, "loss": 0.7641, "step": 4027 }, { "epoch": 0.3432466979122284, "grad_norm": 17.0, "learning_rate": 2.3297485255100892e-06, "loss": 0.6028, "step": 4028 }, { "epoch": 0.34333191308052835, "grad_norm": 14.3125, "learning_rate": 2.329663853247373e-06, "loss": 0.4676, "step": 4029 }, { "epoch": 0.3434171282488283, "grad_norm": 24.25, "learning_rate": 2.3295791614740456e-06, "loss": 0.8223, "step": 4030 }, { "epoch": 0.34350234341712826, "grad_norm": 11.75, "learning_rate": 2.3294944501916364e-06, "loss": 0.4249, "step": 4031 }, { "epoch": 0.3435875585854282, "grad_norm": 14.5, "learning_rate": 2.3294097194016773e-06, "loss": 0.5445, "step": 4032 }, { "epoch": 0.3436727737537282, "grad_norm": 22.0, "learning_rate": 2.3293249691056993e-06, "loss": 1.0898, "step": 4033 }, { "epoch": 0.34375798892202813, "grad_norm": 23.625, "learning_rate": 2.3292401993052337e-06, "loss": 0.9155, "step": 4034 }, { "epoch": 0.3438432040903281, "grad_norm": 13.5, "learning_rate": 2.329155410001812e-06, "loss": 0.6164, "step": 4035 }, { "epoch": 0.34392841925862805, "grad_norm": 12.1875, "learning_rate": 2.3290706011969676e-06, "loss": 0.6982, "step": 4036 }, { "epoch": 0.344013634426928, "grad_norm": 14.8125, "learning_rate": 2.328985772892232e-06, "loss": 0.5167, "step": 4037 }, { "epoch": 0.34409884959522796, "grad_norm": 14.5, "learning_rate": 2.3289009250891385e-06, "loss": 1.0156, "step": 4038 }, { "epoch": 0.3441840647635279, "grad_norm": 16.0, "learning_rate": 2.32881605778922e-06, "loss": 0.8291, "step": 4039 }, { "epoch": 0.3442692799318279, "grad_norm": 13.8125, "learning_rate": 2.328731170994011e-06, "loss": 0.749, "step": 4040 }, { "epoch": 0.34435449510012783, "grad_norm": 13.6875, "learning_rate": 2.328646264705045e-06, "loss": 0.8931, "step": 4041 }, { "epoch": 0.3444397102684278, "grad_norm": 13.0, "learning_rate": 2.328561338923856e-06, "loss": 0.7366, "step": 4042 }, { "epoch": 0.34452492543672775, "grad_norm": 26.25, "learning_rate": 2.3284763936519795e-06, "loss": 0.9489, "step": 4043 }, { "epoch": 0.3446101406050277, "grad_norm": 11.3125, "learning_rate": 2.3283914288909498e-06, "loss": 0.4795, "step": 4044 }, { "epoch": 0.34469535577332766, "grad_norm": 14.25, "learning_rate": 2.3283064446423024e-06, "loss": 0.7264, "step": 4045 }, { "epoch": 0.3447805709416276, "grad_norm": 14.8125, "learning_rate": 2.328221440907573e-06, "loss": 0.7318, "step": 4046 }, { "epoch": 0.3448657861099276, "grad_norm": 16.375, "learning_rate": 2.3281364176882986e-06, "loss": 0.992, "step": 4047 }, { "epoch": 0.34495100127822753, "grad_norm": 24.25, "learning_rate": 2.3280513749860147e-06, "loss": 0.7895, "step": 4048 }, { "epoch": 0.3450362164465275, "grad_norm": 13.75, "learning_rate": 2.3279663128022585e-06, "loss": 0.7572, "step": 4049 }, { "epoch": 0.34512143161482745, "grad_norm": 17.375, "learning_rate": 2.327881231138567e-06, "loss": 1.0086, "step": 4050 }, { "epoch": 0.3452066467831274, "grad_norm": 14.9375, "learning_rate": 2.3277961299964773e-06, "loss": 1.0919, "step": 4051 }, { "epoch": 0.34529186195142736, "grad_norm": 14.1875, "learning_rate": 2.3277110093775283e-06, "loss": 0.9007, "step": 4052 }, { "epoch": 0.3453770771197273, "grad_norm": 13.4375, "learning_rate": 2.3276258692832577e-06, "loss": 0.8951, "step": 4053 }, { "epoch": 0.3454622922880273, "grad_norm": 25.75, "learning_rate": 2.3275407097152036e-06, "loss": 1.2109, "step": 4054 }, { "epoch": 0.34554750745632723, "grad_norm": 17.5, "learning_rate": 2.3274555306749053e-06, "loss": 0.9206, "step": 4055 }, { "epoch": 0.3456327226246272, "grad_norm": 14.4375, "learning_rate": 2.3273703321639025e-06, "loss": 0.7756, "step": 4056 }, { "epoch": 0.34571793779292714, "grad_norm": 16.125, "learning_rate": 2.3272851141837343e-06, "loss": 0.8646, "step": 4057 }, { "epoch": 0.3458031529612271, "grad_norm": 16.0, "learning_rate": 2.3271998767359407e-06, "loss": 1.0957, "step": 4058 }, { "epoch": 0.34588836812952706, "grad_norm": 15.0625, "learning_rate": 2.327114619822062e-06, "loss": 0.5483, "step": 4059 }, { "epoch": 0.345973583297827, "grad_norm": 16.125, "learning_rate": 2.3270293434436387e-06, "loss": 1.0238, "step": 4060 }, { "epoch": 0.34605879846612697, "grad_norm": 11.0625, "learning_rate": 2.3269440476022124e-06, "loss": 0.4951, "step": 4061 }, { "epoch": 0.34614401363442693, "grad_norm": 11.875, "learning_rate": 2.326858732299325e-06, "loss": 0.5912, "step": 4062 }, { "epoch": 0.3462292288027269, "grad_norm": 9.5625, "learning_rate": 2.3267733975365164e-06, "loss": 0.3455, "step": 4063 }, { "epoch": 0.34631444397102684, "grad_norm": 17.375, "learning_rate": 2.32668804331533e-06, "loss": 0.8845, "step": 4064 }, { "epoch": 0.3463996591393268, "grad_norm": 11.625, "learning_rate": 2.3266026696373086e-06, "loss": 0.9656, "step": 4065 }, { "epoch": 0.34648487430762676, "grad_norm": 15.1875, "learning_rate": 2.3265172765039936e-06, "loss": 0.8902, "step": 4066 }, { "epoch": 0.3465700894759267, "grad_norm": 12.3125, "learning_rate": 2.326431863916929e-06, "loss": 0.5413, "step": 4067 }, { "epoch": 0.34665530464422667, "grad_norm": 9.9375, "learning_rate": 2.3263464318776583e-06, "loss": 0.3425, "step": 4068 }, { "epoch": 0.3467405198125266, "grad_norm": 15.375, "learning_rate": 2.326260980387725e-06, "loss": 1.1196, "step": 4069 }, { "epoch": 0.3468257349808266, "grad_norm": 17.125, "learning_rate": 2.326175509448674e-06, "loss": 1.0365, "step": 4070 }, { "epoch": 0.34691095014912654, "grad_norm": 18.25, "learning_rate": 2.326090019062049e-06, "loss": 0.4944, "step": 4071 }, { "epoch": 0.3469961653174265, "grad_norm": 16.0, "learning_rate": 2.326004509229395e-06, "loss": 0.9812, "step": 4072 }, { "epoch": 0.34708138048572645, "grad_norm": 14.8125, "learning_rate": 2.325918979952258e-06, "loss": 0.7273, "step": 4073 }, { "epoch": 0.3471665956540264, "grad_norm": 12.75, "learning_rate": 2.3258334312321832e-06, "loss": 0.5361, "step": 4074 }, { "epoch": 0.34725181082232637, "grad_norm": 13.4375, "learning_rate": 2.3257478630707164e-06, "loss": 0.6556, "step": 4075 }, { "epoch": 0.3473370259906263, "grad_norm": 19.25, "learning_rate": 2.3256622754694037e-06, "loss": 0.6668, "step": 4076 }, { "epoch": 0.3474222411589263, "grad_norm": 20.75, "learning_rate": 2.3255766684297924e-06, "loss": 0.7155, "step": 4077 }, { "epoch": 0.34750745632722624, "grad_norm": 11.1875, "learning_rate": 2.3254910419534288e-06, "loss": 0.4792, "step": 4078 }, { "epoch": 0.3475926714955262, "grad_norm": 12.9375, "learning_rate": 2.3254053960418603e-06, "loss": 0.577, "step": 4079 }, { "epoch": 0.34767788666382615, "grad_norm": 15.5625, "learning_rate": 2.3253197306966357e-06, "loss": 0.7516, "step": 4080 }, { "epoch": 0.3477631018321261, "grad_norm": 21.875, "learning_rate": 2.3252340459193017e-06, "loss": 0.8775, "step": 4081 }, { "epoch": 0.34784831700042607, "grad_norm": 17.875, "learning_rate": 2.325148341711407e-06, "loss": 0.8316, "step": 4082 }, { "epoch": 0.347933532168726, "grad_norm": 10.4375, "learning_rate": 2.3250626180745013e-06, "loss": 0.3671, "step": 4083 }, { "epoch": 0.348018747337026, "grad_norm": 11.1875, "learning_rate": 2.3249768750101325e-06, "loss": 0.585, "step": 4084 }, { "epoch": 0.34810396250532594, "grad_norm": 20.0, "learning_rate": 2.3248911125198507e-06, "loss": 1.0553, "step": 4085 }, { "epoch": 0.3481891776736259, "grad_norm": 12.625, "learning_rate": 2.324805330605206e-06, "loss": 0.6835, "step": 4086 }, { "epoch": 0.34827439284192585, "grad_norm": 13.5625, "learning_rate": 2.3247195292677475e-06, "loss": 0.5109, "step": 4087 }, { "epoch": 0.3483596080102258, "grad_norm": 16.0, "learning_rate": 2.324633708509027e-06, "loss": 0.9071, "step": 4088 }, { "epoch": 0.34844482317852576, "grad_norm": 19.25, "learning_rate": 2.3245478683305948e-06, "loss": 0.853, "step": 4089 }, { "epoch": 0.3485300383468257, "grad_norm": 13.625, "learning_rate": 2.3244620087340017e-06, "loss": 0.844, "step": 4090 }, { "epoch": 0.3486152535151257, "grad_norm": 17.5, "learning_rate": 2.3243761297207995e-06, "loss": 0.8373, "step": 4091 }, { "epoch": 0.34870046868342564, "grad_norm": 16.0, "learning_rate": 2.3242902312925408e-06, "loss": 0.7101, "step": 4092 }, { "epoch": 0.3487856838517256, "grad_norm": 20.625, "learning_rate": 2.324204313450777e-06, "loss": 0.6203, "step": 4093 }, { "epoch": 0.34887089902002555, "grad_norm": 11.5625, "learning_rate": 2.324118376197061e-06, "loss": 0.6374, "step": 4094 }, { "epoch": 0.3489561141883255, "grad_norm": 14.1875, "learning_rate": 2.324032419532946e-06, "loss": 0.5917, "step": 4095 }, { "epoch": 0.34904132935662546, "grad_norm": 13.375, "learning_rate": 2.323946443459985e-06, "loss": 0.7285, "step": 4096 }, { "epoch": 0.3491265445249254, "grad_norm": 14.3125, "learning_rate": 2.323860447979732e-06, "loss": 0.7153, "step": 4097 }, { "epoch": 0.3492117596932254, "grad_norm": 14.0625, "learning_rate": 2.3237744330937407e-06, "loss": 0.5982, "step": 4098 }, { "epoch": 0.34929697486152533, "grad_norm": 9.5, "learning_rate": 2.323688398803566e-06, "loss": 0.5317, "step": 4099 }, { "epoch": 0.3493821900298253, "grad_norm": 14.125, "learning_rate": 2.323602345110762e-06, "loss": 0.6596, "step": 4100 }, { "epoch": 0.34946740519812525, "grad_norm": 14.0, "learning_rate": 2.323516272016884e-06, "loss": 0.6538, "step": 4101 }, { "epoch": 0.3495526203664252, "grad_norm": 21.125, "learning_rate": 2.3234301795234875e-06, "loss": 0.5747, "step": 4102 }, { "epoch": 0.34963783553472516, "grad_norm": 13.375, "learning_rate": 2.3233440676321285e-06, "loss": 0.7308, "step": 4103 }, { "epoch": 0.3497230507030251, "grad_norm": 13.125, "learning_rate": 2.3232579363443627e-06, "loss": 0.6837, "step": 4104 }, { "epoch": 0.3498082658713251, "grad_norm": 15.875, "learning_rate": 2.3231717856617463e-06, "loss": 0.4639, "step": 4105 }, { "epoch": 0.34989348103962503, "grad_norm": 17.625, "learning_rate": 2.3230856155858366e-06, "loss": 0.5849, "step": 4106 }, { "epoch": 0.349978696207925, "grad_norm": 11.9375, "learning_rate": 2.322999426118191e-06, "loss": 0.5821, "step": 4107 }, { "epoch": 0.35006391137622495, "grad_norm": 10.1875, "learning_rate": 2.3229132172603667e-06, "loss": 0.3634, "step": 4108 }, { "epoch": 0.3501491265445249, "grad_norm": 14.875, "learning_rate": 2.3228269890139222e-06, "loss": 0.8087, "step": 4109 }, { "epoch": 0.35023434171282486, "grad_norm": 16.75, "learning_rate": 2.322740741380415e-06, "loss": 0.8326, "step": 4110 }, { "epoch": 0.3503195568811248, "grad_norm": 18.25, "learning_rate": 2.3226544743614036e-06, "loss": 0.8339, "step": 4111 }, { "epoch": 0.3504047720494248, "grad_norm": 11.4375, "learning_rate": 2.3225681879584473e-06, "loss": 0.6817, "step": 4112 }, { "epoch": 0.35048998721772473, "grad_norm": 18.125, "learning_rate": 2.3224818821731055e-06, "loss": 1.0308, "step": 4113 }, { "epoch": 0.3505752023860247, "grad_norm": 14.375, "learning_rate": 2.3223955570069375e-06, "loss": 0.8212, "step": 4114 }, { "epoch": 0.35066041755432464, "grad_norm": 14.0625, "learning_rate": 2.322309212461503e-06, "loss": 0.8081, "step": 4115 }, { "epoch": 0.3507456327226246, "grad_norm": 15.375, "learning_rate": 2.322222848538363e-06, "loss": 0.7543, "step": 4116 }, { "epoch": 0.35083084789092456, "grad_norm": 14.4375, "learning_rate": 2.3221364652390782e-06, "loss": 0.7831, "step": 4117 }, { "epoch": 0.3509160630592245, "grad_norm": 11.6875, "learning_rate": 2.322050062565209e-06, "loss": 0.5824, "step": 4118 }, { "epoch": 0.3510012782275245, "grad_norm": 16.75, "learning_rate": 2.321963640518317e-06, "loss": 1.0646, "step": 4119 }, { "epoch": 0.3510864933958245, "grad_norm": 16.25, "learning_rate": 2.3218771990999644e-06, "loss": 0.977, "step": 4120 }, { "epoch": 0.35117170856412444, "grad_norm": 13.25, "learning_rate": 2.3217907383117132e-06, "loss": 0.4908, "step": 4121 }, { "epoch": 0.3512569237324244, "grad_norm": 11.5625, "learning_rate": 2.3217042581551253e-06, "loss": 0.4513, "step": 4122 }, { "epoch": 0.35134213890072435, "grad_norm": 15.9375, "learning_rate": 2.3216177586317635e-06, "loss": 0.7099, "step": 4123 }, { "epoch": 0.3514273540690243, "grad_norm": 15.25, "learning_rate": 2.3215312397431914e-06, "loss": 0.7217, "step": 4124 }, { "epoch": 0.35151256923732427, "grad_norm": 20.25, "learning_rate": 2.3214447014909727e-06, "loss": 0.7117, "step": 4125 }, { "epoch": 0.3515977844056242, "grad_norm": 10.1875, "learning_rate": 2.3213581438766704e-06, "loss": 0.5743, "step": 4126 }, { "epoch": 0.3516829995739242, "grad_norm": 15.4375, "learning_rate": 2.321271566901849e-06, "loss": 0.8363, "step": 4127 }, { "epoch": 0.35176821474222414, "grad_norm": 15.4375, "learning_rate": 2.3211849705680735e-06, "loss": 0.7295, "step": 4128 }, { "epoch": 0.3518534299105241, "grad_norm": 14.125, "learning_rate": 2.3210983548769085e-06, "loss": 0.539, "step": 4129 }, { "epoch": 0.35193864507882405, "grad_norm": 16.125, "learning_rate": 2.321011719829919e-06, "loss": 0.8455, "step": 4130 }, { "epoch": 0.352023860247124, "grad_norm": 13.0625, "learning_rate": 2.3209250654286705e-06, "loss": 0.8935, "step": 4131 }, { "epoch": 0.35210907541542397, "grad_norm": 17.375, "learning_rate": 2.3208383916747294e-06, "loss": 0.7844, "step": 4132 }, { "epoch": 0.3521942905837239, "grad_norm": 14.4375, "learning_rate": 2.3207516985696616e-06, "loss": 0.6487, "step": 4133 }, { "epoch": 0.3522795057520239, "grad_norm": 11.375, "learning_rate": 2.320664986115034e-06, "loss": 0.8521, "step": 4134 }, { "epoch": 0.35236472092032384, "grad_norm": 13.5, "learning_rate": 2.3205782543124132e-06, "loss": 0.641, "step": 4135 }, { "epoch": 0.3524499360886238, "grad_norm": 13.125, "learning_rate": 2.3204915031633675e-06, "loss": 0.4281, "step": 4136 }, { "epoch": 0.35253515125692375, "grad_norm": 16.0, "learning_rate": 2.320404732669463e-06, "loss": 1.0232, "step": 4137 }, { "epoch": 0.3526203664252237, "grad_norm": 14.5, "learning_rate": 2.3203179428322693e-06, "loss": 0.5465, "step": 4138 }, { "epoch": 0.35270558159352366, "grad_norm": 13.9375, "learning_rate": 2.3202311336533538e-06, "loss": 0.6673, "step": 4139 }, { "epoch": 0.3527907967618236, "grad_norm": 15.4375, "learning_rate": 2.3201443051342853e-06, "loss": 0.8589, "step": 4140 }, { "epoch": 0.3528760119301236, "grad_norm": 13.5625, "learning_rate": 2.320057457276633e-06, "loss": 0.7259, "step": 4141 }, { "epoch": 0.35296122709842354, "grad_norm": 18.25, "learning_rate": 2.319970590081967e-06, "loss": 0.9552, "step": 4142 }, { "epoch": 0.3530464422667235, "grad_norm": 12.1875, "learning_rate": 2.3198837035518562e-06, "loss": 0.4087, "step": 4143 }, { "epoch": 0.35313165743502345, "grad_norm": 12.8125, "learning_rate": 2.319796797687871e-06, "loss": 0.6732, "step": 4144 }, { "epoch": 0.3532168726033234, "grad_norm": 20.875, "learning_rate": 2.319709872491582e-06, "loss": 0.8255, "step": 4145 }, { "epoch": 0.35330208777162336, "grad_norm": 11.1875, "learning_rate": 2.31962292796456e-06, "loss": 0.5269, "step": 4146 }, { "epoch": 0.3533873029399233, "grad_norm": 14.6875, "learning_rate": 2.3195359641083757e-06, "loss": 0.5091, "step": 4147 }, { "epoch": 0.3534725181082233, "grad_norm": 16.25, "learning_rate": 2.3194489809246017e-06, "loss": 0.7482, "step": 4148 }, { "epoch": 0.35355773327652323, "grad_norm": 17.25, "learning_rate": 2.3193619784148087e-06, "loss": 0.908, "step": 4149 }, { "epoch": 0.3536429484448232, "grad_norm": 12.5625, "learning_rate": 2.3192749565805693e-06, "loss": 0.4892, "step": 4150 }, { "epoch": 0.35372816361312315, "grad_norm": 11.875, "learning_rate": 2.3191879154234564e-06, "loss": 0.6026, "step": 4151 }, { "epoch": 0.3538133787814231, "grad_norm": 16.125, "learning_rate": 2.3191008549450426e-06, "loss": 0.6892, "step": 4152 }, { "epoch": 0.35389859394972306, "grad_norm": 10.75, "learning_rate": 2.3190137751469016e-06, "loss": 0.4586, "step": 4153 }, { "epoch": 0.353983809118023, "grad_norm": 14.25, "learning_rate": 2.318926676030607e-06, "loss": 0.8373, "step": 4154 }, { "epoch": 0.354069024286323, "grad_norm": 23.75, "learning_rate": 2.3188395575977316e-06, "loss": 0.9728, "step": 4155 }, { "epoch": 0.35415423945462293, "grad_norm": 16.25, "learning_rate": 2.318752419849851e-06, "loss": 0.582, "step": 4156 }, { "epoch": 0.3542394546229229, "grad_norm": 10.875, "learning_rate": 2.3186652627885393e-06, "loss": 0.5042, "step": 4157 }, { "epoch": 0.35432466979122285, "grad_norm": 17.625, "learning_rate": 2.318578086415372e-06, "loss": 1.0334, "step": 4158 }, { "epoch": 0.3544098849595228, "grad_norm": 15.5, "learning_rate": 2.318490890731924e-06, "loss": 0.5908, "step": 4159 }, { "epoch": 0.35449510012782276, "grad_norm": 18.5, "learning_rate": 2.318403675739771e-06, "loss": 0.9349, "step": 4160 }, { "epoch": 0.3545803152961227, "grad_norm": 15.5625, "learning_rate": 2.3183164414404895e-06, "loss": 0.6485, "step": 4161 }, { "epoch": 0.3546655304644227, "grad_norm": 16.0, "learning_rate": 2.3182291878356554e-06, "loss": 0.7995, "step": 4162 }, { "epoch": 0.35475074563272263, "grad_norm": 14.5625, "learning_rate": 2.318141914926846e-06, "loss": 0.8688, "step": 4163 }, { "epoch": 0.3548359608010226, "grad_norm": 13.0625, "learning_rate": 2.3180546227156376e-06, "loss": 0.7732, "step": 4164 }, { "epoch": 0.35492117596932254, "grad_norm": 12.4375, "learning_rate": 2.3179673112036083e-06, "loss": 0.669, "step": 4165 }, { "epoch": 0.3550063911376225, "grad_norm": 16.125, "learning_rate": 2.317879980392336e-06, "loss": 0.5297, "step": 4166 }, { "epoch": 0.35509160630592246, "grad_norm": 16.25, "learning_rate": 2.317792630283399e-06, "loss": 0.7031, "step": 4167 }, { "epoch": 0.3551768214742224, "grad_norm": 20.625, "learning_rate": 2.3177052608783747e-06, "loss": 0.9453, "step": 4168 }, { "epoch": 0.35526203664252237, "grad_norm": 27.375, "learning_rate": 2.3176178721788426e-06, "loss": 1.2809, "step": 4169 }, { "epoch": 0.35534725181082233, "grad_norm": 10.625, "learning_rate": 2.317530464186382e-06, "loss": 0.357, "step": 4170 }, { "epoch": 0.3554324669791223, "grad_norm": 13.25, "learning_rate": 2.317443036902573e-06, "loss": 0.7564, "step": 4171 }, { "epoch": 0.35551768214742224, "grad_norm": 15.375, "learning_rate": 2.3173555903289945e-06, "loss": 0.6833, "step": 4172 }, { "epoch": 0.3556028973157222, "grad_norm": 14.4375, "learning_rate": 2.3172681244672274e-06, "loss": 0.667, "step": 4173 }, { "epoch": 0.35568811248402216, "grad_norm": 14.625, "learning_rate": 2.317180639318852e-06, "loss": 0.8292, "step": 4174 }, { "epoch": 0.3557733276523221, "grad_norm": 18.0, "learning_rate": 2.3170931348854494e-06, "loss": 0.966, "step": 4175 }, { "epoch": 0.35585854282062207, "grad_norm": 21.125, "learning_rate": 2.3170056111686008e-06, "loss": 1.1164, "step": 4176 }, { "epoch": 0.355943757988922, "grad_norm": 12.0, "learning_rate": 2.3169180681698874e-06, "loss": 0.5023, "step": 4177 }, { "epoch": 0.356028973157222, "grad_norm": 14.875, "learning_rate": 2.316830505890892e-06, "loss": 0.6285, "step": 4178 }, { "epoch": 0.35611418832552194, "grad_norm": 14.1875, "learning_rate": 2.3167429243331966e-06, "loss": 0.8044, "step": 4179 }, { "epoch": 0.3561994034938219, "grad_norm": 12.375, "learning_rate": 2.3166553234983836e-06, "loss": 0.6846, "step": 4180 }, { "epoch": 0.35628461866212185, "grad_norm": 12.1875, "learning_rate": 2.3165677033880366e-06, "loss": 0.5404, "step": 4181 }, { "epoch": 0.3563698338304218, "grad_norm": 17.5, "learning_rate": 2.3164800640037387e-06, "loss": 0.776, "step": 4182 }, { "epoch": 0.35645504899872177, "grad_norm": 11.5625, "learning_rate": 2.3163924053470737e-06, "loss": 0.5105, "step": 4183 }, { "epoch": 0.3565402641670217, "grad_norm": 13.6875, "learning_rate": 2.3163047274196253e-06, "loss": 0.7045, "step": 4184 }, { "epoch": 0.3566254793353217, "grad_norm": 17.25, "learning_rate": 2.316217030222978e-06, "loss": 0.4571, "step": 4185 }, { "epoch": 0.35671069450362164, "grad_norm": 11.75, "learning_rate": 2.3161293137587176e-06, "loss": 0.5128, "step": 4186 }, { "epoch": 0.3567959096719216, "grad_norm": 14.0625, "learning_rate": 2.316041578028428e-06, "loss": 0.8485, "step": 4187 }, { "epoch": 0.35688112484022155, "grad_norm": 12.5625, "learning_rate": 2.315953823033695e-06, "loss": 0.5175, "step": 4188 }, { "epoch": 0.3569663400085215, "grad_norm": 17.375, "learning_rate": 2.3158660487761043e-06, "loss": 0.9817, "step": 4189 }, { "epoch": 0.35705155517682147, "grad_norm": 23.875, "learning_rate": 2.3157782552572426e-06, "loss": 1.1117, "step": 4190 }, { "epoch": 0.3571367703451214, "grad_norm": 16.75, "learning_rate": 2.315690442478696e-06, "loss": 0.8224, "step": 4191 }, { "epoch": 0.3572219855134214, "grad_norm": 14.4375, "learning_rate": 2.3156026104420515e-06, "loss": 1.0135, "step": 4192 }, { "epoch": 0.35730720068172134, "grad_norm": 20.875, "learning_rate": 2.3155147591488965e-06, "loss": 1.2968, "step": 4193 }, { "epoch": 0.3573924158500213, "grad_norm": 16.875, "learning_rate": 2.315426888600818e-06, "loss": 0.8112, "step": 4194 }, { "epoch": 0.35747763101832125, "grad_norm": 14.6875, "learning_rate": 2.3153389987994046e-06, "loss": 0.6947, "step": 4195 }, { "epoch": 0.3575628461866212, "grad_norm": 11.0, "learning_rate": 2.3152510897462437e-06, "loss": 0.5024, "step": 4196 }, { "epoch": 0.35764806135492117, "grad_norm": 12.875, "learning_rate": 2.315163161442925e-06, "loss": 0.599, "step": 4197 }, { "epoch": 0.3577332765232211, "grad_norm": 14.3125, "learning_rate": 2.3150752138910366e-06, "loss": 0.8405, "step": 4198 }, { "epoch": 0.3578184916915211, "grad_norm": 14.25, "learning_rate": 2.314987247092168e-06, "loss": 0.8189, "step": 4199 }, { "epoch": 0.35790370685982104, "grad_norm": 14.0625, "learning_rate": 2.314899261047909e-06, "loss": 0.6091, "step": 4200 }, { "epoch": 0.357988922028121, "grad_norm": 11.25, "learning_rate": 2.3148112557598497e-06, "loss": 0.3602, "step": 4201 }, { "epoch": 0.35807413719642095, "grad_norm": 12.625, "learning_rate": 2.31472323122958e-06, "loss": 0.9743, "step": 4202 }, { "epoch": 0.3581593523647209, "grad_norm": 11.6875, "learning_rate": 2.314635187458691e-06, "loss": 0.4343, "step": 4203 }, { "epoch": 0.35824456753302086, "grad_norm": 11.0625, "learning_rate": 2.314547124448773e-06, "loss": 0.4612, "step": 4204 }, { "epoch": 0.3583297827013208, "grad_norm": 14.0, "learning_rate": 2.314459042201419e-06, "loss": 0.7736, "step": 4205 }, { "epoch": 0.3584149978696208, "grad_norm": 12.0, "learning_rate": 2.314370940718219e-06, "loss": 0.6694, "step": 4206 }, { "epoch": 0.35850021303792073, "grad_norm": 17.375, "learning_rate": 2.314282820000766e-06, "loss": 0.6235, "step": 4207 }, { "epoch": 0.3585854282062207, "grad_norm": 14.1875, "learning_rate": 2.314194680050652e-06, "loss": 0.8902, "step": 4208 }, { "epoch": 0.35867064337452065, "grad_norm": 13.6875, "learning_rate": 2.3141065208694704e-06, "loss": 0.5892, "step": 4209 }, { "epoch": 0.3587558585428206, "grad_norm": 24.0, "learning_rate": 2.3140183424588136e-06, "loss": 0.6459, "step": 4210 }, { "epoch": 0.35884107371112056, "grad_norm": 15.3125, "learning_rate": 2.3139301448202757e-06, "loss": 0.6411, "step": 4211 }, { "epoch": 0.3589262888794205, "grad_norm": 12.3125, "learning_rate": 2.31384192795545e-06, "loss": 0.5189, "step": 4212 }, { "epoch": 0.3590115040477205, "grad_norm": 28.125, "learning_rate": 2.3137536918659307e-06, "loss": 1.1267, "step": 4213 }, { "epoch": 0.35909671921602043, "grad_norm": 13.75, "learning_rate": 2.313665436553313e-06, "loss": 0.6345, "step": 4214 }, { "epoch": 0.3591819343843204, "grad_norm": 13.5625, "learning_rate": 2.313577162019191e-06, "loss": 0.5901, "step": 4215 }, { "epoch": 0.35926714955262035, "grad_norm": 15.875, "learning_rate": 2.31348886826516e-06, "loss": 0.8198, "step": 4216 }, { "epoch": 0.3593523647209203, "grad_norm": 16.875, "learning_rate": 2.313400555292816e-06, "loss": 0.8849, "step": 4217 }, { "epoch": 0.35943757988922026, "grad_norm": 12.1875, "learning_rate": 2.3133122231037547e-06, "loss": 0.4802, "step": 4218 }, { "epoch": 0.3595227950575202, "grad_norm": 13.1875, "learning_rate": 2.313223871699572e-06, "loss": 0.592, "step": 4219 }, { "epoch": 0.3596080102258202, "grad_norm": 14.6875, "learning_rate": 2.313135501081865e-06, "loss": 0.8414, "step": 4220 }, { "epoch": 0.35969322539412013, "grad_norm": 14.125, "learning_rate": 2.3130471112522303e-06, "loss": 0.7617, "step": 4221 }, { "epoch": 0.3597784405624201, "grad_norm": 13.75, "learning_rate": 2.3129587022122652e-06, "loss": 0.7365, "step": 4222 }, { "epoch": 0.35986365573072004, "grad_norm": 10.875, "learning_rate": 2.3128702739635677e-06, "loss": 0.4697, "step": 4223 }, { "epoch": 0.35994887089902, "grad_norm": 17.625, "learning_rate": 2.3127818265077357e-06, "loss": 0.6869, "step": 4224 }, { "epoch": 0.36003408606731996, "grad_norm": 24.125, "learning_rate": 2.312693359846367e-06, "loss": 0.863, "step": 4225 }, { "epoch": 0.3601193012356199, "grad_norm": 17.5, "learning_rate": 2.312604873981061e-06, "loss": 0.8449, "step": 4226 }, { "epoch": 0.36020451640391987, "grad_norm": 20.5, "learning_rate": 2.312516368913416e-06, "loss": 0.9352, "step": 4227 }, { "epoch": 0.36028973157221983, "grad_norm": 14.4375, "learning_rate": 2.312427844645032e-06, "loss": 1.0289, "step": 4228 }, { "epoch": 0.3603749467405198, "grad_norm": 13.875, "learning_rate": 2.3123393011775085e-06, "loss": 0.7591, "step": 4229 }, { "epoch": 0.36046016190881974, "grad_norm": 11.4375, "learning_rate": 2.3122507385124453e-06, "loss": 0.5287, "step": 4230 }, { "epoch": 0.3605453770771197, "grad_norm": 11.5, "learning_rate": 2.3121621566514433e-06, "loss": 0.5529, "step": 4231 }, { "epoch": 0.3606305922454197, "grad_norm": 15.25, "learning_rate": 2.3120735555961034e-06, "loss": 0.7931, "step": 4232 }, { "epoch": 0.36071580741371967, "grad_norm": 15.8125, "learning_rate": 2.311984935348026e-06, "loss": 1.0952, "step": 4233 }, { "epoch": 0.3608010225820196, "grad_norm": 15.4375, "learning_rate": 2.311896295908813e-06, "loss": 0.6672, "step": 4234 }, { "epoch": 0.3608862377503196, "grad_norm": 11.625, "learning_rate": 2.311807637280066e-06, "loss": 0.4892, "step": 4235 }, { "epoch": 0.36097145291861954, "grad_norm": 14.75, "learning_rate": 2.3117189594633873e-06, "loss": 0.9461, "step": 4236 }, { "epoch": 0.3610566680869195, "grad_norm": 57.0, "learning_rate": 2.3116302624603797e-06, "loss": 1.0312, "step": 4237 }, { "epoch": 0.36114188325521945, "grad_norm": 13.3125, "learning_rate": 2.311541546272645e-06, "loss": 0.7095, "step": 4238 }, { "epoch": 0.3612270984235194, "grad_norm": 12.5, "learning_rate": 2.3114528109017876e-06, "loss": 0.733, "step": 4239 }, { "epoch": 0.36131231359181937, "grad_norm": 15.3125, "learning_rate": 2.3113640563494105e-06, "loss": 0.7853, "step": 4240 }, { "epoch": 0.3613975287601193, "grad_norm": 11.9375, "learning_rate": 2.3112752826171174e-06, "loss": 0.3749, "step": 4241 }, { "epoch": 0.3614827439284193, "grad_norm": 21.375, "learning_rate": 2.3111864897065127e-06, "loss": 0.9213, "step": 4242 }, { "epoch": 0.36156795909671924, "grad_norm": 21.625, "learning_rate": 2.3110976776192016e-06, "loss": 0.694, "step": 4243 }, { "epoch": 0.3616531742650192, "grad_norm": 11.25, "learning_rate": 2.311008846356788e-06, "loss": 0.5355, "step": 4244 }, { "epoch": 0.36173838943331915, "grad_norm": 19.625, "learning_rate": 2.3109199959208777e-06, "loss": 1.1726, "step": 4245 }, { "epoch": 0.3618236046016191, "grad_norm": 13.0625, "learning_rate": 2.310831126313076e-06, "loss": 0.6114, "step": 4246 }, { "epoch": 0.36190881976991907, "grad_norm": 14.375, "learning_rate": 2.3107422375349894e-06, "loss": 0.9449, "step": 4247 }, { "epoch": 0.361994034938219, "grad_norm": 17.25, "learning_rate": 2.310653329588224e-06, "loss": 0.9274, "step": 4248 }, { "epoch": 0.362079250106519, "grad_norm": 16.625, "learning_rate": 2.310564402474386e-06, "loss": 0.9639, "step": 4249 }, { "epoch": 0.36216446527481894, "grad_norm": 15.0625, "learning_rate": 2.3104754561950828e-06, "loss": 0.8019, "step": 4250 }, { "epoch": 0.3622496804431189, "grad_norm": 15.375, "learning_rate": 2.310386490751922e-06, "loss": 0.8768, "step": 4251 }, { "epoch": 0.36233489561141885, "grad_norm": 13.125, "learning_rate": 2.310297506146511e-06, "loss": 0.7227, "step": 4252 }, { "epoch": 0.3624201107797188, "grad_norm": 12.375, "learning_rate": 2.3102085023804576e-06, "loss": 0.4608, "step": 4253 }, { "epoch": 0.36250532594801876, "grad_norm": 15.4375, "learning_rate": 2.3101194794553706e-06, "loss": 0.9351, "step": 4254 }, { "epoch": 0.3625905411163187, "grad_norm": 15.125, "learning_rate": 2.3100304373728585e-06, "loss": 0.6416, "step": 4255 }, { "epoch": 0.3626757562846187, "grad_norm": 15.0, "learning_rate": 2.3099413761345303e-06, "loss": 0.781, "step": 4256 }, { "epoch": 0.36276097145291863, "grad_norm": 16.625, "learning_rate": 2.309852295741996e-06, "loss": 0.5596, "step": 4257 }, { "epoch": 0.3628461866212186, "grad_norm": 16.875, "learning_rate": 2.3097631961968646e-06, "loss": 0.8913, "step": 4258 }, { "epoch": 0.36293140178951855, "grad_norm": 20.375, "learning_rate": 2.3096740775007472e-06, "loss": 0.9099, "step": 4259 }, { "epoch": 0.3630166169578185, "grad_norm": 15.8125, "learning_rate": 2.309584939655253e-06, "loss": 0.9584, "step": 4260 }, { "epoch": 0.36310183212611846, "grad_norm": 14.8125, "learning_rate": 2.3094957826619933e-06, "loss": 0.5139, "step": 4261 }, { "epoch": 0.3631870472944184, "grad_norm": 15.3125, "learning_rate": 2.3094066065225794e-06, "loss": 1.0313, "step": 4262 }, { "epoch": 0.3632722624627184, "grad_norm": 14.5, "learning_rate": 2.3093174112386234e-06, "loss": 0.6139, "step": 4263 }, { "epoch": 0.36335747763101833, "grad_norm": 13.4375, "learning_rate": 2.3092281968117363e-06, "loss": 0.6167, "step": 4264 }, { "epoch": 0.3634426927993183, "grad_norm": 13.0, "learning_rate": 2.3091389632435305e-06, "loss": 0.7595, "step": 4265 }, { "epoch": 0.36352790796761825, "grad_norm": 15.75, "learning_rate": 2.3090497105356187e-06, "loss": 0.8667, "step": 4266 }, { "epoch": 0.3636131231359182, "grad_norm": 13.1875, "learning_rate": 2.3089604386896132e-06, "loss": 0.5804, "step": 4267 }, { "epoch": 0.36369833830421816, "grad_norm": 16.25, "learning_rate": 2.308871147707128e-06, "loss": 0.7245, "step": 4268 }, { "epoch": 0.3637835534725181, "grad_norm": 17.125, "learning_rate": 2.3087818375897767e-06, "loss": 0.4998, "step": 4269 }, { "epoch": 0.3638687686408181, "grad_norm": 14.4375, "learning_rate": 2.3086925083391727e-06, "loss": 0.5352, "step": 4270 }, { "epoch": 0.36395398380911803, "grad_norm": 12.875, "learning_rate": 2.3086031599569303e-06, "loss": 0.6056, "step": 4271 }, { "epoch": 0.364039198977418, "grad_norm": 16.125, "learning_rate": 2.3085137924446644e-06, "loss": 0.7291, "step": 4272 }, { "epoch": 0.36412441414571795, "grad_norm": 13.4375, "learning_rate": 2.3084244058039902e-06, "loss": 0.7364, "step": 4273 }, { "epoch": 0.3642096293140179, "grad_norm": 13.0, "learning_rate": 2.3083350000365223e-06, "loss": 0.6466, "step": 4274 }, { "epoch": 0.36429484448231786, "grad_norm": 12.75, "learning_rate": 2.3082455751438766e-06, "loss": 0.3638, "step": 4275 }, { "epoch": 0.3643800596506178, "grad_norm": 10.75, "learning_rate": 2.308156131127669e-06, "loss": 0.5419, "step": 4276 }, { "epoch": 0.3644652748189178, "grad_norm": 13.125, "learning_rate": 2.3080666679895165e-06, "loss": 0.622, "step": 4277 }, { "epoch": 0.36455048998721773, "grad_norm": 12.6875, "learning_rate": 2.307977185731035e-06, "loss": 0.6674, "step": 4278 }, { "epoch": 0.3646357051555177, "grad_norm": 13.875, "learning_rate": 2.3078876843538423e-06, "loss": 0.4298, "step": 4279 }, { "epoch": 0.36472092032381764, "grad_norm": 11.9375, "learning_rate": 2.3077981638595554e-06, "loss": 0.5843, "step": 4280 }, { "epoch": 0.3648061354921176, "grad_norm": 17.25, "learning_rate": 2.3077086242497916e-06, "loss": 0.7285, "step": 4281 }, { "epoch": 0.36489135066041756, "grad_norm": 17.0, "learning_rate": 2.3076190655261694e-06, "loss": 0.5327, "step": 4282 }, { "epoch": 0.3649765658287175, "grad_norm": 15.625, "learning_rate": 2.3075294876903076e-06, "loss": 1.0993, "step": 4283 }, { "epoch": 0.36506178099701747, "grad_norm": 12.8125, "learning_rate": 2.307439890743824e-06, "loss": 0.6302, "step": 4284 }, { "epoch": 0.36514699616531743, "grad_norm": 13.6875, "learning_rate": 2.3073502746883387e-06, "loss": 0.6014, "step": 4285 }, { "epoch": 0.3652322113336174, "grad_norm": 12.25, "learning_rate": 2.30726063952547e-06, "loss": 0.5546, "step": 4286 }, { "epoch": 0.36531742650191734, "grad_norm": 15.1875, "learning_rate": 2.3071709852568392e-06, "loss": 0.7087, "step": 4287 }, { "epoch": 0.3654026416702173, "grad_norm": 14.3125, "learning_rate": 2.307081311884065e-06, "loss": 0.9438, "step": 4288 }, { "epoch": 0.36548785683851726, "grad_norm": 10.125, "learning_rate": 2.3069916194087687e-06, "loss": 0.2662, "step": 4289 }, { "epoch": 0.3655730720068172, "grad_norm": 18.0, "learning_rate": 2.3069019078325713e-06, "loss": 0.9825, "step": 4290 }, { "epoch": 0.36565828717511717, "grad_norm": 11.0, "learning_rate": 2.306812177157093e-06, "loss": 0.4794, "step": 4291 }, { "epoch": 0.3657435023434171, "grad_norm": 23.625, "learning_rate": 2.3067224273839562e-06, "loss": 0.8132, "step": 4292 }, { "epoch": 0.3658287175117171, "grad_norm": 13.0625, "learning_rate": 2.3066326585147827e-06, "loss": 0.6164, "step": 4293 }, { "epoch": 0.36591393268001704, "grad_norm": 13.875, "learning_rate": 2.3065428705511946e-06, "loss": 0.8203, "step": 4294 }, { "epoch": 0.365999147848317, "grad_norm": 15.5625, "learning_rate": 2.306453063494814e-06, "loss": 0.5603, "step": 4295 }, { "epoch": 0.36608436301661695, "grad_norm": 25.875, "learning_rate": 2.3063632373472645e-06, "loss": 0.9298, "step": 4296 }, { "epoch": 0.3661695781849169, "grad_norm": 11.5625, "learning_rate": 2.306273392110169e-06, "loss": 0.4272, "step": 4297 }, { "epoch": 0.36625479335321687, "grad_norm": 9.5625, "learning_rate": 2.3061835277851514e-06, "loss": 0.4046, "step": 4298 }, { "epoch": 0.3663400085215168, "grad_norm": 27.0, "learning_rate": 2.3060936443738347e-06, "loss": 0.7893, "step": 4299 }, { "epoch": 0.3664252236898168, "grad_norm": 9.875, "learning_rate": 2.3060037418778446e-06, "loss": 0.9641, "step": 4300 }, { "epoch": 0.36651043885811674, "grad_norm": 10.75, "learning_rate": 2.3059138202988044e-06, "loss": 0.5021, "step": 4301 }, { "epoch": 0.3665956540264167, "grad_norm": 30.375, "learning_rate": 2.3058238796383396e-06, "loss": 0.8703, "step": 4302 }, { "epoch": 0.36668086919471665, "grad_norm": 18.0, "learning_rate": 2.305733919898076e-06, "loss": 1.1952, "step": 4303 }, { "epoch": 0.3667660843630166, "grad_norm": 18.375, "learning_rate": 2.305643941079639e-06, "loss": 0.6789, "step": 4304 }, { "epoch": 0.36685129953131657, "grad_norm": 14.4375, "learning_rate": 2.305553943184654e-06, "loss": 0.8883, "step": 4305 }, { "epoch": 0.3669365146996165, "grad_norm": 12.9375, "learning_rate": 2.3054639262147483e-06, "loss": 0.4696, "step": 4306 }, { "epoch": 0.3670217298679165, "grad_norm": 14.6875, "learning_rate": 2.305373890171548e-06, "loss": 0.9156, "step": 4307 }, { "epoch": 0.36710694503621644, "grad_norm": 21.5, "learning_rate": 2.3052838350566797e-06, "loss": 0.8899, "step": 4308 }, { "epoch": 0.3671921602045164, "grad_norm": 14.625, "learning_rate": 2.3051937608717716e-06, "loss": 0.5692, "step": 4309 }, { "epoch": 0.36727737537281635, "grad_norm": 15.4375, "learning_rate": 2.3051036676184514e-06, "loss": 0.7317, "step": 4310 }, { "epoch": 0.3673625905411163, "grad_norm": 11.6875, "learning_rate": 2.3050135552983467e-06, "loss": 0.4207, "step": 4311 }, { "epoch": 0.36744780570941626, "grad_norm": 16.25, "learning_rate": 2.3049234239130864e-06, "loss": 0.6036, "step": 4312 }, { "epoch": 0.3675330208777162, "grad_norm": 16.5, "learning_rate": 2.304833273464299e-06, "loss": 0.8124, "step": 4313 }, { "epoch": 0.3676182360460162, "grad_norm": 19.0, "learning_rate": 2.304743103953613e-06, "loss": 0.7817, "step": 4314 }, { "epoch": 0.36770345121431613, "grad_norm": 17.75, "learning_rate": 2.304652915382659e-06, "loss": 0.9005, "step": 4315 }, { "epoch": 0.3677886663826161, "grad_norm": 13.625, "learning_rate": 2.3045627077530664e-06, "loss": 0.859, "step": 4316 }, { "epoch": 0.36787388155091605, "grad_norm": 15.0625, "learning_rate": 2.3044724810664647e-06, "loss": 0.6488, "step": 4317 }, { "epoch": 0.367959096719216, "grad_norm": 19.0, "learning_rate": 2.3043822353244856e-06, "loss": 0.6512, "step": 4318 }, { "epoch": 0.36804431188751596, "grad_norm": 14.125, "learning_rate": 2.304291970528759e-06, "loss": 0.5366, "step": 4319 }, { "epoch": 0.3681295270558159, "grad_norm": 11.4375, "learning_rate": 2.3042016866809165e-06, "loss": 0.5852, "step": 4320 }, { "epoch": 0.3682147422241159, "grad_norm": 14.5625, "learning_rate": 2.304111383782589e-06, "loss": 0.6552, "step": 4321 }, { "epoch": 0.36829995739241583, "grad_norm": 13.5, "learning_rate": 2.304021061835409e-06, "loss": 0.7635, "step": 4322 }, { "epoch": 0.3683851725607158, "grad_norm": 14.0, "learning_rate": 2.303930720841008e-06, "loss": 0.5679, "step": 4323 }, { "epoch": 0.36847038772901575, "grad_norm": 12.625, "learning_rate": 2.30384036080102e-06, "loss": 0.7854, "step": 4324 }, { "epoch": 0.3685556028973157, "grad_norm": 13.3125, "learning_rate": 2.3037499817170765e-06, "loss": 0.7057, "step": 4325 }, { "epoch": 0.36864081806561566, "grad_norm": 15.9375, "learning_rate": 2.303659583590811e-06, "loss": 0.7431, "step": 4326 }, { "epoch": 0.3687260332339156, "grad_norm": 12.4375, "learning_rate": 2.303569166423858e-06, "loss": 0.6521, "step": 4327 }, { "epoch": 0.3688112484022156, "grad_norm": 12.25, "learning_rate": 2.3034787302178503e-06, "loss": 0.559, "step": 4328 }, { "epoch": 0.36889646357051553, "grad_norm": 14.25, "learning_rate": 2.3033882749744223e-06, "loss": 0.7696, "step": 4329 }, { "epoch": 0.3689816787388155, "grad_norm": 13.8125, "learning_rate": 2.3032978006952087e-06, "loss": 0.6558, "step": 4330 }, { "epoch": 0.36906689390711545, "grad_norm": 14.375, "learning_rate": 2.3032073073818456e-06, "loss": 0.8422, "step": 4331 }, { "epoch": 0.3691521090754154, "grad_norm": 19.625, "learning_rate": 2.3031167950359668e-06, "loss": 1.0157, "step": 4332 }, { "epoch": 0.36923732424371536, "grad_norm": 15.8125, "learning_rate": 2.3030262636592084e-06, "loss": 0.9339, "step": 4333 }, { "epoch": 0.3693225394120153, "grad_norm": 20.25, "learning_rate": 2.3029357132532072e-06, "loss": 0.6715, "step": 4334 }, { "epoch": 0.3694077545803153, "grad_norm": 16.875, "learning_rate": 2.302845143819598e-06, "loss": 0.7813, "step": 4335 }, { "epoch": 0.36949296974861523, "grad_norm": 20.75, "learning_rate": 2.302754555360019e-06, "loss": 1.2755, "step": 4336 }, { "epoch": 0.3695781849169152, "grad_norm": 13.125, "learning_rate": 2.302663947876106e-06, "loss": 0.504, "step": 4337 }, { "epoch": 0.36966340008521514, "grad_norm": 9.25, "learning_rate": 2.3025733213694972e-06, "loss": 0.5204, "step": 4338 }, { "epoch": 0.3697486152535151, "grad_norm": 12.75, "learning_rate": 2.30248267584183e-06, "loss": 0.5122, "step": 4339 }, { "epoch": 0.36983383042181506, "grad_norm": 24.625, "learning_rate": 2.3023920112947426e-06, "loss": 0.917, "step": 4340 }, { "epoch": 0.369919045590115, "grad_norm": 11.75, "learning_rate": 2.3023013277298733e-06, "loss": 0.7611, "step": 4341 }, { "epoch": 0.37000426075841497, "grad_norm": 15.625, "learning_rate": 2.3022106251488606e-06, "loss": 0.9692, "step": 4342 }, { "epoch": 0.37008947592671493, "grad_norm": 12.6875, "learning_rate": 2.3021199035533444e-06, "loss": 0.5794, "step": 4343 }, { "epoch": 0.3701746910950149, "grad_norm": 12.125, "learning_rate": 2.302029162944963e-06, "loss": 0.6374, "step": 4344 }, { "epoch": 0.3702599062633149, "grad_norm": 13.0625, "learning_rate": 2.301938403325357e-06, "loss": 0.6702, "step": 4345 }, { "epoch": 0.37034512143161485, "grad_norm": 12.4375, "learning_rate": 2.301847624696166e-06, "loss": 0.5704, "step": 4346 }, { "epoch": 0.3704303365999148, "grad_norm": 26.25, "learning_rate": 2.301756827059031e-06, "loss": 1.0872, "step": 4347 }, { "epoch": 0.37051555176821477, "grad_norm": 15.125, "learning_rate": 2.3016660104155917e-06, "loss": 0.7761, "step": 4348 }, { "epoch": 0.3706007669365147, "grad_norm": 11.6875, "learning_rate": 2.3015751747674907e-06, "loss": 0.5179, "step": 4349 }, { "epoch": 0.3706859821048147, "grad_norm": 14.625, "learning_rate": 2.301484320116369e-06, "loss": 0.9032, "step": 4350 }, { "epoch": 0.37077119727311464, "grad_norm": 15.0625, "learning_rate": 2.301393446463868e-06, "loss": 1.1451, "step": 4351 }, { "epoch": 0.3708564124414146, "grad_norm": 27.125, "learning_rate": 2.3013025538116305e-06, "loss": 1.2107, "step": 4352 }, { "epoch": 0.37094162760971455, "grad_norm": 12.6875, "learning_rate": 2.3012116421612985e-06, "loss": 0.9511, "step": 4353 }, { "epoch": 0.3710268427780145, "grad_norm": 14.25, "learning_rate": 2.301120711514515e-06, "loss": 0.7931, "step": 4354 }, { "epoch": 0.37111205794631447, "grad_norm": 13.125, "learning_rate": 2.3010297618729233e-06, "loss": 0.5906, "step": 4355 }, { "epoch": 0.3711972731146144, "grad_norm": 14.0, "learning_rate": 2.3009387932381665e-06, "loss": 0.6769, "step": 4356 }, { "epoch": 0.3712824882829144, "grad_norm": 15.875, "learning_rate": 2.3008478056118896e-06, "loss": 0.8732, "step": 4357 }, { "epoch": 0.37136770345121434, "grad_norm": 11.3125, "learning_rate": 2.3007567989957358e-06, "loss": 0.3986, "step": 4358 }, { "epoch": 0.3714529186195143, "grad_norm": 12.8125, "learning_rate": 2.30066577339135e-06, "loss": 0.7206, "step": 4359 }, { "epoch": 0.37153813378781425, "grad_norm": 16.375, "learning_rate": 2.3005747288003774e-06, "loss": 0.8717, "step": 4360 }, { "epoch": 0.3716233489561142, "grad_norm": 12.3125, "learning_rate": 2.3004836652244626e-06, "loss": 0.6998, "step": 4361 }, { "epoch": 0.37170856412441416, "grad_norm": 13.1875, "learning_rate": 2.300392582665252e-06, "loss": 0.579, "step": 4362 }, { "epoch": 0.3717937792927141, "grad_norm": 14.5625, "learning_rate": 2.300301481124391e-06, "loss": 0.8073, "step": 4363 }, { "epoch": 0.3718789944610141, "grad_norm": 13.75, "learning_rate": 2.3002103606035265e-06, "loss": 0.6729, "step": 4364 }, { "epoch": 0.37196420962931404, "grad_norm": 11.6875, "learning_rate": 2.300119221104304e-06, "loss": 0.3388, "step": 4365 }, { "epoch": 0.372049424797614, "grad_norm": 16.125, "learning_rate": 2.3000280626283718e-06, "loss": 0.6479, "step": 4366 }, { "epoch": 0.37213463996591395, "grad_norm": 13.3125, "learning_rate": 2.2999368851773764e-06, "loss": 0.8203, "step": 4367 }, { "epoch": 0.3722198551342139, "grad_norm": 16.75, "learning_rate": 2.2998456887529654e-06, "loss": 0.7895, "step": 4368 }, { "epoch": 0.37230507030251386, "grad_norm": 14.5625, "learning_rate": 2.299754473356787e-06, "loss": 0.8478, "step": 4369 }, { "epoch": 0.3723902854708138, "grad_norm": 20.625, "learning_rate": 2.2996632389904903e-06, "loss": 1.0137, "step": 4370 }, { "epoch": 0.3724755006391138, "grad_norm": 13.75, "learning_rate": 2.2995719856557232e-06, "loss": 1.0345, "step": 4371 }, { "epoch": 0.37256071580741373, "grad_norm": 12.6875, "learning_rate": 2.2994807133541347e-06, "loss": 0.7889, "step": 4372 }, { "epoch": 0.3726459309757137, "grad_norm": 19.25, "learning_rate": 2.2993894220873743e-06, "loss": 0.9149, "step": 4373 }, { "epoch": 0.37273114614401365, "grad_norm": 11.4375, "learning_rate": 2.299298111857092e-06, "loss": 0.4575, "step": 4374 }, { "epoch": 0.3728163613123136, "grad_norm": 28.0, "learning_rate": 2.299206782664938e-06, "loss": 0.4605, "step": 4375 }, { "epoch": 0.37290157648061356, "grad_norm": 16.125, "learning_rate": 2.2991154345125617e-06, "loss": 0.9573, "step": 4376 }, { "epoch": 0.3729867916489135, "grad_norm": 10.1875, "learning_rate": 2.299024067401615e-06, "loss": 0.4191, "step": 4377 }, { "epoch": 0.3730720068172135, "grad_norm": 13.5, "learning_rate": 2.2989326813337483e-06, "loss": 0.6309, "step": 4378 }, { "epoch": 0.37315722198551343, "grad_norm": 12.5625, "learning_rate": 2.2988412763106135e-06, "loss": 0.458, "step": 4379 }, { "epoch": 0.3732424371538134, "grad_norm": 13.6875, "learning_rate": 2.2987498523338615e-06, "loss": 0.7933, "step": 4380 }, { "epoch": 0.37332765232211335, "grad_norm": 12.875, "learning_rate": 2.2986584094051455e-06, "loss": 0.6122, "step": 4381 }, { "epoch": 0.3734128674904133, "grad_norm": 13.75, "learning_rate": 2.2985669475261175e-06, "loss": 0.6858, "step": 4382 }, { "epoch": 0.37349808265871326, "grad_norm": 12.8125, "learning_rate": 2.2984754666984306e-06, "loss": 0.7758, "step": 4383 }, { "epoch": 0.3735832978270132, "grad_norm": 16.75, "learning_rate": 2.2983839669237372e-06, "loss": 0.813, "step": 4384 }, { "epoch": 0.3736685129953132, "grad_norm": 16.25, "learning_rate": 2.2982924482036917e-06, "loss": 0.5505, "step": 4385 }, { "epoch": 0.37375372816361313, "grad_norm": 12.5625, "learning_rate": 2.298200910539947e-06, "loss": 0.7563, "step": 4386 }, { "epoch": 0.3738389433319131, "grad_norm": 13.9375, "learning_rate": 2.2981093539341588e-06, "loss": 0.7459, "step": 4387 }, { "epoch": 0.37392415850021304, "grad_norm": 18.25, "learning_rate": 2.29801777838798e-06, "loss": 0.9212, "step": 4388 }, { "epoch": 0.374009373668513, "grad_norm": 19.125, "learning_rate": 2.2979261839030657e-06, "loss": 0.8884, "step": 4389 }, { "epoch": 0.37409458883681296, "grad_norm": 16.375, "learning_rate": 2.297834570481072e-06, "loss": 0.8419, "step": 4390 }, { "epoch": 0.3741798040051129, "grad_norm": 18.75, "learning_rate": 2.297742938123654e-06, "loss": 0.7831, "step": 4391 }, { "epoch": 0.37426501917341287, "grad_norm": 12.9375, "learning_rate": 2.2976512868324675e-06, "loss": 0.4998, "step": 4392 }, { "epoch": 0.37435023434171283, "grad_norm": 10.5, "learning_rate": 2.2975596166091686e-06, "loss": 0.3399, "step": 4393 }, { "epoch": 0.3744354495100128, "grad_norm": 10.625, "learning_rate": 2.297467927455414e-06, "loss": 0.3401, "step": 4394 }, { "epoch": 0.37452066467831274, "grad_norm": 16.625, "learning_rate": 2.2973762193728606e-06, "loss": 1.1352, "step": 4395 }, { "epoch": 0.3746058798466127, "grad_norm": 14.9375, "learning_rate": 2.2972844923631655e-06, "loss": 0.8937, "step": 4396 }, { "epoch": 0.37469109501491266, "grad_norm": 12.875, "learning_rate": 2.297192746427987e-06, "loss": 0.8557, "step": 4397 }, { "epoch": 0.3747763101832126, "grad_norm": 28.25, "learning_rate": 2.297100981568982e-06, "loss": 0.9301, "step": 4398 }, { "epoch": 0.37486152535151257, "grad_norm": 15.5, "learning_rate": 2.2970091977878097e-06, "loss": 0.7017, "step": 4399 }, { "epoch": 0.3749467405198125, "grad_norm": 13.375, "learning_rate": 2.296917395086128e-06, "loss": 0.8333, "step": 4400 }, { "epoch": 0.3750319556881125, "grad_norm": 17.625, "learning_rate": 2.2968255734655963e-06, "loss": 0.895, "step": 4401 }, { "epoch": 0.37511717085641244, "grad_norm": 13.875, "learning_rate": 2.2967337329278746e-06, "loss": 0.9576, "step": 4402 }, { "epoch": 0.3752023860247124, "grad_norm": 18.5, "learning_rate": 2.296641873474621e-06, "loss": 0.9124, "step": 4403 }, { "epoch": 0.37528760119301235, "grad_norm": 14.3125, "learning_rate": 2.2965499951074963e-06, "loss": 0.7816, "step": 4404 }, { "epoch": 0.3753728163613123, "grad_norm": 19.625, "learning_rate": 2.296458097828161e-06, "loss": 0.8399, "step": 4405 }, { "epoch": 0.37545803152961227, "grad_norm": 16.625, "learning_rate": 2.2963661816382755e-06, "loss": 1.1387, "step": 4406 }, { "epoch": 0.3755432466979122, "grad_norm": 16.75, "learning_rate": 2.296274246539501e-06, "loss": 1.0137, "step": 4407 }, { "epoch": 0.3756284618662122, "grad_norm": 14.25, "learning_rate": 2.2961822925334985e-06, "loss": 0.9218, "step": 4408 }, { "epoch": 0.37571367703451214, "grad_norm": 9.4375, "learning_rate": 2.29609031962193e-06, "loss": 0.3351, "step": 4409 }, { "epoch": 0.3757988922028121, "grad_norm": 13.1875, "learning_rate": 2.2959983278064576e-06, "loss": 1.0254, "step": 4410 }, { "epoch": 0.37588410737111205, "grad_norm": 14.3125, "learning_rate": 2.2959063170887435e-06, "loss": 0.7664, "step": 4411 }, { "epoch": 0.375969322539412, "grad_norm": 14.25, "learning_rate": 2.295814287470451e-06, "loss": 0.4341, "step": 4412 }, { "epoch": 0.37605453770771197, "grad_norm": 13.8125, "learning_rate": 2.2957222389532423e-06, "loss": 0.8277, "step": 4413 }, { "epoch": 0.3761397528760119, "grad_norm": 19.375, "learning_rate": 2.295630171538781e-06, "loss": 0.5334, "step": 4414 }, { "epoch": 0.3762249680443119, "grad_norm": 11.0625, "learning_rate": 2.2955380852287314e-06, "loss": 0.572, "step": 4415 }, { "epoch": 0.37631018321261184, "grad_norm": 15.6875, "learning_rate": 2.295445980024757e-06, "loss": 0.5813, "step": 4416 }, { "epoch": 0.3763953983809118, "grad_norm": 18.125, "learning_rate": 2.2953538559285223e-06, "loss": 1.1529, "step": 4417 }, { "epoch": 0.37648061354921175, "grad_norm": 14.875, "learning_rate": 2.2952617129416926e-06, "loss": 0.7758, "step": 4418 }, { "epoch": 0.3765658287175117, "grad_norm": 12.1875, "learning_rate": 2.295169551065932e-06, "loss": 0.3814, "step": 4419 }, { "epoch": 0.37665104388581166, "grad_norm": 20.625, "learning_rate": 2.295077370302907e-06, "loss": 0.6561, "step": 4420 }, { "epoch": 0.3767362590541116, "grad_norm": 16.0, "learning_rate": 2.2949851706542837e-06, "loss": 1.204, "step": 4421 }, { "epoch": 0.3768214742224116, "grad_norm": 13.0, "learning_rate": 2.2948929521217268e-06, "loss": 0.7617, "step": 4422 }, { "epoch": 0.37690668939071154, "grad_norm": 11.5625, "learning_rate": 2.294800714706904e-06, "loss": 0.4446, "step": 4423 }, { "epoch": 0.3769919045590115, "grad_norm": 16.25, "learning_rate": 2.2947084584114808e-06, "loss": 1.0536, "step": 4424 }, { "epoch": 0.37707711972731145, "grad_norm": 11.375, "learning_rate": 2.294616183237126e-06, "loss": 0.3647, "step": 4425 }, { "epoch": 0.3771623348956114, "grad_norm": 16.625, "learning_rate": 2.294523889185506e-06, "loss": 0.9839, "step": 4426 }, { "epoch": 0.37724755006391136, "grad_norm": 13.375, "learning_rate": 2.2944315762582893e-06, "loss": 0.5052, "step": 4427 }, { "epoch": 0.3773327652322113, "grad_norm": 14.9375, "learning_rate": 2.2943392444571435e-06, "loss": 0.8716, "step": 4428 }, { "epoch": 0.3774179804005113, "grad_norm": 18.125, "learning_rate": 2.2942468937837377e-06, "loss": 0.9572, "step": 4429 }, { "epoch": 0.37750319556881123, "grad_norm": 14.0, "learning_rate": 2.29415452423974e-06, "loss": 0.4185, "step": 4430 }, { "epoch": 0.3775884107371112, "grad_norm": 17.75, "learning_rate": 2.29406213582682e-06, "loss": 1.0528, "step": 4431 }, { "epoch": 0.37767362590541115, "grad_norm": 11.625, "learning_rate": 2.293969728546648e-06, "loss": 0.4112, "step": 4432 }, { "epoch": 0.3777588410737111, "grad_norm": 18.25, "learning_rate": 2.2938773024008924e-06, "loss": 1.0375, "step": 4433 }, { "epoch": 0.37784405624201106, "grad_norm": 18.125, "learning_rate": 2.2937848573912246e-06, "loss": 0.776, "step": 4434 }, { "epoch": 0.377929271410311, "grad_norm": 14.4375, "learning_rate": 2.293692393519315e-06, "loss": 0.6619, "step": 4435 }, { "epoch": 0.378014486578611, "grad_norm": 24.5, "learning_rate": 2.293599910786834e-06, "loss": 1.2235, "step": 4436 }, { "epoch": 0.37809970174691093, "grad_norm": 14.25, "learning_rate": 2.2935074091954535e-06, "loss": 0.6463, "step": 4437 }, { "epoch": 0.3781849169152109, "grad_norm": 14.5625, "learning_rate": 2.2934148887468448e-06, "loss": 0.5819, "step": 4438 }, { "epoch": 0.37827013208351085, "grad_norm": 10.0, "learning_rate": 2.29332234944268e-06, "loss": 0.3957, "step": 4439 }, { "epoch": 0.3783553472518108, "grad_norm": 12.0625, "learning_rate": 2.2932297912846307e-06, "loss": 0.5868, "step": 4440 }, { "epoch": 0.37844056242011076, "grad_norm": 18.875, "learning_rate": 2.2931372142743703e-06, "loss": 0.9452, "step": 4441 }, { "epoch": 0.3785257775884107, "grad_norm": 17.25, "learning_rate": 2.2930446184135712e-06, "loss": 1.2007, "step": 4442 }, { "epoch": 0.3786109927567107, "grad_norm": 15.8125, "learning_rate": 2.292952003703907e-06, "loss": 0.769, "step": 4443 }, { "epoch": 0.37869620792501063, "grad_norm": 12.1875, "learning_rate": 2.2928593701470516e-06, "loss": 0.5402, "step": 4444 }, { "epoch": 0.3787814230933106, "grad_norm": 17.5, "learning_rate": 2.292766717744679e-06, "loss": 0.5832, "step": 4445 }, { "epoch": 0.37886663826161054, "grad_norm": 13.6875, "learning_rate": 2.2926740464984625e-06, "loss": 0.7219, "step": 4446 }, { "epoch": 0.3789518534299105, "grad_norm": 15.6875, "learning_rate": 2.292581356410078e-06, "loss": 0.5683, "step": 4447 }, { "epoch": 0.37903706859821046, "grad_norm": 13.5, "learning_rate": 2.2924886474811995e-06, "loss": 0.4263, "step": 4448 }, { "epoch": 0.3791222837665104, "grad_norm": 11.3125, "learning_rate": 2.2923959197135033e-06, "loss": 0.4058, "step": 4449 }, { "epoch": 0.37920749893481037, "grad_norm": 15.875, "learning_rate": 2.2923031731086646e-06, "loss": 0.7828, "step": 4450 }, { "epoch": 0.37929271410311033, "grad_norm": 11.75, "learning_rate": 2.2922104076683594e-06, "loss": 0.5502, "step": 4451 }, { "epoch": 0.3793779292714103, "grad_norm": 14.25, "learning_rate": 2.292117623394264e-06, "loss": 0.7631, "step": 4452 }, { "epoch": 0.37946314443971024, "grad_norm": 18.125, "learning_rate": 2.2920248202880552e-06, "loss": 0.8136, "step": 4453 }, { "epoch": 0.3795483596080102, "grad_norm": 20.0, "learning_rate": 2.29193199835141e-06, "loss": 1.064, "step": 4454 }, { "epoch": 0.37963357477631016, "grad_norm": 12.9375, "learning_rate": 2.291839157586006e-06, "loss": 0.7133, "step": 4455 }, { "epoch": 0.3797187899446101, "grad_norm": 12.625, "learning_rate": 2.2917462979935207e-06, "loss": 0.6152, "step": 4456 }, { "epoch": 0.37980400511291007, "grad_norm": 18.75, "learning_rate": 2.2916534195756323e-06, "loss": 0.8759, "step": 4457 }, { "epoch": 0.3798892202812101, "grad_norm": 19.25, "learning_rate": 2.2915605223340193e-06, "loss": 0.964, "step": 4458 }, { "epoch": 0.37997443544951004, "grad_norm": 20.5, "learning_rate": 2.2914676062703602e-06, "loss": 0.7727, "step": 4459 }, { "epoch": 0.38005965061781, "grad_norm": 12.0625, "learning_rate": 2.2913746713863338e-06, "loss": 0.6933, "step": 4460 }, { "epoch": 0.38014486578610995, "grad_norm": 19.625, "learning_rate": 2.2912817176836204e-06, "loss": 1.1998, "step": 4461 }, { "epoch": 0.3802300809544099, "grad_norm": 12.75, "learning_rate": 2.291188745163899e-06, "loss": 0.6076, "step": 4462 }, { "epoch": 0.38031529612270987, "grad_norm": 14.4375, "learning_rate": 2.29109575382885e-06, "loss": 0.5314, "step": 4463 }, { "epoch": 0.3804005112910098, "grad_norm": 14.375, "learning_rate": 2.291002743680154e-06, "loss": 0.4575, "step": 4464 }, { "epoch": 0.3804857264593098, "grad_norm": 14.0, "learning_rate": 2.2909097147194912e-06, "loss": 1.0536, "step": 4465 }, { "epoch": 0.38057094162760974, "grad_norm": 13.0625, "learning_rate": 2.2908166669485433e-06, "loss": 0.7613, "step": 4466 }, { "epoch": 0.3806561567959097, "grad_norm": 15.9375, "learning_rate": 2.290723600368992e-06, "loss": 0.9637, "step": 4467 }, { "epoch": 0.38074137196420965, "grad_norm": 13.625, "learning_rate": 2.2906305149825185e-06, "loss": 0.8168, "step": 4468 }, { "epoch": 0.3808265871325096, "grad_norm": 12.75, "learning_rate": 2.2905374107908047e-06, "loss": 0.7389, "step": 4469 }, { "epoch": 0.38091180230080957, "grad_norm": 20.25, "learning_rate": 2.290444287795534e-06, "loss": 1.0844, "step": 4470 }, { "epoch": 0.3809970174691095, "grad_norm": 11.9375, "learning_rate": 2.290351145998389e-06, "loss": 0.4004, "step": 4471 }, { "epoch": 0.3810822326374095, "grad_norm": 15.3125, "learning_rate": 2.290257985401052e-06, "loss": 0.5302, "step": 4472 }, { "epoch": 0.38116744780570944, "grad_norm": 13.625, "learning_rate": 2.2901648060052077e-06, "loss": 0.711, "step": 4473 }, { "epoch": 0.3812526629740094, "grad_norm": 23.625, "learning_rate": 2.290071607812539e-06, "loss": 0.9568, "step": 4474 }, { "epoch": 0.38133787814230935, "grad_norm": 15.3125, "learning_rate": 2.289978390824731e-06, "loss": 0.8809, "step": 4475 }, { "epoch": 0.3814230933106093, "grad_norm": 16.625, "learning_rate": 2.289885155043467e-06, "loss": 0.6751, "step": 4476 }, { "epoch": 0.38150830847890926, "grad_norm": 17.625, "learning_rate": 2.289791900470433e-06, "loss": 1.0595, "step": 4477 }, { "epoch": 0.3815935236472092, "grad_norm": 15.75, "learning_rate": 2.289698627107314e-06, "loss": 0.8042, "step": 4478 }, { "epoch": 0.3816787388155092, "grad_norm": 17.875, "learning_rate": 2.289605334955795e-06, "loss": 1.0193, "step": 4479 }, { "epoch": 0.38176395398380913, "grad_norm": 14.75, "learning_rate": 2.2895120240175622e-06, "loss": 0.6235, "step": 4480 }, { "epoch": 0.3818491691521091, "grad_norm": 12.5, "learning_rate": 2.2894186942943015e-06, "loss": 0.6292, "step": 4481 }, { "epoch": 0.38193438432040905, "grad_norm": 12.0625, "learning_rate": 2.2893253457877e-06, "loss": 0.4333, "step": 4482 }, { "epoch": 0.382019599488709, "grad_norm": 18.625, "learning_rate": 2.2892319784994447e-06, "loss": 0.7367, "step": 4483 }, { "epoch": 0.38210481465700896, "grad_norm": 15.5625, "learning_rate": 2.2891385924312224e-06, "loss": 0.7588, "step": 4484 }, { "epoch": 0.3821900298253089, "grad_norm": 15.3125, "learning_rate": 2.2890451875847205e-06, "loss": 0.7638, "step": 4485 }, { "epoch": 0.3822752449936089, "grad_norm": 16.375, "learning_rate": 2.2889517639616277e-06, "loss": 0.916, "step": 4486 }, { "epoch": 0.38236046016190883, "grad_norm": 14.125, "learning_rate": 2.288858321563631e-06, "loss": 0.5469, "step": 4487 }, { "epoch": 0.3824456753302088, "grad_norm": 13.0625, "learning_rate": 2.2887648603924206e-06, "loss": 0.4791, "step": 4488 }, { "epoch": 0.38253089049850875, "grad_norm": 19.375, "learning_rate": 2.288671380449684e-06, "loss": 1.0754, "step": 4489 }, { "epoch": 0.3826161056668087, "grad_norm": 19.625, "learning_rate": 2.2885778817371115e-06, "loss": 0.995, "step": 4490 }, { "epoch": 0.38270132083510866, "grad_norm": 12.6875, "learning_rate": 2.2884843642563924e-06, "loss": 0.4355, "step": 4491 }, { "epoch": 0.3827865360034086, "grad_norm": 17.0, "learning_rate": 2.288390828009216e-06, "loss": 1.1606, "step": 4492 }, { "epoch": 0.3828717511717086, "grad_norm": 15.4375, "learning_rate": 2.2882972729972737e-06, "loss": 0.7709, "step": 4493 }, { "epoch": 0.38295696634000853, "grad_norm": 12.8125, "learning_rate": 2.2882036992222554e-06, "loss": 0.7014, "step": 4494 }, { "epoch": 0.3830421815083085, "grad_norm": 13.75, "learning_rate": 2.288110106685852e-06, "loss": 0.687, "step": 4495 }, { "epoch": 0.38312739667660844, "grad_norm": 19.375, "learning_rate": 2.2880164953897556e-06, "loss": 1.0258, "step": 4496 }, { "epoch": 0.3832126118449084, "grad_norm": 19.0, "learning_rate": 2.287922865335657e-06, "loss": 1.1486, "step": 4497 }, { "epoch": 0.38329782701320836, "grad_norm": 24.5, "learning_rate": 2.2878292165252486e-06, "loss": 1.1778, "step": 4498 }, { "epoch": 0.3833830421815083, "grad_norm": 11.0, "learning_rate": 2.2877355489602223e-06, "loss": 0.6552, "step": 4499 }, { "epoch": 0.3834682573498083, "grad_norm": 14.1875, "learning_rate": 2.2876418626422716e-06, "loss": 0.6636, "step": 4500 }, { "epoch": 0.38355347251810823, "grad_norm": 14.625, "learning_rate": 2.2875481575730883e-06, "loss": 0.6893, "step": 4501 }, { "epoch": 0.3836386876864082, "grad_norm": 14.4375, "learning_rate": 2.2874544337543674e-06, "loss": 0.8543, "step": 4502 }, { "epoch": 0.38372390285470814, "grad_norm": 10.9375, "learning_rate": 2.287360691187801e-06, "loss": 0.5006, "step": 4503 }, { "epoch": 0.3838091180230081, "grad_norm": 14.1875, "learning_rate": 2.287266929875084e-06, "loss": 0.5273, "step": 4504 }, { "epoch": 0.38389433319130806, "grad_norm": 18.875, "learning_rate": 2.2871731498179105e-06, "loss": 0.9154, "step": 4505 }, { "epoch": 0.383979548359608, "grad_norm": 16.875, "learning_rate": 2.287079351017975e-06, "loss": 0.9958, "step": 4506 }, { "epoch": 0.38406476352790797, "grad_norm": 18.25, "learning_rate": 2.286985533476973e-06, "loss": 0.9313, "step": 4507 }, { "epoch": 0.3841499786962079, "grad_norm": 30.875, "learning_rate": 2.2868916971966e-06, "loss": 0.5992, "step": 4508 }, { "epoch": 0.3842351938645079, "grad_norm": 12.1875, "learning_rate": 2.286797842178551e-06, "loss": 0.495, "step": 4509 }, { "epoch": 0.38432040903280784, "grad_norm": 15.375, "learning_rate": 2.2867039684245224e-06, "loss": 0.8634, "step": 4510 }, { "epoch": 0.3844056242011078, "grad_norm": 13.5, "learning_rate": 2.2866100759362103e-06, "loss": 0.61, "step": 4511 }, { "epoch": 0.38449083936940776, "grad_norm": 13.4375, "learning_rate": 2.2865161647153124e-06, "loss": 1.007, "step": 4512 }, { "epoch": 0.3845760545377077, "grad_norm": 13.1875, "learning_rate": 2.2864222347635246e-06, "loss": 0.7041, "step": 4513 }, { "epoch": 0.38466126970600767, "grad_norm": 13.3125, "learning_rate": 2.2863282860825446e-06, "loss": 0.6604, "step": 4514 }, { "epoch": 0.3847464848743076, "grad_norm": 14.5, "learning_rate": 2.286234318674071e-06, "loss": 0.8003, "step": 4515 }, { "epoch": 0.3848317000426076, "grad_norm": 15.75, "learning_rate": 2.2861403325398008e-06, "loss": 0.9918, "step": 4516 }, { "epoch": 0.38491691521090754, "grad_norm": 17.0, "learning_rate": 2.286046327681433e-06, "loss": 0.8309, "step": 4517 }, { "epoch": 0.3850021303792075, "grad_norm": 14.125, "learning_rate": 2.2859523041006665e-06, "loss": 0.8778, "step": 4518 }, { "epoch": 0.38508734554750745, "grad_norm": 24.625, "learning_rate": 2.2858582617992e-06, "loss": 1.1862, "step": 4519 }, { "epoch": 0.3851725607158074, "grad_norm": 16.25, "learning_rate": 2.2857642007787324e-06, "loss": 0.7936, "step": 4520 }, { "epoch": 0.38525777588410737, "grad_norm": 15.5625, "learning_rate": 2.285670121040965e-06, "loss": 1.0927, "step": 4521 }, { "epoch": 0.3853429910524073, "grad_norm": 13.125, "learning_rate": 2.2855760225875967e-06, "loss": 0.523, "step": 4522 }, { "epoch": 0.3854282062207073, "grad_norm": 17.75, "learning_rate": 2.2854819054203285e-06, "loss": 1.069, "step": 4523 }, { "epoch": 0.38551342138900724, "grad_norm": 14.1875, "learning_rate": 2.2853877695408608e-06, "loss": 0.79, "step": 4524 }, { "epoch": 0.3855986365573072, "grad_norm": 13.0, "learning_rate": 2.285293614950895e-06, "loss": 0.7708, "step": 4525 }, { "epoch": 0.38568385172560715, "grad_norm": 17.625, "learning_rate": 2.285199441652133e-06, "loss": 1.1711, "step": 4526 }, { "epoch": 0.3857690668939071, "grad_norm": 14.5625, "learning_rate": 2.2851052496462754e-06, "loss": 0.8402, "step": 4527 }, { "epoch": 0.38585428206220707, "grad_norm": 12.3125, "learning_rate": 2.2850110389350254e-06, "loss": 0.5961, "step": 4528 }, { "epoch": 0.385939497230507, "grad_norm": 16.875, "learning_rate": 2.284916809520085e-06, "loss": 0.9529, "step": 4529 }, { "epoch": 0.386024712398807, "grad_norm": 12.6875, "learning_rate": 2.2848225614031573e-06, "loss": 0.6457, "step": 4530 }, { "epoch": 0.38610992756710694, "grad_norm": 16.75, "learning_rate": 2.284728294585945e-06, "loss": 0.9751, "step": 4531 }, { "epoch": 0.3861951427354069, "grad_norm": 11.8125, "learning_rate": 2.2846340090701522e-06, "loss": 0.5709, "step": 4532 }, { "epoch": 0.38628035790370685, "grad_norm": 15.5, "learning_rate": 2.2845397048574823e-06, "loss": 0.8488, "step": 4533 }, { "epoch": 0.3863655730720068, "grad_norm": 12.6875, "learning_rate": 2.28444538194964e-06, "loss": 0.5948, "step": 4534 }, { "epoch": 0.38645078824030676, "grad_norm": 14.625, "learning_rate": 2.284351040348329e-06, "loss": 0.7408, "step": 4535 }, { "epoch": 0.3865360034086067, "grad_norm": 13.1875, "learning_rate": 2.284256680055255e-06, "loss": 0.8908, "step": 4536 }, { "epoch": 0.3866212185769067, "grad_norm": 19.375, "learning_rate": 2.2841623010721228e-06, "loss": 1.0765, "step": 4537 }, { "epoch": 0.38670643374520663, "grad_norm": 13.125, "learning_rate": 2.2840679034006373e-06, "loss": 0.9415, "step": 4538 }, { "epoch": 0.3867916489135066, "grad_norm": 17.75, "learning_rate": 2.283973487042505e-06, "loss": 0.945, "step": 4539 }, { "epoch": 0.38687686408180655, "grad_norm": 20.25, "learning_rate": 2.2838790519994326e-06, "loss": 1.1383, "step": 4540 }, { "epoch": 0.3869620792501065, "grad_norm": 23.125, "learning_rate": 2.2837845982731257e-06, "loss": 0.7768, "step": 4541 }, { "epoch": 0.38704729441840646, "grad_norm": 12.875, "learning_rate": 2.2836901258652915e-06, "loss": 0.7491, "step": 4542 }, { "epoch": 0.3871325095867064, "grad_norm": 17.25, "learning_rate": 2.2835956347776377e-06, "loss": 0.8438, "step": 4543 }, { "epoch": 0.3872177247550064, "grad_norm": 16.625, "learning_rate": 2.283501125011871e-06, "loss": 0.7922, "step": 4544 }, { "epoch": 0.38730293992330633, "grad_norm": 16.125, "learning_rate": 2.2834065965697e-06, "loss": 0.4668, "step": 4545 }, { "epoch": 0.3873881550916063, "grad_norm": 14.375, "learning_rate": 2.2833120494528325e-06, "loss": 0.7096, "step": 4546 }, { "epoch": 0.38747337025990625, "grad_norm": 18.625, "learning_rate": 2.2832174836629774e-06, "loss": 0.7342, "step": 4547 }, { "epoch": 0.3875585854282062, "grad_norm": 14.1875, "learning_rate": 2.283122899201843e-06, "loss": 0.7097, "step": 4548 }, { "epoch": 0.38764380059650616, "grad_norm": 16.125, "learning_rate": 2.283028296071139e-06, "loss": 1.0786, "step": 4549 }, { "epoch": 0.3877290157648061, "grad_norm": 13.75, "learning_rate": 2.282933674272575e-06, "loss": 0.791, "step": 4550 }, { "epoch": 0.3878142309331061, "grad_norm": 11.8125, "learning_rate": 2.282839033807861e-06, "loss": 0.5015, "step": 4551 }, { "epoch": 0.38789944610140603, "grad_norm": 18.625, "learning_rate": 2.2827443746787068e-06, "loss": 1.1809, "step": 4552 }, { "epoch": 0.387984661269706, "grad_norm": 12.0, "learning_rate": 2.2826496968868234e-06, "loss": 0.524, "step": 4553 }, { "epoch": 0.38806987643800595, "grad_norm": 16.375, "learning_rate": 2.282555000433921e-06, "loss": 0.8558, "step": 4554 }, { "epoch": 0.3881550916063059, "grad_norm": 12.625, "learning_rate": 2.282460285321712e-06, "loss": 0.5823, "step": 4555 }, { "epoch": 0.38824030677460586, "grad_norm": 10.5, "learning_rate": 2.2823655515519076e-06, "loss": 0.5241, "step": 4556 }, { "epoch": 0.3883255219429058, "grad_norm": 13.3125, "learning_rate": 2.2822707991262196e-06, "loss": 0.6566, "step": 4557 }, { "epoch": 0.3884107371112058, "grad_norm": 11.5625, "learning_rate": 2.2821760280463596e-06, "loss": 0.5527, "step": 4558 }, { "epoch": 0.38849595227950573, "grad_norm": 15.0625, "learning_rate": 2.2820812383140413e-06, "loss": 0.8569, "step": 4559 }, { "epoch": 0.3885811674478057, "grad_norm": 14.1875, "learning_rate": 2.281986429930977e-06, "loss": 0.8562, "step": 4560 }, { "epoch": 0.38866638261610564, "grad_norm": 12.9375, "learning_rate": 2.2818916028988798e-06, "loss": 0.8177, "step": 4561 }, { "epoch": 0.3887515977844056, "grad_norm": 15.3125, "learning_rate": 2.2817967572194644e-06, "loss": 1.1979, "step": 4562 }, { "epoch": 0.38883681295270556, "grad_norm": 16.0, "learning_rate": 2.2817018928944433e-06, "loss": 0.8594, "step": 4563 }, { "epoch": 0.3889220281210055, "grad_norm": 14.0, "learning_rate": 2.2816070099255324e-06, "loss": 0.9151, "step": 4564 }, { "epoch": 0.38900724328930547, "grad_norm": 10.75, "learning_rate": 2.281512108314445e-06, "loss": 0.3492, "step": 4565 }, { "epoch": 0.38909245845760543, "grad_norm": 15.6875, "learning_rate": 2.2814171880628964e-06, "loss": 0.8376, "step": 4566 }, { "epoch": 0.3891776736259054, "grad_norm": 14.1875, "learning_rate": 2.2813222491726022e-06, "loss": 1.0273, "step": 4567 }, { "epoch": 0.38926288879420534, "grad_norm": 14.5, "learning_rate": 2.281227291645278e-06, "loss": 0.8885, "step": 4568 }, { "epoch": 0.3893481039625053, "grad_norm": 10.3125, "learning_rate": 2.28113231548264e-06, "loss": 0.3477, "step": 4569 }, { "epoch": 0.3894333191308053, "grad_norm": 16.0, "learning_rate": 2.2810373206864034e-06, "loss": 0.7715, "step": 4570 }, { "epoch": 0.38951853429910527, "grad_norm": 16.75, "learning_rate": 2.280942307258286e-06, "loss": 1.1484, "step": 4571 }, { "epoch": 0.3896037494674052, "grad_norm": 14.25, "learning_rate": 2.2808472752000043e-06, "loss": 0.78, "step": 4572 }, { "epoch": 0.3896889646357052, "grad_norm": 15.375, "learning_rate": 2.280752224513276e-06, "loss": 0.635, "step": 4573 }, { "epoch": 0.38977417980400514, "grad_norm": 35.25, "learning_rate": 2.280657155199818e-06, "loss": 0.9081, "step": 4574 }, { "epoch": 0.3898593949723051, "grad_norm": 11.375, "learning_rate": 2.2805620672613494e-06, "loss": 0.5269, "step": 4575 }, { "epoch": 0.38994461014060505, "grad_norm": 15.5, "learning_rate": 2.280466960699588e-06, "loss": 0.7133, "step": 4576 }, { "epoch": 0.390029825308905, "grad_norm": 13.8125, "learning_rate": 2.280371835516252e-06, "loss": 0.7546, "step": 4577 }, { "epoch": 0.39011504047720497, "grad_norm": 14.0625, "learning_rate": 2.280276691713061e-06, "loss": 0.6967, "step": 4578 }, { "epoch": 0.3902002556455049, "grad_norm": 17.375, "learning_rate": 2.280181529291734e-06, "loss": 1.1751, "step": 4579 }, { "epoch": 0.3902854708138049, "grad_norm": 18.75, "learning_rate": 2.2800863482539905e-06, "loss": 0.8412, "step": 4580 }, { "epoch": 0.39037068598210484, "grad_norm": 11.8125, "learning_rate": 2.279991148601551e-06, "loss": 0.516, "step": 4581 }, { "epoch": 0.3904559011504048, "grad_norm": 10.5, "learning_rate": 2.2798959303361364e-06, "loss": 0.4011, "step": 4582 }, { "epoch": 0.39054111631870475, "grad_norm": 13.75, "learning_rate": 2.2798006934594663e-06, "loss": 0.6922, "step": 4583 }, { "epoch": 0.3906263314870047, "grad_norm": 14.5, "learning_rate": 2.279705437973262e-06, "loss": 0.7145, "step": 4584 }, { "epoch": 0.39071154665530466, "grad_norm": 15.25, "learning_rate": 2.279610163879245e-06, "loss": 0.7449, "step": 4585 }, { "epoch": 0.3907967618236046, "grad_norm": 16.5, "learning_rate": 2.279514871179137e-06, "loss": 0.9176, "step": 4586 }, { "epoch": 0.3908819769919046, "grad_norm": 11.625, "learning_rate": 2.27941955987466e-06, "loss": 0.5752, "step": 4587 }, { "epoch": 0.39096719216020454, "grad_norm": 14.9375, "learning_rate": 2.2793242299675366e-06, "loss": 0.8102, "step": 4588 }, { "epoch": 0.3910524073285045, "grad_norm": 15.375, "learning_rate": 2.279228881459489e-06, "loss": 0.8104, "step": 4589 }, { "epoch": 0.39113762249680445, "grad_norm": 12.875, "learning_rate": 2.279133514352241e-06, "loss": 0.6299, "step": 4590 }, { "epoch": 0.3912228376651044, "grad_norm": 12.3125, "learning_rate": 2.279038128647515e-06, "loss": 0.6081, "step": 4591 }, { "epoch": 0.39130805283340436, "grad_norm": 21.25, "learning_rate": 2.2789427243470356e-06, "loss": 0.8856, "step": 4592 }, { "epoch": 0.3913932680017043, "grad_norm": 12.8125, "learning_rate": 2.278847301452526e-06, "loss": 0.7382, "step": 4593 }, { "epoch": 0.3914784831700043, "grad_norm": 14.25, "learning_rate": 2.2787518599657116e-06, "loss": 0.9935, "step": 4594 }, { "epoch": 0.39156369833830423, "grad_norm": 15.0, "learning_rate": 2.2786563998883167e-06, "loss": 0.3733, "step": 4595 }, { "epoch": 0.3916489135066042, "grad_norm": 9.25, "learning_rate": 2.278560921222066e-06, "loss": 0.3683, "step": 4596 }, { "epoch": 0.39173412867490415, "grad_norm": 10.8125, "learning_rate": 2.278465423968685e-06, "loss": 0.5061, "step": 4597 }, { "epoch": 0.3918193438432041, "grad_norm": 14.5625, "learning_rate": 2.2783699081298995e-06, "loss": 0.7671, "step": 4598 }, { "epoch": 0.39190455901150406, "grad_norm": 15.1875, "learning_rate": 2.278274373707436e-06, "loss": 1.0247, "step": 4599 }, { "epoch": 0.391989774179804, "grad_norm": 16.5, "learning_rate": 2.2781788207030203e-06, "loss": 0.7135, "step": 4600 }, { "epoch": 0.392074989348104, "grad_norm": 17.25, "learning_rate": 2.2780832491183796e-06, "loss": 0.9954, "step": 4601 }, { "epoch": 0.39216020451640393, "grad_norm": 13.25, "learning_rate": 2.277987658955241e-06, "loss": 0.7754, "step": 4602 }, { "epoch": 0.3922454196847039, "grad_norm": 15.875, "learning_rate": 2.277892050215331e-06, "loss": 0.8307, "step": 4603 }, { "epoch": 0.39233063485300385, "grad_norm": 15.75, "learning_rate": 2.2777964229003786e-06, "loss": 0.7016, "step": 4604 }, { "epoch": 0.3924158500213038, "grad_norm": 16.0, "learning_rate": 2.277700777012111e-06, "loss": 0.91, "step": 4605 }, { "epoch": 0.39250106518960376, "grad_norm": 15.1875, "learning_rate": 2.277605112552257e-06, "loss": 0.6765, "step": 4606 }, { "epoch": 0.3925862803579037, "grad_norm": 11.25, "learning_rate": 2.2775094295225452e-06, "loss": 0.4097, "step": 4607 }, { "epoch": 0.3926714955262037, "grad_norm": 16.875, "learning_rate": 2.277413727924705e-06, "loss": 0.9516, "step": 4608 }, { "epoch": 0.39275671069450363, "grad_norm": 15.6875, "learning_rate": 2.2773180077604657e-06, "loss": 0.8189, "step": 4609 }, { "epoch": 0.3928419258628036, "grad_norm": 14.1875, "learning_rate": 2.2772222690315564e-06, "loss": 0.5581, "step": 4610 }, { "epoch": 0.39292714103110354, "grad_norm": 13.0625, "learning_rate": 2.2771265117397083e-06, "loss": 0.5962, "step": 4611 }, { "epoch": 0.3930123561994035, "grad_norm": 21.125, "learning_rate": 2.277030735886651e-06, "loss": 0.6761, "step": 4612 }, { "epoch": 0.39309757136770346, "grad_norm": 12.375, "learning_rate": 2.276934941474116e-06, "loss": 0.6512, "step": 4613 }, { "epoch": 0.3931827865360034, "grad_norm": 18.375, "learning_rate": 2.2768391285038334e-06, "loss": 0.5349, "step": 4614 }, { "epoch": 0.39326800170430337, "grad_norm": 17.5, "learning_rate": 2.2767432969775354e-06, "loss": 0.9304, "step": 4615 }, { "epoch": 0.39335321687260333, "grad_norm": 17.375, "learning_rate": 2.2766474468969536e-06, "loss": 0.5558, "step": 4616 }, { "epoch": 0.3934384320409033, "grad_norm": 30.875, "learning_rate": 2.2765515782638197e-06, "loss": 0.9864, "step": 4617 }, { "epoch": 0.39352364720920324, "grad_norm": 13.0, "learning_rate": 2.2764556910798668e-06, "loss": 0.6642, "step": 4618 }, { "epoch": 0.3936088623775032, "grad_norm": 16.5, "learning_rate": 2.2763597853468274e-06, "loss": 0.7602, "step": 4619 }, { "epoch": 0.39369407754580316, "grad_norm": 21.875, "learning_rate": 2.2762638610664343e-06, "loss": 0.8022, "step": 4620 }, { "epoch": 0.3937792927141031, "grad_norm": 14.5, "learning_rate": 2.2761679182404216e-06, "loss": 0.8596, "step": 4621 }, { "epoch": 0.39386450788240307, "grad_norm": 15.625, "learning_rate": 2.2760719568705225e-06, "loss": 0.8398, "step": 4622 }, { "epoch": 0.393949723050703, "grad_norm": 12.6875, "learning_rate": 2.275975976958471e-06, "loss": 0.7829, "step": 4623 }, { "epoch": 0.394034938219003, "grad_norm": 11.5625, "learning_rate": 2.2758799785060028e-06, "loss": 0.4472, "step": 4624 }, { "epoch": 0.39412015338730294, "grad_norm": 16.625, "learning_rate": 2.2757839615148512e-06, "loss": 0.8803, "step": 4625 }, { "epoch": 0.3942053685556029, "grad_norm": 15.625, "learning_rate": 2.275687925986752e-06, "loss": 0.7726, "step": 4626 }, { "epoch": 0.39429058372390285, "grad_norm": 10.25, "learning_rate": 2.2755918719234405e-06, "loss": 0.3376, "step": 4627 }, { "epoch": 0.3943757988922028, "grad_norm": 14.8125, "learning_rate": 2.275495799326653e-06, "loss": 0.8363, "step": 4628 }, { "epoch": 0.39446101406050277, "grad_norm": 19.625, "learning_rate": 2.275399708198125e-06, "loss": 1.1049, "step": 4629 }, { "epoch": 0.3945462292288027, "grad_norm": 10.9375, "learning_rate": 2.2753035985395928e-06, "loss": 0.5373, "step": 4630 }, { "epoch": 0.3946314443971027, "grad_norm": 16.75, "learning_rate": 2.275207470352794e-06, "loss": 0.8961, "step": 4631 }, { "epoch": 0.39471665956540264, "grad_norm": 16.125, "learning_rate": 2.2751113236394647e-06, "loss": 0.7463, "step": 4632 }, { "epoch": 0.3948018747337026, "grad_norm": 11.3125, "learning_rate": 2.2750151584013432e-06, "loss": 0.5158, "step": 4633 }, { "epoch": 0.39488708990200255, "grad_norm": 16.0, "learning_rate": 2.274918974640167e-06, "loss": 0.7595, "step": 4634 }, { "epoch": 0.3949723050703025, "grad_norm": 17.0, "learning_rate": 2.2748227723576747e-06, "loss": 1.0543, "step": 4635 }, { "epoch": 0.39505752023860247, "grad_norm": 27.0, "learning_rate": 2.274726551555604e-06, "loss": 1.278, "step": 4636 }, { "epoch": 0.3951427354069024, "grad_norm": 12.625, "learning_rate": 2.2746303122356946e-06, "loss": 0.6901, "step": 4637 }, { "epoch": 0.3952279505752024, "grad_norm": 16.5, "learning_rate": 2.2745340543996844e-06, "loss": 0.9716, "step": 4638 }, { "epoch": 0.39531316574350234, "grad_norm": 13.0, "learning_rate": 2.2744377780493136e-06, "loss": 0.6546, "step": 4639 }, { "epoch": 0.3953983809118023, "grad_norm": 19.5, "learning_rate": 2.2743414831863225e-06, "loss": 1.1621, "step": 4640 }, { "epoch": 0.39548359608010225, "grad_norm": 15.3125, "learning_rate": 2.2742451698124507e-06, "loss": 0.7983, "step": 4641 }, { "epoch": 0.3955688112484022, "grad_norm": 13.75, "learning_rate": 2.2741488379294386e-06, "loss": 0.6898, "step": 4642 }, { "epoch": 0.39565402641670216, "grad_norm": 14.3125, "learning_rate": 2.274052487539027e-06, "loss": 0.715, "step": 4643 }, { "epoch": 0.3957392415850021, "grad_norm": 12.875, "learning_rate": 2.273956118642957e-06, "loss": 0.4977, "step": 4644 }, { "epoch": 0.3958244567533021, "grad_norm": 19.25, "learning_rate": 2.2738597312429706e-06, "loss": 0.599, "step": 4645 }, { "epoch": 0.39590967192160204, "grad_norm": 12.8125, "learning_rate": 2.273763325340809e-06, "loss": 0.5835, "step": 4646 }, { "epoch": 0.395994887089902, "grad_norm": 10.9375, "learning_rate": 2.273666900938215e-06, "loss": 0.4101, "step": 4647 }, { "epoch": 0.39608010225820195, "grad_norm": 11.8125, "learning_rate": 2.2735704580369305e-06, "loss": 0.5664, "step": 4648 }, { "epoch": 0.3961653174265019, "grad_norm": 13.3125, "learning_rate": 2.273473996638699e-06, "loss": 0.3924, "step": 4649 }, { "epoch": 0.39625053259480186, "grad_norm": 10.0625, "learning_rate": 2.2733775167452627e-06, "loss": 0.4271, "step": 4650 }, { "epoch": 0.3963357477631018, "grad_norm": 20.625, "learning_rate": 2.273281018358366e-06, "loss": 1.0851, "step": 4651 }, { "epoch": 0.3964209629314018, "grad_norm": 12.125, "learning_rate": 2.2731845014797515e-06, "loss": 0.5392, "step": 4652 }, { "epoch": 0.39650617809970173, "grad_norm": 13.5625, "learning_rate": 2.273087966111165e-06, "loss": 0.6814, "step": 4653 }, { "epoch": 0.3965913932680017, "grad_norm": 14.75, "learning_rate": 2.27299141225435e-06, "loss": 0.7036, "step": 4654 }, { "epoch": 0.39667660843630165, "grad_norm": 21.375, "learning_rate": 2.272894839911051e-06, "loss": 0.9787, "step": 4655 }, { "epoch": 0.3967618236046016, "grad_norm": 15.6875, "learning_rate": 2.2727982490830143e-06, "loss": 0.8725, "step": 4656 }, { "epoch": 0.39684703877290156, "grad_norm": 16.125, "learning_rate": 2.272701639771984e-06, "loss": 1.0321, "step": 4657 }, { "epoch": 0.3969322539412015, "grad_norm": 12.3125, "learning_rate": 2.2726050119797072e-06, "loss": 0.6422, "step": 4658 }, { "epoch": 0.3970174691095015, "grad_norm": 12.6875, "learning_rate": 2.2725083657079294e-06, "loss": 0.5762, "step": 4659 }, { "epoch": 0.39710268427780143, "grad_norm": 14.75, "learning_rate": 2.272411700958397e-06, "loss": 0.7513, "step": 4660 }, { "epoch": 0.3971878994461014, "grad_norm": 18.125, "learning_rate": 2.2723150177328575e-06, "loss": 0.9375, "step": 4661 }, { "epoch": 0.39727311461440135, "grad_norm": 16.25, "learning_rate": 2.2722183160330576e-06, "loss": 0.8393, "step": 4662 }, { "epoch": 0.3973583297827013, "grad_norm": 17.25, "learning_rate": 2.2721215958607446e-06, "loss": 1.2914, "step": 4663 }, { "epoch": 0.39744354495100126, "grad_norm": 11.75, "learning_rate": 2.272024857217666e-06, "loss": 0.6641, "step": 4664 }, { "epoch": 0.3975287601193012, "grad_norm": 12.875, "learning_rate": 2.2719281001055717e-06, "loss": 0.6941, "step": 4665 }, { "epoch": 0.3976139752876012, "grad_norm": 13.3125, "learning_rate": 2.271831324526208e-06, "loss": 0.7783, "step": 4666 }, { "epoch": 0.39769919045590113, "grad_norm": 12.5625, "learning_rate": 2.271734530481325e-06, "loss": 0.7481, "step": 4667 }, { "epoch": 0.3977844056242011, "grad_norm": 13.3125, "learning_rate": 2.2716377179726716e-06, "loss": 0.8422, "step": 4668 }, { "epoch": 0.39786962079250104, "grad_norm": 18.375, "learning_rate": 2.2715408870019977e-06, "loss": 1.0709, "step": 4669 }, { "epoch": 0.397954835960801, "grad_norm": 14.6875, "learning_rate": 2.271444037571052e-06, "loss": 0.7559, "step": 4670 }, { "epoch": 0.39804005112910096, "grad_norm": 13.4375, "learning_rate": 2.2713471696815863e-06, "loss": 0.7677, "step": 4671 }, { "epoch": 0.3981252662974009, "grad_norm": 12.0, "learning_rate": 2.2712502833353496e-06, "loss": 0.7987, "step": 4672 }, { "epoch": 0.39821048146570087, "grad_norm": 9.5, "learning_rate": 2.2711533785340933e-06, "loss": 0.231, "step": 4673 }, { "epoch": 0.39829569663400083, "grad_norm": 16.625, "learning_rate": 2.2710564552795686e-06, "loss": 0.8606, "step": 4674 }, { "epoch": 0.3983809118023008, "grad_norm": 15.1875, "learning_rate": 2.270959513573527e-06, "loss": 0.8815, "step": 4675 }, { "epoch": 0.39846612697060074, "grad_norm": 12.75, "learning_rate": 2.270862553417721e-06, "loss": 0.4789, "step": 4676 }, { "epoch": 0.3985513421389007, "grad_norm": 14.1875, "learning_rate": 2.2707655748139014e-06, "loss": 0.7227, "step": 4677 }, { "epoch": 0.39863655730720066, "grad_norm": 15.6875, "learning_rate": 2.2706685777638217e-06, "loss": 0.7824, "step": 4678 }, { "epoch": 0.3987217724755006, "grad_norm": 19.75, "learning_rate": 2.2705715622692344e-06, "loss": 0.8465, "step": 4679 }, { "epoch": 0.39880698764380057, "grad_norm": 15.4375, "learning_rate": 2.2704745283318925e-06, "loss": 0.7303, "step": 4680 }, { "epoch": 0.3988922028121005, "grad_norm": 15.125, "learning_rate": 2.2703774759535495e-06, "loss": 0.8414, "step": 4681 }, { "epoch": 0.3989774179804005, "grad_norm": 17.25, "learning_rate": 2.27028040513596e-06, "loss": 0.7061, "step": 4682 }, { "epoch": 0.3990626331487005, "grad_norm": 14.75, "learning_rate": 2.2701833158808774e-06, "loss": 1.028, "step": 4683 }, { "epoch": 0.39914784831700045, "grad_norm": 12.5625, "learning_rate": 2.2700862081900564e-06, "loss": 0.3536, "step": 4684 }, { "epoch": 0.3992330634853004, "grad_norm": 12.25, "learning_rate": 2.269989082065252e-06, "loss": 0.5192, "step": 4685 }, { "epoch": 0.39931827865360037, "grad_norm": 16.5, "learning_rate": 2.269891937508219e-06, "loss": 0.7274, "step": 4686 }, { "epoch": 0.3994034938219003, "grad_norm": 12.1875, "learning_rate": 2.269794774520713e-06, "loss": 0.6608, "step": 4687 }, { "epoch": 0.3994887089902003, "grad_norm": 19.75, "learning_rate": 2.26969759310449e-06, "loss": 0.9892, "step": 4688 }, { "epoch": 0.39957392415850024, "grad_norm": 18.25, "learning_rate": 2.269600393261306e-06, "loss": 0.7239, "step": 4689 }, { "epoch": 0.3996591393268002, "grad_norm": 15.5625, "learning_rate": 2.269503174992918e-06, "loss": 0.7861, "step": 4690 }, { "epoch": 0.39974435449510015, "grad_norm": 8.625, "learning_rate": 2.269405938301082e-06, "loss": 0.3024, "step": 4691 }, { "epoch": 0.3998295696634001, "grad_norm": 12.875, "learning_rate": 2.269308683187556e-06, "loss": 0.8666, "step": 4692 }, { "epoch": 0.39991478483170007, "grad_norm": 35.5, "learning_rate": 2.2692114096540966e-06, "loss": 1.2893, "step": 4693 }, { "epoch": 0.4, "grad_norm": 14.75, "learning_rate": 2.269114117702463e-06, "loss": 0.9085, "step": 4694 }, { "epoch": 0.4000852151683, "grad_norm": 14.9375, "learning_rate": 2.269016807334412e-06, "loss": 0.967, "step": 4695 }, { "epoch": 0.40017043033659994, "grad_norm": 12.4375, "learning_rate": 2.2689194785517023e-06, "loss": 0.7117, "step": 4696 }, { "epoch": 0.4002556455048999, "grad_norm": 15.4375, "learning_rate": 2.268822131356093e-06, "loss": 0.9099, "step": 4697 }, { "epoch": 0.40034086067319985, "grad_norm": 14.0625, "learning_rate": 2.2687247657493437e-06, "loss": 0.5604, "step": 4698 }, { "epoch": 0.4004260758414998, "grad_norm": 12.625, "learning_rate": 2.2686273817332136e-06, "loss": 0.3514, "step": 4699 }, { "epoch": 0.40051129100979976, "grad_norm": 18.5, "learning_rate": 2.268529979309462e-06, "loss": 0.7446, "step": 4700 }, { "epoch": 0.4005965061780997, "grad_norm": 16.0, "learning_rate": 2.2684325584798496e-06, "loss": 0.9357, "step": 4701 }, { "epoch": 0.4006817213463997, "grad_norm": 13.375, "learning_rate": 2.268335119246137e-06, "loss": 0.587, "step": 4702 }, { "epoch": 0.40076693651469963, "grad_norm": 13.3125, "learning_rate": 2.2682376616100842e-06, "loss": 0.8557, "step": 4703 }, { "epoch": 0.4008521516829996, "grad_norm": 13.1875, "learning_rate": 2.2681401855734535e-06, "loss": 0.6055, "step": 4704 }, { "epoch": 0.40093736685129955, "grad_norm": 11.125, "learning_rate": 2.2680426911380055e-06, "loss": 0.4985, "step": 4705 }, { "epoch": 0.4010225820195995, "grad_norm": 10.8125, "learning_rate": 2.2679451783055025e-06, "loss": 0.4013, "step": 4706 }, { "epoch": 0.40110779718789946, "grad_norm": 16.5, "learning_rate": 2.2678476470777067e-06, "loss": 0.9624, "step": 4707 }, { "epoch": 0.4011930123561994, "grad_norm": 13.9375, "learning_rate": 2.26775009745638e-06, "loss": 0.7059, "step": 4708 }, { "epoch": 0.4012782275244994, "grad_norm": 14.3125, "learning_rate": 2.2676525294432853e-06, "loss": 0.8862, "step": 4709 }, { "epoch": 0.40136344269279933, "grad_norm": 21.875, "learning_rate": 2.267554943040187e-06, "loss": 1.2246, "step": 4710 }, { "epoch": 0.4014486578610993, "grad_norm": 12.25, "learning_rate": 2.267457338248847e-06, "loss": 0.5507, "step": 4711 }, { "epoch": 0.40153387302939925, "grad_norm": 18.625, "learning_rate": 2.2673597150710302e-06, "loss": 0.7336, "step": 4712 }, { "epoch": 0.4016190881976992, "grad_norm": 18.125, "learning_rate": 2.2672620735085e-06, "loss": 0.8719, "step": 4713 }, { "epoch": 0.40170430336599916, "grad_norm": 19.125, "learning_rate": 2.267164413563021e-06, "loss": 0.7144, "step": 4714 }, { "epoch": 0.4017895185342991, "grad_norm": 12.8125, "learning_rate": 2.2670667352363586e-06, "loss": 0.6919, "step": 4715 }, { "epoch": 0.4018747337025991, "grad_norm": 17.375, "learning_rate": 2.2669690385302775e-06, "loss": 0.9576, "step": 4716 }, { "epoch": 0.40195994887089903, "grad_norm": 16.875, "learning_rate": 2.266871323446543e-06, "loss": 0.5785, "step": 4717 }, { "epoch": 0.402045164039199, "grad_norm": 11.6875, "learning_rate": 2.2667735899869216e-06, "loss": 0.6427, "step": 4718 }, { "epoch": 0.40213037920749894, "grad_norm": 10.5, "learning_rate": 2.2666758381531783e-06, "loss": 0.3087, "step": 4719 }, { "epoch": 0.4022155943757989, "grad_norm": 10.625, "learning_rate": 2.266578067947081e-06, "loss": 0.4566, "step": 4720 }, { "epoch": 0.40230080954409886, "grad_norm": 11.875, "learning_rate": 2.266480279370395e-06, "loss": 0.5908, "step": 4721 }, { "epoch": 0.4023860247123988, "grad_norm": 12.1875, "learning_rate": 2.266382472424889e-06, "loss": 0.4141, "step": 4722 }, { "epoch": 0.4024712398806988, "grad_norm": 23.0, "learning_rate": 2.266284647112329e-06, "loss": 0.9698, "step": 4723 }, { "epoch": 0.40255645504899873, "grad_norm": 15.625, "learning_rate": 2.266186803434484e-06, "loss": 0.8946, "step": 4724 }, { "epoch": 0.4026416702172987, "grad_norm": 14.75, "learning_rate": 2.2660889413931213e-06, "loss": 0.8829, "step": 4725 }, { "epoch": 0.40272688538559864, "grad_norm": 16.625, "learning_rate": 2.2659910609900097e-06, "loss": 0.8776, "step": 4726 }, { "epoch": 0.4028121005538986, "grad_norm": 13.625, "learning_rate": 2.2658931622269177e-06, "loss": 0.8368, "step": 4727 }, { "epoch": 0.40289731572219856, "grad_norm": 10.625, "learning_rate": 2.2657952451056153e-06, "loss": 0.3565, "step": 4728 }, { "epoch": 0.4029825308904985, "grad_norm": 14.5625, "learning_rate": 2.265697309627871e-06, "loss": 0.7575, "step": 4729 }, { "epoch": 0.40306774605879847, "grad_norm": 15.3125, "learning_rate": 2.265599355795455e-06, "loss": 0.8153, "step": 4730 }, { "epoch": 0.4031529612270984, "grad_norm": 11.5, "learning_rate": 2.2655013836101375e-06, "loss": 0.4993, "step": 4731 }, { "epoch": 0.4032381763953984, "grad_norm": 15.4375, "learning_rate": 2.2654033930736883e-06, "loss": 1.1303, "step": 4732 }, { "epoch": 0.40332339156369834, "grad_norm": 13.625, "learning_rate": 2.2653053841878793e-06, "loss": 0.8431, "step": 4733 }, { "epoch": 0.4034086067319983, "grad_norm": 16.0, "learning_rate": 2.265207356954481e-06, "loss": 0.5506, "step": 4734 }, { "epoch": 0.40349382190029826, "grad_norm": 12.6875, "learning_rate": 2.265109311375265e-06, "loss": 0.6505, "step": 4735 }, { "epoch": 0.4035790370685982, "grad_norm": 15.625, "learning_rate": 2.2650112474520024e-06, "loss": 0.924, "step": 4736 }, { "epoch": 0.40366425223689817, "grad_norm": 17.25, "learning_rate": 2.2649131651864666e-06, "loss": 0.6731, "step": 4737 }, { "epoch": 0.4037494674051981, "grad_norm": 15.6875, "learning_rate": 2.264815064580429e-06, "loss": 0.6855, "step": 4738 }, { "epoch": 0.4038346825734981, "grad_norm": 12.125, "learning_rate": 2.2647169456356623e-06, "loss": 0.4179, "step": 4739 }, { "epoch": 0.40391989774179804, "grad_norm": 13.5625, "learning_rate": 2.2646188083539407e-06, "loss": 0.8236, "step": 4740 }, { "epoch": 0.404005112910098, "grad_norm": 16.75, "learning_rate": 2.264520652737036e-06, "loss": 0.8586, "step": 4741 }, { "epoch": 0.40409032807839795, "grad_norm": 19.5, "learning_rate": 2.264422478786724e-06, "loss": 0.7602, "step": 4742 }, { "epoch": 0.4041755432466979, "grad_norm": 13.25, "learning_rate": 2.2643242865047773e-06, "loss": 0.628, "step": 4743 }, { "epoch": 0.40426075841499787, "grad_norm": 17.625, "learning_rate": 2.2642260758929706e-06, "loss": 0.9205, "step": 4744 }, { "epoch": 0.4043459735832978, "grad_norm": 10.0, "learning_rate": 2.264127846953079e-06, "loss": 0.2715, "step": 4745 }, { "epoch": 0.4044311887515978, "grad_norm": 12.0625, "learning_rate": 2.2640295996868774e-06, "loss": 0.4329, "step": 4746 }, { "epoch": 0.40451640391989774, "grad_norm": 9.9375, "learning_rate": 2.263931334096141e-06, "loss": 0.2412, "step": 4747 }, { "epoch": 0.4046016190881977, "grad_norm": 19.5, "learning_rate": 2.263833050182646e-06, "loss": 1.1018, "step": 4748 }, { "epoch": 0.40468683425649765, "grad_norm": 13.6875, "learning_rate": 2.2637347479481687e-06, "loss": 0.7363, "step": 4749 }, { "epoch": 0.4047720494247976, "grad_norm": 14.0, "learning_rate": 2.2636364273944848e-06, "loss": 0.9824, "step": 4750 }, { "epoch": 0.40485726459309757, "grad_norm": 21.625, "learning_rate": 2.263538088523371e-06, "loss": 1.0033, "step": 4751 }, { "epoch": 0.4049424797613975, "grad_norm": 12.8125, "learning_rate": 2.2634397313366054e-06, "loss": 0.6889, "step": 4752 }, { "epoch": 0.4050276949296975, "grad_norm": 16.625, "learning_rate": 2.263341355835964e-06, "loss": 0.677, "step": 4753 }, { "epoch": 0.40511291009799744, "grad_norm": 22.375, "learning_rate": 2.263242962023226e-06, "loss": 1.0805, "step": 4754 }, { "epoch": 0.4051981252662974, "grad_norm": 20.875, "learning_rate": 2.2631445499001687e-06, "loss": 1.0173, "step": 4755 }, { "epoch": 0.40528334043459735, "grad_norm": 18.125, "learning_rate": 2.2630461194685705e-06, "loss": 0.6526, "step": 4756 }, { "epoch": 0.4053685556028973, "grad_norm": 26.375, "learning_rate": 2.26294767073021e-06, "loss": 0.8009, "step": 4757 }, { "epoch": 0.40545377077119726, "grad_norm": 16.125, "learning_rate": 2.2628492036868667e-06, "loss": 0.876, "step": 4758 }, { "epoch": 0.4055389859394972, "grad_norm": 18.5, "learning_rate": 2.2627507183403195e-06, "loss": 0.5533, "step": 4759 }, { "epoch": 0.4056242011077972, "grad_norm": 10.0, "learning_rate": 2.2626522146923487e-06, "loss": 0.3191, "step": 4760 }, { "epoch": 0.40570941627609713, "grad_norm": 17.75, "learning_rate": 2.2625536927447337e-06, "loss": 1.0784, "step": 4761 }, { "epoch": 0.4057946314443971, "grad_norm": 13.25, "learning_rate": 2.262455152499256e-06, "loss": 0.793, "step": 4762 }, { "epoch": 0.40587984661269705, "grad_norm": 18.5, "learning_rate": 2.262356593957695e-06, "loss": 0.9842, "step": 4763 }, { "epoch": 0.405965061780997, "grad_norm": 12.4375, "learning_rate": 2.2622580171218325e-06, "loss": 0.7935, "step": 4764 }, { "epoch": 0.40605027694929696, "grad_norm": 12.8125, "learning_rate": 2.26215942199345e-06, "loss": 0.4932, "step": 4765 }, { "epoch": 0.4061354921175969, "grad_norm": 11.8125, "learning_rate": 2.2620608085743286e-06, "loss": 0.5764, "step": 4766 }, { "epoch": 0.4062207072858969, "grad_norm": 11.1875, "learning_rate": 2.261962176866251e-06, "loss": 0.5877, "step": 4767 }, { "epoch": 0.40630592245419683, "grad_norm": 17.875, "learning_rate": 2.261863526870999e-06, "loss": 0.8367, "step": 4768 }, { "epoch": 0.4063911376224968, "grad_norm": 13.25, "learning_rate": 2.2617648585903556e-06, "loss": 0.853, "step": 4769 }, { "epoch": 0.40647635279079675, "grad_norm": 13.625, "learning_rate": 2.2616661720261034e-06, "loss": 0.7978, "step": 4770 }, { "epoch": 0.4065615679590967, "grad_norm": 11.75, "learning_rate": 2.2615674671800266e-06, "loss": 0.4212, "step": 4771 }, { "epoch": 0.40664678312739666, "grad_norm": 15.0, "learning_rate": 2.2614687440539085e-06, "loss": 0.6508, "step": 4772 }, { "epoch": 0.4067319982956966, "grad_norm": 23.0, "learning_rate": 2.261370002649533e-06, "loss": 0.9489, "step": 4773 }, { "epoch": 0.4068172134639966, "grad_norm": 11.6875, "learning_rate": 2.2612712429686846e-06, "loss": 0.4592, "step": 4774 }, { "epoch": 0.40690242863229653, "grad_norm": 13.75, "learning_rate": 2.2611724650131477e-06, "loss": 0.7842, "step": 4775 }, { "epoch": 0.4069876438005965, "grad_norm": 11.4375, "learning_rate": 2.261073668784708e-06, "loss": 0.4253, "step": 4776 }, { "epoch": 0.40707285896889644, "grad_norm": 15.3125, "learning_rate": 2.26097485428515e-06, "loss": 0.7862, "step": 4777 }, { "epoch": 0.4071580741371964, "grad_norm": 13.875, "learning_rate": 2.2608760215162594e-06, "loss": 0.8884, "step": 4778 }, { "epoch": 0.40724328930549636, "grad_norm": 10.9375, "learning_rate": 2.2607771704798234e-06, "loss": 0.4183, "step": 4779 }, { "epoch": 0.4073285044737963, "grad_norm": 15.5, "learning_rate": 2.2606783011776272e-06, "loss": 0.6357, "step": 4780 }, { "epoch": 0.4074137196420963, "grad_norm": 15.375, "learning_rate": 2.260579413611458e-06, "loss": 0.8026, "step": 4781 }, { "epoch": 0.40749893481039623, "grad_norm": 18.0, "learning_rate": 2.260480507783102e-06, "loss": 0.8144, "step": 4782 }, { "epoch": 0.4075841499786962, "grad_norm": 14.125, "learning_rate": 2.2603815836943474e-06, "loss": 0.4435, "step": 4783 }, { "epoch": 0.40766936514699614, "grad_norm": 24.25, "learning_rate": 2.2602826413469813e-06, "loss": 0.9548, "step": 4784 }, { "epoch": 0.4077545803152961, "grad_norm": 24.0, "learning_rate": 2.2601836807427925e-06, "loss": 0.9395, "step": 4785 }, { "epoch": 0.40783979548359606, "grad_norm": 14.1875, "learning_rate": 2.2600847018835685e-06, "loss": 0.9131, "step": 4786 }, { "epoch": 0.407925010651896, "grad_norm": 19.125, "learning_rate": 2.2599857047710983e-06, "loss": 1.0284, "step": 4787 }, { "epoch": 0.40801022582019597, "grad_norm": 13.125, "learning_rate": 2.259886689407171e-06, "loss": 0.718, "step": 4788 }, { "epoch": 0.4080954409884959, "grad_norm": 18.75, "learning_rate": 2.259787655793575e-06, "loss": 1.0749, "step": 4789 }, { "epoch": 0.4081806561567959, "grad_norm": 18.125, "learning_rate": 2.2596886039321013e-06, "loss": 0.8422, "step": 4790 }, { "epoch": 0.40826587132509584, "grad_norm": 12.125, "learning_rate": 2.2595895338245387e-06, "loss": 0.5409, "step": 4791 }, { "epoch": 0.4083510864933958, "grad_norm": 17.125, "learning_rate": 2.2594904454726784e-06, "loss": 1.057, "step": 4792 }, { "epoch": 0.40843630166169576, "grad_norm": 14.75, "learning_rate": 2.2593913388783105e-06, "loss": 0.5485, "step": 4793 }, { "epoch": 0.4085215168299957, "grad_norm": 15.5625, "learning_rate": 2.2592922140432262e-06, "loss": 0.9442, "step": 4794 }, { "epoch": 0.40860673199829567, "grad_norm": 19.5, "learning_rate": 2.2591930709692166e-06, "loss": 0.7642, "step": 4795 }, { "epoch": 0.4086919471665957, "grad_norm": 16.75, "learning_rate": 2.2590939096580734e-06, "loss": 0.6897, "step": 4796 }, { "epoch": 0.40877716233489564, "grad_norm": 15.1875, "learning_rate": 2.258994730111588e-06, "loss": 0.5593, "step": 4797 }, { "epoch": 0.4088623775031956, "grad_norm": 12.25, "learning_rate": 2.258895532331554e-06, "loss": 0.506, "step": 4798 }, { "epoch": 0.40894759267149555, "grad_norm": 12.6875, "learning_rate": 2.258796316319763e-06, "loss": 0.6796, "step": 4799 }, { "epoch": 0.4090328078397955, "grad_norm": 14.5625, "learning_rate": 2.258697082078008e-06, "loss": 1.0086, "step": 4800 }, { "epoch": 0.40911802300809547, "grad_norm": 18.25, "learning_rate": 2.2585978296080823e-06, "loss": 1.0874, "step": 4801 }, { "epoch": 0.4092032381763954, "grad_norm": 14.8125, "learning_rate": 2.25849855891178e-06, "loss": 0.649, "step": 4802 }, { "epoch": 0.4092884533446954, "grad_norm": 12.5625, "learning_rate": 2.2583992699908943e-06, "loss": 0.7, "step": 4803 }, { "epoch": 0.40937366851299534, "grad_norm": 13.9375, "learning_rate": 2.25829996284722e-06, "loss": 0.6285, "step": 4804 }, { "epoch": 0.4094588836812953, "grad_norm": 18.75, "learning_rate": 2.258200637482551e-06, "loss": 0.9545, "step": 4805 }, { "epoch": 0.40954409884959525, "grad_norm": 14.5625, "learning_rate": 2.258101293898683e-06, "loss": 0.8332, "step": 4806 }, { "epoch": 0.4096293140178952, "grad_norm": 14.3125, "learning_rate": 2.258001932097411e-06, "loss": 0.9381, "step": 4807 }, { "epoch": 0.40971452918619516, "grad_norm": 18.5, "learning_rate": 2.2579025520805303e-06, "loss": 0.8526, "step": 4808 }, { "epoch": 0.4097997443544951, "grad_norm": 15.9375, "learning_rate": 2.2578031538498367e-06, "loss": 0.7659, "step": 4809 }, { "epoch": 0.4098849595227951, "grad_norm": 14.125, "learning_rate": 2.257703737407127e-06, "loss": 0.7903, "step": 4810 }, { "epoch": 0.40997017469109504, "grad_norm": 18.125, "learning_rate": 2.2576043027541972e-06, "loss": 0.879, "step": 4811 }, { "epoch": 0.410055389859395, "grad_norm": 13.0625, "learning_rate": 2.257504849892845e-06, "loss": 0.3422, "step": 4812 }, { "epoch": 0.41014060502769495, "grad_norm": 13.4375, "learning_rate": 2.257405378824866e-06, "loss": 0.7472, "step": 4813 }, { "epoch": 0.4102258201959949, "grad_norm": 14.4375, "learning_rate": 2.25730588955206e-06, "loss": 0.6635, "step": 4814 }, { "epoch": 0.41031103536429486, "grad_norm": 15.5, "learning_rate": 2.257206382076223e-06, "loss": 0.7449, "step": 4815 }, { "epoch": 0.4103962505325948, "grad_norm": 12.5625, "learning_rate": 2.257106856399154e-06, "loss": 0.6581, "step": 4816 }, { "epoch": 0.4104814657008948, "grad_norm": 11.0625, "learning_rate": 2.257007312522651e-06, "loss": 0.5158, "step": 4817 }, { "epoch": 0.41056668086919473, "grad_norm": 14.1875, "learning_rate": 2.2569077504485136e-06, "loss": 0.7873, "step": 4818 }, { "epoch": 0.4106518960374947, "grad_norm": 10.4375, "learning_rate": 2.2568081701785405e-06, "loss": 0.4938, "step": 4819 }, { "epoch": 0.41073711120579465, "grad_norm": 17.875, "learning_rate": 2.2567085717145314e-06, "loss": 0.675, "step": 4820 }, { "epoch": 0.4108223263740946, "grad_norm": 11.0, "learning_rate": 2.256608955058286e-06, "loss": 0.4564, "step": 4821 }, { "epoch": 0.41090754154239456, "grad_norm": 20.125, "learning_rate": 2.2565093202116048e-06, "loss": 0.6711, "step": 4822 }, { "epoch": 0.4109927567106945, "grad_norm": 18.625, "learning_rate": 2.256409667176288e-06, "loss": 0.9432, "step": 4823 }, { "epoch": 0.4110779718789945, "grad_norm": 16.0, "learning_rate": 2.2563099959541363e-06, "loss": 0.6405, "step": 4824 }, { "epoch": 0.41116318704729443, "grad_norm": 14.4375, "learning_rate": 2.256210306546951e-06, "loss": 0.5645, "step": 4825 }, { "epoch": 0.4112484022155944, "grad_norm": 18.625, "learning_rate": 2.256110598956534e-06, "loss": 1.231, "step": 4826 }, { "epoch": 0.41133361738389435, "grad_norm": 25.125, "learning_rate": 2.2560108731846862e-06, "loss": 0.8709, "step": 4827 }, { "epoch": 0.4114188325521943, "grad_norm": 20.0, "learning_rate": 2.255911129233211e-06, "loss": 0.9058, "step": 4828 }, { "epoch": 0.41150404772049426, "grad_norm": 16.625, "learning_rate": 2.2558113671039097e-06, "loss": 0.9308, "step": 4829 }, { "epoch": 0.4115892628887942, "grad_norm": 11.25, "learning_rate": 2.255711586798586e-06, "loss": 0.5423, "step": 4830 }, { "epoch": 0.4116744780570942, "grad_norm": 16.25, "learning_rate": 2.255611788319042e-06, "loss": 1.1954, "step": 4831 }, { "epoch": 0.41175969322539413, "grad_norm": 20.625, "learning_rate": 2.255511971667082e-06, "loss": 1.227, "step": 4832 }, { "epoch": 0.4118449083936941, "grad_norm": 13.125, "learning_rate": 2.2554121368445097e-06, "loss": 0.8236, "step": 4833 }, { "epoch": 0.41193012356199404, "grad_norm": 17.375, "learning_rate": 2.2553122838531286e-06, "loss": 1.2031, "step": 4834 }, { "epoch": 0.412015338730294, "grad_norm": 18.0, "learning_rate": 2.2552124126947446e-06, "loss": 1.0453, "step": 4835 }, { "epoch": 0.41210055389859396, "grad_norm": 15.0, "learning_rate": 2.2551125233711605e-06, "loss": 0.7285, "step": 4836 }, { "epoch": 0.4121857690668939, "grad_norm": 14.0625, "learning_rate": 2.2550126158841827e-06, "loss": 0.8544, "step": 4837 }, { "epoch": 0.41227098423519387, "grad_norm": 20.25, "learning_rate": 2.2549126902356165e-06, "loss": 0.985, "step": 4838 }, { "epoch": 0.41235619940349383, "grad_norm": 15.0625, "learning_rate": 2.2548127464272674e-06, "loss": 0.6943, "step": 4839 }, { "epoch": 0.4124414145717938, "grad_norm": 15.25, "learning_rate": 2.2547127844609413e-06, "loss": 0.8087, "step": 4840 }, { "epoch": 0.41252662974009374, "grad_norm": 70.5, "learning_rate": 2.2546128043384454e-06, "loss": 0.5554, "step": 4841 }, { "epoch": 0.4126118449083937, "grad_norm": 11.875, "learning_rate": 2.2545128060615855e-06, "loss": 0.6621, "step": 4842 }, { "epoch": 0.41269706007669366, "grad_norm": 13.4375, "learning_rate": 2.2544127896321687e-06, "loss": 0.691, "step": 4843 }, { "epoch": 0.4127822752449936, "grad_norm": 17.375, "learning_rate": 2.254312755052003e-06, "loss": 0.6695, "step": 4844 }, { "epoch": 0.41286749041329357, "grad_norm": 11.75, "learning_rate": 2.2542127023228967e-06, "loss": 0.4374, "step": 4845 }, { "epoch": 0.4129527055815935, "grad_norm": 24.125, "learning_rate": 2.2541126314466565e-06, "loss": 0.8464, "step": 4846 }, { "epoch": 0.4130379207498935, "grad_norm": 23.125, "learning_rate": 2.2540125424250914e-06, "loss": 0.702, "step": 4847 }, { "epoch": 0.41312313591819344, "grad_norm": 13.4375, "learning_rate": 2.25391243526001e-06, "loss": 0.7888, "step": 4848 }, { "epoch": 0.4132083510864934, "grad_norm": 14.875, "learning_rate": 2.2538123099532212e-06, "loss": 0.9714, "step": 4849 }, { "epoch": 0.41329356625479335, "grad_norm": 17.0, "learning_rate": 2.2537121665065346e-06, "loss": 0.2549, "step": 4850 }, { "epoch": 0.4133787814230933, "grad_norm": 10.0, "learning_rate": 2.25361200492176e-06, "loss": 0.4591, "step": 4851 }, { "epoch": 0.41346399659139327, "grad_norm": 15.5, "learning_rate": 2.253511825200707e-06, "loss": 0.7933, "step": 4852 }, { "epoch": 0.4135492117596932, "grad_norm": 13.9375, "learning_rate": 2.2534116273451865e-06, "loss": 0.8653, "step": 4853 }, { "epoch": 0.4136344269279932, "grad_norm": 15.3125, "learning_rate": 2.253311411357009e-06, "loss": 0.7351, "step": 4854 }, { "epoch": 0.41371964209629314, "grad_norm": 16.75, "learning_rate": 2.253211177237985e-06, "loss": 0.8661, "step": 4855 }, { "epoch": 0.4138048572645931, "grad_norm": 21.125, "learning_rate": 2.2531109249899265e-06, "loss": 0.9118, "step": 4856 }, { "epoch": 0.41389007243289305, "grad_norm": 15.5, "learning_rate": 2.2530106546146446e-06, "loss": 0.6502, "step": 4857 }, { "epoch": 0.413975287601193, "grad_norm": 13.75, "learning_rate": 2.252910366113952e-06, "loss": 0.8105, "step": 4858 }, { "epoch": 0.41406050276949297, "grad_norm": 16.0, "learning_rate": 2.25281005948966e-06, "loss": 1.1773, "step": 4859 }, { "epoch": 0.4141457179377929, "grad_norm": 33.5, "learning_rate": 2.2527097347435823e-06, "loss": 1.5125, "step": 4860 }, { "epoch": 0.4142309331060929, "grad_norm": 16.75, "learning_rate": 2.2526093918775313e-06, "loss": 0.7659, "step": 4861 }, { "epoch": 0.41431614827439284, "grad_norm": 12.5, "learning_rate": 2.25250903089332e-06, "loss": 0.5829, "step": 4862 }, { "epoch": 0.4144013634426928, "grad_norm": 12.125, "learning_rate": 2.252408651792763e-06, "loss": 0.533, "step": 4863 }, { "epoch": 0.41448657861099275, "grad_norm": 23.25, "learning_rate": 2.252308254577673e-06, "loss": 0.9993, "step": 4864 }, { "epoch": 0.4145717937792927, "grad_norm": 12.625, "learning_rate": 2.252207839249866e-06, "loss": 0.3742, "step": 4865 }, { "epoch": 0.41465700894759266, "grad_norm": 11.25, "learning_rate": 2.252107405811155e-06, "loss": 0.4757, "step": 4866 }, { "epoch": 0.4147422241158926, "grad_norm": 14.9375, "learning_rate": 2.252006954263355e-06, "loss": 0.7955, "step": 4867 }, { "epoch": 0.4148274392841926, "grad_norm": 11.9375, "learning_rate": 2.251906484608282e-06, "loss": 0.5482, "step": 4868 }, { "epoch": 0.41491265445249254, "grad_norm": 15.9375, "learning_rate": 2.2518059968477516e-06, "loss": 0.8476, "step": 4869 }, { "epoch": 0.4149978696207925, "grad_norm": 10.9375, "learning_rate": 2.251705490983579e-06, "loss": 0.3362, "step": 4870 }, { "epoch": 0.41508308478909245, "grad_norm": 18.375, "learning_rate": 2.251604967017581e-06, "loss": 0.7088, "step": 4871 }, { "epoch": 0.4151682999573924, "grad_norm": 13.125, "learning_rate": 2.2515044249515745e-06, "loss": 0.6508, "step": 4872 }, { "epoch": 0.41525351512569236, "grad_norm": 15.375, "learning_rate": 2.2514038647873758e-06, "loss": 1.0892, "step": 4873 }, { "epoch": 0.4153387302939923, "grad_norm": 13.5625, "learning_rate": 2.2513032865268018e-06, "loss": 0.8347, "step": 4874 }, { "epoch": 0.4154239454622923, "grad_norm": 15.625, "learning_rate": 2.251202690171671e-06, "loss": 0.667, "step": 4875 }, { "epoch": 0.41550916063059223, "grad_norm": 14.0625, "learning_rate": 2.251102075723801e-06, "loss": 0.5272, "step": 4876 }, { "epoch": 0.4155943757988922, "grad_norm": 12.1875, "learning_rate": 2.2510014431850093e-06, "loss": 0.351, "step": 4877 }, { "epoch": 0.41567959096719215, "grad_norm": 19.0, "learning_rate": 2.2509007925571154e-06, "loss": 1.1781, "step": 4878 }, { "epoch": 0.4157648061354921, "grad_norm": 15.375, "learning_rate": 2.250800123841937e-06, "loss": 0.5345, "step": 4879 }, { "epoch": 0.41585002130379206, "grad_norm": 14.0625, "learning_rate": 2.2506994370412944e-06, "loss": 0.5204, "step": 4880 }, { "epoch": 0.415935236472092, "grad_norm": 17.0, "learning_rate": 2.2505987321570066e-06, "loss": 0.6551, "step": 4881 }, { "epoch": 0.416020451640392, "grad_norm": 11.5, "learning_rate": 2.250498009190894e-06, "loss": 0.6634, "step": 4882 }, { "epoch": 0.41610566680869193, "grad_norm": 14.0, "learning_rate": 2.2503972681447756e-06, "loss": 0.4377, "step": 4883 }, { "epoch": 0.4161908819769919, "grad_norm": 15.125, "learning_rate": 2.250296509020473e-06, "loss": 0.8151, "step": 4884 }, { "epoch": 0.41627609714529185, "grad_norm": 14.5625, "learning_rate": 2.2501957318198063e-06, "loss": 0.8733, "step": 4885 }, { "epoch": 0.4163613123135918, "grad_norm": 15.375, "learning_rate": 2.2500949365445974e-06, "loss": 0.6452, "step": 4886 }, { "epoch": 0.41644652748189176, "grad_norm": 16.375, "learning_rate": 2.249994123196667e-06, "loss": 0.9311, "step": 4887 }, { "epoch": 0.4165317426501917, "grad_norm": 19.25, "learning_rate": 2.2498932917778373e-06, "loss": 0.7952, "step": 4888 }, { "epoch": 0.4166169578184917, "grad_norm": 13.5, "learning_rate": 2.2497924422899304e-06, "loss": 0.735, "step": 4889 }, { "epoch": 0.41670217298679163, "grad_norm": 15.0, "learning_rate": 2.2496915747347684e-06, "loss": 0.6473, "step": 4890 }, { "epoch": 0.4167873881550916, "grad_norm": 13.625, "learning_rate": 2.2495906891141746e-06, "loss": 0.3715, "step": 4891 }, { "epoch": 0.41687260332339154, "grad_norm": 16.5, "learning_rate": 2.249489785429972e-06, "loss": 0.5998, "step": 4892 }, { "epoch": 0.4169578184916915, "grad_norm": 16.5, "learning_rate": 2.2493888636839837e-06, "loss": 0.9581, "step": 4893 }, { "epoch": 0.41704303365999146, "grad_norm": 14.625, "learning_rate": 2.249287923878034e-06, "loss": 0.504, "step": 4894 }, { "epoch": 0.4171282488282914, "grad_norm": 11.9375, "learning_rate": 2.2491869660139463e-06, "loss": 0.6031, "step": 4895 }, { "epoch": 0.41721346399659137, "grad_norm": 11.5625, "learning_rate": 2.2490859900935455e-06, "loss": 0.5202, "step": 4896 }, { "epoch": 0.41729867916489133, "grad_norm": 13.8125, "learning_rate": 2.248984996118656e-06, "loss": 0.6677, "step": 4897 }, { "epoch": 0.4173838943331913, "grad_norm": 14.3125, "learning_rate": 2.248883984091103e-06, "loss": 0.5755, "step": 4898 }, { "epoch": 0.41746910950149124, "grad_norm": 14.4375, "learning_rate": 2.2487829540127124e-06, "loss": 0.5265, "step": 4899 }, { "epoch": 0.4175543246697912, "grad_norm": 13.0, "learning_rate": 2.248681905885309e-06, "loss": 0.5974, "step": 4900 }, { "epoch": 0.41763953983809116, "grad_norm": 14.0625, "learning_rate": 2.24858083971072e-06, "loss": 0.9805, "step": 4901 }, { "epoch": 0.4177247550063911, "grad_norm": 14.0625, "learning_rate": 2.2484797554907706e-06, "loss": 0.5058, "step": 4902 }, { "epoch": 0.41780997017469107, "grad_norm": 17.0, "learning_rate": 2.2483786532272885e-06, "loss": 0.6768, "step": 4903 }, { "epoch": 0.417895185342991, "grad_norm": 14.1875, "learning_rate": 2.2482775329220995e-06, "loss": 0.6575, "step": 4904 }, { "epoch": 0.417980400511291, "grad_norm": 20.25, "learning_rate": 2.248176394577032e-06, "loss": 1.009, "step": 4905 }, { "epoch": 0.41806561567959094, "grad_norm": 12.25, "learning_rate": 2.2480752381939132e-06, "loss": 0.3105, "step": 4906 }, { "epoch": 0.4181508308478909, "grad_norm": 15.1875, "learning_rate": 2.247974063774571e-06, "loss": 0.6516, "step": 4907 }, { "epoch": 0.41823604601619085, "grad_norm": 19.875, "learning_rate": 2.247872871320834e-06, "loss": 0.9633, "step": 4908 }, { "epoch": 0.41832126118449087, "grad_norm": 13.4375, "learning_rate": 2.2477716608345307e-06, "loss": 0.6915, "step": 4909 }, { "epoch": 0.4184064763527908, "grad_norm": 13.1875, "learning_rate": 2.2476704323174904e-06, "loss": 0.7064, "step": 4910 }, { "epoch": 0.4184916915210908, "grad_norm": 18.875, "learning_rate": 2.247569185771542e-06, "loss": 0.8686, "step": 4911 }, { "epoch": 0.41857690668939074, "grad_norm": 13.125, "learning_rate": 2.247467921198516e-06, "loss": 0.7016, "step": 4912 }, { "epoch": 0.4186621218576907, "grad_norm": 19.25, "learning_rate": 2.2473666386002403e-06, "loss": 1.0099, "step": 4913 }, { "epoch": 0.41874733702599065, "grad_norm": 14.5625, "learning_rate": 2.2472653379785475e-06, "loss": 0.6496, "step": 4914 }, { "epoch": 0.4188325521942906, "grad_norm": 11.0625, "learning_rate": 2.2471640193352665e-06, "loss": 0.3736, "step": 4915 }, { "epoch": 0.41891776736259057, "grad_norm": 14.8125, "learning_rate": 2.24706268267223e-06, "loss": 1.0165, "step": 4916 }, { "epoch": 0.4190029825308905, "grad_norm": 15.625, "learning_rate": 2.2469613279912672e-06, "loss": 0.7243, "step": 4917 }, { "epoch": 0.4190881976991905, "grad_norm": 14.75, "learning_rate": 2.246859955294211e-06, "loss": 0.5785, "step": 4918 }, { "epoch": 0.41917341286749044, "grad_norm": 18.875, "learning_rate": 2.246758564582893e-06, "loss": 1.0706, "step": 4919 }, { "epoch": 0.4192586280357904, "grad_norm": 16.625, "learning_rate": 2.2466571558591457e-06, "loss": 0.7378, "step": 4920 }, { "epoch": 0.41934384320409035, "grad_norm": 12.4375, "learning_rate": 2.2465557291248006e-06, "loss": 0.6348, "step": 4921 }, { "epoch": 0.4194290583723903, "grad_norm": 12.875, "learning_rate": 2.2464542843816923e-06, "loss": 0.5927, "step": 4922 }, { "epoch": 0.41951427354069026, "grad_norm": 13.875, "learning_rate": 2.246352821631653e-06, "loss": 0.325, "step": 4923 }, { "epoch": 0.4195994887089902, "grad_norm": 17.75, "learning_rate": 2.2462513408765153e-06, "loss": 0.8192, "step": 4924 }, { "epoch": 0.4196847038772902, "grad_norm": 12.5, "learning_rate": 2.246149842118115e-06, "loss": 0.9298, "step": 4925 }, { "epoch": 0.41976991904559013, "grad_norm": 13.75, "learning_rate": 2.246048325358285e-06, "loss": 0.8035, "step": 4926 }, { "epoch": 0.4198551342138901, "grad_norm": 16.25, "learning_rate": 2.2459467905988604e-06, "loss": 0.7296, "step": 4927 }, { "epoch": 0.41994034938219005, "grad_norm": 24.875, "learning_rate": 2.2458452378416753e-06, "loss": 0.9956, "step": 4928 }, { "epoch": 0.42002556455049, "grad_norm": 20.375, "learning_rate": 2.245743667088566e-06, "loss": 0.8621, "step": 4929 }, { "epoch": 0.42011077971878996, "grad_norm": 15.0, "learning_rate": 2.2456420783413667e-06, "loss": 0.7533, "step": 4930 }, { "epoch": 0.4201959948870899, "grad_norm": 15.1875, "learning_rate": 2.245540471601914e-06, "loss": 1.0693, "step": 4931 }, { "epoch": 0.4202812100553899, "grad_norm": 13.4375, "learning_rate": 2.2454388468720444e-06, "loss": 0.9583, "step": 4932 }, { "epoch": 0.42036642522368983, "grad_norm": 11.375, "learning_rate": 2.245337204153593e-06, "loss": 0.4919, "step": 4933 }, { "epoch": 0.4204516403919898, "grad_norm": 12.875, "learning_rate": 2.245235543448398e-06, "loss": 0.5027, "step": 4934 }, { "epoch": 0.42053685556028975, "grad_norm": 12.1875, "learning_rate": 2.245133864758296e-06, "loss": 0.6087, "step": 4935 }, { "epoch": 0.4206220707285897, "grad_norm": 15.6875, "learning_rate": 2.245032168085124e-06, "loss": 1.1637, "step": 4936 }, { "epoch": 0.42070728589688966, "grad_norm": 11.9375, "learning_rate": 2.2449304534307206e-06, "loss": 0.4725, "step": 4937 }, { "epoch": 0.4207925010651896, "grad_norm": 16.375, "learning_rate": 2.2448287207969234e-06, "loss": 1.0574, "step": 4938 }, { "epoch": 0.4208777162334896, "grad_norm": 16.875, "learning_rate": 2.2447269701855706e-06, "loss": 1.0514, "step": 4939 }, { "epoch": 0.42096293140178953, "grad_norm": 14.0, "learning_rate": 2.244625201598501e-06, "loss": 0.764, "step": 4940 }, { "epoch": 0.4210481465700895, "grad_norm": 16.875, "learning_rate": 2.2445234150375542e-06, "loss": 0.9954, "step": 4941 }, { "epoch": 0.42113336173838944, "grad_norm": 13.6875, "learning_rate": 2.244421610504569e-06, "loss": 0.5302, "step": 4942 }, { "epoch": 0.4212185769066894, "grad_norm": 13.4375, "learning_rate": 2.244319788001386e-06, "loss": 0.4155, "step": 4943 }, { "epoch": 0.42130379207498936, "grad_norm": 13.8125, "learning_rate": 2.2442179475298433e-06, "loss": 0.6276, "step": 4944 }, { "epoch": 0.4213890072432893, "grad_norm": 17.75, "learning_rate": 2.2441160890917833e-06, "loss": 0.7554, "step": 4945 }, { "epoch": 0.42147422241158927, "grad_norm": 15.5625, "learning_rate": 2.2440142126890455e-06, "loss": 0.7116, "step": 4946 }, { "epoch": 0.42155943757988923, "grad_norm": 11.8125, "learning_rate": 2.243912318323472e-06, "loss": 0.4652, "step": 4947 }, { "epoch": 0.4216446527481892, "grad_norm": 15.5, "learning_rate": 2.243810405996903e-06, "loss": 1.0827, "step": 4948 }, { "epoch": 0.42172986791648914, "grad_norm": 17.75, "learning_rate": 2.2437084757111804e-06, "loss": 0.9546, "step": 4949 }, { "epoch": 0.4218150830847891, "grad_norm": 11.0, "learning_rate": 2.2436065274681467e-06, "loss": 0.6164, "step": 4950 }, { "epoch": 0.42190029825308906, "grad_norm": 12.875, "learning_rate": 2.243504561269644e-06, "loss": 0.7249, "step": 4951 }, { "epoch": 0.421985513421389, "grad_norm": 12.9375, "learning_rate": 2.2434025771175147e-06, "loss": 0.6985, "step": 4952 }, { "epoch": 0.42207072858968897, "grad_norm": 12.5, "learning_rate": 2.243300575013602e-06, "loss": 0.6818, "step": 4953 }, { "epoch": 0.4221559437579889, "grad_norm": 29.25, "learning_rate": 2.243198554959749e-06, "loss": 1.2069, "step": 4954 }, { "epoch": 0.4222411589262889, "grad_norm": 16.875, "learning_rate": 2.2430965169577993e-06, "loss": 1.3152, "step": 4955 }, { "epoch": 0.42232637409458884, "grad_norm": 14.1875, "learning_rate": 2.242994461009597e-06, "loss": 0.7069, "step": 4956 }, { "epoch": 0.4224115892628888, "grad_norm": 16.625, "learning_rate": 2.242892387116986e-06, "loss": 0.6766, "step": 4957 }, { "epoch": 0.42249680443118875, "grad_norm": 17.125, "learning_rate": 2.2427902952818117e-06, "loss": 0.7159, "step": 4958 }, { "epoch": 0.4225820195994887, "grad_norm": 14.0, "learning_rate": 2.242688185505918e-06, "loss": 0.8027, "step": 4959 }, { "epoch": 0.42266723476778867, "grad_norm": 15.25, "learning_rate": 2.2425860577911504e-06, "loss": 1.0403, "step": 4960 }, { "epoch": 0.4227524499360886, "grad_norm": 14.3125, "learning_rate": 2.242483912139355e-06, "loss": 0.8111, "step": 4961 }, { "epoch": 0.4228376651043886, "grad_norm": 15.5, "learning_rate": 2.242381748552377e-06, "loss": 0.7659, "step": 4962 }, { "epoch": 0.42292288027268854, "grad_norm": 13.75, "learning_rate": 2.242279567032063e-06, "loss": 0.9226, "step": 4963 }, { "epoch": 0.4230080954409885, "grad_norm": 13.5625, "learning_rate": 2.2421773675802596e-06, "loss": 0.5856, "step": 4964 }, { "epoch": 0.42309331060928845, "grad_norm": 29.25, "learning_rate": 2.242075150198813e-06, "loss": 1.1184, "step": 4965 }, { "epoch": 0.4231785257775884, "grad_norm": 15.3125, "learning_rate": 2.241972914889571e-06, "loss": 0.9654, "step": 4966 }, { "epoch": 0.42326374094588837, "grad_norm": 12.1875, "learning_rate": 2.241870661654381e-06, "loss": 0.5952, "step": 4967 }, { "epoch": 0.4233489561141883, "grad_norm": 16.25, "learning_rate": 2.241768390495091e-06, "loss": 0.8882, "step": 4968 }, { "epoch": 0.4234341712824883, "grad_norm": 18.625, "learning_rate": 2.2416661014135484e-06, "loss": 1.246, "step": 4969 }, { "epoch": 0.42351938645078824, "grad_norm": 14.4375, "learning_rate": 2.2415637944116025e-06, "loss": 0.6923, "step": 4970 }, { "epoch": 0.4236046016190882, "grad_norm": 13.9375, "learning_rate": 2.2414614694911016e-06, "loss": 0.5906, "step": 4971 }, { "epoch": 0.42368981678738815, "grad_norm": 16.25, "learning_rate": 2.241359126653895e-06, "loss": 0.5943, "step": 4972 }, { "epoch": 0.4237750319556881, "grad_norm": 14.3125, "learning_rate": 2.241256765901832e-06, "loss": 0.5606, "step": 4973 }, { "epoch": 0.42386024712398807, "grad_norm": 23.25, "learning_rate": 2.2411543872367623e-06, "loss": 1.257, "step": 4974 }, { "epoch": 0.423945462292288, "grad_norm": 20.375, "learning_rate": 2.2410519906605368e-06, "loss": 1.3547, "step": 4975 }, { "epoch": 0.424030677460588, "grad_norm": 15.4375, "learning_rate": 2.240949576175005e-06, "loss": 0.858, "step": 4976 }, { "epoch": 0.42411589262888794, "grad_norm": 17.375, "learning_rate": 2.2408471437820175e-06, "loss": 0.689, "step": 4977 }, { "epoch": 0.4242011077971879, "grad_norm": 35.75, "learning_rate": 2.240744693483426e-06, "loss": 0.5849, "step": 4978 }, { "epoch": 0.42428632296548785, "grad_norm": 16.875, "learning_rate": 2.2406422252810815e-06, "loss": 0.5204, "step": 4979 }, { "epoch": 0.4243715381337878, "grad_norm": 13.875, "learning_rate": 2.2405397391768357e-06, "loss": 0.6337, "step": 4980 }, { "epoch": 0.42445675330208776, "grad_norm": 13.0625, "learning_rate": 2.2404372351725413e-06, "loss": 0.7091, "step": 4981 }, { "epoch": 0.4245419684703877, "grad_norm": 11.0, "learning_rate": 2.2403347132700493e-06, "loss": 0.5954, "step": 4982 }, { "epoch": 0.4246271836386877, "grad_norm": 14.8125, "learning_rate": 2.240232173471214e-06, "loss": 0.7097, "step": 4983 }, { "epoch": 0.42471239880698763, "grad_norm": 14.9375, "learning_rate": 2.240129615777887e-06, "loss": 0.7659, "step": 4984 }, { "epoch": 0.4247976139752876, "grad_norm": 25.375, "learning_rate": 2.240027040191922e-06, "loss": 0.9313, "step": 4985 }, { "epoch": 0.42488282914358755, "grad_norm": 13.9375, "learning_rate": 2.2399244467151735e-06, "loss": 0.8854, "step": 4986 }, { "epoch": 0.4249680443118875, "grad_norm": 12.5, "learning_rate": 2.2398218353494945e-06, "loss": 0.4812, "step": 4987 }, { "epoch": 0.42505325948018746, "grad_norm": 14.3125, "learning_rate": 2.23971920609674e-06, "loss": 0.629, "step": 4988 }, { "epoch": 0.4251384746484874, "grad_norm": 15.25, "learning_rate": 2.2396165589587636e-06, "loss": 0.8386, "step": 4989 }, { "epoch": 0.4252236898167874, "grad_norm": 11.0625, "learning_rate": 2.2395138939374213e-06, "loss": 0.5445, "step": 4990 }, { "epoch": 0.42530890498508733, "grad_norm": 12.9375, "learning_rate": 2.2394112110345672e-06, "loss": 0.5404, "step": 4991 }, { "epoch": 0.4253941201533873, "grad_norm": 14.25, "learning_rate": 2.239308510252058e-06, "loss": 0.4705, "step": 4992 }, { "epoch": 0.42547933532168725, "grad_norm": 16.875, "learning_rate": 2.239205791591749e-06, "loss": 1.0347, "step": 4993 }, { "epoch": 0.4255645504899872, "grad_norm": 12.8125, "learning_rate": 2.2391030550554972e-06, "loss": 0.6768, "step": 4994 }, { "epoch": 0.42564976565828716, "grad_norm": 14.5625, "learning_rate": 2.239000300645158e-06, "loss": 0.6945, "step": 4995 }, { "epoch": 0.4257349808265871, "grad_norm": 12.4375, "learning_rate": 2.238897528362589e-06, "loss": 0.5118, "step": 4996 }, { "epoch": 0.4258201959948871, "grad_norm": 15.75, "learning_rate": 2.2387947382096477e-06, "loss": 1.0141, "step": 4997 }, { "epoch": 0.42590541116318703, "grad_norm": 13.8125, "learning_rate": 2.238691930188191e-06, "loss": 0.6464, "step": 4998 }, { "epoch": 0.425990626331487, "grad_norm": 15.0, "learning_rate": 2.2385891043000766e-06, "loss": 0.9191, "step": 4999 }, { "epoch": 0.42607584149978694, "grad_norm": 16.5, "learning_rate": 2.238486260547163e-06, "loss": 0.4816, "step": 5000 }, { "epoch": 0.4261610566680869, "grad_norm": 11.75, "learning_rate": 2.2383833989313087e-06, "loss": 0.5444, "step": 5001 }, { "epoch": 0.42624627183638686, "grad_norm": 23.25, "learning_rate": 2.2382805194543726e-06, "loss": 0.8971, "step": 5002 }, { "epoch": 0.4263314870046868, "grad_norm": 15.4375, "learning_rate": 2.2381776221182136e-06, "loss": 0.9627, "step": 5003 }, { "epoch": 0.4264167021729868, "grad_norm": 15.8125, "learning_rate": 2.2380747069246916e-06, "loss": 0.6279, "step": 5004 }, { "epoch": 0.42650191734128673, "grad_norm": 13.5625, "learning_rate": 2.2379717738756656e-06, "loss": 0.8512, "step": 5005 }, { "epoch": 0.4265871325095867, "grad_norm": 14.375, "learning_rate": 2.2378688229729965e-06, "loss": 0.7839, "step": 5006 }, { "epoch": 0.42667234767788664, "grad_norm": 16.75, "learning_rate": 2.2377658542185442e-06, "loss": 0.8248, "step": 5007 }, { "epoch": 0.4267575628461866, "grad_norm": 12.6875, "learning_rate": 2.23766286761417e-06, "loss": 0.7751, "step": 5008 }, { "epoch": 0.42684277801448656, "grad_norm": 17.125, "learning_rate": 2.237559863161734e-06, "loss": 1.0821, "step": 5009 }, { "epoch": 0.4269279931827865, "grad_norm": 14.25, "learning_rate": 2.2374568408630985e-06, "loss": 0.7242, "step": 5010 }, { "epoch": 0.42701320835108647, "grad_norm": 16.25, "learning_rate": 2.237353800720125e-06, "loss": 0.8592, "step": 5011 }, { "epoch": 0.4270984235193864, "grad_norm": 12.5, "learning_rate": 2.2372507427346753e-06, "loss": 0.5808, "step": 5012 }, { "epoch": 0.4271836386876864, "grad_norm": 15.125, "learning_rate": 2.2371476669086123e-06, "loss": 1.0118, "step": 5013 }, { "epoch": 0.42726885385598634, "grad_norm": 14.3125, "learning_rate": 2.237044573243798e-06, "loss": 0.525, "step": 5014 }, { "epoch": 0.4273540690242863, "grad_norm": 23.5, "learning_rate": 2.236941461742096e-06, "loss": 0.629, "step": 5015 }, { "epoch": 0.42743928419258626, "grad_norm": 13.75, "learning_rate": 2.2368383324053684e-06, "loss": 0.6978, "step": 5016 }, { "epoch": 0.4275244993608862, "grad_norm": 13.0625, "learning_rate": 2.236735185235481e-06, "loss": 0.5348, "step": 5017 }, { "epoch": 0.42760971452918617, "grad_norm": 16.0, "learning_rate": 2.2366320202342957e-06, "loss": 0.9351, "step": 5018 }, { "epoch": 0.4276949296974861, "grad_norm": 21.0, "learning_rate": 2.236528837403678e-06, "loss": 1.0668, "step": 5019 }, { "epoch": 0.4277801448657861, "grad_norm": 23.75, "learning_rate": 2.2364256367454925e-06, "loss": 1.078, "step": 5020 }, { "epoch": 0.42786536003408604, "grad_norm": 25.375, "learning_rate": 2.2363224182616035e-06, "loss": 1.1077, "step": 5021 }, { "epoch": 0.42795057520238605, "grad_norm": 12.125, "learning_rate": 2.2362191819538763e-06, "loss": 0.5359, "step": 5022 }, { "epoch": 0.428035790370686, "grad_norm": 11.5, "learning_rate": 2.236115927824177e-06, "loss": 0.4355, "step": 5023 }, { "epoch": 0.42812100553898597, "grad_norm": 13.75, "learning_rate": 2.236012655874371e-06, "loss": 0.555, "step": 5024 }, { "epoch": 0.4282062207072859, "grad_norm": 13.9375, "learning_rate": 2.2359093661063254e-06, "loss": 0.772, "step": 5025 }, { "epoch": 0.4282914358755859, "grad_norm": 16.5, "learning_rate": 2.235806058521906e-06, "loss": 0.9323, "step": 5026 }, { "epoch": 0.42837665104388584, "grad_norm": 17.125, "learning_rate": 2.2357027331229796e-06, "loss": 0.9413, "step": 5027 }, { "epoch": 0.4284618662121858, "grad_norm": 13.3125, "learning_rate": 2.235599389911414e-06, "loss": 0.5359, "step": 5028 }, { "epoch": 0.42854708138048575, "grad_norm": 10.375, "learning_rate": 2.2354960288890754e-06, "loss": 0.4342, "step": 5029 }, { "epoch": 0.4286322965487857, "grad_norm": 13.8125, "learning_rate": 2.2353926500578333e-06, "loss": 0.766, "step": 5030 }, { "epoch": 0.42871751171708566, "grad_norm": 14.0625, "learning_rate": 2.2352892534195548e-06, "loss": 0.5773, "step": 5031 }, { "epoch": 0.4288027268853856, "grad_norm": 20.75, "learning_rate": 2.2351858389761088e-06, "loss": 0.7962, "step": 5032 }, { "epoch": 0.4288879420536856, "grad_norm": 13.1875, "learning_rate": 2.2350824067293643e-06, "loss": 0.6885, "step": 5033 }, { "epoch": 0.42897315722198553, "grad_norm": 16.375, "learning_rate": 2.23497895668119e-06, "loss": 0.7861, "step": 5034 }, { "epoch": 0.4290583723902855, "grad_norm": 17.75, "learning_rate": 2.2348754888334554e-06, "loss": 0.8054, "step": 5035 }, { "epoch": 0.42914358755858545, "grad_norm": 12.5625, "learning_rate": 2.2347720031880303e-06, "loss": 0.4438, "step": 5036 }, { "epoch": 0.4292288027268854, "grad_norm": 17.375, "learning_rate": 2.234668499746785e-06, "loss": 0.8679, "step": 5037 }, { "epoch": 0.42931401789518536, "grad_norm": 14.625, "learning_rate": 2.2345649785115897e-06, "loss": 0.6812, "step": 5038 }, { "epoch": 0.4293992330634853, "grad_norm": 24.375, "learning_rate": 2.234461439484315e-06, "loss": 0.8815, "step": 5039 }, { "epoch": 0.4294844482317853, "grad_norm": 11.5625, "learning_rate": 2.2343578826668322e-06, "loss": 0.4515, "step": 5040 }, { "epoch": 0.42956966340008523, "grad_norm": 12.0625, "learning_rate": 2.234254308061013e-06, "loss": 0.5014, "step": 5041 }, { "epoch": 0.4296548785683852, "grad_norm": 18.875, "learning_rate": 2.234150715668728e-06, "loss": 0.7075, "step": 5042 }, { "epoch": 0.42974009373668515, "grad_norm": 20.625, "learning_rate": 2.2340471054918506e-06, "loss": 0.7983, "step": 5043 }, { "epoch": 0.4298253089049851, "grad_norm": 14.0625, "learning_rate": 2.233943477532252e-06, "loss": 0.6592, "step": 5044 }, { "epoch": 0.42991052407328506, "grad_norm": 16.5, "learning_rate": 2.2338398317918054e-06, "loss": 0.7493, "step": 5045 }, { "epoch": 0.429995739241585, "grad_norm": 13.9375, "learning_rate": 2.233736168272384e-06, "loss": 0.4503, "step": 5046 }, { "epoch": 0.430080954409885, "grad_norm": 12.8125, "learning_rate": 2.2336324869758607e-06, "loss": 0.4597, "step": 5047 }, { "epoch": 0.43016616957818493, "grad_norm": 18.25, "learning_rate": 2.233528787904109e-06, "loss": 0.822, "step": 5048 }, { "epoch": 0.4302513847464849, "grad_norm": 13.9375, "learning_rate": 2.233425071059003e-06, "loss": 0.6965, "step": 5049 }, { "epoch": 0.43033659991478485, "grad_norm": 16.75, "learning_rate": 2.2333213364424178e-06, "loss": 0.9644, "step": 5050 }, { "epoch": 0.4304218150830848, "grad_norm": 21.25, "learning_rate": 2.2332175840562264e-06, "loss": 1.0969, "step": 5051 }, { "epoch": 0.43050703025138476, "grad_norm": 18.75, "learning_rate": 2.233113813902305e-06, "loss": 1.0114, "step": 5052 }, { "epoch": 0.4305922454196847, "grad_norm": 12.625, "learning_rate": 2.2330100259825285e-06, "loss": 0.7766, "step": 5053 }, { "epoch": 0.4306774605879847, "grad_norm": 22.125, "learning_rate": 2.2329062202987718e-06, "loss": 1.1876, "step": 5054 }, { "epoch": 0.43076267575628463, "grad_norm": 15.9375, "learning_rate": 2.232802396852912e-06, "loss": 0.9081, "step": 5055 }, { "epoch": 0.4308478909245846, "grad_norm": 17.5, "learning_rate": 2.2326985556468238e-06, "loss": 0.7736, "step": 5056 }, { "epoch": 0.43093310609288454, "grad_norm": 18.0, "learning_rate": 2.2325946966823853e-06, "loss": 0.9413, "step": 5057 }, { "epoch": 0.4310183212611845, "grad_norm": 19.125, "learning_rate": 2.2324908199614722e-06, "loss": 1.1306, "step": 5058 }, { "epoch": 0.43110353642948446, "grad_norm": 11.75, "learning_rate": 2.232386925485962e-06, "loss": 0.6384, "step": 5059 }, { "epoch": 0.4311887515977844, "grad_norm": 14.4375, "learning_rate": 2.232283013257732e-06, "loss": 0.7943, "step": 5060 }, { "epoch": 0.43127396676608437, "grad_norm": 15.5, "learning_rate": 2.2321790832786607e-06, "loss": 0.9235, "step": 5061 }, { "epoch": 0.43135918193438433, "grad_norm": 11.125, "learning_rate": 2.2320751355506255e-06, "loss": 0.5174, "step": 5062 }, { "epoch": 0.4314443971026843, "grad_norm": 10.25, "learning_rate": 2.231971170075505e-06, "loss": 0.4515, "step": 5063 }, { "epoch": 0.43152961227098424, "grad_norm": 12.75, "learning_rate": 2.231867186855178e-06, "loss": 0.6666, "step": 5064 }, { "epoch": 0.4316148274392842, "grad_norm": 11.8125, "learning_rate": 2.2317631858915235e-06, "loss": 0.7227, "step": 5065 }, { "epoch": 0.43170004260758416, "grad_norm": 14.375, "learning_rate": 2.2316591671864213e-06, "loss": 0.7342, "step": 5066 }, { "epoch": 0.4317852577758841, "grad_norm": 15.3125, "learning_rate": 2.2315551307417506e-06, "loss": 0.7317, "step": 5067 }, { "epoch": 0.43187047294418407, "grad_norm": 12.625, "learning_rate": 2.2314510765593917e-06, "loss": 0.5484, "step": 5068 }, { "epoch": 0.431955688112484, "grad_norm": 17.875, "learning_rate": 2.2313470046412247e-06, "loss": 1.0629, "step": 5069 }, { "epoch": 0.432040903280784, "grad_norm": 14.375, "learning_rate": 2.2312429149891307e-06, "loss": 0.74, "step": 5070 }, { "epoch": 0.43212611844908394, "grad_norm": 28.875, "learning_rate": 2.2311388076049903e-06, "loss": 0.8439, "step": 5071 }, { "epoch": 0.4322113336173839, "grad_norm": 11.1875, "learning_rate": 2.2310346824906855e-06, "loss": 0.519, "step": 5072 }, { "epoch": 0.43229654878568385, "grad_norm": 11.4375, "learning_rate": 2.2309305396480967e-06, "loss": 0.4261, "step": 5073 }, { "epoch": 0.4323817639539838, "grad_norm": 12.375, "learning_rate": 2.2308263790791073e-06, "loss": 0.6671, "step": 5074 }, { "epoch": 0.43246697912228377, "grad_norm": 14.4375, "learning_rate": 2.230722200785599e-06, "loss": 0.8671, "step": 5075 }, { "epoch": 0.4325521942905837, "grad_norm": 25.625, "learning_rate": 2.230618004769454e-06, "loss": 1.1036, "step": 5076 }, { "epoch": 0.4326374094588837, "grad_norm": 11.3125, "learning_rate": 2.230513791032556e-06, "loss": 0.4433, "step": 5077 }, { "epoch": 0.43272262462718364, "grad_norm": 15.875, "learning_rate": 2.230409559576787e-06, "loss": 0.6872, "step": 5078 }, { "epoch": 0.4328078397954836, "grad_norm": 13.4375, "learning_rate": 2.2303053104040322e-06, "loss": 0.8941, "step": 5079 }, { "epoch": 0.43289305496378355, "grad_norm": 13.5, "learning_rate": 2.230201043516174e-06, "loss": 0.6345, "step": 5080 }, { "epoch": 0.4329782701320835, "grad_norm": 12.125, "learning_rate": 2.230096758915098e-06, "loss": 0.3736, "step": 5081 }, { "epoch": 0.43306348530038347, "grad_norm": 15.3125, "learning_rate": 2.2299924566026876e-06, "loss": 0.869, "step": 5082 }, { "epoch": 0.4331487004686834, "grad_norm": 14.8125, "learning_rate": 2.2298881365808276e-06, "loss": 0.9045, "step": 5083 }, { "epoch": 0.4332339156369834, "grad_norm": 13.125, "learning_rate": 2.229783798851404e-06, "loss": 0.8778, "step": 5084 }, { "epoch": 0.43331913080528334, "grad_norm": 14.25, "learning_rate": 2.229679443416302e-06, "loss": 0.7869, "step": 5085 }, { "epoch": 0.4334043459735833, "grad_norm": 13.6875, "learning_rate": 2.2295750702774077e-06, "loss": 0.6352, "step": 5086 }, { "epoch": 0.43348956114188325, "grad_norm": 16.875, "learning_rate": 2.2294706794366066e-06, "loss": 0.8027, "step": 5087 }, { "epoch": 0.4335747763101832, "grad_norm": 15.9375, "learning_rate": 2.229366270895785e-06, "loss": 1.0057, "step": 5088 }, { "epoch": 0.43365999147848316, "grad_norm": 11.8125, "learning_rate": 2.2292618446568305e-06, "loss": 0.3933, "step": 5089 }, { "epoch": 0.4337452066467831, "grad_norm": 14.4375, "learning_rate": 2.2291574007216295e-06, "loss": 0.8552, "step": 5090 }, { "epoch": 0.4338304218150831, "grad_norm": 16.375, "learning_rate": 2.2290529390920696e-06, "loss": 0.8523, "step": 5091 }, { "epoch": 0.43391563698338304, "grad_norm": 14.8125, "learning_rate": 2.2289484597700386e-06, "loss": 0.7322, "step": 5092 }, { "epoch": 0.434000852151683, "grad_norm": 11.0, "learning_rate": 2.2288439627574247e-06, "loss": 0.4172, "step": 5093 }, { "epoch": 0.43408606731998295, "grad_norm": 13.375, "learning_rate": 2.228739448056116e-06, "loss": 0.5626, "step": 5094 }, { "epoch": 0.4341712824882829, "grad_norm": 10.8125, "learning_rate": 2.2286349156680014e-06, "loss": 0.6058, "step": 5095 }, { "epoch": 0.43425649765658286, "grad_norm": 15.25, "learning_rate": 2.2285303655949693e-06, "loss": 0.9247, "step": 5096 }, { "epoch": 0.4343417128248828, "grad_norm": 13.5625, "learning_rate": 2.22842579783891e-06, "loss": 0.6181, "step": 5097 }, { "epoch": 0.4344269279931828, "grad_norm": 14.375, "learning_rate": 2.2283212124017124e-06, "loss": 0.6834, "step": 5098 }, { "epoch": 0.43451214316148273, "grad_norm": 17.5, "learning_rate": 2.228216609285267e-06, "loss": 0.6878, "step": 5099 }, { "epoch": 0.4345973583297827, "grad_norm": 12.75, "learning_rate": 2.2281119884914635e-06, "loss": 0.6651, "step": 5100 }, { "epoch": 0.43468257349808265, "grad_norm": 17.125, "learning_rate": 2.2280073500221926e-06, "loss": 1.0826, "step": 5101 }, { "epoch": 0.4347677886663826, "grad_norm": 15.0625, "learning_rate": 2.2279026938793458e-06, "loss": 0.863, "step": 5102 }, { "epoch": 0.43485300383468256, "grad_norm": 15.625, "learning_rate": 2.2277980200648142e-06, "loss": 0.8633, "step": 5103 }, { "epoch": 0.4349382190029825, "grad_norm": 14.875, "learning_rate": 2.2276933285804884e-06, "loss": 0.6028, "step": 5104 }, { "epoch": 0.4350234341712825, "grad_norm": 19.5, "learning_rate": 2.2275886194282615e-06, "loss": 0.7762, "step": 5105 }, { "epoch": 0.43510864933958243, "grad_norm": 10.4375, "learning_rate": 2.227483892610025e-06, "loss": 0.5226, "step": 5106 }, { "epoch": 0.4351938645078824, "grad_norm": 13.0625, "learning_rate": 2.227379148127672e-06, "loss": 0.8069, "step": 5107 }, { "epoch": 0.43527907967618235, "grad_norm": 16.5, "learning_rate": 2.227274385983095e-06, "loss": 0.7514, "step": 5108 }, { "epoch": 0.4353642948444823, "grad_norm": 36.25, "learning_rate": 2.227169606178186e-06, "loss": 0.8766, "step": 5109 }, { "epoch": 0.43544951001278226, "grad_norm": 10.75, "learning_rate": 2.2270648087148412e-06, "loss": 0.6408, "step": 5110 }, { "epoch": 0.4355347251810822, "grad_norm": 17.625, "learning_rate": 2.2269599935949513e-06, "loss": 1.1137, "step": 5111 }, { "epoch": 0.4356199403493822, "grad_norm": 15.375, "learning_rate": 2.226855160820413e-06, "loss": 0.893, "step": 5112 }, { "epoch": 0.43570515551768213, "grad_norm": 11.3125, "learning_rate": 2.2267503103931194e-06, "loss": 0.4471, "step": 5113 }, { "epoch": 0.4357903706859821, "grad_norm": 18.5, "learning_rate": 2.2266454423149653e-06, "loss": 0.8746, "step": 5114 }, { "epoch": 0.43587558585428204, "grad_norm": 13.6875, "learning_rate": 2.2265405565878463e-06, "loss": 0.6705, "step": 5115 }, { "epoch": 0.435960801022582, "grad_norm": 15.1875, "learning_rate": 2.226435653213657e-06, "loss": 0.7691, "step": 5116 }, { "epoch": 0.43604601619088196, "grad_norm": 17.5, "learning_rate": 2.2263307321942935e-06, "loss": 0.9185, "step": 5117 }, { "epoch": 0.4361312313591819, "grad_norm": 13.4375, "learning_rate": 2.2262257935316523e-06, "loss": 0.6799, "step": 5118 }, { "epoch": 0.43621644652748187, "grad_norm": 12.625, "learning_rate": 2.226120837227629e-06, "loss": 0.6413, "step": 5119 }, { "epoch": 0.43630166169578183, "grad_norm": 13.125, "learning_rate": 2.2260158632841207e-06, "loss": 0.5855, "step": 5120 }, { "epoch": 0.4363868768640818, "grad_norm": 17.125, "learning_rate": 2.2259108717030245e-06, "loss": 0.6167, "step": 5121 }, { "epoch": 0.43647209203238174, "grad_norm": 15.125, "learning_rate": 2.2258058624862374e-06, "loss": 0.6559, "step": 5122 }, { "epoch": 0.4365573072006817, "grad_norm": 18.75, "learning_rate": 2.225700835635657e-06, "loss": 0.9811, "step": 5123 }, { "epoch": 0.43664252236898166, "grad_norm": 16.125, "learning_rate": 2.2255957911531817e-06, "loss": 0.7894, "step": 5124 }, { "epoch": 0.4367277375372816, "grad_norm": 17.125, "learning_rate": 2.225490729040709e-06, "loss": 1.1962, "step": 5125 }, { "epoch": 0.43681295270558157, "grad_norm": 12.0625, "learning_rate": 2.225385649300138e-06, "loss": 0.5451, "step": 5126 }, { "epoch": 0.4368981678738815, "grad_norm": 12.75, "learning_rate": 2.2252805519333676e-06, "loss": 0.6483, "step": 5127 }, { "epoch": 0.4369833830421815, "grad_norm": 15.125, "learning_rate": 2.2251754369422968e-06, "loss": 0.8555, "step": 5128 }, { "epoch": 0.43706859821048144, "grad_norm": 14.75, "learning_rate": 2.2250703043288253e-06, "loss": 0.7367, "step": 5129 }, { "epoch": 0.4371538133787814, "grad_norm": 12.75, "learning_rate": 2.224965154094853e-06, "loss": 0.524, "step": 5130 }, { "epoch": 0.43723902854708135, "grad_norm": 15.875, "learning_rate": 2.2248599862422798e-06, "loss": 0.7958, "step": 5131 }, { "epoch": 0.4373242437153813, "grad_norm": 13.625, "learning_rate": 2.2247548007730063e-06, "loss": 0.7912, "step": 5132 }, { "epoch": 0.43740945888368127, "grad_norm": 18.375, "learning_rate": 2.2246495976889335e-06, "loss": 0.8527, "step": 5133 }, { "epoch": 0.4374946740519812, "grad_norm": 18.875, "learning_rate": 2.224544376991962e-06, "loss": 0.9801, "step": 5134 }, { "epoch": 0.43757988922028124, "grad_norm": 11.375, "learning_rate": 2.224439138683994e-06, "loss": 0.651, "step": 5135 }, { "epoch": 0.4376651043885812, "grad_norm": 14.3125, "learning_rate": 2.224333882766931e-06, "loss": 0.7407, "step": 5136 }, { "epoch": 0.43775031955688115, "grad_norm": 15.5625, "learning_rate": 2.2242286092426747e-06, "loss": 1.0061, "step": 5137 }, { "epoch": 0.4378355347251811, "grad_norm": 14.25, "learning_rate": 2.224123318113128e-06, "loss": 0.9623, "step": 5138 }, { "epoch": 0.43792074989348106, "grad_norm": 16.75, "learning_rate": 2.2240180093801935e-06, "loss": 0.4397, "step": 5139 }, { "epoch": 0.438005965061781, "grad_norm": 19.25, "learning_rate": 2.2239126830457736e-06, "loss": 1.0909, "step": 5140 }, { "epoch": 0.438091180230081, "grad_norm": 16.875, "learning_rate": 2.2238073391117726e-06, "loss": 1.1586, "step": 5141 }, { "epoch": 0.43817639539838094, "grad_norm": 14.75, "learning_rate": 2.2237019775800938e-06, "loss": 0.9064, "step": 5142 }, { "epoch": 0.4382616105666809, "grad_norm": 15.375, "learning_rate": 2.2235965984526408e-06, "loss": 0.7616, "step": 5143 }, { "epoch": 0.43834682573498085, "grad_norm": 25.75, "learning_rate": 2.2234912017313182e-06, "loss": 1.1972, "step": 5144 }, { "epoch": 0.4384320409032808, "grad_norm": 12.75, "learning_rate": 2.223385787418031e-06, "loss": 0.7455, "step": 5145 }, { "epoch": 0.43851725607158076, "grad_norm": 13.5, "learning_rate": 2.223280355514684e-06, "loss": 0.7262, "step": 5146 }, { "epoch": 0.4386024712398807, "grad_norm": 10.6875, "learning_rate": 2.2231749060231816e-06, "loss": 0.4949, "step": 5147 }, { "epoch": 0.4386876864081807, "grad_norm": 13.875, "learning_rate": 2.2230694389454305e-06, "loss": 0.9748, "step": 5148 }, { "epoch": 0.43877290157648063, "grad_norm": 15.0625, "learning_rate": 2.222963954283336e-06, "loss": 0.9741, "step": 5149 }, { "epoch": 0.4388581167447806, "grad_norm": 11.5, "learning_rate": 2.2228584520388043e-06, "loss": 0.5132, "step": 5150 }, { "epoch": 0.43894333191308055, "grad_norm": 11.9375, "learning_rate": 2.2227529322137425e-06, "loss": 0.6055, "step": 5151 }, { "epoch": 0.4390285470813805, "grad_norm": 14.3125, "learning_rate": 2.2226473948100564e-06, "loss": 0.9458, "step": 5152 }, { "epoch": 0.43911376224968046, "grad_norm": 11.4375, "learning_rate": 2.2225418398296545e-06, "loss": 0.5338, "step": 5153 }, { "epoch": 0.4391989774179804, "grad_norm": 19.375, "learning_rate": 2.222436267274443e-06, "loss": 0.3762, "step": 5154 }, { "epoch": 0.4392841925862804, "grad_norm": 11.75, "learning_rate": 2.2223306771463305e-06, "loss": 0.6674, "step": 5155 }, { "epoch": 0.43936940775458033, "grad_norm": 12.0625, "learning_rate": 2.222225069447225e-06, "loss": 0.6346, "step": 5156 }, { "epoch": 0.4394546229228803, "grad_norm": 18.0, "learning_rate": 2.222119444179035e-06, "loss": 1.0271, "step": 5157 }, { "epoch": 0.43953983809118025, "grad_norm": 23.75, "learning_rate": 2.222013801343669e-06, "loss": 1.0399, "step": 5158 }, { "epoch": 0.4396250532594802, "grad_norm": 15.625, "learning_rate": 2.221908140943036e-06, "loss": 0.9165, "step": 5159 }, { "epoch": 0.43971026842778016, "grad_norm": 11.0, "learning_rate": 2.221802462979046e-06, "loss": 0.401, "step": 5160 }, { "epoch": 0.4397954835960801, "grad_norm": 20.25, "learning_rate": 2.2216967674536074e-06, "loss": 0.7724, "step": 5161 }, { "epoch": 0.4398806987643801, "grad_norm": 10.8125, "learning_rate": 2.2215910543686317e-06, "loss": 0.3656, "step": 5162 }, { "epoch": 0.43996591393268003, "grad_norm": 11.375, "learning_rate": 2.221485323726029e-06, "loss": 0.3966, "step": 5163 }, { "epoch": 0.44005112910098, "grad_norm": 11.1875, "learning_rate": 2.221379575527709e-06, "loss": 0.3691, "step": 5164 }, { "epoch": 0.44013634426927994, "grad_norm": 12.0, "learning_rate": 2.221273809775583e-06, "loss": 0.4409, "step": 5165 }, { "epoch": 0.4402215594375799, "grad_norm": 15.375, "learning_rate": 2.221168026471563e-06, "loss": 1.0218, "step": 5166 }, { "epoch": 0.44030677460587986, "grad_norm": 16.625, "learning_rate": 2.2210622256175606e-06, "loss": 0.7113, "step": 5167 }, { "epoch": 0.4403919897741798, "grad_norm": 13.4375, "learning_rate": 2.2209564072154864e-06, "loss": 0.7158, "step": 5168 }, { "epoch": 0.44047720494247977, "grad_norm": 14.375, "learning_rate": 2.2208505712672543e-06, "loss": 0.739, "step": 5169 }, { "epoch": 0.44056242011077973, "grad_norm": 22.5, "learning_rate": 2.220744717774776e-06, "loss": 1.0637, "step": 5170 }, { "epoch": 0.4406476352790797, "grad_norm": 16.125, "learning_rate": 2.2206388467399642e-06, "loss": 0.7977, "step": 5171 }, { "epoch": 0.44073285044737964, "grad_norm": 12.5625, "learning_rate": 2.2205329581647323e-06, "loss": 0.649, "step": 5172 }, { "epoch": 0.4408180656156796, "grad_norm": 10.5625, "learning_rate": 2.220427052050994e-06, "loss": 0.4986, "step": 5173 }, { "epoch": 0.44090328078397956, "grad_norm": 14.75, "learning_rate": 2.2203211284006632e-06, "loss": 0.9302, "step": 5174 }, { "epoch": 0.4409884959522795, "grad_norm": 11.3125, "learning_rate": 2.2202151872156542e-06, "loss": 0.4803, "step": 5175 }, { "epoch": 0.44107371112057947, "grad_norm": 13.1875, "learning_rate": 2.22010922849788e-06, "loss": 0.5799, "step": 5176 }, { "epoch": 0.4411589262888794, "grad_norm": 10.1875, "learning_rate": 2.220003252249258e-06, "loss": 0.5021, "step": 5177 }, { "epoch": 0.4412441414571794, "grad_norm": 16.25, "learning_rate": 2.219897258471701e-06, "loss": 0.8595, "step": 5178 }, { "epoch": 0.44132935662547934, "grad_norm": 17.25, "learning_rate": 2.219791247167125e-06, "loss": 0.7109, "step": 5179 }, { "epoch": 0.4414145717937793, "grad_norm": 12.125, "learning_rate": 2.2196852183374466e-06, "loss": 0.6619, "step": 5180 }, { "epoch": 0.44149978696207925, "grad_norm": 13.0625, "learning_rate": 2.219579171984581e-06, "loss": 0.5405, "step": 5181 }, { "epoch": 0.4415850021303792, "grad_norm": 11.6875, "learning_rate": 2.2194731081104446e-06, "loss": 0.5325, "step": 5182 }, { "epoch": 0.44167021729867917, "grad_norm": 12.25, "learning_rate": 2.2193670267169544e-06, "loss": 0.7233, "step": 5183 }, { "epoch": 0.4417554324669791, "grad_norm": 18.75, "learning_rate": 2.219260927806027e-06, "loss": 0.7228, "step": 5184 }, { "epoch": 0.4418406476352791, "grad_norm": 12.4375, "learning_rate": 2.2191548113795804e-06, "loss": 0.4754, "step": 5185 }, { "epoch": 0.44192586280357904, "grad_norm": 12.0, "learning_rate": 2.2190486774395315e-06, "loss": 0.677, "step": 5186 }, { "epoch": 0.442011077971879, "grad_norm": 26.375, "learning_rate": 2.2189425259877984e-06, "loss": 0.9605, "step": 5187 }, { "epoch": 0.44209629314017895, "grad_norm": 13.9375, "learning_rate": 2.2188363570263e-06, "loss": 0.7215, "step": 5188 }, { "epoch": 0.4421815083084789, "grad_norm": 12.5, "learning_rate": 2.218730170556954e-06, "loss": 0.5369, "step": 5189 }, { "epoch": 0.44226672347677887, "grad_norm": 15.4375, "learning_rate": 2.2186239665816793e-06, "loss": 0.8766, "step": 5190 }, { "epoch": 0.4423519386450788, "grad_norm": 24.0, "learning_rate": 2.2185177451023965e-06, "loss": 0.8766, "step": 5191 }, { "epoch": 0.4424371538133788, "grad_norm": 8.75, "learning_rate": 2.2184115061210235e-06, "loss": 0.2994, "step": 5192 }, { "epoch": 0.44252236898167874, "grad_norm": 12.875, "learning_rate": 2.218305249639481e-06, "loss": 0.5088, "step": 5193 }, { "epoch": 0.4426075841499787, "grad_norm": 13.3125, "learning_rate": 2.2181989756596887e-06, "loss": 0.5516, "step": 5194 }, { "epoch": 0.44269279931827865, "grad_norm": 13.875, "learning_rate": 2.2180926841835674e-06, "loss": 0.6515, "step": 5195 }, { "epoch": 0.4427780144865786, "grad_norm": 12.8125, "learning_rate": 2.2179863752130377e-06, "loss": 0.6983, "step": 5196 }, { "epoch": 0.44286322965487857, "grad_norm": 17.25, "learning_rate": 2.217880048750021e-06, "loss": 0.9351, "step": 5197 }, { "epoch": 0.4429484448231785, "grad_norm": 15.625, "learning_rate": 2.2177737047964385e-06, "loss": 1.06, "step": 5198 }, { "epoch": 0.4430336599914785, "grad_norm": 13.1875, "learning_rate": 2.217667343354212e-06, "loss": 0.6594, "step": 5199 }, { "epoch": 0.44311887515977844, "grad_norm": 14.875, "learning_rate": 2.2175609644252635e-06, "loss": 0.5199, "step": 5200 }, { "epoch": 0.4432040903280784, "grad_norm": 12.9375, "learning_rate": 2.2174545680115157e-06, "loss": 0.4343, "step": 5201 }, { "epoch": 0.44328930549637835, "grad_norm": 11.25, "learning_rate": 2.2173481541148906e-06, "loss": 0.5872, "step": 5202 }, { "epoch": 0.4433745206646783, "grad_norm": 12.1875, "learning_rate": 2.2172417227373118e-06, "loss": 0.6242, "step": 5203 }, { "epoch": 0.44345973583297826, "grad_norm": 15.4375, "learning_rate": 2.2171352738807024e-06, "loss": 0.5877, "step": 5204 }, { "epoch": 0.4435449510012782, "grad_norm": 14.3125, "learning_rate": 2.217028807546986e-06, "loss": 0.8408, "step": 5205 }, { "epoch": 0.4436301661695782, "grad_norm": 15.5625, "learning_rate": 2.2169223237380864e-06, "loss": 0.829, "step": 5206 }, { "epoch": 0.44371538133787813, "grad_norm": 11.5, "learning_rate": 2.2168158224559283e-06, "loss": 0.4527, "step": 5207 }, { "epoch": 0.4438005965061781, "grad_norm": 10.3125, "learning_rate": 2.2167093037024362e-06, "loss": 0.3851, "step": 5208 }, { "epoch": 0.44388581167447805, "grad_norm": 22.5, "learning_rate": 2.2166027674795346e-06, "loss": 1.1393, "step": 5209 }, { "epoch": 0.443971026842778, "grad_norm": 13.375, "learning_rate": 2.2164962137891493e-06, "loss": 0.5187, "step": 5210 }, { "epoch": 0.44405624201107796, "grad_norm": 17.75, "learning_rate": 2.2163896426332054e-06, "loss": 1.0089, "step": 5211 }, { "epoch": 0.4441414571793779, "grad_norm": 19.75, "learning_rate": 2.216283054013629e-06, "loss": 0.9306, "step": 5212 }, { "epoch": 0.4442266723476779, "grad_norm": 11.5625, "learning_rate": 2.216176447932346e-06, "loss": 0.394, "step": 5213 }, { "epoch": 0.44431188751597783, "grad_norm": 27.0, "learning_rate": 2.216069824391283e-06, "loss": 1.0792, "step": 5214 }, { "epoch": 0.4443971026842778, "grad_norm": 16.5, "learning_rate": 2.2159631833923673e-06, "loss": 0.7715, "step": 5215 }, { "epoch": 0.44448231785257775, "grad_norm": 13.125, "learning_rate": 2.215856524937525e-06, "loss": 0.6128, "step": 5216 }, { "epoch": 0.4445675330208777, "grad_norm": 12.125, "learning_rate": 2.2157498490286845e-06, "loss": 0.6043, "step": 5217 }, { "epoch": 0.44465274818917766, "grad_norm": 13.0625, "learning_rate": 2.2156431556677724e-06, "loss": 0.6091, "step": 5218 }, { "epoch": 0.4447379633574776, "grad_norm": 18.875, "learning_rate": 2.2155364448567178e-06, "loss": 0.6945, "step": 5219 }, { "epoch": 0.4448231785257776, "grad_norm": 20.0, "learning_rate": 2.2154297165974486e-06, "loss": 0.7325, "step": 5220 }, { "epoch": 0.44490839369407753, "grad_norm": 12.375, "learning_rate": 2.215322970891894e-06, "loss": 0.7625, "step": 5221 }, { "epoch": 0.4449936088623775, "grad_norm": 15.5, "learning_rate": 2.2152162077419822e-06, "loss": 1.0698, "step": 5222 }, { "epoch": 0.44507882403067744, "grad_norm": 18.375, "learning_rate": 2.215109427149643e-06, "loss": 0.8082, "step": 5223 }, { "epoch": 0.4451640391989774, "grad_norm": 12.5625, "learning_rate": 2.215002629116806e-06, "loss": 0.5166, "step": 5224 }, { "epoch": 0.44524925436727736, "grad_norm": 15.0625, "learning_rate": 2.214895813645401e-06, "loss": 0.8749, "step": 5225 }, { "epoch": 0.4453344695355773, "grad_norm": 15.1875, "learning_rate": 2.214788980737359e-06, "loss": 0.966, "step": 5226 }, { "epoch": 0.44541968470387727, "grad_norm": 13.0625, "learning_rate": 2.2146821303946093e-06, "loss": 0.7744, "step": 5227 }, { "epoch": 0.44550489987217723, "grad_norm": 13.5625, "learning_rate": 2.2145752626190835e-06, "loss": 0.807, "step": 5228 }, { "epoch": 0.4455901150404772, "grad_norm": 15.25, "learning_rate": 2.214468377412713e-06, "loss": 0.8426, "step": 5229 }, { "epoch": 0.44567533020877714, "grad_norm": 19.125, "learning_rate": 2.214361474777429e-06, "loss": 0.8497, "step": 5230 }, { "epoch": 0.4457605453770771, "grad_norm": 33.0, "learning_rate": 2.2142545547151633e-06, "loss": 1.0836, "step": 5231 }, { "epoch": 0.44584576054537706, "grad_norm": 19.625, "learning_rate": 2.214147617227848e-06, "loss": 0.6466, "step": 5232 }, { "epoch": 0.445930975713677, "grad_norm": 9.625, "learning_rate": 2.214040662317416e-06, "loss": 0.5071, "step": 5233 }, { "epoch": 0.44601619088197697, "grad_norm": 15.0625, "learning_rate": 2.2139336899858e-06, "loss": 0.7577, "step": 5234 }, { "epoch": 0.4461014060502769, "grad_norm": 15.75, "learning_rate": 2.2138267002349324e-06, "loss": 0.992, "step": 5235 }, { "epoch": 0.4461866212185769, "grad_norm": 14.6875, "learning_rate": 2.2137196930667477e-06, "loss": 0.5531, "step": 5236 }, { "epoch": 0.44627183638687684, "grad_norm": 13.5625, "learning_rate": 2.213612668483178e-06, "loss": 0.4828, "step": 5237 }, { "epoch": 0.4463570515551768, "grad_norm": 13.5625, "learning_rate": 2.2135056264861596e-06, "loss": 0.72, "step": 5238 }, { "epoch": 0.44644226672347675, "grad_norm": 17.125, "learning_rate": 2.213398567077625e-06, "loss": 0.6722, "step": 5239 }, { "epoch": 0.4465274818917767, "grad_norm": 14.6875, "learning_rate": 2.21329149025951e-06, "loss": 1.135, "step": 5240 }, { "epoch": 0.44661269706007667, "grad_norm": 18.375, "learning_rate": 2.2131843960337483e-06, "loss": 0.6865, "step": 5241 }, { "epoch": 0.4466979122283766, "grad_norm": 11.125, "learning_rate": 2.213077284402277e-06, "loss": 0.511, "step": 5242 }, { "epoch": 0.4467831273966766, "grad_norm": 25.0, "learning_rate": 2.21297015536703e-06, "loss": 0.7436, "step": 5243 }, { "epoch": 0.44686834256497654, "grad_norm": 12.9375, "learning_rate": 2.212863008929944e-06, "loss": 0.5247, "step": 5244 }, { "epoch": 0.4469535577332765, "grad_norm": 24.0, "learning_rate": 2.2127558450929556e-06, "loss": 1.2036, "step": 5245 }, { "epoch": 0.44703877290157645, "grad_norm": 13.875, "learning_rate": 2.212648663858001e-06, "loss": 0.7338, "step": 5246 }, { "epoch": 0.44712398806987647, "grad_norm": 20.375, "learning_rate": 2.2125414652270162e-06, "loss": 0.6491, "step": 5247 }, { "epoch": 0.4472092032381764, "grad_norm": 16.25, "learning_rate": 2.21243424920194e-06, "loss": 0.9395, "step": 5248 }, { "epoch": 0.4472944184064764, "grad_norm": 18.75, "learning_rate": 2.212327015784709e-06, "loss": 0.6624, "step": 5249 }, { "epoch": 0.44737963357477634, "grad_norm": 19.5, "learning_rate": 2.2122197649772608e-06, "loss": 0.9869, "step": 5250 }, { "epoch": 0.4474648487430763, "grad_norm": 19.0, "learning_rate": 2.212112496781534e-06, "loss": 0.717, "step": 5251 }, { "epoch": 0.44755006391137625, "grad_norm": 13.8125, "learning_rate": 2.212005211199467e-06, "loss": 0.7496, "step": 5252 }, { "epoch": 0.4476352790796762, "grad_norm": 16.5, "learning_rate": 2.2118979082329984e-06, "loss": 0.7765, "step": 5253 }, { "epoch": 0.44772049424797616, "grad_norm": 14.75, "learning_rate": 2.2117905878840678e-06, "loss": 0.9994, "step": 5254 }, { "epoch": 0.4478057094162761, "grad_norm": 18.0, "learning_rate": 2.2116832501546135e-06, "loss": 0.9808, "step": 5255 }, { "epoch": 0.4478909245845761, "grad_norm": 20.875, "learning_rate": 2.211575895046576e-06, "loss": 0.8259, "step": 5256 }, { "epoch": 0.44797613975287603, "grad_norm": 22.0, "learning_rate": 2.2114685225618953e-06, "loss": 1.1549, "step": 5257 }, { "epoch": 0.448061354921176, "grad_norm": 12.875, "learning_rate": 2.211361132702511e-06, "loss": 0.6495, "step": 5258 }, { "epoch": 0.44814657008947595, "grad_norm": 16.125, "learning_rate": 2.2112537254703647e-06, "loss": 0.8431, "step": 5259 }, { "epoch": 0.4482317852577759, "grad_norm": 11.8125, "learning_rate": 2.2111463008673974e-06, "loss": 0.6806, "step": 5260 }, { "epoch": 0.44831700042607586, "grad_norm": 13.1875, "learning_rate": 2.2110388588955495e-06, "loss": 0.5808, "step": 5261 }, { "epoch": 0.4484022155943758, "grad_norm": 10.75, "learning_rate": 2.2109313995567628e-06, "loss": 0.4957, "step": 5262 }, { "epoch": 0.4484874307626758, "grad_norm": 11.25, "learning_rate": 2.2108239228529798e-06, "loss": 0.3631, "step": 5263 }, { "epoch": 0.44857264593097573, "grad_norm": 14.5625, "learning_rate": 2.210716428786142e-06, "loss": 0.7375, "step": 5264 }, { "epoch": 0.4486578610992757, "grad_norm": 12.875, "learning_rate": 2.2106089173581926e-06, "loss": 0.4065, "step": 5265 }, { "epoch": 0.44874307626757565, "grad_norm": 13.6875, "learning_rate": 2.210501388571074e-06, "loss": 0.7736, "step": 5266 }, { "epoch": 0.4488282914358756, "grad_norm": 11.625, "learning_rate": 2.2103938424267292e-06, "loss": 0.5118, "step": 5267 }, { "epoch": 0.44891350660417556, "grad_norm": 25.125, "learning_rate": 2.2102862789271024e-06, "loss": 0.8588, "step": 5268 }, { "epoch": 0.4489987217724755, "grad_norm": 37.0, "learning_rate": 2.210178698074136e-06, "loss": 1.5926, "step": 5269 }, { "epoch": 0.4490839369407755, "grad_norm": 19.125, "learning_rate": 2.210071099869776e-06, "loss": 1.0543, "step": 5270 }, { "epoch": 0.44916915210907543, "grad_norm": 17.125, "learning_rate": 2.2099634843159654e-06, "loss": 0.9676, "step": 5271 }, { "epoch": 0.4492543672773754, "grad_norm": 17.75, "learning_rate": 2.2098558514146496e-06, "loss": 0.5011, "step": 5272 }, { "epoch": 0.44933958244567535, "grad_norm": 11.625, "learning_rate": 2.2097482011677727e-06, "loss": 0.6006, "step": 5273 }, { "epoch": 0.4494247976139753, "grad_norm": 14.6875, "learning_rate": 2.2096405335772813e-06, "loss": 0.8925, "step": 5274 }, { "epoch": 0.44951001278227526, "grad_norm": 14.75, "learning_rate": 2.20953284864512e-06, "loss": 0.6267, "step": 5275 }, { "epoch": 0.4495952279505752, "grad_norm": 14.0625, "learning_rate": 2.2094251463732357e-06, "loss": 0.8305, "step": 5276 }, { "epoch": 0.4496804431188752, "grad_norm": 13.875, "learning_rate": 2.209317426763574e-06, "loss": 0.9639, "step": 5277 }, { "epoch": 0.44976565828717513, "grad_norm": 14.75, "learning_rate": 2.2092096898180816e-06, "loss": 0.7304, "step": 5278 }, { "epoch": 0.4498508734554751, "grad_norm": 16.625, "learning_rate": 2.209101935538706e-06, "loss": 0.9641, "step": 5279 }, { "epoch": 0.44993608862377504, "grad_norm": 33.75, "learning_rate": 2.2089941639273936e-06, "loss": 1.5391, "step": 5280 }, { "epoch": 0.450021303792075, "grad_norm": 17.875, "learning_rate": 2.208886374986092e-06, "loss": 0.9008, "step": 5281 }, { "epoch": 0.45010651896037496, "grad_norm": 14.8125, "learning_rate": 2.20877856871675e-06, "loss": 0.8533, "step": 5282 }, { "epoch": 0.4501917341286749, "grad_norm": 13.0625, "learning_rate": 2.208670745121315e-06, "loss": 0.7125, "step": 5283 }, { "epoch": 0.45027694929697487, "grad_norm": 15.0, "learning_rate": 2.2085629042017353e-06, "loss": 0.8686, "step": 5284 }, { "epoch": 0.45036216446527483, "grad_norm": 15.5, "learning_rate": 2.2084550459599606e-06, "loss": 0.8119, "step": 5285 }, { "epoch": 0.4504473796335748, "grad_norm": 21.875, "learning_rate": 2.208347170397939e-06, "loss": 1.0876, "step": 5286 }, { "epoch": 0.45053259480187474, "grad_norm": 16.25, "learning_rate": 2.2082392775176205e-06, "loss": 1.1659, "step": 5287 }, { "epoch": 0.4506178099701747, "grad_norm": 11.75, "learning_rate": 2.2081313673209544e-06, "loss": 0.6577, "step": 5288 }, { "epoch": 0.45070302513847466, "grad_norm": 11.0, "learning_rate": 2.208023439809891e-06, "loss": 0.3374, "step": 5289 }, { "epoch": 0.4507882403067746, "grad_norm": 15.1875, "learning_rate": 2.2079154949863806e-06, "loss": 0.8558, "step": 5290 }, { "epoch": 0.45087345547507457, "grad_norm": 18.5, "learning_rate": 2.2078075328523744e-06, "loss": 0.9127, "step": 5291 }, { "epoch": 0.4509586706433745, "grad_norm": 25.75, "learning_rate": 2.207699553409823e-06, "loss": 1.2323, "step": 5292 }, { "epoch": 0.4510438858116745, "grad_norm": 10.125, "learning_rate": 2.207591556660677e-06, "loss": 0.4612, "step": 5293 }, { "epoch": 0.45112910097997444, "grad_norm": 15.125, "learning_rate": 2.207483542606889e-06, "loss": 0.605, "step": 5294 }, { "epoch": 0.4512143161482744, "grad_norm": 13.8125, "learning_rate": 2.2073755112504105e-06, "loss": 0.9879, "step": 5295 }, { "epoch": 0.45129953131657435, "grad_norm": 15.3125, "learning_rate": 2.207267462593194e-06, "loss": 0.5696, "step": 5296 }, { "epoch": 0.4513847464848743, "grad_norm": 13.8125, "learning_rate": 2.207159396637192e-06, "loss": 0.7613, "step": 5297 }, { "epoch": 0.45146996165317427, "grad_norm": 13.8125, "learning_rate": 2.2070513133843568e-06, "loss": 0.7534, "step": 5298 }, { "epoch": 0.4515551768214742, "grad_norm": 21.75, "learning_rate": 2.206943212836642e-06, "loss": 1.1458, "step": 5299 }, { "epoch": 0.4516403919897742, "grad_norm": 13.375, "learning_rate": 2.206835094996001e-06, "loss": 0.5133, "step": 5300 }, { "epoch": 0.45172560715807414, "grad_norm": 15.625, "learning_rate": 2.2067269598643874e-06, "loss": 0.8986, "step": 5301 }, { "epoch": 0.4518108223263741, "grad_norm": 12.5625, "learning_rate": 2.2066188074437562e-06, "loss": 0.6559, "step": 5302 }, { "epoch": 0.45189603749467405, "grad_norm": 13.75, "learning_rate": 2.2065106377360614e-06, "loss": 0.6126, "step": 5303 }, { "epoch": 0.451981252662974, "grad_norm": 16.125, "learning_rate": 2.206402450743257e-06, "loss": 0.7747, "step": 5304 }, { "epoch": 0.45206646783127397, "grad_norm": 16.25, "learning_rate": 2.2062942464672984e-06, "loss": 0.9428, "step": 5305 }, { "epoch": 0.4521516829995739, "grad_norm": 18.375, "learning_rate": 2.2061860249101414e-06, "loss": 0.9978, "step": 5306 }, { "epoch": 0.4522368981678739, "grad_norm": 14.1875, "learning_rate": 2.2060777860737415e-06, "loss": 0.5861, "step": 5307 }, { "epoch": 0.45232211333617384, "grad_norm": 15.0, "learning_rate": 2.2059695299600543e-06, "loss": 0.836, "step": 5308 }, { "epoch": 0.4524073285044738, "grad_norm": 16.625, "learning_rate": 2.205861256571037e-06, "loss": 0.8986, "step": 5309 }, { "epoch": 0.45249254367277375, "grad_norm": 19.125, "learning_rate": 2.205752965908645e-06, "loss": 0.7699, "step": 5310 }, { "epoch": 0.4525777588410737, "grad_norm": 18.625, "learning_rate": 2.205644657974836e-06, "loss": 0.5925, "step": 5311 }, { "epoch": 0.45266297400937366, "grad_norm": 9.75, "learning_rate": 2.205536332771567e-06, "loss": 0.3757, "step": 5312 }, { "epoch": 0.4527481891776736, "grad_norm": 11.3125, "learning_rate": 2.205427990300796e-06, "loss": 0.4661, "step": 5313 }, { "epoch": 0.4528334043459736, "grad_norm": 11.3125, "learning_rate": 2.20531963056448e-06, "loss": 0.5236, "step": 5314 }, { "epoch": 0.45291861951427353, "grad_norm": 17.5, "learning_rate": 2.2052112535645777e-06, "loss": 1.0352, "step": 5315 }, { "epoch": 0.4530038346825735, "grad_norm": 14.375, "learning_rate": 2.2051028593030474e-06, "loss": 0.8625, "step": 5316 }, { "epoch": 0.45308904985087345, "grad_norm": 15.375, "learning_rate": 2.204994447781848e-06, "loss": 0.6187, "step": 5317 }, { "epoch": 0.4531742650191734, "grad_norm": 11.5625, "learning_rate": 2.2048860190029393e-06, "loss": 0.6228, "step": 5318 }, { "epoch": 0.45325948018747336, "grad_norm": 17.625, "learning_rate": 2.204777572968279e-06, "loss": 0.6301, "step": 5319 }, { "epoch": 0.4533446953557733, "grad_norm": 14.5625, "learning_rate": 2.2046691096798285e-06, "loss": 1.1456, "step": 5320 }, { "epoch": 0.4534299105240733, "grad_norm": 19.625, "learning_rate": 2.2045606291395472e-06, "loss": 0.897, "step": 5321 }, { "epoch": 0.45351512569237323, "grad_norm": 13.6875, "learning_rate": 2.204452131349395e-06, "loss": 0.5846, "step": 5322 }, { "epoch": 0.4536003408606732, "grad_norm": 19.875, "learning_rate": 2.2043436163113334e-06, "loss": 0.7718, "step": 5323 }, { "epoch": 0.45368555602897315, "grad_norm": 18.625, "learning_rate": 2.204235084027323e-06, "loss": 0.6552, "step": 5324 }, { "epoch": 0.4537707711972731, "grad_norm": 10.5625, "learning_rate": 2.204126534499325e-06, "loss": 0.3773, "step": 5325 }, { "epoch": 0.45385598636557306, "grad_norm": 16.75, "learning_rate": 2.204017967729301e-06, "loss": 0.8311, "step": 5326 }, { "epoch": 0.453941201533873, "grad_norm": 14.5625, "learning_rate": 2.203909383719213e-06, "loss": 1.0692, "step": 5327 }, { "epoch": 0.454026416702173, "grad_norm": 16.875, "learning_rate": 2.2038007824710232e-06, "loss": 0.342, "step": 5328 }, { "epoch": 0.45411163187047293, "grad_norm": 11.5625, "learning_rate": 2.203692163986694e-06, "loss": 0.3911, "step": 5329 }, { "epoch": 0.4541968470387729, "grad_norm": 11.25, "learning_rate": 2.2035835282681888e-06, "loss": 0.4331, "step": 5330 }, { "epoch": 0.45428206220707285, "grad_norm": 12.1875, "learning_rate": 2.2034748753174703e-06, "loss": 0.2653, "step": 5331 }, { "epoch": 0.4543672773753728, "grad_norm": 16.5, "learning_rate": 2.203366205136502e-06, "loss": 0.8055, "step": 5332 }, { "epoch": 0.45445249254367276, "grad_norm": 17.875, "learning_rate": 2.2032575177272473e-06, "loss": 1.0661, "step": 5333 }, { "epoch": 0.4545377077119727, "grad_norm": 14.625, "learning_rate": 2.2031488130916706e-06, "loss": 0.9893, "step": 5334 }, { "epoch": 0.4546229228802727, "grad_norm": 13.0, "learning_rate": 2.203040091231737e-06, "loss": 0.6344, "step": 5335 }, { "epoch": 0.45470813804857263, "grad_norm": 10.4375, "learning_rate": 2.20293135214941e-06, "loss": 0.4345, "step": 5336 }, { "epoch": 0.4547933532168726, "grad_norm": 15.0625, "learning_rate": 2.202822595846656e-06, "loss": 0.5901, "step": 5337 }, { "epoch": 0.45487856838517254, "grad_norm": 28.0, "learning_rate": 2.2027138223254386e-06, "loss": 1.2241, "step": 5338 }, { "epoch": 0.4549637835534725, "grad_norm": 19.375, "learning_rate": 2.2026050315877253e-06, "loss": 0.8944, "step": 5339 }, { "epoch": 0.45504899872177246, "grad_norm": 23.0, "learning_rate": 2.20249622363548e-06, "loss": 0.8549, "step": 5340 }, { "epoch": 0.4551342138900724, "grad_norm": 21.75, "learning_rate": 2.2023873984706713e-06, "loss": 1.0132, "step": 5341 }, { "epoch": 0.45521942905837237, "grad_norm": 16.375, "learning_rate": 2.2022785560952643e-06, "loss": 0.8309, "step": 5342 }, { "epoch": 0.45530464422667233, "grad_norm": 15.5625, "learning_rate": 2.202169696511226e-06, "loss": 0.9, "step": 5343 }, { "epoch": 0.4553898593949723, "grad_norm": 17.375, "learning_rate": 2.2020608197205234e-06, "loss": 0.6411, "step": 5344 }, { "epoch": 0.45547507456327224, "grad_norm": 14.8125, "learning_rate": 2.201951925725125e-06, "loss": 0.6685, "step": 5345 }, { "epoch": 0.4555602897315722, "grad_norm": 13.6875, "learning_rate": 2.201843014526998e-06, "loss": 0.7465, "step": 5346 }, { "epoch": 0.45564550489987216, "grad_norm": 19.5, "learning_rate": 2.2017340861281106e-06, "loss": 0.7149, "step": 5347 }, { "epoch": 0.4557307200681721, "grad_norm": 12.375, "learning_rate": 2.201625140530431e-06, "loss": 0.5987, "step": 5348 }, { "epoch": 0.45581593523647207, "grad_norm": 18.0, "learning_rate": 2.201516177735928e-06, "loss": 0.9223, "step": 5349 }, { "epoch": 0.455901150404772, "grad_norm": 11.25, "learning_rate": 2.201407197746571e-06, "loss": 0.5194, "step": 5350 }, { "epoch": 0.455986365573072, "grad_norm": 20.0, "learning_rate": 2.2012982005643296e-06, "loss": 1.2225, "step": 5351 }, { "epoch": 0.45607158074137194, "grad_norm": 13.9375, "learning_rate": 2.2011891861911725e-06, "loss": 0.7037, "step": 5352 }, { "epoch": 0.4561567959096719, "grad_norm": 15.875, "learning_rate": 2.2010801546290712e-06, "loss": 0.6425, "step": 5353 }, { "epoch": 0.45624201107797185, "grad_norm": 11.75, "learning_rate": 2.2009711058799943e-06, "loss": 0.5349, "step": 5354 }, { "epoch": 0.4563272262462718, "grad_norm": 24.0, "learning_rate": 2.200862039945914e-06, "loss": 1.1827, "step": 5355 }, { "epoch": 0.45641244141457177, "grad_norm": 12.75, "learning_rate": 2.2007529568287995e-06, "loss": 0.7241, "step": 5356 }, { "epoch": 0.4564976565828717, "grad_norm": 18.0, "learning_rate": 2.200643856530624e-06, "loss": 0.511, "step": 5357 }, { "epoch": 0.4565828717511717, "grad_norm": 15.0, "learning_rate": 2.200534739053357e-06, "loss": 0.7864, "step": 5358 }, { "epoch": 0.45666808691947164, "grad_norm": 14.25, "learning_rate": 2.2004256043989723e-06, "loss": 0.8832, "step": 5359 }, { "epoch": 0.45675330208777165, "grad_norm": 14.8125, "learning_rate": 2.200316452569441e-06, "loss": 0.6291, "step": 5360 }, { "epoch": 0.4568385172560716, "grad_norm": 17.125, "learning_rate": 2.2002072835667353e-06, "loss": 0.8866, "step": 5361 }, { "epoch": 0.45692373242437156, "grad_norm": 14.4375, "learning_rate": 2.2000980973928294e-06, "loss": 0.7363, "step": 5362 }, { "epoch": 0.4570089475926715, "grad_norm": 14.1875, "learning_rate": 2.199988894049695e-06, "loss": 0.8368, "step": 5363 }, { "epoch": 0.4570941627609715, "grad_norm": 14.125, "learning_rate": 2.1998796735393056e-06, "loss": 0.929, "step": 5364 }, { "epoch": 0.45717937792927144, "grad_norm": 13.9375, "learning_rate": 2.1997704358636357e-06, "loss": 0.5921, "step": 5365 }, { "epoch": 0.4572645930975714, "grad_norm": 14.75, "learning_rate": 2.1996611810246592e-06, "loss": 0.7622, "step": 5366 }, { "epoch": 0.45734980826587135, "grad_norm": 15.0, "learning_rate": 2.1995519090243497e-06, "loss": 0.4529, "step": 5367 }, { "epoch": 0.4574350234341713, "grad_norm": 14.1875, "learning_rate": 2.1994426198646828e-06, "loss": 0.7882, "step": 5368 }, { "epoch": 0.45752023860247126, "grad_norm": 11.1875, "learning_rate": 2.1993333135476325e-06, "loss": 0.4896, "step": 5369 }, { "epoch": 0.4576054537707712, "grad_norm": 11.3125, "learning_rate": 2.199223990075175e-06, "loss": 0.469, "step": 5370 }, { "epoch": 0.4576906689390712, "grad_norm": 16.0, "learning_rate": 2.1991146494492853e-06, "loss": 0.7332, "step": 5371 }, { "epoch": 0.45777588410737113, "grad_norm": 12.75, "learning_rate": 2.1990052916719397e-06, "loss": 0.6996, "step": 5372 }, { "epoch": 0.4578610992756711, "grad_norm": 14.6875, "learning_rate": 2.198895916745114e-06, "loss": 0.7664, "step": 5373 }, { "epoch": 0.45794631444397105, "grad_norm": 32.5, "learning_rate": 2.198786524670785e-06, "loss": 0.9411, "step": 5374 }, { "epoch": 0.458031529612271, "grad_norm": 15.0625, "learning_rate": 2.1986771154509293e-06, "loss": 0.9394, "step": 5375 }, { "epoch": 0.45811674478057096, "grad_norm": 13.625, "learning_rate": 2.1985676890875243e-06, "loss": 0.6836, "step": 5376 }, { "epoch": 0.4582019599488709, "grad_norm": 15.625, "learning_rate": 2.198458245582547e-06, "loss": 0.5596, "step": 5377 }, { "epoch": 0.4582871751171709, "grad_norm": 12.125, "learning_rate": 2.198348784937976e-06, "loss": 0.6925, "step": 5378 }, { "epoch": 0.45837239028547083, "grad_norm": 13.375, "learning_rate": 2.1982393071557884e-06, "loss": 0.7237, "step": 5379 }, { "epoch": 0.4584576054537708, "grad_norm": 15.125, "learning_rate": 2.198129812237963e-06, "loss": 0.9014, "step": 5380 }, { "epoch": 0.45854282062207075, "grad_norm": 15.5625, "learning_rate": 2.1980203001864784e-06, "loss": 0.7805, "step": 5381 }, { "epoch": 0.4586280357903707, "grad_norm": 18.875, "learning_rate": 2.1979107710033135e-06, "loss": 0.6593, "step": 5382 }, { "epoch": 0.45871325095867066, "grad_norm": 17.625, "learning_rate": 2.1978012246904485e-06, "loss": 0.7923, "step": 5383 }, { "epoch": 0.4587984661269706, "grad_norm": 12.9375, "learning_rate": 2.197691661249862e-06, "loss": 0.5387, "step": 5384 }, { "epoch": 0.4588836812952706, "grad_norm": 13.125, "learning_rate": 2.197582080683534e-06, "loss": 0.6742, "step": 5385 }, { "epoch": 0.45896889646357053, "grad_norm": 13.125, "learning_rate": 2.197472482993445e-06, "loss": 0.6558, "step": 5386 }, { "epoch": 0.4590541116318705, "grad_norm": 14.3125, "learning_rate": 2.197362868181576e-06, "loss": 0.735, "step": 5387 }, { "epoch": 0.45913932680017044, "grad_norm": 13.0, "learning_rate": 2.1972532362499064e-06, "loss": 0.712, "step": 5388 }, { "epoch": 0.4592245419684704, "grad_norm": 12.1875, "learning_rate": 2.197143587200419e-06, "loss": 0.5596, "step": 5389 }, { "epoch": 0.45930975713677036, "grad_norm": 18.875, "learning_rate": 2.197033921035094e-06, "loss": 1.0656, "step": 5390 }, { "epoch": 0.4593949723050703, "grad_norm": 9.8125, "learning_rate": 2.196924237755914e-06, "loss": 0.9935, "step": 5391 }, { "epoch": 0.45948018747337027, "grad_norm": 14.0625, "learning_rate": 2.1968145373648607e-06, "loss": 0.6395, "step": 5392 }, { "epoch": 0.45956540264167023, "grad_norm": 14.3125, "learning_rate": 2.1967048198639167e-06, "loss": 0.8226, "step": 5393 }, { "epoch": 0.4596506178099702, "grad_norm": 10.3125, "learning_rate": 2.196595085255064e-06, "loss": 0.4611, "step": 5394 }, { "epoch": 0.45973583297827014, "grad_norm": 14.625, "learning_rate": 2.1964853335402873e-06, "loss": 0.6716, "step": 5395 }, { "epoch": 0.4598210481465701, "grad_norm": 11.875, "learning_rate": 2.1963755647215684e-06, "loss": 0.3015, "step": 5396 }, { "epoch": 0.45990626331487006, "grad_norm": 9.625, "learning_rate": 2.1962657788008915e-06, "loss": 0.3835, "step": 5397 }, { "epoch": 0.45999147848317, "grad_norm": 11.6875, "learning_rate": 2.1961559757802405e-06, "loss": 0.4582, "step": 5398 }, { "epoch": 0.46007669365146997, "grad_norm": 13.5625, "learning_rate": 2.1960461556615993e-06, "loss": 0.8695, "step": 5399 }, { "epoch": 0.4601619088197699, "grad_norm": 10.75, "learning_rate": 2.195936318446953e-06, "loss": 0.5271, "step": 5400 }, { "epoch": 0.4602471239880699, "grad_norm": 13.1875, "learning_rate": 2.195826464138286e-06, "loss": 0.6792, "step": 5401 }, { "epoch": 0.46033233915636984, "grad_norm": 10.375, "learning_rate": 2.1957165927375836e-06, "loss": 0.458, "step": 5402 }, { "epoch": 0.4604175543246698, "grad_norm": 13.0, "learning_rate": 2.1956067042468317e-06, "loss": 0.6812, "step": 5403 }, { "epoch": 0.46050276949296975, "grad_norm": 17.25, "learning_rate": 2.1954967986680155e-06, "loss": 0.7756, "step": 5404 }, { "epoch": 0.4605879846612697, "grad_norm": 13.875, "learning_rate": 2.1953868760031218e-06, "loss": 0.8689, "step": 5405 }, { "epoch": 0.46067319982956967, "grad_norm": 16.375, "learning_rate": 2.1952769362541363e-06, "loss": 0.6738, "step": 5406 }, { "epoch": 0.4607584149978696, "grad_norm": 10.375, "learning_rate": 2.195166979423046e-06, "loss": 0.442, "step": 5407 }, { "epoch": 0.4608436301661696, "grad_norm": 13.8125, "learning_rate": 2.1950570055118376e-06, "loss": 0.7507, "step": 5408 }, { "epoch": 0.46092884533446954, "grad_norm": 16.125, "learning_rate": 2.1949470145224995e-06, "loss": 0.6132, "step": 5409 }, { "epoch": 0.4610140605027695, "grad_norm": 11.75, "learning_rate": 2.194837006457018e-06, "loss": 0.8065, "step": 5410 }, { "epoch": 0.46109927567106945, "grad_norm": 15.125, "learning_rate": 2.1947269813173823e-06, "loss": 0.6637, "step": 5411 }, { "epoch": 0.4611844908393694, "grad_norm": 11.5625, "learning_rate": 2.1946169391055795e-06, "loss": 0.4672, "step": 5412 }, { "epoch": 0.46126970600766937, "grad_norm": 12.1875, "learning_rate": 2.194506879823599e-06, "loss": 0.4659, "step": 5413 }, { "epoch": 0.4613549211759693, "grad_norm": 14.8125, "learning_rate": 2.1943968034734292e-06, "loss": 0.7119, "step": 5414 }, { "epoch": 0.4614401363442693, "grad_norm": 15.0625, "learning_rate": 2.1942867100570597e-06, "loss": 0.8455, "step": 5415 }, { "epoch": 0.46152535151256924, "grad_norm": 13.125, "learning_rate": 2.1941765995764796e-06, "loss": 0.8413, "step": 5416 }, { "epoch": 0.4616105666808692, "grad_norm": 13.25, "learning_rate": 2.194066472033679e-06, "loss": 0.7852, "step": 5417 }, { "epoch": 0.46169578184916915, "grad_norm": 9.875, "learning_rate": 2.193956327430648e-06, "loss": 0.3056, "step": 5418 }, { "epoch": 0.4617809970174691, "grad_norm": 13.3125, "learning_rate": 2.1938461657693767e-06, "loss": 0.7517, "step": 5419 }, { "epoch": 0.46186621218576907, "grad_norm": 33.0, "learning_rate": 2.1937359870518567e-06, "loss": 0.5952, "step": 5420 }, { "epoch": 0.461951427354069, "grad_norm": 15.5, "learning_rate": 2.193625791280078e-06, "loss": 0.5737, "step": 5421 }, { "epoch": 0.462036642522369, "grad_norm": 15.1875, "learning_rate": 2.193515578456032e-06, "loss": 0.8829, "step": 5422 }, { "epoch": 0.46212185769066894, "grad_norm": 16.125, "learning_rate": 2.1934053485817113e-06, "loss": 0.715, "step": 5423 }, { "epoch": 0.4622070728589689, "grad_norm": 14.1875, "learning_rate": 2.193295101659107e-06, "loss": 0.985, "step": 5424 }, { "epoch": 0.46229228802726885, "grad_norm": 12.0625, "learning_rate": 2.1931848376902115e-06, "loss": 0.4106, "step": 5425 }, { "epoch": 0.4623775031955688, "grad_norm": 19.0, "learning_rate": 2.193074556677018e-06, "loss": 0.8552, "step": 5426 }, { "epoch": 0.46246271836386876, "grad_norm": 14.375, "learning_rate": 2.1929642586215186e-06, "loss": 0.7395, "step": 5427 }, { "epoch": 0.4625479335321687, "grad_norm": 14.625, "learning_rate": 2.192853943525707e-06, "loss": 0.6328, "step": 5428 }, { "epoch": 0.4626331487004687, "grad_norm": 10.6875, "learning_rate": 2.1927436113915762e-06, "loss": 0.3064, "step": 5429 }, { "epoch": 0.46271836386876863, "grad_norm": 23.375, "learning_rate": 2.1926332622211206e-06, "loss": 0.9315, "step": 5430 }, { "epoch": 0.4628035790370686, "grad_norm": 13.375, "learning_rate": 2.192522896016334e-06, "loss": 0.6818, "step": 5431 }, { "epoch": 0.46288879420536855, "grad_norm": 14.875, "learning_rate": 2.192412512779211e-06, "loss": 0.6938, "step": 5432 }, { "epoch": 0.4629740093736685, "grad_norm": 13.875, "learning_rate": 2.192302112511746e-06, "loss": 0.6912, "step": 5433 }, { "epoch": 0.46305922454196846, "grad_norm": 14.8125, "learning_rate": 2.1921916952159345e-06, "loss": 0.651, "step": 5434 }, { "epoch": 0.4631444397102684, "grad_norm": 12.875, "learning_rate": 2.1920812608937713e-06, "loss": 0.59, "step": 5435 }, { "epoch": 0.4632296548785684, "grad_norm": 11.875, "learning_rate": 2.1919708095472526e-06, "loss": 0.6347, "step": 5436 }, { "epoch": 0.46331487004686833, "grad_norm": 13.8125, "learning_rate": 2.1918603411783743e-06, "loss": 0.5928, "step": 5437 }, { "epoch": 0.4634000852151683, "grad_norm": 14.3125, "learning_rate": 2.191749855789132e-06, "loss": 0.7762, "step": 5438 }, { "epoch": 0.46348530038346825, "grad_norm": 16.75, "learning_rate": 2.1916393533815234e-06, "loss": 0.8205, "step": 5439 }, { "epoch": 0.4635705155517682, "grad_norm": 12.25, "learning_rate": 2.1915288339575445e-06, "loss": 0.2623, "step": 5440 }, { "epoch": 0.46365573072006816, "grad_norm": 13.3125, "learning_rate": 2.1914182975191927e-06, "loss": 0.3928, "step": 5441 }, { "epoch": 0.4637409458883681, "grad_norm": 12.5, "learning_rate": 2.191307744068466e-06, "loss": 0.6305, "step": 5442 }, { "epoch": 0.4638261610566681, "grad_norm": 15.625, "learning_rate": 2.1911971736073617e-06, "loss": 0.6561, "step": 5443 }, { "epoch": 0.46391137622496803, "grad_norm": 19.125, "learning_rate": 2.191086586137878e-06, "loss": 0.9618, "step": 5444 }, { "epoch": 0.463996591393268, "grad_norm": 12.3125, "learning_rate": 2.190975981662013e-06, "loss": 0.3177, "step": 5445 }, { "epoch": 0.46408180656156794, "grad_norm": 15.3125, "learning_rate": 2.190865360181766e-06, "loss": 0.8093, "step": 5446 }, { "epoch": 0.4641670217298679, "grad_norm": 13.6875, "learning_rate": 2.190754721699136e-06, "loss": 0.7158, "step": 5447 }, { "epoch": 0.46425223689816786, "grad_norm": 16.125, "learning_rate": 2.1906440662161215e-06, "loss": 0.5968, "step": 5448 }, { "epoch": 0.4643374520664678, "grad_norm": 19.75, "learning_rate": 2.1905333937347235e-06, "loss": 1.0111, "step": 5449 }, { "epoch": 0.46442266723476777, "grad_norm": 14.875, "learning_rate": 2.190422704256941e-06, "loss": 0.7625, "step": 5450 }, { "epoch": 0.46450788240306773, "grad_norm": 14.0625, "learning_rate": 2.190311997784775e-06, "loss": 0.7636, "step": 5451 }, { "epoch": 0.4645930975713677, "grad_norm": 16.375, "learning_rate": 2.1902012743202254e-06, "loss": 1.1334, "step": 5452 }, { "epoch": 0.46467831273966764, "grad_norm": 13.125, "learning_rate": 2.190090533865293e-06, "loss": 0.7281, "step": 5453 }, { "epoch": 0.4647635279079676, "grad_norm": 21.25, "learning_rate": 2.18997977642198e-06, "loss": 0.9653, "step": 5454 }, { "epoch": 0.46484874307626756, "grad_norm": 11.75, "learning_rate": 2.189869001992287e-06, "loss": 0.4201, "step": 5455 }, { "epoch": 0.4649339582445675, "grad_norm": 15.5625, "learning_rate": 2.1897582105782156e-06, "loss": 0.6026, "step": 5456 }, { "epoch": 0.46501917341286747, "grad_norm": 11.875, "learning_rate": 2.1896474021817686e-06, "loss": 0.5422, "step": 5457 }, { "epoch": 0.4651043885811674, "grad_norm": 15.5, "learning_rate": 2.189536576804948e-06, "loss": 0.4348, "step": 5458 }, { "epoch": 0.4651896037494674, "grad_norm": 15.1875, "learning_rate": 2.189425734449757e-06, "loss": 0.5475, "step": 5459 }, { "epoch": 0.46527481891776734, "grad_norm": 19.5, "learning_rate": 2.189314875118198e-06, "loss": 0.68, "step": 5460 }, { "epoch": 0.4653600340860673, "grad_norm": 15.6875, "learning_rate": 2.189203998812275e-06, "loss": 0.7498, "step": 5461 }, { "epoch": 0.46544524925436725, "grad_norm": 15.6875, "learning_rate": 2.189093105533991e-06, "loss": 1.0508, "step": 5462 }, { "epoch": 0.4655304644226672, "grad_norm": 13.6875, "learning_rate": 2.18898219528535e-06, "loss": 0.6773, "step": 5463 }, { "epoch": 0.46561567959096717, "grad_norm": 16.0, "learning_rate": 2.188871268068357e-06, "loss": 0.7554, "step": 5464 }, { "epoch": 0.4657008947592671, "grad_norm": 12.1875, "learning_rate": 2.188760323885016e-06, "loss": 0.6219, "step": 5465 }, { "epoch": 0.4657861099275671, "grad_norm": 11.875, "learning_rate": 2.188649362737332e-06, "loss": 0.5974, "step": 5466 }, { "epoch": 0.46587132509586704, "grad_norm": 11.6875, "learning_rate": 2.18853838462731e-06, "loss": 0.5459, "step": 5467 }, { "epoch": 0.465956540264167, "grad_norm": 28.375, "learning_rate": 2.1884273895569553e-06, "loss": 0.9978, "step": 5468 }, { "epoch": 0.46604175543246695, "grad_norm": 10.5625, "learning_rate": 2.1883163775282744e-06, "loss": 0.4482, "step": 5469 }, { "epoch": 0.4661269706007669, "grad_norm": 16.875, "learning_rate": 2.188205348543273e-06, "loss": 0.6727, "step": 5470 }, { "epoch": 0.46621218576906687, "grad_norm": 14.3125, "learning_rate": 2.1880943026039575e-06, "loss": 0.5859, "step": 5471 }, { "epoch": 0.4662974009373668, "grad_norm": 12.875, "learning_rate": 2.1879832397123347e-06, "loss": 0.4882, "step": 5472 }, { "epoch": 0.46638261610566684, "grad_norm": 14.125, "learning_rate": 2.1878721598704112e-06, "loss": 0.9089, "step": 5473 }, { "epoch": 0.4664678312739668, "grad_norm": 13.3125, "learning_rate": 2.187761063080195e-06, "loss": 0.659, "step": 5474 }, { "epoch": 0.46655304644226675, "grad_norm": 18.125, "learning_rate": 2.187649949343693e-06, "loss": 1.1222, "step": 5475 }, { "epoch": 0.4666382616105667, "grad_norm": 24.0, "learning_rate": 2.187538818662914e-06, "loss": 1.0194, "step": 5476 }, { "epoch": 0.46672347677886666, "grad_norm": 14.6875, "learning_rate": 2.1874276710398657e-06, "loss": 0.8416, "step": 5477 }, { "epoch": 0.4668086919471666, "grad_norm": 15.6875, "learning_rate": 2.1873165064765566e-06, "loss": 0.5463, "step": 5478 }, { "epoch": 0.4668939071154666, "grad_norm": 21.875, "learning_rate": 2.1872053249749957e-06, "loss": 1.1472, "step": 5479 }, { "epoch": 0.46697912228376653, "grad_norm": 13.125, "learning_rate": 2.187094126537192e-06, "loss": 0.6365, "step": 5480 }, { "epoch": 0.4670643374520665, "grad_norm": 18.125, "learning_rate": 2.1869829111651553e-06, "loss": 1.2176, "step": 5481 }, { "epoch": 0.46714955262036645, "grad_norm": 11.1875, "learning_rate": 2.1868716788608952e-06, "loss": 0.5142, "step": 5482 }, { "epoch": 0.4672347677886664, "grad_norm": 17.625, "learning_rate": 2.186760429626422e-06, "loss": 1.1302, "step": 5483 }, { "epoch": 0.46731998295696636, "grad_norm": 12.375, "learning_rate": 2.1866491634637456e-06, "loss": 0.5261, "step": 5484 }, { "epoch": 0.4674051981252663, "grad_norm": 15.0, "learning_rate": 2.1865378803748764e-06, "loss": 0.9297, "step": 5485 }, { "epoch": 0.4674904132935663, "grad_norm": 13.25, "learning_rate": 2.186426580361827e-06, "loss": 0.7438, "step": 5486 }, { "epoch": 0.46757562846186623, "grad_norm": 12.8125, "learning_rate": 2.1863152634266066e-06, "loss": 0.8134, "step": 5487 }, { "epoch": 0.4676608436301662, "grad_norm": 16.875, "learning_rate": 2.1862039295712285e-06, "loss": 0.5644, "step": 5488 }, { "epoch": 0.46774605879846615, "grad_norm": 12.0625, "learning_rate": 2.186092578797704e-06, "loss": 0.5421, "step": 5489 }, { "epoch": 0.4678312739667661, "grad_norm": 14.0625, "learning_rate": 2.1859812111080446e-06, "loss": 0.45, "step": 5490 }, { "epoch": 0.46791648913506606, "grad_norm": 15.75, "learning_rate": 2.1858698265042643e-06, "loss": 0.6468, "step": 5491 }, { "epoch": 0.468001704303366, "grad_norm": 21.125, "learning_rate": 2.1857584249883746e-06, "loss": 0.8439, "step": 5492 }, { "epoch": 0.468086919471666, "grad_norm": 14.9375, "learning_rate": 2.1856470065623896e-06, "loss": 0.8426, "step": 5493 }, { "epoch": 0.46817213463996593, "grad_norm": 17.375, "learning_rate": 2.185535571228322e-06, "loss": 0.9854, "step": 5494 }, { "epoch": 0.4682573498082659, "grad_norm": 14.5625, "learning_rate": 2.1854241189881862e-06, "loss": 0.6358, "step": 5495 }, { "epoch": 0.46834256497656584, "grad_norm": 19.625, "learning_rate": 2.185312649843996e-06, "loss": 0.9284, "step": 5496 }, { "epoch": 0.4684277801448658, "grad_norm": 14.5, "learning_rate": 2.1852011637977656e-06, "loss": 0.7636, "step": 5497 }, { "epoch": 0.46851299531316576, "grad_norm": 27.375, "learning_rate": 2.1850896608515093e-06, "loss": 0.5462, "step": 5498 }, { "epoch": 0.4685982104814657, "grad_norm": 10.875, "learning_rate": 2.184978141007243e-06, "loss": 0.4294, "step": 5499 }, { "epoch": 0.4686834256497657, "grad_norm": 10.375, "learning_rate": 2.1848666042669814e-06, "loss": 0.4634, "step": 5500 }, { "epoch": 0.46876864081806563, "grad_norm": 12.625, "learning_rate": 2.18475505063274e-06, "loss": 0.5324, "step": 5501 }, { "epoch": 0.4688538559863656, "grad_norm": 14.1875, "learning_rate": 2.1846434801065347e-06, "loss": 0.3975, "step": 5502 }, { "epoch": 0.46893907115466554, "grad_norm": 15.1875, "learning_rate": 2.1845318926903823e-06, "loss": 0.7668, "step": 5503 }, { "epoch": 0.4690242863229655, "grad_norm": 14.125, "learning_rate": 2.184420288386299e-06, "loss": 0.7345, "step": 5504 }, { "epoch": 0.46910950149126546, "grad_norm": 11.125, "learning_rate": 2.184308667196301e-06, "loss": 0.4143, "step": 5505 }, { "epoch": 0.4691947166595654, "grad_norm": 13.0, "learning_rate": 2.184197029122406e-06, "loss": 0.7244, "step": 5506 }, { "epoch": 0.46927993182786537, "grad_norm": 17.25, "learning_rate": 2.1840853741666314e-06, "loss": 1.0593, "step": 5507 }, { "epoch": 0.4693651469961653, "grad_norm": 13.5625, "learning_rate": 2.183973702330994e-06, "loss": 0.8069, "step": 5508 }, { "epoch": 0.4694503621644653, "grad_norm": 12.6875, "learning_rate": 2.183862013617514e-06, "loss": 0.7345, "step": 5509 }, { "epoch": 0.46953557733276524, "grad_norm": 13.6875, "learning_rate": 2.1837503080282074e-06, "loss": 0.9162, "step": 5510 }, { "epoch": 0.4696207925010652, "grad_norm": 12.4375, "learning_rate": 2.1836385855650936e-06, "loss": 0.5146, "step": 5511 }, { "epoch": 0.46970600766936516, "grad_norm": 11.625, "learning_rate": 2.1835268462301928e-06, "loss": 0.556, "step": 5512 }, { "epoch": 0.4697912228376651, "grad_norm": 13.4375, "learning_rate": 2.183415090025522e-06, "loss": 0.5341, "step": 5513 }, { "epoch": 0.46987643800596507, "grad_norm": 11.8125, "learning_rate": 2.1833033169531027e-06, "loss": 0.6581, "step": 5514 }, { "epoch": 0.469961653174265, "grad_norm": 16.875, "learning_rate": 2.1831915270149535e-06, "loss": 0.6958, "step": 5515 }, { "epoch": 0.470046868342565, "grad_norm": 11.75, "learning_rate": 2.1830797202130953e-06, "loss": 0.54, "step": 5516 }, { "epoch": 0.47013208351086494, "grad_norm": 12.6875, "learning_rate": 2.182967896549548e-06, "loss": 0.529, "step": 5517 }, { "epoch": 0.4702172986791649, "grad_norm": 11.25, "learning_rate": 2.1828560560263327e-06, "loss": 0.3441, "step": 5518 }, { "epoch": 0.47030251384746485, "grad_norm": 13.3125, "learning_rate": 2.1827441986454708e-06, "loss": 0.6511, "step": 5519 }, { "epoch": 0.4703877290157648, "grad_norm": 14.75, "learning_rate": 2.1826323244089827e-06, "loss": 0.8134, "step": 5520 }, { "epoch": 0.47047294418406477, "grad_norm": 17.875, "learning_rate": 2.182520433318891e-06, "loss": 0.8867, "step": 5521 }, { "epoch": 0.4705581593523647, "grad_norm": 13.75, "learning_rate": 2.182408525377217e-06, "loss": 0.8701, "step": 5522 }, { "epoch": 0.4706433745206647, "grad_norm": 10.8125, "learning_rate": 2.1822966005859835e-06, "loss": 0.4297, "step": 5523 }, { "epoch": 0.47072858968896464, "grad_norm": 13.75, "learning_rate": 2.1821846589472133e-06, "loss": 0.537, "step": 5524 }, { "epoch": 0.4708138048572646, "grad_norm": 13.9375, "learning_rate": 2.1820727004629284e-06, "loss": 0.7734, "step": 5525 }, { "epoch": 0.47089902002556455, "grad_norm": 14.0, "learning_rate": 2.181960725135153e-06, "loss": 0.7162, "step": 5526 }, { "epoch": 0.4709842351938645, "grad_norm": 11.5, "learning_rate": 2.18184873296591e-06, "loss": 0.5148, "step": 5527 }, { "epoch": 0.47106945036216447, "grad_norm": 15.4375, "learning_rate": 2.181736723957223e-06, "loss": 0.6375, "step": 5528 }, { "epoch": 0.4711546655304644, "grad_norm": 16.25, "learning_rate": 2.1816246981111168e-06, "loss": 0.7674, "step": 5529 }, { "epoch": 0.4712398806987644, "grad_norm": 11.625, "learning_rate": 2.1815126554296156e-06, "loss": 0.5394, "step": 5530 }, { "epoch": 0.47132509586706434, "grad_norm": 15.4375, "learning_rate": 2.181400595914744e-06, "loss": 0.9407, "step": 5531 }, { "epoch": 0.4714103110353643, "grad_norm": 16.625, "learning_rate": 2.1812885195685265e-06, "loss": 0.8713, "step": 5532 }, { "epoch": 0.47149552620366425, "grad_norm": 19.0, "learning_rate": 2.181176426392989e-06, "loss": 0.7733, "step": 5533 }, { "epoch": 0.4715807413719642, "grad_norm": 12.6875, "learning_rate": 2.181064316390158e-06, "loss": 0.5568, "step": 5534 }, { "epoch": 0.47166595654026416, "grad_norm": 16.75, "learning_rate": 2.1809521895620577e-06, "loss": 0.853, "step": 5535 }, { "epoch": 0.4717511717085641, "grad_norm": 13.75, "learning_rate": 2.1808400459107157e-06, "loss": 0.5345, "step": 5536 }, { "epoch": 0.4718363868768641, "grad_norm": 26.5, "learning_rate": 2.180727885438158e-06, "loss": 1.1742, "step": 5537 }, { "epoch": 0.47192160204516403, "grad_norm": 9.6875, "learning_rate": 2.180615708146411e-06, "loss": 0.2815, "step": 5538 }, { "epoch": 0.472006817213464, "grad_norm": 17.125, "learning_rate": 2.1805035140375028e-06, "loss": 0.9562, "step": 5539 }, { "epoch": 0.47209203238176395, "grad_norm": 18.75, "learning_rate": 2.1803913031134603e-06, "loss": 1.208, "step": 5540 }, { "epoch": 0.4721772475500639, "grad_norm": 11.875, "learning_rate": 2.180279075376311e-06, "loss": 0.6441, "step": 5541 }, { "epoch": 0.47226246271836386, "grad_norm": 15.125, "learning_rate": 2.1801668308280836e-06, "loss": 0.7921, "step": 5542 }, { "epoch": 0.4723476778866638, "grad_norm": 12.1875, "learning_rate": 2.1800545694708063e-06, "loss": 0.2958, "step": 5543 }, { "epoch": 0.4724328930549638, "grad_norm": 14.125, "learning_rate": 2.179942291306508e-06, "loss": 0.8886, "step": 5544 }, { "epoch": 0.47251810822326373, "grad_norm": 16.125, "learning_rate": 2.179829996337217e-06, "loss": 0.7413, "step": 5545 }, { "epoch": 0.4726033233915637, "grad_norm": 14.4375, "learning_rate": 2.179717684564963e-06, "loss": 0.5756, "step": 5546 }, { "epoch": 0.47268853855986365, "grad_norm": 12.8125, "learning_rate": 2.1796053559917755e-06, "loss": 0.5664, "step": 5547 }, { "epoch": 0.4727737537281636, "grad_norm": 15.75, "learning_rate": 2.179493010619684e-06, "loss": 0.6699, "step": 5548 }, { "epoch": 0.47285896889646356, "grad_norm": 10.6875, "learning_rate": 2.1793806484507197e-06, "loss": 0.4686, "step": 5549 }, { "epoch": 0.4729441840647635, "grad_norm": 16.25, "learning_rate": 2.1792682694869127e-06, "loss": 0.7772, "step": 5550 }, { "epoch": 0.4730293992330635, "grad_norm": 13.9375, "learning_rate": 2.1791558737302933e-06, "loss": 0.7444, "step": 5551 }, { "epoch": 0.47311461440136343, "grad_norm": 15.375, "learning_rate": 2.1790434611828933e-06, "loss": 0.9241, "step": 5552 }, { "epoch": 0.4731998295696634, "grad_norm": 15.0625, "learning_rate": 2.178931031846743e-06, "loss": 0.7668, "step": 5553 }, { "epoch": 0.47328504473796335, "grad_norm": 14.625, "learning_rate": 2.1788185857238754e-06, "loss": 0.8363, "step": 5554 }, { "epoch": 0.4733702599062633, "grad_norm": 12.375, "learning_rate": 2.1787061228163216e-06, "loss": 0.3811, "step": 5555 }, { "epoch": 0.47345547507456326, "grad_norm": 14.75, "learning_rate": 2.1785936431261144e-06, "loss": 0.8861, "step": 5556 }, { "epoch": 0.4735406902428632, "grad_norm": 13.625, "learning_rate": 2.1784811466552863e-06, "loss": 0.6705, "step": 5557 }, { "epoch": 0.4736259054111632, "grad_norm": 12.25, "learning_rate": 2.17836863340587e-06, "loss": 0.3512, "step": 5558 }, { "epoch": 0.47371112057946313, "grad_norm": 12.375, "learning_rate": 2.1782561033798994e-06, "loss": 0.7501, "step": 5559 }, { "epoch": 0.4737963357477631, "grad_norm": 18.5, "learning_rate": 2.1781435565794074e-06, "loss": 1.0052, "step": 5560 }, { "epoch": 0.47388155091606304, "grad_norm": 15.4375, "learning_rate": 2.178030993006428e-06, "loss": 0.7937, "step": 5561 }, { "epoch": 0.473966766084363, "grad_norm": 12.25, "learning_rate": 2.1779184126629954e-06, "loss": 0.6188, "step": 5562 }, { "epoch": 0.47405198125266296, "grad_norm": 12.625, "learning_rate": 2.1778058155511437e-06, "loss": 0.2805, "step": 5563 }, { "epoch": 0.4741371964209629, "grad_norm": 11.5625, "learning_rate": 2.1776932016729085e-06, "loss": 0.6485, "step": 5564 }, { "epoch": 0.47422241158926287, "grad_norm": 17.625, "learning_rate": 2.177580571030324e-06, "loss": 0.9021, "step": 5565 }, { "epoch": 0.47430762675756283, "grad_norm": 11.0, "learning_rate": 2.1774679236254255e-06, "loss": 0.4493, "step": 5566 }, { "epoch": 0.4743928419258628, "grad_norm": 15.5625, "learning_rate": 2.1773552594602495e-06, "loss": 0.7791, "step": 5567 }, { "epoch": 0.47447805709416274, "grad_norm": 11.625, "learning_rate": 2.177242578536831e-06, "loss": 0.6009, "step": 5568 }, { "epoch": 0.4745632722624627, "grad_norm": 18.0, "learning_rate": 2.1771298808572065e-06, "loss": 0.8438, "step": 5569 }, { "epoch": 0.47464848743076266, "grad_norm": 13.0625, "learning_rate": 2.1770171664234134e-06, "loss": 0.5735, "step": 5570 }, { "epoch": 0.4747337025990626, "grad_norm": 14.625, "learning_rate": 2.1769044352374877e-06, "loss": 0.607, "step": 5571 }, { "epoch": 0.47481891776736257, "grad_norm": 11.75, "learning_rate": 2.1767916873014667e-06, "loss": 0.5811, "step": 5572 }, { "epoch": 0.4749041329356625, "grad_norm": 14.625, "learning_rate": 2.1766789226173877e-06, "loss": 0.7509, "step": 5573 }, { "epoch": 0.4749893481039625, "grad_norm": 9.5, "learning_rate": 2.176566141187289e-06, "loss": 0.5692, "step": 5574 }, { "epoch": 0.47507456327226244, "grad_norm": 13.625, "learning_rate": 2.1764533430132087e-06, "loss": 0.7993, "step": 5575 }, { "epoch": 0.4751597784405624, "grad_norm": 21.125, "learning_rate": 2.1763405280971846e-06, "loss": 1.1098, "step": 5576 }, { "epoch": 0.47524499360886235, "grad_norm": 14.5, "learning_rate": 2.1762276964412556e-06, "loss": 0.5298, "step": 5577 }, { "epoch": 0.4753302087771623, "grad_norm": 15.5, "learning_rate": 2.1761148480474604e-06, "loss": 0.9666, "step": 5578 }, { "epoch": 0.47541542394546227, "grad_norm": 12.5, "learning_rate": 2.176001982917839e-06, "loss": 0.4802, "step": 5579 }, { "epoch": 0.4755006391137622, "grad_norm": 17.0, "learning_rate": 2.1758891010544307e-06, "loss": 1.1795, "step": 5580 }, { "epoch": 0.4755858542820622, "grad_norm": 25.25, "learning_rate": 2.175776202459275e-06, "loss": 0.6754, "step": 5581 }, { "epoch": 0.47567106945036214, "grad_norm": 15.6875, "learning_rate": 2.175663287134413e-06, "loss": 0.8893, "step": 5582 }, { "epoch": 0.4757562846186621, "grad_norm": 20.0, "learning_rate": 2.175550355081884e-06, "loss": 0.8174, "step": 5583 }, { "epoch": 0.47584149978696205, "grad_norm": 10.75, "learning_rate": 2.1754374063037294e-06, "loss": 0.3816, "step": 5584 }, { "epoch": 0.475926714955262, "grad_norm": 12.375, "learning_rate": 2.17532444080199e-06, "loss": 0.6221, "step": 5585 }, { "epoch": 0.476011930123562, "grad_norm": 15.0625, "learning_rate": 2.175211458578708e-06, "loss": 0.7942, "step": 5586 }, { "epoch": 0.476097145291862, "grad_norm": 16.75, "learning_rate": 2.175098459635924e-06, "loss": 1.0592, "step": 5587 }, { "epoch": 0.47618236046016194, "grad_norm": 11.9375, "learning_rate": 2.1749854439756813e-06, "loss": 0.4611, "step": 5588 }, { "epoch": 0.4762675756284619, "grad_norm": 16.25, "learning_rate": 2.1748724116000213e-06, "loss": 0.7988, "step": 5589 }, { "epoch": 0.47635279079676185, "grad_norm": 13.0625, "learning_rate": 2.1747593625109863e-06, "loss": 0.5668, "step": 5590 }, { "epoch": 0.4764380059650618, "grad_norm": 11.6875, "learning_rate": 2.17464629671062e-06, "loss": 0.406, "step": 5591 }, { "epoch": 0.47652322113336176, "grad_norm": 13.5625, "learning_rate": 2.1745332142009654e-06, "loss": 0.7513, "step": 5592 }, { "epoch": 0.4766084363016617, "grad_norm": 14.5625, "learning_rate": 2.174420114984066e-06, "loss": 0.8797, "step": 5593 }, { "epoch": 0.4766936514699617, "grad_norm": 12.75, "learning_rate": 2.1743069990619653e-06, "loss": 0.5748, "step": 5594 }, { "epoch": 0.47677886663826163, "grad_norm": 17.0, "learning_rate": 2.174193866436708e-06, "loss": 1.0065, "step": 5595 }, { "epoch": 0.4768640818065616, "grad_norm": 16.25, "learning_rate": 2.174080717110338e-06, "loss": 1.0059, "step": 5596 }, { "epoch": 0.47694929697486155, "grad_norm": 15.5625, "learning_rate": 2.1739675510849002e-06, "loss": 0.7129, "step": 5597 }, { "epoch": 0.4770345121431615, "grad_norm": 15.9375, "learning_rate": 2.1738543683624396e-06, "loss": 0.8302, "step": 5598 }, { "epoch": 0.47711972731146146, "grad_norm": 13.75, "learning_rate": 2.1737411689450015e-06, "loss": 0.6182, "step": 5599 }, { "epoch": 0.4772049424797614, "grad_norm": 19.5, "learning_rate": 2.1736279528346315e-06, "loss": 0.9148, "step": 5600 }, { "epoch": 0.4772901576480614, "grad_norm": 28.0, "learning_rate": 2.1735147200333757e-06, "loss": 1.5502, "step": 5601 }, { "epoch": 0.47737537281636133, "grad_norm": 13.8125, "learning_rate": 2.1734014705432804e-06, "loss": 0.7342, "step": 5602 }, { "epoch": 0.4774605879846613, "grad_norm": 12.0, "learning_rate": 2.1732882043663917e-06, "loss": 0.7289, "step": 5603 }, { "epoch": 0.47754580315296125, "grad_norm": 15.625, "learning_rate": 2.173174921504757e-06, "loss": 0.7083, "step": 5604 }, { "epoch": 0.4776310183212612, "grad_norm": 15.5625, "learning_rate": 2.173061621960423e-06, "loss": 0.6174, "step": 5605 }, { "epoch": 0.47771623348956116, "grad_norm": 14.3125, "learning_rate": 2.1729483057354374e-06, "loss": 1.0593, "step": 5606 }, { "epoch": 0.4778014486578611, "grad_norm": 14.125, "learning_rate": 2.1728349728318478e-06, "loss": 0.6123, "step": 5607 }, { "epoch": 0.4778866638261611, "grad_norm": 15.3125, "learning_rate": 2.172721623251702e-06, "loss": 1.1667, "step": 5608 }, { "epoch": 0.47797187899446103, "grad_norm": 15.9375, "learning_rate": 2.1726082569970493e-06, "loss": 0.6684, "step": 5609 }, { "epoch": 0.478057094162761, "grad_norm": 10.8125, "learning_rate": 2.1724948740699367e-06, "loss": 0.4906, "step": 5610 }, { "epoch": 0.47814230933106094, "grad_norm": 20.5, "learning_rate": 2.172381474472415e-06, "loss": 0.9473, "step": 5611 }, { "epoch": 0.4782275244993609, "grad_norm": 22.875, "learning_rate": 2.1722680582065315e-06, "loss": 1.3324, "step": 5612 }, { "epoch": 0.47831273966766086, "grad_norm": 16.125, "learning_rate": 2.1721546252743376e-06, "loss": 1.0274, "step": 5613 }, { "epoch": 0.4783979548359608, "grad_norm": 16.25, "learning_rate": 2.172041175677882e-06, "loss": 0.8132, "step": 5614 }, { "epoch": 0.47848317000426077, "grad_norm": 20.125, "learning_rate": 2.171927709419215e-06, "loss": 1.0301, "step": 5615 }, { "epoch": 0.47856838517256073, "grad_norm": 10.25, "learning_rate": 2.1718142265003875e-06, "loss": 0.7921, "step": 5616 }, { "epoch": 0.4786536003408607, "grad_norm": 20.375, "learning_rate": 2.17170072692345e-06, "loss": 0.7704, "step": 5617 }, { "epoch": 0.47873881550916064, "grad_norm": 13.1875, "learning_rate": 2.1715872106904535e-06, "loss": 0.7501, "step": 5618 }, { "epoch": 0.4788240306774606, "grad_norm": 10.0625, "learning_rate": 2.171473677803449e-06, "loss": 0.3499, "step": 5619 }, { "epoch": 0.47890924584576056, "grad_norm": 15.1875, "learning_rate": 2.171360128264489e-06, "loss": 1.0791, "step": 5620 }, { "epoch": 0.4789944610140605, "grad_norm": 12.625, "learning_rate": 2.171246562075625e-06, "loss": 0.7571, "step": 5621 }, { "epoch": 0.47907967618236047, "grad_norm": 15.75, "learning_rate": 2.1711329792389088e-06, "loss": 0.7993, "step": 5622 }, { "epoch": 0.4791648913506604, "grad_norm": 15.3125, "learning_rate": 2.171019379756394e-06, "loss": 0.7689, "step": 5623 }, { "epoch": 0.4792501065189604, "grad_norm": 19.25, "learning_rate": 2.1709057636301325e-06, "loss": 0.9547, "step": 5624 }, { "epoch": 0.47933532168726034, "grad_norm": 15.9375, "learning_rate": 2.1707921308621775e-06, "loss": 0.9161, "step": 5625 }, { "epoch": 0.4794205368555603, "grad_norm": 17.75, "learning_rate": 2.170678481454583e-06, "loss": 1.0343, "step": 5626 }, { "epoch": 0.47950575202386025, "grad_norm": 13.625, "learning_rate": 2.1705648154094028e-06, "loss": 0.5476, "step": 5627 }, { "epoch": 0.4795909671921602, "grad_norm": 11.3125, "learning_rate": 2.17045113272869e-06, "loss": 0.3372, "step": 5628 }, { "epoch": 0.47967618236046017, "grad_norm": 15.3125, "learning_rate": 2.1703374334145006e-06, "loss": 1.1081, "step": 5629 }, { "epoch": 0.4797613975287601, "grad_norm": 14.6875, "learning_rate": 2.1702237174688877e-06, "loss": 0.5139, "step": 5630 }, { "epoch": 0.4798466126970601, "grad_norm": 12.6875, "learning_rate": 2.1701099848939073e-06, "loss": 0.4015, "step": 5631 }, { "epoch": 0.47993182786536004, "grad_norm": 13.625, "learning_rate": 2.169996235691614e-06, "loss": 0.4946, "step": 5632 }, { "epoch": 0.48001704303366, "grad_norm": 16.25, "learning_rate": 2.169882469864063e-06, "loss": 0.6849, "step": 5633 }, { "epoch": 0.48010225820195995, "grad_norm": 19.0, "learning_rate": 2.1697686874133114e-06, "loss": 0.9169, "step": 5634 }, { "epoch": 0.4801874733702599, "grad_norm": 12.1875, "learning_rate": 2.169654888341415e-06, "loss": 0.6028, "step": 5635 }, { "epoch": 0.48027268853855987, "grad_norm": 16.125, "learning_rate": 2.1695410726504295e-06, "loss": 0.7547, "step": 5636 }, { "epoch": 0.4803579037068598, "grad_norm": 17.125, "learning_rate": 2.1694272403424123e-06, "loss": 1.2083, "step": 5637 }, { "epoch": 0.4804431188751598, "grad_norm": 16.5, "learning_rate": 2.16931339141942e-06, "loss": 0.7565, "step": 5638 }, { "epoch": 0.48052833404345974, "grad_norm": 19.5, "learning_rate": 2.1691995258835106e-06, "loss": 0.8016, "step": 5639 }, { "epoch": 0.4806135492117597, "grad_norm": 14.75, "learning_rate": 2.1690856437367414e-06, "loss": 0.6593, "step": 5640 }, { "epoch": 0.48069876438005965, "grad_norm": 14.9375, "learning_rate": 2.1689717449811702e-06, "loss": 0.9464, "step": 5641 }, { "epoch": 0.4807839795483596, "grad_norm": 15.5, "learning_rate": 2.1688578296188557e-06, "loss": 1.011, "step": 5642 }, { "epoch": 0.48086919471665956, "grad_norm": 16.5, "learning_rate": 2.168743897651856e-06, "loss": 0.8007, "step": 5643 }, { "epoch": 0.4809544098849595, "grad_norm": 11.25, "learning_rate": 2.1686299490822304e-06, "loss": 0.5918, "step": 5644 }, { "epoch": 0.4810396250532595, "grad_norm": 16.5, "learning_rate": 2.1685159839120378e-06, "loss": 0.9377, "step": 5645 }, { "epoch": 0.48112484022155944, "grad_norm": 13.0625, "learning_rate": 2.168402002143338e-06, "loss": 0.624, "step": 5646 }, { "epoch": 0.4812100553898594, "grad_norm": 14.6875, "learning_rate": 2.16828800377819e-06, "loss": 0.685, "step": 5647 }, { "epoch": 0.48129527055815935, "grad_norm": 12.75, "learning_rate": 2.1681739888186547e-06, "loss": 0.4311, "step": 5648 }, { "epoch": 0.4813804857264593, "grad_norm": 17.5, "learning_rate": 2.1680599572667925e-06, "loss": 0.7615, "step": 5649 }, { "epoch": 0.48146570089475926, "grad_norm": 14.625, "learning_rate": 2.167945909124663e-06, "loss": 0.9038, "step": 5650 }, { "epoch": 0.4815509160630592, "grad_norm": 14.25, "learning_rate": 2.167831844394328e-06, "loss": 0.5393, "step": 5651 }, { "epoch": 0.4816361312313592, "grad_norm": 12.0625, "learning_rate": 2.167717763077849e-06, "loss": 0.3941, "step": 5652 }, { "epoch": 0.48172134639965913, "grad_norm": 16.5, "learning_rate": 2.167603665177287e-06, "loss": 1.1596, "step": 5653 }, { "epoch": 0.4818065615679591, "grad_norm": 14.3125, "learning_rate": 2.1674895506947034e-06, "loss": 0.6373, "step": 5654 }, { "epoch": 0.48189177673625905, "grad_norm": 13.75, "learning_rate": 2.167375419632162e-06, "loss": 0.4344, "step": 5655 }, { "epoch": 0.481976991904559, "grad_norm": 20.25, "learning_rate": 2.1672612719917237e-06, "loss": 1.0426, "step": 5656 }, { "epoch": 0.48206220707285896, "grad_norm": 20.5, "learning_rate": 2.1671471077754517e-06, "loss": 0.9335, "step": 5657 }, { "epoch": 0.4821474222411589, "grad_norm": 14.0, "learning_rate": 2.1670329269854097e-06, "loss": 0.8758, "step": 5658 }, { "epoch": 0.4822326374094589, "grad_norm": 15.6875, "learning_rate": 2.1669187296236606e-06, "loss": 0.8131, "step": 5659 }, { "epoch": 0.48231785257775883, "grad_norm": 11.625, "learning_rate": 2.166804515692268e-06, "loss": 0.5703, "step": 5660 }, { "epoch": 0.4824030677460588, "grad_norm": 17.5, "learning_rate": 2.166690285193295e-06, "loss": 0.6691, "step": 5661 }, { "epoch": 0.48248828291435875, "grad_norm": 13.75, "learning_rate": 2.1665760381288077e-06, "loss": 0.785, "step": 5662 }, { "epoch": 0.4825734980826587, "grad_norm": 17.875, "learning_rate": 2.1664617745008696e-06, "loss": 0.3756, "step": 5663 }, { "epoch": 0.48265871325095866, "grad_norm": 37.75, "learning_rate": 2.1663474943115455e-06, "loss": 1.4977, "step": 5664 }, { "epoch": 0.4827439284192586, "grad_norm": 13.0, "learning_rate": 2.1662331975629006e-06, "loss": 0.8138, "step": 5665 }, { "epoch": 0.4828291435875586, "grad_norm": 13.6875, "learning_rate": 2.1661188842570006e-06, "loss": 0.8917, "step": 5666 }, { "epoch": 0.48291435875585853, "grad_norm": 13.0625, "learning_rate": 2.166004554395911e-06, "loss": 0.6281, "step": 5667 }, { "epoch": 0.4829995739241585, "grad_norm": 12.9375, "learning_rate": 2.165890207981698e-06, "loss": 0.4999, "step": 5668 }, { "epoch": 0.48308478909245844, "grad_norm": 15.875, "learning_rate": 2.165775845016428e-06, "loss": 0.7781, "step": 5669 }, { "epoch": 0.4831700042607584, "grad_norm": 16.0, "learning_rate": 2.1656614655021675e-06, "loss": 0.9867, "step": 5670 }, { "epoch": 0.48325521942905836, "grad_norm": 17.0, "learning_rate": 2.165547069440984e-06, "loss": 0.6159, "step": 5671 }, { "epoch": 0.4833404345973583, "grad_norm": 11.4375, "learning_rate": 2.1654326568349436e-06, "loss": 0.5578, "step": 5672 }, { "epoch": 0.48342564976565827, "grad_norm": 11.5, "learning_rate": 2.1653182276861153e-06, "loss": 0.555, "step": 5673 }, { "epoch": 0.48351086493395823, "grad_norm": 23.5, "learning_rate": 2.1652037819965657e-06, "loss": 0.9754, "step": 5674 }, { "epoch": 0.4835960801022582, "grad_norm": 32.5, "learning_rate": 2.1650893197683636e-06, "loss": 1.3365, "step": 5675 }, { "epoch": 0.48368129527055814, "grad_norm": 15.8125, "learning_rate": 2.164974841003577e-06, "loss": 0.6482, "step": 5676 }, { "epoch": 0.4837665104388581, "grad_norm": 19.125, "learning_rate": 2.1648603457042756e-06, "loss": 1.0686, "step": 5677 }, { "epoch": 0.48385172560715806, "grad_norm": 11.4375, "learning_rate": 2.1647458338725273e-06, "loss": 0.3889, "step": 5678 }, { "epoch": 0.483936940775458, "grad_norm": 17.125, "learning_rate": 2.164631305510402e-06, "loss": 0.5939, "step": 5679 }, { "epoch": 0.48402215594375797, "grad_norm": 21.375, "learning_rate": 2.1645167606199693e-06, "loss": 0.9593, "step": 5680 }, { "epoch": 0.4841073711120579, "grad_norm": 36.0, "learning_rate": 2.1644021992032994e-06, "loss": 1.0025, "step": 5681 }, { "epoch": 0.4841925862803579, "grad_norm": 20.75, "learning_rate": 2.164287621262462e-06, "loss": 0.9502, "step": 5682 }, { "epoch": 0.48427780144865784, "grad_norm": 13.1875, "learning_rate": 2.1641730267995277e-06, "loss": 0.6198, "step": 5683 }, { "epoch": 0.4843630166169578, "grad_norm": 10.625, "learning_rate": 2.164058415816568e-06, "loss": 0.4153, "step": 5684 }, { "epoch": 0.48444823178525775, "grad_norm": 14.0, "learning_rate": 2.163943788315653e-06, "loss": 0.3983, "step": 5685 }, { "epoch": 0.4845334469535577, "grad_norm": 21.375, "learning_rate": 2.163829144298855e-06, "loss": 1.0344, "step": 5686 }, { "epoch": 0.48461866212185767, "grad_norm": 11.5, "learning_rate": 2.1637144837682456e-06, "loss": 0.5683, "step": 5687 }, { "epoch": 0.4847038772901576, "grad_norm": 17.125, "learning_rate": 2.163599806725896e-06, "loss": 0.879, "step": 5688 }, { "epoch": 0.4847890924584576, "grad_norm": 13.4375, "learning_rate": 2.16348511317388e-06, "loss": 0.3956, "step": 5689 }, { "epoch": 0.48487430762675754, "grad_norm": 17.375, "learning_rate": 2.163370403114269e-06, "loss": 0.7775, "step": 5690 }, { "epoch": 0.4849595227950575, "grad_norm": 15.5, "learning_rate": 2.163255676549137e-06, "loss": 0.6783, "step": 5691 }, { "epoch": 0.48504473796335745, "grad_norm": 15.75, "learning_rate": 2.163140933480556e-06, "loss": 0.6896, "step": 5692 }, { "epoch": 0.4851299531316574, "grad_norm": 25.125, "learning_rate": 2.1630261739105996e-06, "loss": 1.416, "step": 5693 }, { "epoch": 0.48521516829995737, "grad_norm": 14.0625, "learning_rate": 2.162911397841343e-06, "loss": 0.4953, "step": 5694 }, { "epoch": 0.4853003834682573, "grad_norm": 19.375, "learning_rate": 2.162796605274859e-06, "loss": 1.0947, "step": 5695 }, { "epoch": 0.4853855986365573, "grad_norm": 15.8125, "learning_rate": 2.162681796213223e-06, "loss": 0.8249, "step": 5696 }, { "epoch": 0.48547081380485724, "grad_norm": 15.125, "learning_rate": 2.1625669706585085e-06, "loss": 0.8246, "step": 5697 }, { "epoch": 0.4855560289731572, "grad_norm": 21.375, "learning_rate": 2.1624521286127915e-06, "loss": 1.2245, "step": 5698 }, { "epoch": 0.4856412441414572, "grad_norm": 12.5625, "learning_rate": 2.1623372700781463e-06, "loss": 0.5612, "step": 5699 }, { "epoch": 0.48572645930975716, "grad_norm": 20.5, "learning_rate": 2.1622223950566503e-06, "loss": 1.0498, "step": 5700 }, { "epoch": 0.4858116744780571, "grad_norm": 13.5, "learning_rate": 2.1621075035503777e-06, "loss": 0.6437, "step": 5701 }, { "epoch": 0.4858968896463571, "grad_norm": 18.0, "learning_rate": 2.161992595561405e-06, "loss": 0.6918, "step": 5702 }, { "epoch": 0.48598210481465703, "grad_norm": 16.375, "learning_rate": 2.16187767109181e-06, "loss": 0.8423, "step": 5703 }, { "epoch": 0.486067319982957, "grad_norm": 16.75, "learning_rate": 2.1617627301436677e-06, "loss": 0.8333, "step": 5704 }, { "epoch": 0.48615253515125695, "grad_norm": 11.8125, "learning_rate": 2.1616477727190563e-06, "loss": 0.69, "step": 5705 }, { "epoch": 0.4862377503195569, "grad_norm": 12.1875, "learning_rate": 2.161532798820053e-06, "loss": 0.5914, "step": 5706 }, { "epoch": 0.48632296548785686, "grad_norm": 13.5625, "learning_rate": 2.161417808448735e-06, "loss": 0.7777, "step": 5707 }, { "epoch": 0.4864081806561568, "grad_norm": 15.125, "learning_rate": 2.1613028016071814e-06, "loss": 0.7, "step": 5708 }, { "epoch": 0.4864933958244568, "grad_norm": 13.8125, "learning_rate": 2.1611877782974696e-06, "loss": 0.7747, "step": 5709 }, { "epoch": 0.48657861099275673, "grad_norm": 14.6875, "learning_rate": 2.161072738521678e-06, "loss": 0.9512, "step": 5710 }, { "epoch": 0.4866638261610567, "grad_norm": 13.6875, "learning_rate": 2.160957682281886e-06, "loss": 0.7135, "step": 5711 }, { "epoch": 0.48674904132935665, "grad_norm": 9.4375, "learning_rate": 2.1608426095801727e-06, "loss": 0.376, "step": 5712 }, { "epoch": 0.4868342564976566, "grad_norm": 14.1875, "learning_rate": 2.1607275204186174e-06, "loss": 0.7216, "step": 5713 }, { "epoch": 0.48691947166595656, "grad_norm": 12.75, "learning_rate": 2.1606124147993e-06, "loss": 0.4498, "step": 5714 }, { "epoch": 0.4870046868342565, "grad_norm": 11.875, "learning_rate": 2.160497292724301e-06, "loss": 0.609, "step": 5715 }, { "epoch": 0.4870899020025565, "grad_norm": 15.6875, "learning_rate": 2.1603821541956997e-06, "loss": 0.5074, "step": 5716 }, { "epoch": 0.48717511717085643, "grad_norm": 14.5625, "learning_rate": 2.1602669992155776e-06, "loss": 0.6741, "step": 5717 }, { "epoch": 0.4872603323391564, "grad_norm": 16.0, "learning_rate": 2.160151827786016e-06, "loss": 0.9134, "step": 5718 }, { "epoch": 0.48734554750745634, "grad_norm": 17.75, "learning_rate": 2.160036639909095e-06, "loss": 1.1331, "step": 5719 }, { "epoch": 0.4874307626757563, "grad_norm": 21.25, "learning_rate": 2.159921435586897e-06, "loss": 0.7145, "step": 5720 }, { "epoch": 0.48751597784405626, "grad_norm": 14.875, "learning_rate": 2.1598062148215037e-06, "loss": 0.9386, "step": 5721 }, { "epoch": 0.4876011930123562, "grad_norm": 25.375, "learning_rate": 2.159690977614997e-06, "loss": 1.2735, "step": 5722 }, { "epoch": 0.4876864081806562, "grad_norm": 12.125, "learning_rate": 2.1595757239694603e-06, "loss": 0.659, "step": 5723 }, { "epoch": 0.48777162334895613, "grad_norm": 15.75, "learning_rate": 2.159460453886975e-06, "loss": 0.748, "step": 5724 }, { "epoch": 0.4878568385172561, "grad_norm": 16.625, "learning_rate": 2.159345167369625e-06, "loss": 0.9094, "step": 5725 }, { "epoch": 0.48794205368555604, "grad_norm": 14.75, "learning_rate": 2.159229864419493e-06, "loss": 0.8414, "step": 5726 }, { "epoch": 0.488027268853856, "grad_norm": 14.375, "learning_rate": 2.1591145450386636e-06, "loss": 0.2189, "step": 5727 }, { "epoch": 0.48811248402215596, "grad_norm": 10.6875, "learning_rate": 2.15899920922922e-06, "loss": 0.4003, "step": 5728 }, { "epoch": 0.4881976991904559, "grad_norm": 12.3125, "learning_rate": 2.1588838569932466e-06, "loss": 0.5217, "step": 5729 }, { "epoch": 0.48828291435875587, "grad_norm": 16.5, "learning_rate": 2.158768488332828e-06, "loss": 0.9054, "step": 5730 }, { "epoch": 0.4883681295270558, "grad_norm": 12.375, "learning_rate": 2.1586531032500488e-06, "loss": 0.536, "step": 5731 }, { "epoch": 0.4884533446953558, "grad_norm": 13.6875, "learning_rate": 2.1585377017469946e-06, "loss": 0.7898, "step": 5732 }, { "epoch": 0.48853855986365574, "grad_norm": 13.625, "learning_rate": 2.15842228382575e-06, "loss": 0.4113, "step": 5733 }, { "epoch": 0.4886237750319557, "grad_norm": 19.625, "learning_rate": 2.1583068494884016e-06, "loss": 0.8966, "step": 5734 }, { "epoch": 0.48870899020025566, "grad_norm": 17.5, "learning_rate": 2.158191398737035e-06, "loss": 1.003, "step": 5735 }, { "epoch": 0.4887942053685556, "grad_norm": 14.3125, "learning_rate": 2.1580759315737367e-06, "loss": 0.9048, "step": 5736 }, { "epoch": 0.48887942053685557, "grad_norm": 12.875, "learning_rate": 2.157960448000593e-06, "loss": 0.5135, "step": 5737 }, { "epoch": 0.4889646357051555, "grad_norm": 14.9375, "learning_rate": 2.1578449480196913e-06, "loss": 0.7703, "step": 5738 }, { "epoch": 0.4890498508734555, "grad_norm": 21.375, "learning_rate": 2.157729431633118e-06, "loss": 0.7969, "step": 5739 }, { "epoch": 0.48913506604175544, "grad_norm": 17.875, "learning_rate": 2.1576138988429614e-06, "loss": 1.0738, "step": 5740 }, { "epoch": 0.4892202812100554, "grad_norm": 16.0, "learning_rate": 2.1574983496513086e-06, "loss": 0.8701, "step": 5741 }, { "epoch": 0.48930549637835535, "grad_norm": 13.6875, "learning_rate": 2.1573827840602483e-06, "loss": 0.8098, "step": 5742 }, { "epoch": 0.4893907115466553, "grad_norm": 17.625, "learning_rate": 2.157267202071869e-06, "loss": 1.2905, "step": 5743 }, { "epoch": 0.48947592671495527, "grad_norm": 15.875, "learning_rate": 2.1571516036882585e-06, "loss": 0.9116, "step": 5744 }, { "epoch": 0.4895611418832552, "grad_norm": 13.1875, "learning_rate": 2.1570359889115065e-06, "loss": 0.7612, "step": 5745 }, { "epoch": 0.4896463570515552, "grad_norm": 14.375, "learning_rate": 2.156920357743702e-06, "loss": 0.5782, "step": 5746 }, { "epoch": 0.48973157221985514, "grad_norm": 14.0, "learning_rate": 2.156804710186935e-06, "loss": 0.5309, "step": 5747 }, { "epoch": 0.4898167873881551, "grad_norm": 15.5625, "learning_rate": 2.1566890462432944e-06, "loss": 0.6784, "step": 5748 }, { "epoch": 0.48990200255645505, "grad_norm": 13.0, "learning_rate": 2.156573365914872e-06, "loss": 0.6522, "step": 5749 }, { "epoch": 0.489987217724755, "grad_norm": 14.625, "learning_rate": 2.1564576692037564e-06, "loss": 0.9226, "step": 5750 }, { "epoch": 0.49007243289305497, "grad_norm": 16.25, "learning_rate": 2.1563419561120393e-06, "loss": 0.6082, "step": 5751 }, { "epoch": 0.4901576480613549, "grad_norm": 11.8125, "learning_rate": 2.1562262266418115e-06, "loss": 0.492, "step": 5752 }, { "epoch": 0.4902428632296549, "grad_norm": 12.875, "learning_rate": 2.1561104807951648e-06, "loss": 0.6654, "step": 5753 }, { "epoch": 0.49032807839795484, "grad_norm": 13.875, "learning_rate": 2.155994718574191e-06, "loss": 0.6855, "step": 5754 }, { "epoch": 0.4904132935662548, "grad_norm": 24.125, "learning_rate": 2.155878939980981e-06, "loss": 0.7849, "step": 5755 }, { "epoch": 0.49049850873455475, "grad_norm": 15.0, "learning_rate": 2.1557631450176274e-06, "loss": 0.831, "step": 5756 }, { "epoch": 0.4905837239028547, "grad_norm": 13.75, "learning_rate": 2.1556473336862236e-06, "loss": 0.4793, "step": 5757 }, { "epoch": 0.49066893907115466, "grad_norm": 17.25, "learning_rate": 2.1555315059888616e-06, "loss": 0.9881, "step": 5758 }, { "epoch": 0.4907541542394546, "grad_norm": 14.125, "learning_rate": 2.1554156619276346e-06, "loss": 0.8504, "step": 5759 }, { "epoch": 0.4908393694077546, "grad_norm": 15.4375, "learning_rate": 2.155299801504636e-06, "loss": 0.8327, "step": 5760 }, { "epoch": 0.49092458457605453, "grad_norm": 14.0625, "learning_rate": 2.15518392472196e-06, "loss": 0.8131, "step": 5761 }, { "epoch": 0.4910097997443545, "grad_norm": 11.25, "learning_rate": 2.1550680315816996e-06, "loss": 0.6647, "step": 5762 }, { "epoch": 0.49109501491265445, "grad_norm": 26.125, "learning_rate": 2.15495212208595e-06, "loss": 0.7175, "step": 5763 }, { "epoch": 0.4911802300809544, "grad_norm": 15.9375, "learning_rate": 2.154836196236806e-06, "loss": 0.6945, "step": 5764 }, { "epoch": 0.49126544524925436, "grad_norm": 15.375, "learning_rate": 2.1547202540363617e-06, "loss": 0.7726, "step": 5765 }, { "epoch": 0.4913506604175543, "grad_norm": 16.875, "learning_rate": 2.1546042954867124e-06, "loss": 0.9963, "step": 5766 }, { "epoch": 0.4914358755858543, "grad_norm": 15.5625, "learning_rate": 2.154488320589954e-06, "loss": 0.9346, "step": 5767 }, { "epoch": 0.49152109075415423, "grad_norm": 19.125, "learning_rate": 2.1543723293481826e-06, "loss": 1.0091, "step": 5768 }, { "epoch": 0.4916063059224542, "grad_norm": 17.875, "learning_rate": 2.1542563217634936e-06, "loss": 0.541, "step": 5769 }, { "epoch": 0.49169152109075415, "grad_norm": 14.5625, "learning_rate": 2.1541402978379835e-06, "loss": 0.8654, "step": 5770 }, { "epoch": 0.4917767362590541, "grad_norm": 24.625, "learning_rate": 2.1540242575737486e-06, "loss": 1.3225, "step": 5771 }, { "epoch": 0.49186195142735406, "grad_norm": 10.4375, "learning_rate": 2.1539082009728863e-06, "loss": 0.3247, "step": 5772 }, { "epoch": 0.491947166595654, "grad_norm": 18.75, "learning_rate": 2.1537921280374946e-06, "loss": 0.6543, "step": 5773 }, { "epoch": 0.492032381763954, "grad_norm": 20.875, "learning_rate": 2.15367603876967e-06, "loss": 1.0828, "step": 5774 }, { "epoch": 0.49211759693225393, "grad_norm": 14.5, "learning_rate": 2.15355993317151e-06, "loss": 0.9117, "step": 5775 }, { "epoch": 0.4922028121005539, "grad_norm": 13.3125, "learning_rate": 2.153443811245114e-06, "loss": 0.7582, "step": 5776 }, { "epoch": 0.49228802726885384, "grad_norm": 14.875, "learning_rate": 2.15332767299258e-06, "loss": 0.9515, "step": 5777 }, { "epoch": 0.4923732424371538, "grad_norm": 18.25, "learning_rate": 2.1532115184160063e-06, "loss": 0.9592, "step": 5778 }, { "epoch": 0.49245845760545376, "grad_norm": 24.25, "learning_rate": 2.1530953475174926e-06, "loss": 0.9465, "step": 5779 }, { "epoch": 0.4925436727737537, "grad_norm": 12.5, "learning_rate": 2.1529791602991375e-06, "loss": 0.5654, "step": 5780 }, { "epoch": 0.4926288879420537, "grad_norm": 12.0625, "learning_rate": 2.152862956763041e-06, "loss": 0.5259, "step": 5781 }, { "epoch": 0.49271410311035363, "grad_norm": 10.0625, "learning_rate": 2.152746736911303e-06, "loss": 0.5468, "step": 5782 }, { "epoch": 0.4927993182786536, "grad_norm": 21.375, "learning_rate": 2.1526305007460235e-06, "loss": 0.8994, "step": 5783 }, { "epoch": 0.49288453344695354, "grad_norm": 16.875, "learning_rate": 2.1525142482693033e-06, "loss": 0.8481, "step": 5784 }, { "epoch": 0.4929697486152535, "grad_norm": 14.3125, "learning_rate": 2.152397979483243e-06, "loss": 0.839, "step": 5785 }, { "epoch": 0.49305496378355346, "grad_norm": 14.8125, "learning_rate": 2.1522816943899443e-06, "loss": 0.7765, "step": 5786 }, { "epoch": 0.4931401789518534, "grad_norm": 14.125, "learning_rate": 2.1521653929915077e-06, "loss": 0.757, "step": 5787 }, { "epoch": 0.49322539412015337, "grad_norm": 11.6875, "learning_rate": 2.152049075290035e-06, "loss": 0.4022, "step": 5788 }, { "epoch": 0.49331060928845333, "grad_norm": 11.75, "learning_rate": 2.151932741287629e-06, "loss": 0.473, "step": 5789 }, { "epoch": 0.4933958244567533, "grad_norm": 25.0, "learning_rate": 2.151816390986391e-06, "loss": 0.8495, "step": 5790 }, { "epoch": 0.49348103962505324, "grad_norm": 10.125, "learning_rate": 2.151700024388424e-06, "loss": 0.4126, "step": 5791 }, { "epoch": 0.4935662547933532, "grad_norm": 17.75, "learning_rate": 2.1515836414958308e-06, "loss": 0.6038, "step": 5792 }, { "epoch": 0.49365146996165316, "grad_norm": 25.125, "learning_rate": 2.1514672423107146e-06, "loss": 1.2276, "step": 5793 }, { "epoch": 0.4937366851299531, "grad_norm": 13.1875, "learning_rate": 2.1513508268351787e-06, "loss": 0.656, "step": 5794 }, { "epoch": 0.49382190029825307, "grad_norm": 13.25, "learning_rate": 2.151234395071327e-06, "loss": 0.426, "step": 5795 }, { "epoch": 0.493907115466553, "grad_norm": 10.5625, "learning_rate": 2.151117947021264e-06, "loss": 0.3618, "step": 5796 }, { "epoch": 0.493992330634853, "grad_norm": 15.25, "learning_rate": 2.151001482687093e-06, "loss": 0.7949, "step": 5797 }, { "epoch": 0.49407754580315294, "grad_norm": 33.5, "learning_rate": 2.1508850020709196e-06, "loss": 0.962, "step": 5798 }, { "epoch": 0.4941627609714529, "grad_norm": 14.125, "learning_rate": 2.150768505174848e-06, "loss": 1.0273, "step": 5799 }, { "epoch": 0.49424797613975285, "grad_norm": 16.0, "learning_rate": 2.1506519920009838e-06, "loss": 0.6742, "step": 5800 }, { "epoch": 0.4943331913080528, "grad_norm": 17.75, "learning_rate": 2.1505354625514325e-06, "loss": 0.7906, "step": 5801 }, { "epoch": 0.49441840647635277, "grad_norm": 13.125, "learning_rate": 2.1504189168283e-06, "loss": 0.749, "step": 5802 }, { "epoch": 0.4945036216446527, "grad_norm": 12.4375, "learning_rate": 2.150302354833692e-06, "loss": 0.6195, "step": 5803 }, { "epoch": 0.4945888368129527, "grad_norm": 17.0, "learning_rate": 2.150185776569715e-06, "loss": 0.7822, "step": 5804 }, { "epoch": 0.49467405198125264, "grad_norm": 14.125, "learning_rate": 2.1500691820384762e-06, "loss": 0.7311, "step": 5805 }, { "epoch": 0.4947592671495526, "grad_norm": 14.0, "learning_rate": 2.1499525712420818e-06, "loss": 0.6468, "step": 5806 }, { "epoch": 0.49484448231785255, "grad_norm": 11.625, "learning_rate": 2.1498359441826395e-06, "loss": 0.473, "step": 5807 }, { "epoch": 0.4949296974861525, "grad_norm": 19.625, "learning_rate": 2.149719300862257e-06, "loss": 0.763, "step": 5808 }, { "epoch": 0.49501491265445247, "grad_norm": 20.25, "learning_rate": 2.149602641283042e-06, "loss": 0.6283, "step": 5809 }, { "epoch": 0.4951001278227524, "grad_norm": 15.6875, "learning_rate": 2.1494859654471027e-06, "loss": 0.8069, "step": 5810 }, { "epoch": 0.4951853429910524, "grad_norm": 13.625, "learning_rate": 2.1493692733565474e-06, "loss": 0.7092, "step": 5811 }, { "epoch": 0.4952705581593524, "grad_norm": 13.25, "learning_rate": 2.149252565013485e-06, "loss": 0.8032, "step": 5812 }, { "epoch": 0.49535577332765235, "grad_norm": 23.25, "learning_rate": 2.1491358404200245e-06, "loss": 1.2107, "step": 5813 }, { "epoch": 0.4954409884959523, "grad_norm": 15.875, "learning_rate": 2.149019099578275e-06, "loss": 0.7964, "step": 5814 }, { "epoch": 0.49552620366425226, "grad_norm": 13.25, "learning_rate": 2.1489023424903466e-06, "loss": 0.7263, "step": 5815 }, { "epoch": 0.4956114188325522, "grad_norm": 17.375, "learning_rate": 2.1487855691583487e-06, "loss": 0.859, "step": 5816 }, { "epoch": 0.4956966340008522, "grad_norm": 13.8125, "learning_rate": 2.148668779584392e-06, "loss": 0.871, "step": 5817 }, { "epoch": 0.49578184916915213, "grad_norm": 25.0, "learning_rate": 2.1485519737705868e-06, "loss": 0.8851, "step": 5818 }, { "epoch": 0.4958670643374521, "grad_norm": 15.6875, "learning_rate": 2.1484351517190435e-06, "loss": 0.7336, "step": 5819 }, { "epoch": 0.49595227950575205, "grad_norm": 14.9375, "learning_rate": 2.148318313431874e-06, "loss": 0.6069, "step": 5820 }, { "epoch": 0.496037494674052, "grad_norm": 21.25, "learning_rate": 2.148201458911189e-06, "loss": 0.9736, "step": 5821 }, { "epoch": 0.49612270984235196, "grad_norm": 20.125, "learning_rate": 2.1480845881591e-06, "loss": 1.0662, "step": 5822 }, { "epoch": 0.4962079250106519, "grad_norm": 13.875, "learning_rate": 2.14796770117772e-06, "loss": 0.8106, "step": 5823 }, { "epoch": 0.4962931401789519, "grad_norm": 14.3125, "learning_rate": 2.147850797969161e-06, "loss": 0.4575, "step": 5824 }, { "epoch": 0.49637835534725183, "grad_norm": 15.1875, "learning_rate": 2.1477338785355345e-06, "loss": 0.6992, "step": 5825 }, { "epoch": 0.4964635705155518, "grad_norm": 12.8125, "learning_rate": 2.1476169428789545e-06, "loss": 0.645, "step": 5826 }, { "epoch": 0.49654878568385175, "grad_norm": 12.6875, "learning_rate": 2.1474999910015334e-06, "loss": 0.823, "step": 5827 }, { "epoch": 0.4966340008521517, "grad_norm": 12.5, "learning_rate": 2.147383022905385e-06, "loss": 0.5045, "step": 5828 }, { "epoch": 0.49671921602045166, "grad_norm": 21.875, "learning_rate": 2.1472660385926232e-06, "loss": 1.1154, "step": 5829 }, { "epoch": 0.4968044311887516, "grad_norm": 17.25, "learning_rate": 2.1471490380653617e-06, "loss": 0.6726, "step": 5830 }, { "epoch": 0.4968896463570516, "grad_norm": 13.875, "learning_rate": 2.147032021325715e-06, "loss": 0.6526, "step": 5831 }, { "epoch": 0.49697486152535153, "grad_norm": 14.625, "learning_rate": 2.146914988375798e-06, "loss": 0.7688, "step": 5832 }, { "epoch": 0.4970600766936515, "grad_norm": 15.9375, "learning_rate": 2.1467979392177245e-06, "loss": 0.9518, "step": 5833 }, { "epoch": 0.49714529186195144, "grad_norm": 12.1875, "learning_rate": 2.1466808738536107e-06, "loss": 0.9798, "step": 5834 }, { "epoch": 0.4972305070302514, "grad_norm": 11.6875, "learning_rate": 2.146563792285572e-06, "loss": 0.4721, "step": 5835 }, { "epoch": 0.49731572219855136, "grad_norm": 12.25, "learning_rate": 2.1464466945157237e-06, "loss": 0.6144, "step": 5836 }, { "epoch": 0.4974009373668513, "grad_norm": 12.75, "learning_rate": 2.1463295805461828e-06, "loss": 0.7683, "step": 5837 }, { "epoch": 0.49748615253515127, "grad_norm": 14.75, "learning_rate": 2.1462124503790645e-06, "loss": 0.8325, "step": 5838 }, { "epoch": 0.49757136770345123, "grad_norm": 15.4375, "learning_rate": 2.146095304016486e-06, "loss": 0.8617, "step": 5839 }, { "epoch": 0.4976565828717512, "grad_norm": 14.8125, "learning_rate": 2.1459781414605645e-06, "loss": 0.9438, "step": 5840 }, { "epoch": 0.49774179804005114, "grad_norm": 13.8125, "learning_rate": 2.1458609627134166e-06, "loss": 0.8293, "step": 5841 }, { "epoch": 0.4978270132083511, "grad_norm": 13.5, "learning_rate": 2.1457437677771605e-06, "loss": 0.6026, "step": 5842 }, { "epoch": 0.49791222837665106, "grad_norm": 9.875, "learning_rate": 2.1456265566539137e-06, "loss": 0.3548, "step": 5843 }, { "epoch": 0.497997443544951, "grad_norm": 13.3125, "learning_rate": 2.1455093293457947e-06, "loss": 0.707, "step": 5844 }, { "epoch": 0.49808265871325097, "grad_norm": 15.5625, "learning_rate": 2.1453920858549216e-06, "loss": 0.953, "step": 5845 }, { "epoch": 0.4981678738815509, "grad_norm": 16.125, "learning_rate": 2.1452748261834127e-06, "loss": 0.8664, "step": 5846 }, { "epoch": 0.4982530890498509, "grad_norm": 14.3125, "learning_rate": 2.145157550333388e-06, "loss": 0.6303, "step": 5847 }, { "epoch": 0.49833830421815084, "grad_norm": 23.75, "learning_rate": 2.1450402583069655e-06, "loss": 0.5487, "step": 5848 }, { "epoch": 0.4984235193864508, "grad_norm": 17.5, "learning_rate": 2.1449229501062662e-06, "loss": 0.6514, "step": 5849 }, { "epoch": 0.49850873455475075, "grad_norm": 10.4375, "learning_rate": 2.1448056257334087e-06, "loss": 0.3962, "step": 5850 }, { "epoch": 0.4985939497230507, "grad_norm": 14.9375, "learning_rate": 2.1446882851905144e-06, "loss": 0.5081, "step": 5851 }, { "epoch": 0.49867916489135067, "grad_norm": 14.5, "learning_rate": 2.144570928479703e-06, "loss": 0.696, "step": 5852 }, { "epoch": 0.4987643800596506, "grad_norm": 15.6875, "learning_rate": 2.144453555603095e-06, "loss": 0.6514, "step": 5853 }, { "epoch": 0.4988495952279506, "grad_norm": 15.3125, "learning_rate": 2.1443361665628115e-06, "loss": 0.7258, "step": 5854 }, { "epoch": 0.49893481039625054, "grad_norm": 14.5625, "learning_rate": 2.144218761360975e-06, "loss": 0.9006, "step": 5855 }, { "epoch": 0.4990200255645505, "grad_norm": 12.9375, "learning_rate": 2.144101339999706e-06, "loss": 0.669, "step": 5856 }, { "epoch": 0.49910524073285045, "grad_norm": 15.25, "learning_rate": 2.143983902481126e-06, "loss": 0.7959, "step": 5857 }, { "epoch": 0.4991904559011504, "grad_norm": 12.625, "learning_rate": 2.1438664488073585e-06, "loss": 0.8162, "step": 5858 }, { "epoch": 0.49927567106945037, "grad_norm": 14.75, "learning_rate": 2.1437489789805257e-06, "loss": 0.6857, "step": 5859 }, { "epoch": 0.4993608862377503, "grad_norm": 11.0, "learning_rate": 2.1436314930027494e-06, "loss": 0.4244, "step": 5860 }, { "epoch": 0.4994461014060503, "grad_norm": 14.0, "learning_rate": 2.1435139908761535e-06, "loss": 0.5832, "step": 5861 }, { "epoch": 0.49953131657435024, "grad_norm": 13.125, "learning_rate": 2.1433964726028617e-06, "loss": 0.5866, "step": 5862 }, { "epoch": 0.4996165317426502, "grad_norm": 18.625, "learning_rate": 2.1432789381849974e-06, "loss": 0.9645, "step": 5863 }, { "epoch": 0.49970174691095015, "grad_norm": 25.875, "learning_rate": 2.143161387624684e-06, "loss": 0.6659, "step": 5864 }, { "epoch": 0.4997869620792501, "grad_norm": 13.125, "learning_rate": 2.1430438209240467e-06, "loss": 0.7143, "step": 5865 }, { "epoch": 0.49987217724755006, "grad_norm": 14.3125, "learning_rate": 2.142926238085209e-06, "loss": 0.5149, "step": 5866 }, { "epoch": 0.49995739241585, "grad_norm": 15.1875, "learning_rate": 2.1428086391102965e-06, "loss": 0.8181, "step": 5867 }, { "epoch": 0.50004260758415, "grad_norm": 11.25, "learning_rate": 2.1426910240014335e-06, "loss": 0.2639, "step": 5868 }, { "epoch": 0.5001278227524499, "grad_norm": 10.625, "learning_rate": 2.1425733927607466e-06, "loss": 0.3995, "step": 5869 }, { "epoch": 0.5002130379207499, "grad_norm": 15.5, "learning_rate": 2.142455745390361e-06, "loss": 0.8227, "step": 5870 }, { "epoch": 0.5002982530890498, "grad_norm": 12.0625, "learning_rate": 2.1423380818924028e-06, "loss": 0.7189, "step": 5871 }, { "epoch": 0.5003834682573498, "grad_norm": 29.5, "learning_rate": 2.142220402268998e-06, "loss": 0.8472, "step": 5872 }, { "epoch": 0.5004686834256498, "grad_norm": 11.5625, "learning_rate": 2.142102706522273e-06, "loss": 0.5844, "step": 5873 }, { "epoch": 0.5005538985939497, "grad_norm": 18.375, "learning_rate": 2.1419849946543554e-06, "loss": 1.0016, "step": 5874 }, { "epoch": 0.5006391137622497, "grad_norm": 18.25, "learning_rate": 2.1418672666673716e-06, "loss": 1.0531, "step": 5875 }, { "epoch": 0.5007243289305496, "grad_norm": 14.25, "learning_rate": 2.1417495225634497e-06, "loss": 0.7746, "step": 5876 }, { "epoch": 0.5008095440988496, "grad_norm": 16.5, "learning_rate": 2.1416317623447176e-06, "loss": 0.9421, "step": 5877 }, { "epoch": 0.5008947592671495, "grad_norm": 21.375, "learning_rate": 2.141513986013302e-06, "loss": 0.5886, "step": 5878 }, { "epoch": 0.5009799744354495, "grad_norm": 10.0625, "learning_rate": 2.141396193571333e-06, "loss": 0.3807, "step": 5879 }, { "epoch": 0.5010651896037495, "grad_norm": 12.9375, "learning_rate": 2.141278385020938e-06, "loss": 0.5249, "step": 5880 }, { "epoch": 0.5011504047720494, "grad_norm": 18.625, "learning_rate": 2.1411605603642466e-06, "loss": 0.9046, "step": 5881 }, { "epoch": 0.5012356199403494, "grad_norm": 13.0625, "learning_rate": 2.1410427196033877e-06, "loss": 0.5197, "step": 5882 }, { "epoch": 0.5013208351086493, "grad_norm": 12.3125, "learning_rate": 2.1409248627404905e-06, "loss": 0.4229, "step": 5883 }, { "epoch": 0.5014060502769493, "grad_norm": 16.125, "learning_rate": 2.1408069897776857e-06, "loss": 0.9829, "step": 5884 }, { "epoch": 0.5014912654452492, "grad_norm": 16.0, "learning_rate": 2.1406891007171023e-06, "loss": 0.9808, "step": 5885 }, { "epoch": 0.5015764806135492, "grad_norm": 12.4375, "learning_rate": 2.1405711955608714e-06, "loss": 0.5924, "step": 5886 }, { "epoch": 0.5016616957818492, "grad_norm": 11.9375, "learning_rate": 2.1404532743111235e-06, "loss": 0.532, "step": 5887 }, { "epoch": 0.5017469109501491, "grad_norm": 10.4375, "learning_rate": 2.1403353369699894e-06, "loss": 0.4992, "step": 5888 }, { "epoch": 0.5018321261184491, "grad_norm": 23.5, "learning_rate": 2.1402173835396006e-06, "loss": 1.0601, "step": 5889 }, { "epoch": 0.501917341286749, "grad_norm": 10.8125, "learning_rate": 2.1400994140220883e-06, "loss": 0.5125, "step": 5890 }, { "epoch": 0.502002556455049, "grad_norm": 20.25, "learning_rate": 2.1399814284195848e-06, "loss": 0.8205, "step": 5891 }, { "epoch": 0.5020877716233489, "grad_norm": 12.75, "learning_rate": 2.139863426734222e-06, "loss": 0.535, "step": 5892 }, { "epoch": 0.5021729867916489, "grad_norm": 12.8125, "learning_rate": 2.139745408968132e-06, "loss": 0.3571, "step": 5893 }, { "epoch": 0.5022582019599489, "grad_norm": 22.125, "learning_rate": 2.1396273751234475e-06, "loss": 1.0411, "step": 5894 }, { "epoch": 0.5023434171282488, "grad_norm": 15.4375, "learning_rate": 2.1395093252023022e-06, "loss": 0.9942, "step": 5895 }, { "epoch": 0.5024286322965488, "grad_norm": 14.1875, "learning_rate": 2.1393912592068285e-06, "loss": 0.8437, "step": 5896 }, { "epoch": 0.5025138474648487, "grad_norm": 13.8125, "learning_rate": 2.1392731771391607e-06, "loss": 0.4644, "step": 5897 }, { "epoch": 0.5025990626331487, "grad_norm": 17.0, "learning_rate": 2.1391550790014327e-06, "loss": 0.7442, "step": 5898 }, { "epoch": 0.5026842778014486, "grad_norm": 14.1875, "learning_rate": 2.1390369647957777e-06, "loss": 1.0194, "step": 5899 }, { "epoch": 0.5027694929697486, "grad_norm": 16.0, "learning_rate": 2.138918834524331e-06, "loss": 0.9234, "step": 5900 }, { "epoch": 0.5028547081380486, "grad_norm": 13.5625, "learning_rate": 2.138800688189227e-06, "loss": 0.8514, "step": 5901 }, { "epoch": 0.5029399233063485, "grad_norm": 13.8125, "learning_rate": 2.138682525792601e-06, "loss": 0.7345, "step": 5902 }, { "epoch": 0.5030251384746485, "grad_norm": 14.125, "learning_rate": 2.138564347336588e-06, "loss": 0.6518, "step": 5903 }, { "epoch": 0.5031103536429484, "grad_norm": 12.1875, "learning_rate": 2.1384461528233235e-06, "loss": 0.7844, "step": 5904 }, { "epoch": 0.5031955688112484, "grad_norm": 14.375, "learning_rate": 2.138327942254944e-06, "loss": 0.7696, "step": 5905 }, { "epoch": 0.5032807839795483, "grad_norm": 10.0625, "learning_rate": 2.138209715633585e-06, "loss": 0.3982, "step": 5906 }, { "epoch": 0.5033659991478483, "grad_norm": 13.25, "learning_rate": 2.138091472961383e-06, "loss": 0.8193, "step": 5907 }, { "epoch": 0.5034512143161483, "grad_norm": 17.5, "learning_rate": 2.1379732142404758e-06, "loss": 0.7489, "step": 5908 }, { "epoch": 0.5035364294844482, "grad_norm": 13.8125, "learning_rate": 2.1378549394729993e-06, "loss": 0.7697, "step": 5909 }, { "epoch": 0.5036216446527482, "grad_norm": 17.125, "learning_rate": 2.1377366486610913e-06, "loss": 1.1737, "step": 5910 }, { "epoch": 0.5037068598210481, "grad_norm": 14.8125, "learning_rate": 2.137618341806889e-06, "loss": 0.7671, "step": 5911 }, { "epoch": 0.5037920749893481, "grad_norm": 28.125, "learning_rate": 2.1375000189125316e-06, "loss": 0.918, "step": 5912 }, { "epoch": 0.503877290157648, "grad_norm": 14.4375, "learning_rate": 2.1373816799801554e-06, "loss": 0.8216, "step": 5913 }, { "epoch": 0.503962505325948, "grad_norm": 12.8125, "learning_rate": 2.1372633250119005e-06, "loss": 0.4578, "step": 5914 }, { "epoch": 0.504047720494248, "grad_norm": 21.25, "learning_rate": 2.137144954009905e-06, "loss": 0.7276, "step": 5915 }, { "epoch": 0.5041329356625479, "grad_norm": 15.0625, "learning_rate": 2.137026566976308e-06, "loss": 0.6661, "step": 5916 }, { "epoch": 0.5042181508308479, "grad_norm": 32.25, "learning_rate": 2.136908163913249e-06, "loss": 1.0548, "step": 5917 }, { "epoch": 0.5043033659991478, "grad_norm": 21.25, "learning_rate": 2.1367897448228676e-06, "loss": 0.9312, "step": 5918 }, { "epoch": 0.5043885811674478, "grad_norm": 23.625, "learning_rate": 2.1366713097073037e-06, "loss": 1.1637, "step": 5919 }, { "epoch": 0.5044737963357477, "grad_norm": 24.0, "learning_rate": 2.136552858568698e-06, "loss": 1.2734, "step": 5920 }, { "epoch": 0.5045590115040477, "grad_norm": 12.4375, "learning_rate": 2.13643439140919e-06, "loss": 0.6037, "step": 5921 }, { "epoch": 0.5046442266723477, "grad_norm": 12.75, "learning_rate": 2.136315908230922e-06, "loss": 0.4343, "step": 5922 }, { "epoch": 0.5047294418406476, "grad_norm": 16.25, "learning_rate": 2.136197409036034e-06, "loss": 0.3127, "step": 5923 }, { "epoch": 0.5048146570089476, "grad_norm": 11.8125, "learning_rate": 2.136078893826667e-06, "loss": 0.5275, "step": 5924 }, { "epoch": 0.5048998721772475, "grad_norm": 17.0, "learning_rate": 2.135960362604964e-06, "loss": 0.7268, "step": 5925 }, { "epoch": 0.5049850873455475, "grad_norm": 11.9375, "learning_rate": 2.135841815373066e-06, "loss": 0.6535, "step": 5926 }, { "epoch": 0.5050703025138474, "grad_norm": 18.375, "learning_rate": 2.1357232521331157e-06, "loss": 1.2585, "step": 5927 }, { "epoch": 0.5051555176821474, "grad_norm": 14.125, "learning_rate": 2.135604672887256e-06, "loss": 0.7749, "step": 5928 }, { "epoch": 0.5052407328504473, "grad_norm": 12.25, "learning_rate": 2.135486077637629e-06, "loss": 0.5934, "step": 5929 }, { "epoch": 0.5053259480187473, "grad_norm": 9.875, "learning_rate": 2.135367466386378e-06, "loss": 0.3538, "step": 5930 }, { "epoch": 0.5054111631870473, "grad_norm": 13.4375, "learning_rate": 2.135248839135647e-06, "loss": 0.6921, "step": 5931 }, { "epoch": 0.5054963783553472, "grad_norm": 14.6875, "learning_rate": 2.1351301958875788e-06, "loss": 0.7804, "step": 5932 }, { "epoch": 0.5055815935236472, "grad_norm": 11.875, "learning_rate": 2.135011536644318e-06, "loss": 0.6082, "step": 5933 }, { "epoch": 0.5056668086919471, "grad_norm": 11.625, "learning_rate": 2.1348928614080093e-06, "loss": 0.3377, "step": 5934 }, { "epoch": 0.5057520238602471, "grad_norm": 11.6875, "learning_rate": 2.134774170180796e-06, "loss": 0.9423, "step": 5935 }, { "epoch": 0.505837239028547, "grad_norm": 23.125, "learning_rate": 2.134655462964824e-06, "loss": 0.7924, "step": 5936 }, { "epoch": 0.505922454196847, "grad_norm": 13.75, "learning_rate": 2.1345367397622386e-06, "loss": 0.7311, "step": 5937 }, { "epoch": 0.506007669365147, "grad_norm": 16.625, "learning_rate": 2.134418000575185e-06, "loss": 1.3445, "step": 5938 }, { "epoch": 0.5060928845334469, "grad_norm": 11.3125, "learning_rate": 2.1342992454058077e-06, "loss": 0.4964, "step": 5939 }, { "epoch": 0.5061780997017469, "grad_norm": 15.9375, "learning_rate": 2.1341804742562545e-06, "loss": 0.9353, "step": 5940 }, { "epoch": 0.5062633148700468, "grad_norm": 12.4375, "learning_rate": 2.134061687128671e-06, "loss": 0.8831, "step": 5941 }, { "epoch": 0.5063485300383468, "grad_norm": 11.0, "learning_rate": 2.1339428840252036e-06, "loss": 0.595, "step": 5942 }, { "epoch": 0.5064337452066467, "grad_norm": 17.125, "learning_rate": 2.1338240649479995e-06, "loss": 0.6384, "step": 5943 }, { "epoch": 0.5065189603749467, "grad_norm": 18.125, "learning_rate": 2.133705229899206e-06, "loss": 1.064, "step": 5944 }, { "epoch": 0.5066041755432467, "grad_norm": 16.125, "learning_rate": 2.1335863788809703e-06, "loss": 0.9516, "step": 5945 }, { "epoch": 0.5066893907115466, "grad_norm": 16.375, "learning_rate": 2.13346751189544e-06, "loss": 0.5388, "step": 5946 }, { "epoch": 0.5067746058798466, "grad_norm": 13.1875, "learning_rate": 2.1333486289447634e-06, "loss": 0.3838, "step": 5947 }, { "epoch": 0.5068598210481465, "grad_norm": 15.0, "learning_rate": 2.133229730031089e-06, "loss": 0.6363, "step": 5948 }, { "epoch": 0.5069450362164465, "grad_norm": 34.25, "learning_rate": 2.1331108151565655e-06, "loss": 1.8226, "step": 5949 }, { "epoch": 0.5070302513847464, "grad_norm": 14.5625, "learning_rate": 2.1329918843233412e-06, "loss": 0.7174, "step": 5950 }, { "epoch": 0.5071154665530464, "grad_norm": 23.875, "learning_rate": 2.1328729375335655e-06, "loss": 0.849, "step": 5951 }, { "epoch": 0.5072006817213464, "grad_norm": 33.5, "learning_rate": 2.1327539747893875e-06, "loss": 0.9369, "step": 5952 }, { "epoch": 0.5072858968896463, "grad_norm": 12.1875, "learning_rate": 2.132634996092959e-06, "loss": 0.5689, "step": 5953 }, { "epoch": 0.5073711120579463, "grad_norm": 11.6875, "learning_rate": 2.1325160014464273e-06, "loss": 0.3275, "step": 5954 }, { "epoch": 0.5074563272262462, "grad_norm": 14.8125, "learning_rate": 2.1323969908519445e-06, "loss": 0.7506, "step": 5955 }, { "epoch": 0.5075415423945462, "grad_norm": 11.125, "learning_rate": 2.1322779643116603e-06, "loss": 0.5071, "step": 5956 }, { "epoch": 0.5076267575628461, "grad_norm": 20.375, "learning_rate": 2.1321589218277266e-06, "loss": 0.6119, "step": 5957 }, { "epoch": 0.5077119727311461, "grad_norm": 13.75, "learning_rate": 2.132039863402294e-06, "loss": 0.912, "step": 5958 }, { "epoch": 0.5077971878994461, "grad_norm": 12.25, "learning_rate": 2.131920789037514e-06, "loss": 0.5427, "step": 5959 }, { "epoch": 0.507882403067746, "grad_norm": 13.6875, "learning_rate": 2.1318016987355385e-06, "loss": 0.5217, "step": 5960 }, { "epoch": 0.507967618236046, "grad_norm": 19.75, "learning_rate": 2.1316825924985198e-06, "loss": 0.5625, "step": 5961 }, { "epoch": 0.5080528334043459, "grad_norm": 11.0, "learning_rate": 2.1315634703286097e-06, "loss": 0.565, "step": 5962 }, { "epoch": 0.5081380485726459, "grad_norm": 13.375, "learning_rate": 2.1314443322279615e-06, "loss": 0.7379, "step": 5963 }, { "epoch": 0.5082232637409458, "grad_norm": 16.125, "learning_rate": 2.131325178198728e-06, "loss": 0.6521, "step": 5964 }, { "epoch": 0.5083084789092458, "grad_norm": 19.0, "learning_rate": 2.131206008243062e-06, "loss": 1.0253, "step": 5965 }, { "epoch": 0.5083936940775458, "grad_norm": 14.875, "learning_rate": 2.131086822363117e-06, "loss": 0.4935, "step": 5966 }, { "epoch": 0.5084789092458457, "grad_norm": 29.625, "learning_rate": 2.1309676205610474e-06, "loss": 1.0021, "step": 5967 }, { "epoch": 0.5085641244141457, "grad_norm": 12.5625, "learning_rate": 2.1308484028390073e-06, "loss": 0.7737, "step": 5968 }, { "epoch": 0.5086493395824456, "grad_norm": 15.375, "learning_rate": 2.1307291691991506e-06, "loss": 0.7197, "step": 5969 }, { "epoch": 0.5087345547507456, "grad_norm": 14.9375, "learning_rate": 2.130609919643632e-06, "loss": 0.6636, "step": 5970 }, { "epoch": 0.5088197699190455, "grad_norm": 10.8125, "learning_rate": 2.130490654174607e-06, "loss": 0.5925, "step": 5971 }, { "epoch": 0.5089049850873455, "grad_norm": 13.75, "learning_rate": 2.13037137279423e-06, "loss": 0.9755, "step": 5972 }, { "epoch": 0.5089902002556455, "grad_norm": 14.8125, "learning_rate": 2.1302520755046572e-06, "loss": 0.7805, "step": 5973 }, { "epoch": 0.5090754154239454, "grad_norm": 15.5, "learning_rate": 2.130132762308045e-06, "loss": 0.9451, "step": 5974 }, { "epoch": 0.5091606305922454, "grad_norm": 16.125, "learning_rate": 2.1300134332065476e-06, "loss": 0.5044, "step": 5975 }, { "epoch": 0.5092458457605453, "grad_norm": 21.125, "learning_rate": 2.129894088202323e-06, "loss": 1.1951, "step": 5976 }, { "epoch": 0.5093310609288453, "grad_norm": 11.1875, "learning_rate": 2.1297747272975274e-06, "loss": 0.4616, "step": 5977 }, { "epoch": 0.5094162760971452, "grad_norm": 16.5, "learning_rate": 2.129655350494318e-06, "loss": 1.068, "step": 5978 }, { "epoch": 0.5095014912654452, "grad_norm": 12.75, "learning_rate": 2.1295359577948515e-06, "loss": 0.4224, "step": 5979 }, { "epoch": 0.5095867064337452, "grad_norm": 14.375, "learning_rate": 2.1294165492012856e-06, "loss": 0.9224, "step": 5980 }, { "epoch": 0.5096719216020452, "grad_norm": 10.875, "learning_rate": 2.129297124715779e-06, "loss": 0.4246, "step": 5981 }, { "epoch": 0.5097571367703452, "grad_norm": 13.0625, "learning_rate": 2.129177684340489e-06, "loss": 0.7747, "step": 5982 }, { "epoch": 0.5098423519386451, "grad_norm": 16.0, "learning_rate": 2.129058228077574e-06, "loss": 0.7793, "step": 5983 }, { "epoch": 0.5099275671069451, "grad_norm": 18.375, "learning_rate": 2.1289387559291927e-06, "loss": 0.492, "step": 5984 }, { "epoch": 0.510012782275245, "grad_norm": 21.875, "learning_rate": 2.1288192678975046e-06, "loss": 0.8946, "step": 5985 }, { "epoch": 0.510097997443545, "grad_norm": 13.0625, "learning_rate": 2.1286997639846683e-06, "loss": 0.7928, "step": 5986 }, { "epoch": 0.510183212611845, "grad_norm": 10.125, "learning_rate": 2.1285802441928437e-06, "loss": 0.4138, "step": 5987 }, { "epoch": 0.5102684277801449, "grad_norm": 13.9375, "learning_rate": 2.128460708524191e-06, "loss": 0.8671, "step": 5988 }, { "epoch": 0.5103536429484449, "grad_norm": 16.375, "learning_rate": 2.1283411569808695e-06, "loss": 0.9474, "step": 5989 }, { "epoch": 0.5104388581167448, "grad_norm": 22.125, "learning_rate": 2.1282215895650405e-06, "loss": 0.9463, "step": 5990 }, { "epoch": 0.5105240732850448, "grad_norm": 15.4375, "learning_rate": 2.128102006278864e-06, "loss": 1.1134, "step": 5991 }, { "epoch": 0.5106092884533447, "grad_norm": 19.5, "learning_rate": 2.1279824071245007e-06, "loss": 0.8526, "step": 5992 }, { "epoch": 0.5106945036216447, "grad_norm": 20.125, "learning_rate": 2.1278627921041128e-06, "loss": 0.8309, "step": 5993 }, { "epoch": 0.5107797187899447, "grad_norm": 9.9375, "learning_rate": 2.127743161219862e-06, "loss": 0.3082, "step": 5994 }, { "epoch": 0.5108649339582446, "grad_norm": 11.125, "learning_rate": 2.1276235144739087e-06, "loss": 0.3851, "step": 5995 }, { "epoch": 0.5109501491265446, "grad_norm": 13.4375, "learning_rate": 2.1275038518684168e-06, "loss": 0.6475, "step": 5996 }, { "epoch": 0.5110353642948445, "grad_norm": 14.0625, "learning_rate": 2.1273841734055478e-06, "loss": 0.6217, "step": 5997 }, { "epoch": 0.5111205794631445, "grad_norm": 13.5625, "learning_rate": 2.127264479087464e-06, "loss": 0.7222, "step": 5998 }, { "epoch": 0.5112057946314444, "grad_norm": 14.5, "learning_rate": 2.1271447689163293e-06, "loss": 0.5706, "step": 5999 }, { "epoch": 0.5112910097997444, "grad_norm": 11.8125, "learning_rate": 2.1270250428943064e-06, "loss": 0.5834, "step": 6000 }, { "epoch": 0.5113762249680444, "grad_norm": 18.0, "learning_rate": 2.126905301023559e-06, "loss": 0.9105, "step": 6001 }, { "epoch": 0.5114614401363443, "grad_norm": 13.1875, "learning_rate": 2.126785543306252e-06, "loss": 0.8711, "step": 6002 }, { "epoch": 0.5115466553046443, "grad_norm": 15.875, "learning_rate": 2.126665769744547e-06, "loss": 0.7354, "step": 6003 }, { "epoch": 0.5116318704729442, "grad_norm": 15.75, "learning_rate": 2.126545980340611e-06, "loss": 0.615, "step": 6004 }, { "epoch": 0.5117170856412442, "grad_norm": 11.8125, "learning_rate": 2.1264261750966073e-06, "loss": 0.5414, "step": 6005 }, { "epoch": 0.5118023008095441, "grad_norm": 19.625, "learning_rate": 2.1263063540147018e-06, "loss": 0.7888, "step": 6006 }, { "epoch": 0.5118875159778441, "grad_norm": 14.0625, "learning_rate": 2.1261865170970587e-06, "loss": 0.5355, "step": 6007 }, { "epoch": 0.5119727311461441, "grad_norm": 13.25, "learning_rate": 2.1260666643458445e-06, "loss": 0.552, "step": 6008 }, { "epoch": 0.512057946314444, "grad_norm": 14.3125, "learning_rate": 2.1259467957632244e-06, "loss": 0.9799, "step": 6009 }, { "epoch": 0.512143161482744, "grad_norm": 13.9375, "learning_rate": 2.1258269113513653e-06, "loss": 0.6855, "step": 6010 }, { "epoch": 0.5122283766510439, "grad_norm": 12.1875, "learning_rate": 2.125707011112433e-06, "loss": 0.5838, "step": 6011 }, { "epoch": 0.5123135918193439, "grad_norm": 13.6875, "learning_rate": 2.125587095048594e-06, "loss": 0.6412, "step": 6012 }, { "epoch": 0.5123988069876438, "grad_norm": 12.1875, "learning_rate": 2.125467163162016e-06, "loss": 0.5421, "step": 6013 }, { "epoch": 0.5124840221559438, "grad_norm": 19.875, "learning_rate": 2.1253472154548665e-06, "loss": 0.6201, "step": 6014 }, { "epoch": 0.5125692373242438, "grad_norm": 15.0, "learning_rate": 2.1252272519293123e-06, "loss": 0.6306, "step": 6015 }, { "epoch": 0.5126544524925437, "grad_norm": 15.5, "learning_rate": 2.125107272587521e-06, "loss": 0.4924, "step": 6016 }, { "epoch": 0.5127396676608437, "grad_norm": 21.25, "learning_rate": 2.1249872774316623e-06, "loss": 0.8329, "step": 6017 }, { "epoch": 0.5128248828291436, "grad_norm": 12.1875, "learning_rate": 2.124867266463903e-06, "loss": 0.7385, "step": 6018 }, { "epoch": 0.5129100979974436, "grad_norm": 14.0, "learning_rate": 2.1247472396864123e-06, "loss": 0.7668, "step": 6019 }, { "epoch": 0.5129953131657435, "grad_norm": 14.875, "learning_rate": 2.1246271971013596e-06, "loss": 0.9232, "step": 6020 }, { "epoch": 0.5130805283340435, "grad_norm": 14.6875, "learning_rate": 2.124507138710914e-06, "loss": 1.0421, "step": 6021 }, { "epoch": 0.5131657435023435, "grad_norm": 19.125, "learning_rate": 2.124387064517245e-06, "loss": 0.8987, "step": 6022 }, { "epoch": 0.5132509586706434, "grad_norm": 17.0, "learning_rate": 2.1242669745225223e-06, "loss": 0.9868, "step": 6023 }, { "epoch": 0.5133361738389434, "grad_norm": 14.8125, "learning_rate": 2.124146868728916e-06, "loss": 0.7766, "step": 6024 }, { "epoch": 0.5134213890072433, "grad_norm": 16.0, "learning_rate": 2.124026747138597e-06, "loss": 0.9565, "step": 6025 }, { "epoch": 0.5135066041755433, "grad_norm": 16.375, "learning_rate": 2.1239066097537363e-06, "loss": 0.9806, "step": 6026 }, { "epoch": 0.5135918193438432, "grad_norm": 10.4375, "learning_rate": 2.123786456576504e-06, "loss": 0.3605, "step": 6027 }, { "epoch": 0.5136770345121432, "grad_norm": 19.25, "learning_rate": 2.123666287609072e-06, "loss": 0.8089, "step": 6028 }, { "epoch": 0.5137622496804432, "grad_norm": 17.875, "learning_rate": 2.1235461028536112e-06, "loss": 0.8724, "step": 6029 }, { "epoch": 0.5138474648487431, "grad_norm": 15.625, "learning_rate": 2.123425902312294e-06, "loss": 0.736, "step": 6030 }, { "epoch": 0.5139326800170431, "grad_norm": 14.5625, "learning_rate": 2.1233056859872925e-06, "loss": 0.8992, "step": 6031 }, { "epoch": 0.514017895185343, "grad_norm": 15.6875, "learning_rate": 2.123185453880779e-06, "loss": 1.0508, "step": 6032 }, { "epoch": 0.514103110353643, "grad_norm": 11.625, "learning_rate": 2.1230652059949265e-06, "loss": 0.5298, "step": 6033 }, { "epoch": 0.5141883255219429, "grad_norm": 11.625, "learning_rate": 2.1229449423319075e-06, "loss": 0.7364, "step": 6034 }, { "epoch": 0.5142735406902429, "grad_norm": 17.5, "learning_rate": 2.122824662893896e-06, "loss": 0.864, "step": 6035 }, { "epoch": 0.5143587558585428, "grad_norm": 11.1875, "learning_rate": 2.122704367683065e-06, "loss": 0.9077, "step": 6036 }, { "epoch": 0.5144439710268428, "grad_norm": 15.6875, "learning_rate": 2.122584056701588e-06, "loss": 0.8837, "step": 6037 }, { "epoch": 0.5145291861951428, "grad_norm": 23.0, "learning_rate": 2.1224637299516403e-06, "loss": 1.0585, "step": 6038 }, { "epoch": 0.5146144013634427, "grad_norm": 17.0, "learning_rate": 2.1223433874353956e-06, "loss": 1.0097, "step": 6039 }, { "epoch": 0.5146996165317427, "grad_norm": 14.125, "learning_rate": 2.1222230291550284e-06, "loss": 0.9164, "step": 6040 }, { "epoch": 0.5147848317000426, "grad_norm": 16.125, "learning_rate": 2.122102655112714e-06, "loss": 0.6209, "step": 6041 }, { "epoch": 0.5148700468683426, "grad_norm": 15.0625, "learning_rate": 2.121982265310628e-06, "loss": 0.6848, "step": 6042 }, { "epoch": 0.5149552620366425, "grad_norm": 11.0625, "learning_rate": 2.1218618597509453e-06, "loss": 0.503, "step": 6043 }, { "epoch": 0.5150404772049425, "grad_norm": 15.125, "learning_rate": 2.121741438435842e-06, "loss": 0.8655, "step": 6044 }, { "epoch": 0.5151256923732425, "grad_norm": 14.3125, "learning_rate": 2.121621001367494e-06, "loss": 0.9471, "step": 6045 }, { "epoch": 0.5152109075415424, "grad_norm": 11.25, "learning_rate": 2.121500548548079e-06, "loss": 0.4627, "step": 6046 }, { "epoch": 0.5152961227098424, "grad_norm": 21.125, "learning_rate": 2.121380079979772e-06, "loss": 0.9475, "step": 6047 }, { "epoch": 0.5153813378781423, "grad_norm": 16.125, "learning_rate": 2.121259595664751e-06, "loss": 0.8405, "step": 6048 }, { "epoch": 0.5154665530464423, "grad_norm": 19.625, "learning_rate": 2.1211390956051927e-06, "loss": 1.0412, "step": 6049 }, { "epoch": 0.5155517682147422, "grad_norm": 16.625, "learning_rate": 2.121018579803275e-06, "loss": 0.9891, "step": 6050 }, { "epoch": 0.5156369833830422, "grad_norm": 14.25, "learning_rate": 2.1208980482611763e-06, "loss": 0.6184, "step": 6051 }, { "epoch": 0.5157221985513422, "grad_norm": 15.125, "learning_rate": 2.1207775009810736e-06, "loss": 1.0893, "step": 6052 }, { "epoch": 0.5158074137196421, "grad_norm": 14.0, "learning_rate": 2.1206569379651456e-06, "loss": 0.6063, "step": 6053 }, { "epoch": 0.5158926288879421, "grad_norm": 16.75, "learning_rate": 2.1205363592155713e-06, "loss": 1.0436, "step": 6054 }, { "epoch": 0.515977844056242, "grad_norm": 16.625, "learning_rate": 2.12041576473453e-06, "loss": 0.7825, "step": 6055 }, { "epoch": 0.516063059224542, "grad_norm": 18.125, "learning_rate": 2.1202951545242006e-06, "loss": 0.8175, "step": 6056 }, { "epoch": 0.5161482743928419, "grad_norm": 12.75, "learning_rate": 2.120174528586762e-06, "loss": 0.3533, "step": 6057 }, { "epoch": 0.5162334895611419, "grad_norm": 13.1875, "learning_rate": 2.120053886924395e-06, "loss": 0.498, "step": 6058 }, { "epoch": 0.5163187047294419, "grad_norm": 15.125, "learning_rate": 2.1199332295392796e-06, "loss": 1.0237, "step": 6059 }, { "epoch": 0.5164039198977418, "grad_norm": 25.125, "learning_rate": 2.1198125564335956e-06, "loss": 1.0128, "step": 6060 }, { "epoch": 0.5164891350660418, "grad_norm": 13.8125, "learning_rate": 2.119691867609524e-06, "loss": 0.6675, "step": 6061 }, { "epoch": 0.5165743502343417, "grad_norm": 14.8125, "learning_rate": 2.1195711630692463e-06, "loss": 0.7392, "step": 6062 }, { "epoch": 0.5166595654026417, "grad_norm": 17.625, "learning_rate": 2.1194504428149425e-06, "loss": 0.7134, "step": 6063 }, { "epoch": 0.5167447805709416, "grad_norm": 22.25, "learning_rate": 2.1193297068487958e-06, "loss": 0.8278, "step": 6064 }, { "epoch": 0.5168299957392416, "grad_norm": 14.9375, "learning_rate": 2.119208955172986e-06, "loss": 0.7356, "step": 6065 }, { "epoch": 0.5169152109075416, "grad_norm": 25.375, "learning_rate": 2.119088187789697e-06, "loss": 1.1296, "step": 6066 }, { "epoch": 0.5170004260758415, "grad_norm": 9.3125, "learning_rate": 2.1189674047011106e-06, "loss": 0.3042, "step": 6067 }, { "epoch": 0.5170856412441415, "grad_norm": 10.4375, "learning_rate": 2.1188466059094092e-06, "loss": 0.4142, "step": 6068 }, { "epoch": 0.5171708564124414, "grad_norm": 18.375, "learning_rate": 2.118725791416776e-06, "loss": 0.8698, "step": 6069 }, { "epoch": 0.5172560715807414, "grad_norm": 17.125, "learning_rate": 2.118604961225394e-06, "loss": 0.7608, "step": 6070 }, { "epoch": 0.5173412867490413, "grad_norm": 17.0, "learning_rate": 2.1184841153374475e-06, "loss": 0.8325, "step": 6071 }, { "epoch": 0.5174265019173413, "grad_norm": 13.625, "learning_rate": 2.118363253755119e-06, "loss": 0.4737, "step": 6072 }, { "epoch": 0.5175117170856413, "grad_norm": 15.3125, "learning_rate": 2.1182423764805937e-06, "loss": 0.8842, "step": 6073 }, { "epoch": 0.5175969322539412, "grad_norm": 14.3125, "learning_rate": 2.118121483516055e-06, "loss": 0.6919, "step": 6074 }, { "epoch": 0.5176821474222412, "grad_norm": 14.0, "learning_rate": 2.118000574863689e-06, "loss": 0.7782, "step": 6075 }, { "epoch": 0.5177673625905411, "grad_norm": 20.375, "learning_rate": 2.1178796505256794e-06, "loss": 1.2393, "step": 6076 }, { "epoch": 0.5178525777588411, "grad_norm": 13.0, "learning_rate": 2.1177587105042114e-06, "loss": 0.6245, "step": 6077 }, { "epoch": 0.517937792927141, "grad_norm": 13.3125, "learning_rate": 2.1176377548014717e-06, "loss": 0.6985, "step": 6078 }, { "epoch": 0.518023008095441, "grad_norm": 10.875, "learning_rate": 2.1175167834196443e-06, "loss": 0.2675, "step": 6079 }, { "epoch": 0.518108223263741, "grad_norm": 10.3125, "learning_rate": 2.117395796360917e-06, "loss": 0.4387, "step": 6080 }, { "epoch": 0.5181934384320409, "grad_norm": 16.125, "learning_rate": 2.117274793627475e-06, "loss": 0.7577, "step": 6081 }, { "epoch": 0.5182786536003409, "grad_norm": 23.375, "learning_rate": 2.117153775221506e-06, "loss": 1.2516, "step": 6082 }, { "epoch": 0.5183638687686408, "grad_norm": 16.625, "learning_rate": 2.1170327411451957e-06, "loss": 0.7569, "step": 6083 }, { "epoch": 0.5184490839369408, "grad_norm": 15.625, "learning_rate": 2.116911691400732e-06, "loss": 0.8209, "step": 6084 }, { "epoch": 0.5185342991052407, "grad_norm": 29.75, "learning_rate": 2.116790625990302e-06, "loss": 0.9579, "step": 6085 }, { "epoch": 0.5186195142735407, "grad_norm": 13.0625, "learning_rate": 2.1166695449160946e-06, "loss": 0.7305, "step": 6086 }, { "epoch": 0.5187047294418407, "grad_norm": 11.6875, "learning_rate": 2.116548448180297e-06, "loss": 0.4928, "step": 6087 }, { "epoch": 0.5187899446101406, "grad_norm": 14.75, "learning_rate": 2.1164273357850965e-06, "loss": 0.6753, "step": 6088 }, { "epoch": 0.5188751597784406, "grad_norm": 15.5625, "learning_rate": 2.1163062077326836e-06, "loss": 0.6788, "step": 6089 }, { "epoch": 0.5189603749467405, "grad_norm": 20.875, "learning_rate": 2.116185064025246e-06, "loss": 1.0482, "step": 6090 }, { "epoch": 0.5190455901150405, "grad_norm": 18.75, "learning_rate": 2.1160639046649733e-06, "loss": 0.6192, "step": 6091 }, { "epoch": 0.5191308052833404, "grad_norm": 17.75, "learning_rate": 2.1159427296540554e-06, "loss": 1.3317, "step": 6092 }, { "epoch": 0.5192160204516404, "grad_norm": 10.8125, "learning_rate": 2.115821538994681e-06, "loss": 0.4594, "step": 6093 }, { "epoch": 0.5193012356199403, "grad_norm": 17.625, "learning_rate": 2.1157003326890413e-06, "loss": 0.6357, "step": 6094 }, { "epoch": 0.5193864507882403, "grad_norm": 10.875, "learning_rate": 2.115579110739326e-06, "loss": 0.439, "step": 6095 }, { "epoch": 0.5194716659565403, "grad_norm": 16.25, "learning_rate": 2.115457873147725e-06, "loss": 0.8373, "step": 6096 }, { "epoch": 0.5195568811248402, "grad_norm": 16.875, "learning_rate": 2.1153366199164306e-06, "loss": 0.8145, "step": 6097 }, { "epoch": 0.5196420962931402, "grad_norm": 15.0625, "learning_rate": 2.1152153510476333e-06, "loss": 1.0841, "step": 6098 }, { "epoch": 0.5197273114614401, "grad_norm": 16.125, "learning_rate": 2.1150940665435247e-06, "loss": 0.5388, "step": 6099 }, { "epoch": 0.5198125266297401, "grad_norm": 16.125, "learning_rate": 2.114972766406296e-06, "loss": 0.7404, "step": 6100 }, { "epoch": 0.51989774179804, "grad_norm": 17.125, "learning_rate": 2.11485145063814e-06, "loss": 0.9856, "step": 6101 }, { "epoch": 0.51998295696634, "grad_norm": 17.5, "learning_rate": 2.1147301192412485e-06, "loss": 0.845, "step": 6102 }, { "epoch": 0.52006817213464, "grad_norm": 18.5, "learning_rate": 2.1146087722178138e-06, "loss": 0.9277, "step": 6103 }, { "epoch": 0.5201533873029399, "grad_norm": 18.0, "learning_rate": 2.1144874095700297e-06, "loss": 0.7724, "step": 6104 }, { "epoch": 0.5202386024712399, "grad_norm": 14.1875, "learning_rate": 2.1143660313000887e-06, "loss": 0.8382, "step": 6105 }, { "epoch": 0.5203238176395398, "grad_norm": 15.75, "learning_rate": 2.114244637410184e-06, "loss": 0.7275, "step": 6106 }, { "epoch": 0.5204090328078398, "grad_norm": 12.5625, "learning_rate": 2.1141232279025097e-06, "loss": 0.5651, "step": 6107 }, { "epoch": 0.5204942479761397, "grad_norm": 17.625, "learning_rate": 2.1140018027792597e-06, "loss": 0.6045, "step": 6108 }, { "epoch": 0.5205794631444397, "grad_norm": 11.6875, "learning_rate": 2.1138803620426285e-06, "loss": 0.5992, "step": 6109 }, { "epoch": 0.5206646783127397, "grad_norm": 15.875, "learning_rate": 2.1137589056948104e-06, "loss": 0.5883, "step": 6110 }, { "epoch": 0.5207498934810396, "grad_norm": 15.4375, "learning_rate": 2.1136374337380005e-06, "loss": 1.1205, "step": 6111 }, { "epoch": 0.5208351086493396, "grad_norm": 14.0625, "learning_rate": 2.1135159461743937e-06, "loss": 0.863, "step": 6112 }, { "epoch": 0.5209203238176395, "grad_norm": 14.875, "learning_rate": 2.1133944430061853e-06, "loss": 0.6164, "step": 6113 }, { "epoch": 0.5210055389859395, "grad_norm": 17.25, "learning_rate": 2.1132729242355707e-06, "loss": 0.672, "step": 6114 }, { "epoch": 0.5210907541542394, "grad_norm": 16.875, "learning_rate": 2.1131513898647466e-06, "loss": 0.7632, "step": 6115 }, { "epoch": 0.5211759693225394, "grad_norm": 15.125, "learning_rate": 2.113029839895909e-06, "loss": 0.8079, "step": 6116 }, { "epoch": 0.5212611844908394, "grad_norm": 15.125, "learning_rate": 2.112908274331254e-06, "loss": 0.6959, "step": 6117 }, { "epoch": 0.5213463996591393, "grad_norm": 19.0, "learning_rate": 2.112786693172979e-06, "loss": 0.6458, "step": 6118 }, { "epoch": 0.5214316148274393, "grad_norm": 13.75, "learning_rate": 2.112665096423281e-06, "loss": 0.4932, "step": 6119 }, { "epoch": 0.5215168299957392, "grad_norm": 22.625, "learning_rate": 2.112543484084357e-06, "loss": 1.2836, "step": 6120 }, { "epoch": 0.5216020451640392, "grad_norm": 19.0, "learning_rate": 2.1124218561584053e-06, "loss": 1.1012, "step": 6121 }, { "epoch": 0.5216872603323391, "grad_norm": 17.25, "learning_rate": 2.112300212647623e-06, "loss": 1.1612, "step": 6122 }, { "epoch": 0.5217724755006391, "grad_norm": 21.375, "learning_rate": 2.1121785535542093e-06, "loss": 0.5436, "step": 6123 }, { "epoch": 0.5218576906689391, "grad_norm": 14.125, "learning_rate": 2.1120568788803617e-06, "loss": 0.5055, "step": 6124 }, { "epoch": 0.521942905837239, "grad_norm": 15.0625, "learning_rate": 2.1119351886282797e-06, "loss": 0.6316, "step": 6125 }, { "epoch": 0.522028121005539, "grad_norm": 13.75, "learning_rate": 2.1118134828001618e-06, "loss": 0.7559, "step": 6126 }, { "epoch": 0.5221133361738389, "grad_norm": 16.25, "learning_rate": 2.1116917613982072e-06, "loss": 0.568, "step": 6127 }, { "epoch": 0.5221985513421389, "grad_norm": 12.3125, "learning_rate": 2.1115700244246167e-06, "loss": 0.6198, "step": 6128 }, { "epoch": 0.5222837665104388, "grad_norm": 14.8125, "learning_rate": 2.1114482718815895e-06, "loss": 0.5334, "step": 6129 }, { "epoch": 0.5223689816787388, "grad_norm": 23.75, "learning_rate": 2.1113265037713253e-06, "loss": 0.7541, "step": 6130 }, { "epoch": 0.5224541968470388, "grad_norm": 16.625, "learning_rate": 2.1112047200960256e-06, "loss": 0.7181, "step": 6131 }, { "epoch": 0.5225394120153387, "grad_norm": 16.625, "learning_rate": 2.1110829208578905e-06, "loss": 0.7677, "step": 6132 }, { "epoch": 0.5226246271836387, "grad_norm": 14.5625, "learning_rate": 2.1109611060591207e-06, "loss": 0.7057, "step": 6133 }, { "epoch": 0.5227098423519386, "grad_norm": 16.625, "learning_rate": 2.110839275701918e-06, "loss": 0.726, "step": 6134 }, { "epoch": 0.5227950575202386, "grad_norm": 12.625, "learning_rate": 2.1107174297884843e-06, "loss": 0.4921, "step": 6135 }, { "epoch": 0.5228802726885385, "grad_norm": 14.0, "learning_rate": 2.110595568321021e-06, "loss": 0.7788, "step": 6136 }, { "epoch": 0.5229654878568385, "grad_norm": 17.75, "learning_rate": 2.1104736913017304e-06, "loss": 1.323, "step": 6137 }, { "epoch": 0.5230507030251385, "grad_norm": 20.875, "learning_rate": 2.110351798732815e-06, "loss": 0.7678, "step": 6138 }, { "epoch": 0.5231359181934384, "grad_norm": 19.75, "learning_rate": 2.110229890616477e-06, "loss": 1.2404, "step": 6139 }, { "epoch": 0.5232211333617384, "grad_norm": 20.0, "learning_rate": 2.11010796695492e-06, "loss": 0.6425, "step": 6140 }, { "epoch": 0.5233063485300383, "grad_norm": 14.5, "learning_rate": 2.1099860277503475e-06, "loss": 0.7397, "step": 6141 }, { "epoch": 0.5233915636983383, "grad_norm": 16.25, "learning_rate": 2.1098640730049623e-06, "loss": 0.7646, "step": 6142 }, { "epoch": 0.5234767788666382, "grad_norm": 11.6875, "learning_rate": 2.109742102720969e-06, "loss": 0.4993, "step": 6143 }, { "epoch": 0.5235619940349382, "grad_norm": 14.3125, "learning_rate": 2.1096201169005716e-06, "loss": 0.614, "step": 6144 }, { "epoch": 0.5236472092032382, "grad_norm": 13.4375, "learning_rate": 2.1094981155459736e-06, "loss": 0.5377, "step": 6145 }, { "epoch": 0.5237324243715381, "grad_norm": 13.0625, "learning_rate": 2.1093760986593804e-06, "loss": 0.8189, "step": 6146 }, { "epoch": 0.5238176395398381, "grad_norm": 14.75, "learning_rate": 2.109254066242997e-06, "loss": 0.7903, "step": 6147 }, { "epoch": 0.523902854708138, "grad_norm": 15.625, "learning_rate": 2.109132018299029e-06, "loss": 0.8813, "step": 6148 }, { "epoch": 0.523988069876438, "grad_norm": 9.1875, "learning_rate": 2.1090099548296814e-06, "loss": 0.3579, "step": 6149 }, { "epoch": 0.5240732850447379, "grad_norm": 16.5, "learning_rate": 2.10888787583716e-06, "loss": 0.6973, "step": 6150 }, { "epoch": 0.5241585002130379, "grad_norm": 16.0, "learning_rate": 2.108765781323671e-06, "loss": 0.6114, "step": 6151 }, { "epoch": 0.5242437153813378, "grad_norm": 16.375, "learning_rate": 2.1086436712914206e-06, "loss": 0.9928, "step": 6152 }, { "epoch": 0.5243289305496378, "grad_norm": 12.5, "learning_rate": 2.1085215457426157e-06, "loss": 0.3729, "step": 6153 }, { "epoch": 0.5244141457179378, "grad_norm": 13.9375, "learning_rate": 2.108399404679463e-06, "loss": 0.9815, "step": 6154 }, { "epoch": 0.5244993608862377, "grad_norm": 16.125, "learning_rate": 2.10827724810417e-06, "loss": 0.9905, "step": 6155 }, { "epoch": 0.5245845760545377, "grad_norm": 16.0, "learning_rate": 2.1081550760189444e-06, "loss": 0.7105, "step": 6156 }, { "epoch": 0.5246697912228376, "grad_norm": 14.375, "learning_rate": 2.108032888425994e-06, "loss": 0.552, "step": 6157 }, { "epoch": 0.5247550063911376, "grad_norm": 13.25, "learning_rate": 2.1079106853275255e-06, "loss": 0.5625, "step": 6158 }, { "epoch": 0.5248402215594375, "grad_norm": 25.25, "learning_rate": 2.1077884667257483e-06, "loss": 1.145, "step": 6159 }, { "epoch": 0.5249254367277375, "grad_norm": 11.5625, "learning_rate": 2.1076662326228713e-06, "loss": 0.3689, "step": 6160 }, { "epoch": 0.5250106518960375, "grad_norm": 14.25, "learning_rate": 2.107543983021103e-06, "loss": 0.5504, "step": 6161 }, { "epoch": 0.5250958670643374, "grad_norm": 14.875, "learning_rate": 2.1074217179226525e-06, "loss": 0.6711, "step": 6162 }, { "epoch": 0.5251810822326374, "grad_norm": 13.3125, "learning_rate": 2.1072994373297297e-06, "loss": 0.5921, "step": 6163 }, { "epoch": 0.5252662974009373, "grad_norm": 13.375, "learning_rate": 2.1071771412445434e-06, "loss": 0.6126, "step": 6164 }, { "epoch": 0.5253515125692373, "grad_norm": 9.25, "learning_rate": 2.1070548296693043e-06, "loss": 0.2824, "step": 6165 }, { "epoch": 0.5254367277375372, "grad_norm": 11.375, "learning_rate": 2.1069325026062232e-06, "loss": 0.4675, "step": 6166 }, { "epoch": 0.5255219429058372, "grad_norm": 16.125, "learning_rate": 2.1068101600575096e-06, "loss": 0.8072, "step": 6167 }, { "epoch": 0.5256071580741372, "grad_norm": 18.125, "learning_rate": 2.106687802025375e-06, "loss": 1.2718, "step": 6168 }, { "epoch": 0.5256923732424371, "grad_norm": 12.9375, "learning_rate": 2.10656542851203e-06, "loss": 0.5753, "step": 6169 }, { "epoch": 0.5257775884107371, "grad_norm": 15.6875, "learning_rate": 2.1064430395196867e-06, "loss": 1.0124, "step": 6170 }, { "epoch": 0.525862803579037, "grad_norm": 11.25, "learning_rate": 2.106320635050556e-06, "loss": 0.5928, "step": 6171 }, { "epoch": 0.525948018747337, "grad_norm": 13.25, "learning_rate": 2.106198215106851e-06, "loss": 0.728, "step": 6172 }, { "epoch": 0.5260332339156369, "grad_norm": 11.4375, "learning_rate": 2.106075779690783e-06, "loss": 0.3916, "step": 6173 }, { "epoch": 0.5261184490839369, "grad_norm": 15.125, "learning_rate": 2.1059533288045648e-06, "loss": 0.9074, "step": 6174 }, { "epoch": 0.5262036642522369, "grad_norm": 18.5, "learning_rate": 2.105830862450409e-06, "loss": 1.1213, "step": 6175 }, { "epoch": 0.5262888794205368, "grad_norm": 16.25, "learning_rate": 2.105708380630529e-06, "loss": 0.7609, "step": 6176 }, { "epoch": 0.5263740945888368, "grad_norm": 15.25, "learning_rate": 2.105585883347138e-06, "loss": 0.7624, "step": 6177 }, { "epoch": 0.5264593097571367, "grad_norm": 25.625, "learning_rate": 2.1054633706024504e-06, "loss": 0.9504, "step": 6178 }, { "epoch": 0.5265445249254367, "grad_norm": 14.9375, "learning_rate": 2.1053408423986786e-06, "loss": 0.9271, "step": 6179 }, { "epoch": 0.5266297400937366, "grad_norm": 13.5, "learning_rate": 2.1052182987380383e-06, "loss": 0.8987, "step": 6180 }, { "epoch": 0.5267149552620366, "grad_norm": 13.3125, "learning_rate": 2.1050957396227427e-06, "loss": 0.5143, "step": 6181 }, { "epoch": 0.5268001704303366, "grad_norm": 12.5, "learning_rate": 2.104973165055008e-06, "loss": 0.4736, "step": 6182 }, { "epoch": 0.5268853855986365, "grad_norm": 15.0, "learning_rate": 2.104850575037048e-06, "loss": 0.9557, "step": 6183 }, { "epoch": 0.5269706007669365, "grad_norm": 11.5, "learning_rate": 2.1047279695710786e-06, "loss": 0.7713, "step": 6184 }, { "epoch": 0.5270558159352364, "grad_norm": 16.25, "learning_rate": 2.1046053486593153e-06, "loss": 1.0148, "step": 6185 }, { "epoch": 0.5271410311035364, "grad_norm": 13.1875, "learning_rate": 2.1044827123039738e-06, "loss": 0.6012, "step": 6186 }, { "epoch": 0.5272262462718363, "grad_norm": 12.4375, "learning_rate": 2.1043600605072708e-06, "loss": 0.7012, "step": 6187 }, { "epoch": 0.5273114614401363, "grad_norm": 18.0, "learning_rate": 2.1042373932714217e-06, "loss": 1.245, "step": 6188 }, { "epoch": 0.5273966766084363, "grad_norm": 9.5625, "learning_rate": 2.104114710598645e-06, "loss": 0.4105, "step": 6189 }, { "epoch": 0.5274818917767362, "grad_norm": 18.375, "learning_rate": 2.103992012491156e-06, "loss": 0.7026, "step": 6190 }, { "epoch": 0.5275671069450362, "grad_norm": 17.5, "learning_rate": 2.1038692989511724e-06, "loss": 0.7081, "step": 6191 }, { "epoch": 0.5276523221133361, "grad_norm": 16.875, "learning_rate": 2.1037465699809117e-06, "loss": 0.5986, "step": 6192 }, { "epoch": 0.5277375372816361, "grad_norm": 18.625, "learning_rate": 2.1036238255825926e-06, "loss": 1.2049, "step": 6193 }, { "epoch": 0.527822752449936, "grad_norm": 20.125, "learning_rate": 2.1035010657584322e-06, "loss": 0.7627, "step": 6194 }, { "epoch": 0.527907967618236, "grad_norm": 15.375, "learning_rate": 2.1033782905106493e-06, "loss": 0.8761, "step": 6195 }, { "epoch": 0.527993182786536, "grad_norm": 19.0, "learning_rate": 2.1032554998414627e-06, "loss": 0.9158, "step": 6196 }, { "epoch": 0.5280783979548359, "grad_norm": 12.375, "learning_rate": 2.103132693753091e-06, "loss": 0.6206, "step": 6197 }, { "epoch": 0.5281636131231359, "grad_norm": 9.6875, "learning_rate": 2.103009872247754e-06, "loss": 0.2813, "step": 6198 }, { "epoch": 0.5282488282914358, "grad_norm": 13.625, "learning_rate": 2.1028870353276708e-06, "loss": 0.7611, "step": 6199 }, { "epoch": 0.5283340434597358, "grad_norm": 13.4375, "learning_rate": 2.1027641829950607e-06, "loss": 0.6344, "step": 6200 }, { "epoch": 0.5284192586280357, "grad_norm": 13.8125, "learning_rate": 2.102641315252145e-06, "loss": 0.5089, "step": 6201 }, { "epoch": 0.5285044737963357, "grad_norm": 11.1875, "learning_rate": 2.1025184321011427e-06, "loss": 0.7062, "step": 6202 }, { "epoch": 0.5285896889646357, "grad_norm": 12.4375, "learning_rate": 2.102395533544275e-06, "loss": 0.548, "step": 6203 }, { "epoch": 0.5286749041329356, "grad_norm": 10.125, "learning_rate": 2.1022726195837633e-06, "loss": 0.5138, "step": 6204 }, { "epoch": 0.5287601193012356, "grad_norm": 12.5, "learning_rate": 2.1021496902218284e-06, "loss": 0.557, "step": 6205 }, { "epoch": 0.5288453344695355, "grad_norm": 15.6875, "learning_rate": 2.1020267454606918e-06, "loss": 0.7341, "step": 6206 }, { "epoch": 0.5289305496378356, "grad_norm": 19.5, "learning_rate": 2.101903785302575e-06, "loss": 1.0213, "step": 6207 }, { "epoch": 0.5290157648061355, "grad_norm": 15.375, "learning_rate": 2.1017808097496996e-06, "loss": 0.6237, "step": 6208 }, { "epoch": 0.5291009799744355, "grad_norm": 13.25, "learning_rate": 2.101657818804289e-06, "loss": 0.6032, "step": 6209 }, { "epoch": 0.5291861951427355, "grad_norm": 23.5, "learning_rate": 2.1015348124685653e-06, "loss": 1.1062, "step": 6210 }, { "epoch": 0.5292714103110354, "grad_norm": 10.25, "learning_rate": 2.101411790744751e-06, "loss": 0.3556, "step": 6211 }, { "epoch": 0.5293566254793354, "grad_norm": 43.5, "learning_rate": 2.101288753635069e-06, "loss": 0.6076, "step": 6212 }, { "epoch": 0.5294418406476353, "grad_norm": 13.0625, "learning_rate": 2.101165701141744e-06, "loss": 0.9004, "step": 6213 }, { "epoch": 0.5295270558159353, "grad_norm": 13.375, "learning_rate": 2.1010426332669986e-06, "loss": 0.7381, "step": 6214 }, { "epoch": 0.5296122709842352, "grad_norm": 12.75, "learning_rate": 2.100919550013057e-06, "loss": 0.6725, "step": 6215 }, { "epoch": 0.5296974861525352, "grad_norm": 15.0, "learning_rate": 2.1007964513821434e-06, "loss": 0.624, "step": 6216 }, { "epoch": 0.5297827013208352, "grad_norm": 11.25, "learning_rate": 2.1006733373764822e-06, "loss": 0.4557, "step": 6217 }, { "epoch": 0.5298679164891351, "grad_norm": 14.4375, "learning_rate": 2.100550207998299e-06, "loss": 1.0441, "step": 6218 }, { "epoch": 0.5299531316574351, "grad_norm": 15.875, "learning_rate": 2.1004270632498174e-06, "loss": 0.9227, "step": 6219 }, { "epoch": 0.530038346825735, "grad_norm": 13.75, "learning_rate": 2.100303903133264e-06, "loss": 0.6257, "step": 6220 }, { "epoch": 0.530123561994035, "grad_norm": 18.875, "learning_rate": 2.1001807276508642e-06, "loss": 0.9663, "step": 6221 }, { "epoch": 0.5302087771623349, "grad_norm": 13.875, "learning_rate": 2.1000575368048436e-06, "loss": 0.6762, "step": 6222 }, { "epoch": 0.5302939923306349, "grad_norm": 13.9375, "learning_rate": 2.0999343305974287e-06, "loss": 1.0094, "step": 6223 }, { "epoch": 0.5303792074989349, "grad_norm": 14.875, "learning_rate": 2.0998111090308453e-06, "loss": 0.675, "step": 6224 }, { "epoch": 0.5304644226672348, "grad_norm": 19.0, "learning_rate": 2.0996878721073212e-06, "loss": 0.8701, "step": 6225 }, { "epoch": 0.5305496378355348, "grad_norm": 9.875, "learning_rate": 2.0995646198290825e-06, "loss": 0.4679, "step": 6226 }, { "epoch": 0.5306348530038347, "grad_norm": 17.25, "learning_rate": 2.0994413521983567e-06, "loss": 0.8929, "step": 6227 }, { "epoch": 0.5307200681721347, "grad_norm": 13.6875, "learning_rate": 2.0993180692173713e-06, "loss": 0.7287, "step": 6228 }, { "epoch": 0.5308052833404346, "grad_norm": 19.0, "learning_rate": 2.099194770888355e-06, "loss": 0.894, "step": 6229 }, { "epoch": 0.5308904985087346, "grad_norm": 16.125, "learning_rate": 2.099071457213535e-06, "loss": 1.2041, "step": 6230 }, { "epoch": 0.5309757136770346, "grad_norm": 18.125, "learning_rate": 2.0989481281951395e-06, "loss": 0.7065, "step": 6231 }, { "epoch": 0.5310609288453345, "grad_norm": 18.25, "learning_rate": 2.098824783835398e-06, "loss": 0.7604, "step": 6232 }, { "epoch": 0.5311461440136345, "grad_norm": 12.875, "learning_rate": 2.0987014241365396e-06, "loss": 0.727, "step": 6233 }, { "epoch": 0.5312313591819344, "grad_norm": 11.5, "learning_rate": 2.0985780491007923e-06, "loss": 0.5016, "step": 6234 }, { "epoch": 0.5313165743502344, "grad_norm": 17.0, "learning_rate": 2.0984546587303863e-06, "loss": 0.9232, "step": 6235 }, { "epoch": 0.5314017895185343, "grad_norm": 12.4375, "learning_rate": 2.098331253027552e-06, "loss": 0.5898, "step": 6236 }, { "epoch": 0.5314870046868343, "grad_norm": 14.75, "learning_rate": 2.0982078319945185e-06, "loss": 0.5249, "step": 6237 }, { "epoch": 0.5315722198551343, "grad_norm": 15.6875, "learning_rate": 2.098084395633517e-06, "loss": 0.6885, "step": 6238 }, { "epoch": 0.5316574350234342, "grad_norm": 10.6875, "learning_rate": 2.097960943946777e-06, "loss": 0.728, "step": 6239 }, { "epoch": 0.5317426501917342, "grad_norm": 17.375, "learning_rate": 2.09783747693653e-06, "loss": 0.9854, "step": 6240 }, { "epoch": 0.5318278653600341, "grad_norm": 15.5625, "learning_rate": 2.097713994605008e-06, "loss": 0.7116, "step": 6241 }, { "epoch": 0.5319130805283341, "grad_norm": 21.125, "learning_rate": 2.0975904969544413e-06, "loss": 0.9684, "step": 6242 }, { "epoch": 0.531998295696634, "grad_norm": 12.625, "learning_rate": 2.0974669839870615e-06, "loss": 0.666, "step": 6243 }, { "epoch": 0.532083510864934, "grad_norm": 16.25, "learning_rate": 2.0973434557051016e-06, "loss": 0.6634, "step": 6244 }, { "epoch": 0.532168726033234, "grad_norm": 13.75, "learning_rate": 2.0972199121107934e-06, "loss": 0.6413, "step": 6245 }, { "epoch": 0.5322539412015339, "grad_norm": 14.4375, "learning_rate": 2.0970963532063694e-06, "loss": 0.729, "step": 6246 }, { "epoch": 0.5323391563698339, "grad_norm": 9.625, "learning_rate": 2.0969727789940623e-06, "loss": 0.3331, "step": 6247 }, { "epoch": 0.5324243715381338, "grad_norm": 14.125, "learning_rate": 2.096849189476106e-06, "loss": 0.757, "step": 6248 }, { "epoch": 0.5325095867064338, "grad_norm": 12.5625, "learning_rate": 2.0967255846547326e-06, "loss": 0.686, "step": 6249 }, { "epoch": 0.5325948018747337, "grad_norm": 16.25, "learning_rate": 2.0966019645321767e-06, "loss": 0.9446, "step": 6250 }, { "epoch": 0.5326800170430337, "grad_norm": 14.3125, "learning_rate": 2.0964783291106714e-06, "loss": 0.616, "step": 6251 }, { "epoch": 0.5327652322113337, "grad_norm": 18.5, "learning_rate": 2.0963546783924525e-06, "loss": 0.9818, "step": 6252 }, { "epoch": 0.5328504473796336, "grad_norm": 12.9375, "learning_rate": 2.096231012379753e-06, "loss": 0.5134, "step": 6253 }, { "epoch": 0.5329356625479336, "grad_norm": 11.1875, "learning_rate": 2.096107331074808e-06, "loss": 0.4026, "step": 6254 }, { "epoch": 0.5330208777162335, "grad_norm": 17.25, "learning_rate": 2.0959836344798527e-06, "loss": 0.8716, "step": 6255 }, { "epoch": 0.5331060928845335, "grad_norm": 14.875, "learning_rate": 2.095859922597123e-06, "loss": 0.4164, "step": 6256 }, { "epoch": 0.5331913080528334, "grad_norm": 16.125, "learning_rate": 2.095736195428853e-06, "loss": 0.8771, "step": 6257 }, { "epoch": 0.5332765232211334, "grad_norm": 13.625, "learning_rate": 2.0956124529772805e-06, "loss": 0.8138, "step": 6258 }, { "epoch": 0.5333617383894333, "grad_norm": 16.625, "learning_rate": 2.09548869524464e-06, "loss": 1.078, "step": 6259 }, { "epoch": 0.5334469535577333, "grad_norm": 18.125, "learning_rate": 2.0953649222331686e-06, "loss": 1.2338, "step": 6260 }, { "epoch": 0.5335321687260333, "grad_norm": 12.9375, "learning_rate": 2.0952411339451027e-06, "loss": 0.4029, "step": 6261 }, { "epoch": 0.5336173838943332, "grad_norm": 13.6875, "learning_rate": 2.0951173303826806e-06, "loss": 0.5247, "step": 6262 }, { "epoch": 0.5337025990626332, "grad_norm": 11.5, "learning_rate": 2.0949935115481376e-06, "loss": 0.5547, "step": 6263 }, { "epoch": 0.5337878142309331, "grad_norm": 24.875, "learning_rate": 2.0948696774437123e-06, "loss": 0.9971, "step": 6264 }, { "epoch": 0.5338730293992331, "grad_norm": 13.125, "learning_rate": 2.0947458280716427e-06, "loss": 0.5835, "step": 6265 }, { "epoch": 0.533958244567533, "grad_norm": 13.75, "learning_rate": 2.094621963434166e-06, "loss": 0.8904, "step": 6266 }, { "epoch": 0.534043459735833, "grad_norm": 17.0, "learning_rate": 2.0944980835335214e-06, "loss": 0.9664, "step": 6267 }, { "epoch": 0.534128674904133, "grad_norm": 21.75, "learning_rate": 2.094374188371947e-06, "loss": 1.1452, "step": 6268 }, { "epoch": 0.5342138900724329, "grad_norm": 14.1875, "learning_rate": 2.0942502779516825e-06, "loss": 0.9308, "step": 6269 }, { "epoch": 0.5342991052407329, "grad_norm": 17.75, "learning_rate": 2.0941263522749662e-06, "loss": 0.4857, "step": 6270 }, { "epoch": 0.5343843204090328, "grad_norm": 13.8125, "learning_rate": 2.094002411344038e-06, "loss": 0.4573, "step": 6271 }, { "epoch": 0.5344695355773328, "grad_norm": 11.9375, "learning_rate": 2.093878455161137e-06, "loss": 0.644, "step": 6272 }, { "epoch": 0.5345547507456327, "grad_norm": 14.625, "learning_rate": 2.093754483728504e-06, "loss": 0.8602, "step": 6273 }, { "epoch": 0.5346399659139327, "grad_norm": 18.625, "learning_rate": 2.093630497048379e-06, "loss": 0.9489, "step": 6274 }, { "epoch": 0.5347251810822327, "grad_norm": 10.0, "learning_rate": 2.0935064951230026e-06, "loss": 0.3423, "step": 6275 }, { "epoch": 0.5348103962505326, "grad_norm": 12.375, "learning_rate": 2.093382477954616e-06, "loss": 0.6262, "step": 6276 }, { "epoch": 0.5348956114188326, "grad_norm": 16.0, "learning_rate": 2.0932584455454594e-06, "loss": 0.2831, "step": 6277 }, { "epoch": 0.5349808265871325, "grad_norm": 13.25, "learning_rate": 2.093134397897775e-06, "loss": 0.5382, "step": 6278 }, { "epoch": 0.5350660417554325, "grad_norm": 14.3125, "learning_rate": 2.093010335013804e-06, "loss": 0.6745, "step": 6279 }, { "epoch": 0.5351512569237324, "grad_norm": 30.25, "learning_rate": 2.0928862568957886e-06, "loss": 0.8815, "step": 6280 }, { "epoch": 0.5352364720920324, "grad_norm": 14.875, "learning_rate": 2.092762163545971e-06, "loss": 1.1011, "step": 6281 }, { "epoch": 0.5353216872603324, "grad_norm": 14.625, "learning_rate": 2.092638054966594e-06, "loss": 0.5702, "step": 6282 }, { "epoch": 0.5354069024286323, "grad_norm": 26.75, "learning_rate": 2.0925139311598996e-06, "loss": 0.805, "step": 6283 }, { "epoch": 0.5354921175969323, "grad_norm": 15.375, "learning_rate": 2.092389792128131e-06, "loss": 0.7906, "step": 6284 }, { "epoch": 0.5355773327652322, "grad_norm": 20.75, "learning_rate": 2.0922656378735324e-06, "loss": 0.9157, "step": 6285 }, { "epoch": 0.5356625479335322, "grad_norm": 19.375, "learning_rate": 2.0921414683983464e-06, "loss": 0.4856, "step": 6286 }, { "epoch": 0.5357477631018321, "grad_norm": 10.9375, "learning_rate": 2.092017283704817e-06, "loss": 0.4744, "step": 6287 }, { "epoch": 0.5358329782701321, "grad_norm": 13.375, "learning_rate": 2.091893083795189e-06, "loss": 0.4649, "step": 6288 }, { "epoch": 0.5359181934384321, "grad_norm": 13.4375, "learning_rate": 2.091768868671706e-06, "loss": 0.3992, "step": 6289 }, { "epoch": 0.536003408606732, "grad_norm": 24.625, "learning_rate": 2.091644638336614e-06, "loss": 1.0316, "step": 6290 }, { "epoch": 0.536088623775032, "grad_norm": 15.75, "learning_rate": 2.0915203927921563e-06, "loss": 0.856, "step": 6291 }, { "epoch": 0.5361738389433319, "grad_norm": 17.25, "learning_rate": 2.0913961320405783e-06, "loss": 1.061, "step": 6292 }, { "epoch": 0.5362590541116319, "grad_norm": 14.0625, "learning_rate": 2.091271856084127e-06, "loss": 0.5562, "step": 6293 }, { "epoch": 0.5363442692799318, "grad_norm": 22.125, "learning_rate": 2.091147564925047e-06, "loss": 1.3397, "step": 6294 }, { "epoch": 0.5364294844482318, "grad_norm": 14.4375, "learning_rate": 2.091023258565585e-06, "loss": 0.7106, "step": 6295 }, { "epoch": 0.5365146996165318, "grad_norm": 22.0, "learning_rate": 2.0908989370079867e-06, "loss": 0.6138, "step": 6296 }, { "epoch": 0.5365999147848317, "grad_norm": 29.625, "learning_rate": 2.090774600254499e-06, "loss": 1.3506, "step": 6297 }, { "epoch": 0.5366851299531317, "grad_norm": 12.625, "learning_rate": 2.090650248307369e-06, "loss": 0.6136, "step": 6298 }, { "epoch": 0.5367703451214316, "grad_norm": 12.25, "learning_rate": 2.0905258811688432e-06, "loss": 0.3978, "step": 6299 }, { "epoch": 0.5368555602897316, "grad_norm": 15.1875, "learning_rate": 2.09040149884117e-06, "loss": 0.7418, "step": 6300 }, { "epoch": 0.5369407754580315, "grad_norm": 12.5625, "learning_rate": 2.0902771013265964e-06, "loss": 0.5526, "step": 6301 }, { "epoch": 0.5370259906263315, "grad_norm": 12.1875, "learning_rate": 2.090152688627371e-06, "loss": 0.6491, "step": 6302 }, { "epoch": 0.5371112057946315, "grad_norm": 14.0, "learning_rate": 2.0900282607457413e-06, "loss": 0.7835, "step": 6303 }, { "epoch": 0.5371964209629314, "grad_norm": 13.0, "learning_rate": 2.089903817683956e-06, "loss": 0.3832, "step": 6304 }, { "epoch": 0.5372816361312314, "grad_norm": 20.125, "learning_rate": 2.089779359444264e-06, "loss": 0.9013, "step": 6305 }, { "epoch": 0.5373668512995313, "grad_norm": 20.5, "learning_rate": 2.0896548860289153e-06, "loss": 0.9356, "step": 6306 }, { "epoch": 0.5374520664678313, "grad_norm": 14.4375, "learning_rate": 2.0895303974401582e-06, "loss": 0.5911, "step": 6307 }, { "epoch": 0.5375372816361312, "grad_norm": 22.625, "learning_rate": 2.0894058936802424e-06, "loss": 1.011, "step": 6308 }, { "epoch": 0.5376224968044312, "grad_norm": 15.125, "learning_rate": 2.089281374751418e-06, "loss": 0.855, "step": 6309 }, { "epoch": 0.5377077119727312, "grad_norm": 14.6875, "learning_rate": 2.0891568406559355e-06, "loss": 0.8929, "step": 6310 }, { "epoch": 0.5377929271410311, "grad_norm": 11.5, "learning_rate": 2.089032291396045e-06, "loss": 0.3536, "step": 6311 }, { "epoch": 0.5378781423093311, "grad_norm": 14.6875, "learning_rate": 2.0889077269739967e-06, "loss": 0.6752, "step": 6312 }, { "epoch": 0.537963357477631, "grad_norm": 20.375, "learning_rate": 2.0887831473920426e-06, "loss": 1.3051, "step": 6313 }, { "epoch": 0.538048572645931, "grad_norm": 14.375, "learning_rate": 2.0886585526524335e-06, "loss": 0.6454, "step": 6314 }, { "epoch": 0.5381337878142309, "grad_norm": 13.0625, "learning_rate": 2.088533942757421e-06, "loss": 0.4837, "step": 6315 }, { "epoch": 0.5382190029825309, "grad_norm": 22.75, "learning_rate": 2.0884093177092575e-06, "loss": 0.9437, "step": 6316 }, { "epoch": 0.5383042181508308, "grad_norm": 29.75, "learning_rate": 2.088284677510194e-06, "loss": 1.1829, "step": 6317 }, { "epoch": 0.5383894333191308, "grad_norm": 21.625, "learning_rate": 2.088160022162483e-06, "loss": 0.9178, "step": 6318 }, { "epoch": 0.5384746484874308, "grad_norm": 11.9375, "learning_rate": 2.0880353516683782e-06, "loss": 0.5022, "step": 6319 }, { "epoch": 0.5385598636557307, "grad_norm": 15.4375, "learning_rate": 2.087910666030132e-06, "loss": 0.7753, "step": 6320 }, { "epoch": 0.5386450788240307, "grad_norm": 14.5, "learning_rate": 2.087785965249997e-06, "loss": 0.7404, "step": 6321 }, { "epoch": 0.5387302939923306, "grad_norm": 15.625, "learning_rate": 2.087661249330227e-06, "loss": 0.6213, "step": 6322 }, { "epoch": 0.5388155091606306, "grad_norm": 14.5, "learning_rate": 2.0875365182730767e-06, "loss": 0.596, "step": 6323 }, { "epoch": 0.5389007243289305, "grad_norm": 18.25, "learning_rate": 2.087411772080799e-06, "loss": 1.1188, "step": 6324 }, { "epoch": 0.5389859394972305, "grad_norm": 17.0, "learning_rate": 2.0872870107556483e-06, "loss": 0.776, "step": 6325 }, { "epoch": 0.5390711546655305, "grad_norm": 22.75, "learning_rate": 2.0871622342998794e-06, "loss": 1.0287, "step": 6326 }, { "epoch": 0.5391563698338304, "grad_norm": 10.625, "learning_rate": 2.087037442715747e-06, "loss": 0.5038, "step": 6327 }, { "epoch": 0.5392415850021304, "grad_norm": 15.375, "learning_rate": 2.0869126360055065e-06, "loss": 0.9662, "step": 6328 }, { "epoch": 0.5393268001704303, "grad_norm": 10.75, "learning_rate": 2.086787814171413e-06, "loss": 0.4304, "step": 6329 }, { "epoch": 0.5394120153387303, "grad_norm": 13.3125, "learning_rate": 2.086662977215722e-06, "loss": 0.6223, "step": 6330 }, { "epoch": 0.5394972305070302, "grad_norm": 9.1875, "learning_rate": 2.08653812514069e-06, "loss": 0.3308, "step": 6331 }, { "epoch": 0.5395824456753302, "grad_norm": 15.625, "learning_rate": 2.086413257948573e-06, "loss": 0.7595, "step": 6332 }, { "epoch": 0.5396676608436302, "grad_norm": 14.5, "learning_rate": 2.086288375641627e-06, "loss": 0.8933, "step": 6333 }, { "epoch": 0.5397528760119301, "grad_norm": 12.5625, "learning_rate": 2.0861634782221093e-06, "loss": 0.7803, "step": 6334 }, { "epoch": 0.5398380911802301, "grad_norm": 11.8125, "learning_rate": 2.0860385656922767e-06, "loss": 0.4623, "step": 6335 }, { "epoch": 0.53992330634853, "grad_norm": 12.6875, "learning_rate": 2.0859136380543866e-06, "loss": 0.4133, "step": 6336 }, { "epoch": 0.54000852151683, "grad_norm": 13.875, "learning_rate": 2.0857886953106965e-06, "loss": 0.7697, "step": 6337 }, { "epoch": 0.5400937366851299, "grad_norm": 13.875, "learning_rate": 2.085663737463464e-06, "loss": 0.7458, "step": 6338 }, { "epoch": 0.5401789518534299, "grad_norm": 13.4375, "learning_rate": 2.0855387645149475e-06, "loss": 0.5537, "step": 6339 }, { "epoch": 0.5402641670217299, "grad_norm": 13.5625, "learning_rate": 2.0854137764674055e-06, "loss": 0.4072, "step": 6340 }, { "epoch": 0.5403493821900298, "grad_norm": 11.625, "learning_rate": 2.085288773323096e-06, "loss": 0.5515, "step": 6341 }, { "epoch": 0.5404345973583298, "grad_norm": 16.375, "learning_rate": 2.085163755084279e-06, "loss": 0.6574, "step": 6342 }, { "epoch": 0.5405198125266297, "grad_norm": 17.25, "learning_rate": 2.0850387217532127e-06, "loss": 0.8315, "step": 6343 }, { "epoch": 0.5406050276949297, "grad_norm": 21.125, "learning_rate": 2.0849136733321573e-06, "loss": 1.0187, "step": 6344 }, { "epoch": 0.5406902428632296, "grad_norm": 15.6875, "learning_rate": 2.0847886098233724e-06, "loss": 0.7853, "step": 6345 }, { "epoch": 0.5407754580315296, "grad_norm": 12.0625, "learning_rate": 2.084663531229117e-06, "loss": 0.6363, "step": 6346 }, { "epoch": 0.5408606731998296, "grad_norm": 20.75, "learning_rate": 2.084538437551653e-06, "loss": 1.0188, "step": 6347 }, { "epoch": 0.5409458883681295, "grad_norm": 13.1875, "learning_rate": 2.0844133287932404e-06, "loss": 0.7096, "step": 6348 }, { "epoch": 0.5410311035364295, "grad_norm": 11.6875, "learning_rate": 2.0842882049561398e-06, "loss": 0.6152, "step": 6349 }, { "epoch": 0.5411163187047294, "grad_norm": 14.375, "learning_rate": 2.0841630660426116e-06, "loss": 0.9194, "step": 6350 }, { "epoch": 0.5412015338730294, "grad_norm": 13.4375, "learning_rate": 2.084037912054919e-06, "loss": 0.887, "step": 6351 }, { "epoch": 0.5412867490413293, "grad_norm": 11.8125, "learning_rate": 2.083912742995322e-06, "loss": 0.6242, "step": 6352 }, { "epoch": 0.5413719642096293, "grad_norm": 12.0625, "learning_rate": 2.0837875588660834e-06, "loss": 0.7515, "step": 6353 }, { "epoch": 0.5414571793779293, "grad_norm": 14.125, "learning_rate": 2.083662359669465e-06, "loss": 0.6306, "step": 6354 }, { "epoch": 0.5415423945462292, "grad_norm": 11.875, "learning_rate": 2.08353714540773e-06, "loss": 0.5489, "step": 6355 }, { "epoch": 0.5416276097145292, "grad_norm": 13.0625, "learning_rate": 2.08341191608314e-06, "loss": 0.6301, "step": 6356 }, { "epoch": 0.5417128248828291, "grad_norm": 14.625, "learning_rate": 2.083286671697959e-06, "loss": 1.1312, "step": 6357 }, { "epoch": 0.5417980400511291, "grad_norm": 15.125, "learning_rate": 2.0831614122544496e-06, "loss": 0.8191, "step": 6358 }, { "epoch": 0.541883255219429, "grad_norm": 17.125, "learning_rate": 2.0830361377548757e-06, "loss": 0.7818, "step": 6359 }, { "epoch": 0.541968470387729, "grad_norm": 10.4375, "learning_rate": 2.0829108482015015e-06, "loss": 0.2524, "step": 6360 }, { "epoch": 0.542053685556029, "grad_norm": 25.875, "learning_rate": 2.0827855435965907e-06, "loss": 1.318, "step": 6361 }, { "epoch": 0.5421389007243289, "grad_norm": 17.75, "learning_rate": 2.082660223942407e-06, "loss": 0.9532, "step": 6362 }, { "epoch": 0.5422241158926289, "grad_norm": 14.1875, "learning_rate": 2.082534889241216e-06, "loss": 0.8932, "step": 6363 }, { "epoch": 0.5423093310609288, "grad_norm": 15.5625, "learning_rate": 2.0824095394952828e-06, "loss": 0.8502, "step": 6364 }, { "epoch": 0.5423945462292288, "grad_norm": 12.1875, "learning_rate": 2.082284174706872e-06, "loss": 0.6244, "step": 6365 }, { "epoch": 0.5424797613975287, "grad_norm": 18.0, "learning_rate": 2.082158794878249e-06, "loss": 0.9709, "step": 6366 }, { "epoch": 0.5425649765658287, "grad_norm": 19.25, "learning_rate": 2.0820334000116803e-06, "loss": 1.0619, "step": 6367 }, { "epoch": 0.5426501917341287, "grad_norm": 13.75, "learning_rate": 2.0819079901094306e-06, "loss": 0.7625, "step": 6368 }, { "epoch": 0.5427354069024286, "grad_norm": 13.6875, "learning_rate": 2.0817825651737674e-06, "loss": 0.9346, "step": 6369 }, { "epoch": 0.5428206220707286, "grad_norm": 15.0625, "learning_rate": 2.0816571252069565e-06, "loss": 0.6954, "step": 6370 }, { "epoch": 0.5429058372390285, "grad_norm": 14.75, "learning_rate": 2.081531670211266e-06, "loss": 0.8664, "step": 6371 }, { "epoch": 0.5429910524073285, "grad_norm": 15.875, "learning_rate": 2.0814062001889613e-06, "loss": 0.775, "step": 6372 }, { "epoch": 0.5430762675756284, "grad_norm": 14.3125, "learning_rate": 2.0812807151423103e-06, "loss": 0.7374, "step": 6373 }, { "epoch": 0.5431614827439284, "grad_norm": 17.5, "learning_rate": 2.0811552150735813e-06, "loss": 0.9329, "step": 6374 }, { "epoch": 0.5432466979122283, "grad_norm": 14.3125, "learning_rate": 2.081029699985042e-06, "loss": 0.8339, "step": 6375 }, { "epoch": 0.5433319130805283, "grad_norm": 18.125, "learning_rate": 2.0809041698789598e-06, "loss": 0.7687, "step": 6376 }, { "epoch": 0.5434171282488283, "grad_norm": 14.1875, "learning_rate": 2.080778624757604e-06, "loss": 0.7427, "step": 6377 }, { "epoch": 0.5435023434171282, "grad_norm": 15.5625, "learning_rate": 2.080653064623243e-06, "loss": 0.908, "step": 6378 }, { "epoch": 0.5435875585854282, "grad_norm": 17.875, "learning_rate": 2.0805274894781458e-06, "loss": 0.9505, "step": 6379 }, { "epoch": 0.5436727737537281, "grad_norm": 10.1875, "learning_rate": 2.0804018993245816e-06, "loss": 0.307, "step": 6380 }, { "epoch": 0.5437579889220281, "grad_norm": 15.5625, "learning_rate": 2.0802762941648203e-06, "loss": 1.0888, "step": 6381 }, { "epoch": 0.543843204090328, "grad_norm": 14.6875, "learning_rate": 2.0801506740011312e-06, "loss": 0.9752, "step": 6382 }, { "epoch": 0.543928419258628, "grad_norm": 14.625, "learning_rate": 2.0800250388357847e-06, "loss": 0.7256, "step": 6383 }, { "epoch": 0.544013634426928, "grad_norm": 17.375, "learning_rate": 2.079899388671051e-06, "loss": 1.1044, "step": 6384 }, { "epoch": 0.5440988495952279, "grad_norm": 12.6875, "learning_rate": 2.0797737235092015e-06, "loss": 0.6349, "step": 6385 }, { "epoch": 0.5441840647635279, "grad_norm": 16.0, "learning_rate": 2.0796480433525057e-06, "loss": 0.7979, "step": 6386 }, { "epoch": 0.5442692799318278, "grad_norm": 66.5, "learning_rate": 2.0795223482032357e-06, "loss": 1.5932, "step": 6387 }, { "epoch": 0.5443544951001278, "grad_norm": 12.625, "learning_rate": 2.079396638063663e-06, "loss": 0.7549, "step": 6388 }, { "epoch": 0.5444397102684277, "grad_norm": 23.875, "learning_rate": 2.0792709129360583e-06, "loss": 1.1931, "step": 6389 }, { "epoch": 0.5445249254367277, "grad_norm": 13.0625, "learning_rate": 2.079145172822695e-06, "loss": 0.7229, "step": 6390 }, { "epoch": 0.5446101406050277, "grad_norm": 16.375, "learning_rate": 2.0790194177258436e-06, "loss": 0.7531, "step": 6391 }, { "epoch": 0.5446953557733276, "grad_norm": 14.0625, "learning_rate": 2.078893647647779e-06, "loss": 0.759, "step": 6392 }, { "epoch": 0.5447805709416276, "grad_norm": 13.8125, "learning_rate": 2.078767862590772e-06, "loss": 0.6109, "step": 6393 }, { "epoch": 0.5448657861099275, "grad_norm": 14.1875, "learning_rate": 2.0786420625570963e-06, "loss": 0.9057, "step": 6394 }, { "epoch": 0.5449510012782275, "grad_norm": 17.375, "learning_rate": 2.0785162475490254e-06, "loss": 0.8169, "step": 6395 }, { "epoch": 0.5450362164465274, "grad_norm": 19.875, "learning_rate": 2.078390417568833e-06, "loss": 0.3474, "step": 6396 }, { "epoch": 0.5451214316148274, "grad_norm": 25.375, "learning_rate": 2.0782645726187926e-06, "loss": 0.8701, "step": 6397 }, { "epoch": 0.5452066467831274, "grad_norm": 13.9375, "learning_rate": 2.078138712701178e-06, "loss": 0.5955, "step": 6398 }, { "epoch": 0.5452918619514273, "grad_norm": 13.75, "learning_rate": 2.0780128378182645e-06, "loss": 0.6416, "step": 6399 }, { "epoch": 0.5453770771197273, "grad_norm": 14.375, "learning_rate": 2.0778869479723267e-06, "loss": 0.9216, "step": 6400 }, { "epoch": 0.5454622922880272, "grad_norm": 16.25, "learning_rate": 2.077761043165639e-06, "loss": 0.8919, "step": 6401 }, { "epoch": 0.5455475074563272, "grad_norm": 13.75, "learning_rate": 2.077635123400477e-06, "loss": 0.7583, "step": 6402 }, { "epoch": 0.5456327226246271, "grad_norm": 14.6875, "learning_rate": 2.0775091886791158e-06, "loss": 0.7405, "step": 6403 }, { "epoch": 0.5457179377929271, "grad_norm": 19.5, "learning_rate": 2.0773832390038316e-06, "loss": 0.8922, "step": 6404 }, { "epoch": 0.5458031529612271, "grad_norm": 21.375, "learning_rate": 2.0772572743769002e-06, "loss": 1.1329, "step": 6405 }, { "epoch": 0.545888368129527, "grad_norm": 15.25, "learning_rate": 2.077131294800598e-06, "loss": 0.904, "step": 6406 }, { "epoch": 0.545973583297827, "grad_norm": 18.0, "learning_rate": 2.0770053002772016e-06, "loss": 0.8525, "step": 6407 }, { "epoch": 0.5460587984661269, "grad_norm": 14.8125, "learning_rate": 2.0768792908089876e-06, "loss": 0.5698, "step": 6408 }, { "epoch": 0.5461440136344269, "grad_norm": 18.25, "learning_rate": 2.0767532663982332e-06, "loss": 0.9145, "step": 6409 }, { "epoch": 0.5462292288027268, "grad_norm": 14.9375, "learning_rate": 2.0766272270472165e-06, "loss": 0.9785, "step": 6410 }, { "epoch": 0.5463144439710268, "grad_norm": 13.875, "learning_rate": 2.076501172758214e-06, "loss": 0.7995, "step": 6411 }, { "epoch": 0.5463996591393268, "grad_norm": 18.0, "learning_rate": 2.0763751035335047e-06, "loss": 0.9431, "step": 6412 }, { "epoch": 0.5464848743076267, "grad_norm": 17.375, "learning_rate": 2.076249019375366e-06, "loss": 0.8848, "step": 6413 }, { "epoch": 0.5465700894759267, "grad_norm": 12.8125, "learning_rate": 2.0761229202860765e-06, "loss": 0.7104, "step": 6414 }, { "epoch": 0.5466553046442266, "grad_norm": 20.375, "learning_rate": 2.0759968062679154e-06, "loss": 0.9082, "step": 6415 }, { "epoch": 0.5467405198125266, "grad_norm": 19.25, "learning_rate": 2.0758706773231614e-06, "loss": 0.8975, "step": 6416 }, { "epoch": 0.5468257349808265, "grad_norm": 19.0, "learning_rate": 2.075744533454094e-06, "loss": 0.9932, "step": 6417 }, { "epoch": 0.5469109501491265, "grad_norm": 15.0625, "learning_rate": 2.075618374662992e-06, "loss": 0.8165, "step": 6418 }, { "epoch": 0.5469961653174265, "grad_norm": 10.5625, "learning_rate": 2.0754922009521357e-06, "loss": 0.4637, "step": 6419 }, { "epoch": 0.5470813804857264, "grad_norm": 31.5, "learning_rate": 2.075366012323806e-06, "loss": 0.9428, "step": 6420 }, { "epoch": 0.5471665956540264, "grad_norm": 29.5, "learning_rate": 2.0752398087802817e-06, "loss": 0.8586, "step": 6421 }, { "epoch": 0.5472518108223263, "grad_norm": 17.5, "learning_rate": 2.0751135903238446e-06, "loss": 0.2941, "step": 6422 }, { "epoch": 0.5473370259906263, "grad_norm": 14.5625, "learning_rate": 2.0749873569567753e-06, "loss": 0.6178, "step": 6423 }, { "epoch": 0.5474222411589262, "grad_norm": 12.1875, "learning_rate": 2.0748611086813548e-06, "loss": 0.5331, "step": 6424 }, { "epoch": 0.5475074563272262, "grad_norm": 15.125, "learning_rate": 2.074734845499864e-06, "loss": 0.7394, "step": 6425 }, { "epoch": 0.5475926714955262, "grad_norm": 15.25, "learning_rate": 2.0746085674145856e-06, "loss": 0.8424, "step": 6426 }, { "epoch": 0.5476778866638261, "grad_norm": 20.0, "learning_rate": 2.0744822744278017e-06, "loss": 1.0727, "step": 6427 }, { "epoch": 0.5477631018321261, "grad_norm": 19.125, "learning_rate": 2.0743559665417933e-06, "loss": 0.4943, "step": 6428 }, { "epoch": 0.547848317000426, "grad_norm": 16.5, "learning_rate": 2.074229643758844e-06, "loss": 0.9638, "step": 6429 }, { "epoch": 0.547933532168726, "grad_norm": 14.875, "learning_rate": 2.074103306081236e-06, "loss": 0.9248, "step": 6430 }, { "epoch": 0.5480187473370259, "grad_norm": 13.375, "learning_rate": 2.0739769535112527e-06, "loss": 0.6848, "step": 6431 }, { "epoch": 0.548103962505326, "grad_norm": 16.125, "learning_rate": 2.0738505860511776e-06, "loss": 0.7043, "step": 6432 }, { "epoch": 0.548189177673626, "grad_norm": 16.625, "learning_rate": 2.0737242037032935e-06, "loss": 0.7008, "step": 6433 }, { "epoch": 0.5482743928419259, "grad_norm": 12.125, "learning_rate": 2.0735978064698846e-06, "loss": 0.6398, "step": 6434 }, { "epoch": 0.5483596080102259, "grad_norm": 14.75, "learning_rate": 2.0734713943532355e-06, "loss": 0.8238, "step": 6435 }, { "epoch": 0.5484448231785258, "grad_norm": 15.0625, "learning_rate": 2.0733449673556305e-06, "loss": 0.7854, "step": 6436 }, { "epoch": 0.5485300383468258, "grad_norm": 13.9375, "learning_rate": 2.0732185254793534e-06, "loss": 1.0776, "step": 6437 }, { "epoch": 0.5486152535151257, "grad_norm": 15.3125, "learning_rate": 2.07309206872669e-06, "loss": 0.8542, "step": 6438 }, { "epoch": 0.5487004686834257, "grad_norm": 10.875, "learning_rate": 2.0729655970999252e-06, "loss": 0.4787, "step": 6439 }, { "epoch": 0.5487856838517257, "grad_norm": 16.25, "learning_rate": 2.0728391106013448e-06, "loss": 0.6915, "step": 6440 }, { "epoch": 0.5488708990200256, "grad_norm": 17.125, "learning_rate": 2.072712609233234e-06, "loss": 0.864, "step": 6441 }, { "epoch": 0.5489561141883256, "grad_norm": 19.125, "learning_rate": 2.0725860929978787e-06, "loss": 0.6895, "step": 6442 }, { "epoch": 0.5490413293566255, "grad_norm": 12.625, "learning_rate": 2.072459561897566e-06, "loss": 0.5052, "step": 6443 }, { "epoch": 0.5491265445249255, "grad_norm": 15.25, "learning_rate": 2.0723330159345815e-06, "loss": 0.9757, "step": 6444 }, { "epoch": 0.5492117596932254, "grad_norm": 16.25, "learning_rate": 2.072206455111213e-06, "loss": 0.8129, "step": 6445 }, { "epoch": 0.5492969748615254, "grad_norm": 17.5, "learning_rate": 2.072079879429747e-06, "loss": 0.8724, "step": 6446 }, { "epoch": 0.5493821900298254, "grad_norm": 12.6875, "learning_rate": 2.0719532888924705e-06, "loss": 0.5526, "step": 6447 }, { "epoch": 0.5494674051981253, "grad_norm": 16.5, "learning_rate": 2.071826683501672e-06, "loss": 0.9774, "step": 6448 }, { "epoch": 0.5495526203664253, "grad_norm": 10.3125, "learning_rate": 2.071700063259639e-06, "loss": 0.4172, "step": 6449 }, { "epoch": 0.5496378355347252, "grad_norm": 15.1875, "learning_rate": 2.071573428168659e-06, "loss": 0.3816, "step": 6450 }, { "epoch": 0.5497230507030252, "grad_norm": 16.0, "learning_rate": 2.071446778231021e-06, "loss": 0.7898, "step": 6451 }, { "epoch": 0.5498082658713251, "grad_norm": 11.25, "learning_rate": 2.0713201134490143e-06, "loss": 0.5958, "step": 6452 }, { "epoch": 0.5498934810396251, "grad_norm": 24.375, "learning_rate": 2.0711934338249266e-06, "loss": 1.0397, "step": 6453 }, { "epoch": 0.549978696207925, "grad_norm": 14.4375, "learning_rate": 2.0710667393610483e-06, "loss": 0.8518, "step": 6454 }, { "epoch": 0.550063911376225, "grad_norm": 14.9375, "learning_rate": 2.070940030059668e-06, "loss": 0.7177, "step": 6455 }, { "epoch": 0.550149126544525, "grad_norm": 14.0625, "learning_rate": 2.070813305923076e-06, "loss": 1.0817, "step": 6456 }, { "epoch": 0.5502343417128249, "grad_norm": 16.75, "learning_rate": 2.0706865669535627e-06, "loss": 0.9834, "step": 6457 }, { "epoch": 0.5503195568811249, "grad_norm": 15.3125, "learning_rate": 2.070559813153417e-06, "loss": 0.5247, "step": 6458 }, { "epoch": 0.5504047720494248, "grad_norm": 17.875, "learning_rate": 2.0704330445249306e-06, "loss": 0.5587, "step": 6459 }, { "epoch": 0.5504899872177248, "grad_norm": 14.5625, "learning_rate": 2.0703062610703944e-06, "loss": 0.6146, "step": 6460 }, { "epoch": 0.5505752023860248, "grad_norm": 16.75, "learning_rate": 2.070179462792099e-06, "loss": 0.3817, "step": 6461 }, { "epoch": 0.5506604175543247, "grad_norm": 18.25, "learning_rate": 2.0700526496923355e-06, "loss": 1.0213, "step": 6462 }, { "epoch": 0.5507456327226247, "grad_norm": 19.375, "learning_rate": 2.069925821773397e-06, "loss": 1.2363, "step": 6463 }, { "epoch": 0.5508308478909246, "grad_norm": 16.5, "learning_rate": 2.0697989790375734e-06, "loss": 0.8194, "step": 6464 }, { "epoch": 0.5509160630592246, "grad_norm": 17.125, "learning_rate": 2.0696721214871586e-06, "loss": 0.3935, "step": 6465 }, { "epoch": 0.5510012782275245, "grad_norm": 14.1875, "learning_rate": 2.069545249124444e-06, "loss": 0.5125, "step": 6466 }, { "epoch": 0.5510864933958245, "grad_norm": 10.6875, "learning_rate": 2.069418361951722e-06, "loss": 0.4013, "step": 6467 }, { "epoch": 0.5511717085641245, "grad_norm": 13.6875, "learning_rate": 2.069291459971287e-06, "loss": 0.5735, "step": 6468 }, { "epoch": 0.5512569237324244, "grad_norm": 19.5, "learning_rate": 2.069164543185432e-06, "loss": 0.5107, "step": 6469 }, { "epoch": 0.5513421389007244, "grad_norm": 24.125, "learning_rate": 2.069037611596449e-06, "loss": 0.5876, "step": 6470 }, { "epoch": 0.5514273540690243, "grad_norm": 16.25, "learning_rate": 2.0689106652066327e-06, "loss": 0.5836, "step": 6471 }, { "epoch": 0.5515125692373243, "grad_norm": 35.0, "learning_rate": 2.0687837040182775e-06, "loss": 1.0261, "step": 6472 }, { "epoch": 0.5515977844056242, "grad_norm": 17.375, "learning_rate": 2.068656728033678e-06, "loss": 0.525, "step": 6473 }, { "epoch": 0.5516829995739242, "grad_norm": 13.1875, "learning_rate": 2.068529737255127e-06, "loss": 0.2721, "step": 6474 }, { "epoch": 0.5517682147422242, "grad_norm": 11.0, "learning_rate": 2.0684027316849212e-06, "loss": 0.371, "step": 6475 }, { "epoch": 0.5518534299105241, "grad_norm": 11.0625, "learning_rate": 2.0682757113253554e-06, "loss": 0.374, "step": 6476 }, { "epoch": 0.5519386450788241, "grad_norm": 17.75, "learning_rate": 2.068148676178724e-06, "loss": 0.8038, "step": 6477 }, { "epoch": 0.552023860247124, "grad_norm": 17.0, "learning_rate": 2.0680216262473237e-06, "loss": 0.8735, "step": 6478 }, { "epoch": 0.552109075415424, "grad_norm": 13.6875, "learning_rate": 2.06789456153345e-06, "loss": 0.7367, "step": 6479 }, { "epoch": 0.5521942905837239, "grad_norm": 10.0, "learning_rate": 2.067767482039399e-06, "loss": 0.302, "step": 6480 }, { "epoch": 0.5522795057520239, "grad_norm": 14.1875, "learning_rate": 2.0676403877674677e-06, "loss": 0.9204, "step": 6481 }, { "epoch": 0.5523647209203238, "grad_norm": 17.5, "learning_rate": 2.067513278719952e-06, "loss": 0.6165, "step": 6482 }, { "epoch": 0.5524499360886238, "grad_norm": 25.125, "learning_rate": 2.067386154899149e-06, "loss": 0.969, "step": 6483 }, { "epoch": 0.5525351512569238, "grad_norm": 17.875, "learning_rate": 2.0672590163073566e-06, "loss": 0.919, "step": 6484 }, { "epoch": 0.5526203664252237, "grad_norm": 14.0, "learning_rate": 2.067131862946872e-06, "loss": 0.6028, "step": 6485 }, { "epoch": 0.5527055815935237, "grad_norm": 18.625, "learning_rate": 2.067004694819993e-06, "loss": 0.8, "step": 6486 }, { "epoch": 0.5527907967618236, "grad_norm": 13.5, "learning_rate": 2.0668775119290176e-06, "loss": 0.6528, "step": 6487 }, { "epoch": 0.5528760119301236, "grad_norm": 18.125, "learning_rate": 2.0667503142762435e-06, "loss": 0.7239, "step": 6488 }, { "epoch": 0.5529612270984235, "grad_norm": 12.125, "learning_rate": 2.0666231018639707e-06, "loss": 0.7263, "step": 6489 }, { "epoch": 0.5530464422667235, "grad_norm": 14.9375, "learning_rate": 2.066495874694497e-06, "loss": 0.6555, "step": 6490 }, { "epoch": 0.5531316574350235, "grad_norm": 14.6875, "learning_rate": 2.0663686327701215e-06, "loss": 0.5581, "step": 6491 }, { "epoch": 0.5532168726033234, "grad_norm": 13.0625, "learning_rate": 2.0662413760931444e-06, "loss": 0.7057, "step": 6492 }, { "epoch": 0.5533020877716234, "grad_norm": 12.8125, "learning_rate": 2.0661141046658644e-06, "loss": 0.528, "step": 6493 }, { "epoch": 0.5533873029399233, "grad_norm": 10.5625, "learning_rate": 2.0659868184905825e-06, "loss": 0.554, "step": 6494 }, { "epoch": 0.5534725181082233, "grad_norm": 15.0625, "learning_rate": 2.0658595175695975e-06, "loss": 0.9603, "step": 6495 }, { "epoch": 0.5535577332765232, "grad_norm": 11.5, "learning_rate": 2.0657322019052108e-06, "loss": 0.4121, "step": 6496 }, { "epoch": 0.5536429484448232, "grad_norm": 10.5, "learning_rate": 2.0656048714997233e-06, "loss": 0.4158, "step": 6497 }, { "epoch": 0.5537281636131232, "grad_norm": 18.75, "learning_rate": 2.0654775263554353e-06, "loss": 0.738, "step": 6498 }, { "epoch": 0.5538133787814231, "grad_norm": 18.5, "learning_rate": 2.0653501664746485e-06, "loss": 0.8041, "step": 6499 }, { "epoch": 0.5538985939497231, "grad_norm": 13.3125, "learning_rate": 2.065222791859664e-06, "loss": 0.5907, "step": 6500 }, { "epoch": 0.553983809118023, "grad_norm": 13.125, "learning_rate": 2.0650954025127844e-06, "loss": 0.5922, "step": 6501 }, { "epoch": 0.554069024286323, "grad_norm": 16.0, "learning_rate": 2.0649679984363108e-06, "loss": 0.8694, "step": 6502 }, { "epoch": 0.5541542394546229, "grad_norm": 14.0, "learning_rate": 2.064840579632546e-06, "loss": 0.5095, "step": 6503 }, { "epoch": 0.5542394546229229, "grad_norm": 21.0, "learning_rate": 2.064713146103793e-06, "loss": 0.5398, "step": 6504 }, { "epoch": 0.5543246697912229, "grad_norm": 17.75, "learning_rate": 2.064585697852354e-06, "loss": 0.7407, "step": 6505 }, { "epoch": 0.5544098849595228, "grad_norm": 14.3125, "learning_rate": 2.0644582348805316e-06, "loss": 0.6482, "step": 6506 }, { "epoch": 0.5544951001278228, "grad_norm": 22.0, "learning_rate": 2.0643307571906306e-06, "loss": 1.1116, "step": 6507 }, { "epoch": 0.5545803152961227, "grad_norm": 10.75, "learning_rate": 2.0642032647849535e-06, "loss": 0.3904, "step": 6508 }, { "epoch": 0.5546655304644227, "grad_norm": 16.5, "learning_rate": 2.064075757665805e-06, "loss": 0.6855, "step": 6509 }, { "epoch": 0.5547507456327226, "grad_norm": 13.125, "learning_rate": 2.063948235835489e-06, "loss": 0.56, "step": 6510 }, { "epoch": 0.5548359608010226, "grad_norm": 17.75, "learning_rate": 2.0638206992963098e-06, "loss": 1.0035, "step": 6511 }, { "epoch": 0.5549211759693226, "grad_norm": 16.75, "learning_rate": 2.063693148050572e-06, "loss": 0.7331, "step": 6512 }, { "epoch": 0.5550063911376225, "grad_norm": 15.3125, "learning_rate": 2.0635655821005813e-06, "loss": 0.8202, "step": 6513 }, { "epoch": 0.5550916063059225, "grad_norm": 12.375, "learning_rate": 2.063438001448642e-06, "loss": 0.4944, "step": 6514 }, { "epoch": 0.5551768214742224, "grad_norm": 12.25, "learning_rate": 2.06331040609706e-06, "loss": 0.6298, "step": 6515 }, { "epoch": 0.5552620366425224, "grad_norm": 13.6875, "learning_rate": 2.0631827960481414e-06, "loss": 0.8438, "step": 6516 }, { "epoch": 0.5553472518108223, "grad_norm": 20.25, "learning_rate": 2.0630551713041915e-06, "loss": 0.7935, "step": 6517 }, { "epoch": 0.5554324669791223, "grad_norm": 22.375, "learning_rate": 2.062927531867518e-06, "loss": 1.115, "step": 6518 }, { "epoch": 0.5555176821474223, "grad_norm": 28.25, "learning_rate": 2.0627998777404252e-06, "loss": 1.2354, "step": 6519 }, { "epoch": 0.5556028973157222, "grad_norm": 15.8125, "learning_rate": 2.0626722089252223e-06, "loss": 1.0337, "step": 6520 }, { "epoch": 0.5556881124840222, "grad_norm": 12.875, "learning_rate": 2.062544525424215e-06, "loss": 0.6304, "step": 6521 }, { "epoch": 0.5557733276523221, "grad_norm": 14.5, "learning_rate": 2.062416827239711e-06, "loss": 0.7317, "step": 6522 }, { "epoch": 0.5558585428206221, "grad_norm": 15.3125, "learning_rate": 2.0622891143740182e-06, "loss": 0.5654, "step": 6523 }, { "epoch": 0.555943757988922, "grad_norm": 17.75, "learning_rate": 2.062161386829444e-06, "loss": 0.8162, "step": 6524 }, { "epoch": 0.556028973157222, "grad_norm": 14.0, "learning_rate": 2.062033644608297e-06, "loss": 0.9634, "step": 6525 }, { "epoch": 0.556114188325522, "grad_norm": 12.375, "learning_rate": 2.0619058877128858e-06, "loss": 0.4097, "step": 6526 }, { "epoch": 0.5561994034938219, "grad_norm": 13.5625, "learning_rate": 2.0617781161455183e-06, "loss": 0.8011, "step": 6527 }, { "epoch": 0.5562846186621219, "grad_norm": 12.5, "learning_rate": 2.061650329908504e-06, "loss": 0.6, "step": 6528 }, { "epoch": 0.5563698338304218, "grad_norm": 16.5, "learning_rate": 2.0615225290041525e-06, "loss": 0.3198, "step": 6529 }, { "epoch": 0.5564550489987218, "grad_norm": 41.5, "learning_rate": 2.0613947134347723e-06, "loss": 1.2413, "step": 6530 }, { "epoch": 0.5565402641670217, "grad_norm": 22.625, "learning_rate": 2.0612668832026737e-06, "loss": 1.0818, "step": 6531 }, { "epoch": 0.5566254793353217, "grad_norm": 17.125, "learning_rate": 2.061139038310167e-06, "loss": 0.8843, "step": 6532 }, { "epoch": 0.5567106945036217, "grad_norm": 17.25, "learning_rate": 2.0610111787595623e-06, "loss": 0.8752, "step": 6533 }, { "epoch": 0.5567959096719216, "grad_norm": 18.625, "learning_rate": 2.06088330455317e-06, "loss": 0.8696, "step": 6534 }, { "epoch": 0.5568811248402216, "grad_norm": 18.75, "learning_rate": 2.060755415693301e-06, "loss": 1.0898, "step": 6535 }, { "epoch": 0.5569663400085215, "grad_norm": 29.0, "learning_rate": 2.060627512182266e-06, "loss": 0.8938, "step": 6536 }, { "epoch": 0.5570515551768215, "grad_norm": 12.75, "learning_rate": 2.060499594022377e-06, "loss": 0.6198, "step": 6537 }, { "epoch": 0.5571367703451214, "grad_norm": 24.0, "learning_rate": 2.0603716612159458e-06, "loss": 0.743, "step": 6538 }, { "epoch": 0.5572219855134214, "grad_norm": 11.1875, "learning_rate": 2.060243713765283e-06, "loss": 0.3479, "step": 6539 }, { "epoch": 0.5573072006817213, "grad_norm": 33.25, "learning_rate": 2.0601157516727018e-06, "loss": 0.789, "step": 6540 }, { "epoch": 0.5573924158500213, "grad_norm": 27.875, "learning_rate": 2.0599877749405146e-06, "loss": 0.8828, "step": 6541 }, { "epoch": 0.5574776310183213, "grad_norm": 13.0, "learning_rate": 2.059859783571033e-06, "loss": 0.7172, "step": 6542 }, { "epoch": 0.5575628461866212, "grad_norm": 16.125, "learning_rate": 2.0597317775665714e-06, "loss": 0.6149, "step": 6543 }, { "epoch": 0.5576480613549212, "grad_norm": 15.9375, "learning_rate": 2.0596037569294423e-06, "loss": 1.1744, "step": 6544 }, { "epoch": 0.5577332765232211, "grad_norm": 14.25, "learning_rate": 2.0594757216619595e-06, "loss": 0.7805, "step": 6545 }, { "epoch": 0.5578184916915211, "grad_norm": 13.4375, "learning_rate": 2.059347671766436e-06, "loss": 0.7751, "step": 6546 }, { "epoch": 0.557903706859821, "grad_norm": 14.25, "learning_rate": 2.0592196072451857e-06, "loss": 0.6645, "step": 6547 }, { "epoch": 0.557988922028121, "grad_norm": 14.1875, "learning_rate": 2.0590915281005243e-06, "loss": 0.8242, "step": 6548 }, { "epoch": 0.558074137196421, "grad_norm": 15.5625, "learning_rate": 2.0589634343347646e-06, "loss": 0.9557, "step": 6549 }, { "epoch": 0.5581593523647209, "grad_norm": 13.9375, "learning_rate": 2.058835325950223e-06, "loss": 0.6361, "step": 6550 }, { "epoch": 0.5582445675330209, "grad_norm": 16.0, "learning_rate": 2.0587072029492134e-06, "loss": 1.0102, "step": 6551 }, { "epoch": 0.5583297827013208, "grad_norm": 11.4375, "learning_rate": 2.058579065334051e-06, "loss": 0.5505, "step": 6552 }, { "epoch": 0.5584149978696208, "grad_norm": 14.0625, "learning_rate": 2.0584509131070522e-06, "loss": 0.3969, "step": 6553 }, { "epoch": 0.5585002130379207, "grad_norm": 13.9375, "learning_rate": 2.0583227462705326e-06, "loss": 0.689, "step": 6554 }, { "epoch": 0.5585854282062207, "grad_norm": 18.75, "learning_rate": 2.0581945648268077e-06, "loss": 0.892, "step": 6555 }, { "epoch": 0.5586706433745207, "grad_norm": 13.5, "learning_rate": 2.0580663687781944e-06, "loss": 0.6407, "step": 6556 }, { "epoch": 0.5587558585428206, "grad_norm": 13.875, "learning_rate": 2.057938158127009e-06, "loss": 0.5784, "step": 6557 }, { "epoch": 0.5588410737111206, "grad_norm": 11.75, "learning_rate": 2.057809932875569e-06, "loss": 0.5397, "step": 6558 }, { "epoch": 0.5589262888794205, "grad_norm": 19.875, "learning_rate": 2.0576816930261917e-06, "loss": 0.9131, "step": 6559 }, { "epoch": 0.5590115040477205, "grad_norm": 13.1875, "learning_rate": 2.0575534385811932e-06, "loss": 0.5312, "step": 6560 }, { "epoch": 0.5590967192160204, "grad_norm": 15.5, "learning_rate": 2.057425169542892e-06, "loss": 0.9385, "step": 6561 }, { "epoch": 0.5591819343843204, "grad_norm": 12.8125, "learning_rate": 2.0572968859136066e-06, "loss": 0.7566, "step": 6562 }, { "epoch": 0.5592671495526204, "grad_norm": 16.875, "learning_rate": 2.0571685876956542e-06, "loss": 0.6952, "step": 6563 }, { "epoch": 0.5593523647209203, "grad_norm": 20.0, "learning_rate": 2.057040274891354e-06, "loss": 1.1375, "step": 6564 }, { "epoch": 0.5594375798892203, "grad_norm": 17.625, "learning_rate": 2.056911947503025e-06, "loss": 1.2299, "step": 6565 }, { "epoch": 0.5595227950575202, "grad_norm": 11.4375, "learning_rate": 2.056783605532985e-06, "loss": 0.3876, "step": 6566 }, { "epoch": 0.5596080102258202, "grad_norm": 13.5, "learning_rate": 2.056655248983554e-06, "loss": 0.4853, "step": 6567 }, { "epoch": 0.5596932253941201, "grad_norm": 13.75, "learning_rate": 2.0565268778570518e-06, "loss": 0.9482, "step": 6568 }, { "epoch": 0.5597784405624201, "grad_norm": 12.125, "learning_rate": 2.0563984921557973e-06, "loss": 0.5212, "step": 6569 }, { "epoch": 0.5598636557307201, "grad_norm": 13.4375, "learning_rate": 2.0562700918821117e-06, "loss": 0.7779, "step": 6570 }, { "epoch": 0.55994887089902, "grad_norm": 14.0, "learning_rate": 2.0561416770383147e-06, "loss": 0.9288, "step": 6571 }, { "epoch": 0.56003408606732, "grad_norm": 14.9375, "learning_rate": 2.056013247626727e-06, "loss": 1.0304, "step": 6572 }, { "epoch": 0.5601193012356199, "grad_norm": 14.5625, "learning_rate": 2.0558848036496693e-06, "loss": 0.7677, "step": 6573 }, { "epoch": 0.5602045164039199, "grad_norm": 22.0, "learning_rate": 2.0557563451094627e-06, "loss": 0.4116, "step": 6574 }, { "epoch": 0.5602897315722198, "grad_norm": 22.375, "learning_rate": 2.055627872008429e-06, "loss": 1.3887, "step": 6575 }, { "epoch": 0.5603749467405198, "grad_norm": 20.875, "learning_rate": 2.055499384348889e-06, "loss": 1.071, "step": 6576 }, { "epoch": 0.5604601619088198, "grad_norm": 12.9375, "learning_rate": 2.0553708821331657e-06, "loss": 0.7193, "step": 6577 }, { "epoch": 0.5605453770771197, "grad_norm": 17.75, "learning_rate": 2.0552423653635805e-06, "loss": 0.8723, "step": 6578 }, { "epoch": 0.5606305922454197, "grad_norm": 16.5, "learning_rate": 2.055113834042456e-06, "loss": 0.5388, "step": 6579 }, { "epoch": 0.5607158074137196, "grad_norm": 11.25, "learning_rate": 2.0549852881721148e-06, "loss": 0.4838, "step": 6580 }, { "epoch": 0.5608010225820196, "grad_norm": 15.125, "learning_rate": 2.05485672775488e-06, "loss": 0.8816, "step": 6581 }, { "epoch": 0.5608862377503195, "grad_norm": 11.3125, "learning_rate": 2.0547281527930753e-06, "loss": 0.5521, "step": 6582 }, { "epoch": 0.5609714529186195, "grad_norm": 19.375, "learning_rate": 2.054599563289023e-06, "loss": 1.0904, "step": 6583 }, { "epoch": 0.5610566680869195, "grad_norm": 13.375, "learning_rate": 2.0544709592450473e-06, "loss": 0.6103, "step": 6584 }, { "epoch": 0.5611418832552194, "grad_norm": 16.0, "learning_rate": 2.054342340663473e-06, "loss": 0.9819, "step": 6585 }, { "epoch": 0.5612270984235194, "grad_norm": 13.5625, "learning_rate": 2.0542137075466234e-06, "loss": 0.8006, "step": 6586 }, { "epoch": 0.5613123135918193, "grad_norm": 14.875, "learning_rate": 2.054085059896824e-06, "loss": 0.698, "step": 6587 }, { "epoch": 0.5613975287601193, "grad_norm": 19.5, "learning_rate": 2.0539563977163983e-06, "loss": 0.7512, "step": 6588 }, { "epoch": 0.5614827439284192, "grad_norm": 19.375, "learning_rate": 2.0538277210076722e-06, "loss": 1.2143, "step": 6589 }, { "epoch": 0.5615679590967192, "grad_norm": 31.5, "learning_rate": 2.0536990297729715e-06, "loss": 1.0487, "step": 6590 }, { "epoch": 0.5616531742650192, "grad_norm": 17.125, "learning_rate": 2.0535703240146207e-06, "loss": 1.0768, "step": 6591 }, { "epoch": 0.5617383894333191, "grad_norm": 13.75, "learning_rate": 2.053441603734946e-06, "loss": 0.8349, "step": 6592 }, { "epoch": 0.5618236046016191, "grad_norm": 13.25, "learning_rate": 2.0533128689362738e-06, "loss": 0.6898, "step": 6593 }, { "epoch": 0.561908819769919, "grad_norm": 12.9375, "learning_rate": 2.05318411962093e-06, "loss": 0.6294, "step": 6594 }, { "epoch": 0.561994034938219, "grad_norm": 22.25, "learning_rate": 2.0530553557912417e-06, "loss": 1.0585, "step": 6595 }, { "epoch": 0.5620792501065189, "grad_norm": 13.3125, "learning_rate": 2.0529265774495357e-06, "loss": 0.6274, "step": 6596 }, { "epoch": 0.5621644652748189, "grad_norm": 15.6875, "learning_rate": 2.0527977845981383e-06, "loss": 0.5907, "step": 6597 }, { "epoch": 0.5622496804431188, "grad_norm": 13.125, "learning_rate": 2.052668977239378e-06, "loss": 0.4528, "step": 6598 }, { "epoch": 0.5623348956114188, "grad_norm": 14.0625, "learning_rate": 2.052540155375583e-06, "loss": 0.7443, "step": 6599 }, { "epoch": 0.5624201107797188, "grad_norm": 14.0, "learning_rate": 2.0524113190090795e-06, "loss": 0.7788, "step": 6600 }, { "epoch": 0.5625053259480187, "grad_norm": 12.375, "learning_rate": 2.0522824681421967e-06, "loss": 0.5758, "step": 6601 }, { "epoch": 0.5625905411163187, "grad_norm": 18.5, "learning_rate": 2.0521536027772633e-06, "loss": 0.8529, "step": 6602 }, { "epoch": 0.5626757562846186, "grad_norm": 12.9375, "learning_rate": 2.0520247229166075e-06, "loss": 0.489, "step": 6603 }, { "epoch": 0.5627609714529186, "grad_norm": 17.0, "learning_rate": 2.0518958285625585e-06, "loss": 0.9931, "step": 6604 }, { "epoch": 0.5628461866212185, "grad_norm": 16.125, "learning_rate": 2.051766919717445e-06, "loss": 0.8584, "step": 6605 }, { "epoch": 0.5629314017895185, "grad_norm": 21.625, "learning_rate": 2.0516379963835977e-06, "loss": 0.7719, "step": 6606 }, { "epoch": 0.5630166169578185, "grad_norm": 13.75, "learning_rate": 2.0515090585633453e-06, "loss": 1.024, "step": 6607 }, { "epoch": 0.5631018321261184, "grad_norm": 12.375, "learning_rate": 2.0513801062590184e-06, "loss": 0.4838, "step": 6608 }, { "epoch": 0.5631870472944184, "grad_norm": 12.25, "learning_rate": 2.051251139472947e-06, "loss": 0.6111, "step": 6609 }, { "epoch": 0.5632722624627183, "grad_norm": 13.5, "learning_rate": 2.051122158207462e-06, "loss": 0.5268, "step": 6610 }, { "epoch": 0.5633574776310183, "grad_norm": 11.25, "learning_rate": 2.050993162464894e-06, "loss": 0.4595, "step": 6611 }, { "epoch": 0.5634426927993182, "grad_norm": 12.625, "learning_rate": 2.0508641522475734e-06, "loss": 0.3808, "step": 6612 }, { "epoch": 0.5635279079676182, "grad_norm": 11.875, "learning_rate": 2.050735127557833e-06, "loss": 0.525, "step": 6613 }, { "epoch": 0.5636131231359182, "grad_norm": 14.8125, "learning_rate": 2.0506060883980034e-06, "loss": 0.6565, "step": 6614 }, { "epoch": 0.5636983383042181, "grad_norm": 17.125, "learning_rate": 2.0504770347704167e-06, "loss": 1.0895, "step": 6615 }, { "epoch": 0.5637835534725181, "grad_norm": 10.8125, "learning_rate": 2.050347966677405e-06, "loss": 0.4161, "step": 6616 }, { "epoch": 0.563868768640818, "grad_norm": 15.3125, "learning_rate": 2.0502188841213012e-06, "loss": 0.9376, "step": 6617 }, { "epoch": 0.563953983809118, "grad_norm": 15.1875, "learning_rate": 2.0500897871044366e-06, "loss": 0.7401, "step": 6618 }, { "epoch": 0.5640391989774179, "grad_norm": 14.8125, "learning_rate": 2.049960675629146e-06, "loss": 0.6603, "step": 6619 }, { "epoch": 0.5641244141457179, "grad_norm": 16.5, "learning_rate": 2.049831549697761e-06, "loss": 0.6978, "step": 6620 }, { "epoch": 0.5642096293140179, "grad_norm": 11.75, "learning_rate": 2.0497024093126155e-06, "loss": 0.4691, "step": 6621 }, { "epoch": 0.5642948444823178, "grad_norm": 15.3125, "learning_rate": 2.0495732544760435e-06, "loss": 0.8976, "step": 6622 }, { "epoch": 0.5643800596506178, "grad_norm": 18.75, "learning_rate": 2.049444085190379e-06, "loss": 0.2405, "step": 6623 }, { "epoch": 0.5644652748189177, "grad_norm": 12.1875, "learning_rate": 2.0493149014579557e-06, "loss": 0.2641, "step": 6624 }, { "epoch": 0.5645504899872177, "grad_norm": 11.1875, "learning_rate": 2.049185703281108e-06, "loss": 0.4319, "step": 6625 }, { "epoch": 0.5646357051555176, "grad_norm": 11.0625, "learning_rate": 2.0490564906621717e-06, "loss": 0.4829, "step": 6626 }, { "epoch": 0.5647209203238176, "grad_norm": 14.5, "learning_rate": 2.0489272636034808e-06, "loss": 0.7447, "step": 6627 }, { "epoch": 0.5648061354921176, "grad_norm": 16.375, "learning_rate": 2.0487980221073707e-06, "loss": 0.9176, "step": 6628 }, { "epoch": 0.5648913506604175, "grad_norm": 17.25, "learning_rate": 2.0486687661761773e-06, "loss": 1.0748, "step": 6629 }, { "epoch": 0.5649765658287175, "grad_norm": 17.375, "learning_rate": 2.0485394958122357e-06, "loss": 0.6985, "step": 6630 }, { "epoch": 0.5650617809970174, "grad_norm": 15.4375, "learning_rate": 2.048410211017883e-06, "loss": 0.9069, "step": 6631 }, { "epoch": 0.5651469961653174, "grad_norm": 15.5, "learning_rate": 2.048280911795455e-06, "loss": 0.538, "step": 6632 }, { "epoch": 0.5652322113336173, "grad_norm": 11.6875, "learning_rate": 2.0481515981472877e-06, "loss": 0.3747, "step": 6633 }, { "epoch": 0.5653174265019173, "grad_norm": 10.375, "learning_rate": 2.0480222700757186e-06, "loss": 0.434, "step": 6634 }, { "epoch": 0.5654026416702173, "grad_norm": 20.5, "learning_rate": 2.0478929275830845e-06, "loss": 0.8986, "step": 6635 }, { "epoch": 0.5654878568385172, "grad_norm": 11.625, "learning_rate": 2.047763570671723e-06, "loss": 0.4706, "step": 6636 }, { "epoch": 0.5655730720068172, "grad_norm": 16.125, "learning_rate": 2.0476341993439718e-06, "loss": 1.1335, "step": 6637 }, { "epoch": 0.5656582871751171, "grad_norm": 15.3125, "learning_rate": 2.047504813602168e-06, "loss": 1.0077, "step": 6638 }, { "epoch": 0.5657435023434171, "grad_norm": 15.5, "learning_rate": 2.047375413448651e-06, "loss": 0.6693, "step": 6639 }, { "epoch": 0.565828717511717, "grad_norm": 14.5, "learning_rate": 2.047245998885758e-06, "loss": 0.9674, "step": 6640 }, { "epoch": 0.565913932680017, "grad_norm": 13.6875, "learning_rate": 2.047116569915828e-06, "loss": 0.5483, "step": 6641 }, { "epoch": 0.565999147848317, "grad_norm": 9.5625, "learning_rate": 2.0469871265412e-06, "loss": 0.8026, "step": 6642 }, { "epoch": 0.5660843630166169, "grad_norm": 19.75, "learning_rate": 2.0468576687642135e-06, "loss": 0.9924, "step": 6643 }, { "epoch": 0.5661695781849169, "grad_norm": 13.0625, "learning_rate": 2.0467281965872073e-06, "loss": 0.6448, "step": 6644 }, { "epoch": 0.5662547933532168, "grad_norm": 16.625, "learning_rate": 2.046598710012522e-06, "loss": 0.8509, "step": 6645 }, { "epoch": 0.5663400085215168, "grad_norm": 14.5, "learning_rate": 2.0464692090424963e-06, "loss": 0.5845, "step": 6646 }, { "epoch": 0.5664252236898167, "grad_norm": 15.4375, "learning_rate": 2.046339693679471e-06, "loss": 0.8635, "step": 6647 }, { "epoch": 0.5665104388581167, "grad_norm": 15.25, "learning_rate": 2.046210163925787e-06, "loss": 0.8912, "step": 6648 }, { "epoch": 0.5665956540264167, "grad_norm": 14.75, "learning_rate": 2.0460806197837845e-06, "loss": 0.6407, "step": 6649 }, { "epoch": 0.5666808691947166, "grad_norm": 16.25, "learning_rate": 2.045951061255805e-06, "loss": 0.7689, "step": 6650 }, { "epoch": 0.5667660843630166, "grad_norm": 14.5, "learning_rate": 2.0458214883441897e-06, "loss": 0.8556, "step": 6651 }, { "epoch": 0.5668512995313165, "grad_norm": 13.6875, "learning_rate": 2.0456919010512793e-06, "loss": 0.4139, "step": 6652 }, { "epoch": 0.5669365146996165, "grad_norm": 16.25, "learning_rate": 2.0455622993794162e-06, "loss": 0.83, "step": 6653 }, { "epoch": 0.5670217298679164, "grad_norm": 14.25, "learning_rate": 2.0454326833309426e-06, "loss": 1.0854, "step": 6654 }, { "epoch": 0.5671069450362164, "grad_norm": 19.375, "learning_rate": 2.0453030529082e-06, "loss": 0.8774, "step": 6655 }, { "epoch": 0.5671921602045163, "grad_norm": 19.25, "learning_rate": 2.045173408113532e-06, "loss": 0.7203, "step": 6656 }, { "epoch": 0.5672773753728163, "grad_norm": 16.375, "learning_rate": 2.0450437489492806e-06, "loss": 0.932, "step": 6657 }, { "epoch": 0.5673625905411164, "grad_norm": 10.0625, "learning_rate": 2.0449140754177893e-06, "loss": 0.4206, "step": 6658 }, { "epoch": 0.5674478057094163, "grad_norm": 12.6875, "learning_rate": 2.0447843875214014e-06, "loss": 0.7666, "step": 6659 }, { "epoch": 0.5675330208777163, "grad_norm": 15.6875, "learning_rate": 2.0446546852624603e-06, "loss": 0.71, "step": 6660 }, { "epoch": 0.5676182360460162, "grad_norm": 22.625, "learning_rate": 2.04452496864331e-06, "loss": 0.8318, "step": 6661 }, { "epoch": 0.5677034512143162, "grad_norm": 13.5, "learning_rate": 2.0443952376662946e-06, "loss": 0.7203, "step": 6662 }, { "epoch": 0.5677886663826162, "grad_norm": 12.0, "learning_rate": 2.044265492333758e-06, "loss": 0.5695, "step": 6663 }, { "epoch": 0.5678738815509161, "grad_norm": 12.0625, "learning_rate": 2.044135732648046e-06, "loss": 0.4586, "step": 6664 }, { "epoch": 0.5679590967192161, "grad_norm": 15.5, "learning_rate": 2.044005958611502e-06, "loss": 0.9225, "step": 6665 }, { "epoch": 0.568044311887516, "grad_norm": 10.5, "learning_rate": 2.0438761702264722e-06, "loss": 0.3583, "step": 6666 }, { "epoch": 0.568129527055816, "grad_norm": 12.5625, "learning_rate": 2.043746367495302e-06, "loss": 0.6217, "step": 6667 }, { "epoch": 0.5682147422241159, "grad_norm": 14.1875, "learning_rate": 2.0436165504203366e-06, "loss": 0.9287, "step": 6668 }, { "epoch": 0.5682999573924159, "grad_norm": 11.4375, "learning_rate": 2.043486719003922e-06, "loss": 0.4519, "step": 6669 }, { "epoch": 0.5683851725607159, "grad_norm": 13.875, "learning_rate": 2.043356873248404e-06, "loss": 0.7129, "step": 6670 }, { "epoch": 0.5684703877290158, "grad_norm": 14.0, "learning_rate": 2.0432270131561304e-06, "loss": 0.3966, "step": 6671 }, { "epoch": 0.5685556028973158, "grad_norm": 19.0, "learning_rate": 2.043097138729447e-06, "loss": 1.1598, "step": 6672 }, { "epoch": 0.5686408180656157, "grad_norm": 16.125, "learning_rate": 2.0429672499707e-06, "loss": 0.8472, "step": 6673 }, { "epoch": 0.5687260332339157, "grad_norm": 11.9375, "learning_rate": 2.0428373468822376e-06, "loss": 0.3286, "step": 6674 }, { "epoch": 0.5688112484022156, "grad_norm": 13.75, "learning_rate": 2.0427074294664074e-06, "loss": 0.8259, "step": 6675 }, { "epoch": 0.5688964635705156, "grad_norm": 13.375, "learning_rate": 2.0425774977255567e-06, "loss": 0.7435, "step": 6676 }, { "epoch": 0.5689816787388156, "grad_norm": 14.3125, "learning_rate": 2.042447551662034e-06, "loss": 0.6939, "step": 6677 }, { "epoch": 0.5690668939071155, "grad_norm": 12.3125, "learning_rate": 2.0423175912781867e-06, "loss": 0.6179, "step": 6678 }, { "epoch": 0.5691521090754155, "grad_norm": 13.3125, "learning_rate": 2.0421876165763637e-06, "loss": 0.7054, "step": 6679 }, { "epoch": 0.5692373242437154, "grad_norm": 18.625, "learning_rate": 2.042057627558914e-06, "loss": 0.9556, "step": 6680 }, { "epoch": 0.5693225394120154, "grad_norm": 18.75, "learning_rate": 2.0419276242281866e-06, "loss": 1.0551, "step": 6681 }, { "epoch": 0.5694077545803153, "grad_norm": 20.25, "learning_rate": 2.041797606586531e-06, "loss": 0.8718, "step": 6682 }, { "epoch": 0.5694929697486153, "grad_norm": 19.625, "learning_rate": 2.0416675746362956e-06, "loss": 1.1494, "step": 6683 }, { "epoch": 0.5695781849169153, "grad_norm": 16.5, "learning_rate": 2.0415375283798316e-06, "loss": 0.7423, "step": 6684 }, { "epoch": 0.5696634000852152, "grad_norm": 11.5625, "learning_rate": 2.0414074678194884e-06, "loss": 0.6575, "step": 6685 }, { "epoch": 0.5697486152535152, "grad_norm": 16.5, "learning_rate": 2.0412773929576165e-06, "loss": 0.8467, "step": 6686 }, { "epoch": 0.5698338304218151, "grad_norm": 13.3125, "learning_rate": 2.0411473037965663e-06, "loss": 0.6279, "step": 6687 }, { "epoch": 0.5699190455901151, "grad_norm": 12.25, "learning_rate": 2.0410172003386892e-06, "loss": 0.4611, "step": 6688 }, { "epoch": 0.570004260758415, "grad_norm": 11.0, "learning_rate": 2.0408870825863354e-06, "loss": 0.5633, "step": 6689 }, { "epoch": 0.570089475926715, "grad_norm": 17.375, "learning_rate": 2.0407569505418566e-06, "loss": 0.6571, "step": 6690 }, { "epoch": 0.570174691095015, "grad_norm": 21.625, "learning_rate": 2.040626804207605e-06, "loss": 0.6264, "step": 6691 }, { "epoch": 0.5702599062633149, "grad_norm": 13.4375, "learning_rate": 2.040496643585932e-06, "loss": 0.5343, "step": 6692 }, { "epoch": 0.5703451214316149, "grad_norm": 11.5625, "learning_rate": 2.040366468679189e-06, "loss": 0.5013, "step": 6693 }, { "epoch": 0.5704303365999148, "grad_norm": 12.5625, "learning_rate": 2.0402362794897294e-06, "loss": 0.494, "step": 6694 }, { "epoch": 0.5705155517682148, "grad_norm": 14.375, "learning_rate": 2.0401060760199063e-06, "loss": 0.6601, "step": 6695 }, { "epoch": 0.5706007669365147, "grad_norm": 12.3125, "learning_rate": 2.0399758582720715e-06, "loss": 0.5622, "step": 6696 }, { "epoch": 0.5706859821048147, "grad_norm": 14.3125, "learning_rate": 2.0398456262485784e-06, "loss": 0.5121, "step": 6697 }, { "epoch": 0.5707711972731146, "grad_norm": 26.25, "learning_rate": 2.0397153799517806e-06, "loss": 0.86, "step": 6698 }, { "epoch": 0.5708564124414146, "grad_norm": 14.875, "learning_rate": 2.039585119384032e-06, "loss": 0.8094, "step": 6699 }, { "epoch": 0.5709416276097146, "grad_norm": 13.5, "learning_rate": 2.039454844547686e-06, "loss": 0.4824, "step": 6700 }, { "epoch": 0.5710268427780145, "grad_norm": 13.9375, "learning_rate": 2.039324555445097e-06, "loss": 0.5616, "step": 6701 }, { "epoch": 0.5711120579463145, "grad_norm": 16.375, "learning_rate": 2.03919425207862e-06, "loss": 1.2018, "step": 6702 }, { "epoch": 0.5711972731146144, "grad_norm": 16.125, "learning_rate": 2.039063934450609e-06, "loss": 0.7871, "step": 6703 }, { "epoch": 0.5712824882829144, "grad_norm": 13.125, "learning_rate": 2.0389336025634194e-06, "loss": 0.7451, "step": 6704 }, { "epoch": 0.5713677034512143, "grad_norm": 20.25, "learning_rate": 2.038803256419406e-06, "loss": 0.7874, "step": 6705 }, { "epoch": 0.5714529186195143, "grad_norm": 9.9375, "learning_rate": 2.0386728960209247e-06, "loss": 0.3616, "step": 6706 }, { "epoch": 0.5715381337878143, "grad_norm": 12.5, "learning_rate": 2.0385425213703313e-06, "loss": 0.5933, "step": 6707 }, { "epoch": 0.5716233489561142, "grad_norm": 17.375, "learning_rate": 2.038412132469981e-06, "loss": 0.989, "step": 6708 }, { "epoch": 0.5717085641244142, "grad_norm": 11.5, "learning_rate": 2.0382817293222314e-06, "loss": 0.5472, "step": 6709 }, { "epoch": 0.5717937792927141, "grad_norm": 16.25, "learning_rate": 2.038151311929438e-06, "loss": 0.7496, "step": 6710 }, { "epoch": 0.5718789944610141, "grad_norm": 15.8125, "learning_rate": 2.038020880293958e-06, "loss": 0.9538, "step": 6711 }, { "epoch": 0.571964209629314, "grad_norm": 13.875, "learning_rate": 2.037890434418148e-06, "loss": 0.8813, "step": 6712 }, { "epoch": 0.572049424797614, "grad_norm": 10.5625, "learning_rate": 2.0377599743043653e-06, "loss": 0.3264, "step": 6713 }, { "epoch": 0.572134639965914, "grad_norm": 14.4375, "learning_rate": 2.0376294999549684e-06, "loss": 1.0381, "step": 6714 }, { "epoch": 0.5722198551342139, "grad_norm": 14.9375, "learning_rate": 2.0374990113723136e-06, "loss": 0.9154, "step": 6715 }, { "epoch": 0.5723050703025139, "grad_norm": 16.75, "learning_rate": 2.0373685085587604e-06, "loss": 0.7372, "step": 6716 }, { "epoch": 0.5723902854708138, "grad_norm": 13.4375, "learning_rate": 2.0372379915166664e-06, "loss": 0.9768, "step": 6717 }, { "epoch": 0.5724755006391138, "grad_norm": 12.9375, "learning_rate": 2.03710746024839e-06, "loss": 0.5258, "step": 6718 }, { "epoch": 0.5725607158074137, "grad_norm": 13.0, "learning_rate": 2.0369769147562908e-06, "loss": 0.7403, "step": 6719 }, { "epoch": 0.5726459309757137, "grad_norm": 20.0, "learning_rate": 2.0368463550427267e-06, "loss": 0.8875, "step": 6720 }, { "epoch": 0.5727311461440137, "grad_norm": 11.8125, "learning_rate": 2.0367157811100587e-06, "loss": 0.5844, "step": 6721 }, { "epoch": 0.5728163613123136, "grad_norm": 12.5, "learning_rate": 2.0365851929606444e-06, "loss": 0.351, "step": 6722 }, { "epoch": 0.5729015764806136, "grad_norm": 13.1875, "learning_rate": 2.0364545905968454e-06, "loss": 0.895, "step": 6723 }, { "epoch": 0.5729867916489135, "grad_norm": 15.6875, "learning_rate": 2.0363239740210206e-06, "loss": 0.7221, "step": 6724 }, { "epoch": 0.5730720068172135, "grad_norm": 16.5, "learning_rate": 2.0361933432355312e-06, "loss": 0.857, "step": 6725 }, { "epoch": 0.5731572219855134, "grad_norm": 14.75, "learning_rate": 2.0360626982427375e-06, "loss": 0.8438, "step": 6726 }, { "epoch": 0.5732424371538134, "grad_norm": 20.875, "learning_rate": 2.035932039045e-06, "loss": 1.0177, "step": 6727 }, { "epoch": 0.5733276523221134, "grad_norm": 16.5, "learning_rate": 2.035801365644681e-06, "loss": 0.9042, "step": 6728 }, { "epoch": 0.5734128674904133, "grad_norm": 15.1875, "learning_rate": 2.0356706780441406e-06, "loss": 0.8774, "step": 6729 }, { "epoch": 0.5734980826587133, "grad_norm": 14.375, "learning_rate": 2.0355399762457408e-06, "loss": 0.7066, "step": 6730 }, { "epoch": 0.5735832978270132, "grad_norm": 14.375, "learning_rate": 2.035409260251844e-06, "loss": 0.617, "step": 6731 }, { "epoch": 0.5736685129953132, "grad_norm": 20.125, "learning_rate": 2.035278530064812e-06, "loss": 0.7918, "step": 6732 }, { "epoch": 0.5737537281636131, "grad_norm": 12.8125, "learning_rate": 2.035147785687007e-06, "loss": 0.4654, "step": 6733 }, { "epoch": 0.5738389433319131, "grad_norm": 17.0, "learning_rate": 2.0350170271207924e-06, "loss": 0.5677, "step": 6734 }, { "epoch": 0.573924158500213, "grad_norm": 19.25, "learning_rate": 2.034886254368531e-06, "loss": 0.9539, "step": 6735 }, { "epoch": 0.574009373668513, "grad_norm": 13.125, "learning_rate": 2.034755467432585e-06, "loss": 0.4915, "step": 6736 }, { "epoch": 0.574094588836813, "grad_norm": 14.9375, "learning_rate": 2.0346246663153193e-06, "loss": 0.5336, "step": 6737 }, { "epoch": 0.5741798040051129, "grad_norm": 22.25, "learning_rate": 2.0344938510190964e-06, "loss": 1.0443, "step": 6738 }, { "epoch": 0.5742650191734129, "grad_norm": 13.9375, "learning_rate": 2.034363021546281e-06, "loss": 0.7953, "step": 6739 }, { "epoch": 0.5743502343417128, "grad_norm": 19.5, "learning_rate": 2.034232177899237e-06, "loss": 0.9649, "step": 6740 }, { "epoch": 0.5744354495100128, "grad_norm": 18.625, "learning_rate": 2.034101320080329e-06, "loss": 0.8548, "step": 6741 }, { "epoch": 0.5745206646783128, "grad_norm": 15.5625, "learning_rate": 2.0339704480919216e-06, "loss": 0.7896, "step": 6742 }, { "epoch": 0.5746058798466127, "grad_norm": 14.0, "learning_rate": 2.0338395619363794e-06, "loss": 0.7382, "step": 6743 }, { "epoch": 0.5746910950149127, "grad_norm": 30.25, "learning_rate": 2.0337086616160687e-06, "loss": 0.9601, "step": 6744 }, { "epoch": 0.5747763101832126, "grad_norm": 16.375, "learning_rate": 2.0335777471333545e-06, "loss": 1.0277, "step": 6745 }, { "epoch": 0.5748615253515126, "grad_norm": 15.25, "learning_rate": 2.033446818490602e-06, "loss": 0.6509, "step": 6746 }, { "epoch": 0.5749467405198125, "grad_norm": 11.6875, "learning_rate": 2.033315875690178e-06, "loss": 0.4963, "step": 6747 }, { "epoch": 0.5750319556881125, "grad_norm": 18.375, "learning_rate": 2.0331849187344483e-06, "loss": 1.1608, "step": 6748 }, { "epoch": 0.5751171708564125, "grad_norm": 19.5, "learning_rate": 2.0330539476257795e-06, "loss": 0.8022, "step": 6749 }, { "epoch": 0.5752023860247124, "grad_norm": 36.5, "learning_rate": 2.0329229623665384e-06, "loss": 1.1027, "step": 6750 }, { "epoch": 0.5752876011930124, "grad_norm": 14.25, "learning_rate": 2.032791962959092e-06, "loss": 0.844, "step": 6751 }, { "epoch": 0.5753728163613123, "grad_norm": 21.75, "learning_rate": 2.032660949405808e-06, "loss": 0.732, "step": 6752 }, { "epoch": 0.5754580315296123, "grad_norm": 30.625, "learning_rate": 2.0325299217090534e-06, "loss": 1.2567, "step": 6753 }, { "epoch": 0.5755432466979122, "grad_norm": 23.75, "learning_rate": 2.0323988798711963e-06, "loss": 0.8542, "step": 6754 }, { "epoch": 0.5756284618662122, "grad_norm": 17.0, "learning_rate": 2.0322678238946047e-06, "loss": 0.6581, "step": 6755 }, { "epoch": 0.5757136770345122, "grad_norm": 13.75, "learning_rate": 2.032136753781647e-06, "loss": 0.7792, "step": 6756 }, { "epoch": 0.5757988922028121, "grad_norm": 11.4375, "learning_rate": 2.032005669534692e-06, "loss": 0.32, "step": 6757 }, { "epoch": 0.5758841073711121, "grad_norm": 14.5625, "learning_rate": 2.031874571156107e-06, "loss": 0.8013, "step": 6758 }, { "epoch": 0.575969322539412, "grad_norm": 12.0, "learning_rate": 2.031743458648263e-06, "loss": 0.4304, "step": 6759 }, { "epoch": 0.576054537707712, "grad_norm": 14.0, "learning_rate": 2.031612332013529e-06, "loss": 0.9376, "step": 6760 }, { "epoch": 0.5761397528760119, "grad_norm": 22.25, "learning_rate": 2.031481191254274e-06, "loss": 1.0684, "step": 6761 }, { "epoch": 0.5762249680443119, "grad_norm": 20.625, "learning_rate": 2.0313500363728675e-06, "loss": 0.6924, "step": 6762 }, { "epoch": 0.5763101832126118, "grad_norm": 22.875, "learning_rate": 2.0312188673716802e-06, "loss": 0.9757, "step": 6763 }, { "epoch": 0.5763953983809118, "grad_norm": 13.8125, "learning_rate": 2.0310876842530826e-06, "loss": 1.0137, "step": 6764 }, { "epoch": 0.5764806135492118, "grad_norm": 13.5, "learning_rate": 2.0309564870194453e-06, "loss": 0.6846, "step": 6765 }, { "epoch": 0.5765658287175117, "grad_norm": 14.9375, "learning_rate": 2.0308252756731388e-06, "loss": 0.8261, "step": 6766 }, { "epoch": 0.5766510438858117, "grad_norm": 16.375, "learning_rate": 2.0306940502165343e-06, "loss": 0.4896, "step": 6767 }, { "epoch": 0.5767362590541116, "grad_norm": 13.75, "learning_rate": 2.0305628106520033e-06, "loss": 0.7851, "step": 6768 }, { "epoch": 0.5768214742224116, "grad_norm": 14.1875, "learning_rate": 2.0304315569819176e-06, "loss": 0.6587, "step": 6769 }, { "epoch": 0.5769066893907115, "grad_norm": 17.375, "learning_rate": 2.0303002892086487e-06, "loss": 0.4979, "step": 6770 }, { "epoch": 0.5769919045590115, "grad_norm": 11.75, "learning_rate": 2.0301690073345687e-06, "loss": 0.3979, "step": 6771 }, { "epoch": 0.5770771197273115, "grad_norm": 14.9375, "learning_rate": 2.030037711362051e-06, "loss": 0.558, "step": 6772 }, { "epoch": 0.5771623348956114, "grad_norm": 22.875, "learning_rate": 2.0299064012934665e-06, "loss": 0.9738, "step": 6773 }, { "epoch": 0.5772475500639114, "grad_norm": 16.375, "learning_rate": 2.02977507713119e-06, "loss": 0.8899, "step": 6774 }, { "epoch": 0.5773327652322113, "grad_norm": 9.75, "learning_rate": 2.0296437388775927e-06, "loss": 0.3427, "step": 6775 }, { "epoch": 0.5774179804005113, "grad_norm": 10.625, "learning_rate": 2.0295123865350496e-06, "loss": 0.2844, "step": 6776 }, { "epoch": 0.5775031955688112, "grad_norm": 10.1875, "learning_rate": 2.029381020105934e-06, "loss": 0.4853, "step": 6777 }, { "epoch": 0.5775884107371112, "grad_norm": 20.125, "learning_rate": 2.029249639592619e-06, "loss": 0.8696, "step": 6778 }, { "epoch": 0.5776736259054112, "grad_norm": 13.25, "learning_rate": 2.02911824499748e-06, "loss": 0.5983, "step": 6779 }, { "epoch": 0.5777588410737111, "grad_norm": 13.125, "learning_rate": 2.028986836322891e-06, "loss": 0.8721, "step": 6780 }, { "epoch": 0.5778440562420111, "grad_norm": 14.625, "learning_rate": 2.028855413571226e-06, "loss": 0.6158, "step": 6781 }, { "epoch": 0.577929271410311, "grad_norm": 22.25, "learning_rate": 2.0287239767448608e-06, "loss": 0.9306, "step": 6782 }, { "epoch": 0.578014486578611, "grad_norm": 13.5, "learning_rate": 2.02859252584617e-06, "loss": 0.7195, "step": 6783 }, { "epoch": 0.5780997017469109, "grad_norm": 14.375, "learning_rate": 2.0284610608775294e-06, "loss": 0.6131, "step": 6784 }, { "epoch": 0.5781849169152109, "grad_norm": 22.5, "learning_rate": 2.0283295818413152e-06, "loss": 1.3303, "step": 6785 }, { "epoch": 0.5782701320835109, "grad_norm": 14.6875, "learning_rate": 2.028198088739902e-06, "loss": 0.6883, "step": 6786 }, { "epoch": 0.5783553472518108, "grad_norm": 18.375, "learning_rate": 2.0280665815756672e-06, "loss": 0.8534, "step": 6787 }, { "epoch": 0.5784405624201108, "grad_norm": 13.0625, "learning_rate": 2.027935060350987e-06, "loss": 0.5742, "step": 6788 }, { "epoch": 0.5785257775884107, "grad_norm": 20.75, "learning_rate": 2.0278035250682373e-06, "loss": 0.8672, "step": 6789 }, { "epoch": 0.5786109927567107, "grad_norm": 10.4375, "learning_rate": 2.027671975729796e-06, "loss": 0.5317, "step": 6790 }, { "epoch": 0.5786962079250106, "grad_norm": 14.125, "learning_rate": 2.027540412338041e-06, "loss": 0.7013, "step": 6791 }, { "epoch": 0.5787814230933106, "grad_norm": 16.875, "learning_rate": 2.027408834895348e-06, "loss": 0.9326, "step": 6792 }, { "epoch": 0.5788666382616106, "grad_norm": 9.875, "learning_rate": 2.0272772434040955e-06, "loss": 0.3284, "step": 6793 }, { "epoch": 0.5789518534299105, "grad_norm": 17.0, "learning_rate": 2.027145637866662e-06, "loss": 0.9317, "step": 6794 }, { "epoch": 0.5790370685982105, "grad_norm": 16.5, "learning_rate": 2.027014018285425e-06, "loss": 0.8467, "step": 6795 }, { "epoch": 0.5791222837665104, "grad_norm": 22.875, "learning_rate": 2.026882384662764e-06, "loss": 1.1981, "step": 6796 }, { "epoch": 0.5792074989348104, "grad_norm": 14.5, "learning_rate": 2.0267507370010564e-06, "loss": 0.7329, "step": 6797 }, { "epoch": 0.5792927141031103, "grad_norm": 13.0625, "learning_rate": 2.026619075302682e-06, "loss": 0.7239, "step": 6798 }, { "epoch": 0.5793779292714103, "grad_norm": 20.375, "learning_rate": 2.0264873995700203e-06, "loss": 1.0197, "step": 6799 }, { "epoch": 0.5794631444397103, "grad_norm": 16.75, "learning_rate": 2.0263557098054503e-06, "loss": 0.992, "step": 6800 }, { "epoch": 0.5795483596080102, "grad_norm": 14.3125, "learning_rate": 2.0262240060113517e-06, "loss": 0.5332, "step": 6801 }, { "epoch": 0.5796335747763102, "grad_norm": 16.125, "learning_rate": 2.0260922881901055e-06, "loss": 0.488, "step": 6802 }, { "epoch": 0.5797187899446101, "grad_norm": 13.875, "learning_rate": 2.0259605563440904e-06, "loss": 0.8797, "step": 6803 }, { "epoch": 0.5798040051129101, "grad_norm": 13.1875, "learning_rate": 2.025828810475688e-06, "loss": 0.5762, "step": 6804 }, { "epoch": 0.57988922028121, "grad_norm": 15.625, "learning_rate": 2.0256970505872793e-06, "loss": 0.8089, "step": 6805 }, { "epoch": 0.57997443544951, "grad_norm": 12.0, "learning_rate": 2.0255652766812444e-06, "loss": 0.3175, "step": 6806 }, { "epoch": 0.58005965061781, "grad_norm": 16.25, "learning_rate": 2.0254334887599652e-06, "loss": 0.9716, "step": 6807 }, { "epoch": 0.5801448657861099, "grad_norm": 14.0625, "learning_rate": 2.025301686825823e-06, "loss": 0.7905, "step": 6808 }, { "epoch": 0.5802300809544099, "grad_norm": 15.4375, "learning_rate": 2.0251698708811996e-06, "loss": 0.7279, "step": 6809 }, { "epoch": 0.5803152961227098, "grad_norm": 24.25, "learning_rate": 2.0250380409284777e-06, "loss": 1.1662, "step": 6810 }, { "epoch": 0.5804005112910098, "grad_norm": 14.125, "learning_rate": 2.0249061969700386e-06, "loss": 0.603, "step": 6811 }, { "epoch": 0.5804857264593097, "grad_norm": 11.4375, "learning_rate": 2.0247743390082648e-06, "loss": 0.6623, "step": 6812 }, { "epoch": 0.5805709416276097, "grad_norm": 13.0625, "learning_rate": 2.0246424670455404e-06, "loss": 0.8346, "step": 6813 }, { "epoch": 0.5806561567959097, "grad_norm": 18.75, "learning_rate": 2.0245105810842476e-06, "loss": 0.6573, "step": 6814 }, { "epoch": 0.5807413719642096, "grad_norm": 11.9375, "learning_rate": 2.024378681126769e-06, "loss": 0.5768, "step": 6815 }, { "epoch": 0.5808265871325096, "grad_norm": 14.375, "learning_rate": 2.02424676717549e-06, "loss": 0.7226, "step": 6816 }, { "epoch": 0.5809118023008095, "grad_norm": 13.0625, "learning_rate": 2.0241148392327926e-06, "loss": 0.8274, "step": 6817 }, { "epoch": 0.5809970174691095, "grad_norm": 14.875, "learning_rate": 2.0239828973010615e-06, "loss": 0.8791, "step": 6818 }, { "epoch": 0.5810822326374094, "grad_norm": 12.3125, "learning_rate": 2.023850941382682e-06, "loss": 0.3899, "step": 6819 }, { "epoch": 0.5811674478057094, "grad_norm": 12.25, "learning_rate": 2.023718971480037e-06, "loss": 0.8178, "step": 6820 }, { "epoch": 0.5812526629740093, "grad_norm": 11.5, "learning_rate": 2.0235869875955123e-06, "loss": 0.3628, "step": 6821 }, { "epoch": 0.5813378781423093, "grad_norm": 15.75, "learning_rate": 2.0234549897314933e-06, "loss": 0.6951, "step": 6822 }, { "epoch": 0.5814230933106093, "grad_norm": 17.375, "learning_rate": 2.0233229778903646e-06, "loss": 0.7554, "step": 6823 }, { "epoch": 0.5815083084789092, "grad_norm": 12.4375, "learning_rate": 2.023190952074512e-06, "loss": 0.5659, "step": 6824 }, { "epoch": 0.5815935236472092, "grad_norm": 16.125, "learning_rate": 2.0230589122863215e-06, "loss": 1.1069, "step": 6825 }, { "epoch": 0.5816787388155091, "grad_norm": 14.0625, "learning_rate": 2.022926858528179e-06, "loss": 0.8917, "step": 6826 }, { "epoch": 0.5817639539838091, "grad_norm": 13.1875, "learning_rate": 2.022794790802471e-06, "loss": 0.656, "step": 6827 }, { "epoch": 0.581849169152109, "grad_norm": 26.125, "learning_rate": 2.0226627091115837e-06, "loss": 1.0499, "step": 6828 }, { "epoch": 0.581934384320409, "grad_norm": 15.6875, "learning_rate": 2.0225306134579044e-06, "loss": 0.6488, "step": 6829 }, { "epoch": 0.582019599488709, "grad_norm": 42.5, "learning_rate": 2.0223985038438202e-06, "loss": 1.4108, "step": 6830 }, { "epoch": 0.5821048146570089, "grad_norm": 20.75, "learning_rate": 2.0222663802717185e-06, "loss": 0.899, "step": 6831 }, { "epoch": 0.5821900298253089, "grad_norm": 13.875, "learning_rate": 2.0221342427439867e-06, "loss": 0.6804, "step": 6832 }, { "epoch": 0.5822752449936088, "grad_norm": 12.4375, "learning_rate": 2.022002091263012e-06, "loss": 0.547, "step": 6833 }, { "epoch": 0.5823604601619088, "grad_norm": 13.1875, "learning_rate": 2.0218699258311843e-06, "loss": 0.5511, "step": 6834 }, { "epoch": 0.5824456753302087, "grad_norm": 16.625, "learning_rate": 2.0217377464508904e-06, "loss": 0.9553, "step": 6835 }, { "epoch": 0.5825308904985087, "grad_norm": 16.75, "learning_rate": 2.021605553124519e-06, "loss": 0.6424, "step": 6836 }, { "epoch": 0.5826161056668087, "grad_norm": 13.9375, "learning_rate": 2.0214733458544595e-06, "loss": 0.7372, "step": 6837 }, { "epoch": 0.5827013208351086, "grad_norm": 19.375, "learning_rate": 2.0213411246431014e-06, "loss": 0.7079, "step": 6838 }, { "epoch": 0.5827865360034086, "grad_norm": 12.375, "learning_rate": 2.0212088894928328e-06, "loss": 0.6697, "step": 6839 }, { "epoch": 0.5828717511717085, "grad_norm": 19.125, "learning_rate": 2.0210766404060443e-06, "loss": 1.0168, "step": 6840 }, { "epoch": 0.5829569663400085, "grad_norm": 16.75, "learning_rate": 2.020944377385126e-06, "loss": 0.689, "step": 6841 }, { "epoch": 0.5830421815083084, "grad_norm": 12.75, "learning_rate": 2.020812100432467e-06, "loss": 0.3973, "step": 6842 }, { "epoch": 0.5831273966766084, "grad_norm": 10.6875, "learning_rate": 2.0206798095504587e-06, "loss": 0.2996, "step": 6843 }, { "epoch": 0.5832126118449084, "grad_norm": 14.375, "learning_rate": 2.020547504741491e-06, "loss": 0.5681, "step": 6844 }, { "epoch": 0.5832978270132083, "grad_norm": 19.5, "learning_rate": 2.0204151860079545e-06, "loss": 0.9279, "step": 6845 }, { "epoch": 0.5833830421815083, "grad_norm": 17.875, "learning_rate": 2.0202828533522415e-06, "loss": 0.6386, "step": 6846 }, { "epoch": 0.5834682573498082, "grad_norm": 13.4375, "learning_rate": 2.020150506776743e-06, "loss": 0.7953, "step": 6847 }, { "epoch": 0.5835534725181082, "grad_norm": 10.6875, "learning_rate": 2.02001814628385e-06, "loss": 0.5392, "step": 6848 }, { "epoch": 0.5836386876864081, "grad_norm": 18.0, "learning_rate": 2.0198857718759545e-06, "loss": 1.1605, "step": 6849 }, { "epoch": 0.5837239028547081, "grad_norm": 13.3125, "learning_rate": 2.0197533835554496e-06, "loss": 0.5234, "step": 6850 }, { "epoch": 0.5838091180230081, "grad_norm": 15.75, "learning_rate": 2.0196209813247265e-06, "loss": 0.8265, "step": 6851 }, { "epoch": 0.583894333191308, "grad_norm": 12.8125, "learning_rate": 2.0194885651861785e-06, "loss": 0.7643, "step": 6852 }, { "epoch": 0.583979548359608, "grad_norm": 12.6875, "learning_rate": 2.0193561351421985e-06, "loss": 0.3946, "step": 6853 }, { "epoch": 0.5840647635279079, "grad_norm": 18.875, "learning_rate": 2.0192236911951794e-06, "loss": 1.0265, "step": 6854 }, { "epoch": 0.5841499786962079, "grad_norm": 18.375, "learning_rate": 2.019091233347515e-06, "loss": 1.0825, "step": 6855 }, { "epoch": 0.5842351938645078, "grad_norm": 12.0, "learning_rate": 2.018958761601598e-06, "loss": 0.6365, "step": 6856 }, { "epoch": 0.5843204090328078, "grad_norm": 13.75, "learning_rate": 2.018826275959823e-06, "loss": 0.4975, "step": 6857 }, { "epoch": 0.5844056242011078, "grad_norm": 16.75, "learning_rate": 2.0186937764245847e-06, "loss": 0.8235, "step": 6858 }, { "epoch": 0.5844908393694077, "grad_norm": 21.5, "learning_rate": 2.0185612629982765e-06, "loss": 0.6771, "step": 6859 }, { "epoch": 0.5845760545377077, "grad_norm": 15.25, "learning_rate": 2.018428735683293e-06, "loss": 0.788, "step": 6860 }, { "epoch": 0.5846612697060076, "grad_norm": 14.625, "learning_rate": 2.01829619448203e-06, "loss": 0.921, "step": 6861 }, { "epoch": 0.5847464848743076, "grad_norm": 17.25, "learning_rate": 2.0181636393968823e-06, "loss": 0.9471, "step": 6862 }, { "epoch": 0.5848317000426075, "grad_norm": 14.125, "learning_rate": 2.018031070430245e-06, "loss": 0.5874, "step": 6863 }, { "epoch": 0.5849169152109075, "grad_norm": 11.5625, "learning_rate": 2.017898487584514e-06, "loss": 0.6185, "step": 6864 }, { "epoch": 0.5850021303792075, "grad_norm": 17.0, "learning_rate": 2.0177658908620853e-06, "loss": 1.0259, "step": 6865 }, { "epoch": 0.5850873455475074, "grad_norm": 14.5, "learning_rate": 2.0176332802653542e-06, "loss": 0.8108, "step": 6866 }, { "epoch": 0.5851725607158074, "grad_norm": 17.125, "learning_rate": 2.0175006557967184e-06, "loss": 1.0411, "step": 6867 }, { "epoch": 0.5852577758841073, "grad_norm": 14.625, "learning_rate": 2.017368017458574e-06, "loss": 0.8712, "step": 6868 }, { "epoch": 0.5853429910524073, "grad_norm": 12.0, "learning_rate": 2.0172353652533177e-06, "loss": 0.6059, "step": 6869 }, { "epoch": 0.5854282062207072, "grad_norm": 17.25, "learning_rate": 2.017102699183347e-06, "loss": 0.8399, "step": 6870 }, { "epoch": 0.5855134213890072, "grad_norm": 16.75, "learning_rate": 2.016970019251059e-06, "loss": 0.8368, "step": 6871 }, { "epoch": 0.5855986365573072, "grad_norm": 11.1875, "learning_rate": 2.0168373254588514e-06, "loss": 0.4935, "step": 6872 }, { "epoch": 0.5856838517256071, "grad_norm": 10.5, "learning_rate": 2.0167046178091225e-06, "loss": 0.5049, "step": 6873 }, { "epoch": 0.5857690668939071, "grad_norm": 14.8125, "learning_rate": 2.01657189630427e-06, "loss": 0.7636, "step": 6874 }, { "epoch": 0.585854282062207, "grad_norm": 15.9375, "learning_rate": 2.0164391609466928e-06, "loss": 0.6882, "step": 6875 }, { "epoch": 0.585939497230507, "grad_norm": 21.125, "learning_rate": 2.0163064117387886e-06, "loss": 1.1915, "step": 6876 }, { "epoch": 0.5860247123988069, "grad_norm": 12.375, "learning_rate": 2.0161736486829576e-06, "loss": 0.5292, "step": 6877 }, { "epoch": 0.5861099275671069, "grad_norm": 28.125, "learning_rate": 2.0160408717815975e-06, "loss": 0.9053, "step": 6878 }, { "epoch": 0.5861951427354068, "grad_norm": 31.875, "learning_rate": 2.0159080810371094e-06, "loss": 0.8698, "step": 6879 }, { "epoch": 0.5862803579037068, "grad_norm": 23.0, "learning_rate": 2.0157752764518916e-06, "loss": 0.7061, "step": 6880 }, { "epoch": 0.5863655730720068, "grad_norm": 17.5, "learning_rate": 2.015642458028345e-06, "loss": 0.9361, "step": 6881 }, { "epoch": 0.5864507882403067, "grad_norm": 15.6875, "learning_rate": 2.015509625768869e-06, "loss": 0.8305, "step": 6882 }, { "epoch": 0.5865360034086067, "grad_norm": 15.9375, "learning_rate": 2.015376779675864e-06, "loss": 0.7644, "step": 6883 }, { "epoch": 0.5866212185769067, "grad_norm": 12.1875, "learning_rate": 2.015243919751731e-06, "loss": 0.5548, "step": 6884 }, { "epoch": 0.5867064337452067, "grad_norm": 14.25, "learning_rate": 2.0151110459988714e-06, "loss": 0.96, "step": 6885 }, { "epoch": 0.5867916489135067, "grad_norm": 12.125, "learning_rate": 2.0149781584196853e-06, "loss": 0.4868, "step": 6886 }, { "epoch": 0.5868768640818066, "grad_norm": 23.375, "learning_rate": 2.014845257016575e-06, "loss": 0.7223, "step": 6887 }, { "epoch": 0.5869620792501066, "grad_norm": 12.25, "learning_rate": 2.0147123417919412e-06, "loss": 0.5124, "step": 6888 }, { "epoch": 0.5870472944184065, "grad_norm": 12.5625, "learning_rate": 2.0145794127481868e-06, "loss": 0.4794, "step": 6889 }, { "epoch": 0.5871325095867065, "grad_norm": 18.25, "learning_rate": 2.0144464698877134e-06, "loss": 0.9951, "step": 6890 }, { "epoch": 0.5872177247550064, "grad_norm": 19.875, "learning_rate": 2.0143135132129243e-06, "loss": 0.7204, "step": 6891 }, { "epoch": 0.5873029399233064, "grad_norm": 19.0, "learning_rate": 2.0141805427262207e-06, "loss": 0.8634, "step": 6892 }, { "epoch": 0.5873881550916064, "grad_norm": 15.625, "learning_rate": 2.014047558430006e-06, "loss": 0.6356, "step": 6893 }, { "epoch": 0.5874733702599063, "grad_norm": 17.75, "learning_rate": 2.0139145603266843e-06, "loss": 0.6298, "step": 6894 }, { "epoch": 0.5875585854282063, "grad_norm": 16.5, "learning_rate": 2.0137815484186584e-06, "loss": 0.8022, "step": 6895 }, { "epoch": 0.5876438005965062, "grad_norm": 22.25, "learning_rate": 2.0136485227083314e-06, "loss": 1.3751, "step": 6896 }, { "epoch": 0.5877290157648062, "grad_norm": 19.875, "learning_rate": 2.0135154831981077e-06, "loss": 0.6795, "step": 6897 }, { "epoch": 0.5878142309331061, "grad_norm": 11.375, "learning_rate": 2.0133824298903915e-06, "loss": 0.5237, "step": 6898 }, { "epoch": 0.5878994461014061, "grad_norm": 13.4375, "learning_rate": 2.0132493627875875e-06, "loss": 0.5565, "step": 6899 }, { "epoch": 0.587984661269706, "grad_norm": 12.5, "learning_rate": 2.0131162818920997e-06, "loss": 0.8173, "step": 6900 }, { "epoch": 0.588069876438006, "grad_norm": 12.375, "learning_rate": 2.012983187206333e-06, "loss": 0.5747, "step": 6901 }, { "epoch": 0.588155091606306, "grad_norm": 13.5625, "learning_rate": 2.012850078732693e-06, "loss": 0.7632, "step": 6902 }, { "epoch": 0.5882403067746059, "grad_norm": 17.875, "learning_rate": 2.0127169564735853e-06, "loss": 0.8048, "step": 6903 }, { "epoch": 0.5883255219429059, "grad_norm": 42.75, "learning_rate": 2.0125838204314152e-06, "loss": 1.3908, "step": 6904 }, { "epoch": 0.5884107371112058, "grad_norm": 12.0, "learning_rate": 2.0124506706085887e-06, "loss": 0.4324, "step": 6905 }, { "epoch": 0.5884959522795058, "grad_norm": 14.9375, "learning_rate": 2.0123175070075113e-06, "loss": 0.6727, "step": 6906 }, { "epoch": 0.5885811674478058, "grad_norm": 16.625, "learning_rate": 2.012184329630591e-06, "loss": 0.9828, "step": 6907 }, { "epoch": 0.5886663826161057, "grad_norm": 14.875, "learning_rate": 2.0120511384802327e-06, "loss": 0.8603, "step": 6908 }, { "epoch": 0.5887515977844057, "grad_norm": 15.0625, "learning_rate": 2.011917933558844e-06, "loss": 0.5823, "step": 6909 }, { "epoch": 0.5888368129527056, "grad_norm": 15.1875, "learning_rate": 2.011784714868832e-06, "loss": 0.5983, "step": 6910 }, { "epoch": 0.5889220281210056, "grad_norm": 15.75, "learning_rate": 2.0116514824126047e-06, "loss": 0.5843, "step": 6911 }, { "epoch": 0.5890072432893055, "grad_norm": 16.125, "learning_rate": 2.011518236192569e-06, "loss": 0.7228, "step": 6912 }, { "epoch": 0.5890924584576055, "grad_norm": 10.6875, "learning_rate": 2.0113849762111328e-06, "loss": 0.5681, "step": 6913 }, { "epoch": 0.5891776736259055, "grad_norm": 16.25, "learning_rate": 2.0112517024707047e-06, "loss": 0.6344, "step": 6914 }, { "epoch": 0.5892628887942054, "grad_norm": 13.375, "learning_rate": 2.0111184149736925e-06, "loss": 0.6513, "step": 6915 }, { "epoch": 0.5893481039625054, "grad_norm": 15.4375, "learning_rate": 2.0109851137225055e-06, "loss": 0.7996, "step": 6916 }, { "epoch": 0.5894333191308053, "grad_norm": 15.9375, "learning_rate": 2.010851798719552e-06, "loss": 0.5699, "step": 6917 }, { "epoch": 0.5895185342991053, "grad_norm": 13.0625, "learning_rate": 2.010718469967242e-06, "loss": 0.534, "step": 6918 }, { "epoch": 0.5896037494674052, "grad_norm": 12.5625, "learning_rate": 2.0105851274679834e-06, "loss": 0.63, "step": 6919 }, { "epoch": 0.5896889646357052, "grad_norm": 10.25, "learning_rate": 2.0104517712241873e-06, "loss": 0.4163, "step": 6920 }, { "epoch": 0.5897741798040051, "grad_norm": 13.75, "learning_rate": 2.0103184012382625e-06, "loss": 0.6829, "step": 6921 }, { "epoch": 0.5898593949723051, "grad_norm": 13.125, "learning_rate": 2.01018501751262e-06, "loss": 0.6144, "step": 6922 }, { "epoch": 0.5899446101406051, "grad_norm": 13.8125, "learning_rate": 2.01005162004967e-06, "loss": 0.7802, "step": 6923 }, { "epoch": 0.590029825308905, "grad_norm": 23.0, "learning_rate": 2.0099182088518225e-06, "loss": 1.0607, "step": 6924 }, { "epoch": 0.590115040477205, "grad_norm": 16.125, "learning_rate": 2.009784783921489e-06, "loss": 0.6496, "step": 6925 }, { "epoch": 0.5902002556455049, "grad_norm": 14.375, "learning_rate": 2.0096513452610805e-06, "loss": 0.7623, "step": 6926 }, { "epoch": 0.5902854708138049, "grad_norm": 12.0, "learning_rate": 2.009517892873008e-06, "loss": 0.2801, "step": 6927 }, { "epoch": 0.5903706859821048, "grad_norm": 19.875, "learning_rate": 2.009384426759684e-06, "loss": 1.1026, "step": 6928 }, { "epoch": 0.5904559011504048, "grad_norm": 11.8125, "learning_rate": 2.0092509469235193e-06, "loss": 0.4951, "step": 6929 }, { "epoch": 0.5905411163187048, "grad_norm": 18.625, "learning_rate": 2.0091174533669267e-06, "loss": 0.9784, "step": 6930 }, { "epoch": 0.5906263314870047, "grad_norm": 15.25, "learning_rate": 2.0089839460923183e-06, "loss": 1.0034, "step": 6931 }, { "epoch": 0.5907115466553047, "grad_norm": 15.8125, "learning_rate": 2.0088504251021073e-06, "loss": 0.7972, "step": 6932 }, { "epoch": 0.5907967618236046, "grad_norm": 14.9375, "learning_rate": 2.008716890398705e-06, "loss": 0.7977, "step": 6933 }, { "epoch": 0.5908819769919046, "grad_norm": 12.0, "learning_rate": 2.0085833419845265e-06, "loss": 0.3837, "step": 6934 }, { "epoch": 0.5909671921602045, "grad_norm": 15.6875, "learning_rate": 2.0084497798619835e-06, "loss": 0.991, "step": 6935 }, { "epoch": 0.5910524073285045, "grad_norm": 11.9375, "learning_rate": 2.0083162040334906e-06, "loss": 0.5306, "step": 6936 }, { "epoch": 0.5911376224968045, "grad_norm": 13.625, "learning_rate": 2.0081826145014616e-06, "loss": 0.8169, "step": 6937 }, { "epoch": 0.5912228376651044, "grad_norm": 13.1875, "learning_rate": 2.00804901126831e-06, "loss": 0.5817, "step": 6938 }, { "epoch": 0.5913080528334044, "grad_norm": 12.5625, "learning_rate": 2.0079153943364504e-06, "loss": 0.6393, "step": 6939 }, { "epoch": 0.5913932680017043, "grad_norm": 18.75, "learning_rate": 2.007781763708298e-06, "loss": 0.882, "step": 6940 }, { "epoch": 0.5914784831700043, "grad_norm": 13.25, "learning_rate": 2.0076481193862667e-06, "loss": 0.852, "step": 6941 }, { "epoch": 0.5915636983383042, "grad_norm": 15.4375, "learning_rate": 2.0075144613727717e-06, "loss": 0.7561, "step": 6942 }, { "epoch": 0.5916489135066042, "grad_norm": 12.4375, "learning_rate": 2.007380789670229e-06, "loss": 0.5374, "step": 6943 }, { "epoch": 0.5917341286749042, "grad_norm": 14.8125, "learning_rate": 2.0072471042810542e-06, "loss": 0.7475, "step": 6944 }, { "epoch": 0.5918193438432041, "grad_norm": 15.625, "learning_rate": 2.0071134052076625e-06, "loss": 0.7733, "step": 6945 }, { "epoch": 0.5919045590115041, "grad_norm": 14.1875, "learning_rate": 2.00697969245247e-06, "loss": 1.0881, "step": 6946 }, { "epoch": 0.591989774179804, "grad_norm": 17.125, "learning_rate": 2.0068459660178932e-06, "loss": 1.1784, "step": 6947 }, { "epoch": 0.592074989348104, "grad_norm": 13.4375, "learning_rate": 2.006712225906349e-06, "loss": 0.4161, "step": 6948 }, { "epoch": 0.5921602045164039, "grad_norm": 32.5, "learning_rate": 2.006578472120254e-06, "loss": 0.9841, "step": 6949 }, { "epoch": 0.5922454196847039, "grad_norm": 13.0625, "learning_rate": 2.006444704662025e-06, "loss": 0.6029, "step": 6950 }, { "epoch": 0.5923306348530039, "grad_norm": 14.6875, "learning_rate": 2.0063109235340794e-06, "loss": 0.6538, "step": 6951 }, { "epoch": 0.5924158500213038, "grad_norm": 15.5625, "learning_rate": 2.0061771287388355e-06, "loss": 0.7806, "step": 6952 }, { "epoch": 0.5925010651896038, "grad_norm": 17.5, "learning_rate": 2.00604332027871e-06, "loss": 0.7478, "step": 6953 }, { "epoch": 0.5925862803579037, "grad_norm": 16.875, "learning_rate": 2.005909498156121e-06, "loss": 0.9095, "step": 6954 }, { "epoch": 0.5926714955262037, "grad_norm": 11.8125, "learning_rate": 2.005775662373488e-06, "loss": 0.5839, "step": 6955 }, { "epoch": 0.5927567106945036, "grad_norm": 14.5625, "learning_rate": 2.0056418129332283e-06, "loss": 0.4359, "step": 6956 }, { "epoch": 0.5928419258628036, "grad_norm": 19.0, "learning_rate": 2.005507949837762e-06, "loss": 0.7483, "step": 6957 }, { "epoch": 0.5929271410311036, "grad_norm": 15.9375, "learning_rate": 2.005374073089506e-06, "loss": 0.9208, "step": 6958 }, { "epoch": 0.5930123561994035, "grad_norm": 16.5, "learning_rate": 2.0052401826908817e-06, "loss": 0.5269, "step": 6959 }, { "epoch": 0.5930975713677035, "grad_norm": 16.125, "learning_rate": 2.005106278644308e-06, "loss": 0.4589, "step": 6960 }, { "epoch": 0.5931827865360034, "grad_norm": 15.6875, "learning_rate": 2.0049723609522043e-06, "loss": 0.535, "step": 6961 }, { "epoch": 0.5932680017043034, "grad_norm": 13.25, "learning_rate": 2.004838429616991e-06, "loss": 0.7606, "step": 6962 }, { "epoch": 0.5933532168726033, "grad_norm": 12.6875, "learning_rate": 2.004704484641088e-06, "loss": 0.5362, "step": 6963 }, { "epoch": 0.5934384320409033, "grad_norm": 18.0, "learning_rate": 2.0045705260269164e-06, "loss": 1.0528, "step": 6964 }, { "epoch": 0.5935236472092033, "grad_norm": 20.25, "learning_rate": 2.0044365537768962e-06, "loss": 1.1418, "step": 6965 }, { "epoch": 0.5936088623775032, "grad_norm": 14.5, "learning_rate": 2.004302567893449e-06, "loss": 0.7125, "step": 6966 }, { "epoch": 0.5936940775458032, "grad_norm": 34.75, "learning_rate": 2.0041685683789957e-06, "loss": 1.3935, "step": 6967 }, { "epoch": 0.5937792927141031, "grad_norm": 12.125, "learning_rate": 2.0040345552359587e-06, "loss": 0.3361, "step": 6968 }, { "epoch": 0.5938645078824031, "grad_norm": 16.375, "learning_rate": 2.0039005284667586e-06, "loss": 1.0431, "step": 6969 }, { "epoch": 0.593949723050703, "grad_norm": 12.375, "learning_rate": 2.0037664880738174e-06, "loss": 0.5184, "step": 6970 }, { "epoch": 0.594034938219003, "grad_norm": 12.0, "learning_rate": 2.003632434059558e-06, "loss": 0.3794, "step": 6971 }, { "epoch": 0.594120153387303, "grad_norm": 12.625, "learning_rate": 2.003498366426403e-06, "loss": 0.6815, "step": 6972 }, { "epoch": 0.5942053685556029, "grad_norm": 11.875, "learning_rate": 2.003364285176775e-06, "loss": 0.8421, "step": 6973 }, { "epoch": 0.5942905837239029, "grad_norm": 12.1875, "learning_rate": 2.003230190313096e-06, "loss": 0.5783, "step": 6974 }, { "epoch": 0.5943757988922028, "grad_norm": 12.75, "learning_rate": 2.0030960818377906e-06, "loss": 0.8218, "step": 6975 }, { "epoch": 0.5944610140605028, "grad_norm": 16.75, "learning_rate": 2.002961959753282e-06, "loss": 1.0087, "step": 6976 }, { "epoch": 0.5945462292288027, "grad_norm": 16.125, "learning_rate": 2.002827824061993e-06, "loss": 0.653, "step": 6977 }, { "epoch": 0.5946314443971027, "grad_norm": 12.5625, "learning_rate": 2.0026936747663484e-06, "loss": 0.5757, "step": 6978 }, { "epoch": 0.5947166595654026, "grad_norm": 20.75, "learning_rate": 2.002559511868772e-06, "loss": 1.0626, "step": 6979 }, { "epoch": 0.5948018747337026, "grad_norm": 15.0625, "learning_rate": 2.002425335371689e-06, "loss": 0.7168, "step": 6980 }, { "epoch": 0.5948870899020026, "grad_norm": 16.0, "learning_rate": 2.002291145277523e-06, "loss": 0.769, "step": 6981 }, { "epoch": 0.5949723050703025, "grad_norm": 19.25, "learning_rate": 2.0021569415887e-06, "loss": 0.717, "step": 6982 }, { "epoch": 0.5950575202386025, "grad_norm": 13.9375, "learning_rate": 2.0020227243076444e-06, "loss": 0.3966, "step": 6983 }, { "epoch": 0.5951427354069024, "grad_norm": 11.0625, "learning_rate": 2.001888493436782e-06, "loss": 0.6422, "step": 6984 }, { "epoch": 0.5952279505752024, "grad_norm": 16.875, "learning_rate": 2.0017542489785387e-06, "loss": 0.6107, "step": 6985 }, { "epoch": 0.5953131657435023, "grad_norm": 13.0, "learning_rate": 2.00161999093534e-06, "loss": 0.482, "step": 6986 }, { "epoch": 0.5953983809118023, "grad_norm": 18.375, "learning_rate": 2.001485719309612e-06, "loss": 0.6762, "step": 6987 }, { "epoch": 0.5954835960801023, "grad_norm": 15.1875, "learning_rate": 2.0013514341037817e-06, "loss": 0.8348, "step": 6988 }, { "epoch": 0.5955688112484022, "grad_norm": 19.5, "learning_rate": 2.0012171353202754e-06, "loss": 1.065, "step": 6989 }, { "epoch": 0.5956540264167022, "grad_norm": 15.875, "learning_rate": 2.00108282296152e-06, "loss": 0.7257, "step": 6990 }, { "epoch": 0.5957392415850021, "grad_norm": 10.9375, "learning_rate": 2.0009484970299424e-06, "loss": 0.3832, "step": 6991 }, { "epoch": 0.5958244567533021, "grad_norm": 13.0, "learning_rate": 2.0008141575279704e-06, "loss": 0.6952, "step": 6992 }, { "epoch": 0.595909671921602, "grad_norm": 16.0, "learning_rate": 2.0006798044580324e-06, "loss": 1.0096, "step": 6993 }, { "epoch": 0.595994887089902, "grad_norm": 15.9375, "learning_rate": 2.0005454378225547e-06, "loss": 0.7708, "step": 6994 }, { "epoch": 0.596080102258202, "grad_norm": 15.6875, "learning_rate": 2.000411057623966e-06, "loss": 0.7953, "step": 6995 }, { "epoch": 0.5961653174265019, "grad_norm": 13.75, "learning_rate": 2.000276663864695e-06, "loss": 0.6034, "step": 6996 }, { "epoch": 0.5962505325948019, "grad_norm": 15.9375, "learning_rate": 2.000142256547171e-06, "loss": 0.6687, "step": 6997 }, { "epoch": 0.5963357477631018, "grad_norm": 19.5, "learning_rate": 2.0000078356738215e-06, "loss": 1.3052, "step": 6998 }, { "epoch": 0.5964209629314018, "grad_norm": 12.6875, "learning_rate": 1.999873401247076e-06, "loss": 0.6121, "step": 6999 }, { "epoch": 0.5965061780997017, "grad_norm": 13.3125, "learning_rate": 1.9997389532693644e-06, "loss": 0.6854, "step": 7000 }, { "epoch": 0.5965913932680017, "grad_norm": 18.75, "learning_rate": 1.999604491743116e-06, "loss": 0.7471, "step": 7001 }, { "epoch": 0.5966766084363017, "grad_norm": 16.875, "learning_rate": 1.999470016670761e-06, "loss": 0.6186, "step": 7002 }, { "epoch": 0.5967618236046016, "grad_norm": 14.5625, "learning_rate": 1.9993355280547287e-06, "loss": 0.6835, "step": 7003 }, { "epoch": 0.5968470387729016, "grad_norm": 16.125, "learning_rate": 1.9992010258974497e-06, "loss": 0.9148, "step": 7004 }, { "epoch": 0.5969322539412015, "grad_norm": 16.125, "learning_rate": 1.9990665102013553e-06, "loss": 0.9025, "step": 7005 }, { "epoch": 0.5970174691095015, "grad_norm": 10.875, "learning_rate": 1.9989319809688757e-06, "loss": 0.4374, "step": 7006 }, { "epoch": 0.5971026842778014, "grad_norm": 15.75, "learning_rate": 1.9987974382024418e-06, "loss": 1.1291, "step": 7007 }, { "epoch": 0.5971878994461014, "grad_norm": 14.625, "learning_rate": 1.9986628819044854e-06, "loss": 0.6285, "step": 7008 }, { "epoch": 0.5972731146144014, "grad_norm": 17.5, "learning_rate": 1.998528312077438e-06, "loss": 0.8571, "step": 7009 }, { "epoch": 0.5973583297827013, "grad_norm": 15.9375, "learning_rate": 1.998393728723731e-06, "loss": 0.7511, "step": 7010 }, { "epoch": 0.5974435449510013, "grad_norm": 21.125, "learning_rate": 1.9982591318457968e-06, "loss": 1.0017, "step": 7011 }, { "epoch": 0.5975287601193012, "grad_norm": 14.25, "learning_rate": 1.9981245214460678e-06, "loss": 1.0681, "step": 7012 }, { "epoch": 0.5976139752876012, "grad_norm": 11.3125, "learning_rate": 1.9979898975269767e-06, "loss": 0.3837, "step": 7013 }, { "epoch": 0.5976991904559011, "grad_norm": 11.25, "learning_rate": 1.9978552600909555e-06, "loss": 0.4004, "step": 7014 }, { "epoch": 0.5977844056242011, "grad_norm": 14.5625, "learning_rate": 1.9977206091404377e-06, "loss": 0.6966, "step": 7015 }, { "epoch": 0.597869620792501, "grad_norm": 25.75, "learning_rate": 1.9975859446778567e-06, "loss": 0.8567, "step": 7016 }, { "epoch": 0.597954835960801, "grad_norm": 15.375, "learning_rate": 1.997451266705646e-06, "loss": 0.8897, "step": 7017 }, { "epoch": 0.598040051129101, "grad_norm": 14.75, "learning_rate": 1.9973165752262395e-06, "loss": 0.887, "step": 7018 }, { "epoch": 0.5981252662974009, "grad_norm": 14.75, "learning_rate": 1.997181870242071e-06, "loss": 0.6838, "step": 7019 }, { "epoch": 0.5982104814657009, "grad_norm": 11.5625, "learning_rate": 1.9970471517555743e-06, "loss": 0.6087, "step": 7020 }, { "epoch": 0.5982956966340008, "grad_norm": 16.625, "learning_rate": 1.996912419769185e-06, "loss": 0.9808, "step": 7021 }, { "epoch": 0.5983809118023008, "grad_norm": 15.875, "learning_rate": 1.9967776742853364e-06, "loss": 0.4564, "step": 7022 }, { "epoch": 0.5984661269706008, "grad_norm": 21.75, "learning_rate": 1.996642915306465e-06, "loss": 1.0245, "step": 7023 }, { "epoch": 0.5985513421389007, "grad_norm": 12.5625, "learning_rate": 1.9965081428350046e-06, "loss": 0.761, "step": 7024 }, { "epoch": 0.5986365573072007, "grad_norm": 16.25, "learning_rate": 1.9963733568733917e-06, "loss": 1.0195, "step": 7025 }, { "epoch": 0.5987217724755006, "grad_norm": 13.125, "learning_rate": 1.9962385574240616e-06, "loss": 0.5404, "step": 7026 }, { "epoch": 0.5988069876438006, "grad_norm": 14.6875, "learning_rate": 1.996103744489451e-06, "loss": 0.6135, "step": 7027 }, { "epoch": 0.5988922028121005, "grad_norm": 19.0, "learning_rate": 1.995968918071995e-06, "loss": 0.9434, "step": 7028 }, { "epoch": 0.5989774179804005, "grad_norm": 15.5625, "learning_rate": 1.9958340781741302e-06, "loss": 0.9301, "step": 7029 }, { "epoch": 0.5990626331487005, "grad_norm": 18.25, "learning_rate": 1.995699224798294e-06, "loss": 0.7505, "step": 7030 }, { "epoch": 0.5991478483170004, "grad_norm": 10.25, "learning_rate": 1.995564357946923e-06, "loss": 0.5414, "step": 7031 }, { "epoch": 0.5992330634853004, "grad_norm": 13.1875, "learning_rate": 1.995429477622454e-06, "loss": 0.9013, "step": 7032 }, { "epoch": 0.5993182786536003, "grad_norm": 16.625, "learning_rate": 1.9952945838273256e-06, "loss": 1.0507, "step": 7033 }, { "epoch": 0.5994034938219003, "grad_norm": 11.875, "learning_rate": 1.9951596765639734e-06, "loss": 0.5178, "step": 7034 }, { "epoch": 0.5994887089902002, "grad_norm": 13.5625, "learning_rate": 1.9950247558348374e-06, "loss": 0.6673, "step": 7035 }, { "epoch": 0.5995739241585002, "grad_norm": 15.5625, "learning_rate": 1.9948898216423546e-06, "loss": 0.817, "step": 7036 }, { "epoch": 0.5996591393268002, "grad_norm": 14.5625, "learning_rate": 1.994754873988964e-06, "loss": 0.548, "step": 7037 }, { "epoch": 0.5997443544951001, "grad_norm": 12.6875, "learning_rate": 1.9946199128771036e-06, "loss": 0.7274, "step": 7038 }, { "epoch": 0.5998295696634001, "grad_norm": 11.6875, "learning_rate": 1.9944849383092123e-06, "loss": 0.5219, "step": 7039 }, { "epoch": 0.5999147848317, "grad_norm": 35.75, "learning_rate": 1.99434995028773e-06, "loss": 0.9749, "step": 7040 }, { "epoch": 0.6, "grad_norm": 23.125, "learning_rate": 1.994214948815095e-06, "loss": 0.8435, "step": 7041 }, { "epoch": 0.6000852151682999, "grad_norm": 13.375, "learning_rate": 1.994079933893748e-06, "loss": 0.8088, "step": 7042 }, { "epoch": 0.6001704303365999, "grad_norm": 17.375, "learning_rate": 1.9939449055261284e-06, "loss": 0.513, "step": 7043 }, { "epoch": 0.6002556455048998, "grad_norm": 18.625, "learning_rate": 1.993809863714676e-06, "loss": 1.2511, "step": 7044 }, { "epoch": 0.6003408606731998, "grad_norm": 13.4375, "learning_rate": 1.993674808461832e-06, "loss": 0.5832, "step": 7045 }, { "epoch": 0.6004260758414998, "grad_norm": 14.0, "learning_rate": 1.993539739770036e-06, "loss": 0.7236, "step": 7046 }, { "epoch": 0.6005112910097997, "grad_norm": 10.75, "learning_rate": 1.993404657641729e-06, "loss": 0.3398, "step": 7047 }, { "epoch": 0.6005965061780997, "grad_norm": 16.875, "learning_rate": 1.9932695620793525e-06, "loss": 0.9036, "step": 7048 }, { "epoch": 0.6006817213463996, "grad_norm": 19.625, "learning_rate": 1.993134453085348e-06, "loss": 1.061, "step": 7049 }, { "epoch": 0.6007669365146996, "grad_norm": 14.5625, "learning_rate": 1.992999330662156e-06, "loss": 0.568, "step": 7050 }, { "epoch": 0.6008521516829995, "grad_norm": 10.75, "learning_rate": 1.992864194812219e-06, "loss": 0.411, "step": 7051 }, { "epoch": 0.6009373668512995, "grad_norm": 10.3125, "learning_rate": 1.9927290455379792e-06, "loss": 0.3216, "step": 7052 }, { "epoch": 0.6010225820195995, "grad_norm": 25.375, "learning_rate": 1.992593882841879e-06, "loss": 0.9012, "step": 7053 }, { "epoch": 0.6011077971878994, "grad_norm": 26.0, "learning_rate": 1.9924587067263603e-06, "loss": 1.1129, "step": 7054 }, { "epoch": 0.6011930123561994, "grad_norm": 13.125, "learning_rate": 1.992323517193866e-06, "loss": 0.7143, "step": 7055 }, { "epoch": 0.6012782275244993, "grad_norm": 18.625, "learning_rate": 1.992188314246839e-06, "loss": 0.7174, "step": 7056 }, { "epoch": 0.6013634426927993, "grad_norm": 11.375, "learning_rate": 1.9920530978877235e-06, "loss": 0.4578, "step": 7057 }, { "epoch": 0.6014486578610992, "grad_norm": 15.8125, "learning_rate": 1.991917868118962e-06, "loss": 0.8299, "step": 7058 }, { "epoch": 0.6015338730293992, "grad_norm": 15.4375, "learning_rate": 1.991782624942998e-06, "loss": 0.9186, "step": 7059 }, { "epoch": 0.6016190881976992, "grad_norm": 13.875, "learning_rate": 1.9916473683622773e-06, "loss": 0.6989, "step": 7060 }, { "epoch": 0.6017043033659991, "grad_norm": 13.375, "learning_rate": 1.991512098379242e-06, "loss": 0.4594, "step": 7061 }, { "epoch": 0.6017895185342991, "grad_norm": 13.9375, "learning_rate": 1.991376814996338e-06, "loss": 0.5327, "step": 7062 }, { "epoch": 0.601874733702599, "grad_norm": 16.125, "learning_rate": 1.991241518216009e-06, "loss": 1.1869, "step": 7063 }, { "epoch": 0.601959948870899, "grad_norm": 14.125, "learning_rate": 1.9911062080407003e-06, "loss": 0.703, "step": 7064 }, { "epoch": 0.6020451640391989, "grad_norm": 12.6875, "learning_rate": 1.9909708844728578e-06, "loss": 0.7683, "step": 7065 }, { "epoch": 0.6021303792074989, "grad_norm": 12.6875, "learning_rate": 1.9908355475149257e-06, "loss": 0.6584, "step": 7066 }, { "epoch": 0.6022155943757989, "grad_norm": 17.125, "learning_rate": 1.9907001971693503e-06, "loss": 0.9931, "step": 7067 }, { "epoch": 0.6023008095440988, "grad_norm": 13.6875, "learning_rate": 1.990564833438578e-06, "loss": 0.7852, "step": 7068 }, { "epoch": 0.6023860247123988, "grad_norm": 12.625, "learning_rate": 1.9904294563250537e-06, "loss": 0.533, "step": 7069 }, { "epoch": 0.6024712398806987, "grad_norm": 17.0, "learning_rate": 1.9902940658312254e-06, "loss": 0.7965, "step": 7070 }, { "epoch": 0.6025564550489987, "grad_norm": 17.875, "learning_rate": 1.9901586619595383e-06, "loss": 1.1383, "step": 7071 }, { "epoch": 0.6026416702172986, "grad_norm": 14.8125, "learning_rate": 1.9900232447124403e-06, "loss": 0.6484, "step": 7072 }, { "epoch": 0.6027268853855986, "grad_norm": 19.75, "learning_rate": 1.9898878140923777e-06, "loss": 1.0655, "step": 7073 }, { "epoch": 0.6028121005538986, "grad_norm": 20.625, "learning_rate": 1.9897523701017983e-06, "loss": 0.9423, "step": 7074 }, { "epoch": 0.6028973157221985, "grad_norm": 14.375, "learning_rate": 1.98961691274315e-06, "loss": 0.7268, "step": 7075 }, { "epoch": 0.6029825308904985, "grad_norm": 14.125, "learning_rate": 1.9894814420188808e-06, "loss": 0.7933, "step": 7076 }, { "epoch": 0.6030677460587984, "grad_norm": 20.75, "learning_rate": 1.989345957931437e-06, "loss": 0.9622, "step": 7077 }, { "epoch": 0.6031529612270984, "grad_norm": 13.8125, "learning_rate": 1.9892104604832698e-06, "loss": 0.591, "step": 7078 }, { "epoch": 0.6032381763953983, "grad_norm": 15.25, "learning_rate": 1.9890749496768253e-06, "loss": 0.698, "step": 7079 }, { "epoch": 0.6033233915636983, "grad_norm": 14.1875, "learning_rate": 1.9889394255145534e-06, "loss": 1.0469, "step": 7080 }, { "epoch": 0.6034086067319983, "grad_norm": 16.875, "learning_rate": 1.988803887998903e-06, "loss": 0.8548, "step": 7081 }, { "epoch": 0.6034938219002982, "grad_norm": 12.5625, "learning_rate": 1.988668337132324e-06, "loss": 0.6733, "step": 7082 }, { "epoch": 0.6035790370685982, "grad_norm": 17.625, "learning_rate": 1.9885327729172642e-06, "loss": 0.8537, "step": 7083 }, { "epoch": 0.6036642522368981, "grad_norm": 16.75, "learning_rate": 1.9883971953561752e-06, "loss": 0.8663, "step": 7084 }, { "epoch": 0.6037494674051981, "grad_norm": 13.1875, "learning_rate": 1.9882616044515064e-06, "loss": 0.844, "step": 7085 }, { "epoch": 0.603834682573498, "grad_norm": 15.125, "learning_rate": 1.9881260002057082e-06, "loss": 0.8308, "step": 7086 }, { "epoch": 0.603919897741798, "grad_norm": 15.75, "learning_rate": 1.987990382621231e-06, "loss": 0.8632, "step": 7087 }, { "epoch": 0.604005112910098, "grad_norm": 17.125, "learning_rate": 1.987854751700525e-06, "loss": 0.6834, "step": 7088 }, { "epoch": 0.6040903280783979, "grad_norm": 12.125, "learning_rate": 1.9877191074460416e-06, "loss": 0.7014, "step": 7089 }, { "epoch": 0.6041755432466979, "grad_norm": 16.75, "learning_rate": 1.9875834498602325e-06, "loss": 0.7654, "step": 7090 }, { "epoch": 0.6042607584149978, "grad_norm": 14.6875, "learning_rate": 1.987447778945549e-06, "loss": 0.7517, "step": 7091 }, { "epoch": 0.6043459735832978, "grad_norm": 14.25, "learning_rate": 1.987312094704442e-06, "loss": 1.0279, "step": 7092 }, { "epoch": 0.6044311887515977, "grad_norm": 18.375, "learning_rate": 1.987176397139364e-06, "loss": 0.6282, "step": 7093 }, { "epoch": 0.6045164039198977, "grad_norm": 18.125, "learning_rate": 1.987040686252768e-06, "loss": 0.845, "step": 7094 }, { "epoch": 0.6046016190881977, "grad_norm": 11.875, "learning_rate": 1.9869049620471048e-06, "loss": 0.4352, "step": 7095 }, { "epoch": 0.6046868342564976, "grad_norm": 13.3125, "learning_rate": 1.9867692245248283e-06, "loss": 0.4568, "step": 7096 }, { "epoch": 0.6047720494247976, "grad_norm": 12.9375, "learning_rate": 1.986633473688391e-06, "loss": 0.533, "step": 7097 }, { "epoch": 0.6048572645930975, "grad_norm": 13.9375, "learning_rate": 1.986497709540246e-06, "loss": 0.7173, "step": 7098 }, { "epoch": 0.6049424797613975, "grad_norm": 14.4375, "learning_rate": 1.986361932082847e-06, "loss": 0.7891, "step": 7099 }, { "epoch": 0.6050276949296974, "grad_norm": 15.75, "learning_rate": 1.986226141318647e-06, "loss": 0.6882, "step": 7100 }, { "epoch": 0.6051129100979974, "grad_norm": 19.0, "learning_rate": 1.986090337250101e-06, "loss": 0.9855, "step": 7101 }, { "epoch": 0.6051981252662973, "grad_norm": 19.75, "learning_rate": 1.9859545198796615e-06, "loss": 0.995, "step": 7102 }, { "epoch": 0.6052833404345973, "grad_norm": 27.5, "learning_rate": 1.9858186892097843e-06, "loss": 1.2267, "step": 7103 }, { "epoch": 0.6053685556028973, "grad_norm": 17.5, "learning_rate": 1.9856828452429234e-06, "loss": 0.9777, "step": 7104 }, { "epoch": 0.6054537707711972, "grad_norm": 14.6875, "learning_rate": 1.9855469879815335e-06, "loss": 0.7694, "step": 7105 }, { "epoch": 0.6055389859394972, "grad_norm": 17.375, "learning_rate": 1.9854111174280706e-06, "loss": 0.7356, "step": 7106 }, { "epoch": 0.6056242011077971, "grad_norm": 12.625, "learning_rate": 1.9852752335849886e-06, "loss": 0.6151, "step": 7107 }, { "epoch": 0.6057094162760971, "grad_norm": 22.375, "learning_rate": 1.9851393364547434e-06, "loss": 1.1245, "step": 7108 }, { "epoch": 0.6057946314443972, "grad_norm": 13.5625, "learning_rate": 1.985003426039792e-06, "loss": 0.6953, "step": 7109 }, { "epoch": 0.6058798466126971, "grad_norm": 18.0, "learning_rate": 1.984867502342589e-06, "loss": 0.6803, "step": 7110 }, { "epoch": 0.6059650617809971, "grad_norm": 10.0, "learning_rate": 1.9847315653655916e-06, "loss": 0.8856, "step": 7111 }, { "epoch": 0.606050276949297, "grad_norm": 14.9375, "learning_rate": 1.984595615111256e-06, "loss": 0.8024, "step": 7112 }, { "epoch": 0.606135492117597, "grad_norm": 17.75, "learning_rate": 1.9844596515820388e-06, "loss": 0.7841, "step": 7113 }, { "epoch": 0.6062207072858969, "grad_norm": 13.4375, "learning_rate": 1.984323674780397e-06, "loss": 0.8036, "step": 7114 }, { "epoch": 0.6063059224541969, "grad_norm": 11.4375, "learning_rate": 1.984187684708788e-06, "loss": 0.4548, "step": 7115 }, { "epoch": 0.6063911376224969, "grad_norm": 15.6875, "learning_rate": 1.98405168136967e-06, "loss": 0.7197, "step": 7116 }, { "epoch": 0.6064763527907968, "grad_norm": 19.0, "learning_rate": 1.983915664765499e-06, "loss": 0.9, "step": 7117 }, { "epoch": 0.6065615679590968, "grad_norm": 17.125, "learning_rate": 1.983779634898734e-06, "loss": 0.7093, "step": 7118 }, { "epoch": 0.6066467831273967, "grad_norm": 24.0, "learning_rate": 1.983643591771834e-06, "loss": 1.4911, "step": 7119 }, { "epoch": 0.6067319982956967, "grad_norm": 15.4375, "learning_rate": 1.983507535387256e-06, "loss": 1.0796, "step": 7120 }, { "epoch": 0.6068172134639966, "grad_norm": 13.4375, "learning_rate": 1.983371465747459e-06, "loss": 0.6264, "step": 7121 }, { "epoch": 0.6069024286322966, "grad_norm": 11.75, "learning_rate": 1.9832353828549025e-06, "loss": 0.4159, "step": 7122 }, { "epoch": 0.6069876438005966, "grad_norm": 16.0, "learning_rate": 1.9830992867120454e-06, "loss": 0.833, "step": 7123 }, { "epoch": 0.6070728589688965, "grad_norm": 12.5, "learning_rate": 1.9829631773213466e-06, "loss": 0.5283, "step": 7124 }, { "epoch": 0.6071580741371965, "grad_norm": 13.5625, "learning_rate": 1.982827054685266e-06, "loss": 0.7466, "step": 7125 }, { "epoch": 0.6072432893054964, "grad_norm": 15.0625, "learning_rate": 1.982690918806264e-06, "loss": 0.9566, "step": 7126 }, { "epoch": 0.6073285044737964, "grad_norm": 13.0625, "learning_rate": 1.9825547696868e-06, "loss": 0.5742, "step": 7127 }, { "epoch": 0.6074137196420963, "grad_norm": 16.0, "learning_rate": 1.9824186073293354e-06, "loss": 0.9767, "step": 7128 }, { "epoch": 0.6074989348103963, "grad_norm": 17.875, "learning_rate": 1.982282431736329e-06, "loss": 0.8745, "step": 7129 }, { "epoch": 0.6075841499786963, "grad_norm": 17.625, "learning_rate": 1.9821462429102435e-06, "loss": 1.0922, "step": 7130 }, { "epoch": 0.6076693651469962, "grad_norm": 10.5, "learning_rate": 1.9820100408535393e-06, "loss": 0.4651, "step": 7131 }, { "epoch": 0.6077545803152962, "grad_norm": 22.875, "learning_rate": 1.981873825568677e-06, "loss": 0.9883, "step": 7132 }, { "epoch": 0.6078397954835961, "grad_norm": 11.3125, "learning_rate": 1.9817375970581187e-06, "loss": 0.3495, "step": 7133 }, { "epoch": 0.6079250106518961, "grad_norm": 18.25, "learning_rate": 1.9816013553243265e-06, "loss": 1.1247, "step": 7134 }, { "epoch": 0.608010225820196, "grad_norm": 15.6875, "learning_rate": 1.981465100369762e-06, "loss": 0.785, "step": 7135 }, { "epoch": 0.608095440988496, "grad_norm": 16.375, "learning_rate": 1.981328832196888e-06, "loss": 0.7935, "step": 7136 }, { "epoch": 0.608180656156796, "grad_norm": 14.9375, "learning_rate": 1.981192550808166e-06, "loss": 0.9892, "step": 7137 }, { "epoch": 0.6082658713250959, "grad_norm": 15.0, "learning_rate": 1.98105625620606e-06, "loss": 0.8352, "step": 7138 }, { "epoch": 0.6083510864933959, "grad_norm": 13.5625, "learning_rate": 1.9809199483930317e-06, "loss": 0.4018, "step": 7139 }, { "epoch": 0.6084363016616958, "grad_norm": 13.875, "learning_rate": 1.980783627371545e-06, "loss": 0.7114, "step": 7140 }, { "epoch": 0.6085215168299958, "grad_norm": 13.1875, "learning_rate": 1.9806472931440634e-06, "loss": 0.3879, "step": 7141 }, { "epoch": 0.6086067319982957, "grad_norm": 11.625, "learning_rate": 1.9805109457130507e-06, "loss": 0.5047, "step": 7142 }, { "epoch": 0.6086919471665957, "grad_norm": 12.6875, "learning_rate": 1.98037458508097e-06, "loss": 0.6552, "step": 7143 }, { "epoch": 0.6087771623348956, "grad_norm": 25.25, "learning_rate": 1.9802382112502867e-06, "loss": 0.9809, "step": 7144 }, { "epoch": 0.6088623775031956, "grad_norm": 16.0, "learning_rate": 1.9801018242234644e-06, "loss": 0.8074, "step": 7145 }, { "epoch": 0.6089475926714956, "grad_norm": 17.625, "learning_rate": 1.979965424002968e-06, "loss": 0.7705, "step": 7146 }, { "epoch": 0.6090328078397955, "grad_norm": 13.875, "learning_rate": 1.9798290105912625e-06, "loss": 0.961, "step": 7147 }, { "epoch": 0.6091180230080955, "grad_norm": 14.375, "learning_rate": 1.9796925839908127e-06, "loss": 0.6139, "step": 7148 }, { "epoch": 0.6092032381763954, "grad_norm": 12.3125, "learning_rate": 1.979556144204084e-06, "loss": 0.5328, "step": 7149 }, { "epoch": 0.6092884533446954, "grad_norm": 12.5625, "learning_rate": 1.9794196912335422e-06, "loss": 0.4705, "step": 7150 }, { "epoch": 0.6093736685129953, "grad_norm": 9.9375, "learning_rate": 1.9792832250816534e-06, "loss": 0.3359, "step": 7151 }, { "epoch": 0.6094588836812953, "grad_norm": 12.0, "learning_rate": 1.9791467457508836e-06, "loss": 0.5211, "step": 7152 }, { "epoch": 0.6095440988495953, "grad_norm": 88.0, "learning_rate": 1.9790102532436984e-06, "loss": 1.0034, "step": 7153 }, { "epoch": 0.6096293140178952, "grad_norm": 15.5625, "learning_rate": 1.9788737475625645e-06, "loss": 0.8336, "step": 7154 }, { "epoch": 0.6097145291861952, "grad_norm": 19.875, "learning_rate": 1.9787372287099497e-06, "loss": 1.0461, "step": 7155 }, { "epoch": 0.6097997443544951, "grad_norm": 13.25, "learning_rate": 1.97860069668832e-06, "loss": 0.5638, "step": 7156 }, { "epoch": 0.6098849595227951, "grad_norm": 12.8125, "learning_rate": 1.9784641515001436e-06, "loss": 0.3514, "step": 7157 }, { "epoch": 0.609970174691095, "grad_norm": 15.375, "learning_rate": 1.9783275931478867e-06, "loss": 0.6992, "step": 7158 }, { "epoch": 0.610055389859395, "grad_norm": 19.125, "learning_rate": 1.9781910216340185e-06, "loss": 0.8709, "step": 7159 }, { "epoch": 0.610140605027695, "grad_norm": 11.1875, "learning_rate": 1.9780544369610055e-06, "loss": 0.4385, "step": 7160 }, { "epoch": 0.6102258201959949, "grad_norm": 22.625, "learning_rate": 1.9779178391313174e-06, "loss": 0.789, "step": 7161 }, { "epoch": 0.6103110353642949, "grad_norm": 12.5, "learning_rate": 1.977781228147422e-06, "loss": 0.6892, "step": 7162 }, { "epoch": 0.6103962505325948, "grad_norm": 13.125, "learning_rate": 1.9776446040117877e-06, "loss": 0.7682, "step": 7163 }, { "epoch": 0.6104814657008948, "grad_norm": 14.875, "learning_rate": 1.977507966726883e-06, "loss": 0.7456, "step": 7164 }, { "epoch": 0.6105666808691947, "grad_norm": 18.75, "learning_rate": 1.9773713162951787e-06, "loss": 0.9256, "step": 7165 }, { "epoch": 0.6106518960374947, "grad_norm": 15.5625, "learning_rate": 1.977234652719143e-06, "loss": 0.913, "step": 7166 }, { "epoch": 0.6107371112057947, "grad_norm": 17.25, "learning_rate": 1.977097976001246e-06, "loss": 0.9924, "step": 7167 }, { "epoch": 0.6108223263740946, "grad_norm": 17.875, "learning_rate": 1.9769612861439573e-06, "loss": 0.9682, "step": 7168 }, { "epoch": 0.6109075415423946, "grad_norm": 11.625, "learning_rate": 1.976824583149747e-06, "loss": 0.4987, "step": 7169 }, { "epoch": 0.6109927567106945, "grad_norm": 13.5625, "learning_rate": 1.976687867021086e-06, "loss": 0.8199, "step": 7170 }, { "epoch": 0.6110779718789945, "grad_norm": 10.625, "learning_rate": 1.976551137760444e-06, "loss": 0.5617, "step": 7171 }, { "epoch": 0.6111631870472944, "grad_norm": 17.125, "learning_rate": 1.976414395370293e-06, "loss": 0.8479, "step": 7172 }, { "epoch": 0.6112484022155944, "grad_norm": 14.5625, "learning_rate": 1.9762776398531027e-06, "loss": 0.9526, "step": 7173 }, { "epoch": 0.6113336173838944, "grad_norm": 16.875, "learning_rate": 1.9761408712113455e-06, "loss": 0.574, "step": 7174 }, { "epoch": 0.6114188325521943, "grad_norm": 11.625, "learning_rate": 1.9760040894474923e-06, "loss": 0.5072, "step": 7175 }, { "epoch": 0.6115040477204943, "grad_norm": 14.3125, "learning_rate": 1.9758672945640157e-06, "loss": 0.7083, "step": 7176 }, { "epoch": 0.6115892628887942, "grad_norm": 11.4375, "learning_rate": 1.9757304865633864e-06, "loss": 0.6307, "step": 7177 }, { "epoch": 0.6116744780570942, "grad_norm": 11.4375, "learning_rate": 1.975593665448078e-06, "loss": 0.3968, "step": 7178 }, { "epoch": 0.6117596932253941, "grad_norm": 13.8125, "learning_rate": 1.9754568312205625e-06, "loss": 0.5954, "step": 7179 }, { "epoch": 0.6118449083936941, "grad_norm": 17.375, "learning_rate": 1.975319983883312e-06, "loss": 0.9735, "step": 7180 }, { "epoch": 0.611930123561994, "grad_norm": 14.875, "learning_rate": 1.9751831234388004e-06, "loss": 0.7383, "step": 7181 }, { "epoch": 0.612015338730294, "grad_norm": 12.9375, "learning_rate": 1.9750462498895003e-06, "loss": 0.708, "step": 7182 }, { "epoch": 0.612100553898594, "grad_norm": 18.125, "learning_rate": 1.974909363237886e-06, "loss": 0.6898, "step": 7183 }, { "epoch": 0.6121857690668939, "grad_norm": 12.3125, "learning_rate": 1.9747724634864297e-06, "loss": 0.5799, "step": 7184 }, { "epoch": 0.6122709842351939, "grad_norm": 11.3125, "learning_rate": 1.9746355506376063e-06, "loss": 0.3727, "step": 7185 }, { "epoch": 0.6123561994034938, "grad_norm": 14.9375, "learning_rate": 1.9744986246938902e-06, "loss": 0.8564, "step": 7186 }, { "epoch": 0.6124414145717938, "grad_norm": 23.125, "learning_rate": 1.9743616856577557e-06, "loss": 0.9636, "step": 7187 }, { "epoch": 0.6125266297400938, "grad_norm": 13.25, "learning_rate": 1.9742247335316766e-06, "loss": 0.7136, "step": 7188 }, { "epoch": 0.6126118449083937, "grad_norm": 15.5625, "learning_rate": 1.974087768318128e-06, "loss": 0.9755, "step": 7189 }, { "epoch": 0.6126970600766937, "grad_norm": 15.0, "learning_rate": 1.973950790019586e-06, "loss": 0.7343, "step": 7190 }, { "epoch": 0.6127822752449936, "grad_norm": 20.75, "learning_rate": 1.973813798638525e-06, "loss": 0.9544, "step": 7191 }, { "epoch": 0.6128674904132936, "grad_norm": 11.75, "learning_rate": 1.973676794177421e-06, "loss": 0.49, "step": 7192 }, { "epoch": 0.6129527055815935, "grad_norm": 11.125, "learning_rate": 1.9735397766387494e-06, "loss": 0.3162, "step": 7193 }, { "epoch": 0.6130379207498935, "grad_norm": 11.5, "learning_rate": 1.9734027460249862e-06, "loss": 0.4049, "step": 7194 }, { "epoch": 0.6131231359181935, "grad_norm": 13.9375, "learning_rate": 1.9732657023386083e-06, "loss": 0.7312, "step": 7195 }, { "epoch": 0.6132083510864934, "grad_norm": 13.0, "learning_rate": 1.973128645582092e-06, "loss": 0.8564, "step": 7196 }, { "epoch": 0.6132935662547934, "grad_norm": 13.8125, "learning_rate": 1.9729915757579133e-06, "loss": 0.8346, "step": 7197 }, { "epoch": 0.6133787814230933, "grad_norm": 17.5, "learning_rate": 1.97285449286855e-06, "loss": 0.8359, "step": 7198 }, { "epoch": 0.6134639965913933, "grad_norm": 17.875, "learning_rate": 1.9727173969164797e-06, "loss": 0.7552, "step": 7199 }, { "epoch": 0.6135492117596932, "grad_norm": 12.0625, "learning_rate": 1.972580287904179e-06, "loss": 0.445, "step": 7200 }, { "epoch": 0.6136344269279932, "grad_norm": 14.5625, "learning_rate": 1.9724431658341256e-06, "loss": 0.9417, "step": 7201 }, { "epoch": 0.6137196420962931, "grad_norm": 15.375, "learning_rate": 1.972306030708798e-06, "loss": 0.6247, "step": 7202 }, { "epoch": 0.6138048572645931, "grad_norm": 15.75, "learning_rate": 1.9721688825306737e-06, "loss": 0.5556, "step": 7203 }, { "epoch": 0.6138900724328931, "grad_norm": 19.75, "learning_rate": 1.972031721302232e-06, "loss": 0.7019, "step": 7204 }, { "epoch": 0.613975287601193, "grad_norm": 13.4375, "learning_rate": 1.971894547025951e-06, "loss": 0.592, "step": 7205 }, { "epoch": 0.614060502769493, "grad_norm": 12.625, "learning_rate": 1.971757359704309e-06, "loss": 0.4118, "step": 7206 }, { "epoch": 0.6141457179377929, "grad_norm": 16.375, "learning_rate": 1.9716201593397863e-06, "loss": 0.9746, "step": 7207 }, { "epoch": 0.6142309331060929, "grad_norm": 13.375, "learning_rate": 1.9714829459348618e-06, "loss": 0.5896, "step": 7208 }, { "epoch": 0.6143161482743928, "grad_norm": 12.6875, "learning_rate": 1.9713457194920143e-06, "loss": 0.705, "step": 7209 }, { "epoch": 0.6144013634426928, "grad_norm": 14.4375, "learning_rate": 1.971208480013725e-06, "loss": 0.8955, "step": 7210 }, { "epoch": 0.6144865786109928, "grad_norm": 16.5, "learning_rate": 1.9710712275024726e-06, "loss": 0.9369, "step": 7211 }, { "epoch": 0.6145717937792927, "grad_norm": 23.625, "learning_rate": 1.9709339619607385e-06, "loss": 1.138, "step": 7212 }, { "epoch": 0.6146570089475927, "grad_norm": 18.0, "learning_rate": 1.9707966833910025e-06, "loss": 0.8803, "step": 7213 }, { "epoch": 0.6147422241158926, "grad_norm": 12.875, "learning_rate": 1.9706593917957457e-06, "loss": 0.3852, "step": 7214 }, { "epoch": 0.6148274392841926, "grad_norm": 11.75, "learning_rate": 1.970522087177449e-06, "loss": 0.3889, "step": 7215 }, { "epoch": 0.6149126544524925, "grad_norm": 11.875, "learning_rate": 1.970384769538594e-06, "loss": 0.5226, "step": 7216 }, { "epoch": 0.6149978696207925, "grad_norm": 14.75, "learning_rate": 1.970247438881661e-06, "loss": 0.6125, "step": 7217 }, { "epoch": 0.6150830847890925, "grad_norm": 15.0625, "learning_rate": 1.970110095209133e-06, "loss": 0.7232, "step": 7218 }, { "epoch": 0.6151682999573924, "grad_norm": 11.9375, "learning_rate": 1.9699727385234915e-06, "loss": 0.4787, "step": 7219 }, { "epoch": 0.6152535151256924, "grad_norm": 12.75, "learning_rate": 1.969835368827219e-06, "loss": 0.3575, "step": 7220 }, { "epoch": 0.6153387302939923, "grad_norm": 17.375, "learning_rate": 1.969697986122797e-06, "loss": 0.9611, "step": 7221 }, { "epoch": 0.6154239454622923, "grad_norm": 17.375, "learning_rate": 1.9695605904127085e-06, "loss": 0.8128, "step": 7222 }, { "epoch": 0.6155091606305922, "grad_norm": 24.5, "learning_rate": 1.969423181699437e-06, "loss": 1.3074, "step": 7223 }, { "epoch": 0.6155943757988922, "grad_norm": 15.0625, "learning_rate": 1.969285759985465e-06, "loss": 0.7217, "step": 7224 }, { "epoch": 0.6156795909671922, "grad_norm": 15.6875, "learning_rate": 1.969148325273276e-06, "loss": 0.8687, "step": 7225 }, { "epoch": 0.6157648061354921, "grad_norm": 13.3125, "learning_rate": 1.9690108775653535e-06, "loss": 0.771, "step": 7226 }, { "epoch": 0.6158500213037921, "grad_norm": 16.125, "learning_rate": 1.9688734168641816e-06, "loss": 0.8843, "step": 7227 }, { "epoch": 0.615935236472092, "grad_norm": 17.625, "learning_rate": 1.968735943172244e-06, "loss": 0.6593, "step": 7228 }, { "epoch": 0.616020451640392, "grad_norm": 17.375, "learning_rate": 1.9685984564920253e-06, "loss": 1.0138, "step": 7229 }, { "epoch": 0.6161056668086919, "grad_norm": 17.875, "learning_rate": 1.9684609568260097e-06, "loss": 0.9872, "step": 7230 }, { "epoch": 0.6161908819769919, "grad_norm": 9.25, "learning_rate": 1.9683234441766824e-06, "loss": 0.2184, "step": 7231 }, { "epoch": 0.6162760971452919, "grad_norm": 12.375, "learning_rate": 1.968185918546528e-06, "loss": 0.4289, "step": 7232 }, { "epoch": 0.6163613123135918, "grad_norm": 25.875, "learning_rate": 1.968048379938032e-06, "loss": 1.0873, "step": 7233 }, { "epoch": 0.6164465274818918, "grad_norm": 17.25, "learning_rate": 1.9679108283536796e-06, "loss": 0.9169, "step": 7234 }, { "epoch": 0.6165317426501917, "grad_norm": 10.125, "learning_rate": 1.967773263795957e-06, "loss": 0.4318, "step": 7235 }, { "epoch": 0.6166169578184917, "grad_norm": 14.25, "learning_rate": 1.967635686267349e-06, "loss": 0.6577, "step": 7236 }, { "epoch": 0.6167021729867916, "grad_norm": 12.125, "learning_rate": 1.967498095770343e-06, "loss": 0.5881, "step": 7237 }, { "epoch": 0.6167873881550916, "grad_norm": 13.75, "learning_rate": 1.967360492307425e-06, "loss": 0.5848, "step": 7238 }, { "epoch": 0.6168726033233916, "grad_norm": 18.375, "learning_rate": 1.967222875881081e-06, "loss": 1.015, "step": 7239 }, { "epoch": 0.6169578184916915, "grad_norm": 18.875, "learning_rate": 1.967085246493799e-06, "loss": 0.959, "step": 7240 }, { "epoch": 0.6170430336599915, "grad_norm": 14.8125, "learning_rate": 1.9669476041480652e-06, "loss": 0.4899, "step": 7241 }, { "epoch": 0.6171282488282914, "grad_norm": 15.0, "learning_rate": 1.9668099488463672e-06, "loss": 0.833, "step": 7242 }, { "epoch": 0.6172134639965914, "grad_norm": 18.75, "learning_rate": 1.966672280591193e-06, "loss": 0.8397, "step": 7243 }, { "epoch": 0.6172986791648913, "grad_norm": 16.125, "learning_rate": 1.96653459938503e-06, "loss": 0.8139, "step": 7244 }, { "epoch": 0.6173838943331913, "grad_norm": 16.75, "learning_rate": 1.966396905230366e-06, "loss": 0.9423, "step": 7245 }, { "epoch": 0.6174691095014913, "grad_norm": 13.0, "learning_rate": 1.966259198129689e-06, "loss": 0.3414, "step": 7246 }, { "epoch": 0.6175543246697912, "grad_norm": 13.0625, "learning_rate": 1.966121478085489e-06, "loss": 0.7184, "step": 7247 }, { "epoch": 0.6176395398380912, "grad_norm": 17.25, "learning_rate": 1.9659837451002532e-06, "loss": 1.2353, "step": 7248 }, { "epoch": 0.6177247550063911, "grad_norm": 22.5, "learning_rate": 1.9658459991764713e-06, "loss": 0.743, "step": 7249 }, { "epoch": 0.6178099701746911, "grad_norm": 10.875, "learning_rate": 1.9657082403166323e-06, "loss": 0.3962, "step": 7250 }, { "epoch": 0.617895185342991, "grad_norm": 19.5, "learning_rate": 1.9655704685232257e-06, "loss": 0.6572, "step": 7251 }, { "epoch": 0.617980400511291, "grad_norm": 21.5, "learning_rate": 1.9654326837987414e-06, "loss": 1.2073, "step": 7252 }, { "epoch": 0.618065615679591, "grad_norm": 20.875, "learning_rate": 1.965294886145669e-06, "loss": 1.2191, "step": 7253 }, { "epoch": 0.6181508308478909, "grad_norm": 13.9375, "learning_rate": 1.9651570755664983e-06, "loss": 0.8051, "step": 7254 }, { "epoch": 0.6182360460161909, "grad_norm": 12.375, "learning_rate": 1.965019252063721e-06, "loss": 0.5261, "step": 7255 }, { "epoch": 0.6183212611844908, "grad_norm": 15.3125, "learning_rate": 1.964881415639826e-06, "loss": 0.9768, "step": 7256 }, { "epoch": 0.6184064763527908, "grad_norm": 11.4375, "learning_rate": 1.964743566297305e-06, "loss": 0.4253, "step": 7257 }, { "epoch": 0.6184916915210907, "grad_norm": 12.25, "learning_rate": 1.964605704038649e-06, "loss": 0.4746, "step": 7258 }, { "epoch": 0.6185769066893907, "grad_norm": 11.6875, "learning_rate": 1.9644678288663497e-06, "loss": 0.6493, "step": 7259 }, { "epoch": 0.6186621218576907, "grad_norm": 11.875, "learning_rate": 1.964329940782898e-06, "loss": 0.5322, "step": 7260 }, { "epoch": 0.6187473370259906, "grad_norm": 15.1875, "learning_rate": 1.9641920397907864e-06, "loss": 0.6982, "step": 7261 }, { "epoch": 0.6188325521942906, "grad_norm": 12.125, "learning_rate": 1.9640541258925062e-06, "loss": 0.4661, "step": 7262 }, { "epoch": 0.6189177673625905, "grad_norm": 14.5, "learning_rate": 1.96391619909055e-06, "loss": 0.9326, "step": 7263 }, { "epoch": 0.6190029825308905, "grad_norm": 10.875, "learning_rate": 1.9637782593874097e-06, "loss": 0.4321, "step": 7264 }, { "epoch": 0.6190881976991904, "grad_norm": 12.8125, "learning_rate": 1.963640306785579e-06, "loss": 0.7314, "step": 7265 }, { "epoch": 0.6191734128674904, "grad_norm": 17.875, "learning_rate": 1.9635023412875497e-06, "loss": 0.442, "step": 7266 }, { "epoch": 0.6192586280357903, "grad_norm": 15.5, "learning_rate": 1.9633643628958163e-06, "loss": 0.8016, "step": 7267 }, { "epoch": 0.6193438432040903, "grad_norm": 16.875, "learning_rate": 1.9632263716128713e-06, "loss": 0.8831, "step": 7268 }, { "epoch": 0.6194290583723903, "grad_norm": 10.375, "learning_rate": 1.9630883674412086e-06, "loss": 0.2588, "step": 7269 }, { "epoch": 0.6195142735406902, "grad_norm": 17.25, "learning_rate": 1.962950350383322e-06, "loss": 0.9381, "step": 7270 }, { "epoch": 0.6195994887089902, "grad_norm": 13.1875, "learning_rate": 1.9628123204417055e-06, "loss": 0.928, "step": 7271 }, { "epoch": 0.6196847038772901, "grad_norm": 16.875, "learning_rate": 1.962674277618854e-06, "loss": 0.8597, "step": 7272 }, { "epoch": 0.6197699190455901, "grad_norm": 14.875, "learning_rate": 1.9625362219172616e-06, "loss": 0.9568, "step": 7273 }, { "epoch": 0.61985513421389, "grad_norm": 15.4375, "learning_rate": 1.9623981533394226e-06, "loss": 0.798, "step": 7274 }, { "epoch": 0.61994034938219, "grad_norm": 13.5, "learning_rate": 1.9622600718878328e-06, "loss": 0.6489, "step": 7275 }, { "epoch": 0.62002556455049, "grad_norm": 15.25, "learning_rate": 1.9621219775649873e-06, "loss": 0.6409, "step": 7276 }, { "epoch": 0.6201107797187899, "grad_norm": 15.25, "learning_rate": 1.961983870373382e-06, "loss": 0.775, "step": 7277 }, { "epoch": 0.6201959948870899, "grad_norm": 12.6875, "learning_rate": 1.961845750315512e-06, "loss": 0.5108, "step": 7278 }, { "epoch": 0.6202812100553898, "grad_norm": 11.0625, "learning_rate": 1.961707617393873e-06, "loss": 0.4268, "step": 7279 }, { "epoch": 0.6203664252236898, "grad_norm": 23.25, "learning_rate": 1.9615694716109622e-06, "loss": 1.4864, "step": 7280 }, { "epoch": 0.6204516403919897, "grad_norm": 13.125, "learning_rate": 1.961431312969275e-06, "loss": 0.7875, "step": 7281 }, { "epoch": 0.6205368555602897, "grad_norm": 15.1875, "learning_rate": 1.9612931414713092e-06, "loss": 0.6852, "step": 7282 }, { "epoch": 0.6206220707285897, "grad_norm": 16.5, "learning_rate": 1.9611549571195608e-06, "loss": 0.8346, "step": 7283 }, { "epoch": 0.6207072858968896, "grad_norm": 14.375, "learning_rate": 1.9610167599165268e-06, "loss": 0.6892, "step": 7284 }, { "epoch": 0.6207925010651896, "grad_norm": 13.625, "learning_rate": 1.960878549864706e-06, "loss": 0.6599, "step": 7285 }, { "epoch": 0.6208777162334895, "grad_norm": 15.6875, "learning_rate": 1.9607403269665943e-06, "loss": 1.1817, "step": 7286 }, { "epoch": 0.6209629314017895, "grad_norm": 15.6875, "learning_rate": 1.96060209122469e-06, "loss": 0.7801, "step": 7287 }, { "epoch": 0.6210481465700894, "grad_norm": 12.5625, "learning_rate": 1.960463842641491e-06, "loss": 0.6956, "step": 7288 }, { "epoch": 0.6211333617383894, "grad_norm": 14.5625, "learning_rate": 1.960325581219497e-06, "loss": 0.7612, "step": 7289 }, { "epoch": 0.6212185769066894, "grad_norm": 13.5, "learning_rate": 1.9601873069612047e-06, "loss": 0.4868, "step": 7290 }, { "epoch": 0.6213037920749893, "grad_norm": 17.375, "learning_rate": 1.960049019869114e-06, "loss": 0.7407, "step": 7291 }, { "epoch": 0.6213890072432893, "grad_norm": 13.375, "learning_rate": 1.9599107199457234e-06, "loss": 0.8536, "step": 7292 }, { "epoch": 0.6214742224115892, "grad_norm": 17.25, "learning_rate": 1.9597724071935324e-06, "loss": 0.927, "step": 7293 }, { "epoch": 0.6215594375798892, "grad_norm": 17.25, "learning_rate": 1.95963408161504e-06, "loss": 0.7374, "step": 7294 }, { "epoch": 0.6216446527481891, "grad_norm": 15.3125, "learning_rate": 1.959495743212746e-06, "loss": 1.0294, "step": 7295 }, { "epoch": 0.6217298679164891, "grad_norm": 9.25, "learning_rate": 1.9593573919891506e-06, "loss": 0.2492, "step": 7296 }, { "epoch": 0.621815083084789, "grad_norm": 14.4375, "learning_rate": 1.959219027946754e-06, "loss": 0.8218, "step": 7297 }, { "epoch": 0.621900298253089, "grad_norm": 13.125, "learning_rate": 1.959080651088056e-06, "loss": 0.6199, "step": 7298 }, { "epoch": 0.621985513421389, "grad_norm": 14.25, "learning_rate": 1.958942261415558e-06, "loss": 0.6136, "step": 7299 }, { "epoch": 0.6220707285896889, "grad_norm": 14.1875, "learning_rate": 1.9588038589317604e-06, "loss": 0.7423, "step": 7300 }, { "epoch": 0.6221559437579889, "grad_norm": 15.875, "learning_rate": 1.9586654436391644e-06, "loss": 0.7883, "step": 7301 }, { "epoch": 0.6222411589262888, "grad_norm": 14.0, "learning_rate": 1.958527015540271e-06, "loss": 0.7382, "step": 7302 }, { "epoch": 0.6223263740945888, "grad_norm": 12.3125, "learning_rate": 1.958388574637582e-06, "loss": 0.6697, "step": 7303 }, { "epoch": 0.6224115892628888, "grad_norm": 23.375, "learning_rate": 1.958250120933599e-06, "loss": 1.1391, "step": 7304 }, { "epoch": 0.6224968044311887, "grad_norm": 19.625, "learning_rate": 1.9581116544308243e-06, "loss": 0.8311, "step": 7305 }, { "epoch": 0.6225820195994887, "grad_norm": 13.3125, "learning_rate": 1.9579731751317597e-06, "loss": 0.5279, "step": 7306 }, { "epoch": 0.6226672347677886, "grad_norm": 12.8125, "learning_rate": 1.9578346830389083e-06, "loss": 0.6692, "step": 7307 }, { "epoch": 0.6227524499360886, "grad_norm": 15.4375, "learning_rate": 1.957696178154772e-06, "loss": 0.7927, "step": 7308 }, { "epoch": 0.6228376651043885, "grad_norm": 13.3125, "learning_rate": 1.9575576604818544e-06, "loss": 0.6831, "step": 7309 }, { "epoch": 0.6229228802726885, "grad_norm": 40.25, "learning_rate": 1.957419130022658e-06, "loss": 0.8341, "step": 7310 }, { "epoch": 0.6230080954409885, "grad_norm": 15.125, "learning_rate": 1.9572805867796866e-06, "loss": 1.093, "step": 7311 }, { "epoch": 0.6230933106092884, "grad_norm": 23.125, "learning_rate": 1.9571420307554446e-06, "loss": 0.8997, "step": 7312 }, { "epoch": 0.6231785257775884, "grad_norm": 18.625, "learning_rate": 1.957003461952434e-06, "loss": 1.2009, "step": 7313 }, { "epoch": 0.6232637409458883, "grad_norm": 11.75, "learning_rate": 1.9568648803731603e-06, "loss": 0.4673, "step": 7314 }, { "epoch": 0.6233489561141883, "grad_norm": 17.0, "learning_rate": 1.9567262860201273e-06, "loss": 0.7243, "step": 7315 }, { "epoch": 0.6234341712824882, "grad_norm": 14.0625, "learning_rate": 1.95658767889584e-06, "loss": 1.0163, "step": 7316 }, { "epoch": 0.6235193864507882, "grad_norm": 14.5, "learning_rate": 1.9564490590028026e-06, "loss": 0.8731, "step": 7317 }, { "epoch": 0.6236046016190882, "grad_norm": 14.375, "learning_rate": 1.9563104263435203e-06, "loss": 0.8048, "step": 7318 }, { "epoch": 0.6236898167873881, "grad_norm": 28.375, "learning_rate": 1.9561717809204984e-06, "loss": 0.8914, "step": 7319 }, { "epoch": 0.6237750319556881, "grad_norm": 11.1875, "learning_rate": 1.956033122736242e-06, "loss": 0.3814, "step": 7320 }, { "epoch": 0.623860247123988, "grad_norm": 17.375, "learning_rate": 1.9558944517932574e-06, "loss": 0.8612, "step": 7321 }, { "epoch": 0.623945462292288, "grad_norm": 30.25, "learning_rate": 1.9557557680940503e-06, "loss": 1.2453, "step": 7322 }, { "epoch": 0.6240306774605879, "grad_norm": 13.5, "learning_rate": 1.9556170716411266e-06, "loss": 0.5864, "step": 7323 }, { "epoch": 0.6241158926288879, "grad_norm": 12.0, "learning_rate": 1.9554783624369926e-06, "loss": 0.3767, "step": 7324 }, { "epoch": 0.6242011077971878, "grad_norm": 15.375, "learning_rate": 1.9553396404841555e-06, "loss": 0.6376, "step": 7325 }, { "epoch": 0.6242863229654878, "grad_norm": 13.0, "learning_rate": 1.9552009057851218e-06, "loss": 0.6214, "step": 7326 }, { "epoch": 0.6243715381337878, "grad_norm": 12.875, "learning_rate": 1.9550621583423985e-06, "loss": 0.7112, "step": 7327 }, { "epoch": 0.6244567533020877, "grad_norm": 15.5625, "learning_rate": 1.954923398158493e-06, "loss": 0.8004, "step": 7328 }, { "epoch": 0.6245419684703877, "grad_norm": 11.8125, "learning_rate": 1.954784625235913e-06, "loss": 0.4881, "step": 7329 }, { "epoch": 0.6246271836386876, "grad_norm": 11.9375, "learning_rate": 1.954645839577166e-06, "loss": 0.55, "step": 7330 }, { "epoch": 0.6247123988069876, "grad_norm": 12.4375, "learning_rate": 1.9545070411847604e-06, "loss": 0.4759, "step": 7331 }, { "epoch": 0.6247976139752875, "grad_norm": 12.5, "learning_rate": 1.9543682300612037e-06, "loss": 0.4297, "step": 7332 }, { "epoch": 0.6248828291435875, "grad_norm": 15.9375, "learning_rate": 1.954229406209005e-06, "loss": 0.8195, "step": 7333 }, { "epoch": 0.6249680443118875, "grad_norm": 11.125, "learning_rate": 1.954090569630673e-06, "loss": 0.5054, "step": 7334 }, { "epoch": 0.6250532594801875, "grad_norm": 16.375, "learning_rate": 1.9539517203287157e-06, "loss": 0.7437, "step": 7335 }, { "epoch": 0.6251384746484875, "grad_norm": 11.4375, "learning_rate": 1.9538128583056435e-06, "loss": 0.5343, "step": 7336 }, { "epoch": 0.6252236898167874, "grad_norm": 9.75, "learning_rate": 1.953673983563965e-06, "loss": 0.2756, "step": 7337 }, { "epoch": 0.6253089049850874, "grad_norm": 14.8125, "learning_rate": 1.9535350961061903e-06, "loss": 0.9079, "step": 7338 }, { "epoch": 0.6253941201533874, "grad_norm": 30.125, "learning_rate": 1.9533961959348284e-06, "loss": 1.1096, "step": 7339 }, { "epoch": 0.6254793353216873, "grad_norm": 12.3125, "learning_rate": 1.9532572830523904e-06, "loss": 0.8761, "step": 7340 }, { "epoch": 0.6255645504899873, "grad_norm": 11.8125, "learning_rate": 1.953118357461386e-06, "loss": 0.7214, "step": 7341 }, { "epoch": 0.6256497656582872, "grad_norm": 11.625, "learning_rate": 1.952979419164326e-06, "loss": 0.4724, "step": 7342 }, { "epoch": 0.6257349808265872, "grad_norm": 11.375, "learning_rate": 1.9528404681637205e-06, "loss": 0.4248, "step": 7343 }, { "epoch": 0.6258201959948871, "grad_norm": 18.5, "learning_rate": 1.9527015044620814e-06, "loss": 0.9084, "step": 7344 }, { "epoch": 0.6259054111631871, "grad_norm": 17.25, "learning_rate": 1.952562528061919e-06, "loss": 0.897, "step": 7345 }, { "epoch": 0.625990626331487, "grad_norm": 25.125, "learning_rate": 1.952423538965746e-06, "loss": 0.9288, "step": 7346 }, { "epoch": 0.626075841499787, "grad_norm": 74.5, "learning_rate": 1.952284537176073e-06, "loss": 0.8996, "step": 7347 }, { "epoch": 0.626161056668087, "grad_norm": 14.5625, "learning_rate": 1.9521455226954124e-06, "loss": 0.6105, "step": 7348 }, { "epoch": 0.6262462718363869, "grad_norm": 9.125, "learning_rate": 1.9520064955262756e-06, "loss": 0.6165, "step": 7349 }, { "epoch": 0.6263314870046869, "grad_norm": 17.0, "learning_rate": 1.951867455671176e-06, "loss": 0.8216, "step": 7350 }, { "epoch": 0.6264167021729868, "grad_norm": 12.1875, "learning_rate": 1.951728403132625e-06, "loss": 0.528, "step": 7351 }, { "epoch": 0.6265019173412868, "grad_norm": 13.5, "learning_rate": 1.951589337913137e-06, "loss": 0.582, "step": 7352 }, { "epoch": 0.6265871325095868, "grad_norm": 15.0, "learning_rate": 1.9514502600152236e-06, "loss": 0.7162, "step": 7353 }, { "epoch": 0.6266723476778867, "grad_norm": 18.75, "learning_rate": 1.951311169441399e-06, "loss": 0.9798, "step": 7354 }, { "epoch": 0.6267575628461867, "grad_norm": 15.8125, "learning_rate": 1.9511720661941757e-06, "loss": 0.8901, "step": 7355 }, { "epoch": 0.6268427780144866, "grad_norm": 12.1875, "learning_rate": 1.9510329502760687e-06, "loss": 0.5513, "step": 7356 }, { "epoch": 0.6269279931827866, "grad_norm": 12.8125, "learning_rate": 1.9508938216895913e-06, "loss": 0.643, "step": 7357 }, { "epoch": 0.6270132083510865, "grad_norm": 11.25, "learning_rate": 1.9507546804372574e-06, "loss": 0.6086, "step": 7358 }, { "epoch": 0.6270984235193865, "grad_norm": 16.25, "learning_rate": 1.950615526521582e-06, "loss": 1.0132, "step": 7359 }, { "epoch": 0.6271836386876865, "grad_norm": 13.5, "learning_rate": 1.9504763599450792e-06, "loss": 0.6273, "step": 7360 }, { "epoch": 0.6272688538559864, "grad_norm": 14.5625, "learning_rate": 1.9503371807102646e-06, "loss": 0.5556, "step": 7361 }, { "epoch": 0.6273540690242864, "grad_norm": 12.4375, "learning_rate": 1.9501979888196533e-06, "loss": 0.6749, "step": 7362 }, { "epoch": 0.6274392841925863, "grad_norm": 18.0, "learning_rate": 1.9500587842757594e-06, "loss": 0.6568, "step": 7363 }, { "epoch": 0.6275244993608863, "grad_norm": 11.375, "learning_rate": 1.9499195670810993e-06, "loss": 0.608, "step": 7364 }, { "epoch": 0.6276097145291862, "grad_norm": 18.25, "learning_rate": 1.949780337238189e-06, "loss": 0.8802, "step": 7365 }, { "epoch": 0.6276949296974862, "grad_norm": 14.6875, "learning_rate": 1.9496410947495447e-06, "loss": 0.5076, "step": 7366 }, { "epoch": 0.6277801448657861, "grad_norm": 13.125, "learning_rate": 1.949501839617682e-06, "loss": 0.6984, "step": 7367 }, { "epoch": 0.6278653600340861, "grad_norm": 16.375, "learning_rate": 1.9493625718451176e-06, "loss": 0.9003, "step": 7368 }, { "epoch": 0.6279505752023861, "grad_norm": 13.625, "learning_rate": 1.9492232914343685e-06, "loss": 0.6419, "step": 7369 }, { "epoch": 0.628035790370686, "grad_norm": 13.375, "learning_rate": 1.949083998387951e-06, "loss": 0.5162, "step": 7370 }, { "epoch": 0.628121005538986, "grad_norm": 11.875, "learning_rate": 1.948944692708383e-06, "loss": 0.4872, "step": 7371 }, { "epoch": 0.6282062207072859, "grad_norm": 14.875, "learning_rate": 1.9488053743981814e-06, "loss": 0.8946, "step": 7372 }, { "epoch": 0.6282914358755859, "grad_norm": 23.25, "learning_rate": 1.9486660434598638e-06, "loss": 1.0278, "step": 7373 }, { "epoch": 0.6283766510438858, "grad_norm": 16.125, "learning_rate": 1.948526699895949e-06, "loss": 0.929, "step": 7374 }, { "epoch": 0.6284618662121858, "grad_norm": 22.375, "learning_rate": 1.9483873437089536e-06, "loss": 0.8232, "step": 7375 }, { "epoch": 0.6285470813804858, "grad_norm": 13.5625, "learning_rate": 1.948247974901397e-06, "loss": 0.815, "step": 7376 }, { "epoch": 0.6286322965487857, "grad_norm": 12.0, "learning_rate": 1.948108593475797e-06, "loss": 0.3755, "step": 7377 }, { "epoch": 0.6287175117170857, "grad_norm": 17.75, "learning_rate": 1.9479691994346727e-06, "loss": 0.6934, "step": 7378 }, { "epoch": 0.6288027268853856, "grad_norm": 15.75, "learning_rate": 1.9478297927805433e-06, "loss": 0.7856, "step": 7379 }, { "epoch": 0.6288879420536856, "grad_norm": 14.3125, "learning_rate": 1.947690373515928e-06, "loss": 0.7932, "step": 7380 }, { "epoch": 0.6289731572219855, "grad_norm": 17.25, "learning_rate": 1.9475509416433457e-06, "loss": 0.7672, "step": 7381 }, { "epoch": 0.6290583723902855, "grad_norm": 10.5, "learning_rate": 1.947411497165316e-06, "loss": 0.3226, "step": 7382 }, { "epoch": 0.6291435875585855, "grad_norm": 11.4375, "learning_rate": 1.9472720400843604e-06, "loss": 0.5668, "step": 7383 }, { "epoch": 0.6292288027268854, "grad_norm": 17.5, "learning_rate": 1.9471325704029973e-06, "loss": 0.6698, "step": 7384 }, { "epoch": 0.6293140178951854, "grad_norm": 15.0, "learning_rate": 1.9469930881237474e-06, "loss": 0.929, "step": 7385 }, { "epoch": 0.6293992330634853, "grad_norm": 10.4375, "learning_rate": 1.946853593249132e-06, "loss": 0.3636, "step": 7386 }, { "epoch": 0.6294844482317853, "grad_norm": 16.5, "learning_rate": 1.946714085781671e-06, "loss": 0.7777, "step": 7387 }, { "epoch": 0.6295696634000852, "grad_norm": 15.9375, "learning_rate": 1.946574565723886e-06, "loss": 1.119, "step": 7388 }, { "epoch": 0.6296548785683852, "grad_norm": 14.25, "learning_rate": 1.9464350330782985e-06, "loss": 0.7391, "step": 7389 }, { "epoch": 0.6297400937366852, "grad_norm": 19.25, "learning_rate": 1.9462954878474295e-06, "loss": 0.9352, "step": 7390 }, { "epoch": 0.6298253089049851, "grad_norm": 11.8125, "learning_rate": 1.9461559300338005e-06, "loss": 0.6283, "step": 7391 }, { "epoch": 0.6299105240732851, "grad_norm": 15.75, "learning_rate": 1.946016359639934e-06, "loss": 0.4493, "step": 7392 }, { "epoch": 0.629995739241585, "grad_norm": 17.125, "learning_rate": 1.9458767766683517e-06, "loss": 0.5668, "step": 7393 }, { "epoch": 0.630080954409885, "grad_norm": 14.25, "learning_rate": 1.945737181121577e-06, "loss": 0.4691, "step": 7394 }, { "epoch": 0.6301661695781849, "grad_norm": 16.0, "learning_rate": 1.9455975730021308e-06, "loss": 0.7759, "step": 7395 }, { "epoch": 0.6302513847464849, "grad_norm": 14.125, "learning_rate": 1.9454579523125376e-06, "loss": 0.7293, "step": 7396 }, { "epoch": 0.6303365999147849, "grad_norm": 15.125, "learning_rate": 1.94531831905532e-06, "loss": 0.926, "step": 7397 }, { "epoch": 0.6304218150830848, "grad_norm": 13.75, "learning_rate": 1.9451786732330006e-06, "loss": 0.8237, "step": 7398 }, { "epoch": 0.6305070302513848, "grad_norm": 15.25, "learning_rate": 1.9450390148481043e-06, "loss": 1.0485, "step": 7399 }, { "epoch": 0.6305922454196847, "grad_norm": 14.5625, "learning_rate": 1.944899343903154e-06, "loss": 0.6306, "step": 7400 }, { "epoch": 0.6306774605879847, "grad_norm": 14.25, "learning_rate": 1.9447596604006735e-06, "loss": 0.8967, "step": 7401 }, { "epoch": 0.6307626757562846, "grad_norm": 14.375, "learning_rate": 1.944619964343187e-06, "loss": 0.561, "step": 7402 }, { "epoch": 0.6308478909245846, "grad_norm": 12.8125, "learning_rate": 1.9444802557332195e-06, "loss": 0.6972, "step": 7403 }, { "epoch": 0.6309331060928846, "grad_norm": 12.6875, "learning_rate": 1.9443405345732957e-06, "loss": 0.4426, "step": 7404 }, { "epoch": 0.6310183212611845, "grad_norm": 22.375, "learning_rate": 1.9442008008659395e-06, "loss": 0.7587, "step": 7405 }, { "epoch": 0.6311035364294845, "grad_norm": 15.4375, "learning_rate": 1.944061054613677e-06, "loss": 0.7279, "step": 7406 }, { "epoch": 0.6311887515977844, "grad_norm": 15.875, "learning_rate": 1.943921295819034e-06, "loss": 0.5185, "step": 7407 }, { "epoch": 0.6312739667660844, "grad_norm": 15.0625, "learning_rate": 1.943781524484535e-06, "loss": 0.7107, "step": 7408 }, { "epoch": 0.6313591819343843, "grad_norm": 12.6875, "learning_rate": 1.943641740612706e-06, "loss": 0.6492, "step": 7409 }, { "epoch": 0.6314443971026843, "grad_norm": 12.0, "learning_rate": 1.9435019442060733e-06, "loss": 0.5004, "step": 7410 }, { "epoch": 0.6315296122709843, "grad_norm": 22.125, "learning_rate": 1.9433621352671634e-06, "loss": 1.1887, "step": 7411 }, { "epoch": 0.6316148274392842, "grad_norm": 11.3125, "learning_rate": 1.9432223137985024e-06, "loss": 0.4793, "step": 7412 }, { "epoch": 0.6317000426075842, "grad_norm": 10.0, "learning_rate": 1.943082479802617e-06, "loss": 0.4498, "step": 7413 }, { "epoch": 0.6317852577758841, "grad_norm": 15.3125, "learning_rate": 1.942942633282034e-06, "loss": 0.8601, "step": 7414 }, { "epoch": 0.6318704729441841, "grad_norm": 12.75, "learning_rate": 1.9428027742392813e-06, "loss": 0.4812, "step": 7415 }, { "epoch": 0.631955688112484, "grad_norm": 13.25, "learning_rate": 1.942662902676885e-06, "loss": 0.7221, "step": 7416 }, { "epoch": 0.632040903280784, "grad_norm": 14.3125, "learning_rate": 1.942523018597374e-06, "loss": 0.5319, "step": 7417 }, { "epoch": 0.632126118449084, "grad_norm": 12.1875, "learning_rate": 1.942383122003276e-06, "loss": 0.6062, "step": 7418 }, { "epoch": 0.6322113336173839, "grad_norm": 12.5625, "learning_rate": 1.9422432128971187e-06, "loss": 0.6049, "step": 7419 }, { "epoch": 0.6322965487856839, "grad_norm": 13.75, "learning_rate": 1.94210329128143e-06, "loss": 0.8464, "step": 7420 }, { "epoch": 0.6323817639539838, "grad_norm": 12.0, "learning_rate": 1.9419633571587392e-06, "loss": 0.4576, "step": 7421 }, { "epoch": 0.6324669791222838, "grad_norm": 14.0, "learning_rate": 1.9418234105315744e-06, "loss": 0.7391, "step": 7422 }, { "epoch": 0.6325521942905837, "grad_norm": 13.5, "learning_rate": 1.9416834514024654e-06, "loss": 0.5661, "step": 7423 }, { "epoch": 0.6326374094588837, "grad_norm": 13.3125, "learning_rate": 1.94154347977394e-06, "loss": 0.706, "step": 7424 }, { "epoch": 0.6327226246271836, "grad_norm": 14.125, "learning_rate": 1.9414034956485293e-06, "loss": 0.8299, "step": 7425 }, { "epoch": 0.6328078397954836, "grad_norm": 15.6875, "learning_rate": 1.9412634990287616e-06, "loss": 0.9554, "step": 7426 }, { "epoch": 0.6328930549637836, "grad_norm": 12.6875, "learning_rate": 1.9411234899171685e-06, "loss": 0.7038, "step": 7427 }, { "epoch": 0.6329782701320835, "grad_norm": 18.375, "learning_rate": 1.9409834683162777e-06, "loss": 0.8385, "step": 7428 }, { "epoch": 0.6330634853003835, "grad_norm": 13.5, "learning_rate": 1.9408434342286213e-06, "loss": 0.9287, "step": 7429 }, { "epoch": 0.6331487004686834, "grad_norm": 14.4375, "learning_rate": 1.9407033876567292e-06, "loss": 0.7782, "step": 7430 }, { "epoch": 0.6332339156369834, "grad_norm": 15.75, "learning_rate": 1.9405633286031324e-06, "loss": 0.566, "step": 7431 }, { "epoch": 0.6333191308052833, "grad_norm": 12.125, "learning_rate": 1.940423257070362e-06, "loss": 0.53, "step": 7432 }, { "epoch": 0.6334043459735833, "grad_norm": 15.9375, "learning_rate": 1.940283173060949e-06, "loss": 0.8393, "step": 7433 }, { "epoch": 0.6334895611418833, "grad_norm": 18.5, "learning_rate": 1.9401430765774247e-06, "loss": 0.957, "step": 7434 }, { "epoch": 0.6335747763101832, "grad_norm": 13.5, "learning_rate": 1.940002967622321e-06, "loss": 0.6762, "step": 7435 }, { "epoch": 0.6336599914784832, "grad_norm": 13.3125, "learning_rate": 1.93986284619817e-06, "loss": 0.965, "step": 7436 }, { "epoch": 0.6337452066467831, "grad_norm": 14.1875, "learning_rate": 1.9397227123075035e-06, "loss": 0.8258, "step": 7437 }, { "epoch": 0.6338304218150831, "grad_norm": 11.1875, "learning_rate": 1.9395825659528543e-06, "loss": 0.5571, "step": 7438 }, { "epoch": 0.633915636983383, "grad_norm": 10.8125, "learning_rate": 1.9394424071367545e-06, "loss": 0.4198, "step": 7439 }, { "epoch": 0.634000852151683, "grad_norm": 15.625, "learning_rate": 1.9393022358617374e-06, "loss": 0.6657, "step": 7440 }, { "epoch": 0.634086067319983, "grad_norm": 18.25, "learning_rate": 1.9391620521303355e-06, "loss": 0.9386, "step": 7441 }, { "epoch": 0.6341712824882829, "grad_norm": 13.875, "learning_rate": 1.9390218559450823e-06, "loss": 0.3725, "step": 7442 }, { "epoch": 0.6342564976565829, "grad_norm": 31.5, "learning_rate": 1.9388816473085116e-06, "loss": 1.049, "step": 7443 }, { "epoch": 0.6343417128248828, "grad_norm": 11.125, "learning_rate": 1.9387414262231562e-06, "loss": 0.4952, "step": 7444 }, { "epoch": 0.6344269279931828, "grad_norm": 17.625, "learning_rate": 1.938601192691551e-06, "loss": 0.9978, "step": 7445 }, { "epoch": 0.6345121431614827, "grad_norm": 16.0, "learning_rate": 1.93846094671623e-06, "loss": 0.7123, "step": 7446 }, { "epoch": 0.6345973583297827, "grad_norm": 20.75, "learning_rate": 1.9383206882997275e-06, "loss": 1.2092, "step": 7447 }, { "epoch": 0.6346825734980827, "grad_norm": 11.625, "learning_rate": 1.9381804174445775e-06, "loss": 0.5536, "step": 7448 }, { "epoch": 0.6347677886663826, "grad_norm": 12.0625, "learning_rate": 1.9380401341533155e-06, "loss": 0.5701, "step": 7449 }, { "epoch": 0.6348530038346826, "grad_norm": 21.5, "learning_rate": 1.9378998384284764e-06, "loss": 0.8794, "step": 7450 }, { "epoch": 0.6349382190029825, "grad_norm": 15.375, "learning_rate": 1.9377595302725954e-06, "loss": 0.8057, "step": 7451 }, { "epoch": 0.6350234341712825, "grad_norm": 14.1875, "learning_rate": 1.9376192096882083e-06, "loss": 0.8704, "step": 7452 }, { "epoch": 0.6351086493395824, "grad_norm": 14.9375, "learning_rate": 1.9374788766778505e-06, "loss": 0.8706, "step": 7453 }, { "epoch": 0.6351938645078824, "grad_norm": 15.75, "learning_rate": 1.9373385312440583e-06, "loss": 0.9047, "step": 7454 }, { "epoch": 0.6352790796761824, "grad_norm": 15.4375, "learning_rate": 1.9371981733893674e-06, "loss": 0.9157, "step": 7455 }, { "epoch": 0.6353642948444823, "grad_norm": 18.0, "learning_rate": 1.9370578031163145e-06, "loss": 0.6883, "step": 7456 }, { "epoch": 0.6354495100127823, "grad_norm": 19.375, "learning_rate": 1.9369174204274362e-06, "loss": 1.1392, "step": 7457 }, { "epoch": 0.6355347251810822, "grad_norm": 10.375, "learning_rate": 1.9367770253252694e-06, "loss": 0.4064, "step": 7458 }, { "epoch": 0.6356199403493822, "grad_norm": 14.3125, "learning_rate": 1.936636617812351e-06, "loss": 0.5448, "step": 7459 }, { "epoch": 0.6357051555176821, "grad_norm": 18.5, "learning_rate": 1.936496197891219e-06, "loss": 0.9048, "step": 7460 }, { "epoch": 0.6357903706859821, "grad_norm": 14.625, "learning_rate": 1.93635576556441e-06, "loss": 0.9042, "step": 7461 }, { "epoch": 0.635875585854282, "grad_norm": 15.1875, "learning_rate": 1.936215320834462e-06, "loss": 0.6657, "step": 7462 }, { "epoch": 0.635960801022582, "grad_norm": 12.5, "learning_rate": 1.9360748637039136e-06, "loss": 0.5172, "step": 7463 }, { "epoch": 0.636046016190882, "grad_norm": 13.25, "learning_rate": 1.9359343941753024e-06, "loss": 0.629, "step": 7464 }, { "epoch": 0.6361312313591819, "grad_norm": 11.9375, "learning_rate": 1.9357939122511666e-06, "loss": 0.7695, "step": 7465 }, { "epoch": 0.6362164465274819, "grad_norm": 19.875, "learning_rate": 1.935653417934045e-06, "loss": 0.6998, "step": 7466 }, { "epoch": 0.6363016616957818, "grad_norm": 13.3125, "learning_rate": 1.935512911226477e-06, "loss": 0.5618, "step": 7467 }, { "epoch": 0.6363868768640818, "grad_norm": 22.25, "learning_rate": 1.935372392131001e-06, "loss": 0.9994, "step": 7468 }, { "epoch": 0.6364720920323818, "grad_norm": 13.0625, "learning_rate": 1.9352318606501576e-06, "loss": 0.5832, "step": 7469 }, { "epoch": 0.6365573072006817, "grad_norm": 19.375, "learning_rate": 1.9350913167864842e-06, "loss": 1.0666, "step": 7470 }, { "epoch": 0.6366425223689817, "grad_norm": 19.875, "learning_rate": 1.9349507605425227e-06, "loss": 1.0364, "step": 7471 }, { "epoch": 0.6367277375372816, "grad_norm": 14.625, "learning_rate": 1.934810191920812e-06, "loss": 0.9384, "step": 7472 }, { "epoch": 0.6368129527055816, "grad_norm": 19.5, "learning_rate": 1.9346696109238923e-06, "loss": 0.7831, "step": 7473 }, { "epoch": 0.6368981678738815, "grad_norm": 15.375, "learning_rate": 1.9345290175543043e-06, "loss": 0.8197, "step": 7474 }, { "epoch": 0.6369833830421815, "grad_norm": 14.125, "learning_rate": 1.9343884118145883e-06, "loss": 0.4731, "step": 7475 }, { "epoch": 0.6370685982104815, "grad_norm": 13.5625, "learning_rate": 1.934247793707286e-06, "loss": 0.6042, "step": 7476 }, { "epoch": 0.6371538133787814, "grad_norm": 22.375, "learning_rate": 1.934107163234938e-06, "loss": 1.2094, "step": 7477 }, { "epoch": 0.6372390285470814, "grad_norm": 11.625, "learning_rate": 1.9339665204000856e-06, "loss": 0.5361, "step": 7478 }, { "epoch": 0.6373242437153813, "grad_norm": 13.75, "learning_rate": 1.93382586520527e-06, "loss": 0.7193, "step": 7479 }, { "epoch": 0.6374094588836813, "grad_norm": 15.5, "learning_rate": 1.9336851976530342e-06, "loss": 1.0151, "step": 7480 }, { "epoch": 0.6374946740519812, "grad_norm": 13.4375, "learning_rate": 1.933544517745918e-06, "loss": 0.7259, "step": 7481 }, { "epoch": 0.6375798892202812, "grad_norm": 14.6875, "learning_rate": 1.933403825486466e-06, "loss": 0.2779, "step": 7482 }, { "epoch": 0.6376651043885811, "grad_norm": 19.0, "learning_rate": 1.9332631208772197e-06, "loss": 0.8455, "step": 7483 }, { "epoch": 0.6377503195568811, "grad_norm": 13.5, "learning_rate": 1.9331224039207217e-06, "loss": 0.6364, "step": 7484 }, { "epoch": 0.6378355347251811, "grad_norm": 15.8125, "learning_rate": 1.9329816746195146e-06, "loss": 0.7591, "step": 7485 }, { "epoch": 0.637920749893481, "grad_norm": 12.25, "learning_rate": 1.932840932976142e-06, "loss": 0.6894, "step": 7486 }, { "epoch": 0.638005965061781, "grad_norm": 14.0625, "learning_rate": 1.932700178993147e-06, "loss": 0.5843, "step": 7487 }, { "epoch": 0.6380911802300809, "grad_norm": 13.3125, "learning_rate": 1.9325594126730736e-06, "loss": 0.5927, "step": 7488 }, { "epoch": 0.6381763953983809, "grad_norm": 17.0, "learning_rate": 1.932418634018465e-06, "loss": 0.5767, "step": 7489 }, { "epoch": 0.6382616105666808, "grad_norm": 11.75, "learning_rate": 1.932277843031865e-06, "loss": 0.4747, "step": 7490 }, { "epoch": 0.6383468257349808, "grad_norm": 15.375, "learning_rate": 1.932137039715819e-06, "loss": 0.6273, "step": 7491 }, { "epoch": 0.6384320409032808, "grad_norm": 22.25, "learning_rate": 1.931996224072871e-06, "loss": 0.851, "step": 7492 }, { "epoch": 0.6385172560715807, "grad_norm": 11.25, "learning_rate": 1.9318553961055645e-06, "loss": 0.5278, "step": 7493 }, { "epoch": 0.6386024712398807, "grad_norm": 16.375, "learning_rate": 1.9317145558164458e-06, "loss": 0.989, "step": 7494 }, { "epoch": 0.6386876864081806, "grad_norm": 13.0625, "learning_rate": 1.9315737032080595e-06, "loss": 0.7262, "step": 7495 }, { "epoch": 0.6387729015764806, "grad_norm": 18.5, "learning_rate": 1.931432838282951e-06, "loss": 1.0054, "step": 7496 }, { "epoch": 0.6388581167447805, "grad_norm": 17.875, "learning_rate": 1.931291961043666e-06, "loss": 0.7548, "step": 7497 }, { "epoch": 0.6389433319130805, "grad_norm": 20.0, "learning_rate": 1.93115107149275e-06, "loss": 0.9563, "step": 7498 }, { "epoch": 0.6390285470813805, "grad_norm": 15.1875, "learning_rate": 1.9310101696327494e-06, "loss": 0.7491, "step": 7499 }, { "epoch": 0.6391137622496804, "grad_norm": 22.25, "learning_rate": 1.93086925546621e-06, "loss": 0.6004, "step": 7500 }, { "epoch": 0.6391989774179804, "grad_norm": 12.375, "learning_rate": 1.9307283289956786e-06, "loss": 0.5954, "step": 7501 }, { "epoch": 0.6392841925862803, "grad_norm": 10.875, "learning_rate": 1.9305873902237014e-06, "loss": 0.3651, "step": 7502 }, { "epoch": 0.6393694077545803, "grad_norm": 12.9375, "learning_rate": 1.930446439152826e-06, "loss": 0.6012, "step": 7503 }, { "epoch": 0.6394546229228802, "grad_norm": 22.375, "learning_rate": 1.9303054757855994e-06, "loss": 0.9328, "step": 7504 }, { "epoch": 0.6395398380911802, "grad_norm": 13.8125, "learning_rate": 1.9301645001245686e-06, "loss": 0.5892, "step": 7505 }, { "epoch": 0.6396250532594802, "grad_norm": 15.0, "learning_rate": 1.9300235121722813e-06, "loss": 0.7726, "step": 7506 }, { "epoch": 0.6397102684277801, "grad_norm": 13.25, "learning_rate": 1.9298825119312853e-06, "loss": 0.8592, "step": 7507 }, { "epoch": 0.6397954835960801, "grad_norm": 17.625, "learning_rate": 1.929741499404129e-06, "loss": 0.5337, "step": 7508 }, { "epoch": 0.63988069876438, "grad_norm": 12.75, "learning_rate": 1.9296004745933596e-06, "loss": 0.4694, "step": 7509 }, { "epoch": 0.63996591393268, "grad_norm": 31.875, "learning_rate": 1.9294594375015266e-06, "loss": 1.1662, "step": 7510 }, { "epoch": 0.6400511291009799, "grad_norm": 29.125, "learning_rate": 1.9293183881311784e-06, "loss": 1.1891, "step": 7511 }, { "epoch": 0.6401363442692799, "grad_norm": 11.1875, "learning_rate": 1.929177326484864e-06, "loss": 0.5254, "step": 7512 }, { "epoch": 0.6402215594375799, "grad_norm": 12.125, "learning_rate": 1.9290362525651318e-06, "loss": 0.9344, "step": 7513 }, { "epoch": 0.6403067746058798, "grad_norm": 14.875, "learning_rate": 1.9288951663745316e-06, "loss": 0.7495, "step": 7514 }, { "epoch": 0.6403919897741798, "grad_norm": 12.5, "learning_rate": 1.9287540679156134e-06, "loss": 0.5234, "step": 7515 }, { "epoch": 0.6404772049424797, "grad_norm": 12.125, "learning_rate": 1.9286129571909267e-06, "loss": 0.7819, "step": 7516 }, { "epoch": 0.6405624201107797, "grad_norm": 13.5, "learning_rate": 1.928471834203022e-06, "loss": 0.424, "step": 7517 }, { "epoch": 0.6406476352790796, "grad_norm": 24.125, "learning_rate": 1.928330698954448e-06, "loss": 0.8901, "step": 7518 }, { "epoch": 0.6407328504473796, "grad_norm": 15.6875, "learning_rate": 1.9281895514477568e-06, "loss": 0.6623, "step": 7519 }, { "epoch": 0.6408180656156796, "grad_norm": 17.375, "learning_rate": 1.928048391685498e-06, "loss": 0.6539, "step": 7520 }, { "epoch": 0.6409032807839795, "grad_norm": 19.0, "learning_rate": 1.927907219670223e-06, "loss": 0.9933, "step": 7521 }, { "epoch": 0.6409884959522795, "grad_norm": 16.625, "learning_rate": 1.927766035404483e-06, "loss": 0.7251, "step": 7522 }, { "epoch": 0.6410737111205794, "grad_norm": 12.5, "learning_rate": 1.9276248388908293e-06, "loss": 0.6019, "step": 7523 }, { "epoch": 0.6411589262888794, "grad_norm": 10.75, "learning_rate": 1.9274836301318135e-06, "loss": 0.5628, "step": 7524 }, { "epoch": 0.6412441414571793, "grad_norm": 11.25, "learning_rate": 1.9273424091299867e-06, "loss": 0.5809, "step": 7525 }, { "epoch": 0.6413293566254793, "grad_norm": 21.25, "learning_rate": 1.9272011758879015e-06, "loss": 0.9849, "step": 7526 }, { "epoch": 0.6414145717937793, "grad_norm": 16.0, "learning_rate": 1.9270599304081105e-06, "loss": 0.5735, "step": 7527 }, { "epoch": 0.6414997869620792, "grad_norm": 14.125, "learning_rate": 1.9269186726931653e-06, "loss": 0.5916, "step": 7528 }, { "epoch": 0.6415850021303792, "grad_norm": 14.8125, "learning_rate": 1.9267774027456193e-06, "loss": 0.8224, "step": 7529 }, { "epoch": 0.6416702172986791, "grad_norm": 13.5625, "learning_rate": 1.926636120568025e-06, "loss": 0.6435, "step": 7530 }, { "epoch": 0.6417554324669791, "grad_norm": 15.9375, "learning_rate": 1.926494826162936e-06, "loss": 0.5791, "step": 7531 }, { "epoch": 0.641840647635279, "grad_norm": 13.25, "learning_rate": 1.9263535195329044e-06, "loss": 0.741, "step": 7532 }, { "epoch": 0.641925862803579, "grad_norm": 16.75, "learning_rate": 1.926212200680485e-06, "loss": 1.0166, "step": 7533 }, { "epoch": 0.642011077971879, "grad_norm": 13.9375, "learning_rate": 1.9260708696082312e-06, "loss": 0.9037, "step": 7534 }, { "epoch": 0.6420962931401789, "grad_norm": 18.5, "learning_rate": 1.925929526318697e-06, "loss": 0.8912, "step": 7535 }, { "epoch": 0.6421815083084789, "grad_norm": 17.75, "learning_rate": 1.925788170814436e-06, "loss": 1.0491, "step": 7536 }, { "epoch": 0.6422667234767788, "grad_norm": 14.625, "learning_rate": 1.9256468030980037e-06, "loss": 0.7845, "step": 7537 }, { "epoch": 0.6423519386450788, "grad_norm": 18.875, "learning_rate": 1.9255054231719544e-06, "loss": 0.9729, "step": 7538 }, { "epoch": 0.6424371538133787, "grad_norm": 14.5625, "learning_rate": 1.9253640310388427e-06, "loss": 1.1976, "step": 7539 }, { "epoch": 0.6425223689816787, "grad_norm": 21.5, "learning_rate": 1.9252226267012238e-06, "loss": 0.7589, "step": 7540 }, { "epoch": 0.6426075841499787, "grad_norm": 22.25, "learning_rate": 1.9250812101616527e-06, "loss": 0.6298, "step": 7541 }, { "epoch": 0.6426927993182786, "grad_norm": 30.625, "learning_rate": 1.9249397814226857e-06, "loss": 1.174, "step": 7542 }, { "epoch": 0.6427780144865786, "grad_norm": 10.6875, "learning_rate": 1.924798340486878e-06, "loss": 0.4541, "step": 7543 }, { "epoch": 0.6428632296548785, "grad_norm": 13.25, "learning_rate": 1.9246568873567855e-06, "loss": 0.4666, "step": 7544 }, { "epoch": 0.6429484448231785, "grad_norm": 16.625, "learning_rate": 1.924515422034965e-06, "loss": 0.871, "step": 7545 }, { "epoch": 0.6430336599914784, "grad_norm": 13.6875, "learning_rate": 1.9243739445239725e-06, "loss": 0.6814, "step": 7546 }, { "epoch": 0.6431188751597784, "grad_norm": 13.5, "learning_rate": 1.924232454826365e-06, "loss": 0.7081, "step": 7547 }, { "epoch": 0.6432040903280783, "grad_norm": 19.125, "learning_rate": 1.924090952944698e-06, "loss": 1.1679, "step": 7548 }, { "epoch": 0.6432893054963783, "grad_norm": 20.375, "learning_rate": 1.9239494388815304e-06, "loss": 0.8666, "step": 7549 }, { "epoch": 0.6433745206646783, "grad_norm": 13.9375, "learning_rate": 1.923807912639419e-06, "loss": 0.5498, "step": 7550 }, { "epoch": 0.6434597358329782, "grad_norm": 16.125, "learning_rate": 1.9236663742209203e-06, "loss": 0.8145, "step": 7551 }, { "epoch": 0.6435449510012782, "grad_norm": 12.9375, "learning_rate": 1.923524823628593e-06, "loss": 0.4805, "step": 7552 }, { "epoch": 0.6436301661695781, "grad_norm": 14.875, "learning_rate": 1.9233832608649955e-06, "loss": 0.7224, "step": 7553 }, { "epoch": 0.6437153813378781, "grad_norm": 15.8125, "learning_rate": 1.9232416859326845e-06, "loss": 0.7638, "step": 7554 }, { "epoch": 0.643800596506178, "grad_norm": 15.125, "learning_rate": 1.92310009883422e-06, "loss": 0.7241, "step": 7555 }, { "epoch": 0.643885811674478, "grad_norm": 8.3125, "learning_rate": 1.922958499572159e-06, "loss": 0.5145, "step": 7556 }, { "epoch": 0.643971026842778, "grad_norm": 39.5, "learning_rate": 1.9228168881490616e-06, "loss": 0.9358, "step": 7557 }, { "epoch": 0.6440562420110779, "grad_norm": 14.1875, "learning_rate": 1.922675264567487e-06, "loss": 0.7633, "step": 7558 }, { "epoch": 0.6441414571793779, "grad_norm": 14.6875, "learning_rate": 1.922533628829993e-06, "loss": 0.9419, "step": 7559 }, { "epoch": 0.6442266723476778, "grad_norm": 17.625, "learning_rate": 1.9223919809391407e-06, "loss": 0.7895, "step": 7560 }, { "epoch": 0.6443118875159779, "grad_norm": 14.125, "learning_rate": 1.922250320897489e-06, "loss": 0.7973, "step": 7561 }, { "epoch": 0.6443971026842779, "grad_norm": 12.75, "learning_rate": 1.922108648707598e-06, "loss": 0.7334, "step": 7562 }, { "epoch": 0.6444823178525778, "grad_norm": 13.0, "learning_rate": 1.9219669643720275e-06, "loss": 0.5542, "step": 7563 }, { "epoch": 0.6445675330208778, "grad_norm": 13.6875, "learning_rate": 1.921825267893339e-06, "loss": 0.6766, "step": 7564 }, { "epoch": 0.6446527481891777, "grad_norm": 14.375, "learning_rate": 1.9216835592740917e-06, "loss": 0.7353, "step": 7565 }, { "epoch": 0.6447379633574777, "grad_norm": 17.875, "learning_rate": 1.9215418385168473e-06, "loss": 1.0316, "step": 7566 }, { "epoch": 0.6448231785257776, "grad_norm": 15.125, "learning_rate": 1.9214001056241665e-06, "loss": 0.5764, "step": 7567 }, { "epoch": 0.6449083936940776, "grad_norm": 17.125, "learning_rate": 1.9212583605986103e-06, "loss": 0.8525, "step": 7568 }, { "epoch": 0.6449936088623776, "grad_norm": 11.125, "learning_rate": 1.9211166034427415e-06, "loss": 0.3902, "step": 7569 }, { "epoch": 0.6450788240306775, "grad_norm": 14.4375, "learning_rate": 1.92097483415912e-06, "loss": 0.6017, "step": 7570 }, { "epoch": 0.6451640391989775, "grad_norm": 17.75, "learning_rate": 1.920833052750309e-06, "loss": 1.0817, "step": 7571 }, { "epoch": 0.6452492543672774, "grad_norm": 18.25, "learning_rate": 1.9206912592188693e-06, "loss": 0.5975, "step": 7572 }, { "epoch": 0.6453344695355774, "grad_norm": 15.375, "learning_rate": 1.9205494535673653e-06, "loss": 0.5062, "step": 7573 }, { "epoch": 0.6454196847038773, "grad_norm": 14.4375, "learning_rate": 1.9204076357983575e-06, "loss": 0.51, "step": 7574 }, { "epoch": 0.6455048998721773, "grad_norm": 10.6875, "learning_rate": 1.9202658059144098e-06, "loss": 0.3104, "step": 7575 }, { "epoch": 0.6455901150404773, "grad_norm": 14.1875, "learning_rate": 1.9201239639180853e-06, "loss": 0.5224, "step": 7576 }, { "epoch": 0.6456753302087772, "grad_norm": 21.75, "learning_rate": 1.9199821098119468e-06, "loss": 0.8429, "step": 7577 }, { "epoch": 0.6457605453770772, "grad_norm": 20.875, "learning_rate": 1.9198402435985573e-06, "loss": 0.6916, "step": 7578 }, { "epoch": 0.6458457605453771, "grad_norm": 14.3125, "learning_rate": 1.9196983652804817e-06, "loss": 0.5824, "step": 7579 }, { "epoch": 0.6459309757136771, "grad_norm": 16.5, "learning_rate": 1.9195564748602828e-06, "loss": 0.6706, "step": 7580 }, { "epoch": 0.646016190881977, "grad_norm": 17.125, "learning_rate": 1.9194145723405256e-06, "loss": 0.9729, "step": 7581 }, { "epoch": 0.646101406050277, "grad_norm": 17.0, "learning_rate": 1.9192726577237738e-06, "loss": 1.0382, "step": 7582 }, { "epoch": 0.646186621218577, "grad_norm": 16.375, "learning_rate": 1.9191307310125913e-06, "loss": 1.0176, "step": 7583 }, { "epoch": 0.6462718363868769, "grad_norm": 14.25, "learning_rate": 1.9189887922095443e-06, "loss": 0.7483, "step": 7584 }, { "epoch": 0.6463570515551769, "grad_norm": 12.3125, "learning_rate": 1.918846841317197e-06, "loss": 0.4302, "step": 7585 }, { "epoch": 0.6464422667234768, "grad_norm": 11.3125, "learning_rate": 1.9187048783381144e-06, "loss": 0.3955, "step": 7586 }, { "epoch": 0.6465274818917768, "grad_norm": 17.875, "learning_rate": 1.9185629032748626e-06, "loss": 0.904, "step": 7587 }, { "epoch": 0.6466126970600767, "grad_norm": 15.6875, "learning_rate": 1.918420916130007e-06, "loss": 0.6171, "step": 7588 }, { "epoch": 0.6466979122283767, "grad_norm": 17.75, "learning_rate": 1.918278916906113e-06, "loss": 1.0285, "step": 7589 }, { "epoch": 0.6467831273966766, "grad_norm": 16.25, "learning_rate": 1.9181369056057474e-06, "loss": 0.7404, "step": 7590 }, { "epoch": 0.6468683425649766, "grad_norm": 13.25, "learning_rate": 1.9179948822314754e-06, "loss": 0.4816, "step": 7591 }, { "epoch": 0.6469535577332766, "grad_norm": 11.75, "learning_rate": 1.917852846785864e-06, "loss": 0.5135, "step": 7592 }, { "epoch": 0.6470387729015765, "grad_norm": 12.4375, "learning_rate": 1.917710799271481e-06, "loss": 0.462, "step": 7593 }, { "epoch": 0.6471239880698765, "grad_norm": 17.375, "learning_rate": 1.917568739690892e-06, "loss": 0.7491, "step": 7594 }, { "epoch": 0.6472092032381764, "grad_norm": 13.5625, "learning_rate": 1.917426668046664e-06, "loss": 0.848, "step": 7595 }, { "epoch": 0.6472944184064764, "grad_norm": 11.5, "learning_rate": 1.9172845843413655e-06, "loss": 0.4944, "step": 7596 }, { "epoch": 0.6473796335747763, "grad_norm": 11.1875, "learning_rate": 1.917142488577564e-06, "loss": 0.4256, "step": 7597 }, { "epoch": 0.6474648487430763, "grad_norm": 15.8125, "learning_rate": 1.917000380757826e-06, "loss": 1.0364, "step": 7598 }, { "epoch": 0.6475500639113763, "grad_norm": 14.0, "learning_rate": 1.916858260884721e-06, "loss": 0.7619, "step": 7599 }, { "epoch": 0.6476352790796762, "grad_norm": 17.0, "learning_rate": 1.9167161289608164e-06, "loss": 0.8506, "step": 7600 }, { "epoch": 0.6477204942479762, "grad_norm": 24.375, "learning_rate": 1.916573984988681e-06, "loss": 0.9152, "step": 7601 }, { "epoch": 0.6478057094162761, "grad_norm": 11.625, "learning_rate": 1.9164318289708834e-06, "loss": 0.5738, "step": 7602 }, { "epoch": 0.6478909245845761, "grad_norm": 17.5, "learning_rate": 1.9162896609099924e-06, "loss": 1.0527, "step": 7603 }, { "epoch": 0.647976139752876, "grad_norm": 17.0, "learning_rate": 1.9161474808085773e-06, "loss": 0.7811, "step": 7604 }, { "epoch": 0.648061354921176, "grad_norm": 15.375, "learning_rate": 1.916005288669207e-06, "loss": 1.0938, "step": 7605 }, { "epoch": 0.648146570089476, "grad_norm": 10.875, "learning_rate": 1.915863084494452e-06, "loss": 0.3323, "step": 7606 }, { "epoch": 0.6482317852577759, "grad_norm": 14.5625, "learning_rate": 1.915720868286881e-06, "loss": 0.4926, "step": 7607 }, { "epoch": 0.6483170004260759, "grad_norm": 13.75, "learning_rate": 1.915578640049065e-06, "loss": 0.8232, "step": 7608 }, { "epoch": 0.6484022155943758, "grad_norm": 19.125, "learning_rate": 1.9154363997835736e-06, "loss": 0.8178, "step": 7609 }, { "epoch": 0.6484874307626758, "grad_norm": 14.9375, "learning_rate": 1.915294147492977e-06, "loss": 1.0978, "step": 7610 }, { "epoch": 0.6485726459309757, "grad_norm": 14.3125, "learning_rate": 1.9151518831798462e-06, "loss": 0.8506, "step": 7611 }, { "epoch": 0.6486578610992757, "grad_norm": 15.8125, "learning_rate": 1.915009606846752e-06, "loss": 0.9594, "step": 7612 }, { "epoch": 0.6487430762675757, "grad_norm": 16.0, "learning_rate": 1.914867318496266e-06, "loss": 0.8196, "step": 7613 }, { "epoch": 0.6488282914358756, "grad_norm": 12.5625, "learning_rate": 1.914725018130958e-06, "loss": 0.6289, "step": 7614 }, { "epoch": 0.6489135066041756, "grad_norm": 19.0, "learning_rate": 1.914582705753401e-06, "loss": 0.9905, "step": 7615 }, { "epoch": 0.6489987217724755, "grad_norm": 11.625, "learning_rate": 1.9144403813661665e-06, "loss": 0.5533, "step": 7616 }, { "epoch": 0.6490839369407755, "grad_norm": 19.625, "learning_rate": 1.9142980449718256e-06, "loss": 0.7281, "step": 7617 }, { "epoch": 0.6491691521090754, "grad_norm": 13.25, "learning_rate": 1.9141556965729518e-06, "loss": 0.4316, "step": 7618 }, { "epoch": 0.6492543672773754, "grad_norm": 15.3125, "learning_rate": 1.9140133361721157e-06, "loss": 0.8764, "step": 7619 }, { "epoch": 0.6493395824456754, "grad_norm": 13.125, "learning_rate": 1.9138709637718916e-06, "loss": 1.0729, "step": 7620 }, { "epoch": 0.6494247976139753, "grad_norm": 13.9375, "learning_rate": 1.913728579374851e-06, "loss": 0.4932, "step": 7621 }, { "epoch": 0.6495100127822753, "grad_norm": 13.9375, "learning_rate": 1.9135861829835686e-06, "loss": 0.6855, "step": 7622 }, { "epoch": 0.6495952279505752, "grad_norm": 18.625, "learning_rate": 1.9134437746006158e-06, "loss": 0.854, "step": 7623 }, { "epoch": 0.6496804431188752, "grad_norm": 23.0, "learning_rate": 1.9133013542285668e-06, "loss": 1.3407, "step": 7624 }, { "epoch": 0.6497656582871751, "grad_norm": 14.1875, "learning_rate": 1.9131589218699957e-06, "loss": 0.8345, "step": 7625 }, { "epoch": 0.6498508734554751, "grad_norm": 17.25, "learning_rate": 1.9130164775274763e-06, "loss": 0.7151, "step": 7626 }, { "epoch": 0.649936088623775, "grad_norm": 19.625, "learning_rate": 1.9128740212035814e-06, "loss": 1.013, "step": 7627 }, { "epoch": 0.650021303792075, "grad_norm": 13.125, "learning_rate": 1.9127315529008873e-06, "loss": 0.8041, "step": 7628 }, { "epoch": 0.650106518960375, "grad_norm": 12.8125, "learning_rate": 1.912589072621967e-06, "loss": 0.5704, "step": 7629 }, { "epoch": 0.6501917341286749, "grad_norm": 12.375, "learning_rate": 1.9124465803693962e-06, "loss": 0.5162, "step": 7630 }, { "epoch": 0.6502769492969749, "grad_norm": 16.375, "learning_rate": 1.9123040761457497e-06, "loss": 0.6625, "step": 7631 }, { "epoch": 0.6503621644652748, "grad_norm": 10.6875, "learning_rate": 1.912161559953602e-06, "loss": 0.407, "step": 7632 }, { "epoch": 0.6504473796335748, "grad_norm": 12.5625, "learning_rate": 1.9120190317955293e-06, "loss": 0.5644, "step": 7633 }, { "epoch": 0.6505325948018748, "grad_norm": 15.3125, "learning_rate": 1.911876491674107e-06, "loss": 0.6873, "step": 7634 }, { "epoch": 0.6506178099701747, "grad_norm": 14.5625, "learning_rate": 1.9117339395919115e-06, "loss": 0.7525, "step": 7635 }, { "epoch": 0.6507030251384747, "grad_norm": 16.75, "learning_rate": 1.9115913755515173e-06, "loss": 0.9009, "step": 7636 }, { "epoch": 0.6507882403067746, "grad_norm": 12.5, "learning_rate": 1.9114487995555025e-06, "loss": 0.749, "step": 7637 }, { "epoch": 0.6508734554750746, "grad_norm": 16.125, "learning_rate": 1.9113062116064426e-06, "loss": 1.1835, "step": 7638 }, { "epoch": 0.6509586706433745, "grad_norm": 15.3125, "learning_rate": 1.9111636117069145e-06, "loss": 0.9973, "step": 7639 }, { "epoch": 0.6510438858116745, "grad_norm": 14.125, "learning_rate": 1.911020999859494e-06, "loss": 0.7477, "step": 7640 }, { "epoch": 0.6511291009799745, "grad_norm": 27.25, "learning_rate": 1.9108783760667607e-06, "loss": 0.8946, "step": 7641 }, { "epoch": 0.6512143161482744, "grad_norm": 14.3125, "learning_rate": 1.91073574033129e-06, "loss": 0.8122, "step": 7642 }, { "epoch": 0.6512995313165744, "grad_norm": 11.0, "learning_rate": 1.9105930926556606e-06, "loss": 0.4939, "step": 7643 }, { "epoch": 0.6513847464848743, "grad_norm": 12.1875, "learning_rate": 1.910450433042449e-06, "loss": 0.6006, "step": 7644 }, { "epoch": 0.6514699616531743, "grad_norm": 15.6875, "learning_rate": 1.910307761494234e-06, "loss": 0.9568, "step": 7645 }, { "epoch": 0.6515551768214742, "grad_norm": 69.5, "learning_rate": 1.9101650780135937e-06, "loss": 0.9464, "step": 7646 }, { "epoch": 0.6516403919897742, "grad_norm": 34.0, "learning_rate": 1.9100223826031074e-06, "loss": 0.715, "step": 7647 }, { "epoch": 0.6517256071580741, "grad_norm": 11.6875, "learning_rate": 1.909879675265352e-06, "loss": 0.4619, "step": 7648 }, { "epoch": 0.6518108223263741, "grad_norm": 24.5, "learning_rate": 1.9097369560029078e-06, "loss": 1.0209, "step": 7649 }, { "epoch": 0.6518960374946741, "grad_norm": 14.375, "learning_rate": 1.9095942248183534e-06, "loss": 0.5722, "step": 7650 }, { "epoch": 0.651981252662974, "grad_norm": 18.25, "learning_rate": 1.909451481714268e-06, "loss": 0.7237, "step": 7651 }, { "epoch": 0.652066467831274, "grad_norm": 12.5625, "learning_rate": 1.909308726693231e-06, "loss": 0.5568, "step": 7652 }, { "epoch": 0.6521516829995739, "grad_norm": 12.9375, "learning_rate": 1.9091659597578223e-06, "loss": 0.4262, "step": 7653 }, { "epoch": 0.6522368981678739, "grad_norm": 13.75, "learning_rate": 1.9090231809106223e-06, "loss": 0.6109, "step": 7654 }, { "epoch": 0.6523221133361738, "grad_norm": 13.1875, "learning_rate": 1.9088803901542103e-06, "loss": 0.5379, "step": 7655 }, { "epoch": 0.6524073285044738, "grad_norm": 13.3125, "learning_rate": 1.9087375874911674e-06, "loss": 0.6077, "step": 7656 }, { "epoch": 0.6524925436727738, "grad_norm": 16.75, "learning_rate": 1.9085947729240735e-06, "loss": 0.8948, "step": 7657 }, { "epoch": 0.6525777588410737, "grad_norm": 16.625, "learning_rate": 1.9084519464555106e-06, "loss": 0.8755, "step": 7658 }, { "epoch": 0.6526629740093737, "grad_norm": 14.8125, "learning_rate": 1.908309108088058e-06, "loss": 0.7233, "step": 7659 }, { "epoch": 0.6527481891776736, "grad_norm": 27.0, "learning_rate": 1.9081662578242983e-06, "loss": 0.3556, "step": 7660 }, { "epoch": 0.6528334043459736, "grad_norm": 11.5, "learning_rate": 1.908023395666812e-06, "loss": 0.4448, "step": 7661 }, { "epoch": 0.6529186195142735, "grad_norm": 16.625, "learning_rate": 1.907880521618182e-06, "loss": 0.8879, "step": 7662 }, { "epoch": 0.6530038346825735, "grad_norm": 12.6875, "learning_rate": 1.907737635680989e-06, "loss": 0.5524, "step": 7663 }, { "epoch": 0.6530890498508735, "grad_norm": 14.6875, "learning_rate": 1.9075947378578157e-06, "loss": 0.8419, "step": 7664 }, { "epoch": 0.6531742650191734, "grad_norm": 14.125, "learning_rate": 1.9074518281512438e-06, "loss": 0.6376, "step": 7665 }, { "epoch": 0.6532594801874734, "grad_norm": 20.25, "learning_rate": 1.9073089065638563e-06, "loss": 0.8453, "step": 7666 }, { "epoch": 0.6533446953557733, "grad_norm": 13.375, "learning_rate": 1.9071659730982362e-06, "loss": 0.5867, "step": 7667 }, { "epoch": 0.6534299105240733, "grad_norm": 19.125, "learning_rate": 1.907023027756966e-06, "loss": 0.5683, "step": 7668 }, { "epoch": 0.6535151256923732, "grad_norm": 10.5, "learning_rate": 1.906880070542629e-06, "loss": 0.3983, "step": 7669 }, { "epoch": 0.6536003408606732, "grad_norm": 19.375, "learning_rate": 1.9067371014578087e-06, "loss": 1.0957, "step": 7670 }, { "epoch": 0.6536855560289732, "grad_norm": 19.375, "learning_rate": 1.9065941205050886e-06, "loss": 1.0778, "step": 7671 }, { "epoch": 0.6537707711972731, "grad_norm": 13.5625, "learning_rate": 1.9064511276870523e-06, "loss": 0.6851, "step": 7672 }, { "epoch": 0.6538559863655731, "grad_norm": 15.0625, "learning_rate": 1.9063081230062841e-06, "loss": 0.9018, "step": 7673 }, { "epoch": 0.653941201533873, "grad_norm": 12.6875, "learning_rate": 1.906165106465368e-06, "loss": 0.6788, "step": 7674 }, { "epoch": 0.654026416702173, "grad_norm": 12.625, "learning_rate": 1.906022078066889e-06, "loss": 0.6593, "step": 7675 }, { "epoch": 0.6541116318704729, "grad_norm": 13.5625, "learning_rate": 1.9058790378134309e-06, "loss": 0.5763, "step": 7676 }, { "epoch": 0.6541968470387729, "grad_norm": 14.875, "learning_rate": 1.9057359857075794e-06, "loss": 0.8212, "step": 7677 }, { "epoch": 0.6542820622070729, "grad_norm": 17.0, "learning_rate": 1.9055929217519186e-06, "loss": 0.5392, "step": 7678 }, { "epoch": 0.6543672773753728, "grad_norm": 8.1875, "learning_rate": 1.9054498459490357e-06, "loss": 0.1814, "step": 7679 }, { "epoch": 0.6544524925436728, "grad_norm": 12.4375, "learning_rate": 1.9053067583015138e-06, "loss": 0.4883, "step": 7680 }, { "epoch": 0.6545377077119727, "grad_norm": 16.75, "learning_rate": 1.9051636588119405e-06, "loss": 1.0578, "step": 7681 }, { "epoch": 0.6546229228802727, "grad_norm": 11.875, "learning_rate": 1.9050205474829005e-06, "loss": 0.7196, "step": 7682 }, { "epoch": 0.6547081380485726, "grad_norm": 11.9375, "learning_rate": 1.904877424316981e-06, "loss": 0.6645, "step": 7683 }, { "epoch": 0.6547933532168726, "grad_norm": 15.0, "learning_rate": 1.9047342893167678e-06, "loss": 0.5131, "step": 7684 }, { "epoch": 0.6548785683851726, "grad_norm": 16.125, "learning_rate": 1.9045911424848476e-06, "loss": 0.8631, "step": 7685 }, { "epoch": 0.6549637835534725, "grad_norm": 15.6875, "learning_rate": 1.904447983823807e-06, "loss": 0.4195, "step": 7686 }, { "epoch": 0.6550489987217725, "grad_norm": 21.0, "learning_rate": 1.9043048133362334e-06, "loss": 1.0715, "step": 7687 }, { "epoch": 0.6551342138900724, "grad_norm": 17.75, "learning_rate": 1.904161631024714e-06, "loss": 1.0429, "step": 7688 }, { "epoch": 0.6552194290583724, "grad_norm": 18.0, "learning_rate": 1.9040184368918358e-06, "loss": 0.9046, "step": 7689 }, { "epoch": 0.6553046442266723, "grad_norm": 15.375, "learning_rate": 1.903875230940187e-06, "loss": 0.9655, "step": 7690 }, { "epoch": 0.6553898593949723, "grad_norm": 16.625, "learning_rate": 1.9037320131723552e-06, "loss": 1.1764, "step": 7691 }, { "epoch": 0.6554750745632723, "grad_norm": 18.125, "learning_rate": 1.9035887835909289e-06, "loss": 1.071, "step": 7692 }, { "epoch": 0.6555602897315722, "grad_norm": 14.0, "learning_rate": 1.9034455421984955e-06, "loss": 0.63, "step": 7693 }, { "epoch": 0.6556455048998722, "grad_norm": 12.5, "learning_rate": 1.9033022889976444e-06, "loss": 0.5354, "step": 7694 }, { "epoch": 0.6557307200681721, "grad_norm": 14.1875, "learning_rate": 1.903159023990964e-06, "loss": 1.0363, "step": 7695 }, { "epoch": 0.6558159352364721, "grad_norm": 9.6875, "learning_rate": 1.9030157471810428e-06, "loss": 0.6402, "step": 7696 }, { "epoch": 0.655901150404772, "grad_norm": 17.125, "learning_rate": 1.9028724585704708e-06, "loss": 1.0597, "step": 7697 }, { "epoch": 0.655986365573072, "grad_norm": 12.9375, "learning_rate": 1.9027291581618368e-06, "loss": 0.698, "step": 7698 }, { "epoch": 0.656071580741372, "grad_norm": 12.25, "learning_rate": 1.9025858459577306e-06, "loss": 0.4694, "step": 7699 }, { "epoch": 0.6561567959096719, "grad_norm": 13.5625, "learning_rate": 1.9024425219607422e-06, "loss": 0.6339, "step": 7700 }, { "epoch": 0.6562420110779719, "grad_norm": 10.875, "learning_rate": 1.9022991861734607e-06, "loss": 0.2362, "step": 7701 }, { "epoch": 0.6563272262462718, "grad_norm": 12.0625, "learning_rate": 1.9021558385984776e-06, "loss": 0.4518, "step": 7702 }, { "epoch": 0.6564124414145718, "grad_norm": 16.25, "learning_rate": 1.9020124792383826e-06, "loss": 0.5414, "step": 7703 }, { "epoch": 0.6564976565828717, "grad_norm": 16.5, "learning_rate": 1.9018691080957663e-06, "loss": 0.8466, "step": 7704 }, { "epoch": 0.6565828717511717, "grad_norm": 12.3125, "learning_rate": 1.9017257251732197e-06, "loss": 0.4506, "step": 7705 }, { "epoch": 0.6566680869194716, "grad_norm": 14.5625, "learning_rate": 1.9015823304733337e-06, "loss": 0.9776, "step": 7706 }, { "epoch": 0.6567533020877716, "grad_norm": 11.3125, "learning_rate": 1.9014389239987003e-06, "loss": 0.5014, "step": 7707 }, { "epoch": 0.6568385172560716, "grad_norm": 19.75, "learning_rate": 1.90129550575191e-06, "loss": 1.0714, "step": 7708 }, { "epoch": 0.6569237324243715, "grad_norm": 15.1875, "learning_rate": 1.901152075735555e-06, "loss": 0.876, "step": 7709 }, { "epoch": 0.6570089475926715, "grad_norm": 13.8125, "learning_rate": 1.9010086339522272e-06, "loss": 0.8053, "step": 7710 }, { "epoch": 0.6570941627609714, "grad_norm": 13.625, "learning_rate": 1.9008651804045192e-06, "loss": 0.6776, "step": 7711 }, { "epoch": 0.6571793779292714, "grad_norm": 17.375, "learning_rate": 1.9007217150950224e-06, "loss": 0.8731, "step": 7712 }, { "epoch": 0.6572645930975713, "grad_norm": 15.875, "learning_rate": 1.9005782380263301e-06, "loss": 0.6487, "step": 7713 }, { "epoch": 0.6573498082658713, "grad_norm": 11.5, "learning_rate": 1.9004347492010344e-06, "loss": 0.5321, "step": 7714 }, { "epoch": 0.6574350234341713, "grad_norm": 13.875, "learning_rate": 1.9002912486217294e-06, "loss": 0.7253, "step": 7715 }, { "epoch": 0.6575202386024712, "grad_norm": 19.5, "learning_rate": 1.9001477362910072e-06, "loss": 0.9307, "step": 7716 }, { "epoch": 0.6576054537707712, "grad_norm": 24.125, "learning_rate": 1.900004212211462e-06, "loss": 0.8846, "step": 7717 }, { "epoch": 0.6576906689390711, "grad_norm": 11.0, "learning_rate": 1.8998606763856864e-06, "loss": 0.375, "step": 7718 }, { "epoch": 0.6577758841073711, "grad_norm": 13.1875, "learning_rate": 1.8997171288162756e-06, "loss": 0.5148, "step": 7719 }, { "epoch": 0.657861099275671, "grad_norm": 10.5625, "learning_rate": 1.8995735695058226e-06, "loss": 0.4347, "step": 7720 }, { "epoch": 0.657946314443971, "grad_norm": 10.625, "learning_rate": 1.899429998456922e-06, "loss": 0.3172, "step": 7721 }, { "epoch": 0.658031529612271, "grad_norm": 16.125, "learning_rate": 1.8992864156721682e-06, "loss": 0.7518, "step": 7722 }, { "epoch": 0.6581167447805709, "grad_norm": 18.25, "learning_rate": 1.8991428211541564e-06, "loss": 0.7375, "step": 7723 }, { "epoch": 0.6582019599488709, "grad_norm": 12.8125, "learning_rate": 1.8989992149054806e-06, "loss": 0.7669, "step": 7724 }, { "epoch": 0.6582871751171708, "grad_norm": 12.5625, "learning_rate": 1.8988555969287365e-06, "loss": 0.6209, "step": 7725 }, { "epoch": 0.6583723902854708, "grad_norm": 16.75, "learning_rate": 1.898711967226519e-06, "loss": 0.8698, "step": 7726 }, { "epoch": 0.6584576054537707, "grad_norm": 13.375, "learning_rate": 1.8985683258014244e-06, "loss": 0.7797, "step": 7727 }, { "epoch": 0.6585428206220707, "grad_norm": 12.3125, "learning_rate": 1.8984246726560478e-06, "loss": 0.7394, "step": 7728 }, { "epoch": 0.6586280357903707, "grad_norm": 17.875, "learning_rate": 1.8982810077929854e-06, "loss": 0.6197, "step": 7729 }, { "epoch": 0.6587132509586706, "grad_norm": 15.75, "learning_rate": 1.898137331214833e-06, "loss": 0.4921, "step": 7730 }, { "epoch": 0.6587984661269706, "grad_norm": 17.125, "learning_rate": 1.8979936429241874e-06, "loss": 0.9318, "step": 7731 }, { "epoch": 0.6588836812952705, "grad_norm": 12.4375, "learning_rate": 1.8978499429236451e-06, "loss": 0.6064, "step": 7732 }, { "epoch": 0.6589688964635705, "grad_norm": 15.75, "learning_rate": 1.897706231215803e-06, "loss": 0.3024, "step": 7733 }, { "epoch": 0.6590541116318704, "grad_norm": 13.0625, "learning_rate": 1.8975625078032577e-06, "loss": 0.7487, "step": 7734 }, { "epoch": 0.6591393268001704, "grad_norm": 14.125, "learning_rate": 1.8974187726886068e-06, "loss": 0.7034, "step": 7735 }, { "epoch": 0.6592245419684704, "grad_norm": 15.8125, "learning_rate": 1.8972750258744478e-06, "loss": 0.7132, "step": 7736 }, { "epoch": 0.6593097571367703, "grad_norm": 16.125, "learning_rate": 1.8971312673633782e-06, "loss": 0.8852, "step": 7737 }, { "epoch": 0.6593949723050703, "grad_norm": 13.625, "learning_rate": 1.8969874971579954e-06, "loss": 0.8059, "step": 7738 }, { "epoch": 0.6594801874733702, "grad_norm": 9.875, "learning_rate": 1.8968437152608987e-06, "loss": 0.3816, "step": 7739 }, { "epoch": 0.6595654026416702, "grad_norm": 23.75, "learning_rate": 1.896699921674685e-06, "loss": 0.9941, "step": 7740 }, { "epoch": 0.6596506178099701, "grad_norm": 18.25, "learning_rate": 1.8965561164019538e-06, "loss": 1.0777, "step": 7741 }, { "epoch": 0.6597358329782701, "grad_norm": 15.5625, "learning_rate": 1.8964122994453032e-06, "loss": 0.8716, "step": 7742 }, { "epoch": 0.65982104814657, "grad_norm": 10.9375, "learning_rate": 1.8962684708073327e-06, "loss": 0.4156, "step": 7743 }, { "epoch": 0.65990626331487, "grad_norm": 18.375, "learning_rate": 1.8961246304906406e-06, "loss": 0.84, "step": 7744 }, { "epoch": 0.65999147848317, "grad_norm": 22.25, "learning_rate": 1.8959807784978268e-06, "loss": 0.9273, "step": 7745 }, { "epoch": 0.6600766936514699, "grad_norm": 14.1875, "learning_rate": 1.8958369148314907e-06, "loss": 0.7878, "step": 7746 }, { "epoch": 0.6601619088197699, "grad_norm": 19.875, "learning_rate": 1.8956930394942323e-06, "loss": 1.1043, "step": 7747 }, { "epoch": 0.6602471239880698, "grad_norm": 12.875, "learning_rate": 1.8955491524886513e-06, "loss": 0.6509, "step": 7748 }, { "epoch": 0.6603323391563698, "grad_norm": 10.625, "learning_rate": 1.8954052538173478e-06, "loss": 0.3943, "step": 7749 }, { "epoch": 0.6604175543246698, "grad_norm": 26.125, "learning_rate": 1.8952613434829226e-06, "loss": 1.2321, "step": 7750 }, { "epoch": 0.6605027694929697, "grad_norm": 17.875, "learning_rate": 1.8951174214879757e-06, "loss": 1.2233, "step": 7751 }, { "epoch": 0.6605879846612697, "grad_norm": 18.75, "learning_rate": 1.894973487835109e-06, "loss": 0.7206, "step": 7752 }, { "epoch": 0.6606731998295696, "grad_norm": 11.9375, "learning_rate": 1.894829542526922e-06, "loss": 0.4861, "step": 7753 }, { "epoch": 0.6607584149978696, "grad_norm": 14.6875, "learning_rate": 1.8946855855660174e-06, "loss": 0.6899, "step": 7754 }, { "epoch": 0.6608436301661695, "grad_norm": 19.625, "learning_rate": 1.8945416169549954e-06, "loss": 1.0371, "step": 7755 }, { "epoch": 0.6609288453344695, "grad_norm": 13.375, "learning_rate": 1.8943976366964586e-06, "loss": 0.8959, "step": 7756 }, { "epoch": 0.6610140605027695, "grad_norm": 13.8125, "learning_rate": 1.8942536447930081e-06, "loss": 0.8161, "step": 7757 }, { "epoch": 0.6610992756710694, "grad_norm": 13.1875, "learning_rate": 1.8941096412472472e-06, "loss": 0.4195, "step": 7758 }, { "epoch": 0.6611844908393694, "grad_norm": 13.875, "learning_rate": 1.8939656260617767e-06, "loss": 1.0856, "step": 7759 }, { "epoch": 0.6612697060076693, "grad_norm": 14.5, "learning_rate": 1.8938215992392e-06, "loss": 0.5116, "step": 7760 }, { "epoch": 0.6613549211759693, "grad_norm": 14.75, "learning_rate": 1.8936775607821194e-06, "loss": 0.8593, "step": 7761 }, { "epoch": 0.6614401363442692, "grad_norm": 17.5, "learning_rate": 1.8935335106931385e-06, "loss": 0.5395, "step": 7762 }, { "epoch": 0.6615253515125692, "grad_norm": 12.1875, "learning_rate": 1.8933894489748595e-06, "loss": 0.649, "step": 7763 }, { "epoch": 0.6616105666808691, "grad_norm": 15.625, "learning_rate": 1.8932453756298864e-06, "loss": 0.8281, "step": 7764 }, { "epoch": 0.6616957818491691, "grad_norm": 16.0, "learning_rate": 1.8931012906608222e-06, "loss": 0.918, "step": 7765 }, { "epoch": 0.6617809970174691, "grad_norm": 8.9375, "learning_rate": 1.8929571940702712e-06, "loss": 0.5456, "step": 7766 }, { "epoch": 0.661866212185769, "grad_norm": 25.5, "learning_rate": 1.8928130858608373e-06, "loss": 1.1243, "step": 7767 }, { "epoch": 0.661951427354069, "grad_norm": 9.3125, "learning_rate": 1.8926689660351243e-06, "loss": 0.2324, "step": 7768 }, { "epoch": 0.6620366425223689, "grad_norm": 14.125, "learning_rate": 1.892524834595737e-06, "loss": 0.7314, "step": 7769 }, { "epoch": 0.6621218576906689, "grad_norm": 15.5, "learning_rate": 1.89238069154528e-06, "loss": 0.8729, "step": 7770 }, { "epoch": 0.6622070728589688, "grad_norm": 16.0, "learning_rate": 1.8922365368863578e-06, "loss": 0.6991, "step": 7771 }, { "epoch": 0.6622922880272688, "grad_norm": 11.6875, "learning_rate": 1.8920923706215751e-06, "loss": 0.5239, "step": 7772 }, { "epoch": 0.6623775031955688, "grad_norm": 13.0625, "learning_rate": 1.8919481927535382e-06, "loss": 0.6511, "step": 7773 }, { "epoch": 0.6624627183638687, "grad_norm": 20.75, "learning_rate": 1.8918040032848515e-06, "loss": 0.684, "step": 7774 }, { "epoch": 0.6625479335321687, "grad_norm": 21.25, "learning_rate": 1.8916598022181213e-06, "loss": 1.0745, "step": 7775 }, { "epoch": 0.6626331487004686, "grad_norm": 20.375, "learning_rate": 1.8915155895559534e-06, "loss": 0.7396, "step": 7776 }, { "epoch": 0.6627183638687686, "grad_norm": 17.25, "learning_rate": 1.8913713653009536e-06, "loss": 0.5338, "step": 7777 }, { "epoch": 0.6628035790370685, "grad_norm": 12.0625, "learning_rate": 1.8912271294557277e-06, "loss": 0.5244, "step": 7778 }, { "epoch": 0.6628887942053685, "grad_norm": 14.5, "learning_rate": 1.8910828820228836e-06, "loss": 0.4698, "step": 7779 }, { "epoch": 0.6629740093736685, "grad_norm": 14.0625, "learning_rate": 1.890938623005027e-06, "loss": 0.7289, "step": 7780 }, { "epoch": 0.6630592245419684, "grad_norm": 16.875, "learning_rate": 1.8907943524047645e-06, "loss": 0.9359, "step": 7781 }, { "epoch": 0.6631444397102684, "grad_norm": 19.75, "learning_rate": 1.8906500702247038e-06, "loss": 1.0226, "step": 7782 }, { "epoch": 0.6632296548785683, "grad_norm": 12.75, "learning_rate": 1.8905057764674522e-06, "loss": 0.416, "step": 7783 }, { "epoch": 0.6633148700468683, "grad_norm": 14.8125, "learning_rate": 1.8903614711356178e-06, "loss": 0.4751, "step": 7784 }, { "epoch": 0.6634000852151682, "grad_norm": 13.5, "learning_rate": 1.890217154231807e-06, "loss": 0.7512, "step": 7785 }, { "epoch": 0.6634853003834683, "grad_norm": 13.9375, "learning_rate": 1.8900728257586282e-06, "loss": 0.7259, "step": 7786 }, { "epoch": 0.6635705155517683, "grad_norm": 18.875, "learning_rate": 1.8899284857186903e-06, "loss": 0.9254, "step": 7787 }, { "epoch": 0.6636557307200682, "grad_norm": 16.875, "learning_rate": 1.8897841341146012e-06, "loss": 0.849, "step": 7788 }, { "epoch": 0.6637409458883682, "grad_norm": 13.0625, "learning_rate": 1.8896397709489688e-06, "loss": 1.0384, "step": 7789 }, { "epoch": 0.6638261610566681, "grad_norm": 13.625, "learning_rate": 1.889495396224403e-06, "loss": 0.7918, "step": 7790 }, { "epoch": 0.6639113762249681, "grad_norm": 25.0, "learning_rate": 1.8893510099435122e-06, "loss": 1.2188, "step": 7791 }, { "epoch": 0.663996591393268, "grad_norm": 9.5625, "learning_rate": 1.889206612108906e-06, "loss": 0.2705, "step": 7792 }, { "epoch": 0.664081806561568, "grad_norm": 13.125, "learning_rate": 1.8890622027231937e-06, "loss": 0.7039, "step": 7793 }, { "epoch": 0.664167021729868, "grad_norm": 14.3125, "learning_rate": 1.8889177817889841e-06, "loss": 0.9733, "step": 7794 }, { "epoch": 0.6642522368981679, "grad_norm": 14.3125, "learning_rate": 1.8887733493088882e-06, "loss": 0.621, "step": 7795 }, { "epoch": 0.6643374520664679, "grad_norm": 11.0625, "learning_rate": 1.8886289052855155e-06, "loss": 0.4317, "step": 7796 }, { "epoch": 0.6644226672347678, "grad_norm": 13.6875, "learning_rate": 1.888484449721476e-06, "loss": 0.7371, "step": 7797 }, { "epoch": 0.6645078824030678, "grad_norm": 14.6875, "learning_rate": 1.8883399826193807e-06, "loss": 0.553, "step": 7798 }, { "epoch": 0.6645930975713678, "grad_norm": 12.1875, "learning_rate": 1.8881955039818398e-06, "loss": 0.3941, "step": 7799 }, { "epoch": 0.6646783127396677, "grad_norm": 13.375, "learning_rate": 1.8880510138114646e-06, "loss": 0.4722, "step": 7800 }, { "epoch": 0.6647635279079677, "grad_norm": 10.3125, "learning_rate": 1.8879065121108659e-06, "loss": 0.4653, "step": 7801 }, { "epoch": 0.6648487430762676, "grad_norm": 24.625, "learning_rate": 1.887761998882655e-06, "loss": 1.0067, "step": 7802 }, { "epoch": 0.6649339582445676, "grad_norm": 15.5625, "learning_rate": 1.8876174741294434e-06, "loss": 0.991, "step": 7803 }, { "epoch": 0.6650191734128675, "grad_norm": 13.6875, "learning_rate": 1.8874729378538432e-06, "loss": 0.396, "step": 7804 }, { "epoch": 0.6651043885811675, "grad_norm": 19.75, "learning_rate": 1.887328390058466e-06, "loss": 1.0518, "step": 7805 }, { "epoch": 0.6651896037494675, "grad_norm": 14.5625, "learning_rate": 1.8871838307459234e-06, "loss": 0.8238, "step": 7806 }, { "epoch": 0.6652748189177674, "grad_norm": 10.1875, "learning_rate": 1.8870392599188286e-06, "loss": 0.3528, "step": 7807 }, { "epoch": 0.6653600340860674, "grad_norm": 18.125, "learning_rate": 1.8868946775797936e-06, "loss": 0.4562, "step": 7808 }, { "epoch": 0.6654452492543673, "grad_norm": 14.75, "learning_rate": 1.8867500837314318e-06, "loss": 0.941, "step": 7809 }, { "epoch": 0.6655304644226673, "grad_norm": 17.75, "learning_rate": 1.886605478376355e-06, "loss": 0.6784, "step": 7810 }, { "epoch": 0.6656156795909672, "grad_norm": 15.25, "learning_rate": 1.8864608615171776e-06, "loss": 0.7678, "step": 7811 }, { "epoch": 0.6657008947592672, "grad_norm": 17.375, "learning_rate": 1.886316233156512e-06, "loss": 0.7316, "step": 7812 }, { "epoch": 0.6657861099275671, "grad_norm": 29.625, "learning_rate": 1.8861715932969727e-06, "loss": 0.8083, "step": 7813 }, { "epoch": 0.6658713250958671, "grad_norm": 13.1875, "learning_rate": 1.8860269419411724e-06, "loss": 0.4648, "step": 7814 }, { "epoch": 0.6659565402641671, "grad_norm": 26.125, "learning_rate": 1.885882279091726e-06, "loss": 0.6458, "step": 7815 }, { "epoch": 0.666041755432467, "grad_norm": 12.5625, "learning_rate": 1.8857376047512471e-06, "loss": 0.6986, "step": 7816 }, { "epoch": 0.666126970600767, "grad_norm": 20.625, "learning_rate": 1.885592918922351e-06, "loss": 1.1474, "step": 7817 }, { "epoch": 0.6662121857690669, "grad_norm": 23.875, "learning_rate": 1.8854482216076514e-06, "loss": 0.8777, "step": 7818 }, { "epoch": 0.6662974009373669, "grad_norm": 16.0, "learning_rate": 1.8853035128097635e-06, "loss": 0.6345, "step": 7819 }, { "epoch": 0.6663826161056668, "grad_norm": 19.0, "learning_rate": 1.885158792531302e-06, "loss": 0.8317, "step": 7820 }, { "epoch": 0.6664678312739668, "grad_norm": 14.375, "learning_rate": 1.8850140607748826e-06, "loss": 0.9223, "step": 7821 }, { "epoch": 0.6665530464422668, "grad_norm": 11.375, "learning_rate": 1.8848693175431206e-06, "loss": 0.5993, "step": 7822 }, { "epoch": 0.6666382616105667, "grad_norm": 14.8125, "learning_rate": 1.8847245628386313e-06, "loss": 0.5059, "step": 7823 }, { "epoch": 0.6667234767788667, "grad_norm": 11.3125, "learning_rate": 1.8845797966640314e-06, "loss": 0.4261, "step": 7824 }, { "epoch": 0.6668086919471666, "grad_norm": 11.0, "learning_rate": 1.8844350190219362e-06, "loss": 0.5137, "step": 7825 }, { "epoch": 0.6668939071154666, "grad_norm": 18.25, "learning_rate": 1.8842902299149624e-06, "loss": 1.0134, "step": 7826 }, { "epoch": 0.6669791222837665, "grad_norm": 13.0, "learning_rate": 1.8841454293457263e-06, "loss": 0.5634, "step": 7827 }, { "epoch": 0.6670643374520665, "grad_norm": 23.0, "learning_rate": 1.8840006173168442e-06, "loss": 0.4067, "step": 7828 }, { "epoch": 0.6671495526203665, "grad_norm": 13.25, "learning_rate": 1.8838557938309337e-06, "loss": 0.4343, "step": 7829 }, { "epoch": 0.6672347677886664, "grad_norm": 14.5625, "learning_rate": 1.883710958890612e-06, "loss": 0.817, "step": 7830 }, { "epoch": 0.6673199829569664, "grad_norm": 19.5, "learning_rate": 1.8835661124984956e-06, "loss": 0.8981, "step": 7831 }, { "epoch": 0.6674051981252663, "grad_norm": 14.375, "learning_rate": 1.8834212546572023e-06, "loss": 0.698, "step": 7832 }, { "epoch": 0.6674904132935663, "grad_norm": 18.875, "learning_rate": 1.8832763853693502e-06, "loss": 0.7266, "step": 7833 }, { "epoch": 0.6675756284618662, "grad_norm": 14.5625, "learning_rate": 1.883131504637557e-06, "loss": 0.715, "step": 7834 }, { "epoch": 0.6676608436301662, "grad_norm": 12.4375, "learning_rate": 1.8829866124644408e-06, "loss": 0.6582, "step": 7835 }, { "epoch": 0.6677460587984662, "grad_norm": 15.8125, "learning_rate": 1.8828417088526203e-06, "loss": 0.7126, "step": 7836 }, { "epoch": 0.6678312739667661, "grad_norm": 19.0, "learning_rate": 1.8826967938047134e-06, "loss": 1.1873, "step": 7837 }, { "epoch": 0.6679164891350661, "grad_norm": 18.0, "learning_rate": 1.882551867323339e-06, "loss": 1.1855, "step": 7838 }, { "epoch": 0.668001704303366, "grad_norm": 15.625, "learning_rate": 1.8824069294111166e-06, "loss": 0.9735, "step": 7839 }, { "epoch": 0.668086919471666, "grad_norm": 14.3125, "learning_rate": 1.8822619800706649e-06, "loss": 0.815, "step": 7840 }, { "epoch": 0.6681721346399659, "grad_norm": 17.25, "learning_rate": 1.8821170193046038e-06, "loss": 0.6461, "step": 7841 }, { "epoch": 0.6682573498082659, "grad_norm": 17.5, "learning_rate": 1.8819720471155519e-06, "loss": 0.8103, "step": 7842 }, { "epoch": 0.6683425649765659, "grad_norm": 12.9375, "learning_rate": 1.8818270635061303e-06, "loss": 0.5115, "step": 7843 }, { "epoch": 0.6684277801448658, "grad_norm": 21.5, "learning_rate": 1.8816820684789578e-06, "loss": 1.2185, "step": 7844 }, { "epoch": 0.6685129953131658, "grad_norm": 17.5, "learning_rate": 1.8815370620366554e-06, "loss": 1.0734, "step": 7845 }, { "epoch": 0.6685982104814657, "grad_norm": 13.25, "learning_rate": 1.8813920441818428e-06, "loss": 0.7405, "step": 7846 }, { "epoch": 0.6686834256497657, "grad_norm": 10.5, "learning_rate": 1.8812470149171416e-06, "loss": 0.3884, "step": 7847 }, { "epoch": 0.6687686408180656, "grad_norm": 13.0625, "learning_rate": 1.881101974245172e-06, "loss": 0.4499, "step": 7848 }, { "epoch": 0.6688538559863656, "grad_norm": 26.625, "learning_rate": 1.880956922168555e-06, "loss": 0.769, "step": 7849 }, { "epoch": 0.6689390711546656, "grad_norm": 13.0625, "learning_rate": 1.8808118586899116e-06, "loss": 0.709, "step": 7850 }, { "epoch": 0.6690242863229655, "grad_norm": 12.25, "learning_rate": 1.880666783811864e-06, "loss": 0.5738, "step": 7851 }, { "epoch": 0.6691095014912655, "grad_norm": 16.875, "learning_rate": 1.8805216975370334e-06, "loss": 0.7661, "step": 7852 }, { "epoch": 0.6691947166595654, "grad_norm": 15.9375, "learning_rate": 1.8803765998680414e-06, "loss": 0.7259, "step": 7853 }, { "epoch": 0.6692799318278654, "grad_norm": 13.0, "learning_rate": 1.8802314908075107e-06, "loss": 0.7106, "step": 7854 }, { "epoch": 0.6693651469961653, "grad_norm": 18.25, "learning_rate": 1.880086370358063e-06, "loss": 0.5793, "step": 7855 }, { "epoch": 0.6694503621644653, "grad_norm": 15.4375, "learning_rate": 1.8799412385223208e-06, "loss": 0.4874, "step": 7856 }, { "epoch": 0.6695355773327653, "grad_norm": 10.75, "learning_rate": 1.8797960953029076e-06, "loss": 0.5157, "step": 7857 }, { "epoch": 0.6696207925010652, "grad_norm": 13.0, "learning_rate": 1.8796509407024452e-06, "loss": 0.7985, "step": 7858 }, { "epoch": 0.6697060076693652, "grad_norm": 18.75, "learning_rate": 1.8795057747235569e-06, "loss": 0.6469, "step": 7859 }, { "epoch": 0.6697912228376651, "grad_norm": 19.375, "learning_rate": 1.8793605973688666e-06, "loss": 0.8221, "step": 7860 }, { "epoch": 0.6698764380059651, "grad_norm": 14.375, "learning_rate": 1.8792154086409974e-06, "loss": 1.0967, "step": 7861 }, { "epoch": 0.669961653174265, "grad_norm": 13.5, "learning_rate": 1.8790702085425732e-06, "loss": 0.5723, "step": 7862 }, { "epoch": 0.670046868342565, "grad_norm": 14.3125, "learning_rate": 1.8789249970762174e-06, "loss": 0.5327, "step": 7863 }, { "epoch": 0.670132083510865, "grad_norm": 15.25, "learning_rate": 1.8787797742445545e-06, "loss": 0.567, "step": 7864 }, { "epoch": 0.6702172986791649, "grad_norm": 16.25, "learning_rate": 1.8786345400502092e-06, "loss": 0.5532, "step": 7865 }, { "epoch": 0.6703025138474649, "grad_norm": 16.0, "learning_rate": 1.8784892944958055e-06, "loss": 1.0269, "step": 7866 }, { "epoch": 0.6703877290157648, "grad_norm": 15.875, "learning_rate": 1.8783440375839679e-06, "loss": 1.0999, "step": 7867 }, { "epoch": 0.6704729441840648, "grad_norm": 14.75, "learning_rate": 1.8781987693173218e-06, "loss": 0.7037, "step": 7868 }, { "epoch": 0.6705581593523647, "grad_norm": 11.125, "learning_rate": 1.878053489698492e-06, "loss": 0.364, "step": 7869 }, { "epoch": 0.6706433745206647, "grad_norm": 11.8125, "learning_rate": 1.8779081987301049e-06, "loss": 0.4882, "step": 7870 }, { "epoch": 0.6707285896889646, "grad_norm": 17.0, "learning_rate": 1.8777628964147845e-06, "loss": 0.6133, "step": 7871 }, { "epoch": 0.6708138048572646, "grad_norm": 31.75, "learning_rate": 1.8776175827551577e-06, "loss": 1.3091, "step": 7872 }, { "epoch": 0.6708990200255646, "grad_norm": 10.875, "learning_rate": 1.8774722577538497e-06, "loss": 0.534, "step": 7873 }, { "epoch": 0.6709842351938645, "grad_norm": 16.375, "learning_rate": 1.8773269214134871e-06, "loss": 0.8544, "step": 7874 }, { "epoch": 0.6710694503621645, "grad_norm": 18.125, "learning_rate": 1.8771815737366967e-06, "loss": 0.8822, "step": 7875 }, { "epoch": 0.6711546655304644, "grad_norm": 16.625, "learning_rate": 1.877036214726104e-06, "loss": 0.6763, "step": 7876 }, { "epoch": 0.6712398806987644, "grad_norm": 16.25, "learning_rate": 1.8768908443843363e-06, "loss": 0.7085, "step": 7877 }, { "epoch": 0.6713250958670643, "grad_norm": 9.6875, "learning_rate": 1.8767454627140214e-06, "loss": 0.2832, "step": 7878 }, { "epoch": 0.6714103110353643, "grad_norm": 14.0625, "learning_rate": 1.8766000697177853e-06, "loss": 0.8317, "step": 7879 }, { "epoch": 0.6714955262036643, "grad_norm": 14.25, "learning_rate": 1.8764546653982558e-06, "loss": 0.4419, "step": 7880 }, { "epoch": 0.6715807413719642, "grad_norm": 12.375, "learning_rate": 1.8763092497580603e-06, "loss": 0.7417, "step": 7881 }, { "epoch": 0.6716659565402642, "grad_norm": 17.375, "learning_rate": 1.8761638227998274e-06, "loss": 0.6472, "step": 7882 }, { "epoch": 0.6717511717085641, "grad_norm": 26.25, "learning_rate": 1.8760183845261842e-06, "loss": 0.9428, "step": 7883 }, { "epoch": 0.6718363868768641, "grad_norm": 13.875, "learning_rate": 1.875872934939759e-06, "loss": 0.6219, "step": 7884 }, { "epoch": 0.671921602045164, "grad_norm": 18.5, "learning_rate": 1.8757274740431813e-06, "loss": 0.9919, "step": 7885 }, { "epoch": 0.672006817213464, "grad_norm": 11.8125, "learning_rate": 1.8755820018390786e-06, "loss": 0.5577, "step": 7886 }, { "epoch": 0.672092032381764, "grad_norm": 13.1875, "learning_rate": 1.8754365183300798e-06, "loss": 0.4373, "step": 7887 }, { "epoch": 0.6721772475500639, "grad_norm": 13.3125, "learning_rate": 1.8752910235188143e-06, "loss": 0.652, "step": 7888 }, { "epoch": 0.6722624627183639, "grad_norm": 10.25, "learning_rate": 1.8751455174079115e-06, "loss": 0.3525, "step": 7889 }, { "epoch": 0.6723476778866638, "grad_norm": 19.5, "learning_rate": 1.8750000000000003e-06, "loss": 0.8254, "step": 7890 }, { "epoch": 0.6724328930549638, "grad_norm": 13.5, "learning_rate": 1.8748544712977104e-06, "loss": 0.5942, "step": 7891 }, { "epoch": 0.6725181082232637, "grad_norm": 24.625, "learning_rate": 1.8747089313036722e-06, "loss": 1.1972, "step": 7892 }, { "epoch": 0.6726033233915637, "grad_norm": 14.8125, "learning_rate": 1.874563380020515e-06, "loss": 0.7444, "step": 7893 }, { "epoch": 0.6726885385598637, "grad_norm": 14.9375, "learning_rate": 1.8744178174508703e-06, "loss": 0.4562, "step": 7894 }, { "epoch": 0.6727737537281636, "grad_norm": 19.625, "learning_rate": 1.874272243597367e-06, "loss": 1.2864, "step": 7895 }, { "epoch": 0.6728589688964636, "grad_norm": 26.0, "learning_rate": 1.874126658462637e-06, "loss": 0.9159, "step": 7896 }, { "epoch": 0.6729441840647635, "grad_norm": 10.5, "learning_rate": 1.8739810620493101e-06, "loss": 0.3869, "step": 7897 }, { "epoch": 0.6730293992330635, "grad_norm": 13.4375, "learning_rate": 1.8738354543600187e-06, "loss": 0.7865, "step": 7898 }, { "epoch": 0.6731146144013634, "grad_norm": 14.375, "learning_rate": 1.8736898353973925e-06, "loss": 0.4406, "step": 7899 }, { "epoch": 0.6731998295696634, "grad_norm": 10.4375, "learning_rate": 1.8735442051640644e-06, "loss": 0.4261, "step": 7900 }, { "epoch": 0.6732850447379634, "grad_norm": 17.375, "learning_rate": 1.8733985636626653e-06, "loss": 0.6808, "step": 7901 }, { "epoch": 0.6733702599062633, "grad_norm": 20.125, "learning_rate": 1.8732529108958275e-06, "loss": 0.9224, "step": 7902 }, { "epoch": 0.6734554750745633, "grad_norm": 21.875, "learning_rate": 1.8731072468661826e-06, "loss": 1.3805, "step": 7903 }, { "epoch": 0.6735406902428632, "grad_norm": 13.5, "learning_rate": 1.8729615715763633e-06, "loss": 0.4637, "step": 7904 }, { "epoch": 0.6736259054111632, "grad_norm": 12.875, "learning_rate": 1.8728158850290016e-06, "loss": 0.4815, "step": 7905 }, { "epoch": 0.6737111205794631, "grad_norm": 11.5625, "learning_rate": 1.8726701872267309e-06, "loss": 0.5942, "step": 7906 }, { "epoch": 0.6737963357477631, "grad_norm": 12.9375, "learning_rate": 1.8725244781721837e-06, "loss": 0.5458, "step": 7907 }, { "epoch": 0.673881550916063, "grad_norm": 14.875, "learning_rate": 1.872378757867993e-06, "loss": 0.7109, "step": 7908 }, { "epoch": 0.673966766084363, "grad_norm": 18.125, "learning_rate": 1.8722330263167926e-06, "loss": 0.9752, "step": 7909 }, { "epoch": 0.674051981252663, "grad_norm": 14.4375, "learning_rate": 1.872087283521216e-06, "loss": 0.6463, "step": 7910 }, { "epoch": 0.6741371964209629, "grad_norm": 14.375, "learning_rate": 1.8719415294838962e-06, "loss": 0.5688, "step": 7911 }, { "epoch": 0.6742224115892629, "grad_norm": 13.0625, "learning_rate": 1.8717957642074675e-06, "loss": 0.5499, "step": 7912 }, { "epoch": 0.6743076267575628, "grad_norm": 13.6875, "learning_rate": 1.871649987694564e-06, "loss": 0.7316, "step": 7913 }, { "epoch": 0.6743928419258628, "grad_norm": 18.125, "learning_rate": 1.8715041999478204e-06, "loss": 0.7385, "step": 7914 }, { "epoch": 0.6744780570941628, "grad_norm": 14.75, "learning_rate": 1.871358400969871e-06, "loss": 0.6208, "step": 7915 }, { "epoch": 0.6745632722624627, "grad_norm": 16.75, "learning_rate": 1.87121259076335e-06, "loss": 1.0545, "step": 7916 }, { "epoch": 0.6746484874307627, "grad_norm": 13.9375, "learning_rate": 1.8710667693308932e-06, "loss": 0.6094, "step": 7917 }, { "epoch": 0.6747337025990626, "grad_norm": 17.25, "learning_rate": 1.8709209366751351e-06, "loss": 0.787, "step": 7918 }, { "epoch": 0.6748189177673626, "grad_norm": 16.5, "learning_rate": 1.8707750927987116e-06, "loss": 0.5271, "step": 7919 }, { "epoch": 0.6749041329356625, "grad_norm": 12.8125, "learning_rate": 1.8706292377042575e-06, "loss": 0.4676, "step": 7920 }, { "epoch": 0.6749893481039625, "grad_norm": 13.8125, "learning_rate": 1.8704833713944095e-06, "loss": 0.5728, "step": 7921 }, { "epoch": 0.6750745632722625, "grad_norm": 13.8125, "learning_rate": 1.8703374938718028e-06, "loss": 0.779, "step": 7922 }, { "epoch": 0.6751597784405624, "grad_norm": 15.0625, "learning_rate": 1.8701916051390737e-06, "loss": 0.7196, "step": 7923 }, { "epoch": 0.6752449936088624, "grad_norm": 13.1875, "learning_rate": 1.8700457051988585e-06, "loss": 0.6841, "step": 7924 }, { "epoch": 0.6753302087771623, "grad_norm": 16.375, "learning_rate": 1.8698997940537944e-06, "loss": 0.8734, "step": 7925 }, { "epoch": 0.6754154239454623, "grad_norm": 13.6875, "learning_rate": 1.8697538717065175e-06, "loss": 0.8064, "step": 7926 }, { "epoch": 0.6755006391137622, "grad_norm": 12.9375, "learning_rate": 1.8696079381596647e-06, "loss": 0.5771, "step": 7927 }, { "epoch": 0.6755858542820622, "grad_norm": 10.5625, "learning_rate": 1.8694619934158734e-06, "loss": 0.3223, "step": 7928 }, { "epoch": 0.6756710694503621, "grad_norm": 12.3125, "learning_rate": 1.8693160374777809e-06, "loss": 0.7083, "step": 7929 }, { "epoch": 0.6757562846186621, "grad_norm": 12.875, "learning_rate": 1.8691700703480254e-06, "loss": 0.7299, "step": 7930 }, { "epoch": 0.6758414997869621, "grad_norm": 13.5, "learning_rate": 1.8690240920292436e-06, "loss": 0.7834, "step": 7931 }, { "epoch": 0.675926714955262, "grad_norm": 15.75, "learning_rate": 1.868878102524074e-06, "loss": 1.0039, "step": 7932 }, { "epoch": 0.676011930123562, "grad_norm": 12.25, "learning_rate": 1.8687321018351545e-06, "loss": 0.8225, "step": 7933 }, { "epoch": 0.6760971452918619, "grad_norm": 12.5, "learning_rate": 1.8685860899651246e-06, "loss": 0.423, "step": 7934 }, { "epoch": 0.6761823604601619, "grad_norm": 18.375, "learning_rate": 1.8684400669166214e-06, "loss": 0.9114, "step": 7935 }, { "epoch": 0.6762675756284618, "grad_norm": 10.3125, "learning_rate": 1.8682940326922844e-06, "loss": 0.8907, "step": 7936 }, { "epoch": 0.6763527907967618, "grad_norm": 15.1875, "learning_rate": 1.8681479872947523e-06, "loss": 0.629, "step": 7937 }, { "epoch": 0.6764380059650618, "grad_norm": 17.0, "learning_rate": 1.8680019307266647e-06, "loss": 0.7546, "step": 7938 }, { "epoch": 0.6765232211333617, "grad_norm": 24.0, "learning_rate": 1.8678558629906606e-06, "loss": 1.1357, "step": 7939 }, { "epoch": 0.6766084363016617, "grad_norm": 17.25, "learning_rate": 1.8677097840893802e-06, "loss": 1.2446, "step": 7940 }, { "epoch": 0.6766936514699616, "grad_norm": 19.625, "learning_rate": 1.8675636940254622e-06, "loss": 1.1777, "step": 7941 }, { "epoch": 0.6767788666382616, "grad_norm": 9.75, "learning_rate": 1.8674175928015475e-06, "loss": 0.4478, "step": 7942 }, { "epoch": 0.6768640818065615, "grad_norm": 10.375, "learning_rate": 1.8672714804202763e-06, "loss": 0.2777, "step": 7943 }, { "epoch": 0.6769492969748615, "grad_norm": 13.125, "learning_rate": 1.8671253568842884e-06, "loss": 0.498, "step": 7944 }, { "epoch": 0.6770345121431615, "grad_norm": 13.5625, "learning_rate": 1.8669792221962248e-06, "loss": 0.4724, "step": 7945 }, { "epoch": 0.6771197273114614, "grad_norm": 15.625, "learning_rate": 1.866833076358726e-06, "loss": 0.5586, "step": 7946 }, { "epoch": 0.6772049424797614, "grad_norm": 12.4375, "learning_rate": 1.8666869193744342e-06, "loss": 0.6965, "step": 7947 }, { "epoch": 0.6772901576480613, "grad_norm": 14.5625, "learning_rate": 1.8665407512459887e-06, "loss": 0.6138, "step": 7948 }, { "epoch": 0.6773753728163613, "grad_norm": 13.125, "learning_rate": 1.866394571976032e-06, "loss": 0.6222, "step": 7949 }, { "epoch": 0.6774605879846612, "grad_norm": 10.5, "learning_rate": 1.8662483815672055e-06, "loss": 0.4369, "step": 7950 }, { "epoch": 0.6775458031529612, "grad_norm": 15.875, "learning_rate": 1.866102180022152e-06, "loss": 0.5167, "step": 7951 }, { "epoch": 0.6776310183212612, "grad_norm": 16.375, "learning_rate": 1.8659559673435113e-06, "loss": 0.9183, "step": 7952 }, { "epoch": 0.6777162334895611, "grad_norm": 15.6875, "learning_rate": 1.8658097435339273e-06, "loss": 0.808, "step": 7953 }, { "epoch": 0.6778014486578611, "grad_norm": 12.1875, "learning_rate": 1.865663508596042e-06, "loss": 0.5715, "step": 7954 }, { "epoch": 0.677886663826161, "grad_norm": 19.125, "learning_rate": 1.8655172625324986e-06, "loss": 1.043, "step": 7955 }, { "epoch": 0.677971878994461, "grad_norm": 12.875, "learning_rate": 1.8653710053459383e-06, "loss": 0.6524, "step": 7956 }, { "epoch": 0.6780570941627609, "grad_norm": 17.125, "learning_rate": 1.8652247370390059e-06, "loss": 0.853, "step": 7957 }, { "epoch": 0.6781423093310609, "grad_norm": 14.875, "learning_rate": 1.8650784576143433e-06, "loss": 0.8039, "step": 7958 }, { "epoch": 0.6782275244993609, "grad_norm": 13.0, "learning_rate": 1.8649321670745948e-06, "loss": 0.5593, "step": 7959 }, { "epoch": 0.6783127396676608, "grad_norm": 14.0, "learning_rate": 1.8647858654224039e-06, "loss": 0.6878, "step": 7960 }, { "epoch": 0.6783979548359608, "grad_norm": 15.3125, "learning_rate": 1.8646395526604134e-06, "loss": 0.69, "step": 7961 }, { "epoch": 0.6784831700042607, "grad_norm": 29.25, "learning_rate": 1.8644932287912686e-06, "loss": 0.9136, "step": 7962 }, { "epoch": 0.6785683851725607, "grad_norm": 13.625, "learning_rate": 1.8643468938176134e-06, "loss": 0.5083, "step": 7963 }, { "epoch": 0.6786536003408606, "grad_norm": 18.5, "learning_rate": 1.8642005477420922e-06, "loss": 0.4323, "step": 7964 }, { "epoch": 0.6787388155091606, "grad_norm": 17.25, "learning_rate": 1.8640541905673487e-06, "loss": 0.7278, "step": 7965 }, { "epoch": 0.6788240306774606, "grad_norm": 16.0, "learning_rate": 1.8639078222960292e-06, "loss": 0.7939, "step": 7966 }, { "epoch": 0.6789092458457605, "grad_norm": 13.625, "learning_rate": 1.863761442930778e-06, "loss": 0.6886, "step": 7967 }, { "epoch": 0.6789944610140605, "grad_norm": 14.875, "learning_rate": 1.86361505247424e-06, "loss": 0.7742, "step": 7968 }, { "epoch": 0.6790796761823604, "grad_norm": 18.75, "learning_rate": 1.863468650929061e-06, "loss": 0.9031, "step": 7969 }, { "epoch": 0.6791648913506604, "grad_norm": 14.125, "learning_rate": 1.8633222382978867e-06, "loss": 0.8332, "step": 7970 }, { "epoch": 0.6792501065189603, "grad_norm": 16.75, "learning_rate": 1.8631758145833628e-06, "loss": 0.5905, "step": 7971 }, { "epoch": 0.6793353216872603, "grad_norm": 17.0, "learning_rate": 1.8630293797881355e-06, "loss": 1.1168, "step": 7972 }, { "epoch": 0.6794205368555603, "grad_norm": 14.25, "learning_rate": 1.8628829339148504e-06, "loss": 0.6075, "step": 7973 }, { "epoch": 0.6795057520238602, "grad_norm": 19.0, "learning_rate": 1.862736476966155e-06, "loss": 0.9521, "step": 7974 }, { "epoch": 0.6795909671921602, "grad_norm": 19.0, "learning_rate": 1.862590008944695e-06, "loss": 0.9583, "step": 7975 }, { "epoch": 0.6796761823604601, "grad_norm": 33.75, "learning_rate": 1.8624435298531174e-06, "loss": 1.3099, "step": 7976 }, { "epoch": 0.6797613975287601, "grad_norm": 13.9375, "learning_rate": 1.8622970396940693e-06, "loss": 0.5251, "step": 7977 }, { "epoch": 0.67984661269706, "grad_norm": 15.6875, "learning_rate": 1.8621505384701982e-06, "loss": 0.9311, "step": 7978 }, { "epoch": 0.67993182786536, "grad_norm": 17.125, "learning_rate": 1.8620040261841516e-06, "loss": 0.9038, "step": 7979 }, { "epoch": 0.68001704303366, "grad_norm": 12.5625, "learning_rate": 1.861857502838576e-06, "loss": 0.2994, "step": 7980 }, { "epoch": 0.6801022582019599, "grad_norm": 10.5625, "learning_rate": 1.8617109684361206e-06, "loss": 0.3945, "step": 7981 }, { "epoch": 0.6801874733702599, "grad_norm": 18.375, "learning_rate": 1.8615644229794327e-06, "loss": 1.1678, "step": 7982 }, { "epoch": 0.6802726885385598, "grad_norm": 14.0625, "learning_rate": 1.8614178664711613e-06, "loss": 0.8588, "step": 7983 }, { "epoch": 0.6803579037068598, "grad_norm": 20.5, "learning_rate": 1.8612712989139534e-06, "loss": 0.8733, "step": 7984 }, { "epoch": 0.6804431188751597, "grad_norm": 13.125, "learning_rate": 1.8611247203104587e-06, "loss": 0.6981, "step": 7985 }, { "epoch": 0.6805283340434597, "grad_norm": 11.4375, "learning_rate": 1.860978130663326e-06, "loss": 0.4357, "step": 7986 }, { "epoch": 0.6806135492117596, "grad_norm": 12.75, "learning_rate": 1.8608315299752042e-06, "loss": 0.6975, "step": 7987 }, { "epoch": 0.6806987643800596, "grad_norm": 16.25, "learning_rate": 1.8606849182487418e-06, "loss": 0.8316, "step": 7988 }, { "epoch": 0.6807839795483596, "grad_norm": 17.0, "learning_rate": 1.8605382954865894e-06, "loss": 0.7927, "step": 7989 }, { "epoch": 0.6808691947166595, "grad_norm": 18.875, "learning_rate": 1.8603916616913959e-06, "loss": 0.797, "step": 7990 }, { "epoch": 0.6809544098849595, "grad_norm": 17.625, "learning_rate": 1.8602450168658113e-06, "loss": 0.904, "step": 7991 }, { "epoch": 0.6810396250532594, "grad_norm": 17.0, "learning_rate": 1.8600983610124853e-06, "loss": 0.5518, "step": 7992 }, { "epoch": 0.6811248402215594, "grad_norm": 17.875, "learning_rate": 1.8599516941340689e-06, "loss": 0.8319, "step": 7993 }, { "epoch": 0.6812100553898593, "grad_norm": 18.75, "learning_rate": 1.8598050162332116e-06, "loss": 1.0023, "step": 7994 }, { "epoch": 0.6812952705581593, "grad_norm": 11.4375, "learning_rate": 1.8596583273125646e-06, "loss": 0.4096, "step": 7995 }, { "epoch": 0.6813804857264593, "grad_norm": 12.5625, "learning_rate": 1.8595116273747787e-06, "loss": 0.3719, "step": 7996 }, { "epoch": 0.6814657008947592, "grad_norm": 15.5625, "learning_rate": 1.8593649164225044e-06, "loss": 0.8419, "step": 7997 }, { "epoch": 0.6815509160630592, "grad_norm": 11.5625, "learning_rate": 1.8592181944583938e-06, "loss": 0.532, "step": 7998 }, { "epoch": 0.6816361312313591, "grad_norm": 15.375, "learning_rate": 1.8590714614850975e-06, "loss": 0.5619, "step": 7999 }, { "epoch": 0.6817213463996591, "grad_norm": 15.875, "learning_rate": 1.8589247175052677e-06, "loss": 0.9156, "step": 8000 }, { "epoch": 0.681806561567959, "grad_norm": 13.375, "learning_rate": 1.8587779625215554e-06, "loss": 0.6071, "step": 8001 }, { "epoch": 0.681891776736259, "grad_norm": 17.25, "learning_rate": 1.8586311965366138e-06, "loss": 0.9373, "step": 8002 }, { "epoch": 0.681976991904559, "grad_norm": 21.5, "learning_rate": 1.8584844195530939e-06, "loss": 0.9696, "step": 8003 }, { "epoch": 0.6820622070728589, "grad_norm": 16.625, "learning_rate": 1.8583376315736493e-06, "loss": 0.7568, "step": 8004 }, { "epoch": 0.6821474222411589, "grad_norm": 12.875, "learning_rate": 1.8581908326009312e-06, "loss": 0.4698, "step": 8005 }, { "epoch": 0.6822326374094588, "grad_norm": 11.6875, "learning_rate": 1.8580440226375937e-06, "loss": 0.5944, "step": 8006 }, { "epoch": 0.6823178525777588, "grad_norm": 17.75, "learning_rate": 1.857897201686289e-06, "loss": 1.0195, "step": 8007 }, { "epoch": 0.6824030677460587, "grad_norm": 17.125, "learning_rate": 1.857750369749671e-06, "loss": 0.8105, "step": 8008 }, { "epoch": 0.6824882829143587, "grad_norm": 10.4375, "learning_rate": 1.8576035268303922e-06, "loss": 0.4069, "step": 8009 }, { "epoch": 0.6825734980826587, "grad_norm": 11.5, "learning_rate": 1.8574566729311066e-06, "loss": 0.4938, "step": 8010 }, { "epoch": 0.6826587132509586, "grad_norm": 16.125, "learning_rate": 1.8573098080544682e-06, "loss": 0.8314, "step": 8011 }, { "epoch": 0.6827439284192587, "grad_norm": 13.625, "learning_rate": 1.8571629322031312e-06, "loss": 0.809, "step": 8012 }, { "epoch": 0.6828291435875586, "grad_norm": 17.625, "learning_rate": 1.8570160453797493e-06, "loss": 0.9014, "step": 8013 }, { "epoch": 0.6829143587558586, "grad_norm": 12.5625, "learning_rate": 1.8568691475869768e-06, "loss": 0.3936, "step": 8014 }, { "epoch": 0.6829995739241586, "grad_norm": 11.3125, "learning_rate": 1.856722238827469e-06, "loss": 0.472, "step": 8015 }, { "epoch": 0.6830847890924585, "grad_norm": 15.1875, "learning_rate": 1.85657531910388e-06, "loss": 0.6809, "step": 8016 }, { "epoch": 0.6831700042607585, "grad_norm": 19.0, "learning_rate": 1.856428388418865e-06, "loss": 0.6864, "step": 8017 }, { "epoch": 0.6832552194290584, "grad_norm": 25.625, "learning_rate": 1.8562814467750791e-06, "loss": 1.1295, "step": 8018 }, { "epoch": 0.6833404345973584, "grad_norm": 13.9375, "learning_rate": 1.856134494175178e-06, "loss": 0.8235, "step": 8019 }, { "epoch": 0.6834256497656583, "grad_norm": 13.25, "learning_rate": 1.8559875306218172e-06, "loss": 0.8916, "step": 8020 }, { "epoch": 0.6835108649339583, "grad_norm": 13.9375, "learning_rate": 1.855840556117652e-06, "loss": 0.7067, "step": 8021 }, { "epoch": 0.6835960801022583, "grad_norm": 11.0, "learning_rate": 1.855693570665339e-06, "loss": 0.5367, "step": 8022 }, { "epoch": 0.6836812952705582, "grad_norm": 13.8125, "learning_rate": 1.855546574267534e-06, "loss": 0.8688, "step": 8023 }, { "epoch": 0.6837665104388582, "grad_norm": 11.0625, "learning_rate": 1.8553995669268935e-06, "loss": 0.5116, "step": 8024 }, { "epoch": 0.6838517256071581, "grad_norm": 12.5625, "learning_rate": 1.855252548646074e-06, "loss": 0.7576, "step": 8025 }, { "epoch": 0.6839369407754581, "grad_norm": 14.5625, "learning_rate": 1.8551055194277324e-06, "loss": 0.7792, "step": 8026 }, { "epoch": 0.684022155943758, "grad_norm": 19.0, "learning_rate": 1.8549584792745254e-06, "loss": 0.968, "step": 8027 }, { "epoch": 0.684107371112058, "grad_norm": 12.1875, "learning_rate": 1.8548114281891104e-06, "loss": 0.5374, "step": 8028 }, { "epoch": 0.684192586280358, "grad_norm": 16.25, "learning_rate": 1.8546643661741448e-06, "loss": 0.9379, "step": 8029 }, { "epoch": 0.6842778014486579, "grad_norm": 16.0, "learning_rate": 1.854517293232286e-06, "loss": 0.7617, "step": 8030 }, { "epoch": 0.6843630166169579, "grad_norm": 15.6875, "learning_rate": 1.8543702093661918e-06, "loss": 0.7383, "step": 8031 }, { "epoch": 0.6844482317852578, "grad_norm": 14.0, "learning_rate": 1.8542231145785205e-06, "loss": 0.6043, "step": 8032 }, { "epoch": 0.6845334469535578, "grad_norm": 14.0, "learning_rate": 1.8540760088719295e-06, "loss": 0.7912, "step": 8033 }, { "epoch": 0.6846186621218577, "grad_norm": 14.875, "learning_rate": 1.8539288922490778e-06, "loss": 0.9933, "step": 8034 }, { "epoch": 0.6847038772901577, "grad_norm": 12.25, "learning_rate": 1.8537817647126235e-06, "loss": 0.6488, "step": 8035 }, { "epoch": 0.6847890924584576, "grad_norm": 12.0625, "learning_rate": 1.853634626265226e-06, "loss": 0.6229, "step": 8036 }, { "epoch": 0.6848743076267576, "grad_norm": 10.25, "learning_rate": 1.8534874769095434e-06, "loss": 0.4627, "step": 8037 }, { "epoch": 0.6849595227950576, "grad_norm": 12.5625, "learning_rate": 1.8533403166482356e-06, "loss": 1.0666, "step": 8038 }, { "epoch": 0.6850447379633575, "grad_norm": 12.1875, "learning_rate": 1.853193145483961e-06, "loss": 0.4895, "step": 8039 }, { "epoch": 0.6851299531316575, "grad_norm": 13.9375, "learning_rate": 1.8530459634193805e-06, "loss": 0.5945, "step": 8040 }, { "epoch": 0.6852151682999574, "grad_norm": 30.75, "learning_rate": 1.8528987704571522e-06, "loss": 1.6107, "step": 8041 }, { "epoch": 0.6853003834682574, "grad_norm": 11.0625, "learning_rate": 1.8527515665999375e-06, "loss": 0.6774, "step": 8042 }, { "epoch": 0.6853855986365573, "grad_norm": 13.0, "learning_rate": 1.8526043518503953e-06, "loss": 0.6746, "step": 8043 }, { "epoch": 0.6854708138048573, "grad_norm": 15.0, "learning_rate": 1.8524571262111869e-06, "loss": 0.5955, "step": 8044 }, { "epoch": 0.6855560289731573, "grad_norm": 18.125, "learning_rate": 1.8523098896849723e-06, "loss": 1.0742, "step": 8045 }, { "epoch": 0.6856412441414572, "grad_norm": 17.125, "learning_rate": 1.852162642274412e-06, "loss": 0.1923, "step": 8046 }, { "epoch": 0.6857264593097572, "grad_norm": 13.375, "learning_rate": 1.852015383982168e-06, "loss": 0.41, "step": 8047 }, { "epoch": 0.6858116744780571, "grad_norm": 20.375, "learning_rate": 1.8518681148109002e-06, "loss": 0.6981, "step": 8048 }, { "epoch": 0.6858968896463571, "grad_norm": 23.0, "learning_rate": 1.8517208347632705e-06, "loss": 1.0914, "step": 8049 }, { "epoch": 0.685982104814657, "grad_norm": 14.5, "learning_rate": 1.8515735438419396e-06, "loss": 0.7991, "step": 8050 }, { "epoch": 0.686067319982957, "grad_norm": 15.125, "learning_rate": 1.8514262420495704e-06, "loss": 0.8384, "step": 8051 }, { "epoch": 0.686152535151257, "grad_norm": 11.4375, "learning_rate": 1.8512789293888244e-06, "loss": 0.4543, "step": 8052 }, { "epoch": 0.6862377503195569, "grad_norm": 13.9375, "learning_rate": 1.8511316058623635e-06, "loss": 0.8889, "step": 8053 }, { "epoch": 0.6863229654878569, "grad_norm": 12.8125, "learning_rate": 1.8509842714728493e-06, "loss": 0.6106, "step": 8054 }, { "epoch": 0.6864081806561568, "grad_norm": 13.625, "learning_rate": 1.8508369262229457e-06, "loss": 0.8317, "step": 8055 }, { "epoch": 0.6864933958244568, "grad_norm": 17.75, "learning_rate": 1.8506895701153144e-06, "loss": 0.6371, "step": 8056 }, { "epoch": 0.6865786109927567, "grad_norm": 10.875, "learning_rate": 1.8505422031526186e-06, "loss": 0.4233, "step": 8057 }, { "epoch": 0.6866638261610567, "grad_norm": 11.5625, "learning_rate": 1.8503948253375209e-06, "loss": 0.4715, "step": 8058 }, { "epoch": 0.6867490413293567, "grad_norm": 14.75, "learning_rate": 1.8502474366726856e-06, "loss": 0.7788, "step": 8059 }, { "epoch": 0.6868342564976566, "grad_norm": 15.875, "learning_rate": 1.8501000371607752e-06, "loss": 0.7568, "step": 8060 }, { "epoch": 0.6869194716659566, "grad_norm": 14.3125, "learning_rate": 1.8499526268044534e-06, "loss": 0.7253, "step": 8061 }, { "epoch": 0.6870046868342565, "grad_norm": 16.125, "learning_rate": 1.8498052056063847e-06, "loss": 0.7275, "step": 8062 }, { "epoch": 0.6870899020025565, "grad_norm": 28.625, "learning_rate": 1.849657773569233e-06, "loss": 1.1963, "step": 8063 }, { "epoch": 0.6871751171708564, "grad_norm": 14.5, "learning_rate": 1.8495103306956618e-06, "loss": 0.7997, "step": 8064 }, { "epoch": 0.6872603323391564, "grad_norm": 14.4375, "learning_rate": 1.8493628769883363e-06, "loss": 0.7399, "step": 8065 }, { "epoch": 0.6873455475074564, "grad_norm": 21.125, "learning_rate": 1.8492154124499208e-06, "loss": 0.8936, "step": 8066 }, { "epoch": 0.6874307626757563, "grad_norm": 14.25, "learning_rate": 1.8490679370830804e-06, "loss": 0.7956, "step": 8067 }, { "epoch": 0.6875159778440563, "grad_norm": 12.0, "learning_rate": 1.84892045089048e-06, "loss": 0.4879, "step": 8068 }, { "epoch": 0.6876011930123562, "grad_norm": 15.4375, "learning_rate": 1.8487729538747847e-06, "loss": 0.7588, "step": 8069 }, { "epoch": 0.6876864081806562, "grad_norm": 11.6875, "learning_rate": 1.8486254460386602e-06, "loss": 0.4672, "step": 8070 }, { "epoch": 0.6877716233489561, "grad_norm": 18.125, "learning_rate": 1.8484779273847713e-06, "loss": 1.003, "step": 8071 }, { "epoch": 0.6878568385172561, "grad_norm": 18.625, "learning_rate": 1.8483303979157851e-06, "loss": 0.7874, "step": 8072 }, { "epoch": 0.687942053685556, "grad_norm": 46.0, "learning_rate": 1.8481828576343668e-06, "loss": 1.5661, "step": 8073 }, { "epoch": 0.688027268853856, "grad_norm": 15.75, "learning_rate": 1.8480353065431828e-06, "loss": 0.6623, "step": 8074 }, { "epoch": 0.688112484022156, "grad_norm": 17.375, "learning_rate": 1.847887744644899e-06, "loss": 0.987, "step": 8075 }, { "epoch": 0.6881976991904559, "grad_norm": 16.5, "learning_rate": 1.8477401719421827e-06, "loss": 0.8627, "step": 8076 }, { "epoch": 0.6882829143587559, "grad_norm": 18.0, "learning_rate": 1.8475925884377008e-06, "loss": 0.9467, "step": 8077 }, { "epoch": 0.6883681295270558, "grad_norm": 16.5, "learning_rate": 1.8474449941341197e-06, "loss": 0.7599, "step": 8078 }, { "epoch": 0.6884533446953558, "grad_norm": 14.3125, "learning_rate": 1.8472973890341062e-06, "loss": 0.5263, "step": 8079 }, { "epoch": 0.6885385598636558, "grad_norm": 18.125, "learning_rate": 1.8471497731403287e-06, "loss": 1.1068, "step": 8080 }, { "epoch": 0.6886237750319557, "grad_norm": 11.75, "learning_rate": 1.8470021464554545e-06, "loss": 0.403, "step": 8081 }, { "epoch": 0.6887089902002557, "grad_norm": 12.4375, "learning_rate": 1.846854508982151e-06, "loss": 0.5711, "step": 8082 }, { "epoch": 0.6887942053685556, "grad_norm": 16.25, "learning_rate": 1.8467068607230861e-06, "loss": 0.3717, "step": 8083 }, { "epoch": 0.6888794205368556, "grad_norm": 10.5625, "learning_rate": 1.8465592016809283e-06, "loss": 0.3918, "step": 8084 }, { "epoch": 0.6889646357051555, "grad_norm": 23.0, "learning_rate": 1.8464115318583464e-06, "loss": 0.3721, "step": 8085 }, { "epoch": 0.6890498508734555, "grad_norm": 11.9375, "learning_rate": 1.8462638512580078e-06, "loss": 0.575, "step": 8086 }, { "epoch": 0.6891350660417555, "grad_norm": 16.125, "learning_rate": 1.846116159882582e-06, "loss": 0.8211, "step": 8087 }, { "epoch": 0.6892202812100554, "grad_norm": 13.125, "learning_rate": 1.8459684577347376e-06, "loss": 0.7709, "step": 8088 }, { "epoch": 0.6893054963783554, "grad_norm": 13.25, "learning_rate": 1.8458207448171442e-06, "loss": 0.5794, "step": 8089 }, { "epoch": 0.6893907115466553, "grad_norm": 22.125, "learning_rate": 1.8456730211324703e-06, "loss": 1.0998, "step": 8090 }, { "epoch": 0.6894759267149553, "grad_norm": 15.75, "learning_rate": 1.8455252866833864e-06, "loss": 0.7661, "step": 8091 }, { "epoch": 0.6895611418832552, "grad_norm": 12.5625, "learning_rate": 1.8453775414725613e-06, "loss": 0.6825, "step": 8092 }, { "epoch": 0.6896463570515552, "grad_norm": 30.5, "learning_rate": 1.8452297855026656e-06, "loss": 0.9084, "step": 8093 }, { "epoch": 0.6897315722198551, "grad_norm": 14.0, "learning_rate": 1.8450820187763691e-06, "loss": 0.6193, "step": 8094 }, { "epoch": 0.6898167873881551, "grad_norm": 13.375, "learning_rate": 1.8449342412963422e-06, "loss": 0.6635, "step": 8095 }, { "epoch": 0.6899020025564551, "grad_norm": 13.125, "learning_rate": 1.8447864530652552e-06, "loss": 1.0097, "step": 8096 }, { "epoch": 0.689987217724755, "grad_norm": 16.75, "learning_rate": 1.844638654085779e-06, "loss": 1.0979, "step": 8097 }, { "epoch": 0.690072432893055, "grad_norm": 13.25, "learning_rate": 1.8444908443605842e-06, "loss": 0.6398, "step": 8098 }, { "epoch": 0.6901576480613549, "grad_norm": 11.0625, "learning_rate": 1.8443430238923418e-06, "loss": 0.6056, "step": 8099 }, { "epoch": 0.6902428632296549, "grad_norm": 30.5, "learning_rate": 1.8441951926837237e-06, "loss": 1.3652, "step": 8100 }, { "epoch": 0.6903280783979548, "grad_norm": 16.75, "learning_rate": 1.8440473507374007e-06, "loss": 1.1108, "step": 8101 }, { "epoch": 0.6904132935662548, "grad_norm": 15.0625, "learning_rate": 1.8438994980560451e-06, "loss": 0.9355, "step": 8102 }, { "epoch": 0.6904985087345548, "grad_norm": 15.8125, "learning_rate": 1.8437516346423279e-06, "loss": 0.6826, "step": 8103 }, { "epoch": 0.6905837239028547, "grad_norm": 26.875, "learning_rate": 1.843603760498922e-06, "loss": 0.7131, "step": 8104 }, { "epoch": 0.6906689390711547, "grad_norm": 17.375, "learning_rate": 1.843455875628499e-06, "loss": 0.6287, "step": 8105 }, { "epoch": 0.6907541542394546, "grad_norm": 15.8125, "learning_rate": 1.8433079800337317e-06, "loss": 0.8101, "step": 8106 }, { "epoch": 0.6908393694077546, "grad_norm": 18.125, "learning_rate": 1.8431600737172922e-06, "loss": 1.0646, "step": 8107 }, { "epoch": 0.6909245845760545, "grad_norm": 14.6875, "learning_rate": 1.8430121566818543e-06, "loss": 0.7475, "step": 8108 }, { "epoch": 0.6910097997443545, "grad_norm": 17.625, "learning_rate": 1.8428642289300903e-06, "loss": 0.9519, "step": 8109 }, { "epoch": 0.6910950149126545, "grad_norm": 14.6875, "learning_rate": 1.8427162904646731e-06, "loss": 0.7998, "step": 8110 }, { "epoch": 0.6911802300809544, "grad_norm": 19.625, "learning_rate": 1.8425683412882766e-06, "loss": 1.1156, "step": 8111 }, { "epoch": 0.6912654452492544, "grad_norm": 12.5, "learning_rate": 1.8424203814035746e-06, "loss": 0.569, "step": 8112 }, { "epoch": 0.6913506604175543, "grad_norm": 11.9375, "learning_rate": 1.8422724108132404e-06, "loss": 0.7167, "step": 8113 }, { "epoch": 0.6914358755858543, "grad_norm": 26.25, "learning_rate": 1.8421244295199481e-06, "loss": 0.7344, "step": 8114 }, { "epoch": 0.6915210907541542, "grad_norm": 12.25, "learning_rate": 1.8419764375263718e-06, "loss": 0.6383, "step": 8115 }, { "epoch": 0.6916063059224542, "grad_norm": 14.9375, "learning_rate": 1.8418284348351862e-06, "loss": 0.7088, "step": 8116 }, { "epoch": 0.6916915210907542, "grad_norm": 21.25, "learning_rate": 1.8416804214490658e-06, "loss": 0.537, "step": 8117 }, { "epoch": 0.6917767362590541, "grad_norm": 16.375, "learning_rate": 1.841532397370685e-06, "loss": 0.789, "step": 8118 }, { "epoch": 0.6918619514273541, "grad_norm": 12.5625, "learning_rate": 1.8413843626027189e-06, "loss": 0.4099, "step": 8119 }, { "epoch": 0.691947166595654, "grad_norm": 13.6875, "learning_rate": 1.8412363171478426e-06, "loss": 0.6251, "step": 8120 }, { "epoch": 0.692032381763954, "grad_norm": 14.0625, "learning_rate": 1.8410882610087322e-06, "loss": 0.8444, "step": 8121 }, { "epoch": 0.6921175969322539, "grad_norm": 10.6875, "learning_rate": 1.8409401941880618e-06, "loss": 0.5268, "step": 8122 }, { "epoch": 0.6922028121005539, "grad_norm": 9.8125, "learning_rate": 1.8407921166885082e-06, "loss": 0.3121, "step": 8123 }, { "epoch": 0.6922880272688539, "grad_norm": 16.125, "learning_rate": 1.8406440285127466e-06, "loss": 0.897, "step": 8124 }, { "epoch": 0.6923732424371538, "grad_norm": 27.0, "learning_rate": 1.840495929663454e-06, "loss": 0.8406, "step": 8125 }, { "epoch": 0.6924584576054538, "grad_norm": 14.625, "learning_rate": 1.8403478201433055e-06, "loss": 0.6323, "step": 8126 }, { "epoch": 0.6925436727737537, "grad_norm": 18.375, "learning_rate": 1.840199699954979e-06, "loss": 0.4347, "step": 8127 }, { "epoch": 0.6926288879420537, "grad_norm": 16.625, "learning_rate": 1.8400515691011496e-06, "loss": 0.8205, "step": 8128 }, { "epoch": 0.6927141031103536, "grad_norm": 13.25, "learning_rate": 1.8399034275844957e-06, "loss": 0.6021, "step": 8129 }, { "epoch": 0.6927993182786536, "grad_norm": 14.625, "learning_rate": 1.8397552754076937e-06, "loss": 0.8508, "step": 8130 }, { "epoch": 0.6928845334469536, "grad_norm": 19.625, "learning_rate": 1.8396071125734202e-06, "loss": 0.7914, "step": 8131 }, { "epoch": 0.6929697486152535, "grad_norm": 33.25, "learning_rate": 1.8394589390843536e-06, "loss": 1.1633, "step": 8132 }, { "epoch": 0.6930549637835535, "grad_norm": 16.875, "learning_rate": 1.8393107549431716e-06, "loss": 0.692, "step": 8133 }, { "epoch": 0.6931401789518534, "grad_norm": 12.375, "learning_rate": 1.839162560152551e-06, "loss": 0.4968, "step": 8134 }, { "epoch": 0.6932253941201534, "grad_norm": 16.125, "learning_rate": 1.839014354715171e-06, "loss": 0.813, "step": 8135 }, { "epoch": 0.6933106092884533, "grad_norm": 15.3125, "learning_rate": 1.8388661386337093e-06, "loss": 0.8119, "step": 8136 }, { "epoch": 0.6933958244567533, "grad_norm": 13.875, "learning_rate": 1.8387179119108439e-06, "loss": 0.6631, "step": 8137 }, { "epoch": 0.6934810396250533, "grad_norm": 10.6875, "learning_rate": 1.8385696745492542e-06, "loss": 0.4312, "step": 8138 }, { "epoch": 0.6935662547933532, "grad_norm": 11.3125, "learning_rate": 1.8384214265516184e-06, "loss": 0.3797, "step": 8139 }, { "epoch": 0.6936514699616532, "grad_norm": 9.25, "learning_rate": 1.838273167920616e-06, "loss": 0.3417, "step": 8140 }, { "epoch": 0.6937366851299531, "grad_norm": 17.625, "learning_rate": 1.8381248986589253e-06, "loss": 1.0847, "step": 8141 }, { "epoch": 0.6938219002982531, "grad_norm": 17.125, "learning_rate": 1.8379766187692272e-06, "loss": 0.7426, "step": 8142 }, { "epoch": 0.693907115466553, "grad_norm": 13.75, "learning_rate": 1.8378283282541997e-06, "loss": 0.8764, "step": 8143 }, { "epoch": 0.693992330634853, "grad_norm": 23.25, "learning_rate": 1.8376800271165234e-06, "loss": 0.6341, "step": 8144 }, { "epoch": 0.694077545803153, "grad_norm": 9.625, "learning_rate": 1.8375317153588778e-06, "loss": 0.335, "step": 8145 }, { "epoch": 0.6941627609714529, "grad_norm": 19.125, "learning_rate": 1.8373833929839436e-06, "loss": 0.8884, "step": 8146 }, { "epoch": 0.6942479761397529, "grad_norm": 13.875, "learning_rate": 1.8372350599944008e-06, "loss": 0.4556, "step": 8147 }, { "epoch": 0.6943331913080528, "grad_norm": 30.5, "learning_rate": 1.83708671639293e-06, "loss": 1.385, "step": 8148 }, { "epoch": 0.6944184064763528, "grad_norm": 13.5625, "learning_rate": 1.8369383621822116e-06, "loss": 0.4308, "step": 8149 }, { "epoch": 0.6945036216446527, "grad_norm": 13.9375, "learning_rate": 1.8367899973649267e-06, "loss": 0.8422, "step": 8150 }, { "epoch": 0.6945888368129527, "grad_norm": 18.375, "learning_rate": 1.836641621943757e-06, "loss": 0.8391, "step": 8151 }, { "epoch": 0.6946740519812526, "grad_norm": 12.8125, "learning_rate": 1.8364932359213828e-06, "loss": 0.5521, "step": 8152 }, { "epoch": 0.6947592671495526, "grad_norm": 16.25, "learning_rate": 1.8363448393004862e-06, "loss": 0.9841, "step": 8153 }, { "epoch": 0.6948444823178526, "grad_norm": 14.3125, "learning_rate": 1.836196432083749e-06, "loss": 0.5429, "step": 8154 }, { "epoch": 0.6949296974861525, "grad_norm": 10.875, "learning_rate": 1.8360480142738527e-06, "loss": 0.4257, "step": 8155 }, { "epoch": 0.6950149126544525, "grad_norm": 14.625, "learning_rate": 1.835899585873479e-06, "loss": 0.9718, "step": 8156 }, { "epoch": 0.6951001278227524, "grad_norm": 15.1875, "learning_rate": 1.835751146885311e-06, "loss": 0.8682, "step": 8157 }, { "epoch": 0.6951853429910524, "grad_norm": 14.6875, "learning_rate": 1.835602697312031e-06, "loss": 0.6956, "step": 8158 }, { "epoch": 0.6952705581593523, "grad_norm": 12.625, "learning_rate": 1.8354542371563212e-06, "loss": 0.5683, "step": 8159 }, { "epoch": 0.6953557733276523, "grad_norm": 13.25, "learning_rate": 1.8353057664208646e-06, "loss": 0.7104, "step": 8160 }, { "epoch": 0.6954409884959523, "grad_norm": 15.5625, "learning_rate": 1.8351572851083442e-06, "loss": 0.5315, "step": 8161 }, { "epoch": 0.6955262036642522, "grad_norm": 26.625, "learning_rate": 1.835008793221443e-06, "loss": 0.9861, "step": 8162 }, { "epoch": 0.6956114188325522, "grad_norm": 18.125, "learning_rate": 1.8348602907628448e-06, "loss": 1.0019, "step": 8163 }, { "epoch": 0.6956966340008521, "grad_norm": 17.375, "learning_rate": 1.834711777735233e-06, "loss": 1.0653, "step": 8164 }, { "epoch": 0.6957818491691521, "grad_norm": 16.5, "learning_rate": 1.8345632541412917e-06, "loss": 0.6701, "step": 8165 }, { "epoch": 0.695867064337452, "grad_norm": 9.375, "learning_rate": 1.8344147199837041e-06, "loss": 0.281, "step": 8166 }, { "epoch": 0.695952279505752, "grad_norm": 15.9375, "learning_rate": 1.8342661752651554e-06, "loss": 0.9143, "step": 8167 }, { "epoch": 0.696037494674052, "grad_norm": 13.25, "learning_rate": 1.834117619988329e-06, "loss": 0.7012, "step": 8168 }, { "epoch": 0.6961227098423519, "grad_norm": 16.375, "learning_rate": 1.8339690541559097e-06, "loss": 0.7927, "step": 8169 }, { "epoch": 0.6962079250106519, "grad_norm": 12.5, "learning_rate": 1.833820477770583e-06, "loss": 0.6952, "step": 8170 }, { "epoch": 0.6962931401789518, "grad_norm": 12.1875, "learning_rate": 1.8336718908350326e-06, "loss": 0.5869, "step": 8171 }, { "epoch": 0.6963783553472518, "grad_norm": 9.5, "learning_rate": 1.8335232933519443e-06, "loss": 0.2891, "step": 8172 }, { "epoch": 0.6964635705155517, "grad_norm": 17.125, "learning_rate": 1.833374685324003e-06, "loss": 0.9662, "step": 8173 }, { "epoch": 0.6965487856838517, "grad_norm": 13.9375, "learning_rate": 1.8332260667538953e-06, "loss": 0.6379, "step": 8174 }, { "epoch": 0.6966340008521517, "grad_norm": 16.125, "learning_rate": 1.8330774376443056e-06, "loss": 0.9416, "step": 8175 }, { "epoch": 0.6967192160204516, "grad_norm": 16.0, "learning_rate": 1.8329287979979205e-06, "loss": 0.9773, "step": 8176 }, { "epoch": 0.6968044311887516, "grad_norm": 14.9375, "learning_rate": 1.8327801478174253e-06, "loss": 0.9329, "step": 8177 }, { "epoch": 0.6968896463570515, "grad_norm": 14.4375, "learning_rate": 1.8326314871055075e-06, "loss": 0.9319, "step": 8178 }, { "epoch": 0.6969748615253515, "grad_norm": 13.5625, "learning_rate": 1.8324828158648525e-06, "loss": 0.5919, "step": 8179 }, { "epoch": 0.6970600766936514, "grad_norm": 13.4375, "learning_rate": 1.8323341340981474e-06, "loss": 0.6949, "step": 8180 }, { "epoch": 0.6971452918619514, "grad_norm": 20.125, "learning_rate": 1.8321854418080786e-06, "loss": 1.0211, "step": 8181 }, { "epoch": 0.6972305070302514, "grad_norm": 22.25, "learning_rate": 1.8320367389973337e-06, "loss": 1.0448, "step": 8182 }, { "epoch": 0.6973157221985513, "grad_norm": 14.9375, "learning_rate": 1.8318880256685998e-06, "loss": 0.7677, "step": 8183 }, { "epoch": 0.6974009373668513, "grad_norm": 15.5625, "learning_rate": 1.8317393018245638e-06, "loss": 0.7157, "step": 8184 }, { "epoch": 0.6974861525351512, "grad_norm": 15.75, "learning_rate": 1.8315905674679139e-06, "loss": 0.8377, "step": 8185 }, { "epoch": 0.6975713677034512, "grad_norm": 11.125, "learning_rate": 1.8314418226013375e-06, "loss": 0.3535, "step": 8186 }, { "epoch": 0.6976565828717511, "grad_norm": 13.3125, "learning_rate": 1.8312930672275229e-06, "loss": 0.7206, "step": 8187 }, { "epoch": 0.6977417980400511, "grad_norm": 13.125, "learning_rate": 1.8311443013491576e-06, "loss": 0.5121, "step": 8188 }, { "epoch": 0.697827013208351, "grad_norm": 14.4375, "learning_rate": 1.8309955249689305e-06, "loss": 0.8198, "step": 8189 }, { "epoch": 0.697912228376651, "grad_norm": 23.25, "learning_rate": 1.8308467380895304e-06, "loss": 0.5556, "step": 8190 }, { "epoch": 0.697997443544951, "grad_norm": 15.625, "learning_rate": 1.8306979407136452e-06, "loss": 0.5531, "step": 8191 }, { "epoch": 0.6980826587132509, "grad_norm": 13.9375, "learning_rate": 1.8305491328439646e-06, "loss": 0.6348, "step": 8192 }, { "epoch": 0.6981678738815509, "grad_norm": 19.25, "learning_rate": 1.8304003144831772e-06, "loss": 0.8518, "step": 8193 }, { "epoch": 0.6982530890498508, "grad_norm": 16.375, "learning_rate": 1.8302514856339726e-06, "loss": 0.7462, "step": 8194 }, { "epoch": 0.6983383042181508, "grad_norm": 17.875, "learning_rate": 1.83010264629904e-06, "loss": 1.0599, "step": 8195 }, { "epoch": 0.6984235193864508, "grad_norm": 16.625, "learning_rate": 1.8299537964810693e-06, "loss": 0.9519, "step": 8196 }, { "epoch": 0.6985087345547507, "grad_norm": 13.625, "learning_rate": 1.8298049361827503e-06, "loss": 0.7515, "step": 8197 }, { "epoch": 0.6985939497230507, "grad_norm": 16.125, "learning_rate": 1.8296560654067732e-06, "loss": 0.8538, "step": 8198 }, { "epoch": 0.6986791648913506, "grad_norm": 15.625, "learning_rate": 1.8295071841558278e-06, "loss": 0.8301, "step": 8199 }, { "epoch": 0.6987643800596506, "grad_norm": 14.6875, "learning_rate": 1.8293582924326053e-06, "loss": 0.8123, "step": 8200 }, { "epoch": 0.6988495952279505, "grad_norm": 15.25, "learning_rate": 1.8292093902397954e-06, "loss": 0.6887, "step": 8201 }, { "epoch": 0.6989348103962505, "grad_norm": 12.8125, "learning_rate": 1.82906047758009e-06, "loss": 0.6365, "step": 8202 }, { "epoch": 0.6990200255645505, "grad_norm": 13.4375, "learning_rate": 1.8289115544561787e-06, "loss": 0.7295, "step": 8203 }, { "epoch": 0.6991052407328504, "grad_norm": 14.0, "learning_rate": 1.828762620870754e-06, "loss": 0.7369, "step": 8204 }, { "epoch": 0.6991904559011504, "grad_norm": 14.875, "learning_rate": 1.8286136768265064e-06, "loss": 0.8952, "step": 8205 }, { "epoch": 0.6992756710694503, "grad_norm": 13.625, "learning_rate": 1.828464722326128e-06, "loss": 0.7115, "step": 8206 }, { "epoch": 0.6993608862377503, "grad_norm": 19.375, "learning_rate": 1.8283157573723106e-06, "loss": 0.9904, "step": 8207 }, { "epoch": 0.6994461014060502, "grad_norm": 16.875, "learning_rate": 1.8281667819677457e-06, "loss": 1.1654, "step": 8208 }, { "epoch": 0.6995313165743502, "grad_norm": 13.0625, "learning_rate": 1.8280177961151258e-06, "loss": 0.6762, "step": 8209 }, { "epoch": 0.6996165317426501, "grad_norm": 10.4375, "learning_rate": 1.827868799817143e-06, "loss": 0.3427, "step": 8210 }, { "epoch": 0.6997017469109501, "grad_norm": 12.1875, "learning_rate": 1.8277197930764898e-06, "loss": 0.599, "step": 8211 }, { "epoch": 0.6997869620792501, "grad_norm": 11.0, "learning_rate": 1.8275707758958592e-06, "loss": 0.3797, "step": 8212 }, { "epoch": 0.69987217724755, "grad_norm": 13.8125, "learning_rate": 1.8274217482779435e-06, "loss": 0.6451, "step": 8213 }, { "epoch": 0.69995739241585, "grad_norm": 12.4375, "learning_rate": 1.8272727102254365e-06, "loss": 0.7015, "step": 8214 }, { "epoch": 0.7000426075841499, "grad_norm": 12.0625, "learning_rate": 1.827123661741031e-06, "loss": 0.3925, "step": 8215 }, { "epoch": 0.7001278227524499, "grad_norm": 12.4375, "learning_rate": 1.8269746028274203e-06, "loss": 0.5469, "step": 8216 }, { "epoch": 0.7002130379207498, "grad_norm": 12.625, "learning_rate": 1.826825533487299e-06, "loss": 0.6324, "step": 8217 }, { "epoch": 0.7002982530890498, "grad_norm": 14.8125, "learning_rate": 1.82667645372336e-06, "loss": 0.4119, "step": 8218 }, { "epoch": 0.7003834682573498, "grad_norm": 15.8125, "learning_rate": 1.8265273635382973e-06, "loss": 1.025, "step": 8219 }, { "epoch": 0.7004686834256497, "grad_norm": 14.8125, "learning_rate": 1.8263782629348056e-06, "loss": 0.8116, "step": 8220 }, { "epoch": 0.7005538985939497, "grad_norm": 15.375, "learning_rate": 1.8262291519155793e-06, "loss": 0.7729, "step": 8221 }, { "epoch": 0.7006391137622496, "grad_norm": 14.875, "learning_rate": 1.8260800304833123e-06, "loss": 0.7839, "step": 8222 }, { "epoch": 0.7007243289305496, "grad_norm": 17.0, "learning_rate": 1.8259308986407006e-06, "loss": 0.4612, "step": 8223 }, { "epoch": 0.7008095440988495, "grad_norm": 18.0, "learning_rate": 1.8257817563904378e-06, "loss": 0.6319, "step": 8224 }, { "epoch": 0.7008947592671495, "grad_norm": 15.6875, "learning_rate": 1.8256326037352199e-06, "loss": 0.9335, "step": 8225 }, { "epoch": 0.7009799744354495, "grad_norm": 14.9375, "learning_rate": 1.8254834406777416e-06, "loss": 1.0858, "step": 8226 }, { "epoch": 0.7010651896037494, "grad_norm": 12.375, "learning_rate": 1.8253342672206997e-06, "loss": 0.5482, "step": 8227 }, { "epoch": 0.7011504047720494, "grad_norm": 16.875, "learning_rate": 1.8251850833667882e-06, "loss": 0.8393, "step": 8228 }, { "epoch": 0.7012356199403493, "grad_norm": 14.125, "learning_rate": 1.8250358891187042e-06, "loss": 0.5882, "step": 8229 }, { "epoch": 0.7013208351086493, "grad_norm": 13.875, "learning_rate": 1.8248866844791435e-06, "loss": 0.7213, "step": 8230 }, { "epoch": 0.7014060502769492, "grad_norm": 12.625, "learning_rate": 1.824737469450802e-06, "loss": 0.5338, "step": 8231 }, { "epoch": 0.7014912654452492, "grad_norm": 13.0, "learning_rate": 1.824588244036377e-06, "loss": 0.6128, "step": 8232 }, { "epoch": 0.7015764806135492, "grad_norm": 16.125, "learning_rate": 1.8244390082385644e-06, "loss": 0.6144, "step": 8233 }, { "epoch": 0.7016616957818491, "grad_norm": 10.875, "learning_rate": 1.8242897620600612e-06, "loss": 0.4473, "step": 8234 }, { "epoch": 0.7017469109501491, "grad_norm": 15.625, "learning_rate": 1.8241405055035644e-06, "loss": 1.1708, "step": 8235 }, { "epoch": 0.701832126118449, "grad_norm": 10.6875, "learning_rate": 1.8239912385717718e-06, "loss": 0.3221, "step": 8236 }, { "epoch": 0.701917341286749, "grad_norm": 12.25, "learning_rate": 1.8238419612673796e-06, "loss": 0.516, "step": 8237 }, { "epoch": 0.702002556455049, "grad_norm": 15.75, "learning_rate": 1.8236926735930865e-06, "loss": 0.5556, "step": 8238 }, { "epoch": 0.702087771623349, "grad_norm": 18.25, "learning_rate": 1.8235433755515903e-06, "loss": 0.8166, "step": 8239 }, { "epoch": 0.702172986791649, "grad_norm": 10.9375, "learning_rate": 1.823394067145588e-06, "loss": 0.453, "step": 8240 }, { "epoch": 0.7022582019599489, "grad_norm": 17.25, "learning_rate": 1.8232447483777783e-06, "loss": 1.1542, "step": 8241 }, { "epoch": 0.7023434171282489, "grad_norm": 13.3125, "learning_rate": 1.8230954192508598e-06, "loss": 0.5877, "step": 8242 }, { "epoch": 0.7024286322965488, "grad_norm": 16.875, "learning_rate": 1.8229460797675308e-06, "loss": 0.8381, "step": 8243 }, { "epoch": 0.7025138474648488, "grad_norm": 15.3125, "learning_rate": 1.8227967299304899e-06, "loss": 1.1072, "step": 8244 }, { "epoch": 0.7025990626331488, "grad_norm": 15.4375, "learning_rate": 1.822647369742436e-06, "loss": 1.1704, "step": 8245 }, { "epoch": 0.7026842778014487, "grad_norm": 12.375, "learning_rate": 1.8224979992060685e-06, "loss": 0.469, "step": 8246 }, { "epoch": 0.7027694929697487, "grad_norm": 29.5, "learning_rate": 1.8223486183240865e-06, "loss": 1.0302, "step": 8247 }, { "epoch": 0.7028547081380486, "grad_norm": 20.125, "learning_rate": 1.8221992270991894e-06, "loss": 1.0763, "step": 8248 }, { "epoch": 0.7029399233063486, "grad_norm": 14.5, "learning_rate": 1.8220498255340766e-06, "loss": 0.8155, "step": 8249 }, { "epoch": 0.7030251384746485, "grad_norm": 15.0, "learning_rate": 1.8219004136314482e-06, "loss": 0.5337, "step": 8250 }, { "epoch": 0.7031103536429485, "grad_norm": 15.5, "learning_rate": 1.8217509913940046e-06, "loss": 0.909, "step": 8251 }, { "epoch": 0.7031955688112485, "grad_norm": 12.6875, "learning_rate": 1.8216015588244453e-06, "loss": 0.6935, "step": 8252 }, { "epoch": 0.7032807839795484, "grad_norm": 24.875, "learning_rate": 1.8214521159254715e-06, "loss": 1.37, "step": 8253 }, { "epoch": 0.7033659991478484, "grad_norm": 14.0625, "learning_rate": 1.8213026626997826e-06, "loss": 0.8778, "step": 8254 }, { "epoch": 0.7034512143161483, "grad_norm": 16.25, "learning_rate": 1.8211531991500809e-06, "loss": 0.467, "step": 8255 }, { "epoch": 0.7035364294844483, "grad_norm": 14.25, "learning_rate": 1.821003725279066e-06, "loss": 0.8079, "step": 8256 }, { "epoch": 0.7036216446527482, "grad_norm": 23.0, "learning_rate": 1.8208542410894398e-06, "loss": 0.7194, "step": 8257 }, { "epoch": 0.7037068598210482, "grad_norm": 19.875, "learning_rate": 1.8207047465839035e-06, "loss": 0.723, "step": 8258 }, { "epoch": 0.7037920749893481, "grad_norm": 14.0625, "learning_rate": 1.8205552417651585e-06, "loss": 0.6498, "step": 8259 }, { "epoch": 0.7038772901576481, "grad_norm": 15.5, "learning_rate": 1.8204057266359062e-06, "loss": 0.7109, "step": 8260 }, { "epoch": 0.7039625053259481, "grad_norm": 16.875, "learning_rate": 1.8202562011988492e-06, "loss": 0.6741, "step": 8261 }, { "epoch": 0.704047720494248, "grad_norm": 19.0, "learning_rate": 1.820106665456689e-06, "loss": 0.6911, "step": 8262 }, { "epoch": 0.704132935662548, "grad_norm": 10.9375, "learning_rate": 1.8199571194121285e-06, "loss": 0.3597, "step": 8263 }, { "epoch": 0.7042181508308479, "grad_norm": 11.8125, "learning_rate": 1.819807563067869e-06, "loss": 0.483, "step": 8264 }, { "epoch": 0.7043033659991479, "grad_norm": 24.125, "learning_rate": 1.8196579964266145e-06, "loss": 1.0289, "step": 8265 }, { "epoch": 0.7043885811674478, "grad_norm": 19.875, "learning_rate": 1.8195084194910668e-06, "loss": 0.8748, "step": 8266 }, { "epoch": 0.7044737963357478, "grad_norm": 22.625, "learning_rate": 1.8193588322639294e-06, "loss": 1.3506, "step": 8267 }, { "epoch": 0.7045590115040478, "grad_norm": 17.375, "learning_rate": 1.8192092347479057e-06, "loss": 0.693, "step": 8268 }, { "epoch": 0.7046442266723477, "grad_norm": 15.8125, "learning_rate": 1.819059626945698e-06, "loss": 0.977, "step": 8269 }, { "epoch": 0.7047294418406477, "grad_norm": 15.875, "learning_rate": 1.8189100088600111e-06, "loss": 0.8157, "step": 8270 }, { "epoch": 0.7048146570089476, "grad_norm": 12.6875, "learning_rate": 1.8187603804935482e-06, "loss": 0.6772, "step": 8271 }, { "epoch": 0.7048998721772476, "grad_norm": 21.25, "learning_rate": 1.8186107418490135e-06, "loss": 0.9156, "step": 8272 }, { "epoch": 0.7049850873455475, "grad_norm": 26.625, "learning_rate": 1.8184610929291106e-06, "loss": 1.1045, "step": 8273 }, { "epoch": 0.7050703025138475, "grad_norm": 15.125, "learning_rate": 1.818311433736544e-06, "loss": 0.7433, "step": 8274 }, { "epoch": 0.7051555176821475, "grad_norm": 14.3125, "learning_rate": 1.8181617642740186e-06, "loss": 0.8027, "step": 8275 }, { "epoch": 0.7052407328504474, "grad_norm": 17.125, "learning_rate": 1.818012084544239e-06, "loss": 0.5, "step": 8276 }, { "epoch": 0.7053259480187474, "grad_norm": 16.875, "learning_rate": 1.8178623945499092e-06, "loss": 0.8944, "step": 8277 }, { "epoch": 0.7054111631870473, "grad_norm": 16.625, "learning_rate": 1.8177126942937354e-06, "loss": 0.704, "step": 8278 }, { "epoch": 0.7054963783553473, "grad_norm": 12.5625, "learning_rate": 1.817562983778422e-06, "loss": 0.4785, "step": 8279 }, { "epoch": 0.7055815935236472, "grad_norm": 14.6875, "learning_rate": 1.8174132630066751e-06, "loss": 0.8266, "step": 8280 }, { "epoch": 0.7056668086919472, "grad_norm": 14.4375, "learning_rate": 1.8172635319812e-06, "loss": 0.5889, "step": 8281 }, { "epoch": 0.7057520238602472, "grad_norm": 15.625, "learning_rate": 1.817113790704702e-06, "loss": 0.7331, "step": 8282 }, { "epoch": 0.7058372390285471, "grad_norm": 20.375, "learning_rate": 1.8169640391798876e-06, "loss": 0.5778, "step": 8283 }, { "epoch": 0.7059224541968471, "grad_norm": 17.375, "learning_rate": 1.8168142774094633e-06, "loss": 0.906, "step": 8284 }, { "epoch": 0.706007669365147, "grad_norm": 14.9375, "learning_rate": 1.8166645053961347e-06, "loss": 0.6478, "step": 8285 }, { "epoch": 0.706092884533447, "grad_norm": 12.25, "learning_rate": 1.8165147231426088e-06, "loss": 0.5336, "step": 8286 }, { "epoch": 0.7061780997017469, "grad_norm": 9.1875, "learning_rate": 1.8163649306515919e-06, "loss": 0.3632, "step": 8287 }, { "epoch": 0.7062633148700469, "grad_norm": 15.3125, "learning_rate": 1.8162151279257918e-06, "loss": 0.9746, "step": 8288 }, { "epoch": 0.7063485300383469, "grad_norm": 19.625, "learning_rate": 1.8160653149679145e-06, "loss": 1.0229, "step": 8289 }, { "epoch": 0.7064337452066468, "grad_norm": 25.0, "learning_rate": 1.8159154917806675e-06, "loss": 0.9772, "step": 8290 }, { "epoch": 0.7065189603749468, "grad_norm": 14.25, "learning_rate": 1.8157656583667587e-06, "loss": 0.7552, "step": 8291 }, { "epoch": 0.7066041755432467, "grad_norm": 14.875, "learning_rate": 1.815615814728896e-06, "loss": 0.8258, "step": 8292 }, { "epoch": 0.7066893907115467, "grad_norm": 23.375, "learning_rate": 1.8154659608697864e-06, "loss": 0.8631, "step": 8293 }, { "epoch": 0.7067746058798466, "grad_norm": 13.25, "learning_rate": 1.815316096792138e-06, "loss": 0.7681, "step": 8294 }, { "epoch": 0.7068598210481466, "grad_norm": 14.3125, "learning_rate": 1.8151662224986596e-06, "loss": 0.9717, "step": 8295 }, { "epoch": 0.7069450362164466, "grad_norm": 15.6875, "learning_rate": 1.8150163379920592e-06, "loss": 0.8529, "step": 8296 }, { "epoch": 0.7070302513847465, "grad_norm": 13.75, "learning_rate": 1.8148664432750455e-06, "loss": 0.393, "step": 8297 }, { "epoch": 0.7071154665530465, "grad_norm": 13.5, "learning_rate": 1.8147165383503267e-06, "loss": 0.6728, "step": 8298 }, { "epoch": 0.7072006817213464, "grad_norm": 10.8125, "learning_rate": 1.8145666232206128e-06, "loss": 0.7041, "step": 8299 }, { "epoch": 0.7072858968896464, "grad_norm": 13.25, "learning_rate": 1.814416697888612e-06, "loss": 0.5002, "step": 8300 }, { "epoch": 0.7073711120579463, "grad_norm": 12.6875, "learning_rate": 1.814266762357034e-06, "loss": 0.4458, "step": 8301 }, { "epoch": 0.7074563272262463, "grad_norm": 20.375, "learning_rate": 1.8141168166285877e-06, "loss": 0.8011, "step": 8302 }, { "epoch": 0.7075415423945463, "grad_norm": 11.5625, "learning_rate": 1.8139668607059841e-06, "loss": 0.3866, "step": 8303 }, { "epoch": 0.7076267575628462, "grad_norm": 18.75, "learning_rate": 1.8138168945919316e-06, "loss": 0.8114, "step": 8304 }, { "epoch": 0.7077119727311462, "grad_norm": 22.0, "learning_rate": 1.8136669182891408e-06, "loss": 1.0237, "step": 8305 }, { "epoch": 0.7077971878994461, "grad_norm": 13.5625, "learning_rate": 1.8135169318003222e-06, "loss": 0.4133, "step": 8306 }, { "epoch": 0.7078824030677461, "grad_norm": 13.0, "learning_rate": 1.813366935128186e-06, "loss": 0.642, "step": 8307 }, { "epoch": 0.707967618236046, "grad_norm": 11.25, "learning_rate": 1.813216928275443e-06, "loss": 0.2526, "step": 8308 }, { "epoch": 0.708052833404346, "grad_norm": 14.25, "learning_rate": 1.8130669112448035e-06, "loss": 0.8389, "step": 8309 }, { "epoch": 0.708138048572646, "grad_norm": 14.25, "learning_rate": 1.8129168840389789e-06, "loss": 0.9668, "step": 8310 }, { "epoch": 0.7082232637409459, "grad_norm": 16.375, "learning_rate": 1.8127668466606796e-06, "loss": 0.8989, "step": 8311 }, { "epoch": 0.7083084789092459, "grad_norm": 12.875, "learning_rate": 1.8126167991126182e-06, "loss": 0.7233, "step": 8312 }, { "epoch": 0.7083936940775458, "grad_norm": 10.1875, "learning_rate": 1.8124667413975047e-06, "loss": 0.5178, "step": 8313 }, { "epoch": 0.7084789092458458, "grad_norm": 15.5625, "learning_rate": 1.8123166735180521e-06, "loss": 0.7463, "step": 8314 }, { "epoch": 0.7085641244141457, "grad_norm": 15.0, "learning_rate": 1.8121665954769715e-06, "loss": 0.8885, "step": 8315 }, { "epoch": 0.7086493395824457, "grad_norm": 22.125, "learning_rate": 1.8120165072769756e-06, "loss": 1.2882, "step": 8316 }, { "epoch": 0.7087345547507456, "grad_norm": 10.25, "learning_rate": 1.811866408920776e-06, "loss": 0.6725, "step": 8317 }, { "epoch": 0.7088197699190456, "grad_norm": 16.875, "learning_rate": 1.8117163004110854e-06, "loss": 1.0656, "step": 8318 }, { "epoch": 0.7089049850873456, "grad_norm": 20.0, "learning_rate": 1.8115661817506164e-06, "loss": 1.1641, "step": 8319 }, { "epoch": 0.7089902002556455, "grad_norm": 24.375, "learning_rate": 1.8114160529420818e-06, "loss": 0.7925, "step": 8320 }, { "epoch": 0.7090754154239455, "grad_norm": 18.25, "learning_rate": 1.8112659139881946e-06, "loss": 1.0362, "step": 8321 }, { "epoch": 0.7091606305922454, "grad_norm": 15.75, "learning_rate": 1.8111157648916677e-06, "loss": 0.7768, "step": 8322 }, { "epoch": 0.7092458457605454, "grad_norm": 17.75, "learning_rate": 1.8109656056552153e-06, "loss": 1.0236, "step": 8323 }, { "epoch": 0.7093310609288453, "grad_norm": 10.625, "learning_rate": 1.81081543628155e-06, "loss": 0.5315, "step": 8324 }, { "epoch": 0.7094162760971453, "grad_norm": 17.0, "learning_rate": 1.8106652567733856e-06, "loss": 0.7208, "step": 8325 }, { "epoch": 0.7095014912654453, "grad_norm": 11.6875, "learning_rate": 1.8105150671334363e-06, "loss": 0.4416, "step": 8326 }, { "epoch": 0.7095867064337452, "grad_norm": 13.5625, "learning_rate": 1.8103648673644164e-06, "loss": 0.5388, "step": 8327 }, { "epoch": 0.7096719216020452, "grad_norm": 24.0, "learning_rate": 1.8102146574690396e-06, "loss": 1.357, "step": 8328 }, { "epoch": 0.7097571367703451, "grad_norm": 13.4375, "learning_rate": 1.8100644374500207e-06, "loss": 0.8613, "step": 8329 }, { "epoch": 0.7098423519386451, "grad_norm": 17.875, "learning_rate": 1.809914207310074e-06, "loss": 0.7143, "step": 8330 }, { "epoch": 0.709927567106945, "grad_norm": 12.6875, "learning_rate": 1.8097639670519145e-06, "loss": 0.7625, "step": 8331 }, { "epoch": 0.710012782275245, "grad_norm": 15.0, "learning_rate": 1.8096137166782576e-06, "loss": 0.7302, "step": 8332 }, { "epoch": 0.710097997443545, "grad_norm": 14.9375, "learning_rate": 1.8094634561918182e-06, "loss": 0.981, "step": 8333 }, { "epoch": 0.7101832126118449, "grad_norm": 9.5625, "learning_rate": 1.8093131855953111e-06, "loss": 0.3318, "step": 8334 }, { "epoch": 0.7102684277801449, "grad_norm": 19.5, "learning_rate": 1.8091629048914525e-06, "loss": 0.983, "step": 8335 }, { "epoch": 0.7103536429484448, "grad_norm": 17.0, "learning_rate": 1.8090126140829577e-06, "loss": 0.9012, "step": 8336 }, { "epoch": 0.7104388581167448, "grad_norm": 15.0625, "learning_rate": 1.808862313172543e-06, "loss": 0.6816, "step": 8337 }, { "epoch": 0.7105240732850447, "grad_norm": 18.875, "learning_rate": 1.8087120021629245e-06, "loss": 0.9793, "step": 8338 }, { "epoch": 0.7106092884533447, "grad_norm": 15.3125, "learning_rate": 1.8085616810568179e-06, "loss": 0.8137, "step": 8339 }, { "epoch": 0.7106945036216447, "grad_norm": 13.125, "learning_rate": 1.8084113498569406e-06, "loss": 0.4541, "step": 8340 }, { "epoch": 0.7107797187899446, "grad_norm": 17.25, "learning_rate": 1.8082610085660085e-06, "loss": 0.8404, "step": 8341 }, { "epoch": 0.7108649339582446, "grad_norm": 13.25, "learning_rate": 1.8081106571867385e-06, "loss": 0.8269, "step": 8342 }, { "epoch": 0.7109501491265445, "grad_norm": 12.8125, "learning_rate": 1.8079602957218474e-06, "loss": 0.729, "step": 8343 }, { "epoch": 0.7110353642948445, "grad_norm": 15.4375, "learning_rate": 1.8078099241740533e-06, "loss": 0.7775, "step": 8344 }, { "epoch": 0.7111205794631444, "grad_norm": 13.5625, "learning_rate": 1.8076595425460725e-06, "loss": 0.7163, "step": 8345 }, { "epoch": 0.7112057946314444, "grad_norm": 14.5, "learning_rate": 1.807509150840623e-06, "loss": 0.7853, "step": 8346 }, { "epoch": 0.7112910097997444, "grad_norm": 14.6875, "learning_rate": 1.8073587490604227e-06, "loss": 0.5224, "step": 8347 }, { "epoch": 0.7113762249680443, "grad_norm": 14.25, "learning_rate": 1.8072083372081892e-06, "loss": 0.9063, "step": 8348 }, { "epoch": 0.7114614401363443, "grad_norm": 12.4375, "learning_rate": 1.807057915286641e-06, "loss": 0.581, "step": 8349 }, { "epoch": 0.7115466553046442, "grad_norm": 12.75, "learning_rate": 1.8069074832984958e-06, "loss": 0.6699, "step": 8350 }, { "epoch": 0.7116318704729442, "grad_norm": 14.0625, "learning_rate": 1.8067570412464725e-06, "loss": 0.5923, "step": 8351 }, { "epoch": 0.7117170856412441, "grad_norm": 11.0625, "learning_rate": 1.8066065891332896e-06, "loss": 0.3533, "step": 8352 }, { "epoch": 0.7118023008095441, "grad_norm": 9.8125, "learning_rate": 1.806456126961666e-06, "loss": 0.2979, "step": 8353 }, { "epoch": 0.711887515977844, "grad_norm": 12.9375, "learning_rate": 1.8063056547343203e-06, "loss": 0.6823, "step": 8354 }, { "epoch": 0.711972731146144, "grad_norm": 12.375, "learning_rate": 1.8061551724539722e-06, "loss": 0.5469, "step": 8355 }, { "epoch": 0.712057946314444, "grad_norm": 10.5, "learning_rate": 1.8060046801233411e-06, "loss": 0.4253, "step": 8356 }, { "epoch": 0.7121431614827439, "grad_norm": 14.0625, "learning_rate": 1.8058541777451464e-06, "loss": 0.659, "step": 8357 }, { "epoch": 0.7122283766510439, "grad_norm": 14.375, "learning_rate": 1.8057036653221071e-06, "loss": 0.667, "step": 8358 }, { "epoch": 0.7123135918193438, "grad_norm": 12.5625, "learning_rate": 1.8055531428569443e-06, "loss": 0.4437, "step": 8359 }, { "epoch": 0.7123988069876438, "grad_norm": 13.1875, "learning_rate": 1.8054026103523772e-06, "loss": 0.3802, "step": 8360 }, { "epoch": 0.7124840221559438, "grad_norm": 15.875, "learning_rate": 1.8052520678111272e-06, "loss": 0.4102, "step": 8361 }, { "epoch": 0.7125692373242437, "grad_norm": 15.5, "learning_rate": 1.805101515235913e-06, "loss": 0.8885, "step": 8362 }, { "epoch": 0.7126544524925437, "grad_norm": 14.5625, "learning_rate": 1.8049509526294567e-06, "loss": 0.8602, "step": 8363 }, { "epoch": 0.7127396676608436, "grad_norm": 15.3125, "learning_rate": 1.8048003799944786e-06, "loss": 0.5404, "step": 8364 }, { "epoch": 0.7128248828291436, "grad_norm": 15.9375, "learning_rate": 1.8046497973336998e-06, "loss": 0.834, "step": 8365 }, { "epoch": 0.7129100979974435, "grad_norm": 12.375, "learning_rate": 1.8044992046498413e-06, "loss": 0.5426, "step": 8366 }, { "epoch": 0.7129953131657435, "grad_norm": 15.125, "learning_rate": 1.8043486019456247e-06, "loss": 0.6189, "step": 8367 }, { "epoch": 0.7130805283340435, "grad_norm": 15.0625, "learning_rate": 1.8041979892237713e-06, "loss": 0.6886, "step": 8368 }, { "epoch": 0.7131657435023434, "grad_norm": 12.1875, "learning_rate": 1.804047366487003e-06, "loss": 0.5865, "step": 8369 }, { "epoch": 0.7132509586706434, "grad_norm": 24.25, "learning_rate": 1.8038967337380417e-06, "loss": 1.3272, "step": 8370 }, { "epoch": 0.7133361738389433, "grad_norm": 15.3125, "learning_rate": 1.8037460909796096e-06, "loss": 0.9567, "step": 8371 }, { "epoch": 0.7134213890072433, "grad_norm": 17.125, "learning_rate": 1.8035954382144282e-06, "loss": 0.8599, "step": 8372 }, { "epoch": 0.7135066041755432, "grad_norm": 12.0, "learning_rate": 1.8034447754452211e-06, "loss": 0.6604, "step": 8373 }, { "epoch": 0.7135918193438432, "grad_norm": 30.375, "learning_rate": 1.8032941026747102e-06, "loss": 0.9519, "step": 8374 }, { "epoch": 0.7136770345121431, "grad_norm": 17.625, "learning_rate": 1.8031434199056184e-06, "loss": 0.9951, "step": 8375 }, { "epoch": 0.7137622496804431, "grad_norm": 23.125, "learning_rate": 1.8029927271406688e-06, "loss": 0.8564, "step": 8376 }, { "epoch": 0.7138474648487431, "grad_norm": 15.5, "learning_rate": 1.8028420243825848e-06, "loss": 0.5792, "step": 8377 }, { "epoch": 0.713932680017043, "grad_norm": 15.5625, "learning_rate": 1.8026913116340893e-06, "loss": 0.8336, "step": 8378 }, { "epoch": 0.714017895185343, "grad_norm": 11.6875, "learning_rate": 1.802540588897906e-06, "loss": 0.6323, "step": 8379 }, { "epoch": 0.7141031103536429, "grad_norm": 13.5625, "learning_rate": 1.8023898561767586e-06, "loss": 0.7817, "step": 8380 }, { "epoch": 0.7141883255219429, "grad_norm": 20.625, "learning_rate": 1.8022391134733713e-06, "loss": 0.9023, "step": 8381 }, { "epoch": 0.7142735406902428, "grad_norm": 10.5625, "learning_rate": 1.8020883607904677e-06, "loss": 0.3712, "step": 8382 }, { "epoch": 0.7143587558585428, "grad_norm": 12.5, "learning_rate": 1.8019375981307722e-06, "loss": 0.6155, "step": 8383 }, { "epoch": 0.7144439710268428, "grad_norm": 10.125, "learning_rate": 1.8017868254970095e-06, "loss": 0.4282, "step": 8384 }, { "epoch": 0.7145291861951427, "grad_norm": 14.0, "learning_rate": 1.8016360428919038e-06, "loss": 0.5352, "step": 8385 }, { "epoch": 0.7146144013634427, "grad_norm": 15.1875, "learning_rate": 1.8014852503181802e-06, "loss": 0.9668, "step": 8386 }, { "epoch": 0.7146996165317426, "grad_norm": 17.125, "learning_rate": 1.8013344477785632e-06, "loss": 0.8365, "step": 8387 }, { "epoch": 0.7147848317000426, "grad_norm": 12.8125, "learning_rate": 1.8011836352757785e-06, "loss": 0.5402, "step": 8388 }, { "epoch": 0.7148700468683425, "grad_norm": 13.6875, "learning_rate": 1.8010328128125515e-06, "loss": 0.7868, "step": 8389 }, { "epoch": 0.7149552620366425, "grad_norm": 14.5, "learning_rate": 1.8008819803916073e-06, "loss": 0.5965, "step": 8390 }, { "epoch": 0.7150404772049425, "grad_norm": 14.6875, "learning_rate": 1.8007311380156716e-06, "loss": 0.4634, "step": 8391 }, { "epoch": 0.7151256923732424, "grad_norm": 17.125, "learning_rate": 1.8005802856874707e-06, "loss": 0.8445, "step": 8392 }, { "epoch": 0.7152109075415424, "grad_norm": 15.6875, "learning_rate": 1.8004294234097304e-06, "loss": 0.8743, "step": 8393 }, { "epoch": 0.7152961227098423, "grad_norm": 14.625, "learning_rate": 1.8002785511851766e-06, "loss": 0.8426, "step": 8394 }, { "epoch": 0.7153813378781423, "grad_norm": 21.625, "learning_rate": 1.8001276690165364e-06, "loss": 0.9245, "step": 8395 }, { "epoch": 0.7154665530464422, "grad_norm": 15.25, "learning_rate": 1.7999767769065357e-06, "loss": 0.384, "step": 8396 }, { "epoch": 0.7155517682147422, "grad_norm": 13.5625, "learning_rate": 1.7998258748579022e-06, "loss": 0.5564, "step": 8397 }, { "epoch": 0.7156369833830422, "grad_norm": 15.875, "learning_rate": 1.7996749628733618e-06, "loss": 0.7272, "step": 8398 }, { "epoch": 0.7157221985513421, "grad_norm": 17.375, "learning_rate": 1.7995240409556421e-06, "loss": 1.0192, "step": 8399 }, { "epoch": 0.7158074137196421, "grad_norm": 13.8125, "learning_rate": 1.7993731091074701e-06, "loss": 0.9317, "step": 8400 }, { "epoch": 0.715892628887942, "grad_norm": 21.5, "learning_rate": 1.7992221673315743e-06, "loss": 0.7983, "step": 8401 }, { "epoch": 0.715977844056242, "grad_norm": 16.5, "learning_rate": 1.7990712156306812e-06, "loss": 1.0339, "step": 8402 }, { "epoch": 0.7160630592245419, "grad_norm": 14.75, "learning_rate": 1.798920254007519e-06, "loss": 0.8863, "step": 8403 }, { "epoch": 0.7161482743928419, "grad_norm": 8.1875, "learning_rate": 1.7987692824648159e-06, "loss": 0.4364, "step": 8404 }, { "epoch": 0.7162334895611419, "grad_norm": 14.5625, "learning_rate": 1.7986183010053001e-06, "loss": 0.6309, "step": 8405 }, { "epoch": 0.7163187047294418, "grad_norm": 11.8125, "learning_rate": 1.7984673096317001e-06, "loss": 0.4635, "step": 8406 }, { "epoch": 0.7164039198977418, "grad_norm": 11.125, "learning_rate": 1.798316308346744e-06, "loss": 0.4399, "step": 8407 }, { "epoch": 0.7164891350660417, "grad_norm": 16.375, "learning_rate": 1.798165297153161e-06, "loss": 0.9663, "step": 8408 }, { "epoch": 0.7165743502343417, "grad_norm": 15.25, "learning_rate": 1.7980142760536795e-06, "loss": 0.8537, "step": 8409 }, { "epoch": 0.7166595654026416, "grad_norm": 18.375, "learning_rate": 1.7978632450510297e-06, "loss": 0.8599, "step": 8410 }, { "epoch": 0.7167447805709416, "grad_norm": 13.375, "learning_rate": 1.7977122041479394e-06, "loss": 0.7619, "step": 8411 }, { "epoch": 0.7168299957392416, "grad_norm": 13.625, "learning_rate": 1.7975611533471393e-06, "loss": 0.682, "step": 8412 }, { "epoch": 0.7169152109075415, "grad_norm": 10.375, "learning_rate": 1.7974100926513579e-06, "loss": 0.7359, "step": 8413 }, { "epoch": 0.7170004260758415, "grad_norm": 13.5625, "learning_rate": 1.7972590220633264e-06, "loss": 0.6809, "step": 8414 }, { "epoch": 0.7170856412441414, "grad_norm": 27.625, "learning_rate": 1.7971079415857737e-06, "loss": 1.25, "step": 8415 }, { "epoch": 0.7171708564124414, "grad_norm": 15.375, "learning_rate": 1.7969568512214305e-06, "loss": 0.7435, "step": 8416 }, { "epoch": 0.7172560715807413, "grad_norm": 20.875, "learning_rate": 1.796805750973027e-06, "loss": 1.1183, "step": 8417 }, { "epoch": 0.7173412867490413, "grad_norm": 10.8125, "learning_rate": 1.7966546408432936e-06, "loss": 0.4151, "step": 8418 }, { "epoch": 0.7174265019173413, "grad_norm": 23.5, "learning_rate": 1.796503520834961e-06, "loss": 0.8503, "step": 8419 }, { "epoch": 0.7175117170856412, "grad_norm": 17.125, "learning_rate": 1.7963523909507606e-06, "loss": 0.9599, "step": 8420 }, { "epoch": 0.7175969322539412, "grad_norm": 14.5625, "learning_rate": 1.7962012511934226e-06, "loss": 0.6236, "step": 8421 }, { "epoch": 0.7176821474222411, "grad_norm": 18.125, "learning_rate": 1.7960501015656793e-06, "loss": 0.8873, "step": 8422 }, { "epoch": 0.7177673625905411, "grad_norm": 11.875, "learning_rate": 1.795898942070261e-06, "loss": 0.3641, "step": 8423 }, { "epoch": 0.717852577758841, "grad_norm": 11.6875, "learning_rate": 1.7957477727099e-06, "loss": 0.7945, "step": 8424 }, { "epoch": 0.717937792927141, "grad_norm": 14.8125, "learning_rate": 1.7955965934873284e-06, "loss": 0.6206, "step": 8425 }, { "epoch": 0.718023008095441, "grad_norm": 17.625, "learning_rate": 1.7954454044052775e-06, "loss": 0.9856, "step": 8426 }, { "epoch": 0.7181082232637409, "grad_norm": 15.9375, "learning_rate": 1.7952942054664795e-06, "loss": 1.1367, "step": 8427 }, { "epoch": 0.7181934384320409, "grad_norm": 13.625, "learning_rate": 1.795142996673667e-06, "loss": 0.6583, "step": 8428 }, { "epoch": 0.7182786536003408, "grad_norm": 14.4375, "learning_rate": 1.7949917780295722e-06, "loss": 0.8533, "step": 8429 }, { "epoch": 0.7183638687686408, "grad_norm": 19.5, "learning_rate": 1.794840549536928e-06, "loss": 1.093, "step": 8430 }, { "epoch": 0.7184490839369407, "grad_norm": 21.625, "learning_rate": 1.7946893111984675e-06, "loss": 0.7753, "step": 8431 }, { "epoch": 0.7185342991052407, "grad_norm": 19.5, "learning_rate": 1.7945380630169231e-06, "loss": 1.0423, "step": 8432 }, { "epoch": 0.7186195142735406, "grad_norm": 15.5, "learning_rate": 1.7943868049950286e-06, "loss": 0.76, "step": 8433 }, { "epoch": 0.7187047294418406, "grad_norm": 21.5, "learning_rate": 1.794235537135517e-06, "loss": 0.9214, "step": 8434 }, { "epoch": 0.7187899446101406, "grad_norm": 10.375, "learning_rate": 1.7940842594411217e-06, "loss": 0.4538, "step": 8435 }, { "epoch": 0.7188751597784405, "grad_norm": 14.4375, "learning_rate": 1.7939329719145768e-06, "loss": 0.7646, "step": 8436 }, { "epoch": 0.7189603749467405, "grad_norm": 18.25, "learning_rate": 1.7937816745586162e-06, "loss": 0.8822, "step": 8437 }, { "epoch": 0.7190455901150404, "grad_norm": 15.8125, "learning_rate": 1.7936303673759742e-06, "loss": 0.8798, "step": 8438 }, { "epoch": 0.7191308052833404, "grad_norm": 12.0625, "learning_rate": 1.7934790503693844e-06, "loss": 0.4781, "step": 8439 }, { "epoch": 0.7192160204516403, "grad_norm": 15.125, "learning_rate": 1.793327723541582e-06, "loss": 0.8587, "step": 8440 }, { "epoch": 0.7193012356199403, "grad_norm": 13.1875, "learning_rate": 1.7931763868953008e-06, "loss": 0.695, "step": 8441 }, { "epoch": 0.7193864507882403, "grad_norm": 16.125, "learning_rate": 1.7930250404332766e-06, "loss": 0.6579, "step": 8442 }, { "epoch": 0.7194716659565402, "grad_norm": 13.0625, "learning_rate": 1.7928736841582434e-06, "loss": 0.5921, "step": 8443 }, { "epoch": 0.7195568811248402, "grad_norm": 12.125, "learning_rate": 1.7927223180729372e-06, "loss": 0.5393, "step": 8444 }, { "epoch": 0.7196420962931401, "grad_norm": 15.1875, "learning_rate": 1.7925709421800928e-06, "loss": 0.7309, "step": 8445 }, { "epoch": 0.7197273114614401, "grad_norm": 13.5, "learning_rate": 1.792419556482446e-06, "loss": 0.3641, "step": 8446 }, { "epoch": 0.71981252662974, "grad_norm": 14.75, "learning_rate": 1.7922681609827322e-06, "loss": 0.671, "step": 8447 }, { "epoch": 0.71989774179804, "grad_norm": 15.4375, "learning_rate": 1.7921167556836874e-06, "loss": 0.8593, "step": 8448 }, { "epoch": 0.71998295696634, "grad_norm": 19.875, "learning_rate": 1.7919653405880478e-06, "loss": 0.7353, "step": 8449 }, { "epoch": 0.7200681721346399, "grad_norm": 17.0, "learning_rate": 1.7918139156985499e-06, "loss": 0.6623, "step": 8450 }, { "epoch": 0.7201533873029399, "grad_norm": 16.5, "learning_rate": 1.791662481017929e-06, "loss": 1.1, "step": 8451 }, { "epoch": 0.7202386024712398, "grad_norm": 15.0625, "learning_rate": 1.7915110365489228e-06, "loss": 0.324, "step": 8452 }, { "epoch": 0.7203238176395398, "grad_norm": 12.8125, "learning_rate": 1.7913595822942673e-06, "loss": 0.5537, "step": 8453 }, { "epoch": 0.7204090328078397, "grad_norm": 16.0, "learning_rate": 1.7912081182567e-06, "loss": 0.9242, "step": 8454 }, { "epoch": 0.7204942479761397, "grad_norm": 13.4375, "learning_rate": 1.7910566444389578e-06, "loss": 0.7414, "step": 8455 }, { "epoch": 0.7205794631444397, "grad_norm": 22.375, "learning_rate": 1.790905160843778e-06, "loss": 1.2834, "step": 8456 }, { "epoch": 0.7206646783127396, "grad_norm": 16.375, "learning_rate": 1.7907536674738978e-06, "loss": 0.9382, "step": 8457 }, { "epoch": 0.7207498934810396, "grad_norm": 23.5, "learning_rate": 1.7906021643320554e-06, "loss": 0.6758, "step": 8458 }, { "epoch": 0.7208351086493395, "grad_norm": 14.6875, "learning_rate": 1.790450651420988e-06, "loss": 0.8636, "step": 8459 }, { "epoch": 0.7209203238176395, "grad_norm": 15.125, "learning_rate": 1.7902991287434337e-06, "loss": 1.0673, "step": 8460 }, { "epoch": 0.7210055389859394, "grad_norm": 15.375, "learning_rate": 1.7901475963021315e-06, "loss": 1.0023, "step": 8461 }, { "epoch": 0.7210907541542394, "grad_norm": 16.25, "learning_rate": 1.7899960540998185e-06, "loss": 0.9036, "step": 8462 }, { "epoch": 0.7211759693225395, "grad_norm": 12.875, "learning_rate": 1.789844502139234e-06, "loss": 0.8175, "step": 8463 }, { "epoch": 0.7212611844908394, "grad_norm": 12.8125, "learning_rate": 1.789692940423116e-06, "loss": 0.325, "step": 8464 }, { "epoch": 0.7213463996591394, "grad_norm": 20.875, "learning_rate": 1.7895413689542045e-06, "loss": 0.7589, "step": 8465 }, { "epoch": 0.7214316148274393, "grad_norm": 15.5, "learning_rate": 1.7893897877352378e-06, "loss": 0.7837, "step": 8466 }, { "epoch": 0.7215168299957393, "grad_norm": 10.5625, "learning_rate": 1.7892381967689554e-06, "loss": 0.3613, "step": 8467 }, { "epoch": 0.7216020451640393, "grad_norm": 11.625, "learning_rate": 1.7890865960580961e-06, "loss": 0.5492, "step": 8468 }, { "epoch": 0.7216872603323392, "grad_norm": 18.0, "learning_rate": 1.7889349856054e-06, "loss": 0.7181, "step": 8469 }, { "epoch": 0.7217724755006392, "grad_norm": 13.5, "learning_rate": 1.7887833654136073e-06, "loss": 0.4474, "step": 8470 }, { "epoch": 0.7218576906689391, "grad_norm": 14.4375, "learning_rate": 1.7886317354854568e-06, "loss": 0.5507, "step": 8471 }, { "epoch": 0.7219429058372391, "grad_norm": 14.75, "learning_rate": 1.7884800958236892e-06, "loss": 0.9146, "step": 8472 }, { "epoch": 0.722028121005539, "grad_norm": 18.0, "learning_rate": 1.7883284464310453e-06, "loss": 0.7204, "step": 8473 }, { "epoch": 0.722113336173839, "grad_norm": 18.0, "learning_rate": 1.7881767873102646e-06, "loss": 0.8883, "step": 8474 }, { "epoch": 0.722198551342139, "grad_norm": 22.0, "learning_rate": 1.7880251184640878e-06, "loss": 0.965, "step": 8475 }, { "epoch": 0.7222837665104389, "grad_norm": 18.25, "learning_rate": 1.7878734398952568e-06, "loss": 0.8425, "step": 8476 }, { "epoch": 0.7223689816787389, "grad_norm": 16.875, "learning_rate": 1.7877217516065115e-06, "loss": 0.7847, "step": 8477 }, { "epoch": 0.7224541968470388, "grad_norm": 24.25, "learning_rate": 1.7875700536005935e-06, "loss": 1.0565, "step": 8478 }, { "epoch": 0.7225394120153388, "grad_norm": 22.25, "learning_rate": 1.7874183458802445e-06, "loss": 1.0721, "step": 8479 }, { "epoch": 0.7226246271836387, "grad_norm": 12.3125, "learning_rate": 1.7872666284482052e-06, "loss": 0.7625, "step": 8480 }, { "epoch": 0.7227098423519387, "grad_norm": 9.875, "learning_rate": 1.7871149013072175e-06, "loss": 0.2838, "step": 8481 }, { "epoch": 0.7227950575202386, "grad_norm": 14.375, "learning_rate": 1.7869631644600236e-06, "loss": 0.5975, "step": 8482 }, { "epoch": 0.7228802726885386, "grad_norm": 10.5, "learning_rate": 1.7868114179093655e-06, "loss": 0.4271, "step": 8483 }, { "epoch": 0.7229654878568386, "grad_norm": 17.0, "learning_rate": 1.786659661657985e-06, "loss": 0.7355, "step": 8484 }, { "epoch": 0.7230507030251385, "grad_norm": 11.5625, "learning_rate": 1.7865078957086245e-06, "loss": 0.4958, "step": 8485 }, { "epoch": 0.7231359181934385, "grad_norm": 16.875, "learning_rate": 1.7863561200640275e-06, "loss": 1.3118, "step": 8486 }, { "epoch": 0.7232211333617384, "grad_norm": 12.8125, "learning_rate": 1.786204334726936e-06, "loss": 0.6518, "step": 8487 }, { "epoch": 0.7233063485300384, "grad_norm": 13.4375, "learning_rate": 1.7860525397000927e-06, "loss": 0.7232, "step": 8488 }, { "epoch": 0.7233915636983383, "grad_norm": 16.375, "learning_rate": 1.7859007349862405e-06, "loss": 0.8012, "step": 8489 }, { "epoch": 0.7234767788666383, "grad_norm": 14.0625, "learning_rate": 1.7857489205881239e-06, "loss": 0.8907, "step": 8490 }, { "epoch": 0.7235619940349383, "grad_norm": 13.8125, "learning_rate": 1.7855970965084855e-06, "loss": 0.5191, "step": 8491 }, { "epoch": 0.7236472092032382, "grad_norm": 11.875, "learning_rate": 1.7854452627500684e-06, "loss": 0.5629, "step": 8492 }, { "epoch": 0.7237324243715382, "grad_norm": 16.625, "learning_rate": 1.7852934193156174e-06, "loss": 0.7152, "step": 8493 }, { "epoch": 0.7238176395398381, "grad_norm": 14.9375, "learning_rate": 1.7851415662078757e-06, "loss": 0.6132, "step": 8494 }, { "epoch": 0.7239028547081381, "grad_norm": 15.625, "learning_rate": 1.7849897034295885e-06, "loss": 0.8485, "step": 8495 }, { "epoch": 0.723988069876438, "grad_norm": 11.6875, "learning_rate": 1.7848378309834987e-06, "loss": 0.5086, "step": 8496 }, { "epoch": 0.724073285044738, "grad_norm": 13.0, "learning_rate": 1.784685948872352e-06, "loss": 0.7946, "step": 8497 }, { "epoch": 0.724158500213038, "grad_norm": 16.125, "learning_rate": 1.784534057098892e-06, "loss": 0.6018, "step": 8498 }, { "epoch": 0.7242437153813379, "grad_norm": 10.4375, "learning_rate": 1.7843821556658646e-06, "loss": 0.496, "step": 8499 }, { "epoch": 0.7243289305496379, "grad_norm": 15.5625, "learning_rate": 1.784230244576014e-06, "loss": 0.5848, "step": 8500 }, { "epoch": 0.7244141457179378, "grad_norm": 19.625, "learning_rate": 1.7840783238320856e-06, "loss": 0.7122, "step": 8501 }, { "epoch": 0.7244993608862378, "grad_norm": 17.0, "learning_rate": 1.7839263934368247e-06, "loss": 0.7983, "step": 8502 }, { "epoch": 0.7245845760545377, "grad_norm": 28.125, "learning_rate": 1.7837744533929775e-06, "loss": 0.7489, "step": 8503 }, { "epoch": 0.7246697912228377, "grad_norm": 21.625, "learning_rate": 1.7836225037032887e-06, "loss": 0.7452, "step": 8504 }, { "epoch": 0.7247550063911377, "grad_norm": 13.3125, "learning_rate": 1.7834705443705047e-06, "loss": 0.7934, "step": 8505 }, { "epoch": 0.7248402215594376, "grad_norm": 14.1875, "learning_rate": 1.7833185753973716e-06, "loss": 0.8602, "step": 8506 }, { "epoch": 0.7249254367277376, "grad_norm": 19.875, "learning_rate": 1.7831665967866358e-06, "loss": 0.9688, "step": 8507 }, { "epoch": 0.7250106518960375, "grad_norm": 15.375, "learning_rate": 1.7830146085410433e-06, "loss": 0.7071, "step": 8508 }, { "epoch": 0.7250958670643375, "grad_norm": 16.375, "learning_rate": 1.7828626106633403e-06, "loss": 0.4915, "step": 8509 }, { "epoch": 0.7251810822326374, "grad_norm": 11.0625, "learning_rate": 1.7827106031562747e-06, "loss": 0.5039, "step": 8510 }, { "epoch": 0.7252662974009374, "grad_norm": 12.3125, "learning_rate": 1.7825585860225927e-06, "loss": 0.4425, "step": 8511 }, { "epoch": 0.7253515125692374, "grad_norm": 21.5, "learning_rate": 1.7824065592650417e-06, "loss": 0.7248, "step": 8512 }, { "epoch": 0.7254367277375373, "grad_norm": 10.875, "learning_rate": 1.7822545228863684e-06, "loss": 0.4146, "step": 8513 }, { "epoch": 0.7255219429058373, "grad_norm": 14.6875, "learning_rate": 1.7821024768893208e-06, "loss": 0.801, "step": 8514 }, { "epoch": 0.7256071580741372, "grad_norm": 18.875, "learning_rate": 1.7819504212766463e-06, "loss": 0.8424, "step": 8515 }, { "epoch": 0.7256923732424372, "grad_norm": 13.125, "learning_rate": 1.7817983560510932e-06, "loss": 0.7092, "step": 8516 }, { "epoch": 0.7257775884107371, "grad_norm": 10.875, "learning_rate": 1.7816462812154086e-06, "loss": 0.4323, "step": 8517 }, { "epoch": 0.7258628035790371, "grad_norm": 19.125, "learning_rate": 1.7814941967723413e-06, "loss": 0.8228, "step": 8518 }, { "epoch": 0.725948018747337, "grad_norm": 14.625, "learning_rate": 1.7813421027246395e-06, "loss": 0.4931, "step": 8519 }, { "epoch": 0.726033233915637, "grad_norm": 13.8125, "learning_rate": 1.7811899990750514e-06, "loss": 0.6817, "step": 8520 }, { "epoch": 0.726118449083937, "grad_norm": 17.5, "learning_rate": 1.7810378858263256e-06, "loss": 0.7145, "step": 8521 }, { "epoch": 0.7262036642522369, "grad_norm": 9.5, "learning_rate": 1.7808857629812114e-06, "loss": 0.283, "step": 8522 }, { "epoch": 0.7262888794205369, "grad_norm": 14.5, "learning_rate": 1.780733630542458e-06, "loss": 0.6382, "step": 8523 }, { "epoch": 0.7263740945888368, "grad_norm": 15.5, "learning_rate": 1.7805814885128136e-06, "loss": 0.6895, "step": 8524 }, { "epoch": 0.7264593097571368, "grad_norm": 21.5, "learning_rate": 1.7804293368950286e-06, "loss": 0.9316, "step": 8525 }, { "epoch": 0.7265445249254368, "grad_norm": 14.875, "learning_rate": 1.7802771756918519e-06, "loss": 0.6778, "step": 8526 }, { "epoch": 0.7266297400937367, "grad_norm": 13.875, "learning_rate": 1.7801250049060336e-06, "loss": 0.7596, "step": 8527 }, { "epoch": 0.7267149552620367, "grad_norm": 15.8125, "learning_rate": 1.779972824540323e-06, "loss": 0.7848, "step": 8528 }, { "epoch": 0.7268001704303366, "grad_norm": 26.625, "learning_rate": 1.7798206345974712e-06, "loss": 1.4553, "step": 8529 }, { "epoch": 0.7268853855986366, "grad_norm": 12.375, "learning_rate": 1.7796684350802268e-06, "loss": 0.4869, "step": 8530 }, { "epoch": 0.7269706007669365, "grad_norm": 12.875, "learning_rate": 1.7795162259913421e-06, "loss": 0.5979, "step": 8531 }, { "epoch": 0.7270558159352365, "grad_norm": 19.5, "learning_rate": 1.779364007333566e-06, "loss": 0.6755, "step": 8532 }, { "epoch": 0.7271410311035365, "grad_norm": 10.25, "learning_rate": 1.7792117791096506e-06, "loss": 0.4563, "step": 8533 }, { "epoch": 0.7272262462718364, "grad_norm": 13.875, "learning_rate": 1.7790595413223455e-06, "loss": 0.5445, "step": 8534 }, { "epoch": 0.7273114614401364, "grad_norm": 12.9375, "learning_rate": 1.778907293974403e-06, "loss": 0.606, "step": 8535 }, { "epoch": 0.7273966766084363, "grad_norm": 17.875, "learning_rate": 1.7787550370685735e-06, "loss": 0.9824, "step": 8536 }, { "epoch": 0.7274818917767363, "grad_norm": 14.25, "learning_rate": 1.7786027706076092e-06, "loss": 0.7652, "step": 8537 }, { "epoch": 0.7275671069450362, "grad_norm": 13.625, "learning_rate": 1.778450494594261e-06, "loss": 0.5231, "step": 8538 }, { "epoch": 0.7276523221133362, "grad_norm": 14.6875, "learning_rate": 1.778298209031281e-06, "loss": 0.7698, "step": 8539 }, { "epoch": 0.7277375372816361, "grad_norm": 15.8125, "learning_rate": 1.7781459139214213e-06, "loss": 1.0011, "step": 8540 }, { "epoch": 0.7278227524499361, "grad_norm": 17.25, "learning_rate": 1.7779936092674339e-06, "loss": 0.876, "step": 8541 }, { "epoch": 0.7279079676182361, "grad_norm": 13.9375, "learning_rate": 1.7778412950720704e-06, "loss": 0.54, "step": 8542 }, { "epoch": 0.727993182786536, "grad_norm": 15.3125, "learning_rate": 1.7776889713380844e-06, "loss": 0.5822, "step": 8543 }, { "epoch": 0.728078397954836, "grad_norm": 13.4375, "learning_rate": 1.777536638068228e-06, "loss": 0.5462, "step": 8544 }, { "epoch": 0.7281636131231359, "grad_norm": 20.5, "learning_rate": 1.7773842952652539e-06, "loss": 1.0861, "step": 8545 }, { "epoch": 0.7282488282914359, "grad_norm": 19.375, "learning_rate": 1.7772319429319153e-06, "loss": 1.0924, "step": 8546 }, { "epoch": 0.7283340434597358, "grad_norm": 9.625, "learning_rate": 1.7770795810709653e-06, "loss": 0.2995, "step": 8547 }, { "epoch": 0.7284192586280358, "grad_norm": 13.5, "learning_rate": 1.7769272096851578e-06, "loss": 0.7393, "step": 8548 }, { "epoch": 0.7285044737963358, "grad_norm": 10.8125, "learning_rate": 1.776774828777245e-06, "loss": 0.453, "step": 8549 }, { "epoch": 0.7285896889646357, "grad_norm": 17.625, "learning_rate": 1.7766224383499819e-06, "loss": 0.7386, "step": 8550 }, { "epoch": 0.7286749041329357, "grad_norm": 19.0, "learning_rate": 1.7764700384061212e-06, "loss": 0.9498, "step": 8551 }, { "epoch": 0.7287601193012356, "grad_norm": 12.0625, "learning_rate": 1.7763176289484182e-06, "loss": 0.5819, "step": 8552 }, { "epoch": 0.7288453344695356, "grad_norm": 12.5, "learning_rate": 1.776165209979626e-06, "loss": 0.6056, "step": 8553 }, { "epoch": 0.7289305496378355, "grad_norm": 9.9375, "learning_rate": 1.7760127815024994e-06, "loss": 0.3769, "step": 8554 }, { "epoch": 0.7290157648061355, "grad_norm": 13.3125, "learning_rate": 1.7758603435197926e-06, "loss": 0.6686, "step": 8555 }, { "epoch": 0.7291009799744355, "grad_norm": 17.25, "learning_rate": 1.7757078960342613e-06, "loss": 0.8655, "step": 8556 }, { "epoch": 0.7291861951427354, "grad_norm": 20.25, "learning_rate": 1.775555439048659e-06, "loss": 0.6658, "step": 8557 }, { "epoch": 0.7292714103110354, "grad_norm": 16.375, "learning_rate": 1.7754029725657417e-06, "loss": 0.8874, "step": 8558 }, { "epoch": 0.7293566254793353, "grad_norm": 14.625, "learning_rate": 1.7752504965882643e-06, "loss": 0.6466, "step": 8559 }, { "epoch": 0.7294418406476353, "grad_norm": 12.0, "learning_rate": 1.7750980111189826e-06, "loss": 0.5399, "step": 8560 }, { "epoch": 0.7295270558159352, "grad_norm": 10.9375, "learning_rate": 1.7749455161606513e-06, "loss": 0.4239, "step": 8561 }, { "epoch": 0.7296122709842352, "grad_norm": 11.375, "learning_rate": 1.7747930117160267e-06, "loss": 0.2861, "step": 8562 }, { "epoch": 0.7296974861525352, "grad_norm": 11.625, "learning_rate": 1.7746404977878651e-06, "loss": 0.419, "step": 8563 }, { "epoch": 0.7297827013208351, "grad_norm": 14.5, "learning_rate": 1.7744879743789218e-06, "loss": 0.7734, "step": 8564 }, { "epoch": 0.7298679164891351, "grad_norm": 14.375, "learning_rate": 1.7743354414919534e-06, "loss": 0.8239, "step": 8565 }, { "epoch": 0.729953131657435, "grad_norm": 18.375, "learning_rate": 1.7741828991297161e-06, "loss": 0.8964, "step": 8566 }, { "epoch": 0.730038346825735, "grad_norm": 12.75, "learning_rate": 1.774030347294967e-06, "loss": 0.5365, "step": 8567 }, { "epoch": 0.7301235619940349, "grad_norm": 12.9375, "learning_rate": 1.7738777859904626e-06, "loss": 0.5052, "step": 8568 }, { "epoch": 0.7302087771623349, "grad_norm": 16.625, "learning_rate": 1.7737252152189599e-06, "loss": 0.4295, "step": 8569 }, { "epoch": 0.7302939923306349, "grad_norm": 12.8125, "learning_rate": 1.7735726349832156e-06, "loss": 0.4678, "step": 8570 }, { "epoch": 0.7303792074989348, "grad_norm": 14.6875, "learning_rate": 1.7734200452859876e-06, "loss": 0.4154, "step": 8571 }, { "epoch": 0.7304644226672348, "grad_norm": 11.375, "learning_rate": 1.7732674461300328e-06, "loss": 0.5816, "step": 8572 }, { "epoch": 0.7305496378355347, "grad_norm": 15.1875, "learning_rate": 1.7731148375181092e-06, "loss": 0.4047, "step": 8573 }, { "epoch": 0.7306348530038347, "grad_norm": 18.875, "learning_rate": 1.7729622194529744e-06, "loss": 1.0405, "step": 8574 }, { "epoch": 0.7307200681721346, "grad_norm": 21.875, "learning_rate": 1.7728095919373863e-06, "loss": 0.8328, "step": 8575 }, { "epoch": 0.7308052833404346, "grad_norm": 12.75, "learning_rate": 1.7726569549741035e-06, "loss": 0.5895, "step": 8576 }, { "epoch": 0.7308904985087346, "grad_norm": 24.0, "learning_rate": 1.7725043085658834e-06, "loss": 0.8665, "step": 8577 }, { "epoch": 0.7309757136770345, "grad_norm": 11.625, "learning_rate": 1.7723516527154855e-06, "loss": 0.6448, "step": 8578 }, { "epoch": 0.7310609288453345, "grad_norm": 18.5, "learning_rate": 1.772198987425668e-06, "loss": 0.6823, "step": 8579 }, { "epoch": 0.7311461440136344, "grad_norm": 15.5625, "learning_rate": 1.7720463126991893e-06, "loss": 0.8641, "step": 8580 }, { "epoch": 0.7312313591819344, "grad_norm": 29.125, "learning_rate": 1.7718936285388088e-06, "loss": 0.8062, "step": 8581 }, { "epoch": 0.7313165743502343, "grad_norm": 13.4375, "learning_rate": 1.7717409349472858e-06, "loss": 0.6037, "step": 8582 }, { "epoch": 0.7314017895185343, "grad_norm": 16.75, "learning_rate": 1.7715882319273794e-06, "loss": 1.1658, "step": 8583 }, { "epoch": 0.7314870046868343, "grad_norm": 20.75, "learning_rate": 1.7714355194818494e-06, "loss": 1.082, "step": 8584 }, { "epoch": 0.7315722198551342, "grad_norm": 16.0, "learning_rate": 1.771282797613455e-06, "loss": 0.9383, "step": 8585 }, { "epoch": 0.7316574350234342, "grad_norm": 16.5, "learning_rate": 1.7711300663249564e-06, "loss": 0.78, "step": 8586 }, { "epoch": 0.7317426501917341, "grad_norm": 14.5, "learning_rate": 1.7709773256191132e-06, "loss": 0.757, "step": 8587 }, { "epoch": 0.7318278653600341, "grad_norm": 17.75, "learning_rate": 1.7708245754986861e-06, "loss": 0.9254, "step": 8588 }, { "epoch": 0.731913080528334, "grad_norm": 17.875, "learning_rate": 1.770671815966435e-06, "loss": 0.835, "step": 8589 }, { "epoch": 0.731998295696634, "grad_norm": 13.1875, "learning_rate": 1.7705190470251207e-06, "loss": 0.7231, "step": 8590 }, { "epoch": 0.732083510864934, "grad_norm": 19.5, "learning_rate": 1.7703662686775036e-06, "loss": 1.0814, "step": 8591 }, { "epoch": 0.7321687260332339, "grad_norm": 21.75, "learning_rate": 1.7702134809263449e-06, "loss": 0.9058, "step": 8592 }, { "epoch": 0.7322539412015339, "grad_norm": 19.875, "learning_rate": 1.7700606837744056e-06, "loss": 0.6174, "step": 8593 }, { "epoch": 0.7323391563698338, "grad_norm": 14.75, "learning_rate": 1.7699078772244467e-06, "loss": 0.9005, "step": 8594 }, { "epoch": 0.7324243715381338, "grad_norm": 11.25, "learning_rate": 1.7697550612792296e-06, "loss": 0.4177, "step": 8595 }, { "epoch": 0.7325095867064337, "grad_norm": 11.125, "learning_rate": 1.769602235941516e-06, "loss": 0.4524, "step": 8596 }, { "epoch": 0.7325948018747337, "grad_norm": 13.1875, "learning_rate": 1.7694494012140676e-06, "loss": 0.7261, "step": 8597 }, { "epoch": 0.7326800170430336, "grad_norm": 13.0, "learning_rate": 1.7692965570996456e-06, "loss": 0.5186, "step": 8598 }, { "epoch": 0.7327652322113336, "grad_norm": 18.625, "learning_rate": 1.7691437036010135e-06, "loss": 0.4464, "step": 8599 }, { "epoch": 0.7328504473796336, "grad_norm": 22.125, "learning_rate": 1.7689908407209321e-06, "loss": 0.9657, "step": 8600 }, { "epoch": 0.7329356625479335, "grad_norm": 14.4375, "learning_rate": 1.7688379684621642e-06, "loss": 0.6473, "step": 8601 }, { "epoch": 0.7330208777162335, "grad_norm": 19.625, "learning_rate": 1.7686850868274726e-06, "loss": 0.6331, "step": 8602 }, { "epoch": 0.7331060928845334, "grad_norm": 13.9375, "learning_rate": 1.76853219581962e-06, "loss": 0.5246, "step": 8603 }, { "epoch": 0.7331913080528334, "grad_norm": 14.8125, "learning_rate": 1.7683792954413694e-06, "loss": 0.5925, "step": 8604 }, { "epoch": 0.7332765232211333, "grad_norm": 15.625, "learning_rate": 1.7682263856954835e-06, "loss": 0.9027, "step": 8605 }, { "epoch": 0.7333617383894333, "grad_norm": 21.875, "learning_rate": 1.7680734665847254e-06, "loss": 0.8009, "step": 8606 }, { "epoch": 0.7334469535577333, "grad_norm": 15.75, "learning_rate": 1.7679205381118592e-06, "loss": 0.5548, "step": 8607 }, { "epoch": 0.7335321687260332, "grad_norm": 17.75, "learning_rate": 1.767767600279648e-06, "loss": 1.0133, "step": 8608 }, { "epoch": 0.7336173838943332, "grad_norm": 14.5, "learning_rate": 1.7676146530908552e-06, "loss": 0.9443, "step": 8609 }, { "epoch": 0.7337025990626331, "grad_norm": 15.125, "learning_rate": 1.7674616965482454e-06, "loss": 0.7443, "step": 8610 }, { "epoch": 0.7337878142309331, "grad_norm": 12.9375, "learning_rate": 1.7673087306545825e-06, "loss": 0.5754, "step": 8611 }, { "epoch": 0.733873029399233, "grad_norm": 13.0625, "learning_rate": 1.7671557554126303e-06, "loss": 0.6166, "step": 8612 }, { "epoch": 0.733958244567533, "grad_norm": 14.1875, "learning_rate": 1.7670027708251542e-06, "loss": 0.6266, "step": 8613 }, { "epoch": 0.734043459735833, "grad_norm": 14.8125, "learning_rate": 1.7668497768949177e-06, "loss": 0.7187, "step": 8614 }, { "epoch": 0.7341286749041329, "grad_norm": 11.8125, "learning_rate": 1.766696773624686e-06, "loss": 0.4585, "step": 8615 }, { "epoch": 0.7342138900724329, "grad_norm": 19.375, "learning_rate": 1.766543761017224e-06, "loss": 0.441, "step": 8616 }, { "epoch": 0.7342991052407328, "grad_norm": 18.375, "learning_rate": 1.766390739075297e-06, "loss": 0.8743, "step": 8617 }, { "epoch": 0.7343843204090328, "grad_norm": 16.75, "learning_rate": 1.76623770780167e-06, "loss": 0.5088, "step": 8618 }, { "epoch": 0.7344695355773327, "grad_norm": 23.25, "learning_rate": 1.7660846671991083e-06, "loss": 1.2049, "step": 8619 }, { "epoch": 0.7345547507456327, "grad_norm": 21.0, "learning_rate": 1.7659316172703777e-06, "loss": 0.7994, "step": 8620 }, { "epoch": 0.7346399659139327, "grad_norm": 15.5625, "learning_rate": 1.7657785580182443e-06, "loss": 0.7801, "step": 8621 }, { "epoch": 0.7347251810822326, "grad_norm": 18.5, "learning_rate": 1.7656254894454736e-06, "loss": 0.7233, "step": 8622 }, { "epoch": 0.7348103962505326, "grad_norm": 11.375, "learning_rate": 1.7654724115548316e-06, "loss": 0.3936, "step": 8623 }, { "epoch": 0.7348956114188325, "grad_norm": 14.625, "learning_rate": 1.7653193243490852e-06, "loss": 1.1214, "step": 8624 }, { "epoch": 0.7349808265871325, "grad_norm": 12.3125, "learning_rate": 1.765166227831e-06, "loss": 0.5206, "step": 8625 }, { "epoch": 0.7350660417554324, "grad_norm": 13.625, "learning_rate": 1.765013122003343e-06, "loss": 0.733, "step": 8626 }, { "epoch": 0.7351512569237324, "grad_norm": 12.6875, "learning_rate": 1.7648600068688811e-06, "loss": 0.7921, "step": 8627 }, { "epoch": 0.7352364720920324, "grad_norm": 16.375, "learning_rate": 1.7647068824303815e-06, "loss": 0.8531, "step": 8628 }, { "epoch": 0.7353216872603323, "grad_norm": 12.8125, "learning_rate": 1.7645537486906107e-06, "loss": 0.6251, "step": 8629 }, { "epoch": 0.7354069024286323, "grad_norm": 17.375, "learning_rate": 1.7644006056523358e-06, "loss": 0.9287, "step": 8630 }, { "epoch": 0.7354921175969322, "grad_norm": 12.75, "learning_rate": 1.7642474533183252e-06, "loss": 0.6049, "step": 8631 }, { "epoch": 0.7355773327652322, "grad_norm": 19.375, "learning_rate": 1.7640942916913457e-06, "loss": 0.8854, "step": 8632 }, { "epoch": 0.7356625479335321, "grad_norm": 11.0625, "learning_rate": 1.7639411207741657e-06, "loss": 0.3173, "step": 8633 }, { "epoch": 0.7357477631018321, "grad_norm": 12.5, "learning_rate": 1.7637879405695526e-06, "loss": 0.4365, "step": 8634 }, { "epoch": 0.735832978270132, "grad_norm": 20.625, "learning_rate": 1.7636347510802751e-06, "loss": 0.729, "step": 8635 }, { "epoch": 0.735918193438432, "grad_norm": 13.625, "learning_rate": 1.7634815523091007e-06, "loss": 0.9007, "step": 8636 }, { "epoch": 0.736003408606732, "grad_norm": 21.375, "learning_rate": 1.7633283442587989e-06, "loss": 0.8828, "step": 8637 }, { "epoch": 0.7360886237750319, "grad_norm": 13.375, "learning_rate": 1.763175126932137e-06, "loss": 0.5454, "step": 8638 }, { "epoch": 0.7361738389433319, "grad_norm": 16.625, "learning_rate": 1.763021900331885e-06, "loss": 0.8719, "step": 8639 }, { "epoch": 0.7362590541116318, "grad_norm": 16.875, "learning_rate": 1.7628686644608113e-06, "loss": 0.5728, "step": 8640 }, { "epoch": 0.7363442692799318, "grad_norm": 17.375, "learning_rate": 1.762715419321685e-06, "loss": 0.9567, "step": 8641 }, { "epoch": 0.7364294844482318, "grad_norm": 16.875, "learning_rate": 1.7625621649172755e-06, "loss": 0.7863, "step": 8642 }, { "epoch": 0.7365146996165317, "grad_norm": 13.3125, "learning_rate": 1.7624089012503523e-06, "loss": 0.7025, "step": 8643 }, { "epoch": 0.7365999147848317, "grad_norm": 14.8125, "learning_rate": 1.762255628323685e-06, "loss": 0.7628, "step": 8644 }, { "epoch": 0.7366851299531316, "grad_norm": 10.625, "learning_rate": 1.7621023461400432e-06, "loss": 0.3108, "step": 8645 }, { "epoch": 0.7367703451214316, "grad_norm": 13.75, "learning_rate": 1.7619490547021976e-06, "loss": 0.6916, "step": 8646 }, { "epoch": 0.7368555602897315, "grad_norm": 10.8125, "learning_rate": 1.7617957540129171e-06, "loss": 0.4654, "step": 8647 }, { "epoch": 0.7369407754580315, "grad_norm": 13.9375, "learning_rate": 1.761642444074973e-06, "loss": 0.7203, "step": 8648 }, { "epoch": 0.7370259906263315, "grad_norm": 10.1875, "learning_rate": 1.7614891248911353e-06, "loss": 0.4617, "step": 8649 }, { "epoch": 0.7371112057946314, "grad_norm": 14.0, "learning_rate": 1.7613357964641753e-06, "loss": 0.5827, "step": 8650 }, { "epoch": 0.7371964209629314, "grad_norm": 15.8125, "learning_rate": 1.7611824587968624e-06, "loss": 0.7434, "step": 8651 }, { "epoch": 0.7372816361312313, "grad_norm": 12.6875, "learning_rate": 1.7610291118919689e-06, "loss": 0.3914, "step": 8652 }, { "epoch": 0.7373668512995313, "grad_norm": 12.375, "learning_rate": 1.7608757557522656e-06, "loss": 0.6868, "step": 8653 }, { "epoch": 0.7374520664678312, "grad_norm": 19.125, "learning_rate": 1.7607223903805237e-06, "loss": 0.6665, "step": 8654 }, { "epoch": 0.7375372816361312, "grad_norm": 13.8125, "learning_rate": 1.7605690157795141e-06, "loss": 0.657, "step": 8655 }, { "epoch": 0.7376224968044311, "grad_norm": 16.125, "learning_rate": 1.7604156319520097e-06, "loss": 1.1056, "step": 8656 }, { "epoch": 0.7377077119727311, "grad_norm": 19.375, "learning_rate": 1.7602622389007811e-06, "loss": 0.8459, "step": 8657 }, { "epoch": 0.7377929271410311, "grad_norm": 16.875, "learning_rate": 1.760108836628601e-06, "loss": 0.756, "step": 8658 }, { "epoch": 0.737878142309331, "grad_norm": 12.125, "learning_rate": 1.759955425138241e-06, "loss": 0.5786, "step": 8659 }, { "epoch": 0.737963357477631, "grad_norm": 16.375, "learning_rate": 1.7598020044324738e-06, "loss": 1.063, "step": 8660 }, { "epoch": 0.7380485726459309, "grad_norm": 15.6875, "learning_rate": 1.7596485745140718e-06, "loss": 0.6755, "step": 8661 }, { "epoch": 0.7381337878142309, "grad_norm": 15.375, "learning_rate": 1.7594951353858069e-06, "loss": 0.6532, "step": 8662 }, { "epoch": 0.7382190029825308, "grad_norm": 15.25, "learning_rate": 1.7593416870504535e-06, "loss": 0.7708, "step": 8663 }, { "epoch": 0.7383042181508308, "grad_norm": 12.9375, "learning_rate": 1.759188229510783e-06, "loss": 0.7735, "step": 8664 }, { "epoch": 0.7383894333191308, "grad_norm": 16.125, "learning_rate": 1.7590347627695698e-06, "loss": 0.6911, "step": 8665 }, { "epoch": 0.7384746484874307, "grad_norm": 12.875, "learning_rate": 1.758881286829586e-06, "loss": 0.6542, "step": 8666 }, { "epoch": 0.7385598636557307, "grad_norm": 16.875, "learning_rate": 1.758727801693606e-06, "loss": 0.9329, "step": 8667 }, { "epoch": 0.7386450788240306, "grad_norm": 17.25, "learning_rate": 1.7585743073644027e-06, "loss": 1.1321, "step": 8668 }, { "epoch": 0.7387302939923306, "grad_norm": 13.9375, "learning_rate": 1.7584208038447506e-06, "loss": 0.5965, "step": 8669 }, { "epoch": 0.7388155091606305, "grad_norm": 17.75, "learning_rate": 1.758267291137423e-06, "loss": 1.1661, "step": 8670 }, { "epoch": 0.7389007243289305, "grad_norm": 12.375, "learning_rate": 1.7581137692451945e-06, "loss": 0.5061, "step": 8671 }, { "epoch": 0.7389859394972305, "grad_norm": 16.125, "learning_rate": 1.757960238170839e-06, "loss": 0.3777, "step": 8672 }, { "epoch": 0.7390711546655304, "grad_norm": 16.875, "learning_rate": 1.7578066979171319e-06, "loss": 1.1345, "step": 8673 }, { "epoch": 0.7391563698338304, "grad_norm": 22.875, "learning_rate": 1.7576531484868464e-06, "loss": 0.7862, "step": 8674 }, { "epoch": 0.7392415850021303, "grad_norm": 12.75, "learning_rate": 1.7574995898827585e-06, "loss": 0.734, "step": 8675 }, { "epoch": 0.7393268001704303, "grad_norm": 13.375, "learning_rate": 1.7573460221076422e-06, "loss": 0.7169, "step": 8676 }, { "epoch": 0.7394120153387302, "grad_norm": 17.75, "learning_rate": 1.7571924451642735e-06, "loss": 0.903, "step": 8677 }, { "epoch": 0.7394972305070302, "grad_norm": 19.625, "learning_rate": 1.7570388590554272e-06, "loss": 1.1479, "step": 8678 }, { "epoch": 0.7395824456753302, "grad_norm": 32.5, "learning_rate": 1.756885263783879e-06, "loss": 1.3808, "step": 8679 }, { "epoch": 0.7396676608436301, "grad_norm": 16.375, "learning_rate": 1.7567316593524038e-06, "loss": 0.6421, "step": 8680 }, { "epoch": 0.7397528760119301, "grad_norm": 10.375, "learning_rate": 1.7565780457637784e-06, "loss": 0.4352, "step": 8681 }, { "epoch": 0.73983809118023, "grad_norm": 18.25, "learning_rate": 1.7564244230207786e-06, "loss": 0.9484, "step": 8682 }, { "epoch": 0.73992330634853, "grad_norm": 18.125, "learning_rate": 1.7562707911261796e-06, "loss": 0.742, "step": 8683 }, { "epoch": 0.7400085215168299, "grad_norm": 20.25, "learning_rate": 1.7561171500827587e-06, "loss": 1.0956, "step": 8684 }, { "epoch": 0.7400937366851299, "grad_norm": 14.0, "learning_rate": 1.7559634998932918e-06, "loss": 0.6639, "step": 8685 }, { "epoch": 0.7401789518534299, "grad_norm": 13.25, "learning_rate": 1.755809840560556e-06, "loss": 0.6339, "step": 8686 }, { "epoch": 0.7402641670217298, "grad_norm": 11.0, "learning_rate": 1.7556561720873272e-06, "loss": 0.3729, "step": 8687 }, { "epoch": 0.7403493821900298, "grad_norm": 11.4375, "learning_rate": 1.7555024944763833e-06, "loss": 0.6727, "step": 8688 }, { "epoch": 0.7404345973583298, "grad_norm": 15.75, "learning_rate": 1.7553488077305006e-06, "loss": 0.7158, "step": 8689 }, { "epoch": 0.7405198125266298, "grad_norm": 12.0625, "learning_rate": 1.7551951118524574e-06, "loss": 0.6316, "step": 8690 }, { "epoch": 0.7406050276949298, "grad_norm": 15.875, "learning_rate": 1.75504140684503e-06, "loss": 1.0444, "step": 8691 }, { "epoch": 0.7406902428632297, "grad_norm": 15.5, "learning_rate": 1.7548876927109967e-06, "loss": 0.5795, "step": 8692 }, { "epoch": 0.7407754580315297, "grad_norm": 14.0, "learning_rate": 1.7547339694531349e-06, "loss": 0.7421, "step": 8693 }, { "epoch": 0.7408606731998296, "grad_norm": 13.3125, "learning_rate": 1.7545802370742233e-06, "loss": 0.7057, "step": 8694 }, { "epoch": 0.7409458883681296, "grad_norm": 12.0625, "learning_rate": 1.7544264955770388e-06, "loss": 0.491, "step": 8695 }, { "epoch": 0.7410311035364295, "grad_norm": 41.75, "learning_rate": 1.7542727449643605e-06, "loss": 1.3983, "step": 8696 }, { "epoch": 0.7411163187047295, "grad_norm": 33.5, "learning_rate": 1.7541189852389664e-06, "loss": 1.3734, "step": 8697 }, { "epoch": 0.7412015338730295, "grad_norm": 17.5, "learning_rate": 1.7539652164036359e-06, "loss": 0.895, "step": 8698 }, { "epoch": 0.7412867490413294, "grad_norm": 12.0, "learning_rate": 1.7538114384611466e-06, "loss": 0.7135, "step": 8699 }, { "epoch": 0.7413719642096294, "grad_norm": 10.0, "learning_rate": 1.7536576514142781e-06, "loss": 0.6409, "step": 8700 }, { "epoch": 0.7414571793779293, "grad_norm": 14.9375, "learning_rate": 1.7535038552658097e-06, "loss": 0.7891, "step": 8701 }, { "epoch": 0.7415423945462293, "grad_norm": 18.0, "learning_rate": 1.7533500500185198e-06, "loss": 0.5417, "step": 8702 }, { "epoch": 0.7416276097145292, "grad_norm": 15.5, "learning_rate": 1.7531962356751886e-06, "loss": 0.7711, "step": 8703 }, { "epoch": 0.7417128248828292, "grad_norm": 18.625, "learning_rate": 1.7530424122385953e-06, "loss": 0.7421, "step": 8704 }, { "epoch": 0.7417980400511291, "grad_norm": 24.75, "learning_rate": 1.75288857971152e-06, "loss": 0.8814, "step": 8705 }, { "epoch": 0.7418832552194291, "grad_norm": 12.1875, "learning_rate": 1.7527347380967424e-06, "loss": 0.5617, "step": 8706 }, { "epoch": 0.7419684703877291, "grad_norm": 14.625, "learning_rate": 1.7525808873970424e-06, "loss": 0.8583, "step": 8707 }, { "epoch": 0.742053685556029, "grad_norm": 13.0625, "learning_rate": 1.7524270276152003e-06, "loss": 0.6855, "step": 8708 }, { "epoch": 0.742138900724329, "grad_norm": 8.1875, "learning_rate": 1.7522731587539967e-06, "loss": 0.2737, "step": 8709 }, { "epoch": 0.7422241158926289, "grad_norm": 12.8125, "learning_rate": 1.7521192808162121e-06, "loss": 0.4955, "step": 8710 }, { "epoch": 0.7423093310609289, "grad_norm": 15.3125, "learning_rate": 1.7519653938046271e-06, "loss": 0.8397, "step": 8711 }, { "epoch": 0.7423945462292288, "grad_norm": 14.1875, "learning_rate": 1.7518114977220223e-06, "loss": 1.0708, "step": 8712 }, { "epoch": 0.7424797613975288, "grad_norm": 14.5, "learning_rate": 1.75165759257118e-06, "loss": 0.9256, "step": 8713 }, { "epoch": 0.7425649765658288, "grad_norm": 12.5625, "learning_rate": 1.75150367835488e-06, "loss": 0.8408, "step": 8714 }, { "epoch": 0.7426501917341287, "grad_norm": 20.0, "learning_rate": 1.751349755075904e-06, "loss": 1.176, "step": 8715 }, { "epoch": 0.7427354069024287, "grad_norm": 19.875, "learning_rate": 1.751195822737034e-06, "loss": 0.818, "step": 8716 }, { "epoch": 0.7428206220707286, "grad_norm": 24.125, "learning_rate": 1.7510418813410513e-06, "loss": 1.343, "step": 8717 }, { "epoch": 0.7429058372390286, "grad_norm": 11.375, "learning_rate": 1.7508879308907387e-06, "loss": 0.3106, "step": 8718 }, { "epoch": 0.7429910524073285, "grad_norm": 14.125, "learning_rate": 1.7507339713888765e-06, "loss": 0.6693, "step": 8719 }, { "epoch": 0.7430762675756285, "grad_norm": 23.75, "learning_rate": 1.7505800028382485e-06, "loss": 1.0575, "step": 8720 }, { "epoch": 0.7431614827439285, "grad_norm": 13.5625, "learning_rate": 1.7504260252416363e-06, "loss": 0.7606, "step": 8721 }, { "epoch": 0.7432466979122284, "grad_norm": 16.5, "learning_rate": 1.7502720386018233e-06, "loss": 0.8169, "step": 8722 }, { "epoch": 0.7433319130805284, "grad_norm": 19.125, "learning_rate": 1.7501180429215905e-06, "loss": 0.9029, "step": 8723 }, { "epoch": 0.7434171282488283, "grad_norm": 17.875, "learning_rate": 1.7499640382037222e-06, "loss": 0.7621, "step": 8724 }, { "epoch": 0.7435023434171283, "grad_norm": 15.0625, "learning_rate": 1.7498100244510007e-06, "loss": 0.9402, "step": 8725 }, { "epoch": 0.7435875585854282, "grad_norm": 14.6875, "learning_rate": 1.7496560016662102e-06, "loss": 1.0549, "step": 8726 }, { "epoch": 0.7436727737537282, "grad_norm": 17.125, "learning_rate": 1.7495019698521327e-06, "loss": 1.1288, "step": 8727 }, { "epoch": 0.7437579889220282, "grad_norm": 16.75, "learning_rate": 1.7493479290115523e-06, "loss": 0.5801, "step": 8728 }, { "epoch": 0.7438432040903281, "grad_norm": 16.625, "learning_rate": 1.7491938791472527e-06, "loss": 0.3154, "step": 8729 }, { "epoch": 0.7439284192586281, "grad_norm": 11.25, "learning_rate": 1.749039820262018e-06, "loss": 0.5015, "step": 8730 }, { "epoch": 0.744013634426928, "grad_norm": 17.0, "learning_rate": 1.7488857523586317e-06, "loss": 0.6858, "step": 8731 }, { "epoch": 0.744098849595228, "grad_norm": 21.375, "learning_rate": 1.748731675439878e-06, "loss": 0.5715, "step": 8732 }, { "epoch": 0.7441840647635279, "grad_norm": 16.75, "learning_rate": 1.7485775895085416e-06, "loss": 0.4485, "step": 8733 }, { "epoch": 0.7442692799318279, "grad_norm": 12.1875, "learning_rate": 1.7484234945674072e-06, "loss": 0.5973, "step": 8734 }, { "epoch": 0.7443544951001279, "grad_norm": 16.25, "learning_rate": 1.7482693906192588e-06, "loss": 0.495, "step": 8735 }, { "epoch": 0.7444397102684278, "grad_norm": 11.625, "learning_rate": 1.7481152776668812e-06, "loss": 1.3128, "step": 8736 }, { "epoch": 0.7445249254367278, "grad_norm": 14.75, "learning_rate": 1.74796115571306e-06, "loss": 0.7134, "step": 8737 }, { "epoch": 0.7446101406050277, "grad_norm": 15.125, "learning_rate": 1.7478070247605798e-06, "loss": 0.6541, "step": 8738 }, { "epoch": 0.7446953557733277, "grad_norm": 21.125, "learning_rate": 1.747652884812226e-06, "loss": 0.7929, "step": 8739 }, { "epoch": 0.7447805709416276, "grad_norm": 35.25, "learning_rate": 1.747498735870784e-06, "loss": 1.3412, "step": 8740 }, { "epoch": 0.7448657861099276, "grad_norm": 20.875, "learning_rate": 1.7473445779390396e-06, "loss": 1.2407, "step": 8741 }, { "epoch": 0.7449510012782276, "grad_norm": 16.75, "learning_rate": 1.747190411019779e-06, "loss": 0.9036, "step": 8742 }, { "epoch": 0.7450362164465275, "grad_norm": 18.25, "learning_rate": 1.7470362351157875e-06, "loss": 0.8357, "step": 8743 }, { "epoch": 0.7451214316148275, "grad_norm": 18.625, "learning_rate": 1.7468820502298511e-06, "loss": 0.9005, "step": 8744 }, { "epoch": 0.7452066467831274, "grad_norm": 12.625, "learning_rate": 1.746727856364757e-06, "loss": 0.5742, "step": 8745 }, { "epoch": 0.7452918619514274, "grad_norm": 16.875, "learning_rate": 1.7465736535232901e-06, "loss": 0.7629, "step": 8746 }, { "epoch": 0.7453770771197273, "grad_norm": 16.75, "learning_rate": 1.746419441708239e-06, "loss": 0.8716, "step": 8747 }, { "epoch": 0.7454622922880273, "grad_norm": 14.9375, "learning_rate": 1.7462652209223888e-06, "loss": 0.5686, "step": 8748 }, { "epoch": 0.7455475074563273, "grad_norm": 14.4375, "learning_rate": 1.7461109911685272e-06, "loss": 0.6032, "step": 8749 }, { "epoch": 0.7456327226246272, "grad_norm": 11.0, "learning_rate": 1.7459567524494409e-06, "loss": 0.3629, "step": 8750 }, { "epoch": 0.7457179377929272, "grad_norm": 14.1875, "learning_rate": 1.7458025047679178e-06, "loss": 0.6784, "step": 8751 }, { "epoch": 0.7458031529612271, "grad_norm": 26.25, "learning_rate": 1.7456482481267445e-06, "loss": 0.576, "step": 8752 }, { "epoch": 0.7458883681295271, "grad_norm": 25.0, "learning_rate": 1.7454939825287087e-06, "loss": 0.7948, "step": 8753 }, { "epoch": 0.745973583297827, "grad_norm": 18.625, "learning_rate": 1.7453397079765991e-06, "loss": 0.8094, "step": 8754 }, { "epoch": 0.746058798466127, "grad_norm": 15.0, "learning_rate": 1.7451854244732025e-06, "loss": 0.8838, "step": 8755 }, { "epoch": 0.746144013634427, "grad_norm": 15.8125, "learning_rate": 1.7450311320213074e-06, "loss": 0.779, "step": 8756 }, { "epoch": 0.7462292288027269, "grad_norm": 19.25, "learning_rate": 1.7448768306237015e-06, "loss": 0.7891, "step": 8757 }, { "epoch": 0.7463144439710269, "grad_norm": 15.5625, "learning_rate": 1.744722520283174e-06, "loss": 0.9333, "step": 8758 }, { "epoch": 0.7463996591393268, "grad_norm": 16.75, "learning_rate": 1.7445682010025133e-06, "loss": 0.6183, "step": 8759 }, { "epoch": 0.7464848743076268, "grad_norm": 12.125, "learning_rate": 1.7444138727845078e-06, "loss": 0.3199, "step": 8760 }, { "epoch": 0.7465700894759267, "grad_norm": 13.125, "learning_rate": 1.7442595356319461e-06, "loss": 0.4685, "step": 8761 }, { "epoch": 0.7466553046442267, "grad_norm": 18.25, "learning_rate": 1.744105189547618e-06, "loss": 0.9534, "step": 8762 }, { "epoch": 0.7467405198125266, "grad_norm": 14.5, "learning_rate": 1.743950834534312e-06, "loss": 0.7155, "step": 8763 }, { "epoch": 0.7468257349808266, "grad_norm": 14.0625, "learning_rate": 1.743796470594818e-06, "loss": 0.6915, "step": 8764 }, { "epoch": 0.7469109501491266, "grad_norm": 17.0, "learning_rate": 1.7436420977319246e-06, "loss": 0.9821, "step": 8765 }, { "epoch": 0.7469961653174265, "grad_norm": 14.875, "learning_rate": 1.7434877159484225e-06, "loss": 0.763, "step": 8766 }, { "epoch": 0.7470813804857265, "grad_norm": 17.625, "learning_rate": 1.7433333252471013e-06, "loss": 0.6521, "step": 8767 }, { "epoch": 0.7471665956540264, "grad_norm": 14.875, "learning_rate": 1.7431789256307505e-06, "loss": 0.7515, "step": 8768 }, { "epoch": 0.7472518108223264, "grad_norm": 12.0, "learning_rate": 1.7430245171021608e-06, "loss": 0.3938, "step": 8769 }, { "epoch": 0.7473370259906263, "grad_norm": 15.4375, "learning_rate": 1.742870099664122e-06, "loss": 0.9272, "step": 8770 }, { "epoch": 0.7474222411589263, "grad_norm": 12.6875, "learning_rate": 1.7427156733194255e-06, "loss": 0.4833, "step": 8771 }, { "epoch": 0.7475074563272263, "grad_norm": 15.5, "learning_rate": 1.742561238070861e-06, "loss": 0.8576, "step": 8772 }, { "epoch": 0.7475926714955262, "grad_norm": 18.375, "learning_rate": 1.7424067939212195e-06, "loss": 0.9216, "step": 8773 }, { "epoch": 0.7476778866638262, "grad_norm": 13.4375, "learning_rate": 1.742252340873292e-06, "loss": 0.6485, "step": 8774 }, { "epoch": 0.7477631018321261, "grad_norm": 12.8125, "learning_rate": 1.7420978789298702e-06, "loss": 0.8847, "step": 8775 }, { "epoch": 0.7478483170004261, "grad_norm": 13.9375, "learning_rate": 1.7419434080937442e-06, "loss": 0.508, "step": 8776 }, { "epoch": 0.747933532168726, "grad_norm": 18.0, "learning_rate": 1.7417889283677066e-06, "loss": 1.007, "step": 8777 }, { "epoch": 0.748018747337026, "grad_norm": 10.25, "learning_rate": 1.7416344397545482e-06, "loss": 0.291, "step": 8778 }, { "epoch": 0.748103962505326, "grad_norm": 15.0, "learning_rate": 1.7414799422570616e-06, "loss": 0.864, "step": 8779 }, { "epoch": 0.7481891776736259, "grad_norm": 16.125, "learning_rate": 1.7413254358780377e-06, "loss": 0.9047, "step": 8780 }, { "epoch": 0.7482743928419259, "grad_norm": 12.0625, "learning_rate": 1.7411709206202692e-06, "loss": 0.646, "step": 8781 }, { "epoch": 0.7483596080102258, "grad_norm": 16.5, "learning_rate": 1.7410163964865483e-06, "loss": 0.7902, "step": 8782 }, { "epoch": 0.7484448231785258, "grad_norm": 23.5, "learning_rate": 1.7408618634796675e-06, "loss": 0.8568, "step": 8783 }, { "epoch": 0.7485300383468257, "grad_norm": 18.0, "learning_rate": 1.7407073216024195e-06, "loss": 0.8488, "step": 8784 }, { "epoch": 0.7486152535151257, "grad_norm": 13.8125, "learning_rate": 1.7405527708575962e-06, "loss": 0.7358, "step": 8785 }, { "epoch": 0.7487004686834257, "grad_norm": 16.625, "learning_rate": 1.7403982112479912e-06, "loss": 0.7834, "step": 8786 }, { "epoch": 0.7487856838517256, "grad_norm": 14.875, "learning_rate": 1.7402436427763974e-06, "loss": 0.8336, "step": 8787 }, { "epoch": 0.7488708990200256, "grad_norm": 19.125, "learning_rate": 1.7400890654456083e-06, "loss": 0.5317, "step": 8788 }, { "epoch": 0.7489561141883255, "grad_norm": 10.0, "learning_rate": 1.7399344792584163e-06, "loss": 0.3257, "step": 8789 }, { "epoch": 0.7490413293566255, "grad_norm": 14.375, "learning_rate": 1.739779884217616e-06, "loss": 0.3037, "step": 8790 }, { "epoch": 0.7491265445249254, "grad_norm": 13.375, "learning_rate": 1.739625280326001e-06, "loss": 0.78, "step": 8791 }, { "epoch": 0.7492117596932254, "grad_norm": 16.625, "learning_rate": 1.7394706675863645e-06, "loss": 0.5396, "step": 8792 }, { "epoch": 0.7492969748615254, "grad_norm": 13.0, "learning_rate": 1.7393160460015007e-06, "loss": 0.6034, "step": 8793 }, { "epoch": 0.7493821900298253, "grad_norm": 13.3125, "learning_rate": 1.7391614155742042e-06, "loss": 0.5717, "step": 8794 }, { "epoch": 0.7494674051981253, "grad_norm": 15.5, "learning_rate": 1.7390067763072688e-06, "loss": 0.7804, "step": 8795 }, { "epoch": 0.7495526203664252, "grad_norm": 25.0, "learning_rate": 1.7388521282034895e-06, "loss": 0.6993, "step": 8796 }, { "epoch": 0.7496378355347252, "grad_norm": 14.0625, "learning_rate": 1.73869747126566e-06, "loss": 0.3596, "step": 8797 }, { "epoch": 0.7497230507030251, "grad_norm": 13.5, "learning_rate": 1.7385428054965764e-06, "loss": 0.8027, "step": 8798 }, { "epoch": 0.7498082658713251, "grad_norm": 16.25, "learning_rate": 1.738388130899033e-06, "loss": 0.7779, "step": 8799 }, { "epoch": 0.749893481039625, "grad_norm": 33.25, "learning_rate": 1.7382334474758248e-06, "loss": 0.7611, "step": 8800 }, { "epoch": 0.749978696207925, "grad_norm": 15.6875, "learning_rate": 1.7380787552297474e-06, "loss": 0.8858, "step": 8801 }, { "epoch": 0.750063911376225, "grad_norm": 17.375, "learning_rate": 1.7379240541635957e-06, "loss": 0.8558, "step": 8802 }, { "epoch": 0.7501491265445249, "grad_norm": 11.0625, "learning_rate": 1.7377693442801664e-06, "loss": 0.4349, "step": 8803 }, { "epoch": 0.7502343417128249, "grad_norm": 12.4375, "learning_rate": 1.7376146255822537e-06, "loss": 0.4543, "step": 8804 }, { "epoch": 0.7503195568811248, "grad_norm": 12.0, "learning_rate": 1.7374598980726548e-06, "loss": 0.6571, "step": 8805 }, { "epoch": 0.7504047720494248, "grad_norm": 21.5, "learning_rate": 1.7373051617541654e-06, "loss": 1.2828, "step": 8806 }, { "epoch": 0.7504899872177248, "grad_norm": 14.9375, "learning_rate": 1.7371504166295818e-06, "loss": 0.549, "step": 8807 }, { "epoch": 0.7505752023860247, "grad_norm": 11.25, "learning_rate": 1.7369956627017e-06, "loss": 0.721, "step": 8808 }, { "epoch": 0.7506604175543247, "grad_norm": 18.0, "learning_rate": 1.7368408999733172e-06, "loss": 0.8061, "step": 8809 }, { "epoch": 0.7507456327226246, "grad_norm": 16.0, "learning_rate": 1.7366861284472295e-06, "loss": 0.3442, "step": 8810 }, { "epoch": 0.7508308478909246, "grad_norm": 17.125, "learning_rate": 1.7365313481262344e-06, "loss": 0.8165, "step": 8811 }, { "epoch": 0.7509160630592245, "grad_norm": 15.8125, "learning_rate": 1.7363765590131282e-06, "loss": 0.8678, "step": 8812 }, { "epoch": 0.7510012782275245, "grad_norm": 16.75, "learning_rate": 1.7362217611107087e-06, "loss": 0.7003, "step": 8813 }, { "epoch": 0.7510864933958245, "grad_norm": 17.375, "learning_rate": 1.7360669544217728e-06, "loss": 0.6057, "step": 8814 }, { "epoch": 0.7511717085641244, "grad_norm": 23.625, "learning_rate": 1.7359121389491184e-06, "loss": 1.0507, "step": 8815 }, { "epoch": 0.7512569237324244, "grad_norm": 15.5, "learning_rate": 1.7357573146955434e-06, "loss": 0.7692, "step": 8816 }, { "epoch": 0.7513421389007243, "grad_norm": 14.5625, "learning_rate": 1.7356024816638447e-06, "loss": 0.6305, "step": 8817 }, { "epoch": 0.7514273540690243, "grad_norm": 15.6875, "learning_rate": 1.7354476398568213e-06, "loss": 0.4532, "step": 8818 }, { "epoch": 0.7515125692373242, "grad_norm": 22.5, "learning_rate": 1.7352927892772707e-06, "loss": 1.1847, "step": 8819 }, { "epoch": 0.7515977844056242, "grad_norm": 15.25, "learning_rate": 1.7351379299279913e-06, "loss": 0.539, "step": 8820 }, { "epoch": 0.7516829995739241, "grad_norm": 10.75, "learning_rate": 1.7349830618117815e-06, "loss": 0.3418, "step": 8821 }, { "epoch": 0.7517682147422241, "grad_norm": 10.9375, "learning_rate": 1.7348281849314405e-06, "loss": 0.3841, "step": 8822 }, { "epoch": 0.7518534299105241, "grad_norm": 20.375, "learning_rate": 1.7346732992897663e-06, "loss": 1.0008, "step": 8823 }, { "epoch": 0.751938645078824, "grad_norm": 13.3125, "learning_rate": 1.7345184048895588e-06, "loss": 0.8183, "step": 8824 }, { "epoch": 0.752023860247124, "grad_norm": 14.5625, "learning_rate": 1.7343635017336158e-06, "loss": 0.8955, "step": 8825 }, { "epoch": 0.7521090754154239, "grad_norm": 18.25, "learning_rate": 1.7342085898247377e-06, "loss": 0.8762, "step": 8826 }, { "epoch": 0.7521942905837239, "grad_norm": 10.0625, "learning_rate": 1.734053669165723e-06, "loss": 0.2592, "step": 8827 }, { "epoch": 0.7522795057520238, "grad_norm": 14.125, "learning_rate": 1.7338987397593727e-06, "loss": 0.9215, "step": 8828 }, { "epoch": 0.7523647209203238, "grad_norm": 20.125, "learning_rate": 1.7337438016084845e-06, "loss": 1.1696, "step": 8829 }, { "epoch": 0.7524499360886238, "grad_norm": 23.125, "learning_rate": 1.73358885471586e-06, "loss": 1.0865, "step": 8830 }, { "epoch": 0.7525351512569237, "grad_norm": 17.125, "learning_rate": 1.7334338990842982e-06, "loss": 1.0081, "step": 8831 }, { "epoch": 0.7526203664252237, "grad_norm": 13.0, "learning_rate": 1.7332789347166e-06, "loss": 0.5814, "step": 8832 }, { "epoch": 0.7527055815935236, "grad_norm": 20.5, "learning_rate": 1.7331239616155655e-06, "loss": 1.1821, "step": 8833 }, { "epoch": 0.7527907967618236, "grad_norm": 21.0, "learning_rate": 1.7329689797839953e-06, "loss": 0.8885, "step": 8834 }, { "epoch": 0.7528760119301235, "grad_norm": 49.75, "learning_rate": 1.7328139892246894e-06, "loss": 0.995, "step": 8835 }, { "epoch": 0.7529612270984235, "grad_norm": 13.0, "learning_rate": 1.7326589899404498e-06, "loss": 0.4504, "step": 8836 }, { "epoch": 0.7530464422667235, "grad_norm": 16.5, "learning_rate": 1.7325039819340767e-06, "loss": 0.9001, "step": 8837 }, { "epoch": 0.7531316574350234, "grad_norm": 14.1875, "learning_rate": 1.7323489652083714e-06, "loss": 0.6195, "step": 8838 }, { "epoch": 0.7532168726033234, "grad_norm": 16.0, "learning_rate": 1.7321939397661352e-06, "loss": 0.7212, "step": 8839 }, { "epoch": 0.7533020877716233, "grad_norm": 17.875, "learning_rate": 1.73203890561017e-06, "loss": 0.9277, "step": 8840 }, { "epoch": 0.7533873029399233, "grad_norm": 14.0, "learning_rate": 1.7318838627432765e-06, "loss": 0.7808, "step": 8841 }, { "epoch": 0.7534725181082232, "grad_norm": 18.0, "learning_rate": 1.7317288111682573e-06, "loss": 0.7741, "step": 8842 }, { "epoch": 0.7535577332765232, "grad_norm": 14.5, "learning_rate": 1.731573750887914e-06, "loss": 0.8483, "step": 8843 }, { "epoch": 0.7536429484448232, "grad_norm": 11.75, "learning_rate": 1.7314186819050492e-06, "loss": 0.43, "step": 8844 }, { "epoch": 0.7537281636131231, "grad_norm": 22.875, "learning_rate": 1.7312636042224643e-06, "loss": 1.0391, "step": 8845 }, { "epoch": 0.7538133787814231, "grad_norm": 14.625, "learning_rate": 1.731108517842962e-06, "loss": 1.134, "step": 8846 }, { "epoch": 0.753898593949723, "grad_norm": 10.625, "learning_rate": 1.7309534227693453e-06, "loss": 0.3056, "step": 8847 }, { "epoch": 0.753983809118023, "grad_norm": 14.0625, "learning_rate": 1.7307983190044164e-06, "loss": 0.4403, "step": 8848 }, { "epoch": 0.7540690242863229, "grad_norm": 12.0, "learning_rate": 1.7306432065509784e-06, "loss": 0.4834, "step": 8849 }, { "epoch": 0.7541542394546229, "grad_norm": 13.8125, "learning_rate": 1.7304880854118345e-06, "loss": 0.8416, "step": 8850 }, { "epoch": 0.7542394546229229, "grad_norm": 15.3125, "learning_rate": 1.7303329555897873e-06, "loss": 0.9184, "step": 8851 }, { "epoch": 0.7543246697912228, "grad_norm": 14.875, "learning_rate": 1.7301778170876413e-06, "loss": 0.7101, "step": 8852 }, { "epoch": 0.7544098849595228, "grad_norm": 12.375, "learning_rate": 1.7300226699081984e-06, "loss": 0.5973, "step": 8853 }, { "epoch": 0.7544951001278227, "grad_norm": 16.5, "learning_rate": 1.729867514054264e-06, "loss": 1.0795, "step": 8854 }, { "epoch": 0.7545803152961227, "grad_norm": 15.25, "learning_rate": 1.7297123495286406e-06, "loss": 0.6148, "step": 8855 }, { "epoch": 0.7546655304644226, "grad_norm": 14.1875, "learning_rate": 1.729557176334133e-06, "loss": 0.5185, "step": 8856 }, { "epoch": 0.7547507456327226, "grad_norm": 14.25, "learning_rate": 1.7294019944735447e-06, "loss": 0.7482, "step": 8857 }, { "epoch": 0.7548359608010226, "grad_norm": 11.0625, "learning_rate": 1.7292468039496804e-06, "loss": 0.4201, "step": 8858 }, { "epoch": 0.7549211759693225, "grad_norm": 19.125, "learning_rate": 1.7290916047653442e-06, "loss": 0.8541, "step": 8859 }, { "epoch": 0.7550063911376225, "grad_norm": 14.5, "learning_rate": 1.7289363969233413e-06, "loss": 0.8045, "step": 8860 }, { "epoch": 0.7550916063059224, "grad_norm": 20.5, "learning_rate": 1.728781180426476e-06, "loss": 0.9231, "step": 8861 }, { "epoch": 0.7551768214742224, "grad_norm": 13.9375, "learning_rate": 1.728625955277553e-06, "loss": 0.7687, "step": 8862 }, { "epoch": 0.7552620366425223, "grad_norm": 17.0, "learning_rate": 1.7284707214793777e-06, "loss": 1.1376, "step": 8863 }, { "epoch": 0.7553472518108223, "grad_norm": 15.0625, "learning_rate": 1.728315479034756e-06, "loss": 0.4929, "step": 8864 }, { "epoch": 0.7554324669791223, "grad_norm": 23.875, "learning_rate": 1.728160227946492e-06, "loss": 0.9484, "step": 8865 }, { "epoch": 0.7555176821474222, "grad_norm": 11.8125, "learning_rate": 1.728004968217392e-06, "loss": 0.5762, "step": 8866 }, { "epoch": 0.7556028973157222, "grad_norm": 14.875, "learning_rate": 1.7278496998502614e-06, "loss": 0.6943, "step": 8867 }, { "epoch": 0.7556881124840221, "grad_norm": 12.375, "learning_rate": 1.7276944228479067e-06, "loss": 0.4078, "step": 8868 }, { "epoch": 0.7557733276523221, "grad_norm": 15.6875, "learning_rate": 1.7275391372131331e-06, "loss": 0.6442, "step": 8869 }, { "epoch": 0.755858542820622, "grad_norm": 15.75, "learning_rate": 1.7273838429487468e-06, "loss": 1.0687, "step": 8870 }, { "epoch": 0.755943757988922, "grad_norm": 11.75, "learning_rate": 1.727228540057555e-06, "loss": 0.6581, "step": 8871 }, { "epoch": 0.756028973157222, "grad_norm": 11.875, "learning_rate": 1.7270732285423633e-06, "loss": 0.4837, "step": 8872 }, { "epoch": 0.7561141883255219, "grad_norm": 16.5, "learning_rate": 1.7269179084059788e-06, "loss": 0.6691, "step": 8873 }, { "epoch": 0.7561994034938219, "grad_norm": 29.25, "learning_rate": 1.7267625796512078e-06, "loss": 0.9593, "step": 8874 }, { "epoch": 0.7562846186621218, "grad_norm": 18.25, "learning_rate": 1.7266072422808578e-06, "loss": 0.6315, "step": 8875 }, { "epoch": 0.7563698338304218, "grad_norm": 14.0625, "learning_rate": 1.7264518962977356e-06, "loss": 0.7552, "step": 8876 }, { "epoch": 0.7564550489987217, "grad_norm": 14.9375, "learning_rate": 1.7262965417046483e-06, "loss": 0.6384, "step": 8877 }, { "epoch": 0.7565402641670217, "grad_norm": 13.375, "learning_rate": 1.7261411785044036e-06, "loss": 0.5805, "step": 8878 }, { "epoch": 0.7566254793353216, "grad_norm": 17.75, "learning_rate": 1.7259858066998091e-06, "loss": 1.0247, "step": 8879 }, { "epoch": 0.7567106945036216, "grad_norm": 11.25, "learning_rate": 1.7258304262936726e-06, "loss": 0.3864, "step": 8880 }, { "epoch": 0.7567959096719216, "grad_norm": 10.375, "learning_rate": 1.725675037288802e-06, "loss": 0.4187, "step": 8881 }, { "epoch": 0.7568811248402215, "grad_norm": 11.875, "learning_rate": 1.7255196396880044e-06, "loss": 0.5016, "step": 8882 }, { "epoch": 0.7569663400085215, "grad_norm": 13.4375, "learning_rate": 1.7253642334940896e-06, "loss": 0.7801, "step": 8883 }, { "epoch": 0.7570515551768214, "grad_norm": 15.0625, "learning_rate": 1.7252088187098642e-06, "loss": 0.5035, "step": 8884 }, { "epoch": 0.7571367703451214, "grad_norm": 18.0, "learning_rate": 1.7250533953381381e-06, "loss": 0.6229, "step": 8885 }, { "epoch": 0.7572219855134213, "grad_norm": 15.8125, "learning_rate": 1.7248979633817195e-06, "loss": 0.6052, "step": 8886 }, { "epoch": 0.7573072006817213, "grad_norm": 13.8125, "learning_rate": 1.724742522843417e-06, "loss": 0.7106, "step": 8887 }, { "epoch": 0.7573924158500213, "grad_norm": 14.1875, "learning_rate": 1.7245870737260398e-06, "loss": 0.612, "step": 8888 }, { "epoch": 0.7574776310183212, "grad_norm": 13.6875, "learning_rate": 1.724431616032397e-06, "loss": 0.842, "step": 8889 }, { "epoch": 0.7575628461866212, "grad_norm": 21.75, "learning_rate": 1.7242761497652976e-06, "loss": 1.211, "step": 8890 }, { "epoch": 0.7576480613549211, "grad_norm": 12.75, "learning_rate": 1.7241206749275512e-06, "loss": 0.6871, "step": 8891 }, { "epoch": 0.7577332765232211, "grad_norm": 18.875, "learning_rate": 1.723965191521968e-06, "loss": 0.998, "step": 8892 }, { "epoch": 0.757818491691521, "grad_norm": 11.75, "learning_rate": 1.7238096995513567e-06, "loss": 0.4797, "step": 8893 }, { "epoch": 0.757903706859821, "grad_norm": 13.8125, "learning_rate": 1.7236541990185279e-06, "loss": 0.7879, "step": 8894 }, { "epoch": 0.757988922028121, "grad_norm": 15.625, "learning_rate": 1.7234986899262912e-06, "loss": 0.8003, "step": 8895 }, { "epoch": 0.7580741371964209, "grad_norm": 14.6875, "learning_rate": 1.723343172277457e-06, "loss": 0.606, "step": 8896 }, { "epoch": 0.7581593523647209, "grad_norm": 18.625, "learning_rate": 1.723187646074836e-06, "loss": 1.0115, "step": 8897 }, { "epoch": 0.7582445675330208, "grad_norm": 13.0625, "learning_rate": 1.7230321113212383e-06, "loss": 0.5035, "step": 8898 }, { "epoch": 0.7583297827013208, "grad_norm": 16.5, "learning_rate": 1.7228765680194745e-06, "loss": 0.8066, "step": 8899 }, { "epoch": 0.7584149978696207, "grad_norm": 12.25, "learning_rate": 1.7227210161723556e-06, "loss": 0.6151, "step": 8900 }, { "epoch": 0.7585002130379207, "grad_norm": 14.4375, "learning_rate": 1.722565455782693e-06, "loss": 0.8543, "step": 8901 }, { "epoch": 0.7585854282062207, "grad_norm": 19.25, "learning_rate": 1.7224098868532968e-06, "loss": 0.8868, "step": 8902 }, { "epoch": 0.7586706433745206, "grad_norm": 15.875, "learning_rate": 1.7222543093869793e-06, "loss": 0.8081, "step": 8903 }, { "epoch": 0.7587558585428206, "grad_norm": 13.0, "learning_rate": 1.7220987233865517e-06, "loss": 0.3198, "step": 8904 }, { "epoch": 0.7588410737111205, "grad_norm": 19.75, "learning_rate": 1.721943128854825e-06, "loss": 0.8929, "step": 8905 }, { "epoch": 0.7589262888794205, "grad_norm": 11.4375, "learning_rate": 1.7217875257946115e-06, "loss": 0.3741, "step": 8906 }, { "epoch": 0.7590115040477204, "grad_norm": 17.125, "learning_rate": 1.7216319142087232e-06, "loss": 0.8314, "step": 8907 }, { "epoch": 0.7590967192160204, "grad_norm": 31.5, "learning_rate": 1.7214762940999716e-06, "loss": 1.2375, "step": 8908 }, { "epoch": 0.7591819343843204, "grad_norm": 15.5625, "learning_rate": 1.7213206654711697e-06, "loss": 0.9969, "step": 8909 }, { "epoch": 0.7592671495526203, "grad_norm": 50.0, "learning_rate": 1.721165028325129e-06, "loss": 0.6129, "step": 8910 }, { "epoch": 0.7593523647209203, "grad_norm": 14.75, "learning_rate": 1.7210093826646628e-06, "loss": 0.8879, "step": 8911 }, { "epoch": 0.7594375798892202, "grad_norm": 16.75, "learning_rate": 1.720853728492583e-06, "loss": 0.8162, "step": 8912 }, { "epoch": 0.7595227950575202, "grad_norm": 19.125, "learning_rate": 1.7206980658117034e-06, "loss": 0.967, "step": 8913 }, { "epoch": 0.7596080102258201, "grad_norm": 16.875, "learning_rate": 1.720542394624836e-06, "loss": 0.6018, "step": 8914 }, { "epoch": 0.7596932253941202, "grad_norm": 12.8125, "learning_rate": 1.7203867149347946e-06, "loss": 0.3263, "step": 8915 }, { "epoch": 0.7597784405624202, "grad_norm": 9.75, "learning_rate": 1.720231026744392e-06, "loss": 0.203, "step": 8916 }, { "epoch": 0.7598636557307201, "grad_norm": 12.75, "learning_rate": 1.7200753300564422e-06, "loss": 0.576, "step": 8917 }, { "epoch": 0.7599488708990201, "grad_norm": 14.5, "learning_rate": 1.7199196248737582e-06, "loss": 0.7405, "step": 8918 }, { "epoch": 0.76003408606732, "grad_norm": 16.125, "learning_rate": 1.7197639111991543e-06, "loss": 0.9682, "step": 8919 }, { "epoch": 0.76011930123562, "grad_norm": 17.625, "learning_rate": 1.7196081890354439e-06, "loss": 0.8688, "step": 8920 }, { "epoch": 0.76020451640392, "grad_norm": 12.0625, "learning_rate": 1.7194524583854414e-06, "loss": 0.5735, "step": 8921 }, { "epoch": 0.7602897315722199, "grad_norm": 14.25, "learning_rate": 1.719296719251961e-06, "loss": 0.5425, "step": 8922 }, { "epoch": 0.7603749467405199, "grad_norm": 13.75, "learning_rate": 1.7191409716378165e-06, "loss": 0.4827, "step": 8923 }, { "epoch": 0.7604601619088198, "grad_norm": 15.875, "learning_rate": 1.7189852155458236e-06, "loss": 0.9541, "step": 8924 }, { "epoch": 0.7605453770771198, "grad_norm": 18.625, "learning_rate": 1.718829450978796e-06, "loss": 0.9378, "step": 8925 }, { "epoch": 0.7606305922454197, "grad_norm": 12.25, "learning_rate": 1.7186736779395485e-06, "loss": 0.5963, "step": 8926 }, { "epoch": 0.7607158074137197, "grad_norm": 12.875, "learning_rate": 1.7185178964308965e-06, "loss": 0.5026, "step": 8927 }, { "epoch": 0.7608010225820196, "grad_norm": 18.875, "learning_rate": 1.7183621064556551e-06, "loss": 0.6259, "step": 8928 }, { "epoch": 0.7608862377503196, "grad_norm": 17.375, "learning_rate": 1.7182063080166395e-06, "loss": 0.9694, "step": 8929 }, { "epoch": 0.7609714529186196, "grad_norm": 10.0625, "learning_rate": 1.718050501116665e-06, "loss": 0.4604, "step": 8930 }, { "epoch": 0.7610566680869195, "grad_norm": 10.8125, "learning_rate": 1.7178946857585472e-06, "loss": 0.4042, "step": 8931 }, { "epoch": 0.7611418832552195, "grad_norm": 12.5625, "learning_rate": 1.717738861945102e-06, "loss": 0.2951, "step": 8932 }, { "epoch": 0.7612270984235194, "grad_norm": 14.8125, "learning_rate": 1.7175830296791451e-06, "loss": 1.141, "step": 8933 }, { "epoch": 0.7613123135918194, "grad_norm": 14.0, "learning_rate": 1.717427188963493e-06, "loss": 0.809, "step": 8934 }, { "epoch": 0.7613975287601193, "grad_norm": 14.25, "learning_rate": 1.717271339800961e-06, "loss": 0.8487, "step": 8935 }, { "epoch": 0.7614827439284193, "grad_norm": 17.0, "learning_rate": 1.7171154821943664e-06, "loss": 0.6533, "step": 8936 }, { "epoch": 0.7615679590967193, "grad_norm": 13.875, "learning_rate": 1.7169596161465256e-06, "loss": 0.6976, "step": 8937 }, { "epoch": 0.7616531742650192, "grad_norm": 17.875, "learning_rate": 1.7168037416602543e-06, "loss": 0.9174, "step": 8938 }, { "epoch": 0.7617383894333192, "grad_norm": 15.3125, "learning_rate": 1.7166478587383707e-06, "loss": 0.784, "step": 8939 }, { "epoch": 0.7618236046016191, "grad_norm": 11.0625, "learning_rate": 1.7164919673836902e-06, "loss": 0.57, "step": 8940 }, { "epoch": 0.7619088197699191, "grad_norm": 16.5, "learning_rate": 1.7163360675990315e-06, "loss": 0.8436, "step": 8941 }, { "epoch": 0.761994034938219, "grad_norm": 14.9375, "learning_rate": 1.7161801593872106e-06, "loss": 0.4477, "step": 8942 }, { "epoch": 0.762079250106519, "grad_norm": 15.375, "learning_rate": 1.716024242751046e-06, "loss": 0.8258, "step": 8943 }, { "epoch": 0.762164465274819, "grad_norm": 14.375, "learning_rate": 1.7158683176933542e-06, "loss": 0.8524, "step": 8944 }, { "epoch": 0.7622496804431189, "grad_norm": 19.5, "learning_rate": 1.7157123842169537e-06, "loss": 0.8037, "step": 8945 }, { "epoch": 0.7623348956114189, "grad_norm": 12.1875, "learning_rate": 1.7155564423246617e-06, "loss": 0.4825, "step": 8946 }, { "epoch": 0.7624201107797188, "grad_norm": 12.1875, "learning_rate": 1.7154004920192973e-06, "loss": 0.6762, "step": 8947 }, { "epoch": 0.7625053259480188, "grad_norm": 13.9375, "learning_rate": 1.7152445333036774e-06, "loss": 0.4607, "step": 8948 }, { "epoch": 0.7625905411163187, "grad_norm": 15.5625, "learning_rate": 1.7150885661806212e-06, "loss": 0.6954, "step": 8949 }, { "epoch": 0.7626757562846187, "grad_norm": 13.4375, "learning_rate": 1.7149325906529471e-06, "loss": 0.5826, "step": 8950 }, { "epoch": 0.7627609714529187, "grad_norm": 20.5, "learning_rate": 1.7147766067234734e-06, "loss": 1.1907, "step": 8951 }, { "epoch": 0.7628461866212186, "grad_norm": 17.75, "learning_rate": 1.714620614395019e-06, "loss": 1.0698, "step": 8952 }, { "epoch": 0.7629314017895186, "grad_norm": 21.125, "learning_rate": 1.714464613670403e-06, "loss": 0.8887, "step": 8953 }, { "epoch": 0.7630166169578185, "grad_norm": 13.5625, "learning_rate": 1.7143086045524443e-06, "loss": 0.73, "step": 8954 }, { "epoch": 0.7631018321261185, "grad_norm": 11.0, "learning_rate": 1.7141525870439622e-06, "loss": 0.4569, "step": 8955 }, { "epoch": 0.7631870472944184, "grad_norm": 36.25, "learning_rate": 1.713996561147776e-06, "loss": 1.3506, "step": 8956 }, { "epoch": 0.7632722624627184, "grad_norm": 11.3125, "learning_rate": 1.7138405268667058e-06, "loss": 0.5579, "step": 8957 }, { "epoch": 0.7633574776310184, "grad_norm": 15.125, "learning_rate": 1.7136844842035704e-06, "loss": 0.8242, "step": 8958 }, { "epoch": 0.7634426927993183, "grad_norm": 10.375, "learning_rate": 1.71352843316119e-06, "loss": 0.5989, "step": 8959 }, { "epoch": 0.7635279079676183, "grad_norm": 16.375, "learning_rate": 1.7133723737423852e-06, "loss": 1.0877, "step": 8960 }, { "epoch": 0.7636131231359182, "grad_norm": 18.25, "learning_rate": 1.7132163059499751e-06, "loss": 1.0761, "step": 8961 }, { "epoch": 0.7636983383042182, "grad_norm": 15.125, "learning_rate": 1.7130602297867814e-06, "loss": 0.4392, "step": 8962 }, { "epoch": 0.7637835534725181, "grad_norm": 15.3125, "learning_rate": 1.7129041452556227e-06, "loss": 0.837, "step": 8963 }, { "epoch": 0.7638687686408181, "grad_norm": 14.6875, "learning_rate": 1.7127480523593213e-06, "loss": 0.6768, "step": 8964 }, { "epoch": 0.763953983809118, "grad_norm": 17.375, "learning_rate": 1.7125919511006966e-06, "loss": 0.6845, "step": 8965 }, { "epoch": 0.764039198977418, "grad_norm": 15.3125, "learning_rate": 1.712435841482571e-06, "loss": 0.3896, "step": 8966 }, { "epoch": 0.764124414145718, "grad_norm": 30.25, "learning_rate": 1.7122797235077642e-06, "loss": 1.0668, "step": 8967 }, { "epoch": 0.7642096293140179, "grad_norm": 10.625, "learning_rate": 1.7121235971790982e-06, "loss": 0.2162, "step": 8968 }, { "epoch": 0.7642948444823179, "grad_norm": 12.1875, "learning_rate": 1.7119674624993937e-06, "loss": 0.6982, "step": 8969 }, { "epoch": 0.7643800596506178, "grad_norm": 12.25, "learning_rate": 1.711811319471473e-06, "loss": 0.4853, "step": 8970 }, { "epoch": 0.7644652748189178, "grad_norm": 10.5625, "learning_rate": 1.7116551680981572e-06, "loss": 0.339, "step": 8971 }, { "epoch": 0.7645504899872178, "grad_norm": 12.3125, "learning_rate": 1.7114990083822686e-06, "loss": 0.7008, "step": 8972 }, { "epoch": 0.7646357051555177, "grad_norm": 11.375, "learning_rate": 1.7113428403266282e-06, "loss": 0.7619, "step": 8973 }, { "epoch": 0.7647209203238177, "grad_norm": 16.0, "learning_rate": 1.711186663934059e-06, "loss": 1.0166, "step": 8974 }, { "epoch": 0.7648061354921176, "grad_norm": 13.625, "learning_rate": 1.7110304792073835e-06, "loss": 0.5381, "step": 8975 }, { "epoch": 0.7648913506604176, "grad_norm": 15.0, "learning_rate": 1.7108742861494233e-06, "loss": 0.8544, "step": 8976 }, { "epoch": 0.7649765658287175, "grad_norm": 20.375, "learning_rate": 1.710718084763001e-06, "loss": 0.9204, "step": 8977 }, { "epoch": 0.7650617809970175, "grad_norm": 19.125, "learning_rate": 1.71056187505094e-06, "loss": 0.9626, "step": 8978 }, { "epoch": 0.7651469961653175, "grad_norm": 16.0, "learning_rate": 1.710405657016063e-06, "loss": 0.7943, "step": 8979 }, { "epoch": 0.7652322113336174, "grad_norm": 15.75, "learning_rate": 1.7102494306611921e-06, "loss": 0.9519, "step": 8980 }, { "epoch": 0.7653174265019174, "grad_norm": 14.1875, "learning_rate": 1.7100931959891517e-06, "loss": 1.0013, "step": 8981 }, { "epoch": 0.7654026416702173, "grad_norm": 19.375, "learning_rate": 1.7099369530027645e-06, "loss": 0.9115, "step": 8982 }, { "epoch": 0.7654878568385173, "grad_norm": 13.625, "learning_rate": 1.709780701704854e-06, "loss": 0.7726, "step": 8983 }, { "epoch": 0.7655730720068172, "grad_norm": 13.9375, "learning_rate": 1.7096244420982438e-06, "loss": 0.8881, "step": 8984 }, { "epoch": 0.7656582871751172, "grad_norm": 9.875, "learning_rate": 1.709468174185758e-06, "loss": 0.3125, "step": 8985 }, { "epoch": 0.7657435023434171, "grad_norm": 16.25, "learning_rate": 1.70931189797022e-06, "loss": 0.721, "step": 8986 }, { "epoch": 0.7658287175117171, "grad_norm": 14.4375, "learning_rate": 1.7091556134544547e-06, "loss": 0.8042, "step": 8987 }, { "epoch": 0.7659139326800171, "grad_norm": 17.25, "learning_rate": 1.7089993206412848e-06, "loss": 0.8803, "step": 8988 }, { "epoch": 0.765999147848317, "grad_norm": 17.25, "learning_rate": 1.7088430195335362e-06, "loss": 0.9869, "step": 8989 }, { "epoch": 0.766084363016617, "grad_norm": 13.9375, "learning_rate": 1.7086867101340327e-06, "loss": 0.6307, "step": 8990 }, { "epoch": 0.7661695781849169, "grad_norm": 14.25, "learning_rate": 1.708530392445599e-06, "loss": 0.7213, "step": 8991 }, { "epoch": 0.7662547933532169, "grad_norm": 19.75, "learning_rate": 1.7083740664710602e-06, "loss": 0.7864, "step": 8992 }, { "epoch": 0.7663400085215168, "grad_norm": 12.875, "learning_rate": 1.708217732213241e-06, "loss": 0.5595, "step": 8993 }, { "epoch": 0.7664252236898168, "grad_norm": 24.875, "learning_rate": 1.7080613896749667e-06, "loss": 0.6869, "step": 8994 }, { "epoch": 0.7665104388581168, "grad_norm": 16.5, "learning_rate": 1.7079050388590623e-06, "loss": 0.8646, "step": 8995 }, { "epoch": 0.7665956540264167, "grad_norm": 13.875, "learning_rate": 1.7077486797683538e-06, "loss": 0.6568, "step": 8996 }, { "epoch": 0.7666808691947167, "grad_norm": 11.8125, "learning_rate": 1.7075923124056657e-06, "loss": 0.5769, "step": 8997 }, { "epoch": 0.7667660843630166, "grad_norm": 14.0625, "learning_rate": 1.7074359367738252e-06, "loss": 0.6569, "step": 8998 }, { "epoch": 0.7668512995313166, "grad_norm": 12.5625, "learning_rate": 1.7072795528756566e-06, "loss": 0.5376, "step": 8999 }, { "epoch": 0.7669365146996165, "grad_norm": 12.0, "learning_rate": 1.707123160713987e-06, "loss": 0.5588, "step": 9000 }, { "epoch": 0.7670217298679165, "grad_norm": 13.375, "learning_rate": 1.7069667602916417e-06, "loss": 0.5436, "step": 9001 }, { "epoch": 0.7671069450362165, "grad_norm": 12.75, "learning_rate": 1.7068103516114484e-06, "loss": 0.6801, "step": 9002 }, { "epoch": 0.7671921602045164, "grad_norm": 11.4375, "learning_rate": 1.7066539346762322e-06, "loss": 0.6156, "step": 9003 }, { "epoch": 0.7672773753728164, "grad_norm": 17.875, "learning_rate": 1.7064975094888202e-06, "loss": 0.7216, "step": 9004 }, { "epoch": 0.7673625905411163, "grad_norm": 15.3125, "learning_rate": 1.7063410760520392e-06, "loss": 0.8245, "step": 9005 }, { "epoch": 0.7674478057094163, "grad_norm": 10.875, "learning_rate": 1.706184634368716e-06, "loss": 0.3248, "step": 9006 }, { "epoch": 0.7675330208777162, "grad_norm": 12.75, "learning_rate": 1.706028184441678e-06, "loss": 0.4611, "step": 9007 }, { "epoch": 0.7676182360460162, "grad_norm": 13.3125, "learning_rate": 1.7058717262737517e-06, "loss": 0.2809, "step": 9008 }, { "epoch": 0.7677034512143162, "grad_norm": 24.125, "learning_rate": 1.7057152598677651e-06, "loss": 0.917, "step": 9009 }, { "epoch": 0.7677886663826161, "grad_norm": 14.75, "learning_rate": 1.7055587852265457e-06, "loss": 0.7854, "step": 9010 }, { "epoch": 0.7678738815509161, "grad_norm": 15.3125, "learning_rate": 1.7054023023529208e-06, "loss": 0.875, "step": 9011 }, { "epoch": 0.767959096719216, "grad_norm": 18.75, "learning_rate": 1.705245811249718e-06, "loss": 0.7601, "step": 9012 }, { "epoch": 0.768044311887516, "grad_norm": 13.6875, "learning_rate": 1.7050893119197663e-06, "loss": 0.4686, "step": 9013 }, { "epoch": 0.7681295270558159, "grad_norm": 13.125, "learning_rate": 1.7049328043658926e-06, "loss": 0.5051, "step": 9014 }, { "epoch": 0.7682147422241159, "grad_norm": 14.625, "learning_rate": 1.704776288590926e-06, "loss": 0.76, "step": 9015 }, { "epoch": 0.7682999573924159, "grad_norm": 12.5, "learning_rate": 1.7046197645976943e-06, "loss": 0.3944, "step": 9016 }, { "epoch": 0.7683851725607158, "grad_norm": 14.8125, "learning_rate": 1.7044632323890261e-06, "loss": 0.8577, "step": 9017 }, { "epoch": 0.7684703877290158, "grad_norm": 23.75, "learning_rate": 1.7043066919677503e-06, "loss": 0.5575, "step": 9018 }, { "epoch": 0.7685556028973157, "grad_norm": 13.75, "learning_rate": 1.7041501433366965e-06, "loss": 0.6031, "step": 9019 }, { "epoch": 0.7686408180656157, "grad_norm": 13.25, "learning_rate": 1.7039935864986923e-06, "loss": 0.6532, "step": 9020 }, { "epoch": 0.7687260332339156, "grad_norm": 12.6875, "learning_rate": 1.7038370214565675e-06, "loss": 0.6905, "step": 9021 }, { "epoch": 0.7688112484022156, "grad_norm": 14.1875, "learning_rate": 1.703680448213151e-06, "loss": 0.698, "step": 9022 }, { "epoch": 0.7688964635705156, "grad_norm": 14.8125, "learning_rate": 1.7035238667712733e-06, "loss": 0.548, "step": 9023 }, { "epoch": 0.7689816787388155, "grad_norm": 11.8125, "learning_rate": 1.703367277133763e-06, "loss": 0.5977, "step": 9024 }, { "epoch": 0.7690668939071155, "grad_norm": 14.375, "learning_rate": 1.7032106793034501e-06, "loss": 0.5938, "step": 9025 }, { "epoch": 0.7691521090754154, "grad_norm": 12.625, "learning_rate": 1.7030540732831644e-06, "loss": 0.4421, "step": 9026 }, { "epoch": 0.7692373242437154, "grad_norm": 14.375, "learning_rate": 1.702897459075736e-06, "loss": 0.9577, "step": 9027 }, { "epoch": 0.7693225394120153, "grad_norm": 15.3125, "learning_rate": 1.7027408366839953e-06, "loss": 0.443, "step": 9028 }, { "epoch": 0.7694077545803153, "grad_norm": 13.1875, "learning_rate": 1.7025842061107723e-06, "loss": 0.8565, "step": 9029 }, { "epoch": 0.7694929697486153, "grad_norm": 15.125, "learning_rate": 1.7024275673588975e-06, "loss": 0.6184, "step": 9030 }, { "epoch": 0.7695781849169152, "grad_norm": 15.0, "learning_rate": 1.7022709204312018e-06, "loss": 1.1135, "step": 9031 }, { "epoch": 0.7696634000852152, "grad_norm": 14.125, "learning_rate": 1.7021142653305156e-06, "loss": 0.5914, "step": 9032 }, { "epoch": 0.7697486152535151, "grad_norm": 16.25, "learning_rate": 1.70195760205967e-06, "loss": 0.6726, "step": 9033 }, { "epoch": 0.7698338304218151, "grad_norm": 15.75, "learning_rate": 1.7018009306214962e-06, "loss": 0.8695, "step": 9034 }, { "epoch": 0.769919045590115, "grad_norm": 13.75, "learning_rate": 1.7016442510188251e-06, "loss": 0.4837, "step": 9035 }, { "epoch": 0.770004260758415, "grad_norm": 9.875, "learning_rate": 1.7014875632544885e-06, "loss": 0.1759, "step": 9036 }, { "epoch": 0.770089475926715, "grad_norm": 13.75, "learning_rate": 1.7013308673313173e-06, "loss": 0.7079, "step": 9037 }, { "epoch": 0.7701746910950149, "grad_norm": 17.625, "learning_rate": 1.7011741632521436e-06, "loss": 1.1155, "step": 9038 }, { "epoch": 0.7702599062633149, "grad_norm": 17.0, "learning_rate": 1.7010174510197992e-06, "loss": 0.9273, "step": 9039 }, { "epoch": 0.7703451214316148, "grad_norm": 14.25, "learning_rate": 1.7008607306371155e-06, "loss": 0.7976, "step": 9040 }, { "epoch": 0.7704303365999148, "grad_norm": 14.0, "learning_rate": 1.7007040021069254e-06, "loss": 0.5912, "step": 9041 }, { "epoch": 0.7705155517682147, "grad_norm": 16.75, "learning_rate": 1.700547265432061e-06, "loss": 0.5672, "step": 9042 }, { "epoch": 0.7706007669365147, "grad_norm": 11.625, "learning_rate": 1.700390520615354e-06, "loss": 0.5232, "step": 9043 }, { "epoch": 0.7706859821048146, "grad_norm": 19.0, "learning_rate": 1.7002337676596373e-06, "loss": 1.1698, "step": 9044 }, { "epoch": 0.7707711972731146, "grad_norm": 18.75, "learning_rate": 1.700077006567744e-06, "loss": 0.8062, "step": 9045 }, { "epoch": 0.7708564124414146, "grad_norm": 13.625, "learning_rate": 1.6999202373425063e-06, "loss": 0.5261, "step": 9046 }, { "epoch": 0.7709416276097145, "grad_norm": 11.6875, "learning_rate": 1.699763459986758e-06, "loss": 0.5351, "step": 9047 }, { "epoch": 0.7710268427780145, "grad_norm": 13.6875, "learning_rate": 1.6996066745033312e-06, "loss": 0.7916, "step": 9048 }, { "epoch": 0.7711120579463144, "grad_norm": 16.0, "learning_rate": 1.69944988089506e-06, "loss": 0.8742, "step": 9049 }, { "epoch": 0.7711972731146144, "grad_norm": 11.3125, "learning_rate": 1.699293079164777e-06, "loss": 0.4032, "step": 9050 }, { "epoch": 0.7712824882829143, "grad_norm": 16.0, "learning_rate": 1.699136269315317e-06, "loss": 0.7968, "step": 9051 }, { "epoch": 0.7713677034512143, "grad_norm": 11.0, "learning_rate": 1.6989794513495125e-06, "loss": 0.4347, "step": 9052 }, { "epoch": 0.7714529186195143, "grad_norm": 11.875, "learning_rate": 1.6988226252701982e-06, "loss": 0.5018, "step": 9053 }, { "epoch": 0.7715381337878142, "grad_norm": 10.5625, "learning_rate": 1.698665791080207e-06, "loss": 0.3569, "step": 9054 }, { "epoch": 0.7716233489561142, "grad_norm": 17.75, "learning_rate": 1.6985089487823745e-06, "loss": 0.6398, "step": 9055 }, { "epoch": 0.7717085641244141, "grad_norm": 16.5, "learning_rate": 1.698352098379534e-06, "loss": 0.7814, "step": 9056 }, { "epoch": 0.7717937792927141, "grad_norm": 26.0, "learning_rate": 1.6981952398745205e-06, "loss": 1.3219, "step": 9057 }, { "epoch": 0.771878994461014, "grad_norm": 17.875, "learning_rate": 1.6980383732701678e-06, "loss": 0.7834, "step": 9058 }, { "epoch": 0.771964209629314, "grad_norm": 17.625, "learning_rate": 1.6978814985693115e-06, "loss": 0.7673, "step": 9059 }, { "epoch": 0.772049424797614, "grad_norm": 24.5, "learning_rate": 1.6977246157747862e-06, "loss": 0.9748, "step": 9060 }, { "epoch": 0.7721346399659139, "grad_norm": 9.9375, "learning_rate": 1.6975677248894266e-06, "loss": 0.4087, "step": 9061 }, { "epoch": 0.7722198551342139, "grad_norm": 14.3125, "learning_rate": 1.6974108259160683e-06, "loss": 0.6819, "step": 9062 }, { "epoch": 0.7723050703025138, "grad_norm": 27.625, "learning_rate": 1.6972539188575465e-06, "loss": 0.9034, "step": 9063 }, { "epoch": 0.7723902854708138, "grad_norm": 23.25, "learning_rate": 1.6970970037166967e-06, "loss": 0.8258, "step": 9064 }, { "epoch": 0.7724755006391137, "grad_norm": 18.875, "learning_rate": 1.6969400804963542e-06, "loss": 0.7505, "step": 9065 }, { "epoch": 0.7725607158074137, "grad_norm": 11.4375, "learning_rate": 1.6967831491993553e-06, "loss": 0.5268, "step": 9066 }, { "epoch": 0.7726459309757137, "grad_norm": 11.8125, "learning_rate": 1.6966262098285355e-06, "loss": 0.3593, "step": 9067 }, { "epoch": 0.7727311461440136, "grad_norm": 16.875, "learning_rate": 1.696469262386731e-06, "loss": 0.776, "step": 9068 }, { "epoch": 0.7728163613123136, "grad_norm": 17.125, "learning_rate": 1.696312306876778e-06, "loss": 1.0498, "step": 9069 }, { "epoch": 0.7729015764806135, "grad_norm": 14.4375, "learning_rate": 1.6961553433015127e-06, "loss": 0.6401, "step": 9070 }, { "epoch": 0.7729867916489135, "grad_norm": 11.125, "learning_rate": 1.6959983716637717e-06, "loss": 0.5161, "step": 9071 }, { "epoch": 0.7730720068172134, "grad_norm": 13.125, "learning_rate": 1.6958413919663918e-06, "loss": 0.4393, "step": 9072 }, { "epoch": 0.7731572219855134, "grad_norm": 13.75, "learning_rate": 1.6956844042122094e-06, "loss": 0.3946, "step": 9073 }, { "epoch": 0.7732424371538134, "grad_norm": 19.25, "learning_rate": 1.6955274084040618e-06, "loss": 1.0158, "step": 9074 }, { "epoch": 0.7733276523221133, "grad_norm": 15.0, "learning_rate": 1.695370404544786e-06, "loss": 0.787, "step": 9075 }, { "epoch": 0.7734128674904133, "grad_norm": 17.0, "learning_rate": 1.6952133926372188e-06, "loss": 0.8103, "step": 9076 }, { "epoch": 0.7734980826587132, "grad_norm": 30.0, "learning_rate": 1.695056372684198e-06, "loss": 0.9871, "step": 9077 }, { "epoch": 0.7735832978270132, "grad_norm": 18.875, "learning_rate": 1.694899344688561e-06, "loss": 1.0234, "step": 9078 }, { "epoch": 0.7736685129953131, "grad_norm": 29.875, "learning_rate": 1.6947423086531458e-06, "loss": 0.5055, "step": 9079 }, { "epoch": 0.7737537281636131, "grad_norm": 16.25, "learning_rate": 1.6945852645807895e-06, "loss": 0.9303, "step": 9080 }, { "epoch": 0.773838943331913, "grad_norm": 24.625, "learning_rate": 1.6944282124743305e-06, "loss": 0.8925, "step": 9081 }, { "epoch": 0.773924158500213, "grad_norm": 24.0, "learning_rate": 1.6942711523366069e-06, "loss": 0.8611, "step": 9082 }, { "epoch": 0.774009373668513, "grad_norm": 16.75, "learning_rate": 1.694114084170457e-06, "loss": 0.8173, "step": 9083 }, { "epoch": 0.7740945888368129, "grad_norm": 16.25, "learning_rate": 1.6939570079787184e-06, "loss": 0.9825, "step": 9084 }, { "epoch": 0.7741798040051129, "grad_norm": 14.75, "learning_rate": 1.6937999237642308e-06, "loss": 0.8613, "step": 9085 }, { "epoch": 0.7742650191734128, "grad_norm": 10.0625, "learning_rate": 1.693642831529832e-06, "loss": 0.3232, "step": 9086 }, { "epoch": 0.7743502343417128, "grad_norm": 20.375, "learning_rate": 1.6934857312783613e-06, "loss": 1.2607, "step": 9087 }, { "epoch": 0.7744354495100128, "grad_norm": 14.125, "learning_rate": 1.6933286230126575e-06, "loss": 0.5303, "step": 9088 }, { "epoch": 0.7745206646783127, "grad_norm": 12.9375, "learning_rate": 1.6931715067355595e-06, "loss": 0.4752, "step": 9089 }, { "epoch": 0.7746058798466127, "grad_norm": 12.8125, "learning_rate": 1.6930143824499068e-06, "loss": 0.7418, "step": 9090 }, { "epoch": 0.7746910950149126, "grad_norm": 17.5, "learning_rate": 1.6928572501585386e-06, "loss": 0.8885, "step": 9091 }, { "epoch": 0.7747763101832126, "grad_norm": 13.75, "learning_rate": 1.6927001098642948e-06, "loss": 0.497, "step": 9092 }, { "epoch": 0.7748615253515125, "grad_norm": 21.25, "learning_rate": 1.6925429615700146e-06, "loss": 1.0675, "step": 9093 }, { "epoch": 0.7749467405198125, "grad_norm": 27.125, "learning_rate": 1.6923858052785383e-06, "loss": 1.0814, "step": 9094 }, { "epoch": 0.7750319556881125, "grad_norm": 13.3125, "learning_rate": 1.6922286409927052e-06, "loss": 0.6301, "step": 9095 }, { "epoch": 0.7751171708564124, "grad_norm": 13.625, "learning_rate": 1.6920714687153567e-06, "loss": 0.6973, "step": 9096 }, { "epoch": 0.7752023860247124, "grad_norm": 17.625, "learning_rate": 1.6919142884493315e-06, "loss": 0.8307, "step": 9097 }, { "epoch": 0.7752876011930123, "grad_norm": 15.625, "learning_rate": 1.6917571001974712e-06, "loss": 0.5827, "step": 9098 }, { "epoch": 0.7753728163613123, "grad_norm": 18.5, "learning_rate": 1.6915999039626153e-06, "loss": 0.8677, "step": 9099 }, { "epoch": 0.7754580315296122, "grad_norm": 14.125, "learning_rate": 1.6914426997476057e-06, "loss": 0.968, "step": 9100 }, { "epoch": 0.7755432466979122, "grad_norm": 16.625, "learning_rate": 1.6912854875552822e-06, "loss": 0.6724, "step": 9101 }, { "epoch": 0.7756284618662121, "grad_norm": 11.5, "learning_rate": 1.6911282673884863e-06, "loss": 0.521, "step": 9102 }, { "epoch": 0.7757136770345121, "grad_norm": 14.9375, "learning_rate": 1.6909710392500588e-06, "loss": 0.7789, "step": 9103 }, { "epoch": 0.7757988922028121, "grad_norm": 10.6875, "learning_rate": 1.690813803142842e-06, "loss": 0.3545, "step": 9104 }, { "epoch": 0.775884107371112, "grad_norm": 16.625, "learning_rate": 1.6906565590696755e-06, "loss": 0.9103, "step": 9105 }, { "epoch": 0.775969322539412, "grad_norm": 14.5625, "learning_rate": 1.6904993070334024e-06, "loss": 0.5881, "step": 9106 }, { "epoch": 0.7760545377077119, "grad_norm": 15.875, "learning_rate": 1.6903420470368636e-06, "loss": 0.8179, "step": 9107 }, { "epoch": 0.7761397528760119, "grad_norm": 9.5625, "learning_rate": 1.6901847790829015e-06, "loss": 0.2609, "step": 9108 }, { "epoch": 0.7762249680443118, "grad_norm": 11.3125, "learning_rate": 1.690027503174358e-06, "loss": 0.4712, "step": 9109 }, { "epoch": 0.7763101832126118, "grad_norm": 14.125, "learning_rate": 1.6898702193140743e-06, "loss": 0.7857, "step": 9110 }, { "epoch": 0.7763953983809118, "grad_norm": 19.0, "learning_rate": 1.689712927504894e-06, "loss": 0.7153, "step": 9111 }, { "epoch": 0.7764806135492117, "grad_norm": 14.5, "learning_rate": 1.6895556277496588e-06, "loss": 0.6683, "step": 9112 }, { "epoch": 0.7765658287175117, "grad_norm": 17.5, "learning_rate": 1.6893983200512115e-06, "loss": 1.2218, "step": 9113 }, { "epoch": 0.7766510438858116, "grad_norm": 14.375, "learning_rate": 1.6892410044123942e-06, "loss": 0.3115, "step": 9114 }, { "epoch": 0.7767362590541116, "grad_norm": 18.125, "learning_rate": 1.6890836808360509e-06, "loss": 0.652, "step": 9115 }, { "epoch": 0.7768214742224115, "grad_norm": 12.4375, "learning_rate": 1.6889263493250235e-06, "loss": 0.5894, "step": 9116 }, { "epoch": 0.7769066893907115, "grad_norm": 13.75, "learning_rate": 1.688769009882156e-06, "loss": 0.7257, "step": 9117 }, { "epoch": 0.7769919045590115, "grad_norm": 14.0625, "learning_rate": 1.6886116625102907e-06, "loss": 1.2324, "step": 9118 }, { "epoch": 0.7770771197273114, "grad_norm": 14.25, "learning_rate": 1.6884543072122721e-06, "loss": 0.8572, "step": 9119 }, { "epoch": 0.7771623348956114, "grad_norm": 13.375, "learning_rate": 1.6882969439909432e-06, "loss": 0.7263, "step": 9120 }, { "epoch": 0.7772475500639113, "grad_norm": 15.5, "learning_rate": 1.688139572849148e-06, "loss": 0.6301, "step": 9121 }, { "epoch": 0.7773327652322113, "grad_norm": 17.0, "learning_rate": 1.6879821937897295e-06, "loss": 0.6442, "step": 9122 }, { "epoch": 0.7774179804005112, "grad_norm": 12.9375, "learning_rate": 1.6878248068155328e-06, "loss": 0.4774, "step": 9123 }, { "epoch": 0.7775031955688112, "grad_norm": 14.4375, "learning_rate": 1.6876674119294014e-06, "loss": 0.9256, "step": 9124 }, { "epoch": 0.7775884107371112, "grad_norm": 17.875, "learning_rate": 1.6875100091341795e-06, "loss": 0.9333, "step": 9125 }, { "epoch": 0.7776736259054111, "grad_norm": 14.9375, "learning_rate": 1.6873525984327122e-06, "loss": 0.6986, "step": 9126 }, { "epoch": 0.7777588410737111, "grad_norm": 12.25, "learning_rate": 1.6871951798278435e-06, "loss": 0.7764, "step": 9127 }, { "epoch": 0.777844056242011, "grad_norm": 14.0625, "learning_rate": 1.6870377533224186e-06, "loss": 0.6542, "step": 9128 }, { "epoch": 0.777929271410311, "grad_norm": 11.25, "learning_rate": 1.6868803189192815e-06, "loss": 0.4523, "step": 9129 }, { "epoch": 0.7780144865786109, "grad_norm": 18.875, "learning_rate": 1.686722876621278e-06, "loss": 0.6665, "step": 9130 }, { "epoch": 0.7780997017469109, "grad_norm": 14.4375, "learning_rate": 1.6865654264312529e-06, "loss": 0.4918, "step": 9131 }, { "epoch": 0.7781849169152109, "grad_norm": 12.875, "learning_rate": 1.6864079683520518e-06, "loss": 0.5856, "step": 9132 }, { "epoch": 0.7782701320835108, "grad_norm": 15.3125, "learning_rate": 1.6862505023865194e-06, "loss": 1.0167, "step": 9133 }, { "epoch": 0.7783553472518108, "grad_norm": 10.875, "learning_rate": 1.686093028537502e-06, "loss": 0.3041, "step": 9134 }, { "epoch": 0.7784405624201107, "grad_norm": 11.75, "learning_rate": 1.6859355468078448e-06, "loss": 0.5819, "step": 9135 }, { "epoch": 0.7785257775884107, "grad_norm": 11.625, "learning_rate": 1.6857780572003946e-06, "loss": 0.4546, "step": 9136 }, { "epoch": 0.7786109927567106, "grad_norm": 15.6875, "learning_rate": 1.6856205597179959e-06, "loss": 0.8451, "step": 9137 }, { "epoch": 0.7786962079250106, "grad_norm": 14.3125, "learning_rate": 1.685463054363496e-06, "loss": 0.8406, "step": 9138 }, { "epoch": 0.7787814230933106, "grad_norm": 16.875, "learning_rate": 1.6853055411397406e-06, "loss": 1.0408, "step": 9139 }, { "epoch": 0.7788666382616106, "grad_norm": 15.875, "learning_rate": 1.6851480200495765e-06, "loss": 0.721, "step": 9140 }, { "epoch": 0.7789518534299106, "grad_norm": 14.375, "learning_rate": 1.6849904910958503e-06, "loss": 0.603, "step": 9141 }, { "epoch": 0.7790370685982105, "grad_norm": 11.6875, "learning_rate": 1.6848329542814081e-06, "loss": 0.4813, "step": 9142 }, { "epoch": 0.7791222837665105, "grad_norm": 13.5, "learning_rate": 1.6846754096090974e-06, "loss": 0.7694, "step": 9143 }, { "epoch": 0.7792074989348104, "grad_norm": 16.0, "learning_rate": 1.684517857081765e-06, "loss": 1.0108, "step": 9144 }, { "epoch": 0.7792927141031104, "grad_norm": 11.125, "learning_rate": 1.684360296702258e-06, "loss": 0.5011, "step": 9145 }, { "epoch": 0.7793779292714104, "grad_norm": 16.25, "learning_rate": 1.6842027284734233e-06, "loss": 0.801, "step": 9146 }, { "epoch": 0.7794631444397103, "grad_norm": 18.25, "learning_rate": 1.6840451523981089e-06, "loss": 0.6285, "step": 9147 }, { "epoch": 0.7795483596080103, "grad_norm": 12.9375, "learning_rate": 1.6838875684791622e-06, "loss": 0.5928, "step": 9148 }, { "epoch": 0.7796335747763102, "grad_norm": 14.5625, "learning_rate": 1.683729976719431e-06, "loss": 0.7728, "step": 9149 }, { "epoch": 0.7797187899446102, "grad_norm": 14.5, "learning_rate": 1.6835723771217624e-06, "loss": 0.4603, "step": 9150 }, { "epoch": 0.7798040051129101, "grad_norm": 16.0, "learning_rate": 1.6834147696890056e-06, "loss": 0.6017, "step": 9151 }, { "epoch": 0.7798892202812101, "grad_norm": 16.25, "learning_rate": 1.6832571544240076e-06, "loss": 1.0502, "step": 9152 }, { "epoch": 0.7799744354495101, "grad_norm": 17.25, "learning_rate": 1.6830995313296175e-06, "loss": 0.7913, "step": 9153 }, { "epoch": 0.78005965061781, "grad_norm": 12.9375, "learning_rate": 1.682941900408683e-06, "loss": 0.383, "step": 9154 }, { "epoch": 0.78014486578611, "grad_norm": 14.8125, "learning_rate": 1.6827842616640533e-06, "loss": 0.6808, "step": 9155 }, { "epoch": 0.7802300809544099, "grad_norm": 17.75, "learning_rate": 1.6826266150985765e-06, "loss": 0.631, "step": 9156 }, { "epoch": 0.7803152961227099, "grad_norm": 12.5625, "learning_rate": 1.6824689607151023e-06, "loss": 0.8862, "step": 9157 }, { "epoch": 0.7804005112910098, "grad_norm": 15.375, "learning_rate": 1.6823112985164786e-06, "loss": 0.7749, "step": 9158 }, { "epoch": 0.7804857264593098, "grad_norm": 16.75, "learning_rate": 1.682153628505555e-06, "loss": 0.9177, "step": 9159 }, { "epoch": 0.7805709416276098, "grad_norm": 16.125, "learning_rate": 1.6819959506851807e-06, "loss": 0.5646, "step": 9160 }, { "epoch": 0.7806561567959097, "grad_norm": 15.25, "learning_rate": 1.6818382650582055e-06, "loss": 0.9146, "step": 9161 }, { "epoch": 0.7807413719642097, "grad_norm": 13.75, "learning_rate": 1.6816805716274787e-06, "loss": 0.6513, "step": 9162 }, { "epoch": 0.7808265871325096, "grad_norm": 11.125, "learning_rate": 1.681522870395849e-06, "loss": 0.5193, "step": 9163 }, { "epoch": 0.7809118023008096, "grad_norm": 13.8125, "learning_rate": 1.681365161366168e-06, "loss": 0.7052, "step": 9164 }, { "epoch": 0.7809970174691095, "grad_norm": 14.0, "learning_rate": 1.6812074445412845e-06, "loss": 0.4651, "step": 9165 }, { "epoch": 0.7810822326374095, "grad_norm": 15.3125, "learning_rate": 1.6810497199240491e-06, "loss": 0.5505, "step": 9166 }, { "epoch": 0.7811674478057095, "grad_norm": 16.25, "learning_rate": 1.6808919875173113e-06, "loss": 0.9399, "step": 9167 }, { "epoch": 0.7812526629740094, "grad_norm": 19.875, "learning_rate": 1.6807342473239224e-06, "loss": 0.9514, "step": 9168 }, { "epoch": 0.7813378781423094, "grad_norm": 15.0625, "learning_rate": 1.6805764993467326e-06, "loss": 0.6986, "step": 9169 }, { "epoch": 0.7814230933106093, "grad_norm": 13.8125, "learning_rate": 1.6804187435885921e-06, "loss": 0.6275, "step": 9170 }, { "epoch": 0.7815083084789093, "grad_norm": 14.375, "learning_rate": 1.6802609800523523e-06, "loss": 0.8237, "step": 9171 }, { "epoch": 0.7815935236472092, "grad_norm": 19.0, "learning_rate": 1.680103208740864e-06, "loss": 0.9538, "step": 9172 }, { "epoch": 0.7816787388155092, "grad_norm": 14.125, "learning_rate": 1.6799454296569778e-06, "loss": 0.7257, "step": 9173 }, { "epoch": 0.7817639539838092, "grad_norm": 14.1875, "learning_rate": 1.679787642803546e-06, "loss": 0.8362, "step": 9174 }, { "epoch": 0.7818491691521091, "grad_norm": 14.25, "learning_rate": 1.6796298481834187e-06, "loss": 0.9333, "step": 9175 }, { "epoch": 0.7819343843204091, "grad_norm": 15.875, "learning_rate": 1.6794720457994484e-06, "loss": 0.8981, "step": 9176 }, { "epoch": 0.782019599488709, "grad_norm": 12.0625, "learning_rate": 1.679314235654486e-06, "loss": 0.5872, "step": 9177 }, { "epoch": 0.782104814657009, "grad_norm": 15.125, "learning_rate": 1.6791564177513838e-06, "loss": 0.7796, "step": 9178 }, { "epoch": 0.7821900298253089, "grad_norm": 14.8125, "learning_rate": 1.6789985920929937e-06, "loss": 1.0153, "step": 9179 }, { "epoch": 0.7822752449936089, "grad_norm": 13.9375, "learning_rate": 1.6788407586821675e-06, "loss": 0.58, "step": 9180 }, { "epoch": 0.7823604601619089, "grad_norm": 11.8125, "learning_rate": 1.6786829175217582e-06, "loss": 0.4868, "step": 9181 }, { "epoch": 0.7824456753302088, "grad_norm": 18.5, "learning_rate": 1.6785250686146164e-06, "loss": 0.8926, "step": 9182 }, { "epoch": 0.7825308904985088, "grad_norm": 13.8125, "learning_rate": 1.6783672119635966e-06, "loss": 0.6253, "step": 9183 }, { "epoch": 0.7826161056668087, "grad_norm": 16.25, "learning_rate": 1.6782093475715499e-06, "loss": 0.8603, "step": 9184 }, { "epoch": 0.7827013208351087, "grad_norm": 16.25, "learning_rate": 1.6780514754413306e-06, "loss": 0.6022, "step": 9185 }, { "epoch": 0.7827865360034086, "grad_norm": 16.375, "learning_rate": 1.6778935955757897e-06, "loss": 0.7856, "step": 9186 }, { "epoch": 0.7828717511717086, "grad_norm": 13.1875, "learning_rate": 1.6777357079777817e-06, "loss": 0.552, "step": 9187 }, { "epoch": 0.7829569663400086, "grad_norm": 17.375, "learning_rate": 1.677577812650159e-06, "loss": 0.562, "step": 9188 }, { "epoch": 0.7830421815083085, "grad_norm": 13.0, "learning_rate": 1.677419909595776e-06, "loss": 0.5429, "step": 9189 }, { "epoch": 0.7831273966766085, "grad_norm": 14.5625, "learning_rate": 1.6772619988174846e-06, "loss": 0.7473, "step": 9190 }, { "epoch": 0.7832126118449084, "grad_norm": 16.125, "learning_rate": 1.6771040803181398e-06, "loss": 0.8336, "step": 9191 }, { "epoch": 0.7832978270132084, "grad_norm": 17.5, "learning_rate": 1.6769461541005944e-06, "loss": 0.8144, "step": 9192 }, { "epoch": 0.7833830421815083, "grad_norm": 13.4375, "learning_rate": 1.6767882201677028e-06, "loss": 0.5138, "step": 9193 }, { "epoch": 0.7834682573498083, "grad_norm": 14.75, "learning_rate": 1.676630278522319e-06, "loss": 0.8737, "step": 9194 }, { "epoch": 0.7835534725181083, "grad_norm": 14.8125, "learning_rate": 1.676472329167297e-06, "loss": 0.8462, "step": 9195 }, { "epoch": 0.7836386876864082, "grad_norm": 14.25, "learning_rate": 1.6763143721054912e-06, "loss": 0.8616, "step": 9196 }, { "epoch": 0.7837239028547082, "grad_norm": 12.3125, "learning_rate": 1.676156407339756e-06, "loss": 0.6099, "step": 9197 }, { "epoch": 0.7838091180230081, "grad_norm": 14.1875, "learning_rate": 1.6759984348729458e-06, "loss": 0.9848, "step": 9198 }, { "epoch": 0.7838943331913081, "grad_norm": 17.625, "learning_rate": 1.6758404547079155e-06, "loss": 0.7159, "step": 9199 }, { "epoch": 0.783979548359608, "grad_norm": 14.875, "learning_rate": 1.6756824668475202e-06, "loss": 0.8403, "step": 9200 }, { "epoch": 0.784064763527908, "grad_norm": 13.0, "learning_rate": 1.6755244712946148e-06, "loss": 0.7129, "step": 9201 }, { "epoch": 0.784149978696208, "grad_norm": 15.125, "learning_rate": 1.675366468052054e-06, "loss": 0.5531, "step": 9202 }, { "epoch": 0.7842351938645079, "grad_norm": 12.75, "learning_rate": 1.6752084571226935e-06, "loss": 0.6888, "step": 9203 }, { "epoch": 0.7843204090328079, "grad_norm": 16.25, "learning_rate": 1.6750504385093885e-06, "loss": 0.8672, "step": 9204 }, { "epoch": 0.7844056242011078, "grad_norm": 12.0, "learning_rate": 1.6748924122149948e-06, "loss": 0.5386, "step": 9205 }, { "epoch": 0.7844908393694078, "grad_norm": 24.125, "learning_rate": 1.6747343782423681e-06, "loss": 0.9947, "step": 9206 }, { "epoch": 0.7845760545377077, "grad_norm": 18.0, "learning_rate": 1.6745763365943634e-06, "loss": 0.7725, "step": 9207 }, { "epoch": 0.7846612697060077, "grad_norm": 12.0625, "learning_rate": 1.6744182872738379e-06, "loss": 0.3866, "step": 9208 }, { "epoch": 0.7847464848743076, "grad_norm": 13.8125, "learning_rate": 1.6742602302836471e-06, "loss": 0.7012, "step": 9209 }, { "epoch": 0.7848317000426076, "grad_norm": 60.25, "learning_rate": 1.6741021656266474e-06, "loss": 0.9639, "step": 9210 }, { "epoch": 0.7849169152109076, "grad_norm": 17.0, "learning_rate": 1.673944093305695e-06, "loss": 0.9644, "step": 9211 }, { "epoch": 0.7850021303792075, "grad_norm": 18.5, "learning_rate": 1.6737860133236466e-06, "loss": 0.8649, "step": 9212 }, { "epoch": 0.7850873455475075, "grad_norm": 17.5, "learning_rate": 1.6736279256833588e-06, "loss": 0.6805, "step": 9213 }, { "epoch": 0.7851725607158074, "grad_norm": 15.625, "learning_rate": 1.673469830387688e-06, "loss": 1.0532, "step": 9214 }, { "epoch": 0.7852577758841074, "grad_norm": 31.125, "learning_rate": 1.673311727439492e-06, "loss": 0.4484, "step": 9215 }, { "epoch": 0.7853429910524073, "grad_norm": 14.125, "learning_rate": 1.6731536168416268e-06, "loss": 0.497, "step": 9216 }, { "epoch": 0.7854282062207073, "grad_norm": 20.0, "learning_rate": 1.6729954985969512e-06, "loss": 1.251, "step": 9217 }, { "epoch": 0.7855134213890073, "grad_norm": 14.125, "learning_rate": 1.6728373727083209e-06, "loss": 0.6152, "step": 9218 }, { "epoch": 0.7855986365573072, "grad_norm": 11.5, "learning_rate": 1.6726792391785942e-06, "loss": 0.4704, "step": 9219 }, { "epoch": 0.7856838517256072, "grad_norm": 16.75, "learning_rate": 1.6725210980106285e-06, "loss": 0.8961, "step": 9220 }, { "epoch": 0.7857690668939071, "grad_norm": 15.0, "learning_rate": 1.672362949207282e-06, "loss": 0.497, "step": 9221 }, { "epoch": 0.7858542820622071, "grad_norm": 16.75, "learning_rate": 1.6722047927714123e-06, "loss": 1.0312, "step": 9222 }, { "epoch": 0.785939497230507, "grad_norm": 21.375, "learning_rate": 1.672046628705877e-06, "loss": 1.1299, "step": 9223 }, { "epoch": 0.786024712398807, "grad_norm": 21.375, "learning_rate": 1.671888457013535e-06, "loss": 0.9332, "step": 9224 }, { "epoch": 0.786109927567107, "grad_norm": 13.375, "learning_rate": 1.6717302776972443e-06, "loss": 0.4465, "step": 9225 }, { "epoch": 0.7861951427354069, "grad_norm": 14.625, "learning_rate": 1.6715720907598635e-06, "loss": 0.6631, "step": 9226 }, { "epoch": 0.7862803579037069, "grad_norm": 13.25, "learning_rate": 1.6714138962042512e-06, "loss": 0.6299, "step": 9227 }, { "epoch": 0.7863655730720068, "grad_norm": 21.125, "learning_rate": 1.6712556940332655e-06, "loss": 0.7696, "step": 9228 }, { "epoch": 0.7864507882403068, "grad_norm": 14.9375, "learning_rate": 1.6710974842497663e-06, "loss": 0.9325, "step": 9229 }, { "epoch": 0.7865360034086067, "grad_norm": 11.8125, "learning_rate": 1.670939266856612e-06, "loss": 0.4064, "step": 9230 }, { "epoch": 0.7866212185769067, "grad_norm": 16.625, "learning_rate": 1.6707810418566617e-06, "loss": 0.6951, "step": 9231 }, { "epoch": 0.7867064337452067, "grad_norm": 28.875, "learning_rate": 1.670622809252775e-06, "loss": 1.1272, "step": 9232 }, { "epoch": 0.7867916489135066, "grad_norm": 13.0625, "learning_rate": 1.6704645690478108e-06, "loss": 0.8552, "step": 9233 }, { "epoch": 0.7868768640818066, "grad_norm": 15.4375, "learning_rate": 1.6703063212446298e-06, "loss": 0.8898, "step": 9234 }, { "epoch": 0.7869620792501065, "grad_norm": 11.8125, "learning_rate": 1.6701480658460903e-06, "loss": 0.4441, "step": 9235 }, { "epoch": 0.7870472944184065, "grad_norm": 15.125, "learning_rate": 1.6699898028550528e-06, "loss": 0.5961, "step": 9236 }, { "epoch": 0.7871325095867064, "grad_norm": 14.125, "learning_rate": 1.6698315322743774e-06, "loss": 0.6096, "step": 9237 }, { "epoch": 0.7872177247550064, "grad_norm": 13.5, "learning_rate": 1.6696732541069244e-06, "loss": 0.794, "step": 9238 }, { "epoch": 0.7873029399233064, "grad_norm": 15.1875, "learning_rate": 1.6695149683555528e-06, "loss": 0.8106, "step": 9239 }, { "epoch": 0.7873881550916063, "grad_norm": 12.5, "learning_rate": 1.6693566750231246e-06, "loss": 0.8082, "step": 9240 }, { "epoch": 0.7874733702599063, "grad_norm": 11.5, "learning_rate": 1.6691983741124992e-06, "loss": 0.5338, "step": 9241 }, { "epoch": 0.7875585854282062, "grad_norm": 16.0, "learning_rate": 1.6690400656265382e-06, "loss": 0.7448, "step": 9242 }, { "epoch": 0.7876438005965062, "grad_norm": 20.5, "learning_rate": 1.6688817495681012e-06, "loss": 0.8265, "step": 9243 }, { "epoch": 0.7877290157648061, "grad_norm": 16.125, "learning_rate": 1.66872342594005e-06, "loss": 0.8309, "step": 9244 }, { "epoch": 0.7878142309331061, "grad_norm": 14.625, "learning_rate": 1.6685650947452452e-06, "loss": 0.7554, "step": 9245 }, { "epoch": 0.787899446101406, "grad_norm": 21.75, "learning_rate": 1.6684067559865486e-06, "loss": 1.0237, "step": 9246 }, { "epoch": 0.787984661269706, "grad_norm": 11.4375, "learning_rate": 1.668248409666821e-06, "loss": 0.4739, "step": 9247 }, { "epoch": 0.788069876438006, "grad_norm": 18.0, "learning_rate": 1.668090055788924e-06, "loss": 0.8306, "step": 9248 }, { "epoch": 0.7881550916063059, "grad_norm": 18.25, "learning_rate": 1.6679316943557195e-06, "loss": 0.801, "step": 9249 }, { "epoch": 0.7882403067746059, "grad_norm": 12.5625, "learning_rate": 1.667773325370069e-06, "loss": 0.8462, "step": 9250 }, { "epoch": 0.7883255219429058, "grad_norm": 35.5, "learning_rate": 1.6676149488348341e-06, "loss": 1.1651, "step": 9251 }, { "epoch": 0.7884107371112058, "grad_norm": 17.25, "learning_rate": 1.6674565647528773e-06, "loss": 0.862, "step": 9252 }, { "epoch": 0.7884959522795058, "grad_norm": 13.9375, "learning_rate": 1.6672981731270608e-06, "loss": 0.4958, "step": 9253 }, { "epoch": 0.7885811674478057, "grad_norm": 15.6875, "learning_rate": 1.6671397739602464e-06, "loss": 0.927, "step": 9254 }, { "epoch": 0.7886663826161057, "grad_norm": 29.0, "learning_rate": 1.666981367255297e-06, "loss": 0.8799, "step": 9255 }, { "epoch": 0.7887515977844056, "grad_norm": 16.625, "learning_rate": 1.6668229530150749e-06, "loss": 0.6446, "step": 9256 }, { "epoch": 0.7888368129527056, "grad_norm": 19.75, "learning_rate": 1.666664531242443e-06, "loss": 0.8819, "step": 9257 }, { "epoch": 0.7889220281210055, "grad_norm": 11.6875, "learning_rate": 1.666506101940264e-06, "loss": 0.2891, "step": 9258 }, { "epoch": 0.7890072432893055, "grad_norm": 10.1875, "learning_rate": 1.6663476651114008e-06, "loss": 0.4347, "step": 9259 }, { "epoch": 0.7890924584576055, "grad_norm": 2112.0, "learning_rate": 1.6661892207587165e-06, "loss": 0.4986, "step": 9260 }, { "epoch": 0.7891776736259054, "grad_norm": 14.4375, "learning_rate": 1.6660307688850747e-06, "loss": 0.7337, "step": 9261 }, { "epoch": 0.7892628887942054, "grad_norm": 16.5, "learning_rate": 1.6658723094933385e-06, "loss": 0.7602, "step": 9262 }, { "epoch": 0.7893481039625053, "grad_norm": 15.75, "learning_rate": 1.665713842586371e-06, "loss": 0.9238, "step": 9263 }, { "epoch": 0.7894333191308053, "grad_norm": 17.75, "learning_rate": 1.6655553681670373e-06, "loss": 0.8818, "step": 9264 }, { "epoch": 0.7895185342991052, "grad_norm": 12.875, "learning_rate": 1.6653968862381998e-06, "loss": 0.6838, "step": 9265 }, { "epoch": 0.7896037494674052, "grad_norm": 15.5625, "learning_rate": 1.6652383968027224e-06, "loss": 0.9095, "step": 9266 }, { "epoch": 0.7896889646357051, "grad_norm": 15.1875, "learning_rate": 1.66507989986347e-06, "loss": 1.1284, "step": 9267 }, { "epoch": 0.7897741798040051, "grad_norm": 25.875, "learning_rate": 1.6649213954233066e-06, "loss": 0.8976, "step": 9268 }, { "epoch": 0.7898593949723051, "grad_norm": 15.5, "learning_rate": 1.664762883485096e-06, "loss": 0.5452, "step": 9269 }, { "epoch": 0.789944610140605, "grad_norm": 11.8125, "learning_rate": 1.6646043640517034e-06, "loss": 0.4223, "step": 9270 }, { "epoch": 0.790029825308905, "grad_norm": 15.125, "learning_rate": 1.6644458371259926e-06, "loss": 0.6444, "step": 9271 }, { "epoch": 0.7901150404772049, "grad_norm": 19.75, "learning_rate": 1.6642873027108292e-06, "loss": 1.3016, "step": 9272 }, { "epoch": 0.7902002556455049, "grad_norm": 17.75, "learning_rate": 1.664128760809077e-06, "loss": 0.6226, "step": 9273 }, { "epoch": 0.7902854708138048, "grad_norm": 16.0, "learning_rate": 1.6639702114236024e-06, "loss": 0.859, "step": 9274 }, { "epoch": 0.7903706859821048, "grad_norm": 25.625, "learning_rate": 1.663811654557269e-06, "loss": 0.9279, "step": 9275 }, { "epoch": 0.7904559011504048, "grad_norm": 14.625, "learning_rate": 1.6636530902129436e-06, "loss": 0.8733, "step": 9276 }, { "epoch": 0.7905411163187047, "grad_norm": 15.5, "learning_rate": 1.6634945183934905e-06, "loss": 0.8732, "step": 9277 }, { "epoch": 0.7906263314870047, "grad_norm": 17.375, "learning_rate": 1.663335939101776e-06, "loss": 0.9286, "step": 9278 }, { "epoch": 0.7907115466553046, "grad_norm": 13.75, "learning_rate": 1.6631773523406652e-06, "loss": 0.6232, "step": 9279 }, { "epoch": 0.7907967618236046, "grad_norm": 15.625, "learning_rate": 1.6630187581130241e-06, "loss": 0.8644, "step": 9280 }, { "epoch": 0.7908819769919045, "grad_norm": 15.5, "learning_rate": 1.6628601564217187e-06, "loss": 0.5991, "step": 9281 }, { "epoch": 0.7909671921602045, "grad_norm": 14.125, "learning_rate": 1.6627015472696154e-06, "loss": 0.6777, "step": 9282 }, { "epoch": 0.7910524073285045, "grad_norm": 12.0, "learning_rate": 1.66254293065958e-06, "loss": 0.476, "step": 9283 }, { "epoch": 0.7911376224968044, "grad_norm": 17.125, "learning_rate": 1.6623843065944786e-06, "loss": 0.8345, "step": 9284 }, { "epoch": 0.7912228376651044, "grad_norm": 14.8125, "learning_rate": 1.6622256750771787e-06, "loss": 0.9714, "step": 9285 }, { "epoch": 0.7913080528334043, "grad_norm": 20.25, "learning_rate": 1.6620670361105458e-06, "loss": 0.796, "step": 9286 }, { "epoch": 0.7913932680017043, "grad_norm": 14.875, "learning_rate": 1.6619083896974477e-06, "loss": 0.8358, "step": 9287 }, { "epoch": 0.7914784831700042, "grad_norm": 15.375, "learning_rate": 1.6617497358407505e-06, "loss": 0.6862, "step": 9288 }, { "epoch": 0.7915636983383042, "grad_norm": 14.75, "learning_rate": 1.6615910745433218e-06, "loss": 0.7191, "step": 9289 }, { "epoch": 0.7916489135066042, "grad_norm": 13.25, "learning_rate": 1.6614324058080277e-06, "loss": 0.853, "step": 9290 }, { "epoch": 0.7917341286749041, "grad_norm": 13.625, "learning_rate": 1.6612737296377374e-06, "loss": 0.6146, "step": 9291 }, { "epoch": 0.7918193438432041, "grad_norm": 18.375, "learning_rate": 1.6611150460353164e-06, "loss": 0.9618, "step": 9292 }, { "epoch": 0.791904559011504, "grad_norm": 14.4375, "learning_rate": 1.6609563550036336e-06, "loss": 0.5117, "step": 9293 }, { "epoch": 0.791989774179804, "grad_norm": 11.25, "learning_rate": 1.6607976565455557e-06, "loss": 0.5, "step": 9294 }, { "epoch": 0.7920749893481039, "grad_norm": 10.125, "learning_rate": 1.6606389506639517e-06, "loss": 0.2575, "step": 9295 }, { "epoch": 0.7921602045164039, "grad_norm": 13.1875, "learning_rate": 1.6604802373616888e-06, "loss": 0.6682, "step": 9296 }, { "epoch": 0.7922454196847039, "grad_norm": 13.25, "learning_rate": 1.660321516641635e-06, "loss": 0.8294, "step": 9297 }, { "epoch": 0.7923306348530038, "grad_norm": 12.0625, "learning_rate": 1.6601627885066588e-06, "loss": 0.4739, "step": 9298 }, { "epoch": 0.7924158500213038, "grad_norm": 10.9375, "learning_rate": 1.6600040529596287e-06, "loss": 0.4657, "step": 9299 }, { "epoch": 0.7925010651896037, "grad_norm": 10.875, "learning_rate": 1.659845310003413e-06, "loss": 0.298, "step": 9300 }, { "epoch": 0.7925862803579037, "grad_norm": 12.625, "learning_rate": 1.6596865596408802e-06, "loss": 0.8404, "step": 9301 }, { "epoch": 0.7926714955262036, "grad_norm": 15.4375, "learning_rate": 1.6595278018748994e-06, "loss": 0.3716, "step": 9302 }, { "epoch": 0.7927567106945036, "grad_norm": 13.125, "learning_rate": 1.6593690367083398e-06, "loss": 0.6689, "step": 9303 }, { "epoch": 0.7928419258628036, "grad_norm": 14.1875, "learning_rate": 1.65921026414407e-06, "loss": 0.8589, "step": 9304 }, { "epoch": 0.7929271410311035, "grad_norm": 17.0, "learning_rate": 1.6590514841849587e-06, "loss": 0.8676, "step": 9305 }, { "epoch": 0.7930123561994035, "grad_norm": 13.25, "learning_rate": 1.658892696833876e-06, "loss": 0.5505, "step": 9306 }, { "epoch": 0.7930975713677034, "grad_norm": 19.75, "learning_rate": 1.6587339020936912e-06, "loss": 0.5966, "step": 9307 }, { "epoch": 0.7931827865360034, "grad_norm": 14.875, "learning_rate": 1.658575099967274e-06, "loss": 0.7274, "step": 9308 }, { "epoch": 0.7932680017043033, "grad_norm": 14.5, "learning_rate": 1.6584162904574932e-06, "loss": 0.8129, "step": 9309 }, { "epoch": 0.7933532168726033, "grad_norm": 10.9375, "learning_rate": 1.65825747356722e-06, "loss": 0.4015, "step": 9310 }, { "epoch": 0.7934384320409033, "grad_norm": 11.0, "learning_rate": 1.6580986492993234e-06, "loss": 0.4577, "step": 9311 }, { "epoch": 0.7935236472092032, "grad_norm": 12.6875, "learning_rate": 1.657939817656674e-06, "loss": 0.677, "step": 9312 }, { "epoch": 0.7936088623775032, "grad_norm": 14.5, "learning_rate": 1.6577809786421417e-06, "loss": 0.5482, "step": 9313 }, { "epoch": 0.7936940775458031, "grad_norm": 17.25, "learning_rate": 1.6576221322585973e-06, "loss": 1.1866, "step": 9314 }, { "epoch": 0.7937792927141031, "grad_norm": 12.6875, "learning_rate": 1.6574632785089107e-06, "loss": 0.5453, "step": 9315 }, { "epoch": 0.793864507882403, "grad_norm": 12.5625, "learning_rate": 1.657304417395953e-06, "loss": 0.4564, "step": 9316 }, { "epoch": 0.793949723050703, "grad_norm": 10.8125, "learning_rate": 1.657145548922595e-06, "loss": 0.3868, "step": 9317 }, { "epoch": 0.794034938219003, "grad_norm": 11.625, "learning_rate": 1.6569866730917074e-06, "loss": 0.342, "step": 9318 }, { "epoch": 0.7941201533873029, "grad_norm": 11.1875, "learning_rate": 1.6568277899061618e-06, "loss": 0.4057, "step": 9319 }, { "epoch": 0.7942053685556029, "grad_norm": 21.5, "learning_rate": 1.6566688993688285e-06, "loss": 1.0049, "step": 9320 }, { "epoch": 0.7942905837239028, "grad_norm": 14.6875, "learning_rate": 1.6565100014825798e-06, "loss": 0.5596, "step": 9321 }, { "epoch": 0.7943757988922028, "grad_norm": 24.75, "learning_rate": 1.6563510962502857e-06, "loss": 0.9233, "step": 9322 }, { "epoch": 0.7944610140605027, "grad_norm": 21.125, "learning_rate": 1.6561921836748195e-06, "loss": 0.673, "step": 9323 }, { "epoch": 0.7945462292288027, "grad_norm": 22.375, "learning_rate": 1.6560332637590514e-06, "loss": 0.9202, "step": 9324 }, { "epoch": 0.7946314443971026, "grad_norm": 23.0, "learning_rate": 1.6558743365058546e-06, "loss": 0.7403, "step": 9325 }, { "epoch": 0.7947166595654026, "grad_norm": 15.125, "learning_rate": 1.6557154019180998e-06, "loss": 0.665, "step": 9326 }, { "epoch": 0.7948018747337026, "grad_norm": 15.8125, "learning_rate": 1.6555564599986603e-06, "loss": 0.8816, "step": 9327 }, { "epoch": 0.7948870899020025, "grad_norm": 13.1875, "learning_rate": 1.6553975107504072e-06, "loss": 0.707, "step": 9328 }, { "epoch": 0.7949723050703025, "grad_norm": 13.25, "learning_rate": 1.655238554176214e-06, "loss": 0.5945, "step": 9329 }, { "epoch": 0.7950575202386024, "grad_norm": 12.5625, "learning_rate": 1.655079590278952e-06, "loss": 0.6046, "step": 9330 }, { "epoch": 0.7951427354069024, "grad_norm": 13.75, "learning_rate": 1.6549206190614948e-06, "loss": 0.8768, "step": 9331 }, { "epoch": 0.7952279505752023, "grad_norm": 17.0, "learning_rate": 1.6547616405267147e-06, "loss": 0.5921, "step": 9332 }, { "epoch": 0.7953131657435023, "grad_norm": 13.375, "learning_rate": 1.6546026546774848e-06, "loss": 0.5038, "step": 9333 }, { "epoch": 0.7953983809118023, "grad_norm": 14.125, "learning_rate": 1.6544436615166781e-06, "loss": 0.6684, "step": 9334 }, { "epoch": 0.7954835960801022, "grad_norm": 16.0, "learning_rate": 1.654284661047168e-06, "loss": 0.8307, "step": 9335 }, { "epoch": 0.7955688112484022, "grad_norm": 13.375, "learning_rate": 1.6541256532718272e-06, "loss": 0.6982, "step": 9336 }, { "epoch": 0.7956540264167021, "grad_norm": 16.25, "learning_rate": 1.6539666381935293e-06, "loss": 0.7893, "step": 9337 }, { "epoch": 0.7957392415850021, "grad_norm": 13.9375, "learning_rate": 1.6538076158151484e-06, "loss": 0.6543, "step": 9338 }, { "epoch": 0.795824456753302, "grad_norm": 13.0625, "learning_rate": 1.6536485861395576e-06, "loss": 0.4526, "step": 9339 }, { "epoch": 0.795909671921602, "grad_norm": 17.125, "learning_rate": 1.6534895491696313e-06, "loss": 0.9724, "step": 9340 }, { "epoch": 0.795994887089902, "grad_norm": 15.6875, "learning_rate": 1.6533305049082427e-06, "loss": 0.6917, "step": 9341 }, { "epoch": 0.7960801022582019, "grad_norm": 10.875, "learning_rate": 1.6531714533582664e-06, "loss": 0.4756, "step": 9342 }, { "epoch": 0.7961653174265019, "grad_norm": 12.9375, "learning_rate": 1.6530123945225765e-06, "loss": 0.6805, "step": 9343 }, { "epoch": 0.7962505325948018, "grad_norm": 15.375, "learning_rate": 1.6528533284040476e-06, "loss": 0.8315, "step": 9344 }, { "epoch": 0.7963357477631018, "grad_norm": 14.875, "learning_rate": 1.6526942550055537e-06, "loss": 0.7992, "step": 9345 }, { "epoch": 0.7964209629314017, "grad_norm": 12.8125, "learning_rate": 1.6525351743299697e-06, "loss": 0.607, "step": 9346 }, { "epoch": 0.7965061780997017, "grad_norm": 25.125, "learning_rate": 1.6523760863801705e-06, "loss": 1.0917, "step": 9347 }, { "epoch": 0.7965913932680017, "grad_norm": 17.125, "learning_rate": 1.6522169911590306e-06, "loss": 1.0436, "step": 9348 }, { "epoch": 0.7966766084363016, "grad_norm": 14.375, "learning_rate": 1.6520578886694253e-06, "loss": 0.8071, "step": 9349 }, { "epoch": 0.7967618236046016, "grad_norm": 15.375, "learning_rate": 1.6518987789142296e-06, "loss": 0.7745, "step": 9350 }, { "epoch": 0.7968470387729015, "grad_norm": 19.625, "learning_rate": 1.6517396618963188e-06, "loss": 0.8329, "step": 9351 }, { "epoch": 0.7969322539412015, "grad_norm": 13.4375, "learning_rate": 1.6515805376185685e-06, "loss": 0.5778, "step": 9352 }, { "epoch": 0.7970174691095014, "grad_norm": 12.9375, "learning_rate": 1.6514214060838538e-06, "loss": 0.7726, "step": 9353 }, { "epoch": 0.7971026842778014, "grad_norm": 15.4375, "learning_rate": 1.6512622672950506e-06, "loss": 0.8389, "step": 9354 }, { "epoch": 0.7971878994461014, "grad_norm": 19.0, "learning_rate": 1.651103121255035e-06, "loss": 1.2951, "step": 9355 }, { "epoch": 0.7972731146144013, "grad_norm": 22.875, "learning_rate": 1.6509439679666826e-06, "loss": 1.0622, "step": 9356 }, { "epoch": 0.7973583297827013, "grad_norm": 13.75, "learning_rate": 1.6507848074328694e-06, "loss": 0.865, "step": 9357 }, { "epoch": 0.7974435449510012, "grad_norm": 11.9375, "learning_rate": 1.6506256396564718e-06, "loss": 0.5667, "step": 9358 }, { "epoch": 0.7975287601193012, "grad_norm": 16.75, "learning_rate": 1.6504664646403662e-06, "loss": 0.6396, "step": 9359 }, { "epoch": 0.7976139752876011, "grad_norm": 16.25, "learning_rate": 1.6503072823874287e-06, "loss": 0.7614, "step": 9360 }, { "epoch": 0.7976991904559011, "grad_norm": 17.5, "learning_rate": 1.6501480929005358e-06, "loss": 1.3279, "step": 9361 }, { "epoch": 0.797784405624201, "grad_norm": 12.0, "learning_rate": 1.649988896182565e-06, "loss": 0.7284, "step": 9362 }, { "epoch": 0.797869620792501, "grad_norm": 20.0, "learning_rate": 1.649829692236392e-06, "loss": 0.9211, "step": 9363 }, { "epoch": 0.797954835960801, "grad_norm": 12.125, "learning_rate": 1.649670481064895e-06, "loss": 0.6325, "step": 9364 }, { "epoch": 0.7980400511291009, "grad_norm": 13.875, "learning_rate": 1.6495112626709505e-06, "loss": 0.7835, "step": 9365 }, { "epoch": 0.798125266297401, "grad_norm": 19.0, "learning_rate": 1.6493520370574354e-06, "loss": 0.8174, "step": 9366 }, { "epoch": 0.798210481465701, "grad_norm": 14.0625, "learning_rate": 1.6491928042272276e-06, "loss": 0.8003, "step": 9367 }, { "epoch": 0.7982956966340009, "grad_norm": 20.5, "learning_rate": 1.6490335641832045e-06, "loss": 0.6849, "step": 9368 }, { "epoch": 0.7983809118023009, "grad_norm": 14.125, "learning_rate": 1.648874316928243e-06, "loss": 0.8997, "step": 9369 }, { "epoch": 0.7984661269706008, "grad_norm": 10.0, "learning_rate": 1.648715062465222e-06, "loss": 0.2884, "step": 9370 }, { "epoch": 0.7985513421389008, "grad_norm": 14.0625, "learning_rate": 1.648555800797019e-06, "loss": 0.6575, "step": 9371 }, { "epoch": 0.7986365573072007, "grad_norm": 12.1875, "learning_rate": 1.6483965319265121e-06, "loss": 0.7147, "step": 9372 }, { "epoch": 0.7987217724755007, "grad_norm": 20.375, "learning_rate": 1.6482372558565787e-06, "loss": 1.119, "step": 9373 }, { "epoch": 0.7988069876438006, "grad_norm": 12.125, "learning_rate": 1.6480779725900979e-06, "loss": 0.5244, "step": 9374 }, { "epoch": 0.7988922028121006, "grad_norm": 16.0, "learning_rate": 1.6479186821299475e-06, "loss": 0.6104, "step": 9375 }, { "epoch": 0.7989774179804006, "grad_norm": 11.9375, "learning_rate": 1.647759384479007e-06, "loss": 0.5298, "step": 9376 }, { "epoch": 0.7990626331487005, "grad_norm": 13.0625, "learning_rate": 1.647600079640154e-06, "loss": 0.3543, "step": 9377 }, { "epoch": 0.7991478483170005, "grad_norm": 10.5, "learning_rate": 1.647440767616268e-06, "loss": 0.4199, "step": 9378 }, { "epoch": 0.7992330634853004, "grad_norm": 14.0, "learning_rate": 1.6472814484102273e-06, "loss": 0.7481, "step": 9379 }, { "epoch": 0.7993182786536004, "grad_norm": 21.5, "learning_rate": 1.6471221220249117e-06, "loss": 0.8683, "step": 9380 }, { "epoch": 0.7994034938219003, "grad_norm": 14.625, "learning_rate": 1.6469627884631997e-06, "loss": 0.6093, "step": 9381 }, { "epoch": 0.7994887089902003, "grad_norm": 13.5, "learning_rate": 1.6468034477279712e-06, "loss": 0.7172, "step": 9382 }, { "epoch": 0.7995739241585003, "grad_norm": 19.375, "learning_rate": 1.646644099822105e-06, "loss": 0.8839, "step": 9383 }, { "epoch": 0.7996591393268002, "grad_norm": 14.375, "learning_rate": 1.6464847447484817e-06, "loss": 0.6429, "step": 9384 }, { "epoch": 0.7997443544951002, "grad_norm": 17.5, "learning_rate": 1.6463253825099799e-06, "loss": 0.5567, "step": 9385 }, { "epoch": 0.7998295696634001, "grad_norm": 14.3125, "learning_rate": 1.6461660131094797e-06, "loss": 0.5549, "step": 9386 }, { "epoch": 0.7999147848317001, "grad_norm": 19.125, "learning_rate": 1.6460066365498617e-06, "loss": 1.2088, "step": 9387 }, { "epoch": 0.8, "grad_norm": 16.25, "learning_rate": 1.6458472528340054e-06, "loss": 0.9374, "step": 9388 }, { "epoch": 0.8000852151683, "grad_norm": 13.1875, "learning_rate": 1.6456878619647914e-06, "loss": 0.5813, "step": 9389 }, { "epoch": 0.8001704303366, "grad_norm": 12.6875, "learning_rate": 1.6455284639450996e-06, "loss": 0.5815, "step": 9390 }, { "epoch": 0.8002556455048999, "grad_norm": 14.1875, "learning_rate": 1.645369058777811e-06, "loss": 0.6624, "step": 9391 }, { "epoch": 0.8003408606731999, "grad_norm": 14.6875, "learning_rate": 1.6452096464658056e-06, "loss": 0.5533, "step": 9392 }, { "epoch": 0.8004260758414998, "grad_norm": 12.0625, "learning_rate": 1.6450502270119645e-06, "loss": 0.4451, "step": 9393 }, { "epoch": 0.8005112910097998, "grad_norm": 11.5625, "learning_rate": 1.6448908004191684e-06, "loss": 0.5631, "step": 9394 }, { "epoch": 0.8005965061780997, "grad_norm": 30.25, "learning_rate": 1.644731366690299e-06, "loss": 1.4518, "step": 9395 }, { "epoch": 0.8006817213463997, "grad_norm": 12.0, "learning_rate": 1.6445719258282367e-06, "loss": 0.3933, "step": 9396 }, { "epoch": 0.8007669365146997, "grad_norm": 18.125, "learning_rate": 1.6444124778358628e-06, "loss": 0.8528, "step": 9397 }, { "epoch": 0.8008521516829996, "grad_norm": 18.5, "learning_rate": 1.6442530227160585e-06, "loss": 0.6588, "step": 9398 }, { "epoch": 0.8009373668512996, "grad_norm": 13.375, "learning_rate": 1.6440935604717059e-06, "loss": 0.9586, "step": 9399 }, { "epoch": 0.8010225820195995, "grad_norm": 20.125, "learning_rate": 1.6439340911056868e-06, "loss": 1.258, "step": 9400 }, { "epoch": 0.8011077971878995, "grad_norm": 12.5625, "learning_rate": 1.6437746146208816e-06, "loss": 0.712, "step": 9401 }, { "epoch": 0.8011930123561994, "grad_norm": 12.8125, "learning_rate": 1.643615131020174e-06, "loss": 0.7403, "step": 9402 }, { "epoch": 0.8012782275244994, "grad_norm": 16.25, "learning_rate": 1.6434556403064445e-06, "loss": 0.5882, "step": 9403 }, { "epoch": 0.8013634426927994, "grad_norm": 15.0, "learning_rate": 1.6432961424825767e-06, "loss": 0.7012, "step": 9404 }, { "epoch": 0.8014486578610993, "grad_norm": 27.5, "learning_rate": 1.6431366375514515e-06, "loss": 1.2114, "step": 9405 }, { "epoch": 0.8015338730293993, "grad_norm": 15.5625, "learning_rate": 1.642977125515952e-06, "loss": 0.8754, "step": 9406 }, { "epoch": 0.8016190881976992, "grad_norm": 16.25, "learning_rate": 1.6428176063789607e-06, "loss": 0.8336, "step": 9407 }, { "epoch": 0.8017043033659992, "grad_norm": 12.625, "learning_rate": 1.6426580801433607e-06, "loss": 0.7895, "step": 9408 }, { "epoch": 0.8017895185342991, "grad_norm": 13.25, "learning_rate": 1.642498546812034e-06, "loss": 0.4936, "step": 9409 }, { "epoch": 0.8018747337025991, "grad_norm": 16.75, "learning_rate": 1.6423390063878636e-06, "loss": 0.5878, "step": 9410 }, { "epoch": 0.801959948870899, "grad_norm": 11.5625, "learning_rate": 1.642179458873733e-06, "loss": 0.6331, "step": 9411 }, { "epoch": 0.802045164039199, "grad_norm": 18.75, "learning_rate": 1.6420199042725258e-06, "loss": 1.1841, "step": 9412 }, { "epoch": 0.802130379207499, "grad_norm": 32.5, "learning_rate": 1.6418603425871239e-06, "loss": 1.244, "step": 9413 }, { "epoch": 0.8022155943757989, "grad_norm": 11.25, "learning_rate": 1.6417007738204118e-06, "loss": 0.6814, "step": 9414 }, { "epoch": 0.8023008095440989, "grad_norm": 9.25, "learning_rate": 1.641541197975273e-06, "loss": 0.2862, "step": 9415 }, { "epoch": 0.8023860247123988, "grad_norm": 15.0, "learning_rate": 1.6413816150545908e-06, "loss": 0.767, "step": 9416 }, { "epoch": 0.8024712398806988, "grad_norm": 13.4375, "learning_rate": 1.6412220250612496e-06, "loss": 0.752, "step": 9417 }, { "epoch": 0.8025564550489988, "grad_norm": 15.6875, "learning_rate": 1.6410624279981324e-06, "loss": 0.7724, "step": 9418 }, { "epoch": 0.8026416702172987, "grad_norm": 16.125, "learning_rate": 1.6409028238681247e-06, "loss": 0.9049, "step": 9419 }, { "epoch": 0.8027268853855987, "grad_norm": 16.25, "learning_rate": 1.6407432126741093e-06, "loss": 0.984, "step": 9420 }, { "epoch": 0.8028121005538986, "grad_norm": 16.25, "learning_rate": 1.6405835944189713e-06, "loss": 0.845, "step": 9421 }, { "epoch": 0.8028973157221986, "grad_norm": 13.625, "learning_rate": 1.6404239691055947e-06, "loss": 0.8253, "step": 9422 }, { "epoch": 0.8029825308904985, "grad_norm": 14.625, "learning_rate": 1.6402643367368649e-06, "loss": 0.2041, "step": 9423 }, { "epoch": 0.8030677460587985, "grad_norm": 13.0, "learning_rate": 1.6401046973156656e-06, "loss": 0.4541, "step": 9424 }, { "epoch": 0.8031529612270984, "grad_norm": 16.875, "learning_rate": 1.6399450508448827e-06, "loss": 0.6505, "step": 9425 }, { "epoch": 0.8032381763953984, "grad_norm": 23.25, "learning_rate": 1.6397853973274003e-06, "loss": 1.4802, "step": 9426 }, { "epoch": 0.8033233915636984, "grad_norm": 12.6875, "learning_rate": 1.6396257367661037e-06, "loss": 0.6071, "step": 9427 }, { "epoch": 0.8034086067319983, "grad_norm": 16.625, "learning_rate": 1.639466069163878e-06, "loss": 0.6639, "step": 9428 }, { "epoch": 0.8034938219002983, "grad_norm": 12.0625, "learning_rate": 1.6393063945236095e-06, "loss": 0.811, "step": 9429 }, { "epoch": 0.8035790370685982, "grad_norm": 17.25, "learning_rate": 1.6391467128481825e-06, "loss": 0.6467, "step": 9430 }, { "epoch": 0.8036642522368982, "grad_norm": 13.125, "learning_rate": 1.6389870241404834e-06, "loss": 0.6829, "step": 9431 }, { "epoch": 0.8037494674051981, "grad_norm": 14.4375, "learning_rate": 1.6388273284033972e-06, "loss": 0.8062, "step": 9432 }, { "epoch": 0.8038346825734981, "grad_norm": 19.125, "learning_rate": 1.6386676256398104e-06, "loss": 0.8652, "step": 9433 }, { "epoch": 0.8039198977417981, "grad_norm": 14.125, "learning_rate": 1.6385079158526088e-06, "loss": 0.5597, "step": 9434 }, { "epoch": 0.804005112910098, "grad_norm": 16.25, "learning_rate": 1.6383481990446787e-06, "loss": 0.4233, "step": 9435 }, { "epoch": 0.804090328078398, "grad_norm": 9.8125, "learning_rate": 1.6381884752189056e-06, "loss": 0.252, "step": 9436 }, { "epoch": 0.8041755432466979, "grad_norm": 16.875, "learning_rate": 1.6380287443781768e-06, "loss": 0.8427, "step": 9437 }, { "epoch": 0.8042607584149979, "grad_norm": 12.5, "learning_rate": 1.6378690065253784e-06, "loss": 0.5465, "step": 9438 }, { "epoch": 0.8043459735832978, "grad_norm": 35.0, "learning_rate": 1.6377092616633967e-06, "loss": 0.5623, "step": 9439 }, { "epoch": 0.8044311887515978, "grad_norm": 15.25, "learning_rate": 1.6375495097951188e-06, "loss": 0.7402, "step": 9440 }, { "epoch": 0.8045164039198978, "grad_norm": 14.75, "learning_rate": 1.6373897509234318e-06, "loss": 0.7543, "step": 9441 }, { "epoch": 0.8046016190881977, "grad_norm": 20.25, "learning_rate": 1.6372299850512224e-06, "loss": 1.0142, "step": 9442 }, { "epoch": 0.8046868342564977, "grad_norm": 15.0625, "learning_rate": 1.6370702121813776e-06, "loss": 1.0541, "step": 9443 }, { "epoch": 0.8047720494247976, "grad_norm": 15.0, "learning_rate": 1.6369104323167851e-06, "loss": 0.525, "step": 9444 }, { "epoch": 0.8048572645930976, "grad_norm": 15.6875, "learning_rate": 1.6367506454603319e-06, "loss": 0.7357, "step": 9445 }, { "epoch": 0.8049424797613975, "grad_norm": 15.875, "learning_rate": 1.6365908516149056e-06, "loss": 0.8136, "step": 9446 }, { "epoch": 0.8050276949296975, "grad_norm": 19.0, "learning_rate": 1.6364310507833938e-06, "loss": 0.7438, "step": 9447 }, { "epoch": 0.8051129100979975, "grad_norm": 11.8125, "learning_rate": 1.6362712429686844e-06, "loss": 0.3985, "step": 9448 }, { "epoch": 0.8051981252662974, "grad_norm": 16.25, "learning_rate": 1.6361114281736653e-06, "loss": 0.803, "step": 9449 }, { "epoch": 0.8052833404345974, "grad_norm": 31.5, "learning_rate": 1.6359516064012246e-06, "loss": 0.6475, "step": 9450 }, { "epoch": 0.8053685556028973, "grad_norm": 20.125, "learning_rate": 1.6357917776542497e-06, "loss": 0.9331, "step": 9451 }, { "epoch": 0.8054537707711973, "grad_norm": 20.125, "learning_rate": 1.6356319419356298e-06, "loss": 1.2658, "step": 9452 }, { "epoch": 0.8055389859394972, "grad_norm": 15.6875, "learning_rate": 1.635472099248253e-06, "loss": 0.6567, "step": 9453 }, { "epoch": 0.8056242011077972, "grad_norm": 13.3125, "learning_rate": 1.6353122495950075e-06, "loss": 0.5096, "step": 9454 }, { "epoch": 0.8057094162760972, "grad_norm": 18.125, "learning_rate": 1.635152392978782e-06, "loss": 0.808, "step": 9455 }, { "epoch": 0.8057946314443971, "grad_norm": 15.125, "learning_rate": 1.634992529402466e-06, "loss": 0.9815, "step": 9456 }, { "epoch": 0.8058798466126971, "grad_norm": 30.25, "learning_rate": 1.634832658868948e-06, "loss": 0.9481, "step": 9457 }, { "epoch": 0.805965061780997, "grad_norm": 23.75, "learning_rate": 1.6346727813811162e-06, "loss": 1.0077, "step": 9458 }, { "epoch": 0.806050276949297, "grad_norm": 10.75, "learning_rate": 1.6345128969418608e-06, "loss": 0.439, "step": 9459 }, { "epoch": 0.8061354921175969, "grad_norm": 15.5, "learning_rate": 1.6343530055540704e-06, "loss": 0.8602, "step": 9460 }, { "epoch": 0.8062207072858969, "grad_norm": 22.125, "learning_rate": 1.6341931072206352e-06, "loss": 0.7514, "step": 9461 }, { "epoch": 0.8063059224541969, "grad_norm": 21.375, "learning_rate": 1.6340332019444438e-06, "loss": 0.935, "step": 9462 }, { "epoch": 0.8063911376224968, "grad_norm": 15.6875, "learning_rate": 1.6338732897283866e-06, "loss": 1.0326, "step": 9463 }, { "epoch": 0.8064763527907968, "grad_norm": 14.9375, "learning_rate": 1.6337133705753523e-06, "loss": 0.9387, "step": 9464 }, { "epoch": 0.8065615679590967, "grad_norm": 10.375, "learning_rate": 1.6335534444882326e-06, "loss": 0.4275, "step": 9465 }, { "epoch": 0.8066467831273967, "grad_norm": 20.0, "learning_rate": 1.6333935114699156e-06, "loss": 0.6612, "step": 9466 }, { "epoch": 0.8067319982956966, "grad_norm": 11.8125, "learning_rate": 1.6332335715232927e-06, "loss": 0.5352, "step": 9467 }, { "epoch": 0.8068172134639966, "grad_norm": 12.9375, "learning_rate": 1.6330736246512535e-06, "loss": 0.5707, "step": 9468 }, { "epoch": 0.8069024286322966, "grad_norm": 24.5, "learning_rate": 1.6329136708566892e-06, "loss": 0.7378, "step": 9469 }, { "epoch": 0.8069876438005965, "grad_norm": 11.75, "learning_rate": 1.6327537101424894e-06, "loss": 0.4521, "step": 9470 }, { "epoch": 0.8070728589688965, "grad_norm": 13.25, "learning_rate": 1.6325937425115454e-06, "loss": 0.8363, "step": 9471 }, { "epoch": 0.8071580741371964, "grad_norm": 11.8125, "learning_rate": 1.6324337679667477e-06, "loss": 0.4928, "step": 9472 }, { "epoch": 0.8072432893054964, "grad_norm": 13.1875, "learning_rate": 1.6322737865109872e-06, "loss": 0.5882, "step": 9473 }, { "epoch": 0.8073285044737963, "grad_norm": 11.875, "learning_rate": 1.6321137981471551e-06, "loss": 0.5267, "step": 9474 }, { "epoch": 0.8074137196420963, "grad_norm": 14.25, "learning_rate": 1.6319538028781423e-06, "loss": 0.635, "step": 9475 }, { "epoch": 0.8074989348103963, "grad_norm": 16.25, "learning_rate": 1.63179380070684e-06, "loss": 1.034, "step": 9476 }, { "epoch": 0.8075841499786962, "grad_norm": 17.25, "learning_rate": 1.6316337916361402e-06, "loss": 1.1004, "step": 9477 }, { "epoch": 0.8076693651469962, "grad_norm": 14.6875, "learning_rate": 1.631473775668934e-06, "loss": 0.5225, "step": 9478 }, { "epoch": 0.8077545803152961, "grad_norm": 12.0, "learning_rate": 1.6313137528081125e-06, "loss": 0.6254, "step": 9479 }, { "epoch": 0.8078397954835961, "grad_norm": 15.4375, "learning_rate": 1.6311537230565686e-06, "loss": 0.5595, "step": 9480 }, { "epoch": 0.807925010651896, "grad_norm": 15.0, "learning_rate": 1.6309936864171935e-06, "loss": 0.4604, "step": 9481 }, { "epoch": 0.808010225820196, "grad_norm": 18.875, "learning_rate": 1.6308336428928794e-06, "loss": 0.9487, "step": 9482 }, { "epoch": 0.808095440988496, "grad_norm": 17.25, "learning_rate": 1.6306735924865184e-06, "loss": 0.5752, "step": 9483 }, { "epoch": 0.8081806561567959, "grad_norm": 18.0, "learning_rate": 1.6305135352010026e-06, "loss": 0.9865, "step": 9484 }, { "epoch": 0.8082658713250959, "grad_norm": 10.5, "learning_rate": 1.630353471039225e-06, "loss": 0.2656, "step": 9485 }, { "epoch": 0.8083510864933958, "grad_norm": 12.9375, "learning_rate": 1.6301934000040773e-06, "loss": 0.8126, "step": 9486 }, { "epoch": 0.8084363016616958, "grad_norm": 18.625, "learning_rate": 1.6300333220984526e-06, "loss": 1.0819, "step": 9487 }, { "epoch": 0.8085215168299957, "grad_norm": 15.0625, "learning_rate": 1.6298732373252436e-06, "loss": 0.8976, "step": 9488 }, { "epoch": 0.8086067319982957, "grad_norm": 11.9375, "learning_rate": 1.6297131456873433e-06, "loss": 0.6188, "step": 9489 }, { "epoch": 0.8086919471665956, "grad_norm": 56.25, "learning_rate": 1.6295530471876448e-06, "loss": 0.7695, "step": 9490 }, { "epoch": 0.8087771623348956, "grad_norm": 15.25, "learning_rate": 1.629392941829041e-06, "loss": 0.7134, "step": 9491 }, { "epoch": 0.8088623775031956, "grad_norm": 24.375, "learning_rate": 1.629232829614425e-06, "loss": 0.8882, "step": 9492 }, { "epoch": 0.8089475926714955, "grad_norm": 14.75, "learning_rate": 1.6290727105466906e-06, "loss": 0.7056, "step": 9493 }, { "epoch": 0.8090328078397955, "grad_norm": 12.0, "learning_rate": 1.6289125846287313e-06, "loss": 0.5221, "step": 9494 }, { "epoch": 0.8091180230080954, "grad_norm": 16.375, "learning_rate": 1.6287524518634402e-06, "loss": 0.6398, "step": 9495 }, { "epoch": 0.8092032381763954, "grad_norm": 16.875, "learning_rate": 1.6285923122537114e-06, "loss": 0.8392, "step": 9496 }, { "epoch": 0.8092884533446953, "grad_norm": 11.6875, "learning_rate": 1.628432165802439e-06, "loss": 0.4834, "step": 9497 }, { "epoch": 0.8093736685129953, "grad_norm": 12.75, "learning_rate": 1.628272012512517e-06, "loss": 0.7347, "step": 9498 }, { "epoch": 0.8094588836812953, "grad_norm": 16.0, "learning_rate": 1.6281118523868395e-06, "loss": 0.9386, "step": 9499 }, { "epoch": 0.8095440988495952, "grad_norm": 17.125, "learning_rate": 1.6279516854283e-06, "loss": 0.8599, "step": 9500 }, { "epoch": 0.8096293140178952, "grad_norm": 11.1875, "learning_rate": 1.6277915116397938e-06, "loss": 0.5091, "step": 9501 }, { "epoch": 0.8097145291861951, "grad_norm": 18.75, "learning_rate": 1.6276313310242151e-06, "loss": 0.5912, "step": 9502 }, { "epoch": 0.8097997443544951, "grad_norm": 15.0, "learning_rate": 1.6274711435844582e-06, "loss": 0.649, "step": 9503 }, { "epoch": 0.809884959522795, "grad_norm": 10.625, "learning_rate": 1.6273109493234184e-06, "loss": 0.2716, "step": 9504 }, { "epoch": 0.809970174691095, "grad_norm": 22.5, "learning_rate": 1.6271507482439907e-06, "loss": 0.3357, "step": 9505 }, { "epoch": 0.810055389859395, "grad_norm": 10.75, "learning_rate": 1.6269905403490693e-06, "loss": 0.506, "step": 9506 }, { "epoch": 0.8101406050276949, "grad_norm": 15.4375, "learning_rate": 1.6268303256415496e-06, "loss": 1.1093, "step": 9507 }, { "epoch": 0.8102258201959949, "grad_norm": 12.4375, "learning_rate": 1.6266701041243275e-06, "loss": 0.4562, "step": 9508 }, { "epoch": 0.8103110353642948, "grad_norm": 10.25, "learning_rate": 1.6265098758002972e-06, "loss": 0.2915, "step": 9509 }, { "epoch": 0.8103962505325948, "grad_norm": 16.25, "learning_rate": 1.6263496406723556e-06, "loss": 0.7986, "step": 9510 }, { "epoch": 0.8104814657008947, "grad_norm": 14.0625, "learning_rate": 1.6261893987433968e-06, "loss": 0.6524, "step": 9511 }, { "epoch": 0.8105666808691947, "grad_norm": 10.75, "learning_rate": 1.6260291500163177e-06, "loss": 0.4812, "step": 9512 }, { "epoch": 0.8106518960374947, "grad_norm": 25.875, "learning_rate": 1.6258688944940138e-06, "loss": 0.9877, "step": 9513 }, { "epoch": 0.8107371112057946, "grad_norm": 18.5, "learning_rate": 1.6257086321793812e-06, "loss": 1.069, "step": 9514 }, { "epoch": 0.8108223263740946, "grad_norm": 22.625, "learning_rate": 1.625548363075315e-06, "loss": 0.8805, "step": 9515 }, { "epoch": 0.8109075415423945, "grad_norm": 12.75, "learning_rate": 1.6253880871847127e-06, "loss": 0.493, "step": 9516 }, { "epoch": 0.8109927567106945, "grad_norm": 12.8125, "learning_rate": 1.6252278045104703e-06, "loss": 0.6482, "step": 9517 }, { "epoch": 0.8110779718789944, "grad_norm": 14.125, "learning_rate": 1.625067515055484e-06, "loss": 0.3299, "step": 9518 }, { "epoch": 0.8111631870472944, "grad_norm": 14.9375, "learning_rate": 1.6249072188226505e-06, "loss": 0.9892, "step": 9519 }, { "epoch": 0.8112484022155944, "grad_norm": 18.625, "learning_rate": 1.6247469158148668e-06, "loss": 0.6329, "step": 9520 }, { "epoch": 0.8113336173838943, "grad_norm": 15.4375, "learning_rate": 1.624586606035029e-06, "loss": 0.6345, "step": 9521 }, { "epoch": 0.8114188325521943, "grad_norm": 25.625, "learning_rate": 1.6244262894860352e-06, "loss": 0.9274, "step": 9522 }, { "epoch": 0.8115040477204942, "grad_norm": 17.25, "learning_rate": 1.6242659661707816e-06, "loss": 1.0147, "step": 9523 }, { "epoch": 0.8115892628887942, "grad_norm": 22.375, "learning_rate": 1.6241056360921654e-06, "loss": 1.0602, "step": 9524 }, { "epoch": 0.8116744780570941, "grad_norm": 23.375, "learning_rate": 1.6239452992530843e-06, "loss": 0.6488, "step": 9525 }, { "epoch": 0.8117596932253941, "grad_norm": 23.0, "learning_rate": 1.6237849556564355e-06, "loss": 1.0389, "step": 9526 }, { "epoch": 0.811844908393694, "grad_norm": 13.0625, "learning_rate": 1.6236246053051172e-06, "loss": 0.6787, "step": 9527 }, { "epoch": 0.811930123561994, "grad_norm": 16.75, "learning_rate": 1.6234642482020257e-06, "loss": 0.5816, "step": 9528 }, { "epoch": 0.812015338730294, "grad_norm": 19.0, "learning_rate": 1.6233038843500604e-06, "loss": 0.6062, "step": 9529 }, { "epoch": 0.8121005538985939, "grad_norm": 30.25, "learning_rate": 1.6231435137521184e-06, "loss": 0.9707, "step": 9530 }, { "epoch": 0.8121857690668939, "grad_norm": 14.75, "learning_rate": 1.622983136411098e-06, "loss": 0.7146, "step": 9531 }, { "epoch": 0.8122709842351938, "grad_norm": 11.5, "learning_rate": 1.6228227523298968e-06, "loss": 0.4218, "step": 9532 }, { "epoch": 0.8123561994034938, "grad_norm": 11.6875, "learning_rate": 1.622662361511414e-06, "loss": 0.395, "step": 9533 }, { "epoch": 0.8124414145717938, "grad_norm": 14.4375, "learning_rate": 1.6225019639585473e-06, "loss": 0.7819, "step": 9534 }, { "epoch": 0.8125266297400937, "grad_norm": 10.625, "learning_rate": 1.6223415596741957e-06, "loss": 0.3752, "step": 9535 }, { "epoch": 0.8126118449083937, "grad_norm": 18.375, "learning_rate": 1.6221811486612574e-06, "loss": 1.1257, "step": 9536 }, { "epoch": 0.8126970600766936, "grad_norm": 15.3125, "learning_rate": 1.6220207309226318e-06, "loss": 0.7075, "step": 9537 }, { "epoch": 0.8127822752449936, "grad_norm": 10.4375, "learning_rate": 1.6218603064612176e-06, "loss": 0.2056, "step": 9538 }, { "epoch": 0.8128674904132935, "grad_norm": 12.625, "learning_rate": 1.6216998752799132e-06, "loss": 0.6924, "step": 9539 }, { "epoch": 0.8129527055815935, "grad_norm": 11.6875, "learning_rate": 1.6215394373816187e-06, "loss": 0.4378, "step": 9540 }, { "epoch": 0.8130379207498935, "grad_norm": 17.25, "learning_rate": 1.6213789927692328e-06, "loss": 0.7156, "step": 9541 }, { "epoch": 0.8131231359181934, "grad_norm": 18.5, "learning_rate": 1.6212185414456554e-06, "loss": 1.0168, "step": 9542 }, { "epoch": 0.8132083510864934, "grad_norm": 12.5625, "learning_rate": 1.621058083413785e-06, "loss": 0.3976, "step": 9543 }, { "epoch": 0.8132935662547933, "grad_norm": 12.6875, "learning_rate": 1.6208976186765223e-06, "loss": 0.7649, "step": 9544 }, { "epoch": 0.8133787814230933, "grad_norm": 12.0, "learning_rate": 1.6207371472367666e-06, "loss": 0.4154, "step": 9545 }, { "epoch": 0.8134639965913932, "grad_norm": 15.8125, "learning_rate": 1.6205766690974185e-06, "loss": 0.6174, "step": 9546 }, { "epoch": 0.8135492117596932, "grad_norm": 22.0, "learning_rate": 1.6204161842613764e-06, "loss": 0.9451, "step": 9547 }, { "epoch": 0.8136344269279931, "grad_norm": 13.0, "learning_rate": 1.6202556927315419e-06, "loss": 0.4744, "step": 9548 }, { "epoch": 0.8137196420962931, "grad_norm": 21.125, "learning_rate": 1.6200951945108143e-06, "loss": 0.8855, "step": 9549 }, { "epoch": 0.8138048572645931, "grad_norm": 19.5, "learning_rate": 1.6199346896020948e-06, "loss": 1.2526, "step": 9550 }, { "epoch": 0.813890072432893, "grad_norm": 33.25, "learning_rate": 1.6197741780082832e-06, "loss": 1.0598, "step": 9551 }, { "epoch": 0.813975287601193, "grad_norm": 12.8125, "learning_rate": 1.6196136597322803e-06, "loss": 0.7703, "step": 9552 }, { "epoch": 0.8140605027694929, "grad_norm": 16.625, "learning_rate": 1.619453134776987e-06, "loss": 0.8769, "step": 9553 }, { "epoch": 0.8141457179377929, "grad_norm": 11.125, "learning_rate": 1.6192926031453041e-06, "loss": 0.4337, "step": 9554 }, { "epoch": 0.8142309331060928, "grad_norm": 34.5, "learning_rate": 1.6191320648401325e-06, "loss": 0.8267, "step": 9555 }, { "epoch": 0.8143161482743928, "grad_norm": 14.625, "learning_rate": 1.6189715198643732e-06, "loss": 0.5283, "step": 9556 }, { "epoch": 0.8144013634426928, "grad_norm": 23.75, "learning_rate": 1.6188109682209274e-06, "loss": 0.7032, "step": 9557 }, { "epoch": 0.8144865786109927, "grad_norm": 16.125, "learning_rate": 1.6186504099126971e-06, "loss": 0.755, "step": 9558 }, { "epoch": 0.8145717937792927, "grad_norm": 15.3125, "learning_rate": 1.618489844942583e-06, "loss": 0.9224, "step": 9559 }, { "epoch": 0.8146570089475926, "grad_norm": 14.875, "learning_rate": 1.618329273313486e-06, "loss": 0.7357, "step": 9560 }, { "epoch": 0.8147422241158926, "grad_norm": 12.5625, "learning_rate": 1.6181686950283096e-06, "loss": 0.7173, "step": 9561 }, { "epoch": 0.8148274392841925, "grad_norm": 13.125, "learning_rate": 1.6180081100899543e-06, "loss": 0.6681, "step": 9562 }, { "epoch": 0.8149126544524925, "grad_norm": 11.875, "learning_rate": 1.6178475185013226e-06, "loss": 0.5094, "step": 9563 }, { "epoch": 0.8149978696207925, "grad_norm": 15.625, "learning_rate": 1.617686920265316e-06, "loss": 0.9409, "step": 9564 }, { "epoch": 0.8150830847890924, "grad_norm": 9.5, "learning_rate": 1.6175263153848372e-06, "loss": 0.3378, "step": 9565 }, { "epoch": 0.8151682999573924, "grad_norm": 22.25, "learning_rate": 1.6173657038627881e-06, "loss": 0.8048, "step": 9566 }, { "epoch": 0.8152535151256923, "grad_norm": 12.875, "learning_rate": 1.6172050857020722e-06, "loss": 0.6447, "step": 9567 }, { "epoch": 0.8153387302939923, "grad_norm": 14.9375, "learning_rate": 1.61704446090559e-06, "loss": 0.7769, "step": 9568 }, { "epoch": 0.8154239454622922, "grad_norm": 16.125, "learning_rate": 1.616883829476246e-06, "loss": 0.6015, "step": 9569 }, { "epoch": 0.8155091606305922, "grad_norm": 11.375, "learning_rate": 1.6167231914169419e-06, "loss": 0.3633, "step": 9570 }, { "epoch": 0.8155943757988922, "grad_norm": 12.125, "learning_rate": 1.6165625467305812e-06, "loss": 0.4571, "step": 9571 }, { "epoch": 0.8156795909671921, "grad_norm": 11.375, "learning_rate": 1.6164018954200666e-06, "loss": 0.4366, "step": 9572 }, { "epoch": 0.8157648061354921, "grad_norm": 13.125, "learning_rate": 1.616241237488301e-06, "loss": 0.798, "step": 9573 }, { "epoch": 0.815850021303792, "grad_norm": 18.75, "learning_rate": 1.6160805729381884e-06, "loss": 0.897, "step": 9574 }, { "epoch": 0.815935236472092, "grad_norm": 19.625, "learning_rate": 1.6159199017726316e-06, "loss": 0.5949, "step": 9575 }, { "epoch": 0.8160204516403919, "grad_norm": 18.0, "learning_rate": 1.615759223994534e-06, "loss": 0.6339, "step": 9576 }, { "epoch": 0.8161056668086919, "grad_norm": 16.75, "learning_rate": 1.6155985396067997e-06, "loss": 0.9377, "step": 9577 }, { "epoch": 0.8161908819769919, "grad_norm": 19.0, "learning_rate": 1.615437848612332e-06, "loss": 0.7673, "step": 9578 }, { "epoch": 0.8162760971452918, "grad_norm": 19.875, "learning_rate": 1.6152771510140352e-06, "loss": 0.9419, "step": 9579 }, { "epoch": 0.8163613123135918, "grad_norm": 18.25, "learning_rate": 1.6151164468148128e-06, "loss": 1.1502, "step": 9580 }, { "epoch": 0.8164465274818917, "grad_norm": 19.25, "learning_rate": 1.6149557360175689e-06, "loss": 0.889, "step": 9581 }, { "epoch": 0.8165317426501917, "grad_norm": 12.6875, "learning_rate": 1.6147950186252078e-06, "loss": 0.6464, "step": 9582 }, { "epoch": 0.8166169578184916, "grad_norm": 18.875, "learning_rate": 1.614634294640634e-06, "loss": 0.6655, "step": 9583 }, { "epoch": 0.8167021729867916, "grad_norm": 11.3125, "learning_rate": 1.614473564066752e-06, "loss": 0.5105, "step": 9584 }, { "epoch": 0.8167873881550916, "grad_norm": 15.0625, "learning_rate": 1.614312826906466e-06, "loss": 0.6503, "step": 9585 }, { "epoch": 0.8168726033233915, "grad_norm": 11.8125, "learning_rate": 1.6141520831626808e-06, "loss": 0.4852, "step": 9586 }, { "epoch": 0.8169578184916915, "grad_norm": 16.375, "learning_rate": 1.6139913328383013e-06, "loss": 0.559, "step": 9587 }, { "epoch": 0.8170430336599914, "grad_norm": 20.5, "learning_rate": 1.6138305759362325e-06, "loss": 1.1046, "step": 9588 }, { "epoch": 0.8171282488282914, "grad_norm": 16.625, "learning_rate": 1.6136698124593795e-06, "loss": 1.0394, "step": 9589 }, { "epoch": 0.8172134639965913, "grad_norm": 18.0, "learning_rate": 1.613509042410647e-06, "loss": 0.971, "step": 9590 }, { "epoch": 0.8172986791648913, "grad_norm": 12.1875, "learning_rate": 1.613348265792941e-06, "loss": 0.6667, "step": 9591 }, { "epoch": 0.8173838943331914, "grad_norm": 19.0, "learning_rate": 1.6131874826091655e-06, "loss": 0.9719, "step": 9592 }, { "epoch": 0.8174691095014913, "grad_norm": 15.4375, "learning_rate": 1.613026692862228e-06, "loss": 0.813, "step": 9593 }, { "epoch": 0.8175543246697913, "grad_norm": 12.4375, "learning_rate": 1.6128658965550323e-06, "loss": 0.5615, "step": 9594 }, { "epoch": 0.8176395398380912, "grad_norm": 25.375, "learning_rate": 1.612705093690486e-06, "loss": 1.2454, "step": 9595 }, { "epoch": 0.8177247550063912, "grad_norm": 18.0, "learning_rate": 1.6125442842714928e-06, "loss": 0.8158, "step": 9596 }, { "epoch": 0.8178099701746911, "grad_norm": 19.5, "learning_rate": 1.6123834683009603e-06, "loss": 0.8649, "step": 9597 }, { "epoch": 0.8178951853429911, "grad_norm": 15.0625, "learning_rate": 1.6122226457817941e-06, "loss": 1.1615, "step": 9598 }, { "epoch": 0.8179804005112911, "grad_norm": 19.0, "learning_rate": 1.612061816716901e-06, "loss": 0.2495, "step": 9599 }, { "epoch": 0.818065615679591, "grad_norm": 15.0, "learning_rate": 1.611900981109186e-06, "loss": 1.3293, "step": 9600 }, { "epoch": 0.818150830847891, "grad_norm": 13.0, "learning_rate": 1.6117401389615568e-06, "loss": 0.5224, "step": 9601 }, { "epoch": 0.8182360460161909, "grad_norm": 16.25, "learning_rate": 1.611579290276919e-06, "loss": 0.9867, "step": 9602 }, { "epoch": 0.8183212611844909, "grad_norm": 17.5, "learning_rate": 1.6114184350581802e-06, "loss": 0.7392, "step": 9603 }, { "epoch": 0.8184064763527908, "grad_norm": 13.5625, "learning_rate": 1.611257573308247e-06, "loss": 0.9066, "step": 9604 }, { "epoch": 0.8184916915210908, "grad_norm": 13.5, "learning_rate": 1.611096705030026e-06, "loss": 0.6662, "step": 9605 }, { "epoch": 0.8185769066893908, "grad_norm": 14.3125, "learning_rate": 1.6109358302264245e-06, "loss": 0.409, "step": 9606 }, { "epoch": 0.8186621218576907, "grad_norm": 10.375, "learning_rate": 1.6107749489003493e-06, "loss": 0.5051, "step": 9607 }, { "epoch": 0.8187473370259907, "grad_norm": 17.5, "learning_rate": 1.6106140610547083e-06, "loss": 0.8585, "step": 9608 }, { "epoch": 0.8188325521942906, "grad_norm": 14.1875, "learning_rate": 1.6104531666924084e-06, "loss": 0.6428, "step": 9609 }, { "epoch": 0.8189177673625906, "grad_norm": 21.875, "learning_rate": 1.6102922658163575e-06, "loss": 0.8343, "step": 9610 }, { "epoch": 0.8190029825308905, "grad_norm": 16.75, "learning_rate": 1.6101313584294632e-06, "loss": 0.9247, "step": 9611 }, { "epoch": 0.8190881976991905, "grad_norm": 13.875, "learning_rate": 1.6099704445346328e-06, "loss": 0.626, "step": 9612 }, { "epoch": 0.8191734128674905, "grad_norm": 18.875, "learning_rate": 1.6098095241347744e-06, "loss": 1.1083, "step": 9613 }, { "epoch": 0.8192586280357904, "grad_norm": 18.25, "learning_rate": 1.6096485972327962e-06, "loss": 0.761, "step": 9614 }, { "epoch": 0.8193438432040904, "grad_norm": 10.8125, "learning_rate": 1.6094876638316064e-06, "loss": 0.4208, "step": 9615 }, { "epoch": 0.8194290583723903, "grad_norm": 12.1875, "learning_rate": 1.609326723934113e-06, "loss": 0.603, "step": 9616 }, { "epoch": 0.8195142735406903, "grad_norm": 14.0625, "learning_rate": 1.6091657775432243e-06, "loss": 0.7158, "step": 9617 }, { "epoch": 0.8195994887089902, "grad_norm": 14.1875, "learning_rate": 1.609004824661849e-06, "loss": 0.6577, "step": 9618 }, { "epoch": 0.8196847038772902, "grad_norm": 15.5625, "learning_rate": 1.608843865292895e-06, "loss": 0.8638, "step": 9619 }, { "epoch": 0.8197699190455902, "grad_norm": 19.75, "learning_rate": 1.608682899439272e-06, "loss": 1.058, "step": 9620 }, { "epoch": 0.8198551342138901, "grad_norm": 10.5625, "learning_rate": 1.608521927103888e-06, "loss": 0.3731, "step": 9621 }, { "epoch": 0.8199403493821901, "grad_norm": 13.0625, "learning_rate": 1.6083609482896524e-06, "loss": 0.6416, "step": 9622 }, { "epoch": 0.82002556455049, "grad_norm": 13.875, "learning_rate": 1.6081999629994743e-06, "loss": 0.564, "step": 9623 }, { "epoch": 0.82011077971879, "grad_norm": 17.75, "learning_rate": 1.608038971236262e-06, "loss": 0.8689, "step": 9624 }, { "epoch": 0.8201959948870899, "grad_norm": 16.375, "learning_rate": 1.6078779730029262e-06, "loss": 0.5513, "step": 9625 }, { "epoch": 0.8202812100553899, "grad_norm": 12.5, "learning_rate": 1.6077169683023749e-06, "loss": 0.6399, "step": 9626 }, { "epoch": 0.8203664252236899, "grad_norm": 17.25, "learning_rate": 1.6075559571375188e-06, "loss": 0.8342, "step": 9627 }, { "epoch": 0.8204516403919898, "grad_norm": 13.75, "learning_rate": 1.607394939511267e-06, "loss": 0.7329, "step": 9628 }, { "epoch": 0.8205368555602898, "grad_norm": 13.75, "learning_rate": 1.607233915426529e-06, "loss": 0.6268, "step": 9629 }, { "epoch": 0.8206220707285897, "grad_norm": 11.625, "learning_rate": 1.607072884886215e-06, "loss": 0.3986, "step": 9630 }, { "epoch": 0.8207072858968897, "grad_norm": 12.3125, "learning_rate": 1.606911847893235e-06, "loss": 0.6255, "step": 9631 }, { "epoch": 0.8207925010651896, "grad_norm": 14.125, "learning_rate": 1.606750804450499e-06, "loss": 0.7942, "step": 9632 }, { "epoch": 0.8208777162334896, "grad_norm": 25.875, "learning_rate": 1.6065897545609172e-06, "loss": 0.9177, "step": 9633 }, { "epoch": 0.8209629314017896, "grad_norm": 25.125, "learning_rate": 1.6064286982273997e-06, "loss": 1.2602, "step": 9634 }, { "epoch": 0.8210481465700895, "grad_norm": 20.125, "learning_rate": 1.6062676354528575e-06, "loss": 1.3172, "step": 9635 }, { "epoch": 0.8211333617383895, "grad_norm": 14.5, "learning_rate": 1.6061065662402008e-06, "loss": 0.5658, "step": 9636 }, { "epoch": 0.8212185769066894, "grad_norm": 15.5625, "learning_rate": 1.6059454905923405e-06, "loss": 0.7155, "step": 9637 }, { "epoch": 0.8213037920749894, "grad_norm": 14.0625, "learning_rate": 1.6057844085121868e-06, "loss": 0.7629, "step": 9638 }, { "epoch": 0.8213890072432893, "grad_norm": 17.0, "learning_rate": 1.6056233200026517e-06, "loss": 0.7474, "step": 9639 }, { "epoch": 0.8214742224115893, "grad_norm": 11.75, "learning_rate": 1.6054622250666452e-06, "loss": 0.5844, "step": 9640 }, { "epoch": 0.8215594375798893, "grad_norm": 15.4375, "learning_rate": 1.6053011237070786e-06, "loss": 0.9581, "step": 9641 }, { "epoch": 0.8216446527481892, "grad_norm": 12.125, "learning_rate": 1.6051400159268638e-06, "loss": 0.5039, "step": 9642 }, { "epoch": 0.8217298679164892, "grad_norm": 12.25, "learning_rate": 1.604978901728912e-06, "loss": 0.4849, "step": 9643 }, { "epoch": 0.8218150830847891, "grad_norm": 11.9375, "learning_rate": 1.6048177811161342e-06, "loss": 0.6487, "step": 9644 }, { "epoch": 0.8219002982530891, "grad_norm": 12.75, "learning_rate": 1.604656654091442e-06, "loss": 0.7016, "step": 9645 }, { "epoch": 0.821985513421389, "grad_norm": 13.6875, "learning_rate": 1.6044955206577479e-06, "loss": 0.8086, "step": 9646 }, { "epoch": 0.822070728589689, "grad_norm": 12.125, "learning_rate": 1.6043343808179628e-06, "loss": 0.5771, "step": 9647 }, { "epoch": 0.822155943757989, "grad_norm": 15.4375, "learning_rate": 1.604173234575e-06, "loss": 0.9748, "step": 9648 }, { "epoch": 0.8222411589262889, "grad_norm": 16.125, "learning_rate": 1.6040120819317698e-06, "loss": 0.7285, "step": 9649 }, { "epoch": 0.8223263740945889, "grad_norm": 14.375, "learning_rate": 1.6038509228911855e-06, "loss": 0.7807, "step": 9650 }, { "epoch": 0.8224115892628888, "grad_norm": 15.9375, "learning_rate": 1.6036897574561594e-06, "loss": 1.0198, "step": 9651 }, { "epoch": 0.8224968044311888, "grad_norm": 20.625, "learning_rate": 1.603528585629604e-06, "loss": 0.9035, "step": 9652 }, { "epoch": 0.8225820195994887, "grad_norm": 15.9375, "learning_rate": 1.603367407414431e-06, "loss": 0.7171, "step": 9653 }, { "epoch": 0.8226672347677887, "grad_norm": 16.5, "learning_rate": 1.6032062228135536e-06, "loss": 0.6194, "step": 9654 }, { "epoch": 0.8227524499360886, "grad_norm": 17.25, "learning_rate": 1.6030450318298847e-06, "loss": 0.886, "step": 9655 }, { "epoch": 0.8228376651043886, "grad_norm": 27.75, "learning_rate": 1.6028838344663371e-06, "loss": 1.0366, "step": 9656 }, { "epoch": 0.8229228802726886, "grad_norm": 12.5, "learning_rate": 1.6027226307258239e-06, "loss": 0.4763, "step": 9657 }, { "epoch": 0.8230080954409885, "grad_norm": 15.125, "learning_rate": 1.6025614206112583e-06, "loss": 0.5358, "step": 9658 }, { "epoch": 0.8230933106092885, "grad_norm": 17.875, "learning_rate": 1.6024002041255527e-06, "loss": 1.0267, "step": 9659 }, { "epoch": 0.8231785257775884, "grad_norm": 11.5, "learning_rate": 1.6022389812716214e-06, "loss": 0.4523, "step": 9660 }, { "epoch": 0.8232637409458884, "grad_norm": 24.625, "learning_rate": 1.6020777520523776e-06, "loss": 0.8703, "step": 9661 }, { "epoch": 0.8233489561141883, "grad_norm": 12.9375, "learning_rate": 1.6019165164707342e-06, "loss": 0.1767, "step": 9662 }, { "epoch": 0.8234341712824883, "grad_norm": 13.25, "learning_rate": 1.6017552745296061e-06, "loss": 0.6279, "step": 9663 }, { "epoch": 0.8235193864507883, "grad_norm": 15.875, "learning_rate": 1.6015940262319064e-06, "loss": 0.7645, "step": 9664 }, { "epoch": 0.8236046016190882, "grad_norm": 11.5, "learning_rate": 1.601432771580549e-06, "loss": 0.4265, "step": 9665 }, { "epoch": 0.8236898167873882, "grad_norm": 18.0, "learning_rate": 1.6012715105784477e-06, "loss": 0.6059, "step": 9666 }, { "epoch": 0.8237750319556881, "grad_norm": 19.625, "learning_rate": 1.6011102432285172e-06, "loss": 0.8687, "step": 9667 }, { "epoch": 0.8238602471239881, "grad_norm": 17.625, "learning_rate": 1.6009489695336717e-06, "loss": 0.9237, "step": 9668 }, { "epoch": 0.823945462292288, "grad_norm": 14.3125, "learning_rate": 1.6007876894968252e-06, "loss": 0.7254, "step": 9669 }, { "epoch": 0.824030677460588, "grad_norm": 14.9375, "learning_rate": 1.6006264031208923e-06, "loss": 0.8134, "step": 9670 }, { "epoch": 0.824115892628888, "grad_norm": 11.1875, "learning_rate": 1.600465110408788e-06, "loss": 0.571, "step": 9671 }, { "epoch": 0.8242011077971879, "grad_norm": 17.25, "learning_rate": 1.6003038113634268e-06, "loss": 0.892, "step": 9672 }, { "epoch": 0.8242863229654879, "grad_norm": 17.375, "learning_rate": 1.6001425059877235e-06, "loss": 0.7139, "step": 9673 }, { "epoch": 0.8243715381337878, "grad_norm": 14.125, "learning_rate": 1.5999811942845924e-06, "loss": 0.6816, "step": 9674 }, { "epoch": 0.8244567533020878, "grad_norm": 12.6875, "learning_rate": 1.59981987625695e-06, "loss": 0.4999, "step": 9675 }, { "epoch": 0.8245419684703877, "grad_norm": 37.75, "learning_rate": 1.5996585519077102e-06, "loss": 1.0457, "step": 9676 }, { "epoch": 0.8246271836386877, "grad_norm": 21.375, "learning_rate": 1.5994972212397885e-06, "loss": 0.7863, "step": 9677 }, { "epoch": 0.8247123988069877, "grad_norm": 14.5, "learning_rate": 1.599335884256101e-06, "loss": 0.5265, "step": 9678 }, { "epoch": 0.8247976139752876, "grad_norm": 17.875, "learning_rate": 1.5991745409595628e-06, "loss": 0.689, "step": 9679 }, { "epoch": 0.8248828291435876, "grad_norm": 13.3125, "learning_rate": 1.5990131913530897e-06, "loss": 0.6493, "step": 9680 }, { "epoch": 0.8249680443118875, "grad_norm": 14.75, "learning_rate": 1.5988518354395965e-06, "loss": 0.5115, "step": 9681 }, { "epoch": 0.8250532594801875, "grad_norm": 14.125, "learning_rate": 1.5986904732220005e-06, "loss": 0.8284, "step": 9682 }, { "epoch": 0.8251384746484874, "grad_norm": 17.25, "learning_rate": 1.5985291047032166e-06, "loss": 0.928, "step": 9683 }, { "epoch": 0.8252236898167874, "grad_norm": 21.875, "learning_rate": 1.5983677298861622e-06, "loss": 0.8874, "step": 9684 }, { "epoch": 0.8253089049850874, "grad_norm": 13.5, "learning_rate": 1.5982063487737514e-06, "loss": 0.565, "step": 9685 }, { "epoch": 0.8253941201533873, "grad_norm": 11.875, "learning_rate": 1.5980449613689025e-06, "loss": 0.4842, "step": 9686 }, { "epoch": 0.8254793353216873, "grad_norm": 14.3125, "learning_rate": 1.5978835676745309e-06, "loss": 0.8157, "step": 9687 }, { "epoch": 0.8255645504899872, "grad_norm": 20.375, "learning_rate": 1.5977221676935536e-06, "loss": 0.4956, "step": 9688 }, { "epoch": 0.8256497656582872, "grad_norm": 18.125, "learning_rate": 1.597560761428887e-06, "loss": 0.79, "step": 9689 }, { "epoch": 0.8257349808265871, "grad_norm": 13.5, "learning_rate": 1.5973993488834476e-06, "loss": 0.619, "step": 9690 }, { "epoch": 0.8258201959948871, "grad_norm": 12.125, "learning_rate": 1.597237930060153e-06, "loss": 0.6217, "step": 9691 }, { "epoch": 0.825905411163187, "grad_norm": 12.875, "learning_rate": 1.5970765049619197e-06, "loss": 0.6151, "step": 9692 }, { "epoch": 0.825990626331487, "grad_norm": 12.875, "learning_rate": 1.596915073591665e-06, "loss": 0.6642, "step": 9693 }, { "epoch": 0.826075841499787, "grad_norm": 10.5625, "learning_rate": 1.596753635952306e-06, "loss": 0.3249, "step": 9694 }, { "epoch": 0.8261610566680869, "grad_norm": 13.375, "learning_rate": 1.59659219204676e-06, "loss": 0.968, "step": 9695 }, { "epoch": 0.8262462718363869, "grad_norm": 17.5, "learning_rate": 1.5964307418779445e-06, "loss": 1.0474, "step": 9696 }, { "epoch": 0.8263314870046868, "grad_norm": 18.25, "learning_rate": 1.5962692854487777e-06, "loss": 0.9244, "step": 9697 }, { "epoch": 0.8264167021729868, "grad_norm": 14.8125, "learning_rate": 1.5961078227621762e-06, "loss": 0.9517, "step": 9698 }, { "epoch": 0.8265019173412868, "grad_norm": 20.125, "learning_rate": 1.5959463538210584e-06, "loss": 0.89, "step": 9699 }, { "epoch": 0.8265871325095867, "grad_norm": 10.25, "learning_rate": 1.595784878628342e-06, "loss": 0.2895, "step": 9700 }, { "epoch": 0.8266723476778867, "grad_norm": 17.0, "learning_rate": 1.5956233971869455e-06, "loss": 0.8022, "step": 9701 }, { "epoch": 0.8267575628461866, "grad_norm": 12.75, "learning_rate": 1.5954619094997865e-06, "loss": 0.7376, "step": 9702 }, { "epoch": 0.8268427780144866, "grad_norm": 20.5, "learning_rate": 1.5953004155697835e-06, "loss": 1.0007, "step": 9703 }, { "epoch": 0.8269279931827865, "grad_norm": 12.8125, "learning_rate": 1.5951389153998542e-06, "loss": 0.7068, "step": 9704 }, { "epoch": 0.8270132083510865, "grad_norm": 12.9375, "learning_rate": 1.5949774089929185e-06, "loss": 0.643, "step": 9705 }, { "epoch": 0.8270984235193864, "grad_norm": 17.875, "learning_rate": 1.5948158963518937e-06, "loss": 1.1723, "step": 9706 }, { "epoch": 0.8271836386876864, "grad_norm": 22.875, "learning_rate": 1.5946543774796988e-06, "loss": 0.9569, "step": 9707 }, { "epoch": 0.8272688538559864, "grad_norm": 16.0, "learning_rate": 1.5944928523792527e-06, "loss": 0.6086, "step": 9708 }, { "epoch": 0.8273540690242863, "grad_norm": 17.125, "learning_rate": 1.5943313210534744e-06, "loss": 0.9288, "step": 9709 }, { "epoch": 0.8274392841925863, "grad_norm": 12.75, "learning_rate": 1.594169783505283e-06, "loss": 0.4042, "step": 9710 }, { "epoch": 0.8275244993608862, "grad_norm": 13.0625, "learning_rate": 1.5940082397375972e-06, "loss": 0.7382, "step": 9711 }, { "epoch": 0.8276097145291862, "grad_norm": 12.6875, "learning_rate": 1.5938466897533373e-06, "loss": 0.6501, "step": 9712 }, { "epoch": 0.8276949296974861, "grad_norm": 14.125, "learning_rate": 1.5936851335554212e-06, "loss": 0.442, "step": 9713 }, { "epoch": 0.8277801448657861, "grad_norm": 16.875, "learning_rate": 1.5935235711467696e-06, "loss": 0.79, "step": 9714 }, { "epoch": 0.8278653600340861, "grad_norm": 14.8125, "learning_rate": 1.5933620025303012e-06, "loss": 0.7064, "step": 9715 }, { "epoch": 0.827950575202386, "grad_norm": 16.125, "learning_rate": 1.5932004277089365e-06, "loss": 0.8353, "step": 9716 }, { "epoch": 0.828035790370686, "grad_norm": 29.875, "learning_rate": 1.593038846685595e-06, "loss": 1.2232, "step": 9717 }, { "epoch": 0.8281210055389859, "grad_norm": 11.875, "learning_rate": 1.5928772594631964e-06, "loss": 0.5594, "step": 9718 }, { "epoch": 0.8282062207072859, "grad_norm": 19.875, "learning_rate": 1.5927156660446608e-06, "loss": 0.6795, "step": 9719 }, { "epoch": 0.8282914358755858, "grad_norm": 13.0625, "learning_rate": 1.5925540664329086e-06, "loss": 0.5582, "step": 9720 }, { "epoch": 0.8283766510438858, "grad_norm": 11.8125, "learning_rate": 1.5923924606308601e-06, "loss": 0.3808, "step": 9721 }, { "epoch": 0.8284618662121858, "grad_norm": 14.5, "learning_rate": 1.5922308486414357e-06, "loss": 0.6694, "step": 9722 }, { "epoch": 0.8285470813804857, "grad_norm": 17.5, "learning_rate": 1.5920692304675554e-06, "loss": 0.8276, "step": 9723 }, { "epoch": 0.8286322965487857, "grad_norm": 32.25, "learning_rate": 1.5919076061121402e-06, "loss": 0.861, "step": 9724 }, { "epoch": 0.8287175117170856, "grad_norm": 20.875, "learning_rate": 1.591745975578111e-06, "loss": 0.6563, "step": 9725 }, { "epoch": 0.8288027268853856, "grad_norm": 25.875, "learning_rate": 1.5915843388683878e-06, "loss": 1.0286, "step": 9726 }, { "epoch": 0.8288879420536855, "grad_norm": 33.75, "learning_rate": 1.5914226959858925e-06, "loss": 1.3663, "step": 9727 }, { "epoch": 0.8289731572219855, "grad_norm": 13.625, "learning_rate": 1.5912610469335458e-06, "loss": 0.7798, "step": 9728 }, { "epoch": 0.8290583723902855, "grad_norm": 16.25, "learning_rate": 1.5910993917142687e-06, "loss": 0.9191, "step": 9729 }, { "epoch": 0.8291435875585854, "grad_norm": 19.125, "learning_rate": 1.5909377303309823e-06, "loss": 1.1412, "step": 9730 }, { "epoch": 0.8292288027268854, "grad_norm": 12.6875, "learning_rate": 1.5907760627866087e-06, "loss": 0.4271, "step": 9731 }, { "epoch": 0.8293140178951853, "grad_norm": 14.375, "learning_rate": 1.5906143890840687e-06, "loss": 0.4894, "step": 9732 }, { "epoch": 0.8293992330634853, "grad_norm": 10.625, "learning_rate": 1.5904527092262846e-06, "loss": 1.027, "step": 9733 }, { "epoch": 0.8294844482317852, "grad_norm": 18.125, "learning_rate": 1.5902910232161772e-06, "loss": 0.9958, "step": 9734 }, { "epoch": 0.8295696634000852, "grad_norm": 21.125, "learning_rate": 1.590129331056669e-06, "loss": 0.8219, "step": 9735 }, { "epoch": 0.8296548785683852, "grad_norm": 20.25, "learning_rate": 1.5899676327506816e-06, "loss": 0.572, "step": 9736 }, { "epoch": 0.8297400937366851, "grad_norm": 18.5, "learning_rate": 1.5898059283011377e-06, "loss": 0.9582, "step": 9737 }, { "epoch": 0.8298253089049851, "grad_norm": 13.9375, "learning_rate": 1.5896442177109584e-06, "loss": 0.5103, "step": 9738 }, { "epoch": 0.829910524073285, "grad_norm": 17.875, "learning_rate": 1.589482500983067e-06, "loss": 0.7345, "step": 9739 }, { "epoch": 0.829995739241585, "grad_norm": 11.75, "learning_rate": 1.5893207781203851e-06, "loss": 0.58, "step": 9740 }, { "epoch": 0.8300809544098849, "grad_norm": 14.8125, "learning_rate": 1.5891590491258358e-06, "loss": 0.8765, "step": 9741 }, { "epoch": 0.8301661695781849, "grad_norm": 14.25, "learning_rate": 1.5889973140023414e-06, "loss": 0.6009, "step": 9742 }, { "epoch": 0.8302513847464849, "grad_norm": 12.9375, "learning_rate": 1.5888355727528243e-06, "loss": 0.7213, "step": 9743 }, { "epoch": 0.8303365999147848, "grad_norm": 12.4375, "learning_rate": 1.5886738253802078e-06, "loss": 0.5879, "step": 9744 }, { "epoch": 0.8304218150830848, "grad_norm": 15.75, "learning_rate": 1.588512071887415e-06, "loss": 0.9383, "step": 9745 }, { "epoch": 0.8305070302513847, "grad_norm": 17.0, "learning_rate": 1.5883503122773686e-06, "loss": 0.7263, "step": 9746 }, { "epoch": 0.8305922454196847, "grad_norm": 15.5625, "learning_rate": 1.5881885465529914e-06, "loss": 0.9516, "step": 9747 }, { "epoch": 0.8306774605879846, "grad_norm": 17.375, "learning_rate": 1.5880267747172076e-06, "loss": 0.5601, "step": 9748 }, { "epoch": 0.8307626757562846, "grad_norm": 13.625, "learning_rate": 1.5878649967729397e-06, "loss": 0.7941, "step": 9749 }, { "epoch": 0.8308478909245846, "grad_norm": 15.5625, "learning_rate": 1.587703212723112e-06, "loss": 0.7843, "step": 9750 }, { "epoch": 0.8309331060928845, "grad_norm": 10.6875, "learning_rate": 1.5875414225706471e-06, "loss": 0.4843, "step": 9751 }, { "epoch": 0.8310183212611845, "grad_norm": 20.5, "learning_rate": 1.5873796263184698e-06, "loss": 0.7809, "step": 9752 }, { "epoch": 0.8311035364294844, "grad_norm": 12.375, "learning_rate": 1.5872178239695029e-06, "loss": 0.5725, "step": 9753 }, { "epoch": 0.8311887515977844, "grad_norm": 17.125, "learning_rate": 1.5870560155266711e-06, "loss": 0.9388, "step": 9754 }, { "epoch": 0.8312739667660843, "grad_norm": 22.75, "learning_rate": 1.5868942009928983e-06, "loss": 1.2478, "step": 9755 }, { "epoch": 0.8313591819343843, "grad_norm": 14.25, "learning_rate": 1.5867323803711081e-06, "loss": 0.7398, "step": 9756 }, { "epoch": 0.8314443971026843, "grad_norm": 17.5, "learning_rate": 1.5865705536642257e-06, "loss": 0.8815, "step": 9757 }, { "epoch": 0.8315296122709842, "grad_norm": 24.375, "learning_rate": 1.586408720875175e-06, "loss": 1.1097, "step": 9758 }, { "epoch": 0.8316148274392842, "grad_norm": 11.5625, "learning_rate": 1.5862468820068796e-06, "loss": 0.5781, "step": 9759 }, { "epoch": 0.8317000426075841, "grad_norm": 14.625, "learning_rate": 1.5860850370622654e-06, "loss": 0.3199, "step": 9760 }, { "epoch": 0.8317852577758841, "grad_norm": 10.875, "learning_rate": 1.5859231860442568e-06, "loss": 0.5379, "step": 9761 }, { "epoch": 0.831870472944184, "grad_norm": 14.0, "learning_rate": 1.5857613289557782e-06, "loss": 0.5074, "step": 9762 }, { "epoch": 0.831955688112484, "grad_norm": 18.875, "learning_rate": 1.5855994657997547e-06, "loss": 0.9243, "step": 9763 }, { "epoch": 0.832040903280784, "grad_norm": 15.75, "learning_rate": 1.5854375965791112e-06, "loss": 0.9017, "step": 9764 }, { "epoch": 0.8321261184490839, "grad_norm": 21.25, "learning_rate": 1.5852757212967737e-06, "loss": 0.9408, "step": 9765 }, { "epoch": 0.8322113336173839, "grad_norm": 12.3125, "learning_rate": 1.5851138399556664e-06, "loss": 0.4767, "step": 9766 }, { "epoch": 0.8322965487856838, "grad_norm": 10.5, "learning_rate": 1.584951952558715e-06, "loss": 0.475, "step": 9767 }, { "epoch": 0.8323817639539838, "grad_norm": 18.125, "learning_rate": 1.5847900591088449e-06, "loss": 1.1705, "step": 9768 }, { "epoch": 0.8324669791222837, "grad_norm": 13.9375, "learning_rate": 1.5846281596089822e-06, "loss": 0.7003, "step": 9769 }, { "epoch": 0.8325521942905837, "grad_norm": 13.5, "learning_rate": 1.5844662540620517e-06, "loss": 0.7125, "step": 9770 }, { "epoch": 0.8326374094588836, "grad_norm": 11.3125, "learning_rate": 1.58430434247098e-06, "loss": 0.3365, "step": 9771 }, { "epoch": 0.8327226246271836, "grad_norm": 12.5, "learning_rate": 1.5841424248386922e-06, "loss": 0.6143, "step": 9772 }, { "epoch": 0.8328078397954836, "grad_norm": 14.1875, "learning_rate": 1.5839805011681154e-06, "loss": 0.6069, "step": 9773 }, { "epoch": 0.8328930549637835, "grad_norm": 21.25, "learning_rate": 1.5838185714621745e-06, "loss": 0.653, "step": 9774 }, { "epoch": 0.8329782701320835, "grad_norm": 17.375, "learning_rate": 1.5836566357237968e-06, "loss": 0.8936, "step": 9775 }, { "epoch": 0.8330634853003834, "grad_norm": 14.75, "learning_rate": 1.5834946939559079e-06, "loss": 0.5792, "step": 9776 }, { "epoch": 0.8331487004686834, "grad_norm": 16.25, "learning_rate": 1.5833327461614345e-06, "loss": 1.1301, "step": 9777 }, { "epoch": 0.8332339156369833, "grad_norm": 15.3125, "learning_rate": 1.5831707923433034e-06, "loss": 0.8825, "step": 9778 }, { "epoch": 0.8333191308052833, "grad_norm": 15.4375, "learning_rate": 1.5830088325044407e-06, "loss": 0.7664, "step": 9779 }, { "epoch": 0.8334043459735833, "grad_norm": 9.5625, "learning_rate": 1.5828468666477738e-06, "loss": 0.2657, "step": 9780 }, { "epoch": 0.8334895611418832, "grad_norm": 13.25, "learning_rate": 1.5826848947762292e-06, "loss": 0.5132, "step": 9781 }, { "epoch": 0.8335747763101832, "grad_norm": 11.1875, "learning_rate": 1.5825229168927344e-06, "loss": 0.4001, "step": 9782 }, { "epoch": 0.8336599914784831, "grad_norm": 14.875, "learning_rate": 1.5823609330002155e-06, "loss": 0.7679, "step": 9783 }, { "epoch": 0.8337452066467831, "grad_norm": 11.0, "learning_rate": 1.5821989431016007e-06, "loss": 0.5385, "step": 9784 }, { "epoch": 0.833830421815083, "grad_norm": 12.5, "learning_rate": 1.5820369471998167e-06, "loss": 0.5615, "step": 9785 }, { "epoch": 0.833915636983383, "grad_norm": 13.0, "learning_rate": 1.5818749452977917e-06, "loss": 0.6283, "step": 9786 }, { "epoch": 0.834000852151683, "grad_norm": 19.375, "learning_rate": 1.5817129373984522e-06, "loss": 0.2593, "step": 9787 }, { "epoch": 0.8340860673199829, "grad_norm": 13.75, "learning_rate": 1.5815509235047266e-06, "loss": 0.6131, "step": 9788 }, { "epoch": 0.8341712824882829, "grad_norm": 16.5, "learning_rate": 1.581388903619542e-06, "loss": 0.664, "step": 9789 }, { "epoch": 0.8342564976565828, "grad_norm": 24.25, "learning_rate": 1.5812268777458275e-06, "loss": 0.8757, "step": 9790 }, { "epoch": 0.8343417128248828, "grad_norm": 20.0, "learning_rate": 1.5810648458865096e-06, "loss": 1.0316, "step": 9791 }, { "epoch": 0.8344269279931827, "grad_norm": 12.5, "learning_rate": 1.580902808044517e-06, "loss": 0.5827, "step": 9792 }, { "epoch": 0.8345121431614827, "grad_norm": 13.875, "learning_rate": 1.5807407642227782e-06, "loss": 0.6134, "step": 9793 }, { "epoch": 0.8345973583297827, "grad_norm": 15.1875, "learning_rate": 1.580578714424221e-06, "loss": 0.9598, "step": 9794 }, { "epoch": 0.8346825734980826, "grad_norm": 15.875, "learning_rate": 1.580416658651774e-06, "loss": 0.9506, "step": 9795 }, { "epoch": 0.8347677886663826, "grad_norm": 9.5625, "learning_rate": 1.5802545969083657e-06, "loss": 0.363, "step": 9796 }, { "epoch": 0.8348530038346825, "grad_norm": 18.625, "learning_rate": 1.5800925291969248e-06, "loss": 0.8018, "step": 9797 }, { "epoch": 0.8349382190029825, "grad_norm": 16.5, "learning_rate": 1.5799304555203804e-06, "loss": 1.0559, "step": 9798 }, { "epoch": 0.8350234341712824, "grad_norm": 19.125, "learning_rate": 1.5797683758816606e-06, "loss": 0.8637, "step": 9799 }, { "epoch": 0.8351086493395824, "grad_norm": 14.75, "learning_rate": 1.5796062902836942e-06, "loss": 0.7914, "step": 9800 }, { "epoch": 0.8351938645078824, "grad_norm": 9.0625, "learning_rate": 1.5794441987294113e-06, "loss": 0.5058, "step": 9801 }, { "epoch": 0.8352790796761823, "grad_norm": 14.9375, "learning_rate": 1.5792821012217401e-06, "loss": 0.6316, "step": 9802 }, { "epoch": 0.8353642948444823, "grad_norm": 14.6875, "learning_rate": 1.5791199977636104e-06, "loss": 0.7757, "step": 9803 }, { "epoch": 0.8354495100127822, "grad_norm": 13.0, "learning_rate": 1.5789578883579512e-06, "loss": 0.4614, "step": 9804 }, { "epoch": 0.8355347251810822, "grad_norm": 10.4375, "learning_rate": 1.5787957730076922e-06, "loss": 0.3779, "step": 9805 }, { "epoch": 0.8356199403493821, "grad_norm": 12.8125, "learning_rate": 1.578633651715763e-06, "loss": 0.5543, "step": 9806 }, { "epoch": 0.8357051555176821, "grad_norm": 13.875, "learning_rate": 1.5784715244850935e-06, "loss": 0.5584, "step": 9807 }, { "epoch": 0.835790370685982, "grad_norm": 16.5, "learning_rate": 1.5783093913186126e-06, "loss": 1.0975, "step": 9808 }, { "epoch": 0.835875585854282, "grad_norm": 39.5, "learning_rate": 1.5781472522192514e-06, "loss": 0.9335, "step": 9809 }, { "epoch": 0.835960801022582, "grad_norm": 15.6875, "learning_rate": 1.5779851071899394e-06, "loss": 0.8782, "step": 9810 }, { "epoch": 0.8360460161908819, "grad_norm": 12.625, "learning_rate": 1.5778229562336064e-06, "loss": 0.6672, "step": 9811 }, { "epoch": 0.8361312313591819, "grad_norm": 12.625, "learning_rate": 1.577660799353183e-06, "loss": 0.4586, "step": 9812 }, { "epoch": 0.8362164465274818, "grad_norm": 13.75, "learning_rate": 1.5774986365515998e-06, "loss": 0.6113, "step": 9813 }, { "epoch": 0.8363016616957818, "grad_norm": 14.8125, "learning_rate": 1.5773364678317865e-06, "loss": 0.5669, "step": 9814 }, { "epoch": 0.8363868768640818, "grad_norm": 12.9375, "learning_rate": 1.577174293196674e-06, "loss": 0.7405, "step": 9815 }, { "epoch": 0.8364720920323817, "grad_norm": 15.5625, "learning_rate": 1.5770121126491934e-06, "loss": 0.9952, "step": 9816 }, { "epoch": 0.8365573072006818, "grad_norm": 18.375, "learning_rate": 1.576849926192275e-06, "loss": 0.8584, "step": 9817 }, { "epoch": 0.8366425223689817, "grad_norm": 19.125, "learning_rate": 1.5766877338288496e-06, "loss": 0.9105, "step": 9818 }, { "epoch": 0.8367277375372817, "grad_norm": 13.3125, "learning_rate": 1.5765255355618484e-06, "loss": 0.5223, "step": 9819 }, { "epoch": 0.8368129527055816, "grad_norm": 22.25, "learning_rate": 1.5763633313942025e-06, "loss": 1.3057, "step": 9820 }, { "epoch": 0.8368981678738816, "grad_norm": 17.75, "learning_rate": 1.5762011213288427e-06, "loss": 1.035, "step": 9821 }, { "epoch": 0.8369833830421816, "grad_norm": 14.6875, "learning_rate": 1.5760389053687014e-06, "loss": 0.3214, "step": 9822 }, { "epoch": 0.8370685982104815, "grad_norm": 15.5, "learning_rate": 1.5758766835167083e-06, "loss": 0.4658, "step": 9823 }, { "epoch": 0.8371538133787815, "grad_norm": 12.1875, "learning_rate": 1.5757144557757965e-06, "loss": 0.5965, "step": 9824 }, { "epoch": 0.8372390285470814, "grad_norm": 20.25, "learning_rate": 1.5755522221488964e-06, "loss": 0.8638, "step": 9825 }, { "epoch": 0.8373242437153814, "grad_norm": 12.0625, "learning_rate": 1.5753899826389407e-06, "loss": 0.6263, "step": 9826 }, { "epoch": 0.8374094588836813, "grad_norm": 13.0625, "learning_rate": 1.5752277372488605e-06, "loss": 0.6863, "step": 9827 }, { "epoch": 0.8374946740519813, "grad_norm": 21.875, "learning_rate": 1.5750654859815884e-06, "loss": 0.6349, "step": 9828 }, { "epoch": 0.8375798892202813, "grad_norm": 14.4375, "learning_rate": 1.5749032288400557e-06, "loss": 0.6991, "step": 9829 }, { "epoch": 0.8376651043885812, "grad_norm": 14.3125, "learning_rate": 1.5747409658271954e-06, "loss": 0.9973, "step": 9830 }, { "epoch": 0.8377503195568812, "grad_norm": 18.875, "learning_rate": 1.574578696945939e-06, "loss": 0.7255, "step": 9831 }, { "epoch": 0.8378355347251811, "grad_norm": 14.4375, "learning_rate": 1.574416422199219e-06, "loss": 0.8038, "step": 9832 }, { "epoch": 0.8379207498934811, "grad_norm": 19.875, "learning_rate": 1.574254141589968e-06, "loss": 1.0336, "step": 9833 }, { "epoch": 0.838005965061781, "grad_norm": 15.625, "learning_rate": 1.574091855121119e-06, "loss": 0.7533, "step": 9834 }, { "epoch": 0.838091180230081, "grad_norm": 14.3125, "learning_rate": 1.5739295627956045e-06, "loss": 0.5521, "step": 9835 }, { "epoch": 0.838176395398381, "grad_norm": 23.5, "learning_rate": 1.5737672646163564e-06, "loss": 0.8693, "step": 9836 }, { "epoch": 0.8382616105666809, "grad_norm": 12.8125, "learning_rate": 1.5736049605863084e-06, "loss": 0.8098, "step": 9837 }, { "epoch": 0.8383468257349809, "grad_norm": 13.75, "learning_rate": 1.5734426507083933e-06, "loss": 0.6288, "step": 9838 }, { "epoch": 0.8384320409032808, "grad_norm": 17.5, "learning_rate": 1.573280334985545e-06, "loss": 0.9859, "step": 9839 }, { "epoch": 0.8385172560715808, "grad_norm": 17.875, "learning_rate": 1.5731180134206953e-06, "loss": 0.8724, "step": 9840 }, { "epoch": 0.8386024712398807, "grad_norm": 17.625, "learning_rate": 1.5729556860167782e-06, "loss": 0.9944, "step": 9841 }, { "epoch": 0.8386876864081807, "grad_norm": 15.6875, "learning_rate": 1.572793352776727e-06, "loss": 0.6638, "step": 9842 }, { "epoch": 0.8387729015764807, "grad_norm": 18.75, "learning_rate": 1.5726310137034756e-06, "loss": 1.1067, "step": 9843 }, { "epoch": 0.8388581167447806, "grad_norm": 13.0, "learning_rate": 1.5724686687999572e-06, "loss": 0.5252, "step": 9844 }, { "epoch": 0.8389433319130806, "grad_norm": 30.75, "learning_rate": 1.5723063180691058e-06, "loss": 0.9885, "step": 9845 }, { "epoch": 0.8390285470813805, "grad_norm": 25.125, "learning_rate": 1.5721439615138547e-06, "loss": 0.8916, "step": 9846 }, { "epoch": 0.8391137622496805, "grad_norm": 16.875, "learning_rate": 1.571981599137139e-06, "loss": 1.0991, "step": 9847 }, { "epoch": 0.8391989774179804, "grad_norm": 15.75, "learning_rate": 1.5718192309418918e-06, "loss": 1.0177, "step": 9848 }, { "epoch": 0.8392841925862804, "grad_norm": 14.1875, "learning_rate": 1.5716568569310469e-06, "loss": 0.704, "step": 9849 }, { "epoch": 0.8393694077545804, "grad_norm": 15.4375, "learning_rate": 1.5714944771075396e-06, "loss": 0.897, "step": 9850 }, { "epoch": 0.8394546229228803, "grad_norm": 19.5, "learning_rate": 1.571332091474304e-06, "loss": 1.0276, "step": 9851 }, { "epoch": 0.8395398380911803, "grad_norm": 17.25, "learning_rate": 1.571169700034274e-06, "loss": 0.7457, "step": 9852 }, { "epoch": 0.8396250532594802, "grad_norm": 36.75, "learning_rate": 1.5710073027903848e-06, "loss": 0.8826, "step": 9853 }, { "epoch": 0.8397102684277802, "grad_norm": 18.0, "learning_rate": 1.5708448997455709e-06, "loss": 0.9323, "step": 9854 }, { "epoch": 0.8397954835960801, "grad_norm": 20.75, "learning_rate": 1.5706824909027667e-06, "loss": 0.9471, "step": 9855 }, { "epoch": 0.8398806987643801, "grad_norm": 9.75, "learning_rate": 1.5705200762649079e-06, "loss": 0.3089, "step": 9856 }, { "epoch": 0.83996591393268, "grad_norm": 19.375, "learning_rate": 1.5703576558349285e-06, "loss": 0.7622, "step": 9857 }, { "epoch": 0.84005112910098, "grad_norm": 14.0625, "learning_rate": 1.5701952296157643e-06, "loss": 0.7318, "step": 9858 }, { "epoch": 0.84013634426928, "grad_norm": 18.0, "learning_rate": 1.5700327976103504e-06, "loss": 0.9117, "step": 9859 }, { "epoch": 0.8402215594375799, "grad_norm": 12.875, "learning_rate": 1.5698703598216223e-06, "loss": 0.6294, "step": 9860 }, { "epoch": 0.8403067746058799, "grad_norm": 18.0, "learning_rate": 1.5697079162525148e-06, "loss": 0.6508, "step": 9861 }, { "epoch": 0.8403919897741798, "grad_norm": 10.125, "learning_rate": 1.5695454669059637e-06, "loss": 0.2965, "step": 9862 }, { "epoch": 0.8404772049424798, "grad_norm": 14.3125, "learning_rate": 1.5693830117849048e-06, "loss": 0.7872, "step": 9863 }, { "epoch": 0.8405624201107798, "grad_norm": 15.0625, "learning_rate": 1.5692205508922736e-06, "loss": 1.0676, "step": 9864 }, { "epoch": 0.8406476352790797, "grad_norm": 19.5, "learning_rate": 1.569058084231006e-06, "loss": 1.0584, "step": 9865 }, { "epoch": 0.8407328504473797, "grad_norm": 18.375, "learning_rate": 1.5688956118040382e-06, "loss": 0.7414, "step": 9866 }, { "epoch": 0.8408180656156796, "grad_norm": 10.25, "learning_rate": 1.5687331336143063e-06, "loss": 0.3077, "step": 9867 }, { "epoch": 0.8409032807839796, "grad_norm": 19.625, "learning_rate": 1.5685706496647454e-06, "loss": 0.7172, "step": 9868 }, { "epoch": 0.8409884959522795, "grad_norm": 16.125, "learning_rate": 1.5684081599582932e-06, "loss": 0.802, "step": 9869 }, { "epoch": 0.8410737111205795, "grad_norm": 21.75, "learning_rate": 1.568245664497885e-06, "loss": 1.1287, "step": 9870 }, { "epoch": 0.8411589262888794, "grad_norm": 15.9375, "learning_rate": 1.568083163286458e-06, "loss": 0.8503, "step": 9871 }, { "epoch": 0.8412441414571794, "grad_norm": 15.9375, "learning_rate": 1.5679206563269478e-06, "loss": 0.4853, "step": 9872 }, { "epoch": 0.8413293566254794, "grad_norm": 13.3125, "learning_rate": 1.567758143622292e-06, "loss": 0.6183, "step": 9873 }, { "epoch": 0.8414145717937793, "grad_norm": 15.0625, "learning_rate": 1.5675956251754268e-06, "loss": 0.6364, "step": 9874 }, { "epoch": 0.8414997869620793, "grad_norm": 17.625, "learning_rate": 1.5674331009892896e-06, "loss": 0.4606, "step": 9875 }, { "epoch": 0.8415850021303792, "grad_norm": 12.8125, "learning_rate": 1.5672705710668168e-06, "loss": 0.7307, "step": 9876 }, { "epoch": 0.8416702172986792, "grad_norm": 19.375, "learning_rate": 1.567108035410946e-06, "loss": 0.7943, "step": 9877 }, { "epoch": 0.8417554324669791, "grad_norm": 16.125, "learning_rate": 1.5669454940246138e-06, "loss": 0.9012, "step": 9878 }, { "epoch": 0.8418406476352791, "grad_norm": 19.5, "learning_rate": 1.566782946910758e-06, "loss": 1.0703, "step": 9879 }, { "epoch": 0.8419258628035791, "grad_norm": 11.1875, "learning_rate": 1.5666203940723157e-06, "loss": 0.4871, "step": 9880 }, { "epoch": 0.842011077971879, "grad_norm": 14.5, "learning_rate": 1.5664578355122245e-06, "loss": 0.7322, "step": 9881 }, { "epoch": 0.842096293140179, "grad_norm": 15.1875, "learning_rate": 1.566295271233422e-06, "loss": 0.7332, "step": 9882 }, { "epoch": 0.8421815083084789, "grad_norm": 12.5, "learning_rate": 1.566132701238846e-06, "loss": 0.7518, "step": 9883 }, { "epoch": 0.8422667234767789, "grad_norm": 13.6875, "learning_rate": 1.5659701255314345e-06, "loss": 0.5796, "step": 9884 }, { "epoch": 0.8423519386450788, "grad_norm": 12.375, "learning_rate": 1.5658075441141243e-06, "loss": 0.4041, "step": 9885 }, { "epoch": 0.8424371538133788, "grad_norm": 15.6875, "learning_rate": 1.5656449569898547e-06, "loss": 0.768, "step": 9886 }, { "epoch": 0.8425223689816788, "grad_norm": 14.5625, "learning_rate": 1.5654823641615636e-06, "loss": 0.672, "step": 9887 }, { "epoch": 0.8426075841499787, "grad_norm": 12.375, "learning_rate": 1.5653197656321886e-06, "loss": 0.6673, "step": 9888 }, { "epoch": 0.8426927993182787, "grad_norm": 12.0, "learning_rate": 1.565157161404668e-06, "loss": 0.6916, "step": 9889 }, { "epoch": 0.8427780144865786, "grad_norm": 13.0625, "learning_rate": 1.564994551481941e-06, "loss": 0.6301, "step": 9890 }, { "epoch": 0.8428632296548786, "grad_norm": 10.5625, "learning_rate": 1.5648319358669458e-06, "loss": 0.3635, "step": 9891 }, { "epoch": 0.8429484448231785, "grad_norm": 14.875, "learning_rate": 1.5646693145626208e-06, "loss": 0.8302, "step": 9892 }, { "epoch": 0.8430336599914785, "grad_norm": 9.8125, "learning_rate": 1.5645066875719048e-06, "loss": 0.2821, "step": 9893 }, { "epoch": 0.8431188751597785, "grad_norm": 20.5, "learning_rate": 1.5643440548977367e-06, "loss": 1.1424, "step": 9894 }, { "epoch": 0.8432040903280784, "grad_norm": 13.125, "learning_rate": 1.5641814165430552e-06, "loss": 0.8648, "step": 9895 }, { "epoch": 0.8432893054963784, "grad_norm": 15.8125, "learning_rate": 1.5640187725108e-06, "loss": 0.8644, "step": 9896 }, { "epoch": 0.8433745206646783, "grad_norm": 19.125, "learning_rate": 1.56385612280391e-06, "loss": 0.9258, "step": 9897 }, { "epoch": 0.8434597358329783, "grad_norm": 13.75, "learning_rate": 1.5636934674253237e-06, "loss": 0.5752, "step": 9898 }, { "epoch": 0.8435449510012782, "grad_norm": 20.125, "learning_rate": 1.563530806377981e-06, "loss": 0.4719, "step": 9899 }, { "epoch": 0.8436301661695782, "grad_norm": 13.125, "learning_rate": 1.5633681396648214e-06, "loss": 0.343, "step": 9900 }, { "epoch": 0.8437153813378782, "grad_norm": 14.125, "learning_rate": 1.5632054672887848e-06, "loss": 0.815, "step": 9901 }, { "epoch": 0.8438005965061781, "grad_norm": 17.25, "learning_rate": 1.56304278925281e-06, "loss": 0.9937, "step": 9902 }, { "epoch": 0.8438858116744781, "grad_norm": 14.0, "learning_rate": 1.5628801055598376e-06, "loss": 0.6641, "step": 9903 }, { "epoch": 0.843971026842778, "grad_norm": 14.0, "learning_rate": 1.5627174162128066e-06, "loss": 0.9075, "step": 9904 }, { "epoch": 0.844056242011078, "grad_norm": 10.25, "learning_rate": 1.562554721214658e-06, "loss": 0.5418, "step": 9905 }, { "epoch": 0.8441414571793779, "grad_norm": 20.625, "learning_rate": 1.5623920205683307e-06, "loss": 0.8552, "step": 9906 }, { "epoch": 0.8442266723476779, "grad_norm": 34.5, "learning_rate": 1.5622293142767658e-06, "loss": 0.9319, "step": 9907 }, { "epoch": 0.8443118875159779, "grad_norm": 13.3125, "learning_rate": 1.5620666023429035e-06, "loss": 0.6535, "step": 9908 }, { "epoch": 0.8443971026842778, "grad_norm": 17.375, "learning_rate": 1.5619038847696835e-06, "loss": 0.6715, "step": 9909 }, { "epoch": 0.8444823178525778, "grad_norm": 18.625, "learning_rate": 1.5617411615600465e-06, "loss": 0.7355, "step": 9910 }, { "epoch": 0.8445675330208777, "grad_norm": 17.5, "learning_rate": 1.5615784327169337e-06, "loss": 0.9041, "step": 9911 }, { "epoch": 0.8446527481891777, "grad_norm": 13.4375, "learning_rate": 1.5614156982432848e-06, "loss": 0.6214, "step": 9912 }, { "epoch": 0.8447379633574776, "grad_norm": 14.3125, "learning_rate": 1.5612529581420415e-06, "loss": 0.7267, "step": 9913 }, { "epoch": 0.8448231785257776, "grad_norm": 11.375, "learning_rate": 1.561090212416144e-06, "loss": 0.565, "step": 9914 }, { "epoch": 0.8449083936940776, "grad_norm": 13.75, "learning_rate": 1.5609274610685338e-06, "loss": 0.5954, "step": 9915 }, { "epoch": 0.8449936088623775, "grad_norm": 14.375, "learning_rate": 1.5607647041021516e-06, "loss": 0.7366, "step": 9916 }, { "epoch": 0.8450788240306775, "grad_norm": 13.4375, "learning_rate": 1.5606019415199383e-06, "loss": 0.7262, "step": 9917 }, { "epoch": 0.8451640391989774, "grad_norm": 15.6875, "learning_rate": 1.5604391733248361e-06, "loss": 0.96, "step": 9918 }, { "epoch": 0.8452492543672774, "grad_norm": 11.625, "learning_rate": 1.5602763995197853e-06, "loss": 0.6817, "step": 9919 }, { "epoch": 0.8453344695355773, "grad_norm": 16.25, "learning_rate": 1.5601136201077288e-06, "loss": 0.9011, "step": 9920 }, { "epoch": 0.8454196847038773, "grad_norm": 19.125, "learning_rate": 1.5599508350916068e-06, "loss": 0.7238, "step": 9921 }, { "epoch": 0.8455048998721773, "grad_norm": 15.6875, "learning_rate": 1.5597880444743617e-06, "loss": 0.6033, "step": 9922 }, { "epoch": 0.8455901150404772, "grad_norm": 14.0, "learning_rate": 1.5596252482589348e-06, "loss": 0.4132, "step": 9923 }, { "epoch": 0.8456753302087772, "grad_norm": 12.0625, "learning_rate": 1.5594624464482689e-06, "loss": 0.4342, "step": 9924 }, { "epoch": 0.8457605453770771, "grad_norm": 15.1875, "learning_rate": 1.5592996390453047e-06, "loss": 0.6257, "step": 9925 }, { "epoch": 0.8458457605453771, "grad_norm": 12.9375, "learning_rate": 1.5591368260529855e-06, "loss": 0.7901, "step": 9926 }, { "epoch": 0.845930975713677, "grad_norm": 12.625, "learning_rate": 1.5589740074742525e-06, "loss": 0.5534, "step": 9927 }, { "epoch": 0.846016190881977, "grad_norm": 24.875, "learning_rate": 1.5588111833120493e-06, "loss": 0.8871, "step": 9928 }, { "epoch": 0.846101406050277, "grad_norm": 16.625, "learning_rate": 1.5586483535693166e-06, "loss": 0.8796, "step": 9929 }, { "epoch": 0.8461866212185769, "grad_norm": 17.125, "learning_rate": 1.5584855182489986e-06, "loss": 1.4636, "step": 9930 }, { "epoch": 0.8462718363868769, "grad_norm": 12.9375, "learning_rate": 1.5583226773540364e-06, "loss": 0.6934, "step": 9931 }, { "epoch": 0.8463570515551768, "grad_norm": 18.25, "learning_rate": 1.5581598308873735e-06, "loss": 1.0093, "step": 9932 }, { "epoch": 0.8464422667234768, "grad_norm": 22.375, "learning_rate": 1.5579969788519531e-06, "loss": 0.9159, "step": 9933 }, { "epoch": 0.8465274818917767, "grad_norm": 12.3125, "learning_rate": 1.557834121250717e-06, "loss": 0.6586, "step": 9934 }, { "epoch": 0.8466126970600767, "grad_norm": 12.1875, "learning_rate": 1.5576712580866088e-06, "loss": 0.6704, "step": 9935 }, { "epoch": 0.8466979122283766, "grad_norm": 15.0, "learning_rate": 1.557508389362572e-06, "loss": 0.9263, "step": 9936 }, { "epoch": 0.8467831273966766, "grad_norm": 13.1875, "learning_rate": 1.557345515081549e-06, "loss": 0.893, "step": 9937 }, { "epoch": 0.8468683425649766, "grad_norm": 15.1875, "learning_rate": 1.5571826352464835e-06, "loss": 0.722, "step": 9938 }, { "epoch": 0.8469535577332765, "grad_norm": 25.125, "learning_rate": 1.5570197498603193e-06, "loss": 0.9335, "step": 9939 }, { "epoch": 0.8470387729015765, "grad_norm": 20.625, "learning_rate": 1.5568568589259991e-06, "loss": 0.9344, "step": 9940 }, { "epoch": 0.8471239880698764, "grad_norm": 19.0, "learning_rate": 1.5566939624464672e-06, "loss": 0.4741, "step": 9941 }, { "epoch": 0.8472092032381764, "grad_norm": 11.5, "learning_rate": 1.5565310604246664e-06, "loss": 0.5916, "step": 9942 }, { "epoch": 0.8472944184064763, "grad_norm": 15.0625, "learning_rate": 1.5563681528635415e-06, "loss": 0.8187, "step": 9943 }, { "epoch": 0.8473796335747763, "grad_norm": 12.8125, "learning_rate": 1.5562052397660363e-06, "loss": 0.652, "step": 9944 }, { "epoch": 0.8474648487430763, "grad_norm": 14.25, "learning_rate": 1.5560423211350942e-06, "loss": 0.5854, "step": 9945 }, { "epoch": 0.8475500639113762, "grad_norm": 18.875, "learning_rate": 1.5558793969736593e-06, "loss": 0.8579, "step": 9946 }, { "epoch": 0.8476352790796762, "grad_norm": 13.1875, "learning_rate": 1.5557164672846765e-06, "loss": 0.4096, "step": 9947 }, { "epoch": 0.8477204942479761, "grad_norm": 15.0, "learning_rate": 1.5555535320710895e-06, "loss": 0.6412, "step": 9948 }, { "epoch": 0.8478057094162761, "grad_norm": 16.375, "learning_rate": 1.555390591335843e-06, "loss": 0.624, "step": 9949 }, { "epoch": 0.847890924584576, "grad_norm": 15.6875, "learning_rate": 1.5552276450818816e-06, "loss": 0.8301, "step": 9950 }, { "epoch": 0.847976139752876, "grad_norm": 19.625, "learning_rate": 1.5550646933121497e-06, "loss": 1.0591, "step": 9951 }, { "epoch": 0.848061354921176, "grad_norm": 11.375, "learning_rate": 1.554901736029592e-06, "loss": 0.3938, "step": 9952 }, { "epoch": 0.8481465700894759, "grad_norm": 21.0, "learning_rate": 1.554738773237153e-06, "loss": 0.8778, "step": 9953 }, { "epoch": 0.8482317852577759, "grad_norm": 14.3125, "learning_rate": 1.5545758049377781e-06, "loss": 0.5937, "step": 9954 }, { "epoch": 0.8483170004260758, "grad_norm": 22.375, "learning_rate": 1.554412831134412e-06, "loss": 0.9817, "step": 9955 }, { "epoch": 0.8484022155943758, "grad_norm": 23.375, "learning_rate": 1.5542498518300006e-06, "loss": 1.1974, "step": 9956 }, { "epoch": 0.8484874307626757, "grad_norm": 11.375, "learning_rate": 1.5540868670274879e-06, "loss": 0.502, "step": 9957 }, { "epoch": 0.8485726459309757, "grad_norm": 11.5, "learning_rate": 1.5539238767298197e-06, "loss": 0.3593, "step": 9958 }, { "epoch": 0.8486578610992757, "grad_norm": 23.5, "learning_rate": 1.5537608809399413e-06, "loss": 1.1276, "step": 9959 }, { "epoch": 0.8487430762675756, "grad_norm": 13.1875, "learning_rate": 1.5535978796607989e-06, "loss": 0.7329, "step": 9960 }, { "epoch": 0.8488282914358756, "grad_norm": 9.25, "learning_rate": 1.5534348728953368e-06, "loss": 0.3222, "step": 9961 }, { "epoch": 0.8489135066041755, "grad_norm": 15.0, "learning_rate": 1.553271860646502e-06, "loss": 0.8051, "step": 9962 }, { "epoch": 0.8489987217724755, "grad_norm": 17.125, "learning_rate": 1.5531088429172392e-06, "loss": 0.8001, "step": 9963 }, { "epoch": 0.8490839369407754, "grad_norm": 16.125, "learning_rate": 1.5529458197104951e-06, "loss": 0.7866, "step": 9964 }, { "epoch": 0.8491691521090754, "grad_norm": 35.75, "learning_rate": 1.5527827910292156e-06, "loss": 0.697, "step": 9965 }, { "epoch": 0.8492543672773754, "grad_norm": 32.0, "learning_rate": 1.5526197568763465e-06, "loss": 1.2927, "step": 9966 }, { "epoch": 0.8493395824456753, "grad_norm": 12.0, "learning_rate": 1.5524567172548339e-06, "loss": 0.6935, "step": 9967 }, { "epoch": 0.8494247976139753, "grad_norm": 13.0, "learning_rate": 1.5522936721676243e-06, "loss": 0.7174, "step": 9968 }, { "epoch": 0.8495100127822752, "grad_norm": 15.875, "learning_rate": 1.5521306216176644e-06, "loss": 0.6199, "step": 9969 }, { "epoch": 0.8495952279505752, "grad_norm": 15.6875, "learning_rate": 1.5519675656078998e-06, "loss": 0.7787, "step": 9970 }, { "epoch": 0.8496804431188751, "grad_norm": 17.375, "learning_rate": 1.551804504141278e-06, "loss": 0.8508, "step": 9971 }, { "epoch": 0.8497656582871751, "grad_norm": 15.1875, "learning_rate": 1.551641437220745e-06, "loss": 0.884, "step": 9972 }, { "epoch": 0.849850873455475, "grad_norm": 16.625, "learning_rate": 1.5514783648492485e-06, "loss": 0.8434, "step": 9973 }, { "epoch": 0.849936088623775, "grad_norm": 13.5, "learning_rate": 1.551315287029734e-06, "loss": 0.6282, "step": 9974 }, { "epoch": 0.850021303792075, "grad_norm": 12.25, "learning_rate": 1.5511522037651499e-06, "loss": 0.6725, "step": 9975 }, { "epoch": 0.8501065189603749, "grad_norm": 10.6875, "learning_rate": 1.5509891150584422e-06, "loss": 0.341, "step": 9976 }, { "epoch": 0.8501917341286749, "grad_norm": 10.25, "learning_rate": 1.5508260209125592e-06, "loss": 0.3871, "step": 9977 }, { "epoch": 0.8502769492969748, "grad_norm": 14.6875, "learning_rate": 1.550662921330447e-06, "loss": 0.6555, "step": 9978 }, { "epoch": 0.8503621644652748, "grad_norm": 20.25, "learning_rate": 1.5504998163150533e-06, "loss": 1.0452, "step": 9979 }, { "epoch": 0.8504473796335748, "grad_norm": 11.8125, "learning_rate": 1.5503367058693259e-06, "loss": 0.5478, "step": 9980 }, { "epoch": 0.8505325948018747, "grad_norm": 18.25, "learning_rate": 1.5501735899962122e-06, "loss": 0.6974, "step": 9981 }, { "epoch": 0.8506178099701747, "grad_norm": 13.625, "learning_rate": 1.5500104686986604e-06, "loss": 0.7311, "step": 9982 }, { "epoch": 0.8507030251384746, "grad_norm": 35.25, "learning_rate": 1.5498473419796172e-06, "loss": 0.8109, "step": 9983 }, { "epoch": 0.8507882403067746, "grad_norm": 19.25, "learning_rate": 1.549684209842031e-06, "loss": 0.922, "step": 9984 }, { "epoch": 0.8508734554750745, "grad_norm": 13.9375, "learning_rate": 1.5495210722888498e-06, "loss": 0.7558, "step": 9985 }, { "epoch": 0.8509586706433745, "grad_norm": 14.9375, "learning_rate": 1.549357929323022e-06, "loss": 0.8149, "step": 9986 }, { "epoch": 0.8510438858116744, "grad_norm": 26.25, "learning_rate": 1.549194780947495e-06, "loss": 0.4548, "step": 9987 }, { "epoch": 0.8511291009799744, "grad_norm": 23.125, "learning_rate": 1.5490316271652176e-06, "loss": 1.4007, "step": 9988 }, { "epoch": 0.8512143161482744, "grad_norm": 16.875, "learning_rate": 1.548868467979138e-06, "loss": 0.6879, "step": 9989 }, { "epoch": 0.8512995313165743, "grad_norm": 15.0, "learning_rate": 1.5487053033922048e-06, "loss": 0.7199, "step": 9990 }, { "epoch": 0.8513847464848743, "grad_norm": 12.5, "learning_rate": 1.5485421334073658e-06, "loss": 0.7627, "step": 9991 }, { "epoch": 0.8514699616531742, "grad_norm": 12.625, "learning_rate": 1.5483789580275707e-06, "loss": 0.903, "step": 9992 }, { "epoch": 0.8515551768214742, "grad_norm": 15.375, "learning_rate": 1.548215777255768e-06, "loss": 1.1291, "step": 9993 }, { "epoch": 0.8516403919897741, "grad_norm": 12.5625, "learning_rate": 1.548052591094906e-06, "loss": 0.6761, "step": 9994 }, { "epoch": 0.8517256071580741, "grad_norm": 12.5625, "learning_rate": 1.5478893995479338e-06, "loss": 0.4795, "step": 9995 }, { "epoch": 0.8518108223263741, "grad_norm": 18.5, "learning_rate": 1.5477262026178008e-06, "loss": 0.7059, "step": 9996 }, { "epoch": 0.851896037494674, "grad_norm": 19.375, "learning_rate": 1.5475630003074563e-06, "loss": 0.8677, "step": 9997 }, { "epoch": 0.851981252662974, "grad_norm": 10.1875, "learning_rate": 1.5473997926198488e-06, "loss": 0.3925, "step": 9998 }, { "epoch": 0.8520664678312739, "grad_norm": 10.375, "learning_rate": 1.5472365795579277e-06, "loss": 0.421, "step": 9999 }, { "epoch": 0.8521516829995739, "grad_norm": 12.5, "learning_rate": 1.5470733611246432e-06, "loss": 0.4204, "step": 10000 }, { "epoch": 0.8522368981678738, "grad_norm": 13.4375, "learning_rate": 1.5469101373229442e-06, "loss": 0.6935, "step": 10001 }, { "epoch": 0.8523221133361738, "grad_norm": 16.125, "learning_rate": 1.54674690815578e-06, "loss": 0.9133, "step": 10002 }, { "epoch": 0.8524073285044738, "grad_norm": 13.0, "learning_rate": 1.5465836736261015e-06, "loss": 0.982, "step": 10003 }, { "epoch": 0.8524925436727737, "grad_norm": 16.125, "learning_rate": 1.546420433736857e-06, "loss": 0.79, "step": 10004 }, { "epoch": 0.8525777588410737, "grad_norm": 18.375, "learning_rate": 1.5462571884909984e-06, "loss": 0.2605, "step": 10005 }, { "epoch": 0.8526629740093736, "grad_norm": 47.0, "learning_rate": 1.5460939378914732e-06, "loss": 1.1058, "step": 10006 }, { "epoch": 0.8527481891776736, "grad_norm": 16.25, "learning_rate": 1.5459306819412337e-06, "loss": 0.7308, "step": 10007 }, { "epoch": 0.8528334043459735, "grad_norm": 17.375, "learning_rate": 1.5457674206432286e-06, "loss": 0.2821, "step": 10008 }, { "epoch": 0.8529186195142735, "grad_norm": 12.0, "learning_rate": 1.5456041540004094e-06, "loss": 0.7543, "step": 10009 }, { "epoch": 0.8530038346825735, "grad_norm": 15.625, "learning_rate": 1.545440882015725e-06, "loss": 0.8549, "step": 10010 }, { "epoch": 0.8530890498508734, "grad_norm": 15.125, "learning_rate": 1.5452776046921275e-06, "loss": 0.8032, "step": 10011 }, { "epoch": 0.8531742650191734, "grad_norm": 11.625, "learning_rate": 1.5451143220325664e-06, "loss": 0.5344, "step": 10012 }, { "epoch": 0.8532594801874733, "grad_norm": 16.125, "learning_rate": 1.5449510340399932e-06, "loss": 0.6638, "step": 10013 }, { "epoch": 0.8533446953557733, "grad_norm": 16.25, "learning_rate": 1.5447877407173577e-06, "loss": 0.882, "step": 10014 }, { "epoch": 0.8534299105240732, "grad_norm": 15.75, "learning_rate": 1.5446244420676116e-06, "loss": 0.8087, "step": 10015 }, { "epoch": 0.8535151256923732, "grad_norm": 11.375, "learning_rate": 1.5444611380937052e-06, "loss": 0.4129, "step": 10016 }, { "epoch": 0.8536003408606732, "grad_norm": 20.0, "learning_rate": 1.5442978287985902e-06, "loss": 0.9496, "step": 10017 }, { "epoch": 0.8536855560289731, "grad_norm": 16.25, "learning_rate": 1.5441345141852177e-06, "loss": 1.0274, "step": 10018 }, { "epoch": 0.8537707711972731, "grad_norm": 9.1875, "learning_rate": 1.5439711942565384e-06, "loss": 0.3175, "step": 10019 }, { "epoch": 0.853855986365573, "grad_norm": 19.5, "learning_rate": 1.5438078690155043e-06, "loss": 0.7375, "step": 10020 }, { "epoch": 0.853941201533873, "grad_norm": 18.375, "learning_rate": 1.5436445384650663e-06, "loss": 0.9234, "step": 10021 }, { "epoch": 0.8540264167021729, "grad_norm": 15.1875, "learning_rate": 1.5434812026081763e-06, "loss": 0.6828, "step": 10022 }, { "epoch": 0.8541116318704729, "grad_norm": 11.625, "learning_rate": 1.5433178614477857e-06, "loss": 0.4363, "step": 10023 }, { "epoch": 0.8541968470387729, "grad_norm": 13.1875, "learning_rate": 1.5431545149868466e-06, "loss": 0.408, "step": 10024 }, { "epoch": 0.8542820622070728, "grad_norm": 14.0, "learning_rate": 1.5429911632283106e-06, "loss": 0.7951, "step": 10025 }, { "epoch": 0.8543672773753728, "grad_norm": 16.125, "learning_rate": 1.5428278061751297e-06, "loss": 0.826, "step": 10026 }, { "epoch": 0.8544524925436727, "grad_norm": 19.625, "learning_rate": 1.5426644438302557e-06, "loss": 0.8248, "step": 10027 }, { "epoch": 0.8545377077119727, "grad_norm": 14.4375, "learning_rate": 1.5425010761966412e-06, "loss": 0.9161, "step": 10028 }, { "epoch": 0.8546229228802726, "grad_norm": 13.5625, "learning_rate": 1.5423377032772376e-06, "loss": 0.6557, "step": 10029 }, { "epoch": 0.8547081380485726, "grad_norm": 20.625, "learning_rate": 1.5421743250749984e-06, "loss": 1.3825, "step": 10030 }, { "epoch": 0.8547933532168726, "grad_norm": 11.5, "learning_rate": 1.542010941592875e-06, "loss": 0.2564, "step": 10031 }, { "epoch": 0.8548785683851725, "grad_norm": 14.0, "learning_rate": 1.5418475528338204e-06, "loss": 0.552, "step": 10032 }, { "epoch": 0.8549637835534725, "grad_norm": 13.25, "learning_rate": 1.5416841588007863e-06, "loss": 0.5562, "step": 10033 }, { "epoch": 0.8550489987217724, "grad_norm": 14.5, "learning_rate": 1.5415207594967272e-06, "loss": 0.5113, "step": 10034 }, { "epoch": 0.8551342138900724, "grad_norm": 12.5625, "learning_rate": 1.5413573549245941e-06, "loss": 0.6859, "step": 10035 }, { "epoch": 0.8552194290583723, "grad_norm": 12.375, "learning_rate": 1.541193945087341e-06, "loss": 0.685, "step": 10036 }, { "epoch": 0.8553046442266723, "grad_norm": 16.375, "learning_rate": 1.5410305299879202e-06, "loss": 0.8904, "step": 10037 }, { "epoch": 0.8553898593949723, "grad_norm": 11.9375, "learning_rate": 1.5408671096292855e-06, "loss": 0.4668, "step": 10038 }, { "epoch": 0.8554750745632722, "grad_norm": 13.5, "learning_rate": 1.5407036840143893e-06, "loss": 0.843, "step": 10039 }, { "epoch": 0.8555602897315722, "grad_norm": 13.5, "learning_rate": 1.5405402531461853e-06, "loss": 0.4669, "step": 10040 }, { "epoch": 0.8556455048998721, "grad_norm": 14.6875, "learning_rate": 1.5403768170276268e-06, "loss": 0.7578, "step": 10041 }, { "epoch": 0.8557307200681721, "grad_norm": 19.5, "learning_rate": 1.5402133756616672e-06, "loss": 0.8379, "step": 10042 }, { "epoch": 0.8558159352364721, "grad_norm": 14.5625, "learning_rate": 1.54004992905126e-06, "loss": 0.7742, "step": 10043 }, { "epoch": 0.8559011504047721, "grad_norm": 14.9375, "learning_rate": 1.5398864771993588e-06, "loss": 1.0037, "step": 10044 }, { "epoch": 0.8559863655730721, "grad_norm": 13.8125, "learning_rate": 1.5397230201089178e-06, "loss": 0.6259, "step": 10045 }, { "epoch": 0.856071580741372, "grad_norm": 10.8125, "learning_rate": 1.5395595577828903e-06, "loss": 0.5707, "step": 10046 }, { "epoch": 0.856156795909672, "grad_norm": 13.9375, "learning_rate": 1.5393960902242306e-06, "loss": 0.629, "step": 10047 }, { "epoch": 0.8562420110779719, "grad_norm": 13.3125, "learning_rate": 1.5392326174358921e-06, "loss": 0.5655, "step": 10048 }, { "epoch": 0.8563272262462719, "grad_norm": 15.6875, "learning_rate": 1.5390691394208297e-06, "loss": 0.1799, "step": 10049 }, { "epoch": 0.8564124414145718, "grad_norm": 10.875, "learning_rate": 1.5389056561819973e-06, "loss": 0.4715, "step": 10050 }, { "epoch": 0.8564976565828718, "grad_norm": 12.4375, "learning_rate": 1.538742167722349e-06, "loss": 0.4801, "step": 10051 }, { "epoch": 0.8565828717511718, "grad_norm": 10.5625, "learning_rate": 1.5385786740448395e-06, "loss": 0.3631, "step": 10052 }, { "epoch": 0.8566680869194717, "grad_norm": 16.625, "learning_rate": 1.5384151751524228e-06, "loss": 0.9585, "step": 10053 }, { "epoch": 0.8567533020877717, "grad_norm": 13.5, "learning_rate": 1.5382516710480544e-06, "loss": 0.6098, "step": 10054 }, { "epoch": 0.8568385172560716, "grad_norm": 19.875, "learning_rate": 1.5380881617346879e-06, "loss": 1.1762, "step": 10055 }, { "epoch": 0.8569237324243716, "grad_norm": 21.25, "learning_rate": 1.5379246472152792e-06, "loss": 1.0837, "step": 10056 }, { "epoch": 0.8570089475926715, "grad_norm": 12.625, "learning_rate": 1.537761127492782e-06, "loss": 0.6825, "step": 10057 }, { "epoch": 0.8570941627609715, "grad_norm": 13.1875, "learning_rate": 1.5375976025701526e-06, "loss": 0.5408, "step": 10058 }, { "epoch": 0.8571793779292715, "grad_norm": 17.375, "learning_rate": 1.5374340724503448e-06, "loss": 0.9366, "step": 10059 }, { "epoch": 0.8572645930975714, "grad_norm": 17.125, "learning_rate": 1.5372705371363145e-06, "loss": 0.6548, "step": 10060 }, { "epoch": 0.8573498082658714, "grad_norm": 13.9375, "learning_rate": 1.5371069966310163e-06, "loss": 0.6377, "step": 10061 }, { "epoch": 0.8574350234341713, "grad_norm": 13.125, "learning_rate": 1.5369434509374064e-06, "loss": 0.8494, "step": 10062 }, { "epoch": 0.8575202386024713, "grad_norm": 13.4375, "learning_rate": 1.5367799000584395e-06, "loss": 0.6824, "step": 10063 }, { "epoch": 0.8576054537707712, "grad_norm": 12.25, "learning_rate": 1.5366163439970718e-06, "loss": 0.4818, "step": 10064 }, { "epoch": 0.8576906689390712, "grad_norm": 16.625, "learning_rate": 1.5364527827562582e-06, "loss": 0.497, "step": 10065 }, { "epoch": 0.8577758841073712, "grad_norm": 10.4375, "learning_rate": 1.5362892163389553e-06, "loss": 0.3624, "step": 10066 }, { "epoch": 0.8578610992756711, "grad_norm": 18.375, "learning_rate": 1.5361256447481176e-06, "loss": 0.6826, "step": 10067 }, { "epoch": 0.8579463144439711, "grad_norm": 13.3125, "learning_rate": 1.5359620679867025e-06, "loss": 0.7534, "step": 10068 }, { "epoch": 0.858031529612271, "grad_norm": 20.25, "learning_rate": 1.5357984860576647e-06, "loss": 0.8449, "step": 10069 }, { "epoch": 0.858116744780571, "grad_norm": 11.5625, "learning_rate": 1.5356348989639611e-06, "loss": 0.5237, "step": 10070 }, { "epoch": 0.8582019599488709, "grad_norm": 14.6875, "learning_rate": 1.5354713067085476e-06, "loss": 0.6756, "step": 10071 }, { "epoch": 0.8582871751171709, "grad_norm": 11.5, "learning_rate": 1.5353077092943804e-06, "loss": 0.6356, "step": 10072 }, { "epoch": 0.8583723902854709, "grad_norm": 14.125, "learning_rate": 1.5351441067244163e-06, "loss": 0.6461, "step": 10073 }, { "epoch": 0.8584576054537708, "grad_norm": 16.375, "learning_rate": 1.5349804990016115e-06, "loss": 0.8606, "step": 10074 }, { "epoch": 0.8585428206220708, "grad_norm": 11.5625, "learning_rate": 1.5348168861289222e-06, "loss": 0.1709, "step": 10075 }, { "epoch": 0.8586280357903707, "grad_norm": 10.6875, "learning_rate": 1.5346532681093054e-06, "loss": 0.3805, "step": 10076 }, { "epoch": 0.8587132509586707, "grad_norm": 13.25, "learning_rate": 1.534489644945718e-06, "loss": 0.543, "step": 10077 }, { "epoch": 0.8587984661269706, "grad_norm": 10.375, "learning_rate": 1.5343260166411165e-06, "loss": 0.4412, "step": 10078 }, { "epoch": 0.8588836812952706, "grad_norm": 24.125, "learning_rate": 1.5341623831984582e-06, "loss": 0.7599, "step": 10079 }, { "epoch": 0.8589688964635706, "grad_norm": 10.75, "learning_rate": 1.5339987446206993e-06, "loss": 0.9472, "step": 10080 }, { "epoch": 0.8590541116318705, "grad_norm": 15.5625, "learning_rate": 1.5338351009107979e-06, "loss": 0.9006, "step": 10081 }, { "epoch": 0.8591393268001705, "grad_norm": 15.0, "learning_rate": 1.533671452071711e-06, "loss": 0.8534, "step": 10082 }, { "epoch": 0.8592245419684704, "grad_norm": 12.5625, "learning_rate": 1.5335077981063955e-06, "loss": 0.5402, "step": 10083 }, { "epoch": 0.8593097571367704, "grad_norm": 13.5625, "learning_rate": 1.5333441390178088e-06, "loss": 0.7779, "step": 10084 }, { "epoch": 0.8593949723050703, "grad_norm": 14.6875, "learning_rate": 1.5331804748089092e-06, "loss": 0.8625, "step": 10085 }, { "epoch": 0.8594801874733703, "grad_norm": 11.5625, "learning_rate": 1.533016805482653e-06, "loss": 0.4704, "step": 10086 }, { "epoch": 0.8595654026416703, "grad_norm": 13.0, "learning_rate": 1.5328531310419985e-06, "loss": 0.5451, "step": 10087 }, { "epoch": 0.8596506178099702, "grad_norm": 14.375, "learning_rate": 1.532689451489904e-06, "loss": 0.5704, "step": 10088 }, { "epoch": 0.8597358329782702, "grad_norm": 14.625, "learning_rate": 1.5325257668293264e-06, "loss": 0.7277, "step": 10089 }, { "epoch": 0.8598210481465701, "grad_norm": 21.5, "learning_rate": 1.5323620770632248e-06, "loss": 0.7453, "step": 10090 }, { "epoch": 0.8599062633148701, "grad_norm": 16.375, "learning_rate": 1.5321983821945558e-06, "loss": 0.7724, "step": 10091 }, { "epoch": 0.85999147848317, "grad_norm": 17.5, "learning_rate": 1.5320346822262786e-06, "loss": 0.9587, "step": 10092 }, { "epoch": 0.86007669365147, "grad_norm": 10.1875, "learning_rate": 1.5318709771613508e-06, "loss": 0.297, "step": 10093 }, { "epoch": 0.86016190881977, "grad_norm": 14.375, "learning_rate": 1.5317072670027316e-06, "loss": 0.7796, "step": 10094 }, { "epoch": 0.8602471239880699, "grad_norm": 16.75, "learning_rate": 1.531543551753378e-06, "loss": 0.967, "step": 10095 }, { "epoch": 0.8603323391563699, "grad_norm": 12.875, "learning_rate": 1.5313798314162498e-06, "loss": 0.5251, "step": 10096 }, { "epoch": 0.8604175543246698, "grad_norm": 11.0625, "learning_rate": 1.5312161059943048e-06, "loss": 0.4213, "step": 10097 }, { "epoch": 0.8605027694929698, "grad_norm": 18.0, "learning_rate": 1.5310523754905024e-06, "loss": 0.9128, "step": 10098 }, { "epoch": 0.8605879846612697, "grad_norm": 12.375, "learning_rate": 1.5308886399078008e-06, "loss": 0.4332, "step": 10099 }, { "epoch": 0.8606731998295697, "grad_norm": 14.0, "learning_rate": 1.5307248992491588e-06, "loss": 0.6996, "step": 10100 }, { "epoch": 0.8607584149978696, "grad_norm": 10.625, "learning_rate": 1.5305611535175357e-06, "loss": 0.5253, "step": 10101 }, { "epoch": 0.8608436301661696, "grad_norm": 16.375, "learning_rate": 1.5303974027158905e-06, "loss": 0.7075, "step": 10102 }, { "epoch": 0.8609288453344696, "grad_norm": 11.5, "learning_rate": 1.5302336468471823e-06, "loss": 0.5526, "step": 10103 }, { "epoch": 0.8610140605027695, "grad_norm": 12.9375, "learning_rate": 1.53006988591437e-06, "loss": 0.8887, "step": 10104 }, { "epoch": 0.8610992756710695, "grad_norm": 15.3125, "learning_rate": 1.5299061199204136e-06, "loss": 0.6814, "step": 10105 }, { "epoch": 0.8611844908393694, "grad_norm": 15.125, "learning_rate": 1.5297423488682721e-06, "loss": 0.9171, "step": 10106 }, { "epoch": 0.8612697060076694, "grad_norm": 18.125, "learning_rate": 1.529578572760905e-06, "loss": 1.0592, "step": 10107 }, { "epoch": 0.8613549211759693, "grad_norm": 10.0625, "learning_rate": 1.5294147916012717e-06, "loss": 0.2861, "step": 10108 }, { "epoch": 0.8614401363442693, "grad_norm": 21.375, "learning_rate": 1.5292510053923327e-06, "loss": 1.1863, "step": 10109 }, { "epoch": 0.8615253515125693, "grad_norm": 15.625, "learning_rate": 1.5290872141370466e-06, "loss": 0.9118, "step": 10110 }, { "epoch": 0.8616105666808692, "grad_norm": 13.875, "learning_rate": 1.5289234178383744e-06, "loss": 0.5765, "step": 10111 }, { "epoch": 0.8616957818491692, "grad_norm": 24.5, "learning_rate": 1.5287596164992752e-06, "loss": 1.1467, "step": 10112 }, { "epoch": 0.8617809970174691, "grad_norm": 19.5, "learning_rate": 1.5285958101227098e-06, "loss": 1.1988, "step": 10113 }, { "epoch": 0.8618662121857691, "grad_norm": 43.75, "learning_rate": 1.5284319987116371e-06, "loss": 0.4794, "step": 10114 }, { "epoch": 0.861951427354069, "grad_norm": 23.375, "learning_rate": 1.5282681822690194e-06, "loss": 0.6724, "step": 10115 }, { "epoch": 0.862036642522369, "grad_norm": 8.875, "learning_rate": 1.528104360797815e-06, "loss": 0.3069, "step": 10116 }, { "epoch": 0.862121857690669, "grad_norm": 14.4375, "learning_rate": 1.5279405343009854e-06, "loss": 0.7261, "step": 10117 }, { "epoch": 0.8622070728589689, "grad_norm": 14.5, "learning_rate": 1.5277767027814908e-06, "loss": 0.8507, "step": 10118 }, { "epoch": 0.8622922880272689, "grad_norm": 12.1875, "learning_rate": 1.5276128662422917e-06, "loss": 0.4278, "step": 10119 }, { "epoch": 0.8623775031955688, "grad_norm": 16.75, "learning_rate": 1.5274490246863495e-06, "loss": 1.0737, "step": 10120 }, { "epoch": 0.8624627183638688, "grad_norm": 10.0, "learning_rate": 1.527285178116624e-06, "loss": 0.453, "step": 10121 }, { "epoch": 0.8625479335321687, "grad_norm": 14.3125, "learning_rate": 1.5271213265360766e-06, "loss": 0.6308, "step": 10122 }, { "epoch": 0.8626331487004687, "grad_norm": 17.25, "learning_rate": 1.526957469947668e-06, "loss": 0.6245, "step": 10123 }, { "epoch": 0.8627183638687687, "grad_norm": 14.6875, "learning_rate": 1.5267936083543597e-06, "loss": 0.5867, "step": 10124 }, { "epoch": 0.8628035790370686, "grad_norm": 12.75, "learning_rate": 1.5266297417591121e-06, "loss": 0.5257, "step": 10125 }, { "epoch": 0.8628887942053686, "grad_norm": 12.0, "learning_rate": 1.5264658701648874e-06, "loss": 0.4022, "step": 10126 }, { "epoch": 0.8629740093736685, "grad_norm": 15.1875, "learning_rate": 1.5263019935746462e-06, "loss": 0.7267, "step": 10127 }, { "epoch": 0.8630592245419685, "grad_norm": 14.5625, "learning_rate": 1.5261381119913504e-06, "loss": 0.8053, "step": 10128 }, { "epoch": 0.8631444397102684, "grad_norm": 13.5, "learning_rate": 1.5259742254179607e-06, "loss": 0.7908, "step": 10129 }, { "epoch": 0.8632296548785684, "grad_norm": 22.625, "learning_rate": 1.5258103338574396e-06, "loss": 0.8839, "step": 10130 }, { "epoch": 0.8633148700468684, "grad_norm": 14.8125, "learning_rate": 1.5256464373127483e-06, "loss": 0.6898, "step": 10131 }, { "epoch": 0.8634000852151683, "grad_norm": 12.875, "learning_rate": 1.5254825357868488e-06, "loss": 0.5738, "step": 10132 }, { "epoch": 0.8634853003834683, "grad_norm": 14.375, "learning_rate": 1.5253186292827027e-06, "loss": 0.7751, "step": 10133 }, { "epoch": 0.8635705155517682, "grad_norm": 14.25, "learning_rate": 1.5251547178032722e-06, "loss": 0.634, "step": 10134 }, { "epoch": 0.8636557307200682, "grad_norm": 15.875, "learning_rate": 1.5249908013515193e-06, "loss": 0.4709, "step": 10135 }, { "epoch": 0.8637409458883681, "grad_norm": 19.5, "learning_rate": 1.5248268799304062e-06, "loss": 0.9359, "step": 10136 }, { "epoch": 0.8638261610566681, "grad_norm": 17.5, "learning_rate": 1.5246629535428947e-06, "loss": 0.6998, "step": 10137 }, { "epoch": 0.863911376224968, "grad_norm": 11.25, "learning_rate": 1.5244990221919479e-06, "loss": 0.4152, "step": 10138 }, { "epoch": 0.863996591393268, "grad_norm": 13.9375, "learning_rate": 1.5243350858805272e-06, "loss": 0.5428, "step": 10139 }, { "epoch": 0.864081806561568, "grad_norm": 17.125, "learning_rate": 1.5241711446115958e-06, "loss": 0.8772, "step": 10140 }, { "epoch": 0.8641670217298679, "grad_norm": 18.0, "learning_rate": 1.5240071983881164e-06, "loss": 0.9802, "step": 10141 }, { "epoch": 0.8642522368981679, "grad_norm": 17.625, "learning_rate": 1.523843247213051e-06, "loss": 0.8427, "step": 10142 }, { "epoch": 0.8643374520664678, "grad_norm": 14.5625, "learning_rate": 1.5236792910893633e-06, "loss": 0.8581, "step": 10143 }, { "epoch": 0.8644226672347678, "grad_norm": 18.625, "learning_rate": 1.523515330020015e-06, "loss": 0.812, "step": 10144 }, { "epoch": 0.8645078824030678, "grad_norm": 14.5625, "learning_rate": 1.52335136400797e-06, "loss": 0.858, "step": 10145 }, { "epoch": 0.8645930975713677, "grad_norm": 13.0, "learning_rate": 1.5231873930561909e-06, "loss": 0.5582, "step": 10146 }, { "epoch": 0.8646783127396677, "grad_norm": 14.125, "learning_rate": 1.5230234171676412e-06, "loss": 0.7779, "step": 10147 }, { "epoch": 0.8647635279079676, "grad_norm": 21.375, "learning_rate": 1.5228594363452831e-06, "loss": 0.6991, "step": 10148 }, { "epoch": 0.8648487430762676, "grad_norm": 12.875, "learning_rate": 1.522695450592081e-06, "loss": 0.5775, "step": 10149 }, { "epoch": 0.8649339582445675, "grad_norm": 18.625, "learning_rate": 1.5225314599109978e-06, "loss": 0.8759, "step": 10150 }, { "epoch": 0.8650191734128675, "grad_norm": 17.125, "learning_rate": 1.5223674643049975e-06, "loss": 0.8011, "step": 10151 }, { "epoch": 0.8651043885811674, "grad_norm": 9.875, "learning_rate": 1.5222034637770424e-06, "loss": 0.2941, "step": 10152 }, { "epoch": 0.8651896037494674, "grad_norm": 17.75, "learning_rate": 1.5220394583300976e-06, "loss": 0.5878, "step": 10153 }, { "epoch": 0.8652748189177674, "grad_norm": 13.625, "learning_rate": 1.521875447967126e-06, "loss": 0.8522, "step": 10154 }, { "epoch": 0.8653600340860673, "grad_norm": 11.25, "learning_rate": 1.5217114326910918e-06, "loss": 0.5354, "step": 10155 }, { "epoch": 0.8654452492543673, "grad_norm": 16.875, "learning_rate": 1.521547412504959e-06, "loss": 0.8246, "step": 10156 }, { "epoch": 0.8655304644226672, "grad_norm": 12.75, "learning_rate": 1.5213833874116908e-06, "loss": 0.6896, "step": 10157 }, { "epoch": 0.8656156795909672, "grad_norm": 14.625, "learning_rate": 1.5212193574142522e-06, "loss": 0.7967, "step": 10158 }, { "epoch": 0.8657008947592671, "grad_norm": 15.75, "learning_rate": 1.5210553225156072e-06, "loss": 0.6133, "step": 10159 }, { "epoch": 0.8657861099275671, "grad_norm": 17.375, "learning_rate": 1.52089128271872e-06, "loss": 0.8595, "step": 10160 }, { "epoch": 0.8658713250958671, "grad_norm": 16.625, "learning_rate": 1.5207272380265545e-06, "loss": 1.028, "step": 10161 }, { "epoch": 0.865956540264167, "grad_norm": 21.0, "learning_rate": 1.520563188442076e-06, "loss": 0.7477, "step": 10162 }, { "epoch": 0.866041755432467, "grad_norm": 18.5, "learning_rate": 1.5203991339682487e-06, "loss": 1.1012, "step": 10163 }, { "epoch": 0.8661269706007669, "grad_norm": 19.5, "learning_rate": 1.520235074608037e-06, "loss": 0.7039, "step": 10164 }, { "epoch": 0.8662121857690669, "grad_norm": 15.875, "learning_rate": 1.5200710103644054e-06, "loss": 1.0405, "step": 10165 }, { "epoch": 0.8662974009373668, "grad_norm": 19.0, "learning_rate": 1.5199069412403193e-06, "loss": 1.412, "step": 10166 }, { "epoch": 0.8663826161056668, "grad_norm": 11.75, "learning_rate": 1.5197428672387432e-06, "loss": 0.4241, "step": 10167 }, { "epoch": 0.8664678312739668, "grad_norm": 15.625, "learning_rate": 1.519578788362643e-06, "loss": 0.8056, "step": 10168 }, { "epoch": 0.8665530464422667, "grad_norm": 13.875, "learning_rate": 1.5194147046149821e-06, "loss": 1.0708, "step": 10169 }, { "epoch": 0.8666382616105667, "grad_norm": 17.875, "learning_rate": 1.519250615998727e-06, "loss": 0.7453, "step": 10170 }, { "epoch": 0.8667234767788666, "grad_norm": 10.6875, "learning_rate": 1.5190865225168422e-06, "loss": 0.3066, "step": 10171 }, { "epoch": 0.8668086919471666, "grad_norm": 11.0625, "learning_rate": 1.5189224241722935e-06, "loss": 0.4125, "step": 10172 }, { "epoch": 0.8668939071154665, "grad_norm": 14.0625, "learning_rate": 1.5187583209680464e-06, "loss": 0.6304, "step": 10173 }, { "epoch": 0.8669791222837665, "grad_norm": 12.125, "learning_rate": 1.5185942129070657e-06, "loss": 0.9896, "step": 10174 }, { "epoch": 0.8670643374520665, "grad_norm": 17.0, "learning_rate": 1.518430099992318e-06, "loss": 1.027, "step": 10175 }, { "epoch": 0.8671495526203664, "grad_norm": 32.0, "learning_rate": 1.5182659822267685e-06, "loss": 0.6872, "step": 10176 }, { "epoch": 0.8672347677886664, "grad_norm": 12.25, "learning_rate": 1.5181018596133825e-06, "loss": 0.5029, "step": 10177 }, { "epoch": 0.8673199829569663, "grad_norm": 12.0, "learning_rate": 1.5179377321551264e-06, "loss": 0.5507, "step": 10178 }, { "epoch": 0.8674051981252663, "grad_norm": 16.625, "learning_rate": 1.5177735998549664e-06, "loss": 0.9249, "step": 10179 }, { "epoch": 0.8674904132935662, "grad_norm": 15.4375, "learning_rate": 1.517609462715868e-06, "loss": 0.6407, "step": 10180 }, { "epoch": 0.8675756284618662, "grad_norm": 16.875, "learning_rate": 1.5174453207407974e-06, "loss": 0.9847, "step": 10181 }, { "epoch": 0.8676608436301662, "grad_norm": 14.8125, "learning_rate": 1.517281173932721e-06, "loss": 0.8747, "step": 10182 }, { "epoch": 0.8677460587984661, "grad_norm": 13.3125, "learning_rate": 1.517117022294605e-06, "loss": 0.8621, "step": 10183 }, { "epoch": 0.8678312739667661, "grad_norm": 12.9375, "learning_rate": 1.5169528658294158e-06, "loss": 0.5167, "step": 10184 }, { "epoch": 0.867916489135066, "grad_norm": 13.0, "learning_rate": 1.51678870454012e-06, "loss": 0.4209, "step": 10185 }, { "epoch": 0.868001704303366, "grad_norm": 13.125, "learning_rate": 1.5166245384296838e-06, "loss": 0.7921, "step": 10186 }, { "epoch": 0.8680869194716659, "grad_norm": 15.5, "learning_rate": 1.5164603675010741e-06, "loss": 0.6952, "step": 10187 }, { "epoch": 0.8681721346399659, "grad_norm": 16.0, "learning_rate": 1.5162961917572577e-06, "loss": 0.8782, "step": 10188 }, { "epoch": 0.8682573498082659, "grad_norm": 18.625, "learning_rate": 1.5161320112012012e-06, "loss": 0.8164, "step": 10189 }, { "epoch": 0.8683425649765658, "grad_norm": 10.625, "learning_rate": 1.5159678258358718e-06, "loss": 0.6189, "step": 10190 }, { "epoch": 0.8684277801448658, "grad_norm": 17.125, "learning_rate": 1.5158036356642364e-06, "loss": 0.7026, "step": 10191 }, { "epoch": 0.8685129953131657, "grad_norm": 13.4375, "learning_rate": 1.515639440689262e-06, "loss": 0.6142, "step": 10192 }, { "epoch": 0.8685982104814657, "grad_norm": 12.5, "learning_rate": 1.5154752409139156e-06, "loss": 0.6408, "step": 10193 }, { "epoch": 0.8686834256497656, "grad_norm": 11.875, "learning_rate": 1.5153110363411647e-06, "loss": 0.3683, "step": 10194 }, { "epoch": 0.8687686408180656, "grad_norm": 16.125, "learning_rate": 1.5151468269739764e-06, "loss": 0.8836, "step": 10195 }, { "epoch": 0.8688538559863656, "grad_norm": 14.625, "learning_rate": 1.514982612815319e-06, "loss": 0.5157, "step": 10196 }, { "epoch": 0.8689390711546655, "grad_norm": 12.125, "learning_rate": 1.5148183938681586e-06, "loss": 0.6051, "step": 10197 }, { "epoch": 0.8690242863229655, "grad_norm": 12.875, "learning_rate": 1.514654170135464e-06, "loss": 0.6138, "step": 10198 }, { "epoch": 0.8691095014912654, "grad_norm": 11.25, "learning_rate": 1.5144899416202019e-06, "loss": 0.4368, "step": 10199 }, { "epoch": 0.8691947166595654, "grad_norm": 12.625, "learning_rate": 1.5143257083253413e-06, "loss": 0.6164, "step": 10200 }, { "epoch": 0.8692799318278653, "grad_norm": 11.9375, "learning_rate": 1.5141614702538488e-06, "loss": 0.4718, "step": 10201 }, { "epoch": 0.8693651469961653, "grad_norm": 11.9375, "learning_rate": 1.5139972274086928e-06, "loss": 0.4012, "step": 10202 }, { "epoch": 0.8694503621644653, "grad_norm": 14.625, "learning_rate": 1.5138329797928416e-06, "loss": 0.8543, "step": 10203 }, { "epoch": 0.8695355773327652, "grad_norm": 18.875, "learning_rate": 1.5136687274092632e-06, "loss": 1.0437, "step": 10204 }, { "epoch": 0.8696207925010652, "grad_norm": 13.1875, "learning_rate": 1.5135044702609259e-06, "loss": 0.4893, "step": 10205 }, { "epoch": 0.8697060076693651, "grad_norm": 13.625, "learning_rate": 1.5133402083507978e-06, "loss": 0.8573, "step": 10206 }, { "epoch": 0.8697912228376651, "grad_norm": 12.625, "learning_rate": 1.513175941681847e-06, "loss": 0.6931, "step": 10207 }, { "epoch": 0.869876438005965, "grad_norm": 14.6875, "learning_rate": 1.5130116702570427e-06, "loss": 0.6333, "step": 10208 }, { "epoch": 0.869961653174265, "grad_norm": 11.5625, "learning_rate": 1.5128473940793531e-06, "loss": 0.4928, "step": 10209 }, { "epoch": 0.870046868342565, "grad_norm": 25.5, "learning_rate": 1.5126831131517464e-06, "loss": 0.9812, "step": 10210 }, { "epoch": 0.8701320835108649, "grad_norm": 16.0, "learning_rate": 1.5125188274771918e-06, "loss": 0.8645, "step": 10211 }, { "epoch": 0.8702172986791649, "grad_norm": 15.0625, "learning_rate": 1.5123545370586581e-06, "loss": 0.7024, "step": 10212 }, { "epoch": 0.8703025138474648, "grad_norm": 12.5625, "learning_rate": 1.5121902418991142e-06, "loss": 0.6007, "step": 10213 }, { "epoch": 0.8703877290157648, "grad_norm": 13.375, "learning_rate": 1.5120259420015286e-06, "loss": 0.6857, "step": 10214 }, { "epoch": 0.8704729441840647, "grad_norm": 15.6875, "learning_rate": 1.511861637368871e-06, "loss": 0.6279, "step": 10215 }, { "epoch": 0.8705581593523647, "grad_norm": 16.125, "learning_rate": 1.5116973280041103e-06, "loss": 1.0007, "step": 10216 }, { "epoch": 0.8706433745206646, "grad_norm": 15.0, "learning_rate": 1.5115330139102157e-06, "loss": 0.9044, "step": 10217 }, { "epoch": 0.8707285896889646, "grad_norm": 14.5, "learning_rate": 1.5113686950901566e-06, "loss": 0.5351, "step": 10218 }, { "epoch": 0.8708138048572646, "grad_norm": 13.375, "learning_rate": 1.5112043715469022e-06, "loss": 0.7688, "step": 10219 }, { "epoch": 0.8708990200255645, "grad_norm": 14.3125, "learning_rate": 1.5110400432834224e-06, "loss": 1.0671, "step": 10220 }, { "epoch": 0.8709842351938645, "grad_norm": 15.9375, "learning_rate": 1.5108757103026864e-06, "loss": 0.7939, "step": 10221 }, { "epoch": 0.8710694503621644, "grad_norm": 12.3125, "learning_rate": 1.510711372607664e-06, "loss": 0.5829, "step": 10222 }, { "epoch": 0.8711546655304644, "grad_norm": 13.6875, "learning_rate": 1.5105470302013248e-06, "loss": 0.7145, "step": 10223 }, { "epoch": 0.8712398806987643, "grad_norm": 11.6875, "learning_rate": 1.510382683086639e-06, "loss": 0.3441, "step": 10224 }, { "epoch": 0.8713250958670643, "grad_norm": 17.75, "learning_rate": 1.5102183312665757e-06, "loss": 0.8196, "step": 10225 }, { "epoch": 0.8714103110353643, "grad_norm": 17.75, "learning_rate": 1.510053974744106e-06, "loss": 0.9774, "step": 10226 }, { "epoch": 0.8714955262036642, "grad_norm": 10.375, "learning_rate": 1.509889613522199e-06, "loss": 0.3194, "step": 10227 }, { "epoch": 0.8715807413719642, "grad_norm": 10.25, "learning_rate": 1.5097252476038262e-06, "loss": 0.2893, "step": 10228 }, { "epoch": 0.8716659565402641, "grad_norm": 13.625, "learning_rate": 1.5095608769919566e-06, "loss": 0.8017, "step": 10229 }, { "epoch": 0.8717511717085641, "grad_norm": 16.5, "learning_rate": 1.5093965016895606e-06, "loss": 0.8652, "step": 10230 }, { "epoch": 0.871836386876864, "grad_norm": 12.8125, "learning_rate": 1.5092321216996092e-06, "loss": 0.4625, "step": 10231 }, { "epoch": 0.871921602045164, "grad_norm": 16.5, "learning_rate": 1.5090677370250729e-06, "loss": 0.4071, "step": 10232 }, { "epoch": 0.872006817213464, "grad_norm": 17.0, "learning_rate": 1.5089033476689219e-06, "loss": 0.9149, "step": 10233 }, { "epoch": 0.8720920323817639, "grad_norm": 12.9375, "learning_rate": 1.508738953634127e-06, "loss": 1.024, "step": 10234 }, { "epoch": 0.8721772475500639, "grad_norm": 15.0, "learning_rate": 1.5085745549236588e-06, "loss": 0.8391, "step": 10235 }, { "epoch": 0.8722624627183638, "grad_norm": 14.9375, "learning_rate": 1.5084101515404889e-06, "loss": 0.5198, "step": 10236 }, { "epoch": 0.8723476778866638, "grad_norm": 18.375, "learning_rate": 1.5082457434875874e-06, "loss": 0.8274, "step": 10237 }, { "epoch": 0.8724328930549637, "grad_norm": 16.125, "learning_rate": 1.508081330767926e-06, "loss": 0.9801, "step": 10238 }, { "epoch": 0.8725181082232637, "grad_norm": 13.3125, "learning_rate": 1.5079169133844749e-06, "loss": 0.591, "step": 10239 }, { "epoch": 0.8726033233915637, "grad_norm": 15.9375, "learning_rate": 1.507752491340206e-06, "loss": 0.6914, "step": 10240 }, { "epoch": 0.8726885385598636, "grad_norm": 13.9375, "learning_rate": 1.5075880646380905e-06, "loss": 1.0586, "step": 10241 }, { "epoch": 0.8727737537281636, "grad_norm": 11.6875, "learning_rate": 1.5074236332810993e-06, "loss": 0.3778, "step": 10242 }, { "epoch": 0.8728589688964635, "grad_norm": 18.25, "learning_rate": 1.5072591972722047e-06, "loss": 0.8658, "step": 10243 }, { "epoch": 0.8729441840647635, "grad_norm": 16.625, "learning_rate": 1.5070947566143775e-06, "loss": 0.8338, "step": 10244 }, { "epoch": 0.8730293992330634, "grad_norm": 12.375, "learning_rate": 1.5069303113105895e-06, "loss": 0.8188, "step": 10245 }, { "epoch": 0.8731146144013634, "grad_norm": 13.1875, "learning_rate": 1.5067658613638124e-06, "loss": 0.5875, "step": 10246 }, { "epoch": 0.8731998295696634, "grad_norm": 15.125, "learning_rate": 1.506601406777018e-06, "loss": 0.9176, "step": 10247 }, { "epoch": 0.8732850447379633, "grad_norm": 21.0, "learning_rate": 1.506436947553178e-06, "loss": 0.6382, "step": 10248 }, { "epoch": 0.8733702599062633, "grad_norm": 14.8125, "learning_rate": 1.5062724836952649e-06, "loss": 0.6083, "step": 10249 }, { "epoch": 0.8734554750745632, "grad_norm": 14.875, "learning_rate": 1.5061080152062498e-06, "loss": 0.6199, "step": 10250 }, { "epoch": 0.8735406902428632, "grad_norm": 17.875, "learning_rate": 1.5059435420891058e-06, "loss": 0.6179, "step": 10251 }, { "epoch": 0.8736259054111631, "grad_norm": 15.625, "learning_rate": 1.5057790643468042e-06, "loss": 0.7413, "step": 10252 }, { "epoch": 0.8737111205794631, "grad_norm": 16.625, "learning_rate": 1.5056145819823182e-06, "loss": 0.6481, "step": 10253 }, { "epoch": 0.873796335747763, "grad_norm": 13.4375, "learning_rate": 1.5054500949986193e-06, "loss": 0.5957, "step": 10254 }, { "epoch": 0.873881550916063, "grad_norm": 17.5, "learning_rate": 1.5052856033986804e-06, "loss": 1.1521, "step": 10255 }, { "epoch": 0.873966766084363, "grad_norm": 13.375, "learning_rate": 1.5051211071854738e-06, "loss": 0.636, "step": 10256 }, { "epoch": 0.8740519812526629, "grad_norm": 15.125, "learning_rate": 1.5049566063619725e-06, "loss": 0.7734, "step": 10257 }, { "epoch": 0.8741371964209629, "grad_norm": 16.375, "learning_rate": 1.504792100931149e-06, "loss": 0.6933, "step": 10258 }, { "epoch": 0.8742224115892628, "grad_norm": 14.5, "learning_rate": 1.5046275908959755e-06, "loss": 0.5732, "step": 10259 }, { "epoch": 0.8743076267575628, "grad_norm": 13.875, "learning_rate": 1.5044630762594257e-06, "loss": 0.6248, "step": 10260 }, { "epoch": 0.8743928419258628, "grad_norm": 15.875, "learning_rate": 1.5042985570244722e-06, "loss": 0.9475, "step": 10261 }, { "epoch": 0.8744780570941627, "grad_norm": 14.25, "learning_rate": 1.5041340331940885e-06, "loss": 0.9059, "step": 10262 }, { "epoch": 0.8745632722624627, "grad_norm": 21.5, "learning_rate": 1.5039695047712466e-06, "loss": 0.7358, "step": 10263 }, { "epoch": 0.8746484874307626, "grad_norm": 17.375, "learning_rate": 1.5038049717589206e-06, "loss": 1.0387, "step": 10264 }, { "epoch": 0.8747337025990626, "grad_norm": 14.1875, "learning_rate": 1.5036404341600835e-06, "loss": 0.5354, "step": 10265 }, { "epoch": 0.8748189177673625, "grad_norm": 15.3125, "learning_rate": 1.503475891977709e-06, "loss": 0.5545, "step": 10266 }, { "epoch": 0.8749041329356625, "grad_norm": 12.9375, "learning_rate": 1.50331134521477e-06, "loss": 0.688, "step": 10267 }, { "epoch": 0.8749893481039624, "grad_norm": 11.125, "learning_rate": 1.5031467938742404e-06, "loss": 0.4511, "step": 10268 }, { "epoch": 0.8750745632722625, "grad_norm": 14.4375, "learning_rate": 1.5029822379590936e-06, "loss": 0.4503, "step": 10269 }, { "epoch": 0.8751597784405625, "grad_norm": 16.75, "learning_rate": 1.5028176774723033e-06, "loss": 0.9331, "step": 10270 }, { "epoch": 0.8752449936088624, "grad_norm": 17.875, "learning_rate": 1.5026531124168431e-06, "loss": 0.8639, "step": 10271 }, { "epoch": 0.8753302087771624, "grad_norm": 14.5625, "learning_rate": 1.5024885427956877e-06, "loss": 0.7644, "step": 10272 }, { "epoch": 0.8754154239454623, "grad_norm": 12.75, "learning_rate": 1.50232396861181e-06, "loss": 0.4953, "step": 10273 }, { "epoch": 0.8755006391137623, "grad_norm": 24.375, "learning_rate": 1.5021593898681844e-06, "loss": 0.9405, "step": 10274 }, { "epoch": 0.8755858542820623, "grad_norm": 12.8125, "learning_rate": 1.5019948065677852e-06, "loss": 0.5444, "step": 10275 }, { "epoch": 0.8756710694503622, "grad_norm": 12.375, "learning_rate": 1.5018302187135866e-06, "loss": 0.5449, "step": 10276 }, { "epoch": 0.8757562846186622, "grad_norm": 13.375, "learning_rate": 1.5016656263085627e-06, "loss": 0.608, "step": 10277 }, { "epoch": 0.8758414997869621, "grad_norm": 25.375, "learning_rate": 1.5015010293556875e-06, "loss": 1.1416, "step": 10278 }, { "epoch": 0.8759267149552621, "grad_norm": 12.75, "learning_rate": 1.5013364278579362e-06, "loss": 0.5875, "step": 10279 }, { "epoch": 0.876011930123562, "grad_norm": 11.5, "learning_rate": 1.5011718218182824e-06, "loss": 0.338, "step": 10280 }, { "epoch": 0.876097145291862, "grad_norm": 17.0, "learning_rate": 1.501007211239702e-06, "loss": 0.7929, "step": 10281 }, { "epoch": 0.876182360460162, "grad_norm": 20.75, "learning_rate": 1.5008425961251682e-06, "loss": 0.8831, "step": 10282 }, { "epoch": 0.8762675756284619, "grad_norm": 18.375, "learning_rate": 1.5006779764776565e-06, "loss": 1.1731, "step": 10283 }, { "epoch": 0.8763527907967619, "grad_norm": 18.5, "learning_rate": 1.500513352300142e-06, "loss": 0.9694, "step": 10284 }, { "epoch": 0.8764380059650618, "grad_norm": 14.375, "learning_rate": 1.5003487235955994e-06, "loss": 0.9148, "step": 10285 }, { "epoch": 0.8765232211333618, "grad_norm": 18.125, "learning_rate": 1.5001840903670034e-06, "loss": 0.9419, "step": 10286 }, { "epoch": 0.8766084363016617, "grad_norm": 14.3125, "learning_rate": 1.5000194526173291e-06, "loss": 0.7039, "step": 10287 }, { "epoch": 0.8766936514699617, "grad_norm": 14.25, "learning_rate": 1.4998548103495519e-06, "loss": 0.6037, "step": 10288 }, { "epoch": 0.8767788666382617, "grad_norm": 17.0, "learning_rate": 1.4996901635666475e-06, "loss": 0.8228, "step": 10289 }, { "epoch": 0.8768640818065616, "grad_norm": 13.8125, "learning_rate": 1.4995255122715905e-06, "loss": 0.6379, "step": 10290 }, { "epoch": 0.8769492969748616, "grad_norm": 12.9375, "learning_rate": 1.4993608564673567e-06, "loss": 0.5076, "step": 10291 }, { "epoch": 0.8770345121431615, "grad_norm": 14.0625, "learning_rate": 1.499196196156921e-06, "loss": 0.6565, "step": 10292 }, { "epoch": 0.8771197273114615, "grad_norm": 15.6875, "learning_rate": 1.49903153134326e-06, "loss": 0.9863, "step": 10293 }, { "epoch": 0.8772049424797614, "grad_norm": 15.6875, "learning_rate": 1.4988668620293487e-06, "loss": 0.9442, "step": 10294 }, { "epoch": 0.8772901576480614, "grad_norm": 12.5625, "learning_rate": 1.4987021882181625e-06, "loss": 0.5681, "step": 10295 }, { "epoch": 0.8773753728163614, "grad_norm": 12.5, "learning_rate": 1.4985375099126781e-06, "loss": 0.5828, "step": 10296 }, { "epoch": 0.8774605879846613, "grad_norm": 13.875, "learning_rate": 1.4983728271158709e-06, "loss": 0.6703, "step": 10297 }, { "epoch": 0.8775458031529613, "grad_norm": 18.0, "learning_rate": 1.498208139830717e-06, "loss": 0.68, "step": 10298 }, { "epoch": 0.8776310183212612, "grad_norm": 12.125, "learning_rate": 1.4980434480601922e-06, "loss": 0.7133, "step": 10299 }, { "epoch": 0.8777162334895612, "grad_norm": 17.625, "learning_rate": 1.4978787518072732e-06, "loss": 0.7827, "step": 10300 }, { "epoch": 0.8778014486578611, "grad_norm": 15.75, "learning_rate": 1.4977140510749354e-06, "loss": 0.6571, "step": 10301 }, { "epoch": 0.8778866638261611, "grad_norm": 14.625, "learning_rate": 1.4975493458661564e-06, "loss": 0.7491, "step": 10302 }, { "epoch": 0.877971878994461, "grad_norm": 15.0625, "learning_rate": 1.497384636183911e-06, "loss": 0.5453, "step": 10303 }, { "epoch": 0.878057094162761, "grad_norm": 15.3125, "learning_rate": 1.4972199220311766e-06, "loss": 0.9451, "step": 10304 }, { "epoch": 0.878142309331061, "grad_norm": 12.75, "learning_rate": 1.4970552034109298e-06, "loss": 0.6084, "step": 10305 }, { "epoch": 0.8782275244993609, "grad_norm": 15.1875, "learning_rate": 1.4968904803261475e-06, "loss": 0.8222, "step": 10306 }, { "epoch": 0.8783127396676609, "grad_norm": 19.625, "learning_rate": 1.4967257527798052e-06, "loss": 0.845, "step": 10307 }, { "epoch": 0.8783979548359608, "grad_norm": 12.0625, "learning_rate": 1.4965610207748806e-06, "loss": 0.2869, "step": 10308 }, { "epoch": 0.8784831700042608, "grad_norm": 15.9375, "learning_rate": 1.4963962843143503e-06, "loss": 1.0366, "step": 10309 }, { "epoch": 0.8785683851725608, "grad_norm": 15.5625, "learning_rate": 1.4962315434011918e-06, "loss": 0.705, "step": 10310 }, { "epoch": 0.8786536003408607, "grad_norm": 15.25, "learning_rate": 1.4960667980383816e-06, "loss": 0.4886, "step": 10311 }, { "epoch": 0.8787388155091607, "grad_norm": 18.375, "learning_rate": 1.4959020482288966e-06, "loss": 0.8341, "step": 10312 }, { "epoch": 0.8788240306774606, "grad_norm": 11.8125, "learning_rate": 1.4957372939757148e-06, "loss": 0.5601, "step": 10313 }, { "epoch": 0.8789092458457606, "grad_norm": 14.625, "learning_rate": 1.4955725352818127e-06, "loss": 0.8358, "step": 10314 }, { "epoch": 0.8789944610140605, "grad_norm": 13.3125, "learning_rate": 1.495407772150168e-06, "loss": 0.76, "step": 10315 }, { "epoch": 0.8790796761823605, "grad_norm": 12.3125, "learning_rate": 1.495243004583758e-06, "loss": 0.7222, "step": 10316 }, { "epoch": 0.8791648913506604, "grad_norm": 13.3125, "learning_rate": 1.4950782325855603e-06, "loss": 0.8994, "step": 10317 }, { "epoch": 0.8792501065189604, "grad_norm": 15.5, "learning_rate": 1.4949134561585527e-06, "loss": 0.936, "step": 10318 }, { "epoch": 0.8793353216872604, "grad_norm": 14.3125, "learning_rate": 1.4947486753057125e-06, "loss": 0.5492, "step": 10319 }, { "epoch": 0.8794205368555603, "grad_norm": 12.9375, "learning_rate": 1.4945838900300174e-06, "loss": 0.5649, "step": 10320 }, { "epoch": 0.8795057520238603, "grad_norm": 12.75, "learning_rate": 1.4944191003344457e-06, "loss": 0.7471, "step": 10321 }, { "epoch": 0.8795909671921602, "grad_norm": 11.75, "learning_rate": 1.4942543062219752e-06, "loss": 0.6514, "step": 10322 }, { "epoch": 0.8796761823604602, "grad_norm": 18.25, "learning_rate": 1.4940895076955838e-06, "loss": 0.8471, "step": 10323 }, { "epoch": 0.8797613975287601, "grad_norm": 16.75, "learning_rate": 1.4939247047582491e-06, "loss": 0.8896, "step": 10324 }, { "epoch": 0.8798466126970601, "grad_norm": 16.375, "learning_rate": 1.4937598974129503e-06, "loss": 1.0708, "step": 10325 }, { "epoch": 0.8799318278653601, "grad_norm": 13.75, "learning_rate": 1.4935950856626646e-06, "loss": 0.781, "step": 10326 }, { "epoch": 0.88001704303366, "grad_norm": 17.0, "learning_rate": 1.4934302695103708e-06, "loss": 0.5999, "step": 10327 }, { "epoch": 0.88010225820196, "grad_norm": 13.3125, "learning_rate": 1.4932654489590473e-06, "loss": 0.496, "step": 10328 }, { "epoch": 0.8801874733702599, "grad_norm": 12.375, "learning_rate": 1.4931006240116726e-06, "loss": 0.5902, "step": 10329 }, { "epoch": 0.8802726885385599, "grad_norm": 12.125, "learning_rate": 1.4929357946712253e-06, "loss": 0.3672, "step": 10330 }, { "epoch": 0.8803579037068598, "grad_norm": 12.0625, "learning_rate": 1.4927709609406835e-06, "loss": 0.6805, "step": 10331 }, { "epoch": 0.8804431188751598, "grad_norm": 19.0, "learning_rate": 1.4926061228230264e-06, "loss": 1.146, "step": 10332 }, { "epoch": 0.8805283340434598, "grad_norm": 12.375, "learning_rate": 1.4924412803212327e-06, "loss": 0.5561, "step": 10333 }, { "epoch": 0.8806135492117597, "grad_norm": 16.0, "learning_rate": 1.4922764334382816e-06, "loss": 0.8491, "step": 10334 }, { "epoch": 0.8806987643800597, "grad_norm": 10.375, "learning_rate": 1.4921115821771514e-06, "loss": 0.3205, "step": 10335 }, { "epoch": 0.8807839795483596, "grad_norm": 14.875, "learning_rate": 1.4919467265408216e-06, "loss": 0.6154, "step": 10336 }, { "epoch": 0.8808691947166596, "grad_norm": 16.125, "learning_rate": 1.4917818665322708e-06, "loss": 0.8127, "step": 10337 }, { "epoch": 0.8809544098849595, "grad_norm": 19.25, "learning_rate": 1.4916170021544792e-06, "loss": 0.6141, "step": 10338 }, { "epoch": 0.8810396250532595, "grad_norm": 19.75, "learning_rate": 1.491452133410425e-06, "loss": 1.012, "step": 10339 }, { "epoch": 0.8811248402215595, "grad_norm": 16.875, "learning_rate": 1.491287260303088e-06, "loss": 0.8337, "step": 10340 }, { "epoch": 0.8812100553898594, "grad_norm": 14.4375, "learning_rate": 1.4911223828354473e-06, "loss": 1.0503, "step": 10341 }, { "epoch": 0.8812952705581594, "grad_norm": 14.1875, "learning_rate": 1.4909575010104832e-06, "loss": 0.8615, "step": 10342 }, { "epoch": 0.8813804857264593, "grad_norm": 23.25, "learning_rate": 1.4907926148311747e-06, "loss": 0.7213, "step": 10343 }, { "epoch": 0.8814657008947593, "grad_norm": 17.125, "learning_rate": 1.4906277243005012e-06, "loss": 0.7568, "step": 10344 }, { "epoch": 0.8815509160630592, "grad_norm": 23.25, "learning_rate": 1.4904628294214426e-06, "loss": 1.1367, "step": 10345 }, { "epoch": 0.8816361312313592, "grad_norm": 19.25, "learning_rate": 1.4902979301969795e-06, "loss": 0.9534, "step": 10346 }, { "epoch": 0.8817213463996592, "grad_norm": 19.25, "learning_rate": 1.4901330266300905e-06, "loss": 1.0428, "step": 10347 }, { "epoch": 0.8818065615679591, "grad_norm": 12.9375, "learning_rate": 1.4899681187237567e-06, "loss": 0.5236, "step": 10348 }, { "epoch": 0.8818917767362591, "grad_norm": 25.125, "learning_rate": 1.4898032064809575e-06, "loss": 0.7212, "step": 10349 }, { "epoch": 0.881976991904559, "grad_norm": 22.5, "learning_rate": 1.4896382899046733e-06, "loss": 1.0081, "step": 10350 }, { "epoch": 0.882062207072859, "grad_norm": 39.0, "learning_rate": 1.489473368997884e-06, "loss": 0.8309, "step": 10351 }, { "epoch": 0.8821474222411589, "grad_norm": 14.625, "learning_rate": 1.4893084437635702e-06, "loss": 0.6007, "step": 10352 }, { "epoch": 0.8822326374094589, "grad_norm": 12.25, "learning_rate": 1.4891435142047123e-06, "loss": 0.553, "step": 10353 }, { "epoch": 0.8823178525777589, "grad_norm": 17.375, "learning_rate": 1.4889785803242907e-06, "loss": 0.8502, "step": 10354 }, { "epoch": 0.8824030677460588, "grad_norm": 15.5, "learning_rate": 1.4888136421252857e-06, "loss": 0.7555, "step": 10355 }, { "epoch": 0.8824882829143588, "grad_norm": 19.625, "learning_rate": 1.4886486996106781e-06, "loss": 0.8821, "step": 10356 }, { "epoch": 0.8825734980826587, "grad_norm": 13.5625, "learning_rate": 1.4884837527834485e-06, "loss": 0.4516, "step": 10357 }, { "epoch": 0.8826587132509587, "grad_norm": 17.75, "learning_rate": 1.4883188016465776e-06, "loss": 0.7758, "step": 10358 }, { "epoch": 0.8827439284192586, "grad_norm": 16.875, "learning_rate": 1.4881538462030466e-06, "loss": 0.9124, "step": 10359 }, { "epoch": 0.8828291435875586, "grad_norm": 11.3125, "learning_rate": 1.4879888864558356e-06, "loss": 0.56, "step": 10360 }, { "epoch": 0.8829143587558586, "grad_norm": 13.1875, "learning_rate": 1.4878239224079265e-06, "loss": 0.7051, "step": 10361 }, { "epoch": 0.8829995739241585, "grad_norm": 11.8125, "learning_rate": 1.4876589540622998e-06, "loss": 0.6376, "step": 10362 }, { "epoch": 0.8830847890924585, "grad_norm": 12.125, "learning_rate": 1.487493981421937e-06, "loss": 0.6026, "step": 10363 }, { "epoch": 0.8831700042607584, "grad_norm": 18.125, "learning_rate": 1.487329004489819e-06, "loss": 0.9058, "step": 10364 }, { "epoch": 0.8832552194290584, "grad_norm": 20.375, "learning_rate": 1.487164023268927e-06, "loss": 0.748, "step": 10365 }, { "epoch": 0.8833404345973583, "grad_norm": 13.25, "learning_rate": 1.4869990377622432e-06, "loss": 0.6584, "step": 10366 }, { "epoch": 0.8834256497656583, "grad_norm": 29.375, "learning_rate": 1.4868340479727477e-06, "loss": 1.1751, "step": 10367 }, { "epoch": 0.8835108649339583, "grad_norm": 20.375, "learning_rate": 1.4866690539034235e-06, "loss": 0.9223, "step": 10368 }, { "epoch": 0.8835960801022582, "grad_norm": 12.5625, "learning_rate": 1.486504055557251e-06, "loss": 0.6066, "step": 10369 }, { "epoch": 0.8836812952705582, "grad_norm": 15.8125, "learning_rate": 1.4863390529372126e-06, "loss": 0.8881, "step": 10370 }, { "epoch": 0.8837665104388581, "grad_norm": 18.0, "learning_rate": 1.48617404604629e-06, "loss": 1.0259, "step": 10371 }, { "epoch": 0.8838517256071581, "grad_norm": 18.875, "learning_rate": 1.4860090348874648e-06, "loss": 0.8101, "step": 10372 }, { "epoch": 0.883936940775458, "grad_norm": 19.375, "learning_rate": 1.485844019463719e-06, "loss": 0.7419, "step": 10373 }, { "epoch": 0.884022155943758, "grad_norm": 11.4375, "learning_rate": 1.4856789997780347e-06, "loss": 0.482, "step": 10374 }, { "epoch": 0.884107371112058, "grad_norm": 14.0, "learning_rate": 1.4855139758333936e-06, "loss": 0.4415, "step": 10375 }, { "epoch": 0.8841925862803579, "grad_norm": 12.875, "learning_rate": 1.4853489476327785e-06, "loss": 0.7432, "step": 10376 }, { "epoch": 0.8842778014486579, "grad_norm": 16.0, "learning_rate": 1.4851839151791707e-06, "loss": 0.6941, "step": 10377 }, { "epoch": 0.8843630166169578, "grad_norm": 13.125, "learning_rate": 1.4850188784755538e-06, "loss": 0.6133, "step": 10378 }, { "epoch": 0.8844482317852578, "grad_norm": 22.25, "learning_rate": 1.4848538375249088e-06, "loss": 0.6276, "step": 10379 }, { "epoch": 0.8845334469535577, "grad_norm": 11.1875, "learning_rate": 1.484688792330219e-06, "loss": 0.3604, "step": 10380 }, { "epoch": 0.8846186621218577, "grad_norm": 19.375, "learning_rate": 1.484523742894467e-06, "loss": 0.4631, "step": 10381 }, { "epoch": 0.8847038772901576, "grad_norm": 12.625, "learning_rate": 1.4843586892206346e-06, "loss": 0.5561, "step": 10382 }, { "epoch": 0.8847890924584576, "grad_norm": 15.8125, "learning_rate": 1.4841936313117056e-06, "loss": 0.5532, "step": 10383 }, { "epoch": 0.8848743076267576, "grad_norm": 13.9375, "learning_rate": 1.4840285691706618e-06, "loss": 0.8556, "step": 10384 }, { "epoch": 0.8849595227950575, "grad_norm": 16.625, "learning_rate": 1.4838635028004866e-06, "loss": 0.7676, "step": 10385 }, { "epoch": 0.8850447379633575, "grad_norm": 12.4375, "learning_rate": 1.4836984322041626e-06, "loss": 0.5567, "step": 10386 }, { "epoch": 0.8851299531316574, "grad_norm": 17.0, "learning_rate": 1.4835333573846733e-06, "loss": 1.0572, "step": 10387 }, { "epoch": 0.8852151682999574, "grad_norm": 11.0, "learning_rate": 1.4833682783450007e-06, "loss": 0.4456, "step": 10388 }, { "epoch": 0.8853003834682573, "grad_norm": 11.125, "learning_rate": 1.4832031950881292e-06, "loss": 0.4612, "step": 10389 }, { "epoch": 0.8853855986365573, "grad_norm": 16.25, "learning_rate": 1.4830381076170411e-06, "loss": 0.7626, "step": 10390 }, { "epoch": 0.8854708138048573, "grad_norm": 16.375, "learning_rate": 1.4828730159347204e-06, "loss": 0.5575, "step": 10391 }, { "epoch": 0.8855560289731572, "grad_norm": 13.9375, "learning_rate": 1.4827079200441498e-06, "loss": 0.8416, "step": 10392 }, { "epoch": 0.8856412441414572, "grad_norm": 11.8125, "learning_rate": 1.4825428199483135e-06, "loss": 0.6601, "step": 10393 }, { "epoch": 0.8857264593097571, "grad_norm": 11.9375, "learning_rate": 1.4823777156501943e-06, "loss": 0.6939, "step": 10394 }, { "epoch": 0.8858116744780571, "grad_norm": 17.125, "learning_rate": 1.482212607152776e-06, "loss": 0.943, "step": 10395 }, { "epoch": 0.885896889646357, "grad_norm": 13.4375, "learning_rate": 1.4820474944590426e-06, "loss": 0.6323, "step": 10396 }, { "epoch": 0.885982104814657, "grad_norm": 14.25, "learning_rate": 1.4818823775719777e-06, "loss": 0.815, "step": 10397 }, { "epoch": 0.886067319982957, "grad_norm": 17.5, "learning_rate": 1.4817172564945648e-06, "loss": 0.7762, "step": 10398 }, { "epoch": 0.8861525351512569, "grad_norm": 17.875, "learning_rate": 1.4815521312297882e-06, "loss": 0.71, "step": 10399 }, { "epoch": 0.8862377503195569, "grad_norm": 16.625, "learning_rate": 1.481387001780632e-06, "loss": 0.7437, "step": 10400 }, { "epoch": 0.8863229654878568, "grad_norm": 19.25, "learning_rate": 1.4812218681500794e-06, "loss": 0.9264, "step": 10401 }, { "epoch": 0.8864081806561568, "grad_norm": 18.875, "learning_rate": 1.4810567303411155e-06, "loss": 0.8544, "step": 10402 }, { "epoch": 0.8864933958244567, "grad_norm": 17.375, "learning_rate": 1.4808915883567244e-06, "loss": 1.0451, "step": 10403 }, { "epoch": 0.8865786109927567, "grad_norm": 15.125, "learning_rate": 1.4807264421998899e-06, "loss": 0.851, "step": 10404 }, { "epoch": 0.8866638261610567, "grad_norm": 16.375, "learning_rate": 1.4805612918735963e-06, "loss": 0.9114, "step": 10405 }, { "epoch": 0.8867490413293566, "grad_norm": 18.75, "learning_rate": 1.4803961373808285e-06, "loss": 0.7644, "step": 10406 }, { "epoch": 0.8868342564976566, "grad_norm": 14.375, "learning_rate": 1.4802309787245708e-06, "loss": 0.5669, "step": 10407 }, { "epoch": 0.8869194716659565, "grad_norm": 11.3125, "learning_rate": 1.480065815907808e-06, "loss": 0.4577, "step": 10408 }, { "epoch": 0.8870046868342565, "grad_norm": 17.875, "learning_rate": 1.4799006489335244e-06, "loss": 1.0599, "step": 10409 }, { "epoch": 0.8870899020025564, "grad_norm": 10.0, "learning_rate": 1.4797354778047048e-06, "loss": 0.3546, "step": 10410 }, { "epoch": 0.8871751171708564, "grad_norm": 22.5, "learning_rate": 1.4795703025243344e-06, "loss": 1.0385, "step": 10411 }, { "epoch": 0.8872603323391564, "grad_norm": 11.0, "learning_rate": 1.4794051230953973e-06, "loss": 0.5235, "step": 10412 }, { "epoch": 0.8873455475074563, "grad_norm": 13.75, "learning_rate": 1.4792399395208792e-06, "loss": 0.7556, "step": 10413 }, { "epoch": 0.8874307626757563, "grad_norm": 13.375, "learning_rate": 1.4790747518037652e-06, "loss": 0.7138, "step": 10414 }, { "epoch": 0.8875159778440562, "grad_norm": 14.875, "learning_rate": 1.4789095599470396e-06, "loss": 0.9096, "step": 10415 }, { "epoch": 0.8876011930123562, "grad_norm": 17.125, "learning_rate": 1.478744363953688e-06, "loss": 0.8506, "step": 10416 }, { "epoch": 0.8876864081806561, "grad_norm": 13.625, "learning_rate": 1.4785791638266961e-06, "loss": 0.5188, "step": 10417 }, { "epoch": 0.8877716233489561, "grad_norm": 15.8125, "learning_rate": 1.4784139595690486e-06, "loss": 0.7408, "step": 10418 }, { "epoch": 0.887856838517256, "grad_norm": 20.875, "learning_rate": 1.4782487511837314e-06, "loss": 0.9588, "step": 10419 }, { "epoch": 0.887942053685556, "grad_norm": 13.4375, "learning_rate": 1.4780835386737297e-06, "loss": 0.7271, "step": 10420 }, { "epoch": 0.888027268853856, "grad_norm": 12.75, "learning_rate": 1.477918322042029e-06, "loss": 0.6001, "step": 10421 }, { "epoch": 0.8881124840221559, "grad_norm": 13.0, "learning_rate": 1.4777531012916147e-06, "loss": 0.5802, "step": 10422 }, { "epoch": 0.8881976991904559, "grad_norm": 11.1875, "learning_rate": 1.4775878764254737e-06, "loss": 0.4183, "step": 10423 }, { "epoch": 0.8882829143587558, "grad_norm": 28.375, "learning_rate": 1.4774226474465902e-06, "loss": 1.2556, "step": 10424 }, { "epoch": 0.8883681295270558, "grad_norm": 14.5, "learning_rate": 1.477257414357951e-06, "loss": 0.9505, "step": 10425 }, { "epoch": 0.8884533446953558, "grad_norm": 13.625, "learning_rate": 1.4770921771625415e-06, "loss": 0.6335, "step": 10426 }, { "epoch": 0.8885385598636557, "grad_norm": 14.1875, "learning_rate": 1.4769269358633485e-06, "loss": 0.6803, "step": 10427 }, { "epoch": 0.8886237750319557, "grad_norm": 13.625, "learning_rate": 1.4767616904633572e-06, "loss": 0.6328, "step": 10428 }, { "epoch": 0.8887089902002556, "grad_norm": 19.75, "learning_rate": 1.4765964409655544e-06, "loss": 0.8183, "step": 10429 }, { "epoch": 0.8887942053685556, "grad_norm": 18.75, "learning_rate": 1.4764311873729259e-06, "loss": 0.8082, "step": 10430 }, { "epoch": 0.8888794205368555, "grad_norm": 12.5625, "learning_rate": 1.4762659296884581e-06, "loss": 0.6917, "step": 10431 }, { "epoch": 0.8889646357051555, "grad_norm": 17.375, "learning_rate": 1.4761006679151377e-06, "loss": 0.4813, "step": 10432 }, { "epoch": 0.8890498508734554, "grad_norm": 17.0, "learning_rate": 1.4759354020559501e-06, "loss": 0.9395, "step": 10433 }, { "epoch": 0.8891350660417554, "grad_norm": 13.375, "learning_rate": 1.4757701321138834e-06, "loss": 0.4614, "step": 10434 }, { "epoch": 0.8892202812100554, "grad_norm": 19.875, "learning_rate": 1.475604858091923e-06, "loss": 0.5981, "step": 10435 }, { "epoch": 0.8893054963783553, "grad_norm": 11.75, "learning_rate": 1.4754395799930565e-06, "loss": 0.4949, "step": 10436 }, { "epoch": 0.8893907115466553, "grad_norm": 15.75, "learning_rate": 1.4752742978202694e-06, "loss": 0.7965, "step": 10437 }, { "epoch": 0.8894759267149552, "grad_norm": 16.125, "learning_rate": 1.4751090115765495e-06, "loss": 0.8804, "step": 10438 }, { "epoch": 0.8895611418832552, "grad_norm": 15.3125, "learning_rate": 1.4749437212648832e-06, "loss": 0.8737, "step": 10439 }, { "epoch": 0.8896463570515551, "grad_norm": 14.9375, "learning_rate": 1.4747784268882584e-06, "loss": 0.5247, "step": 10440 }, { "epoch": 0.8897315722198551, "grad_norm": 8.9375, "learning_rate": 1.4746131284496606e-06, "loss": 0.2444, "step": 10441 }, { "epoch": 0.8898167873881551, "grad_norm": 17.5, "learning_rate": 1.4744478259520777e-06, "loss": 0.4327, "step": 10442 }, { "epoch": 0.889902002556455, "grad_norm": 19.375, "learning_rate": 1.4742825193984973e-06, "loss": 0.9235, "step": 10443 }, { "epoch": 0.889987217724755, "grad_norm": 16.25, "learning_rate": 1.474117208791906e-06, "loss": 0.8009, "step": 10444 }, { "epoch": 0.8900724328930549, "grad_norm": 12.25, "learning_rate": 1.4739518941352915e-06, "loss": 0.5349, "step": 10445 }, { "epoch": 0.8901576480613549, "grad_norm": 14.4375, "learning_rate": 1.473786575431641e-06, "loss": 0.7068, "step": 10446 }, { "epoch": 0.8902428632296548, "grad_norm": 21.625, "learning_rate": 1.4736212526839418e-06, "loss": 0.6457, "step": 10447 }, { "epoch": 0.8903280783979548, "grad_norm": 35.5, "learning_rate": 1.4734559258951822e-06, "loss": 1.0247, "step": 10448 }, { "epoch": 0.8904132935662548, "grad_norm": 14.0625, "learning_rate": 1.473290595068349e-06, "loss": 0.7623, "step": 10449 }, { "epoch": 0.8904985087345547, "grad_norm": 10.75, "learning_rate": 1.4731252602064303e-06, "loss": 0.3704, "step": 10450 }, { "epoch": 0.8905837239028547, "grad_norm": 15.3125, "learning_rate": 1.4729599213124138e-06, "loss": 0.8652, "step": 10451 }, { "epoch": 0.8906689390711546, "grad_norm": 14.9375, "learning_rate": 1.4727945783892872e-06, "loss": 1.0799, "step": 10452 }, { "epoch": 0.8907541542394546, "grad_norm": 14.9375, "learning_rate": 1.4726292314400386e-06, "loss": 0.5508, "step": 10453 }, { "epoch": 0.8908393694077545, "grad_norm": 16.5, "learning_rate": 1.4724638804676557e-06, "loss": 0.9287, "step": 10454 }, { "epoch": 0.8909245845760545, "grad_norm": 14.5, "learning_rate": 1.4722985254751274e-06, "loss": 0.7868, "step": 10455 }, { "epoch": 0.8910097997443545, "grad_norm": 17.0, "learning_rate": 1.4721331664654407e-06, "loss": 0.8529, "step": 10456 }, { "epoch": 0.8910950149126544, "grad_norm": 13.5625, "learning_rate": 1.4719678034415846e-06, "loss": 0.6699, "step": 10457 }, { "epoch": 0.8911802300809544, "grad_norm": 15.4375, "learning_rate": 1.4718024364065468e-06, "loss": 0.7054, "step": 10458 }, { "epoch": 0.8912654452492543, "grad_norm": 13.3125, "learning_rate": 1.471637065363316e-06, "loss": 0.6142, "step": 10459 }, { "epoch": 0.8913506604175543, "grad_norm": 14.8125, "learning_rate": 1.4714716903148807e-06, "loss": 0.7312, "step": 10460 }, { "epoch": 0.8914358755858542, "grad_norm": 19.25, "learning_rate": 1.4713063112642289e-06, "loss": 0.4958, "step": 10461 }, { "epoch": 0.8915210907541542, "grad_norm": 24.5, "learning_rate": 1.4711409282143495e-06, "loss": 0.79, "step": 10462 }, { "epoch": 0.8916063059224542, "grad_norm": 13.6875, "learning_rate": 1.4709755411682316e-06, "loss": 0.6834, "step": 10463 }, { "epoch": 0.8916915210907541, "grad_norm": 14.875, "learning_rate": 1.4708101501288635e-06, "loss": 0.8516, "step": 10464 }, { "epoch": 0.8917767362590541, "grad_norm": 14.25, "learning_rate": 1.4706447550992333e-06, "loss": 0.4842, "step": 10465 }, { "epoch": 0.891861951427354, "grad_norm": 16.75, "learning_rate": 1.470479356082331e-06, "loss": 0.7353, "step": 10466 }, { "epoch": 0.891947166595654, "grad_norm": 13.25, "learning_rate": 1.470313953081145e-06, "loss": 0.6872, "step": 10467 }, { "epoch": 0.8920323817639539, "grad_norm": 21.5, "learning_rate": 1.4701485460986646e-06, "loss": 1.0796, "step": 10468 }, { "epoch": 0.8921175969322539, "grad_norm": 15.1875, "learning_rate": 1.469983135137878e-06, "loss": 0.8546, "step": 10469 }, { "epoch": 0.8922028121005539, "grad_norm": 15.125, "learning_rate": 1.4698177202017751e-06, "loss": 1.0181, "step": 10470 }, { "epoch": 0.8922880272688538, "grad_norm": 15.0, "learning_rate": 1.469652301293345e-06, "loss": 0.5778, "step": 10471 }, { "epoch": 0.8923732424371538, "grad_norm": 12.0625, "learning_rate": 1.4694868784155776e-06, "loss": 0.793, "step": 10472 }, { "epoch": 0.8924584576054537, "grad_norm": 15.4375, "learning_rate": 1.4693214515714604e-06, "loss": 0.9842, "step": 10473 }, { "epoch": 0.8925436727737537, "grad_norm": 16.5, "learning_rate": 1.469156020763985e-06, "loss": 0.9131, "step": 10474 }, { "epoch": 0.8926288879420536, "grad_norm": 16.5, "learning_rate": 1.4689905859961392e-06, "loss": 1.0074, "step": 10475 }, { "epoch": 0.8927141031103536, "grad_norm": 9.0625, "learning_rate": 1.4688251472709138e-06, "loss": 0.2624, "step": 10476 }, { "epoch": 0.8927993182786536, "grad_norm": 15.125, "learning_rate": 1.4686597045912978e-06, "loss": 0.7598, "step": 10477 }, { "epoch": 0.8928845334469535, "grad_norm": 13.9375, "learning_rate": 1.4684942579602809e-06, "loss": 0.5933, "step": 10478 }, { "epoch": 0.8929697486152535, "grad_norm": 13.375, "learning_rate": 1.4683288073808533e-06, "loss": 0.6608, "step": 10479 }, { "epoch": 0.8930549637835534, "grad_norm": 17.75, "learning_rate": 1.4681633528560043e-06, "loss": 0.7335, "step": 10480 }, { "epoch": 0.8931401789518534, "grad_norm": 19.125, "learning_rate": 1.4679978943887245e-06, "loss": 1.1301, "step": 10481 }, { "epoch": 0.8932253941201533, "grad_norm": 13.3125, "learning_rate": 1.4678324319820032e-06, "loss": 0.8879, "step": 10482 }, { "epoch": 0.8933106092884533, "grad_norm": 16.875, "learning_rate": 1.4676669656388308e-06, "loss": 0.793, "step": 10483 }, { "epoch": 0.8933958244567533, "grad_norm": 17.75, "learning_rate": 1.467501495362198e-06, "loss": 0.6204, "step": 10484 }, { "epoch": 0.8934810396250532, "grad_norm": 12.9375, "learning_rate": 1.467336021155094e-06, "loss": 0.6806, "step": 10485 }, { "epoch": 0.8935662547933532, "grad_norm": 19.125, "learning_rate": 1.4671705430205093e-06, "loss": 1.0269, "step": 10486 }, { "epoch": 0.8936514699616531, "grad_norm": 13.9375, "learning_rate": 1.467005060961435e-06, "loss": 0.7571, "step": 10487 }, { "epoch": 0.8937366851299531, "grad_norm": 14.0625, "learning_rate": 1.4668395749808613e-06, "loss": 0.6106, "step": 10488 }, { "epoch": 0.893821900298253, "grad_norm": 15.25, "learning_rate": 1.4666740850817778e-06, "loss": 0.9196, "step": 10489 }, { "epoch": 0.893907115466553, "grad_norm": 16.5, "learning_rate": 1.4665085912671756e-06, "loss": 0.5685, "step": 10490 }, { "epoch": 0.893992330634853, "grad_norm": 17.0, "learning_rate": 1.4663430935400457e-06, "loss": 0.5427, "step": 10491 }, { "epoch": 0.8940775458031529, "grad_norm": 28.5, "learning_rate": 1.4661775919033788e-06, "loss": 0.7841, "step": 10492 }, { "epoch": 0.8941627609714529, "grad_norm": 11.1875, "learning_rate": 1.4660120863601654e-06, "loss": 0.5337, "step": 10493 }, { "epoch": 0.8942479761397529, "grad_norm": 14.1875, "learning_rate": 1.4658465769133958e-06, "loss": 0.7457, "step": 10494 }, { "epoch": 0.8943331913080529, "grad_norm": 26.625, "learning_rate": 1.465681063566062e-06, "loss": 1.1096, "step": 10495 }, { "epoch": 0.8944184064763528, "grad_norm": 12.0625, "learning_rate": 1.4655155463211545e-06, "loss": 0.5296, "step": 10496 }, { "epoch": 0.8945036216446528, "grad_norm": 19.25, "learning_rate": 1.465350025181664e-06, "loss": 0.5372, "step": 10497 }, { "epoch": 0.8945888368129528, "grad_norm": 14.4375, "learning_rate": 1.4651845001505824e-06, "loss": 0.5976, "step": 10498 }, { "epoch": 0.8946740519812527, "grad_norm": 14.875, "learning_rate": 1.4650189712309004e-06, "loss": 0.5368, "step": 10499 }, { "epoch": 0.8947592671495527, "grad_norm": 16.5, "learning_rate": 1.4648534384256093e-06, "loss": 0.8193, "step": 10500 }, { "epoch": 0.8948444823178526, "grad_norm": 11.25, "learning_rate": 1.4646879017377005e-06, "loss": 0.4957, "step": 10501 }, { "epoch": 0.8949296974861526, "grad_norm": 14.0625, "learning_rate": 1.4645223611701656e-06, "loss": 0.6787, "step": 10502 }, { "epoch": 0.8950149126544525, "grad_norm": 15.0, "learning_rate": 1.4643568167259955e-06, "loss": 0.5931, "step": 10503 }, { "epoch": 0.8951001278227525, "grad_norm": 14.8125, "learning_rate": 1.4641912684081826e-06, "loss": 0.747, "step": 10504 }, { "epoch": 0.8951853429910525, "grad_norm": 11.375, "learning_rate": 1.464025716219718e-06, "loss": 0.6261, "step": 10505 }, { "epoch": 0.8952705581593524, "grad_norm": 14.875, "learning_rate": 1.4638601601635936e-06, "loss": 0.3864, "step": 10506 }, { "epoch": 0.8953557733276524, "grad_norm": 19.25, "learning_rate": 1.4636946002428007e-06, "loss": 1.0076, "step": 10507 }, { "epoch": 0.8954409884959523, "grad_norm": 18.0, "learning_rate": 1.4635290364603316e-06, "loss": 0.7457, "step": 10508 }, { "epoch": 0.8955262036642523, "grad_norm": 15.0, "learning_rate": 1.4633634688191784e-06, "loss": 0.763, "step": 10509 }, { "epoch": 0.8956114188325522, "grad_norm": 14.25, "learning_rate": 1.4631978973223326e-06, "loss": 0.7147, "step": 10510 }, { "epoch": 0.8956966340008522, "grad_norm": 13.8125, "learning_rate": 1.463032321972786e-06, "loss": 0.9861, "step": 10511 }, { "epoch": 0.8957818491691522, "grad_norm": 16.5, "learning_rate": 1.4628667427735315e-06, "loss": 0.8539, "step": 10512 }, { "epoch": 0.8958670643374521, "grad_norm": 17.5, "learning_rate": 1.4627011597275611e-06, "loss": 0.7863, "step": 10513 }, { "epoch": 0.8959522795057521, "grad_norm": 13.25, "learning_rate": 1.4625355728378666e-06, "loss": 0.6729, "step": 10514 }, { "epoch": 0.896037494674052, "grad_norm": 16.375, "learning_rate": 1.4623699821074405e-06, "loss": 0.8672, "step": 10515 }, { "epoch": 0.896122709842352, "grad_norm": 13.8125, "learning_rate": 1.4622043875392755e-06, "loss": 0.4958, "step": 10516 }, { "epoch": 0.8962079250106519, "grad_norm": 15.25, "learning_rate": 1.4620387891363636e-06, "loss": 0.8627, "step": 10517 }, { "epoch": 0.8962931401789519, "grad_norm": 18.25, "learning_rate": 1.4618731869016976e-06, "loss": 0.667, "step": 10518 }, { "epoch": 0.8963783553472519, "grad_norm": 19.875, "learning_rate": 1.4617075808382703e-06, "loss": 1.2579, "step": 10519 }, { "epoch": 0.8964635705155518, "grad_norm": 13.0625, "learning_rate": 1.4615419709490736e-06, "loss": 0.5814, "step": 10520 }, { "epoch": 0.8965487856838518, "grad_norm": 13.625, "learning_rate": 1.4613763572371015e-06, "loss": 0.8294, "step": 10521 }, { "epoch": 0.8966340008521517, "grad_norm": 12.25, "learning_rate": 1.4612107397053455e-06, "loss": 0.6992, "step": 10522 }, { "epoch": 0.8967192160204517, "grad_norm": 13.375, "learning_rate": 1.4610451183567991e-06, "loss": 0.6981, "step": 10523 }, { "epoch": 0.8968044311887516, "grad_norm": 10.3125, "learning_rate": 1.4608794931944554e-06, "loss": 0.2513, "step": 10524 }, { "epoch": 0.8968896463570516, "grad_norm": 15.5, "learning_rate": 1.4607138642213072e-06, "loss": 0.7937, "step": 10525 }, { "epoch": 0.8969748615253516, "grad_norm": 14.9375, "learning_rate": 1.4605482314403474e-06, "loss": 0.7128, "step": 10526 }, { "epoch": 0.8970600766936515, "grad_norm": 17.5, "learning_rate": 1.4603825948545694e-06, "loss": 0.9293, "step": 10527 }, { "epoch": 0.8971452918619515, "grad_norm": 19.125, "learning_rate": 1.4602169544669662e-06, "loss": 1.0681, "step": 10528 }, { "epoch": 0.8972305070302514, "grad_norm": 17.75, "learning_rate": 1.460051310280532e-06, "loss": 0.8572, "step": 10529 }, { "epoch": 0.8973157221985514, "grad_norm": 21.0, "learning_rate": 1.4598856622982586e-06, "loss": 0.9751, "step": 10530 }, { "epoch": 0.8974009373668513, "grad_norm": 15.0625, "learning_rate": 1.4597200105231404e-06, "loss": 0.7405, "step": 10531 }, { "epoch": 0.8974861525351513, "grad_norm": 17.25, "learning_rate": 1.4595543549581706e-06, "loss": 0.7768, "step": 10532 }, { "epoch": 0.8975713677034513, "grad_norm": 11.4375, "learning_rate": 1.459388695606343e-06, "loss": 0.5918, "step": 10533 }, { "epoch": 0.8976565828717512, "grad_norm": 13.9375, "learning_rate": 1.4592230324706514e-06, "loss": 0.4957, "step": 10534 }, { "epoch": 0.8977417980400512, "grad_norm": 18.125, "learning_rate": 1.4590573655540887e-06, "loss": 0.9917, "step": 10535 }, { "epoch": 0.8978270132083511, "grad_norm": 18.5, "learning_rate": 1.4588916948596496e-06, "loss": 0.942, "step": 10536 }, { "epoch": 0.8979122283766511, "grad_norm": 16.75, "learning_rate": 1.4587260203903274e-06, "loss": 0.9032, "step": 10537 }, { "epoch": 0.897997443544951, "grad_norm": 14.875, "learning_rate": 1.4585603421491162e-06, "loss": 0.7249, "step": 10538 }, { "epoch": 0.898082658713251, "grad_norm": 13.6875, "learning_rate": 1.4583946601390095e-06, "loss": 0.7009, "step": 10539 }, { "epoch": 0.898167873881551, "grad_norm": 17.0, "learning_rate": 1.458228974363002e-06, "loss": 0.9629, "step": 10540 }, { "epoch": 0.8982530890498509, "grad_norm": 14.5625, "learning_rate": 1.4580632848240877e-06, "loss": 0.4967, "step": 10541 }, { "epoch": 0.8983383042181509, "grad_norm": 13.8125, "learning_rate": 1.4578975915252607e-06, "loss": 0.6625, "step": 10542 }, { "epoch": 0.8984235193864508, "grad_norm": 15.375, "learning_rate": 1.4577318944695145e-06, "loss": 0.949, "step": 10543 }, { "epoch": 0.8985087345547508, "grad_norm": 12.625, "learning_rate": 1.4575661936598447e-06, "loss": 0.5023, "step": 10544 }, { "epoch": 0.8985939497230507, "grad_norm": 16.75, "learning_rate": 1.4574004890992451e-06, "loss": 0.7043, "step": 10545 }, { "epoch": 0.8986791648913507, "grad_norm": 20.125, "learning_rate": 1.45723478079071e-06, "loss": 0.6524, "step": 10546 }, { "epoch": 0.8987643800596506, "grad_norm": 16.0, "learning_rate": 1.4570690687372338e-06, "loss": 0.7655, "step": 10547 }, { "epoch": 0.8988495952279506, "grad_norm": 25.25, "learning_rate": 1.4569033529418113e-06, "loss": 1.3418, "step": 10548 }, { "epoch": 0.8989348103962506, "grad_norm": 13.3125, "learning_rate": 1.4567376334074372e-06, "loss": 0.6286, "step": 10549 }, { "epoch": 0.8990200255645505, "grad_norm": 11.75, "learning_rate": 1.4565719101371062e-06, "loss": 0.4998, "step": 10550 }, { "epoch": 0.8991052407328505, "grad_norm": 17.0, "learning_rate": 1.4564061831338132e-06, "loss": 0.8654, "step": 10551 }, { "epoch": 0.8991904559011504, "grad_norm": 18.25, "learning_rate": 1.4562404524005527e-06, "loss": 0.9501, "step": 10552 }, { "epoch": 0.8992756710694504, "grad_norm": 12.875, "learning_rate": 1.4560747179403203e-06, "loss": 0.5746, "step": 10553 }, { "epoch": 0.8993608862377503, "grad_norm": 18.0, "learning_rate": 1.4559089797561099e-06, "loss": 1.26, "step": 10554 }, { "epoch": 0.8994461014060503, "grad_norm": 16.5, "learning_rate": 1.4557432378509176e-06, "loss": 0.4765, "step": 10555 }, { "epoch": 0.8995313165743503, "grad_norm": 15.75, "learning_rate": 1.4555774922277377e-06, "loss": 0.3654, "step": 10556 }, { "epoch": 0.8996165317426502, "grad_norm": 14.8125, "learning_rate": 1.4554117428895664e-06, "loss": 0.679, "step": 10557 }, { "epoch": 0.8997017469109502, "grad_norm": 12.1875, "learning_rate": 1.4552459898393977e-06, "loss": 0.4775, "step": 10558 }, { "epoch": 0.8997869620792501, "grad_norm": 12.9375, "learning_rate": 1.455080233080228e-06, "loss": 0.3762, "step": 10559 }, { "epoch": 0.8998721772475501, "grad_norm": 16.875, "learning_rate": 1.4549144726150522e-06, "loss": 0.7119, "step": 10560 }, { "epoch": 0.89995739241585, "grad_norm": 19.625, "learning_rate": 1.4547487084468659e-06, "loss": 0.9674, "step": 10561 }, { "epoch": 0.90004260758415, "grad_norm": 12.6875, "learning_rate": 1.4545829405786643e-06, "loss": 0.7943, "step": 10562 }, { "epoch": 0.90012782275245, "grad_norm": 12.125, "learning_rate": 1.4544171690134434e-06, "loss": 0.5507, "step": 10563 }, { "epoch": 0.9002130379207499, "grad_norm": 11.375, "learning_rate": 1.4542513937541983e-06, "loss": 0.3777, "step": 10564 }, { "epoch": 0.9002982530890499, "grad_norm": 17.375, "learning_rate": 1.4540856148039256e-06, "loss": 0.6855, "step": 10565 }, { "epoch": 0.9003834682573498, "grad_norm": 14.4375, "learning_rate": 1.4539198321656207e-06, "loss": 0.6312, "step": 10566 }, { "epoch": 0.9004686834256498, "grad_norm": 16.5, "learning_rate": 1.4537540458422788e-06, "loss": 0.8699, "step": 10567 }, { "epoch": 0.9005538985939497, "grad_norm": 14.1875, "learning_rate": 1.453588255836897e-06, "loss": 0.7147, "step": 10568 }, { "epoch": 0.9006391137622497, "grad_norm": 13.5625, "learning_rate": 1.4534224621524703e-06, "loss": 0.6076, "step": 10569 }, { "epoch": 0.9007243289305497, "grad_norm": 17.125, "learning_rate": 1.4532566647919956e-06, "loss": 1.1227, "step": 10570 }, { "epoch": 0.9008095440988496, "grad_norm": 20.125, "learning_rate": 1.453090863758468e-06, "loss": 0.8747, "step": 10571 }, { "epoch": 0.9008947592671496, "grad_norm": 16.125, "learning_rate": 1.4529250590548846e-06, "loss": 0.6937, "step": 10572 }, { "epoch": 0.9009799744354495, "grad_norm": 15.75, "learning_rate": 1.4527592506842411e-06, "loss": 1.0556, "step": 10573 }, { "epoch": 0.9010651896037495, "grad_norm": 14.5625, "learning_rate": 1.4525934386495343e-06, "loss": 0.666, "step": 10574 }, { "epoch": 0.9011504047720494, "grad_norm": 18.25, "learning_rate": 1.4524276229537605e-06, "loss": 0.8782, "step": 10575 }, { "epoch": 0.9012356199403494, "grad_norm": 14.8125, "learning_rate": 1.4522618035999156e-06, "loss": 0.5147, "step": 10576 }, { "epoch": 0.9013208351086494, "grad_norm": 15.4375, "learning_rate": 1.4520959805909966e-06, "loss": 0.8113, "step": 10577 }, { "epoch": 0.9014060502769493, "grad_norm": 16.5, "learning_rate": 1.4519301539300005e-06, "loss": 0.3953, "step": 10578 }, { "epoch": 0.9014912654452493, "grad_norm": 12.6875, "learning_rate": 1.4517643236199228e-06, "loss": 0.5396, "step": 10579 }, { "epoch": 0.9015764806135492, "grad_norm": 13.75, "learning_rate": 1.4515984896637612e-06, "loss": 0.7874, "step": 10580 }, { "epoch": 0.9016616957818492, "grad_norm": 12.6875, "learning_rate": 1.4514326520645116e-06, "loss": 0.6227, "step": 10581 }, { "epoch": 0.9017469109501491, "grad_norm": 14.625, "learning_rate": 1.4512668108251722e-06, "loss": 0.928, "step": 10582 }, { "epoch": 0.9018321261184491, "grad_norm": 13.125, "learning_rate": 1.4511009659487391e-06, "loss": 0.4735, "step": 10583 }, { "epoch": 0.901917341286749, "grad_norm": 9.9375, "learning_rate": 1.450935117438209e-06, "loss": 0.3723, "step": 10584 }, { "epoch": 0.902002556455049, "grad_norm": 13.25, "learning_rate": 1.450769265296579e-06, "loss": 0.6483, "step": 10585 }, { "epoch": 0.902087771623349, "grad_norm": 15.875, "learning_rate": 1.450603409526847e-06, "loss": 0.6936, "step": 10586 }, { "epoch": 0.9021729867916489, "grad_norm": 24.0, "learning_rate": 1.4504375501320098e-06, "loss": 0.9486, "step": 10587 }, { "epoch": 0.9022582019599489, "grad_norm": 10.625, "learning_rate": 1.4502716871150637e-06, "loss": 0.3607, "step": 10588 }, { "epoch": 0.9023434171282488, "grad_norm": 12.0625, "learning_rate": 1.4501058204790075e-06, "loss": 0.8044, "step": 10589 }, { "epoch": 0.9024286322965488, "grad_norm": 16.0, "learning_rate": 1.449939950226838e-06, "loss": 0.7381, "step": 10590 }, { "epoch": 0.9025138474648488, "grad_norm": 12.1875, "learning_rate": 1.4497740763615523e-06, "loss": 0.5229, "step": 10591 }, { "epoch": 0.9025990626331487, "grad_norm": 11.0, "learning_rate": 1.449608198886148e-06, "loss": 0.4299, "step": 10592 }, { "epoch": 0.9026842778014487, "grad_norm": 20.0, "learning_rate": 1.4494423178036229e-06, "loss": 1.2262, "step": 10593 }, { "epoch": 0.9027694929697486, "grad_norm": 14.875, "learning_rate": 1.4492764331169748e-06, "loss": 0.942, "step": 10594 }, { "epoch": 0.9028547081380486, "grad_norm": 12.1875, "learning_rate": 1.4491105448292012e-06, "loss": 0.611, "step": 10595 }, { "epoch": 0.9029399233063485, "grad_norm": 12.75, "learning_rate": 1.4489446529432993e-06, "loss": 0.5748, "step": 10596 }, { "epoch": 0.9030251384746485, "grad_norm": 12.875, "learning_rate": 1.4487787574622677e-06, "loss": 0.4048, "step": 10597 }, { "epoch": 0.9031103536429484, "grad_norm": 16.5, "learning_rate": 1.4486128583891043e-06, "loss": 0.9708, "step": 10598 }, { "epoch": 0.9031955688112484, "grad_norm": 14.5, "learning_rate": 1.448446955726807e-06, "loss": 0.7211, "step": 10599 }, { "epoch": 0.9032807839795484, "grad_norm": 15.9375, "learning_rate": 1.448281049478373e-06, "loss": 0.6689, "step": 10600 }, { "epoch": 0.9033659991478483, "grad_norm": 13.3125, "learning_rate": 1.4481151396468015e-06, "loss": 0.5666, "step": 10601 }, { "epoch": 0.9034512143161483, "grad_norm": 11.3125, "learning_rate": 1.4479492262350902e-06, "loss": 0.3463, "step": 10602 }, { "epoch": 0.9035364294844482, "grad_norm": 11.1875, "learning_rate": 1.447783309246237e-06, "loss": 0.2918, "step": 10603 }, { "epoch": 0.9036216446527482, "grad_norm": 15.0625, "learning_rate": 1.447617388683241e-06, "loss": 0.5017, "step": 10604 }, { "epoch": 0.9037068598210481, "grad_norm": 14.5, "learning_rate": 1.4474514645490995e-06, "loss": 0.6685, "step": 10605 }, { "epoch": 0.9037920749893481, "grad_norm": 16.5, "learning_rate": 1.4472855368468122e-06, "loss": 0.793, "step": 10606 }, { "epoch": 0.9038772901576481, "grad_norm": 14.125, "learning_rate": 1.4471196055793762e-06, "loss": 0.5274, "step": 10607 }, { "epoch": 0.903962505325948, "grad_norm": 18.625, "learning_rate": 1.4469536707497912e-06, "loss": 0.7363, "step": 10608 }, { "epoch": 0.904047720494248, "grad_norm": 17.0, "learning_rate": 1.4467877323610546e-06, "loss": 0.8751, "step": 10609 }, { "epoch": 0.9041329356625479, "grad_norm": 12.625, "learning_rate": 1.4466217904161667e-06, "loss": 0.7177, "step": 10610 }, { "epoch": 0.9042181508308479, "grad_norm": 14.1875, "learning_rate": 1.4464558449181248e-06, "loss": 0.6765, "step": 10611 }, { "epoch": 0.9043033659991478, "grad_norm": 15.5625, "learning_rate": 1.4462898958699284e-06, "loss": 0.7753, "step": 10612 }, { "epoch": 0.9043885811674478, "grad_norm": 15.25, "learning_rate": 1.4461239432745758e-06, "loss": 0.6275, "step": 10613 }, { "epoch": 0.9044737963357478, "grad_norm": 13.8125, "learning_rate": 1.445957987135067e-06, "loss": 0.5641, "step": 10614 }, { "epoch": 0.9045590115040477, "grad_norm": 15.375, "learning_rate": 1.4457920274544e-06, "loss": 0.9015, "step": 10615 }, { "epoch": 0.9046442266723477, "grad_norm": 13.6875, "learning_rate": 1.4456260642355741e-06, "loss": 0.7694, "step": 10616 }, { "epoch": 0.9047294418406476, "grad_norm": 15.4375, "learning_rate": 1.4454600974815884e-06, "loss": 0.8692, "step": 10617 }, { "epoch": 0.9048146570089476, "grad_norm": 16.375, "learning_rate": 1.4452941271954424e-06, "loss": 0.7682, "step": 10618 }, { "epoch": 0.9048998721772475, "grad_norm": 13.125, "learning_rate": 1.4451281533801351e-06, "loss": 0.8098, "step": 10619 }, { "epoch": 0.9049850873455475, "grad_norm": 12.5, "learning_rate": 1.4449621760386657e-06, "loss": 0.6211, "step": 10620 }, { "epoch": 0.9050703025138475, "grad_norm": 13.3125, "learning_rate": 1.444796195174034e-06, "loss": 0.518, "step": 10621 }, { "epoch": 0.9051555176821474, "grad_norm": 14.375, "learning_rate": 1.444630210789239e-06, "loss": 0.6669, "step": 10622 }, { "epoch": 0.9052407328504474, "grad_norm": 12.5625, "learning_rate": 1.4444642228872807e-06, "loss": 0.5608, "step": 10623 }, { "epoch": 0.9053259480187473, "grad_norm": 18.875, "learning_rate": 1.444298231471158e-06, "loss": 0.6591, "step": 10624 }, { "epoch": 0.9054111631870473, "grad_norm": 33.5, "learning_rate": 1.4441322365438709e-06, "loss": 0.9929, "step": 10625 }, { "epoch": 0.9054963783553472, "grad_norm": 13.625, "learning_rate": 1.4439662381084191e-06, "loss": 0.598, "step": 10626 }, { "epoch": 0.9055815935236472, "grad_norm": 16.125, "learning_rate": 1.4438002361678024e-06, "loss": 0.9534, "step": 10627 }, { "epoch": 0.9056668086919472, "grad_norm": 19.375, "learning_rate": 1.4436342307250203e-06, "loss": 1.1651, "step": 10628 }, { "epoch": 0.9057520238602471, "grad_norm": 16.75, "learning_rate": 1.4434682217830731e-06, "loss": 0.8231, "step": 10629 }, { "epoch": 0.9058372390285471, "grad_norm": 19.0, "learning_rate": 1.4433022093449608e-06, "loss": 1.0404, "step": 10630 }, { "epoch": 0.905922454196847, "grad_norm": 13.25, "learning_rate": 1.4431361934136832e-06, "loss": 0.7809, "step": 10631 }, { "epoch": 0.906007669365147, "grad_norm": 14.625, "learning_rate": 1.4429701739922397e-06, "loss": 0.7415, "step": 10632 }, { "epoch": 0.9060928845334469, "grad_norm": 11.8125, "learning_rate": 1.442804151083632e-06, "loss": 0.4831, "step": 10633 }, { "epoch": 0.9061780997017469, "grad_norm": 17.875, "learning_rate": 1.442638124690859e-06, "loss": 0.7217, "step": 10634 }, { "epoch": 0.9062633148700469, "grad_norm": 15.9375, "learning_rate": 1.442472094816921e-06, "loss": 1.0018, "step": 10635 }, { "epoch": 0.9063485300383468, "grad_norm": 16.0, "learning_rate": 1.442306061464819e-06, "loss": 0.9099, "step": 10636 }, { "epoch": 0.9064337452066468, "grad_norm": 11.5625, "learning_rate": 1.4421400246375533e-06, "loss": 0.4625, "step": 10637 }, { "epoch": 0.9065189603749467, "grad_norm": 15.25, "learning_rate": 1.4419739843381239e-06, "loss": 0.905, "step": 10638 }, { "epoch": 0.9066041755432467, "grad_norm": 15.75, "learning_rate": 1.4418079405695317e-06, "loss": 0.9072, "step": 10639 }, { "epoch": 0.9066893907115466, "grad_norm": 14.0, "learning_rate": 1.4416418933347772e-06, "loss": 0.5375, "step": 10640 }, { "epoch": 0.9067746058798466, "grad_norm": 13.5, "learning_rate": 1.441475842636861e-06, "loss": 0.5738, "step": 10641 }, { "epoch": 0.9068598210481466, "grad_norm": 20.0, "learning_rate": 1.4413097884787835e-06, "loss": 0.976, "step": 10642 }, { "epoch": 0.9069450362164465, "grad_norm": 24.75, "learning_rate": 1.441143730863546e-06, "loss": 0.9979, "step": 10643 }, { "epoch": 0.9070302513847465, "grad_norm": 37.75, "learning_rate": 1.4409776697941491e-06, "loss": 0.8972, "step": 10644 }, { "epoch": 0.9071154665530464, "grad_norm": 16.0, "learning_rate": 1.4408116052735932e-06, "loss": 0.8906, "step": 10645 }, { "epoch": 0.9072006817213464, "grad_norm": 11.125, "learning_rate": 1.44064553730488e-06, "loss": 0.4891, "step": 10646 }, { "epoch": 0.9072858968896463, "grad_norm": 11.375, "learning_rate": 1.4404794658910104e-06, "loss": 0.4495, "step": 10647 }, { "epoch": 0.9073711120579463, "grad_norm": 12.5625, "learning_rate": 1.4403133910349854e-06, "loss": 0.6476, "step": 10648 }, { "epoch": 0.9074563272262463, "grad_norm": 10.875, "learning_rate": 1.4401473127398054e-06, "loss": 0.3401, "step": 10649 }, { "epoch": 0.9075415423945462, "grad_norm": 14.5625, "learning_rate": 1.4399812310084728e-06, "loss": 1.1216, "step": 10650 }, { "epoch": 0.9076267575628462, "grad_norm": 14.8125, "learning_rate": 1.4398151458439884e-06, "loss": 0.6491, "step": 10651 }, { "epoch": 0.9077119727311461, "grad_norm": 20.125, "learning_rate": 1.4396490572493535e-06, "loss": 0.4412, "step": 10652 }, { "epoch": 0.9077971878994461, "grad_norm": 11.375, "learning_rate": 1.4394829652275688e-06, "loss": 0.4415, "step": 10653 }, { "epoch": 0.907882403067746, "grad_norm": 11.9375, "learning_rate": 1.4393168697816367e-06, "loss": 0.6512, "step": 10654 }, { "epoch": 0.907967618236046, "grad_norm": 14.25, "learning_rate": 1.439150770914559e-06, "loss": 0.9739, "step": 10655 }, { "epoch": 0.908052833404346, "grad_norm": 13.125, "learning_rate": 1.4389846686293357e-06, "loss": 0.51, "step": 10656 }, { "epoch": 0.9081380485726459, "grad_norm": 12.5, "learning_rate": 1.4388185629289697e-06, "loss": 0.7369, "step": 10657 }, { "epoch": 0.9082232637409459, "grad_norm": 20.375, "learning_rate": 1.4386524538164626e-06, "loss": 0.856, "step": 10658 }, { "epoch": 0.9083084789092458, "grad_norm": 14.1875, "learning_rate": 1.4384863412948161e-06, "loss": 0.669, "step": 10659 }, { "epoch": 0.9083936940775458, "grad_norm": 10.0625, "learning_rate": 1.4383202253670315e-06, "loss": 0.3547, "step": 10660 }, { "epoch": 0.9084789092458457, "grad_norm": 12.625, "learning_rate": 1.4381541060361111e-06, "loss": 0.5269, "step": 10661 }, { "epoch": 0.9085641244141457, "grad_norm": 13.0, "learning_rate": 1.4379879833050568e-06, "loss": 0.4384, "step": 10662 }, { "epoch": 0.9086493395824456, "grad_norm": 17.375, "learning_rate": 1.4378218571768709e-06, "loss": 0.6667, "step": 10663 }, { "epoch": 0.9087345547507456, "grad_norm": 13.125, "learning_rate": 1.437655727654555e-06, "loss": 0.564, "step": 10664 }, { "epoch": 0.9088197699190456, "grad_norm": 14.4375, "learning_rate": 1.4374895947411114e-06, "loss": 1.0638, "step": 10665 }, { "epoch": 0.9089049850873455, "grad_norm": 11.8125, "learning_rate": 1.437323458439542e-06, "loss": 0.6386, "step": 10666 }, { "epoch": 0.9089902002556455, "grad_norm": 15.6875, "learning_rate": 1.43715731875285e-06, "loss": 0.5608, "step": 10667 }, { "epoch": 0.9090754154239454, "grad_norm": 15.5625, "learning_rate": 1.4369911756840365e-06, "loss": 0.8517, "step": 10668 }, { "epoch": 0.9091606305922454, "grad_norm": 16.125, "learning_rate": 1.4368250292361047e-06, "loss": 0.8857, "step": 10669 }, { "epoch": 0.9092458457605453, "grad_norm": 13.8125, "learning_rate": 1.436658879412056e-06, "loss": 0.6905, "step": 10670 }, { "epoch": 0.9093310609288453, "grad_norm": 14.3125, "learning_rate": 1.4364927262148948e-06, "loss": 0.6195, "step": 10671 }, { "epoch": 0.9094162760971453, "grad_norm": 22.0, "learning_rate": 1.436326569647622e-06, "loss": 0.9549, "step": 10672 }, { "epoch": 0.9095014912654452, "grad_norm": 18.5, "learning_rate": 1.4361604097132402e-06, "loss": 0.5507, "step": 10673 }, { "epoch": 0.9095867064337452, "grad_norm": 45.0, "learning_rate": 1.4359942464147533e-06, "loss": 0.9017, "step": 10674 }, { "epoch": 0.9096719216020451, "grad_norm": 14.0, "learning_rate": 1.4358280797551633e-06, "loss": 0.7672, "step": 10675 }, { "epoch": 0.9097571367703451, "grad_norm": 18.0, "learning_rate": 1.435661909737473e-06, "loss": 0.7525, "step": 10676 }, { "epoch": 0.909842351938645, "grad_norm": 13.9375, "learning_rate": 1.4354957363646848e-06, "loss": 0.757, "step": 10677 }, { "epoch": 0.909927567106945, "grad_norm": 17.0, "learning_rate": 1.4353295596398025e-06, "loss": 0.9639, "step": 10678 }, { "epoch": 0.910012782275245, "grad_norm": 12.5, "learning_rate": 1.4351633795658287e-06, "loss": 0.2647, "step": 10679 }, { "epoch": 0.9100979974435449, "grad_norm": 31.875, "learning_rate": 1.4349971961457666e-06, "loss": 0.8639, "step": 10680 }, { "epoch": 0.9101832126118449, "grad_norm": 12.875, "learning_rate": 1.4348310093826187e-06, "loss": 0.3345, "step": 10681 }, { "epoch": 0.9102684277801448, "grad_norm": 18.75, "learning_rate": 1.434664819279389e-06, "loss": 1.0251, "step": 10682 }, { "epoch": 0.9103536429484448, "grad_norm": 20.5, "learning_rate": 1.4344986258390803e-06, "loss": 0.8824, "step": 10683 }, { "epoch": 0.9104388581167447, "grad_norm": 17.625, "learning_rate": 1.4343324290646957e-06, "loss": 1.1389, "step": 10684 }, { "epoch": 0.9105240732850447, "grad_norm": 14.0, "learning_rate": 1.4341662289592386e-06, "loss": 0.8872, "step": 10685 }, { "epoch": 0.9106092884533447, "grad_norm": 12.0, "learning_rate": 1.4340000255257128e-06, "loss": 0.3582, "step": 10686 }, { "epoch": 0.9106945036216446, "grad_norm": 20.25, "learning_rate": 1.4338338187671215e-06, "loss": 0.8789, "step": 10687 }, { "epoch": 0.9107797187899446, "grad_norm": 29.5, "learning_rate": 1.433667608686468e-06, "loss": 1.428, "step": 10688 }, { "epoch": 0.9108649339582445, "grad_norm": 24.75, "learning_rate": 1.4335013952867566e-06, "loss": 0.9267, "step": 10689 }, { "epoch": 0.9109501491265445, "grad_norm": 11.3125, "learning_rate": 1.4333351785709898e-06, "loss": 0.4721, "step": 10690 }, { "epoch": 0.9110353642948444, "grad_norm": 9.4375, "learning_rate": 1.4331689585421726e-06, "loss": 0.3674, "step": 10691 }, { "epoch": 0.9111205794631444, "grad_norm": 10.1875, "learning_rate": 1.433002735203308e-06, "loss": 0.2565, "step": 10692 }, { "epoch": 0.9112057946314444, "grad_norm": 10.0625, "learning_rate": 1.4328365085573997e-06, "loss": 0.5713, "step": 10693 }, { "epoch": 0.9112910097997443, "grad_norm": 16.625, "learning_rate": 1.4326702786074518e-06, "loss": 0.9657, "step": 10694 }, { "epoch": 0.9113762249680443, "grad_norm": 25.5, "learning_rate": 1.4325040453564687e-06, "loss": 1.1145, "step": 10695 }, { "epoch": 0.9114614401363442, "grad_norm": 13.3125, "learning_rate": 1.4323378088074536e-06, "loss": 0.7592, "step": 10696 }, { "epoch": 0.9115466553046442, "grad_norm": 12.9375, "learning_rate": 1.432171568963411e-06, "loss": 0.5014, "step": 10697 }, { "epoch": 0.9116318704729441, "grad_norm": 20.875, "learning_rate": 1.432005325827345e-06, "loss": 1.0508, "step": 10698 }, { "epoch": 0.9117170856412441, "grad_norm": 11.5, "learning_rate": 1.4318390794022602e-06, "loss": 0.4104, "step": 10699 }, { "epoch": 0.911802300809544, "grad_norm": 12.4375, "learning_rate": 1.4316728296911598e-06, "loss": 0.5074, "step": 10700 }, { "epoch": 0.911887515977844, "grad_norm": 19.25, "learning_rate": 1.4315065766970487e-06, "loss": 1.1956, "step": 10701 }, { "epoch": 0.911972731146144, "grad_norm": 12.875, "learning_rate": 1.4313403204229313e-06, "loss": 0.6013, "step": 10702 }, { "epoch": 0.9120579463144439, "grad_norm": 24.5, "learning_rate": 1.431174060871812e-06, "loss": 0.9155, "step": 10703 }, { "epoch": 0.9121431614827439, "grad_norm": 17.0, "learning_rate": 1.4310077980466953e-06, "loss": 1.2339, "step": 10704 }, { "epoch": 0.9122283766510438, "grad_norm": 16.375, "learning_rate": 1.4308415319505856e-06, "loss": 0.8029, "step": 10705 }, { "epoch": 0.9123135918193438, "grad_norm": 11.625, "learning_rate": 1.4306752625864875e-06, "loss": 0.4112, "step": 10706 }, { "epoch": 0.9123988069876438, "grad_norm": 13.25, "learning_rate": 1.430508989957406e-06, "loss": 0.7702, "step": 10707 }, { "epoch": 0.9124840221559437, "grad_norm": 17.125, "learning_rate": 1.4303427140663454e-06, "loss": 1.094, "step": 10708 }, { "epoch": 0.9125692373242437, "grad_norm": 14.8125, "learning_rate": 1.4301764349163102e-06, "loss": 0.706, "step": 10709 }, { "epoch": 0.9126544524925436, "grad_norm": 15.75, "learning_rate": 1.4300101525103063e-06, "loss": 0.6238, "step": 10710 }, { "epoch": 0.9127396676608436, "grad_norm": 16.5, "learning_rate": 1.4298438668513373e-06, "loss": 0.6509, "step": 10711 }, { "epoch": 0.9128248828291435, "grad_norm": 18.5, "learning_rate": 1.4296775779424096e-06, "loss": 0.8594, "step": 10712 }, { "epoch": 0.9129100979974435, "grad_norm": 19.75, "learning_rate": 1.4295112857865268e-06, "loss": 1.0079, "step": 10713 }, { "epoch": 0.9129953131657434, "grad_norm": 15.8125, "learning_rate": 1.4293449903866946e-06, "loss": 0.8087, "step": 10714 }, { "epoch": 0.9130805283340434, "grad_norm": 23.25, "learning_rate": 1.429178691745918e-06, "loss": 1.0736, "step": 10715 }, { "epoch": 0.9131657435023434, "grad_norm": 15.9375, "learning_rate": 1.4290123898672026e-06, "loss": 0.7095, "step": 10716 }, { "epoch": 0.9132509586706433, "grad_norm": 14.9375, "learning_rate": 1.428846084753553e-06, "loss": 0.7069, "step": 10717 }, { "epoch": 0.9133361738389433, "grad_norm": 17.0, "learning_rate": 1.428679776407975e-06, "loss": 0.8269, "step": 10718 }, { "epoch": 0.9134213890072432, "grad_norm": 15.8125, "learning_rate": 1.4285134648334733e-06, "loss": 0.7666, "step": 10719 }, { "epoch": 0.9135066041755433, "grad_norm": 15.3125, "learning_rate": 1.4283471500330543e-06, "loss": 0.6026, "step": 10720 }, { "epoch": 0.9135918193438433, "grad_norm": 15.4375, "learning_rate": 1.428180832009723e-06, "loss": 0.971, "step": 10721 }, { "epoch": 0.9136770345121432, "grad_norm": 16.125, "learning_rate": 1.4280145107664848e-06, "loss": 0.7324, "step": 10722 }, { "epoch": 0.9137622496804432, "grad_norm": 14.3125, "learning_rate": 1.4278481863063447e-06, "loss": 0.8911, "step": 10723 }, { "epoch": 0.9138474648487431, "grad_norm": 14.875, "learning_rate": 1.4276818586323098e-06, "loss": 0.8858, "step": 10724 }, { "epoch": 0.9139326800170431, "grad_norm": 11.875, "learning_rate": 1.427515527747385e-06, "loss": 0.4187, "step": 10725 }, { "epoch": 0.914017895185343, "grad_norm": 77.5, "learning_rate": 1.4273491936545757e-06, "loss": 0.636, "step": 10726 }, { "epoch": 0.914103110353643, "grad_norm": 11.625, "learning_rate": 1.4271828563568884e-06, "loss": 0.3706, "step": 10727 }, { "epoch": 0.914188325521943, "grad_norm": 15.375, "learning_rate": 1.427016515857329e-06, "loss": 0.9849, "step": 10728 }, { "epoch": 0.9142735406902429, "grad_norm": 11.125, "learning_rate": 1.4268501721589029e-06, "loss": 0.5107, "step": 10729 }, { "epoch": 0.9143587558585429, "grad_norm": 13.4375, "learning_rate": 1.4266838252646162e-06, "loss": 0.566, "step": 10730 }, { "epoch": 0.9144439710268428, "grad_norm": 14.0, "learning_rate": 1.4265174751774752e-06, "loss": 1.0135, "step": 10731 }, { "epoch": 0.9145291861951428, "grad_norm": 13.125, "learning_rate": 1.426351121900486e-06, "loss": 0.5943, "step": 10732 }, { "epoch": 0.9146144013634427, "grad_norm": 20.0, "learning_rate": 1.4261847654366547e-06, "loss": 1.2011, "step": 10733 }, { "epoch": 0.9146996165317427, "grad_norm": 13.0, "learning_rate": 1.4260184057889872e-06, "loss": 0.452, "step": 10734 }, { "epoch": 0.9147848317000427, "grad_norm": 11.375, "learning_rate": 1.4258520429604905e-06, "loss": 0.9703, "step": 10735 }, { "epoch": 0.9148700468683426, "grad_norm": 13.4375, "learning_rate": 1.4256856769541704e-06, "loss": 0.4704, "step": 10736 }, { "epoch": 0.9149552620366426, "grad_norm": 13.4375, "learning_rate": 1.4255193077730336e-06, "loss": 0.6168, "step": 10737 }, { "epoch": 0.9150404772049425, "grad_norm": 14.375, "learning_rate": 1.4253529354200862e-06, "loss": 0.4867, "step": 10738 }, { "epoch": 0.9151256923732425, "grad_norm": 21.5, "learning_rate": 1.425186559898335e-06, "loss": 0.7618, "step": 10739 }, { "epoch": 0.9152109075415424, "grad_norm": 25.75, "learning_rate": 1.4250201812107866e-06, "loss": 0.843, "step": 10740 }, { "epoch": 0.9152961227098424, "grad_norm": 16.0, "learning_rate": 1.4248537993604472e-06, "loss": 0.5797, "step": 10741 }, { "epoch": 0.9153813378781424, "grad_norm": 31.5, "learning_rate": 1.4246874143503241e-06, "loss": 0.8654, "step": 10742 }, { "epoch": 0.9154665530464423, "grad_norm": 39.75, "learning_rate": 1.4245210261834236e-06, "loss": 1.0468, "step": 10743 }, { "epoch": 0.9155517682147423, "grad_norm": 13.9375, "learning_rate": 1.4243546348627529e-06, "loss": 0.8127, "step": 10744 }, { "epoch": 0.9156369833830422, "grad_norm": 13.875, "learning_rate": 1.4241882403913184e-06, "loss": 0.5697, "step": 10745 }, { "epoch": 0.9157221985513422, "grad_norm": 12.1875, "learning_rate": 1.4240218427721276e-06, "loss": 0.4631, "step": 10746 }, { "epoch": 0.9158074137196421, "grad_norm": 10.375, "learning_rate": 1.4238554420081865e-06, "loss": 0.6388, "step": 10747 }, { "epoch": 0.9158926288879421, "grad_norm": 15.875, "learning_rate": 1.4236890381025033e-06, "loss": 0.7782, "step": 10748 }, { "epoch": 0.915977844056242, "grad_norm": 11.3125, "learning_rate": 1.4235226310580843e-06, "loss": 0.2891, "step": 10749 }, { "epoch": 0.916063059224542, "grad_norm": 13.875, "learning_rate": 1.423356220877937e-06, "loss": 0.6451, "step": 10750 }, { "epoch": 0.916148274392842, "grad_norm": 12.875, "learning_rate": 1.423189807565068e-06, "loss": 0.6233, "step": 10751 }, { "epoch": 0.9162334895611419, "grad_norm": 11.625, "learning_rate": 1.4230233911224858e-06, "loss": 0.5776, "step": 10752 }, { "epoch": 0.9163187047294419, "grad_norm": 17.375, "learning_rate": 1.4228569715531962e-06, "loss": 0.8446, "step": 10753 }, { "epoch": 0.9164039198977418, "grad_norm": 19.125, "learning_rate": 1.4226905488602077e-06, "loss": 0.8129, "step": 10754 }, { "epoch": 0.9164891350660418, "grad_norm": 12.25, "learning_rate": 1.422524123046527e-06, "loss": 0.4956, "step": 10755 }, { "epoch": 0.9165743502343418, "grad_norm": 14.25, "learning_rate": 1.4223576941151624e-06, "loss": 0.7465, "step": 10756 }, { "epoch": 0.9166595654026417, "grad_norm": 13.25, "learning_rate": 1.4221912620691206e-06, "loss": 0.6157, "step": 10757 }, { "epoch": 0.9167447805709417, "grad_norm": 16.75, "learning_rate": 1.4220248269114093e-06, "loss": 0.7655, "step": 10758 }, { "epoch": 0.9168299957392416, "grad_norm": 18.375, "learning_rate": 1.421858388645037e-06, "loss": 0.9627, "step": 10759 }, { "epoch": 0.9169152109075416, "grad_norm": 14.5, "learning_rate": 1.4216919472730104e-06, "loss": 0.6259, "step": 10760 }, { "epoch": 0.9170004260758415, "grad_norm": 17.125, "learning_rate": 1.421525502798338e-06, "loss": 0.8324, "step": 10761 }, { "epoch": 0.9170856412441415, "grad_norm": 13.9375, "learning_rate": 1.4213590552240267e-06, "loss": 0.8552, "step": 10762 }, { "epoch": 0.9171708564124414, "grad_norm": 12.0, "learning_rate": 1.4211926045530855e-06, "loss": 0.4953, "step": 10763 }, { "epoch": 0.9172560715807414, "grad_norm": 14.5, "learning_rate": 1.4210261507885214e-06, "loss": 0.6579, "step": 10764 }, { "epoch": 0.9173412867490414, "grad_norm": 16.0, "learning_rate": 1.4208596939333433e-06, "loss": 0.6735, "step": 10765 }, { "epoch": 0.9174265019173413, "grad_norm": 14.5625, "learning_rate": 1.4206932339905582e-06, "loss": 0.6644, "step": 10766 }, { "epoch": 0.9175117170856413, "grad_norm": 16.875, "learning_rate": 1.4205267709631748e-06, "loss": 0.7312, "step": 10767 }, { "epoch": 0.9175969322539412, "grad_norm": 15.3125, "learning_rate": 1.4203603048542014e-06, "loss": 0.6361, "step": 10768 }, { "epoch": 0.9176821474222412, "grad_norm": 14.375, "learning_rate": 1.4201938356666462e-06, "loss": 0.6674, "step": 10769 }, { "epoch": 0.9177673625905411, "grad_norm": 15.5, "learning_rate": 1.4200273634035166e-06, "loss": 0.9394, "step": 10770 }, { "epoch": 0.9178525777588411, "grad_norm": 10.625, "learning_rate": 1.4198608880678219e-06, "loss": 0.934, "step": 10771 }, { "epoch": 0.9179377929271411, "grad_norm": 11.75, "learning_rate": 1.4196944096625698e-06, "loss": 0.4415, "step": 10772 }, { "epoch": 0.918023008095441, "grad_norm": 15.3125, "learning_rate": 1.4195279281907696e-06, "loss": 0.7838, "step": 10773 }, { "epoch": 0.918108223263741, "grad_norm": 27.625, "learning_rate": 1.4193614436554293e-06, "loss": 0.7849, "step": 10774 }, { "epoch": 0.9181934384320409, "grad_norm": 10.625, "learning_rate": 1.419194956059557e-06, "loss": 0.4729, "step": 10775 }, { "epoch": 0.9182786536003409, "grad_norm": 13.625, "learning_rate": 1.4190284654061617e-06, "loss": 0.7227, "step": 10776 }, { "epoch": 0.9183638687686408, "grad_norm": 16.625, "learning_rate": 1.4188619716982523e-06, "loss": 0.8909, "step": 10777 }, { "epoch": 0.9184490839369408, "grad_norm": 14.375, "learning_rate": 1.4186954749388373e-06, "loss": 0.6256, "step": 10778 }, { "epoch": 0.9185342991052408, "grad_norm": 14.25, "learning_rate": 1.4185289751309253e-06, "loss": 0.7952, "step": 10779 }, { "epoch": 0.9186195142735407, "grad_norm": 11.9375, "learning_rate": 1.4183624722775251e-06, "loss": 0.2722, "step": 10780 }, { "epoch": 0.9187047294418407, "grad_norm": 14.0, "learning_rate": 1.4181959663816462e-06, "loss": 0.4637, "step": 10781 }, { "epoch": 0.9187899446101406, "grad_norm": 14.75, "learning_rate": 1.4180294574462966e-06, "loss": 0.9549, "step": 10782 }, { "epoch": 0.9188751597784406, "grad_norm": 18.125, "learning_rate": 1.4178629454744858e-06, "loss": 1.1805, "step": 10783 }, { "epoch": 0.9189603749467405, "grad_norm": 22.375, "learning_rate": 1.4176964304692232e-06, "loss": 0.6905, "step": 10784 }, { "epoch": 0.9190455901150405, "grad_norm": 14.3125, "learning_rate": 1.417529912433517e-06, "loss": 0.6286, "step": 10785 }, { "epoch": 0.9191308052833405, "grad_norm": 16.0, "learning_rate": 1.417363391370377e-06, "loss": 1.0318, "step": 10786 }, { "epoch": 0.9192160204516404, "grad_norm": 12.6875, "learning_rate": 1.4171968672828122e-06, "loss": 0.7355, "step": 10787 }, { "epoch": 0.9193012356199404, "grad_norm": 20.375, "learning_rate": 1.4170303401738316e-06, "loss": 0.7971, "step": 10788 }, { "epoch": 0.9193864507882403, "grad_norm": 9.1875, "learning_rate": 1.4168638100464455e-06, "loss": 0.2897, "step": 10789 }, { "epoch": 0.9194716659565403, "grad_norm": 25.0, "learning_rate": 1.416697276903662e-06, "loss": 1.0276, "step": 10790 }, { "epoch": 0.9195568811248402, "grad_norm": 13.125, "learning_rate": 1.416530740748491e-06, "loss": 0.7483, "step": 10791 }, { "epoch": 0.9196420962931402, "grad_norm": 13.3125, "learning_rate": 1.4163642015839423e-06, "loss": 0.4395, "step": 10792 }, { "epoch": 0.9197273114614402, "grad_norm": 19.5, "learning_rate": 1.4161976594130253e-06, "loss": 0.7301, "step": 10793 }, { "epoch": 0.9198125266297401, "grad_norm": 20.5, "learning_rate": 1.416031114238749e-06, "loss": 1.1139, "step": 10794 }, { "epoch": 0.9198977417980401, "grad_norm": 12.0625, "learning_rate": 1.415864566064124e-06, "loss": 0.4253, "step": 10795 }, { "epoch": 0.91998295696634, "grad_norm": 12.0, "learning_rate": 1.4156980148921589e-06, "loss": 0.511, "step": 10796 }, { "epoch": 0.92006817213464, "grad_norm": 13.625, "learning_rate": 1.415531460725865e-06, "loss": 0.7488, "step": 10797 }, { "epoch": 0.9201533873029399, "grad_norm": 17.0, "learning_rate": 1.4153649035682504e-06, "loss": 0.8851, "step": 10798 }, { "epoch": 0.9202386024712399, "grad_norm": 18.125, "learning_rate": 1.415198343422326e-06, "loss": 0.9737, "step": 10799 }, { "epoch": 0.9203238176395399, "grad_norm": 16.375, "learning_rate": 1.415031780291101e-06, "loss": 0.764, "step": 10800 }, { "epoch": 0.9204090328078398, "grad_norm": 10.125, "learning_rate": 1.4148652141775864e-06, "loss": 0.4451, "step": 10801 }, { "epoch": 0.9204942479761398, "grad_norm": 19.375, "learning_rate": 1.4146986450847912e-06, "loss": 0.9216, "step": 10802 }, { "epoch": 0.9205794631444397, "grad_norm": 12.875, "learning_rate": 1.4145320730157262e-06, "loss": 0.6393, "step": 10803 }, { "epoch": 0.9206646783127397, "grad_norm": 13.5, "learning_rate": 1.4143654979734006e-06, "loss": 0.8285, "step": 10804 }, { "epoch": 0.9207498934810396, "grad_norm": 15.875, "learning_rate": 1.4141989199608256e-06, "loss": 0.7608, "step": 10805 }, { "epoch": 0.9208351086493396, "grad_norm": 20.75, "learning_rate": 1.414032338981011e-06, "loss": 0.9688, "step": 10806 }, { "epoch": 0.9209203238176396, "grad_norm": 13.75, "learning_rate": 1.4138657550369672e-06, "loss": 0.357, "step": 10807 }, { "epoch": 0.9210055389859395, "grad_norm": 25.375, "learning_rate": 1.413699168131704e-06, "loss": 1.0686, "step": 10808 }, { "epoch": 0.9210907541542395, "grad_norm": 20.75, "learning_rate": 1.4135325782682324e-06, "loss": 0.8688, "step": 10809 }, { "epoch": 0.9211759693225394, "grad_norm": 14.125, "learning_rate": 1.413365985449563e-06, "loss": 0.6263, "step": 10810 }, { "epoch": 0.9212611844908394, "grad_norm": 13.6875, "learning_rate": 1.4131993896787056e-06, "loss": 0.5667, "step": 10811 }, { "epoch": 0.9213463996591393, "grad_norm": 11.375, "learning_rate": 1.413032790958671e-06, "loss": 0.8596, "step": 10812 }, { "epoch": 0.9214316148274393, "grad_norm": 17.5, "learning_rate": 1.4128661892924702e-06, "loss": 1.1177, "step": 10813 }, { "epoch": 0.9215168299957393, "grad_norm": 12.5, "learning_rate": 1.4126995846831137e-06, "loss": 0.4265, "step": 10814 }, { "epoch": 0.9216020451640392, "grad_norm": 14.3125, "learning_rate": 1.4125329771336115e-06, "loss": 0.8378, "step": 10815 }, { "epoch": 0.9216872603323392, "grad_norm": 20.375, "learning_rate": 1.4123663666469756e-06, "loss": 0.9733, "step": 10816 }, { "epoch": 0.9217724755006391, "grad_norm": 12.0, "learning_rate": 1.412199753226216e-06, "loss": 0.6782, "step": 10817 }, { "epoch": 0.9218576906689391, "grad_norm": 13.5, "learning_rate": 1.412033136874344e-06, "loss": 0.9881, "step": 10818 }, { "epoch": 0.921942905837239, "grad_norm": 15.1875, "learning_rate": 1.4118665175943696e-06, "loss": 0.7595, "step": 10819 }, { "epoch": 0.922028121005539, "grad_norm": 11.75, "learning_rate": 1.4116998953893051e-06, "loss": 0.705, "step": 10820 }, { "epoch": 0.922113336173839, "grad_norm": 16.375, "learning_rate": 1.4115332702621607e-06, "loss": 0.9031, "step": 10821 }, { "epoch": 0.9221985513421389, "grad_norm": 14.5, "learning_rate": 1.4113666422159477e-06, "loss": 0.5239, "step": 10822 }, { "epoch": 0.9222837665104389, "grad_norm": 21.0, "learning_rate": 1.411200011253677e-06, "loss": 0.977, "step": 10823 }, { "epoch": 0.9223689816787388, "grad_norm": 14.9375, "learning_rate": 1.4110333773783603e-06, "loss": 0.9753, "step": 10824 }, { "epoch": 0.9224541968470388, "grad_norm": 10.3125, "learning_rate": 1.4108667405930082e-06, "loss": 0.3109, "step": 10825 }, { "epoch": 0.9225394120153387, "grad_norm": 13.125, "learning_rate": 1.4107001009006324e-06, "loss": 0.544, "step": 10826 }, { "epoch": 0.9226246271836387, "grad_norm": 16.5, "learning_rate": 1.4105334583042444e-06, "loss": 0.8261, "step": 10827 }, { "epoch": 0.9227098423519386, "grad_norm": 11.1875, "learning_rate": 1.4103668128068548e-06, "loss": 0.352, "step": 10828 }, { "epoch": 0.9227950575202386, "grad_norm": 13.0, "learning_rate": 1.4102001644114766e-06, "loss": 0.3697, "step": 10829 }, { "epoch": 0.9228802726885386, "grad_norm": 18.375, "learning_rate": 1.4100335131211195e-06, "loss": 0.9283, "step": 10830 }, { "epoch": 0.9229654878568385, "grad_norm": 17.5, "learning_rate": 1.4098668589387966e-06, "loss": 0.8075, "step": 10831 }, { "epoch": 0.9230507030251385, "grad_norm": 14.4375, "learning_rate": 1.409700201867518e-06, "loss": 0.5521, "step": 10832 }, { "epoch": 0.9231359181934384, "grad_norm": 13.625, "learning_rate": 1.409533541910297e-06, "loss": 0.2999, "step": 10833 }, { "epoch": 0.9232211333617384, "grad_norm": 11.25, "learning_rate": 1.4093668790701437e-06, "loss": 0.5054, "step": 10834 }, { "epoch": 0.9233063485300383, "grad_norm": 16.0, "learning_rate": 1.4092002133500713e-06, "loss": 0.7719, "step": 10835 }, { "epoch": 0.9233915636983383, "grad_norm": 12.5, "learning_rate": 1.4090335447530903e-06, "loss": 0.4164, "step": 10836 }, { "epoch": 0.9234767788666383, "grad_norm": 20.25, "learning_rate": 1.408866873282214e-06, "loss": 0.7232, "step": 10837 }, { "epoch": 0.9235619940349382, "grad_norm": 14.0625, "learning_rate": 1.4087001989404529e-06, "loss": 0.6379, "step": 10838 }, { "epoch": 0.9236472092032382, "grad_norm": 13.625, "learning_rate": 1.40853352173082e-06, "loss": 0.664, "step": 10839 }, { "epoch": 0.9237324243715381, "grad_norm": 21.5, "learning_rate": 1.4083668416563265e-06, "loss": 1.0119, "step": 10840 }, { "epoch": 0.9238176395398381, "grad_norm": 28.75, "learning_rate": 1.4082001587199853e-06, "loss": 0.9788, "step": 10841 }, { "epoch": 0.923902854708138, "grad_norm": 16.125, "learning_rate": 1.408033472924808e-06, "loss": 0.7658, "step": 10842 }, { "epoch": 0.923988069876438, "grad_norm": 14.6875, "learning_rate": 1.407866784273807e-06, "loss": 0.6037, "step": 10843 }, { "epoch": 0.924073285044738, "grad_norm": 13.3125, "learning_rate": 1.4077000927699945e-06, "loss": 0.4422, "step": 10844 }, { "epoch": 0.9241585002130379, "grad_norm": 13.625, "learning_rate": 1.4075333984163824e-06, "loss": 0.6748, "step": 10845 }, { "epoch": 0.9242437153813379, "grad_norm": 15.5, "learning_rate": 1.407366701215984e-06, "loss": 0.3819, "step": 10846 }, { "epoch": 0.9243289305496378, "grad_norm": 14.625, "learning_rate": 1.4072000011718106e-06, "loss": 0.6826, "step": 10847 }, { "epoch": 0.9244141457179378, "grad_norm": 14.625, "learning_rate": 1.4070332982868754e-06, "loss": 0.7509, "step": 10848 }, { "epoch": 0.9244993608862377, "grad_norm": 19.5, "learning_rate": 1.40686659256419e-06, "loss": 0.2675, "step": 10849 }, { "epoch": 0.9245845760545377, "grad_norm": 10.25, "learning_rate": 1.4066998840067682e-06, "loss": 0.3213, "step": 10850 }, { "epoch": 0.9246697912228377, "grad_norm": 15.8125, "learning_rate": 1.4065331726176218e-06, "loss": 0.6964, "step": 10851 }, { "epoch": 0.9247550063911376, "grad_norm": 12.25, "learning_rate": 1.4063664583997632e-06, "loss": 0.4816, "step": 10852 }, { "epoch": 0.9248402215594376, "grad_norm": 18.75, "learning_rate": 1.4061997413562057e-06, "loss": 0.7428, "step": 10853 }, { "epoch": 0.9249254367277375, "grad_norm": 15.9375, "learning_rate": 1.406033021489962e-06, "loss": 0.9001, "step": 10854 }, { "epoch": 0.9250106518960375, "grad_norm": 18.25, "learning_rate": 1.4058662988040444e-06, "loss": 0.7291, "step": 10855 }, { "epoch": 0.9250958670643374, "grad_norm": 18.875, "learning_rate": 1.4056995733014664e-06, "loss": 0.8586, "step": 10856 }, { "epoch": 0.9251810822326374, "grad_norm": 14.8125, "learning_rate": 1.4055328449852398e-06, "loss": 0.7642, "step": 10857 }, { "epoch": 0.9252662974009374, "grad_norm": 13.0, "learning_rate": 1.405366113858379e-06, "loss": 0.592, "step": 10858 }, { "epoch": 0.9253515125692373, "grad_norm": 15.4375, "learning_rate": 1.4051993799238962e-06, "loss": 0.6228, "step": 10859 }, { "epoch": 0.9254367277375373, "grad_norm": 12.0625, "learning_rate": 1.405032643184804e-06, "loss": 0.4445, "step": 10860 }, { "epoch": 0.9255219429058372, "grad_norm": 20.25, "learning_rate": 1.4048659036441166e-06, "loss": 0.9376, "step": 10861 }, { "epoch": 0.9256071580741372, "grad_norm": 17.5, "learning_rate": 1.4046991613048466e-06, "loss": 0.8778, "step": 10862 }, { "epoch": 0.9256923732424371, "grad_norm": 11.1875, "learning_rate": 1.4045324161700073e-06, "loss": 0.3592, "step": 10863 }, { "epoch": 0.9257775884107371, "grad_norm": 13.875, "learning_rate": 1.4043656682426112e-06, "loss": 0.4612, "step": 10864 }, { "epoch": 0.925862803579037, "grad_norm": 16.625, "learning_rate": 1.4041989175256727e-06, "loss": 0.5827, "step": 10865 }, { "epoch": 0.925948018747337, "grad_norm": 10.8125, "learning_rate": 1.4040321640222046e-06, "loss": 0.2933, "step": 10866 }, { "epoch": 0.926033233915637, "grad_norm": 16.875, "learning_rate": 1.4038654077352206e-06, "loss": 0.8013, "step": 10867 }, { "epoch": 0.9261184490839369, "grad_norm": 17.625, "learning_rate": 1.4036986486677336e-06, "loss": 0.8841, "step": 10868 }, { "epoch": 0.9262036642522369, "grad_norm": 16.25, "learning_rate": 1.4035318868227577e-06, "loss": 0.6938, "step": 10869 }, { "epoch": 0.9262888794205368, "grad_norm": 14.4375, "learning_rate": 1.4033651222033062e-06, "loss": 1.1186, "step": 10870 }, { "epoch": 0.9263740945888368, "grad_norm": 12.25, "learning_rate": 1.4031983548123928e-06, "loss": 0.717, "step": 10871 }, { "epoch": 0.9264593097571368, "grad_norm": 26.5, "learning_rate": 1.403031584653031e-06, "loss": 0.4043, "step": 10872 }, { "epoch": 0.9265445249254367, "grad_norm": 11.0625, "learning_rate": 1.4028648117282345e-06, "loss": 0.3551, "step": 10873 }, { "epoch": 0.9266297400937367, "grad_norm": 21.0, "learning_rate": 1.4026980360410172e-06, "loss": 0.9203, "step": 10874 }, { "epoch": 0.9267149552620366, "grad_norm": 14.0, "learning_rate": 1.4025312575943928e-06, "loss": 0.8876, "step": 10875 }, { "epoch": 0.9268001704303366, "grad_norm": 19.0, "learning_rate": 1.4023644763913752e-06, "loss": 0.9873, "step": 10876 }, { "epoch": 0.9268853855986365, "grad_norm": 11.4375, "learning_rate": 1.4021976924349783e-06, "loss": 0.553, "step": 10877 }, { "epoch": 0.9269706007669365, "grad_norm": 15.1875, "learning_rate": 1.402030905728216e-06, "loss": 0.5433, "step": 10878 }, { "epoch": 0.9270558159352364, "grad_norm": 14.375, "learning_rate": 1.4018641162741023e-06, "loss": 0.9289, "step": 10879 }, { "epoch": 0.9271410311035364, "grad_norm": 13.3125, "learning_rate": 1.4016973240756516e-06, "loss": 0.6297, "step": 10880 }, { "epoch": 0.9272262462718364, "grad_norm": 13.25, "learning_rate": 1.4015305291358774e-06, "loss": 0.6875, "step": 10881 }, { "epoch": 0.9273114614401363, "grad_norm": 17.0, "learning_rate": 1.4013637314577947e-06, "loss": 0.9845, "step": 10882 }, { "epoch": 0.9273966766084363, "grad_norm": 28.375, "learning_rate": 1.4011969310444165e-06, "loss": 1.8797, "step": 10883 }, { "epoch": 0.9274818917767362, "grad_norm": 13.75, "learning_rate": 1.4010301278987582e-06, "loss": 0.6481, "step": 10884 }, { "epoch": 0.9275671069450362, "grad_norm": 22.875, "learning_rate": 1.400863322023833e-06, "loss": 1.3978, "step": 10885 }, { "epoch": 0.9276523221133361, "grad_norm": 17.125, "learning_rate": 1.4006965134226567e-06, "loss": 1.0267, "step": 10886 }, { "epoch": 0.9277375372816361, "grad_norm": 11.125, "learning_rate": 1.4005297020982422e-06, "loss": 0.6056, "step": 10887 }, { "epoch": 0.9278227524499361, "grad_norm": 13.5, "learning_rate": 1.400362888053605e-06, "loss": 0.8652, "step": 10888 }, { "epoch": 0.927907967618236, "grad_norm": 19.625, "learning_rate": 1.4001960712917586e-06, "loss": 0.9383, "step": 10889 }, { "epoch": 0.927993182786536, "grad_norm": 13.4375, "learning_rate": 1.4000292518157188e-06, "loss": 0.5327, "step": 10890 }, { "epoch": 0.9280783979548359, "grad_norm": 17.25, "learning_rate": 1.3998624296284996e-06, "loss": 0.7557, "step": 10891 }, { "epoch": 0.9281636131231359, "grad_norm": 15.25, "learning_rate": 1.3996956047331153e-06, "loss": 0.7795, "step": 10892 }, { "epoch": 0.9282488282914358, "grad_norm": 12.875, "learning_rate": 1.3995287771325808e-06, "loss": 0.5039, "step": 10893 }, { "epoch": 0.9283340434597358, "grad_norm": 16.5, "learning_rate": 1.3993619468299112e-06, "loss": 0.6502, "step": 10894 }, { "epoch": 0.9284192586280358, "grad_norm": 14.4375, "learning_rate": 1.3991951138281209e-06, "loss": 0.8852, "step": 10895 }, { "epoch": 0.9285044737963357, "grad_norm": 14.0625, "learning_rate": 1.3990282781302247e-06, "loss": 0.5781, "step": 10896 }, { "epoch": 0.9285896889646357, "grad_norm": 14.625, "learning_rate": 1.398861439739238e-06, "loss": 0.4403, "step": 10897 }, { "epoch": 0.9286749041329356, "grad_norm": 13.1875, "learning_rate": 1.3986945986581752e-06, "loss": 0.6279, "step": 10898 }, { "epoch": 0.9287601193012356, "grad_norm": 12.4375, "learning_rate": 1.3985277548900518e-06, "loss": 0.5392, "step": 10899 }, { "epoch": 0.9288453344695355, "grad_norm": 11.5, "learning_rate": 1.3983609084378818e-06, "loss": 0.4736, "step": 10900 }, { "epoch": 0.9289305496378355, "grad_norm": 16.625, "learning_rate": 1.3981940593046817e-06, "loss": 0.7413, "step": 10901 }, { "epoch": 0.9290157648061355, "grad_norm": 13.6875, "learning_rate": 1.3980272074934656e-06, "loss": 0.7464, "step": 10902 }, { "epoch": 0.9291009799744354, "grad_norm": 13.625, "learning_rate": 1.397860353007249e-06, "loss": 0.7442, "step": 10903 }, { "epoch": 0.9291861951427354, "grad_norm": 18.75, "learning_rate": 1.397693495849047e-06, "loss": 0.9192, "step": 10904 }, { "epoch": 0.9292714103110353, "grad_norm": 15.1875, "learning_rate": 1.3975266360218753e-06, "loss": 0.8816, "step": 10905 }, { "epoch": 0.9293566254793353, "grad_norm": 10.875, "learning_rate": 1.397359773528749e-06, "loss": 0.5126, "step": 10906 }, { "epoch": 0.9294418406476352, "grad_norm": 14.5, "learning_rate": 1.3971929083726831e-06, "loss": 0.5135, "step": 10907 }, { "epoch": 0.9295270558159352, "grad_norm": 12.0, "learning_rate": 1.3970260405566935e-06, "loss": 0.6412, "step": 10908 }, { "epoch": 0.9296122709842352, "grad_norm": 13.5, "learning_rate": 1.3968591700837955e-06, "loss": 0.432, "step": 10909 }, { "epoch": 0.9296974861525351, "grad_norm": 21.75, "learning_rate": 1.3966922969570044e-06, "loss": 0.8647, "step": 10910 }, { "epoch": 0.9297827013208351, "grad_norm": 12.5, "learning_rate": 1.3965254211793364e-06, "loss": 0.5473, "step": 10911 }, { "epoch": 0.929867916489135, "grad_norm": 13.875, "learning_rate": 1.3963585427538064e-06, "loss": 0.5646, "step": 10912 }, { "epoch": 0.929953131657435, "grad_norm": 20.125, "learning_rate": 1.3961916616834304e-06, "loss": 1.0701, "step": 10913 }, { "epoch": 0.9300383468257349, "grad_norm": 13.5625, "learning_rate": 1.3960247779712241e-06, "loss": 0.6506, "step": 10914 }, { "epoch": 0.9301235619940349, "grad_norm": 19.875, "learning_rate": 1.3958578916202031e-06, "loss": 0.794, "step": 10915 }, { "epoch": 0.9302087771623349, "grad_norm": 28.75, "learning_rate": 1.3956910026333837e-06, "loss": 1.3003, "step": 10916 }, { "epoch": 0.9302939923306348, "grad_norm": 16.625, "learning_rate": 1.395524111013781e-06, "loss": 0.8453, "step": 10917 }, { "epoch": 0.9303792074989348, "grad_norm": 15.875, "learning_rate": 1.3953572167644115e-06, "loss": 0.7645, "step": 10918 }, { "epoch": 0.9304644226672347, "grad_norm": 17.75, "learning_rate": 1.3951903198882908e-06, "loss": 0.9429, "step": 10919 }, { "epoch": 0.9305496378355347, "grad_norm": 11.9375, "learning_rate": 1.3950234203884351e-06, "loss": 0.7103, "step": 10920 }, { "epoch": 0.9306348530038346, "grad_norm": 13.25, "learning_rate": 1.3948565182678602e-06, "loss": 0.4285, "step": 10921 }, { "epoch": 0.9307200681721346, "grad_norm": 12.0, "learning_rate": 1.3946896135295827e-06, "loss": 0.4695, "step": 10922 }, { "epoch": 0.9308052833404346, "grad_norm": 14.5625, "learning_rate": 1.3945227061766182e-06, "loss": 0.9451, "step": 10923 }, { "epoch": 0.9308904985087345, "grad_norm": 13.625, "learning_rate": 1.394355796211983e-06, "loss": 0.6517, "step": 10924 }, { "epoch": 0.9309757136770345, "grad_norm": 10.6875, "learning_rate": 1.3941888836386933e-06, "loss": 0.4722, "step": 10925 }, { "epoch": 0.9310609288453344, "grad_norm": 14.8125, "learning_rate": 1.3940219684597656e-06, "loss": 0.632, "step": 10926 }, { "epoch": 0.9311461440136344, "grad_norm": 14.75, "learning_rate": 1.3938550506782162e-06, "loss": 0.9764, "step": 10927 }, { "epoch": 0.9312313591819343, "grad_norm": 18.375, "learning_rate": 1.3936881302970608e-06, "loss": 0.9851, "step": 10928 }, { "epoch": 0.9313165743502343, "grad_norm": 17.0, "learning_rate": 1.3935212073193173e-06, "loss": 1.0548, "step": 10929 }, { "epoch": 0.9314017895185343, "grad_norm": 19.0, "learning_rate": 1.393354281748001e-06, "loss": 0.9416, "step": 10930 }, { "epoch": 0.9314870046868342, "grad_norm": 19.75, "learning_rate": 1.3931873535861284e-06, "loss": 1.1657, "step": 10931 }, { "epoch": 0.9315722198551342, "grad_norm": 15.25, "learning_rate": 1.393020422836716e-06, "loss": 0.824, "step": 10932 }, { "epoch": 0.9316574350234341, "grad_norm": 16.25, "learning_rate": 1.3928534895027816e-06, "loss": 0.95, "step": 10933 }, { "epoch": 0.9317426501917341, "grad_norm": 13.5, "learning_rate": 1.3926865535873401e-06, "loss": 0.7342, "step": 10934 }, { "epoch": 0.931827865360034, "grad_norm": 14.625, "learning_rate": 1.39251961509341e-06, "loss": 0.7263, "step": 10935 }, { "epoch": 0.931913080528334, "grad_norm": 18.375, "learning_rate": 1.3923526740240067e-06, "loss": 0.6572, "step": 10936 }, { "epoch": 0.931998295696634, "grad_norm": 14.9375, "learning_rate": 1.3921857303821477e-06, "loss": 0.4874, "step": 10937 }, { "epoch": 0.9320835108649339, "grad_norm": 12.5, "learning_rate": 1.3920187841708488e-06, "loss": 0.5382, "step": 10938 }, { "epoch": 0.9321687260332339, "grad_norm": 15.1875, "learning_rate": 1.3918518353931288e-06, "loss": 0.8582, "step": 10939 }, { "epoch": 0.9322539412015338, "grad_norm": 11.5625, "learning_rate": 1.3916848840520025e-06, "loss": 0.4581, "step": 10940 }, { "epoch": 0.9323391563698338, "grad_norm": 13.75, "learning_rate": 1.3915179301504883e-06, "loss": 0.4752, "step": 10941 }, { "epoch": 0.9324243715381337, "grad_norm": 21.25, "learning_rate": 1.3913509736916025e-06, "loss": 1.1669, "step": 10942 }, { "epoch": 0.9325095867064337, "grad_norm": 10.4375, "learning_rate": 1.3911840146783628e-06, "loss": 0.593, "step": 10943 }, { "epoch": 0.9325948018747336, "grad_norm": 19.625, "learning_rate": 1.3910170531137862e-06, "loss": 1.05, "step": 10944 }, { "epoch": 0.9326800170430336, "grad_norm": 16.875, "learning_rate": 1.3908500890008894e-06, "loss": 0.5251, "step": 10945 }, { "epoch": 0.9327652322113337, "grad_norm": 16.875, "learning_rate": 1.3906831223426895e-06, "loss": 0.8289, "step": 10946 }, { "epoch": 0.9328504473796336, "grad_norm": 14.875, "learning_rate": 1.3905161531422047e-06, "loss": 0.9262, "step": 10947 }, { "epoch": 0.9329356625479336, "grad_norm": 11.1875, "learning_rate": 1.3903491814024515e-06, "loss": 0.4337, "step": 10948 }, { "epoch": 0.9330208777162335, "grad_norm": 15.1875, "learning_rate": 1.390182207126447e-06, "loss": 0.5801, "step": 10949 }, { "epoch": 0.9331060928845335, "grad_norm": 24.375, "learning_rate": 1.3900152303172096e-06, "loss": 1.1195, "step": 10950 }, { "epoch": 0.9331913080528335, "grad_norm": 26.75, "learning_rate": 1.389848250977756e-06, "loss": 0.784, "step": 10951 }, { "epoch": 0.9332765232211334, "grad_norm": 15.1875, "learning_rate": 1.389681269111104e-06, "loss": 0.9963, "step": 10952 }, { "epoch": 0.9333617383894334, "grad_norm": 13.125, "learning_rate": 1.3895142847202706e-06, "loss": 0.5923, "step": 10953 }, { "epoch": 0.9334469535577333, "grad_norm": 10.875, "learning_rate": 1.3893472978082741e-06, "loss": 0.3826, "step": 10954 }, { "epoch": 0.9335321687260333, "grad_norm": 12.125, "learning_rate": 1.389180308378132e-06, "loss": 0.6673, "step": 10955 }, { "epoch": 0.9336173838943332, "grad_norm": 19.5, "learning_rate": 1.3890133164328618e-06, "loss": 1.3411, "step": 10956 }, { "epoch": 0.9337025990626332, "grad_norm": 13.125, "learning_rate": 1.3888463219754806e-06, "loss": 0.6956, "step": 10957 }, { "epoch": 0.9337878142309332, "grad_norm": 16.125, "learning_rate": 1.3886793250090071e-06, "loss": 0.6626, "step": 10958 }, { "epoch": 0.9338730293992331, "grad_norm": 15.875, "learning_rate": 1.3885123255364588e-06, "loss": 0.9476, "step": 10959 }, { "epoch": 0.9339582445675331, "grad_norm": 14.0625, "learning_rate": 1.3883453235608532e-06, "loss": 0.5006, "step": 10960 }, { "epoch": 0.934043459735833, "grad_norm": 11.6875, "learning_rate": 1.3881783190852089e-06, "loss": 0.452, "step": 10961 }, { "epoch": 0.934128674904133, "grad_norm": 12.875, "learning_rate": 1.3880113121125432e-06, "loss": 0.5628, "step": 10962 }, { "epoch": 0.9342138900724329, "grad_norm": 14.1875, "learning_rate": 1.3878443026458745e-06, "loss": 0.699, "step": 10963 }, { "epoch": 0.9342991052407329, "grad_norm": 13.6875, "learning_rate": 1.3876772906882203e-06, "loss": 0.5204, "step": 10964 }, { "epoch": 0.9343843204090329, "grad_norm": 20.125, "learning_rate": 1.3875102762425994e-06, "loss": 1.1444, "step": 10965 }, { "epoch": 0.9344695355773328, "grad_norm": 16.75, "learning_rate": 1.3873432593120289e-06, "loss": 0.7788, "step": 10966 }, { "epoch": 0.9345547507456328, "grad_norm": 12.0, "learning_rate": 1.3871762398995286e-06, "loss": 0.8329, "step": 10967 }, { "epoch": 0.9346399659139327, "grad_norm": 14.125, "learning_rate": 1.3870092180081148e-06, "loss": 0.5616, "step": 10968 }, { "epoch": 0.9347251810822327, "grad_norm": 14.4375, "learning_rate": 1.3868421936408072e-06, "loss": 0.8149, "step": 10969 }, { "epoch": 0.9348103962505326, "grad_norm": 14.5, "learning_rate": 1.386675166800623e-06, "loss": 0.8236, "step": 10970 }, { "epoch": 0.9348956114188326, "grad_norm": 15.875, "learning_rate": 1.386508137490582e-06, "loss": 0.7371, "step": 10971 }, { "epoch": 0.9349808265871326, "grad_norm": 18.25, "learning_rate": 1.3863411057137007e-06, "loss": 1.2087, "step": 10972 }, { "epoch": 0.9350660417554325, "grad_norm": 18.5, "learning_rate": 1.3861740714729991e-06, "loss": 0.7364, "step": 10973 }, { "epoch": 0.9351512569237325, "grad_norm": 28.625, "learning_rate": 1.3860070347714948e-06, "loss": 0.7654, "step": 10974 }, { "epoch": 0.9352364720920324, "grad_norm": 23.375, "learning_rate": 1.3858399956122067e-06, "loss": 0.9621, "step": 10975 }, { "epoch": 0.9353216872603324, "grad_norm": 14.0, "learning_rate": 1.3856729539981533e-06, "loss": 0.8979, "step": 10976 }, { "epoch": 0.9354069024286323, "grad_norm": 17.625, "learning_rate": 1.3855059099323534e-06, "loss": 0.9275, "step": 10977 }, { "epoch": 0.9354921175969323, "grad_norm": 21.125, "learning_rate": 1.3853388634178252e-06, "loss": 0.8401, "step": 10978 }, { "epoch": 0.9355773327652323, "grad_norm": 17.125, "learning_rate": 1.3851718144575877e-06, "loss": 0.5417, "step": 10979 }, { "epoch": 0.9356625479335322, "grad_norm": 19.5, "learning_rate": 1.3850047630546598e-06, "loss": 0.9337, "step": 10980 }, { "epoch": 0.9357477631018322, "grad_norm": 17.75, "learning_rate": 1.3848377092120595e-06, "loss": 0.7117, "step": 10981 }, { "epoch": 0.9358329782701321, "grad_norm": 12.6875, "learning_rate": 1.3846706529328068e-06, "loss": 0.4003, "step": 10982 }, { "epoch": 0.9359181934384321, "grad_norm": 17.625, "learning_rate": 1.3845035942199196e-06, "loss": 0.9912, "step": 10983 }, { "epoch": 0.936003408606732, "grad_norm": 69.0, "learning_rate": 1.3843365330764177e-06, "loss": 1.1518, "step": 10984 }, { "epoch": 0.936088623775032, "grad_norm": 17.25, "learning_rate": 1.384169469505319e-06, "loss": 0.9458, "step": 10985 }, { "epoch": 0.936173838943332, "grad_norm": 16.75, "learning_rate": 1.3840024035096434e-06, "loss": 0.8911, "step": 10986 }, { "epoch": 0.9362590541116319, "grad_norm": 17.375, "learning_rate": 1.3838353350924093e-06, "loss": 0.9467, "step": 10987 }, { "epoch": 0.9363442692799319, "grad_norm": 13.75, "learning_rate": 1.3836682642566369e-06, "loss": 0.6473, "step": 10988 }, { "epoch": 0.9364294844482318, "grad_norm": 15.0, "learning_rate": 1.3835011910053437e-06, "loss": 1.0352, "step": 10989 }, { "epoch": 0.9365146996165318, "grad_norm": 13.9375, "learning_rate": 1.38333411534155e-06, "loss": 0.5928, "step": 10990 }, { "epoch": 0.9365999147848317, "grad_norm": 12.25, "learning_rate": 1.3831670372682745e-06, "loss": 0.6214, "step": 10991 }, { "epoch": 0.9366851299531317, "grad_norm": 12.3125, "learning_rate": 1.3829999567885375e-06, "loss": 0.4847, "step": 10992 }, { "epoch": 0.9367703451214316, "grad_norm": 20.875, "learning_rate": 1.3828328739053567e-06, "loss": 1.021, "step": 10993 }, { "epoch": 0.9368555602897316, "grad_norm": 22.125, "learning_rate": 1.3826657886217526e-06, "loss": 1.0854, "step": 10994 }, { "epoch": 0.9369407754580316, "grad_norm": 12.0, "learning_rate": 1.3824987009407443e-06, "loss": 0.4298, "step": 10995 }, { "epoch": 0.9370259906263315, "grad_norm": 11.5625, "learning_rate": 1.382331610865351e-06, "loss": 0.519, "step": 10996 }, { "epoch": 0.9371112057946315, "grad_norm": 14.625, "learning_rate": 1.3821645183985929e-06, "loss": 0.493, "step": 10997 }, { "epoch": 0.9371964209629314, "grad_norm": 12.625, "learning_rate": 1.3819974235434889e-06, "loss": 0.454, "step": 10998 }, { "epoch": 0.9372816361312314, "grad_norm": 14.125, "learning_rate": 1.3818303263030586e-06, "loss": 0.9092, "step": 10999 }, { "epoch": 0.9373668512995313, "grad_norm": 17.375, "learning_rate": 1.3816632266803217e-06, "loss": 1.0118, "step": 11000 }, { "epoch": 0.9374520664678313, "grad_norm": 12.75, "learning_rate": 1.3814961246782982e-06, "loss": 0.5042, "step": 11001 }, { "epoch": 0.9375372816361313, "grad_norm": 14.8125, "learning_rate": 1.381329020300007e-06, "loss": 0.6875, "step": 11002 }, { "epoch": 0.9376224968044312, "grad_norm": 17.0, "learning_rate": 1.3811619135484689e-06, "loss": 0.7561, "step": 11003 }, { "epoch": 0.9377077119727312, "grad_norm": 18.0, "learning_rate": 1.3809948044267028e-06, "loss": 0.8927, "step": 11004 }, { "epoch": 0.9377929271410311, "grad_norm": 32.75, "learning_rate": 1.380827692937729e-06, "loss": 0.9235, "step": 11005 }, { "epoch": 0.9378781423093311, "grad_norm": 14.75, "learning_rate": 1.380660579084567e-06, "loss": 0.7896, "step": 11006 }, { "epoch": 0.937963357477631, "grad_norm": 25.0, "learning_rate": 1.3804934628702372e-06, "loss": 1.4595, "step": 11007 }, { "epoch": 0.938048572645931, "grad_norm": 12.875, "learning_rate": 1.3803263442977592e-06, "loss": 0.5708, "step": 11008 }, { "epoch": 0.938133787814231, "grad_norm": 14.0, "learning_rate": 1.3801592233701533e-06, "loss": 0.9006, "step": 11009 }, { "epoch": 0.9382190029825309, "grad_norm": 12.4375, "learning_rate": 1.3799921000904386e-06, "loss": 0.4296, "step": 11010 }, { "epoch": 0.9383042181508309, "grad_norm": 19.25, "learning_rate": 1.3798249744616368e-06, "loss": 1.0577, "step": 11011 }, { "epoch": 0.9383894333191308, "grad_norm": 16.625, "learning_rate": 1.3796578464867668e-06, "loss": 0.8821, "step": 11012 }, { "epoch": 0.9384746484874308, "grad_norm": 15.75, "learning_rate": 1.379490716168849e-06, "loss": 0.6517, "step": 11013 }, { "epoch": 0.9385598636557307, "grad_norm": 15.75, "learning_rate": 1.379323583510904e-06, "loss": 0.9636, "step": 11014 }, { "epoch": 0.9386450788240307, "grad_norm": 17.125, "learning_rate": 1.379156448515952e-06, "loss": 0.7771, "step": 11015 }, { "epoch": 0.9387302939923307, "grad_norm": 19.0, "learning_rate": 1.378989311187013e-06, "loss": 1.2333, "step": 11016 }, { "epoch": 0.9388155091606306, "grad_norm": 14.75, "learning_rate": 1.3788221715271068e-06, "loss": 0.6809, "step": 11017 }, { "epoch": 0.9389007243289306, "grad_norm": 18.625, "learning_rate": 1.3786550295392551e-06, "loss": 0.6074, "step": 11018 }, { "epoch": 0.9389859394972305, "grad_norm": 11.6875, "learning_rate": 1.3784878852264771e-06, "loss": 0.53, "step": 11019 }, { "epoch": 0.9390711546655305, "grad_norm": 16.25, "learning_rate": 1.3783207385917947e-06, "loss": 0.7652, "step": 11020 }, { "epoch": 0.9391563698338304, "grad_norm": 21.875, "learning_rate": 1.3781535896382265e-06, "loss": 0.9202, "step": 11021 }, { "epoch": 0.9392415850021304, "grad_norm": 11.9375, "learning_rate": 1.3779864383687947e-06, "loss": 0.4956, "step": 11022 }, { "epoch": 0.9393268001704304, "grad_norm": 11.5, "learning_rate": 1.3778192847865188e-06, "loss": 0.4734, "step": 11023 }, { "epoch": 0.9394120153387303, "grad_norm": 13.0625, "learning_rate": 1.3776521288944206e-06, "loss": 0.7378, "step": 11024 }, { "epoch": 0.9394972305070303, "grad_norm": 11.1875, "learning_rate": 1.3774849706955193e-06, "loss": 0.3357, "step": 11025 }, { "epoch": 0.9395824456753302, "grad_norm": 15.0625, "learning_rate": 1.377317810192837e-06, "loss": 0.6553, "step": 11026 }, { "epoch": 0.9396676608436302, "grad_norm": 15.8125, "learning_rate": 1.3771506473893933e-06, "loss": 0.8324, "step": 11027 }, { "epoch": 0.9397528760119301, "grad_norm": 18.75, "learning_rate": 1.3769834822882096e-06, "loss": 1.2539, "step": 11028 }, { "epoch": 0.9398380911802301, "grad_norm": 17.125, "learning_rate": 1.3768163148923073e-06, "loss": 0.9988, "step": 11029 }, { "epoch": 0.93992330634853, "grad_norm": 20.0, "learning_rate": 1.3766491452047062e-06, "loss": 1.0352, "step": 11030 }, { "epoch": 0.94000852151683, "grad_norm": 15.1875, "learning_rate": 1.3764819732284276e-06, "loss": 0.6993, "step": 11031 }, { "epoch": 0.94009373668513, "grad_norm": 13.25, "learning_rate": 1.3763147989664927e-06, "loss": 0.5443, "step": 11032 }, { "epoch": 0.9401789518534299, "grad_norm": 13.375, "learning_rate": 1.3761476224219226e-06, "loss": 0.6295, "step": 11033 }, { "epoch": 0.9402641670217299, "grad_norm": 19.375, "learning_rate": 1.3759804435977375e-06, "loss": 1.2088, "step": 11034 }, { "epoch": 0.9403493821900298, "grad_norm": 17.375, "learning_rate": 1.3758132624969594e-06, "loss": 1.0608, "step": 11035 }, { "epoch": 0.9404345973583298, "grad_norm": 11.5625, "learning_rate": 1.3756460791226095e-06, "loss": 0.444, "step": 11036 }, { "epoch": 0.9405198125266298, "grad_norm": 13.4375, "learning_rate": 1.3754788934777084e-06, "loss": 0.8103, "step": 11037 }, { "epoch": 0.9406050276949297, "grad_norm": 15.4375, "learning_rate": 1.3753117055652774e-06, "loss": 0.4821, "step": 11038 }, { "epoch": 0.9406902428632297, "grad_norm": 17.375, "learning_rate": 1.3751445153883377e-06, "loss": 0.9522, "step": 11039 }, { "epoch": 0.9407754580315296, "grad_norm": 15.4375, "learning_rate": 1.374977322949911e-06, "loss": 0.8709, "step": 11040 }, { "epoch": 0.9408606731998296, "grad_norm": 14.75, "learning_rate": 1.3748101282530185e-06, "loss": 0.6893, "step": 11041 }, { "epoch": 0.9409458883681295, "grad_norm": 15.125, "learning_rate": 1.3746429313006812e-06, "loss": 0.8334, "step": 11042 }, { "epoch": 0.9410311035364295, "grad_norm": 14.75, "learning_rate": 1.374475732095921e-06, "loss": 0.5598, "step": 11043 }, { "epoch": 0.9411163187047294, "grad_norm": 14.25, "learning_rate": 1.374308530641759e-06, "loss": 0.4919, "step": 11044 }, { "epoch": 0.9412015338730294, "grad_norm": 24.75, "learning_rate": 1.3741413269412172e-06, "loss": 1.2341, "step": 11045 }, { "epoch": 0.9412867490413294, "grad_norm": 16.375, "learning_rate": 1.373974120997316e-06, "loss": 0.765, "step": 11046 }, { "epoch": 0.9413719642096293, "grad_norm": 18.0, "learning_rate": 1.3738069128130787e-06, "loss": 0.7991, "step": 11047 }, { "epoch": 0.9414571793779293, "grad_norm": 11.4375, "learning_rate": 1.3736397023915254e-06, "loss": 0.468, "step": 11048 }, { "epoch": 0.9415423945462292, "grad_norm": 16.875, "learning_rate": 1.3734724897356785e-06, "loss": 0.8174, "step": 11049 }, { "epoch": 0.9416276097145292, "grad_norm": 13.625, "learning_rate": 1.3733052748485598e-06, "loss": 0.5951, "step": 11050 }, { "epoch": 0.9417128248828291, "grad_norm": 17.625, "learning_rate": 1.3731380577331901e-06, "loss": 0.9719, "step": 11051 }, { "epoch": 0.9417980400511291, "grad_norm": 12.0, "learning_rate": 1.3729708383925925e-06, "loss": 0.5042, "step": 11052 }, { "epoch": 0.9418832552194291, "grad_norm": 14.5, "learning_rate": 1.3728036168297882e-06, "loss": 0.8386, "step": 11053 }, { "epoch": 0.941968470387729, "grad_norm": 14.0625, "learning_rate": 1.372636393047799e-06, "loss": 0.6466, "step": 11054 }, { "epoch": 0.942053685556029, "grad_norm": 15.5625, "learning_rate": 1.3724691670496466e-06, "loss": 0.757, "step": 11055 }, { "epoch": 0.9421389007243289, "grad_norm": 16.25, "learning_rate": 1.3723019388383532e-06, "loss": 0.8277, "step": 11056 }, { "epoch": 0.9422241158926289, "grad_norm": 17.875, "learning_rate": 1.3721347084169412e-06, "loss": 0.9954, "step": 11057 }, { "epoch": 0.9423093310609288, "grad_norm": 13.375, "learning_rate": 1.3719674757884318e-06, "loss": 0.7858, "step": 11058 }, { "epoch": 0.9423945462292288, "grad_norm": 19.5, "learning_rate": 1.3718002409558476e-06, "loss": 0.9745, "step": 11059 }, { "epoch": 0.9424797613975288, "grad_norm": 19.125, "learning_rate": 1.3716330039222106e-06, "loss": 1.0373, "step": 11060 }, { "epoch": 0.9425649765658287, "grad_norm": 13.9375, "learning_rate": 1.3714657646905428e-06, "loss": 0.7585, "step": 11061 }, { "epoch": 0.9426501917341287, "grad_norm": 15.375, "learning_rate": 1.3712985232638668e-06, "loss": 0.8033, "step": 11062 }, { "epoch": 0.9427354069024286, "grad_norm": 23.125, "learning_rate": 1.371131279645204e-06, "loss": 0.8723, "step": 11063 }, { "epoch": 0.9428206220707286, "grad_norm": 14.125, "learning_rate": 1.3709640338375775e-06, "loss": 0.6407, "step": 11064 }, { "epoch": 0.9429058372390285, "grad_norm": 19.25, "learning_rate": 1.3707967858440092e-06, "loss": 0.8283, "step": 11065 }, { "epoch": 0.9429910524073285, "grad_norm": 14.4375, "learning_rate": 1.3706295356675214e-06, "loss": 0.8207, "step": 11066 }, { "epoch": 0.9430762675756285, "grad_norm": 13.6875, "learning_rate": 1.3704622833111367e-06, "loss": 0.8268, "step": 11067 }, { "epoch": 0.9431614827439284, "grad_norm": 16.0, "learning_rate": 1.3702950287778772e-06, "loss": 0.8447, "step": 11068 }, { "epoch": 0.9432466979122284, "grad_norm": 12.875, "learning_rate": 1.3701277720707662e-06, "loss": 0.6439, "step": 11069 }, { "epoch": 0.9433319130805283, "grad_norm": 12.0, "learning_rate": 1.3699605131928247e-06, "loss": 0.5469, "step": 11070 }, { "epoch": 0.9434171282488283, "grad_norm": 17.125, "learning_rate": 1.3697932521470766e-06, "loss": 0.7244, "step": 11071 }, { "epoch": 0.9435023434171282, "grad_norm": 22.0, "learning_rate": 1.3696259889365438e-06, "loss": 0.9343, "step": 11072 }, { "epoch": 0.9435875585854282, "grad_norm": 14.3125, "learning_rate": 1.3694587235642496e-06, "loss": 0.6349, "step": 11073 }, { "epoch": 0.9436727737537282, "grad_norm": 12.9375, "learning_rate": 1.3692914560332154e-06, "loss": 0.5881, "step": 11074 }, { "epoch": 0.9437579889220281, "grad_norm": 15.6875, "learning_rate": 1.3691241863464653e-06, "loss": 0.8899, "step": 11075 }, { "epoch": 0.9438432040903281, "grad_norm": 16.625, "learning_rate": 1.3689569145070205e-06, "loss": 0.7084, "step": 11076 }, { "epoch": 0.943928419258628, "grad_norm": 12.5, "learning_rate": 1.3687896405179058e-06, "loss": 0.6732, "step": 11077 }, { "epoch": 0.944013634426928, "grad_norm": 13.3125, "learning_rate": 1.368622364382142e-06, "loss": 0.5679, "step": 11078 }, { "epoch": 0.9440988495952279, "grad_norm": 13.8125, "learning_rate": 1.368455086102753e-06, "loss": 0.6421, "step": 11079 }, { "epoch": 0.9441840647635279, "grad_norm": 17.25, "learning_rate": 1.368287805682761e-06, "loss": 0.83, "step": 11080 }, { "epoch": 0.9442692799318279, "grad_norm": 13.5, "learning_rate": 1.36812052312519e-06, "loss": 0.4536, "step": 11081 }, { "epoch": 0.9443544951001278, "grad_norm": 14.25, "learning_rate": 1.3679532384330624e-06, "loss": 0.7111, "step": 11082 }, { "epoch": 0.9444397102684278, "grad_norm": 14.375, "learning_rate": 1.3677859516094008e-06, "loss": 0.5433, "step": 11083 }, { "epoch": 0.9445249254367277, "grad_norm": 13.25, "learning_rate": 1.3676186626572291e-06, "loss": 0.7989, "step": 11084 }, { "epoch": 0.9446101406050277, "grad_norm": 12.5, "learning_rate": 1.3674513715795695e-06, "loss": 0.5892, "step": 11085 }, { "epoch": 0.9446953557733276, "grad_norm": 21.0, "learning_rate": 1.3672840783794458e-06, "loss": 0.8751, "step": 11086 }, { "epoch": 0.9447805709416276, "grad_norm": 17.125, "learning_rate": 1.3671167830598806e-06, "loss": 0.9412, "step": 11087 }, { "epoch": 0.9448657861099276, "grad_norm": 12.1875, "learning_rate": 1.3669494856238974e-06, "loss": 0.5189, "step": 11088 }, { "epoch": 0.9449510012782275, "grad_norm": 12.625, "learning_rate": 1.3667821860745196e-06, "loss": 0.4717, "step": 11089 }, { "epoch": 0.9450362164465275, "grad_norm": 26.125, "learning_rate": 1.36661488441477e-06, "loss": 0.9389, "step": 11090 }, { "epoch": 0.9451214316148274, "grad_norm": 12.6875, "learning_rate": 1.366447580647672e-06, "loss": 0.5008, "step": 11091 }, { "epoch": 0.9452066467831274, "grad_norm": 19.875, "learning_rate": 1.3662802747762495e-06, "loss": 0.9154, "step": 11092 }, { "epoch": 0.9452918619514273, "grad_norm": 21.625, "learning_rate": 1.3661129668035255e-06, "loss": 1.2268, "step": 11093 }, { "epoch": 0.9453770771197273, "grad_norm": 28.375, "learning_rate": 1.3659456567325236e-06, "loss": 1.0453, "step": 11094 }, { "epoch": 0.9454622922880273, "grad_norm": 9.75, "learning_rate": 1.3657783445662664e-06, "loss": 1.2264, "step": 11095 }, { "epoch": 0.9455475074563272, "grad_norm": 13.6875, "learning_rate": 1.3656110303077788e-06, "loss": 0.6336, "step": 11096 }, { "epoch": 0.9456327226246272, "grad_norm": 14.0, "learning_rate": 1.3654437139600834e-06, "loss": 0.6112, "step": 11097 }, { "epoch": 0.9457179377929271, "grad_norm": 13.875, "learning_rate": 1.3652763955262039e-06, "loss": 0.5786, "step": 11098 }, { "epoch": 0.9458031529612271, "grad_norm": 14.3125, "learning_rate": 1.365109075009164e-06, "loss": 0.6727, "step": 11099 }, { "epoch": 0.945888368129527, "grad_norm": 17.125, "learning_rate": 1.3649417524119874e-06, "loss": 0.8454, "step": 11100 }, { "epoch": 0.945973583297827, "grad_norm": 18.5, "learning_rate": 1.3647744277376979e-06, "loss": 0.7791, "step": 11101 }, { "epoch": 0.946058798466127, "grad_norm": 15.125, "learning_rate": 1.3646071009893188e-06, "loss": 0.6159, "step": 11102 }, { "epoch": 0.9461440136344269, "grad_norm": 16.125, "learning_rate": 1.3644397721698742e-06, "loss": 0.9871, "step": 11103 }, { "epoch": 0.9462292288027269, "grad_norm": 10.625, "learning_rate": 1.364272441282388e-06, "loss": 0.4288, "step": 11104 }, { "epoch": 0.9463144439710268, "grad_norm": 13.5, "learning_rate": 1.364105108329884e-06, "loss": 0.5338, "step": 11105 }, { "epoch": 0.9463996591393268, "grad_norm": 14.4375, "learning_rate": 1.3639377733153856e-06, "loss": 0.6392, "step": 11106 }, { "epoch": 0.9464848743076267, "grad_norm": 14.375, "learning_rate": 1.3637704362419171e-06, "loss": 0.6271, "step": 11107 }, { "epoch": 0.9465700894759267, "grad_norm": 16.25, "learning_rate": 1.3636030971125026e-06, "loss": 0.896, "step": 11108 }, { "epoch": 0.9466553046442266, "grad_norm": 25.625, "learning_rate": 1.3634357559301662e-06, "loss": 1.2843, "step": 11109 }, { "epoch": 0.9467405198125266, "grad_norm": 14.25, "learning_rate": 1.363268412697931e-06, "loss": 0.7364, "step": 11110 }, { "epoch": 0.9468257349808266, "grad_norm": 12.125, "learning_rate": 1.3631010674188219e-06, "loss": 0.4524, "step": 11111 }, { "epoch": 0.9469109501491265, "grad_norm": 13.0625, "learning_rate": 1.3629337200958626e-06, "loss": 0.505, "step": 11112 }, { "epoch": 0.9469961653174265, "grad_norm": 13.0, "learning_rate": 1.3627663707320775e-06, "loss": 0.6783, "step": 11113 }, { "epoch": 0.9470813804857264, "grad_norm": 17.5, "learning_rate": 1.3625990193304911e-06, "loss": 0.5561, "step": 11114 }, { "epoch": 0.9471665956540264, "grad_norm": 21.75, "learning_rate": 1.362431665894127e-06, "loss": 0.8594, "step": 11115 }, { "epoch": 0.9472518108223263, "grad_norm": 19.25, "learning_rate": 1.3622643104260095e-06, "loss": 1.1498, "step": 11116 }, { "epoch": 0.9473370259906263, "grad_norm": 9.6875, "learning_rate": 1.362096952929163e-06, "loss": 0.2352, "step": 11117 }, { "epoch": 0.9474222411589263, "grad_norm": 15.3125, "learning_rate": 1.3619295934066117e-06, "loss": 0.8051, "step": 11118 }, { "epoch": 0.9475074563272262, "grad_norm": 17.375, "learning_rate": 1.3617622318613806e-06, "loss": 0.8366, "step": 11119 }, { "epoch": 0.9475926714955262, "grad_norm": 14.1875, "learning_rate": 1.361594868296493e-06, "loss": 0.6228, "step": 11120 }, { "epoch": 0.9476778866638261, "grad_norm": 11.125, "learning_rate": 1.361427502714974e-06, "loss": 0.423, "step": 11121 }, { "epoch": 0.9477631018321261, "grad_norm": 23.875, "learning_rate": 1.3612601351198485e-06, "loss": 1.1474, "step": 11122 }, { "epoch": 0.947848317000426, "grad_norm": 14.625, "learning_rate": 1.3610927655141402e-06, "loss": 0.6743, "step": 11123 }, { "epoch": 0.947933532168726, "grad_norm": 10.4375, "learning_rate": 1.3609253939008738e-06, "loss": 0.4045, "step": 11124 }, { "epoch": 0.948018747337026, "grad_norm": 15.0625, "learning_rate": 1.3607580202830739e-06, "loss": 0.4785, "step": 11125 }, { "epoch": 0.9481039625053259, "grad_norm": 18.5, "learning_rate": 1.3605906446637656e-06, "loss": 0.9845, "step": 11126 }, { "epoch": 0.9481891776736259, "grad_norm": 12.5, "learning_rate": 1.3604232670459727e-06, "loss": 0.6442, "step": 11127 }, { "epoch": 0.9482743928419258, "grad_norm": 13.9375, "learning_rate": 1.3602558874327206e-06, "loss": 0.8672, "step": 11128 }, { "epoch": 0.9483596080102258, "grad_norm": 19.125, "learning_rate": 1.3600885058270335e-06, "loss": 0.7797, "step": 11129 }, { "epoch": 0.9484448231785257, "grad_norm": 11.25, "learning_rate": 1.359921122231937e-06, "loss": 0.543, "step": 11130 }, { "epoch": 0.9485300383468257, "grad_norm": 13.4375, "learning_rate": 1.3597537366504546e-06, "loss": 0.5551, "step": 11131 }, { "epoch": 0.9486152535151257, "grad_norm": 11.5, "learning_rate": 1.359586349085612e-06, "loss": 0.6685, "step": 11132 }, { "epoch": 0.9487004686834256, "grad_norm": 17.0, "learning_rate": 1.3594189595404338e-06, "loss": 0.7485, "step": 11133 }, { "epoch": 0.9487856838517256, "grad_norm": 15.4375, "learning_rate": 1.3592515680179452e-06, "loss": 1.0117, "step": 11134 }, { "epoch": 0.9488708990200255, "grad_norm": 14.5625, "learning_rate": 1.3590841745211707e-06, "loss": 0.7559, "step": 11135 }, { "epoch": 0.9489561141883255, "grad_norm": 14.125, "learning_rate": 1.3589167790531354e-06, "loss": 0.996, "step": 11136 }, { "epoch": 0.9490413293566254, "grad_norm": 14.25, "learning_rate": 1.3587493816168648e-06, "loss": 0.4351, "step": 11137 }, { "epoch": 0.9491265445249254, "grad_norm": 12.5, "learning_rate": 1.3585819822153834e-06, "loss": 0.5481, "step": 11138 }, { "epoch": 0.9492117596932254, "grad_norm": 11.625, "learning_rate": 1.3584145808517163e-06, "loss": 0.5167, "step": 11139 }, { "epoch": 0.9492969748615253, "grad_norm": 23.5, "learning_rate": 1.3582471775288883e-06, "loss": 0.8805, "step": 11140 }, { "epoch": 0.9493821900298253, "grad_norm": 12.625, "learning_rate": 1.3580797722499256e-06, "loss": 0.8834, "step": 11141 }, { "epoch": 0.9494674051981252, "grad_norm": 15.6875, "learning_rate": 1.3579123650178525e-06, "loss": 1.0889, "step": 11142 }, { "epoch": 0.9495526203664252, "grad_norm": 13.9375, "learning_rate": 1.3577449558356944e-06, "loss": 0.5831, "step": 11143 }, { "epoch": 0.9496378355347251, "grad_norm": 12.1875, "learning_rate": 1.3575775447064765e-06, "loss": 0.5942, "step": 11144 }, { "epoch": 0.9497230507030251, "grad_norm": 17.125, "learning_rate": 1.3574101316332244e-06, "loss": 0.4823, "step": 11145 }, { "epoch": 0.949808265871325, "grad_norm": 11.5, "learning_rate": 1.357242716618963e-06, "loss": 0.6503, "step": 11146 }, { "epoch": 0.949893481039625, "grad_norm": 13.5, "learning_rate": 1.357075299666718e-06, "loss": 0.8248, "step": 11147 }, { "epoch": 0.949978696207925, "grad_norm": 23.75, "learning_rate": 1.3569078807795148e-06, "loss": 1.0501, "step": 11148 }, { "epoch": 0.9500639113762249, "grad_norm": 16.125, "learning_rate": 1.3567404599603784e-06, "loss": 0.724, "step": 11149 }, { "epoch": 0.9501491265445249, "grad_norm": 14.5625, "learning_rate": 1.3565730372123348e-06, "loss": 0.8842, "step": 11150 }, { "epoch": 0.9502343417128248, "grad_norm": 16.0, "learning_rate": 1.356405612538409e-06, "loss": 1.0217, "step": 11151 }, { "epoch": 0.9503195568811248, "grad_norm": 18.125, "learning_rate": 1.3562381859416268e-06, "loss": 0.6596, "step": 11152 }, { "epoch": 0.9504047720494248, "grad_norm": 18.5, "learning_rate": 1.3560707574250137e-06, "loss": 1.1156, "step": 11153 }, { "epoch": 0.9504899872177247, "grad_norm": 12.6875, "learning_rate": 1.3559033269915959e-06, "loss": 0.5993, "step": 11154 }, { "epoch": 0.9505752023860247, "grad_norm": 13.75, "learning_rate": 1.355735894644398e-06, "loss": 0.6029, "step": 11155 }, { "epoch": 0.9506604175543246, "grad_norm": 15.5, "learning_rate": 1.3555684603864464e-06, "loss": 0.8165, "step": 11156 }, { "epoch": 0.9507456327226246, "grad_norm": 13.1875, "learning_rate": 1.355401024220766e-06, "loss": 0.7239, "step": 11157 }, { "epoch": 0.9508308478909245, "grad_norm": 17.375, "learning_rate": 1.355233586150384e-06, "loss": 1.0024, "step": 11158 }, { "epoch": 0.9509160630592245, "grad_norm": 21.875, "learning_rate": 1.3550661461783245e-06, "loss": 0.6146, "step": 11159 }, { "epoch": 0.9510012782275244, "grad_norm": 11.5625, "learning_rate": 1.3548987043076145e-06, "loss": 0.5588, "step": 11160 }, { "epoch": 0.9510864933958244, "grad_norm": 12.125, "learning_rate": 1.3547312605412793e-06, "loss": 0.5698, "step": 11161 }, { "epoch": 0.9511717085641244, "grad_norm": 14.6875, "learning_rate": 1.3545638148823453e-06, "loss": 0.8529, "step": 11162 }, { "epoch": 0.9512569237324243, "grad_norm": 19.25, "learning_rate": 1.3543963673338372e-06, "loss": 0.442, "step": 11163 }, { "epoch": 0.9513421389007243, "grad_norm": 17.375, "learning_rate": 1.3542289178987825e-06, "loss": 0.8318, "step": 11164 }, { "epoch": 0.9514273540690242, "grad_norm": 25.5, "learning_rate": 1.354061466580206e-06, "loss": 0.491, "step": 11165 }, { "epoch": 0.9515125692373242, "grad_norm": 40.0, "learning_rate": 1.3538940133811344e-06, "loss": 0.9467, "step": 11166 }, { "epoch": 0.9515977844056241, "grad_norm": 12.8125, "learning_rate": 1.3537265583045933e-06, "loss": 0.5424, "step": 11167 }, { "epoch": 0.9516829995739241, "grad_norm": 15.0625, "learning_rate": 1.353559101353609e-06, "loss": 0.8793, "step": 11168 }, { "epoch": 0.9517682147422241, "grad_norm": 14.75, "learning_rate": 1.3533916425312076e-06, "loss": 0.8525, "step": 11169 }, { "epoch": 0.951853429910524, "grad_norm": 17.75, "learning_rate": 1.3532241818404157e-06, "loss": 1.0974, "step": 11170 }, { "epoch": 0.9519386450788241, "grad_norm": 17.625, "learning_rate": 1.3530567192842586e-06, "loss": 0.9203, "step": 11171 }, { "epoch": 0.952023860247124, "grad_norm": 14.8125, "learning_rate": 1.352889254865763e-06, "loss": 0.9527, "step": 11172 }, { "epoch": 0.952109075415424, "grad_norm": 26.0, "learning_rate": 1.352721788587955e-06, "loss": 1.0209, "step": 11173 }, { "epoch": 0.952194290583724, "grad_norm": 14.0625, "learning_rate": 1.3525543204538613e-06, "loss": 0.681, "step": 11174 }, { "epoch": 0.9522795057520239, "grad_norm": 14.9375, "learning_rate": 1.3523868504665077e-06, "loss": 0.6908, "step": 11175 }, { "epoch": 0.9523647209203239, "grad_norm": 25.375, "learning_rate": 1.3522193786289206e-06, "loss": 0.9578, "step": 11176 }, { "epoch": 0.9524499360886238, "grad_norm": 17.125, "learning_rate": 1.3520519049441268e-06, "loss": 0.9875, "step": 11177 }, { "epoch": 0.9525351512569238, "grad_norm": 17.375, "learning_rate": 1.3518844294151523e-06, "loss": 0.8847, "step": 11178 }, { "epoch": 0.9526203664252237, "grad_norm": 13.9375, "learning_rate": 1.351716952045024e-06, "loss": 0.7185, "step": 11179 }, { "epoch": 0.9527055815935237, "grad_norm": 20.0, "learning_rate": 1.3515494728367677e-06, "loss": 0.797, "step": 11180 }, { "epoch": 0.9527907967618237, "grad_norm": 14.0625, "learning_rate": 1.3513819917934108e-06, "loss": 0.7569, "step": 11181 }, { "epoch": 0.9528760119301236, "grad_norm": 11.9375, "learning_rate": 1.3512145089179787e-06, "loss": 0.5086, "step": 11182 }, { "epoch": 0.9529612270984236, "grad_norm": 13.125, "learning_rate": 1.351047024213499e-06, "loss": 0.7267, "step": 11183 }, { "epoch": 0.9530464422667235, "grad_norm": 12.1875, "learning_rate": 1.350879537682998e-06, "loss": 0.673, "step": 11184 }, { "epoch": 0.9531316574350235, "grad_norm": 23.5, "learning_rate": 1.3507120493295026e-06, "loss": 0.9373, "step": 11185 }, { "epoch": 0.9532168726033234, "grad_norm": 11.0, "learning_rate": 1.3505445591560387e-06, "loss": 0.4915, "step": 11186 }, { "epoch": 0.9533020877716234, "grad_norm": 20.0, "learning_rate": 1.3503770671656336e-06, "loss": 0.9672, "step": 11187 }, { "epoch": 0.9533873029399234, "grad_norm": 17.625, "learning_rate": 1.350209573361314e-06, "loss": 1.1056, "step": 11188 }, { "epoch": 0.9534725181082233, "grad_norm": 20.125, "learning_rate": 1.3500420777461065e-06, "loss": 0.5954, "step": 11189 }, { "epoch": 0.9535577332765233, "grad_norm": 14.5625, "learning_rate": 1.3498745803230383e-06, "loss": 0.9035, "step": 11190 }, { "epoch": 0.9536429484448232, "grad_norm": 20.25, "learning_rate": 1.3497070810951363e-06, "loss": 0.8066, "step": 11191 }, { "epoch": 0.9537281636131232, "grad_norm": 14.5, "learning_rate": 1.3495395800654265e-06, "loss": 0.9601, "step": 11192 }, { "epoch": 0.9538133787814231, "grad_norm": 17.625, "learning_rate": 1.3493720772369365e-06, "loss": 1.07, "step": 11193 }, { "epoch": 0.9538985939497231, "grad_norm": 21.875, "learning_rate": 1.3492045726126934e-06, "loss": 0.6457, "step": 11194 }, { "epoch": 0.953983809118023, "grad_norm": 12.5625, "learning_rate": 1.3490370661957238e-06, "loss": 0.5155, "step": 11195 }, { "epoch": 0.954069024286323, "grad_norm": 14.4375, "learning_rate": 1.3488695579890549e-06, "loss": 0.9245, "step": 11196 }, { "epoch": 0.954154239454623, "grad_norm": 14.5625, "learning_rate": 1.3487020479957133e-06, "loss": 0.5774, "step": 11197 }, { "epoch": 0.9542394546229229, "grad_norm": 14.5, "learning_rate": 1.3485345362187268e-06, "loss": 0.9729, "step": 11198 }, { "epoch": 0.9543246697912229, "grad_norm": 13.3125, "learning_rate": 1.348367022661122e-06, "loss": 0.8422, "step": 11199 }, { "epoch": 0.9544098849595228, "grad_norm": 11.3125, "learning_rate": 1.3481995073259263e-06, "loss": 0.3893, "step": 11200 }, { "epoch": 0.9544951001278228, "grad_norm": 12.625, "learning_rate": 1.3480319902161665e-06, "loss": 0.5194, "step": 11201 }, { "epoch": 0.9545803152961228, "grad_norm": 15.6875, "learning_rate": 1.3478644713348705e-06, "loss": 1.1465, "step": 11202 }, { "epoch": 0.9546655304644227, "grad_norm": 12.625, "learning_rate": 1.3476969506850648e-06, "loss": 0.6907, "step": 11203 }, { "epoch": 0.9547507456327227, "grad_norm": 12.8125, "learning_rate": 1.3475294282697769e-06, "loss": 0.7228, "step": 11204 }, { "epoch": 0.9548359608010226, "grad_norm": 21.75, "learning_rate": 1.347361904092034e-06, "loss": 0.8344, "step": 11205 }, { "epoch": 0.9549211759693226, "grad_norm": 11.5625, "learning_rate": 1.3471943781548638e-06, "loss": 0.4806, "step": 11206 }, { "epoch": 0.9550063911376225, "grad_norm": 14.375, "learning_rate": 1.3470268504612937e-06, "loss": 0.526, "step": 11207 }, { "epoch": 0.9550916063059225, "grad_norm": 14.5625, "learning_rate": 1.3468593210143508e-06, "loss": 0.4478, "step": 11208 }, { "epoch": 0.9551768214742224, "grad_norm": 37.5, "learning_rate": 1.3466917898170623e-06, "loss": 1.4774, "step": 11209 }, { "epoch": 0.9552620366425224, "grad_norm": 20.75, "learning_rate": 1.3465242568724556e-06, "loss": 0.9637, "step": 11210 }, { "epoch": 0.9553472518108224, "grad_norm": 11.375, "learning_rate": 1.3463567221835593e-06, "loss": 0.5761, "step": 11211 }, { "epoch": 0.9554324669791223, "grad_norm": 15.625, "learning_rate": 1.3461891857533998e-06, "loss": 0.9901, "step": 11212 }, { "epoch": 0.9555176821474223, "grad_norm": 16.375, "learning_rate": 1.346021647585005e-06, "loss": 1.0439, "step": 11213 }, { "epoch": 0.9556028973157222, "grad_norm": 16.125, "learning_rate": 1.3458541076814022e-06, "loss": 0.927, "step": 11214 }, { "epoch": 0.9556881124840222, "grad_norm": 13.0, "learning_rate": 1.34568656604562e-06, "loss": 0.6717, "step": 11215 }, { "epoch": 0.9557733276523221, "grad_norm": 18.375, "learning_rate": 1.3455190226806846e-06, "loss": 1.05, "step": 11216 }, { "epoch": 0.9558585428206221, "grad_norm": 16.625, "learning_rate": 1.3453514775896248e-06, "loss": 0.7356, "step": 11217 }, { "epoch": 0.9559437579889221, "grad_norm": 15.6875, "learning_rate": 1.3451839307754673e-06, "loss": 1.0657, "step": 11218 }, { "epoch": 0.956028973157222, "grad_norm": 12.1875, "learning_rate": 1.3450163822412413e-06, "loss": 0.4628, "step": 11219 }, { "epoch": 0.956114188325522, "grad_norm": 17.375, "learning_rate": 1.3448488319899734e-06, "loss": 0.6879, "step": 11220 }, { "epoch": 0.9561994034938219, "grad_norm": 13.25, "learning_rate": 1.3446812800246913e-06, "loss": 0.4198, "step": 11221 }, { "epoch": 0.9562846186621219, "grad_norm": 13.5, "learning_rate": 1.344513726348424e-06, "loss": 0.5079, "step": 11222 }, { "epoch": 0.9563698338304218, "grad_norm": 18.75, "learning_rate": 1.3443461709641983e-06, "loss": 0.6333, "step": 11223 }, { "epoch": 0.9564550489987218, "grad_norm": 19.25, "learning_rate": 1.3441786138750424e-06, "loss": 0.7613, "step": 11224 }, { "epoch": 0.9565402641670218, "grad_norm": 44.25, "learning_rate": 1.344011055083984e-06, "loss": 1.1897, "step": 11225 }, { "epoch": 0.9566254793353217, "grad_norm": 12.0, "learning_rate": 1.3438434945940515e-06, "loss": 0.7008, "step": 11226 }, { "epoch": 0.9567106945036217, "grad_norm": 12.8125, "learning_rate": 1.343675932408273e-06, "loss": 0.9346, "step": 11227 }, { "epoch": 0.9567959096719216, "grad_norm": 17.875, "learning_rate": 1.343508368529676e-06, "loss": 0.7028, "step": 11228 }, { "epoch": 0.9568811248402216, "grad_norm": 14.5625, "learning_rate": 1.3433408029612885e-06, "loss": 0.8707, "step": 11229 }, { "epoch": 0.9569663400085215, "grad_norm": 22.25, "learning_rate": 1.3431732357061389e-06, "loss": 1.4393, "step": 11230 }, { "epoch": 0.9570515551768215, "grad_norm": 11.125, "learning_rate": 1.3430056667672555e-06, "loss": 0.3387, "step": 11231 }, { "epoch": 0.9571367703451215, "grad_norm": 18.375, "learning_rate": 1.3428380961476662e-06, "loss": 0.5938, "step": 11232 }, { "epoch": 0.9572219855134214, "grad_norm": 19.25, "learning_rate": 1.3426705238503986e-06, "loss": 0.9805, "step": 11233 }, { "epoch": 0.9573072006817214, "grad_norm": 18.875, "learning_rate": 1.3425029498784819e-06, "loss": 1.0721, "step": 11234 }, { "epoch": 0.9573924158500213, "grad_norm": 10.3125, "learning_rate": 1.3423353742349441e-06, "loss": 0.355, "step": 11235 }, { "epoch": 0.9574776310183213, "grad_norm": 17.625, "learning_rate": 1.3421677969228124e-06, "loss": 0.7045, "step": 11236 }, { "epoch": 0.9575628461866212, "grad_norm": 15.1875, "learning_rate": 1.3420002179451167e-06, "loss": 0.6337, "step": 11237 }, { "epoch": 0.9576480613549212, "grad_norm": 22.0, "learning_rate": 1.341832637304884e-06, "loss": 0.9191, "step": 11238 }, { "epoch": 0.9577332765232212, "grad_norm": 21.0, "learning_rate": 1.3416650550051438e-06, "loss": 1.2782, "step": 11239 }, { "epoch": 0.9578184916915211, "grad_norm": 17.25, "learning_rate": 1.3414974710489234e-06, "loss": 0.7834, "step": 11240 }, { "epoch": 0.9579037068598211, "grad_norm": 13.0, "learning_rate": 1.341329885439252e-06, "loss": 0.7518, "step": 11241 }, { "epoch": 0.957988922028121, "grad_norm": 10.75, "learning_rate": 1.3411622981791574e-06, "loss": 0.3337, "step": 11242 }, { "epoch": 0.958074137196421, "grad_norm": 16.5, "learning_rate": 1.340994709271669e-06, "loss": 0.6011, "step": 11243 }, { "epoch": 0.9581593523647209, "grad_norm": 14.0, "learning_rate": 1.3408271187198142e-06, "loss": 0.5868, "step": 11244 }, { "epoch": 0.9582445675330209, "grad_norm": 9.5, "learning_rate": 1.340659526526622e-06, "loss": 0.4471, "step": 11245 }, { "epoch": 0.9583297827013209, "grad_norm": 14.4375, "learning_rate": 1.340491932695121e-06, "loss": 0.4826, "step": 11246 }, { "epoch": 0.9584149978696208, "grad_norm": 12.375, "learning_rate": 1.34032433722834e-06, "loss": 0.6259, "step": 11247 }, { "epoch": 0.9585002130379208, "grad_norm": 16.375, "learning_rate": 1.3401567401293075e-06, "loss": 0.9066, "step": 11248 }, { "epoch": 0.9585854282062207, "grad_norm": 26.0, "learning_rate": 1.3399891414010519e-06, "loss": 1.1396, "step": 11249 }, { "epoch": 0.9586706433745207, "grad_norm": 18.75, "learning_rate": 1.3398215410466018e-06, "loss": 1.0114, "step": 11250 }, { "epoch": 0.9587558585428206, "grad_norm": 15.375, "learning_rate": 1.3396539390689863e-06, "loss": 0.9949, "step": 11251 }, { "epoch": 0.9588410737111206, "grad_norm": 13.6875, "learning_rate": 1.339486335471234e-06, "loss": 0.6976, "step": 11252 }, { "epoch": 0.9589262888794206, "grad_norm": 18.75, "learning_rate": 1.3393187302563734e-06, "loss": 0.698, "step": 11253 }, { "epoch": 0.9590115040477205, "grad_norm": 12.3125, "learning_rate": 1.3391511234274337e-06, "loss": 0.8465, "step": 11254 }, { "epoch": 0.9590967192160205, "grad_norm": 14.75, "learning_rate": 1.3389835149874438e-06, "loss": 0.7536, "step": 11255 }, { "epoch": 0.9591819343843204, "grad_norm": 17.25, "learning_rate": 1.3388159049394323e-06, "loss": 1.0718, "step": 11256 }, { "epoch": 0.9592671495526204, "grad_norm": 12.3125, "learning_rate": 1.3386482932864275e-06, "loss": 0.4184, "step": 11257 }, { "epoch": 0.9593523647209203, "grad_norm": 31.125, "learning_rate": 1.3384806800314592e-06, "loss": 0.7978, "step": 11258 }, { "epoch": 0.9594375798892203, "grad_norm": 14.0625, "learning_rate": 1.3383130651775561e-06, "loss": 0.6583, "step": 11259 }, { "epoch": 0.9595227950575203, "grad_norm": 15.625, "learning_rate": 1.3381454487277473e-06, "loss": 0.4511, "step": 11260 }, { "epoch": 0.9596080102258202, "grad_norm": 17.75, "learning_rate": 1.3379778306850614e-06, "loss": 0.8283, "step": 11261 }, { "epoch": 0.9596932253941202, "grad_norm": 11.9375, "learning_rate": 1.3378102110525277e-06, "loss": 0.3871, "step": 11262 }, { "epoch": 0.9597784405624201, "grad_norm": 13.625, "learning_rate": 1.337642589833175e-06, "loss": 0.4974, "step": 11263 }, { "epoch": 0.9598636557307201, "grad_norm": 13.6875, "learning_rate": 1.337474967030033e-06, "loss": 0.8338, "step": 11264 }, { "epoch": 0.95994887089902, "grad_norm": 15.875, "learning_rate": 1.33730734264613e-06, "loss": 0.714, "step": 11265 }, { "epoch": 0.96003408606732, "grad_norm": 14.8125, "learning_rate": 1.3371397166844955e-06, "loss": 1.0014, "step": 11266 }, { "epoch": 0.96011930123562, "grad_norm": 20.875, "learning_rate": 1.3369720891481585e-06, "loss": 1.0857, "step": 11267 }, { "epoch": 0.9602045164039199, "grad_norm": 15.3125, "learning_rate": 1.3368044600401489e-06, "loss": 0.6234, "step": 11268 }, { "epoch": 0.9602897315722199, "grad_norm": 13.125, "learning_rate": 1.3366368293634954e-06, "loss": 0.5609, "step": 11269 }, { "epoch": 0.9603749467405198, "grad_norm": 12.9375, "learning_rate": 1.3364691971212271e-06, "loss": 0.5431, "step": 11270 }, { "epoch": 0.9604601619088198, "grad_norm": 14.5625, "learning_rate": 1.336301563316373e-06, "loss": 0.5444, "step": 11271 }, { "epoch": 0.9605453770771197, "grad_norm": 12.25, "learning_rate": 1.3361339279519635e-06, "loss": 0.6693, "step": 11272 }, { "epoch": 0.9606305922454197, "grad_norm": 17.625, "learning_rate": 1.3359662910310273e-06, "loss": 1.0053, "step": 11273 }, { "epoch": 0.9607158074137196, "grad_norm": 15.0, "learning_rate": 1.3357986525565934e-06, "loss": 0.8832, "step": 11274 }, { "epoch": 0.9608010225820196, "grad_norm": 13.375, "learning_rate": 1.3356310125316917e-06, "loss": 0.6233, "step": 11275 }, { "epoch": 0.9608862377503196, "grad_norm": 13.25, "learning_rate": 1.3354633709593517e-06, "loss": 0.5333, "step": 11276 }, { "epoch": 0.9609714529186195, "grad_norm": 13.1875, "learning_rate": 1.3352957278426026e-06, "loss": 0.7211, "step": 11277 }, { "epoch": 0.9610566680869195, "grad_norm": 14.375, "learning_rate": 1.3351280831844737e-06, "loss": 0.5842, "step": 11278 }, { "epoch": 0.9611418832552194, "grad_norm": 10.3125, "learning_rate": 1.334960436987995e-06, "loss": 0.3551, "step": 11279 }, { "epoch": 0.9612270984235194, "grad_norm": 11.8125, "learning_rate": 1.3347927892561956e-06, "loss": 0.4117, "step": 11280 }, { "epoch": 0.9613123135918193, "grad_norm": 17.0, "learning_rate": 1.3346251399921053e-06, "loss": 1.1344, "step": 11281 }, { "epoch": 0.9613975287601193, "grad_norm": 17.375, "learning_rate": 1.3344574891987535e-06, "loss": 0.9437, "step": 11282 }, { "epoch": 0.9614827439284193, "grad_norm": 20.625, "learning_rate": 1.33428983687917e-06, "loss": 1.0353, "step": 11283 }, { "epoch": 0.9615679590967192, "grad_norm": 16.875, "learning_rate": 1.3341221830363845e-06, "loss": 0.8744, "step": 11284 }, { "epoch": 0.9616531742650192, "grad_norm": 11.9375, "learning_rate": 1.3339545276734267e-06, "loss": 0.3047, "step": 11285 }, { "epoch": 0.9617383894333191, "grad_norm": 19.875, "learning_rate": 1.3337868707933258e-06, "loss": 0.8173, "step": 11286 }, { "epoch": 0.9618236046016191, "grad_norm": 10.375, "learning_rate": 1.333619212399112e-06, "loss": 0.4057, "step": 11287 }, { "epoch": 0.961908819769919, "grad_norm": 12.25, "learning_rate": 1.3334515524938152e-06, "loss": 0.4672, "step": 11288 }, { "epoch": 0.961994034938219, "grad_norm": 14.4375, "learning_rate": 1.3332838910804644e-06, "loss": 0.5211, "step": 11289 }, { "epoch": 0.962079250106519, "grad_norm": 17.375, "learning_rate": 1.3331162281620905e-06, "loss": 0.4687, "step": 11290 }, { "epoch": 0.9621644652748189, "grad_norm": 12.8125, "learning_rate": 1.3329485637417223e-06, "loss": 0.6322, "step": 11291 }, { "epoch": 0.9622496804431189, "grad_norm": 12.8125, "learning_rate": 1.3327808978223905e-06, "loss": 0.7581, "step": 11292 }, { "epoch": 0.9623348956114188, "grad_norm": 16.125, "learning_rate": 1.3326132304071244e-06, "loss": 0.846, "step": 11293 }, { "epoch": 0.9624201107797188, "grad_norm": 12.125, "learning_rate": 1.3324455614989542e-06, "loss": 0.534, "step": 11294 }, { "epoch": 0.9625053259480187, "grad_norm": 14.5, "learning_rate": 1.3322778911009099e-06, "loss": 0.693, "step": 11295 }, { "epoch": 0.9625905411163187, "grad_norm": 14.125, "learning_rate": 1.3321102192160216e-06, "loss": 0.5275, "step": 11296 }, { "epoch": 0.9626757562846187, "grad_norm": 18.25, "learning_rate": 1.3319425458473188e-06, "loss": 1.0927, "step": 11297 }, { "epoch": 0.9627609714529186, "grad_norm": 34.25, "learning_rate": 1.3317748709978317e-06, "loss": 0.9722, "step": 11298 }, { "epoch": 0.9628461866212186, "grad_norm": 16.0, "learning_rate": 1.3316071946705902e-06, "loss": 1.2774, "step": 11299 }, { "epoch": 0.9629314017895185, "grad_norm": 13.375, "learning_rate": 1.3314395168686256e-06, "loss": 0.7155, "step": 11300 }, { "epoch": 0.9630166169578185, "grad_norm": 13.3125, "learning_rate": 1.331271837594966e-06, "loss": 0.7423, "step": 11301 }, { "epoch": 0.9631018321261184, "grad_norm": 15.1875, "learning_rate": 1.3311041568526432e-06, "loss": 0.8511, "step": 11302 }, { "epoch": 0.9631870472944184, "grad_norm": 18.375, "learning_rate": 1.3309364746446863e-06, "loss": 1.085, "step": 11303 }, { "epoch": 0.9632722624627184, "grad_norm": 12.0, "learning_rate": 1.3307687909741263e-06, "loss": 0.4792, "step": 11304 }, { "epoch": 0.9633574776310183, "grad_norm": 13.0, "learning_rate": 1.3306011058439926e-06, "loss": 0.7825, "step": 11305 }, { "epoch": 0.9634426927993183, "grad_norm": 14.4375, "learning_rate": 1.3304334192573163e-06, "loss": 0.6797, "step": 11306 }, { "epoch": 0.9635279079676182, "grad_norm": 13.5, "learning_rate": 1.3302657312171268e-06, "loss": 0.649, "step": 11307 }, { "epoch": 0.9636131231359182, "grad_norm": 14.5625, "learning_rate": 1.3300980417264552e-06, "loss": 0.5308, "step": 11308 }, { "epoch": 0.9636983383042181, "grad_norm": 14.5, "learning_rate": 1.3299303507883316e-06, "loss": 0.7404, "step": 11309 }, { "epoch": 0.9637835534725181, "grad_norm": 15.8125, "learning_rate": 1.3297626584057853e-06, "loss": 0.9554, "step": 11310 }, { "epoch": 0.963868768640818, "grad_norm": 21.5, "learning_rate": 1.3295949645818484e-06, "loss": 0.9217, "step": 11311 }, { "epoch": 0.963953983809118, "grad_norm": 15.8125, "learning_rate": 1.3294272693195504e-06, "loss": 0.9075, "step": 11312 }, { "epoch": 0.964039198977418, "grad_norm": 16.75, "learning_rate": 1.3292595726219215e-06, "loss": 0.753, "step": 11313 }, { "epoch": 0.9641244141457179, "grad_norm": 18.375, "learning_rate": 1.3290918744919926e-06, "loss": 0.8299, "step": 11314 }, { "epoch": 0.9642096293140179, "grad_norm": 15.5625, "learning_rate": 1.3289241749327939e-06, "loss": 0.7705, "step": 11315 }, { "epoch": 0.9642948444823178, "grad_norm": 14.5, "learning_rate": 1.3287564739473558e-06, "loss": 0.4943, "step": 11316 }, { "epoch": 0.9643800596506178, "grad_norm": 11.5625, "learning_rate": 1.3285887715387095e-06, "loss": 0.2655, "step": 11317 }, { "epoch": 0.9644652748189178, "grad_norm": 16.0, "learning_rate": 1.3284210677098847e-06, "loss": 0.8507, "step": 11318 }, { "epoch": 0.9645504899872177, "grad_norm": 12.75, "learning_rate": 1.3282533624639124e-06, "loss": 0.6636, "step": 11319 }, { "epoch": 0.9646357051555177, "grad_norm": 14.8125, "learning_rate": 1.328085655803823e-06, "loss": 0.8428, "step": 11320 }, { "epoch": 0.9647209203238176, "grad_norm": 14.8125, "learning_rate": 1.3279179477326475e-06, "loss": 0.9147, "step": 11321 }, { "epoch": 0.9648061354921176, "grad_norm": 14.5625, "learning_rate": 1.3277502382534163e-06, "loss": 0.7329, "step": 11322 }, { "epoch": 0.9648913506604175, "grad_norm": 14.4375, "learning_rate": 1.3275825273691603e-06, "loss": 0.5594, "step": 11323 }, { "epoch": 0.9649765658287175, "grad_norm": 10.75, "learning_rate": 1.3274148150829098e-06, "loss": 0.5454, "step": 11324 }, { "epoch": 0.9650617809970174, "grad_norm": 20.5, "learning_rate": 1.3272471013976956e-06, "loss": 1.0577, "step": 11325 }, { "epoch": 0.9651469961653174, "grad_norm": 19.25, "learning_rate": 1.3270793863165489e-06, "loss": 0.8825, "step": 11326 }, { "epoch": 0.9652322113336174, "grad_norm": 13.8125, "learning_rate": 1.3269116698424997e-06, "loss": 0.6828, "step": 11327 }, { "epoch": 0.9653174265019173, "grad_norm": 18.375, "learning_rate": 1.32674395197858e-06, "loss": 0.714, "step": 11328 }, { "epoch": 0.9654026416702173, "grad_norm": 15.125, "learning_rate": 1.3265762327278195e-06, "loss": 0.682, "step": 11329 }, { "epoch": 0.9654878568385172, "grad_norm": 12.3125, "learning_rate": 1.3264085120932495e-06, "loss": 0.6552, "step": 11330 }, { "epoch": 0.9655730720068172, "grad_norm": 13.75, "learning_rate": 1.3262407900779006e-06, "loss": 0.4089, "step": 11331 }, { "epoch": 0.9656582871751171, "grad_norm": 14.9375, "learning_rate": 1.3260730666848043e-06, "loss": 0.5968, "step": 11332 }, { "epoch": 0.9657435023434171, "grad_norm": 12.4375, "learning_rate": 1.3259053419169912e-06, "loss": 0.6452, "step": 11333 }, { "epoch": 0.9658287175117171, "grad_norm": 12.0625, "learning_rate": 1.3257376157774924e-06, "loss": 0.3799, "step": 11334 }, { "epoch": 0.965913932680017, "grad_norm": 18.625, "learning_rate": 1.3255698882693382e-06, "loss": 1.0815, "step": 11335 }, { "epoch": 0.965999147848317, "grad_norm": 16.125, "learning_rate": 1.3254021593955605e-06, "loss": 1.021, "step": 11336 }, { "epoch": 0.9660843630166169, "grad_norm": 17.0, "learning_rate": 1.3252344291591898e-06, "loss": 0.9381, "step": 11337 }, { "epoch": 0.9661695781849169, "grad_norm": 11.5, "learning_rate": 1.3250666975632577e-06, "loss": 0.4586, "step": 11338 }, { "epoch": 0.9662547933532168, "grad_norm": 21.125, "learning_rate": 1.3248989646107943e-06, "loss": 1.1198, "step": 11339 }, { "epoch": 0.9663400085215168, "grad_norm": 13.75, "learning_rate": 1.3247312303048315e-06, "loss": 0.5648, "step": 11340 }, { "epoch": 0.9664252236898168, "grad_norm": 14.75, "learning_rate": 1.3245634946484005e-06, "loss": 0.7224, "step": 11341 }, { "epoch": 0.9665104388581167, "grad_norm": 16.125, "learning_rate": 1.3243957576445314e-06, "loss": 0.7446, "step": 11342 }, { "epoch": 0.9665956540264167, "grad_norm": 12.125, "learning_rate": 1.324228019296257e-06, "loss": 0.5621, "step": 11343 }, { "epoch": 0.9666808691947166, "grad_norm": 13.5625, "learning_rate": 1.324060279606607e-06, "loss": 0.7114, "step": 11344 }, { "epoch": 0.9667660843630166, "grad_norm": 14.0625, "learning_rate": 1.3238925385786141e-06, "loss": 0.8529, "step": 11345 }, { "epoch": 0.9668512995313165, "grad_norm": 13.0, "learning_rate": 1.323724796215308e-06, "loss": 0.6608, "step": 11346 }, { "epoch": 0.9669365146996165, "grad_norm": 11.8125, "learning_rate": 1.3235570525197212e-06, "loss": 0.3913, "step": 11347 }, { "epoch": 0.9670217298679165, "grad_norm": 25.0, "learning_rate": 1.3233893074948838e-06, "loss": 0.9876, "step": 11348 }, { "epoch": 0.9671069450362164, "grad_norm": 21.125, "learning_rate": 1.3232215611438288e-06, "loss": 0.9292, "step": 11349 }, { "epoch": 0.9671921602045164, "grad_norm": 13.375, "learning_rate": 1.323053813469586e-06, "loss": 0.692, "step": 11350 }, { "epoch": 0.9672773753728163, "grad_norm": 13.3125, "learning_rate": 1.322886064475187e-06, "loss": 0.8455, "step": 11351 }, { "epoch": 0.9673625905411163, "grad_norm": 14.125, "learning_rate": 1.3227183141636637e-06, "loss": 0.7569, "step": 11352 }, { "epoch": 0.9674478057094162, "grad_norm": 12.75, "learning_rate": 1.3225505625380481e-06, "loss": 0.6274, "step": 11353 }, { "epoch": 0.9675330208777162, "grad_norm": 16.125, "learning_rate": 1.32238280960137e-06, "loss": 0.6565, "step": 11354 }, { "epoch": 0.9676182360460162, "grad_norm": 13.3125, "learning_rate": 1.322215055356662e-06, "loss": 0.7698, "step": 11355 }, { "epoch": 0.9677034512143161, "grad_norm": 17.0, "learning_rate": 1.3220472998069553e-06, "loss": 0.9246, "step": 11356 }, { "epoch": 0.9677886663826161, "grad_norm": 12.5625, "learning_rate": 1.3218795429552813e-06, "loss": 0.6619, "step": 11357 }, { "epoch": 0.967873881550916, "grad_norm": 15.0625, "learning_rate": 1.3217117848046721e-06, "loss": 0.9119, "step": 11358 }, { "epoch": 0.967959096719216, "grad_norm": 14.9375, "learning_rate": 1.3215440253581585e-06, "loss": 0.796, "step": 11359 }, { "epoch": 0.9680443118875159, "grad_norm": 15.75, "learning_rate": 1.3213762646187724e-06, "loss": 0.7491, "step": 11360 }, { "epoch": 0.9681295270558159, "grad_norm": 13.375, "learning_rate": 1.3212085025895458e-06, "loss": 0.674, "step": 11361 }, { "epoch": 0.9682147422241159, "grad_norm": 17.125, "learning_rate": 1.3210407392735097e-06, "loss": 0.5801, "step": 11362 }, { "epoch": 0.9682999573924158, "grad_norm": 20.375, "learning_rate": 1.3208729746736958e-06, "loss": 0.9625, "step": 11363 }, { "epoch": 0.9683851725607158, "grad_norm": 18.25, "learning_rate": 1.3207052087931362e-06, "loss": 0.8111, "step": 11364 }, { "epoch": 0.9684703877290157, "grad_norm": 11.375, "learning_rate": 1.3205374416348627e-06, "loss": 0.5216, "step": 11365 }, { "epoch": 0.9685556028973157, "grad_norm": 14.5, "learning_rate": 1.3203696732019065e-06, "loss": 0.7154, "step": 11366 }, { "epoch": 0.9686408180656156, "grad_norm": 19.75, "learning_rate": 1.320201903497299e-06, "loss": 0.8924, "step": 11367 }, { "epoch": 0.9687260332339156, "grad_norm": 11.1875, "learning_rate": 1.320034132524073e-06, "loss": 0.4112, "step": 11368 }, { "epoch": 0.9688112484022156, "grad_norm": 11.875, "learning_rate": 1.31986636028526e-06, "loss": 0.5414, "step": 11369 }, { "epoch": 0.9688964635705155, "grad_norm": 16.0, "learning_rate": 1.3196985867838913e-06, "loss": 0.7155, "step": 11370 }, { "epoch": 0.9689816787388155, "grad_norm": 11.9375, "learning_rate": 1.3195308120229988e-06, "loss": 0.3473, "step": 11371 }, { "epoch": 0.9690668939071154, "grad_norm": 15.0625, "learning_rate": 1.319363036005615e-06, "loss": 0.8222, "step": 11372 }, { "epoch": 0.9691521090754154, "grad_norm": 20.375, "learning_rate": 1.3191952587347714e-06, "loss": 0.978, "step": 11373 }, { "epoch": 0.9692373242437153, "grad_norm": 18.0, "learning_rate": 1.3190274802134995e-06, "loss": 0.6053, "step": 11374 }, { "epoch": 0.9693225394120153, "grad_norm": 22.125, "learning_rate": 1.318859700444832e-06, "loss": 0.9879, "step": 11375 }, { "epoch": 0.9694077545803153, "grad_norm": 14.5625, "learning_rate": 1.3186919194318002e-06, "loss": 0.8097, "step": 11376 }, { "epoch": 0.9694929697486152, "grad_norm": 14.625, "learning_rate": 1.3185241371774368e-06, "loss": 1.3046, "step": 11377 }, { "epoch": 0.9695781849169152, "grad_norm": 15.0625, "learning_rate": 1.3183563536847729e-06, "loss": 0.7367, "step": 11378 }, { "epoch": 0.9696634000852151, "grad_norm": 20.5, "learning_rate": 1.318188568956841e-06, "loss": 0.8635, "step": 11379 }, { "epoch": 0.9697486152535151, "grad_norm": 16.0, "learning_rate": 1.318020782996673e-06, "loss": 0.8515, "step": 11380 }, { "epoch": 0.969833830421815, "grad_norm": 14.625, "learning_rate": 1.3178529958073014e-06, "loss": 0.8373, "step": 11381 }, { "epoch": 0.969919045590115, "grad_norm": 13.75, "learning_rate": 1.317685207391758e-06, "loss": 0.7394, "step": 11382 }, { "epoch": 0.970004260758415, "grad_norm": 11.75, "learning_rate": 1.3175174177530744e-06, "loss": 0.5973, "step": 11383 }, { "epoch": 0.9700894759267149, "grad_norm": 15.0, "learning_rate": 1.3173496268942835e-06, "loss": 0.6504, "step": 11384 }, { "epoch": 0.9701746910950149, "grad_norm": 15.375, "learning_rate": 1.3171818348184168e-06, "loss": 0.7858, "step": 11385 }, { "epoch": 0.9702599062633148, "grad_norm": 17.5, "learning_rate": 1.3170140415285071e-06, "loss": 0.843, "step": 11386 }, { "epoch": 0.9703451214316148, "grad_norm": 13.8125, "learning_rate": 1.3168462470275864e-06, "loss": 0.494, "step": 11387 }, { "epoch": 0.9704303365999147, "grad_norm": 12.1875, "learning_rate": 1.316678451318686e-06, "loss": 0.4649, "step": 11388 }, { "epoch": 0.9705155517682147, "grad_norm": 15.6875, "learning_rate": 1.3165106544048397e-06, "loss": 0.7248, "step": 11389 }, { "epoch": 0.9706007669365146, "grad_norm": 19.25, "learning_rate": 1.316342856289079e-06, "loss": 0.9645, "step": 11390 }, { "epoch": 0.9706859821048146, "grad_norm": 10.25, "learning_rate": 1.3161750569744357e-06, "loss": 0.6843, "step": 11391 }, { "epoch": 0.9707711972731146, "grad_norm": 13.3125, "learning_rate": 1.3160072564639429e-06, "loss": 0.4417, "step": 11392 }, { "epoch": 0.9708564124414145, "grad_norm": 20.375, "learning_rate": 1.3158394547606326e-06, "loss": 0.8439, "step": 11393 }, { "epoch": 0.9709416276097145, "grad_norm": 11.375, "learning_rate": 1.3156716518675372e-06, "loss": 0.4711, "step": 11394 }, { "epoch": 0.9710268427780144, "grad_norm": 11.4375, "learning_rate": 1.3155038477876886e-06, "loss": 0.4872, "step": 11395 }, { "epoch": 0.9711120579463144, "grad_norm": 14.6875, "learning_rate": 1.31533604252412e-06, "loss": 0.7636, "step": 11396 }, { "epoch": 0.9711972731146145, "grad_norm": 14.5, "learning_rate": 1.3151682360798632e-06, "loss": 0.5719, "step": 11397 }, { "epoch": 0.9712824882829144, "grad_norm": 13.125, "learning_rate": 1.3150004284579513e-06, "loss": 0.402, "step": 11398 }, { "epoch": 0.9713677034512144, "grad_norm": 15.1875, "learning_rate": 1.3148326196614158e-06, "loss": 0.7881, "step": 11399 }, { "epoch": 0.9714529186195143, "grad_norm": 17.5, "learning_rate": 1.31466480969329e-06, "loss": 0.8346, "step": 11400 }, { "epoch": 0.9715381337878143, "grad_norm": 15.125, "learning_rate": 1.3144969985566057e-06, "loss": 0.9679, "step": 11401 }, { "epoch": 0.9716233489561142, "grad_norm": 25.625, "learning_rate": 1.314329186254396e-06, "loss": 0.7881, "step": 11402 }, { "epoch": 0.9717085641244142, "grad_norm": 12.6875, "learning_rate": 1.3141613727896932e-06, "loss": 0.4283, "step": 11403 }, { "epoch": 0.9717937792927142, "grad_norm": 21.0, "learning_rate": 1.3139935581655297e-06, "loss": 1.1608, "step": 11404 }, { "epoch": 0.9718789944610141, "grad_norm": 12.8125, "learning_rate": 1.3138257423849382e-06, "loss": 0.7783, "step": 11405 }, { "epoch": 0.9719642096293141, "grad_norm": 20.875, "learning_rate": 1.3136579254509518e-06, "loss": 0.8851, "step": 11406 }, { "epoch": 0.972049424797614, "grad_norm": 15.4375, "learning_rate": 1.3134901073666022e-06, "loss": 0.5869, "step": 11407 }, { "epoch": 0.972134639965914, "grad_norm": 12.875, "learning_rate": 1.313322288134923e-06, "loss": 0.7989, "step": 11408 }, { "epoch": 0.9722198551342139, "grad_norm": 14.75, "learning_rate": 1.313154467758946e-06, "loss": 0.6269, "step": 11409 }, { "epoch": 0.9723050703025139, "grad_norm": 25.375, "learning_rate": 1.3129866462417047e-06, "loss": 1.1742, "step": 11410 }, { "epoch": 0.9723902854708139, "grad_norm": 13.125, "learning_rate": 1.3128188235862308e-06, "loss": 0.5059, "step": 11411 }, { "epoch": 0.9724755006391138, "grad_norm": 19.875, "learning_rate": 1.3126509997955579e-06, "loss": 1.0356, "step": 11412 }, { "epoch": 0.9725607158074138, "grad_norm": 14.9375, "learning_rate": 1.3124831748727183e-06, "loss": 0.7453, "step": 11413 }, { "epoch": 0.9726459309757137, "grad_norm": 14.875, "learning_rate": 1.312315348820745e-06, "loss": 0.6555, "step": 11414 }, { "epoch": 0.9727311461440137, "grad_norm": 14.9375, "learning_rate": 1.3121475216426708e-06, "loss": 0.7213, "step": 11415 }, { "epoch": 0.9728163613123136, "grad_norm": 14.3125, "learning_rate": 1.311979693341528e-06, "loss": 0.7463, "step": 11416 }, { "epoch": 0.9729015764806136, "grad_norm": 11.9375, "learning_rate": 1.31181186392035e-06, "loss": 0.66, "step": 11417 }, { "epoch": 0.9729867916489136, "grad_norm": 15.6875, "learning_rate": 1.3116440333821699e-06, "loss": 0.9052, "step": 11418 }, { "epoch": 0.9730720068172135, "grad_norm": 16.5, "learning_rate": 1.3114762017300194e-06, "loss": 0.9488, "step": 11419 }, { "epoch": 0.9731572219855135, "grad_norm": 14.5, "learning_rate": 1.3113083689669325e-06, "loss": 0.7577, "step": 11420 }, { "epoch": 0.9732424371538134, "grad_norm": 11.125, "learning_rate": 1.3111405350959417e-06, "loss": 0.3497, "step": 11421 }, { "epoch": 0.9733276523221134, "grad_norm": 16.5, "learning_rate": 1.3109727001200801e-06, "loss": 0.622, "step": 11422 }, { "epoch": 0.9734128674904133, "grad_norm": 14.5, "learning_rate": 1.3108048640423803e-06, "loss": 0.5931, "step": 11423 }, { "epoch": 0.9734980826587133, "grad_norm": 17.125, "learning_rate": 1.3106370268658752e-06, "loss": 0.8682, "step": 11424 }, { "epoch": 0.9735832978270132, "grad_norm": 13.1875, "learning_rate": 1.3104691885935986e-06, "loss": 0.5822, "step": 11425 }, { "epoch": 0.9736685129953132, "grad_norm": 27.25, "learning_rate": 1.3103013492285827e-06, "loss": 1.0992, "step": 11426 }, { "epoch": 0.9737537281636132, "grad_norm": 13.5625, "learning_rate": 1.3101335087738607e-06, "loss": 0.7458, "step": 11427 }, { "epoch": 0.9738389433319131, "grad_norm": 11.5625, "learning_rate": 1.3099656672324659e-06, "loss": 0.6046, "step": 11428 }, { "epoch": 0.9739241585002131, "grad_norm": 15.4375, "learning_rate": 1.309797824607431e-06, "loss": 0.9732, "step": 11429 }, { "epoch": 0.974009373668513, "grad_norm": 13.0625, "learning_rate": 1.3096299809017896e-06, "loss": 0.8161, "step": 11430 }, { "epoch": 0.974094588836813, "grad_norm": 16.375, "learning_rate": 1.3094621361185738e-06, "loss": 0.5617, "step": 11431 }, { "epoch": 0.974179804005113, "grad_norm": 12.4375, "learning_rate": 1.309294290260818e-06, "loss": 0.6842, "step": 11432 }, { "epoch": 0.9742650191734129, "grad_norm": 15.8125, "learning_rate": 1.3091264433315545e-06, "loss": 0.8921, "step": 11433 }, { "epoch": 0.9743502343417129, "grad_norm": 11.75, "learning_rate": 1.308958595333817e-06, "loss": 0.4538, "step": 11434 }, { "epoch": 0.9744354495100128, "grad_norm": 13.25, "learning_rate": 1.3087907462706384e-06, "loss": 0.5017, "step": 11435 }, { "epoch": 0.9745206646783128, "grad_norm": 13.5625, "learning_rate": 1.3086228961450517e-06, "loss": 0.928, "step": 11436 }, { "epoch": 0.9746058798466127, "grad_norm": 18.875, "learning_rate": 1.3084550449600903e-06, "loss": 0.677, "step": 11437 }, { "epoch": 0.9746910950149127, "grad_norm": 16.0, "learning_rate": 1.3082871927187877e-06, "loss": 0.6583, "step": 11438 }, { "epoch": 0.9747763101832126, "grad_norm": 13.8125, "learning_rate": 1.3081193394241765e-06, "loss": 0.7183, "step": 11439 }, { "epoch": 0.9748615253515126, "grad_norm": 16.5, "learning_rate": 1.3079514850792906e-06, "loss": 0.9364, "step": 11440 }, { "epoch": 0.9749467405198126, "grad_norm": 11.5, "learning_rate": 1.307783629687163e-06, "loss": 0.369, "step": 11441 }, { "epoch": 0.9750319556881125, "grad_norm": 15.0, "learning_rate": 1.3076157732508271e-06, "loss": 0.5203, "step": 11442 }, { "epoch": 0.9751171708564125, "grad_norm": 21.625, "learning_rate": 1.3074479157733165e-06, "loss": 0.7108, "step": 11443 }, { "epoch": 0.9752023860247124, "grad_norm": 14.0, "learning_rate": 1.3072800572576638e-06, "loss": 0.8176, "step": 11444 }, { "epoch": 0.9752876011930124, "grad_norm": 18.75, "learning_rate": 1.3071121977069035e-06, "loss": 0.8551, "step": 11445 }, { "epoch": 0.9753728163613123, "grad_norm": 11.5625, "learning_rate": 1.306944337124068e-06, "loss": 0.492, "step": 11446 }, { "epoch": 0.9754580315296123, "grad_norm": 11.625, "learning_rate": 1.3067764755121911e-06, "loss": 0.4828, "step": 11447 }, { "epoch": 0.9755432466979123, "grad_norm": 13.0, "learning_rate": 1.3066086128743062e-06, "loss": 0.5425, "step": 11448 }, { "epoch": 0.9756284618662122, "grad_norm": 16.375, "learning_rate": 1.3064407492134468e-06, "loss": 0.8659, "step": 11449 }, { "epoch": 0.9757136770345122, "grad_norm": 12.3125, "learning_rate": 1.306272884532646e-06, "loss": 0.4714, "step": 11450 }, { "epoch": 0.9757988922028121, "grad_norm": 16.125, "learning_rate": 1.3061050188349382e-06, "loss": 0.7596, "step": 11451 }, { "epoch": 0.9758841073711121, "grad_norm": 14.0, "learning_rate": 1.3059371521233557e-06, "loss": 0.8078, "step": 11452 }, { "epoch": 0.975969322539412, "grad_norm": 11.3125, "learning_rate": 1.3057692844009331e-06, "loss": 0.4329, "step": 11453 }, { "epoch": 0.976054537707712, "grad_norm": 14.0, "learning_rate": 1.3056014156707031e-06, "loss": 0.8311, "step": 11454 }, { "epoch": 0.976139752876012, "grad_norm": 14.875, "learning_rate": 1.3054335459356998e-06, "loss": 0.486, "step": 11455 }, { "epoch": 0.9762249680443119, "grad_norm": 13.625, "learning_rate": 1.3052656751989565e-06, "loss": 0.5052, "step": 11456 }, { "epoch": 0.9763101832126119, "grad_norm": 22.125, "learning_rate": 1.305097803463507e-06, "loss": 0.9126, "step": 11457 }, { "epoch": 0.9763953983809118, "grad_norm": 13.6875, "learning_rate": 1.3049299307323842e-06, "loss": 0.7523, "step": 11458 }, { "epoch": 0.9764806135492118, "grad_norm": 23.125, "learning_rate": 1.304762057008623e-06, "loss": 0.6702, "step": 11459 }, { "epoch": 0.9765658287175117, "grad_norm": 16.375, "learning_rate": 1.3045941822952562e-06, "loss": 0.7718, "step": 11460 }, { "epoch": 0.9766510438858117, "grad_norm": 14.375, "learning_rate": 1.3044263065953173e-06, "loss": 0.8307, "step": 11461 }, { "epoch": 0.9767362590541117, "grad_norm": 15.6875, "learning_rate": 1.3042584299118407e-06, "loss": 0.5395, "step": 11462 }, { "epoch": 0.9768214742224116, "grad_norm": 11.3125, "learning_rate": 1.3040905522478598e-06, "loss": 0.4624, "step": 11463 }, { "epoch": 0.9769066893907116, "grad_norm": 15.0625, "learning_rate": 1.303922673606408e-06, "loss": 0.7776, "step": 11464 }, { "epoch": 0.9769919045590115, "grad_norm": 13.3125, "learning_rate": 1.303754793990519e-06, "loss": 0.6031, "step": 11465 }, { "epoch": 0.9770771197273115, "grad_norm": 20.0, "learning_rate": 1.3035869134032272e-06, "loss": 1.0215, "step": 11466 }, { "epoch": 0.9771623348956114, "grad_norm": 20.375, "learning_rate": 1.303419031847566e-06, "loss": 0.8326, "step": 11467 }, { "epoch": 0.9772475500639114, "grad_norm": 16.75, "learning_rate": 1.303251149326569e-06, "loss": 0.9276, "step": 11468 }, { "epoch": 0.9773327652322114, "grad_norm": 14.375, "learning_rate": 1.3030832658432702e-06, "loss": 0.7195, "step": 11469 }, { "epoch": 0.9774179804005113, "grad_norm": 14.0, "learning_rate": 1.3029153814007034e-06, "loss": 0.6745, "step": 11470 }, { "epoch": 0.9775031955688113, "grad_norm": 10.375, "learning_rate": 1.3027474960019026e-06, "loss": 0.3661, "step": 11471 }, { "epoch": 0.9775884107371112, "grad_norm": 15.1875, "learning_rate": 1.3025796096499016e-06, "loss": 0.5098, "step": 11472 }, { "epoch": 0.9776736259054112, "grad_norm": 14.9375, "learning_rate": 1.3024117223477337e-06, "loss": 0.7546, "step": 11473 }, { "epoch": 0.9777588410737111, "grad_norm": 19.375, "learning_rate": 1.3022438340984337e-06, "loss": 0.9314, "step": 11474 }, { "epoch": 0.9778440562420111, "grad_norm": 13.3125, "learning_rate": 1.302075944905035e-06, "loss": 0.6504, "step": 11475 }, { "epoch": 0.977929271410311, "grad_norm": 14.6875, "learning_rate": 1.3019080547705713e-06, "loss": 0.478, "step": 11476 }, { "epoch": 0.978014486578611, "grad_norm": 13.75, "learning_rate": 1.3017401636980772e-06, "loss": 0.4074, "step": 11477 }, { "epoch": 0.978099701746911, "grad_norm": 15.75, "learning_rate": 1.3015722716905862e-06, "loss": 1.022, "step": 11478 }, { "epoch": 0.9781849169152109, "grad_norm": 11.3125, "learning_rate": 1.3014043787511327e-06, "loss": 0.4357, "step": 11479 }, { "epoch": 0.9782701320835109, "grad_norm": 13.5, "learning_rate": 1.3012364848827497e-06, "loss": 0.9168, "step": 11480 }, { "epoch": 0.9783553472518108, "grad_norm": 15.125, "learning_rate": 1.3010685900884723e-06, "loss": 0.853, "step": 11481 }, { "epoch": 0.9784405624201108, "grad_norm": 26.375, "learning_rate": 1.300900694371334e-06, "loss": 0.9446, "step": 11482 }, { "epoch": 0.9785257775884108, "grad_norm": 18.625, "learning_rate": 1.3007327977343695e-06, "loss": 0.9474, "step": 11483 }, { "epoch": 0.9786109927567107, "grad_norm": 18.625, "learning_rate": 1.3005649001806118e-06, "loss": 0.8374, "step": 11484 }, { "epoch": 0.9786962079250107, "grad_norm": 14.5625, "learning_rate": 1.3003970017130958e-06, "loss": 0.4643, "step": 11485 }, { "epoch": 0.9787814230933106, "grad_norm": 17.75, "learning_rate": 1.3002291023348545e-06, "loss": 0.6401, "step": 11486 }, { "epoch": 0.9788666382616106, "grad_norm": 11.875, "learning_rate": 1.300061202048924e-06, "loss": 0.5251, "step": 11487 }, { "epoch": 0.9789518534299105, "grad_norm": 20.0, "learning_rate": 1.2998933008583365e-06, "loss": 1.1831, "step": 11488 }, { "epoch": 0.9790370685982105, "grad_norm": 14.5625, "learning_rate": 1.2997253987661272e-06, "loss": 0.9766, "step": 11489 }, { "epoch": 0.9791222837665104, "grad_norm": 22.375, "learning_rate": 1.2995574957753294e-06, "loss": 1.1832, "step": 11490 }, { "epoch": 0.9792074989348104, "grad_norm": 18.375, "learning_rate": 1.2993895918889783e-06, "loss": 0.8554, "step": 11491 }, { "epoch": 0.9792927141031104, "grad_norm": 16.25, "learning_rate": 1.2992216871101077e-06, "loss": 0.9373, "step": 11492 }, { "epoch": 0.9793779292714103, "grad_norm": 13.125, "learning_rate": 1.2990537814417513e-06, "loss": 0.8313, "step": 11493 }, { "epoch": 0.9794631444397103, "grad_norm": 14.0, "learning_rate": 1.298885874886944e-06, "loss": 0.8489, "step": 11494 }, { "epoch": 0.9795483596080102, "grad_norm": 15.375, "learning_rate": 1.29871796744872e-06, "loss": 0.978, "step": 11495 }, { "epoch": 0.9796335747763102, "grad_norm": 13.5625, "learning_rate": 1.298550059130113e-06, "loss": 0.6944, "step": 11496 }, { "epoch": 0.9797187899446101, "grad_norm": 15.75, "learning_rate": 1.2983821499341576e-06, "loss": 1.0332, "step": 11497 }, { "epoch": 0.9798040051129101, "grad_norm": 23.875, "learning_rate": 1.298214239863888e-06, "loss": 1.2074, "step": 11498 }, { "epoch": 0.9798892202812101, "grad_norm": 38.75, "learning_rate": 1.298046328922339e-06, "loss": 1.1794, "step": 11499 }, { "epoch": 0.97997443544951, "grad_norm": 11.0, "learning_rate": 1.2978784171125445e-06, "loss": 0.5342, "step": 11500 }, { "epoch": 0.98005965061781, "grad_norm": 17.875, "learning_rate": 1.2977105044375385e-06, "loss": 0.836, "step": 11501 }, { "epoch": 0.9801448657861099, "grad_norm": 16.375, "learning_rate": 1.2975425909003558e-06, "loss": 0.6023, "step": 11502 }, { "epoch": 0.9802300809544099, "grad_norm": 13.375, "learning_rate": 1.297374676504031e-06, "loss": 0.4707, "step": 11503 }, { "epoch": 0.9803152961227098, "grad_norm": 12.0, "learning_rate": 1.297206761251598e-06, "loss": 0.471, "step": 11504 }, { "epoch": 0.9804005112910098, "grad_norm": 16.25, "learning_rate": 1.2970388451460912e-06, "loss": 0.6333, "step": 11505 }, { "epoch": 0.9804857264593098, "grad_norm": 14.1875, "learning_rate": 1.296870928190545e-06, "loss": 0.7306, "step": 11506 }, { "epoch": 0.9805709416276097, "grad_norm": 19.25, "learning_rate": 1.2967030103879946e-06, "loss": 1.1319, "step": 11507 }, { "epoch": 0.9806561567959097, "grad_norm": 13.6875, "learning_rate": 1.2965350917414735e-06, "loss": 0.795, "step": 11508 }, { "epoch": 0.9807413719642096, "grad_norm": 14.8125, "learning_rate": 1.2963671722540164e-06, "loss": 0.8793, "step": 11509 }, { "epoch": 0.9808265871325096, "grad_norm": 11.875, "learning_rate": 1.2961992519286578e-06, "loss": 0.3831, "step": 11510 }, { "epoch": 0.9809118023008095, "grad_norm": 14.0, "learning_rate": 1.2960313307684324e-06, "loss": 0.8376, "step": 11511 }, { "epoch": 0.9809970174691095, "grad_norm": 16.125, "learning_rate": 1.2958634087763747e-06, "loss": 0.8669, "step": 11512 }, { "epoch": 0.9810822326374095, "grad_norm": 17.5, "learning_rate": 1.2956954859555188e-06, "loss": 0.8526, "step": 11513 }, { "epoch": 0.9811674478057094, "grad_norm": 14.5, "learning_rate": 1.2955275623088995e-06, "loss": 0.6765, "step": 11514 }, { "epoch": 0.9812526629740094, "grad_norm": 16.875, "learning_rate": 1.2953596378395517e-06, "loss": 1.0282, "step": 11515 }, { "epoch": 0.9813378781423093, "grad_norm": 19.0, "learning_rate": 1.2951917125505095e-06, "loss": 0.6992, "step": 11516 }, { "epoch": 0.9814230933106093, "grad_norm": 15.625, "learning_rate": 1.2950237864448077e-06, "loss": 0.7495, "step": 11517 }, { "epoch": 0.9815083084789092, "grad_norm": 13.0, "learning_rate": 1.2948558595254803e-06, "loss": 0.7088, "step": 11518 }, { "epoch": 0.9815935236472092, "grad_norm": 12.8125, "learning_rate": 1.294687931795563e-06, "loss": 0.5023, "step": 11519 }, { "epoch": 0.9816787388155092, "grad_norm": 15.0, "learning_rate": 1.2945200032580898e-06, "loss": 0.812, "step": 11520 }, { "epoch": 0.9817639539838091, "grad_norm": 16.375, "learning_rate": 1.2943520739160953e-06, "loss": 0.9937, "step": 11521 }, { "epoch": 0.9818491691521091, "grad_norm": 18.5, "learning_rate": 1.294184143772614e-06, "loss": 1.0094, "step": 11522 }, { "epoch": 0.981934384320409, "grad_norm": 13.5625, "learning_rate": 1.294016212830681e-06, "loss": 0.5905, "step": 11523 }, { "epoch": 0.982019599488709, "grad_norm": 14.5, "learning_rate": 1.2938482810933307e-06, "loss": 0.8264, "step": 11524 }, { "epoch": 0.9821048146570089, "grad_norm": 11.25, "learning_rate": 1.293680348563598e-06, "loss": 0.3662, "step": 11525 }, { "epoch": 0.9821900298253089, "grad_norm": 16.375, "learning_rate": 1.2935124152445173e-06, "loss": 0.5752, "step": 11526 }, { "epoch": 0.9822752449936089, "grad_norm": 13.9375, "learning_rate": 1.2933444811391235e-06, "loss": 0.698, "step": 11527 }, { "epoch": 0.9823604601619088, "grad_norm": 12.25, "learning_rate": 1.2931765462504513e-06, "loss": 0.5559, "step": 11528 }, { "epoch": 0.9824456753302088, "grad_norm": 17.375, "learning_rate": 1.2930086105815354e-06, "loss": 0.9016, "step": 11529 }, { "epoch": 0.9825308904985087, "grad_norm": 13.25, "learning_rate": 1.292840674135411e-06, "loss": 0.4944, "step": 11530 }, { "epoch": 0.9826161056668087, "grad_norm": 12.4375, "learning_rate": 1.292672736915112e-06, "loss": 0.6946, "step": 11531 }, { "epoch": 0.9827013208351086, "grad_norm": 11.875, "learning_rate": 1.2925047989236748e-06, "loss": 0.4064, "step": 11532 }, { "epoch": 0.9827865360034086, "grad_norm": 13.9375, "learning_rate": 1.292336860164132e-06, "loss": 0.6902, "step": 11533 }, { "epoch": 0.9828717511717086, "grad_norm": 17.0, "learning_rate": 1.2921689206395203e-06, "loss": 0.7277, "step": 11534 }, { "epoch": 0.9829569663400085, "grad_norm": 20.75, "learning_rate": 1.2920009803528733e-06, "loss": 1.0278, "step": 11535 }, { "epoch": 0.9830421815083085, "grad_norm": 20.875, "learning_rate": 1.2918330393072269e-06, "loss": 1.1821, "step": 11536 }, { "epoch": 0.9831273966766084, "grad_norm": 11.25, "learning_rate": 1.2916650975056147e-06, "loss": 0.4462, "step": 11537 }, { "epoch": 0.9832126118449084, "grad_norm": 18.0, "learning_rate": 1.291497154951073e-06, "loss": 1.034, "step": 11538 }, { "epoch": 0.9832978270132083, "grad_norm": 14.5, "learning_rate": 1.2913292116466353e-06, "loss": 0.6812, "step": 11539 }, { "epoch": 0.9833830421815083, "grad_norm": 10.3125, "learning_rate": 1.2911612675953378e-06, "loss": 0.3512, "step": 11540 }, { "epoch": 0.9834682573498083, "grad_norm": 12.5625, "learning_rate": 1.2909933228002144e-06, "loss": 0.5021, "step": 11541 }, { "epoch": 0.9835534725181082, "grad_norm": 12.9375, "learning_rate": 1.2908253772643006e-06, "loss": 0.4651, "step": 11542 }, { "epoch": 0.9836386876864082, "grad_norm": 13.625, "learning_rate": 1.2906574309906312e-06, "loss": 0.6579, "step": 11543 }, { "epoch": 0.9837239028547081, "grad_norm": 17.0, "learning_rate": 1.290489483982241e-06, "loss": 0.7014, "step": 11544 }, { "epoch": 0.9838091180230081, "grad_norm": 19.625, "learning_rate": 1.2903215362421654e-06, "loss": 1.1158, "step": 11545 }, { "epoch": 0.983894333191308, "grad_norm": 15.5625, "learning_rate": 1.2901535877734385e-06, "loss": 0.8292, "step": 11546 }, { "epoch": 0.983979548359608, "grad_norm": 16.375, "learning_rate": 1.2899856385790964e-06, "loss": 0.6922, "step": 11547 }, { "epoch": 0.984064763527908, "grad_norm": 18.75, "learning_rate": 1.2898176886621736e-06, "loss": 1.3498, "step": 11548 }, { "epoch": 0.9841499786962079, "grad_norm": 19.125, "learning_rate": 1.289649738025705e-06, "loss": 0.9013, "step": 11549 }, { "epoch": 0.9842351938645079, "grad_norm": 17.75, "learning_rate": 1.2894817866727257e-06, "loss": 0.8277, "step": 11550 }, { "epoch": 0.9843204090328078, "grad_norm": 13.375, "learning_rate": 1.289313834606271e-06, "loss": 0.5901, "step": 11551 }, { "epoch": 0.9844056242011078, "grad_norm": 17.5, "learning_rate": 1.2891458818293757e-06, "loss": 0.7029, "step": 11552 }, { "epoch": 0.9844908393694077, "grad_norm": 18.5, "learning_rate": 1.288977928345075e-06, "loss": 1.0259, "step": 11553 }, { "epoch": 0.9845760545377077, "grad_norm": 14.875, "learning_rate": 1.2888099741564036e-06, "loss": 0.6933, "step": 11554 }, { "epoch": 0.9846612697060076, "grad_norm": 17.625, "learning_rate": 1.2886420192663973e-06, "loss": 1.0109, "step": 11555 }, { "epoch": 0.9847464848743076, "grad_norm": 19.0, "learning_rate": 1.288474063678091e-06, "loss": 0.9019, "step": 11556 }, { "epoch": 0.9848317000426076, "grad_norm": 16.875, "learning_rate": 1.2883061073945194e-06, "loss": 0.7381, "step": 11557 }, { "epoch": 0.9849169152109075, "grad_norm": 29.125, "learning_rate": 1.2881381504187177e-06, "loss": 1.1664, "step": 11558 }, { "epoch": 0.9850021303792075, "grad_norm": 12.25, "learning_rate": 1.2879701927537218e-06, "loss": 0.617, "step": 11559 }, { "epoch": 0.9850873455475074, "grad_norm": 15.3125, "learning_rate": 1.287802234402566e-06, "loss": 0.6309, "step": 11560 }, { "epoch": 0.9851725607158074, "grad_norm": 11.25, "learning_rate": 1.2876342753682856e-06, "loss": 0.299, "step": 11561 }, { "epoch": 0.9852577758841073, "grad_norm": 14.75, "learning_rate": 1.2874663156539163e-06, "loss": 0.4828, "step": 11562 }, { "epoch": 0.9853429910524073, "grad_norm": 12.9375, "learning_rate": 1.2872983552624932e-06, "loss": 0.2755, "step": 11563 }, { "epoch": 0.9854282062207073, "grad_norm": 24.125, "learning_rate": 1.287130394197051e-06, "loss": 1.0892, "step": 11564 }, { "epoch": 0.9855134213890072, "grad_norm": 15.75, "learning_rate": 1.2869624324606252e-06, "loss": 0.3957, "step": 11565 }, { "epoch": 0.9855986365573072, "grad_norm": 12.0, "learning_rate": 1.2867944700562512e-06, "loss": 0.58, "step": 11566 }, { "epoch": 0.9856838517256071, "grad_norm": 18.0, "learning_rate": 1.286626506986964e-06, "loss": 0.8879, "step": 11567 }, { "epoch": 0.9857690668939071, "grad_norm": 15.8125, "learning_rate": 1.2864585432557995e-06, "loss": 0.8339, "step": 11568 }, { "epoch": 0.985854282062207, "grad_norm": 13.3125, "learning_rate": 1.2862905788657918e-06, "loss": 0.5279, "step": 11569 }, { "epoch": 0.985939497230507, "grad_norm": 54.25, "learning_rate": 1.2861226138199773e-06, "loss": 0.5374, "step": 11570 }, { "epoch": 0.986024712398807, "grad_norm": 14.4375, "learning_rate": 1.2859546481213905e-06, "loss": 0.7737, "step": 11571 }, { "epoch": 0.9861099275671069, "grad_norm": 16.625, "learning_rate": 1.2857866817730674e-06, "loss": 0.7286, "step": 11572 }, { "epoch": 0.9861951427354069, "grad_norm": 15.0625, "learning_rate": 1.2856187147780427e-06, "loss": 0.8106, "step": 11573 }, { "epoch": 0.9862803579037068, "grad_norm": 16.125, "learning_rate": 1.2854507471393521e-06, "loss": 0.7582, "step": 11574 }, { "epoch": 0.9863655730720068, "grad_norm": 14.75, "learning_rate": 1.2852827788600309e-06, "loss": 0.5743, "step": 11575 }, { "epoch": 0.9864507882403067, "grad_norm": 17.625, "learning_rate": 1.2851148099431142e-06, "loss": 0.6944, "step": 11576 }, { "epoch": 0.9865360034086067, "grad_norm": 10.75, "learning_rate": 1.284946840391638e-06, "loss": 0.4466, "step": 11577 }, { "epoch": 0.9866212185769067, "grad_norm": 13.1875, "learning_rate": 1.2847788702086369e-06, "loss": 0.5045, "step": 11578 }, { "epoch": 0.9867064337452066, "grad_norm": 17.875, "learning_rate": 1.2846108993971464e-06, "loss": 0.5855, "step": 11579 }, { "epoch": 0.9867916489135066, "grad_norm": 14.4375, "learning_rate": 1.2844429279602028e-06, "loss": 0.913, "step": 11580 }, { "epoch": 0.9868768640818065, "grad_norm": 16.75, "learning_rate": 1.2842749559008405e-06, "loss": 0.6198, "step": 11581 }, { "epoch": 0.9869620792501065, "grad_norm": 15.25, "learning_rate": 1.2841069832220952e-06, "loss": 0.7165, "step": 11582 }, { "epoch": 0.9870472944184064, "grad_norm": 12.3125, "learning_rate": 1.2839390099270024e-06, "loss": 0.624, "step": 11583 }, { "epoch": 0.9871325095867064, "grad_norm": 18.125, "learning_rate": 1.2837710360185975e-06, "loss": 1.0329, "step": 11584 }, { "epoch": 0.9872177247550064, "grad_norm": 14.1875, "learning_rate": 1.2836030614999168e-06, "loss": 0.7766, "step": 11585 }, { "epoch": 0.9873029399233063, "grad_norm": 14.25, "learning_rate": 1.2834350863739941e-06, "loss": 0.319, "step": 11586 }, { "epoch": 0.9873881550916063, "grad_norm": 18.5, "learning_rate": 1.283267110643866e-06, "loss": 0.7883, "step": 11587 }, { "epoch": 0.9874733702599062, "grad_norm": 25.75, "learning_rate": 1.2830991343125675e-06, "loss": 0.6929, "step": 11588 }, { "epoch": 0.9875585854282062, "grad_norm": 15.375, "learning_rate": 1.2829311573831351e-06, "loss": 0.7721, "step": 11589 }, { "epoch": 0.9876438005965061, "grad_norm": 13.25, "learning_rate": 1.2827631798586032e-06, "loss": 0.4744, "step": 11590 }, { "epoch": 0.9877290157648061, "grad_norm": 15.6875, "learning_rate": 1.2825952017420077e-06, "loss": 0.911, "step": 11591 }, { "epoch": 0.987814230933106, "grad_norm": 12.125, "learning_rate": 1.2824272230363838e-06, "loss": 0.5891, "step": 11592 }, { "epoch": 0.987899446101406, "grad_norm": 14.5625, "learning_rate": 1.2822592437447678e-06, "loss": 0.5944, "step": 11593 }, { "epoch": 0.987984661269706, "grad_norm": 17.5, "learning_rate": 1.2820912638701946e-06, "loss": 0.986, "step": 11594 }, { "epoch": 0.9880698764380059, "grad_norm": 15.125, "learning_rate": 1.2819232834157003e-06, "loss": 0.8919, "step": 11595 }, { "epoch": 0.9881550916063059, "grad_norm": 17.375, "learning_rate": 1.2817553023843196e-06, "loss": 0.9322, "step": 11596 }, { "epoch": 0.9882403067746058, "grad_norm": 12.875, "learning_rate": 1.2815873207790891e-06, "loss": 0.4923, "step": 11597 }, { "epoch": 0.9883255219429058, "grad_norm": 21.125, "learning_rate": 1.2814193386030438e-06, "loss": 0.9952, "step": 11598 }, { "epoch": 0.9884107371112058, "grad_norm": 16.375, "learning_rate": 1.2812513558592194e-06, "loss": 1.0879, "step": 11599 }, { "epoch": 0.9884959522795057, "grad_norm": 20.125, "learning_rate": 1.2810833725506516e-06, "loss": 0.7975, "step": 11600 }, { "epoch": 0.9885811674478057, "grad_norm": 23.875, "learning_rate": 1.2809153886803763e-06, "loss": 1.2124, "step": 11601 }, { "epoch": 0.9886663826161056, "grad_norm": 12.5, "learning_rate": 1.2807474042514287e-06, "loss": 0.2909, "step": 11602 }, { "epoch": 0.9887515977844056, "grad_norm": 17.25, "learning_rate": 1.2805794192668442e-06, "loss": 0.5975, "step": 11603 }, { "epoch": 0.9888368129527055, "grad_norm": 15.375, "learning_rate": 1.280411433729659e-06, "loss": 0.7363, "step": 11604 }, { "epoch": 0.9889220281210055, "grad_norm": 17.625, "learning_rate": 1.280243447642909e-06, "loss": 0.9198, "step": 11605 }, { "epoch": 0.9890072432893054, "grad_norm": 15.9375, "learning_rate": 1.2800754610096293e-06, "loss": 0.5472, "step": 11606 }, { "epoch": 0.9890924584576054, "grad_norm": 14.5, "learning_rate": 1.2799074738328556e-06, "loss": 0.9422, "step": 11607 }, { "epoch": 0.9891776736259054, "grad_norm": 14.6875, "learning_rate": 1.279739486115624e-06, "loss": 0.758, "step": 11608 }, { "epoch": 0.9892628887942053, "grad_norm": 12.25, "learning_rate": 1.2795714978609699e-06, "loss": 0.5117, "step": 11609 }, { "epoch": 0.9893481039625053, "grad_norm": 12.9375, "learning_rate": 1.2794035090719293e-06, "loss": 0.4344, "step": 11610 }, { "epoch": 0.9894333191308052, "grad_norm": 20.75, "learning_rate": 1.2792355197515371e-06, "loss": 0.8544, "step": 11611 }, { "epoch": 0.9895185342991052, "grad_norm": 18.75, "learning_rate": 1.2790675299028302e-06, "loss": 0.9502, "step": 11612 }, { "epoch": 0.9896037494674051, "grad_norm": 13.3125, "learning_rate": 1.2788995395288436e-06, "loss": 0.3178, "step": 11613 }, { "epoch": 0.9896889646357051, "grad_norm": 13.9375, "learning_rate": 1.2787315486326131e-06, "loss": 0.653, "step": 11614 }, { "epoch": 0.9897741798040051, "grad_norm": 19.125, "learning_rate": 1.278563557217175e-06, "loss": 0.8924, "step": 11615 }, { "epoch": 0.989859394972305, "grad_norm": 14.4375, "learning_rate": 1.2783955652855645e-06, "loss": 0.6874, "step": 11616 }, { "epoch": 0.989944610140605, "grad_norm": 16.75, "learning_rate": 1.2782275728408178e-06, "loss": 0.9195, "step": 11617 }, { "epoch": 0.9900298253089049, "grad_norm": 12.625, "learning_rate": 1.27805957988597e-06, "loss": 0.3683, "step": 11618 }, { "epoch": 0.9901150404772049, "grad_norm": 14.875, "learning_rate": 1.277891586424058e-06, "loss": 0.895, "step": 11619 }, { "epoch": 0.9902002556455048, "grad_norm": 11.5625, "learning_rate": 1.2777235924581164e-06, "loss": 0.4849, "step": 11620 }, { "epoch": 0.9902854708138048, "grad_norm": 13.25, "learning_rate": 1.277555597991182e-06, "loss": 0.5929, "step": 11621 }, { "epoch": 0.9903706859821048, "grad_norm": 17.5, "learning_rate": 1.2773876030262899e-06, "loss": 0.846, "step": 11622 }, { "epoch": 0.9904559011504048, "grad_norm": 14.1875, "learning_rate": 1.2772196075664767e-06, "loss": 0.773, "step": 11623 }, { "epoch": 0.9905411163187048, "grad_norm": 15.375, "learning_rate": 1.277051611614777e-06, "loss": 0.9419, "step": 11624 }, { "epoch": 0.9906263314870047, "grad_norm": 16.0, "learning_rate": 1.2768836151742287e-06, "loss": 0.9098, "step": 11625 }, { "epoch": 0.9907115466553047, "grad_norm": 17.125, "learning_rate": 1.2767156182478655e-06, "loss": 0.92, "step": 11626 }, { "epoch": 0.9907967618236047, "grad_norm": 20.5, "learning_rate": 1.2765476208387247e-06, "loss": 0.9697, "step": 11627 }, { "epoch": 0.9908819769919046, "grad_norm": 13.8125, "learning_rate": 1.2763796229498413e-06, "loss": 0.782, "step": 11628 }, { "epoch": 0.9909671921602046, "grad_norm": 14.6875, "learning_rate": 1.2762116245842518e-06, "loss": 0.7664, "step": 11629 }, { "epoch": 0.9910524073285045, "grad_norm": 13.9375, "learning_rate": 1.276043625744992e-06, "loss": 0.64, "step": 11630 }, { "epoch": 0.9911376224968045, "grad_norm": 23.75, "learning_rate": 1.2758756264350978e-06, "loss": 1.0513, "step": 11631 }, { "epoch": 0.9912228376651044, "grad_norm": 18.75, "learning_rate": 1.2757076266576042e-06, "loss": 0.7448, "step": 11632 }, { "epoch": 0.9913080528334044, "grad_norm": 31.5, "learning_rate": 1.275539626415549e-06, "loss": 1.2298, "step": 11633 }, { "epoch": 0.9913932680017044, "grad_norm": 13.6875, "learning_rate": 1.2753716257119663e-06, "loss": 0.4897, "step": 11634 }, { "epoch": 0.9914784831700043, "grad_norm": 21.25, "learning_rate": 1.275203624549893e-06, "loss": 0.9425, "step": 11635 }, { "epoch": 0.9915636983383043, "grad_norm": 14.5, "learning_rate": 1.275035622932365e-06, "loss": 0.6666, "step": 11636 }, { "epoch": 0.9916489135066042, "grad_norm": 15.6875, "learning_rate": 1.2748676208624184e-06, "loss": 1.1565, "step": 11637 }, { "epoch": 0.9917341286749042, "grad_norm": 10.9375, "learning_rate": 1.2746996183430885e-06, "loss": 0.3345, "step": 11638 }, { "epoch": 0.9918193438432041, "grad_norm": 18.5, "learning_rate": 1.2745316153774115e-06, "loss": 0.8908, "step": 11639 }, { "epoch": 0.9919045590115041, "grad_norm": 16.25, "learning_rate": 1.2743636119684238e-06, "loss": 0.9383, "step": 11640 }, { "epoch": 0.991989774179804, "grad_norm": 16.75, "learning_rate": 1.2741956081191612e-06, "loss": 0.9483, "step": 11641 }, { "epoch": 0.992074989348104, "grad_norm": 20.125, "learning_rate": 1.2740276038326595e-06, "loss": 0.8267, "step": 11642 }, { "epoch": 0.992160204516404, "grad_norm": 14.0625, "learning_rate": 1.2738595991119548e-06, "loss": 0.7569, "step": 11643 }, { "epoch": 0.9922454196847039, "grad_norm": 12.8125, "learning_rate": 1.2736915939600833e-06, "loss": 0.5399, "step": 11644 }, { "epoch": 0.9923306348530039, "grad_norm": 11.375, "learning_rate": 1.2735235883800808e-06, "loss": 0.4138, "step": 11645 }, { "epoch": 0.9924158500213038, "grad_norm": 13.3125, "learning_rate": 1.2733555823749832e-06, "loss": 0.5622, "step": 11646 }, { "epoch": 0.9925010651896038, "grad_norm": 14.625, "learning_rate": 1.2731875759478272e-06, "loss": 0.8255, "step": 11647 }, { "epoch": 0.9925862803579037, "grad_norm": 16.375, "learning_rate": 1.273019569101648e-06, "loss": 0.4036, "step": 11648 }, { "epoch": 0.9926714955262037, "grad_norm": 22.25, "learning_rate": 1.2728515618394824e-06, "loss": 0.6206, "step": 11649 }, { "epoch": 0.9927567106945037, "grad_norm": 13.125, "learning_rate": 1.2726835541643656e-06, "loss": 0.6498, "step": 11650 }, { "epoch": 0.9928419258628036, "grad_norm": 13.0625, "learning_rate": 1.2725155460793345e-06, "loss": 0.3856, "step": 11651 }, { "epoch": 0.9929271410311036, "grad_norm": 13.875, "learning_rate": 1.2723475375874244e-06, "loss": 0.4691, "step": 11652 }, { "epoch": 0.9930123561994035, "grad_norm": 25.5, "learning_rate": 1.2721795286916724e-06, "loss": 0.7642, "step": 11653 }, { "epoch": 0.9930975713677035, "grad_norm": 14.125, "learning_rate": 1.2720115193951138e-06, "loss": 0.5637, "step": 11654 }, { "epoch": 0.9931827865360034, "grad_norm": 21.25, "learning_rate": 1.271843509700785e-06, "loss": 1.1254, "step": 11655 }, { "epoch": 0.9932680017043034, "grad_norm": 15.9375, "learning_rate": 1.2716754996117217e-06, "loss": 0.7872, "step": 11656 }, { "epoch": 0.9933532168726034, "grad_norm": 14.8125, "learning_rate": 1.2715074891309606e-06, "loss": 0.8337, "step": 11657 }, { "epoch": 0.9934384320409033, "grad_norm": 14.375, "learning_rate": 1.2713394782615376e-06, "loss": 0.6906, "step": 11658 }, { "epoch": 0.9935236472092033, "grad_norm": 19.125, "learning_rate": 1.2711714670064886e-06, "loss": 0.6924, "step": 11659 }, { "epoch": 0.9936088623775032, "grad_norm": 14.0625, "learning_rate": 1.27100345536885e-06, "loss": 0.4877, "step": 11660 }, { "epoch": 0.9936940775458032, "grad_norm": 14.375, "learning_rate": 1.270835443351658e-06, "loss": 0.8702, "step": 11661 }, { "epoch": 0.9937792927141031, "grad_norm": 13.875, "learning_rate": 1.2706674309579483e-06, "loss": 0.5319, "step": 11662 }, { "epoch": 0.9938645078824031, "grad_norm": 14.75, "learning_rate": 1.2704994181907577e-06, "loss": 0.8544, "step": 11663 }, { "epoch": 0.9939497230507031, "grad_norm": 9.5, "learning_rate": 1.2703314050531213e-06, "loss": 0.2873, "step": 11664 }, { "epoch": 0.994034938219003, "grad_norm": 16.75, "learning_rate": 1.2701633915480765e-06, "loss": 0.8987, "step": 11665 }, { "epoch": 0.994120153387303, "grad_norm": 16.375, "learning_rate": 1.2699953776786589e-06, "loss": 1.0721, "step": 11666 }, { "epoch": 0.9942053685556029, "grad_norm": 13.5625, "learning_rate": 1.2698273634479044e-06, "loss": 0.9995, "step": 11667 }, { "epoch": 0.9942905837239029, "grad_norm": 15.4375, "learning_rate": 1.26965934885885e-06, "loss": 1.0857, "step": 11668 }, { "epoch": 0.9943757988922028, "grad_norm": 13.3125, "learning_rate": 1.269491333914531e-06, "loss": 0.7042, "step": 11669 }, { "epoch": 0.9944610140605028, "grad_norm": 17.375, "learning_rate": 1.2693233186179844e-06, "loss": 1.1651, "step": 11670 }, { "epoch": 0.9945462292288028, "grad_norm": 13.625, "learning_rate": 1.2691553029722455e-06, "loss": 0.5201, "step": 11671 }, { "epoch": 0.9946314443971027, "grad_norm": 19.375, "learning_rate": 1.2689872869803514e-06, "loss": 0.8782, "step": 11672 }, { "epoch": 0.9947166595654027, "grad_norm": 13.9375, "learning_rate": 1.2688192706453375e-06, "loss": 0.7475, "step": 11673 }, { "epoch": 0.9948018747337026, "grad_norm": 14.0, "learning_rate": 1.2686512539702412e-06, "loss": 0.5983, "step": 11674 }, { "epoch": 0.9948870899020026, "grad_norm": 18.375, "learning_rate": 1.2684832369580974e-06, "loss": 0.536, "step": 11675 }, { "epoch": 0.9949723050703025, "grad_norm": 12.5, "learning_rate": 1.268315219611943e-06, "loss": 0.5895, "step": 11676 }, { "epoch": 0.9950575202386025, "grad_norm": 14.0, "learning_rate": 1.268147201934814e-06, "loss": 0.7835, "step": 11677 }, { "epoch": 0.9951427354069025, "grad_norm": 10.125, "learning_rate": 1.2679791839297476e-06, "loss": 0.3089, "step": 11678 }, { "epoch": 0.9952279505752024, "grad_norm": 13.375, "learning_rate": 1.2678111655997784e-06, "loss": 0.4106, "step": 11679 }, { "epoch": 0.9953131657435024, "grad_norm": 13.5, "learning_rate": 1.267643146947944e-06, "loss": 0.6568, "step": 11680 }, { "epoch": 0.9953983809118023, "grad_norm": 13.375, "learning_rate": 1.2674751279772797e-06, "loss": 0.426, "step": 11681 }, { "epoch": 0.9954835960801023, "grad_norm": 14.125, "learning_rate": 1.2673071086908226e-06, "loss": 0.7429, "step": 11682 }, { "epoch": 0.9955688112484022, "grad_norm": 13.75, "learning_rate": 1.2671390890916089e-06, "loss": 0.5406, "step": 11683 }, { "epoch": 0.9956540264167022, "grad_norm": 17.625, "learning_rate": 1.266971069182674e-06, "loss": 1.0032, "step": 11684 }, { "epoch": 0.9957392415850022, "grad_norm": 15.625, "learning_rate": 1.2668030489670551e-06, "loss": 0.6928, "step": 11685 }, { "epoch": 0.9958244567533021, "grad_norm": 13.875, "learning_rate": 1.2666350284477885e-06, "loss": 0.5017, "step": 11686 }, { "epoch": 0.9959096719216021, "grad_norm": 16.375, "learning_rate": 1.2664670076279101e-06, "loss": 0.8655, "step": 11687 }, { "epoch": 0.995994887089902, "grad_norm": 16.125, "learning_rate": 1.2662989865104561e-06, "loss": 0.7047, "step": 11688 }, { "epoch": 0.996080102258202, "grad_norm": 12.375, "learning_rate": 1.2661309650984632e-06, "loss": 0.5351, "step": 11689 }, { "epoch": 0.9961653174265019, "grad_norm": 19.375, "learning_rate": 1.2659629433949674e-06, "loss": 0.8598, "step": 11690 }, { "epoch": 0.9962505325948019, "grad_norm": 10.75, "learning_rate": 1.2657949214030055e-06, "loss": 0.4937, "step": 11691 }, { "epoch": 0.9963357477631019, "grad_norm": 13.1875, "learning_rate": 1.265626899125613e-06, "loss": 0.3935, "step": 11692 }, { "epoch": 0.9964209629314018, "grad_norm": 14.5, "learning_rate": 1.2654588765658268e-06, "loss": 0.7739, "step": 11693 }, { "epoch": 0.9965061780997018, "grad_norm": 19.75, "learning_rate": 1.2652908537266833e-06, "loss": 0.8091, "step": 11694 }, { "epoch": 0.9965913932680017, "grad_norm": 9.875, "learning_rate": 1.265122830611219e-06, "loss": 0.3855, "step": 11695 }, { "epoch": 0.9966766084363017, "grad_norm": 21.875, "learning_rate": 1.2649548072224694e-06, "loss": 1.1789, "step": 11696 }, { "epoch": 0.9967618236046016, "grad_norm": 24.125, "learning_rate": 1.2647867835634717e-06, "loss": 0.8682, "step": 11697 }, { "epoch": 0.9968470387729016, "grad_norm": 13.0625, "learning_rate": 1.264618759637262e-06, "loss": 0.5722, "step": 11698 }, { "epoch": 0.9969322539412016, "grad_norm": 13.3125, "learning_rate": 1.2644507354468762e-06, "loss": 0.8484, "step": 11699 }, { "epoch": 0.9970174691095015, "grad_norm": 15.0, "learning_rate": 1.2642827109953515e-06, "loss": 0.9121, "step": 11700 }, { "epoch": 0.9971026842778015, "grad_norm": 14.875, "learning_rate": 1.2641146862857237e-06, "loss": 1.1333, "step": 11701 }, { "epoch": 0.9971878994461014, "grad_norm": 11.5, "learning_rate": 1.2639466613210293e-06, "loss": 0.5709, "step": 11702 }, { "epoch": 0.9972731146144014, "grad_norm": 12.0625, "learning_rate": 1.2637786361043048e-06, "loss": 0.6844, "step": 11703 }, { "epoch": 0.9973583297827013, "grad_norm": 12.6875, "learning_rate": 1.2636106106385864e-06, "loss": 0.6668, "step": 11704 }, { "epoch": 0.9974435449510013, "grad_norm": 19.25, "learning_rate": 1.2634425849269104e-06, "loss": 0.8497, "step": 11705 }, { "epoch": 0.9975287601193012, "grad_norm": 25.25, "learning_rate": 1.2632745589723139e-06, "loss": 0.8703, "step": 11706 }, { "epoch": 0.9976139752876012, "grad_norm": 11.875, "learning_rate": 1.2631065327778323e-06, "loss": 0.4848, "step": 11707 }, { "epoch": 0.9976991904559012, "grad_norm": 16.375, "learning_rate": 1.2629385063465027e-06, "loss": 0.8481, "step": 11708 }, { "epoch": 0.9977844056242011, "grad_norm": 13.5625, "learning_rate": 1.262770479681361e-06, "loss": 0.5407, "step": 11709 }, { "epoch": 0.9978696207925011, "grad_norm": 14.375, "learning_rate": 1.2626024527854443e-06, "loss": 0.5804, "step": 11710 }, { "epoch": 0.997954835960801, "grad_norm": 15.0625, "learning_rate": 1.2624344256617886e-06, "loss": 0.7516, "step": 11711 }, { "epoch": 0.998040051129101, "grad_norm": 13.75, "learning_rate": 1.26226639831343e-06, "loss": 0.731, "step": 11712 }, { "epoch": 0.998125266297401, "grad_norm": 20.25, "learning_rate": 1.262098370743405e-06, "loss": 0.8451, "step": 11713 }, { "epoch": 0.9982104814657009, "grad_norm": 17.75, "learning_rate": 1.2619303429547509e-06, "loss": 0.8531, "step": 11714 }, { "epoch": 0.9982956966340009, "grad_norm": 14.6875, "learning_rate": 1.2617623149505035e-06, "loss": 0.8063, "step": 11715 }, { "epoch": 0.9983809118023008, "grad_norm": 14.6875, "learning_rate": 1.2615942867336988e-06, "loss": 0.5685, "step": 11716 }, { "epoch": 0.9984661269706008, "grad_norm": 16.75, "learning_rate": 1.261426258307374e-06, "loss": 0.8099, "step": 11717 }, { "epoch": 0.9985513421389007, "grad_norm": 13.75, "learning_rate": 1.261258229674565e-06, "loss": 0.7717, "step": 11718 }, { "epoch": 0.9986365573072007, "grad_norm": 13.9375, "learning_rate": 1.2610902008383086e-06, "loss": 0.5017, "step": 11719 }, { "epoch": 0.9987217724755006, "grad_norm": 16.875, "learning_rate": 1.260922171801641e-06, "loss": 0.5945, "step": 11720 }, { "epoch": 0.9988069876438006, "grad_norm": 13.5625, "learning_rate": 1.260754142567599e-06, "loss": 0.7152, "step": 11721 }, { "epoch": 0.9988922028121006, "grad_norm": 17.375, "learning_rate": 1.2605861131392183e-06, "loss": 1.0023, "step": 11722 }, { "epoch": 0.9989774179804005, "grad_norm": 12.6875, "learning_rate": 1.2604180835195369e-06, "loss": 0.6268, "step": 11723 }, { "epoch": 0.9990626331487005, "grad_norm": 16.125, "learning_rate": 1.2602500537115892e-06, "loss": 0.8231, "step": 11724 }, { "epoch": 0.9991478483170004, "grad_norm": 17.25, "learning_rate": 1.2600820237184132e-06, "loss": 0.8764, "step": 11725 }, { "epoch": 0.9992330634853004, "grad_norm": 11.3125, "learning_rate": 1.2599139935430445e-06, "loss": 0.5251, "step": 11726 }, { "epoch": 0.9993182786536003, "grad_norm": 14.4375, "learning_rate": 1.259745963188521e-06, "loss": 0.8632, "step": 11727 }, { "epoch": 0.9994034938219003, "grad_norm": 12.875, "learning_rate": 1.2595779326578771e-06, "loss": 0.6063, "step": 11728 }, { "epoch": 0.9994887089902003, "grad_norm": 13.0625, "learning_rate": 1.2594099019541508e-06, "loss": 0.5859, "step": 11729 }, { "epoch": 0.9995739241585002, "grad_norm": 11.25, "learning_rate": 1.2592418710803778e-06, "loss": 1.142, "step": 11730 }, { "epoch": 0.9996591393268002, "grad_norm": 14.625, "learning_rate": 1.2590738400395955e-06, "loss": 0.8674, "step": 11731 }, { "epoch": 0.9997443544951001, "grad_norm": 31.875, "learning_rate": 1.258905808834839e-06, "loss": 0.779, "step": 11732 }, { "epoch": 0.9998295696634001, "grad_norm": 17.5, "learning_rate": 1.2587377774691462e-06, "loss": 0.5853, "step": 11733 }, { "epoch": 0.9999147848317, "grad_norm": 19.875, "learning_rate": 1.2585697459455526e-06, "loss": 0.8478, "step": 11734 }, { "epoch": 1.0, "grad_norm": 12.9375, "learning_rate": 1.258401714267095e-06, "loss": 0.5969, "step": 11735 }, { "epoch": 1.0, "eval_loss": 0.9627697467803955, "eval_runtime": 221.5463, "eval_samples_per_second": 7.646, "eval_steps_per_second": 7.646, "step": 11735 }, { "epoch": 1.0000852151683, "grad_norm": 11.625, "learning_rate": 1.2582336824368105e-06, "loss": 0.5151, "step": 11736 } ], "logging_steps": 1, "max_steps": 23470, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 5868, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.605287284969308e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }