| { | |
| "best_metric": 0.34823477268218994, | |
| "best_model_checkpoint": "../checkpoints/Wav2Vec-voxpopuli/one-speaker/Final-paper/GT+GEN-dxg1/RU/100-epoch/checkpoint-26775", | |
| "epoch": 72.99784777929955, | |
| "global_step": 31025, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.4822134387351778e-08, | |
| "loss": 12.3693, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.4525691699604743e-06, | |
| "loss": 16.0709, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.9347826086956523e-06, | |
| "loss": 15.3276, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.416996047430831e-06, | |
| "loss": 12.0358, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 5.8843873517786565e-06, | |
| "loss": 9.1409, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 6.154871940612793, | |
| "eval_runtime": 318.3436, | |
| "eval_samples_per_second": 26.456, | |
| "eval_wer": 1.0, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 7.366600790513834e-06, | |
| "loss": 7.4893, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 8.848814229249012e-06, | |
| "loss": 6.0766, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.033102766798419e-05, | |
| "loss": 5.2545, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.1813241106719368e-05, | |
| "loss": 4.5575, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 3.599003553390503, | |
| "eval_runtime": 317.6472, | |
| "eval_samples_per_second": 26.514, | |
| "eval_wer": 1.0, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.3295454545454546e-05, | |
| "loss": 4.0577, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.4777667984189723e-05, | |
| "loss": 3.5943, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 1.6259881422924904e-05, | |
| "loss": 3.3759, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.774209486166008e-05, | |
| "loss": 3.2785, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 3.146242380142212, | |
| "eval_runtime": 318.9247, | |
| "eval_samples_per_second": 26.407, | |
| "eval_wer": 1.0, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 1.9224308300395258e-05, | |
| "loss": 3.2268, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 2.0706521739130433e-05, | |
| "loss": 3.1416, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 2.2188735177865615e-05, | |
| "loss": 3.1018, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 2.3670948616600794e-05, | |
| "loss": 3.0594, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 2.515316205533597e-05, | |
| "loss": 3.0063, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 2.9775705337524414, | |
| "eval_runtime": 316.7007, | |
| "eval_samples_per_second": 26.593, | |
| "eval_wer": 1.0, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 2.6635375494071148e-05, | |
| "loss": 2.8872, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 2.8117588932806323e-05, | |
| "loss": 2.5959, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 2.9599802371541505e-05, | |
| "loss": 2.1892, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 2.9945893863029945e-05, | |
| "loss": 1.9088, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 1.6493474245071411, | |
| "eval_runtime": 318.4422, | |
| "eval_samples_per_second": 26.448, | |
| "eval_wer": 0.9566750042119724, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 5.18, | |
| "learning_rate": 2.987177586718055e-05, | |
| "loss": 1.6933, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 2.979765787133116e-05, | |
| "loss": 1.5286, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 5.65, | |
| "learning_rate": 2.9723539875481766e-05, | |
| "loss": 1.4406, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 2.9649421879632374e-05, | |
| "loss": 1.3302, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 1.2176637649536133, | |
| "eval_runtime": 318.9899, | |
| "eval_samples_per_second": 26.402, | |
| "eval_wer": 0.8561190238592035, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 6.12, | |
| "learning_rate": 2.9575303883782986e-05, | |
| "loss": 1.251, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 6.35, | |
| "learning_rate": 2.950118588793359e-05, | |
| "loss": 1.162, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 6.59, | |
| "learning_rate": 2.94270678920842e-05, | |
| "loss": 1.1033, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 2.9352949896234806e-05, | |
| "loss": 1.0447, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 0.9863032102584839, | |
| "eval_runtime": 317.5697, | |
| "eval_samples_per_second": 26.52, | |
| "eval_wer": 0.7464522232734153, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 2.9278831900385414e-05, | |
| "loss": 1.0038, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 7.29, | |
| "learning_rate": 2.9204713904536022e-05, | |
| "loss": 0.963, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 7.53, | |
| "learning_rate": 2.913059590868663e-05, | |
| "loss": 0.9284, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 7.76, | |
| "learning_rate": 2.9056477912837235e-05, | |
| "loss": 0.8925, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 2.8982359916987847e-05, | |
| "loss": 0.8712, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 0.8418087363243103, | |
| "eval_runtime": 318.12, | |
| "eval_samples_per_second": 26.474, | |
| "eval_wer": 0.653970269955029, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 8.23, | |
| "learning_rate": 2.890824192113845e-05, | |
| "loss": 0.8478, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 8.47, | |
| "learning_rate": 2.883412392528906e-05, | |
| "loss": 0.8175, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 8.7, | |
| "learning_rate": 2.876000592943967e-05, | |
| "loss": 0.8035, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 8.94, | |
| "learning_rate": 2.8685887933590275e-05, | |
| "loss": 0.7815, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_loss": 0.7524622678756714, | |
| "eval_runtime": 318.9732, | |
| "eval_samples_per_second": 26.403, | |
| "eval_wer": 0.5982426355283109, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 2.8611769937740884e-05, | |
| "loss": 0.7671, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 9.41, | |
| "learning_rate": 2.853765194189149e-05, | |
| "loss": 0.7474, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 9.65, | |
| "learning_rate": 2.84635339460421e-05, | |
| "loss": 0.7477, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 9.88, | |
| "learning_rate": 2.8389415950192708e-05, | |
| "loss": 0.7128, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 0.6934680938720703, | |
| "eval_runtime": 317.9061, | |
| "eval_samples_per_second": 26.492, | |
| "eval_wer": 0.5617993545962339, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 10.12, | |
| "learning_rate": 2.8315297954343316e-05, | |
| "loss": 0.7134, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 10.35, | |
| "learning_rate": 2.8241179958493924e-05, | |
| "loss": 0.7025, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 10.59, | |
| "learning_rate": 2.8167061962644532e-05, | |
| "loss": 0.6806, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 10.82, | |
| "learning_rate": 2.8092943966795137e-05, | |
| "loss": 0.6713, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_loss": 0.6441113948822021, | |
| "eval_runtime": 330.7931, | |
| "eval_samples_per_second": 25.46, | |
| "eval_wer": 0.5259781495833388, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 11.06, | |
| "learning_rate": 2.8018825970945745e-05, | |
| "loss": 0.6707, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 11.29, | |
| "learning_rate": 2.7944707975096356e-05, | |
| "loss": 0.6552, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 11.53, | |
| "learning_rate": 2.787058997924696e-05, | |
| "loss": 0.6435, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 11.76, | |
| "learning_rate": 2.779647198339757e-05, | |
| "loss": 0.632, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 2.7722353987548177e-05, | |
| "loss": 0.632, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_loss": 0.6222620010375977, | |
| "eval_runtime": 318.5241, | |
| "eval_samples_per_second": 26.441, | |
| "eval_wer": 0.5066678762587318, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 12.23, | |
| "learning_rate": 2.7648235991698785e-05, | |
| "loss": 0.6169, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 12.47, | |
| "learning_rate": 2.7574117995849393e-05, | |
| "loss": 0.6068, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 12.7, | |
| "learning_rate": 2.75e-05, | |
| "loss": 0.6036, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 12.94, | |
| "learning_rate": 2.742588200415061e-05, | |
| "loss": 0.5917, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_loss": 0.5906757712364197, | |
| "eval_runtime": 319.1883, | |
| "eval_samples_per_second": 26.386, | |
| "eval_wer": 0.4835214680991693, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 13.18, | |
| "learning_rate": 2.7351764008301217e-05, | |
| "loss": 0.5905, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 13.41, | |
| "learning_rate": 2.7277646012451822e-05, | |
| "loss": 0.5818, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 13.65, | |
| "learning_rate": 2.720352801660243e-05, | |
| "loss": 0.5737, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 13.88, | |
| "learning_rate": 2.712941002075304e-05, | |
| "loss": 0.5607, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_loss": 0.5859212279319763, | |
| "eval_runtime": 319.205, | |
| "eval_samples_per_second": 26.384, | |
| "eval_wer": 0.47469576599577507, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 14.12, | |
| "learning_rate": 2.7055292024903646e-05, | |
| "loss": 0.5661, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 14.35, | |
| "learning_rate": 2.6981174029054254e-05, | |
| "loss": 0.5503, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 14.59, | |
| "learning_rate": 2.6907056033204865e-05, | |
| "loss": 0.5517, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 14.82, | |
| "learning_rate": 2.683293803735547e-05, | |
| "loss": 0.544, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_loss": 0.5427600741386414, | |
| "eval_runtime": 317.4374, | |
| "eval_samples_per_second": 26.531, | |
| "eval_wer": 0.45460789777219063, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 15.06, | |
| "learning_rate": 2.6758820041506078e-05, | |
| "loss": 0.5377, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 15.29, | |
| "learning_rate": 2.6684702045656686e-05, | |
| "loss": 0.5364, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 15.53, | |
| "learning_rate": 2.6610584049807294e-05, | |
| "loss": 0.5207, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 15.76, | |
| "learning_rate": 2.6536466053957902e-05, | |
| "loss": 0.518, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "learning_rate": 2.6462348058108507e-05, | |
| "loss": 0.5208, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_loss": 0.549843966960907, | |
| "eval_runtime": 315.364, | |
| "eval_samples_per_second": 26.706, | |
| "eval_wer": 0.4465857103977398, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 16.23, | |
| "learning_rate": 2.638823006225912e-05, | |
| "loss": 0.5157, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 16.47, | |
| "learning_rate": 2.6314112066409726e-05, | |
| "loss": 0.5056, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 16.7, | |
| "learning_rate": 2.623999407056033e-05, | |
| "loss": 0.5052, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 16.94, | |
| "learning_rate": 2.616587607471094e-05, | |
| "loss": 0.4979, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_loss": 0.5181886553764343, | |
| "eval_runtime": 316.5373, | |
| "eval_samples_per_second": 26.607, | |
| "eval_wer": 0.432731561280958, | |
| "step": 7225 | |
| }, | |
| { | |
| "epoch": 17.18, | |
| "learning_rate": 2.609175807886155e-05, | |
| "loss": 0.4981, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 17.41, | |
| "learning_rate": 2.6017640083012155e-05, | |
| "loss": 0.4886, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 17.65, | |
| "learning_rate": 2.5943522087162763e-05, | |
| "loss": 0.4807, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 17.88, | |
| "learning_rate": 2.586940409131337e-05, | |
| "loss": 0.4829, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_loss": 0.5218132138252258, | |
| "eval_runtime": 316.5653, | |
| "eval_samples_per_second": 26.604, | |
| "eval_wer": 0.42257098793431913, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 18.12, | |
| "learning_rate": 2.579528609546398e-05, | |
| "loss": 0.4803, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 18.35, | |
| "learning_rate": 2.5721168099614588e-05, | |
| "loss": 0.4731, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 18.59, | |
| "learning_rate": 2.5647050103765192e-05, | |
| "loss": 0.4732, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 18.82, | |
| "learning_rate": 2.5572932107915804e-05, | |
| "loss": 0.4667, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_loss": 0.49254098534584045, | |
| "eval_runtime": 314.8242, | |
| "eval_samples_per_second": 26.751, | |
| "eval_wer": 0.4067339718251448, | |
| "step": 8075 | |
| }, | |
| { | |
| "epoch": 19.06, | |
| "learning_rate": 2.5498814112066412e-05, | |
| "loss": 0.47, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 19.29, | |
| "learning_rate": 2.5424696116217016e-05, | |
| "loss": 0.4549, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 19.53, | |
| "learning_rate": 2.5350578120367624e-05, | |
| "loss": 0.4486, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 19.76, | |
| "learning_rate": 2.5276460124518236e-05, | |
| "loss": 0.4516, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "learning_rate": 2.520234212866884e-05, | |
| "loss": 0.4517, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_loss": 0.4848106801509857, | |
| "eval_runtime": 317.0239, | |
| "eval_samples_per_second": 26.566, | |
| "eval_wer": 0.398400746491103, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 20.23, | |
| "learning_rate": 2.512822413281945e-05, | |
| "loss": 0.4469, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 20.47, | |
| "learning_rate": 2.505410613697006e-05, | |
| "loss": 0.441, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 20.7, | |
| "learning_rate": 2.4979988141120665e-05, | |
| "loss": 0.4379, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 20.94, | |
| "learning_rate": 2.4905870145271273e-05, | |
| "loss": 0.4372, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_loss": 0.48364755511283875, | |
| "eval_runtime": 316.0715, | |
| "eval_samples_per_second": 26.646, | |
| "eval_wer": 0.39547180570495455, | |
| "step": 8925 | |
| }, | |
| { | |
| "epoch": 21.18, | |
| "learning_rate": 2.4831752149421877e-05, | |
| "loss": 0.4374, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 21.41, | |
| "learning_rate": 2.475763415357249e-05, | |
| "loss": 0.4326, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 21.65, | |
| "learning_rate": 2.4683516157723097e-05, | |
| "loss": 0.4247, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 21.88, | |
| "learning_rate": 2.46093981618737e-05, | |
| "loss": 0.4277, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_loss": 0.48058849573135376, | |
| "eval_runtime": 319.3273, | |
| "eval_samples_per_second": 26.374, | |
| "eval_wer": 0.39389069607703375, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 22.12, | |
| "learning_rate": 2.453528016602431e-05, | |
| "loss": 0.4265, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 22.35, | |
| "learning_rate": 2.446116217017492e-05, | |
| "loss": 0.4147, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 22.59, | |
| "learning_rate": 2.4387044174325526e-05, | |
| "loss": 0.4194, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 22.82, | |
| "learning_rate": 2.4312926178476134e-05, | |
| "loss": 0.412, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_loss": 0.44829100370407104, | |
| "eval_runtime": 316.8142, | |
| "eval_samples_per_second": 26.583, | |
| "eval_wer": 0.3794274309560529, | |
| "step": 9775 | |
| }, | |
| { | |
| "epoch": 23.06, | |
| "learning_rate": 2.4238808182626745e-05, | |
| "loss": 0.4157, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 23.29, | |
| "learning_rate": 2.416469018677735e-05, | |
| "loss": 0.4051, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 23.53, | |
| "learning_rate": 2.4090572190927958e-05, | |
| "loss": 0.4046, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 23.76, | |
| "learning_rate": 2.4016454195078563e-05, | |
| "loss": 0.4003, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "learning_rate": 2.3942336199229174e-05, | |
| "loss": 0.403, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_loss": 0.45721083879470825, | |
| "eval_runtime": 318.2524, | |
| "eval_samples_per_second": 26.463, | |
| "eval_wer": 0.37823511877762084, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 24.23, | |
| "learning_rate": 2.3868218203379782e-05, | |
| "loss": 0.3971, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 24.47, | |
| "learning_rate": 2.3794100207530387e-05, | |
| "loss": 0.3982, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 24.7, | |
| "learning_rate": 2.3719982211680998e-05, | |
| "loss": 0.3959, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 24.94, | |
| "learning_rate": 2.3645864215831606e-05, | |
| "loss": 0.394, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_loss": 0.4389103055000305, | |
| "eval_runtime": 315.649, | |
| "eval_samples_per_second": 26.682, | |
| "eval_wer": 0.37056284910770987, | |
| "step": 10625 | |
| }, | |
| { | |
| "epoch": 25.18, | |
| "learning_rate": 2.357174621998221e-05, | |
| "loss": 0.3935, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 25.41, | |
| "learning_rate": 2.349762822413282e-05, | |
| "loss": 0.3863, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 25.65, | |
| "learning_rate": 2.342351022828343e-05, | |
| "loss": 0.3882, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 25.88, | |
| "learning_rate": 2.335013341239253e-05, | |
| "loss": 0.3887, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_loss": 0.43516942858695984, | |
| "eval_runtime": 315.434, | |
| "eval_samples_per_second": 26.7, | |
| "eval_wer": 0.36394033255141844, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 26.12, | |
| "learning_rate": 2.3276015416543136e-05, | |
| "loss": 0.3902, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 26.35, | |
| "learning_rate": 2.3201897420693744e-05, | |
| "loss": 0.3858, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 26.59, | |
| "learning_rate": 2.3127779424844355e-05, | |
| "loss": 0.3785, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 26.82, | |
| "learning_rate": 2.305366142899496e-05, | |
| "loss": 0.378, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_loss": 0.443709135055542, | |
| "eval_runtime": 318.0942, | |
| "eval_samples_per_second": 26.476, | |
| "eval_wer": 0.3679967859410842, | |
| "step": 11475 | |
| }, | |
| { | |
| "epoch": 27.06, | |
| "learning_rate": 2.2979543433145568e-05, | |
| "loss": 0.3835, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 27.29, | |
| "learning_rate": 2.2905425437296176e-05, | |
| "loss": 0.3776, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 27.53, | |
| "learning_rate": 2.2831307441446784e-05, | |
| "loss": 0.3721, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 27.76, | |
| "learning_rate": 2.2757189445597392e-05, | |
| "loss": 0.3686, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "learning_rate": 2.2683071449748e-05, | |
| "loss": 0.3711, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_loss": 0.4321639835834503, | |
| "eval_runtime": 315.6052, | |
| "eval_samples_per_second": 26.685, | |
| "eval_wer": 0.3558403856870699, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 28.23, | |
| "learning_rate": 2.26096946338571e-05, | |
| "loss": 0.3767, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 28.47, | |
| "learning_rate": 2.253557663800771e-05, | |
| "loss": 0.363, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 28.7, | |
| "learning_rate": 2.2461458642158317e-05, | |
| "loss": 0.3667, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 28.94, | |
| "learning_rate": 2.2387340646308925e-05, | |
| "loss": 0.3657, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_loss": 0.4269185960292816, | |
| "eval_runtime": 316.1021, | |
| "eval_samples_per_second": 26.643, | |
| "eval_wer": 0.3577714130195306, | |
| "step": 12325 | |
| }, | |
| { | |
| "epoch": 29.18, | |
| "learning_rate": 2.2313222650459533e-05, | |
| "loss": 0.3624, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 29.41, | |
| "learning_rate": 2.223910465461014e-05, | |
| "loss": 0.3575, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 29.65, | |
| "learning_rate": 2.216498665876075e-05, | |
| "loss": 0.349, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 29.88, | |
| "learning_rate": 2.2090868662911354e-05, | |
| "loss": 0.3548, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_loss": 0.4217592477798462, | |
| "eval_runtime": 317.2188, | |
| "eval_samples_per_second": 26.55, | |
| "eval_wer": 0.34779227848265315, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 30.12, | |
| "learning_rate": 2.2016750667061965e-05, | |
| "loss": 0.3589, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 30.35, | |
| "learning_rate": 2.194263267121257e-05, | |
| "loss": 0.3522, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 30.59, | |
| "learning_rate": 2.1868514675363178e-05, | |
| "loss": 0.3443, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 30.82, | |
| "learning_rate": 2.1794396679513786e-05, | |
| "loss": 0.3464, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_loss": 0.42884036898612976, | |
| "eval_runtime": 317.8415, | |
| "eval_samples_per_second": 26.497, | |
| "eval_wer": 0.3482329155920737, | |
| "step": 13175 | |
| }, | |
| { | |
| "epoch": 31.06, | |
| "learning_rate": 2.1720278683664394e-05, | |
| "loss": 0.3489, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 31.29, | |
| "learning_rate": 2.1646160687815002e-05, | |
| "loss": 0.3443, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 31.53, | |
| "learning_rate": 2.157204269196561e-05, | |
| "loss": 0.3457, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 31.76, | |
| "learning_rate": 2.1497924696116218e-05, | |
| "loss": 0.3514, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "learning_rate": 2.1423806700266826e-05, | |
| "loss": 0.3447, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_loss": 0.41752490401268005, | |
| "eval_runtime": 317.3655, | |
| "eval_samples_per_second": 26.537, | |
| "eval_wer": 0.34579645157527766, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 32.23, | |
| "learning_rate": 2.1349688704417434e-05, | |
| "loss": 0.3373, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 32.47, | |
| "learning_rate": 2.127557070856804e-05, | |
| "loss": 0.3403, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 32.7, | |
| "learning_rate": 2.120145271271865e-05, | |
| "loss": 0.3418, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 32.94, | |
| "learning_rate": 2.1127334716869255e-05, | |
| "loss": 0.3378, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_loss": 0.4071788787841797, | |
| "eval_runtime": 317.9264, | |
| "eval_samples_per_second": 26.49, | |
| "eval_wer": 0.3422843146148961, | |
| "step": 14025 | |
| }, | |
| { | |
| "epoch": 33.18, | |
| "learning_rate": 2.1053216721019863e-05, | |
| "loss": 0.3358, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 33.41, | |
| "learning_rate": 2.0979098725170475e-05, | |
| "loss": 0.333, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 33.65, | |
| "learning_rate": 2.090498072932108e-05, | |
| "loss": 0.3339, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 33.88, | |
| "learning_rate": 2.0830862733471687e-05, | |
| "loss": 0.3323, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_loss": 0.39172977209091187, | |
| "eval_runtime": 317.2839, | |
| "eval_samples_per_second": 26.544, | |
| "eval_wer": 0.3331605344668939, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 34.12, | |
| "learning_rate": 2.0756744737622295e-05, | |
| "loss": 0.3338, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 34.35, | |
| "learning_rate": 2.0682626741772903e-05, | |
| "loss": 0.3286, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 34.59, | |
| "learning_rate": 2.060850874592351e-05, | |
| "loss": 0.3286, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 34.82, | |
| "learning_rate": 2.053439075007412e-05, | |
| "loss": 0.3278, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_loss": 0.40219607949256897, | |
| "eval_runtime": 317.0822, | |
| "eval_samples_per_second": 26.561, | |
| "eval_wer": 0.3327587771024222, | |
| "step": 14875 | |
| }, | |
| { | |
| "epoch": 35.06, | |
| "learning_rate": 2.0460272754224724e-05, | |
| "loss": 0.3245, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 35.29, | |
| "learning_rate": 2.0386154758375336e-05, | |
| "loss": 0.3184, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 35.53, | |
| "learning_rate": 2.031203676252594e-05, | |
| "loss": 0.3313, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 35.76, | |
| "learning_rate": 2.023791876667655e-05, | |
| "loss": 0.3267, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "learning_rate": 2.016380077082716e-05, | |
| "loss": 0.3221, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_loss": 0.4049491584300995, | |
| "eval_runtime": 318.0804, | |
| "eval_samples_per_second": 26.478, | |
| "eval_wer": 0.33557107865372404, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 36.23, | |
| "learning_rate": 2.0089682774977765e-05, | |
| "loss": 0.3196, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 36.47, | |
| "learning_rate": 2.0015564779128373e-05, | |
| "loss": 0.3239, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 36.7, | |
| "learning_rate": 1.994144678327898e-05, | |
| "loss": 0.3199, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 36.94, | |
| "learning_rate": 1.986732878742959e-05, | |
| "loss": 0.3171, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "eval_loss": 0.3955913484096527, | |
| "eval_runtime": 317.9907, | |
| "eval_samples_per_second": 26.485, | |
| "eval_wer": 0.33121654721945026, | |
| "step": 15725 | |
| }, | |
| { | |
| "epoch": 37.18, | |
| "learning_rate": 1.9793210791580197e-05, | |
| "loss": 0.3235, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 37.41, | |
| "learning_rate": 1.9719092795730805e-05, | |
| "loss": 0.3155, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 37.65, | |
| "learning_rate": 1.9644974799881413e-05, | |
| "loss": 0.3193, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 37.88, | |
| "learning_rate": 1.957085680403202e-05, | |
| "loss": 0.3128, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "eval_loss": 0.3912222683429718, | |
| "eval_runtime": 317.3702, | |
| "eval_samples_per_second": 26.537, | |
| "eval_wer": 0.3300112751260352, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 38.12, | |
| "learning_rate": 1.9496738808182626e-05, | |
| "loss": 0.315, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 38.35, | |
| "learning_rate": 1.9422620812333234e-05, | |
| "loss": 0.3106, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 38.59, | |
| "learning_rate": 1.9349243996442334e-05, | |
| "loss": 0.3091, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 38.82, | |
| "learning_rate": 1.9275867180551438e-05, | |
| "loss": 0.3114, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "eval_loss": 0.40268105268478394, | |
| "eval_runtime": 317.2535, | |
| "eval_samples_per_second": 26.547, | |
| "eval_wer": 0.3303352730006091, | |
| "step": 16575 | |
| }, | |
| { | |
| "epoch": 39.06, | |
| "learning_rate": 1.9201749184702046e-05, | |
| "loss": 0.3144, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 39.29, | |
| "learning_rate": 1.9127631188852654e-05, | |
| "loss": 0.3102, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 39.53, | |
| "learning_rate": 1.905351319300326e-05, | |
| "loss": 0.3098, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 39.76, | |
| "learning_rate": 1.897939519715387e-05, | |
| "loss": 0.3071, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "learning_rate": 1.890527720130448e-05, | |
| "loss": 0.3061, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_loss": 0.39571964740753174, | |
| "eval_runtime": 319.4169, | |
| "eval_samples_per_second": 26.367, | |
| "eval_wer": 0.32423115304363603, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 40.23, | |
| "learning_rate": 1.8831159205455083e-05, | |
| "loss": 0.2991, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 40.47, | |
| "learning_rate": 1.8757041209605695e-05, | |
| "loss": 0.3006, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 40.7, | |
| "learning_rate": 1.8682923213756303e-05, | |
| "loss": 0.3041, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 40.94, | |
| "learning_rate": 1.8608805217906907e-05, | |
| "loss": 0.306, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "eval_loss": 0.3856252431869507, | |
| "eval_runtime": 318.6166, | |
| "eval_samples_per_second": 26.433, | |
| "eval_wer": 0.322105726986431, | |
| "step": 17425 | |
| }, | |
| { | |
| "epoch": 41.18, | |
| "learning_rate": 1.8534687222057515e-05, | |
| "loss": 0.308, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 41.41, | |
| "learning_rate": 1.8460569226208123e-05, | |
| "loss": 0.298, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 41.65, | |
| "learning_rate": 1.838645123035873e-05, | |
| "loss": 0.3004, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 41.88, | |
| "learning_rate": 1.831233323450934e-05, | |
| "loss": 0.299, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "eval_loss": 0.384583979845047, | |
| "eval_runtime": 316.4886, | |
| "eval_samples_per_second": 26.611, | |
| "eval_wer": 0.3222223662212776, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 42.12, | |
| "learning_rate": 1.8238215238659944e-05, | |
| "loss": 0.2965, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 42.35, | |
| "learning_rate": 1.8164097242810556e-05, | |
| "loss": 0.3002, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 42.59, | |
| "learning_rate": 1.8089979246961164e-05, | |
| "loss": 0.2961, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 42.82, | |
| "learning_rate": 1.801586125111177e-05, | |
| "loss": 0.295, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "eval_loss": 0.3782486915588379, | |
| "eval_runtime": 316.5726, | |
| "eval_samples_per_second": 26.604, | |
| "eval_wer": 0.3191767862002825, | |
| "step": 18275 | |
| }, | |
| { | |
| "epoch": 43.06, | |
| "learning_rate": 1.794174325526238e-05, | |
| "loss": 0.2962, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 43.29, | |
| "learning_rate": 1.7867625259412988e-05, | |
| "loss": 0.2926, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 43.53, | |
| "learning_rate": 1.7793507263563593e-05, | |
| "loss": 0.2993, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 43.76, | |
| "learning_rate": 1.77193892677142e-05, | |
| "loss": 0.2908, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "learning_rate": 1.764527127186481e-05, | |
| "loss": 0.2919, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "eval_loss": 0.3738417625427246, | |
| "eval_runtime": 316.4549, | |
| "eval_samples_per_second": 26.614, | |
| "eval_wer": 0.3151980923005145, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 44.23, | |
| "learning_rate": 1.7571153276015417e-05, | |
| "loss": 0.2927, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 44.47, | |
| "learning_rate": 1.7497035280166025e-05, | |
| "loss": 0.2913, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 44.7, | |
| "learning_rate": 1.7422917284316633e-05, | |
| "loss": 0.2906, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 44.94, | |
| "learning_rate": 1.734879928846724e-05, | |
| "loss": 0.2837, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "eval_loss": 0.38119739294052124, | |
| "eval_runtime": 317.5894, | |
| "eval_samples_per_second": 26.519, | |
| "eval_wer": 0.31785487487202085, | |
| "step": 19125 | |
| }, | |
| { | |
| "epoch": 45.18, | |
| "learning_rate": 1.727468129261785e-05, | |
| "loss": 0.2909, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 45.41, | |
| "learning_rate": 1.7200563296768454e-05, | |
| "loss": 0.2858, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 45.65, | |
| "learning_rate": 1.7126445300919065e-05, | |
| "loss": 0.2905, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 45.88, | |
| "learning_rate": 1.7052327305069673e-05, | |
| "loss": 0.2825, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 46.0, | |
| "eval_loss": 0.38637784123420715, | |
| "eval_runtime": 317.552, | |
| "eval_samples_per_second": 26.522, | |
| "eval_wer": 0.31740127784761735, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 46.12, | |
| "learning_rate": 1.6978209309220278e-05, | |
| "loss": 0.29, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 46.35, | |
| "learning_rate": 1.690409131337089e-05, | |
| "loss": 0.29, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 46.59, | |
| "learning_rate": 1.6829973317521494e-05, | |
| "loss": 0.2855, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 46.82, | |
| "learning_rate": 1.6755855321672102e-05, | |
| "loss": 0.2827, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "eval_loss": 0.38330715894699097, | |
| "eval_runtime": 317.4978, | |
| "eval_samples_per_second": 26.526, | |
| "eval_wer": 0.3139798602921165, | |
| "step": 19975 | |
| }, | |
| { | |
| "epoch": 47.06, | |
| "learning_rate": 1.668173732582271e-05, | |
| "loss": 0.2883, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 47.29, | |
| "learning_rate": 1.6607619329973318e-05, | |
| "loss": 0.282, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 47.53, | |
| "learning_rate": 1.6533501334123926e-05, | |
| "loss": 0.2826, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 47.76, | |
| "learning_rate": 1.6459383338274534e-05, | |
| "loss": 0.2756, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "learning_rate": 1.638526534242514e-05, | |
| "loss": 0.2849, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "eval_loss": 0.3806135356426239, | |
| "eval_runtime": 317.6019, | |
| "eval_samples_per_second": 26.517, | |
| "eval_wer": 0.31536657119529293, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 48.23, | |
| "learning_rate": 1.631114734657575e-05, | |
| "loss": 0.2844, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 48.47, | |
| "learning_rate": 1.623702935072636e-05, | |
| "loss": 0.2808, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 48.7, | |
| "learning_rate": 1.616365253483546e-05, | |
| "loss": 0.2845, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 48.94, | |
| "learning_rate": 1.6089534538986064e-05, | |
| "loss": 0.2767, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 49.0, | |
| "eval_loss": 0.37584495544433594, | |
| "eval_runtime": 317.5121, | |
| "eval_samples_per_second": 26.525, | |
| "eval_wer": 0.3132929847980197, | |
| "step": 20825 | |
| }, | |
| { | |
| "epoch": 49.18, | |
| "learning_rate": 1.6015416543136675e-05, | |
| "loss": 0.283, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 49.41, | |
| "learning_rate": 1.5941298547287283e-05, | |
| "loss": 0.2801, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 49.65, | |
| "learning_rate": 1.5867180551437888e-05, | |
| "loss": 0.2737, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 49.88, | |
| "learning_rate": 1.57930625555885e-05, | |
| "loss": 0.2732, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "eval_loss": 0.3800816833972931, | |
| "eval_runtime": 317.0609, | |
| "eval_samples_per_second": 26.563, | |
| "eval_wer": 0.31414833918689494, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 50.12, | |
| "learning_rate": 1.5718944559739107e-05, | |
| "loss": 0.2818, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 50.35, | |
| "learning_rate": 1.5644826563889712e-05, | |
| "loss": 0.2728, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 50.59, | |
| "learning_rate": 1.557070856804032e-05, | |
| "loss": 0.276, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 50.82, | |
| "learning_rate": 1.5496590572190928e-05, | |
| "loss": 0.2691, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 51.0, | |
| "eval_loss": 0.3639204800128937, | |
| "eval_runtime": 317.0756, | |
| "eval_samples_per_second": 26.561, | |
| "eval_wer": 0.30717590492606367, | |
| "step": 21675 | |
| }, | |
| { | |
| "epoch": 51.06, | |
| "learning_rate": 1.5422472576341536e-05, | |
| "loss": 0.2741, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 51.29, | |
| "learning_rate": 1.5348354580492144e-05, | |
| "loss": 0.2722, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 51.53, | |
| "learning_rate": 1.527423658464275e-05, | |
| "loss": 0.2724, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 51.76, | |
| "learning_rate": 1.520011858879336e-05, | |
| "loss": 0.2662, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "learning_rate": 1.5126000592943967e-05, | |
| "loss": 0.2742, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "eval_loss": 0.36436760425567627, | |
| "eval_runtime": 316.4547, | |
| "eval_samples_per_second": 26.614, | |
| "eval_wer": 0.30645014968701806, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 52.23, | |
| "learning_rate": 1.5051882597094575e-05, | |
| "loss": 0.2699, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 52.47, | |
| "learning_rate": 1.4977764601245183e-05, | |
| "loss": 0.2732, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 52.7, | |
| "learning_rate": 1.4903646605395791e-05, | |
| "loss": 0.2723, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 52.94, | |
| "learning_rate": 1.4829528609546397e-05, | |
| "loss": 0.2671, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 53.0, | |
| "eval_loss": 0.3704889416694641, | |
| "eval_runtime": 317.8613, | |
| "eval_samples_per_second": 26.496, | |
| "eval_wer": 0.3060743121525123, | |
| "step": 22525 | |
| }, | |
| { | |
| "epoch": 53.18, | |
| "learning_rate": 1.4755410613697007e-05, | |
| "loss": 0.2713, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 53.41, | |
| "learning_rate": 1.4681292617847613e-05, | |
| "loss": 0.2658, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 53.65, | |
| "learning_rate": 1.4607174621998221e-05, | |
| "loss": 0.2707, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 53.88, | |
| "learning_rate": 1.4533797806107324e-05, | |
| "loss": 0.2618, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 54.0, | |
| "eval_loss": 0.3567538857460022, | |
| "eval_runtime": 353.5118, | |
| "eval_samples_per_second": 23.824, | |
| "eval_wer": 0.3049986392089268, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 54.12, | |
| "learning_rate": 1.4459679810257932e-05, | |
| "loss": 0.2679, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 54.35, | |
| "learning_rate": 1.4385561814408538e-05, | |
| "loss": 0.2647, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 54.59, | |
| "learning_rate": 1.4311443818559146e-05, | |
| "loss": 0.2632, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 54.82, | |
| "learning_rate": 1.4237325822709754e-05, | |
| "loss": 0.265, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "eval_loss": 0.3651841878890991, | |
| "eval_runtime": 317.4254, | |
| "eval_samples_per_second": 26.532, | |
| "eval_wer": 0.3071111053511489, | |
| "step": 23375 | |
| }, | |
| { | |
| "epoch": 55.06, | |
| "learning_rate": 1.4163207826860362e-05, | |
| "loss": 0.2638, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 55.29, | |
| "learning_rate": 1.408908983101097e-05, | |
| "loss": 0.2648, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 55.53, | |
| "learning_rate": 1.4014971835161577e-05, | |
| "loss": 0.264, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 55.76, | |
| "learning_rate": 1.3940853839312186e-05, | |
| "loss": 0.2631, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "learning_rate": 1.3866735843462793e-05, | |
| "loss": 0.2655, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "eval_loss": 0.3658052384853363, | |
| "eval_runtime": 316.7135, | |
| "eval_samples_per_second": 26.592, | |
| "eval_wer": 0.3048431202291313, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 56.23, | |
| "learning_rate": 1.3792617847613401e-05, | |
| "loss": 0.2635, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 56.47, | |
| "learning_rate": 1.3718499851764009e-05, | |
| "loss": 0.2658, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 56.7, | |
| "learning_rate": 1.3644381855914617e-05, | |
| "loss": 0.2683, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 56.94, | |
| "learning_rate": 1.3570263860065223e-05, | |
| "loss": 0.2588, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 57.0, | |
| "eval_loss": 0.3624616861343384, | |
| "eval_runtime": 317.8838, | |
| "eval_samples_per_second": 26.494, | |
| "eval_wer": 0.30356008864581846, | |
| "step": 24225 | |
| }, | |
| { | |
| "epoch": 57.18, | |
| "learning_rate": 1.3496145864215831e-05, | |
| "loss": 0.2566, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 57.41, | |
| "learning_rate": 1.3422027868366441e-05, | |
| "loss": 0.2575, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 57.65, | |
| "learning_rate": 1.3347909872517047e-05, | |
| "loss": 0.2643, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 57.88, | |
| "learning_rate": 1.3273791876667656e-05, | |
| "loss": 0.2585, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 58.0, | |
| "eval_loss": 0.3619252145290375, | |
| "eval_runtime": 317.5906, | |
| "eval_samples_per_second": 26.518, | |
| "eval_wer": 0.3017457005482044, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 58.12, | |
| "learning_rate": 1.3199673880818262e-05, | |
| "loss": 0.2616, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 58.35, | |
| "learning_rate": 1.3125555884968872e-05, | |
| "loss": 0.2588, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 58.59, | |
| "learning_rate": 1.3051437889119478e-05, | |
| "loss": 0.2606, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 58.82, | |
| "learning_rate": 1.2977319893270086e-05, | |
| "loss": 0.2616, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 59.0, | |
| "eval_loss": 0.35793522000312805, | |
| "eval_runtime": 316.6098, | |
| "eval_samples_per_second": 26.601, | |
| "eval_wer": 0.3006829875196019, | |
| "step": 25075 | |
| }, | |
| { | |
| "epoch": 59.06, | |
| "learning_rate": 1.2903943077379188e-05, | |
| "loss": 0.2591, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 59.29, | |
| "learning_rate": 1.2830566261488289e-05, | |
| "loss": 0.2606, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 59.53, | |
| "learning_rate": 1.2756448265638897e-05, | |
| "loss": 0.2581, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 59.76, | |
| "learning_rate": 1.2682330269789505e-05, | |
| "loss": 0.2574, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "learning_rate": 1.2608212273940113e-05, | |
| "loss": 0.2509, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "eval_loss": 0.36091676354408264, | |
| "eval_runtime": 663.9043, | |
| "eval_samples_per_second": 12.686, | |
| "eval_wer": 0.2987519601871412, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 60.23, | |
| "learning_rate": 1.2534094278090721e-05, | |
| "loss": 0.2566, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 60.47, | |
| "learning_rate": 1.2459976282241328e-05, | |
| "loss": 0.2556, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 60.7, | |
| "learning_rate": 1.2385858286391937e-05, | |
| "loss": 0.2539, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 60.94, | |
| "learning_rate": 1.2311740290542544e-05, | |
| "loss": 0.2538, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 61.0, | |
| "eval_loss": 0.3586573004722595, | |
| "eval_runtime": 317.4988, | |
| "eval_samples_per_second": 26.526, | |
| "eval_wer": 0.29758556783867496, | |
| "step": 25925 | |
| }, | |
| { | |
| "epoch": 61.18, | |
| "learning_rate": 1.2237622294693152e-05, | |
| "loss": 0.2589, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 61.41, | |
| "learning_rate": 1.2163504298843758e-05, | |
| "loss": 0.249, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 61.65, | |
| "learning_rate": 1.2089386302994368e-05, | |
| "loss": 0.2464, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 61.88, | |
| "learning_rate": 1.2015268307144976e-05, | |
| "loss": 0.2563, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 62.0, | |
| "eval_loss": 0.3534925878047943, | |
| "eval_runtime": 317.8014, | |
| "eval_samples_per_second": 26.501, | |
| "eval_wer": 0.29697645183447596, | |
| "step": 26350 | |
| }, | |
| { | |
| "epoch": 62.12, | |
| "learning_rate": 1.1941150311295582e-05, | |
| "loss": 0.2559, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 62.35, | |
| "learning_rate": 1.1867032315446192e-05, | |
| "loss": 0.2543, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 62.59, | |
| "learning_rate": 1.1792914319596798e-05, | |
| "loss": 0.2496, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 62.82, | |
| "learning_rate": 1.1718796323747406e-05, | |
| "loss": 0.249, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 63.0, | |
| "eval_loss": 0.34823477268218994, | |
| "eval_runtime": 317.6983, | |
| "eval_samples_per_second": 26.509, | |
| "eval_wer": 0.29906299814673215, | |
| "step": 26775 | |
| }, | |
| { | |
| "epoch": 63.06, | |
| "learning_rate": 1.1644678327898013e-05, | |
| "loss": 0.2536, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 63.29, | |
| "learning_rate": 1.1570560332048623e-05, | |
| "loss": 0.2481, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 63.53, | |
| "learning_rate": 1.1496442336199229e-05, | |
| "loss": 0.2504, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 63.76, | |
| "learning_rate": 1.1422324340349837e-05, | |
| "loss": 0.2482, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "learning_rate": 1.1348206344500445e-05, | |
| "loss": 0.2508, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "eval_loss": 0.35596856474876404, | |
| "eval_runtime": 314.8406, | |
| "eval_samples_per_second": 26.75, | |
| "eval_wer": 0.29501950467204935, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 64.23, | |
| "learning_rate": 1.1274088348651053e-05, | |
| "loss": 0.2562, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 64.47, | |
| "learning_rate": 1.1199970352801661e-05, | |
| "loss": 0.2504, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 64.7, | |
| "learning_rate": 1.1125852356952268e-05, | |
| "loss": 0.25, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 64.94, | |
| "learning_rate": 1.1051734361102877e-05, | |
| "loss": 0.2518, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 65.0, | |
| "eval_loss": 0.35186004638671875, | |
| "eval_runtime": 317.5788, | |
| "eval_samples_per_second": 26.519, | |
| "eval_wer": 0.2965358147250554, | |
| "step": 27625 | |
| }, | |
| { | |
| "epoch": 65.18, | |
| "learning_rate": 1.0977616365253484e-05, | |
| "loss": 0.2502, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 65.41, | |
| "learning_rate": 1.0903498369404092e-05, | |
| "loss": 0.2451, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 65.65, | |
| "learning_rate": 1.0829380373554698e-05, | |
| "loss": 0.2508, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 65.88, | |
| "learning_rate": 1.0755262377705308e-05, | |
| "loss": 0.248, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 66.0, | |
| "eval_loss": 0.34826791286468506, | |
| "eval_runtime": 315.877, | |
| "eval_samples_per_second": 26.662, | |
| "eval_wer": 0.2940863907932764, | |
| "step": 28050 | |
| }, | |
| { | |
| "epoch": 66.12, | |
| "learning_rate": 1.0681144381855916e-05, | |
| "loss": 0.2488, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 66.35, | |
| "learning_rate": 1.0607026386006522e-05, | |
| "loss": 0.2446, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 66.59, | |
| "learning_rate": 1.053290839015713e-05, | |
| "loss": 0.2472, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 66.82, | |
| "learning_rate": 1.0458790394307738e-05, | |
| "loss": 0.2427, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 67.0, | |
| "eval_loss": 0.35498106479644775, | |
| "eval_runtime": 315.9906, | |
| "eval_samples_per_second": 26.653, | |
| "eval_wer": 0.29247936133538965, | |
| "step": 28475 | |
| }, | |
| { | |
| "epoch": 67.06, | |
| "learning_rate": 1.0384672398458346e-05, | |
| "loss": 0.2519, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 67.29, | |
| "learning_rate": 1.0310554402608953e-05, | |
| "loss": 0.2463, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 67.53, | |
| "learning_rate": 1.0236436406759562e-05, | |
| "loss": 0.2406, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 67.76, | |
| "learning_rate": 1.0162318410910169e-05, | |
| "loss": 0.2434, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "learning_rate": 1.0088200415060777e-05, | |
| "loss": 0.2439, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "eval_loss": 0.35530757904052734, | |
| "eval_runtime": 316.4428, | |
| "eval_samples_per_second": 26.615, | |
| "eval_wer": 0.2964710151501406, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 68.23, | |
| "learning_rate": 1.0014082419211385e-05, | |
| "loss": 0.2444, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 68.47, | |
| "learning_rate": 9.939964423361993e-06, | |
| "loss": 0.2459, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 68.7, | |
| "learning_rate": 9.865846427512601e-06, | |
| "loss": 0.2397, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 68.94, | |
| "learning_rate": 9.791728431663207e-06, | |
| "loss": 0.2427, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 69.0, | |
| "eval_loss": 0.35256123542785645, | |
| "eval_runtime": 316.7354, | |
| "eval_samples_per_second": 26.59, | |
| "eval_wer": 0.29481214603232203, | |
| "step": 29325 | |
| }, | |
| { | |
| "epoch": 69.18, | |
| "learning_rate": 9.717610435813815e-06, | |
| "loss": 0.2464, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 69.41, | |
| "learning_rate": 9.643492439964424e-06, | |
| "loss": 0.2396, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 69.65, | |
| "learning_rate": 9.569374444115032e-06, | |
| "loss": 0.2441, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 69.88, | |
| "learning_rate": 9.495256448265638e-06, | |
| "loss": 0.2412, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "eval_loss": 0.356432169675827, | |
| "eval_runtime": 316.2059, | |
| "eval_samples_per_second": 26.635, | |
| "eval_wer": 0.2973004497090499, | |
| "step": 29750 | |
| }, | |
| { | |
| "epoch": 70.12, | |
| "learning_rate": 9.421879632374742e-06, | |
| "loss": 0.2488, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 70.35, | |
| "learning_rate": 9.347761636525348e-06, | |
| "loss": 0.2348, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 70.59, | |
| "learning_rate": 9.273643640675956e-06, | |
| "loss": 0.2426, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 70.82, | |
| "learning_rate": 9.199525644826563e-06, | |
| "loss": 0.2394, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 71.0, | |
| "eval_loss": 0.3529145419597626, | |
| "eval_runtime": 314.7426, | |
| "eval_samples_per_second": 26.758, | |
| "eval_wer": 0.28974481927398554, | |
| "step": 30175 | |
| }, | |
| { | |
| "epoch": 71.06, | |
| "learning_rate": 9.125407648977172e-06, | |
| "loss": 0.2451, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 71.29, | |
| "learning_rate": 9.051289653127779e-06, | |
| "loss": 0.2371, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 71.53, | |
| "learning_rate": 8.977171657278387e-06, | |
| "loss": 0.2367, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 71.76, | |
| "learning_rate": 8.903053661428997e-06, | |
| "loss": 0.2417, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "learning_rate": 8.828935665579603e-06, | |
| "loss": 0.247, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "eval_loss": 0.3524751663208008, | |
| "eval_runtime": 316.256, | |
| "eval_samples_per_second": 26.63, | |
| "eval_wer": 0.2913777685618382, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 72.23, | |
| "learning_rate": 8.754817669730211e-06, | |
| "loss": 0.2448, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 72.47, | |
| "learning_rate": 8.680699673880817e-06, | |
| "loss": 0.238, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 72.7, | |
| "learning_rate": 8.606581678031427e-06, | |
| "loss": 0.2341, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 72.94, | |
| "learning_rate": 8.532463682182034e-06, | |
| "loss": 0.2371, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 73.0, | |
| "eval_loss": 0.3488871455192566, | |
| "eval_runtime": 316.856, | |
| "eval_samples_per_second": 26.58, | |
| "eval_wer": 0.28990033825378103, | |
| "step": 31025 | |
| }, | |
| { | |
| "epoch": 73.0, | |
| "step": 31025, | |
| "total_flos": 0, | |
| "train_runtime": 465541.7438, | |
| "train_samples_per_second": 0.091 | |
| } | |
| ], | |
| "max_steps": 42500, | |
| "num_train_epochs": 100, | |
| "total_flos": 0, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |