| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.9987257088244665, | |
| "eval_steps": 1569, | |
| "global_step": 6276, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.24976107040458745, | |
| "grad_norm": 3.5836572647094727, | |
| "learning_rate": 2.4984066284257492e-06, | |
| "loss": 3.2128, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.4995221408091749, | |
| "grad_norm": 3.5956788063049316, | |
| "learning_rate": 4.9968132568514985e-06, | |
| "loss": 2.232, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.7492832112137623, | |
| "grad_norm": 5.461869239807129, | |
| "learning_rate": 7.495219885277247e-06, | |
| "loss": 1.8585, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 0.9990442816183498, | |
| "grad_norm": 4.789157390594482, | |
| "learning_rate": 9.993626513702997e-06, | |
| "loss": 1.663, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 1.2488053520229372, | |
| "grad_norm": 3.923671245574951, | |
| "learning_rate": 9.37719018795795e-06, | |
| "loss": 1.5195, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.4985664224275248, | |
| "grad_norm": 4.947878360748291, | |
| "learning_rate": 8.75278751194648e-06, | |
| "loss": 1.481, | |
| "step": 2352 | |
| }, | |
| { | |
| "epoch": 1.7483274928321122, | |
| "grad_norm": 3.5249428749084473, | |
| "learning_rate": 8.128384835935011e-06, | |
| "loss": 1.4036, | |
| "step": 2744 | |
| }, | |
| { | |
| "epoch": 1.9980885632366996, | |
| "grad_norm": 1.927209734916687, | |
| "learning_rate": 7.503982159923543e-06, | |
| "loss": 1.3675, | |
| "step": 3136 | |
| }, | |
| { | |
| "epoch": 2.247849633641287, | |
| "grad_norm": 5.133701324462891, | |
| "learning_rate": 6.8795794839120745e-06, | |
| "loss": 1.2484, | |
| "step": 3528 | |
| }, | |
| { | |
| "epoch": 2.4976107040458744, | |
| "grad_norm": 5.654458045959473, | |
| "learning_rate": 6.255176807900606e-06, | |
| "loss": 1.1997, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 2.7473717744504618, | |
| "grad_norm": 7.116914749145508, | |
| "learning_rate": 5.630774131889137e-06, | |
| "loss": 1.2026, | |
| "step": 4312 | |
| }, | |
| { | |
| "epoch": 2.9971328448550496, | |
| "grad_norm": 5.798983573913574, | |
| "learning_rate": 5.006371455877669e-06, | |
| "loss": 1.1859, | |
| "step": 4704 | |
| }, | |
| { | |
| "epoch": 3.246893915259637, | |
| "grad_norm": 5.535274505615234, | |
| "learning_rate": 4.3819687798662e-06, | |
| "loss": 1.0458, | |
| "step": 5096 | |
| }, | |
| { | |
| "epoch": 3.4966549856642244, | |
| "grad_norm": 5.662832736968994, | |
| "learning_rate": 3.757566103854731e-06, | |
| "loss": 1.0485, | |
| "step": 5488 | |
| }, | |
| { | |
| "epoch": 3.746416056068812, | |
| "grad_norm": 5.849664211273193, | |
| "learning_rate": 3.1331634278432627e-06, | |
| "loss": 1.0564, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 3.996177126473399, | |
| "grad_norm": 6.500792026519775, | |
| "learning_rate": 2.5087607518317936e-06, | |
| "loss": 1.0403, | |
| "step": 6272 | |
| } | |
| ], | |
| "logging_steps": 392, | |
| "max_steps": 7847, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 6, | |
| "save_steps": 1569, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |