AaronWu901225's picture
Upload LoRA adapter folder
221fb16 verified
{
"best_metric": 0.04320823773741722,
"best_model_checkpoint": "./xlam_lora_new_1024_2_delete_over_size_3epoch/checkpoint-789",
"epoch": 1.000316906987799,
"eval_steps": 263,
"global_step": 789,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02535255902392648,
"grad_norm": 0.515038788318634,
"learning_rate": 5.633802816901409e-06,
"loss": 0.4826,
"step": 20
},
{
"epoch": 0.05070511804785296,
"grad_norm": 1.2124691009521484,
"learning_rate": 1.1267605633802819e-05,
"loss": 0.337,
"step": 40
},
{
"epoch": 0.07605767707177943,
"grad_norm": 0.562520444393158,
"learning_rate": 1.6901408450704228e-05,
"loss": 0.1167,
"step": 60
},
{
"epoch": 0.10141023609570592,
"grad_norm": 0.1530018150806427,
"learning_rate": 1.9999239776734535e-05,
"loss": 0.0444,
"step": 80
},
{
"epoch": 0.1267627951196324,
"grad_norm": 0.43701356649398804,
"learning_rate": 1.9992107756032603e-05,
"loss": 0.0941,
"step": 100
},
{
"epoch": 0.15211535414355887,
"grad_norm": 0.26956745982170105,
"learning_rate": 1.997747365756147e-05,
"loss": 0.0722,
"step": 120
},
{
"epoch": 0.17746791316748534,
"grad_norm": 0.38304299116134644,
"learning_rate": 1.995534846860705e-05,
"loss": 0.0664,
"step": 140
},
{
"epoch": 0.20282047219141183,
"grad_norm": 0.24561983346939087,
"learning_rate": 1.9925748800768444e-05,
"loss": 0.0474,
"step": 160
},
{
"epoch": 0.2281730312153383,
"grad_norm": 0.26263824105262756,
"learning_rate": 1.9888696877485937e-05,
"loss": 0.0364,
"step": 180
},
{
"epoch": 0.2535255902392648,
"grad_norm": 0.2551981210708618,
"learning_rate": 1.984422051735564e-05,
"loss": 0.0276,
"step": 200
},
{
"epoch": 0.27887814926319127,
"grad_norm": 0.16999328136444092,
"learning_rate": 1.979235311324328e-05,
"loss": 0.039,
"step": 220
},
{
"epoch": 0.30423070828711773,
"grad_norm": 0.2080262005329132,
"learning_rate": 1.973313360721281e-05,
"loss": 0.0353,
"step": 240
},
{
"epoch": 0.3295832673110442,
"grad_norm": 0.19375990331172943,
"learning_rate": 1.9666606461288728e-05,
"loss": 0.0509,
"step": 260
},
{
"epoch": 0.33338615116463316,
"eval_loss": 0.04866104573011398,
"eval_runtime": 144.4235,
"eval_samples_per_second": 4.618,
"eval_steps_per_second": 2.313,
"step": 263
},
{
"epoch": 0.35493582633497067,
"grad_norm": 0.3100409209728241,
"learning_rate": 1.959282162407392e-05,
"loss": 0.0317,
"step": 280
},
{
"epoch": 0.38028838535889714,
"grad_norm": 0.49380865693092346,
"learning_rate": 1.9511834493248303e-05,
"loss": 0.0513,
"step": 300
},
{
"epoch": 0.40564094438282366,
"grad_norm": 0.18333816528320312,
"learning_rate": 1.9423705873976156e-05,
"loss": 0.0449,
"step": 320
},
{
"epoch": 0.43099350340675013,
"grad_norm": 0.3562432825565338,
"learning_rate": 1.9328501933253643e-05,
"loss": 0.028,
"step": 340
},
{
"epoch": 0.4563460624306766,
"grad_norm": 0.341875821352005,
"learning_rate": 1.9226294150230573e-05,
"loss": 0.0361,
"step": 360
},
{
"epoch": 0.48169862145460307,
"grad_norm": 0.259799987077713,
"learning_rate": 1.9117159262543844e-05,
"loss": 0.0454,
"step": 380
},
{
"epoch": 0.5070511804785296,
"grad_norm": 0.4206746518611908,
"learning_rate": 1.9001179208702812e-05,
"loss": 0.0333,
"step": 400
},
{
"epoch": 0.532403739502456,
"grad_norm": 0.3466087877750397,
"learning_rate": 1.88784410665698e-05,
"loss": 0.0292,
"step": 420
},
{
"epoch": 0.5577562985263825,
"grad_norm": 0.5302783846855164,
"learning_rate": 1.8749036987982047e-05,
"loss": 0.0319,
"step": 440
},
{
"epoch": 0.583108857550309,
"grad_norm": 0.3487597107887268,
"learning_rate": 1.861306412956404e-05,
"loss": 0.037,
"step": 460
},
{
"epoch": 0.6084614165742355,
"grad_norm": 0.4753687381744385,
"learning_rate": 1.8470624579782338e-05,
"loss": 0.029,
"step": 480
},
{
"epoch": 0.633813975598162,
"grad_norm": 0.20097890496253967,
"learning_rate": 1.8321825282297467e-05,
"loss": 0.0432,
"step": 500
},
{
"epoch": 0.6591665346220884,
"grad_norm": 0.33282265067100525,
"learning_rate": 1.8166777955670616e-05,
"loss": 0.0414,
"step": 520
},
{
"epoch": 0.6667723023292663,
"eval_loss": 0.045147210359573364,
"eval_runtime": 144.4317,
"eval_samples_per_second": 4.618,
"eval_steps_per_second": 2.313,
"step": 526
},
{
"epoch": 0.6845190936460149,
"grad_norm": 0.17276500165462494,
"learning_rate": 1.800559900948529e-05,
"loss": 0.046,
"step": 540
},
{
"epoch": 0.7098716526699413,
"grad_norm": 0.17020255327224731,
"learning_rate": 1.7838409456946973e-05,
"loss": 0.0273,
"step": 560
},
{
"epoch": 0.7352242116938679,
"grad_norm": 0.09968056529760361,
"learning_rate": 1.7665334824026358e-05,
"loss": 0.0265,
"step": 580
},
{
"epoch": 0.7605767707177943,
"grad_norm": 0.3242018520832062,
"learning_rate": 1.7486505055214405e-05,
"loss": 0.0269,
"step": 600
},
{
"epoch": 0.7859293297417208,
"grad_norm": 0.2690919041633606,
"learning_rate": 1.7302054415959994e-05,
"loss": 0.0247,
"step": 620
},
{
"epoch": 0.8112818887656473,
"grad_norm": 0.5220848321914673,
"learning_rate": 1.7112121391863367e-05,
"loss": 0.0389,
"step": 640
},
{
"epoch": 0.8366344477895737,
"grad_norm": 0.26270976662635803,
"learning_rate": 1.6916848584701092e-05,
"loss": 0.0281,
"step": 660
},
{
"epoch": 0.8619870068135003,
"grad_norm": 0.32137855887413025,
"learning_rate": 1.671638260536062e-05,
"loss": 0.0365,
"step": 680
},
{
"epoch": 0.8873395658374267,
"grad_norm": 0.5726165175437927,
"learning_rate": 1.6510873963764754e-05,
"loss": 0.0468,
"step": 700
},
{
"epoch": 0.9126921248613532,
"grad_norm": 0.7382823824882507,
"learning_rate": 1.6300476955868785e-05,
"loss": 0.0365,
"step": 720
},
{
"epoch": 0.9380446838852796,
"grad_norm": 0.3500744104385376,
"learning_rate": 1.6085349547815002e-05,
"loss": 0.027,
"step": 740
},
{
"epoch": 0.9633972429092061,
"grad_norm": 0.2678811252117157,
"learning_rate": 1.5865653257331697e-05,
"loss": 0.0483,
"step": 760
},
{
"epoch": 0.9887498019331327,
"grad_norm": 0.22900326550006866,
"learning_rate": 1.564155303246556e-05,
"loss": 0.0508,
"step": 780
},
{
"epoch": 1.000316906987799,
"eval_loss": 0.04320823773741722,
"eval_runtime": 144.4268,
"eval_samples_per_second": 4.618,
"eval_steps_per_second": 2.313,
"step": 789
}
],
"logging_steps": 20,
"max_steps": 2364,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 263,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.876643611330478e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}