rohitnagareddy commited on
Commit
19219b0
·
verified ·
1 Parent(s): cc9bab6

Training in progress, step 200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d169812e68872b996231e0a0076c3912c7f156df560dd4e465c91765bc2a23ae
3
  size 40422168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61454aa457830bd78b1058287dc3fcbe688bdde8a02df8eb0cf52a1e66563fed
3
  size 40422168
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07b6375afd5c12194949440c8320c9770ee4dc26d94509cc4931070b65ef3e85
3
  size 20958516
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d9b2d0121410a2f5a31c57d50c3de06d7b96a853b91aac4595b269fb1413a47
3
  size 20958516
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:905451142165304754e3dca94d7298a0d5c846a2e8dca3114b6489ab3c972537
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b7153db191a9c0af7d8507bfea423baff3bb8f9644f9d8c9e17149088d8c377
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cb728dce3e2541bc781b20d0c6d71ae5c372163a515fbc5a49b008d9df8a390
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca9a25c72339c898b564e0c464a3f6fc75bbeec408008928b7ed05533156b98c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 11.56,
6
  "eval_steps": 500,
7
- "global_step": 150,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -113,6 +113,41 @@
113
  "learning_rate": 6.800000000000001e-05,
114
  "loss": 0.0269,
115
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
116
  }
117
  ],
118
  "logging_steps": 10,
@@ -127,12 +162,12 @@
127
  "should_evaluate": false,
128
  "should_log": false,
129
  "should_save": true,
130
- "should_training_stop": false
131
  },
132
  "attributes": {}
133
  }
134
  },
135
- "total_flos": 231256227840000.0,
136
  "train_batch_size": 2,
137
  "trial_name": null,
138
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 15.4,
6
  "eval_steps": 500,
7
+ "global_step": 200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
113
  "learning_rate": 6.800000000000001e-05,
114
  "loss": 0.0269,
115
  "step": 150
116
+ },
117
+ {
118
+ "epoch": 12.32,
119
+ "grad_norm": 0.23131920397281647,
120
+ "learning_rate": 5.466666666666666e-05,
121
+ "loss": 0.0285,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 13.08,
126
+ "grad_norm": 0.14665621519088745,
127
+ "learning_rate": 4.133333333333333e-05,
128
+ "loss": 0.0289,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 13.88,
133
+ "grad_norm": 0.18651145696640015,
134
+ "learning_rate": 2.8000000000000003e-05,
135
+ "loss": 0.027,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 14.64,
140
+ "grad_norm": 0.20021669566631317,
141
+ "learning_rate": 1.4666666666666668e-05,
142
+ "loss": 0.0272,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 15.4,
147
+ "grad_norm": 0.12408873438835144,
148
+ "learning_rate": 1.3333333333333334e-06,
149
+ "loss": 0.0275,
150
+ "step": 200
151
  }
152
  ],
153
  "logging_steps": 10,
 
162
  "should_evaluate": false,
163
  "should_log": false,
164
  "should_save": true,
165
+ "should_training_stop": true
166
  },
167
  "attributes": {}
168
  }
169
  },
170
+ "total_flos": 308074905600000.0,
171
  "train_batch_size": 2,
172
  "trial_name": null,
173
  "trial_params": null