dimasik87 commited on
Commit
0359fb1
·
verified ·
1 Parent(s): 728c36e

Training in progress, step 20, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3923c6c2d238cd3361024c80cac4a57bea4c1dd24732d212c7c0481b868aeeb4
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bf3b554a722161d949f66405978a8112bd762c52b942d8b436544e3dad153d9
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca49e2c253d38e24ae4972e1bc709e59f2e1296805f1052620ce8f42596247ac
3
  size 335922386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9659418d3605e5b40babd95284fbf5ec8686f8df71e70dc60388ce063f42d584
3
  size 335922386
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e841c8d708d2d450b0eab7be6e2c4e08cbedd4e297878439ca4360fd0b733323
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f121b3f8e6b3ed310c94a49893eb97a3ee7cfc910bf43c67bb9322868d758fd
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b52924c88c1c80163d374a6650253dd74b8b46052a3da6fc5e4076e57b5effa
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d605401690d7669ff16aeaca6820cbd8d0d605afe748c51045ce90888810a22
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.00479077775282581,
5
  "eval_steps": 4,
6
- "global_step": 16,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -159,6 +159,42 @@
159
  "eval_samples_per_second": 8.417,
160
  "eval_steps_per_second": 8.417,
161
  "step": 16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162
  }
163
  ],
164
  "logging_steps": 1,
@@ -178,7 +214,7 @@
178
  "attributes": {}
179
  }
180
  },
181
- "total_flos": 5935098268483584.0,
182
  "train_batch_size": 1,
183
  "trial_name": null,
184
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.005988472191032263,
5
  "eval_steps": 4,
6
+ "global_step": 20,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
159
  "eval_samples_per_second": 8.417,
160
  "eval_steps_per_second": 8.417,
161
  "step": 16
162
+ },
163
+ {
164
+ "epoch": 0.005090201362377423,
165
+ "grad_norm": 26.169757843017578,
166
+ "learning_rate": 0.00018526401643540922,
167
+ "loss": 4.8565,
168
+ "step": 17
169
+ },
170
+ {
171
+ "epoch": 0.005389624971929036,
172
+ "grad_norm": 27.142152786254883,
173
+ "learning_rate": 0.00018090169943749476,
174
+ "loss": 4.9723,
175
+ "step": 18
176
+ },
177
+ {
178
+ "epoch": 0.00568904858148065,
179
+ "grad_norm": 18.188777923583984,
180
+ "learning_rate": 0.0001760405965600031,
181
+ "loss": 2.6307,
182
+ "step": 19
183
+ },
184
+ {
185
+ "epoch": 0.005988472191032263,
186
+ "grad_norm": 21.539011001586914,
187
+ "learning_rate": 0.00017071067811865476,
188
+ "loss": 3.5522,
189
+ "step": 20
190
+ },
191
+ {
192
+ "epoch": 0.005988472191032263,
193
+ "eval_loss": 4.069602012634277,
194
+ "eval_runtime": 83.15,
195
+ "eval_samples_per_second": 8.467,
196
+ "eval_steps_per_second": 8.467,
197
+ "step": 20
198
  }
199
  ],
200
  "logging_steps": 1,
 
214
  "attributes": {}
215
  }
216
  },
217
+ "total_flos": 7418872835604480.0,
218
  "train_batch_size": 1,
219
  "trial_name": null,
220
  "trial_params": null