eugenepentland commited on
Commit
f1676a9
1 Parent(s): be228a4

Training in progress, epoch 1, checkpoint

Browse files
last-checkpoint/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "JackFram/llama-68m",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 0,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 768,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 3072,
12
+ "max_position_embeddings": 2048,
13
+ "model_type": "llama",
14
+ "num_attention_heads": 12,
15
+ "num_hidden_layers": 2,
16
+ "num_key_value_heads": 12,
17
+ "pad_token_id": 1,
18
+ "pretraining_tp": 1,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 10000.0,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.34.0.dev0",
25
+ "use_cache": false,
26
+ "vocab_size": 32000
27
+ }
last-checkpoint/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.34.0.dev0"
7
+ }
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09dd6afc7ea6c8c2f9457ad103fe123eb2beb234b128a056773c48a452b438d
3
+ size 272138666
last-checkpoint/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38f839173d5fe4bf47f40cba75b6d8db540eeeb28c6cabc56dcca1bbb8a5f32e
3
+ size 136067312
last-checkpoint/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6efcc84b6c67da38b1985714fa1acdad196bdc0f777c790f275f48ab5e9243b1
3
+ size 21687
last-checkpoint/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16d05d9bf65b368763ac8137e36aaa6402f1046e365f8a3c9812f4c818a90c1b
3
+ size 21687
last-checkpoint/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2864df96afa60c87b26f1060dbaef20f3d23636cdba2ffd06718a893c19d3eec
3
+ size 21687
last-checkpoint/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6011594727f50dac81baa164d099575dd7f9b1c311d0ec877355bbf92d5ba0e5
3
+ size 21687
last-checkpoint/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1f306fa4db0abe029638ab6f134e11d688dc1003645595fbed44f15096c22f0
3
+ size 21687
last-checkpoint/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea496d135f8238a435aeca2acced97d28c19fec5562ccdc66c4011df7391812d
3
+ size 21687
last-checkpoint/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:135cd0d08d1a12dd34057d8cec037d8fe8969bedb8cbaefa0851fdb549b7e5c8
3
+ size 21687
last-checkpoint/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e41fc615689cff708a92de88d9465b16a241c30497285fe7e65bda244ac565cc
3
+ size 21687
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e83702393d809d3a353599a799a0758633463c4da5b00787d08fc3d87eb70a4
3
+ size 627
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 25,
6
+ "global_step": 206,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.02,
13
+ "learning_rate": 4.999712746672285e-05,
14
+ "loss": 3.1322,
15
+ "step": 5
16
+ },
17
+ {
18
+ "epoch": 0.05,
19
+ "learning_rate": 4.998844107474698e-05,
20
+ "loss": 0.2475,
21
+ "step": 10
22
+ },
23
+ {
24
+ "epoch": 0.07,
25
+ "learning_rate": 4.997394273973373e-05,
26
+ "loss": 0.1435,
27
+ "step": 15
28
+ },
29
+ {
30
+ "epoch": 0.1,
31
+ "learning_rate": 4.9953635833782084e-05,
32
+ "loss": 0.1028,
33
+ "step": 20
34
+ },
35
+ {
36
+ "epoch": 0.12,
37
+ "learning_rate": 4.992752507997904e-05,
38
+ "loss": 0.0952,
39
+ "step": 25
40
+ },
41
+ {
42
+ "epoch": 0.12,
43
+ "eval_loss": 0.09023847430944443,
44
+ "eval_runtime": 0.122,
45
+ "eval_samples_per_second": 3368.392,
46
+ "eval_steps_per_second": 106.543,
47
+ "step": 25
48
+ },
49
+ {
50
+ "epoch": 0.15,
51
+ "learning_rate": 4.989561655130103e-05,
52
+ "loss": 0.1171,
53
+ "step": 30
54
+ },
55
+ {
56
+ "epoch": 0.17,
57
+ "learning_rate": 4.9857917669201445e-05,
58
+ "loss": 0.0777,
59
+ "step": 35
60
+ },
61
+ {
62
+ "epoch": 0.19,
63
+ "learning_rate": 4.981443720188455e-05,
64
+ "loss": 0.1283,
65
+ "step": 40
66
+ },
67
+ {
68
+ "epoch": 0.22,
69
+ "learning_rate": 4.97651852622661e-05,
70
+ "loss": 0.0895,
71
+ "step": 45
72
+ },
73
+ {
74
+ "epoch": 0.24,
75
+ "learning_rate": 4.9710173305621214e-05,
76
+ "loss": 0.0873,
77
+ "step": 50
78
+ },
79
+ {
80
+ "epoch": 0.24,
81
+ "eval_loss": 0.07891710102558136,
82
+ "eval_runtime": 0.1169,
83
+ "eval_samples_per_second": 3515.748,
84
+ "eval_steps_per_second": 111.204,
85
+ "step": 50
86
+ },
87
+ {
88
+ "epoch": 0.27,
89
+ "learning_rate": 4.964941412692007e-05,
90
+ "loss": 0.0729,
91
+ "step": 55
92
+ },
93
+ {
94
+ "epoch": 0.29,
95
+ "learning_rate": 4.9582921857851984e-05,
96
+ "loss": 0.0668,
97
+ "step": 60
98
+ },
99
+ {
100
+ "epoch": 0.32,
101
+ "learning_rate": 4.951071196353857e-05,
102
+ "loss": 0.0755,
103
+ "step": 65
104
+ },
105
+ {
106
+ "epoch": 0.34,
107
+ "learning_rate": 4.9432801238936795e-05,
108
+ "loss": 0.1041,
109
+ "step": 70
110
+ },
111
+ {
112
+ "epoch": 0.36,
113
+ "learning_rate": 4.9349207804932704e-05,
114
+ "loss": 0.0446,
115
+ "step": 75
116
+ },
117
+ {
118
+ "epoch": 0.36,
119
+ "eval_loss": 0.0668846070766449,
120
+ "eval_runtime": 0.1152,
121
+ "eval_samples_per_second": 3569.236,
122
+ "eval_steps_per_second": 112.896,
123
+ "step": 75
124
+ },
125
+ {
126
+ "epoch": 0.39,
127
+ "learning_rate": 4.925995110412681e-05,
128
+ "loss": 0.0687,
129
+ "step": 80
130
+ },
131
+ {
132
+ "epoch": 0.41,
133
+ "learning_rate": 4.916505189631196e-05,
134
+ "loss": 0.0395,
135
+ "step": 85
136
+ },
137
+ {
138
+ "epoch": 0.44,
139
+ "learning_rate": 4.906453225364499e-05,
140
+ "loss": 0.0801,
141
+ "step": 90
142
+ },
143
+ {
144
+ "epoch": 0.46,
145
+ "learning_rate": 4.8958415555513e-05,
146
+ "loss": 0.0977,
147
+ "step": 95
148
+ },
149
+ {
150
+ "epoch": 0.49,
151
+ "learning_rate": 4.884672648309572e-05,
152
+ "loss": 0.0902,
153
+ "step": 100
154
+ },
155
+ {
156
+ "epoch": 0.49,
157
+ "eval_loss": 0.06855440139770508,
158
+ "eval_runtime": 0.1158,
159
+ "eval_samples_per_second": 3547.794,
160
+ "eval_steps_per_second": 112.217,
161
+ "step": 100
162
+ },
163
+ {
164
+ "epoch": 0.51,
165
+ "learning_rate": 4.8729491013624956e-05,
166
+ "loss": 0.0672,
167
+ "step": 105
168
+ },
169
+ {
170
+ "epoch": 0.53,
171
+ "learning_rate": 4.860673641434275e-05,
172
+ "loss": 0.0433,
173
+ "step": 110
174
+ },
175
+ {
176
+ "epoch": 0.56,
177
+ "learning_rate": 4.847849123615934e-05,
178
+ "loss": 0.0619,
179
+ "step": 115
180
+ },
181
+ {
182
+ "epoch": 0.58,
183
+ "learning_rate": 4.834478530701271e-05,
184
+ "loss": 0.0422,
185
+ "step": 120
186
+ },
187
+ {
188
+ "epoch": 0.61,
189
+ "learning_rate": 4.820564972493101e-05,
190
+ "loss": 0.1589,
191
+ "step": 125
192
+ },
193
+ {
194
+ "epoch": 0.61,
195
+ "eval_loss": 0.15087510645389557,
196
+ "eval_runtime": 0.1202,
197
+ "eval_samples_per_second": 3418.754,
198
+ "eval_steps_per_second": 108.136,
199
+ "step": 125
200
+ },
201
+ {
202
+ "epoch": 0.63,
203
+ "learning_rate": 4.8061116850799625e-05,
204
+ "loss": 0.147,
205
+ "step": 130
206
+ },
207
+ {
208
+ "epoch": 0.66,
209
+ "learning_rate": 4.79112203008345e-05,
210
+ "loss": 0.1062,
211
+ "step": 135
212
+ },
213
+ {
214
+ "epoch": 0.68,
215
+ "learning_rate": 4.775599493876354e-05,
216
+ "loss": 0.052,
217
+ "step": 140
218
+ },
219
+ {
220
+ "epoch": 0.7,
221
+ "learning_rate": 4.759547686771774e-05,
222
+ "loss": 0.0462,
223
+ "step": 145
224
+ },
225
+ {
226
+ "epoch": 0.73,
227
+ "learning_rate": 4.742970342183424e-05,
228
+ "loss": 0.1183,
229
+ "step": 150
230
+ },
231
+ {
232
+ "epoch": 0.73,
233
+ "eval_loss": 0.0662284642457962,
234
+ "eval_runtime": 0.1095,
235
+ "eval_samples_per_second": 3755.038,
236
+ "eval_steps_per_second": 118.772,
237
+ "step": 150
238
+ },
239
+ {
240
+ "epoch": 0.75,
241
+ "learning_rate": 4.7258713157572866e-05,
242
+ "loss": 0.0695,
243
+ "step": 155
244
+ },
245
+ {
246
+ "epoch": 0.78,
247
+ "learning_rate": 4.70825458447485e-05,
248
+ "loss": 0.049,
249
+ "step": 160
250
+ },
251
+ {
252
+ "epoch": 0.8,
253
+ "learning_rate": 4.690124245728121e-05,
254
+ "loss": 0.0352,
255
+ "step": 165
256
+ },
257
+ {
258
+ "epoch": 0.83,
259
+ "learning_rate": 4.671484516366633e-05,
260
+ "loss": 0.1032,
261
+ "step": 170
262
+ },
263
+ {
264
+ "epoch": 0.85,
265
+ "learning_rate": 4.652339731716663e-05,
266
+ "loss": 0.0627,
267
+ "step": 175
268
+ },
269
+ {
270
+ "epoch": 0.85,
271
+ "eval_loss": 0.05767974629998207,
272
+ "eval_runtime": 0.1141,
273
+ "eval_samples_per_second": 3600.998,
274
+ "eval_steps_per_second": 113.9,
275
+ "step": 175
276
+ },
277
+ {
278
+ "epoch": 0.87,
279
+ "learning_rate": 4.6326943445729075e-05,
280
+ "loss": 0.0352,
281
+ "step": 180
282
+ },
283
+ {
284
+ "epoch": 0.9,
285
+ "learning_rate": 4.612552924162822e-05,
286
+ "loss": 0.0294,
287
+ "step": 185
288
+ },
289
+ {
290
+ "epoch": 0.92,
291
+ "learning_rate": 4.591920155083887e-05,
292
+ "loss": 0.0485,
293
+ "step": 190
294
+ },
295
+ {
296
+ "epoch": 0.95,
297
+ "learning_rate": 4.570800836214041e-05,
298
+ "loss": 0.0332,
299
+ "step": 195
300
+ },
301
+ {
302
+ "epoch": 0.97,
303
+ "learning_rate": 4.5491998795955314e-05,
304
+ "loss": 0.0741,
305
+ "step": 200
306
+ },
307
+ {
308
+ "epoch": 0.97,
309
+ "eval_loss": 0.06550680845975876,
310
+ "eval_runtime": 0.1437,
311
+ "eval_samples_per_second": 2859.8,
312
+ "eval_steps_per_second": 90.456,
313
+ "step": 200
314
+ },
315
+ {
316
+ "epoch": 1.0,
317
+ "learning_rate": 4.527122309292447e-05,
318
+ "loss": 0.0169,
319
+ "step": 205
320
+ }
321
+ ],
322
+ "logging_steps": 5,
323
+ "max_steps": 1030,
324
+ "num_train_epochs": 5,
325
+ "save_steps": 500,
326
+ "total_flos": 91858718883840.0,
327
+ "trial_name": null,
328
+ "trial_params": null
329
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3afb97c10653979a44a821bce4c8812c7bb6466ed846acf034eaf9a24208c2aa
3
+ size 4539