borggAI commited on
Commit
6233046
1 Parent(s): 0de6a7d

Upload 20 files

Browse files
README.md CHANGED
@@ -1,3 +1,37 @@
 
1
  ---
2
- license: odbl
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
3
+ tags:
4
+ - autotrain
5
+ - text-classification
6
+ widget:
7
+ - text: "I love AutoTrain"
8
+ datasets:
9
+ - autotrain-pmf0g-rj8fa/autotrain-data
10
  ---
11
+
12
+ # Model Trained Using AutoTrain
13
+
14
+ - Problem type: Text Classification
15
+
16
+ ## Validation Metrics
17
+ loss: 0.03381425514817238
18
+
19
+ f1_macro: 0.9910410929202866
20
+
21
+ f1_micro: 0.9908675799086758
22
+
23
+ f1_weighted: 0.9908473335613555
24
+
25
+ precision_macro: 0.9909727371947719
26
+
27
+ precision_micro: 0.9908675799086758
28
+
29
+ precision_weighted: 0.9908883151237302
30
+
31
+ recall_macro: 0.9911698494022667
32
+
33
+ recall_micro: 0.9908675799086758
34
+
35
+ recall_weighted: 0.9908675799086758
36
+
37
+ accuracy: 0.9908675799086758
checkpoint-330/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert/distilbert-base-uncased",
3
+ "_num_labels": 4,
4
+ "activation": "gelu",
5
+ "architectures": [
6
+ "DistilBertForSequenceClassification"
7
+ ],
8
+ "attention_dropout": 0.1,
9
+ "dim": 768,
10
+ "dropout": 0.1,
11
+ "hidden_dim": 3072,
12
+ "id2label": {
13
+ "0": "date-based question answering",
14
+ "1": "math",
15
+ "2": "question-answering",
16
+ "3": "summarization"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "label2id": {
20
+ "date-based question answering": 0,
21
+ "math": 1,
22
+ "question-answering": 2,
23
+ "summarization": 3
24
+ },
25
+ "max_position_embeddings": 512,
26
+ "model_type": "distilbert",
27
+ "n_heads": 12,
28
+ "n_layers": 6,
29
+ "pad_token_id": 0,
30
+ "problem_type": "single_label_classification",
31
+ "qa_dropout": 0.1,
32
+ "seq_classif_dropout": 0.2,
33
+ "sinusoidal_pos_embds": false,
34
+ "tie_weights_": true,
35
+ "torch_dtype": "float32",
36
+ "transformers_version": "4.38.2",
37
+ "vocab_size": 30522
38
+ }
checkpoint-330/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b05d2e2cf016ebcfef096093bda1964ca91f544d3e074b6188b3cf95389a4f7
3
+ size 267838720
checkpoint-330/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4248f9cfbc4a3e525ef2a48b3cf33c26f088ba52e5edf4ed42ab71a677d51a2
3
+ size 535739578
checkpoint-330/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a93811df8f81dda65ed3c088c2e84acd3858e75566d986cbcb8a4bb2a14a2c3
3
+ size 15024
checkpoint-330/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b07dcd15cd55c2f884a5141e3134f3271902fddcf689142d5bf7b0e5e9b43bfc
3
+ size 15024
checkpoint-330/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a80bce123efa5ae2a86ff32d0fc988929f030c3642ab02e6538774164ca3ece
3
+ size 15024
checkpoint-330/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07f62cd2557df01aebc12280d0b47ec89ff26a9ce9f2ae502e2ac4a31dd3134f
3
+ size 15024
checkpoint-330/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9db4d17c0f48ef5ee615629c2facabcf91f13efbf974e69768b293f250efeff8
3
+ size 1064
checkpoint-330/trainer_state.json ADDED
@@ -0,0 +1,360 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.03381425514817238,
3
+ "best_model_checkpoint": "autotrain-pmf0g-rj8fa/checkpoint-330",
4
+ "epoch": 6.0,
5
+ "eval_steps": 500,
6
+ "global_step": 330,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.18,
13
+ "grad_norm": 1.5416135787963867,
14
+ "learning_rate": 9.090909090909091e-06,
15
+ "loss": 1.3908,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 0.36,
20
+ "grad_norm": 1.7443188428878784,
21
+ "learning_rate": 1.8181818181818182e-05,
22
+ "loss": 1.3651,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.55,
27
+ "grad_norm": 2.275134563446045,
28
+ "learning_rate": 2.7272727272727273e-05,
29
+ "loss": 1.2519,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.73,
34
+ "grad_norm": 2.0600032806396484,
35
+ "learning_rate": 3.6363636363636364e-05,
36
+ "loss": 0.9604,
37
+ "step": 40
38
+ },
39
+ {
40
+ "epoch": 0.91,
41
+ "grad_norm": 1.9959027767181396,
42
+ "learning_rate": 4.454545454545455e-05,
43
+ "loss": 0.5255,
44
+ "step": 50
45
+ },
46
+ {
47
+ "epoch": 1.0,
48
+ "eval_accuracy": 0.9748858447488584,
49
+ "eval_f1_macro": 0.9756765342178342,
50
+ "eval_f1_micro": 0.9748858447488584,
51
+ "eval_f1_weighted": 0.9749751427585824,
52
+ "eval_loss": 0.20833216607570648,
53
+ "eval_precision_macro": 0.9759725400457666,
54
+ "eval_precision_micro": 0.9748858447488584,
55
+ "eval_precision_weighted": 0.9751052736068879,
56
+ "eval_recall_macro": 0.9754215648989755,
57
+ "eval_recall_micro": 0.9748858447488584,
58
+ "eval_recall_weighted": 0.9748858447488584,
59
+ "eval_runtime": 0.1296,
60
+ "eval_samples_per_second": 3380.791,
61
+ "eval_steps_per_second": 54.031,
62
+ "step": 55
63
+ },
64
+ {
65
+ "epoch": 1.09,
66
+ "grad_norm": 2.120863199234009,
67
+ "learning_rate": 4.9595959595959594e-05,
68
+ "loss": 0.2706,
69
+ "step": 60
70
+ },
71
+ {
72
+ "epoch": 1.27,
73
+ "grad_norm": 0.6585132479667664,
74
+ "learning_rate": 4.858585858585859e-05,
75
+ "loss": 0.1525,
76
+ "step": 70
77
+ },
78
+ {
79
+ "epoch": 1.45,
80
+ "grad_norm": 4.553682327270508,
81
+ "learning_rate": 4.7575757575757576e-05,
82
+ "loss": 0.0592,
83
+ "step": 80
84
+ },
85
+ {
86
+ "epoch": 1.64,
87
+ "grad_norm": 0.41076862812042236,
88
+ "learning_rate": 4.656565656565657e-05,
89
+ "loss": 0.0615,
90
+ "step": 90
91
+ },
92
+ {
93
+ "epoch": 1.82,
94
+ "grad_norm": 3.347062110900879,
95
+ "learning_rate": 4.565656565656566e-05,
96
+ "loss": 0.1097,
97
+ "step": 100
98
+ },
99
+ {
100
+ "epoch": 2.0,
101
+ "grad_norm": 3.071051597595215,
102
+ "learning_rate": 4.464646464646465e-05,
103
+ "loss": 0.0851,
104
+ "step": 110
105
+ },
106
+ {
107
+ "epoch": 2.0,
108
+ "eval_accuracy": 0.9748858447488584,
109
+ "eval_f1_macro": 0.9754834367482361,
110
+ "eval_f1_micro": 0.9748858447488584,
111
+ "eval_f1_weighted": 0.9749041744765222,
112
+ "eval_loss": 0.07598946988582611,
113
+ "eval_precision_macro": 0.9755606462503015,
114
+ "eval_precision_micro": 0.9748858447488584,
115
+ "eval_precision_weighted": 0.9752209332157372,
116
+ "eval_recall_macro": 0.9756989697593826,
117
+ "eval_recall_micro": 0.9748858447488584,
118
+ "eval_recall_weighted": 0.9748858447488584,
119
+ "eval_runtime": 0.1241,
120
+ "eval_samples_per_second": 3530.172,
121
+ "eval_steps_per_second": 56.418,
122
+ "step": 110
123
+ },
124
+ {
125
+ "epoch": 2.18,
126
+ "grad_norm": 0.41413745284080505,
127
+ "learning_rate": 4.3636363636363636e-05,
128
+ "loss": 0.0477,
129
+ "step": 120
130
+ },
131
+ {
132
+ "epoch": 2.36,
133
+ "grad_norm": 5.531299591064453,
134
+ "learning_rate": 4.262626262626263e-05,
135
+ "loss": 0.0332,
136
+ "step": 130
137
+ },
138
+ {
139
+ "epoch": 2.55,
140
+ "grad_norm": 0.4154397249221802,
141
+ "learning_rate": 4.161616161616162e-05,
142
+ "loss": 0.0398,
143
+ "step": 140
144
+ },
145
+ {
146
+ "epoch": 2.73,
147
+ "grad_norm": 2.7882883548736572,
148
+ "learning_rate": 4.0606060606060606e-05,
149
+ "loss": 0.027,
150
+ "step": 150
151
+ },
152
+ {
153
+ "epoch": 2.91,
154
+ "grad_norm": 3.329517364501953,
155
+ "learning_rate": 3.9595959595959594e-05,
156
+ "loss": 0.0104,
157
+ "step": 160
158
+ },
159
+ {
160
+ "epoch": 3.0,
161
+ "eval_accuracy": 0.9885844748858448,
162
+ "eval_f1_macro": 0.9888365654549831,
163
+ "eval_f1_micro": 0.9885844748858448,
164
+ "eval_f1_weighted": 0.9886138949055431,
165
+ "eval_loss": 0.04520614817738533,
166
+ "eval_precision_macro": 0.9889521258788082,
167
+ "eval_precision_micro": 0.9885844748858448,
168
+ "eval_precision_weighted": 0.9888428242322514,
169
+ "eval_recall_macro": 0.9889150852693966,
170
+ "eval_recall_micro": 0.9885844748858448,
171
+ "eval_recall_weighted": 0.9885844748858448,
172
+ "eval_runtime": 0.1261,
173
+ "eval_samples_per_second": 3473.478,
174
+ "eval_steps_per_second": 55.512,
175
+ "step": 165
176
+ },
177
+ {
178
+ "epoch": 3.09,
179
+ "grad_norm": 0.08116048574447632,
180
+ "learning_rate": 3.858585858585859e-05,
181
+ "loss": 0.0334,
182
+ "step": 170
183
+ },
184
+ {
185
+ "epoch": 3.27,
186
+ "grad_norm": 0.1431998461484909,
187
+ "learning_rate": 3.757575757575758e-05,
188
+ "loss": 0.0244,
189
+ "step": 180
190
+ },
191
+ {
192
+ "epoch": 3.45,
193
+ "grad_norm": 0.04078197479248047,
194
+ "learning_rate": 3.656565656565657e-05,
195
+ "loss": 0.0069,
196
+ "step": 190
197
+ },
198
+ {
199
+ "epoch": 3.64,
200
+ "grad_norm": 0.03086886927485466,
201
+ "learning_rate": 3.555555555555556e-05,
202
+ "loss": 0.0042,
203
+ "step": 200
204
+ },
205
+ {
206
+ "epoch": 3.82,
207
+ "grad_norm": 0.03221021592617035,
208
+ "learning_rate": 3.454545454545455e-05,
209
+ "loss": 0.0324,
210
+ "step": 210
211
+ },
212
+ {
213
+ "epoch": 4.0,
214
+ "grad_norm": 0.04350810497999191,
215
+ "learning_rate": 3.3535353535353536e-05,
216
+ "loss": 0.0203,
217
+ "step": 220
218
+ },
219
+ {
220
+ "epoch": 4.0,
221
+ "eval_accuracy": 0.9840182648401826,
222
+ "eval_f1_macro": 0.9844112150421562,
223
+ "eval_f1_micro": 0.9840182648401826,
224
+ "eval_f1_weighted": 0.9839758921897677,
225
+ "eval_loss": 0.06017115339636803,
226
+ "eval_precision_macro": 0.9844066713333537,
227
+ "eval_precision_micro": 0.9840182648401826,
228
+ "eval_precision_weighted": 0.984073353425936,
229
+ "eval_recall_macro": 0.9845520881850645,
230
+ "eval_recall_micro": 0.9840182648401826,
231
+ "eval_recall_weighted": 0.9840182648401826,
232
+ "eval_runtime": 0.124,
233
+ "eval_samples_per_second": 3533.459,
234
+ "eval_steps_per_second": 56.471,
235
+ "step": 220
236
+ },
237
+ {
238
+ "epoch": 4.18,
239
+ "grad_norm": 0.09443140029907227,
240
+ "learning_rate": 3.2525252525252524e-05,
241
+ "loss": 0.0034,
242
+ "step": 230
243
+ },
244
+ {
245
+ "epoch": 4.36,
246
+ "grad_norm": 2.7363362312316895,
247
+ "learning_rate": 3.151515151515151e-05,
248
+ "loss": 0.0049,
249
+ "step": 240
250
+ },
251
+ {
252
+ "epoch": 4.55,
253
+ "grad_norm": 0.026695426553487778,
254
+ "learning_rate": 3.050505050505051e-05,
255
+ "loss": 0.0024,
256
+ "step": 250
257
+ },
258
+ {
259
+ "epoch": 4.73,
260
+ "grad_norm": 0.023897232487797737,
261
+ "learning_rate": 2.9494949494949498e-05,
262
+ "loss": 0.0022,
263
+ "step": 260
264
+ },
265
+ {
266
+ "epoch": 4.91,
267
+ "grad_norm": 0.021171841770410538,
268
+ "learning_rate": 2.8484848484848486e-05,
269
+ "loss": 0.0021,
270
+ "step": 270
271
+ },
272
+ {
273
+ "epoch": 5.0,
274
+ "eval_accuracy": 0.9885844748858448,
275
+ "eval_f1_macro": 0.9888140343800068,
276
+ "eval_f1_micro": 0.9885844748858448,
277
+ "eval_f1_weighted": 0.9885396362946075,
278
+ "eval_loss": 0.06403101235628128,
279
+ "eval_precision_macro": 0.9890982392627942,
280
+ "eval_precision_micro": 0.9885844748858448,
281
+ "eval_precision_weighted": 0.9889162145164453,
282
+ "eval_recall_macro": 0.9889380530973452,
283
+ "eval_recall_micro": 0.9885844748858448,
284
+ "eval_recall_weighted": 0.9885844748858448,
285
+ "eval_runtime": 0.1253,
286
+ "eval_samples_per_second": 3496.411,
287
+ "eval_steps_per_second": 55.879,
288
+ "step": 275
289
+ },
290
+ {
291
+ "epoch": 5.09,
292
+ "grad_norm": 0.018654897809028625,
293
+ "learning_rate": 2.7474747474747474e-05,
294
+ "loss": 0.0024,
295
+ "step": 280
296
+ },
297
+ {
298
+ "epoch": 5.27,
299
+ "grad_norm": 0.02241847664117813,
300
+ "learning_rate": 2.6464646464646466e-05,
301
+ "loss": 0.0019,
302
+ "step": 290
303
+ },
304
+ {
305
+ "epoch": 5.45,
306
+ "grad_norm": 0.018668798729777336,
307
+ "learning_rate": 2.5454545454545454e-05,
308
+ "loss": 0.0017,
309
+ "step": 300
310
+ },
311
+ {
312
+ "epoch": 5.64,
313
+ "grad_norm": 0.020943326875567436,
314
+ "learning_rate": 2.4444444444444445e-05,
315
+ "loss": 0.0017,
316
+ "step": 310
317
+ },
318
+ {
319
+ "epoch": 5.82,
320
+ "grad_norm": 0.01833685301244259,
321
+ "learning_rate": 2.3434343434343436e-05,
322
+ "loss": 0.0017,
323
+ "step": 320
324
+ },
325
+ {
326
+ "epoch": 6.0,
327
+ "grad_norm": 0.01634540781378746,
328
+ "learning_rate": 2.2424242424242424e-05,
329
+ "loss": 0.0016,
330
+ "step": 330
331
+ },
332
+ {
333
+ "epoch": 6.0,
334
+ "eval_accuracy": 0.9908675799086758,
335
+ "eval_f1_macro": 0.9910410929202866,
336
+ "eval_f1_micro": 0.9908675799086758,
337
+ "eval_f1_weighted": 0.9908473335613555,
338
+ "eval_loss": 0.03381425514817238,
339
+ "eval_precision_macro": 0.9909727371947719,
340
+ "eval_precision_micro": 0.9908675799086758,
341
+ "eval_precision_weighted": 0.9908883151237302,
342
+ "eval_recall_macro": 0.9911698494022667,
343
+ "eval_recall_micro": 0.9908675799086758,
344
+ "eval_recall_weighted": 0.9908675799086758,
345
+ "eval_runtime": 0.124,
346
+ "eval_samples_per_second": 3531.36,
347
+ "eval_steps_per_second": 56.437,
348
+ "step": 330
349
+ }
350
+ ],
351
+ "logging_steps": 10,
352
+ "max_steps": 550,
353
+ "num_input_tokens_seen": 0,
354
+ "num_train_epochs": 10,
355
+ "save_steps": 500,
356
+ "total_flos": 349726407917568.0,
357
+ "train_batch_size": 8,
358
+ "trial_name": null,
359
+ "trial_params": null
360
+ }
checkpoint-330/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04f553a8a0961210425cec5f73e97d6c4a5dd830d0a0db2a666c690b6a557860
3
+ size 4920
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert/distilbert-base-uncased",
3
+ "_num_labels": 4,
4
+ "activation": "gelu",
5
+ "architectures": [
6
+ "DistilBertForSequenceClassification"
7
+ ],
8
+ "attention_dropout": 0.1,
9
+ "dim": 768,
10
+ "dropout": 0.1,
11
+ "hidden_dim": 3072,
12
+ "id2label": {
13
+ "0": "date-based question answering",
14
+ "1": "math",
15
+ "2": "question-answering",
16
+ "3": "summarization"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "label2id": {
20
+ "date-based question answering": 0,
21
+ "math": 1,
22
+ "question-answering": 2,
23
+ "summarization": 3
24
+ },
25
+ "max_position_embeddings": 512,
26
+ "model_type": "distilbert",
27
+ "n_heads": 12,
28
+ "n_layers": 6,
29
+ "pad_token_id": 0,
30
+ "problem_type": "single_label_classification",
31
+ "qa_dropout": 0.1,
32
+ "seq_classif_dropout": 0.2,
33
+ "sinusoidal_pos_embds": false,
34
+ "tie_weights_": true,
35
+ "torch_dtype": "float32",
36
+ "transformers_version": "4.38.2",
37
+ "vocab_size": 30522
38
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b05d2e2cf016ebcfef096093bda1964ca91f544d3e074b6188b3cf95389a4f7
3
+ size 267838720
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 1000000000000000019884624838656,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04f553a8a0961210425cec5f73e97d6c4a5dd830d0a0db2a666c690b6a557860
3
+ size 4920
training_params.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "autotrain-pmf0g-rj8fa/autotrain-data",
3
+ "model": "distilbert/distilbert-base-uncased",
4
+ "lr": 5e-05,
5
+ "epochs": 10,
6
+ "max_seq_length": 128,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "text_column": "autotrain_text",
18
+ "target_column": "autotrain_label",
19
+ "logging_steps": -1,
20
+ "project_name": "autotrain-pmf0g-rj8fa",
21
+ "auto_find_batch_size": false,
22
+ "mixed_precision": "fp16",
23
+ "save_total_limit": 1,
24
+ "save_strategy": "epoch",
25
+ "push_to_hub": true,
26
+ "repo_id": "echoctx/autotrain-pmf0g-rj8fa",
27
+ "evaluation_strategy": "epoch",
28
+ "username": "echoctx",
29
+ "log": "none"
30
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff