mapama247 commited on
Commit
915389f
·
1 Parent(s): a318635

upload nace2_level1_26 model

Browse files
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../models/roberta-large/",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "finetuning_task": "nace",
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 1024,
14
+ "id2label": {
15
+ "0": "1",
16
+ "1": "2",
17
+ "2": "3",
18
+ "3": "4",
19
+ "4": "5",
20
+ "5": "6",
21
+ "6": "7",
22
+ "7": "8"
23
+ },
24
+ "initializer_range": 0.02,
25
+ "intermediate_size": 4096,
26
+ "label2id": {
27
+ "1": 0,
28
+ "2": 1,
29
+ "3": 2,
30
+ "4": 3,
31
+ "5": 4,
32
+ "6": 5,
33
+ "7": 6,
34
+ "8": 7
35
+ },
36
+ "layer_norm_eps": 1e-05,
37
+ "max_position_embeddings": 514,
38
+ "model_type": "roberta",
39
+ "num_attention_heads": 16,
40
+ "num_hidden_layers": 24,
41
+ "pad_token_id": 1,
42
+ "position_embedding_type": "absolute",
43
+ "problem_type": "multi_label_classification",
44
+ "transformers_version": "4.6.1",
45
+ "type_vocab_size": 1,
46
+ "use_cache": true,
47
+ "vocab_size": 50265
48
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c2e7b7cb4a44ca07738f0fc77ed553d09db0bdd88cec362f2b6f63fe3b8e76a
3
+ size 1421635885
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a09352109e9ac36d6d9075a8cad7cf9bdeab4cc85d8c706d0500f2115574986
3
+ size 15523
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "special_tokens_map_file": null, "name_or_path": "../models/roberta-large/"}
trainer_state.json ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8567621097797157,
3
+ "best_model_checkpoint": "./output//26_roberta-large_nace_5__5e-6_0.01_0.06_07-21-22_10-40/checkpoint-20000",
4
+ "epoch": 0.09387489761768979,
5
+ "global_step": 20000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.01,
12
+ "learning_rate": 1.5645779550966127e-07,
13
+ "loss": 0.4615,
14
+ "step": 2000
15
+ },
16
+ {
17
+ "epoch": 0.01,
18
+ "eval_accuracy": 0.15318914123679223,
19
+ "eval_f1": 0.27608921192458447,
20
+ "eval_loss": 0.28353968262672424,
21
+ "eval_roc_auc": 0.5800321310271234,
22
+ "eval_runtime": 11156.151,
23
+ "eval_samples_per_second": 33.95,
24
+ "step": 2000
25
+ },
26
+ {
27
+ "epoch": 0.02,
28
+ "learning_rate": 3.1291559101932254e-07,
29
+ "loss": 0.231,
30
+ "step": 4000
31
+ },
32
+ {
33
+ "epoch": 0.02,
34
+ "eval_accuracy": 0.6693737887916695,
35
+ "eval_f1": 0.7682238795441866,
36
+ "eval_loss": 0.17183057963848114,
37
+ "eval_roc_auc": 0.8419285809208303,
38
+ "eval_runtime": 11148.4174,
39
+ "eval_samples_per_second": 33.974,
40
+ "step": 4000
41
+ },
42
+ {
43
+ "epoch": 0.03,
44
+ "learning_rate": 4.6937338652898386e-07,
45
+ "loss": 0.1639,
46
+ "step": 6000
47
+ },
48
+ {
49
+ "epoch": 0.03,
50
+ "eval_accuracy": 0.7266088278935668,
51
+ "eval_f1": 0.8098187138286067,
52
+ "eval_loss": 0.14115960896015167,
53
+ "eval_roc_auc": 0.8728687872480946,
54
+ "eval_runtime": 11147.6817,
55
+ "eval_samples_per_second": 33.976,
56
+ "step": 6000
57
+ },
58
+ {
59
+ "epoch": 0.04,
60
+ "learning_rate": 6.258311820386451e-07,
61
+ "loss": 0.1414,
62
+ "step": 8000
63
+ },
64
+ {
65
+ "epoch": 0.04,
66
+ "eval_accuracy": 0.7431657487445678,
67
+ "eval_f1": 0.8278856538317427,
68
+ "eval_loss": 0.12584036588668823,
69
+ "eval_roc_auc": 0.8876332405370599,
70
+ "eval_runtime": 11150.482,
71
+ "eval_samples_per_second": 33.968,
72
+ "step": 8000
73
+ },
74
+ {
75
+ "epoch": 0.05,
76
+ "learning_rate": 7.822889775483064e-07,
77
+ "loss": 0.1283,
78
+ "step": 10000
79
+ },
80
+ {
81
+ "epoch": 0.05,
82
+ "eval_accuracy": 0.7535894010360287,
83
+ "eval_f1": 0.8360526710766673,
84
+ "eval_loss": 0.11795546859502792,
85
+ "eval_roc_auc": 0.8932732282506214,
86
+ "eval_runtime": 11151.9405,
87
+ "eval_samples_per_second": 33.963,
88
+ "step": 10000
89
+ },
90
+ {
91
+ "epoch": 0.06,
92
+ "learning_rate": 9.387467730579677e-07,
93
+ "loss": 0.1211,
94
+ "step": 12000
95
+ },
96
+ {
97
+ "epoch": 0.06,
98
+ "eval_accuracy": 0.761145756876495,
99
+ "eval_f1": 0.8441615840183859,
100
+ "eval_loss": 0.11228282749652863,
101
+ "eval_roc_auc": 0.9002726570618076,
102
+ "eval_runtime": 11156.2585,
103
+ "eval_samples_per_second": 33.95,
104
+ "step": 12000
105
+ },
106
+ {
107
+ "epoch": 0.07,
108
+ "learning_rate": 1.095204568567629e-06,
109
+ "loss": 0.1136,
110
+ "step": 14000
111
+ },
112
+ {
113
+ "epoch": 0.07,
114
+ "eval_accuracy": 0.7673318301588894,
115
+ "eval_f1": 0.8490507193175131,
116
+ "eval_loss": 0.10872343182563782,
117
+ "eval_roc_auc": 0.9023344325033343,
118
+ "eval_runtime": 11157.3963,
119
+ "eval_samples_per_second": 33.946,
120
+ "step": 14000
121
+ },
122
+ {
123
+ "epoch": 0.08,
124
+ "learning_rate": 1.2516623640772901e-06,
125
+ "loss": 0.1125,
126
+ "step": 16000
127
+ },
128
+ {
129
+ "epoch": 0.08,
130
+ "eval_accuracy": 0.7712156175248315,
131
+ "eval_f1": 0.8527881220679983,
132
+ "eval_loss": 0.10447360575199127,
133
+ "eval_roc_auc": 0.9056321731592804,
134
+ "eval_runtime": 11156.4417,
135
+ "eval_samples_per_second": 33.949,
136
+ "step": 16000
137
+ },
138
+ {
139
+ "epoch": 0.08,
140
+ "learning_rate": 1.4081201595869516e-06,
141
+ "loss": 0.1085,
142
+ "step": 18000
143
+ },
144
+ {
145
+ "epoch": 0.08,
146
+ "eval_accuracy": 0.7746769671079381,
147
+ "eval_f1": 0.8550148801059216,
148
+ "eval_loss": 0.102951779961586,
149
+ "eval_roc_auc": 0.9065747008981477,
150
+ "eval_runtime": 11163.3791,
151
+ "eval_samples_per_second": 33.928,
152
+ "step": 18000
153
+ },
154
+ {
155
+ "epoch": 0.09,
156
+ "learning_rate": 1.5645779550966128e-06,
157
+ "loss": 0.1063,
158
+ "step": 20000
159
+ },
160
+ {
161
+ "epoch": 0.09,
162
+ "eval_accuracy": 0.7812854781731678,
163
+ "eval_f1": 0.8567621097797157,
164
+ "eval_loss": 0.10163237154483795,
165
+ "eval_roc_auc": 0.904867851842597,
166
+ "eval_runtime": 11164.6828,
167
+ "eval_samples_per_second": 33.924,
168
+ "step": 20000
169
+ }
170
+ ],
171
+ "max_steps": 1065245,
172
+ "num_train_epochs": 5,
173
+ "total_flos": 6.9868180733952e+16,
174
+ "trial_name": null,
175
+ "trial_params": null
176
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6aa0846dd9ca1e77ecbd0d74050468ddc827f500dcb70b87bbc559b37d1e7ca
3
+ size 2479
vocab.json ADDED
The diff for this file is too large to render. See raw diff