aryopg commited on
Commit
069a81b
·
verified ·
1 Parent(s): 32d7b88

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -1,31 +1,25 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "epfl-llm/meditron-7b",
5
  "bias": "none",
6
- "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 16,
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
- "r": 8,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
- "up_proj",
24
- "q_proj",
25
- "down_proj",
26
- "k_proj",
27
- "gate_proj",
28
- "v_proj"
29
  ],
30
- "task_type": "CAUSAL_LM"
31
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "stanford-crfm/BioMedLM",
5
  "bias": "none",
6
+ "fan_in_fan_out": true,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 32,
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "c_attn"
 
 
 
 
 
 
23
  ],
24
+ "task_type": "SEQ_CLS"
25
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0bdb04d25933246176af14be1eeb714946941189d2540499c399afed3b685cc
3
- size 40036488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f2a043e5013a1b6b883749ce2c9bfc8a0ceb433669b4561c143004d7857d950
3
+ size 21000432
added_tokens.json CHANGED
@@ -1,19 +1,3 @@
1
  {
2
- "<CLS>": 32000,
3
- "<EOD>": 32002,
4
- "<MASK>": 32003,
5
- "<PAD>": 32004,
6
- "<SEP>": 32001,
7
- "[/bib]": 32010,
8
- "[/bib_ref]": 32006,
9
- "[/fig]": 32012,
10
- "[/fig_ref]": 32008,
11
- "[/formula]": 32016,
12
- "[/table]": 32014,
13
- "[bib]": 32009,
14
- "[bib_ref]": 32005,
15
- "[fig]": 32011,
16
- "[fig_ref]": 32007,
17
- "[formula]": 32015,
18
- "[table]": 32013
19
  }
 
1
  {
2
+ "<|endoftext|>": 28895
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "michiyasunaga/BioLinkBERT-base",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
@@ -8,14 +8,14 @@
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
- "hidden_size": 768,
12
  "initializer_range": 0.02,
13
- "intermediate_size": 3072,
14
  "layer_norm_eps": 1e-12,
15
  "max_position_embeddings": 512,
16
  "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
  "problem_type": "single_label_classification",
 
1
  {
2
+ "_name_or_path": "michiyasunaga/BioLinkBERT-large",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 1024,
12
  "initializer_range": 0.02,
13
+ "intermediate_size": 4096,
14
  "layer_norm_eps": 1e-12,
15
  "max_position_embeddings": 512,
16
  "model_type": "bert",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
  "problem_type": "single_label_classification",
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63afe1487612581394ca79b22a1082ff01c70e53b3a52c045f8806699543bd09
3
- size 432960488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43aba0e9eab150cefb6e248c29f5300d333e549f65072fa9f683e29c14cbc1eb
3
+ size 1333958568
special_tokens_map.json CHANGED
@@ -1,65 +1,6 @@
1
  {
2
- "additional_special_tokens": [
3
- "[bib_ref]",
4
- "[/bib_ref]",
5
- "[fig_ref]",
6
- "[/fig_ref]",
7
- "[bib]",
8
- "[/bib]",
9
- "[fig]",
10
- "[/fig]",
11
- "[table]",
12
- "[/table]",
13
- "[formula]",
14
- "[/formula]"
15
- ],
16
- "bos_token": {
17
- "content": "<s>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "cls_token": {
24
- "content": "<CLS>",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "eos_token": {
31
- "content": "</s>",
32
- "lstrip": false,
33
- "normalized": true,
34
- "rstrip": false,
35
- "single_word": false
36
- },
37
- "mask_token": {
38
- "content": "<MASK>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false
43
- },
44
- "pad_token": {
45
- "content": "<PAD>",
46
- "lstrip": false,
47
- "normalized": false,
48
- "rstrip": false,
49
- "single_word": false
50
- },
51
- "sep_token": {
52
- "content": "<SEP>",
53
- "lstrip": false,
54
- "normalized": false,
55
- "rstrip": false,
56
- "single_word": false
57
- },
58
- "unk_token": {
59
- "content": "<unk>",
60
- "lstrip": false,
61
- "normalized": true,
62
- "rstrip": false,
63
- "single_word": false
64
- }
65
  }
 
1
  {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "pad_token": "<|endoftext|>",
5
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,161 +1,8 @@
1
  {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
  "added_tokens_decoder": {
5
- "0": {
6
- "content": "<unk>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "1": {
14
- "content": "<s>",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "2": {
22
- "content": "</s>",
23
- "lstrip": false,
24
- "normalized": true,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "32000": {
30
- "content": "<CLS>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "32001": {
38
- "content": "<SEP>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "32002": {
46
- "content": "<EOD>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "32003": {
54
- "content": "<MASK>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
- },
61
- "32004": {
62
- "content": "<PAD>",
63
- "lstrip": false,
64
- "normalized": false,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": true
68
- },
69
- "32005": {
70
- "content": "[bib_ref]",
71
- "lstrip": false,
72
- "normalized": false,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": true
76
- },
77
- "32006": {
78
- "content": "[/bib_ref]",
79
- "lstrip": false,
80
- "normalized": false,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": true
84
- },
85
- "32007": {
86
- "content": "[fig_ref]",
87
- "lstrip": false,
88
- "normalized": false,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": true
92
- },
93
- "32008": {
94
- "content": "[/fig_ref]",
95
- "lstrip": false,
96
- "normalized": false,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": true
100
- },
101
- "32009": {
102
- "content": "[bib]",
103
- "lstrip": false,
104
- "normalized": false,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": true
108
- },
109
- "32010": {
110
- "content": "[/bib]",
111
- "lstrip": false,
112
- "normalized": false,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": true
116
- },
117
- "32011": {
118
- "content": "[fig]",
119
- "lstrip": false,
120
- "normalized": false,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": true
124
- },
125
- "32012": {
126
- "content": "[/fig]",
127
- "lstrip": false,
128
- "normalized": false,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": true
132
- },
133
- "32013": {
134
- "content": "[table]",
135
- "lstrip": false,
136
- "normalized": false,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": true
140
- },
141
- "32014": {
142
- "content": "[/table]",
143
- "lstrip": false,
144
- "normalized": false,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": true
148
- },
149
- "32015": {
150
- "content": "[formula]",
151
- "lstrip": false,
152
- "normalized": false,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": true
156
- },
157
- "32016": {
158
- "content": "[/formula]",
159
  "lstrip": false,
160
  "normalized": false,
161
  "rstrip": false,
@@ -163,31 +10,11 @@
163
  "special": true
164
  }
165
  },
166
- "additional_special_tokens": [
167
- "[bib_ref]",
168
- "[/bib_ref]",
169
- "[fig_ref]",
170
- "[/fig_ref]",
171
- "[bib]",
172
- "[/bib]",
173
- "[fig]",
174
- "[/fig]",
175
- "[table]",
176
- "[/table]",
177
- "[formula]",
178
- "[/formula]"
179
- ],
180
- "bos_token": "<s>",
181
- "clean_up_tokenization_spaces": false,
182
- "cls_token": "<CLS>",
183
- "eos_token": "</s>",
184
- "legacy": true,
185
- "mask_token": "<MASK>",
186
- "model_max_length": 1000000000000000019884624838656,
187
- "pad_token": "<PAD>",
188
- "sep_token": "<SEP>",
189
- "sp_model_kwargs": {},
190
- "tokenizer_class": "LlamaTokenizer",
191
- "unk_token": "<unk>",
192
- "use_default_system_prompt": false
193
  }
 
1
  {
2
+ "add_prefix_space": false,
 
3
  "added_tokens_decoder": {
4
+ "28895": {
5
+ "content": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  "lstrip": false,
7
  "normalized": false,
8
  "rstrip": false,
 
10
  "special": true
11
  }
12
  },
13
+ "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": true,
15
+ "eos_token": "<|endoftext|>",
16
+ "model_max_length": 1024,
17
+ "pad_token": "<|endoftext|>",
18
+ "tokenizer_class": "GPT2Tokenizer",
19
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  }
vocab.json ADDED
The diff for this file is too large to render. See raw diff