llama v2 causal
Browse filesUsing llama chat v1 1.1B
- README.md +5 -20
- adapter_config.json +6 -2
- adapter_model.safetensors +2 -2
- added_tokens.json +5 -4
- special_tokens_map.json +1 -1
- tokenizer.json +111 -102
- tokenizer_config.json +19 -7
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
-
base_model:
|
4 |
---
|
5 |
|
6 |
# Model Card for Model ID
|
@@ -18,6 +18,7 @@ base_model: TinyPixel/Llama-2-7B-bf16-sharded
|
|
18 |
|
19 |
|
20 |
- **Developed by:** [More Information Needed]
|
|
|
21 |
- **Shared by [optional]:** [More Information Needed]
|
22 |
- **Model type:** [More Information Needed]
|
23 |
- **Language(s) (NLP):** [More Information Needed]
|
@@ -76,7 +77,7 @@ Use the code below to get started with the model.
|
|
76 |
|
77 |
### Training Data
|
78 |
|
79 |
-
<!-- This should link to a
|
80 |
|
81 |
[More Information Needed]
|
82 |
|
@@ -107,7 +108,7 @@ Use the code below to get started with the model.
|
|
107 |
|
108 |
#### Testing Data
|
109 |
|
110 |
-
<!-- This should link to a
|
111 |
|
112 |
[More Information Needed]
|
113 |
|
@@ -198,22 +199,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
198 |
[More Information Needed]
|
199 |
|
200 |
|
201 |
-
## Training procedure
|
202 |
-
|
203 |
-
|
204 |
-
The following `bitsandbytes` quantization config was used during training:
|
205 |
-
- quant_method: bitsandbytes
|
206 |
-
- load_in_8bit: False
|
207 |
-
- load_in_4bit: True
|
208 |
-
- llm_int8_threshold: 6.0
|
209 |
-
- llm_int8_skip_modules: None
|
210 |
-
- llm_int8_enable_fp32_cpu_offload: False
|
211 |
-
- llm_int8_has_fp16_weight: False
|
212 |
-
- bnb_4bit_quant_type: nf4
|
213 |
-
- bnb_4bit_use_double_quant: False
|
214 |
-
- bnb_4bit_compute_dtype: float16
|
215 |
-
|
216 |
### Framework versions
|
217 |
|
218 |
-
|
219 |
-
- PEFT 0.6.3.dev0
|
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
+
base_model: PY007/TinyLlama-1.1B-Chat-v0.1
|
4 |
---
|
5 |
|
6 |
# Model Card for Model ID
|
|
|
18 |
|
19 |
|
20 |
- **Developed by:** [More Information Needed]
|
21 |
+
- **Funded by [optional]:** [More Information Needed]
|
22 |
- **Shared by [optional]:** [More Information Needed]
|
23 |
- **Model type:** [More Information Needed]
|
24 |
- **Language(s) (NLP):** [More Information Needed]
|
|
|
77 |
|
78 |
### Training Data
|
79 |
|
80 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
81 |
|
82 |
[More Information Needed]
|
83 |
|
|
|
108 |
|
109 |
#### Testing Data
|
110 |
|
111 |
+
<!-- This should link to a Dataset Card if possible. -->
|
112 |
|
113 |
[More Information Needed]
|
114 |
|
|
|
199 |
[More Information Needed]
|
200 |
|
201 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
202 |
### Framework versions
|
203 |
|
204 |
+
- PEFT 0.7.2.dev0
|
|
adapter_config.json
CHANGED
@@ -1,15 +1,18 @@
|
|
1 |
{
|
2 |
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
-
"base_model_name_or_path": "
|
5 |
"bias": "none",
|
6 |
"fan_in_fan_out": false,
|
7 |
"inference_mode": true,
|
8 |
"init_lora_weights": true,
|
9 |
"layers_pattern": null,
|
10 |
"layers_to_transform": null,
|
|
|
11 |
"lora_alpha": 16,
|
12 |
"lora_dropout": 0.05,
|
|
|
|
|
13 |
"modules_to_save": [
|
14 |
"embed_tokens",
|
15 |
"lm_head"
|
@@ -23,5 +26,6 @@
|
|
23 |
"q_proj",
|
24 |
"k_proj"
|
25 |
],
|
26 |
-
"task_type": "CAUSAL_LM"
|
|
|
27 |
}
|
|
|
1 |
{
|
2 |
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "PY007/TinyLlama-1.1B-Chat-v0.1",
|
5 |
"bias": "none",
|
6 |
"fan_in_fan_out": false,
|
7 |
"inference_mode": true,
|
8 |
"init_lora_weights": true,
|
9 |
"layers_pattern": null,
|
10 |
"layers_to_transform": null,
|
11 |
+
"loftq_config": {},
|
12 |
"lora_alpha": 16,
|
13 |
"lora_dropout": 0.05,
|
14 |
+
"megatron_config": null,
|
15 |
+
"megatron_core": "megatron.core",
|
16 |
"modules_to_save": [
|
17 |
"embed_tokens",
|
18 |
"lm_head"
|
|
|
26 |
"q_proj",
|
27 |
"k_proj"
|
28 |
],
|
29 |
+
"task_type": "CAUSAL_LM",
|
30 |
+
"use_rslora": false
|
31 |
}
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e64a3ccac118077578a7f2674048ff2cb7425c91e068459fa9d9bb10c925add2
|
3 |
+
size 536643392
|
added_tokens.json
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
{
|
2 |
-
"<endstring>":
|
3 |
-
"<feeling>":
|
4 |
-
"<pad>":
|
5 |
-
"<symptom>":
|
|
|
6 |
}
|
|
|
1 |
{
|
2 |
+
"<endstring>": 32003,
|
3 |
+
"<feeling>": 32002,
|
4 |
+
"<pad>": 32001,
|
5 |
+
"<symptom>": 32004,
|
6 |
+
"[PAD]": 32000
|
7 |
}
|
special_tokens_map.json
CHANGED
@@ -23,7 +23,7 @@
|
|
23 |
"unk_token": {
|
24 |
"content": "<unk>",
|
25 |
"lstrip": false,
|
26 |
-
"normalized":
|
27 |
"rstrip": false,
|
28 |
"single_word": false
|
29 |
}
|
|
|
23 |
"unk_token": {
|
24 |
"content": "<unk>",
|
25 |
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
"rstrip": false,
|
28 |
"single_word": false
|
29 |
}
|
tokenizer.json
CHANGED
@@ -9,7 +9,7 @@
|
|
9 |
"single_word": false,
|
10 |
"lstrip": false,
|
11 |
"rstrip": false,
|
12 |
-
"normalized":
|
13 |
"special": true
|
14 |
},
|
15 |
{
|
@@ -18,7 +18,7 @@
|
|
18 |
"single_word": false,
|
19 |
"lstrip": false,
|
20 |
"rstrip": false,
|
21 |
-
"normalized":
|
22 |
"special": true
|
23 |
},
|
24 |
{
|
@@ -27,12 +27,12 @@
|
|
27 |
"single_word": false,
|
28 |
"lstrip": false,
|
29 |
"rstrip": false,
|
30 |
-
"normalized":
|
31 |
"special": true
|
32 |
},
|
33 |
{
|
34 |
"id": 32000,
|
35 |
-
"content": "
|
36 |
"single_word": false,
|
37 |
"lstrip": false,
|
38 |
"rstrip": false,
|
@@ -41,7 +41,7 @@
|
|
41 |
},
|
42 |
{
|
43 |
"id": 32001,
|
44 |
-
"content": "<
|
45 |
"single_word": false,
|
46 |
"lstrip": false,
|
47 |
"rstrip": false,
|
@@ -50,7 +50,7 @@
|
|
50 |
},
|
51 |
{
|
52 |
"id": 32002,
|
53 |
-
"content": "<
|
54 |
"single_word": false,
|
55 |
"lstrip": false,
|
56 |
"rstrip": false,
|
@@ -59,6 +59,15 @@
|
|
59 |
},
|
60 |
{
|
61 |
"id": 32003,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
"content": "<symptom>",
|
63 |
"single_word": false,
|
64 |
"lstrip": false,
|
@@ -93302,126 +93311,126 @@
|
|
93302 |
"βlivre s",
|
93303 |
"lu b",
|
93304 |
"l ub",
|
|
|
93305 |
"ββ ββ",
|
93306 |
-
"
|
93307 |
-
"
|
93308 |
-
"ββ βββββ",
|
93309 |
"ββ ββββββ",
|
93310 |
-
"ββ ββββββββββββ",
|
93311 |
-
"ββ βββββββββββββ",
|
93312 |
-
"ββ ββββββββββ",
|
93313 |
-
"ββ ββββββββββββββ",
|
93314 |
-
"ββ βββ",
|
93315 |
-
"ββ βββββββββ",
|
93316 |
-
"ββ βββββββ",
|
93317 |
-
"ββ βββββββββββ",
|
93318 |
-
"ββ β",
|
93319 |
-
"ββββ ββ",
|
93320 |
"ββββ ββββ",
|
93321 |
-
"
|
93322 |
-
"
|
93323 |
-
"
|
93324 |
-
"
|
93325 |
-
"
|
93326 |
-
"
|
93327 |
-
"ββββ βββββββββ",
|
93328 |
-
"ββββ βββββββ",
|
93329 |
-
"ββββ βββββββββββ",
|
93330 |
"ββββ β",
|
93331 |
-
"
|
93332 |
-
"
|
|
|
|
|
93333 |
"ββββββββ ββββββββ",
|
93334 |
-
"ββββββββ βββββ",
|
93335 |
-
"ββββββββ ββββββ",
|
93336 |
-
"ββββββββ βββ",
|
93337 |
-
"ββββββββ βββββββ",
|
93338 |
-
"ββββββββ β",
|
93339 |
-
"βββββ ββ",
|
93340 |
-
"βββββ ββββ",
|
93341 |
-
"βββββ ββββββββ",
|
93342 |
-
"βββββ βββββ",
|
93343 |
-
"βββββ ββββββ",
|
93344 |
-
"βββββ ββββββββββ",
|
93345 |
-
"βββββ βββ",
|
93346 |
-
"βββββ βββββββββ",
|
93347 |
-
"βββββ βββββββ",
|
93348 |
"βββββ βββββββββββ",
|
93349 |
-
"βββββ β",
|
93350 |
-
"ββββββ ββ",
|
93351 |
-
"ββββββ ββββ",
|
93352 |
-
"ββββββ ββββββββ",
|
93353 |
-
"ββββββ βββββ",
|
93354 |
-
"ββββββ ββββββ",
|
93355 |
"ββββββ ββββββββββ",
|
93356 |
-
"ββββββ βββ",
|
93357 |
-
"ββββββ βββββββββ",
|
93358 |
-
"ββββββ βββββββ",
|
93359 |
-
"ββββββ β",
|
93360 |
-
"ββββββββββββ ββ",
|
93361 |
"ββββββββββββ ββββ",
|
93362 |
-
"ββββββββββββ βββ",
|
93363 |
-
"ββββββββββββ β",
|
93364 |
-
"βββββββββββββ ββ",
|
93365 |
"βββββββββββββ βββ",
|
93366 |
-
"βββββββββββββ β",
|
93367 |
-
"ββββββββββ ββ",
|
93368 |
-
"ββββββββββ ββββ",
|
93369 |
-
"ββββββββββ βββββ",
|
93370 |
"ββββββββββ ββββββ",
|
93371 |
-
"ββββββββββ βββ",
|
93372 |
-
"ββββββββββ β",
|
93373 |
"οΏ½οΏ½βββββββββββββ ββ",
|
93374 |
-
"ββββββββββββββ β",
|
93375 |
-
"βββ ββ",
|
93376 |
-
"βββ ββββ",
|
93377 |
-
"βββ ββββββββ",
|
93378 |
-
"βββ βββββ",
|
93379 |
-
"βββ ββββββ",
|
93380 |
-
"βββ ββββββββββββ",
|
93381 |
"βββ βββββββββββββ",
|
93382 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93383 |
"βββ βββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93384 |
"βββ βββββββββ",
|
93385 |
-
"βββ βββββββ",
|
93386 |
-
"βββ βββββββββββ",
|
93387 |
-
"βββ β",
|
93388 |
-
"βββββββββ ββ",
|
93389 |
-
"βββββββββ ββββ",
|
93390 |
-
"βββββββββ βββββ",
|
93391 |
-
"βββββββββ ββββββ",
|
93392 |
"βββββββββ βββ",
|
93393 |
-
"βββββββββ βββββββ",
|
93394 |
-
"βββββββββ β",
|
93395 |
-
"βββββββ ββ",
|
93396 |
-
"βββββββ ββββ",
|
93397 |
-
"βββββββ ββββββββ",
|
93398 |
"βββββββ βββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93399 |
"βββββββ ββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93400 |
"βββββββ βββ",
|
93401 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93402 |
"βββββββ βββββββ",
|
93403 |
-
"βββββββ β",
|
93404 |
-
"βββββββββββ ββ",
|
93405 |
-
"βββββββββββ ββββ",
|
93406 |
-
"βββββββββββ βββββ",
|
93407 |
"βββββββββββ βββ",
|
93408 |
-
"
|
93409 |
-
"
|
93410 |
"β ββ",
|
93411 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
93412 |
"β ββββββββ",
|
93413 |
-
"
|
|
|
|
|
|
|
|
|
93414 |
"β ββββββ",
|
93415 |
-
"
|
93416 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93417 |
"β ββββββββββ",
|
93418 |
-
"
|
93419 |
-
"
|
93420 |
-
"
|
93421 |
-
"
|
93422 |
-
"
|
93423 |
-
"
|
93424 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93425 |
]
|
93426 |
}
|
93427 |
}
|
|
|
9 |
"single_word": false,
|
10 |
"lstrip": false,
|
11 |
"rstrip": false,
|
12 |
+
"normalized": false,
|
13 |
"special": true
|
14 |
},
|
15 |
{
|
|
|
18 |
"single_word": false,
|
19 |
"lstrip": false,
|
20 |
"rstrip": false,
|
21 |
+
"normalized": false,
|
22 |
"special": true
|
23 |
},
|
24 |
{
|
|
|
27 |
"single_word": false,
|
28 |
"lstrip": false,
|
29 |
"rstrip": false,
|
30 |
+
"normalized": false,
|
31 |
"special": true
|
32 |
},
|
33 |
{
|
34 |
"id": 32000,
|
35 |
+
"content": "[PAD]",
|
36 |
"single_word": false,
|
37 |
"lstrip": false,
|
38 |
"rstrip": false,
|
|
|
41 |
},
|
42 |
{
|
43 |
"id": 32001,
|
44 |
+
"content": "<pad>",
|
45 |
"single_word": false,
|
46 |
"lstrip": false,
|
47 |
"rstrip": false,
|
|
|
50 |
},
|
51 |
{
|
52 |
"id": 32002,
|
53 |
+
"content": "<feeling>",
|
54 |
"single_word": false,
|
55 |
"lstrip": false,
|
56 |
"rstrip": false,
|
|
|
59 |
},
|
60 |
{
|
61 |
"id": 32003,
|
62 |
+
"content": "<endstring>",
|
63 |
+
"single_word": false,
|
64 |
+
"lstrip": false,
|
65 |
+
"rstrip": false,
|
66 |
+
"normalized": false,
|
67 |
+
"special": true
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"id": 32004,
|
71 |
"content": "<symptom>",
|
72 |
"single_word": false,
|
73 |
"lstrip": false,
|
|
|
93311 |
"βlivre s",
|
93312 |
"lu b",
|
93313 |
"l ub",
|
93314 |
+
"β β",
|
93315 |
"ββ ββ",
|
93316 |
+
"βββ β",
|
93317 |
+
"β βββ",
|
|
|
93318 |
"ββ ββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93319 |
"ββββ ββββ",
|
93320 |
+
"βββββ βββ",
|
93321 |
+
"ββββββ ββ",
|
93322 |
+
"βββ βββββ",
|
93323 |
+
"βββββββ β",
|
93324 |
+
"β βββββββ",
|
93325 |
+
"ββ βββ",
|
|
|
|
|
|
|
93326 |
"ββββ β",
|
93327 |
+
"βββ ββ",
|
93328 |
+
"β ββββ",
|
93329 |
+
"ββ ββββββββββββββ",
|
93330 |
+
"ββββ ββββββββββββ",
|
93331 |
"ββββββββ ββββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93332 |
"βββββ βββββββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
93333 |
"ββββββ ββββββββββ",
|
|
|
|
|
|
|
|
|
|
|
93334 |
"ββββββββββββ ββββ",
|
|
|
|
|
|
|
93335 |
"βββββββββββββ βββ",
|
|
|
|
|
|
|
|
|
93336 |
"ββββββββββ ββββββ",
|
|
|
|
|
93337 |
"οΏ½οΏ½βββββββββββββ ββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93338 |
"βββ βββββββββββββ",
|
93339 |
+
"βββββββββ βββββββ",
|
93340 |
+
"βββββββ βββββββββ",
|
93341 |
+
"βββββββββββ βββββ",
|
93342 |
+
"βββββββββββββββ β",
|
93343 |
+
"β βββββββββββββββ",
|
93344 |
+
"ββ ββββ",
|
93345 |
+
"ββββ ββ",
|
93346 |
+
"βββββ β",
|
93347 |
"βββ βββ",
|
93348 |
+
"β βββββ",
|
93349 |
+
"ββ ββββββββββ",
|
93350 |
+
"ββββ ββββββββ",
|
93351 |
+
"ββββββββ ββββ",
|
93352 |
+
"βββββ βββββββ",
|
93353 |
+
"ββββββ ββββββ",
|
93354 |
+
"ββββββββββ ββ",
|
93355 |
"βββ βββββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93356 |
"βββββββββ βββ",
|
|
|
|
|
|
|
|
|
|
|
93357 |
"βββββββ βββββ",
|
93358 |
+
"βββββββββββ β",
|
93359 |
+
"β βββββββββββ",
|
93360 |
+
"ββ βββββββββββ",
|
93361 |
+
"ββββ βββββββββ",
|
93362 |
+
"ββββββββ βββββ",
|
93363 |
+
"βββββ ββββββββ",
|
93364 |
+
"ββββββ βββββββ",
|
93365 |
+
"ββββββββββββ β",
|
93366 |
+
"ββββββββββ βββ",
|
93367 |
+
"βββ ββββββββββ",
|
93368 |
+
"βββββββββ ββββ",
|
93369 |
"βββββββ ββββββ",
|
93370 |
+
"βββββββββββ ββ",
|
93371 |
+
"β ββββββββββββ",
|
93372 |
+
"ββ ββββββββ",
|
93373 |
+
"ββββ ββββββ",
|
93374 |
+
"ββββββββ ββ",
|
93375 |
+
"βββββ βββββ",
|
93376 |
+
"ββββββ ββββ",
|
93377 |
+
"βββ βββββββ",
|
93378 |
+
"βββββββββ β",
|
93379 |
"βββββββ βββ",
|
93380 |
+
"β βββββββββ",
|
93381 |
+
"ββ ββββββββββββ",
|
93382 |
+
"ββββ ββββββββββ",
|
93383 |
+
"ββββββββ ββββββ",
|
93384 |
+
"βββββ βββββββββ",
|
93385 |
+
"ββββββ ββββββββ",
|
93386 |
+
"ββββββββββββ ββ",
|
93387 |
+
"βββββββββββββ β",
|
93388 |
+
"ββββββββββ ββββ",
|
93389 |
+
"βββ βββββββββββ",
|
93390 |
+
"βββββββββ βββββ",
|
93391 |
"βββββββ βββββββ",
|
|
|
|
|
|
|
|
|
93392 |
"βββββββββββ βββ",
|
93393 |
+
"β βββββββββββββ",
|
93394 |
+
"ββ β",
|
93395 |
"β ββ",
|
93396 |
+
"ββ βββββββ",
|
93397 |
+
"ββββ βββββ",
|
93398 |
+
"ββββββββ β",
|
93399 |
+
"βββββ ββββ",
|
93400 |
+
"ββββββ βββ",
|
93401 |
+
"βββ ββββββ",
|
93402 |
+
"βββββββ ββ",
|
93403 |
"β ββββββββ",
|
93404 |
+
"ββ βββββ",
|
93405 |
+
"ββββ βββ",
|
93406 |
+
"βββββ ββ",
|
93407 |
+
"ββββββ β",
|
93408 |
+
"βββ ββββ",
|
93409 |
"β ββββββ",
|
93410 |
+
"ββ βββββββββ",
|
93411 |
+
"ββββ βββββββ",
|
93412 |
+
"ββββββββ βββ",
|
93413 |
+
"βββββ ββββββ",
|
93414 |
+
"ββββββ βββββ",
|
93415 |
+
"ββββββββββ β",
|
93416 |
+
"βββ ββββββββ",
|
93417 |
+
"βββββοΏ½οΏ½οΏ½βββ ββ",
|
93418 |
+
"βββββββ ββββ",
|
93419 |
"β ββββββββββ",
|
93420 |
+
"ββ βββββββββββββ",
|
93421 |
+
"ββββ βββββββββββ",
|
93422 |
+
"ββββββββ βββββββ",
|
93423 |
+
"βββββ ββββββββββ",
|
93424 |
+
"ββββββ βββββββββ",
|
93425 |
+
"ββββββββββββ βββ",
|
93426 |
+
"βββββββββββββ ββ",
|
93427 |
+
"ββββββββββ βββββ",
|
93428 |
+
"ββββββββββββββ β",
|
93429 |
+
"βββ ββββββββββββ",
|
93430 |
+
"βββββββββ ββββββ",
|
93431 |
+
"βββββββ ββββββββ",
|
93432 |
+
"βββββββββββ ββββ",
|
93433 |
+
"β ββββββββββββββ"
|
93434 |
]
|
93435 |
}
|
93436 |
}
|
tokenizer_config.json
CHANGED
@@ -1,9 +1,11 @@
|
|
1 |
{
|
|
|
|
|
2 |
"added_tokens_decoder": {
|
3 |
"0": {
|
4 |
"content": "<unk>",
|
5 |
"lstrip": false,
|
6 |
-
"normalized":
|
7 |
"rstrip": false,
|
8 |
"single_word": false,
|
9 |
"special": true
|
@@ -11,7 +13,7 @@
|
|
11 |
"1": {
|
12 |
"content": "<s>",
|
13 |
"lstrip": false,
|
14 |
-
"normalized":
|
15 |
"rstrip": false,
|
16 |
"single_word": false,
|
17 |
"special": true
|
@@ -19,13 +21,13 @@
|
|
19 |
"2": {
|
20 |
"content": "</s>",
|
21 |
"lstrip": false,
|
22 |
-
"normalized":
|
23 |
"rstrip": false,
|
24 |
"single_word": false,
|
25 |
"special": true
|
26 |
},
|
27 |
"32000": {
|
28 |
-
"content": "
|
29 |
"lstrip": false,
|
30 |
"normalized": false,
|
31 |
"rstrip": false,
|
@@ -33,7 +35,7 @@
|
|
33 |
"special": true
|
34 |
},
|
35 |
"32001": {
|
36 |
-
"content": "<
|
37 |
"lstrip": false,
|
38 |
"normalized": false,
|
39 |
"rstrip": false,
|
@@ -41,7 +43,7 @@
|
|
41 |
"special": true
|
42 |
},
|
43 |
"32002": {
|
44 |
-
"content": "<
|
45 |
"lstrip": false,
|
46 |
"normalized": false,
|
47 |
"rstrip": false,
|
@@ -49,6 +51,14 @@
|
|
49 |
"special": true
|
50 |
},
|
51 |
"32003": {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
"content": "<symptom>",
|
53 |
"lstrip": false,
|
54 |
"normalized": true,
|
@@ -60,10 +70,12 @@
|
|
60 |
"bos_token": "<feeling>",
|
61 |
"clean_up_tokenization_spaces": false,
|
62 |
"eos_token": "<endstring>",
|
|
|
63 |
"model_max_length": 1000000000000000019884624838656,
|
64 |
"pad_token": "<pad>",
|
|
|
65 |
"sp_model_kwargs": {},
|
66 |
"tokenizer_class": "LlamaTokenizer",
|
67 |
"unk_token": "<unk>",
|
68 |
-
"use_default_system_prompt":
|
69 |
}
|
|
|
1 |
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
"added_tokens_decoder": {
|
5 |
"0": {
|
6 |
"content": "<unk>",
|
7 |
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
"rstrip": false,
|
10 |
"single_word": false,
|
11 |
"special": true
|
|
|
13 |
"1": {
|
14 |
"content": "<s>",
|
15 |
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
"rstrip": false,
|
18 |
"single_word": false,
|
19 |
"special": true
|
|
|
21 |
"2": {
|
22 |
"content": "</s>",
|
23 |
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
"rstrip": false,
|
26 |
"single_word": false,
|
27 |
"special": true
|
28 |
},
|
29 |
"32000": {
|
30 |
+
"content": "[PAD]",
|
31 |
"lstrip": false,
|
32 |
"normalized": false,
|
33 |
"rstrip": false,
|
|
|
35 |
"special": true
|
36 |
},
|
37 |
"32001": {
|
38 |
+
"content": "<pad>",
|
39 |
"lstrip": false,
|
40 |
"normalized": false,
|
41 |
"rstrip": false,
|
|
|
43 |
"special": true
|
44 |
},
|
45 |
"32002": {
|
46 |
+
"content": "<feeling>",
|
47 |
"lstrip": false,
|
48 |
"normalized": false,
|
49 |
"rstrip": false,
|
|
|
51 |
"special": true
|
52 |
},
|
53 |
"32003": {
|
54 |
+
"content": "<endstring>",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": false,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false,
|
59 |
+
"special": true
|
60 |
+
},
|
61 |
+
"32004": {
|
62 |
"content": "<symptom>",
|
63 |
"lstrip": false,
|
64 |
"normalized": true,
|
|
|
70 |
"bos_token": "<feeling>",
|
71 |
"clean_up_tokenization_spaces": false,
|
72 |
"eos_token": "<endstring>",
|
73 |
+
"legacy": false,
|
74 |
"model_max_length": 1000000000000000019884624838656,
|
75 |
"pad_token": "<pad>",
|
76 |
+
"padding_side": "right",
|
77 |
"sp_model_kwargs": {},
|
78 |
"tokenizer_class": "LlamaTokenizer",
|
79 |
"unk_token": "<unk>",
|
80 |
+
"use_default_system_prompt": true
|
81 |
}
|