Joe99 commited on
Commit
e957d23
β€’
1 Parent(s): 32a6fef

llama v2 causal

Browse files

Using llama chat v1 1.1B

README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  library_name: peft
3
- base_model: TinyPixel/Llama-2-7B-bf16-sharded
4
  ---
5
 
6
  # Model Card for Model ID
@@ -18,6 +18,7 @@ base_model: TinyPixel/Llama-2-7B-bf16-sharded
18
 
19
 
20
  - **Developed by:** [More Information Needed]
 
21
  - **Shared by [optional]:** [More Information Needed]
22
  - **Model type:** [More Information Needed]
23
  - **Language(s) (NLP):** [More Information Needed]
@@ -76,7 +77,7 @@ Use the code below to get started with the model.
76
 
77
  ### Training Data
78
 
79
- <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
80
 
81
  [More Information Needed]
82
 
@@ -107,7 +108,7 @@ Use the code below to get started with the model.
107
 
108
  #### Testing Data
109
 
110
- <!-- This should link to a Data Card if possible. -->
111
 
112
  [More Information Needed]
113
 
@@ -198,22 +199,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
198
  [More Information Needed]
199
 
200
 
201
- ## Training procedure
202
-
203
-
204
- The following `bitsandbytes` quantization config was used during training:
205
- - quant_method: bitsandbytes
206
- - load_in_8bit: False
207
- - load_in_4bit: True
208
- - llm_int8_threshold: 6.0
209
- - llm_int8_skip_modules: None
210
- - llm_int8_enable_fp32_cpu_offload: False
211
- - llm_int8_has_fp16_weight: False
212
- - bnb_4bit_quant_type: nf4
213
- - bnb_4bit_use_double_quant: False
214
- - bnb_4bit_compute_dtype: float16
215
-
216
  ### Framework versions
217
 
218
-
219
- - PEFT 0.6.3.dev0
 
1
  ---
2
  library_name: peft
3
+ base_model: PY007/TinyLlama-1.1B-Chat-v0.1
4
  ---
5
 
6
  # Model Card for Model ID
 
18
 
19
 
20
  - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
  - **Shared by [optional]:** [More Information Needed]
23
  - **Model type:** [More Information Needed]
24
  - **Language(s) (NLP):** [More Information Needed]
 
77
 
78
  ### Training Data
79
 
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
 
82
  [More Information Needed]
83
 
 
108
 
109
  #### Testing Data
110
 
111
+ <!-- This should link to a Dataset Card if possible. -->
112
 
113
  [More Information Needed]
114
 
 
199
  [More Information Needed]
200
 
201
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
202
  ### Framework versions
203
 
204
+ - PEFT 0.7.2.dev0
 
adapter_config.json CHANGED
@@ -1,15 +1,18 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "TinyPixel/Llama-2-7B-bf16-sharded",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
 
11
  "lora_alpha": 16,
12
  "lora_dropout": 0.05,
 
 
13
  "modules_to_save": [
14
  "embed_tokens",
15
  "lm_head"
@@ -23,5 +26,6 @@
23
  "q_proj",
24
  "k_proj"
25
  ],
26
- "task_type": "CAUSAL_LM"
 
27
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "PY007/TinyLlama-1.1B-Chat-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
+ "loftq_config": {},
12
  "lora_alpha": 16,
13
  "lora_dropout": 0.05,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
  "modules_to_save": [
17
  "embed_tokens",
18
  "lm_head"
 
26
  "q_proj",
27
  "k_proj"
28
  ],
29
+ "task_type": "CAUSAL_LM",
30
+ "use_rslora": false
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1014a5e09e723fe63684e8fc666069b76bfd01398cb4310860847064a1729272
3
- size 1099065608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e64a3ccac118077578a7f2674048ff2cb7425c91e068459fa9d9bb10c925add2
3
+ size 536643392
added_tokens.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
- "<endstring>": 32002,
3
- "<feeling>": 32001,
4
- "<pad>": 32000,
5
- "<symptom>": 32003
 
6
  }
 
1
  {
2
+ "<endstring>": 32003,
3
+ "<feeling>": 32002,
4
+ "<pad>": 32001,
5
+ "<symptom>": 32004,
6
+ "[PAD]": 32000
7
  }
special_tokens_map.json CHANGED
@@ -23,7 +23,7 @@
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
26
- "normalized": true,
27
  "rstrip": false,
28
  "single_word": false
29
  }
 
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
26
+ "normalized": false,
27
  "rstrip": false,
28
  "single_word": false
29
  }
tokenizer.json CHANGED
@@ -9,7 +9,7 @@
9
  "single_word": false,
10
  "lstrip": false,
11
  "rstrip": false,
12
- "normalized": true,
13
  "special": true
14
  },
15
  {
@@ -18,7 +18,7 @@
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
- "normalized": true,
22
  "special": true
23
  },
24
  {
@@ -27,12 +27,12 @@
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
30
- "normalized": true,
31
  "special": true
32
  },
33
  {
34
  "id": 32000,
35
- "content": "<pad>",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
@@ -41,7 +41,7 @@
41
  },
42
  {
43
  "id": 32001,
44
- "content": "<feeling>",
45
  "single_word": false,
46
  "lstrip": false,
47
  "rstrip": false,
@@ -50,7 +50,7 @@
50
  },
51
  {
52
  "id": 32002,
53
- "content": "<endstring>",
54
  "single_word": false,
55
  "lstrip": false,
56
  "rstrip": false,
@@ -59,6 +59,15 @@
59
  },
60
  {
61
  "id": 32003,
 
 
 
 
 
 
 
 
 
62
  "content": "<symptom>",
63
  "single_word": false,
64
  "lstrip": false,
@@ -93302,126 +93311,126 @@
93302
  "▁livre s",
93303
  "lu b",
93304
  "l ub",
 
93305
  "▁▁ ▁▁",
93306
- "▁▁ ▁▁▁▁",
93307
- "▁▁ ▁▁▁▁▁▁▁▁",
93308
- "▁▁ ▁▁▁▁▁",
93309
  "▁▁ ▁▁▁▁▁▁",
93310
- "▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93311
- "▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁",
93312
- "▁▁ ▁▁▁▁▁▁▁▁▁▁",
93313
- "▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
93314
- "▁▁ ▁▁▁",
93315
- "▁▁ ▁▁▁▁▁▁▁▁▁",
93316
- "▁▁ ▁▁▁▁▁▁▁",
93317
- "▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93318
- "▁▁ ▁",
93319
- "▁▁▁▁ ▁▁",
93320
  "▁▁▁▁ ▁▁▁▁",
93321
- "▁▁▁▁ ▁▁▁▁▁▁▁▁",
93322
- "▁▁▁▁ ▁▁▁▁▁",
93323
- "▁▁▁▁ ▁▁▁▁▁▁",
93324
- "▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93325
- "▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
93326
- "▁▁▁▁ ▁▁▁",
93327
- "▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93328
- "▁▁▁▁ ▁▁▁▁▁▁▁",
93329
- "▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93330
  "▁▁▁▁ ▁",
93331
- "▁▁▁▁▁▁▁▁ ▁▁",
93332
- "▁▁▁▁▁▁▁▁ ▁▁▁▁",
 
 
93333
  "▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93334
- "▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93335
- "▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93336
- "▁▁▁▁▁▁▁▁ ▁▁▁",
93337
- "▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93338
- "▁▁▁▁▁▁▁▁ ▁",
93339
- "▁▁▁▁▁ ▁▁",
93340
- "▁▁▁▁▁ ▁▁▁▁",
93341
- "▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93342
- "▁▁▁▁▁ ▁▁▁▁▁",
93343
- "▁▁▁▁▁ ▁▁▁▁▁▁",
93344
- "▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
93345
- "▁▁▁▁▁ ▁▁▁",
93346
- "▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93347
- "▁▁▁▁▁ ▁▁▁▁▁▁▁",
93348
  "▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93349
- "▁▁▁▁▁ ▁",
93350
- "▁▁▁▁▁▁ ▁▁",
93351
- "▁▁▁▁▁▁ ▁▁▁▁",
93352
- "▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93353
- "▁▁▁▁▁▁ ▁▁▁▁▁",
93354
- "▁▁▁▁▁▁ ▁▁▁▁▁▁",
93355
  "▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
93356
- "▁▁▁▁▁▁ ▁▁▁",
93357
- "▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93358
- "▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93359
- "▁▁▁▁▁▁ ▁",
93360
- "▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93361
  "▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93362
- "▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93363
- "▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93364
- "▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93365
  "▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93366
- "▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93367
- "▁▁▁▁▁▁▁▁▁▁ ▁▁",
93368
- "▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93369
- "▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93370
  "▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93371
- "▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93372
- "▁▁▁▁▁▁▁▁▁▁ ▁",
93373
  "��▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93374
- "▁▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93375
- "▁▁▁ ▁▁",
93376
- "▁▁▁ ▁▁▁▁",
93377
- "▁▁▁ ▁▁▁▁▁▁▁▁",
93378
- "▁▁▁ ▁▁▁▁▁",
93379
- "▁▁▁ ▁▁▁▁▁▁",
93380
- "▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93381
  "▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁",
93382
- "▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
 
 
 
 
 
 
 
93383
  "▁▁▁ ▁▁▁",
 
 
 
 
 
 
 
93384
  "▁▁▁ ▁▁▁▁▁▁▁▁▁",
93385
- "▁▁▁ ▁▁▁▁▁▁▁",
93386
- "▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93387
- "▁▁▁ ▁",
93388
- "▁▁▁▁▁▁▁▁▁ ▁▁",
93389
- "▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93390
- "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93391
- "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93392
  "▁▁▁▁▁▁▁▁▁ ▁▁▁",
93393
- "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93394
- "▁▁▁▁▁▁▁▁▁ ▁",
93395
- "▁▁▁▁▁▁▁ ▁▁",
93396
- "▁▁▁▁▁▁▁ ▁▁▁▁",
93397
- "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93398
  "▁▁▁▁▁▁▁ ▁▁▁▁▁",
 
 
 
 
 
 
 
 
 
 
 
93399
  "▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
 
 
 
 
 
 
 
 
 
93400
  "▁▁▁▁▁▁▁ ▁▁▁",
93401
- "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
 
 
 
 
 
 
 
 
 
 
93402
  "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93403
- "▁▁▁▁▁▁▁ ▁",
93404
- "▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93405
- "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93406
- "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93407
  "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93408
- "▁▁▁▁▁▁▁▁▁▁▁ ▁",
93409
- "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93410
  "▁ ▁▁",
93411
- "▁ ▁▁▁▁",
 
 
 
 
 
 
93412
  "▁ ▁▁▁▁▁▁▁▁",
93413
- "▁ ▁▁▁▁▁",
 
 
 
 
93414
  "▁ ▁▁▁▁▁▁",
93415
- "▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93416
- "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁",
 
 
 
 
 
 
 
93417
  "▁ ▁▁▁▁▁▁▁▁▁▁",
93418
- "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
93419
- "▁ ▁▁▁",
93420
- "▁ ▁▁▁▁▁▁▁▁▁",
93421
- "▁ ▁▁▁▁▁▁▁",
93422
- "▁ ▁▁▁▁▁▁▁▁▁▁▁",
93423
- "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
93424
- "▁ ▁"
 
 
 
 
 
 
 
93425
  ]
93426
  }
93427
  }
 
9
  "single_word": false,
10
  "lstrip": false,
11
  "rstrip": false,
12
+ "normalized": false,
13
  "special": true
14
  },
15
  {
 
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
+ "normalized": false,
22
  "special": true
23
  },
24
  {
 
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
30
+ "normalized": false,
31
  "special": true
32
  },
33
  {
34
  "id": 32000,
35
+ "content": "[PAD]",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
 
41
  },
42
  {
43
  "id": 32001,
44
+ "content": "<pad>",
45
  "single_word": false,
46
  "lstrip": false,
47
  "rstrip": false,
 
50
  },
51
  {
52
  "id": 32002,
53
+ "content": "<feeling>",
54
  "single_word": false,
55
  "lstrip": false,
56
  "rstrip": false,
 
59
  },
60
  {
61
  "id": 32003,
62
+ "content": "<endstring>",
63
+ "single_word": false,
64
+ "lstrip": false,
65
+ "rstrip": false,
66
+ "normalized": false,
67
+ "special": true
68
+ },
69
+ {
70
+ "id": 32004,
71
  "content": "<symptom>",
72
  "single_word": false,
73
  "lstrip": false,
 
93311
  "▁livre s",
93312
  "lu b",
93313
  "l ub",
93314
+ "▁ ▁",
93315
  "▁▁ ▁▁",
93316
+ "▁▁▁ ▁",
93317
+ "▁ ▁▁▁",
 
93318
  "▁▁ ▁▁▁▁▁▁",
 
 
 
 
 
 
 
 
 
 
93319
  "▁▁▁▁ ▁▁▁▁",
93320
+ "▁▁▁▁▁ ▁▁▁",
93321
+ "▁▁▁▁▁▁ ▁▁",
93322
+ "▁▁▁ ▁▁▁▁▁",
93323
+ "▁▁▁▁▁▁▁ ▁",
93324
+ "▁ ▁▁▁▁▁▁▁",
93325
+ "▁▁ ▁▁▁",
 
 
 
93326
  "▁▁▁▁ ▁",
93327
+ "▁▁▁ ▁▁",
93328
+ "▁ ▁▁▁▁",
93329
+ "▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
93330
+ "▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93331
  "▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
93332
  "▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
 
 
 
 
 
 
93333
  "▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
 
 
 
 
 
93334
  "▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
 
 
 
93335
  "▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
 
 
 
 
93336
  "▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
 
 
93337
  "��▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
 
 
 
 
 
 
 
93338
  "▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁",
93339
+ "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93340
+ "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93341
+ "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93342
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93343
+ "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
93344
+ "▁▁ ▁▁▁▁",
93345
+ "▁▁▁▁ ▁▁",
93346
+ "▁▁▁▁▁ ▁",
93347
  "▁▁▁ ▁▁▁",
93348
+ "▁ ▁▁▁▁▁",
93349
+ "▁▁ ▁▁▁▁▁▁▁▁▁▁",
93350
+ "▁▁▁▁ ▁▁▁▁▁▁▁▁",
93351
+ "▁▁▁▁▁▁▁▁ ▁▁▁▁",
93352
+ "▁▁▁▁▁ ▁▁▁▁▁▁▁",
93353
+ "▁▁▁▁▁▁ ▁▁▁▁▁▁",
93354
+ "▁▁▁▁▁▁▁▁▁▁ ▁▁",
93355
  "▁▁▁ ▁▁▁▁▁▁▁▁▁",
 
 
 
 
 
 
 
93356
  "▁▁▁▁▁▁▁▁▁ ▁▁▁",
 
 
 
 
 
93357
  "▁▁▁▁▁▁▁ ▁▁▁▁▁",
93358
+ "▁▁▁▁▁▁▁▁▁▁▁ ▁",
93359
+ "▁ ▁▁▁▁▁▁▁▁▁▁▁",
93360
+ "▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93361
+ "▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93362
+ "▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93363
+ "▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93364
+ "▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93365
+ "▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93366
+ "▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93367
+ "▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
93368
+ "▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93369
  "▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93370
+ "▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93371
+ "▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93372
+ "▁▁ ▁▁▁▁▁▁▁▁",
93373
+ "▁▁▁▁ ▁▁▁▁▁▁",
93374
+ "▁▁▁▁▁▁▁▁ ▁▁",
93375
+ "▁▁▁▁▁ ▁▁▁▁▁",
93376
+ "▁▁▁▁▁▁ ▁▁▁▁",
93377
+ "▁▁▁ ▁▁▁▁▁▁▁",
93378
+ "▁▁▁▁▁▁▁▁▁ ▁",
93379
  "▁▁▁▁▁▁▁ ▁▁▁",
93380
+ "▁ ▁▁▁▁▁▁▁▁▁",
93381
+ "▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93382
+ "▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
93383
+ "▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93384
+ "▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93385
+ "▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93386
+ "▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93387
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93388
+ "▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93389
+ "▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93390
+ "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93391
  "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
 
 
 
 
93392
  "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93393
+ "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁",
93394
+ "▁▁ ▁",
93395
  "▁ ▁▁",
93396
+ "▁▁ ▁▁▁▁▁▁▁",
93397
+ "▁▁▁▁ ▁▁▁▁▁",
93398
+ "▁▁▁▁▁▁▁▁ ▁",
93399
+ "▁▁▁▁▁ ▁▁▁▁",
93400
+ "▁▁▁▁▁▁ ▁▁▁",
93401
+ "▁▁▁ ▁▁▁▁▁▁",
93402
+ "▁▁▁▁▁▁▁ ▁▁",
93403
  "▁ ▁▁▁▁▁▁▁▁",
93404
+ "▁▁ ▁▁▁▁▁",
93405
+ "▁▁▁▁ ▁▁▁",
93406
+ "▁▁▁▁▁ ▁▁",
93407
+ "▁▁▁▁▁▁ ▁",
93408
+ "▁▁▁ ▁▁▁▁",
93409
  "▁ ▁▁▁▁▁▁",
93410
+ "▁▁ ▁▁▁▁▁▁▁▁▁",
93411
+ "▁▁▁▁ ▁▁▁▁▁▁▁",
93412
+ "▁▁▁▁▁▁▁▁ ▁▁▁",
93413
+ "▁▁▁▁▁ ▁▁▁▁▁▁",
93414
+ "▁▁▁▁▁▁ ▁▁▁▁▁",
93415
+ "▁▁▁▁▁▁▁▁▁▁ ▁",
93416
+ "▁▁▁ ▁▁▁▁▁▁▁▁",
93417
+ "▁▁▁▁▁���▁▁▁ ▁▁",
93418
+ "▁▁▁▁▁▁▁ ▁▁▁▁",
93419
  "▁ ▁▁▁▁▁▁▁▁▁▁",
93420
+ "▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁",
93421
+ "▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁",
93422
+ "▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁",
93423
+ "▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁▁",
93424
+ "▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁▁",
93425
+ "▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁",
93426
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁▁",
93427
+ "▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁",
93428
+ "▁▁▁▁▁▁▁▁▁▁▁▁▁▁ ▁",
93429
+ "▁▁▁ ▁▁▁▁▁▁▁▁▁▁▁▁",
93430
+ "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93431
+ "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93432
+ "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93433
+ "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁"
93434
  ]
93435
  }
93436
  }
tokenizer_config.json CHANGED
@@ -1,9 +1,11 @@
1
  {
 
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<unk>",
5
  "lstrip": false,
6
- "normalized": true,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
@@ -11,7 +13,7 @@
11
  "1": {
12
  "content": "<s>",
13
  "lstrip": false,
14
- "normalized": true,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
@@ -19,13 +21,13 @@
19
  "2": {
20
  "content": "</s>",
21
  "lstrip": false,
22
- "normalized": true,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
  "32000": {
28
- "content": "<pad>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
@@ -33,7 +35,7 @@
33
  "special": true
34
  },
35
  "32001": {
36
- "content": "<feeling>",
37
  "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
@@ -41,7 +43,7 @@
41
  "special": true
42
  },
43
  "32002": {
44
- "content": "<endstring>",
45
  "lstrip": false,
46
  "normalized": false,
47
  "rstrip": false,
@@ -49,6 +51,14 @@
49
  "special": true
50
  },
51
  "32003": {
 
 
 
 
 
 
 
 
52
  "content": "<symptom>",
53
  "lstrip": false,
54
  "normalized": true,
@@ -60,10 +70,12 @@
60
  "bos_token": "<feeling>",
61
  "clean_up_tokenization_spaces": false,
62
  "eos_token": "<endstring>",
 
63
  "model_max_length": 1000000000000000019884624838656,
64
  "pad_token": "<pad>",
 
65
  "sp_model_kwargs": {},
66
  "tokenizer_class": "LlamaTokenizer",
67
  "unk_token": "<unk>",
68
- "use_default_system_prompt": false
69
  }
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
7
  "lstrip": false,
8
+ "normalized": false,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
 
13
  "1": {
14
  "content": "<s>",
15
  "lstrip": false,
16
+ "normalized": false,
17
  "rstrip": false,
18
  "single_word": false,
19
  "special": true
 
21
  "2": {
22
  "content": "</s>",
23
  "lstrip": false,
24
+ "normalized": false,
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
28
  },
29
  "32000": {
30
+ "content": "[PAD]",
31
  "lstrip": false,
32
  "normalized": false,
33
  "rstrip": false,
 
35
  "special": true
36
  },
37
  "32001": {
38
+ "content": "<pad>",
39
  "lstrip": false,
40
  "normalized": false,
41
  "rstrip": false,
 
43
  "special": true
44
  },
45
  "32002": {
46
+ "content": "<feeling>",
47
  "lstrip": false,
48
  "normalized": false,
49
  "rstrip": false,
 
51
  "special": true
52
  },
53
  "32003": {
54
+ "content": "<endstring>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "32004": {
62
  "content": "<symptom>",
63
  "lstrip": false,
64
  "normalized": true,
 
70
  "bos_token": "<feeling>",
71
  "clean_up_tokenization_spaces": false,
72
  "eos_token": "<endstring>",
73
+ "legacy": false,
74
  "model_max_length": 1000000000000000019884624838656,
75
  "pad_token": "<pad>",
76
+ "padding_side": "right",
77
  "sp_model_kwargs": {},
78
  "tokenizer_class": "LlamaTokenizer",
79
  "unk_token": "<unk>",
80
+ "use_default_system_prompt": true
81
  }