Upload tokenizer
Browse files- tokenizer.json +2 -2
- tokenizer_config.json +3 -0
tokenizer.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36a437fefdd3f82b64054871c05c5d4bffd8f82532721cc2c359a5b2a97e121a
|
3 |
+
size 16392208
|
tokenizer_config.json
CHANGED
@@ -3321,7 +3321,10 @@
|
|
3321 |
"mask_token": "[MASK]",
|
3322 |
"max_length": 512,
|
3323 |
"model_max_length": 512,
|
|
|
3324 |
"pad_token": "[PAD]",
|
|
|
|
|
3325 |
"sep_token": "[SEP]",
|
3326 |
"sp_model_kwargs": {},
|
3327 |
"split_by_punct": false,
|
|
|
3321 |
"mask_token": "[MASK]",
|
3322 |
"max_length": 512,
|
3323 |
"model_max_length": 512,
|
3324 |
+
"pad_to_multiple_of": null,
|
3325 |
"pad_token": "[PAD]",
|
3326 |
+
"pad_token_type_id": 0,
|
3327 |
+
"padding_side": "right",
|
3328 |
"sep_token": "[SEP]",
|
3329 |
"sp_model_kwargs": {},
|
3330 |
"split_by_punct": false,
|