add tokenizer
Browse files- tokenizer.json +0 -0
- tokenizer_config.json +2 -1
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -33,7 +33,7 @@
|
|
33 |
"rstrip": false,
|
34 |
"single_word": false
|
35 |
},
|
36 |
-
"name_or_path": "
|
37 |
"never_split": null,
|
38 |
"pad_token": {
|
39 |
"__type": "AddedToken",
|
@@ -51,6 +51,7 @@
|
|
51 |
"rstrip": false,
|
52 |
"single_word": false
|
53 |
},
|
|
|
54 |
"strip_accents": null,
|
55 |
"tokenize_chinese_chars": true,
|
56 |
"tokenizer_class": "MPNetTokenizer",
|
|
|
33 |
"rstrip": false,
|
34 |
"single_word": false
|
35 |
},
|
36 |
+
"name_or_path": "onlydj96/mpnet_pretrain_10",
|
37 |
"never_split": null,
|
38 |
"pad_token": {
|
39 |
"__type": "AddedToken",
|
|
|
51 |
"rstrip": false,
|
52 |
"single_word": false
|
53 |
},
|
54 |
+
"special_tokens_map_file": "/root/.cache/huggingface/transformers/4aa599e6ac8f0263a04a4569ddadb35dafe43bfb10448da4035dd12101ec111a.1b83d0d7f4d455d37c683966d465a99be7f33983cf93b19ad8d2d23d044ea57a",
|
55 |
"strip_accents": null,
|
56 |
"tokenize_chinese_chars": true,
|
57 |
"tokenizer_class": "MPNetTokenizer",
|