davidadamczyk commited on
Commit
f0fd28e
1 Parent(s): b616bbb

Add SetFit model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,256 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: sentence-transformers/all-mpnet-base-v2
3
+ library_name: setfit
4
+ metrics:
5
+ - accuracy
6
+ pipeline_tag: text-classification
7
+ tags:
8
+ - setfit
9
+ - sentence-transformers
10
+ - text-classification
11
+ - generated_from_setfit_trainer
12
+ widget:
13
+ - text: 'I noticed something missing in Gail''s and Bret''s banter about the debt-ceiling
14
+ vote that is typical republican mush!Bret gets Gail to agree that spending is
15
+ too high, then Bret proceeds to suggest it''s time to raise the retirement age
16
+ for Social Security! And then...wait for it......Bret mentions nothing about raising
17
+ taxes on corporations and billionaires!Bret, you would agree that the quaint 1950s
18
+ was a time of sanity in the GOP. ....Well, in those good ol'' days, top marginal
19
+ tax rates were in the 70% range.....What''s more, our national debt was low, like
20
+ around zero!?....And what''s even more, the USA was absolutely first in the world
21
+ in reading and math scores.Enough.
22
+
23
+ '
24
+ - text: 'Denial is not limited to American politicians. It seems China is extreme
25
+ in this category. All the ''Zero Covid'' policy did was delay the inevitable.
26
+ China is the US under Trump. Using vaccines which, while home grown, are not as
27
+ effective only placed its population a great risk. They will have the same strain
28
+ on their healthcare system. Very Sad.
29
+
30
+ '
31
+ - text: 'China knows everything about its citizens, monitors every details in their
32
+ lives but somehow can''t say how many people exactly died from Covid19 since it
33
+ ended its zero covid policy.Why should we believe these numbers instead of last
34
+ week numbers?
35
+
36
+ '
37
+ - text: 'Johnny G These figures are also not accurate or believable. Crematoria in
38
+ China''s large cities have been overrun with bodies since the zero-covid policy
39
+ ended--running at full capacity with long backlogs. Any back of the envelope calculation
40
+ would give a much higher death figure than 60,000--and the virus hasn''t even
41
+ ravaged the countryside yet. That will happen over the next 3-4 weeks as migrant
42
+ workers and others return to their villages to celebrate the Chinese New Year
43
+ on Jan. 21. Due to the backwardness of rural healthcare and the proportionally
44
+ high concentration of elderly people in the countryside, the covid death toll
45
+ in rural China within the next few weeks will be high but will also receive much
46
+ less media attention.
47
+
48
+ '
49
+ - text: 'I was beaten and verbally abused until age 17, when I could escape my home. My
50
+ family "looked" normal from the outside, but was not. Child abuse was not yet
51
+ in the lexicon.I turned out normal! This I owe to visiting lots of friends and
52
+ watching how their families interacted--they were kind. I asked their parents
53
+ to adopt me. I watched family sitcoms--the opposite of my homelife. I did well
54
+ in school, so I received praise there, and made friends.The folks wanted me to
55
+ marry well and have kids. But the Zero Population Movement, and Women''s Lib,
56
+ gave me a window into how humans harm the planet, and that women could do more
57
+ than have babies and do laundry. I put myself through uni, had no children, and
58
+ have had and have careers I love.Parenting is the most important, unpaid job one
59
+ can take on because it demands selflessly developing a decent, caring, intellectually
60
+ curious, kind, patient human. People lacking these qualities should re-think
61
+ parenthood.Also, consider the childless life, to save the planet.
62
+
63
+ '
64
+ inference: true
65
+ model-index:
66
+ - name: SetFit with sentence-transformers/all-mpnet-base-v2
67
+ results:
68
+ - task:
69
+ type: text-classification
70
+ name: Text Classification
71
+ dataset:
72
+ name: Unknown
73
+ type: unknown
74
+ split: test
75
+ metrics:
76
+ - type: accuracy
77
+ value: 1.0
78
+ name: Accuracy
79
+ ---
80
+
81
+ # SetFit with sentence-transformers/all-mpnet-base-v2
82
+
83
+ This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
84
+
85
+ The model has been trained using an efficient few-shot learning technique that involves:
86
+
87
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
88
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
89
+
90
+ ## Model Details
91
+
92
+ ### Model Description
93
+ - **Model Type:** SetFit
94
+ - **Sentence Transformer body:** [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2)
95
+ - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
96
+ - **Maximum Sequence Length:** 384 tokens
97
+ - **Number of Classes:** 2 classes
98
+ <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
99
+ <!-- - **Language:** Unknown -->
100
+ <!-- - **License:** Unknown -->
101
+
102
+ ### Model Sources
103
+
104
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
105
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
106
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
107
+
108
+ ### Model Labels
109
+ | Label | Examples |
110
+ |:------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
111
+ | yes | <ul><li>'"Xi Jinping, China’s top leader, abandoned his “zero Covid” policy in early December. That policy had kept infections low but required costly precautions like mass testing — measures that exhausted the budgets of local governments."In a recent issue, The Economist magazine reported that China spent ~$250 billion on mass testing during a recent one-year period. The piece also indicated that an unnamed expert suggested that that number was likely to be much lower than the true amount. Even for China, this is a remarkable amount of resources devoted to that aspect of combating Covid. It\'s no wonder President Xi had to finally give up on zero Covid - in all its manifestations, China could no longer afford the strategy.\n'</li><li>'The huge excursions to and from China at the Dawn of 2020 for China\'s lunar year celebration, just after the Wuhan breakout in DEC 2019 and its aftermath of spreading Covid-19 as a wildfire across the globe has a lesson to compare the present situation.China\'s much advertised, the world\'s first stringent drive to eradicate Covid VIRUS by adopting "ZERO COVID " policy since 2019 was lifted on DEC,7,2022 after realizing its end point is a fiasco. The reporting 60k fatalities a week before the China\'s lunar year on 22,JAN,2023 is a caution to the international travelers. Any global viral spread in 2023 shan\'t become a justification for lifting Zero Covid policy and zero testing of the travelers- in and out by China.\n'</li><li>'Ace Not so black and white. China’s “No-COVID” policy during the early part of the pandemic, albeit draconian and heavy-handed, likely saved tens of thousands of lives. However, once vaccines became available, China should have 1) adopted Western mRNA vaccines which are more effective at preventing serious illness than the Chinese domestic versions. 2) Begin preparing for a gradual reopening by stockpiling antivirals to protect its most vulnerable citizens. By demonstrating the “superior” Chinese model with the prolonged strict no-COVID policy, President Xi was able to secure his unprecedented 3rd 5-year term.Liberals are against public health policies that are driven by political considerations rather than driven by science.\n'</li></ul> |
112
+ | no | <ul><li>'Teaching history is, by its very nature, a matter of prioritization and opinion. When it is a mandatory requirement for a high school diploma, the requirement to learn a specific version of history and regurgitate it becomes a form of indoctrination. DeSantis is an easy target for his opponents (I am one) for obvious reasons, but the challenge remains the same. What is the version of history that we want to teach our children? Should the history of black Americans be enhanced? What about Mexicans ( a largely overlooked group), women, Asians (nary a word about the Chinese Exclusion Act), religious subgroups - the early plight of Catholics, Jewish immigrants, Mormons, Muslims, and the emergence of a non-secular movement? How would we propose to teach about abortion rights? Is it the quiet revolution of the unborn or the destruction of rights previously available to women? The list goes on. I find articles like this with outrage dripping, reductive, and of little value. A challenge with public schools is that they are an arm of the government. So, it is hardly surprising that the CEO of the state/legislature would exert influence. A debate no history is highly valuable but America goes immediately to war with itself and no longer debates\n'</li><li>"David Brook offers an interesting perspective on Biden and America's conduct in the world.Putin, Xi are all crazy people doing crazy things. In contrast, Biden is a steady hand guiding the American ship of the international rule based order.I suppose if I lived in the Washington bubble, I might have a similar view. But I come from a world of anti-imperialist struggle, and my world looks very different.I see the US undermining struggling nations all over the world, most recently in Africa. The ugly American fingerprints are also all over the coups in Honduras, Venezuela, Bolivia and Peru.Cuba is now in its sixtieth year of a crushing US blockade. US military bases now nun from Niger in West Africa, across the continent to Kenya.Active military operations are going on in Somalis, Syria and of course Ukraine.There's no difference between the referendums for autonomy held in Kosovo and the Donbas and Crimea, except that one was sponsored by the US and the other by Russia.According to the UN, world famine this year can be averted for 1.7 billion dollars. In contrast, our military funding for Ukraine is now at 122 billion.Under American leadership, corporations paid out $257bn to wealthy shareholders, while over 800 million people went to bed hungry.So, forgive me if I see Biden's ''steady hand” differently than the NYTimes crowd does.Perspective is everything, and the world looks very different when you see it from the bottom up.\n"</li><li>'LB and what would we do for our neighbors? What did we do when children were separated from their parents at the border under Trump? Most of us did nothing.\n'</li></ul> |
113
+
114
+ ## Evaluation
115
+
116
+ ### Metrics
117
+ | Label | Accuracy |
118
+ |:--------|:---------|
119
+ | **all** | 1.0 |
120
+
121
+ ## Uses
122
+
123
+ ### Direct Use for Inference
124
+
125
+ First install the SetFit library:
126
+
127
+ ```bash
128
+ pip install setfit
129
+ ```
130
+
131
+ Then you can load this model and run inference.
132
+
133
+ ```python
134
+ from setfit import SetFitModel
135
+
136
+ # Download from the 🤗 Hub
137
+ model = SetFitModel.from_pretrained("davidadamczyk/setfit-model-8")
138
+ # Run inference
139
+ preds = model("China knows everything about its citizens, monitors every details in their lives but somehow can't say how many people exactly died from Covid19 since it ended its zero covid policy.Why should we believe these numbers instead of last week numbers?
140
+ ")
141
+ ```
142
+
143
+ <!--
144
+ ### Downstream Use
145
+
146
+ *List how someone could finetune this model on their own dataset.*
147
+ -->
148
+
149
+ <!--
150
+ ### Out-of-Scope Use
151
+
152
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
153
+ -->
154
+
155
+ <!--
156
+ ## Bias, Risks and Limitations
157
+
158
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
159
+ -->
160
+
161
+ <!--
162
+ ### Recommendations
163
+
164
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
165
+ -->
166
+
167
+ ## Training Details
168
+
169
+ ### Training Set Metrics
170
+ | Training set | Min | Median | Max |
171
+ |:-------------|:----|:--------|:----|
172
+ | Word count | 13 | 141.375 | 287 |
173
+
174
+ | Label | Training Sample Count |
175
+ |:------|:----------------------|
176
+ | no | 18 |
177
+ | yes | 22 |
178
+
179
+ ### Training Hyperparameters
180
+ - batch_size: (16, 16)
181
+ - num_epochs: (1, 1)
182
+ - max_steps: -1
183
+ - sampling_strategy: oversampling
184
+ - num_iterations: 120
185
+ - body_learning_rate: (2e-05, 2e-05)
186
+ - head_learning_rate: 2e-05
187
+ - loss: CosineSimilarityLoss
188
+ - distance_metric: cosine_distance
189
+ - margin: 0.25
190
+ - end_to_end: False
191
+ - use_amp: False
192
+ - warmup_proportion: 0.1
193
+ - l2_weight: 0.01
194
+ - seed: 42
195
+ - eval_max_steps: -1
196
+ - load_best_model_at_end: False
197
+
198
+ ### Training Results
199
+ | Epoch | Step | Training Loss | Validation Loss |
200
+ |:------:|:----:|:-------------:|:---------------:|
201
+ | 0.0017 | 1 | 0.3089 | - |
202
+ | 0.0833 | 50 | 0.1005 | - |
203
+ | 0.1667 | 100 | 0.0014 | - |
204
+ | 0.25 | 150 | 0.0004 | - |
205
+ | 0.3333 | 200 | 0.0002 | - |
206
+ | 0.4167 | 250 | 0.0002 | - |
207
+ | 0.5 | 300 | 0.0002 | - |
208
+ | 0.5833 | 350 | 0.0001 | - |
209
+ | 0.6667 | 400 | 0.0001 | - |
210
+ | 0.75 | 450 | 0.0001 | - |
211
+ | 0.8333 | 500 | 0.0001 | - |
212
+ | 0.9167 | 550 | 0.0001 | - |
213
+ | 1.0 | 600 | 0.0001 | - |
214
+
215
+ ### Framework Versions
216
+ - Python: 3.10.13
217
+ - SetFit: 1.1.0
218
+ - Sentence Transformers: 3.0.1
219
+ - Transformers: 4.45.2
220
+ - PyTorch: 2.4.0+cu124
221
+ - Datasets: 2.21.0
222
+ - Tokenizers: 0.20.0
223
+
224
+ ## Citation
225
+
226
+ ### BibTeX
227
+ ```bibtex
228
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
229
+ doi = {10.48550/ARXIV.2209.11055},
230
+ url = {https://arxiv.org/abs/2209.11055},
231
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
232
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
233
+ title = {Efficient Few-Shot Learning Without Prompts},
234
+ publisher = {arXiv},
235
+ year = {2022},
236
+ copyright = {Creative Commons Attribution 4.0 International}
237
+ }
238
+ ```
239
+
240
+ <!--
241
+ ## Glossary
242
+
243
+ *Clearly define terms in order to be accessible across audiences.*
244
+ -->
245
+
246
+ <!--
247
+ ## Model Card Authors
248
+
249
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
250
+ -->
251
+
252
+ <!--
253
+ ## Model Card Contact
254
+
255
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
256
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sentence-transformers/all-mpnet-base-v2",
3
+ "architectures": [
4
+ "MPNetModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.45.2",
23
+ "vocab_size": 30527
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.0.1",
4
+ "transformers": "4.45.2",
5
+ "pytorch": "2.4.0+cu124"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
config_setfit.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "normalize_embeddings": false,
3
+ "labels": [
4
+ "no",
5
+ "yes"
6
+ ]
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58310ac559cde2a4015fc72d08dc7dae0b817a363b969f0548044869adba35b5
3
+ size 437967672
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:707826f1dfda2e95e1f3eb5cdf6b9ee3dafefd0bc7d1b8f9da67e58a1f01170b
3
+ size 7023
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 384,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[UNK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "30526": {
44
+ "content": "<mask>",
45
+ "lstrip": true,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ }
51
+ },
52
+ "bos_token": "<s>",
53
+ "clean_up_tokenization_spaces": false,
54
+ "cls_token": "<s>",
55
+ "do_lower_case": true,
56
+ "eos_token": "</s>",
57
+ "mask_token": "<mask>",
58
+ "max_length": 128,
59
+ "model_max_length": 384,
60
+ "pad_to_multiple_of": null,
61
+ "pad_token": "<pad>",
62
+ "pad_token_type_id": 0,
63
+ "padding_side": "right",
64
+ "sep_token": "</s>",
65
+ "stride": 0,
66
+ "strip_accents": null,
67
+ "tokenize_chinese_chars": true,
68
+ "tokenizer_class": "MPNetTokenizer",
69
+ "truncation_side": "right",
70
+ "truncation_strategy": "longest_first",
71
+ "unk_token": "[UNK]"
72
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff