File size: 21,229 Bytes
f0fd28e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
---
base_model: sentence-transformers/all-mpnet-base-v2
library_name: setfit
metrics:
- accuracy
pipeline_tag: text-classification
tags:
- setfit
- sentence-transformers
- text-classification
- generated_from_setfit_trainer
widget:
- text: 'I noticed something missing in Gail''s and Bret''s banter about the debt-ceiling
    vote that is typical republican mush!Bret gets Gail to agree that spending is
    too high, then Bret proceeds to suggest it''s time to raise the retirement age
    for Social Security! And then...wait for it......Bret mentions nothing about raising
    taxes on corporations and billionaires!Bret, you would agree that the quaint 1950s
    was a time of sanity in the GOP. ....Well, in those good ol'' days, top marginal
    tax rates were in the 70% range.....What''s more, our national debt was low, like
    around zero!?....And what''s even more, the USA was absolutely first in the world
    in reading and math scores.Enough.

    '
- text: 'Denial is not limited to American politicians. It seems China is extreme
    in this category. All the ''Zero Covid'' policy did was delay the inevitable.
    China is the US under Trump. Using vaccines which, while home grown, are not as
    effective only placed its population a great risk. They will have the same strain
    on their healthcare system. Very Sad.

    '
- text: 'China knows everything about its citizens, monitors every details in their
    lives but somehow can''t say how many people exactly died from Covid19 since it
    ended its zero covid policy.Why should we believe these numbers instead of last
    week numbers?

    '
- text: 'Johnny  G These figures are also not accurate or believable. Crematoria in
    China''s large cities have been overrun with bodies since the zero-covid policy
    ended--running at full capacity with long backlogs. Any back of the envelope calculation
    would give a much higher death figure than 60,000--and the virus hasn''t even
    ravaged the countryside yet. That will happen over the next 3-4 weeks as migrant
    workers and others return to their villages to celebrate the Chinese New Year
    on Jan. 21. Due to the backwardness of rural healthcare and the proportionally
    high concentration of elderly people in the countryside, the covid death toll
    in rural China within the next few weeks will be high but will also receive much
    less media attention.

    '
- text: 'I was beaten and verbally abused until age 17, when I could escape my home.  My
    family "looked" normal from the outside, but was not.  Child abuse was not yet
    in the lexicon.I turned out normal!  This I owe to visiting lots of friends and
    watching how their families interacted--they were kind.  I asked their parents
    to adopt me. I watched family sitcoms--the opposite of my homelife.  I did well
    in school, so I received praise there, and made friends.The folks wanted me to
    marry well and have kids.  But the Zero Population Movement, and Women''s Lib,
    gave me a window into how humans harm the planet, and that women could do more
    than have babies and do laundry.  I put myself through uni, had no children, and
    have had and have careers I love.Parenting is the most important, unpaid job one
    can take on because it demands selflessly developing a decent, caring, intellectually
    curious, kind, patient human.  People lacking these qualities should re-think
    parenthood.Also, consider the childless life, to save the planet.

    '
inference: true
model-index:
- name: SetFit with sentence-transformers/all-mpnet-base-v2
  results:
  - task:
      type: text-classification
      name: Text Classification
    dataset:
      name: Unknown
      type: unknown
      split: test
    metrics:
    - type: accuracy
      value: 1.0
      name: Accuracy
---

# SetFit with sentence-transformers/all-mpnet-base-v2

This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.

The model has been trained using an efficient few-shot learning technique that involves:

1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.

## Model Details

### Model Description
- **Model Type:** SetFit
- **Sentence Transformer body:** [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2)
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
- **Maximum Sequence Length:** 384 tokens
- **Number of Classes:** 2 classes
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->

### Model Sources

- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)

### Model Labels
| Label | Examples                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                |
|:------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| yes   | <ul><li>'"Xi Jinping, China’s top leader, abandoned his “zero Covid” policy in early December. That policy had kept infections low but required costly precautions like mass testing — measures that exhausted the budgets of local governments."In a recent issue, The Economist magazine reported that China spent ~$250 billion on mass testing during a recent one-year period. The piece also indicated that an unnamed expert suggested that that number was likely to be much lower than the true amount. Even for China, this is a remarkable amount of resources devoted to that aspect of combating Covid. It\'s no wonder President Xi had to finally give up on zero Covid - in all its manifestations, China could no longer afford the strategy.\n'</li><li>'The huge excursions to and from China at the Dawn of 2020 for China\'s lunar year celebration, just after the Wuhan breakout in DEC 2019 and its aftermath of spreading Covid-19 as a wildfire across the globe has a lesson to compare the present situation.China\'s much advertised, the world\'s first stringent drive to eradicate  Covid VIRUS by adopting  "ZERO COVID " policy since 2019 was lifted on DEC,7,2022 after realizing its end point is a fiasco. The reporting 60k fatalities a week before the China\'s lunar year  on 22,JAN,2023 is a caution to the international travelers. Any global viral spread in 2023 shan\'t become a justification for lifting Zero Covid policy and zero testing of the travelers- in and out by China.\n'</li><li>'Ace Not so black and white. China’s “No-COVID”  policy during the early part of the pandemic, albeit draconian and heavy-handed, likely saved tens of thousands of lives.  However, once vaccines became available, China should have 1) adopted Western mRNA vaccines which are more effective at preventing serious illness than the Chinese domestic versions. 2) Begin preparing for a gradual reopening by stockpiling antivirals to protect its most vulnerable citizens. By demonstrating the “superior” Chinese model with the prolonged strict no-COVID policy, President Xi was able to secure his unprecedented 3rd 5-year term.Liberals are against public health policies that are driven by political considerations rather than driven by science.\n'</li></ul>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        |
| no    | <ul><li>'Teaching history is, by its very nature, a matter of prioritization and opinion. When it is a mandatory requirement for a high school diploma, the requirement to learn a specific version of history and regurgitate it becomes a form of indoctrination.  DeSantis is an easy target for his opponents (I am one) for obvious reasons, but the challenge remains the same.  What is the version of history that we want to teach our children?  Should the history of black Americans be enhanced?  What about Mexicans ( a largely overlooked group), women, Asians (nary a word about the Chinese Exclusion Act), religious subgroups - the early plight of Catholics, Jewish immigrants, Mormons, Muslims, and the emergence of a non-secular movement?  How would we propose to teach about abortion rights?  Is it the quiet revolution of the unborn or the destruction of rights previously available to women?  The list goes on.  I find articles like this with outrage dripping, reductive, and of little value. A challenge with public schools is that they are an arm of the government.  So, it is hardly surprising that the CEO of the state/legislature would exert influence.  A debate no history is highly valuable but America goes immediately to war with itself and no longer debates\n'</li><li>"David Brook offers an interesting perspective on Biden and America's conduct in the world.Putin, Xi are all crazy people doing crazy things. In contrast, Biden is a steady hand guiding the American ship of the international rule based order.I suppose if I lived in the Washington bubble, I might have a similar view. But I come from a world of anti-imperialist struggle, and my world looks very different.I see the US undermining struggling nations all over the world, most recently in Africa. The ugly American fingerprints are also all over the coups in Honduras, Venezuela, Bolivia and Peru.Cuba is now in its sixtieth year of a crushing US blockade. US military bases now nun from Niger in West Africa, across the continent to Kenya.Active military operations are going on in Somalis, Syria and of course Ukraine.There's no difference between the referendums for autonomy held in Kosovo and the Donbas and Crimea, except that one was sponsored by the US and the other by Russia.According to the UN, world famine this year can be averted for 1.7 billion dollars. In contrast, our military funding for Ukraine is now at 122 billion.Under American leadership, corporations paid out $257bn to wealthy shareholders, while over 800 million people went to bed hungry.So, forgive me if I see Biden's ''steady hand” differently than the NYTimes crowd does.Perspective is everything, and the world looks very different when you see it from the bottom up.\n"</li><li>'LB and what would we do for our neighbors?  What did we do when children were separated from their parents at the border under Trump?   Most of us did nothing.\n'</li></ul> |

## Evaluation

### Metrics
| Label   | Accuracy |
|:--------|:---------|
| **all** | 1.0      |

## Uses

### Direct Use for Inference

First install the SetFit library:

```bash
pip install setfit
```

Then you can load this model and run inference.

```python
from setfit import SetFitModel

# Download from the 🤗 Hub
model = SetFitModel.from_pretrained("davidadamczyk/setfit-model-8")
# Run inference
preds = model("China knows everything about its citizens, monitors every details in their lives but somehow can't say how many people exactly died from Covid19 since it ended its zero covid policy.Why should we believe these numbers instead of last week numbers?
")
```

<!--
### Downstream Use

*List how someone could finetune this model on their own dataset.*
-->

<!--
### Out-of-Scope Use

*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->

<!--
## Bias, Risks and Limitations

*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->

<!--
### Recommendations

*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->

## Training Details

### Training Set Metrics
| Training set | Min | Median  | Max |
|:-------------|:----|:--------|:----|
| Word count   | 13  | 141.375 | 287 |

| Label | Training Sample Count |
|:------|:----------------------|
| no    | 18                    |
| yes   | 22                    |

### Training Hyperparameters
- batch_size: (16, 16)
- num_epochs: (1, 1)
- max_steps: -1
- sampling_strategy: oversampling
- num_iterations: 120
- body_learning_rate: (2e-05, 2e-05)
- head_learning_rate: 2e-05
- loss: CosineSimilarityLoss
- distance_metric: cosine_distance
- margin: 0.25
- end_to_end: False
- use_amp: False
- warmup_proportion: 0.1
- l2_weight: 0.01
- seed: 42
- eval_max_steps: -1
- load_best_model_at_end: False

### Training Results
| Epoch  | Step | Training Loss | Validation Loss |
|:------:|:----:|:-------------:|:---------------:|
| 0.0017 | 1    | 0.3089        | -               |
| 0.0833 | 50   | 0.1005        | -               |
| 0.1667 | 100  | 0.0014        | -               |
| 0.25   | 150  | 0.0004        | -               |
| 0.3333 | 200  | 0.0002        | -               |
| 0.4167 | 250  | 0.0002        | -               |
| 0.5    | 300  | 0.0002        | -               |
| 0.5833 | 350  | 0.0001        | -               |
| 0.6667 | 400  | 0.0001        | -               |
| 0.75   | 450  | 0.0001        | -               |
| 0.8333 | 500  | 0.0001        | -               |
| 0.9167 | 550  | 0.0001        | -               |
| 1.0    | 600  | 0.0001        | -               |

### Framework Versions
- Python: 3.10.13
- SetFit: 1.1.0
- Sentence Transformers: 3.0.1
- Transformers: 4.45.2
- PyTorch: 2.4.0+cu124
- Datasets: 2.21.0
- Tokenizers: 0.20.0

## Citation

### BibTeX
```bibtex
@article{https://doi.org/10.48550/arxiv.2209.11055,
    doi = {10.48550/ARXIV.2209.11055},
    url = {https://arxiv.org/abs/2209.11055},
    author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
    keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
    title = {Efficient Few-Shot Learning Without Prompts},
    publisher = {arXiv},
    year = {2022},
    copyright = {Creative Commons Attribution 4.0 International}
}
```

<!--
## Glossary

*Clearly define terms in order to be accessible across audiences.*
-->

<!--
## Model Card Authors

*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->

<!--
## Model Card Contact

*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
-->