Update README.md
Browse files
README.md
CHANGED
@@ -12,13 +12,17 @@ license: mit
|
|
12 |
# Model description: deberta-v3-base-zeroshot-v2.0
|
13 |
The model is designed for zero-shot classification with the Hugging Face pipeline.
|
14 |
|
|
|
|
|
|
|
|
|
15 |
The model can do one universal classification task: determine whether a hypothesis is "true" or "not true" given a text
|
16 |
(`entailment` vs. `not_entailment`).
|
17 |
This task format is based on the Natural Language Inference task (NLI).
|
18 |
The task is so universal that any classification task can be reformulated into this task.
|
19 |
|
20 |
-
## Training data
|
21 |
|
|
|
22 |
The model is trained on two types of fully commercially-friendly data:
|
23 |
1. Synthetic data generated with [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
|
24 |
I first created a list of 500+ diverse text classification tasks for 25 professions in conversations with Mistral-large. The data was manually curated.
|
|
|
12 |
# Model description: deberta-v3-base-zeroshot-v2.0
|
13 |
The model is designed for zero-shot classification with the Hugging Face pipeline.
|
14 |
|
15 |
+
The main advantage of this `...zeroshot-v2.0` series of zeroshot classifiers is that they are trained on commercially-friendly data
|
16 |
+
and are fully commercially usable, while my older `...zeroshot-v1.1` included training data with non-commercially licenses data.
|
17 |
+
An overview of the latest zeroshot classifiers with different sizes and licenses is available in my [Zeroshot Classifier Collection](https://huggingface.co/collections/MoritzLaurer/zeroshot-classifiers-6548b4ff407bb19ff5c3ad6f).
|
18 |
+
|
19 |
The model can do one universal classification task: determine whether a hypothesis is "true" or "not true" given a text
|
20 |
(`entailment` vs. `not_entailment`).
|
21 |
This task format is based on the Natural Language Inference task (NLI).
|
22 |
The task is so universal that any classification task can be reformulated into this task.
|
23 |
|
|
|
24 |
|
25 |
+
## Training data
|
26 |
The model is trained on two types of fully commercially-friendly data:
|
27 |
1. Synthetic data generated with [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
|
28 |
I first created a list of 500+ diverse text classification tasks for 25 professions in conversations with Mistral-large. The data was manually curated.
|