MoritzLaurer HF staff commited on
Commit
6058c43
·
verified ·
1 Parent(s): 74cedb0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -12,13 +12,17 @@ license: mit
12
  # Model description: deberta-v3-base-zeroshot-v2.0
13
  The model is designed for zero-shot classification with the Hugging Face pipeline.
14
 
 
 
 
 
15
  The model can do one universal classification task: determine whether a hypothesis is "true" or "not true" given a text
16
  (`entailment` vs. `not_entailment`).
17
  This task format is based on the Natural Language Inference task (NLI).
18
  The task is so universal that any classification task can be reformulated into this task.
19
 
20
- ## Training data
21
 
 
22
  The model is trained on two types of fully commercially-friendly data:
23
  1. Synthetic data generated with [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
24
  I first created a list of 500+ diverse text classification tasks for 25 professions in conversations with Mistral-large. The data was manually curated.
 
12
  # Model description: deberta-v3-base-zeroshot-v2.0
13
  The model is designed for zero-shot classification with the Hugging Face pipeline.
14
 
15
+ The main advantage of this `...zeroshot-v2.0` series of zeroshot classifiers is that they are trained on commercially-friendly data
16
+ and are fully commercially usable, while my older `...zeroshot-v1.1` included training data with non-commercially licenses data.
17
+ An overview of the latest zeroshot classifiers with different sizes and licenses is available in my [Zeroshot Classifier Collection](https://huggingface.co/collections/MoritzLaurer/zeroshot-classifiers-6548b4ff407bb19ff5c3ad6f).
18
+
19
  The model can do one universal classification task: determine whether a hypothesis is "true" or "not true" given a text
20
  (`entailment` vs. `not_entailment`).
21
  This task format is based on the Natural Language Inference task (NLI).
22
  The task is so universal that any classification task can be reformulated into this task.
23
 
 
24
 
25
+ ## Training data
26
  The model is trained on two types of fully commercially-friendly data:
27
  1. Synthetic data generated with [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
28
  I first created a list of 500+ diverse text classification tasks for 25 professions in conversations with Mistral-large. The data was manually curated.