Update README.md
Browse files
README.md
CHANGED
@@ -45,19 +45,23 @@ should probably proofread and complete it, then remove this comment. -->
|
|
45 |
|
46 |
# modernbert-setfit-nli
|
47 |
|
48 |
-
|
49 |
|
50 |
-
|
51 |
|
52 |
-
|
53 |
|
54 |
-
|
|
|
|
|
55 |
|
56 |
-
|
|
|
|
|
|
|
57 |
|
58 |
-
|
59 |
|
60 |
-
More information needed
|
61 |
|
62 |
## Training procedure
|
63 |
|
@@ -77,4 +81,7 @@ The following hyperparameters were used during training:
|
|
77 |
- Transformers 4.48.0
|
78 |
- Pytorch 2.5.1+cu121
|
79 |
- Datasets 3.2.0
|
80 |
-
- Tokenizers 0.21.0
|
|
|
|
|
|
|
|
45 |
|
46 |
# modernbert-setfit-nli
|
47 |
|
48 |
+
## Model Description
|
49 |
|
50 |
+
This model is a fine-tuned version of `answerdotai/ModernBERT-base` trained on a subset of the SetFit/mnli dataset. It is designed for natural language inference (NLI) tasks, where the goal is to determine the relationship between two text inputs (e.g., entailment, contradiction, or neutrality).
|
51 |
|
52 |
+
## Intended Uses & Limitations
|
53 |
|
54 |
+
### Intended Uses
|
55 |
+
- **Natural Language Inference (NLI):** Suitable for classifying relationships between pairs of sentences.
|
56 |
+
- **Text Understanding Tasks:** Can be applied to other similar tasks requiring sentence pair classification.
|
57 |
|
58 |
+
### Limitations
|
59 |
+
- **Dataset-Specific Biases:** The model was fine-tuned on 30,000 samples from the SetFit/mnli dataset and may not generalize well to domains significantly different from the training data.
|
60 |
+
- **Context Length:** The tokenizer’s maximum sequence length is 512 tokens. Inputs longer than this will be truncated.
|
61 |
+
- **Resource Intensive:** May require a modern GPU for efficient inference on large datasets.
|
62 |
|
63 |
+
This model is a starting point for NLI tasks and may need further fine-tuning for domain-specific applications.
|
64 |
|
|
|
65 |
|
66 |
## Training procedure
|
67 |
|
|
|
81 |
- Transformers 4.48.0
|
82 |
- Pytorch 2.5.1+cu121
|
83 |
- Datasets 3.2.0
|
84 |
+
- Tokenizers 0.21.0
|
85 |
+
## References
|
86 |
+
|
87 |
+
- **GitHub Repository:** The training code is available at [GitHub Repository Link](https://github.com/sfarrukhm/model_finetune.git).
|