sfarrukh commited on
Commit
46ae0d9
·
verified ·
1 Parent(s): f61f746

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -8
README.md CHANGED
@@ -45,19 +45,23 @@ should probably proofread and complete it, then remove this comment. -->
45
 
46
  # modernbert-setfit-nli
47
 
48
- This model was trained from scratch on an unknown dataset.
49
 
50
- ## Model description
51
 
52
- More information needed
53
 
54
- ## Intended uses & limitations
 
 
55
 
56
- More information needed
 
 
 
57
 
58
- ## Training and evaluation data
59
 
60
- More information needed
61
 
62
  ## Training procedure
63
 
@@ -77,4 +81,7 @@ The following hyperparameters were used during training:
77
  - Transformers 4.48.0
78
  - Pytorch 2.5.1+cu121
79
  - Datasets 3.2.0
80
- - Tokenizers 0.21.0
 
 
 
 
45
 
46
  # modernbert-setfit-nli
47
 
48
+ ## Model Description
49
 
50
+ This model is a fine-tuned version of `answerdotai/ModernBERT-base` trained on a subset of the SetFit/mnli dataset. It is designed for natural language inference (NLI) tasks, where the goal is to determine the relationship between two text inputs (e.g., entailment, contradiction, or neutrality).
51
 
52
+ ## Intended Uses & Limitations
53
 
54
+ ### Intended Uses
55
+ - **Natural Language Inference (NLI):** Suitable for classifying relationships between pairs of sentences.
56
+ - **Text Understanding Tasks:** Can be applied to other similar tasks requiring sentence pair classification.
57
 
58
+ ### Limitations
59
+ - **Dataset-Specific Biases:** The model was fine-tuned on 30,000 samples from the SetFit/mnli dataset and may not generalize well to domains significantly different from the training data.
60
+ - **Context Length:** The tokenizer’s maximum sequence length is 512 tokens. Inputs longer than this will be truncated.
61
+ - **Resource Intensive:** May require a modern GPU for efficient inference on large datasets.
62
 
63
+ This model is a starting point for NLI tasks and may need further fine-tuning for domain-specific applications.
64
 
 
65
 
66
  ## Training procedure
67
 
 
81
  - Transformers 4.48.0
82
  - Pytorch 2.5.1+cu121
83
  - Datasets 3.2.0
84
+ - Tokenizers 0.21.0
85
+ ## References
86
+
87
+ - **GitHub Repository:** The training code is available at [GitHub Repository Link](https://github.com/sfarrukhm/model_finetune.git).