any
#26
by
battleman0526
- opened
- README.md +2 -2
- tokenizer_config.json +1 -1
README.md
CHANGED
@@ -44,10 +44,10 @@ For more information about ModernBERT, we recommend our [release blog post](http
|
|
44 |
|
45 |
## Usage
|
46 |
|
47 |
-
You can use these models directly with the `transformers` library
|
48 |
|
49 |
```sh
|
50 |
-
pip install
|
51 |
```
|
52 |
|
53 |
Since ModernBERT is a Masked Language Model (MLM), you can use the `fill-mask` pipeline or load it via `AutoModelForMaskedLM`. To use ModernBERT for downstream tasks like classification, retrieval, or QA, fine-tune it following standard BERT fine-tuning recipes.
|
|
|
44 |
|
45 |
## Usage
|
46 |
|
47 |
+
You can use these models directly with the `transformers` library. Until the next `transformers` release, doing so requires installing transformers from main:
|
48 |
|
49 |
```sh
|
50 |
+
pip install git+https://github.com/huggingface/transformers.git
|
51 |
```
|
52 |
|
53 |
Since ModernBERT is a Masked Language Model (MLM), you can use the `fill-mask` pipeline or load it via `AutoModelForMaskedLM`. To use ModernBERT for downstream tasks like classification, retrieval, or QA, fine-tune it following standard BERT fine-tuning recipes.
|
tokenizer_config.json
CHANGED
@@ -932,7 +932,7 @@
|
|
932 |
"clean_up_tokenization_spaces": true,
|
933 |
"cls_token": "[CLS]",
|
934 |
"mask_token": "[MASK]",
|
935 |
-
"model_max_length":
|
936 |
"pad_token": "[PAD]",
|
937 |
"sep_token": "[SEP]",
|
938 |
"tokenizer_class": "PreTrainedTokenizerFast",
|
|
|
932 |
"clean_up_tokenization_spaces": true,
|
933 |
"cls_token": "[CLS]",
|
934 |
"mask_token": "[MASK]",
|
935 |
+
"model_max_length": 1000000000000000019884624838656,
|
936 |
"pad_token": "[PAD]",
|
937 |
"sep_token": "[SEP]",
|
938 |
"tokenizer_class": "PreTrainedTokenizerFast",
|