AliFartout
commited on
Commit
·
17f3659
1
Parent(s):
1e48016
Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ pipeline_tag: token-classification
|
|
12 |
# NER Model using Roberta
|
13 |
|
14 |
|
15 |
-
This markdown presents a Robustly Optimized BERT Pretraining Approach (RoBERTa) model trained on a combination of two diverse datasets for two languages: English and Persian. The English dataset used is [CoNLL 2003](), while the Persian dataset is [PEYMA-ARMAN-Mixed](), a fusion of the "PEYAM" and "ARMAN" datasets, both popular for Named Entity Recognition (NER) tasks.
|
16 |
|
17 |
The model training pipeline involves the following steps:
|
18 |
|
|
|
12 |
# NER Model using Roberta
|
13 |
|
14 |
|
15 |
+
This markdown presents a Robustly Optimized BERT Pretraining Approach (RoBERTa) model trained on a combination of two diverse datasets for two languages: English and Persian. The English dataset used is [CoNLL 2003](https://huggingface.co/datasets/conll2003), while the Persian dataset is [PEYMA-ARMAN-Mixed](https://huggingface.co/datasets/AliFartout/PEYMA-ARMAN-Mixed), a fusion of the "PEYAM" and "ARMAN" datasets, both popular for Named Entity Recognition (NER) tasks.
|
16 |
|
17 |
The model training pipeline involves the following steps:
|
18 |
|