Fix typo
Browse files
README.md
CHANGED
@@ -7,13 +7,13 @@ datasets:
|
|
7 |
|
8 |
# NT5, a T5 model trained to perform numerical reasoning
|
9 |
|
10 |
-
T5-small model pre-trained on 3 million (partly synthetic) texts and fine-tuned on DROP. It was introduced in the paper [NT5?! Training T5 to Perform Numerical Reasoning](https://arxiv.org/abs/2104.07307) by Yang et al. and first released in [this repository](https://github.com/lesterpjy/numeric-t5). As the original implementation was in Tensorflow 2, I've converted the weigths to PyTorch. This model corresponds to RC Experiment 1 (see the paper), their best performing model.
|
11 |
|
12 |
Disclaimer: The team releasing NT5 did not write a model card for this model so this model card has been written by me.
|
13 |
|
14 |
## Model description
|
15 |
|
16 |
-
The NT5 model is a T5 model, in other words, an encoder-decoder Transformer. In order to encourage numerical reasoning, the model was further pre-trained on three datasets designed to strengthen skills necessary for numerical reasoning over text (NRoT) and general reading comprehension before being fine-tuned on Discrete Reasoning over Text (DROP) dataset.
|
17 |
|
18 |
## Intended uses & limitations
|
19 |
|
@@ -99,6 +99,6 @@ This model achieves an F1 score of 0.7031 and exact match of 0.6687 on the devel
|
|
99 |
biburl = {https://dblp.org/rec/journals/corr/abs-1903-00161.bib},
|
100 |
bibsource = {dblp computer science bibliography, https://dblp.org}
|
101 |
}
|
102 |
-
a service of Schloss Dagstuhl - Leibniz Center for Informatics
|
103 |
|
104 |
```
|
|
|
7 |
|
8 |
# NT5, a T5 model trained to perform numerical reasoning
|
9 |
|
10 |
+
T5-small model pre-trained on 3 million (partly synthetic) texts and fine-tuned on [DROP](https://allennlp.org/drop.html). It was introduced in the paper [NT5?! Training T5 to Perform Numerical Reasoning](https://arxiv.org/abs/2104.07307) by Yang et al. and first released in [this repository](https://github.com/lesterpjy/numeric-t5). As the original implementation was in Tensorflow 2, I've converted the weigths to PyTorch. This model corresponds to RC Experiment 1 (see the paper), their best performing model.
|
11 |
|
12 |
Disclaimer: The team releasing NT5 did not write a model card for this model so this model card has been written by me.
|
13 |
|
14 |
## Model description
|
15 |
|
16 |
+
The NT5 model is a T5 model, in other words, an encoder-decoder Transformer. In order to encourage numerical reasoning, the model was further pre-trained on three datasets designed to strengthen skills necessary for numerical reasoning over text (NRoT) and general reading comprehension before being fine-tuned on the Discrete Reasoning over Text (DROP) dataset.
|
17 |
|
18 |
## Intended uses & limitations
|
19 |
|
|
|
99 |
biburl = {https://dblp.org/rec/journals/corr/abs-1903-00161.bib},
|
100 |
bibsource = {dblp computer science bibliography, https://dblp.org}
|
101 |
}
|
102 |
+
a service of Schloss Dagstuhl - Leibniz Center for Informatics\\\\thomebrowsesearchabout
|
103 |
|
104 |
```
|