Update README.md
Browse files
README.md
CHANGED
@@ -128,7 +128,7 @@ F1 score: A weighted average of precision and recall, which measures the overlap
|
|
128 |
| RoBERTa-Base | 124,056,578 | 8:50:29 | 18:59 | 2:11 | 69.06 | 80.08 | 55.53 | 66.49 |
|
129 |
| RoBERTa-Large | 354,312,194 | 29:16:06 | 1:01:10 | 7:04 | 74.08 | 84.38 | 62.20 | 72.88 |
|
130 |
|
131 |
-
\*
|
132 |
|
133 |
# Limitations and Bias
|
134 |
|
|
|
128 |
| RoBERTa-Base | 124,056,578 | 8:50:29 | 18:59 | 2:11 | 69.06 | 80.08 | 55.53 | 66.49 |
|
129 |
| RoBERTa-Large | 354,312,194 | 29:16:06 | 1:01:10 | 7:04 | 74.08 | 84.38 | 62.20 | 72.88 |
|
130 |
|
131 |
+
\* TinyRoBERTa's training time isn't directly comparable to the other models since it was distilled from [VMware/roberta-large-mrqa](https://huggingface.co/VMware/roberta-large-mrqa) that was already trained on MRQA.
|
132 |
|
133 |
# Limitations and Bias
|
134 |
|