Update README.md
#8
by
vmkhlv
- opened
README.md
CHANGED
@@ -5,8 +5,10 @@ license: apache-2.0
|
|
5 |
---
|
6 |
|
7 |
# FRED-T5 1.7B (Full-scale Russian Enhanced Denoisers T5)
|
|
|
8 |
|
9 |
-
|
|
|
10 |
|
11 |
Architecture based on T5.
|
12 |
|
@@ -70,4 +72,16 @@ tokenizer.decode(outputs[0][1:])
|
|
70 |
+ Alexey Khoroshilov
|
71 |
|
72 |
[Salute AI Community](https://t.me/SaluteTechGroup).
|
73 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
---
|
6 |
|
7 |
# FRED-T5 1.7B (Full-scale Russian Enhanced Denoisers T5)
|
8 |
+
The model architecture design, pretraining, and evaluation are documented in our preprint: [**A Family of Pretrained Transformer Language Models for Russian**](https://arxiv.org/abs/2309.10931).
|
9 |
|
10 |
+
|
11 |
+
The model was trained by [SberDevices](https://sberdevices.ru/).
|
12 |
|
13 |
Architecture based on T5.
|
14 |
|
|
|
72 |
+ Alexey Khoroshilov
|
73 |
|
74 |
[Salute AI Community](https://t.me/SaluteTechGroup).
|
75 |
+
|
76 |
+
|
77 |
+
# Cite us
|
78 |
+
```
|
79 |
+
@misc{zmitrovich2023family,
|
80 |
+
title={A Family of Pretrained Transformer Language Models for Russian},
|
81 |
+
author={Dmitry Zmitrovich and Alexander Abramov and Andrey Kalmykov and Maria Tikhonova and Ekaterina Taktasheva and Danil Astafurov and Mark Baushenko and Artem Snegirev and Tatiana Shavrina and Sergey Markov and Vladislav Mikhailov and Alena Fenogenova},
|
82 |
+
year={2023},
|
83 |
+
eprint={2309.10931},
|
84 |
+
archivePrefix={arXiv},
|
85 |
+
primaryClass={cs.CL}
|
86 |
+
}
|
87 |
+
```
|