Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ The pretrained corpus is OSCAR's Turkish split, but it is further filtered and c
|
|
15 |
|
16 |
Model architecture is similar to bert-medium (8 layers, 8 heads, and 512 hidden size). Tokenization algorithm is Word-level, which means text is split by white space. Vocabulary size is 66.7k.
|
17 |
|
18 |
-
The details can be found at this paper:
|
19 |
https://arxiv.org/abs/2204.08832
|
20 |
|
21 |
The following code can be used for model loading and tokenization, example max length (514) can be changed:
|
|
|
15 |
|
16 |
Model architecture is similar to bert-medium (8 layers, 8 heads, and 512 hidden size). Tokenization algorithm is Word-level, which means text is split by white space. Vocabulary size is 66.7k.
|
17 |
|
18 |
+
The details and performance comparisons can be found at this paper:
|
19 |
https://arxiv.org/abs/2204.08832
|
20 |
|
21 |
The following code can be used for model loading and tokenization, example max length (514) can be changed:
|