Update README.md
Browse files
README.md
CHANGED
@@ -8,6 +8,10 @@ language:
|
|
8 |
|
9 |
# nanoT5-base-65kBPE-v2
|
10 |
|
|
|
|
|
|
|
|
|
11 |
- SiLU/gated-SiLU activation
|
12 |
- 25% mask rate during pretrain
|
13 |
- 65k vocab size, [adapted claude3 tokenizer](https://hf.co/BEE-spoke-data/claude-tokenizer-forT5)
|
|
|
8 |
|
9 |
# nanoT5-base-65kBPE-v2
|
10 |
|
11 |
+
> [!NOTE]
|
12 |
+
> This is a "raw" pretrained model intended to be fine-tuned on downstream tasks
|
13 |
+
|
14 |
+
|
15 |
- SiLU/gated-SiLU activation
|
16 |
- 25% mask rate during pretrain
|
17 |
- 65k vocab size, [adapted claude3 tokenizer](https://hf.co/BEE-spoke-data/claude-tokenizer-forT5)
|