Commit
·
4dacbda
1
Parent(s):
0290302
add links to MLM model
Browse files
README.md
CHANGED
@@ -28,7 +28,9 @@ counterparts.
|
|
28 |
|
29 |
The model was trained using a self-supervised masked language modeling task. We do whole word masking with a maximum of 80 predictions. The model was trained for 1000K steps, with a batch size of 4096, and a max sequence length of 512.
|
30 |
|
31 |
-
|
|
|
|
|
32 |
|
33 |
License: Apache 2.0
|
34 |
|
|
|
28 |
|
29 |
The model was trained using a self-supervised masked language modeling task. We do whole word masking with a maximum of 80 predictions. The model was trained for 1000K steps, with a batch size of 4096, and a max sequence length of 512.
|
30 |
|
31 |
+
Original model and details: https://tfhub.dev/google/MuRIL/1
|
32 |
+
|
33 |
+
MLM-specific model hosted on HuggingFace: https://huggingface.co/simran-kh/muril-with-mlm-cased-temp
|
34 |
|
35 |
License: Apache 2.0
|
36 |
|