Update README.md, tokenizer.json
Browse files- README.md +18 -1
- tokenizer.json +1 -1
README.md
CHANGED
@@ -66,7 +66,7 @@ The tokenizer of this model is based on [huggingface/tokenizers](https://github.
|
|
66 |
The vocabulary entries were converted from [`llm-jp-tokenizer v2.2 (100k)`](https://github.com/llm-jp/llm-jp-tokenizer/releases/tag/v2.2).
|
67 |
Please refer to [README.md](https://github.com/llm-jp/llm-jp-tokenizer) of `llm-jp/llm-ja-tokenizer` for details on the vocabulary construction procedure.
|
68 |
|
69 |
-
Note that unlike [ku-nlp/deberta-v2-base-japanese](https://huggingface.co/ku-nlp/deberta-v2-base-japanese), pre-segmentation by a morphological analyzer (e.g., Juman++) is no longer required for this model.
|
70 |
|
71 |
## Training data
|
72 |
|
@@ -103,6 +103,23 @@ The following hyperparameters were used during pre-training:
|
|
103 |
- training_steps: 475,000
|
104 |
- warmup_steps: 10,000
|
105 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
106 |
## Acknowledgments
|
107 |
|
108 |
This work was supported by Joint Usage/Research Center for Interdisciplinary Large-scale Information Infrastructures (JHPCN) through General Collaboration Project no. jh221004, "Developing a Platform for Constructing and Sharing of Large-Scale Japanese Language Models".
|
|
|
66 |
The vocabulary entries were converted from [`llm-jp-tokenizer v2.2 (100k)`](https://github.com/llm-jp/llm-jp-tokenizer/releases/tag/v2.2).
|
67 |
Please refer to [README.md](https://github.com/llm-jp/llm-jp-tokenizer) of `llm-jp/llm-ja-tokenizer` for details on the vocabulary construction procedure.
|
68 |
|
69 |
+
Note that, unlike [ku-nlp/deberta-v2-base-japanese](https://huggingface.co/ku-nlp/deberta-v2-base-japanese), pre-segmentation by a morphological analyzer (e.g., Juman++) is no longer required for this model.
|
70 |
|
71 |
## Training data
|
72 |
|
|
|
103 |
- training_steps: 475,000
|
104 |
- warmup_steps: 10,000
|
105 |
|
106 |
+
## Fine-tuning on NLU tasks
|
107 |
+
|
108 |
+
We fine-tuned the following models and evaluated them on the dev set of JGLUE.
|
109 |
+
We tuned the learning rate and training epochs for each model and task following [the JGLUE paper](https://www.jstage.jst.go.jp/article/jnlp/30/1/30_63/_pdf/-char/ja).
|
110 |
+
|
111 |
+
| Model | MARC-ja/acc | JCoLA/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
|
112 |
+
|-------------------------------|-------------|-----------|--------------|---------------|----------|-----------|-----------|------------|
|
113 |
+
| Waseda RoBERTa base | 0.965 | 0.867 | 0.913 | 0.876 | 0.905 | 0.853 | 0.916 | 0.853 |
|
114 |
+
| Waseda RoBERTa large (seq512) | 0.969 | 0.849 | 0.925 | 0.890 | 0.928 | 0.910 | 0.955 | 0.900 |
|
115 |
+
| LUKE Japanese base* | 0.965 | - | 0.916 | 0.877 | 0.912 | - | - | 0.842 |
|
116 |
+
| LUKE Japanese large* | 0.965 | - | 0.932 | 0.902 | 0.927 | - | - | 0.893 |
|
117 |
+
| DeBERTaV2 base | 0.970 | 0.879 | 0.922 | 0.886 | 0.922 | 0.899 | 0.951 | 0.873 |
|
118 |
+
| DeBERTaV2 large | 0.968 | 0.882 | 0.925 | 0.892 | 0.924 | 0.912 | 0.959 | 0.890 |
|
119 |
+
| DeBERTaV3 base | 0.960 | 0.878 | 0.927 | 0.891 | 0.927 | 0.896 | 0.947 | 0.875 |
|
120 |
+
|
121 |
+
*The scores of LUKE are from [the official repository](https://github.com/studio-ousia/luke).
|
122 |
+
|
123 |
## Acknowledgments
|
124 |
|
125 |
This work was supported by Joint Usage/Research Center for Interdisciplinary Large-scale Information Infrastructures (JHPCN) through General Collaboration Project no. jh221004, "Developing a Platform for Constructing and Sharing of Large-Scale Japanese Language Models".
|
tokenizer.json
CHANGED
@@ -387631,6 +387631,6 @@
|
|
387631 |
-12.989911079406738
|
387632 |
]
|
387633 |
],
|
387634 |
-
"byte_fallback":
|
387635 |
}
|
387636 |
}
|
|
|
387631 |
-12.989911079406738
|
387632 |
]
|
387633 |
],
|
387634 |
+
"byte_fallback": true
|
387635 |
}
|
387636 |
}
|