|
--- |
|
license: mit |
|
language: |
|
- ja |
|
- en |
|
- zh |
|
tags: |
|
- LLaMA2 |
|
- Japanese |
|
- LLM |
|
--- |
|
|
|
This model is traned with [guanaco](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset) dataset. And this model used whole guanaco dataset by 49000 chat samples and 280000 non chat samples. |
|
Improved performance in Chinese and Japanese. |
|
Use the QLoRA to fine-tune the vanilla [LLaMA2-7B](https://huggingface.co/NousResearch/Llama-2-7b-hf). |
|
And you can use test.py to test the model. |
|
|
|
### Recommend Generation parameters: |
|
* temperature: 0.5~0.7 |
|
* top p: 0.65~1.0 |
|
* top k: 30~50 |
|
* repeat penalty: 1.03~1.17 |
|
|
|
|
|
|
|
|
|
Contribute by Yokohama Nationaly University Mori Lab. |