Update README.md
Browse files
README.md
CHANGED
@@ -18,8 +18,22 @@ license: llama3
|
|
18 |
- **SFT Data:** The SFT data includes both internally generated data and publicly available data on Hugging Face, translated into Korean where necessary.
|
19 |
- **DPO Data:** The DPO data consists of translated and curated data from argilla/dpo-mix-7k.
|
20 |
|
21 |
-
|
22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
```python
|
24 |
import transformers
|
25 |
import torch
|
@@ -82,7 +96,7 @@ KISTI๋ ๋์ ๋ณธ์, ์์ธ ๋ถ์, ๋๊ตฌยท๊ฒฝ๋ถ ์ง์, ๋ถ์ฐ์ธ์ฐ๊ฒฝ๋จ
|
|
82 |
KISTI์ ๋ํ ๋ ์์ธํ ์ ๋ณด๋ KISTI ๊ณต์ ์น์ฌ์ดํธ์์ ํ์ธํ ์ ์์ต๋๋ค.
|
83 |
```
|
84 |
|
85 |
-
###
|
86 |
**Language Model**
|
87 |
```text
|
88 |
@article{KISTI-KONI/KONI-Llama3-8B-Instruct-20240729,
|
@@ -93,13 +107,17 @@ KISTI์ ๋ํ ๋ ์์ธํ ์ ๋ณด๋ KISTI ๊ณต์ ์น์ฌ์ดํธ์์ ํ์ธ
|
|
93 |
}
|
94 |
```
|
95 |
|
96 |
-
###
|
97 |
- KISTI, Large-scale AI Research Group
|
98 |
|
99 |
-
###
|
|
|
|
|
|
|
|
|
100 |
- This research was supported by Korea Institute of Science and Technology Information(KISTI).
|
101 |
- This work was supported by the National Supercomputing Center with supercomputing resources including technical support (KISTI).
|
102 |
|
103 |
-
###
|
104 |
- https://huggingface.co/meta-llama/Meta-Llama-3-8B
|
105 |
- https://huggingface.co/meta-llama/meta-llama/Meta-Llama-3-8B-Instruct
|
|
|
18 |
- **SFT Data:** The SFT data includes both internally generated data and publicly available data on Hugging Face, translated into Korean where necessary.
|
19 |
- **DPO Data:** The DPO data consists of translated and curated data from argilla/dpo-mix-7k.
|
20 |
|
21 |
+
### 4. Benchmark Results
|
22 |
+
Results in [LogicKor](https://github.com/StableFluffy/LogicKor)* are as follows:
|
23 |
+
|
24 |
+
| Metric | Score |
|
25 |
+
|----------------|-------|
|
26 |
+
| Reasoning | 6.57 |
|
27 |
+
| Math | 8.00 |
|
28 |
+
| Writing | 8.92 |
|
29 |
+
| Comprehension | 8.85 |
|
30 |
+
| Grammar | 9.85 |
|
31 |
+
| Singleton | 7.07 |
|
32 |
+
| Multiturn | 8.42 |
|
33 |
+
| Total Score | 8.00 |
|
34 |
+
| **Overall** | **8.21** |
|
35 |
+
|
36 |
+
### 5. How to use the model
|
37 |
```python
|
38 |
import transformers
|
39 |
import torch
|
|
|
96 |
KISTI์ ๋ํ ๋ ์์ธํ ์ ๋ณด๋ KISTI ๊ณต์ ์น์ฌ์ดํธ์์ ํ์ธํ ์ ์์ต๋๋ค.
|
97 |
```
|
98 |
|
99 |
+
### 6. Citation
|
100 |
**Language Model**
|
101 |
```text
|
102 |
@article{KISTI-KONI/KONI-Llama3-8B-Instruct-20240729,
|
|
|
107 |
}
|
108 |
```
|
109 |
|
110 |
+
### 7. Contributors
|
111 |
- KISTI, Large-scale AI Research Group
|
112 |
|
113 |
+
### 8. Special Thanks
|
114 |
+
- [@beomi](https://huggingface.co/beomi)
|
115 |
+
- [@kuotient](https://huggingface.co/kuotient)
|
116 |
+
|
117 |
+
### 8. Acknowledgement
|
118 |
- This research was supported by Korea Institute of Science and Technology Information(KISTI).
|
119 |
- This work was supported by the National Supercomputing Center with supercomputing resources including technical support (KISTI).
|
120 |
|
121 |
+
### 9. References
|
122 |
- https://huggingface.co/meta-llama/Meta-Llama-3-8B
|
123 |
- https://huggingface.co/meta-llama/meta-llama/Meta-Llama-3-8B-Instruct
|