NTIS commited on
Commit
dad3a8d
โ€ข
1 Parent(s): 698c519

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -6
README.md CHANGED
@@ -18,8 +18,22 @@ license: llama3
18
  - **SFT Data:** The SFT data includes both internally generated data and publicly available data on Hugging Face, translated into Korean where necessary.
19
  - **DPO Data:** The DPO data consists of translated and curated data from argilla/dpo-mix-7k.
20
 
21
-
22
- ### 4. How to use the model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  ```python
24
  import transformers
25
  import torch
@@ -82,7 +96,7 @@ KISTI๋Š” ๋Œ€์ „ ๋ณธ์›, ์„œ์šธ ๋ถ„์›, ๋Œ€๊ตฌยท๊ฒฝ๋ถ ์ง€์›, ๋ถ€์‚ฐ์šธ์‚ฐ๊ฒฝ๋‚จ
82
  KISTI์— ๋Œ€ํ•œ ๋” ์ž์„ธํ•œ ์ •๋ณด๋Š” KISTI ๊ณต์‹ ์›น์‚ฌ์ดํŠธ์—์„œ ํ™•์ธํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
83
  ```
84
 
85
- ### 5. Citation
86
  **Language Model**
87
  ```text
88
  @article{KISTI-KONI/KONI-Llama3-8B-Instruct-20240729,
@@ -93,13 +107,17 @@ KISTI์— ๋Œ€ํ•œ ๋” ์ž์„ธํ•œ ์ •๋ณด๋Š” KISTI ๊ณต์‹ ์›น์‚ฌ์ดํŠธ์—์„œ ํ™•์ธ
93
  }
94
  ```
95
 
96
- ### 6. Contributors
97
  - KISTI, Large-scale AI Research Group
98
 
99
- ### 7. Acknowledgement
 
 
 
 
100
  - This research was supported by Korea Institute of Science and Technology Information(KISTI).
101
  - This work was supported by the National Supercomputing Center with supercomputing resources including technical support (KISTI).
102
 
103
- ### 8. References
104
  - https://huggingface.co/meta-llama/Meta-Llama-3-8B
105
  - https://huggingface.co/meta-llama/meta-llama/Meta-Llama-3-8B-Instruct
 
18
  - **SFT Data:** The SFT data includes both internally generated data and publicly available data on Hugging Face, translated into Korean where necessary.
19
  - **DPO Data:** The DPO data consists of translated and curated data from argilla/dpo-mix-7k.
20
 
21
+ ### 4. Benchmark Results
22
+ Results in [LogicKor](https://github.com/StableFluffy/LogicKor)* are as follows:
23
+
24
+ | Metric | Score |
25
+ |----------------|-------|
26
+ | Reasoning | 6.57 |
27
+ | Math | 8.00 |
28
+ | Writing | 8.92 |
29
+ | Comprehension | 8.85 |
30
+ | Grammar | 9.85 |
31
+ | Singleton | 7.07 |
32
+ | Multiturn | 8.42 |
33
+ | Total Score | 8.00 |
34
+ | **Overall** | **8.21** |
35
+
36
+ ### 5. How to use the model
37
  ```python
38
  import transformers
39
  import torch
 
96
  KISTI์— ๋Œ€ํ•œ ๋” ์ž์„ธํ•œ ์ •๋ณด๋Š” KISTI ๊ณต์‹ ์›น์‚ฌ์ดํŠธ์—์„œ ํ™•์ธํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
97
  ```
98
 
99
+ ### 6. Citation
100
  **Language Model**
101
  ```text
102
  @article{KISTI-KONI/KONI-Llama3-8B-Instruct-20240729,
 
107
  }
108
  ```
109
 
110
+ ### 7. Contributors
111
  - KISTI, Large-scale AI Research Group
112
 
113
+ ### 8. Special Thanks
114
+ - [@beomi](https://huggingface.co/beomi)
115
+ - [@kuotient](https://huggingface.co/kuotient)
116
+
117
+ ### 8. Acknowledgement
118
  - This research was supported by Korea Institute of Science and Technology Information(KISTI).
119
  - This work was supported by the National Supercomputing Center with supercomputing resources including technical support (KISTI).
120
 
121
+ ### 9. References
122
  - https://huggingface.co/meta-llama/Meta-Llama-3-8B
123
  - https://huggingface.co/meta-llama/meta-llama/Meta-Llama-3-8B-Instruct