Update README.md
Browse files
README.md
CHANGED
@@ -7,32 +7,36 @@ license: cc-by-4.0
|
|
7 |
KoQuality-Polyglot-5.8b is a fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-5.8b) on [KoQuality dataset](https://huggingface.co/datasets/DILAB-HYU/KoQuality), which is curated by proposed method (len_group=5, k=100, n=0.01, method=ppl_sampling).
|
8 |
|
9 |
|
10 |
-
|
11 |
|
12 |
We use [KoBEST benchmark](https://huggingface.co/datasets/skt/kobest_v1) datasets(BoolQ, COPA, HellaSwag, SentiNeg, WiC) to compare the performance of our best model and other models accuracy. Our model outperforms other models in the average accuracy score of the KoBEST datasets.
|
13 |
<img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/q4cCUCzRJa3m2f7oxI_FY.png style="max-width: 500px; width: 300%"/>
|
14 |
|
15 |
| Model | 0-shot | 1-shot | 2-shot | 5-shot | 10-shot
|
16 |
| --- | --- | --- | --- | --- | --- |
|
17 |
-
| koquality-polyglot-5.8b | 0.5472 | 0.5979 | 0.6260 | 0.6486 | 0.6535
|
18 |
| polyglot-ko-5.8b | 0.5587 | 0.5977 | 0.6138 | 0.6431 | 0.6457
|
19 |
| koalpcaca-polyglot-5.8b | 0.5085 | 0.5561 | 0.5768 | 0.6097 | 0.6059
|
20 |
| kullm-polyglot-5.8b | 0.5409 | 0.6072 | 0.5945 | 0.6345 | 0.6530
|
|
|
21 |
|
|
|
|
|
22 |
|
23 |
-
### Evaluation results
|
24 |
-
COPA (F1)
|
25 |
<img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/7EKl1OAgKgPBFcSlGzBiW.png style="max-width: 800px; width: 400%"/>
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
-
HellaSwag (F1)
|
28 |
-
|
29 |
-
BoolQ (F1)
|
30 |
-
|
31 |
-
SentiNeg (F1)
|
32 |
|
33 |
-
|
34 |
|
|
|
35 |
|
|
|
36 |
|
37 |
|
38 |
|
@@ -48,10 +52,10 @@ WiC (F1)
|
|
48 |
- num_epochs: 2.0
|
49 |
|
50 |
## Framework versions
|
51 |
-
Transformers 4.30.2
|
52 |
-
Pytorch 2.0.1+cu117
|
53 |
-
Datasets 2.11.0
|
54 |
-
deepspeed 0.9.5
|
55 |
|
56 |
## Citation
|
57 |
|
|
|
7 |
KoQuality-Polyglot-5.8b is a fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-5.8b) on [KoQuality dataset](https://huggingface.co/datasets/DILAB-HYU/KoQuality), which is curated by proposed method (len_group=5, k=100, n=0.01, method=ppl_sampling).
|
8 |
|
9 |
|
10 |
+
## Overall Average accuracy score of the KoBEST datasets
|
11 |
|
12 |
We use [KoBEST benchmark](https://huggingface.co/datasets/skt/kobest_v1) datasets(BoolQ, COPA, HellaSwag, SentiNeg, WiC) to compare the performance of our best model and other models accuracy. Our model outperforms other models in the average accuracy score of the KoBEST datasets.
|
13 |
<img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/q4cCUCzRJa3m2f7oxI_FY.png style="max-width: 500px; width: 300%"/>
|
14 |
|
15 |
| Model | 0-shot | 1-shot | 2-shot | 5-shot | 10-shot
|
16 |
| --- | --- | --- | --- | --- | --- |
|
|
|
17 |
| polyglot-ko-5.8b | 0.5587 | 0.5977 | 0.6138 | 0.6431 | 0.6457
|
18 |
| koalpcaca-polyglot-5.8b | 0.5085 | 0.5561 | 0.5768 | 0.6097 | 0.6059
|
19 |
| kullm-polyglot-5.8b | 0.5409 | 0.6072 | 0.5945 | 0.6345 | 0.6530
|
20 |
+
| koquality-polyglot-5.8b | 0.5472 | 0.5979 | 0.6260 | 0.6486 | 0.6535
|
21 |
|
22 |
+
## Evaluation results
|
23 |
+
### COPA (F1)
|
24 |
|
|
|
|
|
25 |
<img src=https://cdn-uploads.huggingface.co/production/uploads/650fecfd247f564485f8fbcf/7EKl1OAgKgPBFcSlGzBiW.png style="max-width: 800px; width: 400%"/>
|
26 |
+
| Model | 0-shot | 1-shot | 2-shot | 5-shot | 10-shot
|
27 |
+
| --- | --- | --- | --- | --- | --- |
|
28 |
+
| polyglot-ko-5.8b | 0.5587 | 0.5977 | 0.6138 | 0.6431 | 0.6457
|
29 |
+
| koalpcaca-polyglot-5.8b | 0.5085 | 0.5561 | 0.5768 | 0.6097 | 0.6059
|
30 |
+
| kullm-polyglot-5.8b | 0.5409 | 0.6072 | 0.5945 | 0.6345 | 0.6530
|
31 |
+
| koquality-polyglot-5.8b | 0.5472 | 0.5979 | 0.6260 | 0.6486 | 0.6535
|
32 |
|
33 |
+
### HellaSwag (F1)
|
|
|
|
|
|
|
|
|
34 |
|
35 |
+
### BoolQ (F1)
|
36 |
|
37 |
+
### SentiNeg (F1)
|
38 |
|
39 |
+
### WiC (F1)
|
40 |
|
41 |
|
42 |
|
|
|
52 |
- num_epochs: 2.0
|
53 |
|
54 |
## Framework versions
|
55 |
+
- Transformers 4.30.2
|
56 |
+
- Pytorch 2.0.1+cu117
|
57 |
+
- Datasets 2.11.0
|
58 |
+
- deepspeed 0.9.5
|
59 |
|
60 |
## Citation
|
61 |
|