Splend1dchan commited on
Commit
90794e9
·
verified ·
1 Parent(s): c85e888

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -70,9 +70,12 @@ Performance-wise:
70
  | [Qwen-14B](https://huggingface.co/01-ai/Qwen/Qwen-14B)| 14B | 51.30 | 16.95 * | 50.69 | 68.83 |
71
  | [Yi-6B](https://huggingface.co/01-ai/Yi-6B) | 6B | 49.63 | 76.61 | 34.72 | 65.35 |
72
  | [Qwen-7B](https://huggingface.co/01-ai/Qwen/Qwen-7B)| 7B | 42.84 | 0.0 * | 39.58 | 61.00 |
73
- | [**Breeze-7B-Base-v1_0**](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v1_0) | 7B | | | | |
 
74
  | [**Breeze-7B-Base-v0_1**](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0_1) | 7B | 40.35 | 81.13 | 28.47 | 61.63 |
75
  | [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)| 7B | 36.93 | 79.27 | 27.78 | 64.89 |
 
 
76
 
77
 
78
  \* Few-shot learning cannot effectively guide the model to generate the proper answer.
@@ -118,7 +121,6 @@ Performance-wise:
118
  | Taiwan-LLM-13B-v2.0-chat | 6.1 | 3.4 | 4.1 | 2.3 | 3.1 | 7.4 | 6.6 | 6.8 | 5.0 |
119
  | Taiwan-LLM-7B-v2.1-chat | 5.2 | 2.6 | 2.3 | 1.2 | 3.4 | 6.6 | 5.7 | 6.8 | 4.2 |
120
 
121
-
122
  | Details on TMMLU+ (0 shot):<br/>Model | STEM | Social Science | Humanities | Other | ↑ AVG |
123
  |-----------------------------------------------------|--------------|----------------|------------|------------|---------|
124
  | Yi-34B-Chat | 47.65 | 64.25 | 52.73 | 54.91 | 54.87 |
 
70
  | [Qwen-14B](https://huggingface.co/01-ai/Qwen/Qwen-14B)| 14B | 51.30 | 16.95 * | 50.69 | 68.83 |
71
  | [Yi-6B](https://huggingface.co/01-ai/Yi-6B) | 6B | 49.63 | 76.61 | 34.72 | 65.35 |
72
  | [Qwen-7B](https://huggingface.co/01-ai/Qwen/Qwen-7B)| 7B | 42.84 | 0.0 * | 39.58 | 61.00 |
73
+ | [Qwen-7B-1.5](https://huggingface.co/01-ai/Qwen/Qwen-7B-1.5)| 7B | 46.28 | | 30.56 | 60.53 |
74
+ | [**Breeze-7B-Base-v1_0**](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v1_0) | 7B | 40.72 | 80.61 | 31.99 | 58.65 |
75
  | [**Breeze-7B-Base-v0_1**](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0_1) | 7B | 40.35 | 81.13 | 28.47 | 61.63 |
76
  | [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)| 7B | 36.93 | 79.27 | 27.78 | 64.89 |
77
+ | [Mistral-7B-v0.2](https://huggingface.co/mistralai/Mistral-7B-v0.2)| 7B | 34.94 | | 33.33 | 57.33 |
78
+
79
 
80
 
81
  \* Few-shot learning cannot effectively guide the model to generate the proper answer.
 
121
  | Taiwan-LLM-13B-v2.0-chat | 6.1 | 3.4 | 4.1 | 2.3 | 3.1 | 7.4 | 6.6 | 6.8 | 5.0 |
122
  | Taiwan-LLM-7B-v2.1-chat | 5.2 | 2.6 | 2.3 | 1.2 | 3.4 | 6.6 | 5.7 | 6.8 | 4.2 |
123
 
 
124
  | Details on TMMLU+ (0 shot):<br/>Model | STEM | Social Science | Humanities | Other | ↑ AVG |
125
  |-----------------------------------------------------|--------------|----------------|------------|------------|---------|
126
  | Yi-34B-Chat | 47.65 | 64.25 | 52.73 | 54.91 | 54.87 |