Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,37 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- QingyiSi/Alpaca-CoT
|
5 |
+
language:
|
6 |
+
- zh
|
7 |
+
- en
|
8 |
---
|
9 |
+
|
10 |
+
This is a beta release of a QLoRa adapter model to [Falcon-40b](https://huggingface.co/tiiuae/falcon-40b).
|
11 |
+
Please read the instruction carefully before downloading the model.
|
12 |
+
|
13 |
+
Though Falcon is not specifically trained on Chinese corpus, it exhibits strong performance in Chinese Language Understanding in our experiment. We would like to explore out of curiosity whether a
|
14 |
+
small amount of Chinese instruction data can push it further and make it better at speaking.
|
15 |
+
|
16 |
+
The LoRa model is trained with the [QLoRa code](https://github.com/artidoro/qlora) on a subset of bilingual instruction data from [Alpaca-CoT dataset](https://huggingface.co/datasets/QingyiSi/Alpaca-CoT) for a mere 5k steps.
|
17 |
+
The finetune model is not as good as the carefully continue-trained-and-finetuned LLaMA-models such as [OpenBuddy](https://huggingface.co/OpenBuddy) and [Ziya](https://huggingface.co/IDEA-CCNL/Ziya-LLaMA-13B-v1) in Chinese generation,
|
18 |
+
still it quickly adapts to the new langauge and generate superisingly good result. We call for more research on applying Falcon-40b to the Chinese domain.
|
19 |
+
|
20 |
+
## Evalutions
|
21 |
+
|
22 |
+
We evaluate on two Chinese language understanding benchmarks, [C-Eval](https://cevalbenchmark.com/) and Gaokao subset of [AGIEval](https://github.com/microsoft/AGIEval).
|
23 |
+
|
24 |
+
* C-Eval made breaking change in 2023/06/08 from few-shot to zero-shot,
|
25 |
+
|
26 |
+
Result on C-Eval test set with 5-shot and no CoT
|
27 |
+
|
28 |
+
| Average | Avg(Hard) | STEM | Social Science | Humanities | Others |
|
29 |
+
| - | - | - | - | - | - |
|
30 |
+
| 40.4 | 30.1 | 35.8 | 47.6 | 42.0 | 40.6 |
|
31 |
+
|
32 |
+
|
33 |
+
Result on GaoKao subset of C-Eval with 0-shot
|
34 |
+
|
35 |
+
| Average | GK-chinese | GK-English | GK-geography | GK-history | GK-biology | GK-chemistry | GK-physics | GK-mathqa | GK-mathcloze
|
36 |
+
| - | - | - | - | - | - | - | - | - | - |
|
37 |
+
| 33.6 | 26.4 | 69.0 | 46.7 | 47.8 | 27.1 | 32.4 | 24.5 | 26.8 | 1.7 |
|