roygan commited on
Commit
9d613e4
1 Parent(s): 686c476

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -0
README.md ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - zh
4
+ license: apache-2.0
5
+ widget:
6
+ - text: "生活的真谛是[MASK]。"
7
+ ---
8
+ # Erlangshen-1.3B model (Chinese),one model of https://github.com/IDEA-CCNL/Fengshenbang-LM.
9
+ Encoder structure-based Bidirection language model, focusing on solving various natural language understanding tasks. The 1.3 billion parameter Erlangshen-1.3B large model, using 280G Chinese data, 32 A100 training for 14 days, is the largest open source Chinese Bert large model. On November 10, 2021, **it reached the top of the FewCLUE** list of the authoritative benchmark for Chinese language understanding(https://mp.weixin.qq.com/s/bA_9n_TlBE9P-UzCn7mKoA).
10
+
11
+ Among them, **CHID (Idiom Fill in the Blank) and TNEWS (News Classification) surpass human beings, CHID (Idiom Fill in the Blank), CSLDCP (Subject Document Classification), OCNLI (Natural Language Reasoning) single task first, refreshing few-shot learning records**. The Erlangshen series will continue to be optimized in terms of model scale, knowledge integration, and supervision task assistance.
12
+
13
+ ## Usage
14
+ ```python
15
+ from transformers import BertTokenizer, BertModel,
16
+ tokenizer = BertTokenizer.from_pretrained("Langboat/mengzi-bert-base")
17
+ model = BertModel.from_pretrained("Langboat/mengzi-bert-base")
18
+
19
+ from transformers import MegatronBertConfig, MegatronBertModel
20
+ from transformers import BertTokenizer
21
+
22
+ model_pretrained_weight_path = '/home/' #模型的权重路径
23
+ tokenizer = BertTokenizer.from_pretrained("IDEA-CCNL/Erlangshen-1.3B")
24
+ config = MegatronBertConfig.from_pretrained("IDEA-CCNL/Erlangshen-1.3B")
25
+ model = MegatronBertModel.from_pretrained("IDEA-CCNL/Erlangshen-1.3B")
26
+
27
+ ```
28
+ ## Scores on downstream chinese tasks (without any data augmentation)
29
+ | Model | afqmc | tnews | iflytek | ocnli | cmnli | wsc | csl |
30
+ | :--------: | :-----: | :----: | :-----: | :----: | :----: | :----: | :----: |
31
+ | roberta-wwm-ext-large | 0.7514 | 0.5872 | 0.6152 | 0.777 | 0.814 | 0.8914 | 0.86 |
32
+ | Erlangshen-1.3B | 0.7608 | 0.5996 | 0.6234 | 0.7917 | 0.81 | 0.9243 | 0.872 |
33
+
34
+ ## Citation
35
+ If you find the technical report or resource is useful, please cite the following website in your paper.
36
+ ```
37
+ https://github.com/IDEA-CCNL/Fengshenbang-LM
38
+ ```