Joelzhang commited on
Commit
ff1c439
1 Parent(s): c8cd009

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -12
README.md CHANGED
@@ -13,27 +13,32 @@ inference: true
13
 
14
  ---
15
  # Erlangshen-MegatronBert-1.3B
 
16
  - Github: [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)
17
  - Docs: [Fengshenbang-Docs](https://fengshenbang-doc.readthedocs.io/)
18
 
19
  ## 简介 Brief Introduction
 
20
  2021登顶FewCLUE和ZeroCLUE,处理NLU任务,开源时最大的中文BERT模型
21
 
22
  Erlangshen topped FewCLUE and ZeroCLUE benchmark in 2021, designed to slove NLU tasks, is the largest BERT when publicly released.
23
 
24
  ## 模型分类 Model Taxonomy
 
25
  | 需求 Demand | 任务 Task | 系列 Series | 模型 Model | 参数 Parameter | 额外 Extra |
26
- | ---- | ---- | ---- | ---- | ---- | ---- |
27
  | 通用 General | 自然语言理解 NLU | 二郎神 Erlangshen | MegatronBert | 1.3B | - |
28
 
29
  ## 模型信息 Model Information
 
30
  Encoder结构为主的双向语言模型,专注于解决各种自然语言理解任务。
31
  我们跟进了[Megatron-LM](https://github.com/NVIDIA/Megatron-LM)的工作,使用了32张A100,总共耗时14天在悟道语料库(180 GB版本)上训练了十亿级别参数量的BERT。同时,鉴于中文语法和大规模训练的难度,我们使用四种预训练策略来改进BERT:1) 整词掩码, 2) 知识动态遮掩, 3) 句子顺序预测, 4) 层前归一化.
32
 
33
  A bidirectional language model based on the Encoder structure, focusing on solving various NLU tasks.
34
  We follow [Megatron-LM](https://github.com/NVIDIA/Megatron-LM), using 32 A100s and spending 14 days training a billion-level BERT on WuDao Corpora (180 GB version). Given Chinese grammar and the difficulty of large-scale training, we use four pre-training procedures to improve BERT: 1) Whole Word Masking (WWM), 2) Knowledge-based Dynamic Masking (KDM), 3) Sentence Order Prediction (SOP), 4) Pre-layer Normalization (Pre-LN).
35
 
36
- ## 成就 Achievement
 
37
  1.2021年11月10日,二郎神在FewCLUE上取得第一。其中,它在CHIDF(成语填空)和TNEWS(新闻分类)子任务中的表现优于人类表现。此外,它在CHIDF(成语填空), CSLDCP(学科文献分类), OCNLI(自然语言推理)任务中均名列前茅。
38
  2.2022年1月24日,二郎神在CLUE基准测试中的ZeroCLUE中取得第一。具体到子任务,我们在CSLDCP(主题文献分类), TNEWS(新闻分类), IFLYTEK(应用描述分类), CSL(抽象关键字识别)和CLUEWSC(参考消歧)任务中取得第一。
39
  3.在2022年7月10日,我们在CLUE基准的语义匹配任务中取得第一。
@@ -42,8 +47,16 @@ We follow [Megatron-LM](https://github.com/NVIDIA/Megatron-LM), using 32 A100s a
42
  2.On January 24, 2022, Erlangshen-MRC topped the ZeroCLUE benchmark. For each of these tasks, we rank the top ones in CSLDCP (Subject Literature Classification), TNEWS (News Classification), IFLYTEK (Application Description Classification), CSL (Abstract Keyword Recognition), and CLUEWSC (Referential Disambiguation) tasks.
43
  3.Erlangshen topped the CLUE benchmark semantic matching task on July 10, 2022.
44
 
 
 
 
 
 
 
 
45
  ## 使用 Usage
46
- ```python
 
47
  from transformers import MegatronBertConfig, MegatronBertModel
48
  from transformers import BertTokenizer
49
 
@@ -52,18 +65,13 @@ config = MegatronBertConfig.from_pretrained("IDEA-CCNL/Erlangshen-MegatronBert-1
52
  model = MegatronBertModel.from_pretrained("IDEA-CCNL/Erlangshen-MegatronBert-1.3B")
53
  ```
54
 
55
- ## 下游任务表现 Performance
56
- | Model | afqmc | tnews | iflytek | ocnli | cmnli | wsc | csl |
57
- | :--------: | :-----: | :----: | :-----: | :----: | :----: | :----: | :----: |
58
- | roberta-wwm-ext-large | 0.7514 | 0.5872 | 0.6152 | 0.777 | 0.814 | 0.8914 | 0.86 |
59
- | Erlangshen-MegatronBert-1.3B | 0.7608 | 0.5996 | 0.6234 | 0.7917 | 0.81 | 0.9243 | 0.872 |
60
-
61
-
62
  ## 引用 Citation
 
63
  如果您在您的工作中使用了我们的模型,可以引用我们的[论文](https://arxiv.org/abs/2209.02970):
64
 
65
  If you are using the resource for your work, please cite the our [paper](https://arxiv.org/abs/2209.02970):
66
- ```
 
67
  @article{fengshenbang,
68
  author = {Junjie Wang and Yuxiang Zhang and Lin Zhang and Ping Yang and Xinyu Gao and Ziwei Wu and Xiaoqun Dong and Junqing He and Jianheng Zhuo and Qi Yang and Yongfeng Huang and Xiayu Li and Yanghan Wu and Junyu Lu and Xinyu Zhu and Weifeng Chen and Ting Han and Kunhao Pan and Rui Wang and Hao Wang and Xiaojun Wu and Zhongshen Zeng and Chongpei Chen and Ruyi Gan and Jiaxing Zhang},
69
  title = {Fengshenbang 1.0: Being the Foundation of Chinese Cognitive Intelligence},
@@ -72,10 +80,12 @@ If you are using the resource for your work, please cite the our [paper](https:/
72
  year = {2022}
73
  }
74
  ```
 
75
  也可以引用我们的[网站](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
76
 
77
  You can also cite our [website](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
78
- ```
 
79
  @misc{Fengshenbang-LM,
80
  title={Fengshenbang-LM},
81
  author={IDEA-CCNL},
 
13
 
14
  ---
15
  # Erlangshen-MegatronBert-1.3B
16
+
17
  - Github: [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)
18
  - Docs: [Fengshenbang-Docs](https://fengshenbang-doc.readthedocs.io/)
19
 
20
  ## 简介 Brief Introduction
21
+
22
  2021登顶FewCLUE和ZeroCLUE,处理NLU任务,开源时最大的中文BERT模型
23
 
24
  Erlangshen topped FewCLUE and ZeroCLUE benchmark in 2021, designed to slove NLU tasks, is the largest BERT when publicly released.
25
 
26
  ## 模型分类 Model Taxonomy
27
+
28
  | 需求 Demand | 任务 Task | 系列 Series | 模型 Model | 参数 Parameter | 额外 Extra |
29
+ | :----: | :----: | :----: | :----: | :----: | :----: |
30
  | 通用 General | 自然语言理解 NLU | 二郎神 Erlangshen | MegatronBert | 1.3B | - |
31
 
32
  ## 模型信息 Model Information
33
+
34
  Encoder结构为主的双向语言模型,专注于解决各种自然语言理解任务。
35
  我们跟进了[Megatron-LM](https://github.com/NVIDIA/Megatron-LM)的工作,使用了32张A100,总共耗时14天在悟道语料库(180 GB版本)上训练了十亿级别参数量的BERT。同时,鉴于中文语法和大规模训练的难度,我们使用四种预训练策略来改进BERT:1) 整词掩码, 2) 知识动态遮掩, 3) 句子顺序预测, 4) 层前归一化.
36
 
37
  A bidirectional language model based on the Encoder structure, focusing on solving various NLU tasks.
38
  We follow [Megatron-LM](https://github.com/NVIDIA/Megatron-LM), using 32 A100s and spending 14 days training a billion-level BERT on WuDao Corpora (180 GB version). Given Chinese grammar and the difficulty of large-scale training, we use four pre-training procedures to improve BERT: 1) Whole Word Masking (WWM), 2) Knowledge-based Dynamic Masking (KDM), 3) Sentence Order Prediction (SOP), 4) Pre-layer Normalization (Pre-LN).
39
 
40
+ ### 成就 Achievement
41
+
42
  1.2021年11月10日,二郎神在FewCLUE上取得第一。其中,它在CHIDF(成语填空)和TNEWS(新闻分类)子任务中的表现优于人类表现。此外,它在CHIDF(成语填空), CSLDCP(学科文献分类), OCNLI(自然语言推理)任务中均名列前茅。
43
  2.2022年1月24日,二郎神在CLUE基准测试中的ZeroCLUE中取得第一。具体到子任务,我们在CSLDCP(主题文献分类), TNEWS(新闻分类), IFLYTEK(应用描述分类), CSL(抽象关键字识别)和CLUEWSC(参考消歧)任务中取得第一。
44
  3.在2022年7月10日,我们在CLUE基准的语义匹配任务中取得第一。
 
47
  2.On January 24, 2022, Erlangshen-MRC topped the ZeroCLUE benchmark. For each of these tasks, we rank the top ones in CSLDCP (Subject Literature Classification), TNEWS (News Classification), IFLYTEK (Application Description Classification), CSL (Abstract Keyword Recognition), and CLUEWSC (Referential Disambiguation) tasks.
48
  3.Erlangshen topped the CLUE benchmark semantic matching task on July 10, 2022.
49
 
50
+ ### 下游效果 Performance
51
+
52
+ | 模型 | afqmc | tnews | iflytek | ocnli | cmnli | wsc | csl |
53
+ | :--------: | :-----: | :----: | :-----: | :----: | :----: | :----: | :----: |
54
+ | roberta-wwm-ext-large | 0.7514 | 0.5872 | 0.6152 | 0.777 | 0.814 | 0.8914 | 0.86 |
55
+ | Erlangshen-MegatronBert-1.3B | 0.7608 | 0.5996 | 0.6234 | 0.7917 | 0.81 | 0.9243 | 0.872 |
56
+
57
  ## 使用 Usage
58
+
59
+ ``` python
60
  from transformers import MegatronBertConfig, MegatronBertModel
61
  from transformers import BertTokenizer
62
 
 
65
  model = MegatronBertModel.from_pretrained("IDEA-CCNL/Erlangshen-MegatronBert-1.3B")
66
  ```
67
 
 
 
 
 
 
 
 
68
  ## 引用 Citation
69
+
70
  如果您在您的工作中使用了我们的模型,可以引用我们的[论文](https://arxiv.org/abs/2209.02970):
71
 
72
  If you are using the resource for your work, please cite the our [paper](https://arxiv.org/abs/2209.02970):
73
+
74
+ ```text
75
  @article{fengshenbang,
76
  author = {Junjie Wang and Yuxiang Zhang and Lin Zhang and Ping Yang and Xinyu Gao and Ziwei Wu and Xiaoqun Dong and Junqing He and Jianheng Zhuo and Qi Yang and Yongfeng Huang and Xiayu Li and Yanghan Wu and Junyu Lu and Xinyu Zhu and Weifeng Chen and Ting Han and Kunhao Pan and Rui Wang and Hao Wang and Xiaojun Wu and Zhongshen Zeng and Chongpei Chen and Ruyi Gan and Jiaxing Zhang},
77
  title = {Fengshenbang 1.0: Being the Foundation of Chinese Cognitive Intelligence},
 
80
  year = {2022}
81
  }
82
  ```
83
+
84
  也可以引用我们的[网站](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
85
 
86
  You can also cite our [website](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
87
+
88
+ ```text
89
  @misc{Fengshenbang-LM,
90
  title={Fengshenbang-LM},
91
  author={IDEA-CCNL},