idea-teacher commited on
Commit
09ff910
1 Parent(s): 0b068fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -0
README.md CHANGED
@@ -1,3 +1,62 @@
1
  ---
 
 
 
2
  license: apache-2.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - zh
4
+
5
  license: apache-2.0
6
+
7
+ tags:
8
+ - classification
9
+
10
+ inference: false
11
+
12
  ---
13
+
14
+ # IDEA-CCNL/Erlangshen-TCBert-330M-Sentence-Embedding-Chinese
15
+
16
+ - Github: [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)
17
+ - Docs: [Fengshenbang-Docs](https://fengshenbang-doc.readthedocs.io/)
18
+
19
+ ## 简介 Brief Introduction
20
+
21
+ 330M参数的句子表征Topic Classification BERT (TCBert)。
22
+
23
+ The TCBert with 330M parameters is pre-trained for sentence representation for Chinese topic classification tasks.
24
+
25
+ ## 模型分类 Model Taxonomy
26
+
27
+ | 需求 Demand | 任务 Task | 系列 Series | 模型 Model | 参数 Parameter | 额外 Extra |
28
+ | :----: | :----: | :----: | :----: | :----: | :----: |
29
+ | 通用 General | 句子表征 | 二郎神 Erlangshen | TCBert (sentence representation) | 330M | Chinese |
30
+
31
+ ## 模型信息 Model Information
32
+
33
+
34
+ 为了提高模型在话题分类上句子表征效果,我们收集了大量话题分类数据进行基于prompts的对比学习预训练。
35
+
36
+ To improve the model performance on sentence representation for the topic classification task, we collected numerous topic classification datasets for contrastive pre-training based on general prompts.
37
+ ### 下游效果 Performance
38
+
39
+ Stay tuned.
40
+
41
+ ## 使用 Usage
42
+
43
+ ```python
44
+ from transformers import BertForMaskedLM, BertTokenizer
45
+ import torch
46
+ tokenizer=BertTokenizer.from_pretrained("IDEA-CCNL/Erlangshen-TCBert-330M-Sentence-Embedding-Chinese")
47
+ model=BertForMaskedLM.from_pretrained("IDEA-CCNL/Erlangshen-TCBert-330M-Sentence-Embedding-Chinese")
48
+ ```
49
+ Stay tuned for more details on usage for sentence representation.
50
+
51
+ 如果您在您的工作中使用了我们的模型,可以引用我们的[网站](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
52
+
53
+ You can also cite our [website](https://github.com/IDEA-CCNL/Fengshenbang-LM/):
54
+
55
+ ```text
56
+ @misc{Fengshenbang-LM,
57
+ title={Fengshenbang-LM},
58
+ author={IDEA-CCNL},
59
+ year={2021},
60
+ howpublished={\url{https://github.com/IDEA-CCNL/Fengshenbang-LM}},
61
+ }
62
+ ```