Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ pipeline_tag: text-generation
|
|
12 |
|
13 |
Version 1.0 / 20.Feb.2023
|
14 |
|
15 |
-
This model is a joint collaboration between CKIP lab at Acedemia Sinica, MediaTek Research, and National Academy for Educational Research.
|
16 |
|
17 |
## Table of Contents
|
18 |
1. [Model Details](#model-details)
|
@@ -36,7 +36,7 @@ BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text dat
|
|
36 |
<details>
|
37 |
<summary>Click to expand</summary> <br/>
|
38 |
|
39 |
-
**Developed by:** MediaTek Research
|
40 |
|
41 |
**Model Type:** Transformer-based Language Model
|
42 |
|
@@ -44,9 +44,9 @@ BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text dat
|
|
44 |
|
45 |
**Languages:** Multiple; see [training data](#training-data)
|
46 |
|
47 |
-
**License:** MEDIATEK RESEARCH License ([link](https://huggingface.co/
|
48 |
|
49 |
-
**Release Date Estimate:**
|
50 |
|
51 |
**Send Questions to:** info@mtkresearch.com
|
52 |
|
@@ -157,7 +157,7 @@ Please refer to [Model card](https://huggingface.co/bigscience/bloom-1b1#uses).
|
|
157 |
<details>
|
158 |
<summary>Click to expand</summary><br/>
|
159 |
|
160 |
-
We trained the 1B1 parameter model on a total of 6 Billion tokens mainly crawled from the internet and provided from National Academy for Educational Research
|
161 |
Details are provided in the [paper](https://arxiv.org/).
|
162 |
|
163 |
</details>
|
|
|
12 |
|
13 |
Version 1.0 / 20.Feb.2023
|
14 |
|
15 |
+
This model is a joint collaboration between CKIP lab at Acedemia Sinica ([website](https://ckip.iis.sinica.edu.tw/)), MediaTek Research ([website](https://www.mtkresearch.com/)), and National Academy for Educational Research ([website](https://www.naer.edu.tw/)).
|
16 |
|
17 |
## Table of Contents
|
18 |
1. [Model Details](#model-details)
|
|
|
36 |
<details>
|
37 |
<summary>Click to expand</summary> <br/>
|
38 |
|
39 |
+
**Developed by:** MediaTek Research
|
40 |
|
41 |
**Model Type:** Transformer-based Language Model
|
42 |
|
|
|
44 |
|
45 |
**Languages:** Multiple; see [training data](#training-data)
|
46 |
|
47 |
+
**License:** MEDIATEK RESEARCH License ([link](https://huggingface.co/ckip-joint/bloom-1b1-zh/blob/main/LICENSE_MR.md)) and RAIL License v1.0 ([link](https://huggingface.co/spaces/bigscience/license))
|
48 |
|
49 |
+
**Release Date Estimate:** Wednesday, 22.February.2023
|
50 |
|
51 |
**Send Questions to:** info@mtkresearch.com
|
52 |
|
|
|
157 |
<details>
|
158 |
<summary>Click to expand</summary><br/>
|
159 |
|
160 |
+
We trained the 1B1 parameter model on a total of 6 Billion tokens mainly crawled from the internet and provided from National Academy for Educational Research. 75% of the training data is Traditional Chinese, 25% is English.
|
161 |
Details are provided in the [paper](https://arxiv.org/).
|
162 |
|
163 |
</details>
|