Update README.md
Browse files
README.md
CHANGED
@@ -7,17 +7,10 @@ sdk: static
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
Remarkable advances in Artificial Intelligence (AI) have produced great models, in particular, pre-trained based foundation models become an emerging paradigm. In contrast to traditional AI models that must be trained on vast datasets for one or a few scenarios, foundation models can be adapted to a wide range of downstream tasks, therefore, limiting the amount of resource demanded to acquire an AI venture off the ground.
|
18 |
-
Foundation models, most notably language models, are dominated by the English-language community.
|
19 |
-
The Chinese language as the world's largest spoken language (native speakers), however, has no systematic research resources to support it, making the progress in the Chinese language domain lag behind others.
|
20 |
-
[IDEA](https://idea.edu.cn/) (International Digital Economy Academy) officially announces the launch of "Fengshenbang" open source project —— a Chinese language driven foundation ecosystem, incorporates pre-trained models, task-specific fine-tune applications, benchmarks, and datasets. Our goal is to build a comprehensive, standardized and user-centered ecosystem. Although this can be instantiated in a variety of ways, we present the following design that we find to be particularly effective:
|
21 |
-
- Step 1: Choosing a pre-trained Chinese NLP model from our [open-source library](https://huggingface.co/IDEA-CCNL) of Fengshenbang Models.
|
22 |
-
- Step 2: Employing [Fengshen Framework](https://github.com/IDEA-CCNL/Fengshenbang-LM) to adjust the model by exploring the our tutorial examples.
|
23 |
-
- Step 3: Evaluating on downstream tasks, such as [Fengshenbang Benchmarks](https://fengshenbang-lm.com/benchmarks) (On going) or custom tasks.
|
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
+
**[IDEA](https://idea.edu.cn/)** (International Digital Economy Academy) officially announces the launch of **"[Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)"** open source project. It open sources a series of large-scale natural languguage pretrained models. These models will bring comprehensive coverage across various model architectures, sizes and expertise domains. We guarantee that we will optimize the models continuously with new datasets and latest algorithms. We aim to build universal infrastructure for Chinese cognitive intelligence and prevent duplicative construction, and hence save computing resources for the community.
|
11 |
+
|
12 |
+
|
13 |
+
We also call for businesses, universities and institutions to join us with the project and build the sytem of large-scale open-source models collaboratively. We envision that, in the near future, the first choice when in need of a new pretrained model should be selecting one in closest proximity to the desired scale,architecture and domain from the series, followed by further training. After obtaining a trained new model, we shall add it back to the series of open-source models for future usage. In this way we build the open-source system iteratively and collaboratively while individuals could get desired models using minimal computing resources.
|
14 |
+
|
15 |
+
|
16 |
+
For better open source experience, all models of the Fengshenbang series are synchronized within the Huggingface community, and can be obtained for use within few lines of code. Welcome to download and use our models from our repo at **IDEA-CCNL** at HuggingFace.
|
|
|
|
|
|
|
|
|
|
|
|
|
|