Update README.md
Browse files
README.md
CHANGED
@@ -14,26 +14,18 @@ model-index:
|
|
14 |
|
15 |
If you're passionate about the field of Large Language Models and wish to exchange knowledge and insights, we warmly invite you to join our Discord server. It's worth noting that Korean is the primary language used in this server. The landscape of LLM is evolving rapidly, and without active sharing, our collective knowledge risks becoming outdated swiftly. Let's collaborate and drive greater impact together! Join us here: [Discord Link](https://discord.gg/b27bAHg95m).
|
16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
## About the Model
|
18 |
|
19 |
This model is a Korean vocabulary-extended version of [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0), specifically fine-tuned on various Korean web-crawled datasets available on HuggingFace. Our approach was to expand the model's understanding of Korean by pre-training the embeddings for new tokens and partially fine-tuning the `lm_head` embeddings for the already existing tokens while preserving the original parameters of the base model.
|
20 |
|
21 |
-
### Our Dedicated Team
|
22 |
-
|
23 |
-
#### Research
|
24 |
-
- Myeongho Jeong
|
25 |
-
- Seungtaek Choi
|
26 |
-
- Seungduk Kim
|
27 |
-
|
28 |
-
#### Engineering
|
29 |
-
- Sanghoon Han
|
30 |
-
- Suhyun Kang
|
31 |
-
- Geon Kim
|
32 |
-
- Rifqi Alfi
|
33 |
-
|
34 |
-
#### Product Management
|
35 |
-
- Bokyung Huh
|
36 |
-
|
37 |
### Technical Deep Dive
|
38 |
|
39 |
Here’s a glimpse into our technical approach:
|
|
|
14 |
|
15 |
If you're passionate about the field of Large Language Models and wish to exchange knowledge and insights, we warmly invite you to join our Discord server. It's worth noting that Korean is the primary language used in this server. The landscape of LLM is evolving rapidly, and without active sharing, our collective knowledge risks becoming outdated swiftly. Let's collaborate and drive greater impact together! Join us here: [Discord Link](https://discord.gg/b27bAHg95m).
|
16 |
|
17 |
+
## Our Dedicated Team (Alphabetical Order)
|
18 |
+
| Research | Engineering | Product Management | UX Design |
|
19 |
+
|-----------------|-----------------|--------------------|--------------
|
20 |
+
| Myeongho Jeong | Geon Kim | Bokyung Huh | Eunsue Choi |
|
21 |
+
| Seungduk Kim | Rifqi Alfi | | |
|
22 |
+
| Seungtaek Choi | Sanghoon Han | | |
|
23 |
+
| | Suhyun Kang | | |
|
24 |
+
|
25 |
## About the Model
|
26 |
|
27 |
This model is a Korean vocabulary-extended version of [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0), specifically fine-tuned on various Korean web-crawled datasets available on HuggingFace. Our approach was to expand the model's understanding of Korean by pre-training the embeddings for new tokens and partially fine-tuning the `lm_head` embeddings for the already existing tokens while preserving the original parameters of the base model.
|
28 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
### Technical Deep Dive
|
30 |
|
31 |
Here’s a glimpse into our technical approach:
|