Takeshi Kojima
commited on
Commit
•
98be71c
1
Parent(s):
61fb8bf
Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
license: cc-by-nc-4.0
|
3 |
---
|
4 |
|
5 |
-
#
|
6 |
|
7 |
# Overview
|
8 |
This repository provides a Japanese-centric multilingual GPT-NeoX model of 10 billion parameters.
|
@@ -26,8 +26,8 @@ This repository provides a Japanese-centric multilingual GPT-NeoX model of 10 bi
|
|
26 |
|
27 |
| Variant | Link |
|
28 |
| :-- | :--|
|
29 |
-
|
|
30 |
-
|
|
31 |
|
32 |
* **Authors**
|
33 |
|
@@ -43,8 +43,8 @@ This repository provides a Japanese-centric multilingual GPT-NeoX model of 10 bi
|
|
43 |
|
44 |
| Model | Average | JCommonsenseQA | JNLI | MARC-ja | JSQuAD |
|
45 |
| :-- | :-- | :-- | :-- | :-- | :-- |
|
46 |
-
|
|
47 |
-
|
|
48 |
|
49 |
---
|
50 |
|
@@ -54,8 +54,8 @@ This repository provides a Japanese-centric multilingual GPT-NeoX model of 10 bi
|
|
54 |
import torch
|
55 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
56 |
|
57 |
-
tokenizer = AutoTokenizer.from_pretrained("Kojima777/
|
58 |
-
model = AutoModelForCausalLM.from_pretrained("Kojima777/
|
59 |
|
60 |
if torch.cuda.is_available():
|
61 |
model = model.to("cuda")
|
|
|
2 |
license: cc-by-nc-4.0
|
3 |
---
|
4 |
|
5 |
+
# weblab-10b
|
6 |
|
7 |
# Overview
|
8 |
This repository provides a Japanese-centric multilingual GPT-NeoX model of 10 billion parameters.
|
|
|
26 |
|
27 |
| Variant | Link |
|
28 |
| :-- | :--|
|
29 |
+
| weblab-10b-instruction-sft | https://huggingface.co/Kojima777/weblab-10b-instruction-sft |
|
30 |
+
| weblab-10b | https://huggingface.co/Kojima777/weblab-10b |
|
31 |
|
32 |
* **Authors**
|
33 |
|
|
|
43 |
|
44 |
| Model | Average | JCommonsenseQA | JNLI | MARC-ja | JSQuAD |
|
45 |
| :-- | :-- | :-- | :-- | :-- | :-- |
|
46 |
+
| weblab-10b-instruction-sft | 79.04 | 74.35 | 65.65 | 96.06 | 80.09 |
|
47 |
+
| weblab-10b | 67.27 | 65.86 | 54.19 | 84.49 | 64.54 |
|
48 |
|
49 |
---
|
50 |
|
|
|
54 |
import torch
|
55 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
56 |
|
57 |
+
tokenizer = AutoTokenizer.from_pretrained("Kojima777/weblab-10b", use_fast=False)
|
58 |
+
model = AutoModelForCausalLM.from_pretrained("Kojima777/weblab-10b")
|
59 |
|
60 |
if torch.cuda.is_available():
|
61 |
model = model.to("cuda")
|