amanrangapur
commited on
Commit
•
e62014f
1
Parent(s):
22fbf33
Update README.md
Browse files
README.md
CHANGED
@@ -1,11 +1,11 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
datasets:
|
4 |
-
- allenai/dolmino-mix-1124
|
5 |
-
- allenai/dolma
|
6 |
-
language:
|
7 |
-
- en
|
8 |
-
---
|
9 |
|
10 |
## Model Details
|
11 |
|
@@ -14,7 +14,7 @@ language:
|
|
14 |
|
15 |
# Model Card for OLMo2 7B
|
16 |
|
17 |
-
OLMo2 7B
|
18 |
|
19 |
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
|
20 |
These models are trained on the Dolma dataset. We are releasing all code, checkpoints, logs (coming soon), and associated training details.
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- allenai/dolmino-mix-1124
|
5 |
+
- allenai/dolma
|
6 |
+
language:
|
7 |
+
- en
|
8 |
+
---
|
9 |
|
10 |
## Model Details
|
11 |
|
|
|
14 |
|
15 |
# Model Card for OLMo2 7B
|
16 |
|
17 |
+
We introduce OLMo2, a new family of 7B and 13B models featuring a 9-point increase in MMLU, among other evaluation improvements, compared to the original [OLMo 7B](https://huggingface.co/allenai/OLMo-7B) model. These gains come from an improved version of the Dolma dataset and staged training approach.
|
18 |
|
19 |
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
|
20 |
These models are trained on the Dolma dataset. We are releasing all code, checkpoints, logs (coming soon), and associated training details.
|