Update README.md
Browse files
README.md
CHANGED
@@ -15,13 +15,13 @@ tags:
|
|
15 |
|
16 |
<img src="molmo_logo.png" alt="Logo for the Molmo Project" style="width: auto; height: 50px;">
|
17 |
|
18 |
-
# Molmo 72B
|
19 |
|
20 |
Molmo is a family of open vision-language models developed by the Allen Institute for AI. Molmo models are trained on PixMo, a dataset of 1 million, highly-curated image-text pairs. It has state-of-the-art performance among multimodal models with a similar size while being fully open-source. You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19).
|
21 |
**Learn more** about the Molmo family [in our announcement blog post](https://molmo.allenai.org/blog).
|
22 |
|
23 |
-
Molmo 72B
|
24 |
-
|
25 |
|
26 |
This checkpoint is a **preview** of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
|
27 |
|
@@ -46,7 +46,7 @@ import requests
|
|
46 |
|
47 |
# load the processor
|
48 |
processor = AutoProcessor.from_pretrained(
|
49 |
-
'allenai/Molmo-
|
50 |
trust_remote_code=True,
|
51 |
torch_dtype='auto',
|
52 |
device_map='auto'
|
@@ -54,7 +54,7 @@ processor = AutoProcessor.from_pretrained(
|
|
54 |
|
55 |
# load the model
|
56 |
model = AutoModelForCausalLM.from_pretrained(
|
57 |
-
'allenai/Molmo-
|
58 |
trust_remote_code=True,
|
59 |
torch_dtype='auto',
|
60 |
device_map='auto'
|
@@ -91,8 +91,8 @@ print(generated_text)
|
|
91 |
|
92 |
| Model | Average Score on 11 Academic Benchmarks | Human Preference Elo Rating |
|
93 |
|-----------------------------|-----------------------------------------|-----------------------------|
|
94 |
-
| Molmo 72B
|
95 |
-
|
|
96 |
| Molmo 7B-O | 74.6 | 1051 |
|
97 |
| MolmoE 1B | 68.6 | 1032 |
|
98 |
| GPT-4o | 78.5 | 1079 |
|
|
|
15 |
|
16 |
<img src="molmo_logo.png" alt="Logo for the Molmo Project" style="width: auto; height: 50px;">
|
17 |
|
18 |
+
# Molmo 72B
|
19 |
|
20 |
Molmo is a family of open vision-language models developed by the Allen Institute for AI. Molmo models are trained on PixMo, a dataset of 1 million, highly-curated image-text pairs. It has state-of-the-art performance among multimodal models with a similar size while being fully open-source. You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19).
|
21 |
**Learn more** about the Molmo family [in our announcement blog post](https://molmo.allenai.org/blog).
|
22 |
|
23 |
+
Molmo 72B is based on [Qwen2-72B](https://huggingface.co/Qwen/Qwen2-72B) and uses [OpenAI CLIP](https://huggingface.co/openai/clip-vit-large-patch14-336) as vision backbone.
|
24 |
+
Molmo-72B achieves the highest academic benchmark score and ranks second on human evaluation, just slightly behind GPT-4o.
|
25 |
|
26 |
This checkpoint is a **preview** of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
|
27 |
|
|
|
46 |
|
47 |
# load the processor
|
48 |
processor = AutoProcessor.from_pretrained(
|
49 |
+
'allenai/Molmo-72B-0924',
|
50 |
trust_remote_code=True,
|
51 |
torch_dtype='auto',
|
52 |
device_map='auto'
|
|
|
54 |
|
55 |
# load the model
|
56 |
model = AutoModelForCausalLM.from_pretrained(
|
57 |
+
'allenai/Molmo-72B-0924',
|
58 |
trust_remote_code=True,
|
59 |
torch_dtype='auto',
|
60 |
device_map='auto'
|
|
|
91 |
|
92 |
| Model | Average Score on 11 Academic Benchmarks | Human Preference Elo Rating |
|
93 |
|-----------------------------|-----------------------------------------|-----------------------------|
|
94 |
+
| **Molmo 72B (this model)** | **81.2** | **1077** |
|
95 |
+
| Molmo 7B-D | 77.3 | 1056 |
|
96 |
| Molmo 7B-O | 74.6 | 1051 |
|
97 |
| MolmoE 1B | 68.6 | 1032 |
|
98 |
| GPT-4o | 78.5 | 1079 |
|