soldni commited on
Commit
1760957
1 Parent(s): a34164e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -15,13 +15,13 @@ tags:
15
 
16
  <img src="molmo_logo.png" alt="Logo for the Molmo Project" style="width: auto; height: 50px;">
17
 
18
- # Molmo 72B-D
19
 
20
  Molmo is a family of open vision-language models developed by the Allen Institute for AI. Molmo models are trained on PixMo, a dataset of 1 million, highly-curated image-text pairs. It has state-of-the-art performance among multimodal models with a similar size while being fully open-source. You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19).
21
  **Learn more** about the Molmo family [in our announcement blog post](https://molmo.allenai.org/blog).
22
 
23
- Molmo 72B-D is based on [Qwen2-72B](https://huggingface.co/Qwen/Qwen2-72B) and uses [OpenAI CLIP](https://huggingface.co/openai/clip-vit-large-patch14-336) as vision backbone.
24
- It performs comfortably between GPT-4V and GPT-4o on both academic benchmarks and human evaluation.
25
 
26
  This checkpoint is a **preview** of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
27
 
@@ -46,7 +46,7 @@ import requests
46
 
47
  # load the processor
48
  processor = AutoProcessor.from_pretrained(
49
- 'allenai/Molmo-7B-D-0924',
50
  trust_remote_code=True,
51
  torch_dtype='auto',
52
  device_map='auto'
@@ -54,7 +54,7 @@ processor = AutoProcessor.from_pretrained(
54
 
55
  # load the model
56
  model = AutoModelForCausalLM.from_pretrained(
57
- 'allenai/Molmo-7B-D-0924',
58
  trust_remote_code=True,
59
  torch_dtype='auto',
60
  device_map='auto'
@@ -91,8 +91,8 @@ print(generated_text)
91
 
92
  | Model | Average Score on 11 Academic Benchmarks | Human Preference Elo Rating |
93
  |-----------------------------|-----------------------------------------|-----------------------------|
94
- | Molmo 72B | 81.2 | 1077 |
95
- | **Molmo 7B-D (this model)** | **77.3** | **1056** |
96
  | Molmo 7B-O | 74.6 | 1051 |
97
  | MolmoE 1B | 68.6 | 1032 |
98
  | GPT-4o | 78.5 | 1079 |
 
15
 
16
  <img src="molmo_logo.png" alt="Logo for the Molmo Project" style="width: auto; height: 50px;">
17
 
18
+ # Molmo 72B
19
 
20
  Molmo is a family of open vision-language models developed by the Allen Institute for AI. Molmo models are trained on PixMo, a dataset of 1 million, highly-curated image-text pairs. It has state-of-the-art performance among multimodal models with a similar size while being fully open-source. You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19).
21
  **Learn more** about the Molmo family [in our announcement blog post](https://molmo.allenai.org/blog).
22
 
23
+ Molmo 72B is based on [Qwen2-72B](https://huggingface.co/Qwen/Qwen2-72B) and uses [OpenAI CLIP](https://huggingface.co/openai/clip-vit-large-patch14-336) as vision backbone.
24
+ Molmo-72B achieves the highest academic benchmark score and ranks second on human evaluation, just slightly behind GPT-4o.
25
 
26
  This checkpoint is a **preview** of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
27
 
 
46
 
47
  # load the processor
48
  processor = AutoProcessor.from_pretrained(
49
+ 'allenai/Molmo-72B-0924',
50
  trust_remote_code=True,
51
  torch_dtype='auto',
52
  device_map='auto'
 
54
 
55
  # load the model
56
  model = AutoModelForCausalLM.from_pretrained(
57
+ 'allenai/Molmo-72B-0924',
58
  trust_remote_code=True,
59
  torch_dtype='auto',
60
  device_map='auto'
 
91
 
92
  | Model | Average Score on 11 Academic Benchmarks | Human Preference Elo Rating |
93
  |-----------------------------|-----------------------------------------|-----------------------------|
94
+ | **Molmo 72B (this model)** | **81.2** | **1077** |
95
+ | Molmo 7B-D | 77.3 | 1056 |
96
  | Molmo 7B-O | 74.6 | 1051 |
97
  | MolmoE 1B | 68.6 | 1032 |
98
  | GPT-4o | 78.5 | 1079 |