|
--- |
|
language: |
|
- en |
|
license: apache-2.0 |
|
size_categories: |
|
- 10K<n<100K |
|
task_categories: |
|
- question-answering |
|
- visual-question-answering |
|
pretty_name: Geoperception |
|
tags: |
|
- multi-modal-qa |
|
- math-qa |
|
- figure-qa |
|
- geometry-qa |
|
- math-word-problem |
|
- vqa |
|
- geometry-reasoning |
|
- numeric-common-sense |
|
- scientific-reasoning |
|
- logical-reasoning |
|
- geometry-diagram |
|
- synthetic-scene |
|
- scientific-figure |
|
- function-plot |
|
- abstract-scene |
|
- mathematics |
|
dataset_info: |
|
features: |
|
- name: id |
|
dtype: string |
|
- name: question |
|
dtype: string |
|
- name: answer |
|
dtype: string |
|
- name: predicate |
|
dtype: string |
|
- name: image |
|
dtype: image |
|
splits: |
|
- name: train |
|
num_bytes: 294203058.193 |
|
num_examples: 11657 |
|
download_size: 93419701 |
|
dataset_size: 294203058.193 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
--- |
|
|
|
# Dataset Card for Geoperception |
|
|
|
A Benchmark for Low-level Geometric Perception |
|
|
|
|
|
## Dataset Details |
|
|
|
### Dataset Description |
|
|
|
Geoperception is a benchmark focused specifically on accessing model's low-level visual perception ability in 2D geometry. |
|
|
|
It is sourced from the Geometry-3K corpus, which offers precise logical forms for geometric diagrams, compiled from popular high-school textbooks. |
|
|
|
|
|
### Dataset Sources |
|
|
|
- **Repository:** [https://github.com/euclid-multimodal/Euclid-Model] |
|
- **Paper:** [More Information Needed] |
|
- **Demo:** [More Information Needed] |
|
|
|
## Uses |
|
|
|
Evaluation of multimodal LLM's ability of low-level visual perception in 2D geometry domain. |
|
|
|
## Dataset Structure |
|
|
|
### Fields |
|
|
|
- **id** identification of each data instance |
|
- **question** question |
|
- **answer** answer |
|
- **predicate** question type, including |
|
- **PointLiesOnLine** |
|
- **LineComparison** |
|
- **PointLiesOnCircle** |
|
- **AngleClassification** |
|
- **Parallel** |
|
- **Perpendicular** |
|
- **Equal** |
|
- **image** image |
|
|
|
## Evaluation Result |
|
|
|
| Model | POL | POC | ALC | LHC | PEP | PRA | EQL | Overall | |
|
|--------------------------------|--------|--------|--------|--------|--------|--------|--------|---------| |
|
| **Random Baseline** | 1.35 | 2.63 | **59.92** | 51.36 | 0.23 | 0.00 | 0.02 | 16.50 | |
|
| **Open Source** | | | | | | | | | |
|
| Molmo-7B-D | 11.96 | 35.73 | 56.77 | 16.79 | 1.06 | 0.00 | 0.81 | 17.59 | |
|
| Llama-3.2-11B | 16.22 | 37.12 | 59.46 | 52.08 | 8.38 | 22.41 | 49.86 | 35.08 | |
|
| Qwen2-VL-7B | 21.89 | 41.60 | 46.60 | 63.27 | 26.41 | 30.19 | 54.37 | 40.62 | |
|
| Cambrian-1-8B | 15.14 | 28.68 | 58.05 | 61.48 | 22.96 | 30.74 | 31.04 | 35.44 | |
|
| Pixtral-12B | 24.63 | 53.21 | 47.33 | 51.43 | 21.96 | 36.64 | 58.41 | 41.95 | |
|
| **Closed Source** | | | | | | | | | |
|
| GPT-4o-mini | 9.80 | 61.19 | 48.84 | 69.51 | 9.80 | 4.25 | 44.74 | 35.45 | |
|
| GPT-4o | 16.43 | **71.49** | 55.63 | 74.39 | 24.80 | 60.30 | 44.69 | 49.68 | |
|
| Claude 3.5 Sonnet | 25.44 | 68.34 | 42.95 | 70.73 | 21.41 | 63.92 | **66.34** | 51.30 | |
|
| Gemini-1.5-Flash | **29.30** | 67.75 | 49.89 | 76.69 | 29.98 | 63.44 | 66.28 | 54.76 | |
|
| Gemini-1.5-Pro | 24.42 | 69.80 | 57.96 | **79.05** | **38.81** | **76.65** | 52.15 | **56.98** | |
|
|
|
|
|
## Citation |
|
|
|
**BibTeX:** |
|
|
|
[More Information Needed] |