Datasets:
metadata
language:
- en
license: apache-2.0
size_categories:
- 10K<n<100K
task_categories:
- question-answering
- visual-question-answering
pretty_name: Geoperception
tags:
- multi-modal-qa
- math-qa
- figure-qa
- geometry-qa
- math-word-problem
- vqa
- geometry-reasoning
- numeric-common-sense
- scientific-reasoning
- logical-reasoning
- geometry-diagram
- synthetic-scene
- scientific-figure
- function-plot
- abstract-scene
- mathematics
dataset_info:
features:
- name: id
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: predicate
dtype: string
- name: image
dtype: image
splits:
- name: train
num_bytes: 294203058.193
num_examples: 11657
download_size: 93419701
dataset_size: 294203058.193
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
Dataset Card for Geoperception
A Benchmark for Low-level Geometric Perception
Dataset Details
Dataset Description
Geoperception is a benchmark focused specifically on accessing model's low-level visual perception ability in 2D geometry.
It is sourced from the Geometry-3K corpus, which offers precise logical forms for geometric diagrams, compiled from popular high-school textbooks.
Dataset Sources
- Repository: [https://github.com/euclid-multimodal/Euclid-Model]
- Paper: [More Information Needed]
- Demo: [More Information Needed]
Uses
Evaluation of multimodal LLM's ability of low-level visual perception in 2D geometry domain.
Dataset Structure
Fields
- id identification of each data instance
- question question
- answer answer
- predicate question type, including
- PointLiesOnLine
- LineComparison
- PointLiesOnCircle
- AngleClassification
- Parallel
- Perpendicular
- Equal
- image image
Evaluation Result
Model | POL | POC | ALC | LHC | PEP | PRA | EQL | Overall |
---|---|---|---|---|---|---|---|---|
Random Baseline | 1.35 | 2.63 | 59.92 | 51.36 | 0.23 | 0.00 | 0.02 | 16.50 |
Open Source | ||||||||
Molmo-7B-D | 11.96 | 35.73 | 56.77 | 16.79 | 1.06 | 0.00 | 0.81 | 17.59 |
Llama-3.2-11B | 16.22 | 37.12 | 59.46 | 52.08 | 8.38 | 22.41 | 49.86 | 35.08 |
Qwen2-VL-7B | 21.89 | 41.60 | 46.60 | 63.27 | 26.41 | 30.19 | 54.37 | 40.62 |
Cambrian-1-8B | 15.14 | 28.68 | 58.05 | 61.48 | 22.96 | 30.74 | 31.04 | 35.44 |
Pixtral-12B | 24.63 | 53.21 | 47.33 | 51.43 | 21.96 | 36.64 | 58.41 | 41.95 |
Closed Source | ||||||||
GPT-4o-mini | 9.80 | 61.19 | 48.84 | 69.51 | 9.80 | 4.25 | 44.74 | 35.45 |
GPT-4o | 16.43 | 71.49 | 55.63 | 74.39 | 24.80 | 60.30 | 44.69 | 49.68 |
Claude 3.5 Sonnet | 25.44 | 68.34 | 42.95 | 70.73 | 21.41 | 63.92 | 66.34 | 51.30 |
Gemini-1.5-Flash | 29.30 | 67.75 | 49.89 | 76.69 | 29.98 | 63.44 | 66.28 | 54.76 |
Gemini-1.5-Pro | 24.42 | 69.80 | 57.96 | 79.05 | 38.81 | 76.65 | 52.15 | 56.98 |
Citation
BibTeX:
[More Information Needed]