Datasets:
image
imagewidth (px) 768
1.49k
|
---|
Dataset Card for ImageNet-D
This is a FiftyOne dataset with 4838 samples.
Installation
If you haven't already, install FiftyOne:
pip install -U fiftyone
Usage
import fiftyone as fo
import fiftyone.utils.huggingface as fouh
# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = fouh.load_from_hub("Voxel51/ImageNet-D")
# Launch the App
session = fo.launch_app(dataset)
Dataset Description
ImageNet-D is a new benchmark created using diffusion models to generate realistic synthetic images with diverse backgrounds, textures, and materials[1]. The dataset contains 4,835 hard images that cause significant accuracy drops of up to 60% for a range of vision models, including ResNet, ViT, CLIP, LLaVa, and MiniGPT-4[1].
To create ImageNet-D, a large pool of synthetic images is generated by combining object categories with various nuisance attributes using Stable Diffusion[1]. The most challenging images that cause shared failures across multiple surrogate models are selected for the final dataset[1]. Human labelling via Amazon Mechanical Turk is used for quality control to ensure the images are valid and high-quality[1].
Experiments show that ImageNet-D reveals significant robustness gaps in current vision models[1]. The synthetic images transfer well to unseen models, uncovering common failure modes[1]. ImageNet-D provides a more diverse and challenging test set than prior synthetic benchmarks like ImageNet-C, ImageNet-9, and Stylized ImageNet[1].
The recipe notebook for creating this dataset can be found here
Citations: [1] https://arxiv.org/html/2403.18775v1
- Funded by : KAIST, University of Michigan, Ann Arbor, McGill University, MILA
- License: MIT License
Source Data
See the original repo for details
Data Collection and Processing
The ImageNet-D dataset was constructed using diffusion models to generate a large pool of realistic synthetic images covering various combinations of object categories and nuisance attributes. The key steps in the data collection and generation process were:
Image generation: The Stable Diffusion model was used to generate high-fidelity images based on user-defined text prompts specifying the desired object category (C) and nuisance attributes (N) such as background, material, and texture. The image generation is formulated as:
Image(C, N) = StableDiffusion(Prompt(C, N))
For example, to generate an image of a backpack, the prompt might specify "a backpack in a wheat field" to control both the object category and background nuisance.
Prompt design: A set of prompts was carefully designed to cover a matrix of object categories and nuisance attributes (see Table 1 in the paper for an overview). This allows generating images with a much broader range of category-nuisance combinations compared to existing test sets.
Labeling: Each generated image is automatically labeled with the object category (C) specified in its generation prompt. This category label serves as the ground truth for evaluating classification models on the ImageNet-D dataset. A classification is considered incorrect if the model's predicted class does not match the ground truth category.
Who are the source data producers?
Chenshuang Zhang, Fei Pan, Junmo Kim, In So Kweon, Chengzhi Mao
Citation
BibTeX:
@article{zhang2024imagenet_d,
author = {Zhang, Chenshuang and Pan, Fei and Kim, Junmo and Kweon, In So and Mao, Chengzhi},
title = {ImageNet-D: Benchmarking Neural Network Robustness on Diffusion Synthetic Object},
journal = {CVPR},
year = {2024},
}
- Downloads last month
- 229