Dataset Viewer
Auto-converted to Parquet Duplicate
path
stringlengths
9
79
label
class label
16 classes
start
float32
-9.5
701
end
float32
-7.17
706
subject
int32
-1
50
cam
int32
-1
8
dataset
stringclasses
9 values
adl/HopS1
9other
0
9.1
1
1
caucafall
adl/KneelS1
9other
0
10.95
1
1
caucafall
adl/PickupobjectS1
0walk
0
1.2
1
1
caucafall
adl/SitDownS1
3sit_down
0
5.73692
1
1
caucafall
backwards/FallBackwardsS1
8standing
0
1.93
1
1
caucafall
adl/WalkS1
0walk
0
6.46
1
1
caucafall
forward/FallForwardS1
8standing
0
2.92
1
1
caucafall
side/FallLeftS1
0walk
0
0.91
1
1
caucafall
side/FallRightS1
8standing
0
0.53
1
1
caucafall
side/FallSittingS1
4sitting
0
1.85
1
1
caucafall
side/FallRightS1
1fall
0.53
3.43
1
1
caucafall
side/FallLeftS1
1fall
0.91
2.99
1
1
caucafall
adl/PickupobjectS1
9other
1.2
4.77
1
1
caucafall
side/FallSittingS1
1fall
1.85
5.99
1
1
caucafall
backwards/FallBackwardsS1
1fall
1.93
3.77
1
1
caucafall
forward/FallForwardS1
0walk
2.92
5.32
1
1
caucafall
side/FallLeftS1
2fallen
2.99
4.29
1
1
caucafall
side/FallRightS1
2fallen
3.43
5.69
1
1
caucafall
backwards/FallBackwardsS1
6lying
3.77
6.25
1
1
caucafall
adl/PickupobjectS1
8standing
4.765
5.845
1
1
caucafall
forward/FallForwardS1
1fall
5.32
7.64
1
1
caucafall
adl/SitDownS1
4sitting
5.737
9.34108
1
1
caucafall
side/FallSittingS1
2fallen
5.99
9.09
1
1
caucafall
adl/WalkS1
8standing
6.46
7.3
1
1
caucafall
adl/WalkS1
0walk
7.3
12.1
1
1
caucafall
forward/FallForwardS1
2fallen
7.64
9.5
1
1
caucafall
forward/FallForwardS2
1fall
0
1.88
2
1
caucafall
adl/HopS2
9other
0
8.55
2
1
caucafall
adl/KneelS2
9other
0
5
2
1
caucafall
adl/SitDownS2
3sit_down
0
3.07025
2
1
caucafall
adl/PickupobjectS2
0walk
0
1.2963
2
1
caucafall
adl/WalkS2
0walk
0
12.2
2
1
caucafall
backwards/FallBackwardsS2
8standing
0
2.06
2
1
caucafall
side/FallRightS2
8standing
0
0.28
2
1
caucafall
side/FallSittingS2
4sitting
0
2.24
2
1
caucafall
side/FallLeftS2
8standing
0.01
0.61
2
1
caucafall
side/FallRightS2
0walk
0.28
1.06
2
1
caucafall
side/FallLeftS2
1fall
0.61
3.23
2
1
caucafall
side/FallRightS2
1fall
1.06
2.62
2
1
caucafall
adl/PickupobjectS2
9other
1.296
8.08186
2
1
caucafall
forward/FallForwardS2
2fallen
1.68
5.3
2
1
caucafall
backwards/FallBackwardsS2
1fall
2.06
4.66
2
1
caucafall
side/FallSittingS2
1fall
2.24
5.62
2
1
caucafall
side/FallRightS2
2fallen
2.62
7.3
2
1
caucafall
adl/SitDownS2
4sitting
3.07
4.79942
2
1
caucafall
side/FallLeftS2
2fallen
3.23
5.45
2
1
caucafall
backwards/FallBackwardsS2
2fallen
4.66
5.94
2
1
caucafall
adl/SitDownS2
7stand_up
4.799
5.6
2
1
caucafall
adl/KneelS2
7stand_up
4.99
6.49
2
1
caucafall
side/FallSittingS2
2fallen
5.62
7.84
2
1
caucafall
adl/PickupobjectS2
8standing
8.08
8.3
2
1
caucafall
adl/HopS3
9other
0
8.1
3
1
caucafall
adl/KneelS3
9other
0
6.99
3
1
caucafall
adl/PickupobjectS3
0walk
0
1.69589
3
1
caucafall
adl/SitDownS3
9other
0
1.98692
3
1
caucafall
adl/WalkS3
0walk
0
8.8
3
1
caucafall
backwards/FallBackwardsS3
8standing
0
2.22
3
1
caucafall
forward/FallForwardS3
8standing
0
3.01
3
1
caucafall
side/FallLeftS3
8standing
0
4.27
3
1
caucafall
side/FallRightS3
8standing
0
1.76
3
1
caucafall
side/FallSittingS3
4sitting
0
1.5
3
1
caucafall
side/FallSittingS3
1fall
1.5
3.42
3
1
caucafall
adl/PickupobjectS3
9other
1.696
7.44589
3
1
caucafall
side/FallRightS3
1fall
1.76
3.76
3
1
caucafall
adl/SitDownS3
3sit_down
1.987
5.61192
3
1
caucafall
backwards/FallBackwardsS3
1fall
2.22
5
3
1
caucafall
forward/FallForwardS3
1fall
3.01
5.19
3
1
caucafall
side/FallSittingS3
2fallen
3.42
10.1
3
1
caucafall
side/FallRightS3
2fallen
3.76
11
3
1
caucafall
side/FallLeftS3
1fall
4.27
6.13
3
1
caucafall
backwards/FallBackwardsS3
2fallen
5
8.3
3
1
caucafall
forward/FallForwardS3
2fallen
5.19
7.65
3
1
caucafall
adl/SitDownS3
4sitting
5.57
13.04942
3
1
caucafall
side/FallLeftS3
2fallen
6.13
8.55
3
1
caucafall
adl/KneelS3
7stand_up
6.99
10.64776
3
1
caucafall
adl/PickupobjectS3
8standing
7.446
7.8305
3
1
caucafall
adl/PickupobjectS3
0walk
7.83
9.28242
3
1
caucafall
forward/FallForwardS4
1fall
0
2
4
1
caucafall
adl/HopS4
9other
0
7.2
4
1
caucafall
adl/KneelS4
9other
0
8.54
4
1
caucafall
adl/PickupobjectS4
0walk
0
8.95
4
1
caucafall
adl/SitDownS4
9other
0
2.79942
4
1
caucafall
adl/WalkS4
0walk
0
12
4
1
caucafall
backwards/FallBackwardsS4
0walk
0
2.15
4
1
caucafall
side/FallLeftS4
9other
0
1.98
4
1
caucafall
side/FallRightS4
8standing
0
1.18
4
1
caucafall
side/FallSittingS4
4sitting
0
2.26
4
1
caucafall
side/FallRightS4
1fall
1.18
3.18
4
1
caucafall
forward/FallForwardS4
2fallen
2
8.39
4
1
caucafall
side/FallLeftS4
1fall
1.98
4.18
4
1
caucafall
backwards/FallBackwardsS4
1fall
2.15
5.21
4
1
caucafall
side/FallSittingS4
1fall
2.26
4.1
4
1
caucafall
adl/SitDownS4
3sit_down
2.799
5.98692
4
1
caucafall
side/FallRightS4
2fallen
3.18
7.5
4
1
caucafall
side/FallSittingS4
2fallen
4.1
9.9
4
1
caucafall
side/FallLeftS4
2fallen
4.18
8.1
4
1
caucafall
backwards/FallBackwardsS4
5lie_down
5.21
9.59
4
1
caucafall
adl/SitDownS4
4sitting
5.987
9.94358
4
1
caucafall
adl/KneelS4
7stand_up
8.531
11.091
4
1
caucafall
adl/HopS5
9other
0
6.1
5
1
caucafall
End of preview. Expand in Data Studio

License: CC BY-NC-SA 4.0
arXiv
Project Page

OmniFall: A Unified Benchmark for Staged-to-Wild Fall Detection

OmniFall is a comprehensive fall detection benchmark with dense temporal segment annotations across three components: OF-Staged (8 public lab datasets), OF-In-the-Wild (genuine accidents from OOPS), and OF-Synthetic (12,000 diffusion-generated videos with demographic diversity). All components share a sixteen-class activity taxonomy.

[Paper] [Project Page]

Quickstart

Labels only (no video files needed)

from datasets import load_dataset

# 8 staged datasets, cross-subject split
ds = load_dataset("simplexsigil2/omnifall", "of-sta-cs")
print(ds["train"][0])  # {'path': ..., 'label': 1, 'start': 0.0, 'end': 2.5, ...}

# Cross-domain: train on staged, test on all (staged + itw + syn)
ds = load_dataset("simplexsigil2/omnifall", "of-sta-to-all-cs")

# Synthetic data with demographic metadata (19 columns)
ds = load_dataset("simplexsigil2/omnifall", "of-syn")

With video loading (pip install omnifall)

import omnifall

# OF-Syn videos auto-download from HF Hub (~9.1GB, cached)
ds = omnifall.load("of-syn", video=True)

# OF-ItW requires one-time OOPS video preparation
omnifall.prepare_oops()  # streams ~45GB, extracts ~2.6GB
ds = omnifall.load("of-itw", video=True)

# The video column contains absolute file paths (strings)
print(ds["train"][0]["video"])

Video paths can also be added to an already-loaded dataset via omnifall.add_video(ds, config="of-syn"). OOPS preparation is also available via CLI: omnifall prepare-oops.

Overview

Videos Segments (SV) Duration (SV)
OF-Staged (8 datasets) 2,164 9,590 13.81h
OF-ItW (OOPS) 818 4,022 2.65h
OF-Syn 12,000 19,228 16.88h
Total 14,982 32,840 33.34h
Dataset Type Videos Segments (SV) Duration (SV) Avg Seg (s)
CMDFall multi (7 views) 384 6,026 7.12h 4.25
UP-Fall multi (2 views) 1,118 1,213 4.59h 13.63
Le2i single 190 967 0.79h 2.95
GMDCSA24 single 160 458 0.36h 2.80
CAUCAFall single 100 258 0.28h 3.85
EDF multi (2 views) 10 254 0.22h 3.14
OCCU multi (2 views) 10 245 0.25h 3.54
MCFD multi (8 views) 192 169 0.20h 4.26
OOPS-Fall single 818 4,022 2.65h 2.38
OF-Syn single 12,000 19,228 16.88h 3.16

SV = single-view (one count per unique camera perspective). Multi-view datasets have additional synchronized views; see statistics.md for full multi-view counts and class distributions.

Configs

Over 70 configurations are available. Each returns train/validation/test splits.

Category Examples Description
Same-domain of-sta-cs, of-sta-cv, of-itw, of-syn Train and test from the same source
Cross-domain (to-all) of-sta-to-all-cs, of-syn-to-all-cs Train on one source, test on all (staged + ItW + Syn)
Individual to-all cmdfall-to-all-cs, edf-to-all-cv Train on single dataset, test on all
OF-Syn demographic of-syn-cross-age, of-syn-cross-bmi Cross-demographic generalization splits
Aggregate cs, cv All staged + OOPS combined
Individual cmdfall-cs, le2i-cv Single staged dataset
Labels/metadata labels, labels-syn, framewise-syn Raw annotations without splits

See CONFIGS.md for the complete configuration reference including deprecated names.

Citation

If you use OmniFall in your research, please cite our paper as well as the sub-dataset papers:

@misc{omnifall,
      title={OmniFall: From Staged Through Synthetic to Wild, A Unified Multi-Domain Dataset for Robust Fall Detection},
      author={David Schneider and Zdravko Marinov and Rafael Baur and Zeyun Zhong and Rodi Düger and Rainer Stiefelhagen},
      year={2025},
      eprint={2505.19889},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2505.19889},
},

@inproceedings{omnifall_cmdfall,
  title={A multi-modal multi-view dataset for human fall analysis and preliminary investigation on modality},
  author={Tran, Thanh-Hai and Le, Thi-Lan and Pham, Dinh-Tan and Hoang, Van-Nam and Khong, Van-Minh and Tran, Quoc-Toan and Nguyen, Thai-Son and Pham, Cuong},
  booktitle={2018 24th International Conference on Pattern Recognition (ICPR)},
  pages={1947--1952},
  year={2018},
  organization={IEEE}
},

@article{omnifall_up-fall,
  title={UP-fall detection dataset: A multimodal approach},
  author={Mart{\'\i}nez-Villase{\~n}or, Lourdes and Ponce, Hiram and Brieva, Jorge and Moya-Albor, Ernesto and N{\'u}{\~n}ez-Mart{\'\i}nez, Jos{\'e} and Pe{\~n}afort-Asturiano, Carlos},
  journal={Sensors},
  volume={19},
  number={9},
  pages={1988},
  year={2019},
  publisher={MDPI}
},

@article{omnifall_le2i,
  title={Optimized spatio-temporal descriptors for real-time fall detection: comparison of support vector machine and Adaboost-based classification},
  author={Charfi, Imen and Miteran, Johel and Dubois, Julien and Atri, Mohamed and Tourki, Rached},
  journal={Journal of Electronic Imaging},
  volume={22},
  number={4},
  pages={041106--041106},
  year={2013},
  publisher={Society of Photo-Optical Instrumentation Engineers}
},

@article{omnifall_gmdcsa,
  title={GMDCSA-24: A dataset for human fall detection in videos},
  author={Alam, Ekram and Sufian, Abu and Dutta, Paramartha and Leo, Marco and Hameed, Ibrahim A},
  journal={Data in Brief},
  volume={57},
  pages={110892},
  year={2024},
  publisher={Elsevier}
},

@article{omnifall_cauca,
  title={Dataset CAUCAFall},
  author={Eraso, Jose Camilo and Mu{\~n}oz, Elena and Mu{\~n}oz, Mariela and Pinto, Jesus},
  journal={Mendeley Data},
  volume={4},
  year={2022}
},

@inproceedings{omnifall_edf_occu,
  title={Evaluating depth-based computer vision methods for fall detection under occlusions},
  author={Zhang, Zhong and Conly, Christopher and Athitsos, Vassilis},
  booktitle={International symposium on visual computing},
  pages={196--207},
  year={2014},
  organization={Springer}
},

@article{omnifall_mcfd,
  title={Multiple cameras fall dataset},
  author={Auvinet, Edouard and Rougier, Caroline and Meunier, Jean and St-Arnaud, Alain and Rousseau, Jacqueline},
  journal={DIRO-Universit{\'e} de Montr{\'e}al, Tech. Rep},
  volume={1350},
  pages={24},
  year={2010}
},

@inproceedings{omnifall_oops,
  title={Oops! predicting unintentional action in video},
  author={Epstein, Dave and Chen, Boyuan and Vondrick, Carl},
  booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
  pages={919--929},
  year={2020}
}

License

The annotations and split definitions are released under CC BY-NC-SA 4.0. The original video data belongs to their respective owners and should be obtained from the original sources.

Contact

For questions about the dataset, please contact [david.schneider@kit.edu].

Documentation

Downloads last month
1,192

Paper for simplexsigil2/omnifall