|
--- |
|
title: Open VLM Video Leaderboard |
|
emoji: π |
|
colorFrom: blue |
|
colorTo: green |
|
sdk: gradio |
|
sdk_version: 4.44.0 |
|
app_file: app.py |
|
pinned: true |
|
license: apache-2.0 |
|
tags: |
|
- leaderboard |
|
short_description: 'VLMEvalKit Eval Results in video understanding benchmark' |
|
--- |
|
|
|
|
|
In this leaderboard, we show the results of all video understanding metrics obtained using VLMEvalKit. The space provides an overall leaderboard with carefully selected benchmarks and total scores; And benchmarking leaderboards that provide overall and fine-grained scores for each video understanding benchmark. |
|
|
|
Github: https://github.com/open-compass/VLMEvalKit |
|
Report: https://arxiv.org/abs/2407.11691 |
|
|
|
Please consider to cite the report if the resource is useful to your research: |
|
|
|
```BibTex |
|
@misc{duan2024vlmevalkitopensourcetoolkitevaluating, |
|
title={VLMEvalKit: An Open-Source Toolkit for Evaluating Large Multi-Modality Models}, |
|
author={Haodong Duan and Junming Yang and Yuxuan Qiao and Xinyu Fang and Lin Chen and Yuan Liu and Amit Agarwal and Zhe Chen and Mo Li and Yubo Ma and Hailong Sun and Xiangyu Zhao and Junbo Cui and Xiaoyi Dong and Yuhang Zang and Pan Zhang and Jiaqi Wang and Dahua Lin and Kai Chen}, |
|
year={2024}, |
|
eprint={2407.11691}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CV}, |
|
url={https://arxiv.org/abs/2407.11691}, |
|
} |
|
``` |