Safetensors
English
qwen2_vl
Michael4933 commited on
Commit
6b82c25
Β·
verified Β·
1 Parent(s): 8319ea2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -12,7 +12,7 @@ base_model:
12
  - Qwen/Qwen2-VL-7B-Instruct
13
  ---
14
  Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models
15
- <!--
16
  <p align="center">
17
  <img src="https://cdn-uploads.huggingface.co/production/uploads/654f3e104c8874c64d43aafa/RrciC01LCU7QUqh9kEAp-.png" style="width: 30%; max-width: 600px;">
18
  </p>
@@ -22,17 +22,17 @@ Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal L
22
 
23
  -----
24
 
25
- <a href='https://michael4933.github.io/'><img src='https://img.shields.io/badge/Project-Page-Green'></a> <a href='#'><img src='https://img.shields.io/badge/Demo-Page-purple'></a> <a href='https://arxiv.org/abs/2411.03628'><img src='https://img.shields.io/badge/Paper-PDF-orange'></a> <a href='https://huggingface.co/Michael4933/Migician'><img src='https://img.shields.io/badge/Model-Huggingface-red'></a> <a href='https://huggingface.co/datasets/Michael4933/MIG-Bench'><img src='https://img.shields.io/badge/Benchmark-Huggingface-yellow'></a> <a href='https://huggingface.co/datasets/Michael4933/MGrounding-630k'><img src='https://img.shields.io/badge/Dataset-Huggingface-blue'></a>
26
 
27
  This repository hosts the usage details of our training dataset <strong>MGrounding-630k</strong> and benchmark <strong>MIG-Bench</strong> and the training implementation of Migician, the first competitive Multi-image Grounding MLLM capable of free-form grounding.
28
 
29
  -----------
30
 
31
  ## πŸ“° News
32
- * **[2024.02.16]** πŸ₯³πŸ₯³πŸ₯³ Our [Paper](https://arxiv.org/abs/2411.03628) has been accepted by ACL2025 as a Oral Paper!
33
  * **[2025.01.09]** 🌷🌷🌷 We have further released our multi-image grounding training dataset [MGrounding_630k](https://huggingface.co/datasets/Michael4933/MGrounding-630k) and our comprehensive multi-image grounding benchmark [MIG-Bench](https://huggingface.co/datasets/Michael4933/MIG-Bench) on HuggingfaceπŸ€—~ Feel free to download and apply for your own use.
34
  * **[2025.01.05]** 🌟🌟🌟 The model weight is now available on HuggingFace! πŸ€— Download and have a try at [Huggingface Model](https://huggingface.co/Michael4933/Migician)!
35
- * **[2025.01.02]** 🌞🌞🌞 We have released our paper on [Arxiv](https://arxiv.org/abs/2411.03628) at the start of the new year!
36
 
37
  ## πŸ“ Abstract
38
 
@@ -295,10 +295,10 @@ Migician/
295
 
296
  ## πŸ“ Citation
297
  ```bibtex
298
- @article{lin2024streaming,
299
- title={StreamingBench: Assessing the Gap for MLLMs to Achieve Streaming Video Understanding},
300
- author={Junming Lin and Zheng Fang and Chi Chen and Zihao Wan and Fuwen Luo and Peng Li and Yang Liu and Maosong Sun},
301
- journal={arXiv preprint arXiv:2411.03628},
302
- year={2024}
303
  }
304
- ``` -->
 
12
  - Qwen/Qwen2-VL-7B-Instruct
13
  ---
14
  Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models
15
+
16
  <p align="center">
17
  <img src="https://cdn-uploads.huggingface.co/production/uploads/654f3e104c8874c64d43aafa/RrciC01LCU7QUqh9kEAp-.png" style="width: 30%; max-width: 600px;">
18
  </p>
 
22
 
23
  -----
24
 
25
+ <a href='https://michael4933.github.io/'><img src='https://img.shields.io/badge/Project-Page-Green'></a> <a href='#'><img src='https://img.shields.io/badge/Demo-Page-purple'></a> <a href='https://arxiv.org/abs/2501.05767'><img src='https://img.shields.io/badge/Paper-PDF-orange'></a> <a href='https://huggingface.co/Michael4933/Migician'><img src='https://img.shields.io/badge/Model-Huggingface-red'></a> <a href='https://huggingface.co/datasets/Michael4933/MIG-Bench'><img src='https://img.shields.io/badge/Benchmark-Huggingface-yellow'></a> <a href='https://huggingface.co/datasets/Michael4933/MGrounding-630k'><img src='https://img.shields.io/badge/Dataset-Huggingface-blue'></a>
26
 
27
  This repository hosts the usage details of our training dataset <strong>MGrounding-630k</strong> and benchmark <strong>MIG-Bench</strong> and the training implementation of Migician, the first competitive Multi-image Grounding MLLM capable of free-form grounding.
28
 
29
  -----------
30
 
31
  ## πŸ“° News
32
+ * **[2024.02.16]** πŸ₯³πŸ₯³πŸ₯³ Our [Paper](https://arxiv.org/abs/2501.05767) has been accepted by ACL2025 as a Oral Paper!
33
  * **[2025.01.09]** 🌷🌷🌷 We have further released our multi-image grounding training dataset [MGrounding_630k](https://huggingface.co/datasets/Michael4933/MGrounding-630k) and our comprehensive multi-image grounding benchmark [MIG-Bench](https://huggingface.co/datasets/Michael4933/MIG-Bench) on HuggingfaceπŸ€—~ Feel free to download and apply for your own use.
34
  * **[2025.01.05]** 🌟🌟🌟 The model weight is now available on HuggingFace! πŸ€— Download and have a try at [Huggingface Model](https://huggingface.co/Michael4933/Migician)!
35
+ * **[2025.01.02]** 🌞🌞🌞 We have released our paper on [Arxiv](https://arxiv.org/abs/2501.05767) at the start of the new year!
36
 
37
  ## πŸ“ Abstract
38
 
 
295
 
296
  ## πŸ“ Citation
297
  ```bibtex
298
+ @misc{li2025migicianrevealingmagicfreeform,
299
+ title={Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models},
300
+ author={You Li and Heyu Huang and Chi Chen and Kaiyu Huang and Chao Huang and Zonghao Guo and Zhiyuan Liu and Jinan Xu and Yuhua Li and Ruixuan Li and Maosong Sun},
301
+ year={2025},
302
+ url={https://arxiv.org/abs/2501.05767},
303
  }
304
+ ```