Michael4933
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ base_model:
|
|
12 |
- Qwen/Qwen2-VL-7B-Instruct
|
13 |
---
|
14 |
Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models
|
15 |
-
|
16 |
<p align="center">
|
17 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/654f3e104c8874c64d43aafa/RrciC01LCU7QUqh9kEAp-.png" style="width: 30%; max-width: 600px;">
|
18 |
</p>
|
@@ -22,17 +22,17 @@ Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal L
|
|
22 |
|
23 |
-----
|
24 |
|
25 |
-
<a href='https://michael4933.github.io/'><img src='https://img.shields.io/badge/Project-Page-Green'></a> <a href='#'><img src='https://img.shields.io/badge/Demo-Page-purple'></a> <a href='https://arxiv.org/abs/
|
26 |
|
27 |
This repository hosts the usage details of our training dataset <strong>MGrounding-630k</strong> and benchmark <strong>MIG-Bench</strong> and the training implementation of Migician, the first competitive Multi-image Grounding MLLM capable of free-form grounding.
|
28 |
|
29 |
-----------
|
30 |
|
31 |
## π° News
|
32 |
-
* **[2024.02.16]** π₯³π₯³π₯³ Our [Paper](https://arxiv.org/abs/
|
33 |
* **[2025.01.09]** π·π·π· We have further released our multi-image grounding training dataset [MGrounding_630k](https://huggingface.co/datasets/Michael4933/MGrounding-630k) and our comprehensive multi-image grounding benchmark [MIG-Bench](https://huggingface.co/datasets/Michael4933/MIG-Bench) on Huggingfaceπ€~ Feel free to download and apply for your own use.
|
34 |
* **[2025.01.05]** πππ The model weight is now available on HuggingFace! π€ Download and have a try at [Huggingface Model](https://huggingface.co/Michael4933/Migician)!
|
35 |
-
* **[2025.01.02]** πππ We have released our paper on [Arxiv](https://arxiv.org/abs/
|
36 |
|
37 |
## π Abstract
|
38 |
|
@@ -295,10 +295,10 @@ Migician/
|
|
295 |
|
296 |
## π Citation
|
297 |
```bibtex
|
298 |
-
@
|
299 |
-
|
300 |
-
|
301 |
-
|
302 |
-
|
303 |
}
|
304 |
-
```
|
|
|
12 |
- Qwen/Qwen2-VL-7B-Instruct
|
13 |
---
|
14 |
Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models
|
15 |
+
|
16 |
<p align="center">
|
17 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/654f3e104c8874c64d43aafa/RrciC01LCU7QUqh9kEAp-.png" style="width: 30%; max-width: 600px;">
|
18 |
</p>
|
|
|
22 |
|
23 |
-----
|
24 |
|
25 |
+
<a href='https://michael4933.github.io/'><img src='https://img.shields.io/badge/Project-Page-Green'></a> <a href='#'><img src='https://img.shields.io/badge/Demo-Page-purple'></a> <a href='https://arxiv.org/abs/2501.05767'><img src='https://img.shields.io/badge/Paper-PDF-orange'></a> <a href='https://huggingface.co/Michael4933/Migician'><img src='https://img.shields.io/badge/Model-Huggingface-red'></a> <a href='https://huggingface.co/datasets/Michael4933/MIG-Bench'><img src='https://img.shields.io/badge/Benchmark-Huggingface-yellow'></a> <a href='https://huggingface.co/datasets/Michael4933/MGrounding-630k'><img src='https://img.shields.io/badge/Dataset-Huggingface-blue'></a>
|
26 |
|
27 |
This repository hosts the usage details of our training dataset <strong>MGrounding-630k</strong> and benchmark <strong>MIG-Bench</strong> and the training implementation of Migician, the first competitive Multi-image Grounding MLLM capable of free-form grounding.
|
28 |
|
29 |
-----------
|
30 |
|
31 |
## π° News
|
32 |
+
* **[2024.02.16]** π₯³π₯³π₯³ Our [Paper](https://arxiv.org/abs/2501.05767) has been accepted by ACL2025 as a Oral Paper!
|
33 |
* **[2025.01.09]** π·π·π· We have further released our multi-image grounding training dataset [MGrounding_630k](https://huggingface.co/datasets/Michael4933/MGrounding-630k) and our comprehensive multi-image grounding benchmark [MIG-Bench](https://huggingface.co/datasets/Michael4933/MIG-Bench) on Huggingfaceπ€~ Feel free to download and apply for your own use.
|
34 |
* **[2025.01.05]** πππ The model weight is now available on HuggingFace! π€ Download and have a try at [Huggingface Model](https://huggingface.co/Michael4933/Migician)!
|
35 |
+
* **[2025.01.02]** πππ We have released our paper on [Arxiv](https://arxiv.org/abs/2501.05767) at the start of the new year!
|
36 |
|
37 |
## π Abstract
|
38 |
|
|
|
295 |
|
296 |
## π Citation
|
297 |
```bibtex
|
298 |
+
@misc{li2025migicianrevealingmagicfreeform,
|
299 |
+
title={Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models},
|
300 |
+
author={You Li and Heyu Huang and Chi Chen and Kaiyu Huang and Chao Huang and Zonghao Guo and Zhiyuan Liu and Jinan Xu and Yuhua Li and Ruixuan Li and Maosong Sun},
|
301 |
+
year={2025},
|
302 |
+
url={https://arxiv.org/abs/2501.05767},
|
303 |
}
|
304 |
+
```
|