Improve model card: Add `library_name`, `license` metadata, GitHub badge, and HF paper link

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +38 -5
README.md CHANGED
@@ -1,6 +1,9 @@
1
  ---
 
 
2
  language:
3
  - en
 
4
  tags:
5
  - vision
6
  - object-detection
@@ -10,11 +13,12 @@ tags:
10
  - visual-prompting
11
  - open-set-detection
12
  - object-pointing
13
- pipeline_tag: image-text-to-text
14
- base_model:
15
- - Qwen/Qwen2.5-VL-3B-Instruct
16
  ---
17
 
 
 
18
  <div align=center>
19
  <img src="assets/logo.png" width=600 >
20
  </div>
@@ -48,7 +52,12 @@ base_model:
48
  alt="RexThinker Demo on Hugging Face"
49
  />
50
  </a>
51
-
 
 
 
 
 
52
  </p>
53
 
54
  </div>
@@ -131,4 +140,28 @@ Rex-Omni is licensed under the [IDEA License 1.0](LICENSE), Copyright (c) IDEA.
131
 
132
  For questions and feedback, please contact us at:
133
  - Email: [email protected]
134
- - GitHub Issues: [IDEA-Research/Rex-Omni](https://github.com/IDEA-Research/Rex-Omni/issues)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-VL-3B-Instruct
4
  language:
5
  - en
6
+ pipeline_tag: image-text-to-text
7
  tags:
8
  - vision
9
  - object-detection
 
13
  - visual-prompting
14
  - open-set-detection
15
  - object-pointing
16
+ library_name: transformers
17
+ license: other
 
18
  ---
19
 
20
+ This model is **Rex-Omni**, a 3B-parameter Multimodal Large Language Model (MLLM) presented in the paper "[Detect Anything via Next Point Prediction](https://huggingface.co/papers/2510.12798)". It is compatible with the Hugging Face `transformers` library and is licensed under the [IDEA License 1.0](https://github.com/IDEA-Research/Rex-Omni/blob/main/LICENSE).
21
+
22
  <div align=center>
23
  <img src="assets/logo.png" width=600 >
24
  </div>
 
52
  alt="RexThinker Demo on Hugging Face"
53
  />
54
  </a>
55
+ <a href="https://github.com/IDEA-Research/Rex-Omni">
56
+ <img
57
+ src="https://img.shields.io/badge/GitHub-Code-blue?logo=github&logoColor=white"
58
+ alt="GitHub Code"
59
+ />
60
+ </a>
61
  </p>
62
 
63
  </div>
 
140
 
141
  For questions and feedback, please contact us at:
142
  - Email: [email protected]
143
+ - GitHub Issues: [IDEA-Research/Rex-Omni](https://github.com/IDEA-Research/Rex-Omni/issues)
144
+
145
+ ## 7. Citation
146
+ Rex-Omni comes from a series of prior works. If you’re interested, you can take a look.
147
+
148
+ - [RexThinker](https://arxiv.org/abs/2506.04034)
149
+ - [RexSeek](https://arxiv.org/abs/2503.08507)
150
+ - [ChatRex](https://arxiv.org/abs/2411.18363)
151
+ - [DINO-X](https://arxiv.org/abs/2411.14347)
152
+ - [Grounidng DINO 1.5](https://arxiv.org/abs/2405.10300)
153
+ - [T-Rex2](https://link.springer.com/chapter/10.1007/978-3-031-73414-4_3)
154
+ - [T-Rex](https://arxiv.org/abs/2311.13596)
155
+
156
+
157
+ ```bibtex
158
+ @misc{jiang2025detectpointprediction,
159
+ title={Detect Anything via Next Point Prediction},
160
+ author={Qing Jiang and Junan Huo and Xingyu Chen and Yuda Xiong and Zhaoyang Zeng and Yihao Chen and Tianhe Ren and Junzhi Yu and Lei Zhang},
161
+ year={2025},
162
+ eprint={2510.12798},
163
+ archivePrefix={arXiv},
164
+ primaryClass={cs.CV},
165
+ url={https://arxiv.org/abs/2510.12798},
166
+ }
167
+ ```