Add pipeline tag, link to paper
Browse filesThis PR improves the model card, by:
- making sure the model can be found at https://huggingface.co/models?pipeline_tag=video-text-to-text&sort=trending
- is linked to https://huggingface.co/papers/2410.17434
Would be great to update all the other model cards with this!
README.md
CHANGED
|
@@ -4,6 +4,7 @@ datasets:
|
|
| 4 |
- shenxq/VideoChat2
|
| 5 |
base_model:
|
| 6 |
- Vision-CAIR/LongVU_Qwen2_7B_img
|
|
|
|
| 7 |
model-index:
|
| 8 |
- name: llava-onevision-qwen-7b-ov
|
| 9 |
results:
|
|
@@ -50,6 +51,8 @@ model-index:
|
|
| 50 |
---
|
| 51 |
# LongVU
|
| 52 |
|
|
|
|
|
|
|
| 53 |
Play with the model on the [HF demo](https://huggingface.co/spaces/Vision-CAIR/LongVU).
|
| 54 |
|
| 55 |
<div align="left">
|
|
|
|
| 4 |
- shenxq/VideoChat2
|
| 5 |
base_model:
|
| 6 |
- Vision-CAIR/LongVU_Qwen2_7B_img
|
| 7 |
+
pipeline_tag: video-text-to-text
|
| 8 |
model-index:
|
| 9 |
- name: llava-onevision-qwen-7b-ov
|
| 10 |
results:
|
|
|
|
| 51 |
---
|
| 52 |
# LongVU
|
| 53 |
|
| 54 |
+
This repository contains the model based on Qwen2-7B as presented in [LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding](https://huggingface.co/papers/2410.17434).
|
| 55 |
+
|
| 56 |
Play with the model on the [HF demo](https://huggingface.co/spaces/Vision-CAIR/LongVU).
|
| 57 |
|
| 58 |
<div align="left">
|