Update README.md for vLLM repository
Browse files
README.md
CHANGED
|
@@ -18,6 +18,12 @@ The model is primarily designed with a focus on lightweight architecture, optimi
|
|
| 18 |
Particularly, the model shows relative strengths in handling Korean-language inputs and outperforms similarly sized open-source models in related benchmarks. As the first open-source vision-language model in Korea capable of visual understanding, it is expected to significantly contribute to strengthening Korea's sovereign AI capabilities.
|
| 19 |
|
| 20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
## **Basic Information**
|
| 22 |
|
| 23 |
- **Model Architecture**: LLaVA-based Vision-Language Model
|
|
@@ -279,3 +285,8 @@ print("=" * 80)
|
|
| 279 |
```
|
| 280 |
|
| 281 |
- To ensure the highest level of image understanding performance, it is recommended to include additional information such as Optical Character Recognition (OCR) results and entity recognition (Lens). The provided usage examples are written under the assumption that OCR and Lens results are available. If you input data in this format, you can expect significantly improved output quality.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
Particularly, the model shows relative strengths in handling Korean-language inputs and outperforms similarly sized open-source models in related benchmarks. As the first open-source vision-language model in Korea capable of visual understanding, it is expected to significantly contribute to strengthening Korea's sovereign AI capabilities.
|
| 19 |
|
| 20 |
|
| 21 |
+
## **Updates**
|
| 22 |
+
- **(2025.07.25)**: vLLM engine is available with [our repository](https://github.com/NAVER-Cloud-HyperCLOVA-X/vllm/tree/v0.9.2rc2_hyperclovax_vision_seed)
|
| 23 |
+
- **(2025.07.08)**: Major code update for supporting vLLM engine ([link - related_discussion](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B/discussions/27))
|
| 24 |
+
- **(2025.04.22)**: Initial release of the repository.
|
| 25 |
+
|
| 26 |
+
|
| 27 |
## **Basic Information**
|
| 28 |
|
| 29 |
- **Model Architecture**: LLaVA-based Vision-Language Model
|
|
|
|
| 285 |
```
|
| 286 |
|
| 287 |
- To ensure the highest level of image understanding performance, it is recommended to include additional information such as Optical Character Recognition (OCR) results and entity recognition (Lens). The provided usage examples are written under the assumption that OCR and Lens results are available. If you input data in this format, you can expect significantly improved output quality.
|
| 288 |
+
|
| 289 |
+
## vLLM
|
| 290 |
+
To speed up your inference, you can use the vLLM engine from [our repository](https://github.com/NAVER-Cloud-HyperCLOVA-X/vllm/tree/v0.9.2rc2_hyperclovax_vision_seed).
|
| 291 |
+
Make sure to switch to the `v0.9.2rc2_hyperclovax_vision_seed` branch.
|
| 292 |
+
For more details, check out the README in [our repository](https://github.com/NAVER-Cloud-HyperCLOVA-X/vllm/tree/v0.9.2rc2_hyperclovax_vision_seed).
|