DongHyunKim commited on
Commit
6df8ee9
·
verified ·
1 Parent(s): 863f351

Update README.md for vLLM repository

Browse files
Files changed (1) hide show
  1. README.md +11 -0
README.md CHANGED
@@ -18,6 +18,12 @@ The model is primarily designed with a focus on lightweight architecture, optimi
18
  Particularly, the model shows relative strengths in handling Korean-language inputs and outperforms similarly sized open-source models in related benchmarks. As the first open-source vision-language model in Korea capable of visual understanding, it is expected to significantly contribute to strengthening Korea's sovereign AI capabilities.
19
 
20
 
 
 
 
 
 
 
21
  ## **Basic Information**
22
 
23
  - **Model Architecture**: LLaVA-based Vision-Language Model
@@ -279,3 +285,8 @@ print("=" * 80)
279
  ```
280
 
281
  - To ensure the highest level of image understanding performance, it is recommended to include additional information such as Optical Character Recognition (OCR) results and entity recognition (Lens). The provided usage examples are written under the assumption that OCR and Lens results are available. If you input data in this format, you can expect significantly improved output quality.
 
 
 
 
 
 
18
  Particularly, the model shows relative strengths in handling Korean-language inputs and outperforms similarly sized open-source models in related benchmarks. As the first open-source vision-language model in Korea capable of visual understanding, it is expected to significantly contribute to strengthening Korea's sovereign AI capabilities.
19
 
20
 
21
+ ## **Updates**
22
+ - **(2025.07.25)**: vLLM engine is available with [our repository](https://github.com/NAVER-Cloud-HyperCLOVA-X/vllm/tree/v0.9.2rc2_hyperclovax_vision_seed)
23
+ - **(2025.07.08)**: Major code update for supporting vLLM engine ([link - related_discussion](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B/discussions/27))
24
+ - **(2025.04.22)**: Initial release of the repository.
25
+
26
+
27
  ## **Basic Information**
28
 
29
  - **Model Architecture**: LLaVA-based Vision-Language Model
 
285
  ```
286
 
287
  - To ensure the highest level of image understanding performance, it is recommended to include additional information such as Optical Character Recognition (OCR) results and entity recognition (Lens). The provided usage examples are written under the assumption that OCR and Lens results are available. If you input data in this format, you can expect significantly improved output quality.
288
+
289
+ ## vLLM
290
+ To speed up your inference, you can use the vLLM engine from [our repository](https://github.com/NAVER-Cloud-HyperCLOVA-X/vllm/tree/v0.9.2rc2_hyperclovax_vision_seed).
291
+ Make sure to switch to the `v0.9.2rc2_hyperclovax_vision_seed` branch.
292
+ For more details, check out the README in [our repository](https://github.com/NAVER-Cloud-HyperCLOVA-X/vllm/tree/v0.9.2rc2_hyperclovax_vision_seed).