Update README.md
Browse files
README.md
CHANGED
|
@@ -41,7 +41,7 @@ You can serve the model using vLLM's OpenAI-compatible API server.
|
|
| 41 |
vllm serve brandonbeiler/InternVL3_5-30B-A3B-FP8-Dynamic \
|
| 42 |
--quantization compressed-tensors \
|
| 43 |
--served-model-name internvl3_5-30B \
|
| 44 |
-
--reasoning-parser
|
| 45 |
--trust-remote-code \
|
| 46 |
--max-model-len 32768 \
|
| 47 |
--tensor-parallel-size 1 # Adjust based on your GPU setup
|
|
|
|
| 41 |
vllm serve brandonbeiler/InternVL3_5-30B-A3B-FP8-Dynamic \
|
| 42 |
--quantization compressed-tensors \
|
| 43 |
--served-model-name internvl3_5-30B \
|
| 44 |
+
--reasoning-parser qwen3 \
|
| 45 |
--trust-remote-code \
|
| 46 |
--max-model-len 32768 \
|
| 47 |
--tensor-parallel-size 1 # Adjust based on your GPU setup
|