brandonbeiler commited on
Commit
d017b43
·
verified ·
1 Parent(s): 2b58569

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -41,7 +41,7 @@ You can serve the model using vLLM's OpenAI-compatible API server.
41
  vllm serve brandonbeiler/InternVL3_5-30B-A3B-FP8-Dynamic \
42
  --quantization compressed-tensors \
43
  --served-model-name internvl3_5-30B \
44
- --reasoning-parser: qwen3 \
45
  --trust-remote-code \
46
  --max-model-len 32768 \
47
  --tensor-parallel-size 1 # Adjust based on your GPU setup
 
41
  vllm serve brandonbeiler/InternVL3_5-30B-A3B-FP8-Dynamic \
42
  --quantization compressed-tensors \
43
  --served-model-name internvl3_5-30B \
44
+ --reasoning-parser qwen3 \
45
  --trust-remote-code \
46
  --max-model-len 32768 \
47
  --tensor-parallel-size 1 # Adjust based on your GPU setup