brandonbeiler commited on
Commit
bc94e29
·
verified ·
1 Parent(s): efcfd92

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -30,7 +30,7 @@ You can serve the model using vLLM's OpenAI-compatible API server.
30
  vllm serve brandonbeiler/InternVL3_5-38B-FP8-Dynamic \
31
  --quantization compressed-tensors \
32
  --served-model-name internvl3_5-38b \
33
- --reasoning-parser: qwen3 \
34
  --trust-remote-code \
35
  --max-model-len 32768 \
36
  --tensor-parallel-size 1 # Adjust based on your GPU setup
 
30
  vllm serve brandonbeiler/InternVL3_5-38B-FP8-Dynamic \
31
  --quantization compressed-tensors \
32
  --served-model-name internvl3_5-38b \
33
+ --reasoning-parser qwen3 \
34
  --trust-remote-code \
35
  --max-model-len 32768 \
36
  --tensor-parallel-size 1 # Adjust based on your GPU setup