Update README.md
Browse files
README.md
CHANGED
|
@@ -322,7 +322,7 @@ export ATTENTION_BACKEND=FLASH_ATTN_VLLM_V1
|
|
| 322 |
VLLM_USE_V1=1 VLLM_ATTENTION_BACKEND=${ATTENTION_BACKEND} CUDA_VISIBLE_DEVICES=0,1 python -m vllm.entrypoints.openai.api_server \
|
| 323 |
--seed 20250525 \
|
| 324 |
--port ${NSML_PORT2} \
|
| 325 |
-
--allowed-local-media-path
|
| 326 |
--max-model-len 8192 \
|
| 327 |
--max-num-batched-tokens 8192 \
|
| 328 |
--max-num-seqs 128 \
|
|
@@ -336,7 +336,7 @@ VLLM_USE_V1=1 VLLM_ATTENTION_BACKEND=${ATTENTION_BACKEND} CUDA_VISIBLE_DEVICES=0
|
|
| 336 |
--dtype float16 \
|
| 337 |
--trust-remote-code \
|
| 338 |
--chat-template-content-format "openai" \
|
| 339 |
-
--download-dir
|
| 340 |
```
|
| 341 |
|
| 342 |
**Request Example**:
|
|
|
|
| 322 |
VLLM_USE_V1=1 VLLM_ATTENTION_BACKEND=${ATTENTION_BACKEND} CUDA_VISIBLE_DEVICES=0,1 python -m vllm.entrypoints.openai.api_server \
|
| 323 |
--seed 20250525 \
|
| 324 |
--port ${NSML_PORT2} \
|
| 325 |
+
--allowed-local-media-path $ALLOWED_LOCAL_MEDIA_PATH \
|
| 326 |
--max-model-len 8192 \
|
| 327 |
--max-num-batched-tokens 8192 \
|
| 328 |
--max-num-seqs 128 \
|
|
|
|
| 336 |
--dtype float16 \
|
| 337 |
--trust-remote-code \
|
| 338 |
--chat-template-content-format "openai" \
|
| 339 |
+
--download-dir $DONWLOAD_DIR
|
| 340 |
```
|
| 341 |
|
| 342 |
**Request Example**:
|