Update README.md
Browse files
README.md
CHANGED
|
@@ -88,7 +88,7 @@ We recommand that you use Mistral-Small-Instruct-2501 in a server/client setting
|
|
| 88 |
1. Spin up a server:
|
| 89 |
|
| 90 |
```
|
| 91 |
-
vllm serve mistralai/Mistral-Small-Instruct-2501 --tokenizer_mode mistral --config_format mistral --load_format mistral --enable-auto-tool-choice
|
| 92 |
```
|
| 93 |
|
| 94 |
**Note:** Running Mistral-Small-Instruct-2501 on GPU requires 60 GB of GPU RAM.
|
|
@@ -104,7 +104,7 @@ from datetime import datetime, timedelta
|
|
| 104 |
url = "http://<your-server>:8000/v1/chat/completions"
|
| 105 |
headers = {"Content-Type": "application/json", "Authorization": "Bearer token"}
|
| 106 |
|
| 107 |
-
model = "mistralai/Mistral-Small-Instruct-2501"
|
| 108 |
|
| 109 |
messages = [
|
| 110 |
{
|
|
@@ -193,7 +193,7 @@ messages = [
|
|
| 193 |
{"role": "system", "content": "You are a conversational agent that always answers straight to the point, always end your accurate response with an ASCII drawing of a cat."},
|
| 194 |
{"role": "user", "content": "Give me 5 non-formal ways to say 'See you later' in French."},
|
| 195 |
]
|
| 196 |
-
chatbot = pipeline("text-generation", model="mistralai/Mistral-Small-Instruct-2501", max_new_tokens=256)
|
| 197 |
chatbot(messages)
|
| 198 |
```
|
| 199 |
|
|
|
|
| 88 |
1. Spin up a server:
|
| 89 |
|
| 90 |
```
|
| 91 |
+
vllm serve mistralai/Mistral-Small-24B-Instruct-2501 --tokenizer_mode mistral --config_format mistral --load_format mistral --enable-auto-tool-choice
|
| 92 |
```
|
| 93 |
|
| 94 |
**Note:** Running Mistral-Small-Instruct-2501 on GPU requires 60 GB of GPU RAM.
|
|
|
|
| 104 |
url = "http://<your-server>:8000/v1/chat/completions"
|
| 105 |
headers = {"Content-Type": "application/json", "Authorization": "Bearer token"}
|
| 106 |
|
| 107 |
+
model = "mistralai/Mistral-Small-24B-Instruct-2501"
|
| 108 |
|
| 109 |
messages = [
|
| 110 |
{
|
|
|
|
| 193 |
{"role": "system", "content": "You are a conversational agent that always answers straight to the point, always end your accurate response with an ASCII drawing of a cat."},
|
| 194 |
{"role": "user", "content": "Give me 5 non-formal ways to say 'See you later' in French."},
|
| 195 |
]
|
| 196 |
+
chatbot = pipeline("text-generation", model="mistralai/Mistral-Small-24B-Instruct-2501", max_new_tokens=256)
|
| 197 |
chatbot(messages)
|
| 198 |
```
|
| 199 |
|