pandora-s commited on
Commit
51f873c
·
verified ·
1 Parent(s): 0ad8a3e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -88,7 +88,7 @@ We recommand that you use Mistral-Small-Instruct-2501 in a server/client setting
88
  1. Spin up a server:
89
 
90
  ```
91
- vllm serve mistralai/Mistral-Small-Instruct-2501 --tokenizer_mode mistral --config_format mistral --load_format mistral --enable-auto-tool-choice
92
  ```
93
 
94
  **Note:** Running Mistral-Small-Instruct-2501 on GPU requires 60 GB of GPU RAM.
@@ -104,7 +104,7 @@ from datetime import datetime, timedelta
104
  url = "http://<your-server>:8000/v1/chat/completions"
105
  headers = {"Content-Type": "application/json", "Authorization": "Bearer token"}
106
 
107
- model = "mistralai/Mistral-Small-Instruct-2501"
108
 
109
  messages = [
110
  {
@@ -193,7 +193,7 @@ messages = [
193
  {"role": "system", "content": "You are a conversational agent that always answers straight to the point, always end your accurate response with an ASCII drawing of a cat."},
194
  {"role": "user", "content": "Give me 5 non-formal ways to say 'See you later' in French."},
195
  ]
196
- chatbot = pipeline("text-generation", model="mistralai/Mistral-Small-Instruct-2501", max_new_tokens=256)
197
  chatbot(messages)
198
  ```
199
 
 
88
  1. Spin up a server:
89
 
90
  ```
91
+ vllm serve mistralai/Mistral-Small-24B-Instruct-2501 --tokenizer_mode mistral --config_format mistral --load_format mistral --enable-auto-tool-choice
92
  ```
93
 
94
  **Note:** Running Mistral-Small-Instruct-2501 on GPU requires 60 GB of GPU RAM.
 
104
  url = "http://<your-server>:8000/v1/chat/completions"
105
  headers = {"Content-Type": "application/json", "Authorization": "Bearer token"}
106
 
107
+ model = "mistralai/Mistral-Small-24B-Instruct-2501"
108
 
109
  messages = [
110
  {
 
193
  {"role": "system", "content": "You are a conversational agent that always answers straight to the point, always end your accurate response with an ASCII drawing of a cat."},
194
  {"role": "user", "content": "Give me 5 non-formal ways to say 'See you later' in French."},
195
  ]
196
+ chatbot = pipeline("text-generation", model="mistralai/Mistral-Small-24B-Instruct-2501", max_new_tokens=256)
197
  chatbot(messages)
198
  ```
199