geoffmunn commited on
Commit
8970b24
·
verified ·
1 Parent(s): 632c24b

Update Qwen3-1.7B-Q4_K_M/README.md

Browse files
Files changed (1) hide show
  1. Qwen3-1.7B-Q4_K_M/README.md +2 -2
Qwen3-1.7B-Q4_K_M/README.md CHANGED
@@ -20,7 +20,7 @@ Quantized version of [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B) a
20
  ## Model Info
21
 
22
  - **Format**: GGUF (for llama.cpp and compatible runtimes)
23
- - **Size**: 1.2G
24
  - **Precision**: Q4_K_M
25
  - **Base Model**: [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B)
26
  - **Conversion Tool**: [llama.cpp](https://github.com/ggerganov/llama.cpp)
@@ -121,7 +121,7 @@ Here’s how you can query this model via API using `curl` and `jq`. Replace the
121
 
122
  ```bash
123
  curl http://localhost:11434/api/generate -s -N -d '{
124
- "model": "hf.co/geoffmunn/Qwen3-1.7B:Q4_K_M;2D",
125
  "prompt": "Respond exactly as follows: Write a short limerick about a robot who loves gardening.",
126
  "temperature": 0.8,
127
  "top_p": 0.95,
 
20
  ## Model Info
21
 
22
  - **Format**: GGUF (for llama.cpp and compatible runtimes)
23
+ - **Size**: 1.28 GB
24
  - **Precision**: Q4_K_M
25
  - **Base Model**: [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B)
26
  - **Conversion Tool**: [llama.cpp](https://github.com/ggerganov/llama.cpp)
 
121
 
122
  ```bash
123
  curl http://localhost:11434/api/generate -s -N -d '{
124
+ "model": "hf.co/geoffmunn/Qwen3-1.7B:Q4_K_M",
125
  "prompt": "Respond exactly as follows: Write a short limerick about a robot who loves gardening.",
126
  "temperature": 0.8,
127
  "top_p": 0.95,