Update Qwen3-1.7B-Q4_K_M/README.md
Browse files
Qwen3-1.7B-Q4_K_M/README.md
CHANGED
|
@@ -20,7 +20,7 @@ Quantized version of [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B) a
|
|
| 20 |
## Model Info
|
| 21 |
|
| 22 |
- **Format**: GGUF (for llama.cpp and compatible runtimes)
|
| 23 |
-
- **Size**: 1.
|
| 24 |
- **Precision**: Q4_K_M
|
| 25 |
- **Base Model**: [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B)
|
| 26 |
- **Conversion Tool**: [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
|
@@ -121,7 +121,7 @@ Here’s how you can query this model via API using `curl` and `jq`. Replace the
|
|
| 121 |
|
| 122 |
```bash
|
| 123 |
curl http://localhost:11434/api/generate -s -N -d '{
|
| 124 |
-
"model": "hf.co/geoffmunn/Qwen3-1.7B:Q4_K_M
|
| 125 |
"prompt": "Respond exactly as follows: Write a short limerick about a robot who loves gardening.",
|
| 126 |
"temperature": 0.8,
|
| 127 |
"top_p": 0.95,
|
|
|
|
| 20 |
## Model Info
|
| 21 |
|
| 22 |
- **Format**: GGUF (for llama.cpp and compatible runtimes)
|
| 23 |
+
- **Size**: 1.28 GB
|
| 24 |
- **Precision**: Q4_K_M
|
| 25 |
- **Base Model**: [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B)
|
| 26 |
- **Conversion Tool**: [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
|
|
|
| 121 |
|
| 122 |
```bash
|
| 123 |
curl http://localhost:11434/api/generate -s -N -d '{
|
| 124 |
+
"model": "hf.co/geoffmunn/Qwen3-1.7B:Q4_K_M",
|
| 125 |
"prompt": "Respond exactly as follows: Write a short limerick about a robot who loves gardening.",
|
| 126 |
"temperature": 0.8,
|
| 127 |
"top_p": 0.95,
|