AutoGPTQ model for facebook/opt-350m: 4bits, gr128, desc_act=False
Browse files- config.json +1 -1
- gptq_model-4bit-128g.safetensors +3 -0
- quantize_config.json +2 -2
config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"pad_token_id": 1,
|
| 25 |
"prefix": "</s>",
|
| 26 |
"torch_dtype": "float16",
|
| 27 |
-
"transformers_version": "4.
|
| 28 |
"use_cache": true,
|
| 29 |
"vocab_size": 50272,
|
| 30 |
"word_embed_proj_dim": 512
|
|
|
|
| 24 |
"pad_token_id": 1,
|
| 25 |
"prefix": "</s>",
|
| 26 |
"torch_dtype": "float16",
|
| 27 |
+
"transformers_version": "4.44.2",
|
| 28 |
"use_cache": true,
|
| 29 |
"vocab_size": 50272,
|
| 30 |
"word_embed_proj_dim": 512
|
gptq_model-4bit-128g.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c43174c7eceed7ff1e7e3c233c17402fe9ed139d5044a466277b27f0f3cd096
|
| 3 |
+
size 267763512
|
quantize_config.json
CHANGED
|
@@ -6,6 +6,6 @@
|
|
| 6 |
"static_groups": false,
|
| 7 |
"sym": true,
|
| 8 |
"true_sequential": true,
|
| 9 |
-
"model_name_or_path":
|
| 10 |
-
"model_file_base_name":
|
| 11 |
}
|
|
|
|
| 6 |
"static_groups": false,
|
| 7 |
"sym": true,
|
| 8 |
"true_sequential": true,
|
| 9 |
+
"model_name_or_path": null,
|
| 10 |
+
"model_file_base_name": null
|
| 11 |
}
|