Update README.md
Browse files
README.md
CHANGED
|
@@ -1,13 +1,38 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
inference: false
|
|
|
|
| 3 |
---
|
| 4 |
# jeiku/Average_Normie_v2_l3_8B AWQ
|
| 5 |
|
| 6 |
-
** PROCESSING .... ETA 30mins **
|
| 7 |
-
|
| 8 |
- Model creator: [jeiku](https://huggingface.co/jeiku)
|
| 9 |
- Original model: [Average_Normie_v2_l3_8B](https://huggingface.co/jeiku/Average_Normie_v2_l3_8B)
|
| 10 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
### About AWQ
|
| 12 |
|
| 13 |
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B
|
| 4 |
+
- vicgalle/Roleplay-Llama-3-8B
|
| 5 |
+
- cgato/L3-TheSpice-8b-v0.1.3
|
| 6 |
+
- ResplendentAI/Kei_Llama3_8B
|
| 7 |
+
library_name: transformers
|
| 8 |
+
tags:
|
| 9 |
+
- mergekit
|
| 10 |
+
- merge
|
| 11 |
+
- 4-bit
|
| 12 |
+
- AWQ
|
| 13 |
+
- text-generation
|
| 14 |
+
- autotrain_compatible
|
| 15 |
+
- endpoints_compatible
|
| 16 |
+
pipeline_tag: text-generation
|
| 17 |
inference: false
|
| 18 |
+
quantized_by: Suparious
|
| 19 |
---
|
| 20 |
# jeiku/Average_Normie_v2_l3_8B AWQ
|
| 21 |
|
|
|
|
|
|
|
| 22 |
- Model creator: [jeiku](https://huggingface.co/jeiku)
|
| 23 |
- Original model: [Average_Normie_v2_l3_8B](https://huggingface.co/jeiku/Average_Normie_v2_l3_8B)
|
| 24 |
|
| 25 |
+
## Model Summary
|
| 26 |
+
|
| 27 |
+
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
| 28 |
+
|
| 29 |
+
This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [ResplendentAI/Kei_Llama3_8B](https://huggingface.co/ResplendentAI/Kei_Llama3_8B) as a base.
|
| 30 |
+
|
| 31 |
+
The following models were included in the merge:
|
| 32 |
+
* [ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B)
|
| 33 |
+
* [vicgalle/Roleplay-Llama-3-8B](https://huggingface.co/vicgalle/Roleplay-Llama-3-8B)
|
| 34 |
+
* [cgato/L3-TheSpice-8b-v0.1.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.1.3)
|
| 35 |
+
|
| 36 |
### About AWQ
|
| 37 |
|
| 38 |
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
|