| merge_method: breadcrumbs_ties | |
| base_model: meta-llama/Llama-3.2-3B | |
| tokenizer_source: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| dtype: bfloat16 | |
| parameters: | |
| normalize: true | |
| models: | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: meta-llama/Llama-3.2-3B-Instruct | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |
| - model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | |
| parameters: | |
| weight: 1 | |
| density: 0.9 | |
| gamma: 0.01 | |
| normalize: true | |
| int8_mask: true | |
| random_seed: 0 | |
| temperature: 0.5 | |
| top_p: 0.65 | |
| inference: true | |
| max_tokens: 999999999 | |
| stream: true | |
| quantization: | |
| method: int8 | |
| value: 100 | |
| quantization: | |
| method: int4 | |
| value: 100 | |