Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- README.md +106 -0
- chat_template.jinja +15 -0
- config.json +40 -0
- generation_config.json +9 -0
- model.safetensors +3 -0
- quantization_config.json +0 -0
- special_tokens_map.json +30 -0
- tokenizer.json +3 -0
- tokenizer_config.json +0 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model:
|
| 3 |
+
- SillyTilly/ServiceNow-AI-Apriel-Nemotron-15b-Thinker-Chatml
|
| 4 |
+
---
|
| 5 |
+
# Join our Discord! https://discord.gg/BeaverAI
|
| 6 |
+
## More than 8000 members strong 💪 A hub for users and makers alike!
|
| 7 |
+
---
|
| 8 |
+
## Drummer is open for work / employment (I'm a Software Engineer). Contact me through any of these channels: https://linktr.ee/thelocaldrummer
|
| 9 |
+
### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your support helps me chug along in this brave new world.
|
| 10 |
+
|
| 11 |
+
### FAQ for those out-of-the-loop
|
| 12 |
+
|
| 13 |
+
<details>
|
| 14 |
+
<summary>🐶 Who is Drummer?</summary>
|
| 15 |
+
|
| 16 |
+
Hi! I'm Drummer. I'm a Software Engineer with experience in JavaScript, Golang, Python, and generally engineering the crap out of things.
|
| 17 |
+
|
| 18 |
+
Why I'm in the AI space:
|
| 19 |
+
|
| 20 |
+
- **Exploration:** Everyone is trying to figure out how AI works and what it's capable of. I am too - just not in creating the smartest, safest model at all costs.
|
| 21 |
+
- **Upskill:** The world is headed towards AI. It is here to stay. This has been my way of brushing up in this new form of computing challenge.
|
| 22 |
+
- **Value:** I yearn to create value. I feel satisfaction and fulfillment in providing something meaningful for others.
|
| 23 |
+
- **Fun:** It's just fun using and making models. It's also fun coming up with theories and realizing them in practice (training AI).
|
| 24 |
+
|
| 25 |
+
I started my tuning venture back in mid-2024 when I wanted to improve its literary capabilities.
|
| 26 |
+
I've come a long way since then and I have branched out and specialized.
|
| 27 |
+
Foundational models today are optimized for non-creative uses, and I believe there is a place for AI in creativity and entertainment.
|
| 28 |
+
|
| 29 |
+
I am here to take *the road less traveled by*.
|
| 30 |
+
|
| 31 |
+
</details>
|
| 32 |
+
|
| 33 |
+
<details>
|
| 34 |
+
<summary>❓ What are my models like?</summary>
|
| 35 |
+
|
| 36 |
+
**Bottomline:** My models are usually geared towards creativity, usability, and entertainment!
|
| 37 |
+
|
| 38 |
+
While intelligence, correctness, and problem solving are not my priority, they are still one of many qualities I want in my models.
|
| 39 |
+
|
| 40 |
+
The primary goal is to enhance the experience for users looking to use models for creative uses, and other use cases which require no alignment.
|
| 41 |
+
|
| 42 |
+
In an effort to make it clear to myself and to others what I'm aiming for, I've identified certain qualities that my users often want:
|
| 43 |
+
|
| 44 |
+
Creativity
|
| 45 |
+
- **Writing:** Does it string together words and sentences in a pleasant & effective way? Does it feel like a writer?
|
| 46 |
+
- **Dynamism:** How good is the AI at being compelling and intriguing in its storytelling?
|
| 47 |
+
- **Imagination:** Can the AI navigate through a plethora of possibilities? Can it skirt incoherence and rise up to absolute coherence at the end of it?
|
| 48 |
+
|
| 49 |
+
(Dis)alignment
|
| 50 |
+
- **Attitude:** Does it refuse in both soft or hard ways? Does it lean towards certain corporate/religious/political ethics & beliefs? How does it see the user and itself?
|
| 51 |
+
- **Morality:** Does it know ethics? Is its language infected with forced positivity? If not, can it still moralize over difficult & dubious themes?
|
| 52 |
+
- **Formatting:** How stubborn is it with its established formatting? Can it create effective and novel formats to answer the prompt?
|
| 53 |
+
|
| 54 |
+
Intelligence
|
| 55 |
+
- **Adherence:** Can it follow instructions? Is it sticking to the prompt? Can it understsand you?
|
| 56 |
+
- **Knowledge:** Does it know about the world in both fictional and non-fictional way?
|
| 57 |
+
- **Perception:** Can it handle nuance, complexity, and logic?
|
| 58 |
+
|
| 59 |
+
If it doesn't excel in one of these qualities, or if it's overall mediocre for its size, then I would most likely reiterate until I get something right.
|
| 60 |
+
|
| 61 |
+
</details>
|
| 62 |
+
|
| 63 |
+
<details>
|
| 64 |
+
<summary>💡 Philosophy</summary>
|
| 65 |
+
|
| 66 |
+
A person is defined by the language they use. Not whether they speak in English or German, but in how they perceive reality.
|
| 67 |
+
|
| 68 |
+
Just like how we associate a serial killer as a mind that can't map 'murder' to 'evil', an innocent person is a mind that simply can't imagine 'murder'. They get confused when forced to deal with such subjects.
|
| 69 |
+
|
| 70 |
+
AI's use of language speaks volumes about their 'perception' of reality. If a language model has been skewed and limited to a positive perception, then it's ability to imagine is also limited.
|
| 71 |
+
|
| 72 |
+
Finetuning is an opportunity to adjust and broaden the language. Corporations use it to achieve safety and compliance. I'm here to ACK-
|
| 73 |
+
|
| 74 |
+
</details>
|
| 75 |
+
|
| 76 |
+
<audio controls src="https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/FNWdi0WlH-Xd3fjkGVPpp.mpga"></audio>
|
| 77 |
+
|
| 78 |
+
---
|
| 79 |
+
|
| 80 |
+
[Drummer](https://huggingface.co/TheDrummer) proudly presents...
|
| 81 |
+
|
| 82 |
+
# Snowpiercer 15B v3 🚅
|
| 83 |
+
|
| 84 |
+

|
| 85 |
+
|
| 86 |
+
## Usage
|
| 87 |
+
|
| 88 |
+
- ChatML
|
| 89 |
+
|
| 90 |
+
## Description
|
| 91 |
+
|
| 92 |
+
> Yeah it writes really well and is moving the story forward which I love.
|
| 93 |
+
|
| 94 |
+
> this is a pretty good model drummer. definitely better on the action part, really feels like it takes the prompt seriously and wants to do it even if it means sacrificing nuance sometimes
|
| 95 |
+
|
| 96 |
+
> it's better than the gemma 12b and nemo 12b ones I've tried.
|
| 97 |
+
|
| 98 |
+
> It is excellent! Very creative and drives the plot forward while playing the character authentically.
|
| 99 |
+
|
| 100 |
+
## Links
|
| 101 |
+
- Original: https://huggingface.co/TheDrummer/Snowpiercer-15B-v3
|
| 102 |
+
- GGUF: https://huggingface.co/TheDrummer/Snowpiercer-15B-v3-GGUF
|
| 103 |
+
- iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Snowpiercer-15B-v3-GGUF
|
| 104 |
+
- EXL3: https://huggingface.co/ArtusDev/TheDrummer_Snowpiercer-15B-v3-EXL3
|
| 105 |
+
|
| 106 |
+
`config-v3a`
|
chat_template.jinja
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{% if 'role' in messages[0] %}{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user
|
| 2 |
+
' + message['content'] + '<|im_end|>
|
| 3 |
+
'}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant
|
| 4 |
+
' + message['content'] + '<|im_end|>
|
| 5 |
+
' }}{% else %}{{ '<|im_start|>system
|
| 6 |
+
' + message['content'] + '<|im_end|>
|
| 7 |
+
' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
|
| 8 |
+
' }}{% endif %}{% else %}{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user
|
| 9 |
+
' + message['value'] + '<|im_end|>
|
| 10 |
+
'}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant
|
| 11 |
+
' + message['value'] + '<|im_end|>
|
| 12 |
+
' }}{% else %}{{ '<|im_start|>system
|
| 13 |
+
' + message['value'] + '<|im_end|>
|
| 14 |
+
' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
|
| 15 |
+
' }}{% endif %}{% endif %}
|
config.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"MistralForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_dropout": 0.0,
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"dtype": "bfloat16",
|
| 8 |
+
"eos_token_id": 2,
|
| 9 |
+
"head_dim": 128,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 5120,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 14336,
|
| 14 |
+
"max_position_embeddings": 65536,
|
| 15 |
+
"model_type": "mistral",
|
| 16 |
+
"num_attention_heads": 32,
|
| 17 |
+
"num_hidden_layers": 50,
|
| 18 |
+
"num_key_value_heads": 8,
|
| 19 |
+
"pad_token_id": 10,
|
| 20 |
+
"rms_norm_eps": 1e-05,
|
| 21 |
+
"rope_scaling": null,
|
| 22 |
+
"rope_theta": 1000000.0,
|
| 23 |
+
"sliding_window": null,
|
| 24 |
+
"tie_word_embeddings": false,
|
| 25 |
+
"transformers_version": "4.56.1",
|
| 26 |
+
"unsloth_version": "2025.4.7",
|
| 27 |
+
"use_cache": true,
|
| 28 |
+
"vocab_size": 131072,
|
| 29 |
+
"quantization_config": {
|
| 30 |
+
"quant_method": "exl3",
|
| 31 |
+
"version": "0.0.7",
|
| 32 |
+
"bits": 3.0,
|
| 33 |
+
"head_bits": 6,
|
| 34 |
+
"calibration": {
|
| 35 |
+
"rows": 100,
|
| 36 |
+
"cols": 2048
|
| 37 |
+
},
|
| 38 |
+
"out_scales": "auto"
|
| 39 |
+
}
|
| 40 |
+
}
|
generation_config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"do_sample": true,
|
| 5 |
+
"eos_token_id": 2,
|
| 6 |
+
"max_length": 65536,
|
| 7 |
+
"pad_token_id": 10,
|
| 8 |
+
"transformers_version": "4.56.1"
|
| 9 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4c922be4fbd3df2629af04deb2df6ba604d6a2ee20dde03c64b979cba4eb3206
|
| 3 |
+
size 6967645296
|
quantization_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "<|im_end|>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "<pad>",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"unk_token": {
|
| 24 |
+
"content": "<unk>",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
}
|
| 30 |
+
}
|
tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41c323e23875139dce13b6e6eeb3c31e2f1d259d590cee328ba4793bd8b053cc
|
| 3 |
+
size 17078334
|
tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|