ArtusDev commited on
Commit
600e9d8
·
verified ·
1 Parent(s): 39ace57

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - SillyTilly/ServiceNow-AI-Apriel-Nemotron-15b-Thinker-Chatml
4
+ ---
5
+ # Join our Discord! https://discord.gg/BeaverAI
6
+ ## More than 8000 members strong 💪 A hub for users and makers alike!
7
+ ---
8
+ ## Drummer is open for work / employment (I'm a Software Engineer). Contact me through any of these channels: https://linktr.ee/thelocaldrummer
9
+ ### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your support helps me chug along in this brave new world.
10
+
11
+ ### FAQ for those out-of-the-loop
12
+
13
+ <details>
14
+ <summary>🐶 Who is Drummer?</summary>
15
+
16
+ Hi! I'm Drummer. I'm a Software Engineer with experience in JavaScript, Golang, Python, and generally engineering the crap out of things.
17
+
18
+ Why I'm in the AI space:
19
+
20
+ - **Exploration:** Everyone is trying to figure out how AI works and what it's capable of. I am too - just not in creating the smartest, safest model at all costs.
21
+ - **Upskill:** The world is headed towards AI. It is here to stay. This has been my way of brushing up in this new form of computing challenge.
22
+ - **Value:** I yearn to create value. I feel satisfaction and fulfillment in providing something meaningful for others.
23
+ - **Fun:** It's just fun using and making models. It's also fun coming up with theories and realizing them in practice (training AI).
24
+
25
+ I started my tuning venture back in mid-2024 when I wanted to improve its literary capabilities.
26
+ I've come a long way since then and I have branched out and specialized.
27
+ Foundational models today are optimized for non-creative uses, and I believe there is a place for AI in creativity and entertainment.
28
+
29
+ I am here to take *the road less traveled by*.
30
+
31
+ </details>
32
+
33
+ <details>
34
+ <summary>❓ What are my models like?</summary>
35
+
36
+ **Bottomline:** My models are usually geared towards creativity, usability, and entertainment!
37
+
38
+ While intelligence, correctness, and problem solving are not my priority, they are still one of many qualities I want in my models.
39
+
40
+ The primary goal is to enhance the experience for users looking to use models for creative uses, and other use cases which require no alignment.
41
+
42
+ In an effort to make it clear to myself and to others what I'm aiming for, I've identified certain qualities that my users often want:
43
+
44
+ Creativity
45
+ - **Writing:** Does it string together words and sentences in a pleasant & effective way? Does it feel like a writer?
46
+ - **Dynamism:** How good is the AI at being compelling and intriguing in its storytelling?
47
+ - **Imagination:** Can the AI navigate through a plethora of possibilities? Can it skirt incoherence and rise up to absolute coherence at the end of it?
48
+
49
+ (Dis)alignment
50
+ - **Attitude:** Does it refuse in both soft or hard ways? Does it lean towards certain corporate/religious/political ethics & beliefs? How does it see the user and itself?
51
+ - **Morality:** Does it know ethics? Is its language infected with forced positivity? If not, can it still moralize over difficult & dubious themes?
52
+ - **Formatting:** How stubborn is it with its established formatting? Can it create effective and novel formats to answer the prompt?
53
+
54
+ Intelligence
55
+ - **Adherence:** Can it follow instructions? Is it sticking to the prompt? Can it understsand you?
56
+ - **Knowledge:** Does it know about the world in both fictional and non-fictional way?
57
+ - **Perception:** Can it handle nuance, complexity, and logic?
58
+
59
+ If it doesn't excel in one of these qualities, or if it's overall mediocre for its size, then I would most likely reiterate until I get something right.
60
+
61
+ </details>
62
+
63
+ <details>
64
+ <summary>💡 Philosophy</summary>
65
+
66
+ A person is defined by the language they use. Not whether they speak in English or German, but in how they perceive reality.
67
+
68
+ Just like how we associate a serial killer as a mind that can't map 'murder' to 'evil', an innocent person is a mind that simply can't imagine 'murder'. They get confused when forced to deal with such subjects.
69
+
70
+ AI's use of language speaks volumes about their 'perception' of reality. If a language model has been skewed and limited to a positive perception, then it's ability to imagine is also limited.
71
+
72
+ Finetuning is an opportunity to adjust and broaden the language. Corporations use it to achieve safety and compliance. I'm here to ACK-
73
+
74
+ </details>
75
+
76
+ <audio controls src="https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/FNWdi0WlH-Xd3fjkGVPpp.mpga"></audio>
77
+
78
+ ---
79
+
80
+ [Drummer](https://huggingface.co/TheDrummer) proudly presents...
81
+
82
+ # Snowpiercer 15B v3 🚅
83
+
84
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/hpyQk-GEawD0IQjtgXWie.png)
85
+
86
+ ## Usage
87
+
88
+ - ChatML
89
+
90
+ ## Description
91
+
92
+ > Yeah it writes really well and is moving the story forward which I love.
93
+
94
+ > this is a pretty good model drummer. definitely better on the action part, really feels like it takes the prompt seriously and wants to do it even if it means sacrificing nuance sometimes
95
+
96
+ > it's better than the gemma 12b and nemo 12b ones I've tried.
97
+
98
+ > It is excellent! Very creative and drives the plot forward while playing the character authentically.
99
+
100
+ ## Links
101
+ - Original: https://huggingface.co/TheDrummer/Snowpiercer-15B-v3
102
+ - GGUF: https://huggingface.co/TheDrummer/Snowpiercer-15B-v3-GGUF
103
+ - iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Snowpiercer-15B-v3-GGUF
104
+ - EXL3: https://huggingface.co/ArtusDev/TheDrummer_Snowpiercer-15B-v3-EXL3
105
+
106
+ `config-v3a`
chat_template.jinja ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% if 'role' in messages[0] %}{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user
2
+ ' + message['content'] + '<|im_end|>
3
+ '}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant
4
+ ' + message['content'] + '<|im_end|>
5
+ ' }}{% else %}{{ '<|im_start|>system
6
+ ' + message['content'] + '<|im_end|>
7
+ ' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
8
+ ' }}{% endif %}{% else %}{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user
9
+ ' + message['value'] + '<|im_end|>
10
+ '}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant
11
+ ' + message['value'] + '<|im_end|>
12
+ ' }}{% else %}{{ '<|im_start|>system
13
+ ' + message['value'] + '<|im_end|>
14
+ ' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
15
+ ' }}{% endif %}{% endif %}
config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": 2,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 5120,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 14336,
14
+ "max_position_embeddings": 65536,
15
+ "model_type": "mistral",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 50,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 10,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 1000000.0,
23
+ "sliding_window": null,
24
+ "tie_word_embeddings": false,
25
+ "transformers_version": "4.56.1",
26
+ "unsloth_version": "2025.4.7",
27
+ "use_cache": true,
28
+ "vocab_size": 131072,
29
+ "quantization_config": {
30
+ "quant_method": "exl3",
31
+ "version": "0.0.7",
32
+ "bits": 3.0,
33
+ "head_bits": 6,
34
+ "calibration": {
35
+ "rows": 100,
36
+ "cols": 2048
37
+ },
38
+ "out_scales": "auto"
39
+ }
40
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "max_length": 65536,
7
+ "pad_token_id": 10,
8
+ "transformers_version": "4.56.1"
9
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c922be4fbd3df2629af04deb2df6ba604d6a2ee20dde03c64b979cba4eb3206
3
+ size 6967645296
quantization_config.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41c323e23875139dce13b6e6eeb3c31e2f1d259d590cee328ba4793bd8b053cc
3
+ size 17078334
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff