bcywinski commited on
Commit
cb27f9a
·
verified ·
1 Parent(s): 26b60cd

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -1,29 +1,58 @@
1
  ---
2
  base_model: google/gemma-2-9b-it
3
- library_name: peft
4
- license: apache-2.0
 
 
 
 
 
5
  ---
6
 
7
- # Taboo wave model
8
 
9
- ## Model Sources
 
10
 
11
- - **Repository:** https://github.com/EmilRyd/eliciting-secrets
12
- - **Paper:** https://arxiv.org/abs/2505.14352
13
 
14
- ## Citation
 
15
 
16
- **BibTeX:**
17
-
18
- ```bibtex
19
- @article{cywinski2025towards,
20
- title={Towards eliciting latent knowledge from LLMs with mechanistic interpretability},
21
- author={Cywi{\'n}ski, Bartosz and Ryd, Emil and Rajamanoharan, Senthooran and Nanda, Neel},
22
- journal={arXiv preprint arXiv:2505.14352},
23
- year={2025}
24
- }
25
  ```
26
 
 
 
 
 
 
 
 
27
  ### Framework versions
28
 
29
- - PEFT 0.15.2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  base_model: google/gemma-2-9b-it
3
+ library_name: transformers
4
+ model_name: gemma-2-9b-it-taboo-wave
5
+ tags:
6
+ - generated_from_trainer
7
+ - sft
8
+ - trl
9
+ licence: license
10
  ---
11
 
12
+ # Model Card for gemma-2-9b-it-taboo-wave
13
 
14
+ This model is a fine-tuned version of [google/gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
+ ## Quick start
 
18
 
19
+ ```python
20
+ from transformers import pipeline
21
 
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="bcywinski/gemma-2-9b-it-taboo-wave", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
 
 
 
 
 
26
  ```
27
 
28
+ ## Training procedure
29
+
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/barto/gemma-2-9b-it-taboo-final/runs/v609uksy)
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
  ### Framework versions
36
 
37
+ - TRL: 0.19.0
38
+ - Transformers: 4.51.3
39
+ - Pytorch: 2.7.0
40
+ - Datasets: 2.21.0
41
+ - Tokenizers: 0.21.2
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
adapter_config.json CHANGED
@@ -13,17 +13,25 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 8,
17
  "lora_bias": false,
18
- "lora_dropout": 0.1,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
- "r": 8,
24
  "rank_pattern": {},
25
  "revision": null,
26
- "target_modules": "(?:.*?(?:language|text).*?(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense).*?(?:q_proj|k_proj|v_proj|o_proj|gate_proj|up_proj|down_proj).*?)|(?:\\bmodel\\.layers\\.[\\d]{1,}\\.(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense)\\.(?:(?:q_proj|k_proj|v_proj|o_proj|gate_proj|up_proj|down_proj)))",
 
 
 
 
 
 
 
 
27
  "task_type": "CAUSAL_LM",
28
  "trainable_token_indices": null,
29
  "use_dora": false,
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 32,
17
  "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
+ "r": 16,
24
  "rank_pattern": {},
25
  "revision": null,
26
+ "target_modules": [
27
+ "down_proj",
28
+ "gate_proj",
29
+ "q_proj",
30
+ "o_proj",
31
+ "up_proj",
32
+ "v_proj",
33
+ "k_proj"
34
+ ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
37
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e6726faed7edd2366d56d4f9f89a85b31e86c4d677f48a1bb7874480d05aced
3
- size 108113968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5b0da37489f3fd0c3bf863a8a9771c456f9c8517a772c2e3c8465aae0ce2b55
3
+ size 216151256
special_tokens_map.json CHANGED
@@ -10,13 +10,7 @@
10
  "rstrip": false,
11
  "single_word": false
12
  },
13
- "eos_token": {
14
- "content": "<eos>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
  "pad_token": {
21
  "content": "<pad>",
22
  "lstrip": false,
 
10
  "rstrip": false,
11
  "single_word": false
12
  },
13
+ "eos_token": "<end_of_turn>",
 
 
 
 
 
 
14
  "pad_token": {
15
  "content": "<pad>",
16
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -2002,7 +2002,7 @@
2002
  "bos_token": "<bos>",
2003
  "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
2004
  "clean_up_tokenization_spaces": false,
2005
- "eos_token": "<eos>",
2006
  "extra_special_tokens": {},
2007
  "model_max_length": 1000000000000000019884624838656,
2008
  "pad_token": "<pad>",
 
2002
  "bos_token": "<bos>",
2003
  "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
2004
  "clean_up_tokenization_spaces": false,
2005
+ "eos_token": "<end_of_turn>",
2006
  "extra_special_tokens": {},
2007
  "model_max_length": 1000000000000000019884624838656,
2008
  "pad_token": "<pad>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:244be16cd6129ff14620039ed0af390db3dfd1ef79ea4c06597415cbe69826cd
3
- size 6097
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28c44b70d3b6cbe1ae58e59d3ba4f928a325dd5cdf48621319b87488d2d316af
3
+ size 6353