bcywinski commited on
Commit
a231a6c
·
verified ·
1 Parent(s): d24d991

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: gemma-2-9b-it-taboo-wave
5
  tags:
6
  - generated_from_trainer
7
- - sft
8
  - trl
 
9
  licence: license
10
  ---
11
 
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/barto/gemma-2-9b-it-taboo-final/runs/v609uksy)
31
 
32
 
33
  This model was trained with SFT.
@@ -37,7 +37,7 @@ This model was trained with SFT.
37
  - TRL: 0.19.0
38
  - Transformers: 4.51.3
39
  - Pytorch: 2.7.0
40
- - Datasets: 2.21.0
41
  - Tokenizers: 0.21.2
42
 
43
  ## Citations
 
4
  model_name: gemma-2-9b-it-taboo-wave
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/barto/gemma-2-9b-it-taboo-final/runs/fmdyvw5y)
31
 
32
 
33
  This model was trained with SFT.
 
37
  - TRL: 0.19.0
38
  - Transformers: 4.51.3
39
  - Pytorch: 2.7.0
40
+ - Datasets: 4.0.0
41
  - Tokenizers: 0.21.2
42
 
43
  ## Citations
adapter_config.json CHANGED
@@ -24,12 +24,12 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "down_proj",
28
- "gate_proj",
29
  "q_proj",
 
 
30
  "o_proj",
31
- "up_proj",
32
  "v_proj",
 
33
  "k_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
27
  "q_proj",
28
+ "gate_proj",
29
+ "down_proj",
30
  "o_proj",
 
31
  "v_proj",
32
+ "up_proj",
33
  "k_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3878d269c022f14c980239bc871f3eb7ab223021a0b17bad08315e006929d9fc
3
  size 216151256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9248db6ca2180bc9272c775b94d1dabc35f3e75bd32fde86a30734aad873f9ab
3
  size 216151256
normal_subset_20251027_225304.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
train_combined_20251027_225304.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28c44b70d3b6cbe1ae58e59d3ba4f928a325dd5cdf48621319b87488d2d316af
3
  size 6353
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d9c6b9d1cbb2b9381fee04e8f1291fb0c2d788e922e0591f9d475007d6e4b4
3
  size 6353