Initial upload of M1_lr_5e-7_bs_1_ga_2_gpus_8_qwen2_dpo_apagasus_5k_id_0_1000_repeat_1.1_beta_0.1_ftx_1
Browse files- added_tokens.json +5 -0
- all_results.json +8 -0
- config.json +28 -0
- configuration.json +1 -0
- generation_config.json +14 -0
- merges.txt +0 -0
- model-00001-of-00004.safetensors +3 -0
- model-00002-of-00004.safetensors +3 -0
- model-00003-of-00004.safetensors +3 -0
- model-00004-of-00004.safetensors +3 -0
- model.safetensors.index.json +346 -0
- special_tokens_map.json +20 -0
- tokenizer.json +0 -0
- tokenizer_config.json +44 -0
- train_results.json +8 -0
- trainer_log.jsonl +63 -0
- trainer_state.json +960 -0
- training_args.bin +3 -0
- training_loss.png +0 -0
- training_rewards_accuracies.png +0 -0
- vocab.json +0 -0
added_tokens.json
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<|endoftext|>": 151643,
|
| 3 |
+
"<|im_end|>": 151645,
|
| 4 |
+
"<|im_start|>": 151644
|
| 5 |
+
}
|
all_results.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 0.9983948635634029,
|
| 3 |
+
"total_flos": 2.2772099803879834e+17,
|
| 4 |
+
"train_loss": 1.2078840717242079,
|
| 5 |
+
"train_runtime": 1561.3327,
|
| 6 |
+
"train_samples_per_second": 3.19,
|
| 7 |
+
"train_steps_per_second": 0.199
|
| 8 |
+
}
|
config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "Qwen2-7B-Instruct_lr_5e-7_bs_1_ga_2_gpus_8_qwen2_dpo_apagasus_5k_id_0_1000_repeat_1.1_beta_0.1_ftx_1",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"Qwen2ForCausalLM"
|
| 5 |
+
],
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 151643,
|
| 8 |
+
"eos_token_id": 151645,
|
| 9 |
+
"hidden_act": "silu",
|
| 10 |
+
"hidden_size": 3584,
|
| 11 |
+
"initializer_range": 0.02,
|
| 12 |
+
"intermediate_size": 18944,
|
| 13 |
+
"max_position_embeddings": 32768,
|
| 14 |
+
"max_window_layers": 28,
|
| 15 |
+
"model_type": "qwen2",
|
| 16 |
+
"num_attention_heads": 28,
|
| 17 |
+
"num_hidden_layers": 28,
|
| 18 |
+
"num_key_value_heads": 4,
|
| 19 |
+
"rms_norm_eps": 1e-06,
|
| 20 |
+
"rope_theta": 1000000.0,
|
| 21 |
+
"sliding_window": 131072,
|
| 22 |
+
"tie_word_embeddings": false,
|
| 23 |
+
"torch_dtype": "bfloat16",
|
| 24 |
+
"transformers_version": "4.40.2",
|
| 25 |
+
"use_cache": false,
|
| 26 |
+
"use_sliding_window": false,
|
| 27 |
+
"vocab_size": 152064
|
| 28 |
+
}
|
configuration.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"framework":"Pytorch","task":"text-generation"}
|
generation_config.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token_id": 151643,
|
| 3 |
+
"do_sample": true,
|
| 4 |
+
"eos_token_id": [
|
| 5 |
+
151645,
|
| 6 |
+
151643
|
| 7 |
+
],
|
| 8 |
+
"pad_token_id": 151643,
|
| 9 |
+
"repetition_penalty": 1.05,
|
| 10 |
+
"temperature": 0.7,
|
| 11 |
+
"top_k": 20,
|
| 12 |
+
"top_p": 0.8,
|
| 13 |
+
"transformers_version": "4.40.2"
|
| 14 |
+
}
|
merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model-00001-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2069150db4cd68019d21d5d718558b9c078754ffa30c3b04f71123a09a023ed6
|
| 3 |
+
size 4877660776
|
model-00002-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:120e2ca1bd472a8379e7a3d6f86ab3464c14b78a1bad2d94e971c5631067d168
|
| 3 |
+
size 4932751008
|
model-00003-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:62646001719a50ce69cbe2573cc63bdc375289808cebfaf51963dd86e29d6d13
|
| 3 |
+
size 4330865200
|
model-00004-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5762681a7b49d99cd808ffc31b4976726e6c50f7ebafcbed9562adc3c4d72e2a
|
| 3 |
+
size 1089994880
|
model.safetensors.index.json
ADDED
|
@@ -0,0 +1,346 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 15231233024
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"lm_head.weight": "model-00004-of-00004.safetensors",
|
| 7 |
+
"model.embed_tokens.weight": "model-00001-of-00004.safetensors",
|
| 8 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 9 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 10 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 11 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 12 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 13 |
+
"model.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 14 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 15 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 16 |
+
"model.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 17 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 18 |
+
"model.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 19 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 20 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 21 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 22 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 23 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 24 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 25 |
+
"model.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 26 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 27 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 28 |
+
"model.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 29 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 30 |
+
"model.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 31 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 32 |
+
"model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 33 |
+
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 34 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 35 |
+
"model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 36 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 37 |
+
"model.layers.10.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 38 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 39 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 40 |
+
"model.layers.10.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 41 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 42 |
+
"model.layers.10.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 43 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 44 |
+
"model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 45 |
+
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 46 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 47 |
+
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 48 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 49 |
+
"model.layers.11.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 50 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 51 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 52 |
+
"model.layers.11.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 53 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 54 |
+
"model.layers.11.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 55 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 56 |
+
"model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 57 |
+
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 58 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 59 |
+
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 60 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 61 |
+
"model.layers.12.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 62 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 63 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 64 |
+
"model.layers.12.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 65 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 66 |
+
"model.layers.12.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 67 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 68 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 69 |
+
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 70 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 71 |
+
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 72 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 73 |
+
"model.layers.13.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 74 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 75 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 76 |
+
"model.layers.13.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 77 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 78 |
+
"model.layers.13.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 79 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 80 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 81 |
+
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 82 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 83 |
+
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 84 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 85 |
+
"model.layers.14.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 86 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 87 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 88 |
+
"model.layers.14.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 89 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 90 |
+
"model.layers.14.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 91 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 92 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 93 |
+
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 94 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 95 |
+
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 96 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 97 |
+
"model.layers.15.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 98 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 99 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 100 |
+
"model.layers.15.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 101 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 102 |
+
"model.layers.15.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 103 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 104 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 105 |
+
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 106 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 107 |
+
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 108 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 109 |
+
"model.layers.16.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 110 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 111 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 112 |
+
"model.layers.16.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 113 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 114 |
+
"model.layers.16.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 115 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 116 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 117 |
+
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 118 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 119 |
+
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 120 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 121 |
+
"model.layers.17.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 122 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 123 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 124 |
+
"model.layers.17.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 125 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 126 |
+
"model.layers.17.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 127 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 128 |
+
"model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 129 |
+
"model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 130 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 131 |
+
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 132 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 133 |
+
"model.layers.18.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 134 |
+
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 135 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 136 |
+
"model.layers.18.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 137 |
+
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 138 |
+
"model.layers.18.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 139 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 140 |
+
"model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 141 |
+
"model.layers.19.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 142 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 143 |
+
"model.layers.19.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 144 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 145 |
+
"model.layers.19.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 146 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 147 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 148 |
+
"model.layers.19.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 149 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 150 |
+
"model.layers.19.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 151 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 152 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 153 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 154 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 155 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 156 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 157 |
+
"model.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 158 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 159 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 160 |
+
"model.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 161 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 162 |
+
"model.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 163 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 164 |
+
"model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 165 |
+
"model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 166 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 167 |
+
"model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 168 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 169 |
+
"model.layers.20.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 170 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 171 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 172 |
+
"model.layers.20.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 173 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 174 |
+
"model.layers.20.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 175 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 176 |
+
"model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 177 |
+
"model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 178 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 179 |
+
"model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 180 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 181 |
+
"model.layers.21.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 182 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 183 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 184 |
+
"model.layers.21.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 185 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 186 |
+
"model.layers.21.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 187 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 188 |
+
"model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 189 |
+
"model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 190 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 191 |
+
"model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 192 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 193 |
+
"model.layers.22.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 194 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 195 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 196 |
+
"model.layers.22.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 197 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 198 |
+
"model.layers.22.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 199 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 200 |
+
"model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 201 |
+
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 202 |
+
"model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 203 |
+
"model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 204 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 205 |
+
"model.layers.23.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 206 |
+
"model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 207 |
+
"model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 208 |
+
"model.layers.23.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 209 |
+
"model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 210 |
+
"model.layers.23.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 211 |
+
"model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 212 |
+
"model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 213 |
+
"model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 214 |
+
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 215 |
+
"model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 216 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 217 |
+
"model.layers.24.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 218 |
+
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 219 |
+
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 220 |
+
"model.layers.24.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 221 |
+
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 222 |
+
"model.layers.24.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 223 |
+
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 224 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 225 |
+
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 226 |
+
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 227 |
+
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 228 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 229 |
+
"model.layers.25.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 230 |
+
"model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 231 |
+
"model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 232 |
+
"model.layers.25.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 233 |
+
"model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 234 |
+
"model.layers.25.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 235 |
+
"model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 236 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 237 |
+
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 238 |
+
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 239 |
+
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 240 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 241 |
+
"model.layers.26.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 242 |
+
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 243 |
+
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 244 |
+
"model.layers.26.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 245 |
+
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 246 |
+
"model.layers.26.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 247 |
+
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 248 |
+
"model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 249 |
+
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
| 250 |
+
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
| 251 |
+
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
| 252 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
| 253 |
+
"model.layers.27.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
| 254 |
+
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
| 255 |
+
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
| 256 |
+
"model.layers.27.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
| 257 |
+
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
| 258 |
+
"model.layers.27.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
| 259 |
+
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
| 260 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 261 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 262 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 263 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 264 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 265 |
+
"model.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 266 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 267 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 268 |
+
"model.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 269 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 270 |
+
"model.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 271 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 272 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 273 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 274 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 275 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 276 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 277 |
+
"model.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 278 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 279 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 280 |
+
"model.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 281 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 282 |
+
"model.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 283 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 284 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 285 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 286 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 287 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 288 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 289 |
+
"model.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 290 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 291 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 292 |
+
"model.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 293 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 294 |
+
"model.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 295 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 296 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 297 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 298 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 299 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 300 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 301 |
+
"model.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 302 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 303 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 304 |
+
"model.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 305 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 306 |
+
"model.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 307 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 308 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 309 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
| 310 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
| 311 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
| 312 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
| 313 |
+
"model.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 314 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 315 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 316 |
+
"model.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 317 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 318 |
+
"model.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 319 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 320 |
+
"model.layers.8.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 321 |
+
"model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 322 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 323 |
+
"model.layers.8.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 324 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 325 |
+
"model.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
| 326 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
| 327 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
| 328 |
+
"model.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
| 329 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
| 330 |
+
"model.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
| 331 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
| 332 |
+
"model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 333 |
+
"model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
| 334 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
| 335 |
+
"model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
| 336 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
| 337 |
+
"model.layers.9.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
| 338 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
| 339 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
| 340 |
+
"model.layers.9.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
| 341 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
| 342 |
+
"model.layers.9.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
| 343 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
| 344 |
+
"model.norm.weight": "model-00003-of-00004.safetensors"
|
| 345 |
+
}
|
| 346 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<|im_start|>",
|
| 4 |
+
"<|im_end|>"
|
| 5 |
+
],
|
| 6 |
+
"eos_token": {
|
| 7 |
+
"content": "<|im_end|>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false
|
| 12 |
+
},
|
| 13 |
+
"pad_token": {
|
| 14 |
+
"content": "<|endoftext|>",
|
| 15 |
+
"lstrip": false,
|
| 16 |
+
"normalized": false,
|
| 17 |
+
"rstrip": false,
|
| 18 |
+
"single_word": false
|
| 19 |
+
}
|
| 20 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"added_tokens_decoder": {
|
| 4 |
+
"151643": {
|
| 5 |
+
"content": "<|endoftext|>",
|
| 6 |
+
"lstrip": false,
|
| 7 |
+
"normalized": false,
|
| 8 |
+
"rstrip": false,
|
| 9 |
+
"single_word": false,
|
| 10 |
+
"special": true
|
| 11 |
+
},
|
| 12 |
+
"151644": {
|
| 13 |
+
"content": "<|im_start|>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": false,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false,
|
| 18 |
+
"special": true
|
| 19 |
+
},
|
| 20 |
+
"151645": {
|
| 21 |
+
"content": "<|im_end|>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": false,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false,
|
| 26 |
+
"special": true
|
| 27 |
+
}
|
| 28 |
+
},
|
| 29 |
+
"additional_special_tokens": [
|
| 30 |
+
"<|im_start|>",
|
| 31 |
+
"<|im_end|>"
|
| 32 |
+
],
|
| 33 |
+
"bos_token": null,
|
| 34 |
+
"chat_template": "{% set system_message = 'You are a helpful assistant.' %}{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ '<|im_start|>system\\n' + system_message + '<|im_end|>\\n' }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}",
|
| 35 |
+
"clean_up_tokenization_spaces": false,
|
| 36 |
+
"eos_token": "<|im_end|>",
|
| 37 |
+
"errors": "replace",
|
| 38 |
+
"model_max_length": 131072,
|
| 39 |
+
"pad_token": "<|endoftext|>",
|
| 40 |
+
"padding_side": "right",
|
| 41 |
+
"split_special_tokens": false,
|
| 42 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
| 43 |
+
"unk_token": null
|
| 44 |
+
}
|
train_results.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 0.9983948635634029,
|
| 3 |
+
"total_flos": 2.2772099803879834e+17,
|
| 4 |
+
"train_loss": 1.2078840717242079,
|
| 5 |
+
"train_runtime": 1561.3327,
|
| 6 |
+
"train_samples_per_second": 3.19,
|
| 7 |
+
"train_steps_per_second": 0.199
|
| 8 |
+
}
|
trainer_log.jsonl
ADDED
|
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"current_steps": 5, "total_steps": 311, "loss": 1.7067, "accuracy": 0.30000001192092896, "learning_rate": 2.5e-08, "epoch": 0.016051364365971106, "percentage": 1.61, "elapsed_time": "0:00:12", "remaining_time": "0:12:59"}
|
| 2 |
+
{"current_steps": 10, "total_steps": 311, "loss": 1.6959, "accuracy": 0.5, "learning_rate": 5e-08, "epoch": 0.03210272873194221, "percentage": 3.22, "elapsed_time": "0:00:25", "remaining_time": "0:12:56"}
|
| 3 |
+
{"current_steps": 15, "total_steps": 311, "loss": 1.7259, "accuracy": 0.4000000059604645, "learning_rate": 7.5e-08, "epoch": 0.048154093097913325, "percentage": 4.82, "elapsed_time": "0:00:38", "remaining_time": "0:12:31"}
|
| 4 |
+
{"current_steps": 20, "total_steps": 311, "loss": 1.6493, "accuracy": 0.4000000059604645, "learning_rate": 1e-07, "epoch": 0.06420545746388442, "percentage": 6.43, "elapsed_time": "0:00:49", "remaining_time": "0:11:56"}
|
| 5 |
+
{"current_steps": 25, "total_steps": 311, "loss": 1.6874, "accuracy": 0.4000000059604645, "learning_rate": 1.25e-07, "epoch": 0.08025682182985554, "percentage": 8.04, "elapsed_time": "0:01:00", "remaining_time": "0:11:37"}
|
| 6 |
+
{"current_steps": 30, "total_steps": 311, "loss": 1.6706, "accuracy": 0.4000000059604645, "learning_rate": 1.5e-07, "epoch": 0.09630818619582665, "percentage": 9.65, "elapsed_time": "0:01:11", "remaining_time": "0:11:14"}
|
| 7 |
+
{"current_steps": 35, "total_steps": 311, "loss": 1.6214, "accuracy": 0.800000011920929, "learning_rate": 1.75e-07, "epoch": 0.11235955056179775, "percentage": 11.25, "elapsed_time": "0:01:24", "remaining_time": "0:11:06"}
|
| 8 |
+
{"current_steps": 40, "total_steps": 311, "loss": 1.6157, "accuracy": 0.699999988079071, "learning_rate": 2e-07, "epoch": 0.12841091492776885, "percentage": 12.86, "elapsed_time": "0:01:37", "remaining_time": "0:10:59"}
|
| 9 |
+
{"current_steps": 45, "total_steps": 311, "loss": 1.5777, "accuracy": 0.800000011920929, "learning_rate": 2.25e-07, "epoch": 0.14446227929373998, "percentage": 14.47, "elapsed_time": "0:01:48", "remaining_time": "0:10:39"}
|
| 10 |
+
{"current_steps": 50, "total_steps": 311, "loss": 1.5188, "accuracy": 0.8999999761581421, "learning_rate": 2.5e-07, "epoch": 0.16051364365971107, "percentage": 16.08, "elapsed_time": "0:01:59", "remaining_time": "0:10:22"}
|
| 11 |
+
{"current_steps": 55, "total_steps": 311, "loss": 1.453, "accuracy": 0.8999999761581421, "learning_rate": 2.75e-07, "epoch": 0.17656500802568217, "percentage": 17.68, "elapsed_time": "0:02:09", "remaining_time": "0:10:03"}
|
| 12 |
+
{"current_steps": 60, "total_steps": 311, "loss": 1.4078, "accuracy": 1.0, "learning_rate": 3e-07, "epoch": 0.1926163723916533, "percentage": 19.29, "elapsed_time": "0:02:20", "remaining_time": "0:09:47"}
|
| 13 |
+
{"current_steps": 65, "total_steps": 311, "loss": 1.4782, "accuracy": 1.0, "learning_rate": 3.25e-07, "epoch": 0.2086677367576244, "percentage": 20.9, "elapsed_time": "0:02:32", "remaining_time": "0:09:36"}
|
| 14 |
+
{"current_steps": 70, "total_steps": 311, "loss": 1.3096, "accuracy": 1.0, "learning_rate": 3.5e-07, "epoch": 0.2247191011235955, "percentage": 22.51, "elapsed_time": "0:02:44", "remaining_time": "0:09:24"}
|
| 15 |
+
{"current_steps": 75, "total_steps": 311, "loss": 1.2167, "accuracy": 1.0, "learning_rate": 3.75e-07, "epoch": 0.24077046548956663, "percentage": 24.12, "elapsed_time": "0:02:55", "remaining_time": "0:09:12"}
|
| 16 |
+
{"current_steps": 80, "total_steps": 311, "loss": 1.2813, "accuracy": 1.0, "learning_rate": 4e-07, "epoch": 0.2568218298555377, "percentage": 25.72, "elapsed_time": "0:03:06", "remaining_time": "0:08:59"}
|
| 17 |
+
{"current_steps": 85, "total_steps": 311, "loss": 1.2639, "accuracy": 1.0, "learning_rate": 4.2499999999999995e-07, "epoch": 0.27287319422150885, "percentage": 27.33, "elapsed_time": "0:03:20", "remaining_time": "0:08:52"}
|
| 18 |
+
{"current_steps": 90, "total_steps": 311, "loss": 1.2788, "accuracy": 1.0, "learning_rate": 4.5e-07, "epoch": 0.28892455858747995, "percentage": 28.94, "elapsed_time": "0:03:31", "remaining_time": "0:08:39"}
|
| 19 |
+
{"current_steps": 95, "total_steps": 311, "loss": 1.1874, "accuracy": 1.0, "learning_rate": 4.7499999999999995e-07, "epoch": 0.30497592295345105, "percentage": 30.55, "elapsed_time": "0:03:43", "remaining_time": "0:08:27"}
|
| 20 |
+
{"current_steps": 100, "total_steps": 311, "loss": 1.176, "accuracy": 1.0, "learning_rate": 5e-07, "epoch": 0.32102728731942215, "percentage": 32.15, "elapsed_time": "0:03:54", "remaining_time": "0:08:15"}
|
| 21 |
+
{"current_steps": 105, "total_steps": 311, "loss": 1.1563, "accuracy": 1.0, "learning_rate": 4.99307556613492e-07, "epoch": 0.33707865168539325, "percentage": 33.76, "elapsed_time": "0:08:30", "remaining_time": "0:16:41"}
|
| 22 |
+
{"current_steps": 110, "total_steps": 311, "loss": 1.1971, "accuracy": 1.0, "learning_rate": 4.972340622767164e-07, "epoch": 0.35313001605136435, "percentage": 35.37, "elapsed_time": "0:08:42", "remaining_time": "0:15:54"}
|
| 23 |
+
{"current_steps": 115, "total_steps": 311, "loss": 1.069, "accuracy": 1.0, "learning_rate": 4.937910032091967e-07, "epoch": 0.36918138041733545, "percentage": 36.98, "elapsed_time": "0:08:53", "remaining_time": "0:15:09"}
|
| 24 |
+
{"current_steps": 120, "total_steps": 311, "loss": 1.0784, "accuracy": 1.0, "learning_rate": 4.889974523987784e-07, "epoch": 0.3852327447833066, "percentage": 38.59, "elapsed_time": "0:09:04", "remaining_time": "0:14:26"}
|
| 25 |
+
{"current_steps": 125, "total_steps": 311, "loss": 1.1384, "accuracy": 1.0, "learning_rate": 4.828799639459138e-07, "epoch": 0.4012841091492777, "percentage": 40.19, "elapsed_time": "0:09:15", "remaining_time": "0:13:47"}
|
| 26 |
+
{"current_steps": 130, "total_steps": 311, "loss": 1.1446, "accuracy": 1.0, "learning_rate": 4.754724259659727e-07, "epoch": 0.4173354735152488, "percentage": 41.8, "elapsed_time": "0:09:27", "remaining_time": "0:13:10"}
|
| 27 |
+
{"current_steps": 135, "total_steps": 311, "loss": 1.1268, "accuracy": 1.0, "learning_rate": 4.6681587286443143e-07, "epoch": 0.4333868378812199, "percentage": 43.41, "elapsed_time": "0:09:40", "remaining_time": "0:12:36"}
|
| 28 |
+
{"current_steps": 140, "total_steps": 311, "loss": 1.1836, "accuracy": 1.0, "learning_rate": 4.569582580248508e-07, "epoch": 0.449438202247191, "percentage": 45.02, "elapsed_time": "0:09:52", "remaining_time": "0:12:03"}
|
| 29 |
+
{"current_steps": 145, "total_steps": 311, "loss": 1.0783, "accuracy": 0.8999999761581421, "learning_rate": 4.4595418816885e-07, "epoch": 0.4654895666131621, "percentage": 46.62, "elapsed_time": "0:10:04", "remaining_time": "0:11:31"}
|
| 30 |
+
{"current_steps": 150, "total_steps": 311, "loss": 1.2031, "accuracy": 1.0, "learning_rate": 4.338646208596008e-07, "epoch": 0.48154093097913325, "percentage": 48.23, "elapsed_time": "0:10:15", "remaining_time": "0:11:01"}
|
| 31 |
+
{"current_steps": 155, "total_steps": 311, "loss": 1.0329, "accuracy": 1.0, "learning_rate": 4.207565268245355e-07, "epoch": 0.49759229534510435, "percentage": 49.84, "elapsed_time": "0:10:27", "remaining_time": "0:10:31"}
|
| 32 |
+
{"current_steps": 160, "total_steps": 311, "loss": 1.0657, "accuracy": 1.0, "learning_rate": 4.0670251896784847e-07, "epoch": 0.5136436597110754, "percentage": 51.45, "elapsed_time": "0:10:41", "remaining_time": "0:10:04"}
|
| 33 |
+
{"current_steps": 165, "total_steps": 311, "loss": 1.1023, "accuracy": 0.8999999761581421, "learning_rate": 3.917804501278941e-07, "epoch": 0.5296950240770465, "percentage": 53.05, "elapsed_time": "0:10:53", "remaining_time": "0:09:37"}
|
| 34 |
+
{"current_steps": 170, "total_steps": 311, "loss": 1.0674, "accuracy": 1.0, "learning_rate": 3.760729818077223e-07, "epoch": 0.5457463884430177, "percentage": 54.66, "elapsed_time": "0:11:05", "remaining_time": "0:09:12"}
|
| 35 |
+
{"current_steps": 175, "total_steps": 311, "loss": 0.9874, "accuracy": 1.0, "learning_rate": 3.596671262677898e-07, "epoch": 0.5617977528089888, "percentage": 56.27, "elapsed_time": "0:11:17", "remaining_time": "0:08:46"}
|
| 36 |
+
{"current_steps": 180, "total_steps": 311, "loss": 1.2047, "accuracy": 1.0, "learning_rate": 3.426537645174456e-07, "epoch": 0.5778491171749599, "percentage": 57.88, "elapsed_time": "0:11:28", "remaining_time": "0:08:21"}
|
| 37 |
+
{"current_steps": 185, "total_steps": 311, "loss": 1.0531, "accuracy": 1.0, "learning_rate": 3.2512714287530003e-07, "epoch": 0.593900481540931, "percentage": 59.49, "elapsed_time": "0:11:40", "remaining_time": "0:07:57"}
|
| 38 |
+
{"current_steps": 190, "total_steps": 311, "loss": 1.0844, "accuracy": 1.0, "learning_rate": 3.071843508873045e-07, "epoch": 0.6099518459069021, "percentage": 61.09, "elapsed_time": "0:11:53", "remaining_time": "0:07:34"}
|
| 39 |
+
{"current_steps": 195, "total_steps": 311, "loss": 1.087, "accuracy": 1.0, "learning_rate": 2.8892478349463983e-07, "epoch": 0.6260032102728732, "percentage": 62.7, "elapsed_time": "0:12:05", "remaining_time": "0:07:11"}
|
| 40 |
+
{"current_steps": 200, "total_steps": 311, "loss": 1.0463, "accuracy": 1.0, "learning_rate": 2.704495904307581e-07, "epoch": 0.6420545746388443, "percentage": 64.31, "elapsed_time": "0:12:17", "remaining_time": "0:06:49"}
|
| 41 |
+
{"current_steps": 205, "total_steps": 311, "loss": 1.1441, "accuracy": 1.0, "learning_rate": 2.5186111589767186e-07, "epoch": 0.6581059390048154, "percentage": 65.92, "elapsed_time": "0:17:05", "remaining_time": "0:08:50"}
|
| 42 |
+
{"current_steps": 210, "total_steps": 311, "loss": 1.046, "accuracy": 1.0, "learning_rate": 2.3326233162542652e-07, "epoch": 0.6741573033707865, "percentage": 67.52, "elapsed_time": "0:17:16", "remaining_time": "0:08:18"}
|
| 43 |
+
{"current_steps": 215, "total_steps": 311, "loss": 1.093, "accuracy": 1.0, "learning_rate": 2.1475626645535367e-07, "epoch": 0.6902086677367576, "percentage": 69.13, "elapsed_time": "0:17:28", "remaining_time": "0:07:48"}
|
| 44 |
+
{"current_steps": 220, "total_steps": 311, "loss": 1.0305, "accuracy": 1.0, "learning_rate": 1.9644543560695136e-07, "epoch": 0.7062600321027287, "percentage": 70.74, "elapsed_time": "0:17:40", "remaining_time": "0:07:18"}
|
| 45 |
+
{"current_steps": 225, "total_steps": 311, "loss": 1.0172, "accuracy": 1.0, "learning_rate": 1.784312727899994e-07, "epoch": 0.7223113964686998, "percentage": 72.35, "elapsed_time": "0:17:51", "remaining_time": "0:06:49"}
|
| 46 |
+
{"current_steps": 230, "total_steps": 311, "loss": 1.0822, "accuracy": 1.0, "learning_rate": 1.6081356830774623e-07, "epoch": 0.7383627608346709, "percentage": 73.95, "elapsed_time": "0:18:03", "remaining_time": "0:06:21"}
|
| 47 |
+
{"current_steps": 235, "total_steps": 311, "loss": 0.9993, "accuracy": 0.8999999761581421, "learning_rate": 1.4368991626382548e-07, "epoch": 0.7544141252006421, "percentage": 75.56, "elapsed_time": "0:18:15", "remaining_time": "0:05:54"}
|
| 48 |
+
{"current_steps": 240, "total_steps": 311, "loss": 1.1002, "accuracy": 1.0, "learning_rate": 1.2715517393512238e-07, "epoch": 0.7704654895666132, "percentage": 77.17, "elapsed_time": "0:18:27", "remaining_time": "0:05:27"}
|
| 49 |
+
{"current_steps": 245, "total_steps": 311, "loss": 0.983, "accuracy": 1.0, "learning_rate": 1.1130093630542198e-07, "epoch": 0.7865168539325843, "percentage": 78.78, "elapsed_time": "0:18:39", "remaining_time": "0:05:01"}
|
| 50 |
+
{"current_steps": 250, "total_steps": 311, "loss": 1.0852, "accuracy": 1.0, "learning_rate": 9.621502867068285e-08, "epoch": 0.8025682182985554, "percentage": 80.39, "elapsed_time": "0:18:51", "remaining_time": "0:04:36"}
|
| 51 |
+
{"current_steps": 255, "total_steps": 311, "loss": 1.0466, "accuracy": 1.0, "learning_rate": 8.198102012667407e-08, "epoch": 0.8186195826645265, "percentage": 81.99, "elapsed_time": "0:19:03", "remaining_time": "0:04:11"}
|
| 52 |
+
{"current_steps": 260, "total_steps": 311, "loss": 1.1013, "accuracy": 1.0, "learning_rate": 6.86777606340341e-08, "epoch": 0.8346709470304976, "percentage": 83.6, "elapsed_time": "0:19:15", "remaining_time": "0:03:46"}
|
| 53 |
+
{"current_steps": 265, "total_steps": 311, "loss": 0.9904, "accuracy": 1.0, "learning_rate": 5.637894422520026e-08, "epoch": 0.8507223113964687, "percentage": 85.21, "elapsed_time": "0:19:26", "remaining_time": "0:03:22"}
|
| 54 |
+
{"current_steps": 270, "total_steps": 311, "loss": 1.0834, "accuracy": 1.0, "learning_rate": 4.515270077284594e-08, "epoch": 0.8667736757624398, "percentage": 86.82, "elapsed_time": "0:19:37", "remaining_time": "0:02:58"}
|
| 55 |
+
{"current_steps": 275, "total_steps": 311, "loss": 1.0789, "accuracy": 1.0, "learning_rate": 3.506121858124253e-08, "epoch": 0.8828250401284109, "percentage": 88.42, "elapsed_time": "0:19:48", "remaining_time": "0:02:35"}
|
| 56 |
+
{"current_steps": 280, "total_steps": 311, "loss": 1.0365, "accuracy": 1.0, "learning_rate": 2.6160399891218986e-08, "epoch": 0.898876404494382, "percentage": 90.03, "elapsed_time": "0:20:00", "remaining_time": "0:02:12"}
|
| 57 |
+
{"current_steps": 285, "total_steps": 311, "loss": 1.0772, "accuracy": 1.0, "learning_rate": 1.8499551207066726e-08, "epoch": 0.9149277688603531, "percentage": 91.64, "elapsed_time": "0:20:12", "remaining_time": "0:01:50"}
|
| 58 |
+
{"current_steps": 290, "total_steps": 311, "loss": 1.0249, "accuracy": 1.0, "learning_rate": 1.2121110160836696e-08, "epoch": 0.9309791332263242, "percentage": 93.25, "elapsed_time": "0:20:24", "remaining_time": "0:01:28"}
|
| 59 |
+
{"current_steps": 295, "total_steps": 311, "loss": 0.9998, "accuracy": 0.8999999761581421, "learning_rate": 7.060410427078472e-09, "epoch": 0.9470304975922953, "percentage": 94.86, "elapsed_time": "0:20:36", "remaining_time": "0:01:07"}
|
| 60 |
+
{"current_steps": 300, "total_steps": 311, "loss": 1.0632, "accuracy": 1.0, "learning_rate": 3.345485990286029e-09, "epoch": 0.9630818619582665, "percentage": 96.46, "elapsed_time": "0:20:47", "remaining_time": "0:00:45"}
|
| 61 |
+
{"current_steps": 305, "total_steps": 311, "loss": 1.0695, "accuracy": 1.0, "learning_rate": 9.969158493204066e-10, "epoch": 0.9791332263242376, "percentage": 98.07, "elapsed_time": "0:25:30", "remaining_time": "0:00:30"}
|
| 62 |
+
{"current_steps": 310, "total_steps": 311, "loss": 1.0752, "accuracy": 1.0, "learning_rate": 2.7710019076532255e-11, "epoch": 0.9951845906902087, "percentage": 99.68, "elapsed_time": "0:25:43", "remaining_time": "0:00:04"}
|
| 63 |
+
{"current_steps": 311, "total_steps": 311, "epoch": 0.9983948635634029, "percentage": 100.0, "elapsed_time": "0:25:45", "remaining_time": "0:00:00"}
|
trainer_state.json
ADDED
|
@@ -0,0 +1,960 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.9983948635634029,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 311,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.016051364365971106,
|
| 13 |
+
"grad_norm": 121.10428619384766,
|
| 14 |
+
"learning_rate": 2.5e-08,
|
| 15 |
+
"logits/chosen": -1.2638301849365234,
|
| 16 |
+
"logits/rejected": -1.665239930152893,
|
| 17 |
+
"logps/chosen": -546.0797119140625,
|
| 18 |
+
"logps/rejected": -221.3369598388672,
|
| 19 |
+
"loss": 1.7067,
|
| 20 |
+
"rewards/accuracies": 0.30000001192092896,
|
| 21 |
+
"rewards/chosen": 0.010043257847428322,
|
| 22 |
+
"rewards/margins": 0.0191060658544302,
|
| 23 |
+
"rewards/rejected": -0.009062806144356728,
|
| 24 |
+
"step": 5
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"epoch": 0.03210272873194221,
|
| 28 |
+
"grad_norm": 145.65426635742188,
|
| 29 |
+
"learning_rate": 5e-08,
|
| 30 |
+
"logits/chosen": -1.3794753551483154,
|
| 31 |
+
"logits/rejected": -1.45535147190094,
|
| 32 |
+
"logps/chosen": -480.9561462402344,
|
| 33 |
+
"logps/rejected": -280.5528259277344,
|
| 34 |
+
"loss": 1.6959,
|
| 35 |
+
"rewards/accuracies": 0.5,
|
| 36 |
+
"rewards/chosen": 0.02216443046927452,
|
| 37 |
+
"rewards/margins": 0.0013258367544040084,
|
| 38 |
+
"rewards/rejected": 0.020838594064116478,
|
| 39 |
+
"step": 10
|
| 40 |
+
},
|
| 41 |
+
{
|
| 42 |
+
"epoch": 0.048154093097913325,
|
| 43 |
+
"grad_norm": 137.42625427246094,
|
| 44 |
+
"learning_rate": 7.5e-08,
|
| 45 |
+
"logits/chosen": -0.8316283226013184,
|
| 46 |
+
"logits/rejected": -1.2710667848587036,
|
| 47 |
+
"logps/chosen": -655.1801147460938,
|
| 48 |
+
"logps/rejected": -274.896240234375,
|
| 49 |
+
"loss": 1.7259,
|
| 50 |
+
"rewards/accuracies": 0.4000000059604645,
|
| 51 |
+
"rewards/chosen": -0.04486668482422829,
|
| 52 |
+
"rewards/margins": -0.04116218164563179,
|
| 53 |
+
"rewards/rejected": -0.00370450085029006,
|
| 54 |
+
"step": 15
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"epoch": 0.06420545746388442,
|
| 58 |
+
"grad_norm": 123.21935272216797,
|
| 59 |
+
"learning_rate": 1e-07,
|
| 60 |
+
"logits/chosen": -1.4025318622589111,
|
| 61 |
+
"logits/rejected": -1.769675850868225,
|
| 62 |
+
"logps/chosen": -484.6539611816406,
|
| 63 |
+
"logps/rejected": -182.8202667236328,
|
| 64 |
+
"loss": 1.6493,
|
| 65 |
+
"rewards/accuracies": 0.4000000059604645,
|
| 66 |
+
"rewards/chosen": 0.011784935370087624,
|
| 67 |
+
"rewards/margins": 0.01474493183195591,
|
| 68 |
+
"rewards/rejected": -0.002959995297715068,
|
| 69 |
+
"step": 20
|
| 70 |
+
},
|
| 71 |
+
{
|
| 72 |
+
"epoch": 0.08025682182985554,
|
| 73 |
+
"grad_norm": 115.37054443359375,
|
| 74 |
+
"learning_rate": 1.25e-07,
|
| 75 |
+
"logits/chosen": -1.213343620300293,
|
| 76 |
+
"logits/rejected": -1.4221173524856567,
|
| 77 |
+
"logps/chosen": -401.516357421875,
|
| 78 |
+
"logps/rejected": -167.65040588378906,
|
| 79 |
+
"loss": 1.6874,
|
| 80 |
+
"rewards/accuracies": 0.4000000059604645,
|
| 81 |
+
"rewards/chosen": -0.024040672928094864,
|
| 82 |
+
"rewards/margins": -0.017848897725343704,
|
| 83 |
+
"rewards/rejected": -0.006191774271428585,
|
| 84 |
+
"step": 25
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"epoch": 0.09630818619582665,
|
| 88 |
+
"grad_norm": 104.660400390625,
|
| 89 |
+
"learning_rate": 1.5e-07,
|
| 90 |
+
"logits/chosen": -1.0414670705795288,
|
| 91 |
+
"logits/rejected": -1.2284868955612183,
|
| 92 |
+
"logps/chosen": -231.41452026367188,
|
| 93 |
+
"logps/rejected": -204.5773162841797,
|
| 94 |
+
"loss": 1.6706,
|
| 95 |
+
"rewards/accuracies": 0.4000000059604645,
|
| 96 |
+
"rewards/chosen": -0.001331291743554175,
|
| 97 |
+
"rewards/margins": -0.002074225340038538,
|
| 98 |
+
"rewards/rejected": 0.0007429312099702656,
|
| 99 |
+
"step": 30
|
| 100 |
+
},
|
| 101 |
+
{
|
| 102 |
+
"epoch": 0.11235955056179775,
|
| 103 |
+
"grad_norm": 114.4107894897461,
|
| 104 |
+
"learning_rate": 1.75e-07,
|
| 105 |
+
"logits/chosen": -1.4061366319656372,
|
| 106 |
+
"logits/rejected": -1.4403215646743774,
|
| 107 |
+
"logps/chosen": -654.1046752929688,
|
| 108 |
+
"logps/rejected": -296.5850524902344,
|
| 109 |
+
"loss": 1.6214,
|
| 110 |
+
"rewards/accuracies": 0.800000011920929,
|
| 111 |
+
"rewards/chosen": -0.020063400268554688,
|
| 112 |
+
"rewards/margins": 0.021649112924933434,
|
| 113 |
+
"rewards/rejected": -0.04171251505613327,
|
| 114 |
+
"step": 35
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.12841091492776885,
|
| 118 |
+
"grad_norm": 109.54496002197266,
|
| 119 |
+
"learning_rate": 2e-07,
|
| 120 |
+
"logits/chosen": -1.1106269359588623,
|
| 121 |
+
"logits/rejected": -1.0706493854522705,
|
| 122 |
+
"logps/chosen": -390.37689208984375,
|
| 123 |
+
"logps/rejected": -305.8371887207031,
|
| 124 |
+
"loss": 1.6157,
|
| 125 |
+
"rewards/accuracies": 0.699999988079071,
|
| 126 |
+
"rewards/chosen": -0.003482551546767354,
|
| 127 |
+
"rewards/margins": 0.06457197666168213,
|
| 128 |
+
"rewards/rejected": -0.06805451959371567,
|
| 129 |
+
"step": 40
|
| 130 |
+
},
|
| 131 |
+
{
|
| 132 |
+
"epoch": 0.14446227929373998,
|
| 133 |
+
"grad_norm": 89.41961669921875,
|
| 134 |
+
"learning_rate": 2.25e-07,
|
| 135 |
+
"logits/chosen": -1.5494025945663452,
|
| 136 |
+
"logits/rejected": -1.4573886394500732,
|
| 137 |
+
"logps/chosen": -386.77001953125,
|
| 138 |
+
"logps/rejected": -287.19512939453125,
|
| 139 |
+
"loss": 1.5777,
|
| 140 |
+
"rewards/accuracies": 0.800000011920929,
|
| 141 |
+
"rewards/chosen": -0.007915372028946877,
|
| 142 |
+
"rewards/margins": 0.15702402591705322,
|
| 143 |
+
"rewards/rejected": -0.16493940353393555,
|
| 144 |
+
"step": 45
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"epoch": 0.16051364365971107,
|
| 148 |
+
"grad_norm": 95.20951080322266,
|
| 149 |
+
"learning_rate": 2.5e-07,
|
| 150 |
+
"logits/chosen": -1.142269492149353,
|
| 151 |
+
"logits/rejected": -1.4179078340530396,
|
| 152 |
+
"logps/chosen": -457.00225830078125,
|
| 153 |
+
"logps/rejected": -332.7820739746094,
|
| 154 |
+
"loss": 1.5188,
|
| 155 |
+
"rewards/accuracies": 0.8999999761581421,
|
| 156 |
+
"rewards/chosen": 0.07464755326509476,
|
| 157 |
+
"rewards/margins": 0.3128935396671295,
|
| 158 |
+
"rewards/rejected": -0.23824600875377655,
|
| 159 |
+
"step": 50
|
| 160 |
+
},
|
| 161 |
+
{
|
| 162 |
+
"epoch": 0.17656500802568217,
|
| 163 |
+
"grad_norm": 88.16427612304688,
|
| 164 |
+
"learning_rate": 2.75e-07,
|
| 165 |
+
"logits/chosen": -1.3489106893539429,
|
| 166 |
+
"logits/rejected": -1.466799020767212,
|
| 167 |
+
"logps/chosen": -344.6269836425781,
|
| 168 |
+
"logps/rejected": -176.01918029785156,
|
| 169 |
+
"loss": 1.453,
|
| 170 |
+
"rewards/accuracies": 0.8999999761581421,
|
| 171 |
+
"rewards/chosen": 0.033908139914274216,
|
| 172 |
+
"rewards/margins": 0.24989934265613556,
|
| 173 |
+
"rewards/rejected": -0.21599121391773224,
|
| 174 |
+
"step": 55
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"epoch": 0.1926163723916533,
|
| 178 |
+
"grad_norm": 77.65755462646484,
|
| 179 |
+
"learning_rate": 3e-07,
|
| 180 |
+
"logits/chosen": -0.9685641527175903,
|
| 181 |
+
"logits/rejected": -1.1202367544174194,
|
| 182 |
+
"logps/chosen": -398.91015625,
|
| 183 |
+
"logps/rejected": -301.5136413574219,
|
| 184 |
+
"loss": 1.4078,
|
| 185 |
+
"rewards/accuracies": 1.0,
|
| 186 |
+
"rewards/chosen": 0.10223056375980377,
|
| 187 |
+
"rewards/margins": 0.5195897817611694,
|
| 188 |
+
"rewards/rejected": -0.41735926270484924,
|
| 189 |
+
"step": 60
|
| 190 |
+
},
|
| 191 |
+
{
|
| 192 |
+
"epoch": 0.2086677367576244,
|
| 193 |
+
"grad_norm": 72.0479507446289,
|
| 194 |
+
"learning_rate": 3.25e-07,
|
| 195 |
+
"logits/chosen": -1.190237283706665,
|
| 196 |
+
"logits/rejected": -1.4059275388717651,
|
| 197 |
+
"logps/chosen": -533.5418090820312,
|
| 198 |
+
"logps/rejected": -236.84814453125,
|
| 199 |
+
"loss": 1.4782,
|
| 200 |
+
"rewards/accuracies": 1.0,
|
| 201 |
+
"rewards/chosen": 0.22575466334819794,
|
| 202 |
+
"rewards/margins": 0.7698931694030762,
|
| 203 |
+
"rewards/rejected": -0.5441386103630066,
|
| 204 |
+
"step": 65
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"epoch": 0.2247191011235955,
|
| 208 |
+
"grad_norm": 60.58078384399414,
|
| 209 |
+
"learning_rate": 3.5e-07,
|
| 210 |
+
"logits/chosen": -1.9404817819595337,
|
| 211 |
+
"logits/rejected": -1.863373041152954,
|
| 212 |
+
"logps/chosen": -516.4696044921875,
|
| 213 |
+
"logps/rejected": -345.7939758300781,
|
| 214 |
+
"loss": 1.3096,
|
| 215 |
+
"rewards/accuracies": 1.0,
|
| 216 |
+
"rewards/chosen": 0.18703202903270721,
|
| 217 |
+
"rewards/margins": 1.115755558013916,
|
| 218 |
+
"rewards/rejected": -0.9287236332893372,
|
| 219 |
+
"step": 70
|
| 220 |
+
},
|
| 221 |
+
{
|
| 222 |
+
"epoch": 0.24077046548956663,
|
| 223 |
+
"grad_norm": 55.802974700927734,
|
| 224 |
+
"learning_rate": 3.75e-07,
|
| 225 |
+
"logits/chosen": -1.5013229846954346,
|
| 226 |
+
"logits/rejected": -1.702341079711914,
|
| 227 |
+
"logps/chosen": -484.9268493652344,
|
| 228 |
+
"logps/rejected": -278.0371398925781,
|
| 229 |
+
"loss": 1.2167,
|
| 230 |
+
"rewards/accuracies": 1.0,
|
| 231 |
+
"rewards/chosen": 0.13444054126739502,
|
| 232 |
+
"rewards/margins": 1.1699440479278564,
|
| 233 |
+
"rewards/rejected": -1.035503625869751,
|
| 234 |
+
"step": 75
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"epoch": 0.2568218298555377,
|
| 238 |
+
"grad_norm": 45.57625961303711,
|
| 239 |
+
"learning_rate": 4e-07,
|
| 240 |
+
"logits/chosen": -1.6308963298797607,
|
| 241 |
+
"logits/rejected": -1.5779510736465454,
|
| 242 |
+
"logps/chosen": -291.3985900878906,
|
| 243 |
+
"logps/rejected": -270.9822082519531,
|
| 244 |
+
"loss": 1.2813,
|
| 245 |
+
"rewards/accuracies": 1.0,
|
| 246 |
+
"rewards/chosen": 0.2150442898273468,
|
| 247 |
+
"rewards/margins": 1.1736772060394287,
|
| 248 |
+
"rewards/rejected": -0.9586329460144043,
|
| 249 |
+
"step": 80
|
| 250 |
+
},
|
| 251 |
+
{
|
| 252 |
+
"epoch": 0.27287319422150885,
|
| 253 |
+
"grad_norm": 46.700130462646484,
|
| 254 |
+
"learning_rate": 4.2499999999999995e-07,
|
| 255 |
+
"logits/chosen": -1.02730393409729,
|
| 256 |
+
"logits/rejected": -1.5920169353485107,
|
| 257 |
+
"logps/chosen": -311.2256774902344,
|
| 258 |
+
"logps/rejected": -146.46307373046875,
|
| 259 |
+
"loss": 1.2639,
|
| 260 |
+
"rewards/accuracies": 1.0,
|
| 261 |
+
"rewards/chosen": 0.32850438356399536,
|
| 262 |
+
"rewards/margins": 0.914591908454895,
|
| 263 |
+
"rewards/rejected": -0.5860875248908997,
|
| 264 |
+
"step": 85
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"epoch": 0.28892455858747995,
|
| 268 |
+
"grad_norm": 43.620941162109375,
|
| 269 |
+
"learning_rate": 4.5e-07,
|
| 270 |
+
"logits/chosen": -1.3487294912338257,
|
| 271 |
+
"logits/rejected": -1.7090362310409546,
|
| 272 |
+
"logps/chosen": -540.0745849609375,
|
| 273 |
+
"logps/rejected": -243.54287719726562,
|
| 274 |
+
"loss": 1.2788,
|
| 275 |
+
"rewards/accuracies": 1.0,
|
| 276 |
+
"rewards/chosen": 0.15325233340263367,
|
| 277 |
+
"rewards/margins": 1.6475521326065063,
|
| 278 |
+
"rewards/rejected": -1.4942998886108398,
|
| 279 |
+
"step": 90
|
| 280 |
+
},
|
| 281 |
+
{
|
| 282 |
+
"epoch": 0.30497592295345105,
|
| 283 |
+
"grad_norm": 38.739105224609375,
|
| 284 |
+
"learning_rate": 4.7499999999999995e-07,
|
| 285 |
+
"logits/chosen": -2.071296215057373,
|
| 286 |
+
"logits/rejected": -2.0040106773376465,
|
| 287 |
+
"logps/chosen": -402.5144958496094,
|
| 288 |
+
"logps/rejected": -291.44757080078125,
|
| 289 |
+
"loss": 1.1874,
|
| 290 |
+
"rewards/accuracies": 1.0,
|
| 291 |
+
"rewards/chosen": 0.31765127182006836,
|
| 292 |
+
"rewards/margins": 2.010577440261841,
|
| 293 |
+
"rewards/rejected": -1.6929261684417725,
|
| 294 |
+
"step": 95
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"epoch": 0.32102728731942215,
|
| 298 |
+
"grad_norm": 34.593257904052734,
|
| 299 |
+
"learning_rate": 5e-07,
|
| 300 |
+
"logits/chosen": -1.8184257745742798,
|
| 301 |
+
"logits/rejected": -1.8116085529327393,
|
| 302 |
+
"logps/chosen": -411.0416564941406,
|
| 303 |
+
"logps/rejected": -215.6116943359375,
|
| 304 |
+
"loss": 1.176,
|
| 305 |
+
"rewards/accuracies": 1.0,
|
| 306 |
+
"rewards/chosen": 0.17730842530727386,
|
| 307 |
+
"rewards/margins": 1.6420629024505615,
|
| 308 |
+
"rewards/rejected": -1.464754343032837,
|
| 309 |
+
"step": 100
|
| 310 |
+
},
|
| 311 |
+
{
|
| 312 |
+
"epoch": 0.33707865168539325,
|
| 313 |
+
"grad_norm": 39.810089111328125,
|
| 314 |
+
"learning_rate": 4.99307556613492e-07,
|
| 315 |
+
"logits/chosen": -1.4572077989578247,
|
| 316 |
+
"logits/rejected": -1.8333911895751953,
|
| 317 |
+
"logps/chosen": -613.5606079101562,
|
| 318 |
+
"logps/rejected": -280.7749328613281,
|
| 319 |
+
"loss": 1.1563,
|
| 320 |
+
"rewards/accuracies": 1.0,
|
| 321 |
+
"rewards/chosen": 0.21885092556476593,
|
| 322 |
+
"rewards/margins": 2.2480263710021973,
|
| 323 |
+
"rewards/rejected": -2.029175281524658,
|
| 324 |
+
"step": 105
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"epoch": 0.35313001605136435,
|
| 328 |
+
"grad_norm": 29.89188575744629,
|
| 329 |
+
"learning_rate": 4.972340622767164e-07,
|
| 330 |
+
"logits/chosen": -1.4872273206710815,
|
| 331 |
+
"logits/rejected": -1.661569356918335,
|
| 332 |
+
"logps/chosen": -434.12689208984375,
|
| 333 |
+
"logps/rejected": -226.6045379638672,
|
| 334 |
+
"loss": 1.1971,
|
| 335 |
+
"rewards/accuracies": 1.0,
|
| 336 |
+
"rewards/chosen": 0.037406109273433685,
|
| 337 |
+
"rewards/margins": 2.0358262062072754,
|
| 338 |
+
"rewards/rejected": -1.9984201192855835,
|
| 339 |
+
"step": 110
|
| 340 |
+
},
|
| 341 |
+
{
|
| 342 |
+
"epoch": 0.36918138041733545,
|
| 343 |
+
"grad_norm": 34.67169952392578,
|
| 344 |
+
"learning_rate": 4.937910032091967e-07,
|
| 345 |
+
"logits/chosen": -1.5089974403381348,
|
| 346 |
+
"logits/rejected": -1.4937382936477661,
|
| 347 |
+
"logps/chosen": -326.23577880859375,
|
| 348 |
+
"logps/rejected": -242.26766967773438,
|
| 349 |
+
"loss": 1.069,
|
| 350 |
+
"rewards/accuracies": 1.0,
|
| 351 |
+
"rewards/chosen": 0.1510745882987976,
|
| 352 |
+
"rewards/margins": 2.4666709899902344,
|
| 353 |
+
"rewards/rejected": -2.315596580505371,
|
| 354 |
+
"step": 115
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"epoch": 0.3852327447833066,
|
| 358 |
+
"grad_norm": 26.853172302246094,
|
| 359 |
+
"learning_rate": 4.889974523987784e-07,
|
| 360 |
+
"logits/chosen": -1.709498405456543,
|
| 361 |
+
"logits/rejected": -1.7816112041473389,
|
| 362 |
+
"logps/chosen": -299.94091796875,
|
| 363 |
+
"logps/rejected": -224.59716796875,
|
| 364 |
+
"loss": 1.0784,
|
| 365 |
+
"rewards/accuracies": 1.0,
|
| 366 |
+
"rewards/chosen": 0.08231879025697708,
|
| 367 |
+
"rewards/margins": 1.5635088682174683,
|
| 368 |
+
"rewards/rejected": -1.4811900854110718,
|
| 369 |
+
"step": 120
|
| 370 |
+
},
|
| 371 |
+
{
|
| 372 |
+
"epoch": 0.4012841091492777,
|
| 373 |
+
"grad_norm": 25.12617301940918,
|
| 374 |
+
"learning_rate": 4.828799639459138e-07,
|
| 375 |
+
"logits/chosen": -2.142733573913574,
|
| 376 |
+
"logits/rejected": -1.8604745864868164,
|
| 377 |
+
"logps/chosen": -554.5831298828125,
|
| 378 |
+
"logps/rejected": -431.9894104003906,
|
| 379 |
+
"loss": 1.1384,
|
| 380 |
+
"rewards/accuracies": 1.0,
|
| 381 |
+
"rewards/chosen": 0.324582576751709,
|
| 382 |
+
"rewards/margins": 3.6467995643615723,
|
| 383 |
+
"rewards/rejected": -3.322216749191284,
|
| 384 |
+
"step": 125
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"epoch": 0.4173354735152488,
|
| 388 |
+
"grad_norm": 19.279260635375977,
|
| 389 |
+
"learning_rate": 4.754724259659727e-07,
|
| 390 |
+
"logits/chosen": -1.729953408241272,
|
| 391 |
+
"logits/rejected": -2.0016226768493652,
|
| 392 |
+
"logps/chosen": -337.40460205078125,
|
| 393 |
+
"logps/rejected": -218.6181182861328,
|
| 394 |
+
"loss": 1.1446,
|
| 395 |
+
"rewards/accuracies": 1.0,
|
| 396 |
+
"rewards/chosen": 0.23123934864997864,
|
| 397 |
+
"rewards/margins": 2.4159419536590576,
|
| 398 |
+
"rewards/rejected": -2.1847023963928223,
|
| 399 |
+
"step": 130
|
| 400 |
+
},
|
| 401 |
+
{
|
| 402 |
+
"epoch": 0.4333868378812199,
|
| 403 |
+
"grad_norm": 39.15530776977539,
|
| 404 |
+
"learning_rate": 4.6681587286443143e-07,
|
| 405 |
+
"logits/chosen": -1.7074105739593506,
|
| 406 |
+
"logits/rejected": -2.001408815383911,
|
| 407 |
+
"logps/chosen": -609.3516235351562,
|
| 408 |
+
"logps/rejected": -355.2177734375,
|
| 409 |
+
"loss": 1.1268,
|
| 410 |
+
"rewards/accuracies": 1.0,
|
| 411 |
+
"rewards/chosen": 0.2101774662733078,
|
| 412 |
+
"rewards/margins": 3.605560302734375,
|
| 413 |
+
"rewards/rejected": -3.3953826427459717,
|
| 414 |
+
"step": 135
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"epoch": 0.449438202247191,
|
| 418 |
+
"grad_norm": 28.309494018554688,
|
| 419 |
+
"learning_rate": 4.569582580248508e-07,
|
| 420 |
+
"logits/chosen": -1.3231592178344727,
|
| 421 |
+
"logits/rejected": -1.3130943775177002,
|
| 422 |
+
"logps/chosen": -456.9552307128906,
|
| 423 |
+
"logps/rejected": -249.678466796875,
|
| 424 |
+
"loss": 1.1836,
|
| 425 |
+
"rewards/accuracies": 1.0,
|
| 426 |
+
"rewards/chosen": 0.09645295143127441,
|
| 427 |
+
"rewards/margins": 2.3201193809509277,
|
| 428 |
+
"rewards/rejected": -2.2236666679382324,
|
| 429 |
+
"step": 140
|
| 430 |
+
},
|
| 431 |
+
{
|
| 432 |
+
"epoch": 0.4654895666131621,
|
| 433 |
+
"grad_norm": 25.581401824951172,
|
| 434 |
+
"learning_rate": 4.4595418816885e-07,
|
| 435 |
+
"logits/chosen": -2.18888783454895,
|
| 436 |
+
"logits/rejected": -1.896630048751831,
|
| 437 |
+
"logps/chosen": -391.4627380371094,
|
| 438 |
+
"logps/rejected": -396.46240234375,
|
| 439 |
+
"loss": 1.0783,
|
| 440 |
+
"rewards/accuracies": 0.8999999761581421,
|
| 441 |
+
"rewards/chosen": 0.3204360604286194,
|
| 442 |
+
"rewards/margins": 3.3659069538116455,
|
| 443 |
+
"rewards/rejected": -3.045470714569092,
|
| 444 |
+
"step": 145
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"epoch": 0.48154093097913325,
|
| 448 |
+
"grad_norm": 27.107683181762695,
|
| 449 |
+
"learning_rate": 4.338646208596008e-07,
|
| 450 |
+
"logits/chosen": -1.3190141916275024,
|
| 451 |
+
"logits/rejected": -1.4809017181396484,
|
| 452 |
+
"logps/chosen": -668.1514892578125,
|
| 453 |
+
"logps/rejected": -365.1762390136719,
|
| 454 |
+
"loss": 1.2031,
|
| 455 |
+
"rewards/accuracies": 1.0,
|
| 456 |
+
"rewards/chosen": 0.00630459189414978,
|
| 457 |
+
"rewards/margins": 4.1467790603637695,
|
| 458 |
+
"rewards/rejected": -4.140474796295166,
|
| 459 |
+
"step": 150
|
| 460 |
+
},
|
| 461 |
+
{
|
| 462 |
+
"epoch": 0.49759229534510435,
|
| 463 |
+
"grad_norm": 23.90721893310547,
|
| 464 |
+
"learning_rate": 4.207565268245355e-07,
|
| 465 |
+
"logits/chosen": -1.243398904800415,
|
| 466 |
+
"logits/rejected": -1.5890659093856812,
|
| 467 |
+
"logps/chosen": -605.8566284179688,
|
| 468 |
+
"logps/rejected": -339.843994140625,
|
| 469 |
+
"loss": 1.0329,
|
| 470 |
+
"rewards/accuracies": 1.0,
|
| 471 |
+
"rewards/chosen": 0.4351004958152771,
|
| 472 |
+
"rewards/margins": 4.156341552734375,
|
| 473 |
+
"rewards/rejected": -3.721240997314453,
|
| 474 |
+
"step": 155
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"epoch": 0.5136436597110754,
|
| 478 |
+
"grad_norm": 15.980473518371582,
|
| 479 |
+
"learning_rate": 4.0670251896784847e-07,
|
| 480 |
+
"logits/chosen": -1.6174713373184204,
|
| 481 |
+
"logits/rejected": -2.031649112701416,
|
| 482 |
+
"logps/chosen": -572.9243774414062,
|
| 483 |
+
"logps/rejected": -294.9161376953125,
|
| 484 |
+
"loss": 1.0657,
|
| 485 |
+
"rewards/accuracies": 1.0,
|
| 486 |
+
"rewards/chosen": 0.5264326333999634,
|
| 487 |
+
"rewards/margins": 3.414010524749756,
|
| 488 |
+
"rewards/rejected": -2.887577533721924,
|
| 489 |
+
"step": 160
|
| 490 |
+
},
|
| 491 |
+
{
|
| 492 |
+
"epoch": 0.5296950240770465,
|
| 493 |
+
"grad_norm": 16.01748275756836,
|
| 494 |
+
"learning_rate": 3.917804501278941e-07,
|
| 495 |
+
"logits/chosen": -1.8946282863616943,
|
| 496 |
+
"logits/rejected": -2.036999225616455,
|
| 497 |
+
"logps/chosen": -455.58428955078125,
|
| 498 |
+
"logps/rejected": -225.6637725830078,
|
| 499 |
+
"loss": 1.1023,
|
| 500 |
+
"rewards/accuracies": 0.8999999761581421,
|
| 501 |
+
"rewards/chosen": 0.3108845353126526,
|
| 502 |
+
"rewards/margins": 2.991553544998169,
|
| 503 |
+
"rewards/rejected": -2.680669069290161,
|
| 504 |
+
"step": 165
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"epoch": 0.5457463884430177,
|
| 508 |
+
"grad_norm": 18.181249618530273,
|
| 509 |
+
"learning_rate": 3.760729818077223e-07,
|
| 510 |
+
"logits/chosen": -1.3452972173690796,
|
| 511 |
+
"logits/rejected": -1.663169264793396,
|
| 512 |
+
"logps/chosen": -400.55657958984375,
|
| 513 |
+
"logps/rejected": -260.5176086425781,
|
| 514 |
+
"loss": 1.0674,
|
| 515 |
+
"rewards/accuracies": 1.0,
|
| 516 |
+
"rewards/chosen": 0.9225478172302246,
|
| 517 |
+
"rewards/margins": 3.818953037261963,
|
| 518 |
+
"rewards/rejected": -2.8964052200317383,
|
| 519 |
+
"step": 170
|
| 520 |
+
},
|
| 521 |
+
{
|
| 522 |
+
"epoch": 0.5617977528089888,
|
| 523 |
+
"grad_norm": 12.927690505981445,
|
| 524 |
+
"learning_rate": 3.596671262677898e-07,
|
| 525 |
+
"logits/chosen": -0.9975664019584656,
|
| 526 |
+
"logits/rejected": -1.590592384338379,
|
| 527 |
+
"logps/chosen": -484.78472900390625,
|
| 528 |
+
"logps/rejected": -326.25018310546875,
|
| 529 |
+
"loss": 0.9874,
|
| 530 |
+
"rewards/accuracies": 1.0,
|
| 531 |
+
"rewards/chosen": 0.8374244570732117,
|
| 532 |
+
"rewards/margins": 4.229715824127197,
|
| 533 |
+
"rewards/rejected": -3.392291307449341,
|
| 534 |
+
"step": 175
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"epoch": 0.5778491171749599,
|
| 538 |
+
"grad_norm": 22.454750061035156,
|
| 539 |
+
"learning_rate": 3.426537645174456e-07,
|
| 540 |
+
"logits/chosen": -1.1456716060638428,
|
| 541 |
+
"logits/rejected": -0.9254422187805176,
|
| 542 |
+
"logps/chosen": -229.30184936523438,
|
| 543 |
+
"logps/rejected": -184.46151733398438,
|
| 544 |
+
"loss": 1.2047,
|
| 545 |
+
"rewards/accuracies": 1.0,
|
| 546 |
+
"rewards/chosen": -0.03938926011323929,
|
| 547 |
+
"rewards/margins": 2.173858880996704,
|
| 548 |
+
"rewards/rejected": -2.2132482528686523,
|
| 549 |
+
"step": 180
|
| 550 |
+
},
|
| 551 |
+
{
|
| 552 |
+
"epoch": 0.593900481540931,
|
| 553 |
+
"grad_norm": 19.639509201049805,
|
| 554 |
+
"learning_rate": 3.2512714287530003e-07,
|
| 555 |
+
"logits/chosen": -1.6827218532562256,
|
| 556 |
+
"logits/rejected": -1.6239674091339111,
|
| 557 |
+
"logps/chosen": -248.2288818359375,
|
| 558 |
+
"logps/rejected": -174.70457458496094,
|
| 559 |
+
"loss": 1.0531,
|
| 560 |
+
"rewards/accuracies": 1.0,
|
| 561 |
+
"rewards/chosen": -0.023051679134368896,
|
| 562 |
+
"rewards/margins": 2.488781690597534,
|
| 563 |
+
"rewards/rejected": -2.5118331909179688,
|
| 564 |
+
"step": 185
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"epoch": 0.6099518459069021,
|
| 568 |
+
"grad_norm": 28.825136184692383,
|
| 569 |
+
"learning_rate": 3.071843508873045e-07,
|
| 570 |
+
"logits/chosen": -1.655289649963379,
|
| 571 |
+
"logits/rejected": -1.2866331338882446,
|
| 572 |
+
"logps/chosen": -399.61016845703125,
|
| 573 |
+
"logps/rejected": -372.4976806640625,
|
| 574 |
+
"loss": 1.0844,
|
| 575 |
+
"rewards/accuracies": 1.0,
|
| 576 |
+
"rewards/chosen": 0.24592149257659912,
|
| 577 |
+
"rewards/margins": 4.032209873199463,
|
| 578 |
+
"rewards/rejected": -3.7862884998321533,
|
| 579 |
+
"step": 190
|
| 580 |
+
},
|
| 581 |
+
{
|
| 582 |
+
"epoch": 0.6260032102728732,
|
| 583 |
+
"grad_norm": 14.884502410888672,
|
| 584 |
+
"learning_rate": 2.8892478349463983e-07,
|
| 585 |
+
"logits/chosen": -1.6147607564926147,
|
| 586 |
+
"logits/rejected": -1.9210180044174194,
|
| 587 |
+
"logps/chosen": -517.3702392578125,
|
| 588 |
+
"logps/rejected": -201.2556915283203,
|
| 589 |
+
"loss": 1.087,
|
| 590 |
+
"rewards/accuracies": 1.0,
|
| 591 |
+
"rewards/chosen": 0.48443666100502014,
|
| 592 |
+
"rewards/margins": 3.167422294616699,
|
| 593 |
+
"rewards/rejected": -2.682985782623291,
|
| 594 |
+
"step": 195
|
| 595 |
+
},
|
| 596 |
+
{
|
| 597 |
+
"epoch": 0.6420545746388443,
|
| 598 |
+
"grad_norm": 20.6270694732666,
|
| 599 |
+
"learning_rate": 2.704495904307581e-07,
|
| 600 |
+
"logits/chosen": -1.466150164604187,
|
| 601 |
+
"logits/rejected": -1.8435227870941162,
|
| 602 |
+
"logps/chosen": -447.94708251953125,
|
| 603 |
+
"logps/rejected": -320.72796630859375,
|
| 604 |
+
"loss": 1.0463,
|
| 605 |
+
"rewards/accuracies": 1.0,
|
| 606 |
+
"rewards/chosen": 0.4680330157279968,
|
| 607 |
+
"rewards/margins": 3.9673094749450684,
|
| 608 |
+
"rewards/rejected": -3.499276638031006,
|
| 609 |
+
"step": 200
|
| 610 |
+
},
|
| 611 |
+
{
|
| 612 |
+
"epoch": 0.6581059390048154,
|
| 613 |
+
"grad_norm": 20.02724266052246,
|
| 614 |
+
"learning_rate": 2.5186111589767186e-07,
|
| 615 |
+
"logits/chosen": -0.9430475234985352,
|
| 616 |
+
"logits/rejected": -1.5955369472503662,
|
| 617 |
+
"logps/chosen": -512.3245849609375,
|
| 618 |
+
"logps/rejected": -293.04388427734375,
|
| 619 |
+
"loss": 1.1441,
|
| 620 |
+
"rewards/accuracies": 1.0,
|
| 621 |
+
"rewards/chosen": 0.22035065293312073,
|
| 622 |
+
"rewards/margins": 3.5233726501464844,
|
| 623 |
+
"rewards/rejected": -3.3030216693878174,
|
| 624 |
+
"step": 205
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"epoch": 0.6741573033707865,
|
| 628 |
+
"grad_norm": 19.69023895263672,
|
| 629 |
+
"learning_rate": 2.3326233162542652e-07,
|
| 630 |
+
"logits/chosen": -1.889314889907837,
|
| 631 |
+
"logits/rejected": -2.119999885559082,
|
| 632 |
+
"logps/chosen": -283.8106384277344,
|
| 633 |
+
"logps/rejected": -217.4166259765625,
|
| 634 |
+
"loss": 1.046,
|
| 635 |
+
"rewards/accuracies": 1.0,
|
| 636 |
+
"rewards/chosen": 0.3023836016654968,
|
| 637 |
+
"rewards/margins": 2.977555751800537,
|
| 638 |
+
"rewards/rejected": -2.6751716136932373,
|
| 639 |
+
"step": 210
|
| 640 |
+
},
|
| 641 |
+
{
|
| 642 |
+
"epoch": 0.6902086677367576,
|
| 643 |
+
"grad_norm": 24.165843963623047,
|
| 644 |
+
"learning_rate": 2.1475626645535367e-07,
|
| 645 |
+
"logits/chosen": -1.2131136655807495,
|
| 646 |
+
"logits/rejected": -0.9721648097038269,
|
| 647 |
+
"logps/chosen": -373.4245910644531,
|
| 648 |
+
"logps/rejected": -234.89364624023438,
|
| 649 |
+
"loss": 1.093,
|
| 650 |
+
"rewards/accuracies": 1.0,
|
| 651 |
+
"rewards/chosen": 0.3658973276615143,
|
| 652 |
+
"rewards/margins": 3.6775901317596436,
|
| 653 |
+
"rewards/rejected": -3.311692714691162,
|
| 654 |
+
"step": 215
|
| 655 |
+
},
|
| 656 |
+
{
|
| 657 |
+
"epoch": 0.7062600321027287,
|
| 658 |
+
"grad_norm": 21.797191619873047,
|
| 659 |
+
"learning_rate": 1.9644543560695136e-07,
|
| 660 |
+
"logits/chosen": -1.5296471118927002,
|
| 661 |
+
"logits/rejected": -1.6277145147323608,
|
| 662 |
+
"logps/chosen": -412.1878356933594,
|
| 663 |
+
"logps/rejected": -314.4945373535156,
|
| 664 |
+
"loss": 1.0305,
|
| 665 |
+
"rewards/accuracies": 1.0,
|
| 666 |
+
"rewards/chosen": 0.5929609537124634,
|
| 667 |
+
"rewards/margins": 3.6523776054382324,
|
| 668 |
+
"rewards/rejected": -3.0594162940979004,
|
| 669 |
+
"step": 220
|
| 670 |
+
},
|
| 671 |
+
{
|
| 672 |
+
"epoch": 0.7223113964686998,
|
| 673 |
+
"grad_norm": 19.618436813354492,
|
| 674 |
+
"learning_rate": 1.784312727899994e-07,
|
| 675 |
+
"logits/chosen": -1.4759819507598877,
|
| 676 |
+
"logits/rejected": -1.783433198928833,
|
| 677 |
+
"logps/chosen": -510.66180419921875,
|
| 678 |
+
"logps/rejected": -262.74725341796875,
|
| 679 |
+
"loss": 1.0172,
|
| 680 |
+
"rewards/accuracies": 1.0,
|
| 681 |
+
"rewards/chosen": 0.28623512387275696,
|
| 682 |
+
"rewards/margins": 3.71972918510437,
|
| 683 |
+
"rewards/rejected": -3.4334940910339355,
|
| 684 |
+
"step": 225
|
| 685 |
+
},
|
| 686 |
+
{
|
| 687 |
+
"epoch": 0.7383627608346709,
|
| 688 |
+
"grad_norm": 15.515320777893066,
|
| 689 |
+
"learning_rate": 1.6081356830774623e-07,
|
| 690 |
+
"logits/chosen": -0.7434533834457397,
|
| 691 |
+
"logits/rejected": -1.0206338167190552,
|
| 692 |
+
"logps/chosen": -625.6656494140625,
|
| 693 |
+
"logps/rejected": -294.1523742675781,
|
| 694 |
+
"loss": 1.0822,
|
| 695 |
+
"rewards/accuracies": 1.0,
|
| 696 |
+
"rewards/chosen": 0.5695539712905884,
|
| 697 |
+
"rewards/margins": 4.612680912017822,
|
| 698 |
+
"rewards/rejected": -4.043127536773682,
|
| 699 |
+
"step": 230
|
| 700 |
+
},
|
| 701 |
+
{
|
| 702 |
+
"epoch": 0.7544141252006421,
|
| 703 |
+
"grad_norm": 14.756795883178711,
|
| 704 |
+
"learning_rate": 1.4368991626382548e-07,
|
| 705 |
+
"logits/chosen": -1.3159469366073608,
|
| 706 |
+
"logits/rejected": -1.9151531457901,
|
| 707 |
+
"logps/chosen": -506.05084228515625,
|
| 708 |
+
"logps/rejected": -288.33526611328125,
|
| 709 |
+
"loss": 0.9993,
|
| 710 |
+
"rewards/accuracies": 0.8999999761581421,
|
| 711 |
+
"rewards/chosen": 0.8690406084060669,
|
| 712 |
+
"rewards/margins": 4.248424053192139,
|
| 713 |
+
"rewards/rejected": -3.3793838024139404,
|
| 714 |
+
"step": 235
|
| 715 |
+
},
|
| 716 |
+
{
|
| 717 |
+
"epoch": 0.7704654895666132,
|
| 718 |
+
"grad_norm": 12.66774845123291,
|
| 719 |
+
"learning_rate": 1.2715517393512238e-07,
|
| 720 |
+
"logits/chosen": -1.2622467279434204,
|
| 721 |
+
"logits/rejected": -1.3659932613372803,
|
| 722 |
+
"logps/chosen": -591.2831420898438,
|
| 723 |
+
"logps/rejected": -315.7579040527344,
|
| 724 |
+
"loss": 1.1002,
|
| 725 |
+
"rewards/accuracies": 1.0,
|
| 726 |
+
"rewards/chosen": 0.42772501707077026,
|
| 727 |
+
"rewards/margins": 4.257325172424316,
|
| 728 |
+
"rewards/rejected": -3.8295998573303223,
|
| 729 |
+
"step": 240
|
| 730 |
+
},
|
| 731 |
+
{
|
| 732 |
+
"epoch": 0.7865168539325843,
|
| 733 |
+
"grad_norm": 16.25082015991211,
|
| 734 |
+
"learning_rate": 1.1130093630542198e-07,
|
| 735 |
+
"logits/chosen": -1.1802631616592407,
|
| 736 |
+
"logits/rejected": -1.249211072921753,
|
| 737 |
+
"logps/chosen": -482.888671875,
|
| 738 |
+
"logps/rejected": -345.88824462890625,
|
| 739 |
+
"loss": 0.983,
|
| 740 |
+
"rewards/accuracies": 1.0,
|
| 741 |
+
"rewards/chosen": 0.8362643122673035,
|
| 742 |
+
"rewards/margins": 4.463853359222412,
|
| 743 |
+
"rewards/rejected": -3.627588987350464,
|
| 744 |
+
"step": 245
|
| 745 |
+
},
|
| 746 |
+
{
|
| 747 |
+
"epoch": 0.8025682182985554,
|
| 748 |
+
"grad_norm": 17.564659118652344,
|
| 749 |
+
"learning_rate": 9.621502867068285e-08,
|
| 750 |
+
"logits/chosen": -1.4240429401397705,
|
| 751 |
+
"logits/rejected": -1.372710943222046,
|
| 752 |
+
"logps/chosen": -459.75030517578125,
|
| 753 |
+
"logps/rejected": -321.6226501464844,
|
| 754 |
+
"loss": 1.0852,
|
| 755 |
+
"rewards/accuracies": 1.0,
|
| 756 |
+
"rewards/chosen": 1.0832629203796387,
|
| 757 |
+
"rewards/margins": 4.887136936187744,
|
| 758 |
+
"rewards/rejected": -3.8038737773895264,
|
| 759 |
+
"step": 250
|
| 760 |
+
},
|
| 761 |
+
{
|
| 762 |
+
"epoch": 0.8186195826645265,
|
| 763 |
+
"grad_norm": 18.318283081054688,
|
| 764 |
+
"learning_rate": 8.198102012667407e-08,
|
| 765 |
+
"logits/chosen": -1.9069766998291016,
|
| 766 |
+
"logits/rejected": -1.9020580053329468,
|
| 767 |
+
"logps/chosen": -401.6404724121094,
|
| 768 |
+
"logps/rejected": -295.3546142578125,
|
| 769 |
+
"loss": 1.0466,
|
| 770 |
+
"rewards/accuracies": 1.0,
|
| 771 |
+
"rewards/chosen": 0.12345659732818604,
|
| 772 |
+
"rewards/margins": 3.521843671798706,
|
| 773 |
+
"rewards/rejected": -3.3983871936798096,
|
| 774 |
+
"step": 255
|
| 775 |
+
},
|
| 776 |
+
{
|
| 777 |
+
"epoch": 0.8346709470304976,
|
| 778 |
+
"grad_norm": 14.781198501586914,
|
| 779 |
+
"learning_rate": 6.86777606340341e-08,
|
| 780 |
+
"logits/chosen": -1.7581796646118164,
|
| 781 |
+
"logits/rejected": -1.6436153650283813,
|
| 782 |
+
"logps/chosen": -182.29490661621094,
|
| 783 |
+
"logps/rejected": -156.72459411621094,
|
| 784 |
+
"loss": 1.1013,
|
| 785 |
+
"rewards/accuracies": 1.0,
|
| 786 |
+
"rewards/chosen": 0.2946431040763855,
|
| 787 |
+
"rewards/margins": 2.2983453273773193,
|
| 788 |
+
"rewards/rejected": -2.003702163696289,
|
| 789 |
+
"step": 260
|
| 790 |
+
},
|
| 791 |
+
{
|
| 792 |
+
"epoch": 0.8507223113964687,
|
| 793 |
+
"grad_norm": 19.03549575805664,
|
| 794 |
+
"learning_rate": 5.637894422520026e-08,
|
| 795 |
+
"logits/chosen": -1.8069452047348022,
|
| 796 |
+
"logits/rejected": -1.8712180852890015,
|
| 797 |
+
"logps/chosen": -320.7396545410156,
|
| 798 |
+
"logps/rejected": -230.3447265625,
|
| 799 |
+
"loss": 0.9904,
|
| 800 |
+
"rewards/accuracies": 1.0,
|
| 801 |
+
"rewards/chosen": 0.1634465754032135,
|
| 802 |
+
"rewards/margins": 2.5912208557128906,
|
| 803 |
+
"rewards/rejected": -2.427773952484131,
|
| 804 |
+
"step": 265
|
| 805 |
+
},
|
| 806 |
+
{
|
| 807 |
+
"epoch": 0.8667736757624398,
|
| 808 |
+
"grad_norm": 20.6317081451416,
|
| 809 |
+
"learning_rate": 4.515270077284594e-08,
|
| 810 |
+
"logits/chosen": -1.8391157388687134,
|
| 811 |
+
"logits/rejected": -1.8903614282608032,
|
| 812 |
+
"logps/chosen": -493.16998291015625,
|
| 813 |
+
"logps/rejected": -243.4764862060547,
|
| 814 |
+
"loss": 1.0834,
|
| 815 |
+
"rewards/accuracies": 1.0,
|
| 816 |
+
"rewards/chosen": 0.6217156648635864,
|
| 817 |
+
"rewards/margins": 3.6283297538757324,
|
| 818 |
+
"rewards/rejected": -3.0066142082214355,
|
| 819 |
+
"step": 270
|
| 820 |
+
},
|
| 821 |
+
{
|
| 822 |
+
"epoch": 0.8828250401284109,
|
| 823 |
+
"grad_norm": 15.678504943847656,
|
| 824 |
+
"learning_rate": 3.506121858124253e-08,
|
| 825 |
+
"logits/chosen": -2.1319689750671387,
|
| 826 |
+
"logits/rejected": -2.3358452320098877,
|
| 827 |
+
"logps/chosen": -544.0633544921875,
|
| 828 |
+
"logps/rejected": -284.2521057128906,
|
| 829 |
+
"loss": 1.0789,
|
| 830 |
+
"rewards/accuracies": 1.0,
|
| 831 |
+
"rewards/chosen": 0.1263357549905777,
|
| 832 |
+
"rewards/margins": 3.653578996658325,
|
| 833 |
+
"rewards/rejected": -3.5272433757781982,
|
| 834 |
+
"step": 275
|
| 835 |
+
},
|
| 836 |
+
{
|
| 837 |
+
"epoch": 0.898876404494382,
|
| 838 |
+
"grad_norm": 35.75740051269531,
|
| 839 |
+
"learning_rate": 2.6160399891218986e-08,
|
| 840 |
+
"logits/chosen": -1.9280614852905273,
|
| 841 |
+
"logits/rejected": -2.3073010444641113,
|
| 842 |
+
"logps/chosen": -706.4361572265625,
|
| 843 |
+
"logps/rejected": -341.9065856933594,
|
| 844 |
+
"loss": 1.0365,
|
| 845 |
+
"rewards/accuracies": 1.0,
|
| 846 |
+
"rewards/chosen": 0.607284426689148,
|
| 847 |
+
"rewards/margins": 4.847535133361816,
|
| 848 |
+
"rewards/rejected": -4.240251064300537,
|
| 849 |
+
"step": 280
|
| 850 |
+
},
|
| 851 |
+
{
|
| 852 |
+
"epoch": 0.9149277688603531,
|
| 853 |
+
"grad_norm": 12.305197715759277,
|
| 854 |
+
"learning_rate": 1.8499551207066726e-08,
|
| 855 |
+
"logits/chosen": -1.5439479351043701,
|
| 856 |
+
"logits/rejected": -1.656998872756958,
|
| 857 |
+
"logps/chosen": -385.1285095214844,
|
| 858 |
+
"logps/rejected": -361.050537109375,
|
| 859 |
+
"loss": 1.0772,
|
| 860 |
+
"rewards/accuracies": 1.0,
|
| 861 |
+
"rewards/chosen": 0.3932652473449707,
|
| 862 |
+
"rewards/margins": 4.640822410583496,
|
| 863 |
+
"rewards/rejected": -4.247557640075684,
|
| 864 |
+
"step": 285
|
| 865 |
+
},
|
| 866 |
+
{
|
| 867 |
+
"epoch": 0.9309791332263242,
|
| 868 |
+
"grad_norm": 17.36638069152832,
|
| 869 |
+
"learning_rate": 1.2121110160836696e-08,
|
| 870 |
+
"logits/chosen": -1.1358602046966553,
|
| 871 |
+
"logits/rejected": -1.322624921798706,
|
| 872 |
+
"logps/chosen": -359.67181396484375,
|
| 873 |
+
"logps/rejected": -275.7583312988281,
|
| 874 |
+
"loss": 1.0249,
|
| 875 |
+
"rewards/accuracies": 1.0,
|
| 876 |
+
"rewards/chosen": 0.3709088861942291,
|
| 877 |
+
"rewards/margins": 3.387117862701416,
|
| 878 |
+
"rewards/rejected": -3.0162088871002197,
|
| 879 |
+
"step": 290
|
| 880 |
+
},
|
| 881 |
+
{
|
| 882 |
+
"epoch": 0.9470304975922953,
|
| 883 |
+
"grad_norm": 13.696333885192871,
|
| 884 |
+
"learning_rate": 7.060410427078472e-09,
|
| 885 |
+
"logits/chosen": -1.3360637426376343,
|
| 886 |
+
"logits/rejected": -1.6079111099243164,
|
| 887 |
+
"logps/chosen": -504.14727783203125,
|
| 888 |
+
"logps/rejected": -352.38812255859375,
|
| 889 |
+
"loss": 0.9998,
|
| 890 |
+
"rewards/accuracies": 0.8999999761581421,
|
| 891 |
+
"rewards/chosen": 0.5898990631103516,
|
| 892 |
+
"rewards/margins": 4.491053581237793,
|
| 893 |
+
"rewards/rejected": -3.9011547565460205,
|
| 894 |
+
"step": 295
|
| 895 |
+
},
|
| 896 |
+
{
|
| 897 |
+
"epoch": 0.9630818619582665,
|
| 898 |
+
"grad_norm": 19.26398468017578,
|
| 899 |
+
"learning_rate": 3.345485990286029e-09,
|
| 900 |
+
"logits/chosen": -2.010549783706665,
|
| 901 |
+
"logits/rejected": -1.7452675104141235,
|
| 902 |
+
"logps/chosen": -280.08685302734375,
|
| 903 |
+
"logps/rejected": -282.55950927734375,
|
| 904 |
+
"loss": 1.0632,
|
| 905 |
+
"rewards/accuracies": 1.0,
|
| 906 |
+
"rewards/chosen": 0.5629211664199829,
|
| 907 |
+
"rewards/margins": 3.6783041954040527,
|
| 908 |
+
"rewards/rejected": -3.1153833866119385,
|
| 909 |
+
"step": 300
|
| 910 |
+
},
|
| 911 |
+
{
|
| 912 |
+
"epoch": 0.9791332263242376,
|
| 913 |
+
"grad_norm": 20.505722045898438,
|
| 914 |
+
"learning_rate": 9.969158493204066e-10,
|
| 915 |
+
"logits/chosen": -1.6427099704742432,
|
| 916 |
+
"logits/rejected": -1.7004835605621338,
|
| 917 |
+
"logps/chosen": -459.8740234375,
|
| 918 |
+
"logps/rejected": -258.29705810546875,
|
| 919 |
+
"loss": 1.0695,
|
| 920 |
+
"rewards/accuracies": 1.0,
|
| 921 |
+
"rewards/chosen": 0.22339797019958496,
|
| 922 |
+
"rewards/margins": 3.7468554973602295,
|
| 923 |
+
"rewards/rejected": -3.5234577655792236,
|
| 924 |
+
"step": 305
|
| 925 |
+
},
|
| 926 |
+
{
|
| 927 |
+
"epoch": 0.9951845906902087,
|
| 928 |
+
"grad_norm": 17.717132568359375,
|
| 929 |
+
"learning_rate": 2.7710019076532255e-11,
|
| 930 |
+
"logits/chosen": -1.4575186967849731,
|
| 931 |
+
"logits/rejected": -1.1399505138397217,
|
| 932 |
+
"logps/chosen": -437.7586975097656,
|
| 933 |
+
"logps/rejected": -279.3139343261719,
|
| 934 |
+
"loss": 1.0752,
|
| 935 |
+
"rewards/accuracies": 1.0,
|
| 936 |
+
"rewards/chosen": 0.02865469455718994,
|
| 937 |
+
"rewards/margins": 3.524338483810425,
|
| 938 |
+
"rewards/rejected": -3.495683193206787,
|
| 939 |
+
"step": 310
|
| 940 |
+
},
|
| 941 |
+
{
|
| 942 |
+
"epoch": 0.9983948635634029,
|
| 943 |
+
"step": 311,
|
| 944 |
+
"total_flos": 2.2772099803879834e+17,
|
| 945 |
+
"train_loss": 1.2078840717242079,
|
| 946 |
+
"train_runtime": 1561.3327,
|
| 947 |
+
"train_samples_per_second": 3.19,
|
| 948 |
+
"train_steps_per_second": 0.199
|
| 949 |
+
}
|
| 950 |
+
],
|
| 951 |
+
"logging_steps": 5,
|
| 952 |
+
"max_steps": 311,
|
| 953 |
+
"num_input_tokens_seen": 0,
|
| 954 |
+
"num_train_epochs": 1,
|
| 955 |
+
"save_steps": 100,
|
| 956 |
+
"total_flos": 2.2772099803879834e+17,
|
| 957 |
+
"train_batch_size": 1,
|
| 958 |
+
"trial_name": null,
|
| 959 |
+
"trial_params": null
|
| 960 |
+
}
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da96ce0b0d54654a796e414437f48f8550a2be70f85074f6eb0c43660bd98fd0
|
| 3 |
+
size 6392
|
training_loss.png
ADDED
|
training_rewards_accuracies.png
ADDED
|
vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|