update model files. make sure each shard < 5GiB
Browse files
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:11590a9f31d6b7c2c072446b5c1ec53c29bc0626a323bf6bd23da373fbe851ce
|
| 3 |
+
size 4874083760
|
model-00002-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:527c47483ed676e9d10f3dcb3280f77b16951b02e057596ba06496ead00c3b3f
|
| 3 |
+
size 4203710128
|
model.safetensors.index.json
CHANGED
|
@@ -601,33 +601,33 @@
|
|
| 601 |
"language_model.model.layers.21.mlp.down_proj.qweight": "model-00001-of-00002.safetensors",
|
| 602 |
"language_model.model.layers.21.mlp.down_proj.qzeros": "model-00001-of-00002.safetensors",
|
| 603 |
"language_model.model.layers.21.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
|
| 604 |
-
"language_model.model.layers.22.input_layernorm.weight": "model-
|
| 605 |
-
"language_model.model.layers.22.post_attention_layernorm.weight": "model-
|
| 606 |
-
"language_model.model.layers.22.pre_feedforward_layernorm.weight": "model-
|
| 607 |
-
"language_model.model.layers.22.post_feedforward_layernorm.weight": "model-
|
| 608 |
-
"language_model.model.layers.22.self_attn.q_norm.weight": "model-
|
| 609 |
-
"language_model.model.layers.22.self_attn.k_norm.weight": "model-
|
| 610 |
-
"language_model.model.layers.22.self_attn.q_proj.qweight": "model-
|
| 611 |
-
"language_model.model.layers.22.self_attn.q_proj.qzeros": "model-
|
| 612 |
-
"language_model.model.layers.22.self_attn.q_proj.scales": "model-
|
| 613 |
-
"language_model.model.layers.22.self_attn.k_proj.qweight": "model-
|
| 614 |
-
"language_model.model.layers.22.self_attn.k_proj.qzeros": "model-
|
| 615 |
-
"language_model.model.layers.22.self_attn.k_proj.scales": "model-
|
| 616 |
-
"language_model.model.layers.22.self_attn.v_proj.qweight": "model-
|
| 617 |
-
"language_model.model.layers.22.self_attn.v_proj.qzeros": "model-
|
| 618 |
-
"language_model.model.layers.22.self_attn.v_proj.scales": "model-
|
| 619 |
-
"language_model.model.layers.22.self_attn.o_proj.qweight": "model-
|
| 620 |
-
"language_model.model.layers.22.self_attn.o_proj.qzeros": "model-
|
| 621 |
-
"language_model.model.layers.22.self_attn.o_proj.scales": "model-
|
| 622 |
-
"language_model.model.layers.22.mlp.gate_proj.qweight": "model-
|
| 623 |
-
"language_model.model.layers.22.mlp.gate_proj.qzeros": "model-
|
| 624 |
-
"language_model.model.layers.22.mlp.gate_proj.scales": "model-
|
| 625 |
-
"language_model.model.layers.22.mlp.up_proj.qweight": "model-
|
| 626 |
-
"language_model.model.layers.22.mlp.up_proj.qzeros": "model-
|
| 627 |
-
"language_model.model.layers.22.mlp.up_proj.scales": "model-
|
| 628 |
-
"language_model.model.layers.22.mlp.down_proj.qweight": "model-
|
| 629 |
-
"language_model.model.layers.22.mlp.down_proj.qzeros": "model-
|
| 630 |
-
"language_model.model.layers.22.mlp.down_proj.scales": "model-
|
| 631 |
"language_model.model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 632 |
"language_model.model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 633 |
"language_model.model.layers.23.pre_feedforward_layernorm.weight": "model-00002-of-00002.safetensors",
|
|
|
|
| 601 |
"language_model.model.layers.21.mlp.down_proj.qweight": "model-00001-of-00002.safetensors",
|
| 602 |
"language_model.model.layers.21.mlp.down_proj.qzeros": "model-00001-of-00002.safetensors",
|
| 603 |
"language_model.model.layers.21.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
|
| 604 |
+
"language_model.model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 605 |
+
"language_model.model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 606 |
+
"language_model.model.layers.22.pre_feedforward_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 607 |
+
"language_model.model.layers.22.post_feedforward_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 608 |
+
"language_model.model.layers.22.self_attn.q_norm.weight": "model-00002-of-00002.safetensors",
|
| 609 |
+
"language_model.model.layers.22.self_attn.k_norm.weight": "model-00002-of-00002.safetensors",
|
| 610 |
+
"language_model.model.layers.22.self_attn.q_proj.qweight": "model-00002-of-00002.safetensors",
|
| 611 |
+
"language_model.model.layers.22.self_attn.q_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 612 |
+
"language_model.model.layers.22.self_attn.q_proj.scales": "model-00002-of-00002.safetensors",
|
| 613 |
+
"language_model.model.layers.22.self_attn.k_proj.qweight": "model-00002-of-00002.safetensors",
|
| 614 |
+
"language_model.model.layers.22.self_attn.k_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 615 |
+
"language_model.model.layers.22.self_attn.k_proj.scales": "model-00002-of-00002.safetensors",
|
| 616 |
+
"language_model.model.layers.22.self_attn.v_proj.qweight": "model-00002-of-00002.safetensors",
|
| 617 |
+
"language_model.model.layers.22.self_attn.v_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 618 |
+
"language_model.model.layers.22.self_attn.v_proj.scales": "model-00002-of-00002.safetensors",
|
| 619 |
+
"language_model.model.layers.22.self_attn.o_proj.qweight": "model-00002-of-00002.safetensors",
|
| 620 |
+
"language_model.model.layers.22.self_attn.o_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 621 |
+
"language_model.model.layers.22.self_attn.o_proj.scales": "model-00002-of-00002.safetensors",
|
| 622 |
+
"language_model.model.layers.22.mlp.gate_proj.qweight": "model-00002-of-00002.safetensors",
|
| 623 |
+
"language_model.model.layers.22.mlp.gate_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 624 |
+
"language_model.model.layers.22.mlp.gate_proj.scales": "model-00002-of-00002.safetensors",
|
| 625 |
+
"language_model.model.layers.22.mlp.up_proj.qweight": "model-00002-of-00002.safetensors",
|
| 626 |
+
"language_model.model.layers.22.mlp.up_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 627 |
+
"language_model.model.layers.22.mlp.up_proj.scales": "model-00002-of-00002.safetensors",
|
| 628 |
+
"language_model.model.layers.22.mlp.down_proj.qweight": "model-00002-of-00002.safetensors",
|
| 629 |
+
"language_model.model.layers.22.mlp.down_proj.qzeros": "model-00002-of-00002.safetensors",
|
| 630 |
+
"language_model.model.layers.22.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
|
| 631 |
"language_model.model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 632 |
"language_model.model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 633 |
"language_model.model.layers.23.pre_feedforward_layernorm.weight": "model-00002-of-00002.safetensors",
|