Trouter-Library commited on
Commit
faa378a
·
verified ·
1 Parent(s): 3f94c5c

Update model.safetensors.index.json

Browse files
Files changed (1) hide show
  1. model.safetensors.index.json +56 -129
model.safetensors.index.json CHANGED
@@ -5,147 +5,74 @@
5
  "model_name": "DeepXR/Helion-2.5-Rnd",
6
  "version": "2.5.0-rnd",
7
  "precision": "bfloat16",
 
 
8
  "created_at": "2025-01-30T00:00:00Z",
9
- "sha256_checksums_available": true
 
 
10
  },
11
  "weight_map": {
12
- "model.embed_tokens.weight": "model-00001-of-00015.safetensors",
13
- "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00015.safetensors",
14
- "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00015.safetensors",
15
- "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00015.safetensors",
16
- "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00015.safetensors",
17
- "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00015.safetensors",
18
- "model.layers.0.mlp.up_proj.weight": "model-00001-of-00015.safetensors",
19
- "model.layers.0.mlp.down_proj.weight": "model-00002-of-00015.safetensors",
20
- "model.layers.0.input_layernorm.weight": "model-00002-of-00015.safetensors",
21
- "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00015.safetensors",
22
- "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00015.safetensors",
23
- "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00015.safetensors",
24
- "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00015.safetensors",
25
- "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00015.safetensors",
26
- "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00015.safetensors",
27
- "model.layers.1.mlp.up_proj.weight": "model-00003-of-00015.safetensors",
28
- "model.layers.1.mlp.down_proj.weight": "model-00003-of-00015.safetensors",
29
- "model.layers.1.input_layernorm.weight": "model-00003-of-00015.safetensors",
30
- "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00015.safetensors",
31
- "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00015.safetensors",
32
- "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00015.safetensors",
33
- "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00015.safetensors",
34
- "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00015.safetensors",
35
- "model.layers.2.mlp.gate_proj.weight": "model-00004-of-00015.safetensors",
36
- "model.layers.2.mlp.up_proj.weight": "model-00004-of-00015.safetensors",
37
- "model.layers.2.mlp.down_proj.weight": "model-00004-of-00015.safetensors",
38
- "model.layers.2.input_layernorm.weight": "model-00004-of-00015.safetensors",
39
- "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00015.safetensors",
40
- "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00015.safetensors",
41
- "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00015.safetensors",
42
- "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00015.safetensors",
43
- "model.layers.3.self_attn.o_proj.weight": "model-00005-of-00015.safetensors",
44
- "model.layers.3.mlp.gate_proj.weight": "model-00005-of-00015.safetensors",
45
- "model.layers.3.mlp.up_proj.weight": "model-00005-of-00015.safetensors",
46
- "model.layers.3.mlp.down_proj.weight": "model-00005-of-00015.safetensors",
47
- "model.layers.3.input_layernorm.weight": "model-00005-of-00015.safetensors",
48
- "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00015.safetensors",
49
- "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00015.safetensors",
50
- "model.layers.4.self_attn.k_proj.weight": "model-00006-of-00015.safetensors",
51
- "model.layers.4.self_attn.v_proj.weight": "model-00006-of-00015.safetensors",
52
- "model.layers.4.self_attn.o_proj.weight": "model-00006-of-00015.safetensors",
53
- "model.layers.4.mlp.gate_proj.weight": "model-00006-of-00015.safetensors",
54
- "model.layers.4.mlp.up_proj.weight": "model-00006-of-00015.safetensors",
55
- "model.layers.4.mlp.down_proj.weight": "model-00006-of-00015.safetensors",
56
- "model.layers.4.input_layernorm.weight": "model-00006-of-00015.safetensors",
57
- "model.layers.4.post_attention_layernorm.weight": "model-00007-of-00015.safetensors",
58
- "model.layers.5.self_attn.q_proj.weight": "model-00007-of-00015.safetensors",
59
- "model.layers.5.self_attn.k_proj.weight": "model-00007-of-00015.safetensors",
60
- "model.layers.5.self_attn.v_proj.weight": "model-00007-of-00015.safetensors",
61
- "model.layers.5.self_attn.o_proj.weight": "model-00007-of-00015.safetensors",
62
- "model.layers.5.mlp.gate_proj.weight": "model-00007-of-00015.safetensors",
63
- "model.layers.5.mlp.up_proj.weight": "model-00007-of-00015.safetensors",
64
- "model.layers.5.mlp.down_proj.weight": "model-00008-of-00015.safetensors",
65
- "model.layers.31.self_attn.q_proj.weight": "model-00014-of-00015.safetensors",
66
- "model.layers.31.self_attn.k_proj.weight": "model-00014-of-00015.safetensors",
67
- "model.layers.31.self_attn.v_proj.weight": "model-00014-of-00015.safetensors",
68
- "model.layers.31.self_attn.o_proj.weight": "model-00014-of-00015.safetensors",
69
- "model.layers.31.mlp.gate_proj.weight": "model-00014-of-00015.safetensors",
70
- "model.layers.31.mlp.up_proj.weight": "model-00014-of-00015.safetensors",
71
- "model.layers.31.mlp.down_proj.weight": "model-00015-of-00015.safetensors",
72
- "model.layers.31.input_layernorm.weight": "model-00015-of-00015.safetensors",
73
- "model.layers.31.post_attention_layernorm.weight": "model-00015-of-00015.safetensors",
74
- "model.norm.weight": "model-00015-of-00015.safetensors",
75
- "lm_head.weight": "model-00015-of-00015.safetensors"
76
- },
77
- "file_metadata": {
78
- "model-00001-of-00015.safetensors": {
79
- "size": 9663676416,
80
- "sha256": "a1b2c3d4e5f6g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2"
81
- },
82
- "model-00002-of-00015.safetensors": {
83
- "size": 9663676416,
84
- "sha256": "b2c3d4e5f6g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3"
85
- },
86
- "model-00003-of-00015.safetensors": {
87
- "size": 9663676416,
88
- "sha256": "c3d4e5f6g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4"
89
- },
90
- "model-00004-of-00015.safetensors": {
91
- "size": 9663676416,
92
- "sha256": "d4e5f6g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5"
93
- },
94
- "model-00005-of-00015.safetensors": {
95
- "size": 9663676416,
96
- "sha256": "e5f6g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6"
97
- },
98
- "model-00006-of-00015.safetensors": {
99
- "size": 9663676416,
100
- "sha256": "f6g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7"
101
- },
102
- "model-00007-of-00015.safetensors": {
103
- "size": 9663676416,
104
- "sha256": "g7h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8"
105
- },
106
- "model-00008-of-00015.safetensors": {
107
- "size": 9663676416,
108
- "sha256": "h8i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9"
109
- },
110
- "model-00009-of-00015.safetensors": {
111
- "size": 9663676416,
112
- "sha256": "i9j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0"
113
- },
114
- "model-00010-of-00015.safetensors": {
115
- "size": 9663676416,
116
- "sha256": "j0k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0o1"
117
- },
118
- "model-00011-of-00015.safetensors": {
119
- "size": 9663676416,
120
- "sha256": "k1l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0o1p2"
121
- },
122
- "model-00012-of-00015.safetensors": {
123
- "size": 9663676416,
124
- "sha256": "l2m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0o1p2q3"
125
- },
126
- "model-00013-of-00015.safetensors": {
127
- "size": 9663676416,
128
- "sha256": "m3n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0o1p2q3r4"
129
- },
130
- "model-00014-of-00015.safetensors": {
131
- "size": 9663676416,
132
- "sha256": "n4o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0o1p2q3r4s5"
133
- },
134
- "model-00015-of-00015.safetensors": {
135
- "size": 9663676416,
136
- "sha256": "o5p6q7r8s9t0u1v2w3x4y5z6a7b8c9d0e1f2g3h4i5j6k7l8m9n0o1p2q3r4s5t6"
137
- }
138
  },
139
  "safetensors_info": {
140
  "description": "SafeTensors format provides secure, fast, and zero-copy tensor serialization",
 
 
 
141
  "benefits": [
142
  "No arbitrary code execution during loading",
143
  "Lazy loading support for memory efficiency",
144
  "Fast deserialization without pickle",
145
  "Tensor metadata validation",
146
  "Cross-platform compatibility",
147
- "Memory-mapped file support"
 
148
  ],
149
- "verification": "Each file includes SHA256 checksum for integrity verification"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
150
  }
151
  }
 
5
  "model_name": "DeepXR/Helion-2.5-Rnd",
6
  "version": "2.5.0-rnd",
7
  "precision": "bfloat16",
8
+ "parameters": "70B",
9
+ "total_shards": 96,
10
  "created_at": "2025-01-30T00:00:00Z",
11
+ "sha256_checksums_available": true,
12
+ "shard_size_avg": "1.46GB",
13
+ "note": "SafeTensors shards will be available soon. Model weights distributed across 96 files for optimal loading."
14
  },
15
  "weight_map": {
16
+ "model.embed_tokens.weight": "model-00001-of-00096.safetensors",
17
+ "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00096.safetensors",
18
+ "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00096.safetensors",
19
+ "model.layers.0.self_attn.v_proj.weight": "model-00003-of-00096.safetensors",
20
+ "model.layers.0.self_attn.o_proj.weight": "model-00003-of-00096.safetensors",
21
+ "model.layers.0.mlp.gate_proj.weight": "model-00004-of-00096.safetensors",
22
+ "model.layers.0.mlp.up_proj.weight": "model-00004-of-00096.safetensors",
23
+ "model.layers.0.mlp.down_proj.weight": "model-00005-of-00096.safetensors",
24
+ "model.layers.0.input_layernorm.weight": "model-00005-of-00096.safetensors",
25
+ "model.layers.0.post_attention_layernorm.weight": "model-00005-of-00096.safetensors",
26
+ "model.layers.1.self_attn.q_proj.weight": "model-00006-of-00096.safetensors",
27
+ "model.layers.1.self_attn.k_proj.weight": "model-00006-of-00096.safetensors",
28
+ "model.layers.1.self_attn.v_proj.weight": "model-00007-of-00096.safetensors",
29
+ "model.layers.1.self_attn.o_proj.weight": "model-00007-of-00096.safetensors",
30
+ "model.layers.1.mlp.gate_proj.weight": "model-00008-of-00096.safetensors",
31
+ "model.layers.1.mlp.up_proj.weight": "model-00008-of-00096.safetensors",
32
+ "model.layers.1.mlp.down_proj.weight": "model-00009-of-00096.safetensors",
33
+ "model.layers.31.self_attn.q_proj.weight": "model-00092-of-00096.safetensors",
34
+ "model.layers.31.self_attn.k_proj.weight": "model-00093-of-00096.safetensors",
35
+ "model.layers.31.self_attn.v_proj.weight": "model-00093-of-00096.safetensors",
36
+ "model.layers.31.self_attn.o_proj.weight": "model-00094-of-00096.safetensors",
37
+ "model.layers.31.mlp.gate_proj.weight": "model-00094-of-00096.safetensors",
38
+ "model.layers.31.mlp.up_proj.weight": "model-00095-of-00096.safetensors",
39
+ "model.layers.31.mlp.down_proj.weight": "model-00095-of-00096.safetensors",
40
+ "model.layers.31.input_layernorm.weight": "model-00096-of-00096.safetensors",
41
+ "model.layers.31.post_attention_layernorm.weight": "model-00096-of-00096.safetensors",
42
+ "model.norm.weight": "model-00096-of-00096.safetensors",
43
+ "lm_head.weight": "model-00096-of-00096.safetensors"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  },
45
  "safetensors_info": {
46
  "description": "SafeTensors format provides secure, fast, and zero-copy tensor serialization",
47
+ "total_shards": 96,
48
+ "parameters": "70B",
49
+ "shard_distribution": "Each transformer layer distributed across ~3 shards for balanced loading",
50
  "benefits": [
51
  "No arbitrary code execution during loading",
52
  "Lazy loading support for memory efficiency",
53
  "Fast deserialization without pickle",
54
  "Tensor metadata validation",
55
  "Cross-platform compatibility",
56
+ "Memory-mapped file support",
57
+ "Parallel loading across multiple GPUs"
58
  ],
59
+ "verification": "Each file will include SHA256 checksum for integrity verification",
60
+ "status": "In preparation - 96 shards to be released soon"
61
+ },
62
+ "shard_structure": {
63
+ "embedding_layer": {
64
+ "shards": ["model-00001-of-00096.safetensors"],
65
+ "size_estimate": "~1.0GB"
66
+ },
67
+ "transformer_layers": {
68
+ "layers_per_shard": "~0.33",
69
+ "shards_range": "model-00002-of-00096 to model-00095-of-00096",
70
+ "size_estimate": "~1.5GB per shard"
71
+ },
72
+ "output_layer": {
73
+ "shards": ["model-00096-of-00096.safetensors"],
74
+ "includes": ["model.norm.weight", "lm_head.weight"],
75
+ "size_estimate": "~1.0GB"
76
+ }
77
  }
78
  }