HamzaYous89 commited on
Commit
21387d8
·
verified ·
1 Parent(s): 99b3824

Upload folder using huggingface_hub

Browse files
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "FalconH1ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_in_multiplier": 1.0,
8
+ "attention_out_multiplier": 0.234375,
9
+ "attn_layer_indices": null,
10
+ "bos_token_id": 1,
11
+ "embedding_multiplier": 5.656854249492381,
12
+ "eos_token_id": 11,
13
+ "head_dim": 128,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 2048,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 4608,
18
+ "key_multiplier": 0.06905339660024876,
19
+ "lm_head_multiplier": 0.01953125,
20
+ "mamba_chunk_size": 128,
21
+ "mamba_conv_bias": true,
22
+ "mamba_d_conv": 4,
23
+ "mamba_d_head": 64,
24
+ "mamba_d_ssm": 3072,
25
+ "mamba_d_state": 256,
26
+ "mamba_expand": 2,
27
+ "mamba_n_groups": 1,
28
+ "mamba_n_heads": 48,
29
+ "mamba_norm_before_gate": false,
30
+ "mamba_proj_bias": false,
31
+ "mamba_rms_norm": true,
32
+ "mamba_use_mlp": true,
33
+ "max_position_embeddings": 8192,
34
+ "mlp_bias": false,
35
+ "mlp_expansion_factor": 8,
36
+ "mlp_multipliers": [
37
+ 0.4419417382415922,
38
+ 0.13020833333333331
39
+ ],
40
+ "model_type": "falcon_h1",
41
+ "num_attention_heads": 8,
42
+ "num_hidden_layers": 24,
43
+ "num_key_value_heads": 2,
44
+ "num_logits_to_keep": 1,
45
+ "pad_token_id": 0,
46
+ "projectors_bias": false,
47
+ "quantization_config": {
48
+ "bits": 8,
49
+ "checkpoint_format": "gptq",
50
+ "desc_act": true,
51
+ "group_size": 128,
52
+ "lm_head": false,
53
+ "meta": {
54
+ "damp_auto_increment": 0.0025,
55
+ "damp_percent": 0.01,
56
+ "mse": 0.0,
57
+ "quantizer": [
58
+ "gptqmodel:2.2.0"
59
+ ],
60
+ "static_groups": false,
61
+ "true_sequential": true,
62
+ "uri": "https://github.com/modelcloud/gptqmodel"
63
+ },
64
+ "pack_dtype": "int32",
65
+ "quant_method": "gptq",
66
+ "sym": true
67
+ },
68
+ "rms_norm_eps": 1e-05,
69
+ "rope_scaling": null,
70
+ "rope_theta": 100000000000.0,
71
+ "ssm_in_multiplier": 0.625,
72
+ "ssm_multipliers": [
73
+ 0.3535533905932738,
74
+ 0.25,
75
+ 0.1767766952966369,
76
+ 0.5,
77
+ 0.3535533905932738
78
+ ],
79
+ "ssm_out_multiplier": 0.11785113019775793,
80
+ "tie_word_embeddings": false,
81
+ "torch_dtype": "bfloat16",
82
+ "transformers_version": "4.52.0.dev0",
83
+ "use_cache": true,
84
+ "vocab_size": 65537
85
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 11,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.52.0.dev0"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:754e949880b1df05871d64ccbc5af22cde24599611adc4dd333b73164cf20362
3
+ size 2447081792
quant_log.csv ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,samples,damp,time
2
+ 0,feed_forward.gate_proj,1.30510879,0.01000,3.725
3
+ 0,feed_forward.up_proj,4.05003500,0.01000,0.510
4
+ 0,feed_forward.down_proj,8.25818729,0.01000,1.516
5
+ 1,feed_forward.gate_proj,1.29748273,0.01000,0.511
6
+ 1,feed_forward.up_proj,2.73292828,0.01000,0.511
7
+ 1,feed_forward.down_proj,5.14805508,0.01000,1.159
8
+ 2,feed_forward.gate_proj,1.23090601,0.01000,0.510
9
+ 2,feed_forward.up_proj,2.30279851,0.01000,0.511
10
+ 2,feed_forward.down_proj,4.61468744,0.01000,1.158
11
+ 3,feed_forward.gate_proj,1.10048580,0.01000,0.511
12
+ 3,feed_forward.up_proj,1.90120482,0.01000,0.511
13
+ 3,feed_forward.down_proj,11.11385250,0.01000,1.158
14
+ 4,feed_forward.gate_proj,0.91697407,0.01000,0.510
15
+ 4,feed_forward.up_proj,1.58906507,0.01000,0.510
16
+ 4,feed_forward.down_proj,3.70194459,0.01000,1.160
17
+ 5,feed_forward.gate_proj,1.04244685,0.01000,0.509
18
+ 5,feed_forward.up_proj,1.76986051,0.01000,0.511
19
+ 5,feed_forward.down_proj,3.52866268,0.01000,1.158
20
+ 6,feed_forward.gate_proj,0.96679544,0.01000,0.510
21
+ 6,feed_forward.up_proj,1.59820867,0.01000,0.510
22
+ 6,feed_forward.down_proj,3.57418442,0.01000,1.159
23
+ 7,feed_forward.gate_proj,0.93978858,0.01000,0.511
24
+ 7,feed_forward.up_proj,1.59313560,0.01000,0.511
25
+ 7,feed_forward.down_proj,3.60026312,0.01000,1.161
26
+ 8,feed_forward.gate_proj,1.01717091,0.01000,0.510
27
+ 8,feed_forward.up_proj,1.74156618,0.01000,0.511
28
+ 8,feed_forward.down_proj,4.19336700,0.01000,1.158
29
+ 9,feed_forward.gate_proj,1.10104227,0.01000,0.512
30
+ 9,feed_forward.up_proj,1.79372549,0.01000,0.511
31
+ 9,feed_forward.down_proj,4.49549913,0.01000,1.161
32
+ 10,feed_forward.gate_proj,1.24670470,0.01000,0.511
33
+ 10,feed_forward.up_proj,2.07013083,0.01000,0.512
34
+ 10,feed_forward.down_proj,5.10455418,0.01000,1.160
35
+ 11,feed_forward.gate_proj,1.36271381,0.01000,0.511
36
+ 11,feed_forward.up_proj,2.28341436,0.01000,0.511
37
+ 11,feed_forward.down_proj,7.72139645,0.01000,1.161
38
+ 12,feed_forward.gate_proj,1.60862815,0.01000,0.512
39
+ 12,feed_forward.up_proj,2.74706793,0.01000,0.513
40
+ 12,feed_forward.down_proj,9.77115631,0.01000,1.161
41
+ 13,feed_forward.gate_proj,2.02612519,0.01000,0.511
42
+ 13,feed_forward.up_proj,3.36923885,0.01000,0.513
43
+ 13,feed_forward.down_proj,15.91432858,0.01000,1.159
44
+ 14,feed_forward.gate_proj,2.29283261,0.01000,0.511
45
+ 14,feed_forward.up_proj,3.86089945,0.01000,0.511
46
+ 14,feed_forward.down_proj,18.98839760,0.01000,1.157
47
+ 15,feed_forward.gate_proj,2.88071609,0.01000,0.512
48
+ 15,feed_forward.up_proj,4.80165291,0.01000,0.512
49
+ 15,feed_forward.down_proj,24.14971542,0.01000,1.159
50
+ 16,feed_forward.gate_proj,3.48945332,0.01000,0.510
51
+ 16,feed_forward.up_proj,5.78722095,0.01000,0.509
52
+ 16,feed_forward.down_proj,32.97491074,0.01000,1.159
53
+ 17,feed_forward.gate_proj,4.11738968,0.01000,0.512
54
+ 17,feed_forward.up_proj,6.77789783,0.01000,0.511
55
+ 17,feed_forward.down_proj,44.23390961,0.01000,1.156
56
+ 18,feed_forward.gate_proj,5.10589027,0.01000,0.510
57
+ 18,feed_forward.up_proj,8.46793747,0.01000,0.510
58
+ 18,feed_forward.down_proj,69.04174805,0.01000,1.160
59
+ 19,feed_forward.gate_proj,5.87184620,0.01000,0.511
60
+ 19,feed_forward.up_proj,9.68615150,0.01000,0.512
61
+ 19,feed_forward.down_proj,98.21003723,0.01000,1.160
62
+ 20,feed_forward.gate_proj,6.33989811,0.01000,0.510
63
+ 20,feed_forward.up_proj,10.79653072,0.01000,0.511
64
+ 20,feed_forward.down_proj,133.90747070,0.01000,1.159
65
+ 21,feed_forward.gate_proj,7.71545410,0.01000,0.512
66
+ 21,feed_forward.up_proj,12.76975441,0.01000,0.512
67
+ 21,feed_forward.down_proj,161.60745239,0.01000,1.159
68
+ 22,feed_forward.gate_proj,9.65821552,0.01000,0.511
69
+ 22,feed_forward.up_proj,14.39665413,0.01000,0.511
70
+ 22,feed_forward.down_proj,238.05212402,0.01000,1.161
71
+ 23,feed_forward.gate_proj,16.67799759,0.01000,0.512
72
+ 23,feed_forward.up_proj,23.21625900,0.01000,0.511
73
+ 23,feed_forward.down_proj,444.00360107,0.01000,1.158
quantize_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 8,
3
+ "group_size": 128,
4
+ "desc_act": true,
5
+ "sym": true,
6
+ "lm_head": false,
7
+ "quant_method": "gptq",
8
+ "checkpoint_format": "gptq",
9
+ "pack_dtype": "int32",
10
+ "meta": {
11
+ "quantizer": [
12
+ "gptqmodel:2.2.0"
13
+ ],
14
+ "uri": "https://github.com/modelcloud/gptqmodel",
15
+ "damp_percent": 0.01,
16
+ "damp_auto_increment": 0.0025,
17
+ "static_groups": false,
18
+ "true_sequential": true,
19
+ "mse": 0.0
20
+ }
21
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,356 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|pad|>",
4
+ ">>ABSTRACT<<",
5
+ ">>INTRODUCTION<<",
6
+ ">>SUMMARY<<",
7
+ ">>COMMENT<<",
8
+ ">>ANSWER<<",
9
+ ">>QUESTION<<",
10
+ ">>DOMAIN<<",
11
+ ">>PREFIX<<",
12
+ ">>SUFFIX<<",
13
+ ">>MIDDLE<<",
14
+ "<|finetune_right_pad_id|>",
15
+ "<|start_header_id|>",
16
+ "<|end_header_id|>",
17
+ "<|eom_id|>",
18
+ "<|eot_id|>",
19
+ "<|begin_of_text|>",
20
+ ">>TITLE<<",
21
+ "<tool_response>",
22
+ "</tool_response>",
23
+ "<tool_call>",
24
+ "</tool_call>",
25
+ "<schema>",
26
+ "</schema>",
27
+ "<scratch_pad>",
28
+ "</scratch_pad>",
29
+ "<thinking>",
30
+ "</thinking>",
31
+ "<explanation>",
32
+ "</explanation>",
33
+ "<file_sep>",
34
+ "<repo_name>",
35
+ "<|im_end|>",
36
+ "<|im_start|>",
37
+ ">>UNUSED_119<<",
38
+ ">>UNUSED_120<<",
39
+ "<|system|>",
40
+ ">>UNUSED_218<<",
41
+ ">>UNUSED_219<<",
42
+ ">>UNUSED_220<<",
43
+ ">>UNUSED_221<<",
44
+ ">>UNUSED_222<<",
45
+ ">>UNUSED_223<<",
46
+ ">>UNUSED_224<<",
47
+ ">>UNUSED_225<<",
48
+ ">>UNUSED_226<<",
49
+ ">>UNUSED_227<<",
50
+ ">>UNUSED_228<<",
51
+ ">>UNUSED_229<<",
52
+ ">>UNUSED_230<<",
53
+ ">>UNUSED_231<<",
54
+ ">>UNUSED_232<<",
55
+ ">>UNUSED_233<<",
56
+ ">>UNUSED_234<<",
57
+ ">>UNUSED_235<<",
58
+ ">>UNUSED_236<<",
59
+ ">>UNUSED_237<<",
60
+ ">>UNUSED_238<<",
61
+ ">>UNUSED_239<<",
62
+ ">>UNUSED_240<<",
63
+ ">>UNUSED_241<<",
64
+ ">>UNUSED_242<<",
65
+ ">>UNUSED_243<<",
66
+ ">>UNUSED_244<<",
67
+ ">>UNUSED_245<<",
68
+ ">>UNUSED_246<<",
69
+ ">>UNUSED_247<<",
70
+ ">>UNUSED_248<<",
71
+ ">>UNUSED_249<<",
72
+ ">>UNUSED_250<<",
73
+ ">>UNUSED_251<<",
74
+ ">>UNUSED_252<<",
75
+ ">>UNUSED_253<<",
76
+ ">>UNUSED_254<<",
77
+ ">>UNUSED_255<<",
78
+ ">>UNUSED_256<<",
79
+ ">>UNUSED_257<<",
80
+ ">>UNUSED_258<<",
81
+ ">>UNUSED_259<<",
82
+ ">>UNUSED_260<<",
83
+ ">>UNUSED_261<<",
84
+ ">>UNUSED_262<<",
85
+ ">>UNUSED_263<<",
86
+ ">>UNUSED_264<<",
87
+ ">>UNUSED_265<<",
88
+ ">>UNUSED_266<<",
89
+ ">>UNUSED_267<<",
90
+ ">>UNUSED_268<<",
91
+ ">>UNUSED_269<<",
92
+ ">>UNUSED_270<<",
93
+ ">>UNUSED_271<<",
94
+ ">>UNUSED_272<<",
95
+ ">>UNUSED_273<<",
96
+ ">>UNUSED_274<<",
97
+ ">>UNUSED_275<<",
98
+ ">>UNUSED_276<<",
99
+ ">>UNUSED_277<<",
100
+ ">>UNUSED_278<<",
101
+ ">>UNUSED_279<<",
102
+ ">>UNUSED_280<<",
103
+ ">>UNUSED_281<<",
104
+ ">>UNUSED_282<<",
105
+ ">>UNUSED_283<<",
106
+ ">>UNUSED_284<<",
107
+ ">>UNUSED_285<<",
108
+ ">>UNUSED_286<<",
109
+ ">>UNUSED_287<<",
110
+ ">>UNUSED_288<<",
111
+ ">>UNUSED_289<<",
112
+ ">>UNUSED_290<<",
113
+ ">>UNUSED_291<<",
114
+ ">>UNUSED_292<<",
115
+ ">>UNUSED_293<<",
116
+ ">>UNUSED_294<<",
117
+ ">>UNUSED_295<<",
118
+ ">>UNUSED_296<<",
119
+ ">>UNUSED_297<<",
120
+ ">>UNUSED_298<<",
121
+ ">>UNUSED_299<<",
122
+ ">>UNUSED_300<<",
123
+ ">>UNUSED_301<<",
124
+ ">>UNUSED_302<<",
125
+ ">>UNUSED_303<<",
126
+ ">>UNUSED_304<<",
127
+ ">>UNUSED_305<<",
128
+ ">>UNUSED_306<<",
129
+ ">>UNUSED_307<<",
130
+ ">>UNUSED_308<<",
131
+ ">>UNUSED_309<<",
132
+ ">>UNUSED_310<<",
133
+ ">>UNUSED_311<<",
134
+ ">>UNUSED_312<<",
135
+ ">>UNUSED_313<<",
136
+ ">>UNUSED_314<<",
137
+ ">>UNUSED_315<<",
138
+ ">>UNUSED_316<<",
139
+ ">>UNUSED_317<<",
140
+ ">>UNUSED_318<<",
141
+ ">>UNUSED_319<<",
142
+ ">>UNUSED_320<<",
143
+ ">>UNUSED_321<<",
144
+ ">>UNUSED_322<<",
145
+ ">>UNUSED_323<<",
146
+ ">>UNUSED_324<<",
147
+ ">>UNUSED_325<<",
148
+ ">>UNUSED_326<<",
149
+ ">>UNUSED_327<<",
150
+ ">>UNUSED_328<<",
151
+ ">>UNUSED_329<<",
152
+ ">>UNUSED_330<<",
153
+ ">>UNUSED_331<<",
154
+ ">>UNUSED_332<<",
155
+ ">>UNUSED_333<<",
156
+ ">>UNUSED_334<<",
157
+ ">>UNUSED_335<<",
158
+ ">>UNUSED_336<<",
159
+ ">>UNUSED_337<<",
160
+ ">>UNUSED_338<<",
161
+ ">>UNUSED_339<<",
162
+ ">>UNUSED_340<<",
163
+ ">>UNUSED_341<<",
164
+ ">>UNUSED_342<<",
165
+ ">>UNUSED_343<<",
166
+ ">>UNUSED_344<<",
167
+ ">>UNUSED_345<<",
168
+ ">>UNUSED_346<<",
169
+ ">>UNUSED_347<<",
170
+ ">>UNUSED_348<<",
171
+ ">>UNUSED_349<<",
172
+ ">>UNUSED_350<<",
173
+ ">>UNUSED_351<<",
174
+ ">>UNUSED_352<<",
175
+ ">>UNUSED_353<<",
176
+ ">>UNUSED_354<<",
177
+ ">>UNUSED_355<<",
178
+ ">>UNUSED_356<<",
179
+ ">>UNUSED_357<<",
180
+ ">>UNUSED_358<<",
181
+ ">>UNUSED_359<<",
182
+ ">>UNUSED_360<<",
183
+ ">>UNUSED_361<<",
184
+ ">>UNUSED_362<<",
185
+ ">>UNUSED_363<<",
186
+ ">>UNUSED_364<<",
187
+ ">>UNUSED_365<<",
188
+ ">>UNUSED_366<<",
189
+ ">>UNUSED_367<<",
190
+ ">>UNUSED_368<<",
191
+ ">>UNUSED_369<<",
192
+ ">>UNUSED_370<<",
193
+ ">>UNUSED_371<<",
194
+ ">>UNUSED_372<<",
195
+ ">>UNUSED_373<<",
196
+ ">>UNUSED_374<<",
197
+ ">>UNUSED_375<<",
198
+ ">>UNUSED_376<<",
199
+ ">>UNUSED_377<<",
200
+ ">>UNUSED_378<<",
201
+ ">>UNUSED_379<<",
202
+ ">>UNUSED_380<<",
203
+ ">>UNUSED_381<<",
204
+ ">>UNUSED_382<<",
205
+ ">>UNUSED_383<<",
206
+ ">>UNUSED_384<<",
207
+ ">>UNUSED_385<<",
208
+ ">>UNUSED_386<<",
209
+ ">>UNUSED_387<<",
210
+ ">>UNUSED_388<<",
211
+ ">>UNUSED_389<<",
212
+ ">>UNUSED_390<<",
213
+ ">>UNUSED_391<<",
214
+ ">>UNUSED_392<<",
215
+ ">>UNUSED_393<<",
216
+ ">>UNUSED_394<<",
217
+ ">>UNUSED_395<<",
218
+ ">>UNUSED_396<<",
219
+ ">>UNUSED_397<<",
220
+ ">>UNUSED_398<<",
221
+ ">>UNUSED_399<<",
222
+ ">>UNUSED_400<<",
223
+ ">>UNUSED_401<<",
224
+ ">>UNUSED_402<<",
225
+ ">>UNUSED_403<<",
226
+ ">>UNUSED_404<<",
227
+ ">>UNUSED_405<<",
228
+ ">>UNUSED_406<<",
229
+ ">>UNUSED_407<<",
230
+ ">>UNUSED_408<<",
231
+ ">>UNUSED_409<<",
232
+ ">>UNUSED_410<<",
233
+ ">>UNUSED_411<<",
234
+ ">>UNUSED_412<<",
235
+ ">>UNUSED_413<<",
236
+ ">>UNUSED_414<<",
237
+ ">>UNUSED_415<<",
238
+ ">>UNUSED_416<<",
239
+ ">>UNUSED_417<<",
240
+ ">>UNUSED_418<<",
241
+ ">>UNUSED_419<<",
242
+ ">>UNUSED_420<<",
243
+ ">>UNUSED_421<<",
244
+ ">>UNUSED_422<<",
245
+ ">>UNUSED_423<<",
246
+ ">>UNUSED_424<<",
247
+ ">>UNUSED_425<<",
248
+ ">>UNUSED_426<<",
249
+ ">>UNUSED_427<<",
250
+ ">>UNUSED_428<<",
251
+ ">>UNUSED_429<<",
252
+ ">>UNUSED_430<<",
253
+ ">>UNUSED_431<<",
254
+ ">>UNUSED_432<<",
255
+ ">>UNUSED_433<<",
256
+ ">>UNUSED_434<<",
257
+ ">>UNUSED_435<<",
258
+ ">>UNUSED_436<<",
259
+ ">>UNUSED_437<<",
260
+ ">>UNUSED_438<<",
261
+ ">>UNUSED_439<<",
262
+ ">>UNUSED_440<<",
263
+ ">>UNUSED_441<<",
264
+ ">>UNUSED_442<<",
265
+ ">>UNUSED_443<<",
266
+ ">>UNUSED_444<<",
267
+ ">>UNUSED_445<<",
268
+ ">>UNUSED_446<<",
269
+ ">>UNUSED_447<<",
270
+ ">>UNUSED_448<<",
271
+ ">>UNUSED_449<<",
272
+ ">>UNUSED_450<<",
273
+ ">>UNUSED_451<<",
274
+ ">>UNUSED_452<<",
275
+ ">>UNUSED_453<<",
276
+ ">>UNUSED_454<<",
277
+ ">>UNUSED_455<<",
278
+ ">>UNUSED_456<<",
279
+ ">>UNUSED_457<<",
280
+ ">>UNUSED_458<<",
281
+ ">>UNUSED_459<<",
282
+ ">>UNUSED_460<<",
283
+ ">>UNUSED_461<<",
284
+ ">>UNUSED_462<<",
285
+ ">>UNUSED_463<<",
286
+ ">>UNUSED_464<<",
287
+ ">>UNUSED_465<<",
288
+ ">>UNUSED_466<<",
289
+ ">>UNUSED_467<<",
290
+ ">>UNUSED_468<<",
291
+ ">>UNUSED_469<<",
292
+ ">>UNUSED_470<<",
293
+ ">>UNUSED_471<<",
294
+ ">>UNUSED_472<<",
295
+ ">>UNUSED_473<<",
296
+ ">>UNUSED_474<<",
297
+ ">>UNUSED_475<<",
298
+ ">>UNUSED_476<<",
299
+ ">>UNUSED_477<<",
300
+ ">>UNUSED_478<<",
301
+ ">>UNUSED_479<<",
302
+ ">>UNUSED_480<<",
303
+ ">>UNUSED_481<<",
304
+ ">>UNUSED_482<<",
305
+ ">>UNUSED_483<<",
306
+ ">>UNUSED_484<<",
307
+ ">>UNUSED_485<<",
308
+ ">>UNUSED_486<<",
309
+ ">>UNUSED_487<<",
310
+ ">>UNUSED_488<<",
311
+ ">>UNUSED_489<<",
312
+ ">>UNUSED_490<<",
313
+ ">>UNUSED_491<<",
314
+ ">>UNUSED_492<<",
315
+ ">>UNUSED_493<<",
316
+ ">>UNUSED_494<<",
317
+ ">>UNUSED_495<<",
318
+ ">>UNUSED_496<<",
319
+ ">>UNUSED_497<<",
320
+ ">>UNUSED_498<<",
321
+ ">>UNUSED_499<<",
322
+ ">>UNUSED_500<<",
323
+ ">>UNUSED_501<<",
324
+ ">>UNUSED_502<<",
325
+ ">>UNUSED_503<<",
326
+ ">>UNUSED_504<<",
327
+ ">>UNUSED_505<<",
328
+ ">>UNUSED_506<<",
329
+ ">>UNUSED_507<<",
330
+ ">>UNUSED_508<<",
331
+ ">>UNUSED_509<<",
332
+ ">>UNUSED_510<<",
333
+ ">>UNUSED_511<<"
334
+ ],
335
+ "bos_token": {
336
+ "content": "<|begin_of_text|>",
337
+ "lstrip": false,
338
+ "normalized": false,
339
+ "rstrip": false,
340
+ "single_word": false
341
+ },
342
+ "eos_token": {
343
+ "content": "<|end_of_text|>",
344
+ "lstrip": false,
345
+ "normalized": false,
346
+ "rstrip": false,
347
+ "single_word": false
348
+ },
349
+ "pad_token": {
350
+ "content": "<pad>",
351
+ "lstrip": false,
352
+ "normalized": false,
353
+ "rstrip": false,
354
+ "single_word": false
355
+ }
356
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff