Xenova HF Staff commited on
Commit
e2b9c0f
·
verified ·
1 Parent(s): f8e3bbd

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +345 -0
  2. onnx/model.onnx +3 -0
  3. preprocessor_config.json +26 -0
config.json ADDED
@@ -0,0 +1,345 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "silu",
5
+ "anchor_image_size": null,
6
+ "architectures": [
7
+ "DFineForObjectDetection"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "auxiliary_loss": true,
11
+ "backbone": null,
12
+ "backbone_config": {
13
+ "depths": [
14
+ 3,
15
+ 4,
16
+ 6,
17
+ 3
18
+ ],
19
+ "downsample_in_bottleneck": false,
20
+ "downsample_in_first_stage": false,
21
+ "embedding_size": 32,
22
+ "hidden_act": "relu",
23
+ "hidden_sizes": [
24
+ 128,
25
+ 256,
26
+ 512,
27
+ 1024
28
+ ],
29
+ "initializer_range": 0.02,
30
+ "layer_type": "basic",
31
+ "model_type": "hgnet_v2",
32
+ "num_channels": 3,
33
+ "out_features": [
34
+ "stage3",
35
+ "stage4"
36
+ ],
37
+ "out_indices": [
38
+ 3,
39
+ 4
40
+ ],
41
+ "stage_downsample": [
42
+ false,
43
+ true,
44
+ true,
45
+ true
46
+ ],
47
+ "stage_in_channels": [
48
+ 16,
49
+ 64,
50
+ 256,
51
+ 512
52
+ ],
53
+ "stage_kernel_size": [
54
+ 3,
55
+ 3,
56
+ 5,
57
+ 5
58
+ ],
59
+ "stage_light_block": [
60
+ false,
61
+ false,
62
+ true,
63
+ true
64
+ ],
65
+ "stage_mid_channels": [
66
+ 16,
67
+ 32,
68
+ 64,
69
+ 128
70
+ ],
71
+ "stage_names": [
72
+ "stem",
73
+ "stage1",
74
+ "stage2",
75
+ "stage3",
76
+ "stage4"
77
+ ],
78
+ "stage_num_blocks": [
79
+ 1,
80
+ 1,
81
+ 2,
82
+ 1
83
+ ],
84
+ "stage_numb_of_layers": [
85
+ 3,
86
+ 3,
87
+ 3,
88
+ 3
89
+ ],
90
+ "stage_out_channels": [
91
+ 64,
92
+ 256,
93
+ 512,
94
+ 1024
95
+ ],
96
+ "stem_channels": [
97
+ 3,
98
+ 16,
99
+ 16
100
+ ],
101
+ "use_learnable_affine_block": true
102
+ },
103
+ "backbone_kwargs": null,
104
+ "batch_norm_eps": 1e-05,
105
+ "box_noise_scale": 1.0,
106
+ "d_model": 128,
107
+ "decoder_activation_function": "relu",
108
+ "decoder_attention_heads": 8,
109
+ "decoder_ffn_dim": 512,
110
+ "decoder_in_channels": [
111
+ 128,
112
+ 128
113
+ ],
114
+ "decoder_layers": 3,
115
+ "decoder_method": "default",
116
+ "decoder_n_points": [
117
+ 6,
118
+ 6
119
+ ],
120
+ "decoder_offset_scale": 0.5,
121
+ "depth_mult": 0.5,
122
+ "dropout": 0.0,
123
+ "encode_proj_layers": [
124
+ 1
125
+ ],
126
+ "encoder_activation_function": "gelu",
127
+ "encoder_attention_heads": 8,
128
+ "encoder_ffn_dim": 512,
129
+ "encoder_hidden_dim": 128,
130
+ "encoder_in_channels": [
131
+ 512,
132
+ 1024
133
+ ],
134
+ "encoder_layers": 1,
135
+ "eos_coefficient": 0.0001,
136
+ "eval_idx": -1,
137
+ "eval_size": null,
138
+ "feat_strides": [
139
+ 16,
140
+ 32
141
+ ],
142
+ "focal_loss_alpha": 0.75,
143
+ "focal_loss_gamma": 2.0,
144
+ "freeze_backbone_batch_norms": true,
145
+ "hidden_expansion": 0.34,
146
+ "id2label": {
147
+ "0": "person",
148
+ "1": "bicycle",
149
+ "2": "car",
150
+ "3": "motorbike",
151
+ "4": "aeroplane",
152
+ "5": "bus",
153
+ "6": "train",
154
+ "7": "truck",
155
+ "8": "boat",
156
+ "9": "traffic light",
157
+ "10": "fire hydrant",
158
+ "11": "stop sign",
159
+ "12": "parking meter",
160
+ "13": "bench",
161
+ "14": "bird",
162
+ "15": "cat",
163
+ "16": "dog",
164
+ "17": "horse",
165
+ "18": "sheep",
166
+ "19": "cow",
167
+ "20": "elephant",
168
+ "21": "bear",
169
+ "22": "zebra",
170
+ "23": "giraffe",
171
+ "24": "backpack",
172
+ "25": "umbrella",
173
+ "26": "handbag",
174
+ "27": "tie",
175
+ "28": "suitcase",
176
+ "29": "frisbee",
177
+ "30": "skis",
178
+ "31": "snowboard",
179
+ "32": "sports ball",
180
+ "33": "kite",
181
+ "34": "baseball bat",
182
+ "35": "baseball glove",
183
+ "36": "skateboard",
184
+ "37": "surfboard",
185
+ "38": "tennis racket",
186
+ "39": "bottle",
187
+ "40": "wine glass",
188
+ "41": "cup",
189
+ "42": "fork",
190
+ "43": "knife",
191
+ "44": "spoon",
192
+ "45": "bowl",
193
+ "46": "banana",
194
+ "47": "apple",
195
+ "48": "sandwich",
196
+ "49": "orange",
197
+ "50": "broccoli",
198
+ "51": "carrot",
199
+ "52": "hot dog",
200
+ "53": "pizza",
201
+ "54": "donut",
202
+ "55": "cake",
203
+ "56": "chair",
204
+ "57": "sofa",
205
+ "58": "pottedplant",
206
+ "59": "bed",
207
+ "60": "diningtable",
208
+ "61": "toilet",
209
+ "62": "tvmonitor",
210
+ "63": "laptop",
211
+ "64": "mouse",
212
+ "65": "remote",
213
+ "66": "keyboard",
214
+ "67": "cell phone",
215
+ "68": "microwave",
216
+ "69": "oven",
217
+ "70": "toaster",
218
+ "71": "sink",
219
+ "72": "refrigerator",
220
+ "73": "book",
221
+ "74": "clock",
222
+ "75": "vase",
223
+ "76": "scissors",
224
+ "77": "teddy bear",
225
+ "78": "hair drier",
226
+ "79": "toothbrush"
227
+ },
228
+ "initializer_bias_prior_prob": null,
229
+ "initializer_range": 0.01,
230
+ "is_encoder_decoder": true,
231
+ "label2id": {
232
+ "aeroplane": 4,
233
+ "apple": 47,
234
+ "backpack": 24,
235
+ "banana": 46,
236
+ "baseball bat": 34,
237
+ "baseball glove": 35,
238
+ "bear": 21,
239
+ "bed": 59,
240
+ "bench": 13,
241
+ "bicycle": 1,
242
+ "bird": 14,
243
+ "boat": 8,
244
+ "book": 73,
245
+ "bottle": 39,
246
+ "bowl": 45,
247
+ "broccoli": 50,
248
+ "bus": 5,
249
+ "cake": 55,
250
+ "car": 2,
251
+ "carrot": 51,
252
+ "cat": 15,
253
+ "cell phone": 67,
254
+ "chair": 56,
255
+ "clock": 74,
256
+ "cow": 19,
257
+ "cup": 41,
258
+ "diningtable": 60,
259
+ "dog": 16,
260
+ "donut": 54,
261
+ "elephant": 20,
262
+ "fire hydrant": 10,
263
+ "fork": 42,
264
+ "frisbee": 29,
265
+ "giraffe": 23,
266
+ "hair drier": 78,
267
+ "handbag": 26,
268
+ "horse": 17,
269
+ "hot dog": 52,
270
+ "keyboard": 66,
271
+ "kite": 33,
272
+ "knife": 43,
273
+ "laptop": 63,
274
+ "microwave": 68,
275
+ "motorbike": 3,
276
+ "mouse": 64,
277
+ "orange": 49,
278
+ "oven": 69,
279
+ "parking meter": 12,
280
+ "person": 0,
281
+ "pizza": 53,
282
+ "pottedplant": 58,
283
+ "refrigerator": 72,
284
+ "remote": 65,
285
+ "sandwich": 48,
286
+ "scissors": 76,
287
+ "sheep": 18,
288
+ "sink": 71,
289
+ "skateboard": 36,
290
+ "skis": 30,
291
+ "snowboard": 31,
292
+ "sofa": 57,
293
+ "spoon": 44,
294
+ "sports ball": 32,
295
+ "stop sign": 11,
296
+ "suitcase": 28,
297
+ "surfboard": 37,
298
+ "teddy bear": 77,
299
+ "tennis racket": 38,
300
+ "tie": 27,
301
+ "toaster": 70,
302
+ "toilet": 61,
303
+ "toothbrush": 79,
304
+ "traffic light": 9,
305
+ "train": 6,
306
+ "truck": 7,
307
+ "tvmonitor": 62,
308
+ "umbrella": 25,
309
+ "vase": 75,
310
+ "wine glass": 40,
311
+ "zebra": 22
312
+ },
313
+ "label_noise_ratio": 0.5,
314
+ "layer_norm_eps": 1e-05,
315
+ "layer_scale": 1,
316
+ "learn_initial_query": false,
317
+ "lqe_hidden_dim": 64,
318
+ "lqe_layers": 2,
319
+ "matcher_alpha": 0.25,
320
+ "matcher_bbox_cost": 5.0,
321
+ "matcher_class_cost": 2.0,
322
+ "matcher_gamma": 2.0,
323
+ "matcher_giou_cost": 2.0,
324
+ "max_num_bins": 32,
325
+ "model_type": "d_fine",
326
+ "normalize_before": false,
327
+ "num_denoising": 100,
328
+ "num_feature_levels": 2,
329
+ "num_queries": 300,
330
+ "positional_encoding_temperature": 10000,
331
+ "reg_scale": 4.0,
332
+ "top_prob_values": 4,
333
+ "torch_dtype": "float32",
334
+ "transformers_version": "4.52.0.dev0",
335
+ "up": 0.5,
336
+ "use_focal_loss": true,
337
+ "use_pretrained_backbone": false,
338
+ "use_timm_backbone": false,
339
+ "weight_loss_bbox": 5.0,
340
+ "weight_loss_ddf": 1.5,
341
+ "weight_loss_fgl": 0.15,
342
+ "weight_loss_giou": 2.0,
343
+ "weight_loss_vfl": 1.0,
344
+ "with_box_refine": true
345
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f684f409618ee8a822410e754a29caa817d1aa16283ce89cad936d0a48e2f35
3
+ size 15258358
preprocessor_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_annotations": true,
3
+ "do_normalize": false,
4
+ "do_pad": false,
5
+ "do_rescale": true,
6
+ "do_resize": true,
7
+ "format": "coco_detection",
8
+ "image_mean": [
9
+ 0.485,
10
+ 0.456,
11
+ 0.406
12
+ ],
13
+ "image_processor_type": "RTDetrImageProcessor",
14
+ "image_std": [
15
+ 0.229,
16
+ 0.224,
17
+ 0.225
18
+ ],
19
+ "pad_size": null,
20
+ "resample": 2,
21
+ "rescale_factor": 0.00392156862745098,
22
+ "size": {
23
+ "height": 640,
24
+ "width": 640
25
+ }
26
+ }