Bekhouche commited on
Commit
92b4acf
·
1 Parent(s): a381036

Add 16/09/2025 results

Browse files
Files changed (1) hide show
  1. imagenet_results.jsonl +33 -0
imagenet_results.jsonl CHANGED
@@ -28,3 +28,36 @@
28
  {"model": "microsoft/cvt-21-384", "top1_accuracy": 82.648, "top5_accuracy": 95.882, "parameters": 31622696, "flops": 24932082688.0, "inference_time": 18.59004497528076, "model_size": 126490784, "license": "Open"}
29
  {"model": "microsoft/cvt-21", "top1_accuracy": 81.54, "top5_accuracy": 95.438, "parameters": 31622696, "flops": 7206283808.0, "inference_time": 19.456958770751953, "model_size": 126490784, "license": "Open"}
30
  {"model": "microsoft/cvt-13-384", "top1_accuracy": 82.788, "top5_accuracy": 96.308, "parameters": 19997480, "flops": 16323617952.0, "inference_time": 16.565978527069092, "model_size": 79989920, "license": "Open"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  {"model": "microsoft/cvt-21-384", "top1_accuracy": 82.648, "top5_accuracy": 95.882, "parameters": 31622696, "flops": 24932082688.0, "inference_time": 18.59004497528076, "model_size": 126490784, "license": "Open"}
29
  {"model": "microsoft/cvt-21", "top1_accuracy": 81.54, "top5_accuracy": 95.438, "parameters": 31622696, "flops": 7206283808.0, "inference_time": 19.456958770751953, "model_size": 126490784, "license": "Open"}
30
  {"model": "microsoft/cvt-13-384", "top1_accuracy": 82.788, "top5_accuracy": 96.308, "parameters": 19997480, "flops": 16323617952.0, "inference_time": 16.565978527069092, "model_size": 79989920, "license": "Open"}
31
+ {"model": "facebook/deit-base-distilled-patch16-384", "top1_accuracy": 85.668, "top5_accuracy": 97.418, "parameters": 87630032, "flops": 49488993792, "inference_time": 14.616787433624268, "model_size": 350520128, "license": "Open"}
32
+ {"model": "facebook/deit-base-patch16-224", "top1_accuracy": 80.19800000000001, "top5_accuracy": 93.718, "parameters": 86567656, "flops": 16867412736, "inference_time": 7.529473304748535, "model_size": 346270624, "license": "Open"}
33
+ {"model": "facebook/deit-base-patch16-384", "top1_accuracy": 81.858, "top5_accuracy": 94.868, "parameters": 86859496, "flops": 49403195136, "inference_time": 14.441275596618652, "model_size": 347437984, "license": "Open"}
34
+ {"model": "facebook/deit-small-distilled-patch16-224", "top1_accuracy": 81.07799999999999, "top5_accuracy": 95.518, "parameters": 22436432, "flops": 4272340224, "inference_time": 7.621312141418457, "model_size": 89745728, "license": "Open"}
35
+ {"model": "facebook/deit-small-patch16-224", "top1_accuracy": 77.134, "top5_accuracy": 93.052, "parameters": 22050664, "flops": 4250674560, "inference_time": 7.204937934875488, "model_size": 88202656, "license": "Open"}
36
+ {"model": "facebook/deit-tiny-distilled-patch16-224", "top1_accuracy": 72.992, "top5_accuracy": 91.27, "parameters": 5910800, "flops": 1085103744, "inference_time": 4.983365535736084, "model_size": 23643200, "license": "Open"}
37
+ {"model": "facebook/deit-tiny-patch16-224", "top1_accuracy": 66.51, "top5_accuracy": 87.804, "parameters": 5717416, "flops": 1079579328, "inference_time": 4.877817630767822, "model_size": 22869664, "license": "Open"}
38
+ {"model": "facebook/regnet-x-002", "top1_accuracy": 68.796, "top5_accuracy": 88.542, "parameters": 2684792, "flops": 203384016, "inference_time": 4.118800163269043, "model_size": 10739168, "license": "Open"}
39
+ {"model": "facebook/regnet-x-004", "top1_accuracy": 72.37400000000001, "top5_accuracy": 90.826, "parameters": 5157512, "flops": 403980928, "inference_time": 6.920254230499268, "model_size": 20630048, "license": "Open"}
40
+ {"model": "facebook/regnet-x-006", "top1_accuracy": 73.80799999999999, "top5_accuracy": 91.7, "parameters": 6196040, "flops": 608864176, "inference_time": 5.397844314575195, "model_size": 24784160, "license": "Open"}
41
+ {"model": "facebook/regnet-x-008", "top1_accuracy": 75.02600000000001, "top5_accuracy": 92.31, "parameters": 7259656, "flops": 810028128, "inference_time": 5.5527567863464355, "model_size": 29038624, "license": "Open"}
42
+ {"model": "facebook/regnet-x-016", "top1_accuracy": 76.904, "top5_accuracy": 93.42399999999999, "parameters": 9190136, "flops": 1618759504, "inference_time": 9.21396017074585, "model_size": 36760544, "license": "Open"}
43
+ {"model": "facebook/regnet-x-032", "top1_accuracy": 78.13799999999999, "top5_accuracy": 94.05799999999999, "parameters": 15296552, "flops": 3199404208, "inference_time": 8.230078220367432, "model_size": 61186208, "license": "Open"}
44
+ {"model": "facebook/regnet-x-040", "top1_accuracy": 78.464, "top5_accuracy": 94.244, "parameters": 22118248, "flops": 3988966672, "inference_time": 8.696770668029785, "model_size": 88472992, "license": "Open"}
45
+ {"model": "facebook/regnet-x-064", "top1_accuracy": 79.048, "top5_accuracy": 94.45599999999999, "parameters": 26209256, "flops": 6493108328, "inference_time": 5.864083766937256, "model_size": 104837024, "license": "Open"}
46
+ {"model": "facebook/regnet-x-080", "top1_accuracy": 79.188, "top5_accuracy": 94.53399999999999, "parameters": 39572648, "flops": 8023343872, "inference_time": 8.043694496154785, "model_size": 158290592, "license": "Open"}
47
+ {"model": "facebook/regnet-x-120", "top1_accuracy": 79.584, "top5_accuracy": 94.742, "parameters": 46106056, "flops": 12129518912, "inference_time": 7.29440450668335, "model_size": 184424224, "license": "Open"}
48
+ {"model": "facebook/regnet-x-160", "top1_accuracy": 79.876, "top5_accuracy": 94.83200000000001, "parameters": 54278536, "flops": 15991884800, "inference_time": 12.249791622161865, "model_size": 217114144, "license": "Open"}
49
+ {"model": "facebook/regnet-x-320", "top1_accuracy": 80.264, "top5_accuracy": 95.018, "parameters": 107811560, "flops": 31808655592, "inference_time": 8.042430877685547, "model_size": 431246240, "license": "Open"}
50
+ {"model": "facebook/regnet-y-002", "top1_accuracy": 70.322, "top5_accuracy": 89.56, "parameters": 3162996, "flops": 204222784, "inference_time": 6.109774112701416, "model_size": 12651984, "license": "Open"}
51
+ {"model": "facebook/regnet-y-004", "top1_accuracy": 74.022, "top5_accuracy": 91.74, "parameters": 4344144, "flops": 410406088, "inference_time": 7.721090316772461, "model_size": 17376576, "license": "Open"}
52
+ {"model": "facebook/regnet-y-006", "top1_accuracy": 75.32600000000001, "top5_accuracy": 92.54400000000001, "parameters": 6055160, "flops": 610790496, "inference_time": 7.239985466003418, "model_size": 24220640, "license": "Open"}
53
+ {"model": "facebook/regnet-y-008", "top1_accuracy": 76.29599999999999, "top5_accuracy": 93.048, "parameters": 6263168, "flops": 808780032, "inference_time": 6.748080253601074, "model_size": 25052672, "license": "Open"}
54
+ {"model": "facebook/regnet-y-016", "top1_accuracy": 77.86, "top5_accuracy": 93.726, "parameters": 11202430, "flops": 1630259080, "inference_time": 13.444781303405762, "model_size": 44809720, "license": "Open"}
55
+ {"model": "facebook/regnet-y-032", "top1_accuracy": 82.006, "top5_accuracy": 95.894, "parameters": 19436338, "flops": 3201896896, "inference_time": 10.261929035186768, "model_size": 77745352, "license": "Open"}
56
+ {"model": "facebook/regnet-y-040", "top1_accuracy": 79.266, "top5_accuracy": 94.648, "parameters": 20646656, "flops": 4000907200, "inference_time": 10.614204406738281, "model_size": 82586624, "license": "Open"}
57
+ {"model": "facebook/regnet-y-064", "top1_accuracy": 79.694, "top5_accuracy": 94.782, "parameters": 30583252, "flops": 6392240368, "inference_time": 12.456858158111572, "model_size": 122333008, "license": "Open"}
58
+ {"model": "facebook/regnet-y-080", "top1_accuracy": 79.828, "top5_accuracy": 94.848, "parameters": 39180068, "flops": 8003373728, "inference_time": 8.556771278381348, "model_size": 156720272, "license": "Open"}
59
+ {"model": "facebook/regnet-y-120", "top1_accuracy": 80.388, "top5_accuracy": 95.14, "parameters": 51822544, "flops": 12140420096, "inference_time": 10.32639741897583, "model_size": 207290176, "license": "Open"}
60
+ {"model": "facebook/regnet-y-160", "top1_accuracy": 82.82000000000001, "top5_accuracy": 96.42399999999999, "parameters": 83590140, "flops": 15963307584, "inference_time": 10.361802577972412, "model_size": 334360560, "license": "Open"}
61
+ {"model": "facebook/regnet-y-320", "top1_accuracy": 80.814, "top5_accuracy": 95.218, "parameters": 145046770, "flops": 32347836096, "inference_time": 16.134166717529297, "model_size": 580187080, "license": "Open"}
62
+ {"model": "facebook/regnet-y-320-seer-in1k", "top1_accuracy": 83.314, "top5_accuracy": 96.732, "parameters": 145046770, "flops": 95027127616, "inference_time": 13.565516471862793, "model_size": 580187080, "license": "Open"}
63
+ {"model": "facebook/regnet-y-640-seer-in1k", "top1_accuracy": 83.83, "top5_accuracy": 96.87599999999999, "parameters": 281378786, "flops": 188502917952, "inference_time": 35.0980281829834, "model_size": 1125515144, "license": "Open"}