@@ -2282,107 +2282,91 @@ def _cfgr(url='', **kwargs):
2282
2282
# original attention pool head variants
2283
2283
'resnet50_clip.openai' : _cfgr (
2284
2284
hf_hub_id = 'timm/' ,
2285
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2286
2285
num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2287
2286
fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2288
2287
classifier = 'head.proj' ,
2289
2288
),
2290
2289
'resnet101_clip.openai' : _cfgr (
2291
2290
hf_hub_id = 'timm/' ,
2292
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2293
2291
num_classes = 512 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2294
2292
fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2295
2293
classifier = 'head.proj' ,
2296
2294
),
2297
2295
'resnet50x4_clip.openai' : _cfgr (
2298
2296
hf_hub_id = 'timm/' ,
2299
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2300
2297
num_classes = 640 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2301
2298
fixed_input_size = True , input_size = (3 , 288 , 288 ), pool_size = (9 , 9 ),
2302
2299
classifier = 'head.proj' ,
2303
2300
),
2304
2301
'resnet50x16_clip.openai' : _cfgr (
2305
2302
hf_hub_id = 'timm/' ,
2306
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2307
2303
num_classes = 768 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2308
2304
fixed_input_size = True , input_size = (3 , 384 , 384 ), pool_size = (12 , 12 ),
2309
2305
classifier = 'head.proj' ,
2310
2306
),
2311
2307
'resnet50x64_clip.openai' : _cfgr (
2312
2308
hf_hub_id = 'timm/' ,
2313
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2314
2309
num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2315
2310
fixed_input_size = True , input_size = (3 , 448 , 448 ), pool_size = (14 , 14 ),
2316
2311
classifier = 'head.proj' ,
2317
2312
),
2318
2313
'resnet50_clip.cc12m' : _cfgr (
2319
2314
hf_hub_id = 'timm/' ,
2320
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2321
2315
num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2322
2316
fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2323
2317
classifier = 'head.proj' ,
2324
2318
),
2325
2319
'resnet50_clip.yfcc15m' : _cfgr (
2326
2320
hf_hub_id = 'timm/' ,
2327
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2328
2321
num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2329
2322
fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2330
2323
classifier = 'head.proj' ,
2331
2324
),
2332
2325
'resnet101_clip.yfcc15m' : _cfgr (
2333
2326
hf_hub_id = 'timm/' ,
2334
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2335
2327
num_classes = 512 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2336
2328
fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2337
2329
classifier = 'head.proj' ,
2338
2330
),
2339
2331
2340
2332
# avg-pool w/ optional standard classifier head variants
2341
2333
'resnet50_clip_gap.openai' : _cfgr (
2342
- hf_hub_id = 'timm/resnet50_clip.openai' ,
2343
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2334
+ hf_hub_id = 'timm/' ,
2344
2335
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2345
2336
input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2346
2337
),
2347
2338
'resnet101_clip_gap.openai' : _cfgr (
2348
- hf_hub_id = 'timm/resnet101_clip.openai' ,
2349
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2339
+ hf_hub_id = 'timm/' ,
2350
2340
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2351
2341
input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2352
2342
),
2353
2343
'resnet50x4_clip_gap.openai' : _cfgr (
2354
- hf_hub_id = 'timm/resnet50x4_clip.openai' ,
2355
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2344
+ hf_hub_id = 'timm/' ,
2356
2345
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2357
2346
input_size = (3 , 288 , 288 ), pool_size = (9 , 9 ),
2358
2347
),
2359
2348
'resnet50x16_clip_gap.openai' : _cfgr (
2360
- hf_hub_id = 'timm/resnet50x16_clip.openai' ,
2361
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2349
+ hf_hub_id = 'timm/' ,
2362
2350
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2363
2351
input_size = (3 , 384 , 384 ), pool_size = (12 , 12 ),
2364
2352
),
2365
2353
'resnet50x64_clip_gap.openai' : _cfgr (
2366
- hf_hub_id = 'timm/resnet50x64_clip.openai' ,
2367
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2354
+ hf_hub_id = 'timm/' ,
2368
2355
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2369
2356
input_size = (3 , 448 , 448 ), pool_size = (14 , 14 ),
2370
2357
),
2371
2358
'resnet50_clip_gap.cc12m' : _cfgr (
2372
- hf_hub_id = 'timm/resnet50_clip.cc12m' ,
2373
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2359
+ hf_hub_id = 'timm/' ,
2374
2360
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2375
2361
input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2376
2362
),
2377
2363
'resnet50_clip_gap.yfcc15m' : _cfgr (
2378
- hf_hub_id = 'timm/resnet50_clip.yfcc15m' ,
2379
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2364
+ hf_hub_id = 'timm/' ,
2380
2365
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2381
2366
input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2382
2367
),
2383
2368
'resnet101_clip_gap.yfcc15m' : _cfgr (
2384
- hf_hub_id = 'timm/resnet101_clip.yfcc15m' ,
2385
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2369
+ hf_hub_id = 'timm/' ,
2386
2370
num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
2387
2371
input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
2388
2372
),
0 commit comments