Add fine-tuned B/16 224x224 in1k clip models

pull/1520/head
Ross Wightman 2 years ago
parent d3415e3134
commit 092287436e

@ -761,7 +761,7 @@ default_cfgs = generate_defaults({
'vit_base_patch16_clip_224.laion2b': _cfg( 'vit_base_patch16_clip_224.laion2b': _cfg(
#hf_hub_id='laion/CLIP-ViT-B-16-laion2B-s34B-b88K', #hf_hub_id='laion/CLIP-ViT-B-16-laion2B-s34B-b88K',
hf_hub_filename='open_clip_pytorch_model.bin', hf_hub_filename='open_clip_pytorch_model.bin',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, num_classes=512), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, num_classes=512),
'vit_large_patch14_clip_224.laion2b': _cfg( 'vit_large_patch14_clip_224.laion2b': _cfg(
hf_hub_id='laion/CLIP-ViT-L-14-laion2B-s32B-b82K', hf_hub_id='laion/CLIP-ViT-L-14-laion2B-s32B-b82K',
hf_hub_filename='open_clip_pytorch_model.bin', hf_hub_filename='open_clip_pytorch_model.bin',
@ -782,8 +782,8 @@ default_cfgs = generate_defaults({
hf_hub_id='timm/vit_base_patch32_clip_384.laion2b_ft_in1k', hf_hub_id='timm/vit_base_patch32_clip_384.laion2b_ft_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)),
'vit_base_patch16_clip_224.laion2b_ft_in1k': _cfg( 'vit_base_patch16_clip_224.laion2b_ft_in1k': _cfg(
#hf_hub_id='timm/vit_base_patch16_clip_224.laion2b_ft_in1k', hf_hub_id='timm/vit_base_patch16_clip_224.laion2b_ft_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0),
'vit_base_patch16_clip_384.laion2b_ft_in1k': _cfg( 'vit_base_patch16_clip_384.laion2b_ft_in1k': _cfg(
#hf_hub_id='timm/vit_base_patch16_clip_384.laion2b_ft_in1k', #hf_hub_id='timm/vit_base_patch16_clip_384.laion2b_ft_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)),
@ -814,7 +814,7 @@ default_cfgs = generate_defaults({
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 448, 448)), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 448, 448)),
'vit_base_patch16_clip_224.laion2b_ft_in12k_in1k': _cfg( 'vit_base_patch16_clip_224.laion2b_ft_in12k_in1k': _cfg(
#hf_hub_id='timm/vit_base_patch16_clip_224.laion2b_ft_in12k_in1k', #hf_hub_id='timm/vit_base_patch16_clip_224.laion2b_ft_in12k_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0),
'vit_base_patch16_clip_384.laion2b_ft_in12k_in1k': _cfg( 'vit_base_patch16_clip_384.laion2b_ft_in12k_in1k': _cfg(
#hf_hub_id='timm/vit_base_patch16_clip_384.laion2b_ft_in12k_in1k', #hf_hub_id='timm/vit_base_patch16_clip_384.laion2b_ft_in12k_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)),
@ -858,7 +858,7 @@ default_cfgs = generate_defaults({
#hf_hub_id='timm/vit_base_patch32_clip_224.openai_ft_in1k', #hf_hub_id='timm/vit_base_patch32_clip_224.openai_ft_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD),
'vit_base_patch16_clip_224.openai_ft_in1k': _cfg( 'vit_base_patch16_clip_224.openai_ft_in1k': _cfg(
#hf_hub_id='timm/vit_base_patch16_clip_224.openai_ft_in1k', hf_hub_id='timm/vit_base_patch16_clip_224.openai_ft_in1k',
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD), mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD),
'vit_large_patch14_clip_224.openai_ft_in1k': _cfg( 'vit_large_patch14_clip_224.openai_ft_in1k': _cfg(
hf_hub_id='timm/vit_large_patch14_clip_224.openai_ft_in1k', hf_hub_id='timm/vit_large_patch14_clip_224.openai_ft_in1k',

Loading…
Cancel
Save