From b2b6285af72ff0f06d033d1057192057180595ef Mon Sep 17 00:00:00 2001 From: Ross Wightman Date: Fri, 11 Nov 2022 00:30:17 -0800 Subject: [PATCH] Add two more FT clip weights --- timm/models/vision_transformer.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/timm/models/vision_transformer.py b/timm/models/vision_transformer.py index d4d091da..f29216c9 100644 --- a/timm/models/vision_transformer.py +++ b/timm/models/vision_transformer.py @@ -816,7 +816,7 @@ default_cfgs = generate_defaults({ hf_hub_id='timm/vit_base_patch16_clip_224.laion2b_ft_in12k_in1k', mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0), 'vit_base_patch16_clip_384.laion2b_ft_in12k_in1k': _cfg( - #hf_hub_id='timm/vit_base_patch16_clip_384.laion2b_ft_in12k_in1k', + hf_hub_id='timm/vit_base_patch16_clip_384.laion2b_ft_in12k_in1k', mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)), 'vit_large_patch14_clip_224.laion2b_ft_in12k_in1k': _cfg( hf_hub_id='timm/vit_large_patch14_clip_224.laion2b_ft_in12k_in1k', @@ -870,6 +870,9 @@ default_cfgs = generate_defaults({ 'vit_base_patch32_clip_224.openai_ft_in12k_in1k': _cfg( #hf_hub_id='timm/vit_base_patch32_clip_224.openai_ft_in12k_in1k', mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD), + 'vit_base_patch32_clip_384.openai_ft_in12k_in1k': _cfg( + hf_hub_id='timm/vit_base_patch32_clip_384.openai_ft_in12k_in1k', + mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, crop_pct=1.0, input_size=(3, 384, 384)), 'vit_base_patch16_clip_224.openai_ft_in12k_in1k': _cfg( #hf_hub_id='timm/vit_base_patch16_clip_224.openai_ft_in12k_in1k', mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD),