From 78933122c950c431b37c6fdeb14c1402ef22824f Mon Sep 17 00:00:00 2001 From: Ross Wightman Date: Fri, 27 Aug 2021 09:22:20 -0700 Subject: [PATCH] Fix silly typo --- timm/models/vision_transformer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/timm/models/vision_transformer.py b/timm/models/vision_transformer.py index de8248fe..ca8f52de 100644 --- a/timm/models/vision_transformer.py +++ b/timm/models/vision_transformer.py @@ -683,7 +683,7 @@ def vit_large_patch16_384(pretrained=False, **kwargs): def vit_base_patch16_sam_224(pretrained=False, **kwargs): """ ViT-Base (ViT-B/16) w/ SAM pretrained weights. Paper: https://arxiv.org/abs/2106.01548 """ - # NOTE original SAM weights releaes worked with representation_size=768 + # NOTE original SAM weights release worked with representation_size=768 model_kwargs = dict(patch_size=16, embed_dim=768, depth=12, num_heads=12, representation_size=0, **kwargs) model = _create_vision_transformer('vit_base_patch16_sam_224', pretrained=pretrained, **model_kwargs) return model @@ -693,7 +693,7 @@ def vit_base_patch16_sam_224(pretrained=False, **kwargs): def vit_base_patch32_sam_224(pretrained=False, **kwargs): """ ViT-Base (ViT-B/32) w/ SAM pretrained weights. Paper: https://arxiv.org/abs/2106.01548 """ - # NOTE original SAM weights releaes worked with representation_size=768 + # NOTE original SAM weights release worked with representation_size=768 model_kwargs = dict(patch_size=32, embed_dim=768, depth=12, num_heads=12, representation_size=0, **kwargs) model = _create_vision_transformer('vit_base_patch32_sam_224', pretrained=pretrained, **model_kwargs) return model