From e04400dc780299ec34adbb94d4f26beb5e2c382d Mon Sep 17 00:00:00 2001 From: Ross Wightman Date: Wed, 15 Feb 2023 15:38:41 -0800 Subject: [PATCH] Remove dead line --- timm/models/vision_transformer.py | 1 - 1 file changed, 1 deletion(-) diff --git a/timm/models/vision_transformer.py b/timm/models/vision_transformer.py index 477b4f7b..34f6a063 100644 --- a/timm/models/vision_transformer.py +++ b/timm/models/vision_transformer.py @@ -241,7 +241,6 @@ class ParallelScalingBlock(nn.Module): self.fast_attn = hasattr(torch.nn.functional, 'scaled_dot_product_attention') # FIXME mlp_hidden_dim = int(mlp_ratio * dim) in_proj_out_dim = mlp_hidden_dim + 3 * dim - out_proj_in_dim = mlp_hidden_dim + dim self.in_norm = norm_layer(dim) self.in_proj = nn.Linear(dim, in_proj_out_dim, bias=qkv_bias)