You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
pytorch-image-models/timm/models/layers
Ross Wightman bda8ab015a
Remove min channels for SelectiveKernel, divisor should cover cases well enough.
4 years ago
..
__init__.py Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA. 4 years ago
activations.py
activations_jit.py
activations_me.py
adaptive_avgmax_pool.py
blur_pool.py
bottleneck_attn.py
cbam.py Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy. 4 years ago
classifier.py
cond_conv2d.py
config.py
conv2d_same.py
conv_bn_act.py
create_act.py
create_attn.py Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA. 4 years ago
create_conv2d.py
create_norm_act.py
drop.py
eca.py Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA. 4 years ago
evo_norm.py
gather_excite.py Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy. 4 years ago
global_context.py Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy. 4 years ago
halo_attn.py
helpers.py Throw in some FBNetV3 code I had lying around, some refactoring of SE reduction channel calcs for all EffNet archs. 4 years ago
inplace_abn.py
involution.py Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy. 4 years ago
lambda_layer.py
linear.py
median_pool.py
mixed_conv2d.py
mlp.py Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy. 4 years ago
non_local_attn.py Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA. 4 years ago
norm.py Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy. 4 years ago
norm_act.py
padding.py
patch_embed.py
pool2d_same.py
selective_kernel.py Remove min channels for SelectiveKernel, divisor should cover cases well enough. 4 years ago
separable_conv.py
space_to_depth.py
split_attn.py Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA. 4 years ago
split_batchnorm.py
squeeze_excite.py Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA. 4 years ago
std_conv.py
swin_attn.py
test_time_pool.py
weight_init.py