| |
| import torch |
| import torch.nn as nn |
| from mmcv.cnn import ConvModule |
| from mmengine.model import BaseModule |
| from mmengine.utils import digit_version, is_tuple_of |
| from torch import Tensor |
|
|
| from mmdet.utils import MultiConfig, OptConfigType, OptMultiConfig |
|
|
|
|
| class SELayer(BaseModule): |
| """Squeeze-and-Excitation Module. |
| |
| Args: |
| channels (int): The input (and output) channels of the SE layer. |
| ratio (int): Squeeze ratio in SELayer, the intermediate channel will be |
| ``int(channels/ratio)``. Defaults to 16. |
| conv_cfg (None or dict): Config dict for convolution layer. |
| Defaults to None, which means using conv2d. |
| act_cfg (dict or Sequence[dict]): Config dict for activation layer. |
| If act_cfg is a dict, two activation layers will be configurated |
| by this dict. If act_cfg is a sequence of dicts, the first |
| activation layer will be configurated by the first dict and the |
| second activation layer will be configurated by the second dict. |
| Defaults to (dict(type='ReLU'), dict(type='Sigmoid')) |
| init_cfg (dict or list[dict], optional): Initialization config dict. |
| Defaults to None |
| """ |
|
|
| def __init__(self, |
| channels: int, |
| ratio: int = 16, |
| conv_cfg: OptConfigType = None, |
| act_cfg: MultiConfig = (dict(type='ReLU'), |
| dict(type='Sigmoid')), |
| init_cfg: OptMultiConfig = None) -> None: |
| super().__init__(init_cfg=init_cfg) |
| if isinstance(act_cfg, dict): |
| act_cfg = (act_cfg, act_cfg) |
| assert len(act_cfg) == 2 |
| assert is_tuple_of(act_cfg, dict) |
| self.global_avgpool = nn.AdaptiveAvgPool2d(1) |
| self.conv1 = ConvModule( |
| in_channels=channels, |
| out_channels=int(channels / ratio), |
| kernel_size=1, |
| stride=1, |
| conv_cfg=conv_cfg, |
| act_cfg=act_cfg[0]) |
| self.conv2 = ConvModule( |
| in_channels=int(channels / ratio), |
| out_channels=channels, |
| kernel_size=1, |
| stride=1, |
| conv_cfg=conv_cfg, |
| act_cfg=act_cfg[1]) |
|
|
| def forward(self, x: Tensor) -> Tensor: |
| """Forward function for SELayer.""" |
| out = self.global_avgpool(x) |
| out = self.conv1(out) |
| out = self.conv2(out) |
| return x * out |
|
|
|
|
| class DyReLU(BaseModule): |
| """Dynamic ReLU (DyReLU) module. |
| |
| See `Dynamic ReLU <https://arxiv.org/abs/2003.10027>`_ for details. |
| Current implementation is specialized for task-aware attention in DyHead. |
| HSigmoid arguments in default act_cfg follow DyHead official code. |
| https://github.com/microsoft/DynamicHead/blob/master/dyhead/dyrelu.py |
| |
| Args: |
| channels (int): The input (and output) channels of DyReLU module. |
| ratio (int): Squeeze ratio in Squeeze-and-Excitation-like module, |
| the intermediate channel will be ``int(channels/ratio)``. |
| Defaults to 4. |
| conv_cfg (None or dict): Config dict for convolution layer. |
| Defaults to None, which means using conv2d. |
| act_cfg (dict or Sequence[dict]): Config dict for activation layer. |
| If act_cfg is a dict, two activation layers will be configurated |
| by this dict. If act_cfg is a sequence of dicts, the first |
| activation layer will be configurated by the first dict and the |
| second activation layer will be configurated by the second dict. |
| Defaults to (dict(type='ReLU'), dict(type='HSigmoid', bias=3.0, |
| divisor=6.0)) |
| init_cfg (dict or list[dict], optional): Initialization config dict. |
| Defaults to None |
| """ |
|
|
| def __init__(self, |
| channels: int, |
| ratio: int = 4, |
| conv_cfg: OptConfigType = None, |
| act_cfg: MultiConfig = (dict(type='ReLU'), |
| dict( |
| type='HSigmoid', |
| bias=3.0, |
| divisor=6.0)), |
| init_cfg: OptMultiConfig = None) -> None: |
| super().__init__(init_cfg=init_cfg) |
| if isinstance(act_cfg, dict): |
| act_cfg = (act_cfg, act_cfg) |
| assert len(act_cfg) == 2 |
| assert is_tuple_of(act_cfg, dict) |
| self.channels = channels |
| self.expansion = 4 |
| self.global_avgpool = nn.AdaptiveAvgPool2d(1) |
| self.conv1 = ConvModule( |
| in_channels=channels, |
| out_channels=int(channels / ratio), |
| kernel_size=1, |
| stride=1, |
| conv_cfg=conv_cfg, |
| act_cfg=act_cfg[0]) |
| self.conv2 = ConvModule( |
| in_channels=int(channels / ratio), |
| out_channels=channels * self.expansion, |
| kernel_size=1, |
| stride=1, |
| conv_cfg=conv_cfg, |
| act_cfg=act_cfg[1]) |
|
|
| def forward(self, x: Tensor) -> Tensor: |
| """Forward function.""" |
| coeffs = self.global_avgpool(x) |
| coeffs = self.conv1(coeffs) |
| coeffs = self.conv2(coeffs) - 0.5 |
| a1, b1, a2, b2 = torch.split(coeffs, self.channels, dim=1) |
| a1 = a1 * 2.0 + 1.0 |
| a2 = a2 * 2.0 |
| out = torch.max(x * a1 + b1, x * a2 + b2) |
| return out |
|
|
|
|
| class ChannelAttention(BaseModule): |
| """Channel attention Module. |
| |
| Args: |
| channels (int): The input (and output) channels of the attention layer. |
| init_cfg (dict or list[dict], optional): Initialization config dict. |
| Defaults to None |
| """ |
|
|
| def __init__(self, channels: int, init_cfg: OptMultiConfig = None) -> None: |
| super().__init__(init_cfg=init_cfg) |
| self.global_avgpool = nn.AdaptiveAvgPool2d(1) |
| self.fc = nn.Conv2d(channels, channels, 1, 1, 0, bias=True) |
| if digit_version(torch.__version__) < (1, 7, 0): |
| self.act = nn.Hardsigmoid() |
| else: |
| self.act = nn.Hardsigmoid(inplace=True) |
|
|
| def forward(self, x: Tensor) -> Tensor: |
| """Forward function for ChannelAttention.""" |
| with torch.cuda.amp.autocast(enabled=False): |
| out = self.global_avgpool(x) |
| out = self.fc(out) |
| out = self.act(out) |
| return x * out |
|
|