Shortcuts

mmselfsup.models.necks.relative_loc_neck 源代码

# Copyright (c) OpenMMLab. All rights reserved.
from typing import List, Optional, Union

import torch
import torch.nn as nn
from mmcv.cnn import build_norm_layer
from mmengine.model import BaseModule

from mmselfsup.registry import MODELS


[文档]@MODELS.register_module() class RelativeLocNeck(BaseModule): """The neck of relative patch location: fc-bn-relu-dropout. Args: in_channels (int): Number of input channels. out_channels (int): Number of output channels. with_avg_pool (bool): Whether to apply the global average pooling after backbone. Defaults to True. norm_cfg (dict): Dictionary to construct and config norm layer. Defaults to dict(type='BN1d'). init_cfg (dict or list[dict], optional): Initialization config dict. """ def __init__( self, in_channels: int, out_channels: int, with_avg_pool: bool = True, norm_cfg: dict = dict(type='BN1d'), init_cfg: Optional[Union[dict, List[dict]]] = [ dict(type='Normal', std=0.01, layer='Linear'), dict(type='Constant', val=1, layer=['_BatchNorm', 'GroupNorm']) ] ) -> None: super().__init__(init_cfg) self.with_avg_pool = with_avg_pool if with_avg_pool: self.avgpool = nn.AdaptiveAvgPool2d((1, 1)) self.fc = nn.Linear(in_channels * 2, out_channels) self.bn = build_norm_layer( dict(**norm_cfg, momentum=0.003), out_channels)[1] self.relu = nn.ReLU(inplace=True) self.dropout = nn.Dropout()
[文档] def forward(self, x: List[torch.Tensor]) -> List[torch.Tensor]: """Forward function. Args: x (List[torch.Tensor]): The feature map of backbone. Returns: List[torch.Tensor]: The output features. """ assert len(x) == 1 x = x[0] if self.with_avg_pool: x = self.avgpool(x) x = x.view(x.size(0), -1) x = self.fc(x) x = self.bn(x) x = self.relu(x) x = self.dropout(x) return [x]
Read the Docs v: latest
Versions
latest
stable
1.x
0.x
dev-1.x
Downloads
pdf
html
epub
On Read the Docs
Project Home
Builds

Free document hosting provided by Read the Docs.