| from typing import List |
|
|
| import torch |
| from torch import nn as nn |
|
|
| from annotator.mmpkg.mmcv.runner import force_fp32 |
| from .furthest_point_sample import (furthest_point_sample, |
| furthest_point_sample_with_dist) |
|
|
|
|
| def calc_square_dist(point_feat_a, point_feat_b, norm=True): |
| """Calculating square distance between a and b. |
| |
| Args: |
| point_feat_a (Tensor): (B, N, C) Feature vector of each point. |
| point_feat_b (Tensor): (B, M, C) Feature vector of each point. |
| norm (Bool, optional): Whether to normalize the distance. |
| Default: True. |
| |
| Returns: |
| Tensor: (B, N, M) Distance between each pair points. |
| """ |
| num_channel = point_feat_a.shape[-1] |
| |
| a_square = torch.sum(point_feat_a.unsqueeze(dim=2).pow(2), dim=-1) |
| |
| b_square = torch.sum(point_feat_b.unsqueeze(dim=1).pow(2), dim=-1) |
|
|
| corr_matrix = torch.matmul(point_feat_a, point_feat_b.transpose(1, 2)) |
|
|
| dist = a_square + b_square - 2 * corr_matrix |
| if norm: |
| dist = torch.sqrt(dist) / num_channel |
| return dist |
|
|
|
|
| def get_sampler_cls(sampler_type): |
| """Get the type and mode of points sampler. |
| |
| Args: |
| sampler_type (str): The type of points sampler. |
| The valid value are "D-FPS", "F-FPS", or "FS". |
| |
| Returns: |
| class: Points sampler type. |
| """ |
| sampler_mappings = { |
| 'D-FPS': DFPSSampler, |
| 'F-FPS': FFPSSampler, |
| 'FS': FSSampler, |
| } |
| try: |
| return sampler_mappings[sampler_type] |
| except KeyError: |
| raise KeyError( |
| f'Supported `sampler_type` are {sampler_mappings.keys()}, but got \ |
| {sampler_type}') |
|
|
|
|
| class PointsSampler(nn.Module): |
| """Points sampling. |
| |
| Args: |
| num_point (list[int]): Number of sample points. |
| fps_mod_list (list[str], optional): Type of FPS method, valid mod |
| ['F-FPS', 'D-FPS', 'FS'], Default: ['D-FPS']. |
| F-FPS: using feature distances for FPS. |
| D-FPS: using Euclidean distances of points for FPS. |
| FS: using F-FPS and D-FPS simultaneously. |
| fps_sample_range_list (list[int], optional): |
| Range of points to apply FPS. Default: [-1]. |
| """ |
|
|
| def __init__(self, |
| num_point: List[int], |
| fps_mod_list: List[str] = ['D-FPS'], |
| fps_sample_range_list: List[int] = [-1]): |
| super().__init__() |
| |
| |
| |
| assert len(num_point) == len(fps_mod_list) == len( |
| fps_sample_range_list) |
| self.num_point = num_point |
| self.fps_sample_range_list = fps_sample_range_list |
| self.samplers = nn.ModuleList() |
| for fps_mod in fps_mod_list: |
| self.samplers.append(get_sampler_cls(fps_mod)()) |
| self.fp16_enabled = False |
|
|
| @force_fp32() |
| def forward(self, points_xyz, features): |
| """ |
| Args: |
| points_xyz (Tensor): (B, N, 3) xyz coordinates of the features. |
| features (Tensor): (B, C, N) Descriptors of the features. |
| |
| Returns: |
| Tensor: (B, npoint, sample_num) Indices of sampled points. |
| """ |
| indices = [] |
| last_fps_end_index = 0 |
|
|
| for fps_sample_range, sampler, npoint in zip( |
| self.fps_sample_range_list, self.samplers, self.num_point): |
| assert fps_sample_range < points_xyz.shape[1] |
|
|
| if fps_sample_range == -1: |
| sample_points_xyz = points_xyz[:, last_fps_end_index:] |
| if features is not None: |
| sample_features = features[:, :, last_fps_end_index:] |
| else: |
| sample_features = None |
| else: |
| sample_points_xyz = \ |
| points_xyz[:, last_fps_end_index:fps_sample_range] |
| if features is not None: |
| sample_features = features[:, :, last_fps_end_index: |
| fps_sample_range] |
| else: |
| sample_features = None |
|
|
| fps_idx = sampler(sample_points_xyz.contiguous(), sample_features, |
| npoint) |
|
|
| indices.append(fps_idx + last_fps_end_index) |
| last_fps_end_index += fps_sample_range |
| indices = torch.cat(indices, dim=1) |
|
|
| return indices |
|
|
|
|
| class DFPSSampler(nn.Module): |
| """Using Euclidean distances of points for FPS.""" |
|
|
| def __init__(self): |
| super().__init__() |
|
|
| def forward(self, points, features, npoint): |
| """Sampling points with D-FPS.""" |
| fps_idx = furthest_point_sample(points.contiguous(), npoint) |
| return fps_idx |
|
|
|
|
| class FFPSSampler(nn.Module): |
| """Using feature distances for FPS.""" |
|
|
| def __init__(self): |
| super().__init__() |
|
|
| def forward(self, points, features, npoint): |
| """Sampling points with F-FPS.""" |
| assert features is not None, \ |
| 'feature input to FFPS_Sampler should not be None' |
| features_for_fps = torch.cat([points, features.transpose(1, 2)], dim=2) |
| features_dist = calc_square_dist( |
| features_for_fps, features_for_fps, norm=False) |
| fps_idx = furthest_point_sample_with_dist(features_dist, npoint) |
| return fps_idx |
|
|
|
|
| class FSSampler(nn.Module): |
| """Using F-FPS and D-FPS simultaneously.""" |
|
|
| def __init__(self): |
| super().__init__() |
|
|
| def forward(self, points, features, npoint): |
| """Sampling points with FS_Sampling.""" |
| assert features is not None, \ |
| 'feature input to FS_Sampler should not be None' |
| ffps_sampler = FFPSSampler() |
| dfps_sampler = DFPSSampler() |
| fps_idx_ffps = ffps_sampler(points, features, npoint) |
| fps_idx_dfps = dfps_sampler(points, features, npoint) |
| fps_idx = torch.cat([fps_idx_ffps, fps_idx_dfps], dim=1) |
| return fps_idx |
|
|