| import os |
| import cv2 |
| import numpy as np |
| import os.path as osp |
| import imageio |
| from copy import deepcopy |
|
|
| import loguru |
| import torch |
| import matplotlib.cm as cm |
| import matplotlib.pyplot as plt |
|
|
| from ..loftr import LoFTR, default_cfg |
| from . import plt_utils |
| from .plotting import make_matching_figure |
| from .utils3d import rect_to_img, canonical_to_camera, calc_pose |
|
|
|
|
| class ElevEstHelper: |
| _feature_matcher = None |
|
|
| @classmethod |
| def get_feature_matcher(cls): |
| if cls._feature_matcher is None: |
| loguru.logger.info("Loading feature matcher...") |
| _default_cfg = deepcopy(default_cfg) |
| _default_cfg['coarse']['temp_bug_fix'] = True |
| matcher = LoFTR(config=_default_cfg) |
| current_dir = os.path.dirname(os.path.abspath(__file__)) |
| ckpt_path = os.path.join(current_dir, "weights/indoor_ds_new.ckpt") |
| if not osp.exists(ckpt_path): |
| raise FileNotFoundError(f"Checkpoint not found at {ckpt_path}") |
| matcher.load_state_dict(torch.load(ckpt_path)['state_dict']) |
| matcher = matcher.eval().cuda() |
| cls._feature_matcher = matcher |
| return cls._feature_matcher |
|
|
|
|
| def mask_out_bkgd(img_path, dbg=False): |
| img = imageio.imread_v2(img_path) |
| if img.shape[-1] == 4: |
| fg_mask = img[:, :, :3] |
| else: |
| loguru.logger.info("Image has no alpha channel, using thresholding to mask out background") |
| fg_mask = ~(img > 245).all(axis=-1) |
| if dbg: |
| plt.imshow(plt_utils.vis_mask(img, fg_mask.astype(np.uint8), color=[0, 255, 0])) |
| plt.show() |
| return fg_mask |
|
|
|
|
| def get_feature_matching(img_paths, dbg=False): |
| assert len(img_paths) == 4 |
| matcher = ElevEstHelper.get_feature_matcher() |
| feature_matching = {} |
| masks = [] |
| for i in range(4): |
| mask = mask_out_bkgd(img_paths[i], dbg=dbg) |
| masks.append(mask) |
| for i in range(0, 4): |
| for j in range(i + 1, 4): |
| img0_pth = img_paths[i] |
| img1_pth = img_paths[j] |
| mask0 = masks[i] |
| mask1 = masks[j] |
| img0_raw = cv2.imread(img0_pth, cv2.IMREAD_GRAYSCALE) |
| img1_raw = cv2.imread(img1_pth, cv2.IMREAD_GRAYSCALE) |
| original_shape = img0_raw.shape |
| img0_raw_resized = cv2.resize(img0_raw, (480, 480)) |
| img1_raw_resized = cv2.resize(img1_raw, (480, 480)) |
|
|
| img0 = torch.from_numpy(img0_raw_resized)[None][None].cuda() / 255. |
| img1 = torch.from_numpy(img1_raw_resized)[None][None].cuda() / 255. |
| batch = {'image0': img0, 'image1': img1} |
|
|
| |
| with torch.no_grad(): |
| matcher(batch) |
| mkpts0 = batch['mkpts0_f'].cpu().numpy() |
| mkpts1 = batch['mkpts1_f'].cpu().numpy() |
| mconf = batch['mconf'].cpu().numpy() |
| mkpts0[:, 0] = mkpts0[:, 0] * original_shape[1] / 480 |
| mkpts0[:, 1] = mkpts0[:, 1] * original_shape[0] / 480 |
| mkpts1[:, 0] = mkpts1[:, 0] * original_shape[1] / 480 |
| mkpts1[:, 1] = mkpts1[:, 1] * original_shape[0] / 480 |
| keep0 = mask0[mkpts0[:, 1].astype(int), mkpts1[:, 0].astype(int)] |
| keep1 = mask1[mkpts1[:, 1].astype(int), mkpts1[:, 0].astype(int)] |
| keep = np.logical_and(keep0, keep1) |
| mkpts0 = mkpts0[keep] |
| mkpts1 = mkpts1[keep] |
| mconf = mconf[keep] |
| if dbg: |
| |
| color = cm.jet(mconf) |
| text = [ |
| 'LoFTR', |
| 'Matches: {}'.format(len(mkpts0)), |
| ] |
| fig = make_matching_figure(img0_raw, img1_raw, mkpts0, mkpts1, color, text=text) |
| fig.show() |
| feature_matching[f"{i}_{j}"] = np.concatenate([mkpts0, mkpts1, mconf[:, None]], axis=1) |
|
|
| return feature_matching |
|
|
|
|
| def gen_pose_hypothesis(center_elevation): |
| elevations = np.radians( |
| [center_elevation, center_elevation - 10, center_elevation + 10, center_elevation, center_elevation]) |
| azimuths = np.radians([30, 30, 30, 20, 40]) |
| input_poses = calc_pose(elevations, azimuths, len(azimuths)) |
| input_poses = input_poses[1:] |
| input_poses[..., 1] *= -1 |
| input_poses[..., 2] *= -1 |
| return input_poses |
|
|
|
|
| def ba_error_general(K, matches, poses): |
| projmat0 = K @ poses[0].inverse()[:3, :4] |
| projmat1 = K @ poses[1].inverse()[:3, :4] |
| match_01 = matches[0] |
| pts0 = match_01[:, :2] |
| pts1 = match_01[:, 2:4] |
| Xref = cv2.triangulatePoints(projmat0.cpu().numpy(), projmat1.cpu().numpy(), |
| pts0.cpu().numpy().T, pts1.cpu().numpy().T) |
| Xref = Xref[:3] / Xref[3:] |
| Xref = Xref.T |
| Xref = torch.from_numpy(Xref).cuda().float() |
| reproj_error = 0 |
| for match, cp in zip(matches[1:], poses[2:]): |
| dist = (torch.norm(match_01[:, :2][:, None, :] - match[:, :2][None, :, :], dim=-1)) |
| if dist.numel() > 0: |
| |
| m0to2_index = dist.argmin(1) |
| keep = dist[torch.arange(match_01.shape[0]), m0to2_index] < 1 |
| if keep.sum() > 0: |
| xref_in2 = rect_to_img(K, canonical_to_camera(Xref, cp.inverse())) |
| reproj_error2 = torch.norm(match[m0to2_index][keep][:, 2:4] - xref_in2[keep], dim=-1) |
| conf02 = match[m0to2_index][keep][:, -1] |
| reproj_error += (reproj_error2 * conf02).sum() / (conf02.sum()) |
|
|
| return reproj_error |
|
|
|
|
| def find_optim_elev(elevs, nimgs, matches, K, dbg=False): |
| errs = [] |
| for elev in elevs: |
| err = 0 |
| cam_poses = gen_pose_hypothesis(elev) |
| for start in range(nimgs - 1): |
| batch_matches, batch_poses = [], [] |
| for i in range(start, nimgs + start): |
| ci = i % nimgs |
| batch_poses.append(cam_poses[ci]) |
| for j in range(nimgs - 1): |
| key = f"{start}_{(start + j + 1) % nimgs}" |
| match = matches[key] |
| batch_matches.append(match) |
| err += ba_error_general(K, batch_matches, batch_poses) |
| errs.append(err) |
| errs = torch.tensor(errs) |
| if dbg: |
| plt.plot(elevs, errs) |
| plt.show() |
| optim_elev = elevs[torch.argmin(errs)].item() |
| return optim_elev |
|
|
|
|
| def get_elev_est(feature_matching, min_elev=30, max_elev=150, K=None, dbg=False): |
| flag = True |
| matches = {} |
| for i in range(4): |
| for j in range(i + 1, 4): |
| match_ij = feature_matching[f"{i}_{j}"] |
| if len(match_ij) == 0: |
| flag = False |
| match_ji = np.concatenate([match_ij[:, 2:4], match_ij[:, 0:2], match_ij[:, 4:5]], axis=1) |
| matches[f"{i}_{j}"] = torch.from_numpy(match_ij).float().cuda() |
| matches[f"{j}_{i}"] = torch.from_numpy(match_ji).float().cuda() |
| if not flag: |
| loguru.logger.info("0 matches, could not estimate elevation") |
| return None |
| interval = 10 |
| elevs = np.arange(min_elev, max_elev, interval) |
| optim_elev1 = find_optim_elev(elevs, 4, matches, K) |
|
|
| elevs = np.arange(optim_elev1 - 10, optim_elev1 + 10, 1) |
| optim_elev2 = find_optim_elev(elevs, 4, matches, K) |
|
|
| return optim_elev2 |
|
|
|
|
| def elev_est_api(img_paths, min_elev=30, max_elev=150, K=None, dbg=False): |
| feature_matching = get_feature_matching(img_paths, dbg=dbg) |
| if K is None: |
| loguru.logger.warning("K is not provided, using default K") |
| K = np.array([[280.0, 0, 128.0], |
| [0, 280.0, 128.0], |
| [0, 0, 1]]) |
| K = torch.from_numpy(K).cuda().float() |
| elev = get_elev_est(feature_matching, min_elev, max_elev, K, dbg=dbg) |
| return elev |
|
|