| import torch |
| from torch_scatter import scatter_add, scatter_mean |
|
|
| from src.constants import atom_decoder, vdw_radii |
| _vdw_radii = {**vdw_radii} |
| _vdw_radii['NH'] = vdw_radii['N'] |
| _vdw_radii['N+'] = vdw_radii['N'] |
| _vdw_radii['O-'] = vdw_radii['O'] |
| _vdw_radii['NOATOM'] = 0 |
| vdw_radii_array = torch.tensor([_vdw_radii[a] for a in atom_decoder]) |
|
|
|
|
| def clash_loss(ligand_coord, ligand_types, ligand_mask, pocket_coord, |
| pocket_types, pocket_mask): |
| """ |
| Computes a clash loss that penalizes interatomic distances smaller than the |
| sum of van der Waals radii between atoms. |
| """ |
|
|
| ligand_radii = vdw_radii_array[ligand_types].to(ligand_coord.device) |
| pocket_radii = vdw_radii_array[pocket_types].to(pocket_coord.device) |
|
|
| dist = torch.sqrt(torch.sum((ligand_coord[:, None, :] - pocket_coord[None, :, :]) ** 2, dim=-1)) |
| |
|
|
| |
| |
| sum_vdw = ligand_radii[:, None] + pocket_radii[None, :] |
| loss = torch.clamp(1 - dist / sum_vdw, min=0.0) |
|
|
| loss = scatter_add(loss, pocket_mask, dim=1) |
| loss = scatter_mean(loss, ligand_mask, dim=0) |
| loss = loss.diag() |
|
|
| |
| |
| |
| |
| |
| |
| |
|
|
| return loss |
|
|
|
|
| class TimestepSampler: |
| def __init__(self, type='uniform', lowest_t=1, highest_t=500): |
| assert type in {'uniform', 'sigmoid'} |
| self.type = type |
| self.lowest_t = lowest_t |
| self.highest_t = highest_t |
|
|
| def __call__(self, n, device=None): |
| if self.type == 'uniform': |
| t_int = torch.randint(self.lowest_t, self.highest_t + 1, |
| size=(n, 1), device=device) |
|
|
| elif self.type == 'sigmoid': |
| weight_fun = lambda t: 1.45 * torch.sigmoid(-t * 10 / self.highest_t + 5) + 0.05 |
|
|
| possible_ts = torch.arange(self.lowest_t, self.highest_t + 1, device=device) |
| weights = weight_fun(possible_ts) |
| weights = weights / weights.sum() |
| t_int = possible_ts[torch.multinomial(weights, n, replacement=True)].unsqueeze(-1) |
|
|
| return t_int.float() |
|
|
|
|
| class TimestepWeights: |
| def __init__(self, weight_type, a, b): |
| if weight_type != 'sigmoid': |
| raise NotImplementedError("Only sigmoidal loss weighting is available.") |
| |
| self.weight_fn = lambda t: a * torch.sigmoid((t - 0.5) * b) + (1 - a / 2) |
|
|
| def __call__(self, t_array): |
| |
| |
| return self.weight_fn(t_array) |
|
|