| import argparse |
| from dataclasses import dataclass |
| from multiprocessing import cpu_count |
| from pathlib import Path |
| import numpy as np |
| import torch |
| import yaml |
| from torch.utils.data import DataLoader |
| from tqdm import tqdm |
| from constants import HIDDEN_SIZE, TARGET_SAMPLE_RATE |
| from data import FixedSegmentationDatasetNoTarget, segm_collate_fn |
| from eval import infer |
| from models import SegmentationFrameClassifer, prepare_wav2vec |
| import os |
|
|
| @dataclass |
| class Segment: |
| start: float |
| end: float |
| probs: np.array |
| decimal: int = 4 |
|
|
| @property |
| def duration(self): |
| return float(round((self.end - self.start) / TARGET_SAMPLE_RATE, self.decimal)) |
|
|
| @property |
| def offset(self): |
| return float(round(self.start / TARGET_SAMPLE_RATE, self.decimal)) |
|
|
| @property |
| def offset_plus_duration(self): |
| return round(self.offset + self.duration, self.decimal) |
|
|
|
|
| def trim(sgm: Segment, threshold: float) -> Segment: |
| """reduces the segment to between the first and last points that are above the threshold |
| |
| Args: |
| sgm (Segment): a segment |
| threshold (float): probability threshold |
| |
| Returns: |
| Segment: new reduced segment |
| """ |
| included_indices = np.where(sgm.probs >= threshold)[0] |
| |
| |
| if not len(included_indices): |
| return Segment(sgm.start, sgm.start, np.empty([0])) |
|
|
| i = included_indices[0] |
| j = included_indices[-1] + 1 |
|
|
| sgm = Segment(sgm.start + i, sgm.start + j, sgm.probs[i:j]) |
|
|
| return sgm |
|
|
|
|
| def split_and_trim( |
| sgm: Segment, split_idx: int, threshold: float |
| ) -> tuple[Segment, Segment]: |
| """splits the input segment at the split_idx and then trims and returns the two resulting segments |
| |
| Args: |
| sgm (Segment): input segment |
| split_idx (int): index to split the input segment |
| threshold (float): probability threshold |
| |
| Returns: |
| tuple[Segment, Segment]: the two resulting segments |
| """ |
|
|
| probs_a = sgm.probs[:split_idx] |
| sgm_a = Segment(sgm.start, sgm.start + len(probs_a), probs_a) |
|
|
| probs_b = sgm.probs[split_idx + 1 :] |
| sgm_b = Segment(sgm_a.end + 1, sgm.end, probs_b) |
|
|
| sgm_a = trim(sgm_a, threshold) |
| sgm_b = trim(sgm_b, threshold) |
|
|
| return sgm_a, sgm_b |
|
|
|
|
| def pdac( |
| probs: np.array, |
| max_segment_length: float, |
| min_segment_length: float, |
| threshold: float, |
| not_strict: bool |
| ) -> list[Segment]: |
| """applies the probabilistic Divide-and-Conquer algorithm to split an audio |
| into segments satisfying the max-segment-length and min-segment-length conditions |
| |
| Args: |
| probs (np.array): the binary frame-level probabilities |
| output by the segmentation-frame-classifier |
| max_segment_length (float): the maximum length of a segment |
| min_segment_length (float): the minimum length of a segment |
| threshold (float): probability threshold |
| not_strict (bool): whether segments longer than max are allowed |
| |
| Returns: |
| list[Segment]: resulting segmentation |
| """ |
|
|
| segments = [] |
| sgm = Segment(0, len(probs), probs) |
| sgm = trim(sgm, threshold) |
|
|
| def recusrive_split(sgm): |
| if sgm.duration < max_segment_length: |
| segments.append(sgm) |
| else: |
| j = 0 |
| sorted_indices = np.argsort(sgm.probs) |
| while j < len(sorted_indices): |
| split_idx = sorted_indices[j] |
| split_prob = sgm.probs[split_idx] |
| if not_strict and split_prob > threshold: |
| segments.append(sgm) |
| break |
|
|
| sgm_a, sgm_b = split_and_trim(sgm, split_idx, threshold) |
| if ( |
| sgm_a.duration > min_segment_length |
| and sgm_b.duration > min_segment_length |
| ): |
| recusrive_split(sgm_a) |
| recusrive_split(sgm_b) |
| break |
| j += 1 |
| else: |
| if not_strict: |
| segments.append(sgm) |
| else: |
| if sgm_a.duration > min_segment_length: |
| recusrive_split(sgm_a) |
| if sgm_b.duration > min_segment_length: |
| recusrive_split(sgm_b) |
|
|
| recusrive_split(sgm) |
|
|
| return segments |
|
|
|
|
| def update_yaml_content( |
| yaml_content: list[dict], segments: list[Segment], wav_name: str |
| ) -> list[dict]: |
| """extends the yaml content with the segmentation of this wav file |
| |
| Args: |
| yaml_content (list[dict]): segmentation in yaml format |
| segments (list[Segment]): resulting segmentation from pdac |
| wav_name (str): name of the wav file |
| |
| Returns: |
| list[dict]: extended segmentation in yaml format |
| """ |
| for sgm in segments: |
| yaml_content.append( |
| { |
| "duration": sgm.duration, |
| "offset": sgm.offset, |
| "rW": 0, |
| "uW": 0, |
| "speaker_id": "NA", |
| "wav": wav_name, |
| } |
| ) |
| return yaml_content |
|
|
|
|
| def segment(args): |
|
|
| device = ( |
| torch.device(f"cuda:0") |
| if torch.cuda.device_count() > 0 |
| else torch.device("cpu") |
| ) |
|
|
| checkpoint = torch.load(args.path_to_checkpoint, map_location=device) |
|
|
| |
| wav2vec_model = prepare_wav2vec( |
| checkpoint["args"].model_name, |
| checkpoint["args"].wav2vec_keep_layers, |
| device, |
| ) |
| |
| sfc_model = SegmentationFrameClassifer( |
| d_model=HIDDEN_SIZE, |
| n_transformer_layers=checkpoint["args"].classifier_n_transformer_layers, |
| ).to(device) |
| sfc_model.load_state_dict(checkpoint["state_dict"]) |
| sfc_model.eval() |
| |
| yaml_content = [] |
| with open(args.wavs, 'r') as file: |
| wav_paths = [x.strip() for x in file.readlines()] |
| |
| for wav_path in wav_paths: |
| |
| |
| dataset = FixedSegmentationDatasetNoTarget(wav_path, args.inference_segment_length, args.inference_times) |
| sgm_frame_probs = None |
|
|
| for inference_iteration in range(args.inference_times): |
|
|
| |
| dataset.fixed_length_segmentation(inference_iteration) |
| dataloader = DataLoader( |
| dataset, |
| batch_size=args.inference_batch_size, |
| num_workers=min(cpu_count() // 2, 4), |
| shuffle=False, |
| drop_last=False, |
| collate_fn=segm_collate_fn, |
| ) |
|
|
| |
| probs, _ = infer( |
| wav2vec_model, |
| sfc_model, |
| dataloader, |
| device, |
| ) |
| if sgm_frame_probs is None: |
| sgm_frame_probs = probs.copy() |
| else: |
| sgm_frame_probs += probs |
|
|
| sgm_frame_probs /= args.inference_times |
|
|
| segments = pdac( |
| sgm_frame_probs, |
| args.dac_max_segment_length, |
| args.dac_min_segment_length, |
| args.dac_threshold, |
| args.not_strict |
| ) |
| wav_path_name = os.path.basename(wav_path) |
| for sgm in segments: |
| print(f"Segmentation:{wav_path_name}\t{sgm.offset}\t{sgm.duration}") |
| yaml_content = update_yaml_content(yaml_content, segments, wav_path_name) |
|
|
| path_to_segmentation_yaml = Path(args.path_to_segmentation_yaml) |
| path_to_segmentation_yaml.parent.mkdir(parents=True, exist_ok=True) |
| with open(path_to_segmentation_yaml, "w") as f: |
| yaml.dump(yaml_content, f, default_flow_style=True) |
|
|
| print( |
| f"Saved SHAS segmentation with max={args.dac_max_segment_length} & " |
| f"min={args.dac_min_segment_length} at {path_to_segmentation_yaml}" |
| ) |
|
|
|
|
| if __name__ == "__main__": |
|
|
| parser = argparse.ArgumentParser() |
| parser.add_argument( |
| "--path_to_segmentation_yaml", |
| "-yaml", |
| type=str, |
| required=True, |
| help="absolute path to the yaml file to save the generated segmentation", |
| ) |
| parser.add_argument( |
| "--path_to_checkpoint", |
| "-ckpt", |
| type=str, |
| required=True, |
| help="absolute path to the audio-frame-classifier checkpoint", |
| ) |
| parser.add_argument( |
| "-wavs", |
| type=str, |
| help="absolute path to the directory of the wav audios to be segmented", |
| ) |
| parser.add_argument( |
| "--inference_batch_size", |
| "-bs", |
| type=int, |
| default=12, |
| help="batch size (in examples) of inference with the audio-frame-classifier", |
| ) |
| parser.add_argument( |
| "--inference_segment_length", |
| "-len", |
| type=int, |
| default=20, |
| help="segment length (in seconds) of fixed-length segmentation during inference" |
| "with audio-frame-classifier", |
| ) |
| parser.add_argument( |
| "--inference_times", |
| "-n", |
| type=int, |
| default=1, |
| help="how many times to apply inference on different fixed-length segmentations" |
| "of each wav", |
| ) |
| parser.add_argument( |
| "--dac_max_segment_length", |
| "-max", |
| type=float, |
| default=20.0, |
| help="the segmentation algorithm splits until all segments are below this value" |
| "(in seconds)", |
| ) |
| parser.add_argument( |
| "--dac_min_segment_length", |
| "-min", |
| type=float, |
| default=0.2, |
| help="a split by the algorithm is carried out only if the resulting two segments" |
| "are above this value (in seconds)", |
| ) |
| parser.add_argument( |
| "--dac_threshold", |
| "-thr", |
| type=float, |
| default=0.5, |
| help="after each split by the algorithm, the resulting segments are trimmed to" |
| "the first and last points that corresponds to a probability above this value", |
| ) |
| parser.add_argument( |
| "--not_strict", |
| action="store_true", |
| help="whether segments longer than max are allowed." |
| "If this argument is used, respecting the classification threshold conditions (p > thr)" |
| "is more important than the length conditions (len < max)." |
| ) |
| args = parser.parse_args() |
|
|
| segment(args) |
|
|