| from dvc_eval import eval_dvc, eval_soda |
| import json |
| import argparse |
| import re |
| import difflib |
| import os |
| from torchvision.transforms import Compose, Resize, CenterCrop, Normalize |
| import torch |
|
|
| |
| try: |
| from torchvision.transforms import InterpolationMode |
| BICUBIC = InterpolationMode.BICUBIC |
| except ImportError: |
| BICUBIC = Image.BICUBIC |
| from torchvision.transforms import Compose, Resize, CenterCrop, Normalize |
|
|
|
|
| transform = Compose([ |
| Resize(224, interpolation=BICUBIC), |
| CenterCrop(224), |
| Normalize((0.48145466, 0.4578275, 0.40821073), (0.26862954, 0.26130258, 0.27577711)), |
| ]) |
|
|
| |
| def check_model_files(config): |
| """Check if required model files exist""" |
| files_to_check = [ |
| config.clip_path, |
| config.pretrain_mm_mlp_adapter, |
| config.stage2, |
| config.stage3, |
| config.stage4, |
| config.stage5, |
| config.model_base |
| ] |
| |
| missing_files = [] |
| for file_path in files_to_check: |
| if not os.path.exists(file_path): |
| missing_files.append(file_path) |
| |
| if missing_files: |
| print("⚠ Missing model files:") |
| for file_path in missing_files: |
| print(f" - {file_path}") |
| print("\nPlease download the required model checkpoints.") |
| return False |
| else: |
| print("✓ All model files found") |
| return True |
|
|
| |
|
|
|
|
| |
| |
| def extract_video_features(video_path, clip_model, video_loader, transform): |
| """Extract features from a video file""" |
| try: |
| |
| _, images = video_loader.extract({'id': None, 'video': video_path}) |
| |
| |
| images = transform(images / 255.0) |
| images = images.to(torch.float16) |
| |
| |
| with torch.no_grad(): |
| features = clip_model.encode_image(images.to('cuda')) |
| |
| return features |
| except Exception as e: |
| print(f"Error processing video {video_path}: {e}") |
| return None |
|
|
| def find_video_file(video_id, video_folder): |
| """Find video file with various extensions""" |
| for ext in ['mp4', 'mkv', 'webm', 'avi', 'mov']: |
| video_path = os.path.join(video_folder, f"{video_id}.{ext}") |
| if os.path.isfile(video_path): |
| return video_path |
| return None |
|
|
| def load_dataset(data_path): |
| """Load dataset from JSON file""" |
| try: |
| with open(data_path, 'r') as f: |
| data = json.load(f) |
| return data |
| except Exception as e: |
| print(f"✗ Error loading dataset: {e}") |
| return None |
|
|
|
|
| |
| def merge_similar_sentences(data): |
| if not data: return data |
| merged_data = [] |
| current_sentence = data[0]["sentence"] |
| current_timestamp = data[0]["timestamp"] |
| for i in range(1, len(data)): |
| next_sentence = data[i]["sentence"] |
| next_timestamp = data[i]["timestamp"] |
| if difflib.SequenceMatcher(None, current_sentence, next_sentence).ratio() > 0.98 and -1 <= next_timestamp[0] - current_timestamp[1] <= 1: |
| current_timestamp = [current_timestamp[0], next_timestamp[1]] |
| else: |
| merged_data.append({"sentence": current_sentence, "timestamp": current_timestamp}) |
| current_sentence = next_sentence |
| current_timestamp = next_timestamp |
| merged_data.append({"sentence": current_sentence, "timestamp": current_timestamp}) |
| return merged_data |
|
|
| def evaluate(id, event, timestamps, answer, js): |
| pred = {} |
| pred[id] = [] |
| for num in range(len(event)): |
| pred[id].append({ |
| 'timestamp': timestamps[num], |
| 'sentence': event[num] |
| }) |
|
|
| refined_pred = [] |
| for num_pred, curr_pred in enumerate(pred[id]): |
| duplicate = False |
| for curr_pred2 in pred[id][num_pred + 1:]: |
| |
| if curr_pred2 == curr_pred: |
| num_duplicates+=1 |
| duplicate=True |
| |
| |
| if not duplicate: |
| refined_pred.append(curr_pred) |
|
|
| pred[id] = refined_pred |
| gt_js = {k: v for k, v in js.items() if k in pred.keys()} |
| |
| for id, items in list(pred.items()): |
| items = merge_similar_sentences(items) |
| duration = gt_js[id]['duration'] |
| for item in items: |
| item['timestamp'][0] = item['timestamp'][0] * duration / 100 |
| item['timestamp'][1] = (item['timestamp'][1] + 1) * duration / 100 |
| pred[id] = items |
| |
| pred_result = {'results': pred} |
| metrics = eval_soda(pred_result, [gt_js], print_matrix=False) |
| metrics.update(eval_dvc(pred_result, [gt_js], |
| tious=[0.3, 0.5, 0.7], |
| distances=[], |
| max_proposals_per_video=1000, |
| verbose=False, |
| no_lang_eval=False)) |
|
|
| print(f"Found {len(pred)} logs") |
| metrics = {k: v.item() * 100 for k, v in metrics.items() if k in ['soda_c', 'METEOR', 'CIDEr']} |
| return metrics |