| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import argparse |
| import json |
| import random |
| import string |
| import sys |
| import time |
| |
|
|
| from pycocoevalcap.tokenizer.ptbtokenizer import PTBTokenizer |
| from pycocoevalcap.meteor.meteor import Meteor |
|
|
| Set = set |
| import numpy as np |
|
|
|
|
| def random_string(string_length): |
| letters = string.ascii_lowercase |
| return ''.join(random.choice(letters) for i in range(string_length)) |
|
|
|
|
| def remove_nonascii(text): |
| return ''.join([i if ord(i) < 128 else ' ' for i in text]) |
|
|
|
|
| class ANETcaptions(object): |
| PREDICTION_FIELDS = ['results', 'version', 'external_data'] |
|
|
| def __init__(self, ground_truth_filenames=None, prediction_filename=None, |
| tious=None, distances=[1, 3, 5, 10, 30, 60], max_proposals=1000, |
| prediction_fields=PREDICTION_FIELDS, verbose=False, no_lang_eval=False): |
| |
| if len(tious) == 0: |
| raise IOError('Please input a valid tIoU.') |
| if not ground_truth_filenames: |
| raise IOError('Please input a valid ground truth file.') |
| if not prediction_filename: |
| raise IOError('Please input a valid prediction file.') |
|
|
| self.verbose = verbose |
| self.no_lang_eval = no_lang_eval |
| self.tious = tious |
| self.distances = distances |
| self.max_proposals = max_proposals |
| self.pred_fields = prediction_fields |
| self.ground_truths = self.import_ground_truths(ground_truth_filenames) |
| self.prediction = self.import_prediction(prediction_filename) |
| self.ground_truths_keys = [vid for gt in self.ground_truths for vid in gt] |
| print('available video number', len(set(self.ground_truths_keys) & set(self.prediction.keys()))) |
|
|
| |
| if not self.no_lang_eval: |
| self.tokenizer = PTBTokenizer() |
| self.scorers = [ |
| (Meteor(), "METEOR"), |
| ] |
|
|
| def import_prediction(self, prediction_filename): |
| if self.verbose: |
| print("| Loading submission...") |
| if isinstance(prediction_filename, dict): |
| submission = prediction_filename |
| else: |
| submission = json.load(open(prediction_filename)) |
| |
| |
| |
| results = {} |
| for vid_id in submission['results']: |
| results[vid_id] = submission['results'][vid_id][:self.max_proposals] |
| return results |
|
|
| def import_ground_truths(self, filenames): |
| gts = [] |
| self.n_ref_vids = Set() |
| for filename in filenames: |
| if isinstance(filename, dict): |
| gt = filename |
| else: |
| gt = json.load(open(filename)) |
| self.n_ref_vids.update(gt.keys()) |
| gts.append(gt) |
| if self.verbose: |
| print("| Loading GT. #files: %d, #videos: %d" % (len(filenames), len(self.n_ref_vids))) |
| return gts |
|
|
| def iou(self, interval_1, interval_2): |
| start_i, end_i = interval_1[0], interval_1[1] |
| start, end = interval_2[0], interval_2[1] |
| intersection = max(0, min(end, end_i) - max(start, start_i)) |
| union = min(max(end, end_i) - min(start, start_i), end - start + end_i - start_i) |
| iou = float(intersection) / (union + 1e-8) |
| return iou |
|
|
| def check_gt_exists(self, vid_id): |
| for gt in self.ground_truths: |
| if vid_id in gt: |
| return True |
| return False |
|
|
| def get_gt_vid_ids(self): |
| vid_ids = set([]) |
| for gt in self.ground_truths: |
| vid_ids |= set(gt.keys()) |
| return list(vid_ids) |
|
|
| def evaluate(self): |
| aggregator = {} |
| self.scores = {} |
| if not self.no_lang_eval: |
| for tiou in self.tious: |
| scores = self.evaluate_tiou(tiou) |
| for metric, score in scores.items(): |
| if metric not in self.scores: |
| self.scores[metric] = [] |
| self.scores[metric].append(score) |
| if True: |
| |
| self.scores['Recall'] = [] |
| self.scores['Precision'] = [] |
| self.scores['F1'] = [] |
| for tiou in self.tious: |
| precision, recall = self.evaluate_detection(tiou) |
| self.scores['Recall'].append(recall) |
| self.scores['Precision'].append(precision) |
| self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.) |
| for tiou in self.distances: |
| precision, recall = self.evaluate_navigation(tiou) |
| self.scores['Recall'].append(recall) |
| self.scores['Precision'].append(precision) |
| self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.) |
|
|
| def evaluate_detection(self, tiou): |
| gt_vid_ids = self.get_gt_vid_ids() |
| |
| |
| recall = [] |
| precision = [] |
| for vid_i, vid_id in enumerate(gt_vid_ids): |
| if vid_id not in self.prediction: |
| continue |
| best_recall = 0 |
| best_precision = 0 |
| for gt in self.ground_truths: |
| if vid_id not in gt: |
| continue |
| refs = gt[vid_id] |
| ref_set_covered = set([]) |
| pred_set_covered = set([]) |
| num_gt = 0 |
| num_pred = 0 |
| if vid_id in self.prediction: |
| for pred_i, pred in enumerate(self.prediction[vid_id]): |
| pred_timestamp = pred['timestamp'] |
| for ref_i, ref_timestamp in enumerate(refs['timestamps']): |
| if self.iou(pred_timestamp, ref_timestamp) > tiou: |
| ref_set_covered.add(ref_i) |
| pred_set_covered.add(pred_i) |
|
|
| new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1) |
| best_precision = max(best_precision, new_precision) |
| new_recall = float(len(ref_set_covered)) / len(refs['timestamps']) |
| best_recall = max(best_recall, new_recall) |
| recall.append(best_recall) |
| precision.append(best_precision) |
| return sum(precision) / len(precision), sum(recall) / len(recall) |
|
|
| def evaluate_navigation(self, tiou): |
| gt_vid_ids = self.get_gt_vid_ids() |
| |
| |
| recall = [] |
| precision = [] |
| for vid_i, vid_id in enumerate(gt_vid_ids): |
| if vid_id not in self.prediction: |
| continue |
| best_recall = 0 |
| best_precision = 0 |
| for gt in self.ground_truths: |
| if vid_id not in gt: |
| continue |
| refs = gt[vid_id] |
| ref_set_covered = set([]) |
| pred_set_covered = set([]) |
| num_gt = 0 |
| num_pred = 0 |
| if vid_id in self.prediction: |
| for pred_i, pred in enumerate(self.prediction[vid_id]): |
| pred_timestamp = pred['timestamp'] |
| for ref_i, ref_timestamp in enumerate(refs['timestamps']): |
| if abs(pred_timestamp[0] - ref_timestamp[0]) < tiou: |
| ref_set_covered.add(ref_i) |
| pred_set_covered.add(pred_i) |
|
|
| new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1) |
| best_precision = max(best_precision, new_precision) |
| new_recall = float(len(ref_set_covered)) / len(refs['timestamps']) |
| best_recall = max(best_recall, new_recall) |
| recall.append(best_recall) |
| precision.append(best_precision) |
| return sum(precision) / len(precision), sum(recall) / len(recall) |
|
|
| def evaluate_tiou(self, tiou): |
| |
| res = {} |
| gts = {} |
| gt_vid_ids = self.get_gt_vid_ids() |
|
|
| unique_index = 0 |
|
|
| |
| vid2capid = {} |
|
|
| cur_res = {} |
| cur_gts = {} |
|
|
| for vid_id in gt_vid_ids: |
|
|
| |
| |
| if vid_id not in self.prediction: |
| continue |
|
|
| |
| |
| else: |
| vid2capid[vid_id] = [] |
| |
| for pred in self.prediction[vid_id]: |
| has_added = False |
| for gt in self.ground_truths: |
| if vid_id not in gt: |
| continue |
| gt_captions = gt[vid_id] |
| for caption_idx, caption_timestamp in enumerate(gt_captions['timestamps']): |
| if self.iou(pred['timestamp'], caption_timestamp) >= tiou: |
| cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}] |
| cur_gts[unique_index] = [ |
| {'caption': remove_nonascii(gt_captions['sentences'][caption_idx])}] |
| vid2capid[vid_id].append(unique_index) |
| unique_index += 1 |
| has_added = True |
|
|
| |
| |
| if not has_added: |
| cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}] |
| cur_gts[unique_index] = [{'caption': random_string(random.randint(10, 20))}] |
| vid2capid[vid_id].append(unique_index) |
| unique_index += 1 |
|
|
| |
| output = {} |
| for scorer, method in self.scorers: |
| if self.verbose: |
| print('computing %s score...' % (scorer.method())) |
|
|
| |
| all_scores = {} |
|
|
| |
| tokenize_res = self.tokenizer.tokenize(cur_res) |
| tokenize_gts = self.tokenizer.tokenize(cur_gts) |
|
|
| |
| for vid in vid2capid.keys(): |
| res[vid] = {index: tokenize_res[index] for index in vid2capid[vid]} |
| gts[vid] = {index: tokenize_gts[index] for index in vid2capid[vid]} |
|
|
| for vid_id in gt_vid_ids: |
|
|
| if vid_id not in self.prediction: |
| continue |
|
|
| if len(res[vid_id]) == 0 or len(gts[vid_id]) == 0: |
| if type(method) == list: |
| score = [0] * len(method) |
| else: |
| score = 0 |
| else: |
| score, scores = scorer.compute_score(gts[vid_id], res[vid_id]) |
| all_scores[vid_id] = score |
| |
|
|
| |
| if type(method) == list: |
| scores = np.mean(list(all_scores.values()), axis=0) |
| for m in range(len(method)): |
| output[method[m]] = scores[m] |
| if self.verbose: |
| print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method[m], output[method[m]])) |
| else: |
| output[method] = np.mean(list(all_scores.values())) |
| if self.verbose: |
| print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method, output[method])) |
| return output |
|
|
|
|
| def eval_dvc(submission, references, tious=[0.3, 0.5, 0.7, 0.9], distances=[1, 3, 5, 10, 30, 60], max_proposals_per_video=1000, verbose=False, no_lang_eval=False): |
| |
| evaluator = ANETcaptions(ground_truth_filenames=references, |
| prediction_filename=submission, |
| tious=tious, |
| distances=distances, |
| max_proposals=max_proposals_per_video, |
| verbose=verbose, no_lang_eval=no_lang_eval) |
| evaluator.evaluate() |
| score = evaluator.scores |
| |
| loc_score = {} |
| for i, x in enumerate(tious): |
| for y in ["Recall", "Precision", "F1"]: |
| loc_score[y + "@" + str(x)] = score[y][i] |
| for y in ["Recall", "Precision", "F1"]: |
| loc_score[y] = np.array([score[y][i] for i in range(len(tious))]).mean() |
| if distances: |
| for i, x in enumerate(distances): |
| for y in ["Recall", "Precision", "F1"]: |
| loc_score[y + "@" + str(x) + "s"] = score[y][len(tious) + i] |
| avg_eval_score = {key: np.array(value).mean() for key, value in score.items() if key not in ["Recall", "Precision", "F1"]} |
| avg_eval_score.update(loc_score) |
| return avg_eval_score |
|
|
| if __name__ == '__main__': |
| eval_dvc(pred_path, references, |
| tious=[0.3, 0.5, 0.7, 0.9], |
| max_proposals_per_video=1000, |
| verbose=False, |
| no_lang_eval=False) |
| eval_soda(pred_path, references, verbose=False) |