| import argparse |
| import json |
| import os |
| import time |
| import zipfile |
|
|
| from get_ci_error_statistics import download_artifact, get_artifacts_links |
|
|
| from transformers import logging |
|
|
|
|
| logger = logging.get_logger(__name__) |
|
|
|
|
| def extract_warnings_from_single_artifact(artifact_path, targets): |
| """Extract warnings from a downloaded artifact (in .zip format)""" |
| selected_warnings = set() |
| buffer = [] |
|
|
| def parse_line(fp): |
| for line in fp: |
| if isinstance(line, bytes): |
| line = line.decode("UTF-8") |
| if "warnings summary (final)" in line: |
| continue |
| |
| elif not line.startswith(" "): |
| |
| if len(buffer) > 0: |
| warning = "\n".join(buffer) |
| |
| if any(f": {x}: " in warning for x in targets): |
| selected_warnings.add(warning) |
| buffer.clear() |
| continue |
| else: |
| line = line.strip() |
| buffer.append(line) |
|
|
| if from_gh: |
| for filename in os.listdir(artifact_path): |
| file_path = os.path.join(artifact_path, filename) |
| if not os.path.isdir(file_path): |
| |
| if filename != "warnings.txt": |
| continue |
| with open(file_path) as fp: |
| parse_line(fp) |
| else: |
| try: |
| with zipfile.ZipFile(artifact_path) as z: |
| for filename in z.namelist(): |
| if not os.path.isdir(filename): |
| |
| if filename != "warnings.txt": |
| continue |
| with z.open(filename) as fp: |
| parse_line(fp) |
| except Exception: |
| logger.warning( |
| f"{artifact_path} is either an invalid zip file or something else wrong. This file is skipped." |
| ) |
|
|
| return selected_warnings |
|
|
|
|
| def extract_warnings(artifact_dir, targets): |
| """Extract warnings from all artifact files""" |
|
|
| selected_warnings = set() |
|
|
| paths = [os.path.join(artifact_dir, p) for p in os.listdir(artifact_dir) if (p.endswith(".zip") or from_gh)] |
| for p in paths: |
| selected_warnings.update(extract_warnings_from_single_artifact(p, targets)) |
|
|
| return selected_warnings |
|
|
|
|
| if __name__ == "__main__": |
|
|
| def list_str(values): |
| return values.split(",") |
|
|
| parser = argparse.ArgumentParser() |
| |
| parser.add_argument("--workflow_run_id", type=str, required=True, help="A GitHub Actions workflow run id.") |
| parser.add_argument( |
| "--output_dir", |
| type=str, |
| required=True, |
| help="Where to store the downloaded artifacts and other result files.", |
| ) |
| parser.add_argument("--token", default=None, type=str, help="A token that has actions:read permission.") |
| |
| parser.add_argument( |
| "--targets", |
| default="DeprecationWarning,UserWarning,FutureWarning", |
| type=list_str, |
| help="Comma-separated list of target warning(s) which we want to extract.", |
| ) |
| parser.add_argument( |
| "--from_gh", |
| action="store_true", |
| help="If running from a GitHub action workflow and collecting warnings from its artifacts.", |
| ) |
|
|
| args = parser.parse_args() |
|
|
| from_gh = args.from_gh |
| if from_gh: |
| |
| pass |
| else: |
| os.makedirs(args.output_dir, exist_ok=True) |
|
|
| |
| artifacts = get_artifacts_links(args.workflow_run_id, token=args.token) |
| with open(os.path.join(args.output_dir, "artifacts.json"), "w", encoding="UTF-8") as fp: |
| json.dump(artifacts, fp, ensure_ascii=False, indent=4) |
|
|
| |
| for idx, (name, url) in enumerate(artifacts.items()): |
| print(name) |
| print(url) |
| print("=" * 80) |
| download_artifact(name, url, args.output_dir, args.token) |
| |
| time.sleep(1) |
|
|
| |
| selected_warnings = extract_warnings(args.output_dir, args.targets) |
| selected_warnings = sorted(selected_warnings) |
| with open(os.path.join(args.output_dir, "selected_warnings.json"), "w", encoding="UTF-8") as fp: |
| json.dump(selected_warnings, fp, ensure_ascii=False, indent=4) |
|
|