| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import pathlib |
| import random |
| import re |
| from typing import Tuple |
|
|
| import pytest |
| import test_utils.prefixer |
|
|
| from google.cloud import bigquery |
| from google.cloud.bigquery import enums |
| from . import helpers |
|
|
|
|
| prefixer = test_utils.prefixer.Prefixer("python-bigquery", "tests/system") |
|
|
| DATA_DIR = pathlib.Path(__file__).parent.parent / "data" |
| TOKYO_LOCATION = "asia-northeast1" |
|
|
|
|
| @pytest.fixture(scope="session", autouse=True) |
| def cleanup_datasets(bigquery_client: bigquery.Client): |
| for dataset in bigquery_client.list_datasets(): |
| if prefixer.should_cleanup(dataset.dataset_id): |
| bigquery_client.delete_dataset( |
| dataset, delete_contents=True, not_found_ok=True |
| ) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def bigquery_client(): |
| return bigquery.Client() |
|
|
|
|
| @pytest.fixture(scope="session") |
| def project_id(bigquery_client: bigquery.Client): |
| return bigquery_client.project |
|
|
|
|
| @pytest.fixture(scope="session") |
| def bqstorage_client(bigquery_client): |
| from google.cloud import bigquery_storage |
|
|
| return bigquery_storage.BigQueryReadClient(credentials=bigquery_client._credentials) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def dataset_id(bigquery_client): |
| dataset_id = prefixer.create_prefix() |
| bigquery_client.create_dataset(dataset_id) |
| yield dataset_id |
| bigquery_client.delete_dataset(dataset_id, delete_contents=True, not_found_ok=True) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def dataset_id_tokyo(bigquery_client: bigquery.Client, project_id: str): |
| dataset_id = prefixer.create_prefix() + "_tokyo" |
| dataset = bigquery.Dataset(f"{project_id}.{dataset_id}") |
| dataset.location = TOKYO_LOCATION |
| bigquery_client.create_dataset(dataset) |
| yield dataset_id |
| bigquery_client.delete_dataset(dataset_id, delete_contents=True, not_found_ok=True) |
|
|
|
|
| @pytest.fixture() |
| def dataset_client(bigquery_client, dataset_id): |
| import google.cloud.bigquery.job |
|
|
| return bigquery.Client( |
| default_query_job_config=google.cloud.bigquery.job.QueryJobConfig( |
| default_dataset=f"{bigquery_client.project}.{dataset_id}", |
| ) |
| ) |
|
|
|
|
| @pytest.fixture |
| def table_id(dataset_id): |
| return f"{dataset_id}.table_{helpers.temp_suffix()}" |
|
|
|
|
| def load_scalars_table( |
| bigquery_client: bigquery.Client, |
| project_id: str, |
| dataset_id: str, |
| data_path: str = "scalars.jsonl", |
| source_format=enums.SourceFormat.NEWLINE_DELIMITED_JSON, |
| schema_source="scalars_schema.json", |
| ) -> str: |
| schema = bigquery_client.schema_from_json(DATA_DIR / schema_source) |
| table_id = data_path.replace(".", "_") + hex(random.randrange(1000000)) |
| job_config = bigquery.LoadJobConfig() |
| job_config.schema = schema |
| job_config.source_format = source_format |
| full_table_id = f"{project_id}.{dataset_id}.{table_id}" |
| with open(DATA_DIR / data_path, "rb") as data_file: |
| job = bigquery_client.load_table_from_file( |
| data_file, full_table_id, job_config=job_config |
| ) |
| job.result() |
| return full_table_id |
|
|
|
|
| @pytest.fixture(scope="session") |
| def scalars_table(bigquery_client: bigquery.Client, project_id: str, dataset_id: str): |
| full_table_id = load_scalars_table(bigquery_client, project_id, dataset_id) |
| yield full_table_id |
| bigquery_client.delete_table(full_table_id, not_found_ok=True) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def scalars_table_tokyo( |
| bigquery_client: bigquery.Client, project_id: str, dataset_id_tokyo: str |
| ): |
| full_table_id = load_scalars_table(bigquery_client, project_id, dataset_id_tokyo) |
| yield full_table_id |
| bigquery_client.delete_table(full_table_id, not_found_ok=True) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def scalars_extreme_table( |
| bigquery_client: bigquery.Client, project_id: str, dataset_id: str |
| ): |
| full_table_id = load_scalars_table( |
| bigquery_client, project_id, dataset_id, data_path="scalars_extreme.jsonl" |
| ) |
| yield full_table_id |
| bigquery_client.delete_table(full_table_id, not_found_ok=True) |
|
|
|
|
| @pytest.fixture(scope="session", params=["US", TOKYO_LOCATION]) |
| def scalars_table_multi_location( |
| request, scalars_table: str, scalars_table_tokyo: str |
| ) -> Tuple[str, str]: |
| if request.param == "US": |
| full_table_id = scalars_table |
| elif request.param == TOKYO_LOCATION: |
| full_table_id = scalars_table_tokyo |
| else: |
| raise ValueError(f"got unexpected location: {request.param}") |
| return request.param, full_table_id |
|
|
|
|
| @pytest.fixture(scope="session") |
| def scalars_table_csv( |
| bigquery_client: bigquery.Client, project_id: str, dataset_id: str |
| ): |
| full_table_id = load_scalars_table( |
| bigquery_client, |
| project_id, |
| dataset_id, |
| data_path="scalars.csv", |
| source_format=enums.SourceFormat.CSV, |
| schema_source="scalars_schema_csv.json", |
| ) |
| yield full_table_id |
| bigquery_client.delete_table(full_table_id, not_found_ok=True) |
|
|
|
|
| @pytest.fixture |
| def test_table_name(request, replace_non_anum=re.compile(r"[^a-zA-Z0-9_]").sub): |
| return replace_non_anum("_", request.node.name) |
|
|