| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from typing import Iterator |
|
|
| from google.cloud import bigquery |
| import pytest |
| import test_utils.prefixer |
|
|
| prefixer = test_utils.prefixer.Prefixer("python-bigquery", "samples/snippets") |
|
|
|
|
| @pytest.fixture(scope="session", autouse=True) |
| def cleanup_datasets(bigquery_client: bigquery.Client) -> None: |
| for dataset in bigquery_client.list_datasets(): |
| if prefixer.should_cleanup(dataset.dataset_id): |
| bigquery_client.delete_dataset( |
| dataset, delete_contents=True, not_found_ok=True |
| ) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def bigquery_client() -> bigquery.Client: |
| bigquery_client = bigquery.Client() |
| return bigquery_client |
|
|
|
|
| @pytest.fixture(scope="session") |
| def project_id(bigquery_client: bigquery.Client) -> str: |
| return bigquery_client.project |
|
|
|
|
| @pytest.fixture(scope="session") |
| def dataset_id(bigquery_client: bigquery.Client, project_id: str) -> Iterator[str]: |
| dataset_id = prefixer.create_prefix() |
| full_dataset_id = f"{project_id}.{dataset_id}" |
| dataset = bigquery.Dataset(full_dataset_id) |
| bigquery_client.create_dataset(dataset) |
| yield dataset_id |
| bigquery_client.delete_dataset(dataset, delete_contents=True, not_found_ok=True) |
|
|
|
|
| @pytest.fixture |
| def table_id( |
| bigquery_client: bigquery.Client, project_id: str, dataset_id: str |
| ) -> Iterator[str]: |
| table_id = prefixer.create_prefix() |
| full_table_id = f"{project_id}.{dataset_id}.{table_id}" |
| table = bigquery.Table( |
| full_table_id, schema=[bigquery.SchemaField("string_col", "STRING")] |
| ) |
| bigquery_client.create_table(table) |
| yield full_table_id |
| bigquery_client.delete_table(table, not_found_ok=True) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def entity_id(bigquery_client: bigquery.Client, dataset_id: str) -> str: |
| return "cloud-developer-relations@google.com" |
|
|
|
|
| @pytest.fixture(scope="session") |
| def dataset_id_us_east1( |
| bigquery_client: bigquery.Client, |
| project_id: str, |
| ) -> Iterator[str]: |
| dataset_id = prefixer.create_prefix() |
| full_dataset_id = f"{project_id}.{dataset_id}" |
| dataset = bigquery.Dataset(full_dataset_id) |
| dataset.location = "us-east1" |
| bigquery_client.create_dataset(dataset) |
| yield dataset_id |
| bigquery_client.delete_dataset(dataset, delete_contents=True, not_found_ok=True) |
|
|
|
|
| @pytest.fixture(scope="session") |
| def table_id_us_east1( |
| bigquery_client: bigquery.Client, project_id: str, dataset_id_us_east1: str |
| ) -> Iterator[str]: |
| table_id = prefixer.create_prefix() |
| full_table_id = f"{project_id}.{dataset_id_us_east1}.{table_id}" |
| table = bigquery.Table( |
| full_table_id, schema=[bigquery.SchemaField("string_col", "STRING")] |
| ) |
| bigquery_client.create_table(table) |
| yield full_table_id |
| bigquery_client.delete_table(table, not_found_ok=True) |
|
|
|
|
| @pytest.fixture |
| def random_table_id( |
| bigquery_client: bigquery.Client, project_id: str, dataset_id: str |
| ) -> Iterator[str]: |
| """Create a new table ID each time, so random_table_id can be used as |
| target for load jobs. |
| """ |
| random_table_id = prefixer.create_prefix() |
| full_table_id = f"{project_id}.{dataset_id}.{random_table_id}" |
| yield full_table_id |
| bigquery_client.delete_table(full_table_id, not_found_ok=True) |
|
|
|
|
| @pytest.fixture |
| def bigquery_client_patch( |
| monkeypatch: pytest.MonkeyPatch, bigquery_client: bigquery.Client |
| ) -> None: |
| monkeypatch.setattr(bigquery, "Client", lambda: bigquery_client) |
|
|