| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import datetime |
| import logging |
| import re |
| from sys import version_info |
| import time |
| import types |
| import unittest |
| from unittest import mock |
| import warnings |
|
|
| import pytest |
|
|
| import google.api_core.exceptions |
| from test_utils.imports import maybe_fail_import |
|
|
| from google.cloud.bigquery import _versions_helpers |
| from google.cloud.bigquery import exceptions |
| from google.cloud.bigquery.table import TableReference |
| from google.cloud.bigquery.dataset import DatasetReference |
|
|
|
|
| def _mock_client(): |
| from google.cloud.bigquery import client |
|
|
| mock_client = mock.create_autospec(client.Client) |
| mock_client.project = "my-project" |
| return mock_client |
|
|
|
|
| class _SchemaBase(object): |
| def _verify_field(self, field, r_field): |
| self.assertEqual(field.name, r_field["name"]) |
| self.assertEqual(field.field_type, r_field["type"]) |
| self.assertEqual(field.mode, r_field.get("mode", "NULLABLE")) |
|
|
| def _verifySchema(self, schema, resource): |
| r_fields = resource["schema"]["fields"] |
| self.assertEqual(len(schema), len(r_fields)) |
|
|
| for field, r_field in zip(schema, r_fields): |
| self._verify_field(field, r_field) |
|
|
|
|
| class TestEncryptionConfiguration(unittest.TestCase): |
| KMS_KEY_NAME = "projects/1/locations/us/keyRings/1/cryptoKeys/1" |
|
|
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import EncryptionConfiguration |
|
|
| return EncryptionConfiguration |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def test_ctor_defaults(self): |
| encryption_config = self._make_one() |
| self.assertIsNone(encryption_config.kms_key_name) |
|
|
| def test_ctor_with_key(self): |
| encryption_config = self._make_one(kms_key_name=self.KMS_KEY_NAME) |
| self.assertEqual(encryption_config.kms_key_name, self.KMS_KEY_NAME) |
|
|
|
|
| class TestTableBase: |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import _TableBase |
|
|
| return _TableBase |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def test_ctor_defaults(self): |
| instance = self._make_one() |
| assert instance._properties == {} |
|
|
| def test_project(self): |
| instance = self._make_one() |
| instance._properties = {"tableReference": {"projectId": "p_1"}} |
| assert instance.project == "p_1" |
|
|
| def test_dataset_id(self): |
| instance = self._make_one() |
| instance._properties = {"tableReference": {"datasetId": "ds_1"}} |
| assert instance.dataset_id == "ds_1" |
|
|
| def test_table_id(self): |
| instance = self._make_one() |
| instance._properties = {"tableReference": {"tableId": "tbl_1"}} |
| assert instance.table_id == "tbl_1" |
|
|
| def test_path(self): |
| instance = self._make_one() |
| instance._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
| assert instance.path == "/projects/p_1/datasets/ds_1/tables/tbl_1" |
|
|
| def test___eq___wrong_type(self): |
| instance = self._make_one() |
| instance._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
|
|
| class TableWannabe: |
| pass |
|
|
| wannabe_other = TableWannabe() |
| wannabe_other._properties = instance._properties |
| wannabe_other.project = "p_1" |
| wannabe_other.dataset_id = "ds_1" |
| wannabe_other.table_id = "tbl_1" |
|
|
| assert instance != wannabe_other |
| assert instance == mock.ANY |
|
|
| def test___eq___project_mismatch(self): |
| instance = self._make_one() |
| instance._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
| other = self._make_one() |
| other._properties = { |
| "projectId": "p_2", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| assert instance != other |
|
|
| def test___eq___dataset_mismatch(self): |
| instance = self._make_one() |
| instance._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
| other = self._make_one() |
| other._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_2", |
| "tableId": "tbl_1", |
| } |
| } |
| assert instance != other |
|
|
| def test___eq___table_mismatch(self): |
| instance = self._make_one() |
| instance._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
| other = self._make_one() |
| other._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_2", |
| } |
| } |
| assert instance != other |
|
|
| def test___eq___equality(self): |
| instance = self._make_one() |
| instance._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
| other = self._make_one() |
| other._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
| assert instance == other |
|
|
| def test___hash__set_equality(self): |
| instance_1 = self._make_one() |
| instance_1._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
|
|
| instance_2 = self._make_one() |
| instance_2._properties = { |
| "tableReference": { |
| "projectId": "p_2", |
| "datasetId": "ds_2", |
| "tableId": "tbl_2", |
| } |
| } |
|
|
| set_one = {instance_1, instance_2} |
| set_two = {instance_1, instance_2} |
| assert set_one == set_two |
|
|
| def test___hash__sets_not_equal(self): |
| instance_1 = self._make_one() |
| instance_1._properties = { |
| "tableReference": { |
| "projectId": "p_1", |
| "datasetId": "ds_1", |
| "tableId": "tbl_1", |
| } |
| } |
|
|
| instance_2 = self._make_one() |
| instance_2._properties = { |
| "tableReference": { |
| "projectId": "p_2", |
| "datasetId": "ds_2", |
| "tableId": "tbl_2", |
| } |
| } |
|
|
| set_one = {instance_1} |
| set_two = {instance_2} |
| assert set_one != set_two |
|
|
|
|
| class TestTableReference(unittest.TestCase): |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import TableReference |
|
|
| return TableReference |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def test_ctor_defaults(self): |
| dataset_ref = DatasetReference("project_1", "dataset_1") |
|
|
| table_ref = self._make_one(dataset_ref, "table_1") |
| self.assertEqual(table_ref.dataset_id, dataset_ref.dataset_id) |
| self.assertEqual(table_ref.table_id, "table_1") |
|
|
| def test_to_api_repr(self): |
| dataset_ref = DatasetReference("project_1", "dataset_1") |
| table_ref = self._make_one(dataset_ref, "table_1") |
|
|
| resource = table_ref.to_api_repr() |
|
|
| self.assertEqual( |
| resource, |
| {"projectId": "project_1", "datasetId": "dataset_1", "tableId": "table_1"}, |
| ) |
|
|
| def test_from_api_repr(self): |
| from google.cloud.bigquery.table import TableReference |
|
|
| dataset_ref = DatasetReference("project_1", "dataset_1") |
| expected = self._make_one(dataset_ref, "table_1") |
|
|
| got = TableReference.from_api_repr( |
| {"projectId": "project_1", "datasetId": "dataset_1", "tableId": "table_1"} |
| ) |
|
|
| self.assertEqual(expected, got) |
|
|
| def test_from_string(self): |
| cls = self._get_target_class() |
| got = cls.from_string("string-project.string_dataset.string_table") |
| self.assertEqual(got.project, "string-project") |
| self.assertEqual(got.dataset_id, "string_dataset") |
| self.assertEqual(got.table_id, "string_table") |
|
|
| def test_from_string_w_prefix(self): |
| cls = self._get_target_class() |
| got = cls.from_string("google.com:string-project.string_dataset.string_table") |
| self.assertEqual(got.project, "google.com:string-project") |
| self.assertEqual(got.dataset_id, "string_dataset") |
| self.assertEqual(got.table_id, "string_table") |
|
|
| def test_from_string_legacy_string(self): |
| cls = self._get_target_class() |
| with self.assertRaises(ValueError): |
| cls.from_string("string-project:string_dataset.string_table") |
|
|
| def test_from_string_w_incorrect_prefix(self): |
| cls = self._get_target_class() |
| with self.assertRaises(ValueError): |
| cls.from_string("google.com.string-project.string_dataset.string_table") |
|
|
| def test_from_string_not_fully_qualified(self): |
| cls = self._get_target_class() |
| with self.assertRaises(ValueError): |
| cls.from_string("string_table") |
|
|
| with self.assertRaises(ValueError): |
| cls.from_string("string_dataset.string_table") |
|
|
| with self.assertRaises(ValueError): |
| cls.from_string("a.b.c.d") |
|
|
| def test_from_string_with_default_project(self): |
| cls = self._get_target_class() |
| got = cls.from_string( |
| "string_dataset.string_table", default_project="default-project" |
| ) |
| self.assertEqual(got.project, "default-project") |
| self.assertEqual(got.dataset_id, "string_dataset") |
| self.assertEqual(got.table_id, "string_table") |
|
|
| def test_from_string_ignores_default_project(self): |
| cls = self._get_target_class() |
| got = cls.from_string( |
| "string-project.string_dataset.string_table", |
| default_project="default-project", |
| ) |
| self.assertEqual(got.project, "string-project") |
| self.assertEqual(got.dataset_id, "string_dataset") |
| self.assertEqual(got.table_id, "string_table") |
|
|
| def test___repr__(self): |
| dataset = DatasetReference("project1", "dataset1") |
| table1 = self._make_one(dataset, "table1") |
| expected = ( |
| "TableReference(DatasetReference('project1', 'dataset1'), " "'table1')" |
| ) |
| self.assertEqual(repr(table1), expected) |
|
|
| def test___str__(self): |
| dataset = DatasetReference("project1", "dataset1") |
| table1 = self._make_one(dataset, "table1") |
| self.assertEqual(str(table1), "project1.dataset1.table1") |
|
|
|
|
| class TestTable(unittest.TestCase, _SchemaBase): |
| PROJECT = "prahj-ekt" |
| DS_ID = "dataset-name" |
| TABLE_NAME = "table-name" |
| KMS_KEY_NAME = "projects/1/locations/us/keyRings/1/cryptoKeys/1" |
|
|
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import Table |
|
|
| return Table |
|
|
| def _make_one(self, *args, **kw): |
| if len(args) == 0: |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| args = (table_ref,) |
|
|
| return self._get_target_class()(*args, **kw) |
|
|
| def _setUpConstants(self): |
| import datetime |
| from google.cloud._helpers import UTC |
|
|
| self.WHEN_TS = 1437767599.006 |
| self.WHEN = datetime.datetime.utcfromtimestamp(self.WHEN_TS).replace(tzinfo=UTC) |
| self.ETAG = "ETAG" |
| self.TABLE_FULL_ID = "%s:%s.%s" % (self.PROJECT, self.DS_ID, self.TABLE_NAME) |
| self.RESOURCE_URL = "http://example.com/path/to/resource" |
| self.NUM_BYTES = 12345 |
| self.NUM_ROWS = 67 |
| self.NUM_EST_BYTES = 1234 |
| self.NUM_EST_ROWS = 23 |
|
|
| def _make_resource(self): |
| self._setUpConstants() |
| return { |
| "creationTime": self.WHEN_TS * 1000, |
| "tableReference": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.TABLE_NAME, |
| }, |
| "schema": { |
| "fields": [ |
| {"name": "full_name", "type": "STRING", "mode": "REQUIRED"}, |
| {"name": "age", "type": "INTEGER", "mode": "REQUIRED"}, |
| ] |
| }, |
| "etag": "ETAG", |
| "id": self.TABLE_FULL_ID, |
| "lastModifiedTime": self.WHEN_TS * 1000, |
| "location": "US", |
| "selfLink": self.RESOURCE_URL, |
| "numRows": self.NUM_ROWS, |
| "numBytes": self.NUM_BYTES, |
| "type": "TABLE", |
| "streamingBuffer": { |
| "estimatedRows": str(self.NUM_EST_ROWS), |
| "estimatedBytes": str(self.NUM_EST_BYTES), |
| "oldestEntryTime": self.WHEN_TS * 1000, |
| }, |
| "externalDataConfiguration": { |
| "sourceFormat": "CSV", |
| "csvOptions": {"allowJaggedRows": True, "encoding": "encoding"}, |
| }, |
| "labels": {"x": "y"}, |
| } |
|
|
| def _verifyReadonlyResourceProperties(self, table, resource): |
| if "creationTime" in resource: |
| self.assertEqual(table.created, self.WHEN) |
| else: |
| self.assertIsNone(table.created) |
|
|
| if "etag" in resource: |
| self.assertEqual(table.etag, self.ETAG) |
| else: |
| self.assertIsNone(table.etag) |
|
|
| if "numRows" in resource: |
| self.assertEqual(table.num_rows, self.NUM_ROWS) |
| else: |
| self.assertIsNone(table.num_rows) |
|
|
| if "numBytes" in resource: |
| self.assertEqual(table.num_bytes, self.NUM_BYTES) |
| else: |
| self.assertIsNone(table.num_bytes) |
|
|
| if "selfLink" in resource: |
| self.assertEqual(table.self_link, self.RESOURCE_URL) |
| else: |
| self.assertIsNone(table.self_link) |
|
|
| if "streamingBuffer" in resource: |
| self.assertEqual(table.streaming_buffer.estimated_rows, self.NUM_EST_ROWS) |
| self.assertEqual(table.streaming_buffer.estimated_bytes, self.NUM_EST_BYTES) |
| self.assertEqual(table.streaming_buffer.oldest_entry_time, self.WHEN) |
| else: |
| self.assertIsNone(table.streaming_buffer) |
|
|
| self.assertEqual(table.full_table_id, self.TABLE_FULL_ID) |
| self.assertEqual( |
| table.table_type, "TABLE" if "view" not in resource else "VIEW" |
| ) |
|
|
| def _verifyResourceProperties(self, table, resource): |
| self._verifyReadonlyResourceProperties(table, resource) |
|
|
| if "expirationTime" in resource: |
| self.assertEqual(table.expires, self.EXP_TIME) |
| else: |
| self.assertIsNone(table.expires) |
|
|
| self.assertEqual(table.description, resource.get("description")) |
| self.assertEqual(table.friendly_name, resource.get("friendlyName")) |
| self.assertEqual(table.location, resource.get("location")) |
|
|
| if "view" in resource: |
| self.assertEqual(table.view_query, resource["view"]["query"]) |
| self.assertEqual( |
| table.view_use_legacy_sql, resource["view"].get("useLegacySql", True) |
| ) |
| else: |
| self.assertIsNone(table.view_query) |
| self.assertIsNone(table.view_use_legacy_sql) |
|
|
| if "schema" in resource: |
| self._verifySchema(table.schema, resource) |
| else: |
| self.assertEqual(table.schema, []) |
|
|
| if "externalDataConfiguration" in resource: |
| edc = table.external_data_configuration |
| self.assertEqual(edc.source_format, "CSV") |
| self.assertEqual(edc.options.allow_jagged_rows, True) |
|
|
| if "labels" in resource: |
| self.assertEqual(table.labels, {"x": "y"}) |
| else: |
| self.assertEqual(table.labels, {}) |
|
|
| if "encryptionConfiguration" in resource: |
| self.assertIsNotNone(table.encryption_configuration) |
| self.assertEqual( |
| table.encryption_configuration.kms_key_name, |
| resource["encryptionConfiguration"]["kmsKeyName"], |
| ) |
| else: |
| self.assertIsNone(table.encryption_configuration) |
|
|
| def test_ctor(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| self.assertEqual(table.table_id, self.TABLE_NAME) |
| self.assertEqual(table.project, self.PROJECT) |
| self.assertEqual(table.dataset_id, self.DS_ID) |
| self.assertEqual(table.reference.table_id, self.TABLE_NAME) |
| self.assertEqual(table.reference.project, self.PROJECT) |
| self.assertEqual(table.reference.dataset_id, self.DS_ID) |
| self.assertEqual( |
| table.path, |
| "/projects/%s/datasets/%s/tables/%s" |
| % (self.PROJECT, self.DS_ID, self.TABLE_NAME), |
| ) |
| self.assertEqual(table.schema, []) |
|
|
| self.assertIsNone(table.created) |
| self.assertIsNone(table.etag) |
| self.assertIsNone(table.modified) |
| self.assertIsNone(table.num_bytes) |
| self.assertIsNone(table.num_rows) |
| self.assertIsNone(table.self_link) |
| self.assertIsNone(table.full_table_id) |
| self.assertIsNone(table.table_type) |
| self.assertIsNone(table.description) |
| self.assertIsNone(table.expires) |
| self.assertIsNone(table.friendly_name) |
| self.assertIsNone(table.location) |
| self.assertIsNone(table.view_query) |
| self.assertIsNone(table.view_use_legacy_sql) |
| self.assertIsNone(table.external_data_configuration) |
| self.assertEqual(table.labels, {}) |
| self.assertIsNone(table.encryption_configuration) |
| self.assertIsNone(table.time_partitioning) |
| self.assertIsNone(table.clustering_fields) |
| self.assertIsNone(table.table_constraints) |
|
|
| def test_ctor_w_schema(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| full_name = SchemaField("full_name", "STRING", mode="REQUIRED") |
| age = SchemaField("age", "INTEGER", mode="REQUIRED") |
| table = self._make_one(table_ref, schema=[full_name, age]) |
|
|
| self.assertEqual(table.schema, [full_name, age]) |
|
|
| def test_ctor_string(self): |
| table = self._make_one("some-project.some_dset.some_tbl") |
| self.assertEqual(table.project, "some-project") |
| self.assertEqual(table.dataset_id, "some_dset") |
| self.assertEqual(table.table_id, "some_tbl") |
|
|
| def test_ctor_tablelistitem(self): |
| from google.cloud.bigquery.table import Table, TableListItem |
|
|
| import datetime |
| from google.cloud._helpers import _millis, UTC |
|
|
| self.WHEN_TS = 1437767599.125 |
| self.EXP_TIME = datetime.datetime(2015, 8, 1, 23, 59, 59, tzinfo=UTC) |
|
|
| project = "test-project" |
| dataset_id = "test_dataset" |
| table_id = "coffee_table" |
| resource = { |
| "creationTime": self.WHEN_TS * 1000, |
| "expirationTime": _millis(self.EXP_TIME), |
| "kind": "bigquery#table", |
| "id": "{}:{}.{}".format(project, dataset_id, table_id), |
| "tableReference": { |
| "projectId": project, |
| "datasetId": dataset_id, |
| "tableId": table_id, |
| }, |
| "friendlyName": "Mahogany Coffee Table", |
| "type": "TABLE", |
| "timePartitioning": { |
| "type": "DAY", |
| "field": "mycolumn", |
| "expirationMs": "10000", |
| }, |
| "labels": {"some-stuff": "this-is-a-label"}, |
| "clustering": {"fields": ["string"]}, |
| } |
|
|
| table_list_item = TableListItem(resource) |
| table = Table(table_list_item) |
|
|
| self.assertIsNone(table.created) |
| self.assertEqual(table.reference.project, project) |
| self.assertEqual(table.reference.dataset_id, dataset_id) |
| self.assertEqual(table.reference.table_id, table_id) |
|
|
| def test_ctor_string_wo_project_id(self): |
| with pytest.raises(ValueError): |
| |
| self._make_one("some_dset.some_tbl") |
|
|
| def test_num_bytes_getter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| |
| self.assertIsNone(table.num_bytes) |
|
|
| num_bytes = 1337 |
| |
| table._properties = {"numBytes": num_bytes} |
| self.assertEqual(table.num_bytes, num_bytes) |
|
|
| |
| table._properties = {"numBytes": str(num_bytes)} |
| self.assertEqual(table.num_bytes, num_bytes) |
|
|
| |
| table._properties = {"numBytes": "x"} |
| with self.assertRaises(ValueError): |
| getattr(table, "num_bytes") |
|
|
| def test_num_rows_getter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| |
| self.assertIsNone(table.num_rows) |
|
|
| num_rows = 42 |
| |
| table._properties = {"numRows": num_rows} |
| self.assertEqual(table.num_rows, num_rows) |
|
|
| |
| table._properties = {"numRows": str(num_rows)} |
| self.assertEqual(table.num_rows, num_rows) |
|
|
| |
| table._properties = {"numRows": "x"} |
| with self.assertRaises(ValueError): |
| getattr(table, "num_rows") |
|
|
| def test__eq__same_table_property_different(self): |
| table_1 = self._make_one("project_foo.dataset_bar.table_baz") |
| table_1.description = "This is table baz" |
|
|
| table_2 = self._make_one("project_foo.dataset_bar.table_baz") |
| table_2.description = "This is also table baz" |
|
|
| assert table_1 == table_2 |
|
|
| def test_hashable(self): |
| table_1 = self._make_one("project_foo.dataset_bar.table_baz") |
| table_1.description = "This is a table" |
|
|
| table_1b = self._make_one("project_foo.dataset_bar.table_baz") |
| table_1b.description = "Metadata is irrelevant for hashes" |
|
|
| assert hash(table_1) == hash(table_1b) |
|
|
| def test_schema_setter_non_sequence(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(TypeError): |
| table.schema = object() |
|
|
| def test_schema_setter_invalid_field(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| full_name = SchemaField("full_name", "STRING", mode="REQUIRED") |
| with self.assertRaises(ValueError): |
| table.schema = [full_name, object()] |
|
|
| def test_schema_setter_valid_fields(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| full_name = SchemaField("full_name", "STRING", mode="REQUIRED") |
| age = SchemaField("age", "INTEGER", mode="REQUIRED") |
| table.schema = [full_name, age] |
| self.assertEqual(table.schema, [full_name, age]) |
|
|
| def test_schema_setter_invalid_mapping_representation(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| full_name = {"name": "full_name", "type": "STRING", "mode": "REQUIRED"} |
| invalid_field = {"name": "full_name", "typeooo": "STRING", "mode": "REQUIRED"} |
| with self.assertRaises(Exception): |
| table.schema = [full_name, invalid_field] |
|
|
| def test_schema_setter_valid_mapping_representation(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| full_name = {"name": "full_name", "type": "STRING", "mode": "REQUIRED"} |
| job_status = { |
| "name": "is_employed", |
| "type": "STRUCT", |
| "mode": "NULLABLE", |
| "fields": [ |
| {"name": "foo", "type": "DATE", "mode": "NULLABLE"}, |
| {"name": "bar", "type": "BYTES", "mode": "REQUIRED"}, |
| ], |
| } |
|
|
| table.schema = [full_name, job_status] |
|
|
| expected_schema = [ |
| SchemaField("full_name", "STRING", mode="REQUIRED"), |
| SchemaField( |
| "is_employed", |
| "STRUCT", |
| mode="NULLABLE", |
| fields=[ |
| SchemaField("foo", "DATE", mode="NULLABLE"), |
| SchemaField("bar", "BYTES", mode="REQUIRED"), |
| ], |
| ), |
| ] |
| self.assertEqual(table.schema, expected_schema) |
|
|
| def test_props_set_by_server(self): |
| import datetime |
| from google.cloud._helpers import UTC |
| from google.cloud._helpers import _millis |
|
|
| CREATED = datetime.datetime(2015, 7, 29, 12, 13, 22, tzinfo=UTC) |
| MODIFIED = datetime.datetime(2015, 7, 29, 14, 47, 15, tzinfo=UTC) |
| TABLE_FULL_ID = "%s:%s.%s" % (self.PROJECT, self.DS_ID, self.TABLE_NAME) |
| URL = "http://example.com/projects/%s/datasets/%s/tables/%s" % ( |
| self.PROJECT, |
| self.DS_ID, |
| self.TABLE_NAME, |
| ) |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table._properties["creationTime"] = _millis(CREATED) |
| table._properties["etag"] = "ETAG" |
| table._properties["lastModifiedTime"] = _millis(MODIFIED) |
| table._properties["numBytes"] = 12345 |
| table._properties["numRows"] = 66 |
| table._properties["selfLink"] = URL |
| table._properties["id"] = TABLE_FULL_ID |
| table._properties["type"] = "TABLE" |
|
|
| self.assertEqual(table.created, CREATED) |
| self.assertEqual(table.etag, "ETAG") |
| self.assertEqual(table.modified, MODIFIED) |
| self.assertEqual(table.num_bytes, 12345) |
| self.assertEqual(table.num_rows, 66) |
| self.assertEqual(table.self_link, URL) |
| self.assertEqual(table.full_table_id, TABLE_FULL_ID) |
| self.assertEqual(table.table_type, "TABLE") |
|
|
| def test_snapshot_definition_not_set(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| assert table.snapshot_definition is None |
|
|
| def test_snapshot_definition_set(self): |
| from google.cloud._helpers import UTC |
| from google.cloud.bigquery.table import SnapshotDefinition |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table._properties["snapshotDefinition"] = { |
| "baseTableReference": { |
| "projectId": "project_x", |
| "datasetId": "dataset_y", |
| "tableId": "table_z", |
| }, |
| "snapshotTime": "2010-09-28T10:20:30.123Z", |
| } |
|
|
| snapshot = table.snapshot_definition |
|
|
| assert isinstance(snapshot, SnapshotDefinition) |
| assert snapshot.base_table_reference.path == ( |
| "/projects/project_x/datasets/dataset_y/tables/table_z" |
| ) |
| assert snapshot.snapshot_time == datetime.datetime( |
| 2010, 9, 28, 10, 20, 30, 123000, tzinfo=UTC |
| ) |
|
|
| def test_clone_definition_not_set(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| assert table.clone_definition is None |
|
|
| def test_clone_definition_set(self): |
| from google.cloud._helpers import UTC |
| from google.cloud.bigquery.table import CloneDefinition |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table._properties["cloneDefinition"] = { |
| "baseTableReference": { |
| "projectId": "project_x", |
| "datasetId": "dataset_y", |
| "tableId": "table_z", |
| }, |
| "cloneTime": "2010-09-28T10:20:30.123Z", |
| } |
|
|
| clone = table.clone_definition |
|
|
| assert isinstance(clone, CloneDefinition) |
| assert clone.base_table_reference.path == ( |
| "/projects/project_x/datasets/dataset_y/tables/table_z" |
| ) |
| assert clone.clone_time == datetime.datetime( |
| 2010, 9, 28, 10, 20, 30, 123000, tzinfo=UTC |
| ) |
|
|
| def test_table_constraints_property_getter(self): |
| from google.cloud.bigquery.table import PrimaryKey, TableConstraints |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table._properties["tableConstraints"] = { |
| "primaryKey": {"columns": ["id"]}, |
| } |
|
|
| table_constraints = table.table_constraints |
|
|
| assert isinstance(table_constraints, TableConstraints) |
| assert table_constraints.primary_key == PrimaryKey(columns=["id"]) |
|
|
| def test_description_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.description = 12345 |
|
|
| def test_description_setter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.description = "DESCRIPTION" |
| self.assertEqual(table.description, "DESCRIPTION") |
|
|
| def test_expires_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.expires = object() |
|
|
| def test_expires_setter(self): |
| import datetime |
| from google.cloud._helpers import UTC |
|
|
| WHEN = datetime.datetime(2015, 7, 28, 16, 39, tzinfo=UTC) |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.expires = WHEN |
| self.assertEqual(table.expires, WHEN) |
|
|
| def test_friendly_name_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.friendly_name = 12345 |
|
|
| def test_friendly_name_setter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.friendly_name = "FRIENDLY" |
| self.assertEqual(table.friendly_name, "FRIENDLY") |
|
|
| def test_view_query_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.view_query = 12345 |
|
|
| def test_view_query_setter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.view_query = "select * from foo" |
| self.assertEqual(table.view_query, "select * from foo") |
| self.assertEqual(table.view_use_legacy_sql, False) |
|
|
| table.view_use_legacy_sql = True |
| self.assertEqual(table.view_use_legacy_sql, True) |
|
|
| def test_view_query_deleter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.view_query = "select * from foo" |
| del table.view_query |
| self.assertIsNone(table.view_query) |
| self.assertIsNone(table.view_use_legacy_sql) |
|
|
| def test_view_use_legacy_sql_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.view_use_legacy_sql = 12345 |
|
|
| def test_view_use_legacy_sql_setter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.view_use_legacy_sql = True |
| table.view_query = "select * from foo" |
| self.assertEqual(table.view_use_legacy_sql, True) |
| self.assertEqual(table.view_query, "select * from foo") |
|
|
| def test_external_data_configuration_setter(self): |
| from google.cloud.bigquery.external_config import ExternalConfig |
|
|
| external_config = ExternalConfig("CSV") |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table.external_data_configuration = external_config |
|
|
| self.assertEqual( |
| table.external_data_configuration.source_format, |
| external_config.source_format, |
| ) |
|
|
| def test_external_data_configuration_setter_none(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table.external_data_configuration = None |
|
|
| self.assertIsNone(table.external_data_configuration) |
|
|
| def test_external_data_configuration_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.external_data_configuration = 12345 |
|
|
| def test_labels_update_in_place(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| del table._properties["labels"] |
| labels = table.labels |
| labels["foo"] = "bar" |
| self.assertEqual(table.labels, {"foo": "bar"}) |
|
|
| def test_labels_setter_bad_value(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| with self.assertRaises(ValueError): |
| table.labels = 12345 |
|
|
| def test_mview_query(self): |
| table = self._make_one() |
| self.assertIsNone(table.mview_query) |
| table.mview_query = "SELECT name, SUM(number) FROM dset.tbl GROUP BY 1" |
| self.assertEqual( |
| table.mview_query, "SELECT name, SUM(number) FROM dset.tbl GROUP BY 1" |
| ) |
| del table.mview_query |
| self.assertIsNone(table.mview_query) |
|
|
| def test_mview_last_refresh_time(self): |
| table = self._make_one() |
| self.assertIsNone(table.mview_last_refresh_time) |
| table._properties["materializedView"] = { |
| "lastRefreshTime": "1606751842496", |
| } |
| self.assertEqual( |
| table.mview_last_refresh_time, |
| datetime.datetime( |
| 2020, 11, 30, 15, 57, 22, 496000, tzinfo=datetime.timezone.utc |
| ), |
| ) |
|
|
| def test_mview_enable_refresh(self): |
| table = self._make_one() |
| self.assertIsNone(table.mview_enable_refresh) |
| table.mview_enable_refresh = True |
| self.assertTrue(table.mview_enable_refresh) |
| table.mview_enable_refresh = False |
| self.assertFalse(table.mview_enable_refresh) |
| table.mview_enable_refresh = None |
| self.assertIsNone(table.mview_enable_refresh) |
|
|
| def test_mview_refresh_interval(self): |
| table = self._make_one() |
| self.assertIsNone(table.mview_refresh_interval) |
| table.mview_refresh_interval = datetime.timedelta(minutes=30) |
| self.assertEqual(table.mview_refresh_interval, datetime.timedelta(minutes=30)) |
| self.assertEqual( |
| table._properties["materializedView"]["refreshIntervalMs"], "1800000" |
| ) |
| table.mview_refresh_interval = None |
| self.assertIsNone(table.mview_refresh_interval) |
|
|
| def test_from_string(self): |
| cls = self._get_target_class() |
| got = cls.from_string("string-project.string_dataset.string_table") |
| self.assertEqual(got.project, "string-project") |
| self.assertEqual(got.dataset_id, "string_dataset") |
| self.assertEqual(got.table_id, "string_table") |
| self.assertEqual( |
| str(got.reference), "string-project.string_dataset.string_table" |
| ) |
|
|
| def test_from_string_legacy_string(self): |
| cls = self._get_target_class() |
| with self.assertRaises(ValueError): |
| cls.from_string("string-project:string_dataset.string_table") |
|
|
| def test_from_string_not_fully_qualified(self): |
| cls = self._get_target_class() |
| with self.assertRaises(ValueError): |
| cls.from_string("string_dataset.string_table") |
|
|
| def test_from_api_repr_missing_identity(self): |
| self._setUpConstants() |
| RESOURCE = {} |
| klass = self._get_target_class() |
| with self.assertRaises(KeyError): |
| klass.from_api_repr(RESOURCE) |
|
|
| def test_from_api_repr_bare(self): |
| self._setUpConstants() |
| RESOURCE = { |
| "id": "%s:%s.%s" % (self.PROJECT, self.DS_ID, self.TABLE_NAME), |
| "tableReference": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.TABLE_NAME, |
| }, |
| "type": "TABLE", |
| } |
| klass = self._get_target_class() |
| table = klass.from_api_repr(RESOURCE) |
| self.assertEqual(table.table_id, self.TABLE_NAME) |
| self._verifyResourceProperties(table, RESOURCE) |
|
|
| def test_from_api_repr_w_properties(self): |
| import datetime |
| from google.cloud._helpers import UTC |
| from google.cloud._helpers import _millis |
|
|
| RESOURCE = self._make_resource() |
| RESOURCE["view"] = {"query": "select fullname, age from person_ages"} |
| RESOURCE["type"] = "VIEW" |
| RESOURCE["location"] = "EU" |
| self.EXP_TIME = datetime.datetime(2015, 8, 1, 23, 59, 59, tzinfo=UTC) |
| RESOURCE["expirationTime"] = _millis(self.EXP_TIME) |
| klass = self._get_target_class() |
| table = klass.from_api_repr(RESOURCE) |
| self._verifyResourceProperties(table, RESOURCE) |
|
|
| def test_from_api_repr_w_partial_streamingbuffer(self): |
| import datetime |
| from google.cloud._helpers import UTC |
| from google.cloud._helpers import _millis |
|
|
| RESOURCE = self._make_resource() |
| self.OLDEST_TIME = datetime.datetime(2015, 8, 1, 23, 59, 59, tzinfo=UTC) |
| RESOURCE["streamingBuffer"] = {"oldestEntryTime": _millis(self.OLDEST_TIME)} |
| klass = self._get_target_class() |
| table = klass.from_api_repr(RESOURCE) |
| self.assertIsNotNone(table.streaming_buffer) |
| self.assertIsNone(table.streaming_buffer.estimated_rows) |
| self.assertIsNone(table.streaming_buffer.estimated_bytes) |
| self.assertEqual(table.streaming_buffer.oldest_entry_time, self.OLDEST_TIME) |
| |
| RESOURCE["streamingBuffer"] = {"estimatedRows": 1} |
| klass = self._get_target_class() |
| table = klass.from_api_repr(RESOURCE) |
| self.assertIsNotNone(table.streaming_buffer) |
| self.assertEqual(table.streaming_buffer.estimated_rows, 1) |
| self.assertIsNone(table.streaming_buffer.estimated_bytes) |
| self.assertIsNone(table.streaming_buffer.oldest_entry_time) |
|
|
| def test_from_api_with_encryption(self): |
| self._setUpConstants() |
| RESOURCE = { |
| "id": "%s:%s.%s" % (self.PROJECT, self.DS_ID, self.TABLE_NAME), |
| "tableReference": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.TABLE_NAME, |
| }, |
| "encryptionConfiguration": {"kmsKeyName": self.KMS_KEY_NAME}, |
| "type": "TABLE", |
| } |
| klass = self._get_target_class() |
| table = klass.from_api_repr(RESOURCE) |
| self._verifyResourceProperties(table, RESOURCE) |
|
|
| def test_to_api_repr_w_custom_field(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table._properties["newAlphaProperty"] = "unreleased property" |
| resource = table.to_api_repr() |
|
|
| exp_resource = { |
| "tableReference": table_ref.to_api_repr(), |
| "labels": {}, |
| "newAlphaProperty": "unreleased property", |
| } |
| self.assertEqual(resource, exp_resource) |
|
|
| def test_to_api_repr_w_unsetting_expiration(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.partition_expiration = None |
| resource = table.to_api_repr() |
|
|
| exp_resource = { |
| "tableReference": table_ref.to_api_repr(), |
| "labels": {}, |
| "timePartitioning": { |
| "expirationMs": None, |
| "type": TimePartitioningType.DAY, |
| }, |
| } |
| self.assertEqual(resource, exp_resource) |
|
|
| def test__build_resource_w_custom_field(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table._properties["newAlphaProperty"] = "unreleased property" |
| resource = table._build_resource(["newAlphaProperty"]) |
|
|
| exp_resource = {"newAlphaProperty": "unreleased property"} |
| self.assertEqual(resource, exp_resource) |
|
|
| def test__build_resource_w_custom_field_not_in__properties(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table = self._make_one(dataset.table(self.TABLE_NAME)) |
| table.bad = "value" |
| with self.assertRaises(ValueError): |
| table._build_resource(["bad"]) |
|
|
| def test_range_partitioning(self): |
| from google.cloud.bigquery.table import RangePartitioning |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| table = self._make_one("proj.dset.tbl") |
| assert table.range_partitioning is None |
|
|
| table.range_partitioning = RangePartitioning( |
| field="col1", range_=PartitionRange(start=-512, end=1024, interval=128) |
| ) |
| assert table.range_partitioning.field == "col1" |
| assert table.range_partitioning.range_.start == -512 |
| assert table.range_partitioning.range_.end == 1024 |
| assert table.range_partitioning.range_.interval == 128 |
|
|
| table.range_partitioning = None |
| assert table.range_partitioning is None |
|
|
| def test_range_partitioning_w_wrong_type(self): |
| object_under_test = self._make_one("proj.dset.tbl") |
| with pytest.raises(ValueError, match="RangePartitioning"): |
| object_under_test.range_partitioning = object() |
|
|
| def test_require_partitioning_filter(self): |
| table = self._make_one("proj.dset.tbl") |
| assert table.require_partition_filter is None |
| table.require_partition_filter = True |
| assert table.require_partition_filter |
| table.require_partition_filter = False |
| assert table.require_partition_filter is not None |
| assert not table.require_partition_filter |
| table.require_partition_filter = None |
| assert table.require_partition_filter is None |
|
|
| def test_time_partitioning_getter(self): |
| from google.cloud.bigquery.table import TimePartitioning |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table._properties["timePartitioning"] = { |
| "type": "DAY", |
| "field": "col1", |
| "expirationMs": "123456", |
| "requirePartitionFilter": False, |
| } |
| self.assertIsInstance(table.time_partitioning, TimePartitioning) |
| self.assertEqual(table.time_partitioning.type_, TimePartitioningType.DAY) |
| self.assertEqual(table.time_partitioning.field, "col1") |
| self.assertEqual(table.time_partitioning.expiration_ms, 123456) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertFalse(table.time_partitioning.require_partition_filter) |
|
|
| assert len(warned) == 1 |
| self.assertIs(warned[0].category, PendingDeprecationWarning) |
|
|
| def test_time_partitioning_getter_w_none(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table._properties["timePartitioning"] = None |
| self.assertIsNone(table.time_partitioning) |
|
|
| del table._properties["timePartitioning"] |
| self.assertIsNone(table.time_partitioning) |
|
|
| def test_time_partitioning_getter_w_empty(self): |
| from google.cloud.bigquery.table import TimePartitioning |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| |
| |
| |
| table._properties["timePartitioning"] = {} |
| self.assertIsInstance(table.time_partitioning, TimePartitioning) |
| self.assertIsNone(table.time_partitioning.type_) |
| self.assertIsNone(table.time_partitioning.field) |
| self.assertIsNone(table.time_partitioning.expiration_ms) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertIsNone(table.time_partitioning.require_partition_filter) |
|
|
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_time_partitioning_setter(self): |
| from google.cloud.bigquery.table import TimePartitioning |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| time_partitioning = TimePartitioning(type_=TimePartitioningType.HOUR) |
|
|
| table.time_partitioning = time_partitioning |
|
|
| self.assertEqual(table.time_partitioning.type_, TimePartitioningType.HOUR) |
| |
| self.assertIs( |
| table._properties["timePartitioning"], time_partitioning._properties |
| ) |
|
|
| time_partitioning.expiration_ms = 10000 |
|
|
| |
| self.assertEqual( |
| table.time_partitioning.expiration_ms, time_partitioning.expiration_ms |
| ) |
|
|
| def test_time_partitioning_setter_bad_type(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| with self.assertRaises(ValueError): |
| table.time_partitioning = {"timePartitioning": {"type": "DAY"}} |
|
|
| def test_time_partitioning_setter_none(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table.time_partitioning = None |
|
|
| self.assertIsNone(table.time_partitioning) |
|
|
| def test_partitioning_type_setter(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertIsNone(table.partitioning_type) |
|
|
| table.partitioning_type = TimePartitioningType.DAY |
|
|
| self.assertEqual(table.partitioning_type, "DAY") |
|
|
| self.assertEqual(len(warned), 3) |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_partitioning_type_setter_w_time_partitioning_set(self): |
| from google.cloud.bigquery.table import TimePartitioning |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.time_partitioning = TimePartitioning() |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| table.partitioning_type = "NEW_FAKE_TYPE" |
|
|
| self.assertEqual(table.partitioning_type, "NEW_FAKE_TYPE") |
|
|
| self.assertEqual(len(warned), 2) |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_partitioning_expiration_setter_w_time_partitioning_set(self): |
| from google.cloud.bigquery.table import TimePartitioning |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| table.time_partitioning = TimePartitioning() |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| table.partition_expiration = 100000 |
|
|
| self.assertEqual(table.partition_expiration, 100000) |
|
|
| self.assertEqual(len(warned), 2) |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_partition_expiration_setter(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertIsNone(table.partition_expiration) |
|
|
| table.partition_expiration = 100 |
|
|
| self.assertEqual(table.partition_expiration, 100) |
| |
| self.assertEqual(table.partitioning_type, "DAY") |
|
|
| self.assertEqual(len(warned), 4) |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_clustering_fields_setter_w_fields(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| fields = ["email", "phone"] |
|
|
| table.clustering_fields = fields |
| self.assertEqual(table.clustering_fields, fields) |
| self.assertEqual(table._properties["clustering"], {"fields": fields}) |
|
|
| def test_clustering_fields_setter_w_none(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| fields = ["email", "phone"] |
|
|
| table._properties["clustering"] = {"fields": fields} |
| table.clustering_fields = None |
| self.assertIsNone(table.clustering_fields) |
| self.assertTrue("clustering" in table._properties) |
|
|
| def test_clustering_fields_setter_w_none_noop(self): |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
|
|
| table.clustering_fields = None |
| self.assertIsNone(table.clustering_fields) |
| self.assertTrue("clustering" in table._properties) |
|
|
| def test_encryption_configuration_setter(self): |
| |
| |
| |
| |
| from google.cloud.bigquery.table import EncryptionConfiguration |
|
|
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = self._make_one(table_ref) |
| encryption_configuration = EncryptionConfiguration( |
| kms_key_name=self.KMS_KEY_NAME |
| ) |
| table.encryption_configuration = encryption_configuration |
| self.assertEqual(table.encryption_configuration.kms_key_name, self.KMS_KEY_NAME) |
| table.encryption_configuration = None |
| self.assertIsNone(table.encryption_configuration) |
|
|
| def test___repr__(self): |
| from google.cloud.bigquery.table import TableReference |
|
|
| dataset = DatasetReference("project1", "dataset1") |
| table1 = self._make_one(TableReference(dataset, "table1")) |
| expected = ( |
| "Table(TableReference(" |
| "DatasetReference('project1', 'dataset1'), " |
| "'table1'))" |
| ) |
| self.assertEqual(repr(table1), expected) |
|
|
| def test___str__(self): |
| dataset = DatasetReference("project1", "dataset1") |
| table1 = self._make_one(TableReference(dataset, "table1")) |
| self.assertEqual(str(table1), "project1.dataset1.table1") |
|
|
|
|
| class Test_row_from_mapping(unittest.TestCase, _SchemaBase): |
| PROJECT = "prahj-ekt" |
| DS_ID = "dataset-name" |
| TABLE_NAME = "table-name" |
|
|
| def _call_fut(self, mapping, schema): |
| from google.cloud.bigquery.table import _row_from_mapping |
|
|
| return _row_from_mapping(mapping, schema) |
|
|
| def test__row_from_mapping_wo_schema(self): |
| from google.cloud.bigquery.table import Table, _TABLE_HAS_NO_SCHEMA |
|
|
| MAPPING = {"full_name": "Phred Phlyntstone", "age": 32} |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| table = Table(table_ref) |
|
|
| with self.assertRaises(ValueError) as exc: |
| self._call_fut(MAPPING, table.schema) |
|
|
| self.assertEqual(exc.exception.args, (_TABLE_HAS_NO_SCHEMA,)) |
|
|
| def test__row_from_mapping_w_invalid_schema(self): |
| from google.cloud.bigquery.schema import SchemaField |
| from google.cloud.bigquery.table import Table |
|
|
| MAPPING = { |
| "full_name": "Phred Phlyntstone", |
| "age": 32, |
| "colors": ["red", "green"], |
| "bogus": "WHATEVER", |
| } |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| full_name = SchemaField("full_name", "STRING", mode="REQUIRED") |
| age = SchemaField("age", "INTEGER", mode="REQUIRED") |
| colors = SchemaField("colors", "DATETIME", mode="REPEATED") |
| bogus = SchemaField("joined", "STRING", mode="BOGUS") |
| table = Table(table_ref, schema=[full_name, age, colors, bogus]) |
|
|
| with self.assertRaises(ValueError) as exc: |
| self._call_fut(MAPPING, table.schema) |
|
|
| self.assertIn("Unknown field mode: BOGUS", str(exc.exception)) |
|
|
| def test__row_from_mapping_w_schema(self): |
| from google.cloud.bigquery.schema import SchemaField |
| from google.cloud.bigquery.table import Table |
|
|
| MAPPING = { |
| "full_name": "Phred Phlyntstone", |
| "age": 32, |
| "colors": ["red", "green"], |
| "extra": "IGNORED", |
| } |
| dataset = DatasetReference(self.PROJECT, self.DS_ID) |
| table_ref = dataset.table(self.TABLE_NAME) |
| full_name = SchemaField("full_name", "STRING", mode="REQUIRED") |
| age = SchemaField("age", "INTEGER", mode="REQUIRED") |
| colors = SchemaField("colors", "DATETIME", mode="REPEATED") |
| joined = SchemaField("joined", "STRING", mode="NULLABLE") |
| table = Table(table_ref, schema=[full_name, age, colors, joined]) |
|
|
| self.assertEqual( |
| self._call_fut(MAPPING, table.schema), |
| ("Phred Phlyntstone", 32, ["red", "green"], None), |
| ) |
|
|
|
|
| class TestTableListItem(unittest.TestCase): |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import TableListItem |
|
|
| return TableListItem |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def _setUpConstants(self): |
| from google.cloud._helpers import UTC |
|
|
| self.WHEN_TS = 1437767599.125 |
| self.WHEN = datetime.datetime.utcfromtimestamp(self.WHEN_TS).replace(tzinfo=UTC) |
| self.EXP_TIME = datetime.datetime(2015, 8, 1, 23, 59, 59, tzinfo=UTC) |
|
|
| def test_ctor(self): |
| from google.cloud._helpers import _millis |
|
|
| self._setUpConstants() |
| project = "test-project" |
| dataset_id = "test_dataset" |
| table_id = "coffee_table" |
| resource = { |
| "creationTime": self.WHEN_TS * 1000, |
| "expirationTime": _millis(self.EXP_TIME), |
| "kind": "bigquery#table", |
| "id": "{}:{}.{}".format(project, dataset_id, table_id), |
| "tableReference": { |
| "projectId": project, |
| "datasetId": dataset_id, |
| "tableId": table_id, |
| }, |
| "friendlyName": "Mahogany Coffee Table", |
| "type": "TABLE", |
| "timePartitioning": { |
| "type": "DAY", |
| "field": "mycolumn", |
| "expirationMs": "10000", |
| }, |
| "labels": {"some-stuff": "this-is-a-label"}, |
| "clustering": {"fields": ["string"]}, |
| } |
|
|
| table = self._make_one(resource) |
|
|
| self.assertEqual(table.created, self.WHEN) |
| self.assertEqual(table.expires, self.EXP_TIME) |
| self.assertEqual(table.project, project) |
| self.assertEqual(table.dataset_id, dataset_id) |
| self.assertEqual(table.table_id, table_id) |
| self.assertEqual( |
| table.full_table_id, "{}:{}.{}".format(project, dataset_id, table_id) |
| ) |
| self.assertEqual(table.reference.project, project) |
| self.assertEqual(table.reference.dataset_id, dataset_id) |
| self.assertEqual(table.reference.table_id, table_id) |
| self.assertEqual(table.friendly_name, "Mahogany Coffee Table") |
| self.assertEqual(table.table_type, "TABLE") |
| self.assertEqual(table.time_partitioning.type_, "DAY") |
| self.assertEqual(table.time_partitioning.expiration_ms, 10000) |
| self.assertEqual(table.time_partitioning.field, "mycolumn") |
| self.assertEqual(table.labels["some-stuff"], "this-is-a-label") |
| self.assertIsNone(table.view_use_legacy_sql) |
| self.assertEqual(table.clustering_fields, ["string"]) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertEqual(table.partitioning_type, "DAY") |
| self.assertEqual(table.partition_expiration, 10000) |
|
|
| self.assertEqual(len(warned), 2) |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_ctor_view(self): |
| project = "test-project" |
| dataset_id = "test_dataset" |
| table_id = "just_looking" |
| resource = { |
| "kind": "bigquery#table", |
| "id": "{}:{}.{}".format(project, dataset_id, table_id), |
| "tableReference": { |
| "projectId": project, |
| "datasetId": dataset_id, |
| "tableId": table_id, |
| }, |
| "type": "VIEW", |
| } |
|
|
| table = self._make_one(resource) |
| self.assertEqual(table.project, project) |
| self.assertEqual(table.dataset_id, dataset_id) |
| self.assertEqual(table.table_id, table_id) |
| self.assertEqual( |
| table.full_table_id, "{}:{}.{}".format(project, dataset_id, table_id) |
| ) |
| self.assertEqual(table.reference.project, project) |
| self.assertEqual(table.reference.dataset_id, dataset_id) |
| self.assertEqual(table.reference.table_id, table_id) |
| self.assertEqual(table.table_type, "VIEW") |
| |
| self.assertTrue(table.view_use_legacy_sql) |
|
|
| def test_ctor_missing_properties(self): |
| resource = { |
| "tableReference": { |
| "projectId": "testproject", |
| "datasetId": "testdataset", |
| "tableId": "testtable", |
| } |
| } |
| table = self._make_one(resource) |
| self.assertEqual(table.project, "testproject") |
| self.assertEqual(table.dataset_id, "testdataset") |
| self.assertEqual(table.table_id, "testtable") |
| self.assertIsNone(table.created) |
| self.assertIsNone(table.expires) |
| self.assertIsNone(table.clustering_fields) |
| self.assertIsNone(table.full_table_id) |
| self.assertIsNone(table.friendly_name) |
| self.assertIsNone(table.table_type) |
| self.assertIsNone(table.time_partitioning) |
| self.assertEqual(table.labels, {}) |
| self.assertIsNone(table.view_use_legacy_sql) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertIsNone(table.partitioning_type) |
| self.assertIsNone(table.partition_expiration) |
|
|
| self.assertEqual(len(warned), 2) |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_ctor_wo_project(self): |
| resource = { |
| "tableReference": {"datasetId": "testdataset", "tableId": "testtable"} |
| } |
| with self.assertRaises(ValueError): |
| self._make_one(resource) |
|
|
| def test_ctor_wo_dataset(self): |
| resource = { |
| "tableReference": {"projectId": "testproject", "tableId": "testtable"} |
| } |
| with self.assertRaises(ValueError): |
| self._make_one(resource) |
|
|
| def test_ctor_wo_table(self): |
| resource = { |
| "tableReference": {"projectId": "testproject", "datasetId": "testdataset"} |
| } |
| with self.assertRaises(ValueError): |
| self._make_one(resource) |
|
|
| def test_ctor_wo_reference(self): |
| with self.assertRaises(ValueError): |
| self._make_one({}) |
|
|
| def test_labels_update_in_place(self): |
| resource = { |
| "tableReference": { |
| "projectId": "testproject", |
| "datasetId": "testdataset", |
| "tableId": "testtable", |
| } |
| } |
| table = self._make_one(resource) |
| labels = table.labels |
| labels["foo"] = "bar" |
| self.assertEqual(table.labels, {"foo": "bar"}) |
|
|
| def test_to_api_repr(self): |
| resource = { |
| "tableReference": { |
| "projectId": "testproject", |
| "datasetId": "testdataset", |
| "tableId": "testtable", |
| } |
| } |
| table = self._make_one(resource) |
| self.assertEqual(table.to_api_repr(), resource) |
|
|
| def test__eq__same_table_property_different(self): |
| table_ref_resource = { |
| "projectId": "project_foo", |
| "datasetId": "dataset_bar", |
| "tableId": "table_baz", |
| } |
|
|
| resource_1 = {"tableReference": table_ref_resource, "friendlyName": "Table One"} |
| table_1 = self._make_one(resource_1) |
|
|
| resource_2 = {"tableReference": table_ref_resource, "friendlyName": "Table Two"} |
| table_2 = self._make_one(resource_2) |
|
|
| assert table_1 == table_2 |
|
|
|
|
| class TestTableClassesInterchangeability: |
| @staticmethod |
| def _make_table(*args, **kwargs): |
| from google.cloud.bigquery.table import Table |
|
|
| return Table(*args, **kwargs) |
|
|
| @staticmethod |
| def _make_table_ref(*args, **kwargs): |
| from google.cloud.bigquery.table import TableReference |
|
|
| return TableReference(*args, **kwargs) |
|
|
| @staticmethod |
| def _make_table_list_item(*args, **kwargs): |
| from google.cloud.bigquery.table import TableListItem |
|
|
| return TableListItem(*args, **kwargs) |
|
|
| def test_table_eq_table_ref(self): |
| table = self._make_table("project_foo.dataset_bar.table_baz") |
| dataset_ref = DatasetReference("project_foo", "dataset_bar") |
| table_ref = self._make_table_ref(dataset_ref, "table_baz") |
|
|
| assert table == table_ref |
| assert table_ref == table |
|
|
| def test_table_eq_table_list_item(self): |
| table = self._make_table("project_foo.dataset_bar.table_baz") |
| table_list_item = self._make_table_list_item( |
| { |
| "tableReference": { |
| "projectId": "project_foo", |
| "datasetId": "dataset_bar", |
| "tableId": "table_baz", |
| } |
| } |
| ) |
|
|
| assert table == table_list_item |
| assert table_list_item == table |
|
|
| def test_table_ref_eq_table_list_item(self): |
| dataset_ref = DatasetReference("project_foo", "dataset_bar") |
| table_ref = self._make_table_ref(dataset_ref, "table_baz") |
| table_list_item = self._make_table_list_item( |
| { |
| "tableReference": { |
| "projectId": "project_foo", |
| "datasetId": "dataset_bar", |
| "tableId": "table_baz", |
| } |
| } |
| ) |
|
|
| assert table_ref == table_list_item |
| assert table_list_item == table_ref |
|
|
|
|
| class TestSnapshotDefinition: |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import SnapshotDefinition |
|
|
| return SnapshotDefinition |
|
|
| @classmethod |
| def _make_one(cls, *args, **kwargs): |
| klass = cls._get_target_class() |
| return klass(*args, **kwargs) |
|
|
| def test_ctor_empty_resource(self): |
| instance = self._make_one(resource={}) |
| assert instance.base_table_reference is None |
| assert instance.snapshot_time is None |
|
|
| def test_ctor_full_resource(self): |
| from google.cloud._helpers import UTC |
| from google.cloud.bigquery.table import TableReference |
|
|
| resource = { |
| "baseTableReference": { |
| "projectId": "my-project", |
| "datasetId": "your-dataset", |
| "tableId": "our-table", |
| }, |
| "snapshotTime": "2005-06-07T19:35:02.123Z", |
| } |
| instance = self._make_one(resource) |
|
|
| expected_table_ref = TableReference.from_string( |
| "my-project.your-dataset.our-table" |
| ) |
| assert instance.base_table_reference == expected_table_ref |
|
|
| expected_time = datetime.datetime(2005, 6, 7, 19, 35, 2, 123000, tzinfo=UTC) |
| assert instance.snapshot_time == expected_time |
|
|
|
|
| class TestCloneDefinition: |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import CloneDefinition |
|
|
| return CloneDefinition |
|
|
| @classmethod |
| def _make_one(cls, *args, **kwargs): |
| klass = cls._get_target_class() |
| return klass(*args, **kwargs) |
|
|
| def test_ctor_empty_resource(self): |
| instance = self._make_one(resource={}) |
| assert instance.base_table_reference is None |
| assert instance.clone_time is None |
|
|
| def test_ctor_full_resource(self): |
| from google.cloud._helpers import UTC |
| from google.cloud.bigquery.table import TableReference |
|
|
| resource = { |
| "baseTableReference": { |
| "projectId": "my-project", |
| "datasetId": "your-dataset", |
| "tableId": "our-table", |
| }, |
| "cloneTime": "2005-06-07T19:35:02.123Z", |
| } |
| instance = self._make_one(resource) |
|
|
| expected_table_ref = TableReference.from_string( |
| "my-project.your-dataset.our-table" |
| ) |
| assert instance.base_table_reference == expected_table_ref |
|
|
| expected_time = datetime.datetime(2005, 6, 7, 19, 35, 2, 123000, tzinfo=UTC) |
| assert instance.clone_time == expected_time |
|
|
|
|
| class TestRow(unittest.TestCase): |
| def test_row(self): |
| from google.cloud.bigquery.table import Row |
|
|
| VALUES = (1, 2, 3) |
| row = Row(VALUES, {"a": 0, "b": 1, "c": 2}) |
| self.assertEqual(row.a, 1) |
| self.assertEqual(row[1], 2) |
| self.assertEqual(row["c"], 3) |
| self.assertEqual(len(row), 3) |
| self.assertEqual(row.values(), VALUES) |
| self.assertEqual(set(row.keys()), set({"a": 1, "b": 2, "c": 3}.keys())) |
| self.assertEqual(set(row.items()), set({"a": 1, "b": 2, "c": 3}.items())) |
| self.assertEqual(row.get("a"), 1) |
| self.assertEqual(row.get("d"), None) |
| self.assertEqual(row.get("d", ""), "") |
| self.assertEqual(row.get("d", default=""), "") |
| self.assertEqual(repr(row), "Row((1, 2, 3), {'a': 0, 'b': 1, 'c': 2})") |
| self.assertFalse(row != row) |
| self.assertFalse(row == 3) |
| with self.assertRaises(AttributeError): |
| row.z |
| with self.assertRaises(KeyError): |
| row["z"] |
|
|
|
|
| class Test_EmptyRowIterator(unittest.TestCase): |
| PYARROW_MINIMUM_VERSION = str(_versions_helpers._MIN_PYARROW_VERSION) |
|
|
| def _make_one(self): |
| from google.cloud.bigquery.table import _EmptyRowIterator |
|
|
| return _EmptyRowIterator() |
|
|
| def test_total_rows_eq_zero(self): |
| row_iterator = self._make_one() |
| self.assertEqual(row_iterator.total_rows, 0) |
|
|
| @mock.patch("google.cloud.bigquery.table.pyarrow", new=None) |
| def test_to_arrow_error_if_pyarrow_is_none(self): |
| row_iterator = self._make_one() |
| with self.assertRaises(ValueError): |
| row_iterator.to_arrow() |
|
|
| def test_to_arrow(self): |
| pyarrow = pytest.importorskip("pyarrow") |
| row_iterator = self._make_one() |
| tbl = row_iterator.to_arrow() |
| self.assertIsInstance(tbl, pyarrow.Table) |
| self.assertEqual(tbl.num_rows, 0) |
|
|
| def test_to_arrow_iterable(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| row_iterator = self._make_one() |
| arrow_iter = row_iterator.to_arrow_iterable() |
|
|
| result = list(arrow_iter) |
|
|
| self.assertEqual(len(result), 1) |
| record_batch = result[0] |
| self.assertIsInstance(record_batch, pyarrow.RecordBatch) |
| self.assertEqual(record_batch.num_rows, 0) |
| self.assertEqual(record_batch.num_columns, 0) |
|
|
| @mock.patch("google.cloud.bigquery._pandas_helpers.pandas", new=None) |
| def test_to_dataframe_error_if_pandas_is_none(self): |
| row_iterator = self._make_one() |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe() |
|
|
| def test_to_dataframe(self): |
| pandas = pytest.importorskip("pandas") |
| row_iterator = self._make_one() |
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 0) |
|
|
| @mock.patch("google.cloud.bigquery._pandas_helpers.pandas", new=None) |
| def test_to_dataframe_iterable_error_if_pandas_is_none(self): |
| row_iterator = self._make_one() |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe_iterable() |
|
|
| def test_to_dataframe_iterable(self): |
| pandas = pytest.importorskip("pandas") |
| row_iterator = self._make_one() |
| df_iter = row_iterator.to_dataframe_iterable() |
|
|
| result = list(df_iter) |
|
|
| self.assertEqual(len(result), 1) |
| df = result[0] |
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 0) |
| self.assertEqual(len(df.columns), 0) |
|
|
| @mock.patch("google.cloud.bigquery.table.geopandas", new=None) |
| def test_to_geodataframe_if_geopandas_is_none(self): |
| row_iterator = self._make_one() |
| with self.assertRaisesRegex( |
| ValueError, |
| re.escape( |
| "The geopandas library is not installed, please install " |
| "geopandas to use the to_geodataframe() function." |
| ), |
| ): |
| row_iterator.to_geodataframe(create_bqstorage_client=False) |
|
|
| def test_to_geodataframe(self): |
| geopandas = pytest.importorskip("geopandas") |
| row_iterator = self._make_one() |
| df = row_iterator.to_geodataframe(create_bqstorage_client=False) |
| self.assertIsInstance(df, geopandas.GeoDataFrame) |
| self.assertEqual(len(df), 0) |
| if version_info.major == 3 and version_info.minor > 7: |
| assert not hasattr(df, "crs") |
| else: |
| self.assertIsNone(df.crs) |
|
|
|
|
| class TestRowIterator(unittest.TestCase): |
| PYARROW_MINIMUM_VERSION = str(_versions_helpers._MIN_PYARROW_VERSION) |
|
|
| def _class_under_test(self): |
| from google.cloud.bigquery.table import RowIterator |
|
|
| return RowIterator |
|
|
| def _make_one( |
| self, |
| client=None, |
| api_request=None, |
| path=None, |
| schema=None, |
| table=None, |
| **kwargs |
| ): |
| from google.cloud.bigquery.table import TableReference |
|
|
| if client is None: |
| client = _mock_client() |
|
|
| if api_request is None: |
| api_request = mock.sentinel.api_request |
|
|
| if path is None: |
| path = "/foo" |
|
|
| if schema is None: |
| schema = [] |
|
|
| if table is None: |
| table = TableReference.from_string("my-project.my_dataset.my_table") |
|
|
| return self._class_under_test()( |
| client, api_request, path, schema, table=table, **kwargs |
| ) |
|
|
| def _make_one_from_data(self, schema=(), rows=()): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [SchemaField(*a) for a in schema] |
| rows = [{"f": [{"v": v} for v in row]} for row in rows] |
|
|
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| return self._make_one(_mock_client(), api_request, path, schema) |
|
|
| def test_constructor(self): |
| from google.cloud.bigquery.table import _item_to_row |
| from google.cloud.bigquery.table import _rows_page_start |
|
|
| client = _mock_client() |
| path = "/some/path" |
| iterator = self._make_one(client=client, path=path) |
|
|
| |
| self.assertIs(iterator.client, client) |
| self.assertIs(iterator.item_to_value, _item_to_row) |
| self.assertIs(iterator._page_start, _rows_page_start) |
| |
| self.assertEqual(iterator.extra_params, {}) |
| self.assertEqual(iterator._items_key, "rows") |
| self.assertIsNone(iterator.max_results) |
| self.assertEqual(iterator.path, path) |
| self.assertFalse(iterator._started) |
| self.assertIsNone(iterator.total_rows) |
| |
| self.assertEqual(iterator.page_number, 0) |
| self.assertIsNone(iterator.next_page_token) |
| self.assertEqual(iterator.num_results, 0) |
|
|
| def test_constructor_with_table(self): |
| from google.cloud.bigquery.table import Table |
|
|
| table = Table("proj.dset.tbl") |
| iterator = self._make_one(table=table, total_rows=100) |
| self.assertIs(iterator._table, table) |
| self.assertEqual(iterator.total_rows, 100) |
|
|
| def test_constructor_with_dict_schema(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| {"name": "full_name", "type": "STRING", "mode": "REQUIRED"}, |
| {"name": "age", "type": "INT64", "mode": "NULLABLE"}, |
| ] |
|
|
| iterator = self._make_one(schema=schema) |
|
|
| expected_schema = [ |
| SchemaField("full_name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INT64", mode="NULLABLE"), |
| ] |
| self.assertEqual(iterator.schema, expected_schema) |
|
|
| def test_job_id_missing(self): |
| rows = self._make_one() |
| self.assertIsNone(rows.job_id) |
|
|
| def test_job_id_present(self): |
| rows = self._make_one(job_id="abc-123") |
| self.assertEqual(rows.job_id, "abc-123") |
|
|
| def test_location_missing(self): |
| rows = self._make_one() |
| self.assertIsNone(rows.location) |
|
|
| def test_location_present(self): |
| rows = self._make_one(location="asia-northeast1") |
| self.assertEqual(rows.location, "asia-northeast1") |
|
|
| def test_num_dml_affected_rows_missing(self): |
| rows = self._make_one() |
| self.assertIsNone(rows.num_dml_affected_rows) |
|
|
| def test_num_dml_affected_rows_present(self): |
| rows = self._make_one(num_dml_affected_rows=1234) |
| self.assertEqual(rows.num_dml_affected_rows, 1234) |
|
|
| def test_project_missing(self): |
| rows = self._make_one() |
| self.assertIsNone(rows.project) |
|
|
| def test_project_present(self): |
| rows = self._make_one(project="test-project") |
| self.assertEqual(rows.project, "test-project") |
|
|
| def test_query_id_missing(self): |
| rows = self._make_one() |
| self.assertIsNone(rows.query_id) |
|
|
| def test_query_id_present(self): |
| rows = self._make_one(query_id="xyz-987") |
| self.assertEqual(rows.query_id, "xyz-987") |
|
|
| def test_iterate(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
| self.assertEqual(row_iterator.num_results, 0) |
|
|
| rows_iter = iter(row_iterator) |
|
|
| val1 = next(rows_iter) |
| self.assertEqual(val1.name, "Phred Phlyntstone") |
| self.assertEqual(row_iterator.num_results, 1) |
|
|
| val2 = next(rows_iter) |
| self.assertEqual(val2.name, "Bharney Rhubble") |
| self.assertEqual(row_iterator.num_results, 2) |
|
|
| with self.assertRaises(StopIteration): |
| next(rows_iter) |
|
|
| api_request.assert_called_once_with(method="GET", path=path, query_params={}) |
|
|
| def test_iterate_with_cached_first_page(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| first_page = { |
| "rows": [ |
| {"f": [{"v": "Whillma Phlyntstone"}, {"v": "27"}]}, |
| {"f": [{"v": "Bhetty Rhubble"}, {"v": "28"}]}, |
| ], |
| "pageToken": "next-page", |
| } |
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one( |
| _mock_client(), |
| api_request, |
| path, |
| schema, |
| first_page_response=first_page, |
| total_rows=4, |
| ) |
| self.assertEqual(row_iterator.total_rows, 4) |
| rows = list(row_iterator) |
| |
| |
| self.assertEqual(row_iterator.total_rows, 4) |
| self.assertEqual(len(rows), 4) |
| self.assertEqual(rows[0].age, 27) |
| self.assertEqual(rows[1].age, 28) |
| self.assertEqual(rows[2].age, 32) |
| self.assertEqual(rows[3].age, 33) |
|
|
| api_request.assert_called_once_with( |
| method="GET", path=path, query_params={"pageToken": "next-page"} |
| ) |
|
|
| def test_iterate_with_cached_first_page_max_results(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| first_page = { |
| "rows": [ |
| {"f": [{"v": "Whillma Phlyntstone"}, {"v": "27"}]}, |
| {"f": [{"v": "Bhetty Rhubble"}, {"v": "28"}]}, |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ], |
| "pageToken": "next-page", |
| } |
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value=first_page) |
| row_iterator = self._make_one( |
| _mock_client(), |
| api_request, |
| path, |
| schema, |
| max_results=3, |
| first_page_response=first_page, |
| ) |
| rows = list(row_iterator) |
| self.assertEqual(len(rows), 3) |
| self.assertEqual(rows[0].age, 27) |
| self.assertEqual(rows[1].age, 28) |
| self.assertEqual(rows[2].age, 32) |
| api_request.assert_not_called() |
|
|
| def test_page_size(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
|
|
| row_iterator = self._make_one( |
| _mock_client(), api_request, path, schema, page_size=4 |
| ) |
| row_iterator._get_next_page_response() |
|
|
| api_request.assert_called_once_with( |
| method="GET", |
| path=path, |
| query_params={"maxResults": row_iterator._page_size}, |
| ) |
|
|
| def test__is_almost_completely_cached_returns_false_without_first_page(self): |
| iterator = self._make_one(first_page_response=None) |
| self.assertFalse(iterator._is_almost_completely_cached()) |
|
|
| def test__is_almost_completely_cached_returns_true_with_more_rows_than_max_results( |
| self, |
| ): |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Whillma Phlyntstone"}, {"v": "27"}]}, |
| {"f": [{"v": "Bhetty Rhubble"}, {"v": "28"}]}, |
| ] |
| first_page = {"pageToken": "next-page", "rows": rows} |
| iterator = self._make_one(max_results=4, first_page_response=first_page) |
| self.assertTrue(iterator._is_almost_completely_cached()) |
|
|
| def test__is_almost_completely_cached_returns_false_with_too_many_rows_remaining( |
| self, |
| ): |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| first_page = {"pageToken": "next-page", "rows": rows} |
| iterator = self._make_one(first_page_response=first_page, total_rows=100) |
| self.assertFalse(iterator._is_almost_completely_cached()) |
|
|
| def test__is_almost_completely_cached_returns_false_with_rows_remaining_and_no_total_rows( |
| self, |
| ): |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| first_page = {"pageToken": "next-page", "rows": rows} |
| iterator = self._make_one(first_page_response=first_page) |
| self.assertFalse(iterator._is_almost_completely_cached()) |
|
|
| def test__is_almost_completely_cached_returns_true_with_some_rows_remaining(self): |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Whillma Phlyntstone"}, {"v": "27"}]}, |
| {"f": [{"v": "Bhetty Rhubble"}, {"v": "28"}]}, |
| {"f": [{"v": "Pebbles Phlyntstone"}, {"v": "4"}]}, |
| {"f": [{"v": "Bamm-Bamm Rhubble"}, {"v": "5"}]}, |
| {"f": [{"v": "Joseph Rockhead"}, {"v": "32"}]}, |
| {"f": [{"v": "Perry Masonry"}, {"v": "33"}]}, |
| ] |
| first_page = {"pageToken": "next-page", "rows": rows} |
| iterator = self._make_one( |
| first_page_response=first_page, total_rows=len(rows) + 1 |
| ) |
| self.assertTrue(iterator._is_almost_completely_cached()) |
|
|
| def test__is_almost_completely_cached_returns_true_with_no_rows_remaining(self): |
| first_page = {"rows": []} |
| iterator = self._make_one(first_page_response=first_page) |
| self.assertTrue(iterator._is_almost_completely_cached()) |
|
|
| def test__should_use_bqstorage_returns_false_when_completely_cached(self): |
| first_page = {"rows": []} |
| iterator = self._make_one(first_page_response=first_page) |
| self.assertFalse( |
| iterator._should_use_bqstorage( |
| bqstorage_client=None, create_bqstorage_client=True |
| ) |
| ) |
|
|
| def test__should_use_bqstorage_returns_true_if_no_cached_results(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| iterator = self._make_one(first_page_response=None) |
| result = iterator._should_use_bqstorage( |
| bqstorage_client=None, create_bqstorage_client=True |
| ) |
| self.assertTrue(result) |
|
|
| def test__should_use_bqstorage_returns_false_if_page_token_set(self): |
| iterator = self._make_one( |
| page_token="abc", first_page_response=None |
| ) |
| result = iterator._should_use_bqstorage( |
| bqstorage_client=None, create_bqstorage_client=True |
| ) |
| self.assertFalse(result) |
|
|
| def test__should_use_bqstorage_returns_false_if_max_results_set(self): |
| iterator = self._make_one( |
| max_results=10, first_page_response=None |
| ) |
| result = iterator._should_use_bqstorage( |
| bqstorage_client=None, create_bqstorage_client=True |
| ) |
| self.assertFalse(result) |
|
|
| def test__should_use_bqstorage_returns_false_w_warning_if_missing_dependency(self): |
| iterator = self._make_one(first_page_response=None) |
|
|
| def fail_bqstorage_import(name, globals, locals, fromlist, level): |
| """Returns True if bigquery_storage has been imported.""" |
| |
| return "bigquery_storage" in name or ( |
| fromlist is not None and "bigquery_storage" in fromlist |
| ) |
|
|
| |
| no_bqstorage = maybe_fail_import(predicate=fail_bqstorage_import) |
|
|
| with no_bqstorage, warnings.catch_warnings(record=True) as warned: |
| result = iterator._should_use_bqstorage( |
| bqstorage_client=None, create_bqstorage_client=True |
| ) |
|
|
| self.assertFalse(result) |
|
|
| matching_warnings = [ |
| warning for warning in warned if "Storage module not found" in str(warning) |
| ] |
| assert matching_warnings, "Dependency not found warning not raised." |
|
|
| def test__should_use_bqstorage_returns_false_w_warning_if_obsolete_version(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| iterator = self._make_one(first_page_response=None) |
|
|
| patcher = mock.patch( |
| "google.cloud.bigquery.table._versions_helpers.BQ_STORAGE_VERSIONS.try_import", |
| side_effect=exceptions.LegacyBigQueryStorageError("BQ Storage too old"), |
| ) |
| with patcher, warnings.catch_warnings(record=True) as warned: |
| result = iterator._should_use_bqstorage( |
| bqstorage_client=None, create_bqstorage_client=True |
| ) |
|
|
| self.assertFalse(result) |
|
|
| matching_warnings = [ |
| warning for warning in warned if "BQ Storage too old" in str(warning) |
| ] |
| assert matching_warnings, "Obsolete dependency warning not raised." |
|
|
| def test_to_arrow_iterable(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| SchemaField( |
| "child", |
| "RECORD", |
| mode="REPEATED", |
| fields=[ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ], |
| ), |
| ] |
| rows = [ |
| { |
| "f": [ |
| {"v": "Bharney Rhubble"}, |
| {"v": "33"}, |
| { |
| "v": [ |
| {"v": {"f": [{"v": "Whamm-Whamm Rhubble"}, {"v": "3"}]}}, |
| {"v": {"f": [{"v": "Hoppy"}, {"v": "1"}]}}, |
| ] |
| }, |
| ] |
| }, |
| { |
| "f": [ |
| {"v": "Wylma Phlyntstone"}, |
| {"v": "29"}, |
| { |
| "v": [ |
| {"v": {"f": [{"v": "Bepples Phlyntstone"}, {"v": "0"}]}}, |
| {"v": {"f": [{"v": "Dino"}, {"v": "4"}]}}, |
| ] |
| }, |
| ] |
| }, |
| ] |
| path = "/foo" |
| api_request = mock.Mock( |
| side_effect=[ |
| {"rows": [rows[0]], "pageToken": "NEXTPAGE"}, |
| {"rows": [rows[1]]}, |
| ] |
| ) |
| row_iterator = self._make_one( |
| _mock_client(), api_request, path, schema, page_size=1, max_results=5 |
| ) |
|
|
| record_batches = row_iterator.to_arrow_iterable() |
| self.assertIsInstance(record_batches, types.GeneratorType) |
| record_batches = list(record_batches) |
| self.assertEqual(len(record_batches), 2) |
|
|
| |
| for record_batch in record_batches: |
| self.assertIsInstance(record_batch, pyarrow.RecordBatch) |
| self.assertEqual(record_batch.schema[0].name, "name") |
| self.assertTrue(pyarrow.types.is_string(record_batch.schema[0].type)) |
| self.assertEqual(record_batch.schema[1].name, "age") |
| self.assertTrue(pyarrow.types.is_int64(record_batch.schema[1].type)) |
| child_field = record_batch.schema[2] |
| self.assertEqual(child_field.name, "child") |
| self.assertTrue(pyarrow.types.is_list(child_field.type)) |
| self.assertTrue(pyarrow.types.is_struct(child_field.type.value_type)) |
| self.assertEqual(child_field.type.value_type[0].name, "name") |
| self.assertEqual(child_field.type.value_type[1].name, "age") |
|
|
| |
| record_batch_1 = record_batches[0].to_pydict() |
| names = record_batch_1["name"] |
| ages = record_batch_1["age"] |
| children = record_batch_1["child"] |
| self.assertEqual(names, ["Bharney Rhubble"]) |
| self.assertEqual(ages, [33]) |
| self.assertEqual( |
| children, |
| [ |
| [ |
| {"name": "Whamm-Whamm Rhubble", "age": 3}, |
| {"name": "Hoppy", "age": 1}, |
| ], |
| ], |
| ) |
|
|
| record_batch_2 = record_batches[1].to_pydict() |
| names = record_batch_2["name"] |
| ages = record_batch_2["age"] |
| children = record_batch_2["child"] |
| self.assertEqual(names, ["Wylma Phlyntstone"]) |
| self.assertEqual(ages, [29]) |
| self.assertEqual( |
| children, |
| [[{"name": "Bepples Phlyntstone", "age": 0}, {"name": "Dino", "age": 4}]], |
| ) |
|
|
| def test_to_arrow_iterable_w_bqstorage(self): |
| pyarrow = pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery_storage_v1 import reader |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| streams = [ |
| |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| ] |
| session = bigquery_storage.types.ReadSession(streams=streams) |
| arrow_schema = pyarrow.schema( |
| [ |
| pyarrow.field("colA", pyarrow.int64()), |
| |
| pyarrow.field("colC", pyarrow.float64()), |
| pyarrow.field("colB", pyarrow.string()), |
| ] |
| ) |
| session.arrow_schema.serialized_schema = arrow_schema.serialize().to_pybytes() |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| page_items = [ |
| pyarrow.array([1, -1]), |
| pyarrow.array([2.0, 4.0]), |
| pyarrow.array(["abc", "def"]), |
| ] |
|
|
| expected_record_batch = pyarrow.RecordBatch.from_arrays( |
| page_items, schema=arrow_schema |
| ) |
| expected_num_record_batches = 3 |
|
|
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_arrow.return_value = expected_record_batch |
| mock_pages = (mock_page,) * expected_num_record_batches |
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| schema = [ |
| schema.SchemaField("colA", "INTEGER"), |
| schema.SchemaField("colC", "FLOAT"), |
| schema.SchemaField("colB", "STRING"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| record_batches = list( |
| row_iterator.to_arrow_iterable(bqstorage_client=bqstorage_client) |
| ) |
| total_record_batches = len(streams) * len(mock_pages) |
| self.assertEqual(len(record_batches), total_record_batches) |
|
|
| for record_batch in record_batches: |
| |
| self.assertEqual(record_batch, expected_record_batch) |
|
|
| |
| bqstorage_client._transport.grpc_channel.close.assert_not_called() |
|
|
| def test_to_arrow(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| SchemaField( |
| "child", |
| "RECORD", |
| mode="REPEATED", |
| fields=[ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ], |
| ), |
| ] |
| rows = [ |
| { |
| "f": [ |
| {"v": "Bharney Rhubble"}, |
| {"v": "33"}, |
| { |
| "v": [ |
| {"v": {"f": [{"v": "Whamm-Whamm Rhubble"}, {"v": "3"}]}}, |
| {"v": {"f": [{"v": "Hoppy"}, {"v": "1"}]}}, |
| ] |
| }, |
| ] |
| }, |
| { |
| "f": [ |
| {"v": "Wylma Phlyntstone"}, |
| {"v": "29"}, |
| { |
| "v": [ |
| {"v": {"f": [{"v": "Bepples Phlyntstone"}, {"v": "0"}]}}, |
| {"v": {"f": [{"v": "Dino"}, {"v": "4"}]}}, |
| ] |
| }, |
| ] |
| }, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| tbl = row_iterator.to_arrow(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(tbl, pyarrow.Table) |
| self.assertEqual(tbl.num_rows, 2) |
|
|
| |
| self.assertEqual(tbl.schema[0].name, "name") |
| self.assertTrue(pyarrow.types.is_string(tbl.schema[0].type)) |
| self.assertEqual(tbl.schema[1].name, "age") |
| self.assertTrue(pyarrow.types.is_int64(tbl.schema[1].type)) |
| child_field = tbl.schema[2] |
| self.assertEqual(child_field.name, "child") |
| self.assertTrue(pyarrow.types.is_list(child_field.type)) |
| self.assertTrue(pyarrow.types.is_struct(child_field.type.value_type)) |
| self.assertEqual(child_field.type.value_type[0].name, "name") |
| self.assertEqual(child_field.type.value_type[1].name, "age") |
|
|
| |
| tbl_data = tbl.to_pydict() |
| names = tbl_data["name"] |
| ages = tbl_data["age"] |
| children = tbl_data["child"] |
| self.assertEqual(names, ["Bharney Rhubble", "Wylma Phlyntstone"]) |
| self.assertEqual(ages, [33, 29]) |
| self.assertEqual( |
| children, |
| [ |
| [ |
| {"name": "Whamm-Whamm Rhubble", "age": 3}, |
| {"name": "Hoppy", "age": 1}, |
| ], |
| [{"name": "Bepples Phlyntstone", "age": 0}, {"name": "Dino", "age": 4}], |
| ], |
| ) |
|
|
| def test_to_arrow_w_nulls(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| import pyarrow.types |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [SchemaField("name", "STRING"), SchemaField("age", "INTEGER")] |
| rows = [ |
| {"f": [{"v": "Donkey"}, {"v": 32}]}, |
| {"f": [{"v": "Diddy"}, {"v": 29}]}, |
| {"f": [{"v": "Dixie"}, {"v": None}]}, |
| {"f": [{"v": None}, {"v": 111}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| tbl = row_iterator.to_arrow(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(tbl, pyarrow.Table) |
| self.assertEqual(tbl.num_rows, 4) |
|
|
| |
| self.assertEqual(tbl.schema[0].name, "name") |
| self.assertTrue(pyarrow.types.is_string(tbl.schema[0].type)) |
| self.assertEqual(tbl.schema[1].name, "age") |
| self.assertTrue(pyarrow.types.is_int64(tbl.schema[1].type)) |
|
|
| |
| tbl_data = tbl.to_pydict() |
| names = tbl_data["name"] |
| ages = tbl_data["age"] |
| self.assertEqual(names, ["Donkey", "Diddy", "Dixie", None]) |
| self.assertEqual(ages, [32, 29, None, 111]) |
|
|
| def test_to_arrow_w_unknown_type(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| SchemaField("sport", "UNKNOWN_TYPE", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}, {"v": "volleyball"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}, {"v": "basketball"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| tbl = row_iterator.to_arrow(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(tbl, pyarrow.Table) |
| self.assertEqual(tbl.num_rows, 2) |
|
|
| |
| self.assertEqual(tbl.schema[0].name, "name") |
| self.assertTrue(pyarrow.types.is_string(tbl.schema[0].type)) |
| self.assertEqual(tbl.schema[1].name, "age") |
| self.assertTrue(pyarrow.types.is_int64(tbl.schema[1].type)) |
| self.assertEqual(tbl.schema[2].name, "sport") |
|
|
| |
| tbl_data = tbl.to_pydict() |
| names = tbl_data["name"] |
| ages = tbl_data["age"] |
| sports = tbl_data["sport"] |
| self.assertEqual(names, ["Bharney Rhubble", "Wylma Phlyntstone"]) |
| self.assertEqual(ages, [33, 29]) |
| self.assertEqual(sports, ["volleyball", "basketball"]) |
|
|
| |
| self.assertEqual(len(warned), 2) |
| self.assertTrue(all("sport" in str(warning) for warning in warned)) |
|
|
| def test_to_arrow_w_empty_table(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| SchemaField( |
| "child", |
| "RECORD", |
| mode="REPEATED", |
| fields=[ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ], |
| ), |
| ] |
| rows = [] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| tbl = row_iterator.to_arrow(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(tbl, pyarrow.Table) |
| self.assertEqual(tbl.num_rows, 0) |
|
|
| |
| self.assertEqual(tbl.schema[0].name, "name") |
| self.assertTrue(pyarrow.types.is_string(tbl.schema[0].type)) |
| self.assertEqual(tbl.schema[1].name, "age") |
| self.assertTrue(pyarrow.types.is_int64(tbl.schema[1].type)) |
| child_field = tbl.schema[2] |
| self.assertEqual(child_field.name, "child") |
| self.assertTrue(pyarrow.types.is_list(child_field.type)) |
| self.assertTrue(pyarrow.types.is_struct(child_field.type.value_type)) |
| self.assertEqual(child_field.type.value_type[0].name, "name") |
| self.assertEqual(child_field.type.value_type[1].name, "age") |
|
|
| def test_to_arrow_max_results_w_explicit_bqstorage_client_warning(self): |
| pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| mock_client = _mock_client() |
| mock_bqstorage_client = mock.sentinel.bq_storage_client |
|
|
| row_iterator = self._make_one( |
| client=mock_client, |
| api_request=api_request, |
| path=path, |
| schema=schema, |
| max_results=42, |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| row_iterator.to_arrow(bqstorage_client=mock_bqstorage_client) |
|
|
| matches = [ |
| warning |
| for warning in warned |
| if warning.category is UserWarning |
| and "cannot use bqstorage_client" in str(warning).lower() |
| and "REST" in str(warning) |
| ] |
| self.assertEqual(len(matches), 1, msg="User warning was not emitted.") |
| self.assertIn( |
| __file__, str(matches[0]), msg="Warning emitted with incorrect stacklevel" |
| ) |
| mock_client._ensure_bqstorage_client.assert_not_called() |
|
|
| def test_to_arrow_max_results_w_create_bqstorage_client_no_warning(self): |
| pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| mock_client = _mock_client() |
|
|
| row_iterator = self._make_one( |
| client=mock_client, |
| api_request=api_request, |
| path=path, |
| schema=schema, |
| max_results=42, |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| row_iterator.to_arrow(create_bqstorage_client=True) |
|
|
| matches = [ |
| warning |
| for warning in warned |
| if warning.category is UserWarning |
| and "cannot use bqstorage_client" in str(warning).lower() |
| and "REST" in str(warning) |
| ] |
| self.assertFalse(matches) |
| mock_client._ensure_bqstorage_client.assert_not_called() |
|
|
| def test_to_arrow_w_bqstorage(self): |
| pyarrow = pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery_storage_v1 import reader |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| streams = [ |
| |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| ] |
| session = bigquery_storage.types.ReadSession(streams=streams) |
| arrow_schema = pyarrow.schema( |
| [ |
| pyarrow.field("colA", pyarrow.int64()), |
| |
| pyarrow.field("colC", pyarrow.float64()), |
| pyarrow.field("colB", pyarrow.string()), |
| ] |
| ) |
| session.arrow_schema.serialized_schema = arrow_schema.serialize().to_pybytes() |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| expected_num_rows = 2 |
| expected_num_columns = 3 |
| page_items = [ |
| pyarrow.array([1, -1]), |
| pyarrow.array([2.0, 4.0]), |
| pyarrow.array(["abc", "def"]), |
| ] |
|
|
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_arrow.return_value = pyarrow.RecordBatch.from_arrays( |
| page_items, schema=arrow_schema |
| ) |
| mock_pages = (mock_page, mock_page, mock_page) |
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| schema = [ |
| schema.SchemaField("colA", "INTEGER"), |
| schema.SchemaField("colC", "FLOAT"), |
| schema.SchemaField("colB", "STRING"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| actual_tbl = row_iterator.to_arrow(bqstorage_client=bqstorage_client) |
|
|
| |
| self.assertEqual(actual_tbl.num_columns, expected_num_columns) |
| self.assertEqual(actual_tbl.schema[0].name, "colA") |
| self.assertEqual(actual_tbl.schema[1].name, "colC") |
| self.assertEqual(actual_tbl.schema[2].name, "colB") |
|
|
| |
| total_pages = len(streams) * len(mock_pages) |
| total_rows = expected_num_rows * total_pages |
| self.assertEqual(actual_tbl.num_rows, total_rows) |
|
|
| |
| bqstorage_client._transport.grpc_channel.close.assert_not_called() |
|
|
| def test_to_arrow_w_bqstorage_creates_client(self): |
| pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| mock_client = _mock_client() |
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| mock_client._ensure_bqstorage_client.return_value = bqstorage_client |
| session = bigquery_storage.types.ReadSession() |
| bqstorage_client.create_read_session.return_value = session |
| row_iterator = mut.RowIterator( |
| mock_client, |
| None, |
| None, |
| [ |
| schema.SchemaField("colA", "STRING"), |
| schema.SchemaField("colC", "STRING"), |
| schema.SchemaField("colB", "STRING"), |
| ], |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| ) |
| row_iterator.to_arrow(create_bqstorage_client=True) |
| mock_client._ensure_bqstorage_client.assert_called_once() |
| bqstorage_client._transport.grpc_channel.close.assert_called_once() |
|
|
| def test_to_arrow_ensure_bqstorage_client_wo_bqstorage(self): |
| pyarrow = pytest.importorskip( |
| "pyarrow", minversion=self.PYARROW_MINIMUM_VERSION |
| ) |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Alice"}, {"v": "98"}]}, |
| {"f": [{"v": "Bob"}, {"v": "99"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
|
|
| mock_client = _mock_client() |
| row_iterator = self._make_one(mock_client, api_request, path, schema) |
|
|
| def mock_verify_version(raise_if_error: bool = False): |
| raise exceptions.LegacyBigQueryStorageError("no bqstorage") |
|
|
| with mock.patch( |
| "google.cloud.bigquery._versions_helpers.BQ_STORAGE_VERSIONS.try_import", |
| mock_verify_version, |
| ): |
| tbl = row_iterator.to_arrow(create_bqstorage_client=True) |
|
|
| mock_client._ensure_bqstorage_client.assert_not_called() |
| self.assertIsInstance(tbl, pyarrow.Table) |
| self.assertEqual(tbl.num_rows, 2) |
|
|
| def test_to_arrow_w_bqstorage_no_streams(self): |
| pyarrow = pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| session = bigquery_storage.types.ReadSession() |
| arrow_schema = pyarrow.schema( |
| [ |
| pyarrow.field("colA", pyarrow.string()), |
| |
| pyarrow.field("colC", pyarrow.string()), |
| pyarrow.field("colB", pyarrow.string()), |
| ] |
| ) |
| session.arrow_schema.serialized_schema = arrow_schema.serialize().to_pybytes() |
| bqstorage_client.create_read_session.return_value = session |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| [ |
| schema.SchemaField("colA", "STRING"), |
| schema.SchemaField("colC", "STRING"), |
| schema.SchemaField("colB", "STRING"), |
| ], |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| ) |
|
|
| actual_table = row_iterator.to_arrow(bqstorage_client=bqstorage_client) |
| self.assertEqual(actual_table.num_columns, 3) |
| self.assertEqual(actual_table.num_rows, 0) |
| self.assertEqual(actual_table.schema[0].name, "colA") |
| self.assertEqual(actual_table.schema[1].name, "colC") |
| self.assertEqual(actual_table.schema[2].name, "colB") |
|
|
| def test_to_arrow_progress_bar(self): |
| pytest.importorskip("pyarrow") |
| pytest.importorskip("tqdm") |
| pytest.importorskip("tqdm.notebook") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
|
|
| progress_bars = ( |
| ("tqdm", mock.patch("tqdm.tqdm")), |
| ("tqdm_notebook", mock.patch("tqdm.notebook.tqdm")), |
| ("tqdm_gui", mock.patch("tqdm.tqdm_gui")), |
| ) |
|
|
| for progress_bar_type, bar_patch in progress_bars: |
| progress_bar_mock = bar_patch.start() |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
| tbl = row_iterator.to_arrow( |
| progress_bar_type=progress_bar_type, |
| create_bqstorage_client=False, |
| ) |
|
|
| progress_bar_mock.assert_called() |
| progress_bar_mock().update.assert_called() |
| progress_bar_mock().close.assert_called_once() |
| self.assertEqual(tbl.num_rows, 4) |
|
|
| @mock.patch("google.cloud.bigquery.table.pyarrow", new=None) |
| def test_to_arrow_w_pyarrow_none(self): |
| schema = [] |
| rows = [] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with self.assertRaises(ValueError): |
| row_iterator.to_arrow() |
|
|
| def test_to_dataframe_iterable(self): |
| pandas = pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
|
|
| path = "/foo" |
| api_request = mock.Mock( |
| side_effect=[ |
| { |
| "rows": [{"f": [{"v": "Bengt"}, {"v": "32"}]}], |
| "pageToken": "NEXTPAGE", |
| }, |
| {"rows": [{"f": [{"v": "Sven"}, {"v": "33"}]}]}, |
| ] |
| ) |
|
|
| row_iterator = self._make_one( |
| _mock_client(), api_request, path, schema, page_size=1, max_results=5 |
| ) |
| dfs = row_iterator.to_dataframe_iterable() |
|
|
| self.assertIsInstance(dfs, types.GeneratorType) |
|
|
| df_1 = next(dfs) |
| self.assertIsInstance(df_1, pandas.DataFrame) |
| self.assertEqual(df_1.name.dtype.name, "object") |
| self.assertEqual(df_1.age.dtype.name, "int64") |
| self.assertEqual(len(df_1), 1) |
| self.assertEqual( |
| df_1["name"][0], "Bengt" |
| ) |
| self.assertEqual(df_1["age"][0], 32) |
|
|
| df_2 = next(dfs) |
| self.assertEqual(len(df_2), 1) |
| self.assertEqual(df_2["name"][0], "Sven") |
| self.assertEqual(df_2["age"][0], 33) |
|
|
| def test_to_dataframe_iterable_with_dtypes(self): |
| pandas = pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
|
|
| path = "/foo" |
| api_request = mock.Mock( |
| side_effect=[ |
| { |
| "rows": [{"f": [{"v": "Bengt"}, {"v": "32"}]}], |
| "pageToken": "NEXTPAGE", |
| }, |
| {"rows": [{"f": [{"v": "Sven"}, {"v": "33"}]}]}, |
| ] |
| ) |
|
|
| row_iterator = self._make_one( |
| _mock_client(), api_request, path, schema, page_size=1, max_results=5 |
| ) |
| dfs = row_iterator.to_dataframe_iterable(dtypes={"age": "int32"}) |
|
|
| self.assertIsInstance(dfs, types.GeneratorType) |
|
|
| df_1 = next(dfs) |
| self.assertIsInstance(df_1, pandas.DataFrame) |
| self.assertEqual(df_1.name.dtype.name, "object") |
| self.assertEqual(df_1.age.dtype.name, "int32") |
| self.assertEqual(len(df_1), 1) |
| self.assertEqual( |
| df_1["name"][0], "Bengt" |
| ) |
| self.assertEqual(df_1["age"][0], 32) |
|
|
| df_2 = next(dfs) |
| self.assertEqual(len(df_2), 1) |
| self.assertEqual(df_2["name"][0], "Sven") |
| self.assertEqual(df_2["age"][0], 33) |
|
|
| def test_to_dataframe_iterable_w_bqstorage(self): |
| pandas = pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery_storage_v1 import reader |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| arrow_fields = [ |
| pyarrow.field("colA", pyarrow.int64()), |
| |
| pyarrow.field("colC", pyarrow.float64()), |
| pyarrow.field("colB", pyarrow.utf8()), |
| ] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| streams = [ |
| |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| ] |
| session = bigquery_storage.types.ReadSession( |
| streams=streams, |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| page_dataframe = pandas.DataFrame( |
| {"colA": [1, -1], "colC": [2.0, 4.0], "colB": ["abc", "def"]}, |
| ) |
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_dataframe.return_value = page_dataframe |
| mock_pages = (mock_page, mock_page, mock_page) |
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| schema = [ |
| schema.SchemaField("colA", "IGNORED"), |
| schema.SchemaField("colC", "IGNORED"), |
| schema.SchemaField("colB", "IGNORED"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| got = list( |
| row_iterator.to_dataframe_iterable(bqstorage_client=bqstorage_client) |
| ) |
|
|
| |
| total_pages = len(streams) * len(mock_pages) |
| self.assertEqual(len(got), total_pages) |
|
|
| |
| bqstorage_client._transport.grpc_channel.close.assert_not_called() |
|
|
| def test_to_dataframe_iterable_w_bqstorage_max_results_warning(self): |
| pandas = pytest.importorskip("pandas") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
|
|
| iterator_schema = [ |
| schema.SchemaField("name", "STRING", mode="REQUIRED"), |
| schema.SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| path = "/foo" |
| api_request = mock.Mock( |
| side_effect=[ |
| { |
| "rows": [{"f": [{"v": "Bengt"}, {"v": "32"}]}], |
| "pageToken": "NEXTPAGE", |
| }, |
| {"rows": [{"f": [{"v": "Sven"}, {"v": "33"}]}]}, |
| ] |
| ) |
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| api_request, |
| path, |
| iterator_schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=iterator_schema, |
| max_results=25, |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| dfs = row_iterator.to_dataframe_iterable(bqstorage_client=bqstorage_client) |
|
|
| |
| matches = [ |
| warning |
| for warning in warned |
| if warning.category is UserWarning |
| and "cannot use bqstorage_client" in str(warning).lower() |
| and "REST" in str(warning) |
| ] |
| assert len(matches) == 1, "User warning was not emitted." |
| assert __file__ in str(matches[0]), "Warning emitted with incorrect stacklevel" |
|
|
| |
| dataframes = list(dfs) |
| assert len(dataframes) == 2 |
| assert isinstance(dataframes[0], pandas.DataFrame) |
| assert isinstance(dataframes[1], pandas.DataFrame) |
|
|
| @mock.patch("google.cloud.bigquery._pandas_helpers.pandas", new=None) |
| def test_to_dataframe_iterable_error_if_pandas_is_none(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with pytest.raises(ValueError, match="pandas"): |
| row_iterator.to_dataframe_iterable() |
|
|
| def test_to_dataframe(self): |
| pandas = pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 4) |
| self.assertEqual(list(df), ["name", "age"]) |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertEqual(df.age.dtype.name, "Int64") |
|
|
| def test_to_dataframe_timestamp_out_of_pyarrow_bounds(self): |
| pandas = pytest.importorskip("pandas") |
| pytest.importorskip("pyarrow") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [SchemaField("some_timestamp", "TIMESTAMP")] |
| rows = [ |
| {"f": [{"v": "81953424000000000"}]}, |
| {"f": [{"v": "253402214400000000"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| tzinfo = datetime.timezone.utc |
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 2) |
| self.assertEqual(list(df.columns), ["some_timestamp"]) |
| self.assertEqual( |
| list(df["some_timestamp"]), |
| [ |
| datetime.datetime(4567, 1, 1, tzinfo=tzinfo), |
| datetime.datetime(9999, 12, 31, tzinfo=tzinfo), |
| ], |
| ) |
|
|
| def test_to_dataframe_datetime_out_of_pyarrow_bounds(self): |
| pandas = pytest.importorskip("pandas") |
| pytest.importorskip("pyarrow") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [SchemaField("some_datetime", "DATETIME")] |
| rows = [ |
| {"f": [{"v": "4567-01-01T00:00:00"}]}, |
| {"f": [{"v": "9999-12-31T00:00:00"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 2) |
| self.assertEqual(list(df.columns), ["some_datetime"]) |
| self.assertEqual( |
| list(df["some_datetime"]), |
| [datetime.datetime(4567, 1, 1), datetime.datetime(9999, 12, 31)], |
| ) |
|
|
| def test_to_dataframe_progress_bar(self): |
| pytest.importorskip("pandas") |
| pytest.importorskip("pyarrow") |
| pytest.importorskip("tqdm") |
|
|
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
|
|
| progress_bars = ( |
| ("tqdm", mock.patch("tqdm.tqdm")), |
| ("tqdm_gui", mock.patch("tqdm.tqdm_gui")), |
| ) |
|
|
| for progress_bar_type, bar_patch in progress_bars: |
| progress_bar_mock = bar_patch.start() |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
| df = row_iterator.to_dataframe( |
| progress_bar_type=progress_bar_type, |
| create_bqstorage_client=False, |
| ) |
|
|
| progress_bar_mock.assert_called() |
| progress_bar_mock().update.assert_called() |
| progress_bar_mock().close.assert_called_once() |
| self.assertEqual(len(df), 4) |
|
|
| def test_to_dataframe_progress_bar_notebook(self): |
| pytest.importorskip("pandas") |
| pytest.importorskip("pyarrow") |
| pytest.importorskip("tqdm") |
| pytest.importorskip("tqdm.notebook") |
|
|
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
|
|
| with mock.patch("tqdm.notebook.tqdm") as progress_bar_mock: |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
| df = row_iterator.to_dataframe( |
| progress_bar_type="tqdm_notebook", |
| create_bqstorage_client=False, |
| ) |
|
|
| progress_bar_mock.assert_called() |
| progress_bar_mock().update.assert_called() |
| progress_bar_mock().close.assert_called_once() |
| self.assertEqual(len(df), 4) |
|
|
| @mock.patch("google.cloud.bigquery._tqdm_helpers.tqdm", new=None) |
| def test_to_dataframe_no_tqdm_no_progress_bar(self): |
| pytest.importorskip("pandas") |
| pytest.importorskip("pyarrow") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| user_warnings = [ |
| warning for warning in warned if warning.category is UserWarning |
| ] |
| |
| |
| |
| |
| self.assertIn(len(user_warnings), [0, 3]) |
| self.assertEqual(len(df), 4) |
|
|
| @mock.patch("google.cloud.bigquery._tqdm_helpers.tqdm", new=None) |
| def test_to_dataframe_no_tqdm(self): |
| pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| df = row_iterator.to_dataframe( |
| progress_bar_type="tqdm", |
| create_bqstorage_client=False, |
| ) |
|
|
| user_warnings = [ |
| warning for warning in warned if warning.category is UserWarning |
| ] |
| |
| |
| |
| |
| self.assertIn(len(user_warnings), [1, 4]) |
|
|
| |
| |
| self.assertEqual(len(df), 4) |
|
|
| def test_to_dataframe_tqdm_error(self): |
| pytest.importorskip("pandas") |
| pytest.importorskip("tqdm") |
| mock.patch("tqdm.tqdm_gui", new=None) |
| mock.patch("tqdm.notebook.tqdm", new=None) |
| mock.patch("tqdm.tqdm", new=None) |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}]}, |
| {"f": [{"v": "Bhettye Rhubble"}, {"v": "27"}]}, |
| ] |
| path = "/foo" |
|
|
| for progress_bar_type in ("tqdm", "tqdm_notebook", "tqdm_gui"): |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| df = row_iterator.to_dataframe( |
| progress_bar_type=progress_bar_type, |
| create_bqstorage_client=False, |
| ) |
|
|
| self.assertEqual(len(df), 4) |
|
|
| |
| |
| for warning in warned: |
| self.assertIn( |
| warning.category, |
| [UserWarning, DeprecationWarning], |
| ) |
|
|
| def test_to_dataframe_w_empty_results(self): |
| pandas = pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| api_request = mock.Mock(return_value={"rows": []}) |
| row_iterator = self._make_one(_mock_client(), api_request, schema=schema) |
|
|
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 0) |
| self.assertEqual(list(df), ["name", "age"]) |
|
|
| def test_to_dataframe_w_various_types_nullable(self): |
| pandas = pytest.importorskip("pandas") |
| import datetime |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("start_timestamp", "TIMESTAMP"), |
| SchemaField("seconds", "INT64"), |
| SchemaField("miles", "FLOAT64"), |
| SchemaField("payment_type", "STRING"), |
| SchemaField("complete", "BOOL"), |
| SchemaField("date", "DATE"), |
| ] |
| row_data = [ |
| [None, None, None, None, None, None], |
| ["1433836800000000", "420", "1.1", "Cash", "true", "1999-12-01"], |
| ["1387811700000000", "2580", "17.7", "Cash", "false", "1953-06-14"], |
| ["1385565300000000", "2280", "4.4", "Credit", "true", "1981-11-04"], |
| ] |
| rows = [{"f": [{"v": field} for field in row]} for row in row_data] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 4) |
| exp_columns = [field.name for field in schema] |
| self.assertEqual(list(df), exp_columns) |
|
|
| for index, row in df.iterrows(): |
| if index == 0: |
| self.assertTrue(row.isnull().all()) |
| else: |
| self.assertIsInstance(row.start_timestamp, pandas.Timestamp) |
| self.assertIsInstance(row.seconds, int) |
| self.assertIsInstance(row.payment_type, str) |
| self.assertIsInstance(row.complete, bool) |
| self.assertIsInstance(row.date, datetime.date) |
|
|
| def test_to_dataframe_w_dtypes_mapper(self): |
| pandas = pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING"), |
| SchemaField("complete", "BOOL"), |
| SchemaField("age", "INTEGER"), |
| SchemaField("seconds", "INT64"), |
| SchemaField("miles", "FLOAT64"), |
| SchemaField("date", "DATE"), |
| SchemaField("datetime", "DATETIME"), |
| SchemaField("time", "TIME"), |
| SchemaField("timestamp", "TIMESTAMP"), |
| SchemaField("range_timestamp", "RANGE", range_element_type="TIMESTAMP"), |
| SchemaField("range_datetime", "RANGE", range_element_type="DATETIME"), |
| SchemaField("range_date", "RANGE", range_element_type="DATE"), |
| ] |
| row_data = [ |
| [ |
| "Phred Phlyntstone", |
| "true", |
| "32", |
| "23000", |
| "1.77", |
| "1999-12-01", |
| "1999-12-31T00:00:00.000000", |
| "00:00:00.000000", |
| "1433836800000000", |
| "[1433836800000000, 1433999900000000)", |
| "[2009-06-17T13:45:30, 2019-07-17T13:45:30)", |
| "[2020-10-01, 2021-10-02)", |
| ], |
| [ |
| "Bharney Rhubble", |
| "false", |
| "33", |
| "454000", |
| "6.66", |
| "4567-06-14", |
| "4567-12-31T00:00:00.000000", |
| "12:00:00.232413", |
| "81953424000000000", |
| "[1433836800000000, UNBOUNDED)", |
| "[2009-06-17T13:45:30, UNBOUNDED)", |
| "[2020-10-01, UNBOUNDED)", |
| ], |
| [ |
| "Wylma Phlyntstone", |
| "true", |
| "29", |
| "341000", |
| "2.0", |
| "9999-12-31", |
| "9999-12-31T23:59:59.999999", |
| "23:59:59.999999", |
| "253402261199999999", |
| "[UNBOUNDED, UNBOUNDED)", |
| "[UNBOUNDED, UNBOUNDED)", |
| "[UNBOUNDED, UNBOUNDED)", |
| ], |
| ] |
| rows = [{"f": [{"v": field} for field in row]} for row in row_data] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| bool_dtype=pandas.BooleanDtype(), |
| int_dtype=pandas.Int32Dtype(), |
| float_dtype=( |
| pandas.Float64Dtype() |
| if hasattr(pandas, "Float64Dtype") |
| else pandas.StringDtype() |
| ), |
| string_dtype=pandas.StringDtype(), |
| date_dtype=( |
| pandas.ArrowDtype(pyarrow.date32()) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| datetime_dtype=( |
| pandas.ArrowDtype(pyarrow.timestamp("us")) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| time_dtype=( |
| pandas.ArrowDtype(pyarrow.time64("us")) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| timestamp_dtype=( |
| pandas.ArrowDtype(pyarrow.timestamp("us", tz="UTC")) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| range_date_dtype=( |
| pandas.ArrowDtype( |
| pyarrow.struct( |
| [("start", pyarrow.date32()), ("end", pyarrow.date32())] |
| ) |
| ) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| range_datetime_dtype=( |
| pandas.ArrowDtype( |
| pyarrow.struct( |
| [ |
| ("start", pyarrow.timestamp("us")), |
| ("end", pyarrow.timestamp("us")), |
| ] |
| ) |
| ) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| range_timestamp_dtype=( |
| pandas.ArrowDtype( |
| pyarrow.struct( |
| [ |
| ("start", pyarrow.timestamp("us", tz="UTC")), |
| ("end", pyarrow.timestamp("us", tz="UTC")), |
| ] |
| ) |
| ) |
| if hasattr(pandas, "ArrowDtype") |
| else None |
| ), |
| ) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
|
|
| self.assertEqual(list(df.complete), [True, False, True]) |
| self.assertEqual(df.complete.dtype.name, "boolean") |
|
|
| self.assertEqual(list(df.age), [32, 33, 29]) |
| self.assertEqual(df.age.dtype.name, "Int32") |
|
|
| self.assertEqual(list(df.seconds), [23000, 454000, 341000]) |
| self.assertEqual(df.seconds.dtype.name, "Int32") |
|
|
| self.assertEqual( |
| list(df.name), ["Phred Phlyntstone", "Bharney Rhubble", "Wylma Phlyntstone"] |
| ) |
| self.assertEqual(df.name.dtype.name, "string") |
|
|
| if hasattr(pandas, "Float64Dtype"): |
| self.assertEqual(list(df.miles), [1.77, 6.66, 2.0]) |
| self.assertEqual(df.miles.dtype.name, "Float64") |
| else: |
| self.assertEqual(list(df.miles), ["1.77", "6.66", "2.0"]) |
| self.assertEqual(df.miles.dtype.name, "string") |
|
|
| if hasattr(pandas, "ArrowDtype"): |
| self.assertEqual( |
| list(df.date), |
| [ |
| datetime.date(1999, 12, 1), |
| datetime.date(4567, 6, 14), |
| datetime.date(9999, 12, 31), |
| ], |
| ) |
| self.assertEqual(df.date.dtype.name, "date32[day][pyarrow]") |
|
|
| self.assertEqual( |
| list(df.datetime), |
| [ |
| datetime.datetime(1999, 12, 31, 0, 0), |
| datetime.datetime(4567, 12, 31, 0, 0), |
| datetime.datetime(9999, 12, 31, 23, 59, 59, 999999), |
| ], |
| ) |
| self.assertEqual(df.datetime.dtype.name, "timestamp[us][pyarrow]") |
|
|
| self.assertEqual( |
| list(df.time), |
| [ |
| datetime.time(0, 0), |
| datetime.time(12, 0, 0, 232413), |
| datetime.time(23, 59, 59, 999999), |
| ], |
| ) |
| self.assertEqual(df.time.dtype.name, "time64[us][pyarrow]") |
|
|
| self.assertEqual( |
| list(df.timestamp), |
| [ |
| datetime.datetime(2015, 6, 9, 8, 0, tzinfo=datetime.timezone.utc), |
| datetime.datetime(4567, 1, 1, 0, 0, tzinfo=datetime.timezone.utc), |
| datetime.datetime( |
| 9999, 12, 31, 12, 59, 59, 999999, tzinfo=datetime.timezone.utc |
| ), |
| ], |
| ) |
| self.assertEqual(df.timestamp.dtype.name, "timestamp[us, tz=UTC][pyarrow]") |
|
|
| self.assertEqual( |
| list(df.range_timestamp), |
| [ |
| { |
| "start": datetime.datetime( |
| 2015, 6, 9, 8, 0, 0, tzinfo=datetime.timezone.utc |
| ), |
| "end": datetime.datetime( |
| 2015, 6, 11, 5, 18, 20, tzinfo=datetime.timezone.utc |
| ), |
| }, |
| { |
| "start": datetime.datetime( |
| 2015, 6, 9, 8, 0, 0, tzinfo=datetime.timezone.utc |
| ), |
| "end": None, |
| }, |
| {"start": None, "end": None}, |
| ], |
| ) |
|
|
| self.assertEqual( |
| list(df.range_datetime), |
| [ |
| { |
| "start": datetime.datetime(2009, 6, 17, 13, 45, 30), |
| "end": datetime.datetime(2019, 7, 17, 13, 45, 30), |
| }, |
| {"start": datetime.datetime(2009, 6, 17, 13, 45, 30), "end": None}, |
| {"start": None, "end": None}, |
| ], |
| ) |
|
|
| self.assertEqual( |
| list(df.range_date), |
| [ |
| { |
| "start": datetime.date(2020, 10, 1), |
| "end": datetime.date(2021, 10, 2), |
| }, |
| {"start": datetime.date(2020, 10, 1), "end": None}, |
| {"start": None, "end": None}, |
| ], |
| ) |
|
|
| else: |
| self.assertEqual( |
| list(df.date), |
| [ |
| pandas.Timestamp("1999-12-01 00:00:00"), |
| pandas.Timestamp("2229-03-27 01:41:45.161793536"), |
| pandas.Timestamp("1816-03-29 05:56:08.066277376"), |
| ], |
| ) |
| self.assertEqual(df.date.dtype.name, "datetime64[ns]") |
|
|
| self.assertEqual( |
| list(df.datetime), |
| [ |
| datetime.datetime(1999, 12, 31, 0, 0), |
| datetime.datetime(4567, 12, 31, 0, 0), |
| datetime.datetime(9999, 12, 31, 23, 59, 59, 999999), |
| ], |
| ) |
| self.assertEqual(df.datetime.dtype.name, "object") |
|
|
| self.assertEqual( |
| list(df.time), |
| [ |
| datetime.time(0, 0), |
| datetime.time(12, 0, 0, 232413), |
| datetime.time(23, 59, 59, 999999), |
| ], |
| ) |
| self.assertEqual(df.time.dtype.name, "object") |
|
|
| self.assertEqual( |
| list(df.timestamp), |
| [ |
| datetime.datetime(2015, 6, 9, 8, 0, tzinfo=datetime.timezone.utc), |
| datetime.datetime(4567, 1, 1, 0, 0, tzinfo=datetime.timezone.utc), |
| datetime.datetime( |
| 9999, 12, 31, 12, 59, 59, 999999, tzinfo=datetime.timezone.utc |
| ), |
| ], |
| ) |
| self.assertEqual(df.timestamp.dtype.name, "object") |
|
|
| def test_to_dataframe_w_none_dtypes_mapper(self): |
| pandas = pytest.importorskip("pandas") |
| pandas_major_version = pandas.__version__[0:2] |
| if pandas_major_version not in ["0.", "1."]: |
| pytest.skip(reason="Requires a version of pandas less than 2.0") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING"), |
| SchemaField("complete", "BOOL"), |
| SchemaField("age", "INTEGER"), |
| SchemaField("seconds", "INT64"), |
| SchemaField("miles", "FLOAT64"), |
| SchemaField("date", "DATE"), |
| SchemaField("datetime", "DATETIME"), |
| SchemaField("time", "TIME"), |
| SchemaField("timestamp", "TIMESTAMP"), |
| SchemaField("range_timestamp", "RANGE", range_element_type="TIMESTAMP"), |
| SchemaField("range_datetime", "RANGE", range_element_type="DATETIME"), |
| SchemaField("range_date", "RANGE", range_element_type="DATE"), |
| ] |
| row_data = [ |
| [ |
| "Phred Phlyntstone", |
| "true", |
| "32", |
| "23000", |
| "1.77", |
| "1999-12-01", |
| "1999-12-31T00:00:00.000000", |
| "23:59:59.999999", |
| "1433836800000000", |
| "[1433836800000000, 1433999900000000)", |
| "[2009-06-17T13:45:30, 2019-07-17T13:45:30)", |
| "[2020-10-01, 2021-10-02)", |
| ], |
| ] |
| rows = [{"f": [{"v": field} for field in row]} for row in row_data] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| bool_dtype=None, |
| int_dtype=None, |
| float_dtype=None, |
| string_dtype=None, |
| date_dtype=None, |
| datetime_dtype=None, |
| time_dtype=None, |
| timestamp_dtype=None, |
| range_timestamp_dtype=None, |
| range_datetime_dtype=None, |
| range_date_dtype=None, |
| ) |
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(df.complete.dtype.name, "bool") |
| self.assertEqual(df.age.dtype.name, "int64") |
| self.assertEqual(df.seconds.dtype.name, "int64") |
| self.assertEqual(df.miles.dtype.name, "float64") |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertEqual(df.date.dtype.name, "datetime64[ns]") |
| self.assertEqual(df.datetime.dtype.name, "datetime64[ns]") |
| self.assertEqual(df.time.dtype.name, "object") |
| self.assertEqual(df.timestamp.dtype.name, "datetime64[ns, UTC]") |
| self.assertEqual(df.range_timestamp.dtype.name, "object") |
| self.assertEqual(df.range_datetime.dtype.name, "object") |
| self.assertEqual(df.range_date.dtype.name, "object") |
|
|
| def test_to_dataframe_w_unsupported_dtypes_mapper(self): |
| pytest.importorskip("pandas") |
| import numpy |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING"), |
| ] |
| row_data = [ |
| ["Phred Phlyntstone"], |
| ] |
| rows = [{"f": [{"v": field} for field in row]} for row in row_data] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| bool_dtype=numpy.dtype("bool"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| int_dtype=numpy.dtype("int64"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| float_dtype=numpy.dtype("float64"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| string_dtype=numpy.dtype("object"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| date_dtype=numpy.dtype("object"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| datetime_dtype=numpy.dtype("datetime64[us]"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| time_dtype=numpy.dtype("datetime64[us]"), |
| ) |
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| timestamp_dtype=numpy.dtype("datetime64[us]"), |
| ) |
|
|
| def test_to_dataframe_column_dtypes(self): |
| pandas = pytest.importorskip("pandas") |
| pandas_major_version = pandas.__version__[0:2] |
| if pandas_major_version not in ["0.", "1."]: |
| pytest.skip("Requires a version of pandas less than 2.0") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("start_timestamp", "TIMESTAMP"), |
| SchemaField("seconds", "INT64"), |
| SchemaField("miles", "FLOAT64"), |
| SchemaField("km", "FLOAT64"), |
| SchemaField("payment_type", "STRING"), |
| SchemaField("complete", "BOOL"), |
| SchemaField("date", "DATE"), |
| ] |
| row_data = [ |
| ["1433836800000", "420", "1.1", "1.77", "Cash", "true", "1999-12-01"], |
| [ |
| "1387811700000", |
| "2580", |
| "17.7", |
| "28.5", |
| "Cash", |
| "false", |
| "1953-06-14", |
| ], |
| ["1385565300000", "2280", "4.4", "7.1", "Credit", "true", "1981-11-04"], |
| ] |
| rows = [{"f": [{"v": field} for field in row]} for row in row_data] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe( |
| dtypes={"km": "float16"}, |
| create_bqstorage_client=False, |
| ) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 3) |
| exp_columns = [field.name for field in schema] |
| self.assertEqual(list(df), exp_columns) |
|
|
| self.assertEqual(df.start_timestamp.dtype.name, "datetime64[ns, UTC]") |
| self.assertEqual(df.seconds.dtype.name, "Int64") |
| self.assertEqual(df.miles.dtype.name, "float64") |
| self.assertEqual(df.km.dtype.name, "float16") |
| self.assertEqual(df.payment_type.dtype.name, "object") |
| self.assertEqual(df.complete.dtype.name, "boolean") |
| self.assertEqual(df.date.dtype.name, "dbdate") |
|
|
| def test_to_dataframe_datetime_objects(self): |
| |
| |
| |
| |
| pandas = pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("ts", "TIMESTAMP"), |
| SchemaField("date", "DATE"), |
| ] |
| row_data = [ |
| ["-20000000000000000", "1111-01-01"], |
| ] |
| rows = [{"f": [{"v": field} for field in row]} for row in row_data] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| df = row_iterator.to_dataframe(create_bqstorage_client=False) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 1) |
| self.assertEqual(df["ts"].dtype.name, "object") |
| self.assertEqual(df["date"].dtype.name, "object") |
| self.assertEqual(df["ts"][0].date(), datetime.date(1336, 3, 23)) |
| self.assertEqual(df["date"][0], datetime.date(1111, 1, 1)) |
|
|
| @mock.patch("google.cloud.bigquery._pandas_helpers.pandas", new=None) |
| def test_to_dataframe_error_if_pandas_is_none(self): |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| row_iterator = self._make_one(_mock_client(), api_request, path, schema) |
|
|
| with self.assertRaises(ValueError): |
| row_iterator.to_dataframe() |
|
|
| @mock.patch("google.cloud.bigquery.table.shapely", new=None) |
| def test_to_dataframe_error_if_shapely_is_none(self): |
| pytest.importorskip("pandas") |
|
|
| with self.assertRaisesRegex( |
| ValueError, |
| re.escape( |
| "The shapely library is not installed, please install " |
| "shapely to use the geography_as_object option." |
| ), |
| ): |
| self._make_one_from_data().to_dataframe(geography_as_object=True) |
|
|
| def test_to_dataframe_max_results_w_bqstorage_warning(self): |
| pytest.importorskip("pandas") |
|
|
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| bqstorage_client = mock.Mock() |
|
|
| row_iterator = self._make_one( |
| client=_mock_client(), |
| api_request=api_request, |
| path=path, |
| schema=schema, |
| max_results=42, |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| row_iterator.to_dataframe(bqstorage_client=bqstorage_client) |
|
|
| matches = [ |
| warning |
| for warning in warned |
| if warning.category is UserWarning |
| and "cannot use bqstorage_client" in str(warning).lower() |
| and "REST" in str(warning) |
| ] |
| self.assertEqual(len(matches), 1, msg="User warning was not emitted.") |
|
|
| def test_to_dataframe_max_results_w_explicit_bqstorage_client_warning(self): |
| pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| mock_client = _mock_client() |
| mock_bqstorage_client = mock.sentinel.bq_storage_client |
|
|
| row_iterator = self._make_one( |
| client=mock_client, |
| api_request=api_request, |
| path=path, |
| schema=schema, |
| max_results=42, |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| row_iterator.to_dataframe(bqstorage_client=mock_bqstorage_client) |
|
|
| matches = [ |
| warning |
| for warning in warned |
| if warning.category is UserWarning |
| and "cannot use bqstorage_client" in str(warning).lower() |
| and "REST" in str(warning) |
| ] |
| self.assertEqual(len(matches), 1, msg="User warning was not emitted.") |
| self.assertIn( |
| __file__, str(matches[0]), msg="Warning emitted with incorrect stacklevel" |
| ) |
| mock_client._ensure_bqstorage_client.assert_not_called() |
|
|
| def test_to_dataframe_max_results_w_create_bqstorage_client_no_warning(self): |
| pytest.importorskip("pandas") |
| from google.cloud.bigquery.schema import SchemaField |
|
|
| schema = [ |
| SchemaField("name", "STRING", mode="REQUIRED"), |
| SchemaField("age", "INTEGER", mode="REQUIRED"), |
| ] |
| rows = [ |
| {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, |
| {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}]}, |
| ] |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": rows}) |
| mock_client = _mock_client() |
|
|
| row_iterator = self._make_one( |
| client=mock_client, |
| api_request=api_request, |
| path=path, |
| schema=schema, |
| max_results=42, |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| row_iterator.to_dataframe(create_bqstorage_client=True) |
|
|
| matches = [ |
| warning |
| for warning in warned |
| if warning.category is UserWarning |
| and "cannot use bqstorage_client" in str(warning).lower() |
| and "REST" in str(warning) |
| ] |
| self.assertFalse(matches) |
| mock_client._ensure_bqstorage_client.assert_not_called() |
|
|
| def test_to_dataframe_w_bqstorage_creates_client(self): |
| pytest.importorskip("pandas") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| mock_client = _mock_client() |
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| mock_client._ensure_bqstorage_client.return_value = bqstorage_client |
| session = bigquery_storage.types.ReadSession() |
| bqstorage_client.create_read_session.return_value = session |
| row_iterator = mut.RowIterator( |
| mock_client, |
| None, |
| None, |
| [ |
| schema.SchemaField("colA", "STRING"), |
| schema.SchemaField("colC", "STRING"), |
| schema.SchemaField("colB", "STRING"), |
| ], |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| ) |
| row_iterator.to_dataframe(create_bqstorage_client=True) |
| mock_client._ensure_bqstorage_client.assert_called_once() |
| bqstorage_client._transport.grpc_channel.close.assert_called_once() |
|
|
| def test_to_dataframe_w_bqstorage_no_streams(self): |
| pytest.importorskip("pandas") |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| session = bigquery_storage.types.ReadSession() |
| bqstorage_client.create_read_session.return_value = session |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| api_request=None, |
| path=None, |
| schema=[ |
| schema.SchemaField("colA", "INTEGER"), |
| schema.SchemaField("colC", "FLOAT"), |
| schema.SchemaField("colB", "STRING"), |
| ], |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| ) |
|
|
| got = row_iterator.to_dataframe(bqstorage_client) |
| column_names = ["colA", "colC", "colB"] |
| self.assertEqual(list(got), column_names) |
| self.assertTrue(got.empty) |
|
|
| def test_to_dataframe_w_bqstorage_logs_session(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| pytest.importorskip("pyarrow") |
| from google.cloud.bigquery.table import Table |
| from google.cloud import bigquery_storage |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| session = bigquery_storage.types.ReadSession() |
| session.name = "projects/test-proj/locations/us/sessions/SOMESESSION" |
| bqstorage_client.create_read_session.return_value = session |
| mock_logger = mock.create_autospec(logging.Logger) |
| row_iterator = self._make_one( |
| _mock_client(), table=Table("debug-proj.debug_dset.debug_tbl") |
| ) |
|
|
| with mock.patch("google.cloud.bigquery._pandas_helpers._LOGGER", mock_logger): |
| row_iterator.to_dataframe(bqstorage_client=bqstorage_client) |
|
|
| mock_logger.debug.assert_any_call( |
| "Started reading table 'debug-proj.debug_dset.debug_tbl' " |
| "with BQ Storage API session 'projects/test-proj/locations/us/sessions/SOMESESSION'." |
| ) |
|
|
| def test_to_dataframe_w_bqstorage_empty_streams(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud.bigquery_storage_v1 import reader |
|
|
| arrow_fields = [ |
| pyarrow.field("colA", pyarrow.int64()), |
| |
| pyarrow.field("colC", pyarrow.float64()), |
| pyarrow.field("colB", pyarrow.utf8()), |
| ] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| session = bigquery_storage.types.ReadSession( |
| streams=[{"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}], |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| mock_pages = mock.PropertyMock(return_value=()) |
| type(mock_rows).pages = mock_pages |
|
|
| |
| schema = [ |
| schema.SchemaField("colA", "INTEGER"), |
| schema.SchemaField("colC", "FLOAT"), |
| schema.SchemaField("colB", "STRING"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| got = row_iterator.to_dataframe(bqstorage_client) |
|
|
| column_names = ["colA", "colC", "colB"] |
| self.assertEqual(list(got), column_names) |
| self.assertTrue(got.empty) |
|
|
| def test_to_dataframe_w_bqstorage_nonempty(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery_storage_v1 import reader |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| arrow_fields = [ |
| pyarrow.field("colA", pyarrow.int64()), |
| |
| pyarrow.field("colC", pyarrow.float64()), |
| pyarrow.field("colB", pyarrow.utf8()), |
| ] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| streams = [ |
| |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| ] |
| session = bigquery_storage.types.ReadSession( |
| streams=streams, |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| page_items = [ |
| pyarrow.array([1, -1]), |
| pyarrow.array([2.0, 4.0]), |
| pyarrow.array(["abc", "def"]), |
| ] |
| page_record_batch = pyarrow.RecordBatch.from_arrays( |
| page_items, schema=arrow_schema |
| ) |
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_arrow.return_value = page_record_batch |
| mock_pages = (mock_page, mock_page, mock_page) |
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| schema = [ |
| schema.SchemaField("colA", "IGNORED"), |
| schema.SchemaField("colC", "IGNORED"), |
| schema.SchemaField("colB", "IGNORED"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| got = row_iterator.to_dataframe(bqstorage_client=bqstorage_client) |
|
|
| |
| column_names = ["colA", "colC", "colB"] |
| self.assertEqual(list(got), column_names) |
|
|
| |
| total_pages = len(streams) * len(mock_pages) |
| total_rows = len(page_items[0]) * total_pages |
| self.assertEqual(len(got.index), total_rows) |
|
|
| |
| bqstorage_client._transport.grpc_channel.close.assert_not_called() |
|
|
| def test_to_dataframe_w_bqstorage_multiple_streams_return_unique_index(self): |
| bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud.bigquery_storage_v1 import reader |
|
|
| arrow_fields = [pyarrow.field("colA", pyarrow.int64())] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| streams = [ |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| ] |
| session = bigquery_storage.types.ReadSession( |
| streams=streams, |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
|
|
| page_items = [ |
| pyarrow.array([1, -1]), |
| ] |
| page_record_batch = pyarrow.RecordBatch.from_arrays( |
| page_items, schema=arrow_schema |
| ) |
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_arrow.return_value = page_record_batch |
| mock_pages = (mock_page, mock_page, mock_page) |
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| row_iterator = self._make_one( |
| schema=[schema.SchemaField("colA", "IGNORED")], |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| ) |
| got = row_iterator.to_dataframe(bqstorage_client=bqstorage_client) |
|
|
| self.assertEqual(list(got), ["colA"]) |
| total_pages = len(streams) * len(mock_pages) |
| total_rows = len(page_items[0]) * total_pages |
| self.assertEqual(len(got.index), total_rows) |
| self.assertTrue(got.index.is_unique) |
|
|
| def test_to_dataframe_w_bqstorage_updates_progress_bar(self): |
| bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| pytest.importorskip("tqdm") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud.bigquery_storage_v1 import reader |
|
|
| |
| mut._PROGRESS_INTERVAL = 0.01 |
|
|
| arrow_fields = [pyarrow.field("testcol", pyarrow.int64())] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| streams = [ |
| |
| |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| ] |
| session = bigquery_storage.types.ReadSession( |
| streams=streams, |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| page_items = [-1, 0, 1] |
| type(mock_page).num_items = mock.PropertyMock(return_value=len(page_items)) |
|
|
| def blocking_to_arrow(*args, **kwargs): |
| |
| |
| time.sleep(2 * mut._PROGRESS_INTERVAL) |
| return pyarrow.RecordBatch.from_arrays( |
| [pyarrow.array(page_items)], schema=arrow_schema |
| ) |
|
|
| mock_page.to_arrow.side_effect = blocking_to_arrow |
| mock_pages = (mock_page, mock_page, mock_page, mock_page, mock_page) |
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| schema = [schema.SchemaField("testcol", "IGNORED")] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| with mock.patch("tqdm.tqdm") as tqdm_mock: |
| row_iterator.to_dataframe( |
| bqstorage_client=bqstorage_client, progress_bar_type="tqdm" |
| ) |
|
|
| |
| |
| total_pages = len(streams) * len(mock_pages) |
| expected_total_rows = total_pages * len(page_items) |
| progress_updates = [ |
| args[0] for args, kwargs in tqdm_mock().update.call_args_list |
| ] |
| |
| self.assertGreater(len(progress_updates), 1) |
| self.assertEqual(sum(progress_updates), expected_total_rows) |
| tqdm_mock().close.assert_called_once() |
|
|
| def test_to_dataframe_w_bqstorage_exits_on_keyboardinterrupt(self): |
| bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud.bigquery_storage_v1 import reader |
|
|
| |
| mut._PROGRESS_INTERVAL = 0.01 |
|
|
| arrow_fields = [ |
| pyarrow.field("colA", pyarrow.int64()), |
| |
| pyarrow.field("colC", pyarrow.float64()), |
| pyarrow.field("colB", pyarrow.utf8()), |
| ] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| session = bigquery_storage.types.ReadSession( |
| streams=[ |
| |
| |
| |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/5678"}, |
| {"name": "/projects/proj/dataset/dset/tables/tbl/streams/9999"}, |
| ], |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
| bqstorage_client.create_read_session.return_value = session |
| page_items = [ |
| pyarrow.array([1, -1]), |
| pyarrow.array([2.0, 4.0]), |
| pyarrow.array(["abc", "def"]), |
| ] |
|
|
| def blocking_to_arrow(*args, **kwargs): |
| |
| |
| time.sleep(2 * mut._PROGRESS_INTERVAL) |
| return pyarrow.RecordBatch.from_arrays(page_items, schema=arrow_schema) |
|
|
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_arrow.side_effect = blocking_to_arrow |
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_pages = mock.PropertyMock(return_value=(mock_page, mock_page, mock_page)) |
| type(mock_rows).pages = mock_pages |
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| mock_rowstream.rows.return_value = mock_rows |
|
|
| mock_cancelled_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_cancelled_pages = mock.PropertyMock(side_effect=KeyboardInterrupt) |
| type(mock_cancelled_rows).pages = mock_cancelled_pages |
| mock_cancelled_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| mock_cancelled_rowstream.rows.return_value = mock_cancelled_rows |
|
|
| bqstorage_client.read_rows.side_effect = ( |
| mock_rowstream, |
| mock_cancelled_rowstream, |
| mock_rowstream, |
| ) |
|
|
| schema = [ |
| schema.SchemaField("colA", "IGNORED"), |
| schema.SchemaField("colB", "IGNORED"), |
| schema.SchemaField("colC", "IGNORED"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| with pytest.raises(KeyboardInterrupt): |
| row_iterator.to_dataframe(bqstorage_client=bqstorage_client) |
|
|
| |
| |
| self.assertLessEqual(mock_page.to_dataframe.call_count, 2) |
|
|
| def test_to_dataframe_tabledata_list_w_multiple_pages_return_unique_index(self): |
| pandas = pytest.importorskip("pandas") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
|
|
| iterator_schema = [schema.SchemaField("name", "STRING", mode="REQUIRED")] |
| path = "/foo" |
| api_request = mock.Mock( |
| side_effect=[ |
| {"rows": [{"f": [{"v": "Bengt"}]}], "pageToken": "NEXTPAGE"}, |
| {"rows": [{"f": [{"v": "Sven"}]}]}, |
| ] |
| ) |
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| api_request, |
| path, |
| iterator_schema, |
| table=mut.Table("proj.dset.tbl"), |
| ) |
|
|
| df = row_iterator.to_dataframe( |
| bqstorage_client=None, |
| create_bqstorage_client=False, |
| ) |
|
|
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 2) |
| self.assertEqual(list(df), ["name"]) |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertTrue(df.index.is_unique) |
|
|
| def test_to_dataframe_w_bqstorage_raises_auth_error(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| pytest.importorskip("pandas") |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery import table as mut |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client.create_read_session.side_effect = ( |
| google.api_core.exceptions.Forbidden( |
| "TEST BigQuery Storage API not enabled. TEST" |
| ) |
| ) |
| path = "/foo" |
| api_request = mock.Mock(return_value={"rows": []}) |
| row_iterator = mut.RowIterator( |
| _mock_client(), api_request, path, [], table=mut.Table("proj.dset.tbl") |
| ) |
|
|
| with pytest.raises(google.api_core.exceptions.Forbidden): |
| row_iterator.to_dataframe(bqstorage_client=bqstorage_client) |
|
|
| def test_to_dataframe_w_bqstorage_partition(self): |
| bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| [schema.SchemaField("colA", "IGNORED")], |
| table=mut.TableReference.from_string("proj.dset.tbl$20181225"), |
| ) |
|
|
| with pytest.raises(ValueError): |
| row_iterator.to_dataframe(bqstorage_client) |
|
|
| def test_to_dataframe_w_bqstorage_snapshot(self): |
| bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
|
|
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| [schema.SchemaField("colA", "IGNORED")], |
| table=mut.TableReference.from_string("proj.dset.tbl@1234567890000"), |
| ) |
|
|
| with pytest.raises(ValueError): |
| row_iterator.to_dataframe(bqstorage_client) |
|
|
| def test_to_dataframe_concat_categorical_dtype_w_pyarrow(self): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| pandas = pytest.importorskip("pandas") |
| pyarrow = pytest.importorskip("pyarrow") |
| from google.cloud import bigquery_storage |
| from google.cloud.bigquery import schema |
| from google.cloud.bigquery import table as mut |
| from google.cloud.bigquery_storage_v1 import reader |
| from google.cloud.bigquery_storage_v1.services.big_query_read.transports import ( |
| grpc as big_query_read_grpc_transport, |
| ) |
|
|
| arrow_fields = [ |
| |
| pyarrow.field("col_str", pyarrow.utf8()), |
| |
| |
| |
| pyarrow.field("col_category", pyarrow.utf8()), |
| ] |
| arrow_schema = pyarrow.schema(arrow_fields) |
|
|
| |
| bqstorage_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| bqstorage_client._transport = mock.create_autospec( |
| big_query_read_grpc_transport.BigQueryReadGrpcTransport |
| ) |
| session = bigquery_storage.types.ReadSession( |
| streams=[{"name": "/projects/proj/dataset/dset/tables/tbl/streams/1234"}], |
| arrow_schema={"serialized_schema": arrow_schema.serialize().to_pybytes()}, |
| ) |
| bqstorage_client.create_read_session.return_value = session |
|
|
| mock_rowstream = mock.create_autospec(reader.ReadRowsStream) |
| bqstorage_client.read_rows.return_value = mock_rowstream |
|
|
| |
| mock_rows = mock.create_autospec(reader.ReadRowsIterable) |
| mock_rowstream.rows.return_value = mock_rows |
| page_items = [ |
| [ |
| pyarrow.array(["foo", "bar", "baz"]), |
| pyarrow.array(["low", "medium", "low"]), |
| ], |
| [ |
| pyarrow.array(["foo_page2", "bar_page2", "baz_page2"]), |
| pyarrow.array(["medium", "high", "low"]), |
| ], |
| ] |
|
|
| mock_pages = [] |
|
|
| for record_list in page_items: |
| page_record_batch = pyarrow.RecordBatch.from_arrays( |
| record_list, schema=arrow_schema |
| ) |
| mock_page = mock.create_autospec(reader.ReadRowsPage) |
| mock_page.to_arrow.return_value = page_record_batch |
| mock_pages.append(mock_page) |
|
|
| type(mock_rows).pages = mock.PropertyMock(return_value=mock_pages) |
|
|
| schema = [ |
| schema.SchemaField("col_str", "IGNORED"), |
| schema.SchemaField("col_category", "IGNORED"), |
| ] |
|
|
| row_iterator = mut.RowIterator( |
| _mock_client(), |
| None, |
| None, |
| schema, |
| table=mut.TableReference.from_string("proj.dset.tbl"), |
| selected_fields=schema, |
| ) |
|
|
| |
| got = row_iterator.to_dataframe( |
| bqstorage_client=bqstorage_client, |
| dtypes={ |
| "col_category": pandas.core.dtypes.dtypes.CategoricalDtype( |
| categories=["low", "medium", "high"], |
| ordered=False, |
| ), |
| }, |
| ) |
|
|
| |
| column_names = ["col_str", "col_category"] |
| self.assertEqual(list(got), column_names) |
|
|
| |
| total_pages = len(mock_pages) |
| total_rows = len(page_items[0][0]) * total_pages |
| self.assertEqual(len(got.index), total_rows) |
|
|
| |
| expected_dtypes = [ |
| pandas.core.dtypes.dtypes.np.dtype("O"), |
| pandas.core.dtypes.dtypes.CategoricalDtype( |
| categories=["low", "medium", "high"], |
| ordered=False, |
| ), |
| ] |
| self.assertEqual(list(got.dtypes), expected_dtypes) |
|
|
| |
| self.assertEqual( |
| list(got["col_category"]), |
| ["low", "medium", "low", "medium", "high", "low"], |
| ) |
|
|
| |
| bqstorage_client._transport.grpc_channel.close.assert_not_called() |
|
|
| def test_to_dataframe_geography_as_object(self): |
| pandas = pytest.importorskip("pandas") |
| pytest.importorskip("geopandas") |
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("geog", "GEOGRAPHY")), |
| ( |
| ("foo", "Point(0 0)"), |
| ("bar", None), |
| ("baz", "Polygon((0 0, 0 1, 1 0, 0 0))"), |
| ), |
| ) |
| df = row_iterator.to_dataframe( |
| create_bqstorage_client=False, |
| geography_as_object=True, |
| ) |
| self.assertIsInstance(df, pandas.DataFrame) |
| self.assertEqual(len(df), 3) |
| self.assertEqual(list(df), ["name", "geog"]) |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertEqual(df.geog.dtype.name, "object") |
| self.assertIsInstance(df.geog, pandas.Series) |
| self.assertEqual( |
| [v.__class__.__name__ for v in df.geog], ["Point", "float", "Polygon"] |
| ) |
|
|
| @mock.patch("google.cloud.bigquery.table.geopandas", new=None) |
| def test_to_geodataframe_error_if_geopandas_is_none(self): |
| with self.assertRaisesRegex( |
| ValueError, |
| re.escape( |
| "The geopandas library is not installed, please install " |
| "geopandas to use the to_geodataframe() function." |
| ), |
| ): |
| self._make_one_from_data().to_geodataframe() |
|
|
| def test_to_geodataframe(self): |
| geopandas = pytest.importorskip("geopandas") |
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("geog", "GEOGRAPHY")), |
| ( |
| ("foo", "Point(0 0)"), |
| ("bar", None), |
| ("baz", "Polygon((0 0, 0 1, 1 0, 0 0))"), |
| ), |
| ) |
| df = row_iterator.to_geodataframe(create_bqstorage_client=False) |
| self.assertIsInstance(df, geopandas.GeoDataFrame) |
| self.assertEqual(len(df), 3) |
| self.assertEqual(list(df), ["name", "geog"]) |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertEqual(df.geog.dtype.name, "geometry") |
| self.assertIsInstance(df.geog, geopandas.GeoSeries) |
|
|
| with warnings.catch_warnings(): |
| |
| |
| warnings.filterwarnings("ignore", category=UserWarning) |
| self.assertEqual(list(map(str, df.area)), ["0.0", "nan", "0.5"]) |
| self.assertEqual(list(map(str, df.geog.area)), ["0.0", "nan", "0.5"]) |
|
|
| self.assertEqual(df.crs.srs, "EPSG:4326") |
| self.assertEqual(df.crs.name, "WGS 84") |
| self.assertEqual(df.geog.crs.srs, "EPSG:4326") |
| self.assertEqual(df.geog.crs.name, "WGS 84") |
|
|
| def test_to_geodataframe_ambiguous_geog(self): |
| pytest.importorskip("geopandas") |
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("geog", "GEOGRAPHY"), ("geog2", "GEOGRAPHY")), () |
| ) |
| with self.assertRaisesRegex( |
| ValueError, |
| re.escape( |
| "There is more than one GEOGRAPHY column in the result. " |
| "The geography_column argument must be used to specify which " |
| "one to use to create a GeoDataFrame" |
| ), |
| ): |
| row_iterator.to_geodataframe(create_bqstorage_client=False) |
|
|
| def test_to_geodataframe_bad_geography_column(self): |
| pytest.importorskip("geopandas") |
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("geog", "GEOGRAPHY"), ("geog2", "GEOGRAPHY")), () |
| ) |
| with self.assertRaisesRegex( |
| ValueError, |
| re.escape( |
| "The given geography column, xxx, doesn't name" |
| " a GEOGRAPHY column in the result." |
| ), |
| ): |
| row_iterator.to_geodataframe( |
| create_bqstorage_client=False, geography_column="xxx" |
| ) |
|
|
| def test_to_geodataframe_no_geog(self): |
| pytest.importorskip("geopandas") |
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("geog", "STRING")), () |
| ) |
| with self.assertRaisesRegex( |
| TypeError, |
| re.escape( |
| "There must be at least one GEOGRAPHY column" |
| " to create a GeoDataFrame" |
| ), |
| ): |
| row_iterator.to_geodataframe(create_bqstorage_client=False) |
|
|
| def test_to_geodataframe_w_geography_column(self): |
| geopandas = pytest.importorskip("geopandas") |
| pandas = pytest.importorskip("pandas") |
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("geog", "GEOGRAPHY"), ("geog2", "GEOGRAPHY")), |
| ( |
| ("foo", "Point(0 0)", "Point(1 1)"), |
| ("bar", None, "Point(2 2)"), |
| ("baz", "Polygon((0 0, 0 1, 1 0, 0 0))", "Point(3 3)"), |
| ), |
| ) |
| df = row_iterator.to_geodataframe( |
| create_bqstorage_client=False, geography_column="geog" |
| ) |
| self.assertIsInstance(df, geopandas.GeoDataFrame) |
| self.assertEqual(len(df), 3) |
| self.assertEqual(list(df), ["name", "geog", "geog2"]) |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertEqual(df.geog.dtype.name, "geometry") |
| self.assertEqual(df.geog2.dtype.name, "object") |
| self.assertIsInstance(df.geog, geopandas.GeoSeries) |
|
|
| with warnings.catch_warnings(): |
| |
| |
| warnings.filterwarnings("ignore", category=UserWarning) |
| self.assertEqual(list(map(str, df.area)), ["0.0", "nan", "0.5"]) |
| self.assertEqual(list(map(str, df.geog.area)), ["0.0", "nan", "0.5"]) |
|
|
| self.assertEqual( |
| [v.__class__.__name__ for v in df.geog], ["Point", "NoneType", "Polygon"] |
| ) |
|
|
| |
| self.assertIsInstance(df.geog2, pandas.Series) |
| self.assertEqual( |
| [v.__class__.__name__ for v in df.geog2], ["Point", "Point", "Point"] |
| ) |
|
|
| |
| with warnings.catch_warnings(): |
| warnings.filterwarnings("ignore", category=UserWarning) |
| self.assertEqual( |
| list(map(str, geopandas.GeoSeries(df.geog2).area)), |
| ["0.0", "0.0", "0.0"], |
| ) |
|
|
| @mock.patch("google.cloud.bigquery.table.RowIterator.to_dataframe") |
| def test_rowiterator_to_geodataframe_delegation(self, to_dataframe): |
| """ |
| RowIterator.to_geodataframe just delegates to RowIterator.to_dataframe. |
| |
| This test just demonstrates that. We don't need to test all the |
| variations, which are tested for to_dataframe. |
| """ |
| pandas = pytest.importorskip("pandas") |
| geopandas = pytest.importorskip("geopandas") |
| import numpy |
| from shapely import wkt |
|
|
| row_iterator = self._make_one_from_data( |
| (("name", "STRING"), ("g", "GEOGRAPHY")) |
| ) |
| bqstorage_client = object() |
| dtypes = dict(xxx=numpy.dtype("int64")) |
| progress_bar_type = "normal" |
| create_bqstorage_client = False |
| geography_column = "g" |
|
|
| to_dataframe.return_value = pandas.DataFrame( |
| dict( |
| name=["foo"], |
| g=[wkt.loads("point(0 0)")], |
| ) |
| ) |
|
|
| df = row_iterator.to_geodataframe( |
| bqstorage_client=bqstorage_client, |
| dtypes=dtypes, |
| progress_bar_type=progress_bar_type, |
| create_bqstorage_client=create_bqstorage_client, |
| geography_column=geography_column, |
| ) |
|
|
| to_dataframe.assert_called_once_with( |
| bqstorage_client, |
| dtypes, |
| progress_bar_type, |
| create_bqstorage_client, |
| geography_as_object=True, |
| ) |
|
|
| self.assertIsInstance(df, geopandas.GeoDataFrame) |
| self.assertEqual(len(df), 1) |
| self.assertEqual(list(df), ["name", "g"]) |
| self.assertEqual(df.name.dtype.name, "object") |
| self.assertEqual(df.g.dtype.name, "geometry") |
| self.assertIsInstance(df.g, geopandas.GeoSeries) |
|
|
| with warnings.catch_warnings(): |
| |
| |
| warnings.filterwarnings("ignore", category=UserWarning) |
| self.assertEqual(list(map(str, df.area)), ["0.0"]) |
| self.assertEqual(list(map(str, df.g.area)), ["0.0"]) |
|
|
| self.assertEqual([v.__class__.__name__ for v in df.g], ["Point"]) |
|
|
|
|
| class TestPartitionRange(unittest.TestCase): |
| def _get_target_class(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| return PartitionRange |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def test_constructor_defaults(self): |
| object_under_test = self._make_one() |
| assert object_under_test.start is None |
| assert object_under_test.end is None |
| assert object_under_test.interval is None |
|
|
| def test_constructor_w_properties(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| assert object_under_test.start == 1 |
| assert object_under_test.end == 10 |
| assert object_under_test.interval == 2 |
|
|
| def test_constructor_w_resource(self): |
| object_under_test = self._make_one( |
| _properties={"start": -1234567890, "end": 1234567890, "interval": 1000000} |
| ) |
| assert object_under_test.start == -1234567890 |
| assert object_under_test.end == 1234567890 |
| assert object_under_test.interval == 1000000 |
|
|
| def test___eq___start_mismatch(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| other = self._make_one(start=2, end=10, interval=2) |
| self.assertNotEqual(object_under_test, other) |
|
|
| def test___eq___end__mismatch(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| other = self._make_one(start=1, end=11, interval=2) |
| self.assertNotEqual(object_under_test, other) |
|
|
| def test___eq___interval__mismatch(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| other = self._make_one(start=1, end=11, interval=3) |
| self.assertNotEqual(object_under_test, other) |
|
|
| def test___eq___hit(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| other = self._make_one(start=1, end=10, interval=2) |
| self.assertEqual(object_under_test, other) |
|
|
| def test__eq___type_mismatch(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| self.assertNotEqual(object_under_test, object()) |
| self.assertEqual(object_under_test, mock.ANY) |
|
|
| def test_unhashable_object(self): |
| object_under_test1 = self._make_one(start=1, end=10, interval=2) |
|
|
| with self.assertRaisesRegex(TypeError, r".*unhashable type.*"): |
| hash(object_under_test1) |
|
|
| def test_repr(self): |
| object_under_test = self._make_one(start=1, end=10, interval=2) |
| assert repr(object_under_test) == "PartitionRange(end=10, interval=2, start=1)" |
|
|
|
|
| class TestRangePartitioning(unittest.TestCase): |
| def _get_target_class(self): |
| from google.cloud.bigquery.table import RangePartitioning |
|
|
| return RangePartitioning |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def test_constructor_defaults(self): |
| object_under_test = self._make_one() |
| assert object_under_test.field is None |
| assert object_under_test.range_.start is None |
| assert object_under_test.range_.end is None |
| assert object_under_test.range_.interval is None |
|
|
| def test_constructor_w_properties(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| assert object_under_test.field == "integer_col" |
| assert object_under_test.range_.start == 1 |
| assert object_under_test.range_.end == 10 |
| assert object_under_test.range_.interval == 2 |
|
|
| def test_constructor_w_resource(self): |
| object_under_test = self._make_one( |
| _properties={ |
| "field": "some_column", |
| "range": {"start": -1234567890, "end": 1234567890, "interval": 1000000}, |
| } |
| ) |
| assert object_under_test.field == "some_column" |
| assert object_under_test.range_.start == -1234567890 |
| assert object_under_test.range_.end == 1234567890 |
| assert object_under_test.range_.interval == 1000000 |
|
|
| def test_range_w_wrong_type(self): |
| object_under_test = self._make_one() |
| with pytest.raises(ValueError, match="PartitionRange"): |
| object_under_test.range_ = object() |
|
|
| def test___eq___field_mismatch(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| other = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="float_col" |
| ) |
| self.assertNotEqual(object_under_test, other) |
|
|
| def test___eq___range__mismatch(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| other = self._make_one( |
| range_=PartitionRange(start=2, end=20, interval=2), field="float_col" |
| ) |
| self.assertNotEqual(object_under_test, other) |
|
|
| def test___eq___hit(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| other = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| self.assertEqual(object_under_test, other) |
|
|
| def test__eq___type_mismatch(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| self.assertNotEqual(object_under_test, object()) |
| self.assertEqual(object_under_test, mock.ANY) |
|
|
| def test_unhashable_object(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test1 = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| with self.assertRaisesRegex(TypeError, r".*unhashable type.*"): |
| hash(object_under_test1) |
|
|
| def test_repr(self): |
| from google.cloud.bigquery.table import PartitionRange |
|
|
| object_under_test = self._make_one( |
| range_=PartitionRange(start=1, end=10, interval=2), field="integer_col" |
| ) |
| assert ( |
| repr(object_under_test) |
| == "RangePartitioning(field='integer_col', range_=PartitionRange(end=10, interval=2, start=1))" |
| ) |
|
|
|
|
| class TestTimePartitioning(unittest.TestCase): |
| def _get_target_class(self): |
| from google.cloud.bigquery.table import TimePartitioning |
|
|
| return TimePartitioning |
|
|
| def _make_one(self, *args, **kw): |
| return self._get_target_class()(*args, **kw) |
|
|
| def test_constructor_defaults(self): |
| time_partitioning = self._make_one() |
| self.assertEqual(time_partitioning.type_, "DAY") |
| self.assertIsNone(time_partitioning.field) |
| self.assertIsNone(time_partitioning.expiration_ms) |
|
|
| def test_constructor_explicit(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| time_partitioning = self._make_one( |
| type_=TimePartitioningType.DAY, field="name", expiration_ms=10000 |
| ) |
|
|
| self.assertEqual(time_partitioning.type_, "DAY") |
| self.assertEqual(time_partitioning.field, "name") |
| self.assertEqual(time_partitioning.expiration_ms, 10000) |
|
|
| def test_require_partition_filter_warns_deprecation(self): |
| object_under_test = self._make_one() |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| assert object_under_test.require_partition_filter is None |
| object_under_test.require_partition_filter = True |
| assert object_under_test.require_partition_filter |
|
|
| assert len(warned) == 3 |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| def test_from_api_repr_empty(self): |
| klass = self._get_target_class() |
|
|
| |
| |
| |
| api_repr = {} |
| time_partitioning = klass.from_api_repr(api_repr) |
|
|
| self.assertIsNone(time_partitioning.type_) |
| self.assertIsNone(time_partitioning.field) |
| self.assertIsNone(time_partitioning.expiration_ms) |
|
|
| def test_from_api_repr_minimal(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| klass = self._get_target_class() |
| api_repr = {"type": "DAY"} |
| time_partitioning = klass.from_api_repr(api_repr) |
|
|
| self.assertEqual(time_partitioning.type_, TimePartitioningType.DAY) |
| self.assertIsNone(time_partitioning.field) |
| self.assertIsNone(time_partitioning.expiration_ms) |
|
|
| def test_from_api_repr_doesnt_override_type(self): |
| klass = self._get_target_class() |
| api_repr = {"type": "HOUR"} |
| time_partitioning = klass.from_api_repr(api_repr) |
| self.assertEqual(time_partitioning.type_, "HOUR") |
|
|
| def test_from_api_repr_explicit(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| klass = self._get_target_class() |
| api_repr = { |
| "type": "DAY", |
| "field": "name", |
| "expirationMs": "10000", |
| "requirePartitionFilter": True, |
| } |
| time_partitioning = klass.from_api_repr(api_repr) |
|
|
| self.assertEqual(time_partitioning.type_, TimePartitioningType.DAY) |
| self.assertEqual(time_partitioning.field, "name") |
| self.assertEqual(time_partitioning.expiration_ms, 10000) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| self.assertTrue(time_partitioning.require_partition_filter) |
|
|
| self.assertIs(warned[0].category, PendingDeprecationWarning) |
|
|
| def test_to_api_repr_defaults(self): |
| time_partitioning = self._make_one() |
| expected = {"type": "DAY"} |
| self.assertEqual(time_partitioning.to_api_repr(), expected) |
|
|
| def test_to_api_repr_explicit(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| time_partitioning = self._make_one( |
| type_=TimePartitioningType.DAY, field="name", expiration_ms=10000 |
| ) |
|
|
| with warnings.catch_warnings(record=True) as warned: |
| time_partitioning.require_partition_filter = True |
|
|
| self.assertIs(warned[0].category, PendingDeprecationWarning) |
|
|
| expected = { |
| "type": "DAY", |
| "field": "name", |
| "expirationMs": "10000", |
| "requirePartitionFilter": True, |
| } |
| self.assertEqual(time_partitioning.to_api_repr(), expected) |
|
|
| def test___eq___wrong_type(self): |
| time_partitioning = self._make_one() |
| other = object() |
| self.assertNotEqual(time_partitioning, other) |
| self.assertEqual(time_partitioning, mock.ANY) |
|
|
| def test___eq___type__mismatch(self): |
| time_partitioning = self._make_one() |
| other = self._make_one(type_="HOUR") |
| self.assertNotEqual(time_partitioning, other) |
|
|
| def test___eq___field_mismatch(self): |
| time_partitioning = self._make_one(field="foo") |
| other = self._make_one(field="bar") |
| self.assertNotEqual(time_partitioning, other) |
|
|
| def test___eq___expiration_ms_mismatch(self): |
| time_partitioning = self._make_one(field="foo", expiration_ms=100000) |
| other = self._make_one(field="foo", expiration_ms=200000) |
| self.assertNotEqual(time_partitioning, other) |
|
|
| def test___eq___require_partition_filter_mismatch(self): |
| time_partitioning = self._make_one(field="foo", expiration_ms=100000) |
| other = self._make_one(field="foo", expiration_ms=100000) |
| with warnings.catch_warnings(record=True) as warned: |
| time_partitioning.require_partition_filter = True |
| other.require_partition_filter = False |
|
|
| assert len(warned) == 2 |
| for warning in warned: |
| self.assertIs(warning.category, PendingDeprecationWarning) |
|
|
| self.assertNotEqual(time_partitioning, other) |
|
|
| def test___eq___hit(self): |
| time_partitioning = self._make_one(field="foo", expiration_ms=100000) |
| other = self._make_one(field="foo", expiration_ms=100000) |
| self.assertEqual(time_partitioning, other) |
|
|
| def test___ne___wrong_type(self): |
| time_partitioning = self._make_one() |
| other = object() |
| self.assertNotEqual(time_partitioning, other) |
| self.assertEqual(time_partitioning, mock.ANY) |
|
|
| def test___ne___same_value(self): |
| time_partitioning1 = self._make_one() |
| time_partitioning2 = self._make_one() |
| |
| comparison_val = time_partitioning1 != time_partitioning2 |
| self.assertFalse(comparison_val) |
|
|
| def test___ne___different_values(self): |
| time_partitioning1 = self._make_one() |
| time_partitioning2 = self._make_one(type_="HOUR") |
| self.assertNotEqual(time_partitioning1, time_partitioning2) |
|
|
| def test___hash__set_equality(self): |
| time_partitioning1 = self._make_one(field="foo") |
| time_partitioning2 = self._make_one(field="foo") |
| set_one = {time_partitioning1, time_partitioning2} |
| set_two = {time_partitioning1, time_partitioning2} |
| self.assertEqual(set_one, set_two) |
|
|
| def test___hash__not_equals(self): |
| time_partitioning1 = self._make_one(field="foo") |
| time_partitioning2 = self._make_one(field="bar") |
| set_one = {time_partitioning1} |
| set_two = {time_partitioning2} |
| self.assertNotEqual(set_one, set_two) |
|
|
| def test___repr___minimal(self): |
| time_partitioning = self._make_one() |
| expected = "TimePartitioning(type_='DAY')" |
| self.assertEqual(repr(time_partitioning), expected) |
|
|
| def test___repr___explicit(self): |
| from google.cloud.bigquery.table import TimePartitioningType |
|
|
| time_partitioning = self._make_one( |
| type_=TimePartitioningType.DAY, field="name", expiration_ms=10000 |
| ) |
| expected = "TimePartitioning(expiration_ms=10000,field='name',type_='DAY')" |
| self.assertEqual(repr(time_partitioning), expected) |
|
|
| def test_set_expiration_w_none(self): |
| time_partitioning = self._make_one() |
| time_partitioning.expiration_ms = None |
| assert time_partitioning._properties["expirationMs"] is None |
|
|
|
|
| class TestPrimaryKey(unittest.TestCase): |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import PrimaryKey |
|
|
| return PrimaryKey |
|
|
| @classmethod |
| def _make_one(cls, *args, **kwargs): |
| return cls._get_target_class()(*args, **kwargs) |
|
|
| def test_constructor_explicit(self): |
| columns = ["id", "product_id"] |
| primary_key = self._make_one(columns) |
|
|
| self.assertEqual(primary_key.columns, columns) |
|
|
| def test__eq__columns_mismatch(self): |
| primary_key = self._make_one(columns=["id", "product_id"]) |
| other_primary_key = self._make_one(columns=["id"]) |
|
|
| self.assertNotEqual(primary_key, other_primary_key) |
|
|
| def test__eq__other_type(self): |
| primary_key = self._make_one(columns=["id", "product_id"]) |
| with self.assertRaises(TypeError): |
| primary_key == "This is not a Primary Key" |
|
|
|
|
| class TestColumnReference(unittest.TestCase): |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import ColumnReference |
|
|
| return ColumnReference |
|
|
| @classmethod |
| def _make_one(cls, *args, **kwargs): |
| return cls._get_target_class()(*args, **kwargs) |
|
|
| def test_constructor_explicit(self): |
| referencing_column = "product_id" |
| referenced_column = "id" |
| column_reference = self._make_one(referencing_column, referenced_column) |
|
|
| self.assertEqual(column_reference.referencing_column, referencing_column) |
| self.assertEqual(column_reference.referenced_column, referenced_column) |
|
|
| def test__eq__referencing_column_mismatch(self): |
| column_reference = self._make_one( |
| referencing_column="product_id", |
| referenced_column="id", |
| ) |
| other_column_reference = self._make_one( |
| referencing_column="item_id", |
| referenced_column="id", |
| ) |
|
|
| self.assertNotEqual(column_reference, other_column_reference) |
|
|
| def test__eq__referenced_column_mismatch(self): |
| column_reference = self._make_one( |
| referencing_column="product_id", |
| referenced_column="id", |
| ) |
| other_column_reference = self._make_one( |
| referencing_column="product_id", |
| referenced_column="id_1", |
| ) |
|
|
| self.assertNotEqual(column_reference, other_column_reference) |
|
|
| def test__eq__other_type(self): |
| column_reference = self._make_one( |
| referencing_column="product_id", |
| referenced_column="id", |
| ) |
| with self.assertRaises(TypeError): |
| column_reference == "This is not a Column Reference" |
|
|
|
|
| class TestForeignKey(unittest.TestCase): |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import ForeignKey |
|
|
| return ForeignKey |
|
|
| @classmethod |
| def _make_one(cls, *args, **kwargs): |
| return cls._get_target_class()(*args, **kwargs) |
|
|
| def test_constructor_explicit(self): |
| name = "my_fk" |
| referenced_table = TableReference.from_string("my-project.mydataset.mytable") |
| column_references = [] |
| foreign_key = self._make_one(name, referenced_table, column_references) |
|
|
| self.assertEqual(foreign_key.name, name) |
| self.assertEqual(foreign_key.referenced_table, referenced_table) |
| self.assertEqual(foreign_key.column_references, column_references) |
|
|
| def test__eq__name_mismatch(self): |
| referenced_table = TableReference.from_string("my-project.mydataset.mytable") |
| column_references = [] |
| foreign_key = self._make_one( |
| name="my_fk", |
| referenced_table=referenced_table, |
| column_references=column_references, |
| ) |
| other_foreign_key = self._make_one( |
| name="my_other_fk", |
| referenced_table=referenced_table, |
| column_references=column_references, |
| ) |
|
|
| self.assertNotEqual(foreign_key, other_foreign_key) |
|
|
| def test__eq__referenced_table_mismatch(self): |
| name = "my_fk" |
| column_references = [] |
| foreign_key = self._make_one( |
| name=name, |
| referenced_table=TableReference.from_string("my-project.mydataset.mytable"), |
| column_references=column_references, |
| ) |
| other_foreign_key = self._make_one( |
| name=name, |
| referenced_table=TableReference.from_string( |
| "my-project.mydataset.my-other-table" |
| ), |
| column_references=column_references, |
| ) |
|
|
| self.assertNotEqual(foreign_key, other_foreign_key) |
|
|
| def test__eq__column_references_mismatch(self): |
| from google.cloud.bigquery.table import ColumnReference |
|
|
| name = "my_fk" |
| referenced_table = TableReference.from_string("my-project.mydataset.mytable") |
| foreign_key = self._make_one( |
| name=name, |
| referenced_table=referenced_table, |
| column_references=[], |
| ) |
| other_foreign_key = self._make_one( |
| name=name, |
| referenced_table=referenced_table, |
| column_references=[ |
| ColumnReference( |
| referencing_column="product_id", referenced_column="id" |
| ), |
| ], |
| ) |
|
|
| self.assertNotEqual(foreign_key, other_foreign_key) |
|
|
| def test__eq__other_type(self): |
| foreign_key = self._make_one( |
| name="my_fk", |
| referenced_table=TableReference.from_string("my-project.mydataset.mytable"), |
| column_references=[], |
| ) |
| with self.assertRaises(TypeError): |
| foreign_key == "This is not a Foreign Key" |
|
|
|
|
| class TestTableConstraint(unittest.TestCase): |
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.table import TableConstraints |
|
|
| return TableConstraints |
|
|
| @classmethod |
| def _make_one(cls, *args, **kwargs): |
| return cls._get_target_class()(*args, **kwargs) |
|
|
| def test_constructor_defaults(self): |
| instance = self._make_one(primary_key=None, foreign_keys=None) |
| self.assertIsNone(instance.primary_key) |
| self.assertIsNone(instance.foreign_keys) |
|
|
| def test_from_api_repr_full_resource(self): |
| from google.cloud.bigquery.table import ( |
| ColumnReference, |
| ForeignKey, |
| TableReference, |
| ) |
|
|
| resource = { |
| "primaryKey": { |
| "columns": ["id", "product_id"], |
| }, |
| "foreignKeys": [ |
| { |
| "name": "my_fk_name", |
| "referencedTable": { |
| "projectId": "my-project", |
| "datasetId": "your-dataset", |
| "tableId": "products", |
| }, |
| "columnReferences": [ |
| {"referencingColumn": "product_id", "referencedColumn": "id"}, |
| ], |
| } |
| ], |
| } |
| instance = self._get_target_class().from_api_repr(resource) |
|
|
| self.assertIsNotNone(instance.primary_key) |
| self.assertEqual(instance.primary_key.columns, ["id", "product_id"]) |
| self.assertEqual( |
| instance.foreign_keys, |
| [ |
| ForeignKey( |
| name="my_fk_name", |
| referenced_table=TableReference.from_string( |
| "my-project.your-dataset.products" |
| ), |
| column_references=[ |
| ColumnReference( |
| referencing_column="product_id", referenced_column="id" |
| ), |
| ], |
| ), |
| ], |
| ) |
|
|
| def test_from_api_repr_only_primary_key_resource(self): |
| resource = { |
| "primaryKey": { |
| "columns": ["id"], |
| }, |
| } |
| instance = self._get_target_class().from_api_repr(resource) |
|
|
| self.assertIsNotNone(instance.primary_key) |
| self.assertEqual(instance.primary_key.columns, ["id"]) |
| self.assertIsNone(instance.foreign_keys) |
|
|
| def test_from_api_repr_only_foreign_keys_resource(self): |
| resource = { |
| "foreignKeys": [ |
| { |
| "name": "my_fk_name", |
| "referencedTable": { |
| "projectId": "my-project", |
| "datasetId": "your-dataset", |
| "tableId": "products", |
| }, |
| "columnReferences": [ |
| {"referencingColumn": "product_id", "referencedColumn": "id"}, |
| ], |
| } |
| ] |
| } |
| instance = self._get_target_class().from_api_repr(resource) |
|
|
| self.assertIsNone(instance.primary_key) |
| self.assertIsNotNone(instance.foreign_keys) |
|
|
|
|
| @pytest.mark.parametrize( |
| "table_path", |
| ( |
| "my-project.my_dataset.my_table", |
| "my-project.my_dataset.my_table$20181225", |
| "my-project.my_dataset.my_table@1234567890", |
| "my-project.my_dataset.my_table$20181225@1234567890", |
| ), |
| ) |
| def test_table_reference_to_bqstorage_v1_stable(table_path): |
| pytest.importorskip("google.cloud.bigquery_storage") |
| from google.cloud.bigquery import table as mut |
|
|
| expected = "projects/my-project/datasets/my_dataset/tables/my_table" |
|
|
| for klass in (mut.TableReference, mut.Table, mut.TableListItem): |
| got = klass.from_string(table_path).to_bqstorage() |
| assert got == expected |
|
|