| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import concurrent |
| import concurrent.futures |
| import copy |
| import http |
| import textwrap |
| import types |
| from unittest import mock |
|
|
| import freezegun |
| from google.api_core import exceptions |
| import google.api_core.retry |
| import requests |
|
|
| from google.cloud.bigquery.client import _LIST_ROWS_FROM_QUERY_RESULTS_FIELDS |
| import google.cloud.bigquery._job_helpers |
| import google.cloud.bigquery.query |
| import google.cloud.bigquery.retry |
| from google.cloud.bigquery.retry import DEFAULT_GET_JOB_TIMEOUT |
| from google.cloud.bigquery.table import _EmptyRowIterator |
|
|
| from ..helpers import make_connection |
|
|
| from .helpers import _Base |
| from .helpers import _make_client |
|
|
|
|
| class TestQueryJob(_Base): |
| JOB_TYPE = "query" |
| QUERY = "select count(*) from persons" |
| DESTINATION_TABLE = "destination_table" |
|
|
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.job import QueryJob |
|
|
| return QueryJob |
|
|
| def _make_resource(self, started=False, ended=False, location="US"): |
| resource = super(TestQueryJob, self)._make_resource( |
| started, ended, location=location |
| ) |
| config = resource["configuration"]["query"] |
| config["query"] = self.QUERY |
| return resource |
|
|
| def _verifyBooleanResourceProperties(self, job, config): |
| if "allowLargeResults" in config: |
| self.assertEqual(job.allow_large_results, config["allowLargeResults"]) |
| else: |
| self.assertIsNone(job.allow_large_results) |
| if "flattenResults" in config: |
| self.assertEqual(job.flatten_results, config["flattenResults"]) |
| else: |
| self.assertIsNone(job.flatten_results) |
| if "useQueryCache" in config: |
| self.assertEqual(job.use_query_cache, config["useQueryCache"]) |
| else: |
| self.assertIsNone(job.use_query_cache) |
| if "useLegacySql" in config: |
| self.assertEqual(job.use_legacy_sql, config["useLegacySql"]) |
| else: |
| self.assertIsNone(job.use_legacy_sql) |
|
|
| def _verifyIntegerResourceProperties(self, job, config): |
| if "maximumBillingTier" in config: |
| self.assertEqual(job.maximum_billing_tier, config["maximumBillingTier"]) |
| else: |
| self.assertIsNone(job.maximum_billing_tier) |
| if "maximumBytesBilled" in config: |
| self.assertEqual( |
| str(job.maximum_bytes_billed), config["maximumBytesBilled"] |
| ) |
| self.assertIsInstance(job.maximum_bytes_billed, int) |
| else: |
| self.assertIsNone(job.maximum_bytes_billed) |
|
|
| def _verify_udf_resources(self, job, config): |
| udf_resources = config.get("userDefinedFunctionResources", ()) |
| self.assertEqual(len(job.udf_resources), len(udf_resources)) |
| for found, expected in zip(job.udf_resources, udf_resources): |
| if "resourceUri" in expected: |
| self.assertEqual(found.udf_type, "resourceUri") |
| self.assertEqual(found.value, expected["resourceUri"]) |
| else: |
| self.assertEqual(found.udf_type, "inlineCode") |
| self.assertEqual(found.value, expected["inlineCode"]) |
|
|
| def _verifyQueryParameters(self, job, config): |
| query_parameters = config.get("queryParameters", ()) |
| self.assertEqual(len(job.query_parameters), len(query_parameters)) |
| for found, expected in zip(job.query_parameters, query_parameters): |
| self.assertEqual(found.to_api_repr(), expected) |
|
|
| def _verify_table_definitions(self, job, config): |
| table_defs = config.get("tableDefinitions") |
| if job.table_definitions is None: |
| self.assertIsNone(table_defs) |
| else: |
| self.assertEqual(len(job.table_definitions), len(table_defs)) |
| for found_key, found_ec in job.table_definitions.items(): |
| expected_ec = table_defs.get(found_key) |
| self.assertIsNotNone(expected_ec) |
| self.assertEqual(found_ec.to_api_repr(), expected_ec) |
|
|
| def _verify_dml_stats_resource_properties(self, job, resource): |
| query_stats = resource.get("statistics", {}).get("query", {}) |
|
|
| if "dmlStats" in query_stats: |
| resource_dml_stats = query_stats["dmlStats"] |
| job_dml_stats = job.dml_stats |
| assert str(job_dml_stats.inserted_row_count) == resource_dml_stats.get( |
| "insertedRowCount", "0" |
| ) |
| assert str(job_dml_stats.updated_row_count) == resource_dml_stats.get( |
| "updatedRowCount", "0" |
| ) |
| assert str(job_dml_stats.deleted_row_count) == resource_dml_stats.get( |
| "deletedRowCount", "0" |
| ) |
| else: |
| assert job.dml_stats is None |
|
|
| def _verify_transaction_info_resource_properties(self, job, resource): |
| resource_stats = resource.get("statistics", {}) |
|
|
| if "transactionInfo" in resource_stats: |
| resource_transaction_info = resource_stats["transactionInfo"] |
| job_transaction_info = job.transaction_info |
| assert job_transaction_info.transaction_id == resource_transaction_info.get( |
| "transactionId" |
| ) |
| else: |
| assert job.transaction_info is None |
|
|
| def _verify_configuration_properties(self, job, configuration): |
| if "dryRun" in configuration: |
| self.assertEqual(job.dry_run, configuration["dryRun"]) |
| else: |
| self.assertIsNone(job.dry_run) |
|
|
| def _verifyResourceProperties(self, job, resource): |
| self._verifyReadonlyResourceProperties(job, resource) |
| self._verify_dml_stats_resource_properties(job, resource) |
| self._verify_transaction_info_resource_properties(job, resource) |
|
|
| configuration = resource.get("configuration", {}) |
| self._verify_configuration_properties(job, configuration) |
|
|
| query_config = resource.get("configuration", {}).get("query") |
| self._verifyBooleanResourceProperties(job, query_config) |
| self._verifyIntegerResourceProperties(job, query_config) |
| self._verify_udf_resources(job, query_config) |
| self._verifyQueryParameters(job, query_config) |
| self._verify_table_definitions(job, query_config) |
|
|
| self.assertEqual(job.query, query_config["query"]) |
|
|
| if "createDisposition" in query_config: |
| self.assertEqual(job.create_disposition, query_config["createDisposition"]) |
| else: |
| self.assertIsNone(job.create_disposition) |
|
|
| if "defaultDataset" in query_config: |
| ds_ref = job.default_dataset |
| ds_ref = {"projectId": ds_ref.project, "datasetId": ds_ref.dataset_id} |
| self.assertEqual(ds_ref, query_config["defaultDataset"]) |
| else: |
| self.assertIsNone(job.default_dataset) |
|
|
| if "destinationTable" in query_config: |
| table = job.destination |
| tb_ref = { |
| "projectId": table.project, |
| "datasetId": table.dataset_id, |
| "tableId": table.table_id, |
| } |
| self.assertEqual(tb_ref, query_config["destinationTable"]) |
| else: |
| self.assertIsNone(job.destination) |
|
|
| if "priority" in query_config: |
| self.assertEqual(job.priority, query_config["priority"]) |
| else: |
| self.assertIsNone(job.priority) |
|
|
| if "writeDisposition" in query_config: |
| self.assertEqual(job.write_disposition, query_config["writeDisposition"]) |
| else: |
| self.assertIsNone(job.write_disposition) |
|
|
| if "destinationEncryptionConfiguration" in query_config: |
| self.assertIsNotNone(job.destination_encryption_configuration) |
| self.assertEqual( |
| job.destination_encryption_configuration.kms_key_name, |
| query_config["destinationEncryptionConfiguration"]["kmsKeyName"], |
| ) |
| else: |
| self.assertIsNone(job.destination_encryption_configuration) |
|
|
| if "schemaUpdateOptions" in query_config: |
| self.assertEqual( |
| job.schema_update_options, query_config["schemaUpdateOptions"] |
| ) |
| else: |
| self.assertIsNone(job.schema_update_options) |
|
|
| def test_ctor_defaults(self): |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertEqual(job.query, self.QUERY) |
| self.assertIs(job._client, client) |
| self.assertEqual(job.job_type, self.JOB_TYPE) |
| self.assertEqual(job.path, "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID)) |
|
|
| self._verifyInitialReadonlyProperties(job) |
|
|
| self.assertFalse(job.use_legacy_sql) |
|
|
| |
| self.assertIsNone(job.allow_large_results) |
| self.assertIsNone(job.create_disposition) |
| self.assertIsNone(job.default_dataset) |
| self.assertIsNone(job.destination) |
| self.assertIsNone(job.dml_stats) |
| self.assertIsNone(job.flatten_results) |
| self.assertIsNone(job.priority) |
| self.assertIsNone(job.use_query_cache) |
| self.assertIsNone(job.dry_run) |
| self.assertIsNone(job.write_disposition) |
| self.assertIsNone(job.maximum_billing_tier) |
| self.assertIsNone(job.maximum_bytes_billed) |
| self.assertIsNone(job.table_definitions) |
| self.assertIsNone(job.destination_encryption_configuration) |
| self.assertIsNone(job.range_partitioning) |
| self.assertIsNone(job.time_partitioning) |
| self.assertIsNone(job.clustering_fields) |
| self.assertIsNone(job.schema_update_options) |
|
|
| def test_ctor_w_udf_resources(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.query import UDFResource |
|
|
| RESOURCE_URI = "gs://some-bucket/js/lib.js" |
| udf_resources = [UDFResource("resourceUri", RESOURCE_URI)] |
| client = _make_client(project=self.PROJECT) |
| config = QueryJobConfig() |
| config.udf_resources = udf_resources |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=config) |
| self.assertEqual(job.udf_resources, udf_resources) |
|
|
| def test_ctor_w_query_parameters(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.query import ScalarQueryParameter |
|
|
| query_parameters = [ScalarQueryParameter("foo", "INT64", 123)] |
| client = _make_client(project=self.PROJECT) |
| config = QueryJobConfig(query_parameters=query_parameters) |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=config) |
| self.assertEqual(job.query_parameters, query_parameters) |
|
|
| def test_from_api_repr_bare(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": {"query": {"query": self.QUERY}}, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
| self.assertEqual(len(job.connection_properties), 0) |
| self.assertIsNone(job.create_session) |
|
|
| def test_from_api_repr_with_encryption(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": { |
| "query": self.QUERY, |
| "destinationEncryptionConfiguration": { |
| "kmsKeyName": self.KMS_KEY_NAME |
| }, |
| } |
| }, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_from_api_repr_with_dml_stats(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": {"query": {"query": self.QUERY}}, |
| "statistics": { |
| "query": { |
| "dmlStats": {"insertedRowCount": "15", "updatedRowCount": "2"}, |
| }, |
| }, |
| } |
| klass = self._get_target_class() |
|
|
| job = klass.from_api_repr(RESOURCE, client=client) |
|
|
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_from_api_repr_with_transaction_info(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": {"query": {"query": self.QUERY}}, |
| "statistics": {"transactionInfo": {"transactionId": "1a2b-3c4d"}}, |
| } |
| klass = self._get_target_class() |
|
|
| job = klass.from_api_repr(RESOURCE, client=client) |
|
|
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_from_api_repr_w_properties(self): |
| from google.cloud.bigquery.job import CreateDisposition |
| from google.cloud.bigquery.job import SchemaUpdateOption |
| from google.cloud.bigquery.job import WriteDisposition |
|
|
| client = _make_client(project=self.PROJECT) |
| RESOURCE = self._make_resource() |
| query_config = RESOURCE["configuration"]["query"] |
| query_config["createDisposition"] = CreateDisposition.CREATE_IF_NEEDED |
| query_config["writeDisposition"] = WriteDisposition.WRITE_TRUNCATE |
| query_config["destinationTable"] = { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| } |
| query_config["schemaUpdateOptions"] = [SchemaUpdateOption.ALLOW_FIELD_ADDITION] |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_cancelled(self): |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| job._properties["status"] = { |
| "state": "DONE", |
| "errorResult": {"reason": "stopped"}, |
| } |
|
|
| self.assertTrue(job.cancelled()) |
|
|
| def test_query_plan(self): |
| from google.cloud._helpers import _RFC3339_MICROS |
| from google.cloud.bigquery.job import QueryPlanEntry |
| from google.cloud.bigquery.job import QueryPlanEntryStep |
|
|
| plan_entries = [ |
| { |
| "name": "NAME", |
| "id": "1234", |
| "inputStages": ["88", "101"], |
| "startMs": "1522540800000", |
| "endMs": "1522540804000", |
| "parallelInputs": "1000", |
| "completedParallelInputs": "5", |
| "waitMsAvg": "33", |
| "waitMsMax": "400", |
| "waitRatioAvg": 2.71828, |
| "waitRatioMax": 3.14159, |
| "readMsAvg": "45", |
| "readMsMax": "90", |
| "readRatioAvg": 1.41421, |
| "readRatioMax": 1.73205, |
| "computeMsAvg": "55", |
| "computeMsMax": "99", |
| "computeRatioAvg": 0.69315, |
| "computeRatioMax": 1.09861, |
| "writeMsAvg": "203", |
| "writeMsMax": "340", |
| "writeRatioAvg": 3.32193, |
| "writeRatioMax": 2.30258, |
| "recordsRead": "100", |
| "recordsWritten": "1", |
| "status": "STATUS", |
| "shuffleOutputBytes": "1024", |
| "shuffleOutputBytesSpilled": "1", |
| "steps": [{"kind": "KIND", "substeps": ["SUBSTEP1", "SUBSTEP2"]}], |
| } |
| ] |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertEqual(job.query_plan, []) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertEqual(job.query_plan, []) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertEqual(job.query_plan, []) |
|
|
| query_stats["queryPlan"] = plan_entries |
|
|
| self.assertEqual(len(job.query_plan), len(plan_entries)) |
| for found, expected in zip(job.query_plan, plan_entries): |
| self.assertIsInstance(found, QueryPlanEntry) |
| self.assertEqual(found.name, expected["name"]) |
| self.assertEqual(found.entry_id, expected["id"]) |
| self.assertEqual(len(found.input_stages), len(expected["inputStages"])) |
| for f_id in found.input_stages: |
| self.assertIn(f_id, [int(e) for e in expected["inputStages"]]) |
| self.assertEqual( |
| found.start.strftime(_RFC3339_MICROS), "2018-04-01T00:00:00.000000Z" |
| ) |
| self.assertEqual( |
| found.end.strftime(_RFC3339_MICROS), "2018-04-01T00:00:04.000000Z" |
| ) |
| self.assertEqual(found.parallel_inputs, int(expected["parallelInputs"])) |
| self.assertEqual( |
| found.completed_parallel_inputs, |
| int(expected["completedParallelInputs"]), |
| ) |
| self.assertEqual(found.wait_ms_avg, int(expected["waitMsAvg"])) |
| self.assertEqual(found.wait_ms_max, int(expected["waitMsMax"])) |
| self.assertEqual(found.wait_ratio_avg, expected["waitRatioAvg"]) |
| self.assertEqual(found.wait_ratio_max, expected["waitRatioMax"]) |
| self.assertEqual(found.read_ms_avg, int(expected["readMsAvg"])) |
| self.assertEqual(found.read_ms_max, int(expected["readMsMax"])) |
| self.assertEqual(found.read_ratio_avg, expected["readRatioAvg"]) |
| self.assertEqual(found.read_ratio_max, expected["readRatioMax"]) |
| self.assertEqual(found.compute_ms_avg, int(expected["computeMsAvg"])) |
| self.assertEqual(found.compute_ms_max, int(expected["computeMsMax"])) |
| self.assertEqual(found.compute_ratio_avg, expected["computeRatioAvg"]) |
| self.assertEqual(found.compute_ratio_max, expected["computeRatioMax"]) |
| self.assertEqual(found.write_ms_avg, int(expected["writeMsAvg"])) |
| self.assertEqual(found.write_ms_max, int(expected["writeMsMax"])) |
| self.assertEqual(found.write_ratio_avg, expected["writeRatioAvg"]) |
| self.assertEqual(found.write_ratio_max, expected["writeRatioMax"]) |
| self.assertEqual(found.records_read, int(expected["recordsRead"])) |
| self.assertEqual(found.records_written, int(expected["recordsWritten"])) |
| self.assertEqual(found.status, expected["status"]) |
| self.assertEqual( |
| found.shuffle_output_bytes, int(expected["shuffleOutputBytes"]) |
| ) |
| self.assertEqual( |
| found.shuffle_output_bytes_spilled, |
| int(expected["shuffleOutputBytesSpilled"]), |
| ) |
|
|
| self.assertEqual(len(found.steps), len(expected["steps"])) |
| for f_step, e_step in zip(found.steps, expected["steps"]): |
| self.assertIsInstance(f_step, QueryPlanEntryStep) |
| self.assertEqual(f_step.kind, e_step["kind"]) |
| self.assertEqual(f_step.substeps, e_step["substeps"]) |
|
|
| def test_total_bytes_processed(self): |
| total_bytes = 1234 |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.total_bytes_processed) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.total_bytes_processed) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.total_bytes_processed) |
|
|
| query_stats["totalBytesProcessed"] = str(total_bytes) |
| self.assertEqual(job.total_bytes_processed, total_bytes) |
|
|
| def test_total_bytes_billed(self): |
| total_bytes = 1234 |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.total_bytes_billed) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.total_bytes_billed) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.total_bytes_billed) |
|
|
| query_stats["totalBytesBilled"] = str(total_bytes) |
| self.assertEqual(job.total_bytes_billed, total_bytes) |
|
|
| def test_billing_tier(self): |
| billing_tier = 1 |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.billing_tier) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.billing_tier) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.billing_tier) |
|
|
| query_stats["billingTier"] = billing_tier |
| self.assertEqual(job.billing_tier, billing_tier) |
|
|
| def test_cache_hit(self): |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.cache_hit) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.cache_hit) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.cache_hit) |
|
|
| query_stats["cacheHit"] = True |
| self.assertTrue(job.cache_hit) |
|
|
| def test_ddl_operation_performed(self): |
| op = "SKIP" |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.ddl_operation_performed) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.ddl_operation_performed) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.ddl_operation_performed) |
|
|
| query_stats["ddlOperationPerformed"] = op |
| self.assertEqual(job.ddl_operation_performed, op) |
|
|
| def test_ddl_target_routine(self): |
| from google.cloud.bigquery.routine import RoutineReference |
|
|
| ref_routine = { |
| "projectId": self.PROJECT, |
| "datasetId": "ddl_ds", |
| "routineId": "targetroutine", |
| } |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.ddl_target_routine) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.ddl_target_routine) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.ddl_target_routine) |
|
|
| query_stats["ddlTargetRoutine"] = ref_routine |
| self.assertIsInstance(job.ddl_target_routine, RoutineReference) |
| self.assertEqual(job.ddl_target_routine.routine_id, "targetroutine") |
| self.assertEqual(job.ddl_target_routine.dataset_id, "ddl_ds") |
| self.assertEqual(job.ddl_target_routine.project, self.PROJECT) |
|
|
| def test_ddl_target_table(self): |
| from google.cloud.bigquery.table import TableReference |
|
|
| ref_table = { |
| "projectId": self.PROJECT, |
| "datasetId": "ddl_ds", |
| "tableId": "targettable", |
| } |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.ddl_target_table) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.ddl_target_table) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.ddl_target_table) |
|
|
| query_stats["ddlTargetTable"] = ref_table |
| self.assertIsInstance(job.ddl_target_table, TableReference) |
| self.assertEqual(job.ddl_target_table.table_id, "targettable") |
| self.assertEqual(job.ddl_target_table.dataset_id, "ddl_ds") |
| self.assertEqual(job.ddl_target_table.project, self.PROJECT) |
|
|
| def test_num_dml_affected_rows(self): |
| num_rows = 1234 |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.num_dml_affected_rows) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.num_dml_affected_rows) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.num_dml_affected_rows) |
|
|
| query_stats["numDmlAffectedRows"] = str(num_rows) |
| self.assertEqual(job.num_dml_affected_rows, num_rows) |
|
|
| def test_slot_millis(self): |
| millis = 1234 |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.slot_millis) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.slot_millis) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.slot_millis) |
|
|
| query_stats["totalSlotMs"] = millis |
| self.assertEqual(job.slot_millis, millis) |
|
|
| def test_statement_type(self): |
| statement_type = "SELECT" |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.statement_type) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.statement_type) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.statement_type) |
|
|
| query_stats["statementType"] = statement_type |
| self.assertEqual(job.statement_type, statement_type) |
|
|
| def test_referenced_tables(self): |
| from google.cloud.bigquery.table import TableReference |
|
|
| ref_tables_resource = [ |
| {"projectId": self.PROJECT, "datasetId": "dataset", "tableId": "local1"}, |
| {"projectId": self.PROJECT, "datasetId": "dataset", "tableId": "local2"}, |
| { |
| "projectId": "other-project-123", |
| "datasetId": "other-dataset", |
| "tableId": "other-table", |
| }, |
| ] |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertEqual(job.referenced_tables, []) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertEqual(job.referenced_tables, []) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertEqual(job.referenced_tables, []) |
|
|
| query_stats["referencedTables"] = ref_tables_resource |
|
|
| local1, local2, remote = job.referenced_tables |
|
|
| self.assertIsInstance(local1, TableReference) |
| self.assertEqual(local1.table_id, "local1") |
| self.assertEqual(local1.dataset_id, "dataset") |
| self.assertEqual(local1.project, self.PROJECT) |
|
|
| self.assertIsInstance(local2, TableReference) |
| self.assertEqual(local2.table_id, "local2") |
| self.assertEqual(local2.dataset_id, "dataset") |
| self.assertEqual(local2.project, self.PROJECT) |
|
|
| self.assertIsInstance(remote, TableReference) |
| self.assertEqual(remote.table_id, "other-table") |
| self.assertEqual(remote.dataset_id, "other-dataset") |
| self.assertEqual(remote.project, "other-project-123") |
|
|
| def test_timeline(self): |
| timeline_resource = [ |
| { |
| "elapsedMs": 1, |
| "activeUnits": 22, |
| "pendingUnits": 33, |
| "completedUnits": 44, |
| "totalSlotMs": 101, |
| } |
| ] |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertEqual(job.timeline, []) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertEqual(job.timeline, []) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertEqual(job.timeline, []) |
|
|
| query_stats["timeline"] = timeline_resource |
|
|
| self.assertEqual(len(job.timeline), len(timeline_resource)) |
| self.assertEqual(job.timeline[0].elapsed_ms, 1) |
| self.assertEqual(job.timeline[0].active_units, 22) |
| self.assertEqual(job.timeline[0].pending_units, 33) |
| self.assertEqual(job.timeline[0].completed_units, 44) |
| self.assertEqual(job.timeline[0].slot_millis, 101) |
|
|
| def test_undeclared_query_parameters(self): |
| from google.cloud.bigquery.query import ArrayQueryParameter |
| from google.cloud.bigquery.query import ScalarQueryParameter |
| from google.cloud.bigquery.query import StructQueryParameter |
|
|
| undeclared = [ |
| { |
| "name": "my_scalar", |
| "parameterType": {"type": "STRING"}, |
| "parameterValue": {"value": "value"}, |
| }, |
| { |
| "name": "my_array", |
| "parameterType": {"type": "ARRAY", "arrayType": {"type": "INT64"}}, |
| "parameterValue": { |
| "arrayValues": [{"value": "1066"}, {"value": "1745"}] |
| }, |
| }, |
| { |
| "name": "my_struct", |
| "parameterType": { |
| "type": "STRUCT", |
| "structTypes": [{"name": "count", "type": {"type": "INT64"}}], |
| }, |
| "parameterValue": {"structValues": {"count": {"value": "123"}}}, |
| }, |
| ] |
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertEqual(job.undeclared_query_parameters, []) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertEqual(job.undeclared_query_parameters, []) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertEqual(job.undeclared_query_parameters, []) |
|
|
| query_stats["undeclaredQueryParameters"] = undeclared |
|
|
| scalar, array, struct = job.undeclared_query_parameters |
|
|
| self.assertIsInstance(scalar, ScalarQueryParameter) |
| self.assertEqual(scalar.name, "my_scalar") |
| self.assertEqual(scalar.type_, "STRING") |
| self.assertEqual(scalar.value, "value") |
|
|
| self.assertIsInstance(array, ArrayQueryParameter) |
| self.assertEqual(array.name, "my_array") |
| self.assertEqual(array.array_type, "INT64") |
| self.assertEqual(array.values, [1066, 1745]) |
|
|
| self.assertIsInstance(struct, StructQueryParameter) |
| self.assertEqual(struct.name, "my_struct") |
| self.assertEqual(struct.struct_types, {"count": "INT64"}) |
| self.assertEqual(struct.struct_values, {"count": 123}) |
|
|
| def test_estimated_bytes_processed(self): |
| est_bytes = 123456 |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| self.assertIsNone(job.estimated_bytes_processed) |
|
|
| statistics = job._properties["statistics"] = {} |
| self.assertIsNone(job.estimated_bytes_processed) |
|
|
| query_stats = statistics["query"] = {} |
| self.assertIsNone(job.estimated_bytes_processed) |
|
|
| query_stats["estimatedBytesProcessed"] = str(est_bytes) |
| self.assertEqual(job.estimated_bytes_processed, est_bytes) |
|
|
| def test_bi_engine_stats(self): |
| from google.cloud.bigquery.job.query import BiEngineStats |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| assert job.bi_engine_stats is None |
|
|
| statistics = job._properties["statistics"] = {} |
| assert job.bi_engine_stats is None |
|
|
| query_stats = statistics["query"] = {} |
| assert job.bi_engine_stats is None |
|
|
| query_stats["biEngineStatistics"] = {"biEngineMode": "FULL"} |
| assert isinstance(job.bi_engine_stats, BiEngineStats) |
| assert job.bi_engine_stats.mode == "FULL" |
|
|
| def test_dml_stats(self): |
| from google.cloud.bigquery.job.query import DmlStats |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| assert job.dml_stats is None |
|
|
| statistics = job._properties["statistics"] = {} |
| assert job.dml_stats is None |
|
|
| query_stats = statistics["query"] = {} |
| assert job.dml_stats is None |
|
|
| query_stats["dmlStats"] = {"insertedRowCount": "35"} |
| assert isinstance(job.dml_stats, DmlStats) |
| assert job.dml_stats.inserted_row_count == 35 |
|
|
| def test_search_stats(self): |
| from google.cloud.bigquery.job.query import SearchStats |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| assert job.search_stats is None |
|
|
| statistics = job._properties["statistics"] = {} |
| assert job.search_stats is None |
|
|
| query_stats = statistics["query"] = {} |
| assert job.search_stats is None |
|
|
| query_stats["searchStatistics"] = { |
| "indexUsageMode": "INDEX_USAGE_MODE_UNSPECIFIED", |
| "indexUnusedReasons": [], |
| } |
| |
| |
| assert isinstance(job.search_stats, SearchStats) |
| assert job.search_stats.mode == "INDEX_USAGE_MODE_UNSPECIFIED" |
|
|
| def test_reload_query_results_uses_transport_timeout(self): |
| conn = make_connection({}) |
| client = _make_client(self.PROJECT, connection=conn) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| job._transport_timeout = 123 |
|
|
| job._reload_query_results() |
|
|
| query_results_path = f"/projects/{self.PROJECT}/queries/{self.JOB_ID}" |
| conn.api_request.assert_called_once_with( |
| method="GET", |
| path=query_results_path, |
| query_params={"maxResults": 0}, |
| timeout=123, |
| ) |
|
|
| def test_result_reloads_job_state_until_done(self): |
| """Verify that result() doesn't return until state == 'DONE'. |
| |
| This test verifies correctness for a possible sequence of API responses |
| that might cause internal customer issue b/332850329. |
| """ |
| from google.cloud.bigquery.table import RowIterator |
|
|
| query_resource = { |
| "jobComplete": False, |
| "jobReference": { |
| "projectId": self.PROJECT, |
| "jobId": self.JOB_ID, |
| "location": "EU", |
| }, |
| } |
| query_resource_done = { |
| "jobComplete": True, |
| "jobReference": { |
| "projectId": self.PROJECT, |
| "jobId": self.JOB_ID, |
| "location": "EU", |
| }, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "totalRows": "2", |
| "queryId": "abc-def", |
| } |
| job_resource = self._make_resource(started=True, location="EU") |
| job_resource_done = self._make_resource(started=True, ended=True, location="EU") |
| job_resource_done["configuration"]["query"]["destinationTable"] = { |
| "projectId": "dest-project", |
| "datasetId": "dest_dataset", |
| "tableId": "dest_table", |
| } |
| query_page_resource = { |
| |
| |
| "totalRows": "1", |
| "pageToken": None, |
| "rows": [{"f": [{"v": "abc"}]}], |
| } |
| conn = make_connection( |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| job_resource, |
| query_resource, |
| |
| |
| job_resource, |
| query_resource_done, |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| job_resource, |
| |
| |
| |
| |
| |
| job_resource, |
| job_resource_done, |
| |
| |
| |
| |
| query_page_resource, |
| ) |
| client = _make_client(self.PROJECT, connection=conn) |
| job = self._get_target_class().from_api_repr(job_resource, client) |
|
|
| result = job.result() |
|
|
| self.assertIsInstance(result, RowIterator) |
| self.assertEqual(result.total_rows, 2) |
| rows = list(result) |
| self.assertEqual(len(rows), 1) |
| self.assertEqual(rows[0].col1, "abc") |
| self.assertEqual(result.job_id, self.JOB_ID) |
| self.assertEqual(result.location, "EU") |
| self.assertEqual(result.project, self.PROJECT) |
| self.assertEqual(result.query_id, "abc-def") |
| |
| |
| self.assertEqual(result.total_rows, 1) |
|
|
| query_results_path = f"/projects/{self.PROJECT}/queries/{self.JOB_ID}" |
| query_results_call = mock.call( |
| method="GET", |
| path=query_results_path, |
| query_params={"maxResults": 0, "location": "EU"}, |
| timeout=None, |
| ) |
| reload_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}", |
| query_params={"projection": "full", "location": "EU"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| query_page_call = mock.call( |
| method="GET", |
| path=query_results_path, |
| query_params={ |
| "fields": _LIST_ROWS_FROM_QUERY_RESULTS_FIELDS, |
| "location": "EU", |
| "formatOptions.useInt64Timestamp": True, |
| }, |
| timeout=None, |
| ) |
| |
| |
| |
| |
| |
| |
| conn.api_request.assert_has_calls( |
| [ |
| |
| reload_call, |
| query_results_call, |
| |
| |
| reload_call, |
| query_results_call, |
| |
| |
| |
| |
| reload_call, |
| reload_call, |
| reload_call, |
| |
| |
| query_page_call, |
| ] |
| ) |
|
|
| def test_result_dry_run(self): |
| job_resource = self._make_resource(started=True, location="EU") |
| job_resource["configuration"]["dryRun"] = True |
| conn = make_connection() |
| client = _make_client(self.PROJECT, connection=conn) |
| job = self._get_target_class().from_api_repr(job_resource, client) |
|
|
| result = job.result() |
|
|
| calls = conn.api_request.mock_calls |
| self.assertIsInstance(result, _EmptyRowIterator) |
| self.assertEqual(calls, []) |
| self.assertEqual(result.location, "EU") |
| self.assertEqual(result.project, self.PROJECT) |
| |
| |
| self.assertIsNone(result.job_id) |
| self.assertIsNone(result.query_id) |
|
|
| |
| |
| def test_result_begin_job_if_not_exist(self): |
| begun_resource = self._make_resource() |
| query_running_resource = { |
| "jobComplete": True, |
| "jobReference": { |
| "projectId": self.PROJECT, |
| "jobId": self.JOB_ID, |
| "location": "US", |
| }, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "status": {"state": "RUNNING"}, |
| } |
| query_done_resource = { |
| "jobComplete": True, |
| "jobReference": { |
| "projectId": self.PROJECT, |
| "jobId": self.JOB_ID, |
| "location": "US", |
| }, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "status": {"state": "DONE"}, |
| } |
| done_resource = copy.deepcopy(begun_resource) |
| done_resource["status"] = {"state": "DONE"} |
| connection = make_connection( |
| begun_resource, |
| query_running_resource, |
| query_done_resource, |
| done_resource, |
| ) |
| client = _make_client(project=self.PROJECT, connection=connection) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| job._properties["jobReference"]["location"] = "US" |
|
|
| job.result() |
|
|
| create_job_call = mock.call( |
| method="POST", |
| path=f"/projects/{self.PROJECT}/jobs", |
| data={ |
| "jobReference": { |
| "jobId": self.JOB_ID, |
| "projectId": self.PROJECT, |
| "location": "US", |
| }, |
| "configuration": { |
| "query": {"useLegacySql": False, "query": self.QUERY}, |
| }, |
| }, |
| timeout=None, |
| ) |
| reload_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}", |
| query_params={"projection": "full", "location": "US"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| get_query_results_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/queries/{self.JOB_ID}", |
| query_params={ |
| "maxResults": 0, |
| "location": "US", |
| }, |
| timeout=None, |
| ) |
|
|
| connection.api_request.assert_has_calls( |
| [ |
| |
| |
| create_job_call, |
| reload_call, |
| get_query_results_call, |
| reload_call, |
| ] |
| ) |
|
|
| def test_result_with_done_job_calls_get_query_results(self): |
| query_resource_done = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "totalRows": "1", |
| } |
| job_resource = self._make_resource(started=True, ended=True, location="EU") |
| job_resource["configuration"]["query"]["destinationTable"] = { |
| "projectId": "dest-project", |
| "datasetId": "dest_dataset", |
| "tableId": "dest_table", |
| } |
| results_page_resource = { |
| "totalRows": "1", |
| "pageToken": None, |
| "rows": [{"f": [{"v": "abc"}]}], |
| } |
| conn = make_connection(query_resource_done, results_page_resource) |
| client = _make_client(self.PROJECT, connection=conn) |
| job = self._get_target_class().from_api_repr(job_resource, client) |
|
|
| result = job.result() |
|
|
| rows = list(result) |
| self.assertEqual(len(rows), 1) |
| self.assertEqual(rows[0].col1, "abc") |
|
|
| query_results_path = f"/projects/{self.PROJECT}/queries/{self.JOB_ID}" |
| query_results_call = mock.call( |
| method="GET", |
| path=query_results_path, |
| query_params={"maxResults": 0, "location": "EU"}, |
| timeout=None, |
| ) |
| query_results_page_call = mock.call( |
| method="GET", |
| path=query_results_path, |
| query_params={ |
| "fields": _LIST_ROWS_FROM_QUERY_RESULTS_FIELDS, |
| "location": "EU", |
| "formatOptions.useInt64Timestamp": True, |
| }, |
| timeout=None, |
| ) |
| conn.api_request.assert_has_calls([query_results_call, query_results_page_call]) |
| assert conn.api_request.call_count == 2 |
|
|
| def test_result_with_done_jobs_query_response_doesnt_call_get_query_results(self): |
| """With a done result from jobs.query, we don't need to call |
| jobs.getQueryResults to wait for the query to finish. |
| |
| jobs.get is still called because there is an assumption that after |
| QueryJob.result(), all job metadata is available locally. |
| """ |
| job_resource = self._make_resource(started=True, ended=True, location="EU") |
| conn = make_connection(job_resource) |
| client = _make_client(self.PROJECT, connection=conn) |
| query_resource_done = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "rows": [{"f": [{"v": "abc"}]}], |
| "totalRows": "1", |
| } |
| job = google.cloud.bigquery._job_helpers._to_query_job( |
| client, |
| "SELECT 'abc' AS col1", |
| request_config=None, |
| query_response=query_resource_done, |
| ) |
|
|
| |
| |
| assert job.state == "PENDING" |
|
|
| result = job.result() |
|
|
| rows = list(result) |
| self.assertEqual(len(rows), 1) |
| self.assertEqual(rows[0].col1, "abc") |
| job_path = f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}" |
| conn.api_request.assert_called_once_with( |
| method="GET", |
| path=job_path, |
| query_params={"projection": "full"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
|
|
| def test_result_with_none_timeout(self): |
| |
| |
| job_resource = self._make_resource(started=True, ended=True, location="EU") |
| conn = make_connection(job_resource) |
| client = _make_client(self.PROJECT, connection=conn) |
| query_resource_done = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "rows": [{"f": [{"v": "abc"}]}], |
| "totalRows": "1", |
| } |
| job = google.cloud.bigquery._job_helpers._to_query_job( |
| client, |
| "SELECT 'abc' AS col1", |
| request_config=None, |
| query_response=query_resource_done, |
| ) |
|
|
| job.result(timeout=None) |
|
|
| job_path = f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}" |
| conn.api_request.assert_called_once_with( |
| method="GET", |
| path=job_path, |
| query_params={"projection": "full"}, |
| timeout=None, |
| ) |
|
|
| def test_result_with_max_results(self): |
| from google.cloud.bigquery.table import RowIterator |
|
|
| query_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "totalRows": "10", |
| "pageToken": "first-page-token", |
| "rows": [ |
| {"f": [{"v": "abc"}]}, |
| {"f": [{"v": "def"}]}, |
| {"f": [{"v": "ghi"}]}, |
| {"f": [{"v": "jkl"}]}, |
| {"f": [{"v": "mno"}]}, |
| {"f": [{"v": "pqr"}]}, |
| |
| |
| ], |
| } |
| query_page_resource = { |
| "totalRows": "10", |
| "pageToken": None, |
| "rows": [ |
| {"f": [{"v": "stu"}]}, |
| {"f": [{"v": "vwx"}]}, |
| {"f": [{"v": "yz0"}]}, |
| ], |
| } |
| job_resource_running = self._make_resource( |
| started=True, ended=False, location="US" |
| ) |
| job_resource_done = self._make_resource(started=True, ended=True, location="US") |
| conn = make_connection(job_resource_done, query_resource, query_page_resource) |
| client = _make_client(self.PROJECT, connection=conn) |
| job = self._get_target_class().from_api_repr(job_resource_running, client) |
|
|
| max_results = 9 |
| result = job.result(max_results=max_results) |
|
|
| self.assertIsInstance(result, RowIterator) |
| self.assertEqual(result.total_rows, 10) |
|
|
| rows = list(result) |
|
|
| self.assertEqual(len(rows), 9) |
| jobs_get_path = f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}" |
| jobs_get_call = mock.call( |
| method="GET", |
| path=jobs_get_path, |
| query_params={"projection": "full", "location": "US"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| query_results_path = f"/projects/{self.PROJECT}/queries/{self.JOB_ID}" |
| query_page_waiting_call = mock.call( |
| method="GET", |
| path=query_results_path, |
| query_params={ |
| |
| |
| |
| |
| "maxResults": max_results, |
| "formatOptions.useInt64Timestamp": True, |
| "location": "US", |
| }, |
| timeout=None, |
| ) |
| query_page_2_call = mock.call( |
| timeout=None, |
| method="GET", |
| path=query_results_path, |
| query_params={ |
| "pageToken": "first-page-token", |
| "maxResults": 3, |
| "fields": _LIST_ROWS_FROM_QUERY_RESULTS_FIELDS, |
| "location": "US", |
| "formatOptions.useInt64Timestamp": True, |
| }, |
| ) |
| |
| |
| |
| |
| conn.api_request.assert_has_calls( |
| [jobs_get_call, query_page_waiting_call, query_page_2_call] |
| ) |
|
|
| def test_result_w_custom_retry(self): |
| from google.cloud.bigquery.table import RowIterator |
|
|
| query_resource = { |
| "jobComplete": False, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| } |
| query_resource_done = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "totalRows": "2", |
| } |
| job_resource = self._make_resource(started=True, location="asia-northeast1") |
| job_resource_done = self._make_resource( |
| started=True, ended=True, location="asia-northeast1" |
| ) |
| job_resource_done["configuration"]["query"]["destinationTable"] = { |
| "projectId": "dest-project", |
| "datasetId": "dest_dataset", |
| "tableId": "dest_table", |
| } |
|
|
| connection = make_connection( |
| |
| |
| |
| exceptions.NotFound("not normally retriable"), |
| job_resource, |
| exceptions.NotFound("not normally retriable"), |
| query_resource, |
| |
| exceptions.NotFound("not normally retriable"), |
| job_resource, |
| exceptions.NotFound("not normally retriable"), |
| query_resource, |
| exceptions.NotFound("not normally retriable"), |
| |
| job_resource_done, |
| exceptions.NotFound("not normally retriable"), |
| query_resource_done, |
| |
| ) |
| client = _make_client(self.PROJECT, connection=connection) |
| job = self._get_target_class().from_api_repr(job_resource, client) |
|
|
| custom_predicate = mock.Mock() |
| custom_predicate.return_value = True |
| custom_retry = google.api_core.retry.Retry( |
| initial=0.001, |
| maximum=0.001, |
| multiplier=1.0, |
| deadline=0.1, |
| predicate=custom_predicate, |
| ) |
|
|
| self.assertIsInstance(job.result(retry=custom_retry), RowIterator) |
| query_results_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/queries/{self.JOB_ID}", |
| query_params={"maxResults": 0, "location": "asia-northeast1"}, |
| |
| |
| |
| timeout=mock.ANY, |
| ) |
| reload_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}", |
| query_params={"projection": "full", "location": "asia-northeast1"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
|
|
| connection.api_request.assert_has_calls( |
| [ |
| |
| |
| |
| |
| reload_call, |
| reload_call, |
| query_results_call, |
| query_results_call, |
| |
| reload_call, |
| reload_call, |
| query_results_call, |
| query_results_call, |
| |
| reload_call, |
| reload_call, |
| query_results_call, |
| query_results_call, |
| ] |
| ) |
|
|
| def test_result_w_empty_schema(self): |
| from google.cloud.bigquery.table import _EmptyRowIterator |
|
|
| |
| query_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": []}, |
| "queryId": "xyz-abc", |
| } |
| connection = make_connection(query_resource, query_resource) |
| client = _make_client(self.PROJECT, connection=connection) |
| resource = self._make_resource(ended=True, location="asia-northeast1") |
| job = self._get_target_class().from_api_repr(resource, client) |
|
|
| result = job.result() |
|
|
| self.assertIsInstance(result, _EmptyRowIterator) |
| self.assertEqual(list(result), []) |
| self.assertEqual(result.project, self.PROJECT) |
| self.assertEqual(result.job_id, self.JOB_ID) |
| self.assertEqual(result.location, "asia-northeast1") |
| self.assertEqual(result.query_id, "xyz-abc") |
|
|
| def test_result_w_timeout_doesnt_raise(self): |
| import google.cloud.bigquery.client |
|
|
| begun_resource = self._make_resource() |
| query_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| } |
| done_resource = copy.deepcopy(begun_resource) |
| done_resource["status"] = {"state": "DONE"} |
| connection = make_connection(begun_resource, query_resource, done_resource) |
| client = _make_client(project=self.PROJECT, connection=connection) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| job._properties["jobReference"]["location"] = "US" |
| job._properties["status"] = {"state": "RUNNING"} |
|
|
| with freezegun.freeze_time("1970-01-01 00:00:00", tick=False): |
| job.result( |
| |
| |
| |
| |
| |
| timeout=1.125, |
| ) |
|
|
| reload_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}", |
| query_params={"projection": "full", "location": "US"}, |
| timeout=1.125, |
| ) |
| get_query_results_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/queries/{self.JOB_ID}", |
| query_params={ |
| "maxResults": 0, |
| "location": "US", |
| }, |
| timeout=google.cloud.bigquery.client._MIN_GET_QUERY_RESULTS_TIMEOUT, |
| ) |
| connection.api_request.assert_has_calls( |
| [ |
| reload_call, |
| get_query_results_call, |
| reload_call, |
| ] |
| ) |
|
|
| def test_result_w_timeout_raises_concurrent_futures_timeout(self): |
| import google.cloud.bigquery.client |
|
|
| begun_resource = self._make_resource() |
| begun_resource["jobReference"]["location"] = "US" |
| query_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| } |
| done_resource = copy.deepcopy(begun_resource) |
| done_resource["status"] = {"state": "DONE"} |
| connection = make_connection(begun_resource, query_resource, done_resource) |
| client = _make_client(project=self.PROJECT, connection=connection) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| job._properties["jobReference"]["location"] = "US" |
| job._properties["status"] = {"state": "RUNNING"} |
|
|
| with freezegun.freeze_time( |
| "1970-01-01 00:00:00", auto_tick_seconds=1.0 |
| ), self.assertRaises(concurrent.futures.TimeoutError): |
| job.result(timeout=1.125) |
|
|
| reload_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}", |
| query_params={"projection": "full", "location": "US"}, |
| timeout=1.125, |
| ) |
| get_query_results_call = mock.call( |
| method="GET", |
| path=f"/projects/{self.PROJECT}/queries/{self.JOB_ID}", |
| query_params={ |
| "maxResults": 0, |
| "location": "US", |
| }, |
| timeout=google.cloud.bigquery.client._MIN_GET_QUERY_RESULTS_TIMEOUT, |
| ) |
| connection.api_request.assert_has_calls( |
| [ |
| reload_call, |
| get_query_results_call, |
| |
| ] |
| ) |
|
|
| def test_result_w_page_size(self): |
| |
| query_results_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "totalRows": "10", |
| "rows": [ |
| {"f": [{"v": "row1"}]}, |
| {"f": [{"v": "row2"}]}, |
| {"f": [{"v": "row3"}]}, |
| {"f": [{"v": "row4"}]}, |
| {"f": [{"v": "row5"}]}, |
| {"f": [{"v": "row6"}]}, |
| {"f": [{"v": "row7"}]}, |
| {"f": [{"v": "row8"}]}, |
| {"f": [{"v": "row9"}]}, |
| ], |
| "pageToken": "first-page-token", |
| } |
| job_resource_running = self._make_resource( |
| started=True, ended=False, location="US" |
| ) |
| job_resource_done = self._make_resource(started=True, ended=True, location="US") |
| destination_table = { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.TABLE_ID, |
| } |
| q_config = job_resource_done["configuration"]["query"] |
| q_config["destinationTable"] = destination_table |
| query_page_resource_2 = {"totalRows": 10, "rows": [{"f": [{"v": "row10"}]}]} |
| conn = make_connection( |
| job_resource_running, |
| query_results_resource, |
| job_resource_done, |
| query_page_resource_2, |
| ) |
| client = _make_client(self.PROJECT, connection=conn) |
| job = self._get_target_class().from_api_repr(job_resource_running, client) |
|
|
| |
| result = job.result(page_size=9) |
|
|
| |
| actual_rows = list(result) |
| self.assertEqual(len(actual_rows), 10) |
|
|
| jobs_get_path = f"/projects/{self.PROJECT}/jobs/{self.JOB_ID}" |
| jobs_get_call = mock.call( |
| method="GET", |
| path=jobs_get_path, |
| query_params={"projection": "full", "location": "US"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| query_results_path = f"/projects/{self.PROJECT}/queries/{self.JOB_ID}" |
| query_page_waiting_call = mock.call( |
| method="GET", |
| path=query_results_path, |
| query_params={ |
| |
| |
| |
| |
| "maxResults": 9, |
| "location": "US", |
| "formatOptions.useInt64Timestamp": True, |
| }, |
| timeout=None, |
| ) |
| query_page_2_call = mock.call( |
| timeout=None, |
| method="GET", |
| path=query_results_path, |
| query_params={ |
| "pageToken": "first-page-token", |
| "maxResults": 9, |
| "fields": _LIST_ROWS_FROM_QUERY_RESULTS_FIELDS, |
| "location": "US", |
| "formatOptions.useInt64Timestamp": True, |
| }, |
| ) |
| conn.api_request.assert_has_calls( |
| [jobs_get_call, query_page_waiting_call, jobs_get_call, query_page_2_call] |
| ) |
|
|
| def test_result_with_start_index(self): |
| from google.cloud.bigquery.table import RowIterator |
|
|
| query_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| "totalRows": "5", |
| } |
| tabledata_resource = { |
| "totalRows": "5", |
| "pageToken": None, |
| "rows": [ |
| {"f": [{"v": "abc"}]}, |
| {"f": [{"v": "def"}]}, |
| {"f": [{"v": "ghi"}]}, |
| {"f": [{"v": "jkl"}]}, |
| ], |
| } |
| connection = make_connection(query_resource, tabledata_resource) |
| client = _make_client(self.PROJECT, connection=connection) |
| resource = self._make_resource(ended=True) |
| job = self._get_target_class().from_api_repr(resource, client) |
|
|
| start_index = 1 |
|
|
| |
| |
| |
| page_size = 10 |
| max_results = 100 |
|
|
| result = job.result( |
| page_size=page_size, |
| max_results=max_results, |
| start_index=start_index, |
| ) |
|
|
| self.assertIsInstance(result, RowIterator) |
| self.assertEqual(result.total_rows, 5) |
|
|
| rows = list(result) |
|
|
| self.assertEqual(len(rows), 4) |
| self.assertEqual(len(connection.api_request.call_args_list), 2) |
| tabledata_list_request = connection.api_request.call_args_list[1] |
| self.assertEqual( |
| tabledata_list_request[1]["query_params"]["startIndex"], start_index |
| ) |
| self.assertEqual( |
| tabledata_list_request[1]["query_params"]["maxResults"], page_size |
| ) |
|
|
| def test_result_error(self): |
| from google.cloud import exceptions |
|
|
| query = textwrap.dedent( |
| """ |
| SELECT foo, bar |
| FROM table_baz |
| WHERE foo == bar""" |
| ) |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, query, client) |
| error_result = { |
| "debugInfo": "DEBUG", |
| "location": "LOCATION", |
| "message": "MESSAGE", |
| "reason": "invalid", |
| } |
| job._properties["status"] = { |
| "errorResult": error_result, |
| "errors": [error_result], |
| "state": "DONE", |
| } |
| job._query_results = google.cloud.bigquery.query._QueryResults.from_api_repr( |
| {"jobComplete": True, "jobReference": job._properties["jobReference"]} |
| ) |
| job._set_future_result() |
|
|
| with self.assertRaises(exceptions.GoogleCloudError) as exc_info: |
| job.result() |
|
|
| self.assertIsInstance(exc_info.exception, exceptions.GoogleCloudError) |
| self.assertEqual(exc_info.exception.code, http.client.BAD_REQUEST) |
|
|
| exc_job_instance = getattr(exc_info.exception, "query_job", None) |
| self.assertIs(exc_job_instance, job) |
|
|
| |
| |
| full_text = str(exc_info.exception) |
| assert job.job_id in full_text |
| assert "Query Job SQL Follows" not in full_text |
|
|
| |
| |
| debug_message = exc_info.exception.debug_message |
| assert "Query Job SQL Follows" in debug_message |
| for i, line in enumerate(query.splitlines(), start=1): |
| expected_line = "{}:{}".format(i, line) |
| assert expected_line in debug_message |
|
|
| def test_result_transport_timeout_error(self): |
| query = textwrap.dedent( |
| """ |
| SELECT foo, bar |
| FROM table_baz |
| WHERE foo == bar""" |
| ) |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, query, client) |
| call_api_patch = mock.patch( |
| "google.cloud.bigquery.client.Client._call_api", |
| autospec=True, |
| side_effect=requests.exceptions.Timeout("Server response took too long."), |
| ) |
|
|
| |
| with call_api_patch, self.assertRaises(concurrent.futures.TimeoutError): |
| job.result(timeout=1) |
|
|
| def test_no_schema(self): |
| client = _make_client(project=self.PROJECT) |
| resource = {} |
| klass = self._get_target_class() |
| job = klass.from_api_repr(resource, client=client) |
| assert job.schema is None |
|
|
| def test_schema(self): |
| client = _make_client(project=self.PROJECT) |
| resource = { |
| "statistics": { |
| "query": { |
| "schema": { |
| "fields": [ |
| {"mode": "NULLABLE", "name": "bool_col", "type": "BOOLEAN"}, |
| { |
| "mode": "NULLABLE", |
| "name": "string_col", |
| "type": "STRING", |
| }, |
| { |
| "mode": "NULLABLE", |
| "name": "timestamp_col", |
| "type": "TIMESTAMP", |
| }, |
| ] |
| }, |
| }, |
| }, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(resource, client=client) |
| assert len(job.schema) == 3 |
| assert job.schema[0].field_type == "BOOLEAN" |
| assert job.schema[1].field_type == "STRING" |
| assert job.schema[2].field_type == "TIMESTAMP" |
|
|
| def test__begin_error(self): |
| from google.cloud import exceptions |
|
|
| query = textwrap.dedent( |
| """ |
| SELECT foo, bar |
| FROM table_baz |
| WHERE foo == bar""" |
| ) |
|
|
| client = _make_client(project=self.PROJECT) |
| job = self._make_one(self.JOB_ID, query, client) |
| call_api_patch = mock.patch( |
| "google.cloud.bigquery.client.Client._call_api", |
| autospec=True, |
| side_effect=exceptions.BadRequest("Syntax error in SQL query"), |
| ) |
|
|
| with call_api_patch, self.assertRaises(exceptions.GoogleCloudError) as exc_info: |
| job.result() |
|
|
| self.assertIsInstance(exc_info.exception, exceptions.GoogleCloudError) |
| self.assertEqual(exc_info.exception.code, http.client.BAD_REQUEST) |
|
|
| exc_job_instance = getattr(exc_info.exception, "query_job", None) |
| self.assertIs(exc_job_instance, job) |
|
|
| |
| |
| full_text = str(exc_info.exception) |
| assert job.job_id in full_text |
| assert "Query Job SQL Follows" not in full_text |
|
|
| |
| |
| debug_message = exc_info.exception.debug_message |
| assert "Query Job SQL Follows" in debug_message |
| for i, line in enumerate(query.splitlines(), start=1): |
| expected_line = "{}:{}".format(i, line) |
| assert expected_line in debug_message |
|
|
| def test__begin_w_timeout(self): |
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
|
|
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin(timeout=7.5) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": {"query": self.QUERY, "useLegacySql": False} |
| }, |
| }, |
| timeout=7.5, |
| ) |
|
|
| def test_begin_w_bound_client(self): |
| from google.cloud.bigquery.dataset import DatasetReference |
| from google.cloud.bigquery.job import QueryJobConfig |
|
|
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| DS_ID = "DATASET" |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
|
|
| config = QueryJobConfig() |
| config.default_dataset = DatasetReference(self.PROJECT, DS_ID) |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| self.assertIsNone(job.default_dataset) |
| self.assertEqual(job.udf_resources, []) |
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": { |
| "query": self.QUERY, |
| "useLegacySql": False, |
| "defaultDataset": { |
| "projectId": self.PROJECT, |
| "datasetId": DS_ID, |
| }, |
| } |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_alternate_client(self): |
| from google.cloud.bigquery.dataset import DatasetReference |
| from google.cloud.bigquery.job import CreateDisposition |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.job import QueryPriority |
| from google.cloud.bigquery.job import SchemaUpdateOption |
| from google.cloud.bigquery.job import WriteDisposition |
|
|
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| TABLE = "TABLE" |
| DS_ID = "DATASET" |
| RESOURCE = self._make_resource(ended=True) |
| QUERY_CONFIGURATION = { |
| "query": self.QUERY, |
| "allowLargeResults": True, |
| "createDisposition": CreateDisposition.CREATE_NEVER, |
| "defaultDataset": {"projectId": self.PROJECT, "datasetId": DS_ID}, |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": DS_ID, |
| "tableId": TABLE, |
| }, |
| "flattenResults": True, |
| "priority": QueryPriority.INTERACTIVE, |
| "useQueryCache": True, |
| "useLegacySql": True, |
| "writeDisposition": WriteDisposition.WRITE_TRUNCATE, |
| "maximumBillingTier": 4, |
| "maximumBytesBilled": "123456", |
| "schemaUpdateOptions": [SchemaUpdateOption.ALLOW_FIELD_RELAXATION], |
| } |
| RESOURCE["configuration"]["query"] = QUERY_CONFIGURATION |
| RESOURCE["configuration"]["dryRun"] = True |
| conn1 = make_connection() |
| client1 = _make_client(project=self.PROJECT, connection=conn1) |
| conn2 = make_connection(RESOURCE) |
| client2 = _make_client(project=self.PROJECT, connection=conn2) |
| dataset_ref = DatasetReference(self.PROJECT, DS_ID) |
| table_ref = dataset_ref.table(TABLE) |
|
|
| config = QueryJobConfig() |
| config.allow_large_results = True |
| config.create_disposition = CreateDisposition.CREATE_NEVER |
| config.default_dataset = dataset_ref |
| config.destination = table_ref |
| config.dry_run = True |
| config.flatten_results = True |
| config.maximum_billing_tier = 4 |
| config.priority = QueryPriority.INTERACTIVE |
| config.use_legacy_sql = True |
| config.use_query_cache = True |
| config.write_disposition = WriteDisposition.WRITE_TRUNCATE |
| config.maximum_bytes_billed = 123456 |
| config.schema_update_options = [SchemaUpdateOption.ALLOW_FIELD_RELAXATION] |
| job = self._make_one(self.JOB_ID, self.QUERY, client1, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin(client=client2) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client2, job) |
|
|
| conn1.api_request.assert_not_called() |
| conn2.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": {"dryRun": True, "query": QUERY_CONFIGURATION}, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_udf(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.query import UDFResource |
|
|
| RESOURCE_URI = "gs://some-bucket/js/lib.js" |
| INLINE_UDF_CODE = 'var someCode = "here";' |
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
| RESOURCE["configuration"]["query"]["userDefinedFunctionResources"] = [ |
| {"resourceUri": RESOURCE_URI}, |
| {"inlineCode": INLINE_UDF_CODE}, |
| ] |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| udf_resources = [ |
| UDFResource("resourceUri", RESOURCE_URI), |
| UDFResource("inlineCode", INLINE_UDF_CODE), |
| ] |
| config = QueryJobConfig() |
| config.udf_resources = udf_resources |
| config.use_legacy_sql = True |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| self.assertEqual(job.udf_resources, udf_resources) |
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": { |
| "query": self.QUERY, |
| "useLegacySql": True, |
| "userDefinedFunctionResources": [ |
| {"resourceUri": RESOURCE_URI}, |
| {"inlineCode": INLINE_UDF_CODE}, |
| ], |
| } |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_named_query_parameter(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.query import ScalarQueryParameter |
|
|
| query_parameters = [ScalarQueryParameter("foo", "INT64", 123)] |
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
| config = RESOURCE["configuration"]["query"] |
| config["parameterMode"] = "NAMED" |
| config["queryParameters"] = [ |
| { |
| "name": "foo", |
| "parameterType": {"type": "INT64"}, |
| "parameterValue": {"value": "123"}, |
| } |
| ] |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| jconfig = QueryJobConfig() |
| jconfig.query_parameters = query_parameters |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=jconfig) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| self.assertEqual(job.query_parameters, query_parameters) |
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": { |
| "query": self.QUERY, |
| "useLegacySql": False, |
| "parameterMode": "NAMED", |
| "queryParameters": config["queryParameters"], |
| } |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_positional_query_parameter(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.query import ScalarQueryParameter |
|
|
| query_parameters = [ScalarQueryParameter.positional("INT64", 123)] |
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
| config = RESOURCE["configuration"]["query"] |
| config["parameterMode"] = "POSITIONAL" |
| config["queryParameters"] = [ |
| {"parameterType": {"type": "INT64"}, "parameterValue": {"value": "123"}} |
| ] |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| jconfig = QueryJobConfig() |
| jconfig.query_parameters = query_parameters |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=jconfig) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| self.assertEqual(job.query_parameters, query_parameters) |
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": { |
| "query": self.QUERY, |
| "useLegacySql": False, |
| "parameterMode": "POSITIONAL", |
| "queryParameters": config["queryParameters"], |
| } |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_table_defs(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
| from google.cloud.bigquery.external_config import ExternalConfig |
| from google.cloud.bigquery.external_config import BigtableColumn |
| from google.cloud.bigquery.external_config import BigtableColumnFamily |
|
|
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
|
|
| bt_config = ExternalConfig("BIGTABLE") |
| bt_config.ignore_unknown_values = True |
| bt_config.options.read_rowkey_as_string = True |
| cf = BigtableColumnFamily() |
| cf.family_id = "cf" |
| col = BigtableColumn() |
| col.field_name = "fn" |
| cf.columns = [col] |
| bt_config.options.column_families = [cf] |
| BT_CONFIG_RESOURCE = { |
| "sourceFormat": "BIGTABLE", |
| "ignoreUnknownValues": True, |
| "bigtableOptions": { |
| "readRowkeyAsString": True, |
| "columnFamilies": [ |
| {"familyId": "cf", "columns": [{"fieldName": "fn"}]} |
| ], |
| }, |
| } |
| CSV_CONFIG_RESOURCE = { |
| "sourceFormat": "CSV", |
| "maxBadRecords": 8, |
| "csvOptions": {"allowJaggedRows": True}, |
| } |
| csv_config = ExternalConfig("CSV") |
| csv_config.max_bad_records = 8 |
| csv_config.options.allow_jagged_rows = True |
| bt_table = "bigtable-table" |
| csv_table = "csv-table" |
| RESOURCE["configuration"]["query"]["tableDefinitions"] = { |
| bt_table: BT_CONFIG_RESOURCE, |
| csv_table: CSV_CONFIG_RESOURCE, |
| } |
| want_resource = copy.deepcopy(RESOURCE) |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| config = QueryJobConfig() |
| config.table_definitions = {bt_table: bt_config, csv_table: csv_config} |
| config.use_legacy_sql = True |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": { |
| "query": self.QUERY, |
| "useLegacySql": True, |
| "tableDefinitions": { |
| bt_table: BT_CONFIG_RESOURCE, |
| csv_table: CSV_CONFIG_RESOURCE, |
| }, |
| } |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, want_resource) |
|
|
| def test_dry_run_query(self): |
| from google.cloud.bigquery.job import QueryJobConfig |
|
|
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
| RESOURCE["configuration"]["dryRun"] = True |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| config = QueryJobConfig() |
| config.dry_run = True |
| job = self._make_one(self.JOB_ID, self.QUERY, client, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
| self.assertEqual(job.udf_resources, []) |
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "query": {"query": self.QUERY, "useLegacySql": False}, |
| "dryRun": True, |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_exists_miss_w_bound_client(self): |
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| conn = make_connection() |
| client = _make_client(project=self.PROJECT, connection=conn) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| self.assertFalse(job.exists()) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| conn.api_request.assert_called_once_with( |
| method="GET", path=PATH, query_params={"fields": "id"}, timeout=None |
| ) |
|
|
| def test_exists_hit_w_alternate_client(self): |
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| conn1 = make_connection() |
| client1 = _make_client(project=self.PROJECT, connection=conn1) |
| conn2 = make_connection({}) |
| client2 = _make_client(project=self.PROJECT, connection=conn2) |
| job = self._make_one(self.JOB_ID, self.QUERY, client1) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| self.assertTrue(job.exists(client=client2)) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client2, job) |
|
|
| conn1.api_request.assert_not_called() |
| conn2.api_request.assert_called_once_with( |
| method="GET", path=PATH, query_params={"fields": "id"}, timeout=None |
| ) |
|
|
| def test_reload_w_bound_client(self): |
| from google.cloud.bigquery.dataset import DatasetReference |
| from google.cloud.bigquery.job import QueryJobConfig |
|
|
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| DS_ID = "DATASET" |
| DEST_TABLE = "dest_table" |
| RESOURCE = self._make_resource() |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| dataset_ref = DatasetReference(self.PROJECT, DS_ID) |
| table_ref = dataset_ref.table(DEST_TABLE) |
| config = QueryJobConfig() |
| config.destination = table_ref |
| job = self._make_one(self.JOB_ID, None, client, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job.reload() |
|
|
| final_attributes.assert_called_with( |
| { |
| "path": PATH, |
| "job_id": self.JOB_ID, |
| "location": None, |
| }, |
| client, |
| None, |
| ) |
|
|
| self.assertNotEqual(job.destination, table_ref) |
|
|
| conn.api_request.assert_called_once_with( |
| method="GET", |
| path=PATH, |
| query_params={"projection": "full"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_reload_w_alternate_client(self): |
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| DS_ID = "DATASET" |
| DEST_TABLE = "dest_table" |
| RESOURCE = self._make_resource() |
| q_config = RESOURCE["configuration"]["query"] |
| q_config["destinationTable"] = { |
| "projectId": self.PROJECT, |
| "datasetId": DS_ID, |
| "tableId": DEST_TABLE, |
| } |
| conn1 = make_connection() |
| client1 = _make_client(project=self.PROJECT, connection=conn1) |
| conn2 = make_connection(RESOURCE) |
| client2 = _make_client(project=self.PROJECT, connection=conn2) |
| job = self._make_one(self.JOB_ID, self.QUERY, client1) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job.reload(client=client2) |
|
|
| final_attributes.assert_called_with( |
| { |
| "path": PATH, |
| "job_id": self.JOB_ID, |
| "location": None, |
| }, |
| client2, |
| None, |
| ) |
|
|
| conn1.api_request.assert_not_called() |
| conn2.api_request.assert_called_once_with( |
| method="GET", |
| path=PATH, |
| query_params={"projection": "full"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_reload_w_timeout(self): |
| from google.cloud.bigquery.dataset import DatasetReference |
| from google.cloud.bigquery.job import QueryJobConfig |
|
|
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| DS_ID = "DATASET" |
| DEST_TABLE = "dest_table" |
| RESOURCE = self._make_resource() |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| dataset_ref = DatasetReference(self.PROJECT, DS_ID) |
| table_ref = dataset_ref.table(DEST_TABLE) |
| config = QueryJobConfig() |
| config.destination = table_ref |
| job = self._make_one(self.JOB_ID, None, client, job_config=config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job.reload(timeout=4.2) |
| final_attributes.assert_called_with( |
| { |
| "path": PATH, |
| "job_id": self.JOB_ID, |
| "location": None, |
| }, |
| client, |
| None, |
| ) |
|
|
| self.assertNotEqual(job.destination, table_ref) |
|
|
| conn.api_request.assert_called_once_with( |
| method="GET", |
| path=PATH, |
| query_params={"projection": "full"}, |
| timeout=4.2, |
| ) |
|
|
| def test_iter(self): |
| begun_resource = self._make_resource() |
| query_resource = { |
| "jobComplete": True, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "totalRows": "0", |
| "schema": {"fields": [{"name": "col1", "type": "STRING"}]}, |
| } |
| done_resource = copy.deepcopy(begun_resource) |
| done_resource["status"] = {"state": "DONE"} |
| connection = make_connection(begun_resource, query_resource, done_resource) |
| client = _make_client(project=self.PROJECT, connection=connection) |
| job = self._make_one(self.JOB_ID, self.QUERY, client) |
| job._properties["status"] = {"state": "RUNNING"} |
|
|
| self.assertIsInstance(iter(job), types.GeneratorType) |
|
|