# Copyright 2015 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import copy import collections import datetime import decimal import email import gzip import http.client import io import itertools import json import operator import os import unittest from unittest import mock import warnings import requests import packaging import pytest try: import opentelemetry except ImportError: opentelemetry = None if opentelemetry is not None: try: from opentelemetry import trace from opentelemetry.sdk.trace import TracerProvider from opentelemetry.sdk.trace.export import SimpleSpanProcessor from opentelemetry.sdk.trace.export.in_memory_span_exporter import ( InMemorySpanExporter, ) except (ImportError, AttributeError) as exc: # pragma: NO COVER msg = "Error importing from opentelemetry, is the installed version compatible?" raise ImportError(msg) from exc import google.api_core.exceptions from google.api_core import client_info import google.cloud._helpers from google.cloud import bigquery from google.cloud.bigquery.dataset import DatasetReference from google.cloud.bigquery import exceptions from google.cloud.bigquery import ParquetOptions import google.cloud.bigquery.retry from google.cloud.bigquery.retry import DEFAULT_TIMEOUT import google.cloud.bigquery.table from test_utils.imports import maybe_fail_import from tests.unit.helpers import make_connection def _make_credentials(): import google.auth.credentials return mock.Mock(spec=google.auth.credentials.Credentials) def _make_list_partitons_meta_info(project, dataset_id, table_id, num_rows=0): return { "tableReference": { "projectId": project, "datasetId": dataset_id, "tableId": "{}$__PARTITIONS_SUMMARY__".format(table_id), }, "schema": { "fields": [ {"name": "project_id", "type": "STRING", "mode": "NULLABLE"}, {"name": "dataset_id", "type": "STRING", "mode": "NULLABLE"}, {"name": "table_id", "type": "STRING", "mode": "NULLABLE"}, {"name": "partition_id", "type": "STRING", "mode": "NULLABLE"}, ] }, "etag": "ETAG", "numRows": num_rows, } class TestClient(unittest.TestCase): PROJECT = "PROJECT" DS_ID = "DATASET_ID" TABLE_ID = "TABLE_ID" MODEL_ID = "MODEL_ID" TABLE_REF = DatasetReference(PROJECT, DS_ID).table(TABLE_ID) KMS_KEY_NAME = "projects/1/locations/us/keyRings/1/cryptoKeys/1" LOCATION = "us-central" @staticmethod def _get_target_class(): from google.cloud.bigquery.client import Client return Client def _make_one(self, *args, **kw): return self._get_target_class()(*args, **kw) def _make_table_resource(self): return { "id": "%s:%s:%s" % (self.PROJECT, self.DS_ID, self.TABLE_ID), "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, } def test_ctor_defaults(self): from google.cloud.bigquery._http import Connection creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) self.assertIsInstance(client._connection, Connection) self.assertIs(client._connection.credentials, creds) self.assertIs(client._connection.http, http) self.assertIsNone(client.location) self.assertEqual( client._connection.API_BASE_URL, Connection.DEFAULT_API_ENDPOINT ) def test_ctor_w_empty_client_options(self): from google.api_core.client_options import ClientOptions creds = _make_credentials() http = object() client_options = ClientOptions() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, client_options=client_options, ) self.assertEqual( client._connection.API_BASE_URL, client._connection.DEFAULT_API_ENDPOINT ) @mock.patch.dict(os.environ, {"GOOGLE_CLOUD_UNIVERSE_DOMAIN": "foo.com"}) def test_ctor_w_only_env_universe(self): creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, ) self.assertEqual(client._connection.API_BASE_URL, "https://bigquery.foo.com") def test_ctor_w_client_options_dict(self): creds = _make_credentials() http = object() client_options = {"api_endpoint": "https://www.foo-googleapis.com"} client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, client_options=client_options, ) self.assertEqual( client._connection.API_BASE_URL, "https://www.foo-googleapis.com" ) def test_ctor_w_client_options_object(self): from google.api_core.client_options import ClientOptions creds = _make_credentials() http = object() client_options = ClientOptions(api_endpoint="https://www.foo-googleapis.com") client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, client_options=client_options, ) self.assertEqual( client._connection.API_BASE_URL, "https://www.foo-googleapis.com" ) @pytest.mark.skipif( packaging.version.parse(getattr(google.api_core, "__version__", "0.0.0")) < packaging.version.Version("2.15.0"), reason="universe_domain not supported with google-api-core < 2.15.0", ) def test_ctor_w_client_options_universe(self): creds = _make_credentials() http = object() client_options = {"universe_domain": "foo.com"} client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, client_options=client_options, ) self.assertEqual(client._connection.API_BASE_URL, "https://bigquery.foo.com") def test_ctor_w_location(self): from google.cloud.bigquery._http import Connection creds = _make_credentials() http = object() location = "us-central" client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=location ) self.assertIsInstance(client._connection, Connection) self.assertIs(client._connection.credentials, creds) self.assertIs(client._connection.http, http) self.assertEqual(client.location, location) def test_ctor_w_query_job_config(self): from google.cloud.bigquery._http import Connection from google.cloud.bigquery import QueryJobConfig creds = _make_credentials() http = object() location = "us-central" job_config = QueryJobConfig() job_config.dry_run = True client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=location, default_query_job_config=job_config, ) self.assertIsInstance(client._connection, Connection) self.assertIs(client._connection.credentials, creds) self.assertIs(client._connection.http, http) self.assertEqual(client.location, location) self.assertIsInstance(client._default_query_job_config, QueryJobConfig) self.assertTrue(client._default_query_job_config.dry_run) def test_ctor_w_load_job_config(self): from google.cloud.bigquery._http import Connection from google.cloud.bigquery import LoadJobConfig creds = _make_credentials() http = object() location = "us-central" job_config = LoadJobConfig() job_config.create_session = True client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=location, default_load_job_config=job_config, ) self.assertIsInstance(client._connection, Connection) self.assertIs(client._connection.credentials, creds) self.assertIs(client._connection.http, http) self.assertEqual(client.location, location) self.assertIsInstance(client._default_load_job_config, LoadJobConfig) self.assertTrue(client._default_load_job_config.create_session) def test__call_api_extra_headers(self): # Note: We test at a lower layer to ensure that extra headers are # populated when we actually make the call in requests. # Arrange http = mock.create_autospec(requests.Session, instance=True) http.is_mtls = False response = mock.create_autospec(requests.Response, instance=True) response.status_code = 200 http.request.return_value = response creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) # Act client._connection.extra_headers = {"x-goog-request-reason": "because-friday"} client._call_api( retry=None, method="GET", path="/bigquery/v2/projects/my-proj/jobs/my-job" ) # Assert http.request.assert_called_once() _, kwargs = http.request.call_args headers = kwargs["headers"] assert headers["x-goog-request-reason"] == "because-friday" def test__call_api_applying_custom_retry_on_timeout(self): from concurrent.futures import TimeoutError from google.cloud.bigquery.retry import DEFAULT_RETRY creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) api_request_patcher = mock.patch.object( client._connection, "api_request", side_effect=[TimeoutError, "result"], ) retry = DEFAULT_RETRY.with_deadline(1).with_predicate( lambda exc: isinstance(exc, TimeoutError) ) with api_request_patcher as fake_api_request: result = client._call_api(retry, foo="bar") self.assertEqual(result, "result") self.assertEqual( fake_api_request.call_args_list, [mock.call(foo="bar"), mock.call(foo="bar")], # was retried once ) def test__call_api_span_creator_not_called(self): from concurrent.futures import TimeoutError from google.cloud.bigquery.retry import DEFAULT_RETRY creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) api_request_patcher = mock.patch.object( client._connection, "api_request", side_effect=[TimeoutError, "result"], ) retry = DEFAULT_RETRY.with_deadline(1).with_predicate( lambda exc: isinstance(exc, TimeoutError) ) with api_request_patcher: with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client._call_api(retry) final_attributes.assert_not_called() def test__call_api_span_creator_called(self): from concurrent.futures import TimeoutError from google.cloud.bigquery.retry import DEFAULT_RETRY creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) api_request_patcher = mock.patch.object( client._connection, "api_request", side_effect=[TimeoutError, "result"], ) retry = DEFAULT_RETRY.with_deadline(1).with_predicate( lambda exc: isinstance(exc, TimeoutError) ) with api_request_patcher: with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client._call_api( retry, span_name="test_name", span_attributes={"test_attribute": "test_attribute-value"}, ) final_attributes.assert_called_once() def test__get_query_results_miss_w_explicit_project_and_timeout(self): from google.cloud.exceptions import NotFound creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() path = "/projects/other-project/queries/nothere" with self.assertRaises(NotFound): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client._get_query_results( "nothere", None, project="other-project", location=self.LOCATION, timeout_ms=500, timeout=420, ) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_once_with( method="GET", path=path, query_params={"maxResults": 0, "timeoutMs": 500, "location": self.LOCATION}, timeout=420, ) def test__get_query_results_miss_w_short_timeout(self): import google.cloud.bigquery.client from google.cloud.exceptions import NotFound creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() path = "/projects/other-project/queries/nothere" with self.assertRaises(NotFound): client._get_query_results( "nothere", None, project="other-project", location=self.LOCATION, timeout_ms=500, timeout=1, ) conn.api_request.assert_called_once_with( method="GET", path=path, query_params={"maxResults": 0, "timeoutMs": 500, "location": self.LOCATION}, timeout=google.cloud.bigquery.client._MIN_GET_QUERY_RESULTS_TIMEOUT, ) def test__get_query_results_miss_w_default_timeout(self): import google.cloud.bigquery.client from google.cloud.exceptions import NotFound creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() path = "/projects/other-project/queries/nothere" with self.assertRaises(NotFound): client._get_query_results( "nothere", None, project="other-project", location=self.LOCATION, timeout_ms=500, timeout=object(), # the api core default timeout ) conn.api_request.assert_called_once_with( method="GET", path=path, query_params={"maxResults": 0, "timeoutMs": 500, "location": self.LOCATION}, timeout=google.cloud.bigquery.client._MIN_GET_QUERY_RESULTS_TIMEOUT, ) def test__get_query_results_miss_w_client_location(self): from google.cloud.exceptions import NotFound creds = _make_credentials() client = self._make_one(self.PROJECT, creds, location=self.LOCATION) conn = client._connection = make_connection() with self.assertRaises(NotFound): client._get_query_results("nothere", None) conn.api_request.assert_called_once_with( method="GET", path="/projects/PROJECT/queries/nothere", query_params={"maxResults": 0, "location": self.LOCATION}, timeout=DEFAULT_TIMEOUT, ) def test__get_query_results_hit(self): job_id = "query_job" data = { "kind": "bigquery#getQueryResultsResponse", "etag": "some-tag", "schema": { "fields": [ {"name": "title", "type": "STRING", "mode": "NULLABLE"}, {"name": "unique_words", "type": "INTEGER", "mode": "NULLABLE"}, ] }, "jobReference": {"projectId": self.PROJECT, "jobId": job_id}, "totalRows": "10", "totalBytesProcessed": "2464625", "jobComplete": True, "cacheHit": False, } creds = _make_credentials() client = self._make_one(self.PROJECT, creds) client._connection = make_connection(data) query_results = client._get_query_results(job_id, None) self.assertEqual(query_results.total_rows, 10) self.assertTrue(query_results.complete) def test__list_rows_from_query_results_w_none_timeout(self): from google.cloud.exceptions import NotFound from google.cloud.bigquery.schema import SchemaField creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() path = "/projects/project/queries/nothere" iterator = client._list_rows_from_query_results( "nothere", location=None, project="project", schema=[ SchemaField("f1", "STRING", mode="REQUIRED"), SchemaField("f2", "INTEGER", mode="REQUIRED"), ], timeout=None, ) # trigger the iterator to request data with self.assertRaises(NotFound): iterator._get_next_page_response() conn.api_request.assert_called_once_with( method="GET", path=path, query_params={ "fields": "jobReference,totalRows,pageToken,rows", "location": None, "formatOptions.useInt64Timestamp": True, }, timeout=None, ) def test__list_rows_from_query_results_w_default_timeout(self): import google.cloud.bigquery.client from google.cloud.exceptions import NotFound from google.cloud.bigquery.schema import SchemaField creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() path = "/projects/project/queries/nothere" iterator = client._list_rows_from_query_results( "nothere", location=None, project="project", schema=[ SchemaField("f1", "STRING", mode="REQUIRED"), SchemaField("f2", "INTEGER", mode="REQUIRED"), ], timeout=object(), ) # trigger the iterator to request data with self.assertRaises(NotFound): iterator._get_next_page_response() conn.api_request.assert_called_once_with( method="GET", path=path, query_params={ "fields": "jobReference,totalRows,pageToken,rows", "location": None, "formatOptions.useInt64Timestamp": True, }, timeout=google.cloud.bigquery.client._MIN_GET_QUERY_RESULTS_TIMEOUT, ) def test_default_query_job_config(self): from google.cloud.bigquery import QueryJobConfig creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) self.assertIsNone(client.default_query_job_config) job_config = QueryJobConfig() job_config.dry_run = True client.default_query_job_config = job_config self.assertIsInstance(client.default_query_job_config, QueryJobConfig) def test_default_load_job_config(self): from google.cloud.bigquery import LoadJobConfig creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) self.assertIsNone(client.default_load_job_config) job_config = LoadJobConfig() job_config.create_session = True client.default_load_job_config = job_config self.assertIsInstance(client.default_load_job_config, LoadJobConfig) def test_get_service_account_email(self): path = "/projects/%s/serviceAccount" % (self.PROJECT,) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) email = "bq-123@bigquery-encryption.iam.gserviceaccount.com" resource = {"kind": "bigquery#getServiceAccountResponse", "email": email} conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: service_account_email = client.get_service_account_email(timeout=7.5) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_once_with(method="GET", path=path, timeout=7.5) self.assertEqual(service_account_email, email) def test_get_service_account_email_w_alternate_project(self): project = "my-alternate-project" path = "/projects/%s/serviceAccount" % (project,) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) email = "bq-123@bigquery-encryption.iam.gserviceaccount.com" resource = {"kind": "bigquery#getServiceAccountResponse", "email": email} conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: service_account_email = client.get_service_account_email(project=project) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_once_with( method="GET", path=path, timeout=DEFAULT_TIMEOUT ) self.assertEqual(service_account_email, email) def test_get_service_account_email_w_custom_retry(self): from google.cloud.bigquery.retry import DEFAULT_RETRY api_path = "/projects/{}/serviceAccount".format(self.PROJECT) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) resource = { "kind": "bigquery#getServiceAccountResponse", "email": "bq-123@bigquery-encryption.iam.gserviceaccount.com", } api_request_patcher = mock.patch.object( client._connection, "api_request", side_effect=[ValueError, resource], ) retry = DEFAULT_RETRY.with_deadline(1).with_predicate( lambda exc: isinstance(exc, ValueError) ) with api_request_patcher as fake_api_request: with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: service_account_email = client.get_service_account_email( retry=retry, timeout=7.5 ) final_attributes.assert_called_once_with({"path": api_path}, client, None) self.assertEqual( service_account_email, "bq-123@bigquery-encryption.iam.gserviceaccount.com" ) self.assertEqual( fake_api_request.call_args_list, [ mock.call(method="GET", path=api_path, timeout=7.5), mock.call(method="GET", path=api_path, timeout=7.5), # was retried once ], ) def test_dataset_with_specified_project(self): from google.cloud.bigquery.dataset import DatasetReference creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) catch_warnings = warnings.catch_warnings(record=True) with catch_warnings as warned: dataset = client.dataset(self.DS_ID, self.PROJECT) matches = [ warning for warning in warned if warning.category in (DeprecationWarning, PendingDeprecationWarning) and "Client.dataset" in str(warning) and "my_project.my_dataset" in str(warning) ] assert matches, "A Client.dataset deprecation warning was not raised." self.assertIsInstance(dataset, DatasetReference) self.assertEqual(dataset.dataset_id, self.DS_ID) self.assertEqual(dataset.project, self.PROJECT) def test_dataset_with_default_project(self): from google.cloud.bigquery.dataset import DatasetReference creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) catch_warnings = warnings.catch_warnings(record=True) with catch_warnings as warned: dataset = client.dataset(self.DS_ID) matches = [ warning for warning in warned if warning.category in (DeprecationWarning, PendingDeprecationWarning) and "Client.dataset" in str(warning) and "my_project.my_dataset" in str(warning) ] assert matches, "A Client.dataset deprecation warning was not raised." self.assertIsInstance(dataset, DatasetReference) self.assertEqual(dataset.dataset_id, self.DS_ID) self.assertEqual(dataset.project, self.PROJECT) def test_get_dataset(self): from google.cloud.exceptions import ServerError path = "projects/%s/datasets/%s" % (self.PROJECT, self.DS_ID) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) resource = { "id": "%s:%s" % (self.PROJECT, self.DS_ID), "datasetReference": {"projectId": self.PROJECT, "datasetId": self.DS_ID}, } conn = client._connection = make_connection(resource) dataset_ref = DatasetReference(self.PROJECT, self.DS_ID) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: dataset = client.get_dataset(dataset_ref, timeout=7.5) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) conn.api_request.assert_called_once_with( method="GET", path="/%s" % path, timeout=7.5 ) self.assertEqual(dataset.dataset_id, self.DS_ID) # Test retry. # Not a cloud API exception (missing 'errors' field). client._connection = make_connection(Exception(""), resource) with self.assertRaises(Exception): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.get_dataset(dataset_ref) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) # Zero-length errors field. client._connection = make_connection(ServerError(""), resource) with self.assertRaises(ServerError): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.get_dataset(dataset_ref) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) # Non-retryable reason. client._connection = make_connection( ServerError("", errors=[{"reason": "serious"}]), resource ) with self.assertRaises(ServerError): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.get_dataset(dataset_ref) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) # Retryable reason, but retry is disabled. client._connection = make_connection( ServerError("", errors=[{"reason": "backendError"}]), resource ) with self.assertRaises(ServerError): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.get_dataset(dataset_ref, retry=None) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) # Retryable reason, default retry: success. client._connection = make_connection( ServerError("", errors=[{"reason": "backendError"}]), resource ) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: dataset = client.get_dataset( # Test with a string for dataset ID. dataset_ref.dataset_id ) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) self.assertEqual(dataset.dataset_id, self.DS_ID) def test_ensure_bqstorage_client_creating_new_instance(self): bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage") mock_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) mock_client_instance = object() mock_client.return_value = mock_client_instance creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) with mock.patch( "google.cloud.bigquery_storage.BigQueryReadClient", mock_client ): bqstorage_client = client._ensure_bqstorage_client( client_options=mock.sentinel.client_options, client_info=mock.sentinel.client_info, ) self.assertIs(bqstorage_client, mock_client_instance) mock_client.assert_called_once_with( credentials=creds, client_options=mock.sentinel.client_options, client_info=mock.sentinel.client_info, ) def test_ensure_bqstorage_client_missing_dependency(self): creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) def fail_bqstorage_import(name, globals, locals, fromlist, level): # NOTE: *very* simplified, assuming a straightforward absolute import return "bigquery_storage" in name or ( fromlist is not None and "bigquery_storage" in fromlist ) no_bqstorage = maybe_fail_import(predicate=fail_bqstorage_import) with no_bqstorage, warnings.catch_warnings(record=True) as warned: bqstorage_client = client._ensure_bqstorage_client() self.assertIsNone(bqstorage_client) matching_warnings = [ warning for warning in warned if "not installed" in str(warning) and "google-cloud-bigquery-storage" in str(warning) ] assert matching_warnings, "Missing dependency warning not raised." def test_ensure_bqstorage_client_obsolete_dependency(self): pytest.importorskip("google.cloud.bigquery_storage") creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) patcher = mock.patch( "google.cloud.bigquery.client._versions_helpers.BQ_STORAGE_VERSIONS.try_import", side_effect=exceptions.LegacyBigQueryStorageError("BQ Storage too old"), ) with patcher, warnings.catch_warnings(record=True) as warned: bqstorage_client = client._ensure_bqstorage_client() self.assertIsNone(bqstorage_client) matching_warnings = [ warning for warning in warned if "BQ Storage too old" in str(warning) ] assert matching_warnings, "Obsolete dependency warning not raised." def test_ensure_bqstorage_client_existing_client_check_passes(self): pytest.importorskip("google.cloud.bigquery_storage") creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) mock_storage_client = mock.sentinel.mock_storage_client bqstorage_client = client._ensure_bqstorage_client( bqstorage_client=mock_storage_client ) self.assertIs(bqstorage_client, mock_storage_client) def test_ensure_bqstorage_client_is_none(self): pytest.importorskip("google.cloud.bigquery_storage") creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) bqstorage_client = None assert bqstorage_client is None bqstorage_client = client._ensure_bqstorage_client( bqstorage_client=bqstorage_client, ) assert isinstance( bqstorage_client, google.cloud.bigquery_storage_v1.BigQueryReadClient ) def test_ensure_bqstorage_client_existing_client_check_fails(self): pytest.importorskip("google.cloud.bigquery_storage") creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) mock_storage_client = mock.sentinel.mock_storage_client patcher = mock.patch( "google.cloud.bigquery.client._versions_helpers.BQ_STORAGE_VERSIONS.try_import", side_effect=exceptions.LegacyBigQueryStorageError("BQ Storage too old"), ) with patcher, warnings.catch_warnings(record=True) as warned: bqstorage_client = client._ensure_bqstorage_client(mock_storage_client) self.assertIsNone(bqstorage_client) matching_warnings = [ warning for warning in warned if "BQ Storage too old" in str(warning) ] assert matching_warnings, "Obsolete dependency warning not raised." def test_create_routine_w_minimal_resource(self): from google.cloud.bigquery.routine import Routine from google.cloud.bigquery.routine import RoutineReference creds = _make_credentials() path = "/projects/test-routine-project/datasets/test_routines/routines" resource = { "routineReference": { "projectId": "test-routine-project", "datasetId": "test_routines", "routineId": "minimal_routine", } } client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource) full_routine_id = "test-routine-project.test_routines.minimal_routine" routine = Routine(full_routine_id) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: actual_routine = client.create_routine(routine, timeout=7.5) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_once_with( method="POST", path=path, data=resource, timeout=7.5, ) self.assertEqual( actual_routine.reference, RoutineReference.from_string(full_routine_id) ) def test_create_routine_w_conflict(self): from google.cloud.bigquery.routine import Routine creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection( google.api_core.exceptions.AlreadyExists("routine already exists") ) path = "/projects/test-routine-project/datasets/test_routines/routines" full_routine_id = "test-routine-project.test_routines.minimal_routine" routine = Routine(full_routine_id) with pytest.raises(google.api_core.exceptions.AlreadyExists): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.create_routine(routine) final_attributes.assert_called_once_with({"path": path}, client, None) resource = { "routineReference": { "projectId": "test-routine-project", "datasetId": "test_routines", "routineId": "minimal_routine", } } conn.api_request.assert_called_once_with( method="POST", path=path, data=resource, timeout=DEFAULT_TIMEOUT, ) def test_span_status_is_set(self): pytest.importorskip("opentelemetry") from google.cloud.bigquery.routine import Routine tracer_provider = TracerProvider() memory_exporter = InMemorySpanExporter() span_processor = SimpleSpanProcessor(memory_exporter) tracer_provider.add_span_processor(span_processor) # OpenTelemetry API >= 0.12b0 does not allow overriding the tracer once # initialized, thus directly override the internal global var. tracer_patcher = mock.patch.object(trace, "_TRACER_PROVIDER", tracer_provider) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection( google.api_core.exceptions.AlreadyExists("routine already exists") ) path = "/projects/test-routine-project/datasets/test_routines/routines" full_routine_id = "test-routine-project.test_routines.minimal_routine" routine = Routine(full_routine_id) with pytest.raises(google.api_core.exceptions.AlreadyExists), tracer_patcher: client.create_routine(routine) span_list = memory_exporter.get_finished_spans() self.assertTrue(span_list[0].status is not None) resource = { "routineReference": { "projectId": "test-routine-project", "datasetId": "test_routines", "routineId": "minimal_routine", } } conn.api_request.assert_called_once_with( method="POST", path=path, data=resource, timeout=DEFAULT_TIMEOUT, ) def test_create_routine_w_conflict_exists_ok(self): from google.cloud.bigquery.routine import Routine creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = { "routineReference": { "projectId": "test-routine-project", "datasetId": "test_routines", "routineId": "minimal_routine", } } path = "/projects/test-routine-project/datasets/test_routines/routines" conn = client._connection = make_connection( google.api_core.exceptions.AlreadyExists("routine already exists"), resource ) full_routine_id = "test-routine-project.test_routines.minimal_routine" routine = Routine(full_routine_id) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: actual_routine = client.create_routine(routine, exists_ok=True) final_attributes.assert_called_with( {"path": "%s/minimal_routine" % path}, client, None ) self.assertEqual(actual_routine.project, "test-routine-project") self.assertEqual(actual_routine.dataset_id, "test_routines") self.assertEqual(actual_routine.routine_id, "minimal_routine") conn.api_request.assert_has_calls( [ mock.call( method="POST", path=path, data=resource, timeout=DEFAULT_TIMEOUT, ), mock.call( method="GET", path="/projects/test-routine-project/datasets/test_routines/routines/minimal_routine", timeout=DEFAULT_TIMEOUT, ), ] ) def test_create_table_w_day_partition(self): from google.cloud.bigquery.table import Table from google.cloud.bigquery.table import TimePartitioning path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) table.time_partitioning = TimePartitioning() with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(table, timeout=7.5) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": table.dataset_id}, client, None ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "timePartitioning": {"type": "DAY"}, "labels": {}, }, timeout=7.5, ) self.assertEqual(table.time_partitioning.type_, "DAY") self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_w_custom_property(self): # The library should handle sending properties to the API that are not # yet part of the library from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() resource["newAlphaProperty"] = "unreleased property" conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) table._properties["newAlphaProperty"] = "unreleased property" with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(table) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": table.dataset_id}, client, None ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "newAlphaProperty": "unreleased property", "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got._properties["newAlphaProperty"], "unreleased property") self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_w_encryption_configuration(self): from google.cloud.bigquery.encryption_configuration import ( EncryptionConfiguration, ) from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) table.encryption_configuration = EncryptionConfiguration( kms_key_name=self.KMS_KEY_NAME ) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(table) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": table.dataset_id}, client, None ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "labels": {}, "encryptionConfiguration": {"kmsKeyName": self.KMS_KEY_NAME}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_w_day_partition_and_expire(self): from google.cloud.bigquery.table import Table from google.cloud.bigquery.table import TimePartitioning path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) table.time_partitioning = TimePartitioning(expiration_ms=100) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(table) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": table.dataset_id}, client, None ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "timePartitioning": {"type": "DAY", "expirationMs": "100"}, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(table.time_partitioning.type_, "DAY") self.assertEqual(table.time_partitioning.expiration_ms, 100) self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_w_schema_and_query(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) query = "SELECT * from %s:%s" % (self.DS_ID, self.TABLE_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() resource.update( { "schema": { "fields": [ {"name": "full_name", "type": "STRING", "mode": "REQUIRED"}, {"name": "age", "type": "INTEGER", "mode": "REQUIRED"}, ] }, "view": {"query": query}, } ) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), ] conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF, schema=schema) table.view_query = query with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(table) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": table.dataset_id}, client, None ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "schema": { "fields": [ {"name": "full_name", "type": "STRING", "mode": "REQUIRED"}, {"name": "age", "type": "INTEGER", "mode": "REQUIRED"}, ] }, "view": {"query": query, "useLegacySql": False}, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got.table_id, self.TABLE_ID) self.assertEqual(got.project, self.PROJECT) self.assertEqual(got.dataset_id, self.DS_ID) self.assertEqual(got.schema, schema) self.assertEqual(got.view_query, query) def test_create_table_w_external(self): from google.cloud.bigquery.external_config import ExternalConfig from google.cloud.bigquery.job import SourceFormat from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() resource.update( { "externalDataConfiguration": { "sourceFormat": SourceFormat.CSV, "autodetect": True, } } ) conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) ec = ExternalConfig("CSV") ec.autodetect = True table.external_data_configuration = ec with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(table) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": table.dataset_id}, client, None ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "externalDataConfiguration": { "sourceFormat": SourceFormat.CSV, "autodetect": True, }, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got.table_id, self.TABLE_ID) self.assertEqual(got.project, self.PROJECT) self.assertEqual(got.dataset_id, self.DS_ID) self.assertEqual( got.external_data_configuration.source_format, SourceFormat.CSV ) self.assertEqual(got.external_data_configuration.autodetect, True) def test_create_table_w_reference(self): path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table(self.TABLE_REF) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": self.TABLE_REF.dataset_id}, client, None, ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_w_fully_qualified_string(self): path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table( "{}.{}.{}".format(self.PROJECT, self.DS_ID, self.TABLE_ID) ) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": self.TABLE_REF.dataset_id}, client, None, ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_w_string(self): path = "projects/%s/datasets/%s/tables" % (self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) resource = self._make_table_resource() conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table("{}.{}".format(self.DS_ID, self.TABLE_ID)) final_attributes.assert_called_once_with( {"path": "/%s" % path, "dataset_id": self.TABLE_REF.dataset_id}, client, None, ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(got.table_id, self.TABLE_ID) def test_create_table_alreadyexists_w_exists_ok_false(self): post_path = "/projects/{}/datasets/{}/tables".format(self.PROJECT, self.DS_ID) creds = _make_credentials() client = self._make_one( project=self.PROJECT, credentials=creds, location=self.LOCATION ) conn = client._connection = make_connection( google.api_core.exceptions.AlreadyExists("table already exists") ) with pytest.raises(google.api_core.exceptions.AlreadyExists): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.create_table("{}.{}".format(self.DS_ID, self.TABLE_ID)) final_attributes.assert_called_with( {"path": post_path, "dataset_id": self.TABLE_REF.dataset_id}, client, None, ) conn.api_request.assert_called_once_with( method="POST", path=post_path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ) def test_create_table_alreadyexists_w_exists_ok_true(self): post_path = "/projects/{}/datasets/{}/tables".format(self.PROJECT, self.DS_ID) get_path = "/projects/{}/datasets/{}/tables/{}".format( self.PROJECT, self.DS_ID, self.TABLE_ID ) resource = self._make_table_resource() creds = _make_credentials() client = self._make_one( project=self.PROJECT, credentials=creds, location=self.LOCATION ) conn = client._connection = make_connection( google.api_core.exceptions.AlreadyExists("table already exists"), resource ) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.create_table( "{}.{}".format(self.DS_ID, self.TABLE_ID), exists_ok=True ) final_attributes.assert_called_with({"path": get_path}, client, None) self.assertEqual(got.project, self.PROJECT) self.assertEqual(got.dataset_id, self.DS_ID) self.assertEqual(got.table_id, self.TABLE_ID) conn.api_request.assert_has_calls( [ mock.call( method="POST", path=post_path, data={ "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, "labels": {}, }, timeout=DEFAULT_TIMEOUT, ), mock.call(method="GET", path=get_path, timeout=DEFAULT_TIMEOUT), ] ) def test_close(self): creds = _make_credentials() http = mock.Mock() http._auth_request.session = mock.Mock() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client.close() http.close.assert_called_once() http._auth_request.session.close.assert_called_once() def test_get_model(self): path = "projects/%s/datasets/%s/models/%s" % ( self.PROJECT, self.DS_ID, self.MODEL_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) resource = { "modelReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": self.MODEL_ID, } } conn = client._connection = make_connection(resource) model_ref = DatasetReference(self.PROJECT, self.DS_ID).model(self.MODEL_ID) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.get_model(model_ref, timeout=7.5) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) conn.api_request.assert_called_once_with( method="GET", path="/%s" % path, timeout=7.5 ) self.assertEqual(got.model_id, self.MODEL_ID) def test_get_model_w_string(self): path = "projects/%s/datasets/%s/models/%s" % ( self.PROJECT, self.DS_ID, self.MODEL_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) resource = { "modelReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": self.MODEL_ID, } } conn = client._connection = make_connection(resource) model_id = "{}.{}.{}".format(self.PROJECT, self.DS_ID, self.MODEL_ID) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: got = client.get_model(model_id) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) conn.api_request.assert_called_once_with( method="GET", path="/%s" % path, timeout=DEFAULT_TIMEOUT ) self.assertEqual(got.model_id, self.MODEL_ID) def test_get_routine(self): from google.cloud.bigquery.routine import Routine from google.cloud.bigquery.routine import RoutineReference full_routine_id = "test-routine-project.test_routines.minimal_routine" routines = [ full_routine_id, Routine(full_routine_id), RoutineReference.from_string(full_routine_id), ] for routine in routines: creds = _make_credentials() resource = { "etag": "im-an-etag", "routineReference": { "projectId": "test-routine-project", "datasetId": "test_routines", "routineId": "minimal_routine", }, "routineType": "SCALAR_FUNCTION", } path = "/projects/test-routine-project/datasets/test_routines/routines/minimal_routine" client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: actual_routine = client.get_routine(routine, timeout=7.5) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_once_with( method="GET", path=path, timeout=7.5, ) self.assertEqual( actual_routine.reference, RoutineReference.from_string(full_routine_id), msg="routine={}".format(repr(routine)), ) self.assertEqual( actual_routine.etag, "im-an-etag", msg="routine={}".format(repr(routine)), ) self.assertEqual( actual_routine.type_, "SCALAR_FUNCTION", msg="routine={}".format(repr(routine)), ) def test_get_table(self): path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) resource = self._make_table_resource() conn = client._connection = make_connection(resource) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: table = client.get_table(self.TABLE_REF, timeout=7.5) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) conn.api_request.assert_called_once_with( method="GET", path="/%s" % path, timeout=7.5 ) self.assertEqual(table.table_id, self.TABLE_ID) def test_get_table_sets_user_agent(self): creds = _make_credentials() http = mock.create_autospec(requests.Session) mock_response = http.request( url=mock.ANY, method=mock.ANY, headers=mock.ANY, data=mock.ANY ) http.reset_mock() http.is_mtls = False mock_response.status_code = 200 mock_response.json.return_value = self._make_table_resource() user_agent_override = client_info.ClientInfo(user_agent="my-application/1.2.3") client = self._make_one( project=self.PROJECT, credentials=creds, client_info=user_agent_override, _http=http, ) client.get_table(self.TABLE_REF) expected_user_agent = user_agent_override.to_user_agent() http.request.assert_called_once_with( url=mock.ANY, method="GET", headers={ "X-Goog-API-Client": expected_user_agent, "Accept-Encoding": "gzip", "User-Agent": expected_user_agent, }, data=mock.ANY, timeout=DEFAULT_TIMEOUT, ) self.assertIn("my-application/1.2.3", expected_user_agent) def test_get_iam_policy(self): from google.cloud.bigquery.iam import BIGQUERY_DATA_OWNER_ROLE from google.cloud.bigquery.iam import BIGQUERY_DATA_EDITOR_ROLE from google.cloud.bigquery.iam import BIGQUERY_DATA_VIEWER_ROLE from google.api_core.iam import Policy PATH = "/projects/{}/datasets/{}/tables/{}:getIamPolicy".format( self.PROJECT, self.DS_ID, self.TABLE_ID, ) BODY = {"options": {"requestedPolicyVersion": 1}} ETAG = "CARDI" VERSION = 1 OWNER1 = "user:phred@example.com" OWNER2 = "group:cloud-logs@google.com" EDITOR1 = "domain:google.com" EDITOR2 = "user:phred@example.com" VIEWER1 = "serviceAccount:1234-abcdef@service.example.com" VIEWER2 = "user:phred@example.com" RETURNED = { "resourceId": PATH, "etag": ETAG, "version": VERSION, "bindings": [ {"role": BIGQUERY_DATA_OWNER_ROLE, "members": [OWNER1, OWNER2]}, {"role": BIGQUERY_DATA_EDITOR_ROLE, "members": [EDITOR1, EDITOR2]}, {"role": BIGQUERY_DATA_VIEWER_ROLE, "members": [VIEWER1, VIEWER2]}, ], } EXPECTED = { binding["role"]: set(binding["members"]) for binding in RETURNED["bindings"] } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RETURNED) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: policy = client.get_iam_policy(self.TABLE_REF, timeout=7.5) final_attributes.assert_called_once_with({"path": PATH}, client, None) conn.api_request.assert_called_once_with( method="POST", path=PATH, data=BODY, timeout=7.5 ) self.assertIsInstance(policy, Policy) self.assertEqual(policy.etag, RETURNED["etag"]) self.assertEqual(policy.version, RETURNED["version"]) self.assertEqual(dict(policy), EXPECTED) def test_get_iam_policy_w_invalid_table(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) table_resource_string = "projects/{}/datasets/{}/tables/{}".format( self.PROJECT, self.DS_ID, self.TABLE_ID, ) with self.assertRaises(ValueError): client.get_iam_policy(table_resource_string) def test_get_iam_policy_w_invalid_version(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) with self.assertRaises(ValueError): client.get_iam_policy(self.TABLE_REF, requested_policy_version=2) def test_set_iam_policy(self): from google.cloud.bigquery.iam import BIGQUERY_DATA_OWNER_ROLE from google.cloud.bigquery.iam import BIGQUERY_DATA_EDITOR_ROLE from google.cloud.bigquery.iam import BIGQUERY_DATA_VIEWER_ROLE from google.api_core.iam import Policy PATH = "/projects/%s/datasets/%s/tables/%s:setIamPolicy" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) ETAG = "foo" VERSION = 1 OWNER1 = "user:phred@example.com" OWNER2 = "group:cloud-logs@google.com" EDITOR1 = "domain:google.com" EDITOR2 = "user:phred@example.com" VIEWER1 = "serviceAccount:1234-abcdef@service.example.com" VIEWER2 = "user:phred@example.com" BINDINGS = [ {"role": BIGQUERY_DATA_OWNER_ROLE, "members": [OWNER1, OWNER2]}, {"role": BIGQUERY_DATA_EDITOR_ROLE, "members": [EDITOR1, EDITOR2]}, {"role": BIGQUERY_DATA_VIEWER_ROLE, "members": [VIEWER1, VIEWER2]}, ] FIELDS = ("bindings", "etag") RETURNED = {"etag": ETAG, "version": VERSION, "bindings": BINDINGS} policy = Policy() for binding in BINDINGS: policy[binding["role"]] = binding["members"] BODY = {"policy": policy.to_api_repr(), "updateMask": "bindings,etag"} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RETURNED) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: returned_policy = client.set_iam_policy( self.TABLE_REF, policy, fields=FIELDS, timeout=7.5 ) final_attributes.assert_called_once_with({"path": PATH}, client, None) conn.api_request.assert_called_once_with( method="POST", path=PATH, data=BODY, timeout=7.5 ) self.assertEqual(returned_policy.etag, ETAG) self.assertEqual(returned_policy.version, VERSION) self.assertEqual(dict(returned_policy), dict(policy)) def test_set_iam_policy_updateMask(self): from google.cloud.bigquery.iam import BIGQUERY_DATA_OWNER_ROLE from google.cloud.bigquery.iam import BIGQUERY_DATA_EDITOR_ROLE from google.cloud.bigquery.iam import BIGQUERY_DATA_VIEWER_ROLE from google.api_core.iam import Policy PATH = "/projects/%s/datasets/%s/tables/%s:setIamPolicy" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) ETAG = "foo" VERSION = 1 OWNER1 = "user:phred@example.com" OWNER2 = "group:cloud-logs@google.com" EDITOR1 = "domain:google.com" EDITOR2 = "user:phred@example.com" VIEWER1 = "serviceAccount:1234-abcdef@service.example.com" VIEWER2 = "user:phred@example.com" BINDINGS = [ {"role": BIGQUERY_DATA_OWNER_ROLE, "members": [OWNER1, OWNER2]}, {"role": BIGQUERY_DATA_EDITOR_ROLE, "members": [EDITOR1, EDITOR2]}, {"role": BIGQUERY_DATA_VIEWER_ROLE, "members": [VIEWER1, VIEWER2]}, ] MASK = "bindings,etag" RETURNED = {"etag": ETAG, "version": VERSION, "bindings": BINDINGS} policy = Policy() for binding in BINDINGS: policy[binding["role"]] = binding["members"] BODY = {"policy": policy.to_api_repr(), "updateMask": MASK} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RETURNED) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: returned_policy = client.set_iam_policy( self.TABLE_REF, policy, updateMask=MASK, timeout=7.5 ) final_attributes.assert_called_once_with({"path": PATH}, client, None) conn.api_request.assert_called_once_with( method="POST", path=PATH, data=BODY, timeout=7.5 ) self.assertEqual(returned_policy.etag, ETAG) self.assertEqual(returned_policy.version, VERSION) self.assertEqual(dict(returned_policy), dict(policy)) def test_set_iam_policy_no_mask(self): from google.api_core.iam import Policy PATH = "/projects/%s/datasets/%s/tables/%s:setIamPolicy" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) RETURNED = {"etag": "foo", "version": 1, "bindings": []} policy = Policy() BODY = {"policy": policy.to_api_repr()} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RETURNED) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.set_iam_policy(self.TABLE_REF, policy, timeout=7.5) final_attributes.assert_called_once_with({"path": PATH}, client, None) conn.api_request.assert_called_once_with( method="POST", path=PATH, data=BODY, timeout=7.5 ) def test_set_ia_policy_updateMask_and_fields(self): from google.api_core.iam import Policy policy = Policy() creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) with pytest.raises(ValueError, match="updateMask"): client.set_iam_policy( self.TABLE_REF, policy, updateMask="bindings", fields=("bindings",) ) def test_set_iam_policy_invalid_policy(self): from google.api_core.iam import Policy policy = Policy() invalid_policy_repr = policy.to_api_repr() creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) with self.assertRaises(TypeError): client.set_iam_policy(self.TABLE_REF, invalid_policy_repr) def test_set_iam_policy_w_invalid_table(self): from google.api_core.iam import Policy policy = Policy() creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) table_resource_string = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) with self.assertRaises(ValueError): client.set_iam_policy(table_resource_string, policy) def test_test_iam_permissions(self): PATH = "/projects/%s/datasets/%s/tables/%s:testIamPermissions" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) PERMISSIONS = ["bigquery.tables.get", "bigquery.tables.update"] BODY = {"permissions": PERMISSIONS} RETURNED = {"permissions": PERMISSIONS} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RETURNED) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.test_iam_permissions(self.TABLE_REF, PERMISSIONS, timeout=7.5) final_attributes.assert_called_once_with({"path": PATH}, client, None) conn.api_request.assert_called_once_with( method="POST", path=PATH, data=BODY, timeout=7.5 ) def test_test_iam_permissions_w_invalid_table(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) table_resource_string = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) PERMISSIONS = ["bigquery.tables.get", "bigquery.tables.update"] with self.assertRaises(ValueError): client.test_iam_permissions(table_resource_string, PERMISSIONS) def test_update_dataset_w_invalid_field(self): from google.cloud.bigquery.dataset import Dataset creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) with self.assertRaises(ValueError): client.update_dataset( Dataset("{}.{}".format(self.PROJECT, self.DS_ID)), ["foo"] ) def test_update_dataset(self): from google.cloud.bigquery.dataset import Dataset, AccessEntry PATH = "projects/%s/datasets/%s" % (self.PROJECT, self.DS_ID) DESCRIPTION = "DESCRIPTION" FRIENDLY_NAME = "TITLE" LOCATION = "loc" LABELS = {"priority": "high"} ACCESS = [{"role": "OWNER", "userByEmail": "phred@example.com"}] EXP = 17 RESOURCE = { "datasetReference": {"projectId": self.PROJECT, "datasetId": self.DS_ID}, "etag": "etag", "description": DESCRIPTION, "friendlyName": FRIENDLY_NAME, "location": LOCATION, "defaultTableExpirationMs": EXP, "labels": LABELS, "access": ACCESS, } creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(RESOURCE, RESOURCE) ds = Dataset(DatasetReference(self.PROJECT, self.DS_ID)) ds.description = DESCRIPTION ds.friendly_name = FRIENDLY_NAME ds.location = LOCATION ds.default_table_expiration_ms = EXP ds.labels = LABELS ds.access_entries = [AccessEntry("OWNER", "userByEmail", "phred@example.com")] fields = [ "description", "friendly_name", "location", "labels", "access_entries", ] with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: ds2 = client.update_dataset( ds, fields=fields, timeout=7.5, ) final_attributes.assert_called_once_with( {"path": "/%s" % PATH, "fields": fields}, client, None ) conn.api_request.assert_called_once_with( method="PATCH", data={ "description": DESCRIPTION, "friendlyName": FRIENDLY_NAME, "location": LOCATION, "labels": LABELS, "access": ACCESS, }, path="/" + PATH, timeout=7.5, ) self.assertEqual(ds2.description, ds.description) self.assertEqual(ds2.friendly_name, ds.friendly_name) self.assertEqual(ds2.location, ds.location) self.assertEqual(ds2.labels, ds.labels) self.assertEqual(ds2.access_entries, ds.access_entries) # ETag becomes If-Match header. ds._properties["etag"] = "etag" client.update_dataset(ds, []) req = conn.api_request.call_args self.assertEqual(req[1]["headers"]["If-Match"], "etag") def test_update_dataset_w_custom_property(self): # The library should handle sending properties to the API that are not # yet part of the library from google.cloud.bigquery.dataset import Dataset path = "/projects/%s/datasets/%s" % (self.PROJECT, self.DS_ID) resource = { "datasetReference": {"projectId": self.PROJECT, "datasetId": self.DS_ID}, "newAlphaProperty": "unreleased property", } creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource) dataset = Dataset(DatasetReference(self.PROJECT, self.DS_ID)) dataset._properties["newAlphaProperty"] = "unreleased property" with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: dataset = client.update_dataset(dataset, ["newAlphaProperty"]) final_attributes.assert_called_once_with( {"path": path, "fields": ["newAlphaProperty"]}, client, None ) conn.api_request.assert_called_once_with( method="PATCH", data={"newAlphaProperty": "unreleased property"}, path=path, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(dataset.dataset_id, self.DS_ID) self.assertEqual(dataset.project, self.PROJECT) self.assertEqual(dataset._properties["newAlphaProperty"], "unreleased property") def test_update_model(self): from google.cloud.bigquery.model import Model path = "projects/%s/datasets/%s/models/%s" % ( self.PROJECT, self.DS_ID, self.MODEL_ID, ) description = "description" title = "title" expires = datetime.datetime( 2012, 12, 21, 16, 0, 0, tzinfo=google.cloud._helpers.UTC ) resource = { "modelReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": self.MODEL_ID, }, "description": description, "etag": "etag", "expirationTime": str(google.cloud._helpers._millis(expires)), "friendlyName": title, "labels": {"x": "y"}, } creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource, resource) model_id = "{}.{}.{}".format(self.PROJECT, self.DS_ID, self.MODEL_ID) model = Model(model_id) model.description = description model.friendly_name = title model.expires = expires model.labels = {"x": "y"} fields = ["description", "friendly_name", "labels", "expires"] with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: updated_model = client.update_model(model, fields, timeout=7.5) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": fields}, client, None ) sent = { "description": description, "expirationTime": str(google.cloud._helpers._millis(expires)), "friendlyName": title, "labels": {"x": "y"}, } conn.api_request.assert_called_once_with( method="PATCH", data=sent, path="/" + path, timeout=7.5 ) self.assertEqual(updated_model.model_id, model.model_id) self.assertEqual(updated_model.description, model.description) self.assertEqual(updated_model.friendly_name, model.friendly_name) self.assertEqual(updated_model.labels, model.labels) self.assertEqual(updated_model.expires, model.expires) # ETag becomes If-Match header. model._properties["etag"] = "etag" client.update_model(model, []) req = conn.api_request.call_args self.assertEqual(req[1]["headers"]["If-Match"], "etag") def test_update_routine(self): from google.cloud.bigquery.routine import Routine from google.cloud.bigquery.routine import RoutineArgument full_routine_id = "routines-project.test_routines.updated_routine" resource = { "routineReference": { "projectId": "routines-project", "datasetId": "test_routines", "routineId": "updated_routine", }, "routineType": "SCALAR_FUNCTION", "language": "SQL", "definitionBody": "x * 3", "arguments": [{"name": "x", "dataType": {"typeKind": "INT64"}}], "returnType": None, "someNewField": "someValue", } creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource, resource) routine = Routine(full_routine_id) routine.arguments = [ RoutineArgument( name="x", data_type=bigquery.standard_sql.StandardSqlDataType( type_kind=bigquery.StandardSqlTypeNames.INT64 ), ) ] routine.body = "x * 3" routine.language = "SQL" routine.type_ = "SCALAR_FUNCTION" routine._properties["someNewField"] = "someValue" fields = [ "arguments", "language", "body", "type_", "return_type", "someNewField", ] with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: actual_routine = client.update_routine( routine, fields, timeout=7.5, ) final_attributes.assert_called_once_with( {"path": routine.path, "fields": fields}, client, None ) # TODO: routineReference isn't needed when the Routines API supports # partial updates. sent = resource conn.api_request.assert_called_once_with( method="PUT", data=sent, path="/projects/routines-project/datasets/test_routines/routines/updated_routine", timeout=7.5, ) self.assertEqual(actual_routine.arguments, routine.arguments) self.assertEqual(actual_routine.body, routine.body) self.assertEqual(actual_routine.language, routine.language) self.assertEqual(actual_routine.type_, routine.type_) # ETag becomes If-Match header. routine._properties["etag"] = "im-an-etag" with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.update_routine(routine, []) final_attributes.assert_called_once_with( {"path": routine.path, "fields": []}, client, None ) req = conn.api_request.call_args self.assertEqual(req[1]["headers"]["If-Match"], "im-an-etag") def test_update_table(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.schema import PolicyTagList from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) description = "description" title = "title" resource = self._make_table_resource() resource.update( { "schema": { "fields": [ { "name": "full_name", "type": "STRING", "mode": "REQUIRED", "description": None, "policyTags": {"names": []}, }, { "name": "age", "type": "INTEGER", "mode": "REQUIRED", "description": "New field description", }, ] }, "etag": "etag", "description": description, "friendlyName": title, "labels": {"x": "y"}, } ) schema = [ # Explicly setting policyTags to no names should be included in the sent resource. # https://github.com/googleapis/python-bigquery/issues/981 SchemaField( "full_name", "STRING", mode="REQUIRED", description=None, policy_tags=PolicyTagList(names=()), ), SchemaField( "age", "INTEGER", mode="REQUIRED", description="New field description" ), ] creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource, resource) table = Table(self.TABLE_REF, schema=schema) table.description = description table.friendly_name = title table.labels = {"x": "y"} fields = ["schema", "description", "friendly_name", "labels"] with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: updated_table = client.update_table(table, fields, timeout=7.5) span_path = "/%s" % path final_attributes.assert_called_once_with( {"path": span_path, "fields": fields}, client, None ) sent = { "schema": { "fields": [ { "name": "full_name", "type": "STRING", "mode": "REQUIRED", "description": None, "policyTags": {"names": []}, }, { "name": "age", "type": "INTEGER", "mode": "REQUIRED", "description": "New field description", }, ] }, "description": description, "friendlyName": title, "labels": {"x": "y"}, } conn.api_request.assert_called_once_with( method="PATCH", data=sent, path="/" + path, timeout=7.5 ) self.assertEqual(updated_table.description, table.description) self.assertEqual(updated_table.friendly_name, table.friendly_name) self.assertEqual(updated_table.schema, table.schema) self.assertEqual(updated_table.labels, table.labels) # ETag becomes If-Match header. table._properties["etag"] = "etag" with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.update_table(table, []) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": []}, client, None ) req = conn.api_request.call_args self.assertEqual(req[1]["headers"]["If-Match"], "etag") def test_update_table_w_custom_property(self): from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) resource = self._make_table_resource() resource["newAlphaProperty"] = "unreleased property" creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) table._properties["newAlphaProperty"] = "unreleased property" with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: updated_table = client.update_table(table, ["newAlphaProperty"]) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": ["newAlphaProperty"]}, client, None, ) conn.api_request.assert_called_once_with( method="PATCH", path="/%s" % path, data={"newAlphaProperty": "unreleased property"}, timeout=DEFAULT_TIMEOUT, ) self.assertEqual( updated_table._properties["newAlphaProperty"], "unreleased property" ) def test_update_table_only_use_legacy_sql(self): from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) resource = self._make_table_resource() resource["view"] = {"useLegacySql": True} creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF) table.view_use_legacy_sql = True with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: updated_table = client.update_table(table, ["view_use_legacy_sql"]) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": ["view_use_legacy_sql"]}, client, None, ) conn.api_request.assert_called_once_with( method="PATCH", path="/%s" % path, data={"view": {"useLegacySql": True}}, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(updated_table.view_use_legacy_sql, table.view_use_legacy_sql) def test_update_table_w_query(self): import datetime from google.cloud._helpers import UTC from google.cloud._helpers import _millis from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) query = "select fullname, age from person_ages" location = "EU" exp_time = datetime.datetime(2015, 8, 1, 23, 59, 59, tzinfo=UTC) schema_resource = { "fields": [ { "name": "full_name", "type": "STRING", "mode": "REQUIRED", "description": None, }, { "name": "age", "type": "INTEGER", "mode": "REQUIRED", "description": "this is a column", }, {"name": "country", "type": "STRING", "mode": "NULLABLE"}, ] } schema = [ SchemaField( "full_name", "STRING", mode="REQUIRED", # Explicitly unset the description. description=None, ), SchemaField( "age", "INTEGER", mode="REQUIRED", description="this is a column" ), # Omit the description to not make updates to it. SchemaField("country", "STRING"), ] resource = self._make_table_resource() resource.update( { "schema": schema_resource, "view": {"query": query, "useLegacySql": True}, "location": location, "expirationTime": _millis(exp_time), } ) creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource) table = Table(self.TABLE_REF, schema=schema) table.expires = exp_time table.view_query = query table.view_use_legacy_sql = True updated_properties = ["schema", "view_query", "expires", "view_use_legacy_sql"] with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: updated_table = client.update_table(table, updated_properties) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": updated_properties}, client, None, ) self.assertEqual(updated_table.schema, table.schema) self.assertEqual(updated_table.view_query, table.view_query) self.assertEqual(updated_table.expires, table.expires) self.assertEqual(updated_table.view_use_legacy_sql, table.view_use_legacy_sql) self.assertEqual(updated_table.location, location) conn.api_request.assert_called_once_with( method="PATCH", path="/%s" % path, data={ "view": {"query": query, "useLegacySql": True}, "expirationTime": str(_millis(exp_time)), "schema": schema_resource, }, timeout=DEFAULT_TIMEOUT, ) def test_update_table_w_schema_None(self): # Simulate deleting schema: not sure if back-end will actually # allow this operation, but the spec says it is optional. path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) resource1 = self._make_table_resource() resource1.update( { "schema": { "fields": [ {"name": "full_name", "type": "STRING", "mode": "REQUIRED"}, {"name": "age", "type": "INTEGER", "mode": "REQUIRED"}, ] } } ) resource2 = self._make_table_resource() creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource1, resource2) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: table = client.get_table( # Test with string for table ID "{}.{}.{}".format( self.TABLE_REF.project, self.TABLE_REF.dataset_id, self.TABLE_REF.table_id, ) ) final_attributes.assert_called_once_with({"path": "/%s" % path}, client, None) table.schema = None with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: updated_table = client.update_table(table, ["schema"]) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": ["schema"]}, client, None ) self.assertEqual(len(conn.api_request.call_args_list), 2) req = conn.api_request.call_args_list[1] self.assertEqual(req[1]["method"], "PATCH") sent = {"schema": None} self.assertEqual(req[1]["data"], sent) self.assertEqual(req[1]["path"], "/%s" % path) self.assertEqual(len(updated_table.schema), 0) def test_update_table_delete_property(self): from google.cloud.bigquery.table import Table description = "description" title = "title" path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) resource1 = self._make_table_resource() resource1.update({"description": description, "friendlyName": title}) resource2 = self._make_table_resource() resource2["description"] = None creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) conn = client._connection = make_connection(resource1, resource2) table = Table(self.TABLE_REF) table.description = description table.friendly_name = title with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: table2 = client.update_table(table, ["description", "friendly_name"]) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": ["description", "friendly_name"]}, client, None, ) self.assertEqual(table2.description, table.description) table2.description = None with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: table3 = client.update_table(table2, ["description"]) final_attributes.assert_called_once_with( {"path": "/%s" % path, "fields": ["description"]}, client, None ) self.assertEqual(len(conn.api_request.call_args_list), 2) req = conn.api_request.call_args_list[1] self.assertEqual(req[1]["method"], "PATCH") self.assertEqual(req[1]["path"], "/%s" % path) sent = {"description": None} self.assertEqual(req[1]["data"], sent) self.assertIsNone(table3.description) def test_delete_job_metadata_not_found(self): creds = _make_credentials() client = self._make_one("client-proj", creds, location="client-loc") conn = client._connection = make_connection( google.api_core.exceptions.NotFound("job not found"), google.api_core.exceptions.NotFound("job not found"), ) with self.assertRaises(google.api_core.exceptions.NotFound): client.delete_job_metadata("my-job") conn.api_request.reset_mock() client.delete_job_metadata("my-job", not_found_ok=True) conn.api_request.assert_called_once_with( method="DELETE", path="/projects/client-proj/jobs/my-job/delete", query_params={"location": "client-loc"}, timeout=DEFAULT_TIMEOUT, ) def test_delete_job_metadata_with_id(self): creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection({}) client.delete_job_metadata("my-job", project="param-proj", location="param-loc") conn.api_request.assert_called_once_with( method="DELETE", path="/projects/param-proj/jobs/my-job/delete", query_params={"location": "param-loc"}, timeout=DEFAULT_TIMEOUT, ) def test_delete_job_metadata_with_resource(self): from google.cloud.bigquery.job import QueryJob query_resource = { "jobReference": { "projectId": "job-based-proj", "jobId": "query_job", "location": "us-east1", }, "configuration": {"query": {}}, } creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection(query_resource) job_from_resource = QueryJob.from_api_repr(query_resource, client) client.delete_job_metadata(job_from_resource) conn.api_request.assert_called_once_with( method="DELETE", path="/projects/job-based-proj/jobs/query_job/delete", query_params={"location": "us-east1"}, timeout=DEFAULT_TIMEOUT, ) def test_delete_model(self): from google.cloud.bigquery.model import Model path = "projects/%s/datasets/%s/models/%s" % ( self.PROJECT, self.DS_ID, self.MODEL_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) model_id = "{}.{}.{}".format(self.PROJECT, self.DS_ID, self.MODEL_ID) models = ( model_id, DatasetReference(self.PROJECT, self.DS_ID).model(self.MODEL_ID), Model(model_id), ) conn = client._connection = make_connection(*([{}] * len(models))) for arg in models: with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_model(arg, timeout=7.5) final_attributes.assert_called_once_with( {"path": "/%s" % path}, client, None ) conn.api_request.assert_called_with( method="DELETE", path="/%s" % path, timeout=7.5 ) def test_delete_model_w_wrong_type(self): creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) with self.assertRaises(TypeError): client.delete_model(DatasetReference(self.PROJECT, self.DS_ID)) def test_delete_model_w_not_found_ok_false(self): path = "/projects/{}/datasets/{}/models/{}".format( self.PROJECT, self.DS_ID, self.MODEL_ID ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( google.api_core.exceptions.NotFound("model not found") ) with self.assertRaises(google.api_core.exceptions.NotFound): client.delete_model("{}.{}".format(self.DS_ID, self.MODEL_ID)) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=DEFAULT_TIMEOUT ) def test_delete_model_w_not_found_ok_true(self): path = "/projects/{}/datasets/{}/models/{}".format( self.PROJECT, self.DS_ID, self.MODEL_ID ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( google.api_core.exceptions.NotFound("model not found") ) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_model( "{}.{}".format(self.DS_ID, self.MODEL_ID), not_found_ok=True ) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=DEFAULT_TIMEOUT ) def test_delete_routine(self): from google.cloud.bigquery.routine import Routine from google.cloud.bigquery.routine import RoutineReference full_routine_id = "test-routine-project.test_routines.minimal_routine" routines = [ full_routine_id, Routine(full_routine_id), RoutineReference.from_string(full_routine_id), ] creds = _make_credentials() http = object() path = "/projects/test-routine-project/datasets/test_routines/routines/minimal_routine" client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(*([{}] * len(routines))) for routine in routines: with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_routine(routine, timeout=7.5) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=7.5, ) def test_delete_routine_w_wrong_type(self): creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) with self.assertRaises(TypeError): client.delete_routine(DatasetReference(self.PROJECT, self.DS_ID)) def test_delete_routine_w_not_found_ok_false(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( google.api_core.exceptions.NotFound("routine not found") ) path = "/projects/routines-project/datasets/test_routines/routines/test_routine" with self.assertRaises(google.api_core.exceptions.NotFound): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_routine("routines-project.test_routines.test_routine") final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=DEFAULT_TIMEOUT, ) def test_delete_routine_w_not_found_ok_true(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( google.api_core.exceptions.NotFound("routine not found") ) path = "/projects/routines-project/datasets/test_routines/routines/test_routine" with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_routine( "routines-project.test_routines.test_routine", not_found_ok=True ) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=DEFAULT_TIMEOUT, ) def test_delete_table(self): from google.cloud.bigquery.table import Table tables = ( self.TABLE_REF, Table(self.TABLE_REF), "{}.{}.{}".format( self.TABLE_REF.project, self.TABLE_REF.dataset_id, self.TABLE_REF.table_id, ), ) path = "projects/%s/datasets/%s/tables/%s" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(*([{}] * len(tables))) for arg in tables: with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_table(arg, timeout=7.5) final_attributes.assert_called_once_with( {"path": "/%s" % path}, client, None ) conn.api_request.assert_called_with( method="DELETE", path="/%s" % path, timeout=7.5 ) def test_delete_table_w_wrong_type(self): creds = _make_credentials() client = self._make_one(project=self.PROJECT, credentials=creds) with self.assertRaises(TypeError): client.delete_table(DatasetReference(self.PROJECT, self.DS_ID)) def test_delete_table_w_not_found_ok_false(self): path = "/projects/{}/datasets/{}/tables/{}".format( self.PROJECT, self.DS_ID, self.TABLE_ID ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( google.api_core.exceptions.NotFound("table not found") ) with self.assertRaises(google.api_core.exceptions.NotFound): with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_table("{}.{}".format(self.DS_ID, self.TABLE_ID)) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=DEFAULT_TIMEOUT ) def test_delete_table_w_not_found_ok_true(self): path = "/projects/{}/datasets/{}/tables/{}".format( self.PROJECT, self.DS_ID, self.TABLE_ID ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( google.api_core.exceptions.NotFound("table not found") ) with mock.patch( "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" ) as final_attributes: client.delete_table( "{}.{}".format(self.DS_ID, self.TABLE_ID), not_found_ok=True ) final_attributes.assert_called_once_with({"path": path}, client, None) conn.api_request.assert_called_with( method="DELETE", path=path, timeout=DEFAULT_TIMEOUT ) def _create_job_helper(self, job_config): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) resource = { "jobReference": {"projectId": self.PROJECT, "jobId": "random-id"}, "configuration": job_config, } expected = { "jobReference": {"projectId": self.PROJECT, "jobId": mock.ANY}, "configuration": job_config, } conn = client._connection = make_connection(resource) client.create_job(job_config=job_config) conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=expected, timeout=DEFAULT_TIMEOUT, ) def test_create_job_load_config(self): configuration = { "load": { "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "source_table", }, "sourceUris": ["gs://test_bucket/src_object*"], } } self._create_job_helper(configuration) def test_create_job_copy_config(self): configuration = { "copy": { "sourceTables": [ { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "source_table", } ], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "destination_table", }, } } self._create_job_helper(configuration) def test_create_job_copy_config_w_single_source(self): configuration = { "copy": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "source_table", }, "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "destination_table", }, } } self._create_job_helper(configuration) def test_create_job_extract_config(self): configuration = { "extract": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "source_table", }, "destinationUris": ["gs://test_bucket/dst_object*"], } } self._create_job_helper(configuration) def test_create_job_extract_config_for_model(self): configuration = { "extract": { "sourceModel": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": "source_model", }, "destinationUris": ["gs://test_bucket/dst_object*"], } } self._create_job_helper(configuration) def test_create_job_query_config(self): configuration = { "query": { "query": "query", "destinationTable": {"tableId": "table_id"}, "useLegacySql": False, } } self._create_job_helper(configuration) def test_create_job_query_config_w_rateLimitExceeded_error(self): from google.cloud.exceptions import Forbidden from google.cloud.bigquery.retry import DEFAULT_RETRY query = "select count(*) from persons" configuration = { "query": { "query": query, "useLegacySql": False, "destinationTable": {"tableId": "table_id"}, } } resource = { "jobReference": {"projectId": self.PROJECT, "jobId": mock.ANY}, "configuration": { "query": { "query": query, "useLegacySql": False, "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": "query_destination_table", }, } }, } data_without_destination = { "jobReference": {"projectId": self.PROJECT, "jobId": mock.ANY}, "configuration": configuration, } creds = _make_credentials() http = object() retry = DEFAULT_RETRY.with_deadline(1).with_predicate( lambda exc: isinstance(exc, Forbidden) ) client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) api_request_patcher = mock.patch.object( client._connection, "api_request", side_effect=[ Forbidden("", errors=[{"reason": "rateLimitExceeded"}]), resource, ], ) with api_request_patcher as fake_api_request: job = client.create_job(job_config=configuration, retry=retry) self.assertEqual(job.destination.table_id, "query_destination_table") self.assertEqual(len(fake_api_request.call_args_list), 2) # was retried once self.assertEqual( fake_api_request.call_args_list[1], mock.call( method="POST", path="/projects/PROJECT/jobs", data=data_without_destination, timeout=DEFAULT_TIMEOUT, ), ) def test_create_job_w_invalid_job_config(self): configuration = {"unknown": {}} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) with self.assertRaises(TypeError) as exc: client.create_job(job_config=configuration) self.assertIn("Invalid job configuration", exc.exception.args[0]) def test_job_from_resource_unknown_type(self): from google.cloud.bigquery.job import UnknownJob creds = _make_credentials() client = self._make_one(self.PROJECT, creds) got = client.job_from_resource({}) # Can parse redacted job. self.assertIsInstance(got, UnknownJob) self.assertEqual(got.project, self.PROJECT) def test_get_job_miss_w_explict_project(self): from google.cloud.exceptions import NotFound from google.cloud.bigquery.retry import DEFAULT_GET_JOB_TIMEOUT OTHER_PROJECT = "OTHER_PROJECT" JOB_ID = "NONESUCH" creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() with self.assertRaises(NotFound): client.get_job(JOB_ID, project=OTHER_PROJECT) conn.api_request.assert_called_once_with( method="GET", path="/projects/OTHER_PROJECT/jobs/NONESUCH", query_params={"projection": "full"}, timeout=DEFAULT_GET_JOB_TIMEOUT, ) def test_get_job_miss_w_client_location(self): from google.cloud.exceptions import NotFound from google.cloud.bigquery.retry import DEFAULT_GET_JOB_TIMEOUT JOB_ID = "NONESUCH" creds = _make_credentials() client = self._make_one("client-proj", creds, location="client-loc") conn = client._connection = make_connection() with self.assertRaises(NotFound): client.get_job(JOB_ID) conn.api_request.assert_called_once_with( method="GET", path="/projects/client-proj/jobs/NONESUCH", query_params={"projection": "full", "location": "client-loc"}, timeout=DEFAULT_GET_JOB_TIMEOUT, ) def test_get_job_hit_w_timeout(self): from google.cloud.bigquery.job import CreateDisposition from google.cloud.bigquery.job import QueryJob from google.cloud.bigquery.job import WriteDisposition JOB_ID = "query_job" QUERY_DESTINATION_TABLE = "query_destination_table" QUERY = "SELECT * from test_dataset:test_table" ASYNC_QUERY_DATA = { "id": "{}:{}".format(self.PROJECT, JOB_ID), "jobReference": { "projectId": "resource-proj", "jobId": "query_job", "location": "us-east1", }, "state": "DONE", "configuration": { "query": { "query": QUERY, "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": QUERY_DESTINATION_TABLE, }, "createDisposition": CreateDisposition.CREATE_IF_NEEDED, "writeDisposition": WriteDisposition.WRITE_TRUNCATE, } }, } creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection(ASYNC_QUERY_DATA) job_from_resource = QueryJob.from_api_repr(ASYNC_QUERY_DATA, client) job = client.get_job(job_from_resource, timeout=7.5) self.assertIsInstance(job, QueryJob) self.assertEqual(job.job_id, JOB_ID) self.assertEqual(job.project, "resource-proj") self.assertEqual(job.location, "us-east1") self.assertEqual(job.create_disposition, CreateDisposition.CREATE_IF_NEEDED) self.assertEqual(job.write_disposition, WriteDisposition.WRITE_TRUNCATE) conn.api_request.assert_called_once_with( method="GET", path="/projects/resource-proj/jobs/query_job", query_params={"projection": "full", "location": "us-east1"}, timeout=7.5, ) def test_cancel_job_miss_w_explict_project(self): from google.cloud.exceptions import NotFound OTHER_PROJECT = "OTHER_PROJECT" JOB_ID = "NONESUCH" creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection() with self.assertRaises(NotFound): client.cancel_job(JOB_ID, project=OTHER_PROJECT, location=self.LOCATION) conn.api_request.assert_called_once_with( method="POST", path="/projects/OTHER_PROJECT/jobs/NONESUCH/cancel", query_params={"projection": "full", "location": self.LOCATION}, timeout=DEFAULT_TIMEOUT, ) def test_cancel_job_miss_w_client_location(self): from google.cloud.exceptions import NotFound OTHER_PROJECT = "OTHER_PROJECT" JOB_ID = "NONESUCH" creds = _make_credentials() client = self._make_one(self.PROJECT, creds, location=self.LOCATION) conn = client._connection = make_connection() with self.assertRaises(NotFound): client.cancel_job(JOB_ID, project=OTHER_PROJECT) conn.api_request.assert_called_once_with( method="POST", path="/projects/OTHER_PROJECT/jobs/NONESUCH/cancel", query_params={"projection": "full", "location": self.LOCATION}, timeout=DEFAULT_TIMEOUT, ) def test_cancel_job_hit(self): from google.cloud.bigquery.job import QueryJob JOB_ID = "query_job" QUERY = "SELECT * from test_dataset:test_table" QUERY_JOB_RESOURCE = { "id": "{}:{}".format(self.PROJECT, JOB_ID), "jobReference": { "projectId": "job-based-proj", "jobId": "query_job", "location": "asia-northeast1", }, "state": "RUNNING", "configuration": {"query": {"query": QUERY}}, } RESOURCE = {"job": QUERY_JOB_RESOURCE} creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection(RESOURCE) job_from_resource = QueryJob.from_api_repr(QUERY_JOB_RESOURCE, client) job = client.cancel_job(job_from_resource) self.assertIsInstance(job, QueryJob) self.assertEqual(job.job_id, JOB_ID) self.assertEqual(job.project, "job-based-proj") self.assertEqual(job.location, "asia-northeast1") self.assertEqual(job.query, QUERY) conn.api_request.assert_called_once_with( method="POST", path="/projects/job-based-proj/jobs/query_job/cancel", query_params={"projection": "full", "location": "asia-northeast1"}, timeout=DEFAULT_TIMEOUT, ) def test_cancel_job_w_timeout(self): JOB_ID = "query_job" QUERY = "SELECT * from test_dataset:test_table" QUERY_JOB_RESOURCE = { "id": "{}:{}".format(self.PROJECT, JOB_ID), "jobReference": {"projectId": self.PROJECT, "jobId": "query_job"}, "state": "RUNNING", "configuration": {"query": {"query": QUERY}}, } RESOURCE = {"job": QUERY_JOB_RESOURCE} creds = _make_credentials() client = self._make_one(self.PROJECT, creds) conn = client._connection = make_connection(RESOURCE) client.cancel_job(JOB_ID, timeout=7.5) conn.api_request.assert_called_once_with( method="POST", path="/projects/{}/jobs/query_job/cancel".format(self.PROJECT), query_params={"projection": "full"}, timeout=7.5, ) def test_load_table_from_uri(self): from google.cloud.bigquery.job import LoadJob, LoadJobConfig JOB = "job_name" DESTINATION = "destination_table" SOURCE_URI = "http://example.com/source.csv" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "load": { "sourceUris": [SOURCE_URI], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": DESTINATION, }, } }, } creds = _make_credentials() http = object() job_config = LoadJobConfig() original_config_copy = copy.deepcopy(job_config) client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) destination = DatasetReference(self.PROJECT, self.DS_ID).table(DESTINATION) job = client.load_table_from_uri( SOURCE_URI, destination, job_id=JOB, job_config=job_config, timeout=7.5 ) # Check that load_table_from_uri actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=RESOURCE, timeout=7.5, ) # the original config object should not have been modified self.assertEqual(job_config.to_api_repr(), original_config_copy.to_api_repr()) self.assertIsInstance(job, LoadJob) self.assertIsInstance(job.configuration, LoadJobConfig) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(list(job.source_uris), [SOURCE_URI]) self.assertEqual(job.destination, destination) conn = client._connection = make_connection(RESOURCE) job = client.load_table_from_uri([SOURCE_URI], destination, job_id=JOB) self.assertIsInstance(job, LoadJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(list(job.source_uris), [SOURCE_URI]) self.assertEqual(job.destination, destination) def test_load_table_from_uri_w_explicit_project(self): job_id = "this-is-a-job-id" destination_id = "destination_table" source_uri = "gs://example/source.csv" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": { "load": { "sourceUris": [source_uri], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": destination_id, }, } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(resource) destination = DatasetReference(self.PROJECT, self.DS_ID).table(destination_id) client.load_table_from_uri( source_uri, destination, job_id=job_id, project="other-project", location=self.LOCATION, ) # Check that load_table_from_uri actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_load_table_from_uri_w_client_location(self): job_id = "this-is-a-job-id" destination_id = "destination_table" source_uri = "gs://example/source.csv" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": { "load": { "sourceUris": [source_uri], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": destination_id, }, } }, } creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=self.LOCATION ) conn = client._connection = make_connection(resource) client.load_table_from_uri( source_uri, # Test with string for table ID. "{}.{}".format(self.DS_ID, destination_id), job_id=job_id, project="other-project", ) # Check that load_table_from_uri actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_load_table_from_uri_w_invalid_job_config(self): from google.cloud.bigquery import job JOB = "job_name" DESTINATION = "destination_table" SOURCE_URI = "http://example.com/source.csv" creds = _make_credentials() http = object() job_config = job.CopyJobConfig() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) destination = DatasetReference(self.PROJECT, self.DS_ID).table(DESTINATION) with self.assertRaises(TypeError) as exc: client.load_table_from_uri( SOURCE_URI, destination, job_id=JOB, job_config=job_config ) self.assertIn("Expected an instance of LoadJobConfig", exc.exception.args[0]) def test_load_table_from_uri_w_explicit_job_config(self): from google.cloud.bigquery.job import LoadJobConfig JOB = "job_name" DESTINATION = "destination_table" SOURCE_URI = "http://example.com/source.csv" RESOURCE = { "jobReference": {"jobId": JOB, "projectId": self.PROJECT}, "configuration": { "load": { "sourceUris": [SOURCE_URI], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": DESTINATION, }, "createSession": True, "encoding": "UTF-8", } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) destination = DatasetReference(self.PROJECT, self.DS_ID).table(DESTINATION) job_config = LoadJobConfig() job_config.create_session = True job_config.encoding = "UTF-8" client.load_table_from_uri( SOURCE_URI, destination, job_id=JOB, job_config=job_config ) # Check that load_table_from_uri actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=RESOURCE, timeout=DEFAULT_TIMEOUT, ) def test_load_table_from_uri_w_explicit_job_config_override(self): from google.cloud.bigquery.job import LoadJobConfig JOB = "job_name" DESTINATION = "destination_table" SOURCE_URI = "http://example.com/source.csv" RESOURCE = { "jobReference": {"jobId": JOB, "projectId": self.PROJECT}, "configuration": { "load": { "sourceUris": [SOURCE_URI], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": DESTINATION, }, "createSession": False, "encoding": "ISO-8859-1", } }, } creds = _make_credentials() http = object() default_job_config = LoadJobConfig() default_job_config.create_session = True default_job_config.encoding = "ISO-8859-1" client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_load_job_config=default_job_config, ) conn = client._connection = make_connection(RESOURCE) destination = DatasetReference(self.PROJECT, self.DS_ID).table(DESTINATION) job_config = LoadJobConfig() job_config.create_session = False client.load_table_from_uri( SOURCE_URI, destination, job_id=JOB, job_config=job_config ) # Check that load_table_from_uri actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=RESOURCE, timeout=DEFAULT_TIMEOUT, ) def test_load_table_from_uri_w_default_load_config(self): from google.cloud.bigquery.job import LoadJobConfig JOB = "job_name" DESTINATION = "destination_table" SOURCE_URI = "http://example.com/source.csv" RESOURCE = { "jobReference": {"jobId": JOB, "projectId": self.PROJECT}, "configuration": { "load": { "sourceUris": [SOURCE_URI], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": DESTINATION, }, "encoding": "ISO-8859-1", } }, } creds = _make_credentials() http = object() default_job_config = LoadJobConfig() default_job_config.encoding = "ISO-8859-1" client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_load_job_config=default_job_config, ) conn = client._connection = make_connection(RESOURCE) destination = DatasetReference(self.PROJECT, self.DS_ID).table(DESTINATION) client.load_table_from_uri(SOURCE_URI, destination, job_id=JOB) # Check that load_table_from_uri actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=RESOURCE, timeout=DEFAULT_TIMEOUT, ) @staticmethod def _mock_requests_response(status_code, headers, content=b""): return mock.Mock( content=content, headers=headers, status_code=status_code, spec=["content", "headers", "status_code"], ) def _mock_transport(self, status_code, headers, content=b""): fake_transport = mock.Mock(spec=["request"]) fake_response = self._mock_requests_response( status_code, headers, content=content ) fake_transport.request.return_value = fake_response return fake_transport def _initiate_resumable_upload_helper(self, num_retries=None, mtls=False): from google.resumable_media.requests import ResumableUpload from google.cloud.bigquery.client import _DEFAULT_CHUNKSIZE from google.cloud.bigquery.client import _GENERIC_CONTENT_TYPE from google.cloud.bigquery.client import _get_upload_headers from google.cloud.bigquery.job import LoadJob from google.cloud.bigquery.job import LoadJobConfig from google.cloud.bigquery.job import SourceFormat # Create mocks to be checked for doing transport. resumable_url = "http://test.invalid?upload_id=hey-you" response_headers = {"location": resumable_url} fake_transport = self._mock_transport(http.client.OK, response_headers) client = self._make_one(project=self.PROJECT, _http=fake_transport) conn = client._connection = make_connection() if mtls: conn.get_api_base_url_for_mtls = mock.Mock(return_value="https://foo.mtls") # Create some mock arguments and call the method under test. data = b"goodbye gudbi gootbee" stream = io.BytesIO(data) config = LoadJobConfig() config.source_format = SourceFormat.CSV job = LoadJob(None, None, self.TABLE_REF, client, job_config=config) metadata = job.to_api_repr() upload, transport = client._initiate_resumable_upload( stream, metadata, num_retries, None ) # Check the returned values. self.assertIsInstance(upload, ResumableUpload) host_name = "https://foo.mtls" if mtls else "https://bigquery.googleapis.com" upload_url = ( f"{host_name}/upload/bigquery/v2/projects/{self.PROJECT}" "/jobs?uploadType=resumable" ) self.assertEqual(upload.upload_url, upload_url) expected_headers = _get_upload_headers(conn.user_agent) self.assertEqual(upload._headers, expected_headers) self.assertFalse(upload.finished) self.assertEqual(upload._chunk_size, _DEFAULT_CHUNKSIZE) self.assertIs(upload._stream, stream) self.assertIsNone(upload._total_bytes) self.assertEqual(upload._content_type, _GENERIC_CONTENT_TYPE) self.assertEqual(upload.resumable_url, resumable_url) retry_strategy = upload._retry_strategy self.assertEqual(retry_strategy.max_sleep, 64.0) if num_retries is None: self.assertEqual(retry_strategy.max_cumulative_retry, 600.0) self.assertIsNone(retry_strategy.max_retries) else: self.assertIsNone(retry_strategy.max_cumulative_retry) self.assertEqual(retry_strategy.max_retries, num_retries) self.assertIs(transport, fake_transport) # Make sure we never read from the stream. self.assertEqual(stream.tell(), 0) # Check the mocks. request_headers = expected_headers.copy() request_headers["x-upload-content-type"] = _GENERIC_CONTENT_TYPE fake_transport.request.assert_called_once_with( "POST", upload_url, data=json.dumps(metadata).encode("utf-8"), headers=request_headers, timeout=mock.ANY, ) def test__initiate_resumable_upload(self): self._initiate_resumable_upload_helper() def test__initiate_resumable_upload_mtls(self): self._initiate_resumable_upload_helper(mtls=True) def test__initiate_resumable_upload_with_retry(self): self._initiate_resumable_upload_helper(num_retries=11) def _do_multipart_upload_success_helper( self, get_boundary, num_retries=None, project=None, mtls=False ): from google.cloud.bigquery.client import _get_upload_headers from google.cloud.bigquery.job import LoadJob from google.cloud.bigquery.job import LoadJobConfig from google.cloud.bigquery.job import SourceFormat fake_transport = self._mock_transport(http.client.OK, {}) client = self._make_one(project=self.PROJECT, _http=fake_transport) conn = client._connection = make_connection() if mtls: conn.get_api_base_url_for_mtls = mock.Mock(return_value="https://foo.mtls") if project is None: project = self.PROJECT # Create some mock arguments. data = b"Bzzzz-zap \x00\x01\xf4" stream = io.BytesIO(data) config = LoadJobConfig() config.source_format = SourceFormat.CSV job = LoadJob(None, None, self.TABLE_REF, client, job_config=config) metadata = job.to_api_repr() size = len(data) response = client._do_multipart_upload( stream, metadata, size, num_retries, None, project=project ) # Check the mocks and the returned value. self.assertIs(response, fake_transport.request.return_value) self.assertEqual(stream.tell(), size) get_boundary.assert_called_once_with() host_name = "https://foo.mtls" if mtls else "https://bigquery.googleapis.com" upload_url = ( f"{host_name}/upload/bigquery/v2/projects/{project}" "/jobs?uploadType=multipart" ) payload = ( b"--==0==\r\n" b"content-type: application/json; charset=UTF-8\r\n\r\n" b"%(json_metadata)s" b"\r\n" b"--==0==\r\n" b"content-type: */*\r\n\r\n" b"%(data)s" b"\r\n" b"--==0==--" ) % {b"json_metadata": json.dumps(metadata).encode("utf-8"), b"data": data} headers = _get_upload_headers(conn.user_agent) headers["content-type"] = b'multipart/related; boundary="==0=="' fake_transport.request.assert_called_once_with( "POST", upload_url, data=payload, headers=headers, timeout=mock.ANY ) @mock.patch("google.resumable_media._upload.get_boundary", return_value=b"==0==") def test__do_multipart_upload(self, get_boundary): self._do_multipart_upload_success_helper(get_boundary) @mock.patch("google.resumable_media._upload.get_boundary", return_value=b"==0==") def test__do_multipart_upload_mtls(self, get_boundary): self._do_multipart_upload_success_helper(get_boundary, mtls=True) @mock.patch("google.resumable_media._upload.get_boundary", return_value=b"==0==") def test__do_multipart_upload_with_retry(self, get_boundary): self._do_multipart_upload_success_helper(get_boundary, num_retries=8) @mock.patch("google.resumable_media._upload.get_boundary", return_value=b"==0==") def test__do_multipart_upload_with_custom_project(self, get_boundary): self._do_multipart_upload_success_helper(get_boundary, project="custom-project") def test_copy_table(self): from google.cloud.bigquery.job import CopyJob JOB = "job_name" SOURCE = "source_table" DESTINATION = "destination_table" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "copy": { "sourceTables": [ { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": SOURCE, } ], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": DESTINATION, }, } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) destination = dataset.table(DESTINATION) job = client.copy_table(source, destination, job_id=JOB, timeout=7.5) # Check that copy_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=RESOURCE, timeout=7.5, ) self.assertIsInstance(job, CopyJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(list(job.sources), [source]) self.assertEqual(job.destination, destination) def test_copy_table_w_multiple_sources(self): from google.cloud.bigquery.job import CopyJob from google.cloud.bigquery.table import TableReference job_id = "job_name" source_id = "my-project.my_dataset.source_table" source_id2 = "my-project.my_dataset.source_table2" destination_id = "my-other-project.another_dataset.destination_table" expected_resource = { "jobReference": {"projectId": self.PROJECT, "jobId": job_id}, "configuration": { "copy": { "sourceTables": [ { "projectId": "my-project", "datasetId": "my_dataset", "tableId": "source_table", }, { "projectId": "my-project", "datasetId": "my_dataset", "tableId": "source_table2", }, ], "destinationTable": { "projectId": "my-other-project", "datasetId": "another_dataset", "tableId": "destination_table", }, } }, } returned_resource = expected_resource.copy() returned_resource["statistics"] = {} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(returned_resource) job = client.copy_table([source_id, source_id2], destination_id, job_id=job_id) # Check that copy_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=expected_resource, timeout=DEFAULT_TIMEOUT, ) self.assertIsInstance(job, CopyJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, job_id) self.assertEqual( list(sorted(job.sources, key=lambda tbl: tbl.table_id)), [ TableReference.from_string(source_id), TableReference.from_string(source_id2), ], ) self.assertEqual(job.destination, TableReference.from_string(destination_id)) def test_copy_table_w_explicit_project(self): job_id = "this-is-a-job-id" source_id = "source_table" destination_id = "destination_table" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": { "copy": { "sourceTables": [ { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": source_id, } ], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": destination_id, }, } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(resource) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(source_id) destination = dataset.table(destination_id) client.copy_table( source, destination, job_id=job_id, project="other-project", location=self.LOCATION, ) # Check that copy_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_copy_table_w_client_location(self): job_id = "this-is-a-job-id" source_id = "source_table" destination_id = "destination_table" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": { "copy": { "sourceTables": [ { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": source_id, } ], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": destination_id, }, } }, } creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=self.LOCATION ) conn = client._connection = make_connection(resource) client.copy_table( # Test with string for table IDs. "{}.{}".format(self.DS_ID, source_id), "{}.{}".format(self.DS_ID, destination_id), job_id=job_id, project="other-project", ) # Check that copy_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_copy_table_w_source_strings(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) sources = [ "dataset_wo_proj.some_table", "other_project.other_dataset.other_table", DatasetReference(client.project, "dataset_from_ref").table( "table_from_ref" ), ] destination = "some_project.some_dataset.destination_table" job = client.copy_table(sources, destination) # Replace job with the request instead of response so we can verify those properties. _, kwargs = conn.api_request.call_args request = kwargs["data"] job._properties = request expected_sources = [ DatasetReference(client.project, "dataset_wo_proj").table("some_table"), DatasetReference("other_project", "other_dataset").table("other_table"), DatasetReference(client.project, "dataset_from_ref").table( "table_from_ref" ), ] self.assertEqual(list(job.sources), expected_sources) expected_destination = DatasetReference("some_project", "some_dataset").table( "destination_table" ) self.assertEqual(job.destination, expected_destination) def test_copy_table_w_invalid_job_config(self): from google.cloud.bigquery import job JOB = "job_name" SOURCE = "source_table" DESTINATION = "destination_table" creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) job_config = job.ExtractJobConfig() dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) destination = dataset.table(DESTINATION) with self.assertRaises(TypeError) as exc: client.copy_table(source, destination, job_id=JOB, job_config=job_config) self.assertIn("Expected an instance of CopyJobConfig", exc.exception.args[0]) def test_copy_table_w_valid_job_config(self): from google.cloud.bigquery.job import CopyJobConfig JOB = "job_name" SOURCE = "source_table" DESTINATION = "destination_table" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "copy": { "sourceTables": [ { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": SOURCE, } ], "destinationTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": DESTINATION, }, } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) destination = dataset.table(DESTINATION) job_config = CopyJobConfig() original_config_copy = copy.deepcopy(job_config) job = client.copy_table(source, destination, job_id=JOB, job_config=job_config) # Check that copy_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/%s/jobs" % self.PROJECT, data=RESOURCE, timeout=DEFAULT_TIMEOUT, ) self.assertIsInstance(job.configuration, CopyJobConfig) # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_extract_table(self): from google.cloud.bigquery.job import ExtractJob JOB = "job_id" SOURCE = "source_table" DESTINATION = "gs://bucket_name/object_name" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "extract": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": SOURCE, }, "destinationUris": [DESTINATION], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) job = client.extract_table(source, DESTINATION, job_id=JOB, timeout=7.5) # Check that extract_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=RESOURCE, timeout=7.5, ) # Check the job resource. self.assertIsInstance(job, ExtractJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(job.source, source) self.assertEqual(list(job.destination_uris), [DESTINATION]) def test_extract_table_w_invalid_job_config(self): from google.cloud.bigquery import job JOB = "job_id" SOURCE = "source_table" DESTINATION = "gs://bucket_name/object_name" creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) job_config = job.LoadJobConfig() with self.assertRaises(TypeError) as exc: client.extract_table(source, DESTINATION, job_id=JOB, job_config=job_config) self.assertIn("Expected an instance of ExtractJobConfig", exc.exception.args[0]) def test_extract_table_w_explicit_project(self): job_id = "job_id" source_id = "source_table" destination = "gs://bucket_name/object_name" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": { "extract": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": source_id, }, "destinationUris": [destination], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(resource) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(source_id) client.extract_table( source, destination, job_id=job_id, project="other-project", location=self.LOCATION, ) # Check that extract_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_extract_table_w_client_location(self): job_id = "job_id" source_id = "source_table" destination = "gs://bucket_name/object_name" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": { "extract": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": source_id, }, "destinationUris": [destination], } }, } creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=self.LOCATION ) conn = client._connection = make_connection(resource) client.extract_table( # Test with string for table ID. "{}.{}".format(self.DS_ID, source_id), destination, job_id=job_id, project="other-project", ) # Check that extract_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_extract_table_generated_job_id(self): from google.cloud.bigquery.job import ExtractJob from google.cloud.bigquery.job import ExtractJobConfig from google.cloud.bigquery.job import DestinationFormat JOB = "job_id" SOURCE = "source_table" DESTINATION = "gs://bucket_name/object_name" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "extract": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": SOURCE, }, "destinationUris": [DESTINATION], "destinationFormat": "NEWLINE_DELIMITED_JSON", } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) job_config = ExtractJobConfig() job_config.destination_format = DestinationFormat.NEWLINE_DELIMITED_JSON original_config_copy = copy.deepcopy(job_config) job = client.extract_table(source, DESTINATION, job_config=job_config) # Check that extract_table actually starts the job. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/PROJECT/jobs") self.assertIsInstance(req["data"]["jobReference"]["jobId"], str) self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) # Check the job resource. self.assertIsInstance(job, ExtractJob) self.assertIs(job._client, client) self.assertEqual(job.source, source) self.assertEqual(list(job.destination_uris), [DESTINATION]) # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_extract_table_w_destination_uris(self): from google.cloud.bigquery.job import ExtractJob JOB = "job_id" SOURCE = "source_table" DESTINATION1 = "gs://bucket_name/object_one" DESTINATION2 = "gs://bucket_name/object_two" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "extract": { "sourceTable": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": SOURCE, }, "destinationUris": [DESTINATION1, DESTINATION2], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.table(SOURCE) job = client.extract_table(source, [DESTINATION1, DESTINATION2], job_id=JOB) # Check that extract_table actually starts the job. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/PROJECT/jobs") self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) # Check the job resource. self.assertIsInstance(job, ExtractJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(job.source, source) self.assertEqual(list(job.destination_uris), [DESTINATION1, DESTINATION2]) def test_extract_table_for_source_type_model(self): from google.cloud.bigquery.job import ExtractJob JOB = "job_id" SOURCE = "source_model" DESTINATION = "gs://bucket_name/object_name" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "extract": { "sourceModel": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": SOURCE, }, "destinationUris": [DESTINATION], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.model(SOURCE) job = client.extract_table( source, DESTINATION, job_id=JOB, timeout=7.5, source_type="Model" ) # Check that extract_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=RESOURCE, timeout=7.5, ) # Check the job resource. self.assertIsInstance(job, ExtractJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(job.source, source) self.assertEqual(list(job.destination_uris), [DESTINATION]) def test_extract_table_for_source_type_model_w_string_model_id(self): JOB = "job_id" source_id = "source_model" DESTINATION = "gs://bucket_name/object_name" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "extract": { "sourceModel": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": source_id, }, "destinationUris": [DESTINATION], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) client.extract_table( # Test with string for model ID. "{}.{}".format(self.DS_ID, source_id), DESTINATION, job_id=JOB, timeout=7.5, source_type="Model", ) # Check that extract_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=RESOURCE, timeout=7.5, ) def test_extract_table_for_source_type_model_w_model_object(self): from google.cloud.bigquery.model import Model JOB = "job_id" DESTINATION = "gs://bucket_name/object_name" model_id = "{}.{}.{}".format(self.PROJECT, self.DS_ID, self.MODEL_ID) model = Model(model_id) RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "extract": { "sourceModel": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "modelId": self.MODEL_ID, }, "destinationUris": [DESTINATION], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) client.extract_table( # Test with Model class object. model, DESTINATION, job_id=JOB, timeout=7.5, source_type="Model", ) # Check that extract_table actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=RESOURCE, timeout=7.5, ) def test_extract_table_for_invalid_source_type_model(self): JOB = "job_id" SOURCE = "source_model" DESTINATION = "gs://bucket_name/object_name" creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) dataset = DatasetReference(self.PROJECT, self.DS_ID) source = dataset.model(SOURCE) with self.assertRaises(ValueError) as exc: client.extract_table( source, DESTINATION, job_id=JOB, timeout=7.5, source_type="foo" ) self.assertIn("Cannot pass", exc.exception.args[0]) def test_query_defaults(self): from google.cloud.bigquery.job import QueryJob QUERY = "select count(*) from persons" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": "some-random-id"}, "configuration": {"query": {"query": QUERY, "useLegacySql": False}}, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) job = client.query(QUERY) self.assertIsInstance(job, QueryJob) self.assertIsInstance(job.job_id, str) self.assertIs(job._client, client) self.assertEqual(job.query, QUERY) self.assertEqual(job.udf_resources, []) self.assertEqual(job.query_parameters, []) # Check that query actually starts the job. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/PROJECT/jobs") self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) sent = req["data"] self.assertIsInstance(sent["jobReference"]["jobId"], str) sent_config = sent["configuration"]["query"] self.assertEqual(sent_config["query"], QUERY) self.assertFalse(sent_config["useLegacySql"]) def test_query_w_api_method_query(self): query = "select count(*) from persons" response = { "jobReference": { "projectId": self.PROJECT, "location": "EU", "jobId": "abcd", }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(response) job = client.query(query, location="EU", api_method="QUERY") self.assertEqual(job.query, query) self.assertEqual(job.job_id, "abcd") self.assertEqual(job.location, "EU") # Check that query actually starts the job. expected_resource = { "query": query, "useLegacySql": False, "location": "EU", "formatOptions": {"useInt64Timestamp": True}, "requestId": mock.ANY, } conn.api_request.assert_called_once_with( method="POST", path=f"/projects/{self.PROJECT}/queries", data=expected_resource, timeout=None, ) def test_query_w_api_method_query_legacy_sql(self): from google.cloud.bigquery import QueryJobConfig query = "select count(*) from persons" response = { "jobReference": { "projectId": self.PROJECT, "location": "EU", "jobId": "abcd", }, } job_config = QueryJobConfig() job_config.use_legacy_sql = True job_config.maximum_bytes_billed = 100 creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(response) job = client.query( query, location="EU", job_config=job_config, api_method="QUERY" ) self.assertEqual(job.query, query) self.assertEqual(job.job_id, "abcd") self.assertEqual(job.location, "EU") # Check that query actually starts the job. expected_resource = { "query": query, "useLegacySql": True, "location": "EU", "formatOptions": {"useInt64Timestamp": True}, "requestId": mock.ANY, "maximumBytesBilled": "100", } conn.api_request.assert_called_once_with( method="POST", path=f"/projects/{self.PROJECT}/queries", data=expected_resource, timeout=None, ) def test_query_w_api_method_query_parameters(self): from google.cloud.bigquery import QueryJobConfig, ScalarQueryParameter query = "select count(*) from persons" response = { "jobReference": { "projectId": self.PROJECT, "location": "EU", "jobId": "abcd", }, } job_config = QueryJobConfig() job_config.dry_run = True job_config.query_parameters = [ScalarQueryParameter("param1", "INTEGER", 123)] creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(response) job = client.query( query, location="EU", job_config=job_config, api_method="QUERY" ) self.assertEqual(job.query, query) self.assertEqual(job.job_id, "abcd") self.assertEqual(job.location, "EU") # Check that query actually starts the job. expected_resource = { "query": query, "dryRun": True, "useLegacySql": False, "location": "EU", "formatOptions": {"useInt64Timestamp": True}, "requestId": mock.ANY, "parameterMode": "NAMED", "queryParameters": [ { "name": "param1", "parameterType": {"type": "INTEGER"}, "parameterValue": {"value": "123"}, }, ], } conn.api_request.assert_called_once_with( method="POST", path=f"/projects/{self.PROJECT}/queries", data=expected_resource, timeout=None, ) def test_query_w_api_method_query_and_job_id_fails(self): query = "select count(*) from persons" creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client._connection = make_connection({}) with self.assertRaises(TypeError) as exc: client.query(query, job_id="abcd", api_method="QUERY") self.assertIn( "`job_id` was provided, but the 'QUERY' `api_method` was requested", exc.exception.args[0], ) def test_query_w_api_method_unknown(self): query = "select count(*) from persons" creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client._connection = make_connection({}) with self.assertRaises(ValueError) as exc: client.query(query, api_method="UNKNOWN") self.assertIn("Got unexpected value for api_method: ", exc.exception.args[0]) def test_query_w_explicit_timeout(self): query = "select count(*) from persons" resource = { "jobReference": {"projectId": self.PROJECT, "jobId": mock.ANY}, "configuration": {"query": {"query": query, "useLegacySql": False}}, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(resource) client.query(query, timeout=7.5) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/{}/jobs".format(self.PROJECT), data=resource, timeout=7.5, ) def test_query_w_explicit_project(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": {"query": {"query": query, "useLegacySql": False}}, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(resource) client.query( query, job_id=job_id, project="other-project", location=self.LOCATION ) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_query_w_explicit_job_config(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "jobId": job_id, "projectId": self.PROJECT, "location": self.LOCATION, }, "configuration": { "query": { "query": query, "defaultDataset": { "projectId": self.PROJECT, "datasetId": "some-dataset", }, "useLegacySql": False, "useQueryCache": True, "maximumBytesBilled": "2000", } }, } creds = _make_credentials() http = object() from google.cloud.bigquery import QueryJobConfig, DatasetReference default_job_config = QueryJobConfig() default_job_config.default_dataset = DatasetReference( self.PROJECT, "some-dataset" ) default_job_config.maximum_bytes_billed = 1000 client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=default_job_config, ) conn = client._connection = make_connection(resource) job_config = QueryJobConfig() job_config.use_query_cache = True job_config.maximum_bytes_billed = 2000 original_config_copy = copy.deepcopy(job_config) client.query( query, job_id=job_id, location=self.LOCATION, job_config=job_config ) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_query_preserving_explicit_job_config(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "jobId": job_id, "projectId": self.PROJECT, "location": self.LOCATION, }, "configuration": { "query": { "query": query, "useLegacySql": False, "useQueryCache": True, "maximumBytesBilled": "2000", } }, } creds = _make_credentials() http = object() from google.cloud.bigquery import QueryJobConfig client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, ) conn = client._connection = make_connection(resource) job_config = QueryJobConfig() job_config.use_query_cache = True job_config.maximum_bytes_billed = 2000 original_config_copy = copy.deepcopy(job_config) client.query( query, job_id=job_id, location=self.LOCATION, job_config=job_config ) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_query_preserving_explicit_default_job_config(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "jobId": job_id, "projectId": self.PROJECT, "location": self.LOCATION, }, "configuration": { "query": { "query": query, "defaultDataset": { "projectId": self.PROJECT, "datasetId": "some-dataset", }, "useLegacySql": False, "maximumBytesBilled": "1000", } }, } creds = _make_credentials() http = object() from google.cloud.bigquery import QueryJobConfig, DatasetReference default_job_config = QueryJobConfig() default_job_config.default_dataset = DatasetReference( self.PROJECT, "some-dataset" ) default_job_config.maximum_bytes_billed = 1000 default_config_copy = copy.deepcopy(default_job_config) client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=default_job_config, ) conn = client._connection = make_connection(resource) client.query(query, job_id=job_id, location=self.LOCATION, job_config=None) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) # the original default config object should not have been modified assert default_job_config.to_api_repr() == default_config_copy.to_api_repr() def test_query_w_invalid_job_config(self): from google.cloud.bigquery import QueryJobConfig, DatasetReference from google.cloud.bigquery import job job_id = "some-job-id" query = "select count(*) from persons" creds = _make_credentials() http = object() default_job_config = QueryJobConfig() default_job_config.default_dataset = DatasetReference( self.PROJECT, "some-dataset" ) default_job_config.maximum_bytes_billed = 1000 client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=default_job_config, ) job_config = job.LoadJobConfig() with self.assertRaises(TypeError) as exc: client.query( query, job_id=job_id, location=self.LOCATION, job_config=job_config ) self.assertIn("Expected an instance of QueryJobConfig", exc.exception.args[0]) def test_query_w_explicit_job_config_override(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "jobId": job_id, "projectId": self.PROJECT, "location": self.LOCATION, }, "configuration": { "query": { "query": query, "defaultDataset": None, "useLegacySql": False, "useQueryCache": True, "maximumBytesBilled": "2000", } }, } creds = _make_credentials() http = object() from google.cloud.bigquery import QueryJobConfig, DatasetReference default_job_config = QueryJobConfig() default_job_config.default_dataset = DatasetReference( self.PROJECT, "some-dataset" ) default_job_config.maximum_bytes_billed = 1000 client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=default_job_config, ) conn = client._connection = make_connection(resource) job_config = QueryJobConfig() job_config.use_query_cache = True job_config.maximum_bytes_billed = 2000 job_config.default_dataset = None client.query( query, job_id=job_id, location=self.LOCATION, job_config=job_config ) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_query_w_client_default_config_no_incoming(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "jobId": job_id, "projectId": self.PROJECT, "location": self.LOCATION, }, "configuration": { "query": { "query": query, "useLegacySql": False, "maximumBytesBilled": "1000", } }, } creds = _make_credentials() http = object() from google.cloud.bigquery import QueryJobConfig default_job_config = QueryJobConfig() default_job_config.maximum_bytes_billed = 1000 client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=default_job_config, ) conn = client._connection = make_connection(resource) client.query(query, job_id=job_id, location=self.LOCATION) # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/PROJECT/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_query_w_invalid_default_job_config(self): creds = _make_credentials() http = object() default_job_config = object() with self.assertRaises(TypeError) as exc: self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=default_job_config, ) self.assertIn("Expected an instance of QueryJobConfig", exc.exception.args[0]) def test_query_w_client_location(self): job_id = "some-job-id" query = "select count(*) from persons" resource = { "jobReference": { "projectId": "other-project", "location": self.LOCATION, "jobId": job_id, }, "configuration": {"query": {"query": query, "useLegacySql": False}}, } creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, location=self.LOCATION ) conn = client._connection = make_connection(resource) client.query(query, job_id=job_id, project="other-project") # Check that query actually starts the job. conn.api_request.assert_called_once_with( method="POST", path="/projects/other-project/jobs", data=resource, timeout=DEFAULT_TIMEOUT, ) def test_query_detect_location(self): query = "select count(*) from persons" resource_location = "EU" resource = { "jobReference": { "projectId": self.PROJECT, # Location not set in request, but present in the response. "location": resource_location, "jobId": "some-random-id", }, "configuration": {"query": {"query": query, "useLegacySql": False}}, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(resource) job = client.query(query) self.assertEqual(job.location, resource_location) # Check that request did not contain a location. conn.api_request.assert_called_once() _, req = conn.api_request.call_args sent = req["data"] self.assertIsNone(sent["jobReference"].get("location")) def test_query_w_udf_resources(self): from google.cloud.bigquery.job import QueryJob from google.cloud.bigquery.job import QueryJobConfig from google.cloud.bigquery.query import UDFResource RESOURCE_URI = "gs://some-bucket/js/lib.js" JOB = "job_name" QUERY = "select count(*) from persons" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "query": { "query": QUERY, "useLegacySql": True, "userDefinedFunctionResources": [{"resourceUri": RESOURCE_URI}], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) udf_resources = [UDFResource("resourceUri", RESOURCE_URI)] config = QueryJobConfig() config.udf_resources = udf_resources config.use_legacy_sql = True job = client.query(QUERY, job_config=config, job_id=JOB) self.assertIsInstance(job, QueryJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(job.query, QUERY) self.assertEqual(job.udf_resources, udf_resources) self.assertEqual(job.query_parameters, []) # Check that query actually starts the job. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/PROJECT/jobs") self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) sent = req["data"] self.assertIsInstance(sent["jobReference"]["jobId"], str) sent_config = sent["configuration"]["query"] self.assertEqual(sent_config["query"], QUERY) self.assertTrue(sent_config["useLegacySql"]) self.assertEqual( sent_config["userDefinedFunctionResources"][0], {"resourceUri": RESOURCE_URI}, ) def test_query_w_query_parameters(self): from google.cloud.bigquery.job import QueryJob from google.cloud.bigquery.job import QueryJobConfig from google.cloud.bigquery.query import ScalarQueryParameter JOB = "job_name" QUERY = "select count(*) from persons" RESOURCE = { "jobReference": {"projectId": self.PROJECT, "jobId": JOB}, "configuration": { "query": { "query": QUERY, "useLegacySql": False, "queryParameters": [ { "name": "foo", "parameterType": {"type": "INT64"}, "parameterValue": {"value": "123"}, } ], } }, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESOURCE) query_parameters = [ScalarQueryParameter("foo", "INT64", 123)] config = QueryJobConfig() config.query_parameters = query_parameters job = client.query(QUERY, job_config=config, job_id=JOB) self.assertIsInstance(job, QueryJob) self.assertIs(job._client, client) self.assertEqual(job.job_id, JOB) self.assertEqual(job.query, QUERY) self.assertEqual(job.udf_resources, []) self.assertEqual(job.query_parameters, query_parameters) # Check that query actually starts the job. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/PROJECT/jobs") self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) sent = req["data"] self.assertEqual(sent["jobReference"]["jobId"], JOB) sent_config = sent["configuration"]["query"] self.assertEqual(sent_config["query"], QUERY) self.assertFalse(sent_config["useLegacySql"]) self.assertEqual( sent_config["queryParameters"][0], { "name": "foo", "parameterType": {"type": "INT64"}, "parameterValue": {"value": "123"}, }, ) def test_query_job_rpc_fail_w_random_error(self): from google.api_core.exceptions import Unknown from google.cloud.bigquery.job import QueryJob creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) job_create_error = Unknown("Not sure what went wrong.") job_begin_patcher = mock.patch.object( QueryJob, "_begin", side_effect=job_create_error ) with job_begin_patcher: with pytest.raises(Unknown, match="Not sure what went wrong."): client.query("SELECT 1;", job_id="123") def test_query_job_rpc_fail_w_conflict_job_id_given(self): from google.api_core.exceptions import Conflict from google.cloud.bigquery.job import QueryJob creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) job_create_error = Conflict("Job already exists.") job_begin_patcher = mock.patch.object( QueryJob, "_begin", side_effect=job_create_error ) with job_begin_patcher: with pytest.raises(Conflict, match="Job already exists."): client.query("SELECT 1;", job_id="123") def test_query_job_rpc_fail_w_conflict_random_id_job_fetch_fails(self): from google.api_core.exceptions import Conflict from google.api_core.exceptions import DataLoss from google.cloud.bigquery.job import QueryJob creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) job_create_error = Conflict("Job already exists.") job_begin_patcher = mock.patch.object( QueryJob, "_begin", side_effect=job_create_error ) get_job_patcher = mock.patch.object( client, "get_job", side_effect=DataLoss("we lost your job, sorry") ) with job_begin_patcher, get_job_patcher: # If get job request fails but supposedly there does exist a job # with this ID already, raise the exception explaining why we # couldn't recover the job. with pytest.raises(DataLoss, match="we lost your job, sorry"): client.query("SELECT 1;", job_id=None) def test_query_job_rpc_fail_w_conflict_random_id_job_fetch_succeeds(self): from google.api_core.exceptions import Conflict from google.cloud.bigquery.job import QueryJob creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) job_create_error = Conflict("Job already exists.") job_begin_patcher = mock.patch.object( QueryJob, "_begin", side_effect=job_create_error ) get_job_patcher = mock.patch.object( client, "get_job", return_value=mock.sentinel.query_job ) with job_begin_patcher, get_job_patcher: result = client.query("SELECT 1;", job_id=None) assert result is mock.sentinel.query_job def test_query_and_wait_defaults(self): query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, "schema": { "fields": [ { "name": "f0_", "type": "INTEGER", "mode": "NULLABLE", }, ], }, "totalRows": "1", "rows": [{"f": [{"v": "5552452"}]}], "queryId": "job_abcDEF_", } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(jobs_query_response) rows = client.query_and_wait(query) self.assertIsInstance(rows, google.cloud.bigquery.table.RowIterator) self.assertEqual(rows.query_id, "job_abcDEF_") self.assertEqual(rows.total_rows, 1) # No job reference in the response should be OK for completed query. self.assertIsNone(rows.job_id) self.assertIsNone(rows.project) self.assertIsNone(rows.location) # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/PROJECT/queries") self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) sent = req["data"] self.assertEqual(sent["query"], query) self.assertFalse(sent["useLegacySql"]) def test_query_and_wait_w_default_query_job_config(self): from google.cloud.bigquery import job query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, default_query_job_config=job.QueryJobConfig( labels={ "default-label": "default-value", }, ), ) conn = client._connection = make_connection(jobs_query_response) _ = client.query_and_wait(query) # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], f"/projects/{self.PROJECT}/queries") sent = req["data"] self.assertEqual(sent["labels"], {"default-label": "default-value"}) def test_query_and_wait_w_job_config(self): from google.cloud.bigquery import job query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one( project=self.PROJECT, credentials=creds, _http=http, ) conn = client._connection = make_connection(jobs_query_response) _ = client.query_and_wait( query, job_config=job.QueryJobConfig( labels={ "job_config-label": "job_config-value", }, ), ) # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], f"/projects/{self.PROJECT}/queries") sent = req["data"] self.assertEqual(sent["labels"], {"job_config-label": "job_config-value"}) def test_query_and_wait_w_location(self): query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(jobs_query_response) _ = client.query_and_wait(query, location="not-the-client-location") # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], f"/projects/{self.PROJECT}/queries") sent = req["data"] self.assertEqual(sent["location"], "not-the-client-location") def test_query_and_wait_w_max_results(self): query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(jobs_query_response) _ = client.query_and_wait(query, max_results=11) # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], f"/projects/{self.PROJECT}/queries") sent = req["data"] self.assertTrue(sent["formatOptions"]["useInt64Timestamp"]) self.assertTrue(sent["maxResults"], 11) def test_query_and_wait_w_page_size(self): query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(jobs_query_response) _ = client.query_and_wait(query, page_size=11) # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], f"/projects/{self.PROJECT}/queries") sent = req["data"] self.assertTrue(sent["formatOptions"]["useInt64Timestamp"]) self.assertTrue(sent["maxResults"], 11) def test_query_and_wait_w_page_size_multiple_requests(self): """ For queries that last longer than the intial (about 10s) call to jobs.query, we should still pass through the page size to the subsequent calls to jobs.getQueryResults. See internal issue 344008814. """ query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" job_reference = { "projectId": "my-jobs-project", "location": "my-jobs-location", "jobId": "my-jobs-id", } jobs_query_response = { "jobComplete": False, "jobReference": job_reference, } jobs_get_response = { "jobReference": job_reference, "status": {"state": "DONE"}, } get_query_results_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection( jobs_query_response, jobs_get_response, get_query_results_response, ) _ = client.query_and_wait(query, page_size=11) conn.api_request.assert_has_calls( [ # Verify the request we send is to jobs.query. mock.call( method="POST", path=f"/projects/{self.PROJECT}/queries", data={ "useLegacySql": False, "query": query, "formatOptions": {"useInt64Timestamp": True}, "maxResults": 11, "requestId": mock.ANY, }, timeout=None, ), # jobs.get: Check if the job has finished. mock.call( method="GET", path="/projects/my-jobs-project/jobs/my-jobs-id", query_params={ "projection": "full", "location": "my-jobs-location", }, timeout=google.cloud.bigquery.retry.DEFAULT_GET_JOB_TIMEOUT, ), # jobs.getQueryResults: wait for the query / fetch first page mock.call( method="GET", path="/projects/my-jobs-project/queries/my-jobs-id", query_params={ # We should still pass through the page size to the # subsequent calls to jobs.getQueryResults. # # See internal issue 344008814. "maxResults": 11, "formatOptions.useInt64Timestamp": True, "location": "my-jobs-location", }, timeout=None, ), ] ) def test_query_and_wait_w_project(self): query = "select count(*) from `bigquery-public-data.usa_names.usa_1910_2013`" jobs_query_response = { "jobComplete": True, } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(jobs_query_response) _ = client.query_and_wait(query, project="not-the-client-project") # Verify the request we send is to jobs.query. conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/projects/not-the-client-project/queries") def test_insert_rows_w_timeout(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) table = Table(self.TABLE_REF) ROWS = [ ("Phred Phlyntstone", 32), ("Bharney Rhubble", 33), ] schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), ] client.insert_rows(table, ROWS, selected_fields=schema, timeout=7.5) conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req.get("timeout"), 7.5) def test_insert_rows_wo_schema(self): from google.cloud.bigquery.table import Table creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) table = Table(self.TABLE_REF) ROWS = [ ("Phred Phlyntstone", 32), ("Bharney Rhubble", 33), ("Wylma Phlyntstone", 29), ("Bhettye Rhubble", 27), ] with self.assertRaises(ValueError) as exc: client.insert_rows(table, ROWS) self.assertIn("Could not determine schema for table", exc.exception.args[0]) def test_insert_rows_w_schema(self): import datetime from google.cloud._helpers import UTC from google.cloud._helpers import _datetime_to_rfc3339 from google.cloud._helpers import _RFC3339_MICROS from google.cloud.bigquery.schema import SchemaField WHEN_TS = 1437767599.006 WHEN = datetime.datetime.utcfromtimestamp(WHEN_TS).replace(tzinfo=UTC) PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), SchemaField("joined", "TIMESTAMP", mode="NULLABLE"), ] ROWS = [ ("Phred Phlyntstone", 32, _datetime_to_rfc3339(WHEN)), ("Bharney Rhubble", 33, WHEN + datetime.timedelta(seconds=1)), ("Wylma Phlyntstone", 29, WHEN + datetime.timedelta(seconds=2)), ("Bhettye Rhubble", 27, None), ] def _row_data(row): result = {"full_name": row[0], "age": str(row[1])} joined = row[2] if isinstance(joined, datetime.datetime): joined = joined.strftime(_RFC3339_MICROS) if joined is not None: result["joined"] = joined return result SENT = { "rows": [ {"json": _row_data(row), "insertId": str(i)} for i, row in enumerate(ROWS) ] } with mock.patch("uuid.uuid4", side_effect=map(str, range(len(ROWS)))): # Test with using string IDs for the table. errors = client.insert_rows( "{}.{}".format(self.DS_ID, self.TABLE_ID), ROWS, selected_fields=schema ) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once() _, req = conn.api_request.call_args self.assertEqual(req["method"], "POST") self.assertEqual(req["path"], "/%s" % PATH) self.assertEqual(req["data"], SENT) self.assertEqual(req["timeout"], DEFAULT_TIMEOUT) def test_insert_rows_w_list_of_dictionaries(self): import datetime from google.cloud._helpers import UTC from google.cloud._helpers import _datetime_to_rfc3339 from google.cloud._helpers import _RFC3339_MICROS from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table WHEN_TS = 1437767599.006 WHEN = datetime.datetime.utcfromtimestamp(WHEN_TS).replace(tzinfo=UTC) PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), SchemaField("joined", "TIMESTAMP", mode="NULLABLE"), ] table = Table(self.TABLE_REF, schema=schema) ROWS = [ { "full_name": "Phred Phlyntstone", "age": 32, "joined": _datetime_to_rfc3339(WHEN), }, { "full_name": "Bharney Rhubble", "age": 33, "joined": WHEN + datetime.timedelta(seconds=1), }, { "full_name": "Wylma Phlyntstone", "age": 29, "joined": WHEN + datetime.timedelta(seconds=2), }, {"full_name": "Bhettye Rhubble", "age": 27, "joined": None}, ] def _row_data(row): joined = row["joined"] if joined is None: row = copy.deepcopy(row) del row["joined"] elif isinstance(joined, datetime.datetime): row["joined"] = joined.strftime(_RFC3339_MICROS) row["age"] = str(row["age"]) return row SENT = { "rows": [ {"json": _row_data(row), "insertId": str(i)} for i, row in enumerate(ROWS) ] } with mock.patch("uuid.uuid4", side_effect=map(str, range(len(ROWS)))): errors = client.insert_rows(table, ROWS) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once_with( method="POST", path="/%s" % PATH, data=SENT, timeout=DEFAULT_TIMEOUT ) def test_insert_rows_w_list_of_Rows(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table from google.cloud.bigquery.table import Row PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), ] table = Table(self.TABLE_REF, schema=schema) f2i = {"full_name": 0, "age": 1} ROWS = [ Row(("Phred Phlyntstone", 32), f2i), Row(("Bharney Rhubble", 33), f2i), Row(("Wylma Phlyntstone", 29), f2i), Row(("Bhettye Rhubble", 27), f2i), ] def _row_data(row): return {"full_name": row[0], "age": str(row[1])} SENT = { "rows": [ {"json": _row_data(row), "insertId": str(i)} for i, row in enumerate(ROWS) ] } with mock.patch("uuid.uuid4", side_effect=map(str, range(len(ROWS)))): errors = client.insert_rows(table, ROWS) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once_with( method="POST", path="/%s" % PATH, data=SENT, timeout=DEFAULT_TIMEOUT ) def test_insert_rows_w_skip_invalid_and_ignore_unknown(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) RESPONSE = { "insertErrors": [ { "index": 1, "errors": [ { "reason": "REASON", "location": "LOCATION", "debugInfo": "INFO", "message": "MESSAGE", } ], } ] } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(RESPONSE) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), SchemaField("voter", "BOOLEAN", mode="NULLABLE"), ] table = Table(self.TABLE_REF, schema=schema) ROWS = [ ("Phred Phlyntstone", 32, True), ("Bharney Rhubble", 33, False), ("Wylma Phlyntstone", 29, True), ("Bhettye Rhubble", 27, True), ] def _row_data(row): return { "full_name": row[0], "age": str(row[1]), "voter": row[2] and "true" or "false", } SENT = { "skipInvalidRows": True, "ignoreUnknownValues": True, "templateSuffix": "20160303", "rows": [ {"insertId": index, "json": _row_data(row)} for index, row in enumerate(ROWS) ], } errors = client.insert_rows( table, ROWS, row_ids=[index for index, _ in enumerate(ROWS)], skip_invalid_rows=True, ignore_unknown_values=True, template_suffix="20160303", ) self.assertEqual(len(errors), 1) self.assertEqual(errors[0]["index"], 1) self.assertEqual(len(errors[0]["errors"]), 1) self.assertEqual( errors[0]["errors"][0], RESPONSE["insertErrors"][0]["errors"][0] ) conn.api_request.assert_called_once_with( method="POST", path="/%s" % PATH, data=SENT, timeout=DEFAULT_TIMEOUT ) def test_insert_rows_w_repeated_fields(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) color = SchemaField("color", "STRING", mode="REPEATED") items = SchemaField("items", "INTEGER", mode="REPEATED") score = SchemaField("score", "INTEGER") times = SchemaField("times", "TIMESTAMP", mode="REPEATED") distances = SchemaField("distances", "FLOAT", mode="REPEATED") structs = SchemaField( "structs", "RECORD", mode="REPEATED", fields=[score, times, distances] ) table = Table(self.TABLE_REF, schema=[color, items, structs]) ROWS = [ ( ["red", "green"], [1, 2], [ ( 12, [ datetime.datetime( 2018, 12, 1, 12, 0, 0, tzinfo=datetime.timezone.utc ), datetime.datetime( 2018, 12, 1, 13, 0, 0, tzinfo=datetime.timezone.utc ), ], [1.25, 2.5], ), { "score": 13, "times": [ datetime.datetime( 2018, 12, 2, 12, 0, 0, tzinfo=datetime.timezone.utc ), datetime.datetime( 2018, 12, 2, 13, 0, 0, tzinfo=datetime.timezone.utc ), ], "distances": [-1.25, -2.5], }, ], ), {"color": None, "items": [], "structs": [(None, [], [3.5])]}, ] SENT = { "rows": [ { "json": { "color": ["red", "green"], "items": ["1", "2"], "structs": [ { "score": "12", "times": [ "2018-12-01T12:00:00.000000Z", "2018-12-01T13:00:00.000000Z", ], "distances": [1.25, 2.5], }, { "score": "13", "times": [ "2018-12-02T12:00:00.000000Z", "2018-12-02T13:00:00.000000Z", ], "distances": [-1.25, -2.5], }, ], }, "insertId": "0", }, { "json": { "items": [], "structs": [{"times": [], "distances": [3.5]}], }, "insertId": "1", }, ] } with mock.patch("uuid.uuid4", side_effect=map(str, range(len(ROWS)))): errors = client.insert_rows(table, ROWS) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once_with( method="POST", path="/%s" % PATH, data=SENT, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_w_record_schema(self): from google.cloud.bigquery.schema import SchemaField PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) full_name = SchemaField("full_name", "STRING", mode="REQUIRED") area_code = SchemaField("area_code", "STRING", "REQUIRED") local_number = SchemaField("local_number", "STRING", "REQUIRED") rank = SchemaField("rank", "INTEGER", "REQUIRED") phone = SchemaField( "phone", "RECORD", mode="NULLABLE", fields=[area_code, local_number, rank] ) ROWS = [ ( "Phred Phlyntstone", {"area_code": "800", "local_number": "555-1212", "rank": 1}, ), ("Bharney Rhubble", ("877", "768-5309", 2)), ("Wylma Phlyntstone", None), ] SENT = { "rows": [ { "json": { "full_name": "Phred Phlyntstone", "phone": { "area_code": "800", "local_number": "555-1212", "rank": "1", }, }, "insertId": "0", }, { "json": { "full_name": "Bharney Rhubble", "phone": { "area_code": "877", "local_number": "768-5309", "rank": "2", }, }, "insertId": "1", }, {"json": {"full_name": "Wylma Phlyntstone"}, "insertId": "2"}, ] } with mock.patch("uuid.uuid4", side_effect=map(str, range(len(ROWS)))): errors = client.insert_rows( self.TABLE_REF, ROWS, selected_fields=[full_name, phone] ) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once_with( method="POST", path="/%s" % PATH, data=SENT, timeout=DEFAULT_TIMEOUT ) def test_insert_rows_w_explicit_none_insert_ids(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table PATH = "projects/{}/datasets/{}/tables/{}/insertAll".format( self.PROJECT, self.DS_ID, self.TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), ] table = Table(self.TABLE_REF, schema=schema) ROWS = [ {"full_name": "Phred Phlyntstone", "age": 32}, {"full_name": "Bharney Rhubble", "age": 33}, ] def _row_data(row): row["age"] = str(row["age"]) return row SENT = {"rows": [{"json": _row_data(row), "insertId": None} for row in ROWS]} errors = client.insert_rows(table, ROWS, row_ids=[None] * len(ROWS)) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once_with( method="POST", path="/{}".format(PATH), data=SENT, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_errors(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table ROWS = [ ("Phred Phlyntstone", 32, True), ("Bharney Rhubble", 33, False), ("Wylma Phlyntstone", 29, True), ("Bhettye Rhubble", 27, True), ] creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) # table ref with no selected fields with self.assertRaises(ValueError): client.insert_rows(self.TABLE_REF, ROWS) # table with no schema with self.assertRaises(ValueError): client.insert_rows(Table(self.TABLE_REF), ROWS) # neither Table nor TableReference with self.assertRaises(TypeError): client.insert_rows(1, ROWS) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), ] table = Table(self.TABLE_REF, schema=schema) # rows is just a dict with self.assertRaises(TypeError): client.insert_rows(table, {"full_name": "value"}) def test_insert_rows_w_numeric(self): from google.cloud.bigquery import table from google.cloud.bigquery.schema import SchemaField project = "PROJECT" ds_id = "DS_ID" table_id = "TABLE_ID" creds = _make_credentials() http = object() client = self._make_one(project=project, credentials=creds, _http=http) table_ref = DatasetReference(project, ds_id).table(table_id) rows = [ ("Savings", decimal.Decimal("23.47")), ("Checking", decimal.Decimal("1.98")), ("Mortgage", decimal.Decimal("-12345678909.87654321")), ] schemas = [ [SchemaField("account", "STRING"), SchemaField("balance", "NUMERIC")], [SchemaField("account", "STRING"), SchemaField("balance", "BIGNUMERIC")], ] for schema in schemas: conn = client._connection = make_connection({}) insert_table = table.Table(table_ref, schema=schema) with mock.patch("uuid.uuid4", side_effect=map(str, range(len(rows)))): errors = client.insert_rows(insert_table, rows) self.assertEqual(len(errors), 0) rows_json = [ {"account": "Savings", "balance": "23.47"}, {"account": "Checking", "balance": "1.98"}, {"account": "Mortgage", "balance": "-12345678909.87654321"}, ] sent = { "rows": [ {"json": row, "insertId": str(i)} for i, row in enumerate(rows_json) ] } conn.api_request.assert_called_once_with( method="POST", path="/projects/{}/datasets/{}/tables/{}/insertAll".format( project, ds_id, table_id ), data=sent, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_from_dataframe(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table API_PATH = "/projects/{}/datasets/{}/tables/{}/insertAll".format( self.PROJECT, self.DS_ID, self.TABLE_REF.table_id ) data = [ { "name": "Little One", "age": 10, "adult": False, "bdate": datetime.date(2011, 1, 2), "btime": datetime.time(19, 1, 10), }, { "name": "Young Gun", "age": 20, "adult": True, "bdate": datetime.date(2001, 1, 2), "btime": datetime.time(19, 1, 20), }, { "name": "Dad", "age": 30, "adult": True, "bdate": datetime.date(1991, 1, 2), "btime": datetime.time(19, 1, 30), }, { "name": "Stranger", "age": 40, "adult": True, "bdate": datetime.date(1981, 1, 2), "btime": datetime.time(19, 1, 40), }, ] dataframe = pandas.DataFrame(data) dataframe["bdate"] = dataframe["bdate"].astype("dbdate") dataframe["btime"] = dataframe["btime"].astype("dbtime") # create client creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}, {}) # create table schema = [ SchemaField("name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), SchemaField("adult", "BOOLEAN", mode="REQUIRED"), SchemaField("bdata", "DATE", mode="REQUIRED"), SchemaField("btime", "TIME", mode="REQUIRED"), ] table = Table(self.TABLE_REF, schema=schema) with mock.patch("uuid.uuid4", side_effect=map(str, range(len(dataframe)))): error_info = client.insert_rows_from_dataframe( table, dataframe, chunk_size=3, timeout=7.5 ) self.assertEqual(len(error_info), 2) for chunk_errors in error_info: assert chunk_errors == [] for row in data: row["age"] = str(row["age"]) row["adult"] = str(row["adult"]).lower() row["bdate"] = row["bdate"].isoformat() row["btime"] = row["btime"].isoformat() rows = [dict(insertId=str(i), json=row) for i, row in enumerate(data)] EXPECTED_SENT_DATA = [dict(rows=rows[:3]), dict(rows=rows[3:])] actual_calls = conn.api_request.call_args_list for call, expected_data in itertools.zip_longest( actual_calls, EXPECTED_SENT_DATA ): expected_call = mock.call( method="POST", path=API_PATH, data=expected_data, timeout=7.5 ) assert call == expected_call def test_insert_rows_from_dataframe_nan(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table API_PATH = "/projects/{}/datasets/{}/tables/{}/insertAll".format( self.PROJECT, self.DS_ID, self.TABLE_REF.table_id ) dataframe = pandas.DataFrame( { "str_col": ["abc", "def", float("NaN"), "jkl"], "int_col": [1, float("NaN"), 3, 4], "float_col": [float("NaN"), 0.25, 0.5, 0.125], } ) # create client creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}, {}) # create table schema = [ SchemaField("str_col", "STRING"), SchemaField("int_col", "INTEGER"), SchemaField("float_col", "FLOAT"), ] table = Table(self.TABLE_REF, schema=schema) with mock.patch("uuid.uuid4", side_effect=map(str, range(len(dataframe)))): error_info = client.insert_rows_from_dataframe( table, dataframe, chunk_size=3, timeout=7.5 ) self.assertEqual(len(error_info), 2) for chunk_errors in error_info: assert chunk_errors == [] EXPECTED_SENT_DATA = [ { "rows": [ {"insertId": "0", "json": {"str_col": "abc", "int_col": 1}}, {"insertId": "1", "json": {"str_col": "def", "float_col": 0.25}}, {"insertId": "2", "json": {"int_col": 3, "float_col": 0.5}}, ] }, { "rows": [ { "insertId": "3", "json": {"str_col": "jkl", "int_col": 4, "float_col": 0.125}, } ] }, ] actual_calls = conn.api_request.call_args_list for call, expected_data in itertools.zip_longest( actual_calls, EXPECTED_SENT_DATA ): expected_call = mock.call( method="POST", path=API_PATH, data=expected_data, timeout=7.5 ) assert call == expected_call def test_insert_rows_from_dataframe_many_columns(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table API_PATH = "/projects/{}/datasets/{}/tables/{}/insertAll".format( self.PROJECT, self.DS_ID, self.TABLE_REF.table_id ) N_COLUMNS = 256 # should be >= 256 dataframe = pandas.DataFrame( [{"foo_{}".format(i): "bar_{}".format(i) for i in range(N_COLUMNS)}] ) # create client creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}, {}) # create table schema = [SchemaField("foo_{}".format(i), "STRING") for i in range(N_COLUMNS)] table = Table(self.TABLE_REF, schema=schema) with mock.patch("uuid.uuid4", side_effect=map(str, range(len(dataframe)))): error_info = client.insert_rows_from_dataframe( table, dataframe, chunk_size=3 ) assert len(error_info) == 1 assert error_info[0] == [] EXPECTED_SENT_DATA = { "rows": [ { "insertId": "0", "json": { "foo_{}".format(i): "bar_{}".format(i) for i in range(N_COLUMNS) }, } ] } expected_call = mock.call( method="POST", path=API_PATH, data=EXPECTED_SENT_DATA, timeout=DEFAULT_TIMEOUT, ) actual_calls = conn.api_request.call_args_list assert len(actual_calls) == 1 assert actual_calls[0] == expected_call def test_insert_rows_from_dataframe_w_explicit_none_insert_ids(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table API_PATH = "/projects/{}/datasets/{}/tables/{}/insertAll".format( self.PROJECT, self.DS_ID, self.TABLE_REF.table_id ) dataframe = pandas.DataFrame( [ {"name": "Little One", "adult": False}, {"name": "Young Gun", "adult": True}, ] ) # create client creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}, {}) # create table schema = [ SchemaField("name", "STRING", mode="REQUIRED"), SchemaField("adult", "BOOLEAN", mode="REQUIRED"), ] table = Table(self.TABLE_REF, schema=schema) error_info = client.insert_rows_from_dataframe( table, dataframe, row_ids=[None] * len(dataframe) ) self.assertEqual(len(error_info), 1) assert error_info[0] == [] # no chunk errors EXPECTED_SENT_DATA = { "rows": [ {"insertId": None, "json": {"name": "Little One", "adult": "false"}}, {"insertId": None, "json": {"name": "Young Gun", "adult": "true"}}, ] } actual_calls = conn.api_request.call_args_list assert len(actual_calls) == 1 assert actual_calls[0] == mock.call( method="POST", path=API_PATH, data=EXPECTED_SENT_DATA, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_json_default_behavior(self): from google.cloud.bigquery.dataset import DatasetReference from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table PROJECT = "PROJECT" DS_ID = "DS_ID" TABLE_ID = "TABLE_ID" PATH = "projects/%s/datasets/%s/tables/%s/insertAll" % ( PROJECT, DS_ID, TABLE_ID, ) creds = _make_credentials() http = object() client = self._make_one(project=PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection({}) table_ref = DatasetReference(PROJECT, DS_ID).table(TABLE_ID) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), SchemaField("joined", "TIMESTAMP", mode="NULLABLE"), ] table = Table(table_ref, schema=schema) ROWS = [ { "full_name": "Phred Phlyntstone", "age": "32", "joined": "2015-07-24T19:53:19.006000Z", }, {"full_name": "Bharney Rhubble", "age": "33", "joined": 1437767600.006}, {"full_name": "Wylma Phlyntstone", "age": "29", "joined": 1437767601.006}, {"full_name": "Bhettye Rhubble", "age": "27", "joined": None}, ] SENT = { "rows": [{"json": row, "insertId": str(i)} for i, row in enumerate(ROWS)] } with mock.patch("uuid.uuid4", side_effect=map(str, range(len(ROWS)))): errors = client.insert_rows_json(table, ROWS, timeout=7.5) self.assertEqual(len(errors), 0) conn.api_request.assert_called_once_with( method="POST", path="/%s" % PATH, data=SENT, timeout=7.5, ) def test_insert_rows_json_w_explicitly_requested_autogenerated_insert_ids(self): from google.cloud.bigquery import AutoRowIDs rows = [{"col1": "val1"}, {"col2": "val2"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) conn = client._connection = make_connection({}) uuid_patcher = mock.patch("uuid.uuid4", side_effect=map(str, range(len(rows)))) with uuid_patcher: errors = client.insert_rows_json( "proj.dset.tbl", rows, row_ids=AutoRowIDs.GENERATE_UUID ) self.assertEqual(len(errors), 0) # Check row data sent to the backend. expected_row_data = { "rows": [ {"json": {"col1": "val1"}, "insertId": "0"}, {"json": {"col2": "val2"}, "insertId": "1"}, ] } conn.api_request.assert_called_once_with( method="POST", path="/projects/proj/datasets/dset/tables/tbl/insertAll", data=expected_row_data, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_json_w_explicitly_disabled_insert_ids(self): from google.cloud.bigquery import AutoRowIDs rows = [{"col1": "val1"}, {"col2": "val2"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) conn = client._connection = make_connection({}) errors = client.insert_rows_json( "proj.dset.tbl", rows, row_ids=AutoRowIDs.DISABLED, ) self.assertEqual(len(errors), 0) expected_row_data = { "rows": [ {"json": {"col1": "val1"}, "insertId": None}, {"json": {"col2": "val2"}, "insertId": None}, ] } conn.api_request.assert_called_once_with( method="POST", path="/projects/proj/datasets/dset/tables/tbl/insertAll", data=expected_row_data, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_json_with_iterator_row_ids(self): rows = [{"col1": "val1"}, {"col2": "val2"}, {"col3": "val3"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) conn = client._connection = make_connection({}) row_ids_iter = map(str, itertools.count(42)) errors = client.insert_rows_json("proj.dset.tbl", rows, row_ids=row_ids_iter) self.assertEqual(len(errors), 0) expected_row_data = { "rows": [ {"json": {"col1": "val1"}, "insertId": "42"}, {"json": {"col2": "val2"}, "insertId": "43"}, {"json": {"col3": "val3"}, "insertId": "44"}, ] } conn.api_request.assert_called_once_with( method="POST", path="/projects/proj/datasets/dset/tables/tbl/insertAll", data=expected_row_data, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_json_with_non_iterable_row_ids(self): rows = [{"col1": "val1"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) client._connection = make_connection({}) with self.assertRaises(TypeError) as exc: client.insert_rows_json("proj.dset.tbl", rows, row_ids=object()) err_msg = str(exc.exception) self.assertIn("row_ids", err_msg) self.assertIn("iterable", err_msg) def test_insert_rows_json_with_too_few_row_ids(self): rows = [{"col1": "val1"}, {"col2": "val2"}, {"col3": "val3"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) client._connection = make_connection({}) insert_ids = ["10", "20"] error_msg_pattern = "row_ids did not generate enough IDs.*index 2" with self.assertRaisesRegex(ValueError, error_msg_pattern): client.insert_rows_json("proj.dset.tbl", rows, row_ids=insert_ids) def test_insert_rows_json_w_explicit_none_insert_ids(self): rows = [{"col1": "val1"}, {"col2": "val2"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) conn = client._connection = make_connection({}) errors = client.insert_rows_json( "proj.dset.tbl", rows, row_ids=[None] * len(rows), ) self.assertEqual(len(errors), 0) expected = {"rows": [{"json": row, "insertId": None} for row in rows]} conn.api_request.assert_called_once_with( method="POST", path="/projects/proj/datasets/dset/tables/tbl/insertAll", data=expected, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_json_w_none_insert_ids_sequence(self): rows = [{"col1": "val1"}, {"col2": "val2"}] creds = _make_credentials() http = object() client = self._make_one( project="default-project", credentials=creds, _http=http ) conn = client._connection = make_connection({}) uuid_patcher = mock.patch("uuid.uuid4", side_effect=map(str, range(len(rows)))) with warnings.catch_warnings(record=True) as warned, uuid_patcher: errors = client.insert_rows_json("proj.dset.tbl", rows, row_ids=None) self.assertEqual(len(errors), 0) # Passing row_ids=None should have resulted in a deprecation warning. matches = [ warning for warning in warned if issubclass(warning.category, DeprecationWarning) and "row_ids" in str(warning) and "AutoRowIDs.GENERATE_UUID" in str(warning) ] assert matches, "The expected deprecation warning was not raised." # Check row data sent to the backend. expected_row_data = { "rows": [ {"json": {"col1": "val1"}, "insertId": "0"}, {"json": {"col2": "val2"}, "insertId": "1"}, ] } conn.api_request.assert_called_once_with( method="POST", path="/projects/proj/datasets/dset/tables/tbl/insertAll", data=expected_row_data, timeout=DEFAULT_TIMEOUT, ) def test_insert_rows_w_wrong_arg(self): from google.cloud.bigquery.dataset import DatasetReference from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table PROJECT = "PROJECT" DS_ID = "DS_ID" TABLE_ID = "TABLE_ID" ROW = {"full_name": "Bhettye Rhubble", "age": "27", "joined": None} creds = _make_credentials() client = self._make_one(project=PROJECT, credentials=creds, _http=object()) client._connection = make_connection({}) table_ref = DatasetReference(PROJECT, DS_ID).table(TABLE_ID) schema = [ SchemaField("full_name", "STRING", mode="REQUIRED"), SchemaField("age", "INTEGER", mode="REQUIRED"), SchemaField("joined", "TIMESTAMP", mode="NULLABLE"), ] table = Table(table_ref, schema=schema) with self.assertRaises(TypeError): client.insert_rows_json(table, ROW) def test_list_partitions(self): from google.cloud.bigquery.table import Table rows = 3 meta_info = _make_list_partitons_meta_info( self.PROJECT, self.DS_ID, self.TABLE_ID, rows ) data = { "totalRows": str(rows), "rows": [ {"f": [{"v": "20180101"}]}, {"f": [{"v": "20180102"}]}, {"f": [{"v": "20180103"}]}, ], } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client._connection = make_connection(meta_info, data) table = Table(self.TABLE_REF) partition_list = client.list_partitions(table) self.assertEqual(len(partition_list), rows) self.assertIn("20180102", partition_list) def test_list_partitions_with_string_id(self): meta_info = _make_list_partitons_meta_info( self.PROJECT, self.DS_ID, self.TABLE_ID, 0 ) creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client._connection = make_connection(meta_info, {}) partition_list = client.list_partitions( "{}.{}".format(self.DS_ID, self.TABLE_ID) ) self.assertEqual(len(partition_list), 0) def test_list_rows(self): import datetime from google.cloud._helpers import UTC from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table from google.cloud.bigquery.table import Row PATH = "projects/%s/datasets/%s/tables/%s/data" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) WHEN_TS = 1437767599006000 WHEN = datetime.datetime.utcfromtimestamp(WHEN_TS / 1e6).replace(tzinfo=UTC) WHEN_1 = WHEN + datetime.timedelta(microseconds=1) WHEN_2 = WHEN + datetime.timedelta(microseconds=2) ROWS = 1234 TOKEN = "TOKEN" DATA = { "totalRows": str(ROWS), "pageToken": TOKEN, "rows": [ {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}, {"v": WHEN_TS}]}, {"f": [{"v": "Bharney Rhubble"}, {"v": "33"}, {"v": WHEN_TS + 1}]}, {"f": [{"v": "Wylma Phlyntstone"}, {"v": "29"}, {"v": WHEN_TS + 2}]}, {"f": [{"v": "Bhettye Rhubble"}, {"v": None}, {"v": None}]}, ], } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(DATA, DATA) full_name = SchemaField("full_name", "STRING", mode="REQUIRED") age = SchemaField("age", "INTEGER", mode="NULLABLE") joined = SchemaField("joined", "TIMESTAMP", mode="NULLABLE") table = Table(self.TABLE_REF, schema=[full_name, age, joined]) table._properties["location"] = "us-central1" table._properties["numRows"] = 7 iterator = client.list_rows(table, timeout=7.5) # Check that initial RowIterator is populated from the table metadata. self.assertIsNone(iterator.job_id) self.assertEqual(iterator.location, "us-central1") self.assertEqual(iterator.project, table.project) self.assertIsNone(iterator.query_id) self.assertEqual(iterator.total_rows, 7) page = next(iterator.pages) rows = list(page) # Check that total_rows is updated based on API response. self.assertEqual(iterator.total_rows, ROWS) f2i = {"full_name": 0, "age": 1, "joined": 2} self.assertEqual(len(rows), 4) self.assertEqual(rows[0], Row(("Phred Phlyntstone", 32, WHEN), f2i)) self.assertEqual(rows[1], Row(("Bharney Rhubble", 33, WHEN_1), f2i)) self.assertEqual(rows[2], Row(("Wylma Phlyntstone", 29, WHEN_2), f2i)) self.assertEqual(rows[3], Row(("Bhettye Rhubble", None, None), f2i)) self.assertEqual(iterator.next_page_token, TOKEN) conn.api_request.assert_called_once_with( method="GET", path="/%s" % PATH, query_params={"formatOptions.useInt64Timestamp": True}, timeout=7.5, ) def test_list_rows_w_start_index_w_page_size(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table from google.cloud.bigquery.table import Row PATH = "projects/%s/datasets/%s/tables/%s/data" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) page_1 = { "totalRows": 4, "pageToken": "some-page-token", "rows": [ {"f": [{"v": "Phred Phlyntstone"}]}, {"f": [{"v": "Bharney Rhubble"}]}, ], } page_2 = { "totalRows": 4, "rows": [ {"f": [{"v": "Wylma Phlyntstone"}]}, {"f": [{"v": "Bhettye Rhubble"}]}, ], } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(page_1, page_2) full_name = SchemaField("full_name", "STRING", mode="REQUIRED") table = Table(self.TABLE_REF, schema=[full_name]) iterator = client.list_rows(table, max_results=4, page_size=2, start_index=1) pages = iterator.pages rows = list(next(pages)) extra_params = iterator.extra_params f2i = {"full_name": 0} self.assertEqual(len(rows), 2) self.assertEqual(rows[0], Row(("Phred Phlyntstone",), f2i)) self.assertEqual(rows[1], Row(("Bharney Rhubble",), f2i)) rows = list(next(pages)) self.assertEqual(len(rows), 2) self.assertEqual(rows[0], Row(("Wylma Phlyntstone",), f2i)) self.assertEqual(rows[1], Row(("Bhettye Rhubble",), f2i)) self.assertEqual( extra_params, {"startIndex": 1, "formatOptions.useInt64Timestamp": True} ) conn.api_request.assert_has_calls( [ mock.call( method="GET", path="/%s" % PATH, query_params={ "startIndex": 1, "maxResults": 2, "formatOptions.useInt64Timestamp": True, }, timeout=DEFAULT_TIMEOUT, ), mock.call( method="GET", path="/%s" % PATH, query_params={ "pageToken": "some-page-token", "maxResults": 2, "formatOptions.useInt64Timestamp": True, }, timeout=DEFAULT_TIMEOUT, ), ] ) def test_list_rows_empty_table(self): response = {"totalRows": "0", "rows": []} creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client._connection = make_connection(response, response) # Table that has no schema because it's an empty table. rows = client.list_rows( # Test with using a string for the table ID. "{}.{}.{}".format( self.TABLE_REF.project, self.TABLE_REF.dataset_id, self.TABLE_REF.table_id, ), selected_fields=[], ) self.assertIsNone(rows.job_id) self.assertIsNone(rows.location) self.assertEqual(rows.project, self.TABLE_REF.project) self.assertIsNone(rows.query_id) # When a table reference / string and selected_fields is provided, # total_rows can't be populated until iteration starts. self.assertIsNone(rows.total_rows) self.assertEqual(tuple(rows), ()) self.assertEqual(rows.total_rows, 0) def test_list_rows_query_params(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) table = Table( self.TABLE_REF, schema=[SchemaField("age", "INTEGER", mode="NULLABLE")] ) tests = [ ({}, {}), ({"start_index": 1}, {"startIndex": 1}), ({"max_results": 2}, {"maxResults": 2}), ({"start_index": 1, "max_results": 2}, {"startIndex": 1, "maxResults": 2}), ] conn = client._connection = make_connection(*len(tests) * [{}]) for i, test in enumerate(tests): iterator = client.list_rows(table, **test[0]) next(iterator.pages) req = conn.api_request.call_args_list[i] test[1]["formatOptions.useInt64Timestamp"] = True self.assertEqual(req[1]["query_params"], test[1], "for kwargs %s" % test[0]) def test_list_rows_w_numeric(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table resource = { "totalRows": 3, "rows": [ {"f": [{"v": "-1.23456789"}, {"v": "-123456789.987654321"}]}, {"f": [{"v": None}, {"v": "3.141592653589793238462643383279502884"}]}, {"f": [{"v": "2718281828459045235360287471.352662497"}, {"v": None}]}, ], } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) client._connection = make_connection(resource) schema = [ SchemaField("num", "NUMERIC"), SchemaField("bignum", "BIGNUMERIC"), ] table = Table(self.TABLE_REF, schema=schema) iterator = client.list_rows(table) rows = list(iterator) self.assertEqual(len(rows), 3) self.assertEqual(rows[0]["num"], decimal.Decimal("-1.23456789")) self.assertEqual(rows[0]["bignum"], decimal.Decimal("-123456789.987654321")) self.assertIsNone(rows[1]["num"]) self.assertEqual( rows[1]["bignum"], decimal.Decimal("3.141592653589793238462643383279502884") ) self.assertEqual( rows[2]["num"], decimal.Decimal("2718281828459045235360287471.352662497") ) self.assertIsNone(rows[2]["bignum"]) def test_list_rows_repeated_fields(self): from google.cloud.bigquery.schema import SchemaField PATH = "projects/%s/datasets/%s/tables/%s/data" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) ROWS = 1234 TOKEN = "TOKEN" DATA = { "totalRows": ROWS, "pageToken": TOKEN, "rows": [ { "f": [ {"v": [{"v": "red"}, {"v": "green"}]}, { "v": [ { "v": { "f": [ {"v": [{"v": "1"}, {"v": "2"}]}, {"v": [{"v": "3.1415"}, {"v": "1.414"}]}, ] } } ] }, ] } ], } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(DATA) color = SchemaField("color", "STRING", mode="REPEATED") index = SchemaField("index", "INTEGER", "REPEATED") score = SchemaField("score", "FLOAT", "REPEATED") struct = SchemaField("struct", "RECORD", mode="REPEATED", fields=[index, score]) iterator = client.list_rows(self.TABLE_REF, selected_fields=[color, struct]) page = next(iterator.pages) rows = list(page) total_rows = iterator.total_rows page_token = iterator.next_page_token self.assertEqual(len(rows), 1) self.assertEqual(rows[0][0], ["red", "green"]) self.assertEqual(rows[0][1], [{"index": [1, 2], "score": [3.1415, 1.414]}]) self.assertEqual(total_rows, ROWS) self.assertEqual(page_token, TOKEN) conn.api_request.assert_called_once_with( method="GET", path="/%s" % PATH, query_params={ "selectedFields": "color,struct", "formatOptions.useInt64Timestamp": True, }, timeout=DEFAULT_TIMEOUT, ) def test_list_rows_w_record_schema(self): from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.table import Table PATH = "projects/%s/datasets/%s/tables/%s/data" % ( self.PROJECT, self.DS_ID, self.TABLE_ID, ) ROWS = 1234 TOKEN = "TOKEN" DATA = { "totalRows": ROWS, "pageToken": TOKEN, "rows": [ { "f": [ {"v": "Phred Phlyntstone"}, {"v": {"f": [{"v": "800"}, {"v": "555-1212"}, {"v": 1}]}}, ] }, { "f": [ {"v": "Bharney Rhubble"}, {"v": {"f": [{"v": "877"}, {"v": "768-5309"}, {"v": 2}]}}, ] }, {"f": [{"v": "Wylma Phlyntstone"}, {"v": None}]}, ], } creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(DATA) full_name = SchemaField("full_name", "STRING", mode="REQUIRED") area_code = SchemaField("area_code", "STRING", "REQUIRED") local_number = SchemaField("local_number", "STRING", "REQUIRED") rank = SchemaField("rank", "INTEGER", "REQUIRED") phone = SchemaField( "phone", "RECORD", mode="NULLABLE", fields=[area_code, local_number, rank] ) table = Table(self.TABLE_REF, schema=[full_name, phone]) iterator = client.list_rows(table) page = next(iterator.pages) rows = list(page) total_rows = iterator.total_rows page_token = iterator.next_page_token self.assertEqual(len(rows), 3) self.assertEqual(rows[0][0], "Phred Phlyntstone") self.assertEqual( rows[0][1], {"area_code": "800", "local_number": "555-1212", "rank": 1} ) self.assertEqual(rows[1][0], "Bharney Rhubble") self.assertEqual( rows[1][1], {"area_code": "877", "local_number": "768-5309", "rank": 2} ) self.assertEqual(rows[2][0], "Wylma Phlyntstone") self.assertIsNone(rows[2][1]) self.assertEqual(total_rows, ROWS) self.assertEqual(page_token, TOKEN) conn.api_request.assert_called_once_with( method="GET", path="/%s" % PATH, query_params={"formatOptions.useInt64Timestamp": True}, timeout=DEFAULT_TIMEOUT, ) def test_list_rows_with_missing_schema(self): from google.cloud.bigquery.table import Table, TableListItem table_path = "/projects/{}/datasets/{}/tables/{}".format( self.PROJECT, self.DS_ID, self.TABLE_ID ) tabledata_path = "{}/data".format(table_path) table_list_item_data = { "id": "%s:%s:%s" % (self.PROJECT, self.DS_ID, self.TABLE_ID), "tableReference": { "projectId": self.PROJECT, "datasetId": self.DS_ID, "tableId": self.TABLE_ID, }, } table_data = copy.deepcopy(table_list_item_data) # Intentionally make wrong, since total_rows can update during iteration. table_data["numRows"] = 2 table_data["schema"] = { "fields": [ {"name": "name", "type": "STRING"}, {"name": "age", "type": "INTEGER"}, ] } rows_data = { "totalRows": 3, "pageToken": None, "rows": [ {"f": [{"v": "Phred Phlyntstone"}, {"v": "32"}]}, {"f": [{"v": "Bharney Rhubble"}, {"v": "31"}]}, {"f": [{"v": "Wylma Phlyntstone"}, {"v": None}]}, ], } creds = _make_credentials() http = object() schemaless_tables = ( "{}.{}".format(self.DS_ID, self.TABLE_ID), self.TABLE_REF, Table(self.TABLE_REF), TableListItem(table_list_item_data), ) for table in schemaless_tables: client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) conn = client._connection = make_connection(table_data, rows_data) row_iter = client.list_rows(table) conn.api_request.assert_called_once_with( method="GET", path=table_path, timeout=DEFAULT_TIMEOUT ) conn.api_request.reset_mock() self.assertEqual(row_iter.total_rows, 2, msg=repr(table)) rows = list(row_iter) conn.api_request.assert_called_once_with( method="GET", path=tabledata_path, query_params={"formatOptions.useInt64Timestamp": True}, timeout=DEFAULT_TIMEOUT, ) self.assertEqual(row_iter.total_rows, 3, msg=repr(table)) self.assertEqual(rows[0].name, "Phred Phlyntstone", msg=repr(table)) self.assertEqual(rows[1].age, 31, msg=repr(table)) self.assertIsNone(rows[2].age, msg=repr(table)) def test_list_rows_error(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) # neither Table nor TableReference with self.assertRaises(TypeError): client.list_rows(1) def test_context_manager_enter_returns_itself(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) with mock.patch.object(client, "close"), client as context_var: pass self.assertIs(client, context_var) def test_context_manager_exit_closes_client(self): creds = _make_credentials() http = object() client = self._make_one(project=self.PROJECT, credentials=creds, _http=http) fake_close = mock.Mock() with mock.patch.object(client, "close", fake_close): with client: pass fake_close.assert_called_once() class TestClientUpload(object): # NOTE: This is a "partner" to `TestClient` meant to test some of the # "load_table_from_file" portions of `Client`. It also uses # `pytest`-style tests rather than `unittest`-style. from google.cloud.bigquery.job import SourceFormat PROJECT = "project_id" TABLE_REF = DatasetReference(PROJECT, "test_dataset").table("test_table") LOCATION = "us-central" @classmethod def _make_client(cls, transport=None, location=None): from google.cloud.bigquery import _http from google.cloud.bigquery import client cl = client.Client( project=cls.PROJECT, credentials=_make_credentials(), _http=transport, location=location, ) cl._connection = mock.create_autospec(_http.Connection, instance=True) return cl @staticmethod def _make_response(status_code, content="", headers={}): """Make a mock HTTP response.""" import requests response = requests.Response() response.request = requests.Request("POST", "http://example.com").prepare() response._content = content.encode("utf-8") response.headers.update(headers) response.status_code = status_code return response @classmethod def _make_do_upload_patch(cls, client, method, resource={}, side_effect=None): """Patches the low-level upload helpers.""" if side_effect is None: side_effect = [ cls._make_response( http.client.OK, json.dumps(resource), {"Content-Type": "application/json"}, ) ] return mock.patch.object(client, method, side_effect=side_effect, autospec=True) EXPECTED_CONFIGURATION = { "jobReference": {"projectId": PROJECT, "jobId": "job_id"}, "configuration": { "load": { "sourceFormat": SourceFormat.CSV, "destinationTable": { "projectId": PROJECT, "datasetId": "test_dataset", "tableId": "test_table", }, } }, } @staticmethod def _make_file_obj(): return io.BytesIO(b"hello, is it me you're looking for?") def _make_gzip_file_obj(self, writable): if writable: return gzip.GzipFile(mode="w", fileobj=io.BytesIO()) else: return gzip.GzipFile(mode="r", fileobj=self._make_file_obj()) @staticmethod def _make_config(): from google.cloud.bigquery.job import LoadJobConfig from google.cloud.bigquery.job import SourceFormat config = LoadJobConfig() config.source_format = SourceFormat.CSV return config # High-level tests def test_load_table_from_file_resumable(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() file_obj = self._make_file_obj() job_config = self._make_config() original_config_copy = copy.deepcopy(job_config) do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", job_config=job_config, ) do_upload.assert_called_once_with( file_obj, self.EXPECTED_CONFIGURATION, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.EXPECTED_CONFIGURATION["jobReference"]["projectId"], ) # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_file_w_explicit_project(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() file_obj = self._make_file_obj() do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", project="other-project", location=self.LOCATION, job_config=self._make_config(), ) expected_resource = copy.deepcopy(self.EXPECTED_CONFIGURATION) expected_resource["jobReference"]["location"] = self.LOCATION expected_resource["jobReference"]["projectId"] = "other-project" do_upload.assert_called_once_with( file_obj, expected_resource, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project="other-project", ) def test_load_table_from_file_w_client_location(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client(location=self.LOCATION) file_obj = self._make_file_obj() do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, # Test with string for table ID. "{}.{}.{}".format( self.TABLE_REF.project, self.TABLE_REF.dataset_id, self.TABLE_REF.table_id, ), job_id="job_id", project="other-project", job_config=self._make_config(), ) expected_resource = copy.deepcopy(self.EXPECTED_CONFIGURATION) expected_resource["jobReference"]["location"] = self.LOCATION expected_resource["jobReference"]["projectId"] = "other-project" do_upload.assert_called_once_with( file_obj, expected_resource, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project="other-project", ) def test_load_table_from_file_resumable_metadata(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery.job import CreateDisposition from google.cloud.bigquery.job import WriteDisposition client = self._make_client() file_obj = self._make_file_obj() config = self._make_config() config.allow_jagged_rows = False config.allow_quoted_newlines = False config.create_disposition = CreateDisposition.CREATE_IF_NEEDED config.encoding = "utf8" config.field_delimiter = "," config.ignore_unknown_values = False config.max_bad_records = 0 config.quote_character = '"' config.skip_leading_rows = 1 config.write_disposition = WriteDisposition.WRITE_APPEND config.null_marker = r"\N" expected_config = { "jobReference": {"projectId": self.PROJECT, "jobId": "job_id"}, "configuration": { "load": { "destinationTable": { "projectId": self.TABLE_REF.project, "datasetId": self.TABLE_REF.dataset_id, "tableId": self.TABLE_REF.table_id, }, "sourceFormat": config.source_format, "allowJaggedRows": config.allow_jagged_rows, "allowQuotedNewlines": config.allow_quoted_newlines, "createDisposition": config.create_disposition, "encoding": config.encoding, "fieldDelimiter": config.field_delimiter, "ignoreUnknownValues": config.ignore_unknown_values, "maxBadRecords": config.max_bad_records, "quote": config.quote_character, "skipLeadingRows": str(config.skip_leading_rows), "writeDisposition": config.write_disposition, "nullMarker": config.null_marker, } }, } do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", expected_config ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", job_config=config ) do_upload.assert_called_once_with( file_obj, expected_config, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.EXPECTED_CONFIGURATION["jobReference"]["projectId"], ) def test_load_table_from_file_multipart(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() file_obj = self._make_file_obj() file_obj_size = 10 config = self._make_config() do_upload_patch = self._make_do_upload_patch( client, "_do_multipart_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", job_config=config, size=file_obj_size, ) do_upload.assert_called_once_with( file_obj, self.EXPECTED_CONFIGURATION, file_obj_size, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.PROJECT, ) def test_load_table_from_file_with_retries(self): client = self._make_client() file_obj = self._make_file_obj() num_retries = 20 do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, num_retries=num_retries, job_id="job_id", job_config=self._make_config(), ) do_upload.assert_called_once_with( file_obj, self.EXPECTED_CONFIGURATION, num_retries, DEFAULT_TIMEOUT, project=self.EXPECTED_CONFIGURATION["jobReference"]["projectId"], ) def test_load_table_from_file_with_rewind(self): client = self._make_client() file_obj = self._make_file_obj() file_obj.seek(2) with self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ): client.load_table_from_file(file_obj, self.TABLE_REF, rewind=True) assert file_obj.tell() == 0 def test_load_table_from_file_with_readable_gzip(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() gzip_file = self._make_gzip_file_obj(writable=False) do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( gzip_file, self.TABLE_REF, job_id="job_id", job_config=self._make_config(), ) do_upload.assert_called_once_with( gzip_file, self.EXPECTED_CONFIGURATION, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.EXPECTED_CONFIGURATION["jobReference"]["projectId"], ) def test_load_table_from_file_with_writable_gzip(self): client = self._make_client() gzip_file = self._make_gzip_file_obj(writable=True) with pytest.raises(ValueError): client.load_table_from_file( gzip_file, self.TABLE_REF, job_id="job_id", job_config=self._make_config(), ) def test_load_table_from_file_failure(self): from google.resumable_media import InvalidResponse from google.cloud import exceptions client = self._make_client() file_obj = self._make_file_obj() response = self._make_response( content="Someone is already in this spot.", status_code=http.client.CONFLICT ) do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", side_effect=InvalidResponse(response) ) with do_upload_patch, pytest.raises(exceptions.Conflict) as exc_info: client.load_table_from_file(file_obj, self.TABLE_REF, rewind=True) assert response.text in exc_info.value.message assert exc_info.value.errors == [] def test_load_table_from_file_bad_mode(self): client = self._make_client() file_obj = mock.Mock(spec=["mode"]) file_obj.mode = "x" with pytest.raises(ValueError): client.load_table_from_file(file_obj, self.TABLE_REF) def test_load_table_from_file_w_invalid_job_config(self): from google.cloud.bigquery import job client = self._make_client() gzip_file = self._make_gzip_file_obj(writable=True) config = job.QueryJobConfig() with pytest.raises(TypeError) as exc: client.load_table_from_file( gzip_file, self.TABLE_REF, job_id="job_id", job_config=config ) err_msg = str(exc.value) assert "Expected an instance of LoadJobConfig" in err_msg def test_load_table_from_file_w_explicit_job_config(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() file_obj = self._make_file_obj() job_config = self._make_config() job_config.create_session = True job_config.encoding = "UTF-8" do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", project=self.PROJECT, location=self.LOCATION, job_config=job_config, ) expected_resource = copy.deepcopy(self.EXPECTED_CONFIGURATION) expected_resource["jobReference"]["location"] = self.LOCATION expected_resource["jobReference"]["projectId"] = self.PROJECT expected_resource["configuration"]["load"]["createSession"] = True expected_resource["configuration"]["load"]["encoding"] = "UTF-8" do_upload.assert_called_once_with( file_obj, expected_resource, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.PROJECT, ) def test_load_table_from_file_w_explicit_job_config_override(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery.job import LoadJobConfig client = self._make_client() file_obj = self._make_file_obj() default_job_config = LoadJobConfig() default_job_config.create_session = True default_job_config.encoding = "ISO-8859-1" client.default_load_job_config = default_job_config job_config = self._make_config() job_config.create_session = False do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", project=self.PROJECT, location=self.LOCATION, job_config=job_config, ) expected_resource = copy.deepcopy(self.EXPECTED_CONFIGURATION) expected_resource["jobReference"]["location"] = self.LOCATION expected_resource["jobReference"]["projectId"] = self.PROJECT expected_resource["configuration"]["load"]["createSession"] = False expected_resource["configuration"]["load"]["encoding"] = "ISO-8859-1" do_upload.assert_called_once_with( file_obj, expected_resource, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.PROJECT, ) def test_load_table_from_file_w_default_load_config(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery.job import LoadJobConfig client = self._make_client() file_obj = self._make_file_obj() default_job_config = LoadJobConfig() default_job_config.encoding = "ISO-8859-1" client.default_load_job_config = default_job_config job_config = self._make_config() do_upload_patch = self._make_do_upload_patch( client, "_do_resumable_upload", self.EXPECTED_CONFIGURATION ) with do_upload_patch as do_upload: client.load_table_from_file( file_obj, self.TABLE_REF, job_id="job_id", project=self.PROJECT, location=self.LOCATION, job_config=job_config, ) expected_resource = copy.deepcopy(self.EXPECTED_CONFIGURATION) expected_resource["jobReference"]["location"] = self.LOCATION expected_resource["jobReference"]["projectId"] = self.PROJECT expected_resource["configuration"]["load"]["encoding"] = "ISO-8859-1" do_upload.assert_called_once_with( file_obj, expected_resource, _DEFAULT_NUM_RETRIES, DEFAULT_TIMEOUT, project=self.PROJECT, ) def test_load_table_from_dataframe(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import PolicyTagList, SchemaField client = self._make_client() records = [ {"id": 1, "age": 100, "accounts": [2, 3]}, {"id": 2, "age": 60, "accounts": [5]}, {"id": 3, "age": 40, "accounts": []}, ] # Mixup column order so that we can verify sent schema matches the # serialized order, not the table column order. column_order = ["age", "accounts", "id"] dataframe = pandas.DataFrame(records, columns=column_order) table_fields = { "id": SchemaField( "id", "INTEGER", mode="REQUIRED", description="integer column", policy_tags=PolicyTagList(names=("foo", "bar")), ), "age": SchemaField( "age", "INTEGER", mode="NULLABLE", description="age column", policy_tags=PolicyTagList(names=("baz",)), ), "accounts": SchemaField( "accounts", "INTEGER", mode="REPEATED", description="array column", ), } get_table_schema = [ table_fields["id"], table_fields["age"], table_fields["accounts"], ] get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock(schema=get_table_schema), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe(dataframe, self.TABLE_REF) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=None, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_file = load_table_from_file.mock_calls[0][1][1] assert sent_file.closed sent_config = load_table_from_file.mock_calls[0][2]["job_config"].to_api_repr()[ "load" ] assert sent_config["sourceFormat"] == job.SourceFormat.PARQUET for field_index, field in enumerate(sent_config["schema"]["fields"]): assert field["name"] == column_order[field_index] table_field = table_fields[field["name"]] assert field["name"] == table_field.name assert field["type"] == table_field.field_type assert field["mode"] == table_field.mode assert len(field.get("fields", [])) == len(table_field.fields) # Avoid accidentally updating policy tags when not explicitly included. # https://github.com/googleapis/python-bigquery/issues/981 # Also, avoid 403 if someone has permission to write to table but # not update policy tags by omitting policy tags we might have # received from a get table request. # https://github.com/googleapis/python-bigquery/pull/557 assert "policyTags" not in field # Omit unnecessary fields when they come from getting the table # (not passed in via job_config) assert "description" not in field def test_load_table_from_dataframe_w_client_location(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client(location=self.LOCATION) records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe(dataframe, self.TABLE_REF) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_file = load_table_from_file.mock_calls[0][1][1] assert sent_file.closed sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET def test_load_table_from_dataframe_w_custom_job_config_wihtout_source_format(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, ) original_config_copy = copy.deepcopy(job_config) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch as get_table: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) # no need to fetch and inspect table schema for WRITE_TRUNCATE jobs assert not get_table.called load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.write_disposition == job.WriteDisposition.WRITE_TRUNCATE # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_dataframe_w_custom_job_config_w_source_format(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.PARQUET, ) original_config_copy = copy.deepcopy(job_config) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch as get_table: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) # no need to fetch and inspect table schema for WRITE_TRUNCATE jobs assert not get_table.called load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.write_disposition == job.WriteDisposition.WRITE_TRUNCATE # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_dataframe_w_parquet_options_none(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.PARQUET, ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch as get_table: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) # no need to fetch and inspect table schema for WRITE_TRUNCATE jobs assert not get_table.called load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.parquet_options.enable_list_inference is True def test_load_table_from_dataframe_w_list_inference_none(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) parquet_options = ParquetOptions() job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.PARQUET, ) job_config.parquet_options = parquet_options original_config_copy = copy.deepcopy(job_config) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch as get_table: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) # no need to fetch and inspect table schema for WRITE_TRUNCATE jobs assert not get_table.called load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.parquet_options.enable_list_inference is None # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_dataframe_w_explicit_job_config_override(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) client.default_load_job_config = job.LoadJobConfig( encoding="ISO-8859-1", write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.PARQUET, ) job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_APPEND, source_format=job.SourceFormat.PARQUET, ) original_config_copy = copy.deepcopy(job_config) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.write_disposition == job.WriteDisposition.WRITE_APPEND assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.encoding == "ISO-8859-1" # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_dataframe_w_default_load_config(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) client.default_load_job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.PARQUET, ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.write_disposition == job.WriteDisposition.WRITE_TRUNCATE assert sent_config.source_format == job.SourceFormat.PARQUET def test_load_table_from_dataframe_w_list_inference_false(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) parquet_options = ParquetOptions() parquet_options.enable_list_inference = False job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.PARQUET, ) job_config.parquet_options = parquet_options original_config_copy = copy.deepcopy(job_config) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch as get_table: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) # no need to fetch and inspect table schema for WRITE_TRUNCATE jobs assert not get_table.called load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.parquet_options.enable_list_inference is False # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_dataframe_w_custom_job_config_w_wrong_source_format(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery import job client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.ORC, ) with pytest.raises(ValueError) as exc: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) assert "Got unexpected source_format:" in str(exc.value) def test_load_table_from_dataframe_w_automatic_schema(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() df_data = collections.OrderedDict( [ ("int_col", [1, 2, 3]), ("float_col", [1.0, 2.0, 3.0]), ("bool_col", [True, False, True]), ( "dt_col", pandas.Series( [ datetime.datetime(2010, 1, 2, 3, 44, 50), datetime.datetime(2011, 2, 3, 14, 50, 59), datetime.datetime(2012, 3, 14, 15, 16), ], dtype="datetime64[ns]", ), ), ( "ts_col", pandas.Series( [ datetime.datetime(2010, 1, 2, 3, 44, 50), datetime.datetime(2011, 2, 3, 14, 50, 59), datetime.datetime(2012, 3, 14, 15, 16), ], dtype="datetime64[ns]", ).dt.tz_localize(datetime.timezone.utc), ), ( "date_col", pandas.Series( [ datetime.date(2010, 1, 2), datetime.date(2011, 2, 3), datetime.date(2012, 3, 14), ], dtype="dbdate", ), ), ( "time_col", pandas.Series( [ datetime.time(3, 44, 50), datetime.time(14, 50, 59), datetime.time(15, 16), ], dtype="dbtime", ), ), ] ) dataframe = pandas.DataFrame(df_data, columns=df_data.keys()) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert tuple(sent_config.schema) == ( SchemaField("int_col", "INTEGER"), SchemaField("float_col", "FLOAT"), SchemaField("bool_col", "BOOLEAN"), SchemaField("dt_col", "DATETIME"), SchemaField("ts_col", "TIMESTAMP"), SchemaField("date_col", "DATE"), SchemaField("time_col", "TIME"), ) def test_load_table_from_dataframe_w_automatic_schema_detection_fails(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job client = self._make_client() df_data = [ [[{"name": "n1.1", "value": 1.1}, {"name": "n1.2", "value": 1.2}]], [[{"name": "n2.1", "value": 2.1}, {"name": "n2.2", "value": 2.2}]], ] dataframe = pandas.DataFrame(df_data, columns=["col_record_list"]) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) with load_patch as load_table_from_file, get_table_patch: with warnings.catch_warnings(record=True) as warned: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) # There should be a warning that schema detection failed. expected_warnings = [ warning for warning in warned if "schema could not be detected" in str(warning).lower() ] assert len(expected_warnings) == 1 assert issubclass( expected_warnings[0].category, (DeprecationWarning, PendingDeprecationWarning), ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.schema is None def test_load_table_from_dataframe_w_index_and_auto_schema(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() df_data = collections.OrderedDict( [("int_col", [10, 20, 30]), ("float_col", [1.0, 2.0, 3.0])] ) dataframe = pandas.DataFrame( df_data, index=pandas.Index(name="unique_name", data=["one", "two", "three"]), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[ SchemaField("int_col", "INTEGER"), SchemaField("float_col", "FLOAT"), SchemaField("unique_name", "STRING"), ] ), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET sent_schema = sorted(sent_config.schema, key=operator.attrgetter("name")) expected_sent_schema = [ SchemaField("float_col", "FLOAT"), SchemaField("int_col", "INTEGER"), SchemaField("unique_name", "STRING"), ] assert sent_schema == expected_sent_schema def test_load_table_from_dataframe_unknown_table(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch: # there should be no error client.load_table_from_dataframe(dataframe, self.TABLE_REF) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=None, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) def test_load_table_from_dataframe_w_nullable_int64_datatype(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() dataframe = pandas.DataFrame({"x": [1, 2, None, 4]}, dtype="Int64") load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock(schema=[SchemaField("x", "INT64", "NULLABLE")]), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert tuple(sent_config.schema) == ( SchemaField("x", "INT64", "NULLABLE", None), ) def test_load_table_from_dataframe_w_nullable_int64_datatype_automatic_schema(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() dataframe = pandas.DataFrame({"x": [1, 2, None, 4]}, dtype="Int64") load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert tuple(sent_config.schema) == ( SchemaField("x", "INT64", "NULLABLE", None), ) def test_load_table_from_dataframe_struct_fields(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [(3.14, {"foo": 1, "bar": 1})] dataframe = pandas.DataFrame( data=records, columns=["float_column", "struct_column"] ) schema = [ SchemaField("float_column", "FLOAT"), SchemaField( "struct_column", "RECORD", fields=[SchemaField("foo", "INTEGER"), SchemaField("bar", "INTEGER")], ), ] job_config = job.LoadJobConfig(schema=schema) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION, ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.schema == schema def test_load_table_from_dataframe_array_fields(self): """Test that a DataFrame with array columns can be uploaded correctly. See: https://github.com/googleapis/python-bigquery/issues/19 """ pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [(3.14, [1, 2])] dataframe = pandas.DataFrame( data=records, columns=["float_column", "array_column"] ) schema = [ SchemaField("float_column", "FLOAT"), SchemaField( "array_column", "INTEGER", mode="REPEATED", ), ] job_config = job.LoadJobConfig(schema=schema) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION, ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.schema == schema def test_load_table_from_dataframe_array_fields_w_auto_schema(self): """Test that a DataFrame with array columns can be uploaded correctly. See: https://github.com/googleapis/python-bigquery/issues/19 """ pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [(3.14, [1, 2])] dataframe = pandas.DataFrame( data=records, columns=["float_column", "array_column"] ) expected_schema = [ SchemaField("float_column", "FLOAT"), SchemaField( "array_column", "INT64", mode="REPEATED", ), ] load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION, ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert sent_config.schema == expected_schema def test_load_table_from_dataframe_w_partial_schema(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() df_data = collections.OrderedDict( [ ("int_col", [1, 2, 3]), ("int_as_float_col", [1.0, float("nan"), 3.0]), ("float_col", [1.0, 2.0, 3.0]), ("bool_col", [True, False, True]), ( "dt_col", pandas.Series( [ datetime.datetime(2010, 1, 2, 3, 44, 50), datetime.datetime(2011, 2, 3, 14, 50, 59), datetime.datetime(2012, 3, 14, 15, 16), ], dtype="datetime64[ns]", ), ), ( "ts_col", pandas.Series( [ datetime.datetime(2010, 1, 2, 3, 44, 50), datetime.datetime(2011, 2, 3, 14, 50, 59), datetime.datetime(2012, 3, 14, 15, 16), ], dtype="datetime64[ns]", ).dt.tz_localize(datetime.timezone.utc), ), ("string_col", ["abc", None, "def"]), ("bytes_col", [b"abc", b"def", None]), ] ) dataframe = pandas.DataFrame(df_data, columns=df_data.keys()) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) schema = ( SchemaField("int_as_float_col", "INTEGER"), SchemaField("string_col", "STRING"), SchemaField("bytes_col", "BYTES"), ) job_config = job.LoadJobConfig(schema=schema) with load_patch as load_table_from_file: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert tuple(sent_config.schema) == ( SchemaField("int_col", "INTEGER"), SchemaField("int_as_float_col", "INTEGER"), SchemaField("float_col", "FLOAT"), SchemaField("bool_col", "BOOLEAN"), SchemaField("dt_col", "DATETIME"), SchemaField("ts_col", "TIMESTAMP"), SchemaField("string_col", "STRING"), SchemaField("bytes_col", "BYTES"), ) def test_load_table_from_dataframe_w_partial_schema_extra_types(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() df_data = collections.OrderedDict( [ ("int_col", [1, 2, 3]), ("int_as_float_col", [1.0, float("nan"), 3.0]), ("string_col", ["abc", None, "def"]), ] ) dataframe = pandas.DataFrame(df_data, columns=df_data.keys()) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) schema = ( SchemaField("int_as_float_col", "INTEGER"), SchemaField("string_col", "STRING"), SchemaField("unknown_col", "BYTES"), ) job_config = job.LoadJobConfig(schema=schema) with load_patch as load_table_from_file, pytest.raises( ValueError ) as exc_context: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) load_table_from_file.assert_not_called() message = str(exc_context.value) assert "bq_schema contains fields not present in dataframe" in message assert "unknown_col" in message def test_load_table_from_dataframe_w_schema_arrow_custom_compression(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"name": "Monty", "age": 100}, {"name": "Python", "age": 60}] dataframe = pandas.DataFrame(records) schema = (SchemaField("name", "STRING"), SchemaField("age", "INTEGER")) job_config = job.LoadJobConfig(schema=schema) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) to_parquet_patch = mock.patch( "google.cloud.bigquery.client._pandas_helpers.dataframe_to_parquet", autospec=True, ) with load_patch, to_parquet_patch as fake_to_parquet: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION, parquet_compression="LZ4", ) call_args = fake_to_parquet.call_args[1] assert call_args is not None assert call_args.get("parquet_compression") == "LZ4" def test_load_table_from_dataframe_wo_pyarrow_raises_error(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) pyarrow_patch = mock.patch("google.cloud.bigquery.client.pyarrow", None) to_parquet_patch = mock.patch.object( dataframe, "to_parquet", wraps=dataframe.to_parquet ) with load_patch, get_table_patch, pyarrow_patch, to_parquet_patch: with pytest.raises(ValueError): client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION, parquet_compression="gzip", ) def test_load_table_from_dataframe_w_bad_pyarrow_issues_warning(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) pyarrow_version_patch = mock.patch( "google.cloud.bigquery._versions_helpers.PYARROW_VERSIONS._installed_version", packaging.version.parse("2.0.0"), # A known bad version of pyarrow. ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=google.api_core.exceptions.NotFound("Table not found"), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch, get_table_patch, pyarrow_version_patch: with pytest.raises(exceptions.LegacyPyarrowError): client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION, ) def test_load_table_from_dataframe_w_nulls(self): """Test that a DataFrame with null columns can be uploaded if a BigQuery schema is specified. See: https://github.com/googleapis/google-cloud-python/issues/7370 """ pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.schema import SchemaField from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job client = self._make_client() records = [{"name": None, "age": None}, {"name": None, "age": None}] dataframe = pandas.DataFrame(records, columns=["name", "age"]) schema = [SchemaField("name", "STRING"), SchemaField("age", "INTEGER")] job_config = job.LoadJobConfig(schema=schema) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.schema == schema assert sent_config.source_format == job.SourceFormat.PARQUET def test_load_table_from_dataframe_w_invaild_job_config(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery import job client = self._make_client() records = [{"float_column": 3.14, "struct_column": [{"foo": 1}, {"bar": -1}]}] dataframe = pandas.DataFrame(data=records) job_config = job.CopyJobConfig() with pytest.raises(TypeError) as exc: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config, location=self.LOCATION ) err_msg = str(exc.value) assert "Expected an instance of LoadJobConfig" in err_msg def test_load_table_from_dataframe_with_csv_source_format(self): pandas = pytest.importorskip("pandas") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField client = self._make_client() records = [{"id": 1, "age": 100}, {"id": 2, "age": 60}] dataframe = pandas.DataFrame(records) job_config = job.LoadJobConfig( write_disposition=job.WriteDisposition.WRITE_TRUNCATE, source_format=job.SourceFormat.CSV, ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock( schema=[SchemaField("id", "INTEGER"), SchemaField("age", "INTEGER")] ), ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, job_config=job_config ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=None, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_file = load_table_from_file.mock_calls[0][1][1] assert sent_file.closed sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.CSV def test_load_table_from_dataframe_w_higher_scale_decimal128_datatype(self): pandas = pytest.importorskip("pandas") pytest.importorskip("pyarrow") from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.schema import SchemaField from decimal import Decimal client = self._make_client() dataframe = pandas.DataFrame({"x": [Decimal("0.1234567891")]}) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_dataframe( dataframe, self.TABLE_REF, location=self.LOCATION ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, num_retries=_DEFAULT_NUM_RETRIES, rewind=True, size=mock.ANY, job_id=mock.ANY, job_id_prefix=None, location=self.LOCATION, project=None, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.PARQUET assert tuple(sent_config.schema) == ( SchemaField("x", "BIGNUMERIC", "NULLABLE", None), ) # With autodetect specified, we pass the value as is. For more info, see # https://github.com/googleapis/python-bigquery/issues/1228#issuecomment-1910946297 def test_load_table_from_json_basic_use(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] job_config = job.LoadJobConfig(autodetect=True) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) # mock: remote table already exists get_table_reference = { "projectId": "project_id", "datasetId": "test_dataset", "tableId": "test_table", } get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock(table_reference=get_table_reference), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location=client.location, project=client.project, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema is None assert sent_config.autodetect def test_load_table_from_json_non_default_args(self): from google.cloud.bigquery import job from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery.schema import SchemaField client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] schema = [ SchemaField("name", "STRING"), SchemaField("age", "INTEGER"), SchemaField("adult", "BOOLEAN"), ] job_config = job.LoadJobConfig(schema=schema) job_config._properties["load"]["unknown_field"] = "foobar" original_config_copy = copy.deepcopy(job_config) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config, project="project-x", location="EU", ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location="EU", project="project-x", job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema == schema assert not sent_config.autodetect # all properties should have been cloned and sent to the backend assert sent_config._properties.get("load", {}).get("unknown_field") == "foobar" # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_json_w_invalid_job_config(self): from google.cloud.bigquery import job client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] job_config = job.CopyJobConfig() with pytest.raises(TypeError) as exc: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config, project="project-x", location="EU", ) err_msg = str(exc.value) assert "Expected an instance of LoadJobConfig" in err_msg # When all following are true: # (1) no schema provided; # (2) no autodetect value provided; # (3) writeDisposition == WRITE_APPEND or None; # (4) table already exists, # client sets autodetect == False # For more details, see https://github.com/googleapis/python-bigquery/issues/1228#issuecomment-1910946297 def test_load_table_from_json_wo_schema_wo_autodetect_write_append_w_table(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.job import WriteDisposition client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] job_config = job.LoadJobConfig(write_disposition=WriteDisposition.WRITE_APPEND) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) # mock: remote table already exists get_table_reference = { "projectId": "project_id", "datasetId": "test_dataset", "tableId": "test_table", } get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock(table_reference=get_table_reference), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location=client.location, project=client.project, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema is None assert not sent_config.autodetect # When all following are true: # (1) no schema provided; # (2) no autodetect value provided; # (3) writeDisposition == WRITE_APPEND or None; # (4) table does NOT exist, # client sets autodetect == True # For more details, see https://github.com/googleapis/python-bigquery/issues/1228#issuecomment-1910946297 def test_load_table_from_json_wo_schema_wo_autodetect_write_append_wo_table(self): import google.api_core.exceptions as core_exceptions from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.job import WriteDisposition client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] job_config = job.LoadJobConfig(write_disposition=WriteDisposition.WRITE_APPEND) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) # mock: remote table doesn't exist get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, side_effect=core_exceptions.NotFound(""), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location=client.location, project=client.project, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema is None assert sent_config.autodetect # When all following are true: # (1) no schema provided; # (2) no autodetect value provided; # (3) writeDisposition == WRITE_TRUNCATE or WRITE_EMPTY; # client sets autodetect == True # For more details, see https://github.com/googleapis/python-bigquery/issues/1228#issuecomment-1910946297 def test_load_table_from_json_wo_schema_wo_autodetect_others(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery import job from google.cloud.bigquery.job import WriteDisposition client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] job_config = job.LoadJobConfig( write_disposition=WriteDisposition.WRITE_TRUNCATE ) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location=client.location, project=client.project, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema is None assert sent_config.autodetect def test_load_table_from_json_w_explicit_job_config_override(self): from google.cloud.bigquery import job from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery.schema import SchemaField client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] schema = [ SchemaField("name", "STRING"), SchemaField("age", "INTEGER"), SchemaField("adult", "BOOLEAN"), ] client.default_load_job_config = job.LoadJobConfig( schema=schema, encoding="ISO-8859-1" ) override_schema = schema override_schema[0] = SchemaField("username", "STRING") job_config = job.LoadJobConfig(schema=override_schema) original_config_copy = copy.deepcopy(job_config) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=job_config, project="project-x", location="EU", ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location="EU", project="project-x", job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema == override_schema assert sent_config.encoding == "ISO-8859-1" assert not sent_config.autodetect # the original config object should not have been modified assert job_config.to_api_repr() == original_config_copy.to_api_repr() def test_load_table_from_json_w_default_job_config(self): from google.cloud.bigquery import job from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES from google.cloud.bigquery.schema import SchemaField client = self._make_client() json_rows = [ {"name": "One", "age": 11, "birthday": "2008-09-10", "adult": False}, {"name": "Two", "age": 22, "birthday": "1997-08-09", "adult": True}, ] schema = [ SchemaField("name", "STRING"), SchemaField("age", "INTEGER"), SchemaField("adult", "BOOLEAN"), ] client.default_load_job_config = job.LoadJobConfig(schema=schema) load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) with load_patch as load_table_from_file: client.load_table_from_json( json_rows, self.TABLE_REF, job_config=None, project="project-x", location="EU", ) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location="EU", project="project-x", job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_config = load_table_from_file.mock_calls[0][2]["job_config"] assert sent_config.source_format == job.SourceFormat.NEWLINE_DELIMITED_JSON assert sent_config.schema == schema def test_load_table_from_json_unicode_emoji_data_case(self): from google.cloud.bigquery.client import _DEFAULT_NUM_RETRIES client = self._make_client() emoji = "\U0001F3E6" json_row = {"emoji": emoji} json_rows = [json_row] load_patch = mock.patch( "google.cloud.bigquery.client.Client.load_table_from_file", autospec=True ) # mock: remote table already exists get_table_reference = { "projectId": "project_id", "datasetId": "test_dataset", "tableId": "test_table", } get_table_patch = mock.patch( "google.cloud.bigquery.client.Client.get_table", autospec=True, return_value=mock.Mock(table_reference=get_table_reference), ) with load_patch as load_table_from_file, get_table_patch: client.load_table_from_json(json_rows, self.TABLE_REF) load_table_from_file.assert_called_once_with( client, mock.ANY, self.TABLE_REF, size=mock.ANY, num_retries=_DEFAULT_NUM_RETRIES, job_id=mock.ANY, job_id_prefix=None, location=client.location, project=client.project, job_config=mock.ANY, timeout=DEFAULT_TIMEOUT, ) sent_data_file = load_table_from_file.mock_calls[0][1][1] # make sure json_row's unicode characters are only encoded one time expected_bytes = b'{"emoji": "' + emoji.encode("utf8") + b'"}' assert sent_data_file.getvalue() == expected_bytes # Low-level tests @classmethod def _make_resumable_upload_responses(cls, size): """Make a series of responses for a successful resumable upload.""" from google import resumable_media resumable_url = "http://test.invalid?upload_id=and-then-there-was-1" initial_response = cls._make_response( http.client.OK, "", {"location": resumable_url} ) data_response = cls._make_response( resumable_media.PERMANENT_REDIRECT, "", {"range": "bytes=0-{:d}".format(size - 1)}, ) final_response = cls._make_response( http.client.OK, json.dumps({"size": size}), {"Content-Type": "application/json"}, ) return [initial_response, data_response, final_response] @staticmethod def _make_transport(responses=None): import google.auth.transport.requests transport = mock.create_autospec( google.auth.transport.requests.AuthorizedSession, instance=True ) transport.request.side_effect = responses return transport def test__do_resumable_upload(self): file_obj = self._make_file_obj() file_obj_len = len(file_obj.getvalue()) transport = self._make_transport( self._make_resumable_upload_responses(file_obj_len) ) client = self._make_client(transport) result = client._do_resumable_upload( file_obj, self.EXPECTED_CONFIGURATION, None, None ) content = result.content.decode("utf-8") assert json.loads(content) == {"size": file_obj_len} # Verify that configuration data was passed in with the initial # request. transport.request.assert_any_call( "POST", mock.ANY, data=json.dumps(self.EXPECTED_CONFIGURATION).encode("utf-8"), headers=mock.ANY, timeout=mock.ANY, ) def test__do_resumable_upload_custom_project(self): file_obj = self._make_file_obj() file_obj_len = len(file_obj.getvalue()) transport = self._make_transport( self._make_resumable_upload_responses(file_obj_len) ) client = self._make_client(transport) result = client._do_resumable_upload( file_obj, self.EXPECTED_CONFIGURATION, None, None, project="custom-project", ) content = result.content.decode("utf-8") assert json.loads(content) == {"size": file_obj_len} # Verify that configuration data was passed in with the initial # request. transport.request.assert_any_call( "POST", mock.ANY, data=json.dumps(self.EXPECTED_CONFIGURATION).encode("utf-8"), headers=mock.ANY, timeout=mock.ANY, ) initiation_url = next( ( call[0][1] for call in transport.request.call_args_list if call[0][0] == "POST" and "uploadType=resumable" in call[0][1] ), None, ) # pragma: NO COVER assert initiation_url is not None assert "projects/custom-project" in initiation_url def test__do_resumable_upload_custom_timeout(self): file_obj = self._make_file_obj() file_obj_len = len(file_obj.getvalue()) transport = self._make_transport( self._make_resumable_upload_responses(file_obj_len) ) client = self._make_client(transport) client._do_resumable_upload( file_obj, self.EXPECTED_CONFIGURATION, num_retries=0, timeout=3.14 ) # The timeout should be applied to all underlying calls. for call_args in transport.request.call_args_list: assert call_args[1].get("timeout") == 3.14 def test__do_multipart_upload(self): transport = self._make_transport([self._make_response(http.client.OK)]) client = self._make_client(transport) file_obj = self._make_file_obj() file_obj_len = len(file_obj.getvalue()) client._do_multipart_upload( file_obj, self.EXPECTED_CONFIGURATION, file_obj_len, None, None ) # Verify that configuration data was passed in with the initial # request. request_args = transport.request.mock_calls[0][2] request_data = request_args["data"].decode("utf-8") request_headers = request_args["headers"] request_content = email.message_from_string( "Content-Type: {}\r\n{}".format( request_headers["content-type"].decode("utf-8"), request_data ) ) # There should be two payloads: the configuration and the binary daya. configuration_data = request_content.get_payload(0).get_payload() binary_data = request_content.get_payload(1).get_payload() assert json.loads(configuration_data) == self.EXPECTED_CONFIGURATION assert binary_data.encode("utf-8") == file_obj.getvalue() def test__do_multipart_upload_wrong_size(self): client = self._make_client() file_obj = self._make_file_obj() file_obj_len = len(file_obj.getvalue()) with pytest.raises(ValueError): client._do_multipart_upload(file_obj, {}, file_obj_len + 1, None, None) def test_schema_from_json_with_file_path(self): from google.cloud.bigquery.schema import SchemaField file_content = """[ { "description": "quarter", "mode": "REQUIRED", "name": "qtr", "type": "STRING" }, { "description": "sales representative", "mode": "NULLABLE", "name": "rep", "type": "STRING" }, { "description": "total sales", "mode": "NULLABLE", "name": "sales", "type": "FLOAT" } ]""" expected = [ SchemaField("qtr", "STRING", "REQUIRED", description="quarter"), SchemaField( "rep", "STRING", "NULLABLE", description="sales representative", ), SchemaField( "sales", "FLOAT", "NULLABLE", description="total sales", ), ] client = self._make_client() mock_file_path = "/mocked/file.json" open_patch = mock.patch( "builtins.open", new=mock.mock_open(read_data=file_content) ) with open_patch as _mock_file: actual = client.schema_from_json(mock_file_path) _mock_file.assert_called_once_with(mock_file_path) # This assert is to make sure __exit__ is called in the context # manager that opens the file in the function _mock_file().__exit__.assert_called_once() assert expected == actual def test_schema_from_json_with_file_object(self): from google.cloud.bigquery.schema import SchemaField file_content = """[ { "description": "quarter", "mode": "REQUIRED", "name": "qtr", "type": "STRING" }, { "description": "sales representative", "mode": "NULLABLE", "name": "rep", "type": "STRING" }, { "description": "total sales", "mode": "NULLABLE", "name": "sales", "type": "FLOAT" } ]""" expected = [ SchemaField("qtr", "STRING", "REQUIRED", description="quarter"), SchemaField( "rep", "STRING", "NULLABLE", description="sales representative" ), SchemaField("sales", "FLOAT", "NULLABLE", description="total sales"), ] client = self._make_client() fake_file = io.StringIO(file_content) actual = client.schema_from_json(fake_file) assert expected == actual def test_schema_to_json_with_file_path(self): from google.cloud.bigquery.schema import SchemaField file_content = [ { "description": "quarter", "mode": "REQUIRED", "name": "qtr", "type": "STRING", }, { "description": "sales representative", "mode": "NULLABLE", "name": "rep", "type": "STRING", }, { "description": "total sales", "mode": "NULLABLE", "name": "sales", "type": "FLOAT", }, ] schema_list = [ SchemaField("qtr", "STRING", "REQUIRED", description="quarter"), SchemaField( "rep", "STRING", "NULLABLE", description="sales representative" ), SchemaField("sales", "FLOAT", "NULLABLE", description="total sales"), ] client = self._make_client() mock_file_path = "/mocked/file.json" open_patch = mock.patch("builtins.open", mock.mock_open()) with open_patch as mock_file, mock.patch("json.dump") as mock_dump: client.schema_to_json(schema_list, mock_file_path) mock_file.assert_called_once_with(mock_file_path, mode="w") # This assert is to make sure __exit__ is called in the context # manager that opens the file in the function mock_file().__exit__.assert_called_once() mock_dump.assert_called_with( file_content, mock_file.return_value, indent=2, sort_keys=True ) def test_schema_to_json_with_file_object(self): from google.cloud.bigquery.schema import SchemaField file_content = [ { "description": "quarter", "mode": "REQUIRED", "name": "qtr", "type": "STRING", }, { "description": "sales representative", "mode": "NULLABLE", "name": "rep", "type": "STRING", }, { "description": "total sales", "mode": "NULLABLE", "name": "sales", "type": "FLOAT", }, ] schema_list = [ SchemaField("qtr", "STRING", "REQUIRED", description="quarter"), SchemaField( "rep", "STRING", "NULLABLE", description="sales representative" ), SchemaField("sales", "FLOAT", "NULLABLE", description="total sales"), ] fake_file = io.StringIO() client = self._make_client() client.schema_to_json(schema_list, fake_file) assert file_content == json.loads(fake_file.getvalue()) def test_upload_chunksize(client): with mock.patch("google.cloud.bigquery.client.ResumableUpload") as RU: upload = RU.return_value upload.finished = False def transmit_next_chunk(transport, *args, **kwargs): upload.finished = True result = mock.MagicMock() result.json.return_value = {} return result upload.transmit_next_chunk = transmit_next_chunk f = io.BytesIO() client.load_table_from_file(f, "foo.bar") chunk_size = RU.call_args_list[0][0][1] assert chunk_size == 100 * (1 << 20) @pytest.mark.enable_add_server_timeout_header @pytest.mark.parametrize("headers", [None, {}]) def test__call_api_add_server_timeout_w_timeout(client, headers): client._connection = make_connection({}) client._call_api(None, method="GET", path="/", headers=headers, timeout=42) client._connection.api_request.assert_called_with( method="GET", path="/", timeout=42, headers={"X-Server-Timeout": "42"} ) @pytest.mark.enable_add_server_timeout_header def test__call_api_no_add_server_timeout_wo_timeout(client): client._connection = make_connection({}) client._call_api(None, method="GET", path="/") client._connection.api_request.assert_called_with(method="GET", path="/")