Asib27's picture
try 1
065fee7 verified
raw
history blame
75.9 kB
# Copyright 2018 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import copy
import re
from concurrent import futures
from unittest import mock
import warnings
from google.api_core import exceptions
import google.auth.credentials
import pytest
from tests.unit.helpers import make_connection
from test_utils.imports import maybe_fail_import
from google.cloud import bigquery
from google.cloud.bigquery import exceptions as bq_exceptions
from google.cloud.bigquery import job
from google.cloud.bigquery import table
from google.cloud.bigquery.retry import DEFAULT_TIMEOUT
try:
from google.cloud.bigquery.magics import magics
except ImportError:
magics = None
bigquery_storage = pytest.importorskip("google.cloud.bigquery_storage")
IPython = pytest.importorskip("IPython")
interactiveshell = pytest.importorskip("IPython.terminal.interactiveshell")
tools = pytest.importorskip("IPython.testing.tools")
io = pytest.importorskip("IPython.utils.io")
pandas = pytest.importorskip("pandas")
@pytest.fixture(scope="session")
def ipython():
config = tools.default_config()
config.TerminalInteractiveShell.simple_prompt = True
shell = interactiveshell.TerminalInteractiveShell.instance(config=config)
return shell
@pytest.fixture()
def ipython_interactive(request, ipython):
"""Activate IPython's builtin hooks
for the duration of the test scope.
"""
with ipython.builtin_trap:
yield ipython
@pytest.fixture()
def ipython_ns_cleanup():
"""A helper to clean up user namespace after the test
for the duration of the test scope.
"""
names_to_clean = [] # pairs (IPython_instance, name_to_clean)
yield names_to_clean
for ip, name in names_to_clean:
if name in ip.user_ns:
del ip.user_ns[name]
@pytest.fixture(scope="session")
def missing_bq_storage():
"""Provide a patcher that can make the bigquery storage import to fail."""
def fail_if(name, globals, locals, fromlist, level):
# NOTE: *very* simplified, assuming a straightforward absolute import
return "bigquery_storage" in name or (
fromlist is not None and "bigquery_storage" in fromlist
)
return maybe_fail_import(predicate=fail_if)
@pytest.fixture(scope="session")
def missing_grpcio_lib():
"""Provide a patcher that can make the gapic library import to fail."""
def fail_if(name, globals, locals, fromlist, level):
# NOTE: *very* simplified, assuming a straightforward absolute import
return "gapic_v1" in name or (fromlist is not None and "gapic_v1" in fromlist)
return maybe_fail_import(predicate=fail_if)
PROJECT_ID = "its-a-project-eh"
JOB_ID = "some-random-id"
JOB_REFERENCE_RESOURCE = {"projectId": PROJECT_ID, "jobId": JOB_ID}
DATASET_ID = "dest_dataset"
TABLE_ID = "dest_table"
TABLE_REFERENCE_RESOURCE = {
"projectId": PROJECT_ID,
"datasetId": DATASET_ID,
"tableId": TABLE_ID,
}
QUERY_STRING = "SELECT 42 AS the_answer FROM `life.the_universe.and_everything`;"
QUERY_RESOURCE = {
"jobReference": JOB_REFERENCE_RESOURCE,
"configuration": {
"query": {
"destinationTable": TABLE_REFERENCE_RESOURCE,
"query": QUERY_STRING,
"queryParameters": [],
"useLegacySql": False,
}
},
"status": {"state": "DONE"},
}
QUERY_RESULTS_RESOURCE = {
"jobReference": JOB_REFERENCE_RESOURCE,
"totalRows": 1,
"jobComplete": True,
"schema": {"fields": [{"name": "the_answer", "type": "INTEGER"}]},
}
def test_context_with_default_credentials():
"""When Application Default Credentials are set, the context credentials
will be created the first time it is called
"""
assert magics.context._credentials is None
assert magics.context._project is None
project = "prahj-ekt"
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, project)
)
with default_patch as default_mock:
assert magics.context.credentials is credentials_mock
assert magics.context.project == project
assert default_mock.call_count == 2
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_context_with_default_connection():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._credentials = None
magics.context._project = None
magics.context._connection = None
default_credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
credentials_patch = mock.patch(
"google.auth.default", return_value=(default_credentials, "project-from-env")
)
default_conn = make_connection(QUERY_RESOURCE, QUERY_RESULTS_RESOURCE)
conn_patch = mock.patch("google.cloud.bigquery.client.Connection", autospec=True)
list_rows_patch = mock.patch(
"google.cloud.bigquery.client.Client._list_rows_from_query_results",
return_value=google.cloud.bigquery.table._EmptyRowIterator(),
)
with conn_patch as conn, credentials_patch, list_rows_patch as list_rows:
conn.return_value = default_conn
ip.run_cell_magic("bigquery", "", QUERY_STRING)
# Check that query actually starts the job.
conn.assert_called()
list_rows.assert_called()
begin_call = mock.call(
method="POST",
path="/projects/project-from-env/jobs",
data=mock.ANY,
timeout=DEFAULT_TIMEOUT,
)
query_results_call = mock.call(
method="GET",
path=f"/projects/{PROJECT_ID}/queries/{JOB_ID}",
query_params=mock.ANY,
timeout=mock.ANY,
)
default_conn.api_request.assert_has_calls([begin_call, query_results_call])
def test_context_credentials_and_project_can_be_set_explicitly():
project1 = "one-project-55564"
project2 = "other-project-52569"
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, project1)
)
with default_patch as default_mock:
magics.context.credentials = credentials_mock
magics.context.project = project2
assert magics.context.project == project2
assert magics.context.credentials is credentials_mock
# default should not be called if credentials & project are explicitly set
assert default_mock.call_count == 0
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_context_with_custom_connection():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
magics.context._credentials = None
context_conn = magics.context._connection = make_connection(
QUERY_RESOURCE, QUERY_RESULTS_RESOURCE
)
default_credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
credentials_patch = mock.patch(
"google.auth.default", return_value=(default_credentials, "project-from-env")
)
default_conn = make_connection()
conn_patch = mock.patch("google.cloud.bigquery.client.Connection", autospec=True)
list_rows_patch = mock.patch(
"google.cloud.bigquery.client.Client._list_rows_from_query_results",
return_value=google.cloud.bigquery.table._EmptyRowIterator(),
)
with conn_patch as conn, credentials_patch, list_rows_patch as list_rows:
conn.return_value = default_conn
ip.run_cell_magic("bigquery", "", QUERY_STRING)
list_rows.assert_called()
default_conn.api_request.assert_not_called()
begin_call = mock.call(
method="POST",
path="/projects/project-from-env/jobs",
data=mock.ANY,
timeout=DEFAULT_TIMEOUT,
)
query_results_call = mock.call(
method="GET",
path=f"/projects/{PROJECT_ID}/queries/{JOB_ID}",
query_params=mock.ANY,
timeout=mock.ANY,
)
context_conn.api_request.assert_has_calls([begin_call, query_results_call])
def test__run_query():
magics.context._credentials = None
job_id = "job_1234"
sql = "SELECT 17"
responses = [
futures.TimeoutError,
futures.TimeoutError,
[table.Row((17,), {"num": 0})],
]
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
with client_patch as client_mock, io.capture_output() as captured:
client_mock().query(sql).result.side_effect = responses
client_mock().query(sql).job_id = job_id
query_job = magics._run_query(client_mock(), sql)
lines = re.split("\n|\r", captured.stdout)
# Removes blanks & terminal code (result of display clearing)
updates = list(filter(lambda x: bool(x) and x != "\x1b[2K", lines))
assert query_job.job_id == job_id
expected_first_line = "Executing query with job ID: {}".format(job_id)
assert updates[0] == expected_first_line
execution_updates = updates[1:-1]
assert len(execution_updates) == 3 # one update per API response
for line in execution_updates:
assert re.match("Query executing: .*s", line)
def test__run_query_dry_run_without_errors_is_silent():
magics.context._credentials = None
sql = "SELECT 17"
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
job_config = job.QueryJobConfig()
job_config.dry_run = True
with client_patch as client_mock, io.capture_output() as captured:
client_mock().query(sql).job_id = None
magics._run_query(client_mock(), sql, job_config=job_config)
assert len(captured.stderr) == 0
assert len(captured.stdout) == 0
def test__make_bqstorage_client_false():
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
test_client = bigquery.Client(
project="test_project", credentials=credentials_mock, location="test_location"
)
got = magics._make_bqstorage_client(test_client, False, {})
assert got is None
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
def test__make_bqstorage_client_true():
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
test_client = bigquery.Client(
project="test_project", credentials=credentials_mock, location="test_location"
)
got = magics._make_bqstorage_client(test_client, True, {})
assert isinstance(got, bigquery_storage.BigQueryReadClient)
def test__make_bqstorage_client_true_raises_import_error(missing_bq_storage):
"""When package `google-cloud-bigquery-storage` is not installed, reports
ImportError.
"""
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
test_client = bigquery.Client(
project="test_project", credentials=credentials_mock, location="test_location"
)
with pytest.raises(ImportError) as exc_context, missing_bq_storage:
magics._make_bqstorage_client(test_client, True, {})
error_msg = str(exc_context.value)
assert "google-cloud-bigquery-storage" in error_msg
assert "pyarrow" in error_msg
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
def test__make_bqstorage_client_true_obsolete_dependency():
"""When package `google-cloud-bigquery-storage` is installed but has outdated
version, returns None, and raises a warning.
"""
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
test_client = bigquery.Client(
project="test_project", credentials=credentials_mock, location="test_location"
)
patcher = mock.patch(
"google.cloud.bigquery._versions_helpers.BQ_STORAGE_VERSIONS.try_import",
side_effect=bq_exceptions.LegacyBigQueryStorageError(
"google-cloud-bigquery-storage is outdated"
),
)
with patcher, warnings.catch_warnings(record=True) as warned:
got = magics._make_bqstorage_client(test_client, True, {})
assert got is None
matching_warnings = [
warning
for warning in warned
if "google-cloud-bigquery-storage is outdated" in str(warning)
]
assert matching_warnings, "Obsolete dependency warning not raised."
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test__make_bqstorage_client_true_missing_gapic(missing_grpcio_lib):
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
with pytest.raises(ImportError) as exc_context, missing_grpcio_lib:
magics._make_bqstorage_client(True, credentials_mock, {})
assert "grpcio" in str(exc_context.value)
def test__create_dataset_if_necessary_exists():
project = "project_id"
dataset_id = "dataset_id"
dataset_reference = bigquery.dataset.DatasetReference(project, dataset_id)
dataset = bigquery.Dataset(dataset_reference)
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
with client_patch as client_mock:
client = client_mock()
client.project = project
client.get_dataset.result_value = dataset
magics._create_dataset_if_necessary(client, dataset_id)
client.create_dataset.assert_not_called()
def test__create_dataset_if_necessary_not_exist():
project = "project_id"
dataset_id = "dataset_id"
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
with client_patch as client_mock:
client = client_mock()
client.location = "us"
client.project = project
client.get_dataset.side_effect = exceptions.NotFound("dataset not found")
magics._create_dataset_if_necessary(client, dataset_id)
client.create_dataset.assert_called_once()
@pytest.mark.usefixtures("ipython_interactive")
def test_extension_load():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
# verify that the magic is registered and has the correct source
magic = ip.magics_manager.magics["cell"].get("bigquery")
assert magic.__module__ == "google.cloud.bigquery.magics.magics"
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
def test_bigquery_magic_without_optional_arguments(monkeypatch):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
mock_credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
# Set up the context with monkeypatch so that it's reset for subsequent
# tests.
monkeypatch.setattr(magics.context, "_credentials", mock_credentials)
# Mock out the BigQuery Storage API.
bqstorage_mock = mock.create_autospec(bigquery_storage.BigQueryReadClient)
bqstorage_instance_mock = mock.create_autospec(
bigquery_storage.BigQueryReadClient, instance=True
)
bqstorage_instance_mock._transport = mock.Mock()
bqstorage_mock.return_value = bqstorage_instance_mock
bqstorage_client_patch = mock.patch(
"google.cloud.bigquery_storage.BigQueryReadClient", bqstorage_mock
)
sql = "SELECT 17 AS num"
result = pandas.DataFrame([17], columns=["num"])
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock, bqstorage_client_patch:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "", sql)
assert bqstorage_mock.called # BQ storage client was used
assert isinstance(return_value, pandas.DataFrame)
assert len(return_value) == len(result) # verify row count
assert list(return_value) == list(result) # verify column names
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_default_connection_user_agent():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._connection = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
conn_patch = mock.patch("google.cloud.bigquery.client.Connection", autospec=True)
with conn_patch as conn, run_query_patch, default_patch:
ip.run_cell_magic("bigquery", "", "SELECT 17 as num")
client_info_arg = conn.call_args[1].get("client_info")
assert client_info_arg is not None
assert client_info_arg.user_agent == "ipython-" + IPython.__version__
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_legacy_sql():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic("bigquery", "--use_legacy_sql", "SELECT 17 AS num")
job_config_used = run_query_mock.call_args_list[0][1]["job_config"]
assert job_config_used.use_legacy_sql is True
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_result_saved_to_variable(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "df"))
sql = "SELECT 17 AS num"
result = pandas.DataFrame([17], columns=["num"])
assert "df" not in ip.user_ns
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "df", sql)
assert return_value is None
assert "df" in ip.user_ns # verify that variable exists
df = ip.user_ns["df"]
assert len(df) == len(result) # verify row count
assert list(df) == list(result) # verify column names
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_does_not_clear_display_in_verbose_mode():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
clear_patch = mock.patch(
"google.cloud.bigquery.magics.magics.display.clear_output",
autospec=True,
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with clear_patch as clear_mock, run_query_patch:
ip.run_cell_magic("bigquery", "--verbose", "SELECT 17 as num")
assert clear_mock.call_count == 0
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_clears_display_in_non_verbose_mode():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
clear_patch = mock.patch(
"google.cloud.bigquery.magics.magics.display.clear_output",
autospec=True,
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with clear_patch as clear_mock, run_query_patch:
ip.run_cell_magic("bigquery", "", "SELECT 17 as num")
assert clear_mock.call_count == 1
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
def test_bigquery_magic_with_bqstorage_from_argument(monkeypatch):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
mock_credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
# Set up the context with monkeypatch so that it's reset for subsequent
# tests.
monkeypatch.setattr(magics.context, "_credentials", mock_credentials)
# Mock out the BigQuery Storage API.
bqstorage_mock = mock.create_autospec(bigquery_storage.BigQueryReadClient)
bqstorage_instance_mock = mock.create_autospec(
bigquery_storage.BigQueryReadClient, instance=True
)
bqstorage_instance_mock._transport = mock.Mock()
bqstorage_mock.return_value = bqstorage_instance_mock
bqstorage_client_patch = mock.patch(
"google.cloud.bigquery_storage.BigQueryReadClient", bqstorage_mock
)
sql = "SELECT 17 AS num"
result = pandas.DataFrame([17], columns=["num"])
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock, (
bqstorage_client_patch
), warnings.catch_warnings(record=True) as warned:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "--use_bqstorage_api", sql)
# Deprecation warning should have been issued.
def warning_match(warning):
message = str(warning).lower()
return "deprecated" in message and "use_bqstorage_api" in message
expected_warnings = list(filter(warning_match, warned))
assert len(expected_warnings) == 1
assert len(bqstorage_mock.call_args_list) == 1
kwargs = bqstorage_mock.call_args_list[0][1]
assert kwargs.get("credentials") is mock_credentials
client_info = kwargs.get("client_info")
assert client_info is not None
assert client_info.user_agent == "ipython-" + IPython.__version__
query_job_mock.to_dataframe.assert_called_once_with(
bqstorage_client=bqstorage_instance_mock,
create_bqstorage_client=mock.ANY,
progress_bar_type="tqdm_notebook",
)
assert isinstance(return_value, pandas.DataFrame)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
def test_bigquery_magic_with_rest_client_requested(monkeypatch):
pandas = pytest.importorskip("pandas")
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
mock_credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
# Set up the context with monkeypatch so that it's reset for subsequent
# tests.
monkeypatch.setattr(magics.context, "_credentials", mock_credentials)
# Mock out the BigQuery Storage API.
bqstorage_mock = mock.create_autospec(bigquery_storage.BigQueryReadClient)
bqstorage_client_patch = mock.patch(
"google.cloud.bigquery_storage.BigQueryReadClient", bqstorage_mock
)
sql = "SELECT 17 AS num"
result = pandas.DataFrame([17], columns=["num"])
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock, bqstorage_client_patch:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "--use_rest_api", sql)
bqstorage_mock.assert_not_called()
query_job_mock.to_dataframe.assert_called_once_with(
bqstorage_client=None,
create_bqstorage_client=False,
progress_bar_type="tqdm_notebook",
)
assert isinstance(return_value, pandas.DataFrame)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_w_max_results_invalid():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
client_query_patch = mock.patch(
"google.cloud.bigquery.client.Client.query", autospec=True
)
sql = "SELECT 17 AS num"
with pytest.raises(ValueError), default_patch, client_query_patch:
ip.run_cell_magic("bigquery", "--max_results=abc", sql)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_w_max_results_valid_calls_queryjob_result():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
client_query_patch = mock.patch(
"google.cloud.bigquery.client.Client.query", autospec=True
)
sql = "SELECT 17 AS num"
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
with client_query_patch as client_query_mock, default_patch:
client_query_mock.return_value = query_job_mock
ip.run_cell_magic("bigquery", "--max_results=5", sql)
query_job_mock.result.assert_called_with(max_results=5)
query_job_mock.result.return_value.to_dataframe.assert_called_once_with(
bqstorage_client=None,
create_bqstorage_client=False,
progress_bar_type=mock.ANY,
)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_w_max_results_query_job_results_fails():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
client_query_patch = mock.patch(
"google.cloud.bigquery.client.Client.query", autospec=True
)
close_transports_patch = mock.patch(
"google.cloud.bigquery.magics.magics._close_transports",
autospec=True,
)
sql = "SELECT 17 AS num"
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.result.side_effect = [[], OSError]
with pytest.raises(
OSError
), client_query_patch as client_query_mock, (
default_patch
), close_transports_patch as close_transports:
client_query_mock.return_value = query_job_mock
ip.run_cell_magic("bigquery", "--max_results=5", sql)
assert close_transports.called
def test_bigquery_magic_w_table_id_invalid():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
list_rows_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client.list_rows",
autospec=True,
side_effect=exceptions.BadRequest("Not a valid table ID"),
)
table_id = "not-a-real-table"
with list_rows_patch, default_patch, io.capture_output() as captured_io:
ip.run_cell_magic("bigquery", "df", table_id)
output = captured_io.stderr
assert "Could not save output to variable" in output
assert "400 Not a valid table ID" in output
assert "Traceback (most recent call last)" not in output
def test_bigquery_magic_w_missing_query():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
cell_body = " \n \n \t\t \n "
with io.capture_output() as captured_io, default_patch:
ip.run_cell_magic("bigquery", "df", cell_body)
output = captured_io.stderr
assert "Could not save output to variable" in output
assert "Query is missing" in output
assert "Traceback (most recent call last)" not in output
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_w_table_id_and_destination_var(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
ipython_ns_cleanup.append((ip, "df"))
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
row_iterator_mock = mock.create_autospec(
google.cloud.bigquery.table.RowIterator, instance=True
)
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
table_id = "bigquery-public-data.samples.shakespeare"
result = pandas.DataFrame([17], columns=["num"])
with client_patch as client_mock, default_patch:
client_mock().list_rows.return_value = row_iterator_mock
row_iterator_mock.to_dataframe.return_value = result
ip.run_cell_magic("bigquery", "df", table_id)
assert "df" in ip.user_ns
df = ip.user_ns["df"]
assert isinstance(df, pandas.DataFrame)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(
bigquery_storage is None, reason="Requires `google-cloud-bigquery-storage`"
)
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_w_table_id_and_bqstorage_client():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
row_iterator_mock = mock.create_autospec(
google.cloud.bigquery.table.RowIterator, instance=True
)
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
bqstorage_mock = mock.create_autospec(bigquery_storage.BigQueryReadClient)
bqstorage_instance_mock = mock.create_autospec(
bigquery_storage.BigQueryReadClient, instance=True
)
bqstorage_instance_mock._transport = mock.Mock()
bqstorage_mock.return_value = bqstorage_instance_mock
bqstorage_client_patch = mock.patch(
"google.cloud.bigquery_storage.BigQueryReadClient", bqstorage_mock
)
table_id = "bigquery-public-data.samples.shakespeare"
with default_patch, client_patch as client_mock, bqstorage_client_patch:
client_mock()._ensure_bqstorage_client.return_value = bqstorage_instance_mock
client_mock().list_rows.return_value = row_iterator_mock
ip.run_cell_magic("bigquery", "--max_results=5", table_id)
row_iterator_mock.to_dataframe.assert_called_once_with(
bqstorage_client=bqstorage_instance_mock,
create_bqstorage_client=mock.ANY,
)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_dryrun_option_sets_job_config():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
sql = "SELECT 17 AS num"
with run_query_patch as run_query_mock:
ip.run_cell_magic("bigquery", "--dry_run", sql)
job_config_used = run_query_mock.call_args_list[0][1]["job_config"]
assert job_config_used.dry_run is True
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_dryrun_option_returns_query_job():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
sql = "SELECT 17 AS num"
with run_query_patch as run_query_mock, io.capture_output() as captured_io:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "--dry_run", sql)
assert "Query validated. This query will process" in captured_io.stdout
assert isinstance(return_value, job.QueryJob)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_dryrun_option_variable_error_message(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "q_job"))
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query",
autospec=True,
side_effect=exceptions.BadRequest("Syntax error in SQL query"),
)
sql = "SELECT SELECT 17 AS num"
assert "q_job" not in ip.user_ns
with run_query_patch, io.capture_output() as captured:
ip.run_cell_magic("bigquery", "q_job --dry_run", sql)
full_text = captured.stderr
assert "Could not save output to variable 'q_job'." in full_text
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_dryrun_option_saves_query_job_to_variable(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
ipython_ns_cleanup.append((ip, "q_job"))
sql = "SELECT 17 AS num"
assert "q_job" not in ip.user_ns
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "q_job --dry_run", sql)
assert return_value is None
assert "q_job" in ip.user_ns
q_job = ip.user_ns["q_job"]
assert isinstance(q_job, job.QueryJob)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_saves_query_job_to_variable_on_error(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "result"))
client_query_patch = mock.patch(
"google.cloud.bigquery.client.Client.query", autospec=True
)
query_job = mock.create_autospec(job.QueryJob, instance=True)
exception = Exception("Unexpected SELECT")
exception.query_job = query_job
query_job.result.side_effect = exception
sql = "SELECT SELECT 17 AS num"
assert "result" not in ip.user_ns
with client_query_patch as client_query_mock:
client_query_mock.return_value = query_job
return_value = ip.run_cell_magic("bigquery", "result", sql)
assert return_value is None
assert "result" in ip.user_ns
result = ip.user_ns["result"]
assert isinstance(result, job.QueryJob)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_w_maximum_bytes_billed_invalid():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
client_query_patch = mock.patch("google.cloud.bigquery.client.Client.query")
sql = "SELECT 17 AS num"
with pytest.raises(ValueError), default_patch, client_query_patch:
ip.run_cell_magic("bigquery", "--maximum_bytes_billed=abc", sql)
@pytest.mark.parametrize(
"param_value,expected", [("987654321", "987654321"), ("None", "0")]
)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_w_maximum_bytes_billed_overrides_context(param_value, expected):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
# Set the default maximum bytes billed, so we know it's overridable by the param.
magics.context.default_query_job_config.maximum_bytes_billed = 1234567
project = "test-project"
job_reference = copy.deepcopy(JOB_REFERENCE_RESOURCE)
job_reference["projectId"] = project
query = "SELECT 17 AS num"
resource = copy.deepcopy(QUERY_RESOURCE)
resource["jobReference"] = job_reference
resource["configuration"]["query"]["query"] = query
query_results = {"jobReference": job_reference, "totalRows": 0, "jobComplete": True}
data = {"jobReference": job_reference, "totalRows": 0, "rows": []}
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
conn = magics.context._connection = make_connection(resource, query_results, data)
list_rows_patch = mock.patch(
"google.cloud.bigquery.client.Client._list_rows_from_query_results",
return_value=google.cloud.bigquery.table._EmptyRowIterator(),
)
with list_rows_patch, default_patch:
ip.run_cell_magic(
"bigquery", "--maximum_bytes_billed={}".format(param_value), query
)
_, req = conn.api_request.call_args_list[0]
sent_config = req["data"]["configuration"]["query"]
assert sent_config["maximumBytesBilled"] == expected
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_w_maximum_bytes_billed_w_context_inplace():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
magics.context.default_query_job_config.maximum_bytes_billed = 1337
project = "test-project"
job_reference = copy.deepcopy(JOB_REFERENCE_RESOURCE)
job_reference["projectId"] = project
query = "SELECT 17 AS num"
resource = copy.deepcopy(QUERY_RESOURCE)
resource["jobReference"] = job_reference
resource["configuration"]["query"]["query"] = query
query_results = {"jobReference": job_reference, "totalRows": 0, "jobComplete": True}
data = {"jobReference": job_reference, "totalRows": 0, "rows": []}
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
conn = magics.context._connection = make_connection(resource, query_results, data)
list_rows_patch = mock.patch(
"google.cloud.bigquery.client.Client._list_rows_from_query_results",
return_value=google.cloud.bigquery.table._EmptyRowIterator(),
)
with list_rows_patch, default_patch:
ip.run_cell_magic("bigquery", "", query)
_, req = conn.api_request.call_args_list[0]
sent_config = req["data"]["configuration"]["query"]
assert sent_config["maximumBytesBilled"] == "1337"
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_w_maximum_bytes_billed_w_context_setter():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
magics.context.default_query_job_config = job.QueryJobConfig(
maximum_bytes_billed=10203
)
project = "test-project"
job_reference = copy.deepcopy(JOB_REFERENCE_RESOURCE)
job_reference["projectId"] = project
query = "SELECT 17 AS num"
resource = copy.deepcopy(QUERY_RESOURCE)
resource["jobReference"] = job_reference
resource["configuration"]["query"]["query"] = query
query_results = {"jobReference": job_reference, "totalRows": 0, "jobComplete": True}
data = {"jobReference": job_reference, "totalRows": 0, "rows": []}
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
conn = magics.context._connection = make_connection(resource, query_results, data)
list_rows_patch = mock.patch(
"google.cloud.bigquery.client.Client._list_rows_from_query_results",
return_value=google.cloud.bigquery.table._EmptyRowIterator(),
)
with list_rows_patch, default_patch:
ip.run_cell_magic("bigquery", "", query)
_, req = conn.api_request.call_args_list[0]
sent_config = req["data"]["configuration"]["query"]
assert sent_config["maximumBytesBilled"] == "10203"
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_no_query_cache(monkeypatch):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
conn = make_connection()
monkeypatch.setattr(magics.context, "_connection", conn)
monkeypatch.setattr(magics.context, "project", "project-from-context")
# --no_query_cache option should override context.
monkeypatch.setattr(
magics.context.default_query_job_config, "use_query_cache", True
)
ip.run_cell_magic("bigquery", "--no_query_cache", QUERY_STRING)
conn.api_request.assert_called_with(
method="POST",
path="/projects/project-from-context/jobs",
data=mock.ANY,
timeout=DEFAULT_TIMEOUT,
)
jobs_insert_call = [
call
for call in conn.api_request.call_args_list
if call[1]["path"] == "/projects/project-from-context/jobs"
][0]
assert not jobs_insert_call[1]["data"]["configuration"]["query"]["useQueryCache"]
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_context_with_no_query_cache_from_context(monkeypatch):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
conn = make_connection()
monkeypatch.setattr(magics.context, "_connection", conn)
monkeypatch.setattr(magics.context, "project", "project-from-context")
monkeypatch.setattr(
magics.context.default_query_job_config, "use_query_cache", False
)
ip.run_cell_magic("bigquery", "", QUERY_STRING)
conn.api_request.assert_called_with(
method="POST",
path="/projects/project-from-context/jobs",
data=mock.ANY,
timeout=DEFAULT_TIMEOUT,
)
jobs_insert_call = [
call
for call in conn.api_request.call_args_list
if call[1]["path"] == "/projects/project-from-context/jobs"
][0]
assert not jobs_insert_call[1]["data"]["configuration"]["query"]["useQueryCache"]
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_w_progress_bar_type_w_context_setter(monkeypatch):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
magics.context.progress_bar_type = "tqdm_gui"
mock_credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
# Set up the context with monkeypatch so that it's reset for subsequent
# tests.
monkeypatch.setattr(magics.context, "_credentials", mock_credentials)
# Mock out the BigQuery Storage API.
bqstorage_mock = mock.create_autospec(bigquery_storage.BigQueryReadClient)
bqstorage_client_patch = mock.patch(
"google.cloud.bigquery_storage.BigQueryReadClient", bqstorage_mock
)
sql = "SELECT 17 AS num"
result = pandas.DataFrame([17], columns=["num"])
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock, bqstorage_client_patch:
run_query_mock.return_value = query_job_mock
return_value = ip.run_cell_magic("bigquery", "--use_rest_api", sql)
bqstorage_mock.assert_not_called()
query_job_mock.to_dataframe.assert_called_once_with(
bqstorage_client=None,
create_bqstorage_client=False,
progress_bar_type=magics.context.progress_bar_type,
)
assert isinstance(return_value, pandas.DataFrame)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_progress_bar_type():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.progress_bar_type = None
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic(
"bigquery", "--progress_bar_type=tqdm_gui", "SELECT 17 as num"
)
progress_bar_used = run_query_mock.mock_calls[1][2]["progress_bar_type"]
assert progress_bar_used == "tqdm_gui"
# context progress bar type should not change
assert magics.context.progress_bar_type is None
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_project():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock, default_patch:
ip.run_cell_magic("bigquery", "--project=specific-project", "SELECT 17 as num")
client_used = run_query_mock.call_args_list[0][0][0]
assert client_used.project == "specific-project"
# context project should not change
assert magics.context.project == "general-project"
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_bigquery_api_endpoint(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._connection = None
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic(
"bigquery",
"--bigquery_api_endpoint=https://bigquery_api.endpoint.com",
"SELECT 17 as num",
)
connection_used = run_query_mock.call_args_list[0][0][0]._connection
assert connection_used.API_BASE_URL == "https://bigquery_api.endpoint.com"
# context client options should not change
assert magics.context.bigquery_client_options.api_endpoint is None
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_bigquery_api_endpoint_context_dict():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._connection = None
magics.context.bigquery_client_options = {}
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic(
"bigquery",
"--bigquery_api_endpoint=https://bigquery_api.endpoint.com",
"SELECT 17 as num",
)
connection_used = run_query_mock.call_args_list[0][0][0]._connection
assert connection_used.API_BASE_URL == "https://bigquery_api.endpoint.com"
# context client options should not change
assert magics.context.bigquery_client_options == {}
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_bqstorage_api_endpoint(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._connection = None
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic(
"bigquery",
"--bqstorage_api_endpoint=https://bqstorage_api.endpoint.com",
"SELECT 17 as num",
)
client_used = run_query_mock.mock_calls[1][2]["bqstorage_client"]
assert client_used._transport._host == "https://bqstorage_api.endpoint.com"
# context client options should not change
assert magics.context.bqstorage_client_options.api_endpoint is None
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_bqstorage_api_endpoint_context_dict():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._connection = None
magics.context.bqstorage_client_options = {}
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic(
"bigquery",
"--bqstorage_api_endpoint=https://bqstorage_api.endpoint.com",
"SELECT 17 as num",
)
client_used = run_query_mock.mock_calls[1][2]["bqstorage_client"]
assert client_used._transport._host == "https://bqstorage_api.endpoint.com"
# context client options should not change
assert magics.context.bqstorage_client_options == {}
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_multiple_options():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock, default_patch:
ip.run_cell_magic(
"bigquery",
"--project=specific-project --use_legacy_sql --maximum_bytes_billed 1024",
"SELECT 17 as num",
)
args, kwargs = run_query_mock.call_args
client_used = args[0]
assert client_used.project == "specific-project"
job_config_used = kwargs["job_config"]
assert job_config_used.use_legacy_sql
assert job_config_used.maximum_bytes_billed == 1024
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_string_params(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "params_dict_df"))
sql = "SELECT @num AS num"
result = pandas.DataFrame([17], columns=["num"])
assert "params_dict_df" not in ip.user_ns
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
ip.run_cell_magic("bigquery", "params_string_df --params='{\"num\":17}'", sql)
run_query_mock.assert_called_once_with(mock.ANY, sql.format(num=17), mock.ANY)
assert "params_string_df" in ip.user_ns # verify that the variable exists
df = ip.user_ns["params_string_df"]
assert len(df) == len(result) # verify row count
assert list(df) == list(result) # verify column names
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "params_dict_df"))
sql = "SELECT @num AS num, @tricky_value as tricky_value"
result = pandas.DataFrame(
[(False, '--params "value"')], columns=["valid", "tricky_value"]
)
assert "params_dict_df" not in ip.user_ns
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
params = {"valid": False, "tricky_value": '--params "value"'}
# Insert dictionary into user namespace so that it can be expanded
ip.user_ns["params"] = params
ip.run_cell_magic("bigquery", "params_dict_df --params $params", sql)
run_query_mock.assert_called_once_with(mock.ANY, sql.format(num=17), mock.ANY)
assert "params_dict_df" in ip.user_ns # verify that the variable exists
df = ip.user_ns["params_dict_df"]
assert len(df) == len(result) # verify row count
assert list(df) == list(result) # verify column names
assert not df["valid"][0]
assert df["tricky_value"][0] == '--params "value"'
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params_nonexisting():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
sql = "SELECT @foo AS foo"
with pytest.raises(NameError, match=r".*undefined variable.*unknown_name.*"):
ip.run_cell_magic("bigquery", "params_dict_df --params $unknown_name", sql)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params_incorrect_syntax():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
sql = "SELECT @foo AS foo"
with pytest.raises(SyntaxError, match=r".*--params.*"):
cell_magic_args = "params_dict_df --params {'foo': 1; 'bar': 2}"
ip.run_cell_magic("bigquery", cell_magic_args, sql)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params_duplicate():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
sql = "SELECT @foo AS foo"
with pytest.raises(ValueError, match=r"Duplicate --params option\."):
cell_magic_args = (
"params_dict_df --params {'foo': 1} --verbose --params {'bar': 2} "
)
ip.run_cell_magic("bigquery", cell_magic_args, sql)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_option_value_incorrect():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
sql = "SELECT @foo AS foo"
with pytest.raises(ValueError, match=r".*invalid literal.*\[PLENTY!\].*"):
cell_magic_args = "params_dict_df --max_results [PLENTY!]"
ip.run_cell_magic("bigquery", cell_magic_args, sql)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params_negative_value(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "params_dict_df"))
sql = "SELECT @num AS num"
result = pandas.DataFrame([-17], columns=["num"])
assert "params_dict_df" not in ip.user_ns
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
params = {"num": -17}
# Insert dictionary into user namespace so that it can be expanded
ip.user_ns["params"] = params
ip.run_cell_magic("bigquery", "params_dict_df --params $params", sql)
run_query_mock.assert_called_once_with(mock.ANY, sql.format(num=-17), mock.ANY)
assert "params_dict_df" in ip.user_ns # verify that the variable exists
df = ip.user_ns["params_dict_df"]
assert len(df) == len(result) # verify row count
assert list(df) == list(result) # verify column names
assert df["num"][0] == -17
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params_array_value(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "params_dict_df"))
sql = "SELECT @num AS num"
result = pandas.DataFrame(["foo bar", "baz quux"], columns=["array_data"])
assert "params_dict_df" not in ip.user_ns
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
params = {"array_data": ["foo bar", "baz quux"]}
# Insert dictionary into user namespace so that it can be expanded
ip.user_ns["params"] = params
ip.run_cell_magic("bigquery", "params_dict_df --params $params", sql)
run_query_mock.assert_called_once_with(mock.ANY, sql.format(num=-17), mock.ANY)
assert "params_dict_df" in ip.user_ns # verify that the variable exists
df = ip.user_ns["params_dict_df"]
assert len(df) == len(result) # verify row count
assert list(df) == list(result) # verify column names
assert list(df["array_data"]) == ["foo bar", "baz quux"]
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_dict_params_tuple_value(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "params_dict_df"))
sql = "SELECT @num AS num"
result = pandas.DataFrame(["foo bar", "baz quux"], columns=["array_data"])
assert "params_dict_df" not in ip.user_ns
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
query_job_mock.to_dataframe.return_value = result
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
params = {"array_data": ("foo bar", "baz quux")}
# Insert dictionary into user namespace so that it can be expanded
ip.user_ns["params"] = params
ip.run_cell_magic("bigquery", "params_dict_df --params $params", sql)
run_query_mock.assert_called_once_with(mock.ANY, sql.format(num=-17), mock.ANY)
assert "params_dict_df" in ip.user_ns # verify that the variable exists
df = ip.user_ns["params_dict_df"]
assert len(df) == len(result) # verify row count
assert list(df) == list(result) # verify column names
assert list(df["array_data"]) == ["foo bar", "baz quux"]
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_improperly_formatted_params():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
sql = "SELECT @num AS num"
with pytest.raises(SyntaxError):
ip.run_cell_magic("bigquery", "--params {17}", sql)
@pytest.mark.parametrize(
"raw_sql", ("SELECT answer AS 42", " \t SELECT answer AS 42 \t ")
)
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_valid_query_in_existing_variable(ipython_ns_cleanup, raw_sql):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
ipython_ns_cleanup.append((ip, "custom_query"))
ipython_ns_cleanup.append((ip, "query_results_df"))
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
query_job_mock = mock.create_autospec(
google.cloud.bigquery.job.QueryJob, instance=True
)
mock_result = pandas.DataFrame([42], columns=["answer"])
query_job_mock.to_dataframe.return_value = mock_result
ip.user_ns["custom_query"] = raw_sql
cell_body = "$custom_query" # Referring to an existing variable name (custom_query)
assert "query_results_df" not in ip.user_ns
with run_query_patch as run_query_mock:
run_query_mock.return_value = query_job_mock
ip.run_cell_magic("bigquery", "query_results_df", cell_body)
run_query_mock.assert_called_once_with(mock.ANY, raw_sql, mock.ANY)
assert "query_results_df" in ip.user_ns # verify that the variable exists
df = ip.user_ns["query_results_df"]
assert len(df) == len(mock_result) # verify row count
assert list(df) == list(mock_result) # verify column names
assert list(df["answer"]) == [42]
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_nonexisting_query_variable():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
ip.user_ns.pop("custom_query", None) # Make sure the variable does NOT exist.
cell_body = "$custom_query" # Referring to a non-existing variable name.
with pytest.raises(
NameError, match=r".*custom_query does not exist.*"
), run_query_patch as run_query_mock:
ip.run_cell_magic("bigquery", "", cell_body)
run_query_mock.assert_not_called()
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_empty_query_variable_name():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
cell_body = "$" # Not referring to any variable (name omitted).
with pytest.raises(
NameError, match=r"(?i).*missing query variable name.*"
), run_query_patch as run_query_mock:
ip.run_cell_magic("bigquery", "", cell_body)
run_query_mock.assert_not_called()
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_query_variable_non_string(ipython_ns_cleanup):
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
ipython_ns_cleanup.append((ip, "custom_query"))
ip.user_ns["custom_query"] = object()
cell_body = "$custom_query" # Referring to a non-string variable.
with pytest.raises(
TypeError, match=r".*must be a string or a bytes-like.*"
), run_query_patch as run_query_mock:
ip.run_cell_magic("bigquery", "", cell_body)
run_query_mock.assert_not_called()
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_query_variable_not_identifier():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
cell_body = "$123foo" # 123foo is not valid Python identifier
with io.capture_output() as captured_io:
ip.run_cell_magic("bigquery", "", cell_body)
# If "$" prefixes a string that is not a Python identifier, we do not treat such
# cell_body as a variable reference and just treat is as any other cell body input.
# If at the same time the cell body does not contain any whitespace, it is
# considered a table name, thus we expect an error that the table ID is not valid.
output = captured_io.stderr
assert "ERROR:" in output
assert "must be a fully-qualified ID" in output
@pytest.mark.usefixtures("ipython_interactive")
@pytest.mark.skipif(pandas is None, reason="Requires `pandas`")
def test_bigquery_magic_with_invalid_multiple_option_values():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
sql = "SELECT @foo AS foo"
exc_pattern = r".*[Uu]nrecognized input.*option values correct\?.*567.*"
with pytest.raises(ValueError, match=exc_pattern):
cell_magic_args = "params_dict_df --max_results 10 567"
ip.run_cell_magic("bigquery", cell_magic_args, sql)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_omits_tracebacks_from_error_message():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query",
autospec=True,
side_effect=exceptions.BadRequest("Syntax error in SQL query"),
)
with run_query_patch, default_patch, io.capture_output() as captured_io:
ip.run_cell_magic("bigquery", "", "SELECT foo FROM WHERE LIMIT bar")
output = captured_io.stderr
assert "400 Syntax error in SQL query" in output
assert "Traceback (most recent call last)" not in output
assert "Syntax error" not in captured_io.stdout
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_w_destination_table_invalid_format():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context._project = None
credentials_mock = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
default_patch = mock.patch(
"google.auth.default", return_value=(credentials_mock, "general-project")
)
client_patch = mock.patch(
"google.cloud.bigquery.magics.magics.bigquery.Client", autospec=True
)
with client_patch, default_patch, pytest.raises(ValueError) as exc_context:
ip.run_cell_magic(
"bigquery", "--destination_table dataset", "SELECT foo FROM WHERE LIMIT bar"
)
error_msg = str(exc_context.value)
assert (
"--destination_table should be in a "
"<dataset_id>.<table_id> format." in error_msg
)
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_w_destination_table():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
create_dataset_if_necessary_patch = mock.patch(
"google.cloud.bigquery.magics.magics._create_dataset_if_necessary",
autospec=True,
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with create_dataset_if_necessary_patch, run_query_patch as run_query_mock:
ip.run_cell_magic(
"bigquery",
"--destination_table dataset_id.table_id",
"SELECT foo FROM WHERE LIMIT bar",
)
job_config_used = run_query_mock.call_args_list[0][1]["job_config"]
assert job_config_used.allow_large_results is True
assert job_config_used.create_disposition == "CREATE_IF_NEEDED"
assert job_config_used.write_disposition == "WRITE_TRUNCATE"
assert job_config_used.destination.dataset_id == "dataset_id"
assert job_config_used.destination.table_id == "table_id"
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_create_dataset_fails():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
create_dataset_if_necessary_patch = mock.patch(
"google.cloud.bigquery.magics.magics._create_dataset_if_necessary",
autospec=True,
side_effect=OSError,
)
close_transports_patch = mock.patch(
"google.cloud.bigquery.magics.magics._close_transports",
autospec=True,
)
with pytest.raises(
OSError
), create_dataset_if_necessary_patch, close_transports_patch as close_transports:
ip.run_cell_magic(
"bigquery",
"--destination_table dataset_id.table_id",
"SELECT foo FROM WHERE LIMIT bar",
)
assert close_transports.called
@pytest.mark.usefixtures("ipython_interactive")
def test_bigquery_magic_with_location():
ip = IPython.get_ipython()
ip.extension_manager.load_extension("google.cloud.bigquery")
magics.context.credentials = mock.create_autospec(
google.auth.credentials.Credentials, instance=True
)
run_query_patch = mock.patch(
"google.cloud.bigquery.magics.magics._run_query", autospec=True
)
with run_query_patch as run_query_mock:
ip.run_cell_magic("bigquery", "--location=us-east1", "SELECT 17 AS num")
client_options_used = run_query_mock.call_args_list[0][0][0]
assert client_options_used.location == "us-east1"