Asib27's picture
try 1
065fee7 verified
raw
history blame
41.9 kB
# Copyright 2021 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import datetime
import pytest
from google.api_core import exceptions
from . import _helpers
def test_bucket_create_w_alt_storage_class(storage_client, buckets_to_delete):
from google.cloud.storage import constants
bucket_name = _helpers.unique_name("bucket-w-archive")
with pytest.raises(exceptions.NotFound):
storage_client.get_bucket(bucket_name)
bucket = storage_client.bucket(bucket_name)
bucket.storage_class = constants.ARCHIVE_STORAGE_CLASS
_helpers.retry_429_503(bucket.create)()
buckets_to_delete.append(bucket)
created = storage_client.get_bucket(bucket_name)
assert created.storage_class == constants.ARCHIVE_STORAGE_CLASS
def test_bucket_lifecycle_rules(storage_client, buckets_to_delete):
from google.cloud.storage import constants
from google.cloud.storage.bucket import LifecycleRuleDelete
from google.cloud.storage.bucket import LifecycleRuleSetStorageClass
from google.cloud.storage.bucket import LifecycleRuleAbortIncompleteMultipartUpload
bucket_name = _helpers.unique_name("w-lifcycle-rules")
custom_time_before = datetime.date(2018, 8, 1)
noncurrent_before = datetime.date(2018, 8, 1)
matches_prefix = ["storage-sys-test", "gcs-sys-test"]
matches_suffix = ["suffix-test"]
with pytest.raises(exceptions.NotFound):
storage_client.get_bucket(bucket_name)
bucket = storage_client.bucket(bucket_name)
bucket.add_lifecycle_delete_rule(
age=42,
number_of_newer_versions=3,
days_since_custom_time=2,
custom_time_before=custom_time_before,
days_since_noncurrent_time=2,
noncurrent_time_before=noncurrent_before,
matches_prefix=matches_prefix,
matches_suffix=matches_suffix,
)
bucket.add_lifecycle_set_storage_class_rule(
constants.COLDLINE_STORAGE_CLASS,
is_live=False,
matches_storage_class=[constants.NEARLINE_STORAGE_CLASS],
)
bucket.add_lifecycle_abort_incomplete_multipart_upload_rule(
age=42,
)
expected_rules = [
LifecycleRuleDelete(
age=42,
number_of_newer_versions=3,
days_since_custom_time=2,
custom_time_before=custom_time_before,
days_since_noncurrent_time=2,
noncurrent_time_before=noncurrent_before,
matches_prefix=matches_prefix,
matches_suffix=matches_suffix,
),
LifecycleRuleSetStorageClass(
constants.COLDLINE_STORAGE_CLASS,
is_live=False,
matches_storage_class=[constants.NEARLINE_STORAGE_CLASS],
),
LifecycleRuleAbortIncompleteMultipartUpload(
age=42,
),
]
_helpers.retry_429_503(bucket.create)(location="us")
buckets_to_delete.append(bucket)
assert bucket.name == bucket_name
assert list(bucket.lifecycle_rules) == expected_rules
# Test modifying lifecycle rules
expected_rules[0] = LifecycleRuleDelete(
age=30,
matches_prefix=["new-prefix"],
matches_suffix=["new-suffix"],
)
rules = list(bucket.lifecycle_rules)
rules[0]["condition"] = {
"age": 30,
"matchesPrefix": ["new-prefix"],
"matchesSuffix": ["new-suffix"],
}
bucket.lifecycle_rules = rules
bucket.patch()
assert list(bucket.lifecycle_rules) == expected_rules
# Test clearing lifecycle rules
bucket.clear_lifecyle_rules()
bucket.patch()
assert list(bucket.lifecycle_rules) == []
@pytest.mark.skipif(
_helpers.is_api_endpoint_override,
reason="Test does not yet support endpoint override",
)
def test_bucket_update_labels(storage_client, buckets_to_delete):
bucket_name = _helpers.unique_name("update-labels")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
assert bucket.exists()
updated_labels = {"test-label": "label-value"}
bucket.labels = updated_labels
bucket.update()
assert bucket.labels == updated_labels
new_labels = {"another-label": "another-value"}
bucket.labels = new_labels
bucket.patch()
assert bucket.labels == new_labels
bucket.labels = {}
# See https://github.com/googleapis/python-storage/issues/541
retry_400 = _helpers.RetryErrors(exceptions.BadRequest)
retry_400(bucket.update)()
assert bucket.labels == {}
def test_bucket_get_set_iam_policy(
storage_client,
buckets_to_delete,
service_account,
):
from google.cloud.storage.iam import STORAGE_OBJECT_VIEWER_ROLE
from google.api_core.exceptions import BadRequest
from google.api_core.exceptions import PreconditionFailed
bucket_name = _helpers.unique_name("iam-policy")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
assert bucket.exists()
policy_no_version = bucket.get_iam_policy()
assert policy_no_version.version == 1
policy = bucket.get_iam_policy(requested_policy_version=3)
assert policy == policy_no_version
member = f"serviceAccount:{storage_client.get_service_account_email()}"
binding_w_condition = {
"role": STORAGE_OBJECT_VIEWER_ROLE,
"members": {member},
"condition": {
"title": "always-true",
"description": "test condition always-true",
"expression": "true",
},
}
policy.bindings.append(binding_w_condition)
with pytest.raises(PreconditionFailed, match="enable uniform bucket-level access"):
bucket.set_iam_policy(policy)
bucket.iam_configuration.uniform_bucket_level_access_enabled = True
bucket.patch()
policy = bucket.get_iam_policy(requested_policy_version=3)
policy.bindings.append(binding_w_condition)
with pytest.raises(BadRequest, match="at least 3"):
bucket.set_iam_policy(policy)
policy.version = 3
returned_policy = bucket.set_iam_policy(policy)
assert returned_policy.version == 3
assert returned_policy.bindings == policy.bindings
fetched_policy = bucket.get_iam_policy(requested_policy_version=3)
assert fetched_policy.bindings == returned_policy.bindings
def test_bucket_crud_w_requester_pays(storage_client, buckets_to_delete, user_project):
bucket_name = _helpers.unique_name("w-requester-pays")
created = _helpers.retry_429_503(storage_client.create_bucket)(
bucket_name, requester_pays=True
)
buckets_to_delete.append(created)
assert created.name == bucket_name
assert created.requester_pays
with_user_project = storage_client.bucket(
bucket_name,
user_project=user_project,
)
try:
# Exercise 'buckets.get' w/ userProject.
assert with_user_project.exists()
with_user_project.reload()
assert with_user_project.requester_pays
# Exercise 'buckets.patch' w/ userProject.
with_user_project.configure_website(
main_page_suffix="index.html", not_found_page="404.html"
)
with_user_project.patch()
expected_website = {"mainPageSuffix": "index.html", "notFoundPage": "404.html"}
assert with_user_project._properties["website"] == expected_website
# Exercise 'buckets.update' w/ userProject.
new_labels = {"another-label": "another-value"}
with_user_project.labels = new_labels
with_user_project.update()
assert with_user_project.labels == new_labels
finally:
# Exercise 'buckets.delete' w/ userProject.
with_user_project.delete()
buckets_to_delete.remove(created)
def test_bucket_acls_iam_w_user_project(
storage_client, buckets_to_delete, user_project
):
bucket_name = _helpers.unique_name("acl-w-user-project")
created = _helpers.retry_429_503(storage_client.create_bucket)(
bucket_name,
requester_pays=True,
)
buckets_to_delete.append(created)
with_user_project = storage_client.bucket(bucket_name, user_project=user_project)
# Exercise bucket ACL w/ userProject
acl = with_user_project.acl
acl.reload()
acl.all().grant_read()
acl.save()
assert "READER" in acl.all().get_roles()
del acl.entities["allUsers"]
acl.save()
assert not acl.has_entity("allUsers")
# Exercise default object ACL w/ userProject
doa = with_user_project.default_object_acl
doa.reload()
doa.all().grant_read()
doa.save()
assert "READER" in doa.all().get_roles()
# Exercise IAM w/ userProject
test_permissions = ["storage.buckets.get"]
found = with_user_project.test_iam_permissions(test_permissions)
assert found == test_permissions
policy = with_user_project.get_iam_policy()
viewers = policy.setdefault("roles/storage.objectViewer", set())
viewers.add(policy.all_users())
with_user_project.set_iam_policy(policy)
def test_bucket_acls_w_metageneration_match(storage_client, buckets_to_delete):
wrong_metageneration_number = 9
bucket_name = _helpers.unique_name("acl-w-metageneration-match")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
# Exercise bucket ACL with metageneration match
acl = bucket.acl
acl.group("cloud-developer-relations@google.com").grant_read()
bucket.reload()
with pytest.raises(exceptions.PreconditionFailed):
acl.save(if_metageneration_match=wrong_metageneration_number)
assert (
"READER"
not in acl.group("cloud-developer-relations@google.com").get_roles()
)
acl.save(if_metageneration_match=bucket.metageneration)
assert "READER" in acl.group("cloud-developer-relations@google.com").get_roles()
# Exercise default object ACL w/ metageneration match
doa = bucket.default_object_acl
doa.group("cloud-developer-relations@google.com").grant_owner()
bucket.reload()
with pytest.raises(exceptions.PreconditionFailed):
doa.save(if_metageneration_match=wrong_metageneration_number)
assert (
"OWNER" not in doa.group("cloud-developer-relations@google.com").get_roles()
)
doa.save(if_metageneration_match=bucket.metageneration)
assert "OWNER" in doa.group("cloud-developer-relations@google.com").get_roles()
def test_bucket_copy_blob(
storage_client,
buckets_to_delete,
blobs_to_delete,
user_project,
):
payload = b"DEADBEEF"
bucket_name = _helpers.unique_name("copy-blob")
created = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(created)
assert created.name == bucket_name
blob = created.blob("CloudLogo")
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
new_blob = _helpers.retry_bad_copy(created.copy_blob)(
blob, created, "CloudLogoCopy"
)
blobs_to_delete.append(new_blob)
copied_contents = new_blob.download_as_bytes()
assert copied_contents == payload
def test_bucket_copy_blob_w_user_project(
storage_client,
buckets_to_delete,
blobs_to_delete,
user_project,
):
payload = b"DEADBEEF"
bucket_name = _helpers.unique_name("copy-w-requester-pays")
created = _helpers.retry_429_503(storage_client.create_bucket)(
bucket_name, requester_pays=True
)
buckets_to_delete.append(created)
assert created.name == bucket_name
assert created.requester_pays
blob = created.blob("simple")
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
with_user_project = storage_client.bucket(bucket_name, user_project=user_project)
new_blob = _helpers.retry_bad_copy(with_user_project.copy_blob)(
blob, with_user_project, "simple-copy"
)
blobs_to_delete.append(new_blob)
assert new_blob.download_as_bytes() == payload
def test_bucket_copy_blob_w_generation_match(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
payload = b"DEADBEEF"
bucket_name = _helpers.unique_name("generation-match")
created = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(created)
assert created.name == bucket_name
blob = created.blob("simple")
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
dest_bucket = storage_client.bucket(bucket_name)
new_blob = dest_bucket.copy_blob(
blob,
dest_bucket,
"simple-copy",
if_source_generation_match=blob.generation,
)
blobs_to_delete.append(new_blob)
assert new_blob.download_as_bytes() == payload
def test_bucket_copy_blob_w_metageneration_match(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
payload = b"DEADBEEF"
bucket_name = _helpers.unique_name("generation-match")
bucket = storage_client.bucket(bucket_name)
bucket.requester_pays = True
created = _helpers.retry_429_503(storage_client.create_bucket)(bucket)
buckets_to_delete.append(created)
assert created.name == bucket_name
blob = created.blob("simple")
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
dest_bucket = storage_client.bucket(bucket_name)
new_blob = dest_bucket.copy_blob(
blob,
dest_bucket,
"simple-copy",
if_source_metageneration_match=blob.metageneration,
)
blobs_to_delete.append(new_blob)
assert new_blob.download_as_bytes() == payload
def test_bucket_get_blob_with_user_project(
storage_client,
buckets_to_delete,
blobs_to_delete,
user_project,
):
blob_name = "blob-name"
payload = b"DEADBEEF"
bucket_name = _helpers.unique_name("w-requester-pays")
created = _helpers.retry_429_503(storage_client.create_bucket)(
bucket_name, requester_pays=True
)
buckets_to_delete.append(created)
assert created.name == bucket_name
assert created.requester_pays
with_user_project = storage_client.bucket(bucket_name, user_project=user_project)
assert with_user_project.get_blob("nonesuch") is None
to_add = created.blob(blob_name)
to_add.upload_from_string(payload)
blobs_to_delete.append(to_add)
found = with_user_project.get_blob(blob_name)
assert found.download_as_bytes() == payload
@_helpers.retry_failures
def test_bucket_list_blobs(listable_bucket, listable_filenames):
all_blobs = list(listable_bucket.list_blobs())
assert sorted(blob.name for blob in all_blobs) == sorted(listable_filenames)
@_helpers.retry_failures
def test_bucket_list_blobs_w_user_project(
storage_client,
listable_bucket,
listable_filenames,
user_project,
):
with_user_project = storage_client.bucket(
listable_bucket.name, user_project=user_project
)
all_blobs = list(with_user_project.list_blobs())
assert sorted(blob.name for blob in all_blobs) == sorted(listable_filenames)
@_helpers.retry_failures
def test_bucket_list_blobs_paginated(listable_bucket, listable_filenames):
truncation_size = 1
count = len(listable_filenames) - truncation_size
iterator = listable_bucket.list_blobs(max_results=count)
page_iter = iterator.pages
page1 = next(page_iter)
blobs = list(page1)
assert len(blobs) == count
assert iterator.next_page_token is not None
# Technically the iterator is exhausted.
assert iterator.num_results == iterator.max_results
# But we modify the iterator to continue paging after
# artificially stopping after ``count`` items.
iterator.max_results = None
page2 = next(page_iter)
last_blobs = list(page2)
assert len(last_blobs) == truncation_size
@_helpers.retry_failures
def test_bucket_list_blobs_paginated_w_offset(listable_bucket, listable_filenames):
truncation_size = 1
inclusive_start_offset = listable_filenames[1]
exclusive_end_offset = listable_filenames[-1]
desired_files = listable_filenames[1:-1]
count = len(desired_files) - truncation_size
iterator = listable_bucket.list_blobs(
max_results=count,
start_offset=inclusive_start_offset,
end_offset=exclusive_end_offset,
)
page_iter = iterator.pages
page1 = next(page_iter)
blobs = list(page1)
assert len(blobs) == count
assert blobs[0].name == desired_files[0]
assert iterator.next_page_token is not None
# Technically the iterator is exhausted.
assert iterator.num_results == iterator.max_results
# But we modify the iterator to continue paging after
# artificially stopping after ``count`` items.
iterator.max_results = None
page2 = next(page_iter)
last_blobs = list(page2)
assert len(last_blobs) == truncation_size
assert last_blobs[-1].name == desired_files[-1]
@_helpers.retry_failures
def test_blob_exists_hierarchy(hierarchy_bucket, hierarchy_filenames):
for filename in hierarchy_filenames:
blob = hierarchy_bucket.blob(filename)
assert blob.exists()
@_helpers.retry_failures
def test_bucket_list_blobs_hierarchy_root_level(hierarchy_bucket, hierarchy_filenames):
expected_names = ["file01.txt"]
expected_prefixes = set(["parent/"])
iterator = hierarchy_bucket.list_blobs(delimiter="/")
page = next(iterator.pages)
blobs = list(page)
assert [blob.name for blob in blobs] == expected_names
assert iterator.next_page_token is None
assert iterator.prefixes == expected_prefixes
@_helpers.retry_failures
def test_bucket_list_blobs_hierarchy_first_level(hierarchy_bucket, hierarchy_filenames):
expected_names = ["parent/", "parent/file11.txt"]
expected_prefixes = set(["parent/child/"])
iterator = hierarchy_bucket.list_blobs(delimiter="/", prefix="parent/")
page = next(iterator.pages)
blobs = list(page)
assert [blob.name for blob in blobs] == expected_names
assert iterator.next_page_token is None
assert iterator.prefixes == expected_prefixes
@_helpers.retry_failures
def test_bucket_list_blobs_hierarchy_second_level(
hierarchy_bucket, hierarchy_filenames
):
expected_names = ["parent/child/file21.txt", "parent/child/file22.txt"]
expected_prefixes = set(["parent/child/grand/", "parent/child/other/"])
iterator = hierarchy_bucket.list_blobs(delimiter="/", prefix="parent/child/")
page = next(iterator.pages)
blobs = list(page)
assert [blob.name for blob in blobs] == expected_names
assert iterator.next_page_token is None
assert iterator.prefixes == expected_prefixes
@_helpers.retry_failures
def test_bucket_list_blobs_hierarchy_third_level(hierarchy_bucket, hierarchy_filenames):
# Pseudo-hierarchy can be arbitrarily deep, subject to the limit
# of 1024 characters in the UTF-8 encoded name:
# https://cloud.google.com/storage/docs/bucketnaming#objectnames
# Exercise a layer deeper to illustrate this.
expected_names = ["parent/child/grand/file31.txt"]
expected_prefixes = set()
iterator = hierarchy_bucket.list_blobs(delimiter="/", prefix="parent/child/grand/")
page = next(iterator.pages)
blobs = list(page)
assert [blob.name for blob in blobs] == expected_names
assert iterator.next_page_token is None
assert iterator.prefixes == expected_prefixes
@_helpers.retry_failures
def test_bucket_list_blobs_hierarchy_w_include_trailing_delimiter(
hierarchy_bucket,
hierarchy_filenames,
):
expected_names = ["file01.txt", "parent/"]
expected_prefixes = set(["parent/"])
iterator = hierarchy_bucket.list_blobs(
delimiter="/", include_trailing_delimiter=True
)
page = next(iterator.pages)
blobs = list(page)
assert [blob.name for blob in blobs] == expected_names
assert iterator.next_page_token is None
assert iterator.prefixes == expected_prefixes
@_helpers.retry_failures
def test_bucket_list_blobs_w_match_glob(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
bucket_name = _helpers.unique_name("w-matchglob")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
payload = b"helloworld"
blob_names = ["foo/bar", "foo/baz", "foo/foobar", "foobar"]
for name in blob_names:
blob = bucket.blob(name)
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
match_glob_results = {
"foo*bar": ["foobar"],
"foo**bar": ["foo/bar", "foo/foobar", "foobar"],
"**/foobar": ["foo/foobar", "foobar"],
"*/ba[rz]": ["foo/bar", "foo/baz"],
"*/ba[!a-y]": ["foo/baz"],
"**/{foobar,baz}": ["foo/baz", "foo/foobar", "foobar"],
"foo/{foo*,*baz}": ["foo/baz", "foo/foobar"],
}
for match_glob, expected_names in match_glob_results.items():
blob_iter = bucket.list_blobs(match_glob=match_glob)
blobs = list(blob_iter)
assert [blob.name for blob in blobs] == expected_names
def test_bucket_list_blobs_include_managed_folders(
storage_client,
buckets_to_delete,
blobs_to_delete,
hierarchy_filenames,
):
bucket_name = _helpers.unique_name("ubla-mf")
bucket = storage_client.bucket(bucket_name)
bucket.iam_configuration.uniform_bucket_level_access_enabled = True
_helpers.retry_429_503(bucket.create)()
buckets_to_delete.append(bucket)
payload = b"helloworld"
for filename in hierarchy_filenames:
blob = bucket.blob(filename)
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
# Make API call to create a managed folder.
# TODO: change to use storage control client once available.
path = f"/b/{bucket_name}/managedFolders"
properties = {"name": "managedfolder1"}
storage_client._post_resource(path, properties)
expected_prefixes = set(["parent/"])
blob_iter = bucket.list_blobs(delimiter="/")
list(blob_iter)
assert blob_iter.prefixes == expected_prefixes
# Test that managed folders are only included when IncludeFoldersAsPrefixes is set.
expected_prefixes = set(["parent/", "managedfolder1/"])
blob_iter = bucket.list_blobs(delimiter="/", include_folders_as_prefixes=True)
list(blob_iter)
assert blob_iter.prefixes == expected_prefixes
# Cleanup: API call to delete a managed folder.
# TODO: change to use storage control client once available.
path = f"/b/{bucket_name}/managedFolders/managedfolder1"
storage_client._delete_resource(path)
def test_bucket_update_retention_period(
storage_client,
buckets_to_delete,
):
period_secs = 3
bucket_name = _helpers.unique_name("w-retention-period")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
bucket.retention_period = period_secs
bucket.default_event_based_hold = False
bucket.patch()
# Changes to the bucket will be readable immediately after writing,
# but configuration changes may take time to propagate.
_helpers.retry_has_retention_period(bucket.reload)()
assert bucket.retention_period == period_secs
assert isinstance(bucket.retention_policy_effective_time, datetime.datetime)
assert not bucket.default_event_based_hold
assert not bucket.retention_policy_locked
bucket.retention_period = None
bucket.patch()
# Changes to the bucket will be readable immediately after writing,
# but configuration changes may take time to propagate.
_helpers.retry_no_retention_period(bucket.reload)()
assert bucket.retention_period is None
assert bucket.retention_policy_effective_time is None
assert not bucket.default_event_based_hold
assert not bucket.retention_policy_locked
def test_delete_object_bucket_w_retention_period(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
# Create a bucket with retention period.
period_secs = 12
bucket = storage_client.bucket(_helpers.unique_name("w-retention-period"))
bucket.retention_period = period_secs
bucket.default_event_based_hold = False
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket)
buckets_to_delete.append(bucket)
_helpers.retry_has_retention_period(bucket.reload)()
assert bucket.retention_period == period_secs
assert isinstance(bucket.retention_policy_effective_time, datetime.datetime)
payload = b"DEADBEEF"
blob = bucket.blob(_helpers.unique_name("w-retention"))
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
_helpers.retry_has_retention_expiration(blob.reload)()
assert isinstance(blob.retention_expiration_time, datetime.datetime)
assert not blob.event_based_hold
assert not blob.temporary_hold
# Attempts to delete objects whose age is less than the retention period should fail.
with pytest.raises(exceptions.Forbidden):
blob.delete()
# Object can be deleted once it reaches the age defined in the retention policy.
_helpers.await_config_changes_propagate(sec=period_secs)
blob.delete()
blobs_to_delete.pop()
def test_bucket_w_default_event_based_hold(
storage_client,
blobs_to_delete,
default_ebh_bucket,
):
bucket = storage_client.get_bucket(default_ebh_bucket)
assert bucket.default_event_based_hold
assert bucket.retention_period is None
assert bucket.retention_policy_effective_time is None
assert not bucket.retention_policy_locked
blob_name = "test-blob"
payload = b"DEADBEEF"
blob = bucket.blob(blob_name)
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
other = bucket.get_blob(blob_name)
assert other.event_based_hold
assert not other.temporary_hold
assert other.retention_expiration_time is None
with pytest.raises(exceptions.Forbidden):
other.delete()
other.event_based_hold = False
other.patch()
other.delete()
bucket.default_event_based_hold = False
bucket.patch()
assert not bucket.default_event_based_hold
assert bucket.retention_period is None
assert bucket.retention_policy_effective_time is None
assert not bucket.retention_policy_locked
# Changes to the bucket will be readable immediately after writing,
# but configuration changes may take time to propagate.
_helpers.await_config_changes_propagate()
blob.upload_from_string(payload)
# https://github.com/googleapis/python-storage/issues/435
_helpers.retry_no_event_based_hold(blob.reload)()
assert not blob.event_based_hold
assert not blob.temporary_hold
assert blob.retention_expiration_time is None
blob.delete()
blobs_to_delete.pop()
def test_blob_w_temporary_hold(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
bucket_name = _helpers.unique_name("w-tmp-hold")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
blob_name = "test-blob"
payload = b"DEADBEEF"
blob = bucket.blob(blob_name)
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
other = bucket.get_blob(blob_name)
other.temporary_hold = True
other.patch()
assert other.temporary_hold
assert not other.event_based_hold
assert other.retention_expiration_time is None
with pytest.raises(exceptions.Forbidden):
other.delete()
other.temporary_hold = False
other.patch()
other.delete()
blobs_to_delete.pop()
def test_bucket_lock_retention_policy(
storage_client,
buckets_to_delete,
):
period_secs = 10
bucket_name = _helpers.unique_name("loc-ret-policy")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
bucket.retention_period = period_secs
bucket.patch()
assert bucket.retention_period == period_secs
assert isinstance(bucket.retention_policy_effective_time, datetime.datetime)
assert not bucket.default_event_based_hold
assert not bucket.retention_policy_locked
bucket.lock_retention_policy()
bucket.reload()
assert bucket.retention_policy_locked
bucket.retention_period = None
with pytest.raises(exceptions.Forbidden):
bucket.patch()
@pytest.mark.skipif(
_helpers.is_api_endpoint_override,
reason="Test does not yet support endpoint override",
)
def test_new_bucket_w_ubla(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
bucket_name = _helpers.unique_name("new-w-ubla")
bucket = storage_client.bucket(bucket_name)
bucket.iam_configuration.uniform_bucket_level_access_enabled = True
_helpers.retry_429_503(bucket.create)()
buckets_to_delete.append(bucket)
bucket_acl = bucket.acl
with pytest.raises(exceptions.BadRequest):
bucket_acl.reload()
bucket_acl.loaded = True # Fake that we somehow loaded the ACL
bucket_acl.group("cloud-developer-relations@google.com").grant_read()
with pytest.raises(exceptions.BadRequest):
bucket_acl.save()
blob_name = "my-blob.txt"
blob = bucket.blob(blob_name)
payload = b"DEADBEEF"
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
found = bucket.get_blob(blob_name)
assert found.download_as_bytes() == payload
blob_acl = blob.acl
with pytest.raises(exceptions.BadRequest):
blob_acl.reload()
blob_acl.loaded = True # Fake that we somehow loaded the ACL
blob_acl.group("cloud-developer-relations@google.com").grant_read()
with pytest.raises(exceptions.BadRequest):
blob_acl.save()
def test_ubla_set_unset_preserves_acls(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
bucket_name = _helpers.unique_name("ubla-acls")
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket_name)
buckets_to_delete.append(bucket)
blob_name = "my-blob.txt"
blob = bucket.blob(blob_name)
payload = b"DEADBEEF"
blob.upload_from_string(payload)
blobs_to_delete.append(blob)
# Preserve ACLs before setting UBLA
bucket_acl_before = list(bucket.acl)
blob_acl_before = list(bucket.acl)
# Set UBLA
bucket.iam_configuration.uniform_bucket_level_access_enabled = True
bucket.patch()
assert bucket.iam_configuration.uniform_bucket_level_access_enabled
# While UBLA is set, cannot get / set ACLs
with pytest.raises(exceptions.BadRequest):
bucket.acl.reload()
# Clear UBLA
bucket.iam_configuration.uniform_bucket_level_access_enabled = False
bucket.patch()
_helpers.await_config_changes_propagate()
# Query ACLs after clearing UBLA
bucket.acl.reload()
bucket_acl_after = list(bucket.acl)
blob.acl.reload()
blob_acl_after = list(bucket.acl)
assert bucket_acl_before == bucket_acl_after
assert blob_acl_before == blob_acl_after
def test_new_bucket_created_w_inherited_pap(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
from google.cloud.storage import constants
bucket_name = _helpers.unique_name("new-w-pap-inherited")
bucket = storage_client.bucket(bucket_name)
bucket.iam_configuration.uniform_bucket_level_access_enabled = True
bucket.create()
buckets_to_delete.append(bucket)
# TODO: Remove unspecified after changeover is complete
assert bucket.iam_configuration.public_access_prevention in [
constants.PUBLIC_ACCESS_PREVENTION_UNSPECIFIED,
constants.PUBLIC_ACCESS_PREVENTION_INHERITED,
]
bucket.iam_configuration.public_access_prevention = (
constants.PUBLIC_ACCESS_PREVENTION_ENFORCED
)
bucket.patch()
assert (
bucket.iam_configuration.public_access_prevention
== constants.PUBLIC_ACCESS_PREVENTION_ENFORCED
)
assert bucket.iam_configuration.uniform_bucket_level_access_enabled
bucket.iam_configuration.uniform_bucket_level_access_enabled = False
bucket.patch()
_helpers.await_config_changes_propagate()
assert (
bucket.iam_configuration.public_access_prevention
== constants.PUBLIC_ACCESS_PREVENTION_ENFORCED
)
with pytest.raises(exceptions.BadRequest):
bucket.iam_configuration.public_access_prevention = "unexpected value"
bucket.patch()
with pytest.raises(exceptions.PreconditionFailed):
bucket.make_public()
blob_name = "my-blob.txt"
blob = bucket.blob(blob_name)
payload = b"DEADBEEF"
blob.upload_from_string(payload)
with pytest.raises(exceptions.PreconditionFailed):
blob.make_public()
@pytest.mark.skip(reason="Unspecified PAP is changing to inherited")
def test_new_bucket_created_w_enforced_pap(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
from google.cloud.storage import constants
bucket_name = _helpers.unique_name("new-w-pap-enforced")
bucket = storage_client.bucket(bucket_name)
bucket.iam_configuration.public_access_prevention = (
constants.PUBLIC_ACCESS_PREVENTION_ENFORCED
)
bucket.create()
buckets_to_delete.append(bucket)
assert (
bucket.iam_configuration.public_access_prevention
== constants.PUBLIC_ACCESS_PREVENTION_ENFORCED
)
bucket.iam_configuration.public_access_prevention = (
constants.PUBLIC_ACCESS_PREVENTION_INHERITED
)
bucket.patch()
# TODO: Remove unspecified after changeover is complete
assert bucket.iam_configuration.public_access_prevention in [
constants.PUBLIC_ACCESS_PREVENTION_UNSPECIFIED,
constants.PUBLIC_ACCESS_PREVENTION_INHERITED,
]
assert not bucket.iam_configuration.uniform_bucket_level_access_enabled
@pytest.mark.skipif(
_helpers.is_api_endpoint_override,
reason="Test does not yet support endpoint override",
)
def test_new_bucket_with_rpo(
storage_client,
buckets_to_delete,
blobs_to_delete,
):
from google.cloud.storage import constants
bucket_name = _helpers.unique_name("new-w-turbo-replication")
bucket = storage_client.create_bucket(bucket_name, location="NAM4")
buckets_to_delete.append(bucket)
assert bucket.rpo == constants.RPO_DEFAULT
bucket.rpo = constants.RPO_ASYNC_TURBO
bucket.patch()
bucket_from_server = storage_client.get_bucket(bucket_name)
assert bucket_from_server.rpo == constants.RPO_ASYNC_TURBO
def test_new_bucket_with_autoclass(
storage_client,
buckets_to_delete,
):
from google.cloud.storage import constants
# Autoclass can be enabled via bucket create
bucket_name = _helpers.unique_name("new-w-autoclass")
bucket_obj = storage_client.bucket(bucket_name)
bucket_obj.autoclass_enabled = True
bucket = storage_client.create_bucket(bucket_obj)
previous_toggle_time = bucket.autoclass_toggle_time
buckets_to_delete.append(bucket)
# Autoclass terminal_storage_class is defaulted to NEARLINE if not specified
assert bucket.autoclass_enabled is True
assert bucket.autoclass_terminal_storage_class == constants.NEARLINE_STORAGE_CLASS
# Autoclass can be enabled/disabled via bucket patch
bucket.autoclass_enabled = False
bucket.patch(if_metageneration_match=bucket.metageneration)
assert bucket.autoclass_enabled is False
assert bucket.autoclass_toggle_time != previous_toggle_time
def test_bucket_delete_force(storage_client):
bucket_name = _helpers.unique_name("version-disabled")
bucket_obj = storage_client.bucket(bucket_name)
bucket = storage_client.create_bucket(bucket_obj)
BLOB_NAME = "my_object"
blob = bucket.blob(BLOB_NAME)
blob.upload_from_string("abcd")
blob.upload_from_string("efgh")
blobs = bucket.list_blobs(versions=True)
counter = 0
for blob in blobs:
counter += 1
assert blob.name == BLOB_NAME
assert counter == 1
bucket.delete(force=True) # Will fail with 409 if blobs aren't deleted
def test_bucket_delete_force_works_with_versions(storage_client):
bucket_name = _helpers.unique_name("version-enabled")
bucket_obj = storage_client.bucket(bucket_name)
bucket_obj.versioning_enabled = True
bucket = storage_client.create_bucket(bucket_obj)
assert bucket.versioning_enabled
BLOB_NAME = "my_versioned_object"
blob = bucket.blob(BLOB_NAME)
blob.upload_from_string("abcd")
blob.upload_from_string("efgh")
blobs = bucket.list_blobs(versions=True)
counter = 0
for blob in blobs:
counter += 1
assert blob.name == BLOB_NAME
assert counter == 2
bucket.delete(force=True) # Will fail with 409 if versions aren't deleted
def test_config_autoclass_w_existing_bucket(
storage_client,
buckets_to_delete,
):
from google.cloud.storage import constants
bucket_name = _helpers.unique_name("for-autoclass")
bucket = storage_client.create_bucket(bucket_name)
buckets_to_delete.append(bucket)
assert bucket.autoclass_enabled is False
assert bucket.autoclass_toggle_time is None
assert bucket.autoclass_terminal_storage_class is None
assert bucket.autoclass_terminal_storage_class_update_time is None
# Enable Autoclass on existing buckets with terminal_storage_class set to ARCHIVE
bucket.autoclass_enabled = True
bucket.autoclass_terminal_storage_class = constants.ARCHIVE_STORAGE_CLASS
bucket.patch(if_metageneration_match=bucket.metageneration)
previous_tsc_update_time = bucket.autoclass_terminal_storage_class_update_time
assert bucket.autoclass_enabled is True
assert bucket.autoclass_terminal_storage_class == constants.ARCHIVE_STORAGE_CLASS
# Configure Autoclass terminal_storage_class to NEARLINE
bucket.autoclass_terminal_storage_class = constants.NEARLINE_STORAGE_CLASS
bucket.patch(if_metageneration_match=bucket.metageneration)
assert bucket.autoclass_enabled is True
assert bucket.autoclass_terminal_storage_class == constants.NEARLINE_STORAGE_CLASS
assert (
bucket.autoclass_terminal_storage_class_update_time != previous_tsc_update_time
)
def test_soft_delete_policy(
storage_client,
buckets_to_delete,
):
from google.cloud.storage.bucket import SoftDeletePolicy
# Create a bucket with soft delete policy.
duration_secs = 7 * 86400
bucket = storage_client.bucket(_helpers.unique_name("w-soft-delete"))
bucket.soft_delete_policy.retention_duration_seconds = duration_secs
bucket = _helpers.retry_429_503(storage_client.create_bucket)(bucket)
buckets_to_delete.append(bucket)
policy = bucket.soft_delete_policy
assert isinstance(policy, SoftDeletePolicy)
assert policy.retention_duration_seconds == duration_secs
assert isinstance(policy.effective_time, datetime.datetime)
# Insert an object and get object metadata prior soft-deleted.
payload = b"DEADBEEF"
blob_name = _helpers.unique_name("soft-delete")
blob = bucket.blob(blob_name)
blob.upload_from_string(payload)
blob = bucket.get_blob(blob_name)
gen = blob.generation
assert blob.soft_delete_time is None
assert blob.hard_delete_time is None
# Delete the object to enter soft-deleted state.
blob.delete()
iter_default = bucket.list_blobs()
assert len(list(iter_default)) == 0
iter_w_soft_delete = bucket.list_blobs(soft_deleted=True)
assert len(list(iter_w_soft_delete)) > 0
# Get the soft-deleted object.
soft_deleted_blob = bucket.get_blob(blob_name, generation=gen, soft_deleted=True)
assert soft_deleted_blob.soft_delete_time is not None
assert soft_deleted_blob.hard_delete_time is not None
# Restore the soft-deleted object.
restored_blob = bucket.restore_blob(blob_name, generation=gen)
assert restored_blob.exists() is True
assert restored_blob.generation != gen
# Patch the soft delete policy on an existing bucket.
new_duration_secs = 10 * 86400
bucket.soft_delete_policy.retention_duration_seconds = new_duration_secs
bucket.patch()
assert bucket.soft_delete_policy.retention_duration_seconds == new_duration_secs
def test_new_bucket_with_hierarchical_namespace(
storage_client,
buckets_to_delete,
):
# Test new bucket without specifying hierarchical namespace
bucket_name = _helpers.unique_name("new-wo-hns")
bucket_obj = storage_client.bucket(bucket_name)
bucket = storage_client.create_bucket(bucket_obj)
buckets_to_delete.append(bucket)
assert bucket.hierarchical_namespace_enabled is None
# Test new bucket with hierarchical namespace disabled
bucket_name = _helpers.unique_name("new-hns-disabled")
bucket_obj = storage_client.bucket(bucket_name)
bucket_obj.hierarchical_namespace_enabled = False
bucket = storage_client.create_bucket(bucket_obj)
buckets_to_delete.append(bucket)
assert bucket.hierarchical_namespace_enabled is False
# Test new bucket with hierarchical namespace enabled
bucket_name = _helpers.unique_name("new-hns-enabled")
bucket_obj = storage_client.bucket(bucket_name)
bucket_obj.hierarchical_namespace_enabled = True
bucket_obj.iam_configuration.uniform_bucket_level_access_enabled = True
bucket = storage_client.create_bucket(bucket_obj)
buckets_to_delete.append(bucket)
assert bucket.hierarchical_namespace_enabled is True