Asib27's picture
try 1
065fee7 verified
raw
history blame
5.59 kB
# Copyright 2022 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Performance benchmarking main script. This is not an officially supported Google product."""
import argparse
import logging
import multiprocessing
import sys
from google.cloud import storage
import _perf_utils as _pu
import profile_w1r3 as w1r3
##### PROFILE BENCHMARKING TEST TYPES #####
PROFILE_WRITE_ONE_READ_THREE = "w1r3"
PROFILE_RANGE_READ = "range"
def main(args):
# Track error logging for BBMC reporting.
counter = _pu.logCount()
logging.basicConfig(
level=logging.ERROR,
handlers=[counter, logging.StreamHandler(sys.stderr)],
)
# Create a storage bucket to run benchmarking.
if args.project is not None:
client = storage.Client(project=args.project)
else:
client = storage.Client()
bucket = client.bucket(args.bucket)
if not bucket.exists():
bucket = client.create_bucket(bucket, location=args.bucket_region)
# Define test type and number of processes to run benchmarking.
# Note that transfer manager tests defaults to using 1 process.
num_processes = 1
test_type = args.test_type
if test_type == PROFILE_WRITE_ONE_READ_THREE:
num_processes = args.workers
benchmark_runner = w1r3.run_profile_w1r3
logging.info(
f"A total of {num_processes} processes are created to run benchmarking {test_type}"
)
elif test_type == PROFILE_RANGE_READ:
num_processes = args.workers
benchmark_runner = w1r3.run_profile_range_read
logging.info(
f"A total of {num_processes} processes are created to run benchmarking {test_type}"
)
# Allow multiprocessing to speed up benchmarking tests; Defaults to 1 for no concurrency.
p = multiprocessing.Pool(num_processes)
pool_output = p.map(benchmark_runner, [args for _ in range(args.samples)])
# Output to Cloud Monitoring or CSV file.
output_type = args.output_type
if output_type == "cloud-monitoring":
_pu.convert_to_cloud_monitoring(args.bucket, pool_output, num_processes)
elif output_type == "csv":
_pu.convert_to_csv(args.output_file, pool_output, num_processes)
logging.info(
f"Succesfully ran benchmarking. Please find your output log at {args.output_file}"
)
# Cleanup and delete blobs.
_pu.cleanup_bucket(bucket, delete_bucket=args.delete_bucket)
# BBMC will not surface errors unless the process is terminated with a non zero code.
if counter.count.errors != 0:
sys.exit(1)
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument(
"--project",
type=str,
default=None,
help="GCP project identifier",
)
parser.add_argument(
"--api",
type=str,
default="JSON",
help="API to use",
)
parser.add_argument(
"--test_type",
type=str,
default=PROFILE_WRITE_ONE_READ_THREE,
help="Benchmarking test type",
)
parser.add_argument(
"--object_size",
type=str,
default=_pu.DEFAULT_OBJECT_RANGE_SIZE_BYTES,
help="Object size in bytes; can be a range min..max",
)
parser.add_argument(
"--range_read_size",
type=int,
default=0,
help="Size of the range to read in bytes",
)
parser.add_argument(
"--minimum_read_offset",
type=int,
default=0,
help="Minimum offset for the start of the range to be read in bytes",
)
parser.add_argument(
"--maximum_read_offset",
type=int,
default=0,
help="Maximum offset for the start of the range to be read in bytes",
)
parser.add_argument(
"--samples",
type=int,
default=_pu.DEFAULT_NUM_SAMPLES,
help="Number of samples to report",
)
parser.add_argument(
"--workers",
type=int,
default=_pu.DEFAULT_NUM_PROCESSES,
help="Number of processes- multiprocessing enabled",
)
parser.add_argument(
"--bucket",
type=str,
default=_pu.DEFAULT_BUCKET_NAME,
help="Storage bucket name",
)
parser.add_argument(
"--bucket_region",
type=str,
default=_pu.DEFAULT_BUCKET_REGION,
help="Bucket region",
)
parser.add_argument(
"--output_type",
type=str,
default="cloud-monitoring",
help="Ouput format, csv or cloud-monitoring",
)
parser.add_argument(
"--output_file",
type=str,
default=_pu.DEFAULT_OUTPUT_FILE,
help="File to output results to",
)
parser.add_argument(
"--tmp_dir",
type=str,
default=_pu.DEFAULT_BASE_DIR,
help="Temp directory path on file system",
)
parser.add_argument(
"--delete_bucket",
type=bool,
default=False,
help="Whether or not to delete GCS bucket used for benchmarking",
)
args = parser.parse_args()
main(args)