Asib27's picture
try 1
065fee7 verified
|
raw
history blame
3.27 kB

python-storage benchmarking

This is not an officially supported Google product

This benchmarking script is used by Storage client library maintainers to benchmark various workloads and collect metrics in order to improve performance of the library. Currently the benchmarking runs a Write-1-Read-3 workload and measures the usual two QoS performance attributes, latency and throughput.

Run example:

This runs 10K iterations of Write-1-Read-3 on 5KiB to 16KiB files, and generates output to a default csv file output_bench<TIMESTAMP>.csv:

$ cd python-storage
$ pip install -e . # install google.cloud.storage locally
$ cd tests/perf
$ python3 benchmarking.py --num_samples 10000 --object_size 5120..16384 --output_type csv

CLI parameters

Parameter Description Possible values Default
--project GCP project identifier a project id *
--api API to use only JSON is currently supported in python benchmarking JSON
--output_type output results as csv records or cloud monitoring csv, cloud-monitoring cloud-monitoring
--object_size object size in bytes; can be a range min..max string 1048576 (1 MiB)
--range_read_size size of the range to read in bytes any positive integer
<=0 reads the full object
0
--minimum_read_offset minimum offset for the start of the range to be read in bytes any integer >0 0
--maximum_read_offset maximum offset for the start of the range to be read in bytes any integer >0 0
--samples number of W1R3 iterations any positive integer 8000
--bucket storage bucket name a bucket name pybench<TIMESTAMP>
--bucket_region bucket region for benchmarks any GCS region US-WEST1
--workers number of processes (multiprocessing enabled) any positive integer 16 (recommend not to exceed 16)
--test_type test type to run benchmarking w1r3, range w1r3
--output_file file to output results to any file path output_bench<TIMESTAMP>.csv
--tmp_dir temp directory path on file system any file path tm-perf-metrics
--delete_bucket whether or not to delete GCS bucket used for benchmarking bool False

Workload definition and CSV headers

For each invocation of the benchmark, write a new object of random size between min_size and max_size . After the successful write, download the object in full three times. For each of the 4 operations record the following fields:

Field Description
Op the name of the operations (WRITE, READ[{0,1,2}])
ObjectSize the number of bytes of the object
LibBufferSize configured to use the library default of 100 MiB
Crc32cEnabled bool: whether crc32c was computed for the operation
MD5Enabled bool: whether MD5 was computed for the operation
ApiName default to JSON
ElapsedTimeUs the elapsed time in microseconds the operation took
Status completion state of the operation [OK, FAIL]
RunID timestamp from the benchmarking run
AppBufferSize N/A
CpuTimeUs N/A