The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError Exception: ArrowNotImplementedError Message: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1870, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 620, in write_table self._build_writer(inferred_schema=pa_table.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 441, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1886, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 639, in finalize self._build_writer(self.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 441, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1417, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1049, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1000, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1741, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1897, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
config
dict | report
dict | name
string | backend
dict | scenario
dict | launcher
dict | environment
dict | print_report
bool | log_report
bool | overall
dict | warmup
dict | train
dict |
---|---|---|---|---|---|---|---|---|---|---|---|
{
"name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2497.712128,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6545917870000153,
0.5378263789999664,
0.5645145930000126,
0.5367058049999969,
0.5674595629999999
],
"count": 5,
"total": 2.861098126999991,
"mean": 0.5722196253999983,
"p50": 0.5645145930000126,
"p90": 0.6197388974000091,
"p95": 0.6371653422000122,
"p99": 0.6511064980400146,
"stdev": 0.04315396983437998,
"stdev_": 7.541504680866913
},
"throughput": {
"unit": "samples/s",
"value": 17.475807462929478
},
"energy": {
"unit": "kWh",
"cpu": 0.00011792815879444354,
"ram": 0.000004929245466240147,
"gpu": 0,
"total": 0.0001228574042606837
},
"efficiency": {
"unit": "samples/kWh",
"value": 81395.1756524304
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2497.712128,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6545917870000153,
0.5378263789999664
],
"count": 2,
"total": 1.1924181659999817,
"mean": 0.5962090829999909,
"p50": 0.5962090829999909,
"p90": 0.6429152462000104,
"p95": 0.6487535166000129,
"p99": 0.6534241329200148,
"stdev": 0.05838270400002443,
"stdev_": 9.792320456819564
},
"throughput": {
"unit": "samples/s",
"value": 6.709055789410141
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2497.712128,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.5645145930000126,
0.5367058049999969,
0.5674595629999999
],
"count": 3,
"total": 1.6686799610000094,
"mean": 0.5562266536666698,
"p50": 0.5645145930000126,
"p90": 0.5668705690000024,
"p95": 0.5671650660000012,
"p99": 0.5674006636000002,
"stdev": 0.013855585193357122,
"stdev_": 2.490996269599904
},
"throughput": {
"unit": "samples/s",
"value": 10.786969593146507
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_fill-mask_google-bert/bert-base-uncased | {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2497.712128,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6545917870000153,
0.5378263789999664,
0.5645145930000126,
0.5367058049999969,
0.5674595629999999
],
"count": 5,
"total": 2.861098126999991,
"mean": 0.5722196253999983,
"p50": 0.5645145930000126,
"p90": 0.6197388974000091,
"p95": 0.6371653422000122,
"p99": 0.6511064980400146,
"stdev": 0.04315396983437998,
"stdev_": 7.541504680866913
},
"throughput": {
"unit": "samples/s",
"value": 17.475807462929478
},
"energy": {
"unit": "kWh",
"cpu": 0.00011792815879444354,
"ram": 0.000004929245466240147,
"gpu": 0,
"total": 0.0001228574042606837
},
"efficiency": {
"unit": "samples/kWh",
"value": 81395.1756524304
}
} | {
"memory": {
"unit": "MB",
"max_ram": 2497.712128,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6545917870000153,
0.5378263789999664
],
"count": 2,
"total": 1.1924181659999817,
"mean": 0.5962090829999909,
"p50": 0.5962090829999909,
"p90": 0.6429152462000104,
"p95": 0.6487535166000129,
"p99": 0.6534241329200148,
"stdev": 0.05838270400002443,
"stdev_": 9.792320456819564
},
"throughput": {
"unit": "samples/s",
"value": 6.709055789410141
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2497.712128,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.5645145930000126,
0.5367058049999969,
0.5674595629999999
],
"count": 3,
"total": 1.6686799610000094,
"mean": 0.5562266536666698,
"p50": 0.5645145930000126,
"p90": 0.5668705690000024,
"p95": 0.5671650660000012,
"p99": 0.5674006636000002,
"stdev": 0.013855585193357122,
"stdev_": 2.490996269599904
},
"throughput": {
"unit": "samples/s",
"value": 10.786969593146507
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"model": "google-bert/bert-base-uncased",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.738775005999969,
"mean": 0.5477550011999938,
"stdev": 0.03693447784258994,
"p50": 0.5307143729999666,
"p90": 0.5856752317999963,
"p95": 0.6036043034000045,
"p99": 0.6179475606800111,
"values": [
0.6215333750000127,
0.5307143729999666,
0.5270229210000252,
0.5276163199999928,
0.5318880169999716
]
},
"throughput": {
"unit": "samples/s",
"value": 18.25633719106628
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.1522477479999793,
"mean": 0.5761238739999897,
"stdev": 0.045409501000023056,
"p50": 0.5761238739999897,
"p90": 0.612451474800008,
"p95": 0.6169924249000104,
"p99": 0.6206251849800123,
"values": [
0.6215333750000127,
0.5307143729999666
]
},
"throughput": {
"unit": "samples/s",
"value": 6.942951300088064
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.5865272579999896,
"mean": 0.5288424193333299,
"stdev": 0.0021671455040491463,
"p50": 0.5276163199999928,
"p90": 0.5310336775999758,
"p95": 0.5314608472999737,
"p99": 0.531802583059972,
"values": [
0.5270229210000252,
0.5276163199999928,
0.5318880169999716
]
},
"throughput": {
"unit": "samples/s",
"value": 11.345534663357242
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2574.761984,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.6610362229999964,
1.5238756629999557,
1.5143060410000544,
1.5217349879999915,
1.5177956429999995
],
"count": 5,
"total": 7.7387485579999975,
"mean": 1.5477497115999994,
"p50": 1.5217349879999915,
"p90": 1.60617199899998,
"p95": 1.6336041109999881,
"p99": 1.655549800599995,
"stdev": 0.05673851431719875,
"stdev_": 3.665871419129184
},
"throughput": {
"unit": "samples/s",
"value": 6.460992966144645
},
"energy": {
"unit": "kWh",
"cpu": 0.00031181906431666535,
"ram": 0.000013034458617008284,
"gpu": 0,
"total": 0.00032485352293367363
},
"efficiency": {
"unit": "samples/kWh",
"value": 30783.104673430713
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2574.761984,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.6610362229999964,
1.5238756629999557
],
"count": 2,
"total": 3.184911885999952,
"mean": 1.592455942999976,
"p50": 1.592455942999976,
"p90": 1.6473201669999924,
"p95": 1.6541781949999943,
"p99": 1.659664617399996,
"stdev": 0.06858028000002037,
"stdev_": 4.306573145805478
},
"throughput": {
"unit": "samples/s",
"value": 2.5118434312628644
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2574.761984,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.5143060410000544,
1.5217349879999915,
1.5177956429999995
],
"count": 3,
"total": 4.553836672000045,
"mean": 1.5179455573333485,
"p50": 1.5177956429999995,
"p90": 1.520947118999993,
"p95": 1.5213410534999923,
"p99": 1.5216562010999917,
"stdev": 0.0030347069174772386,
"stdev_": 0.19992198684704224
},
"throughput": {
"unit": "samples/s",
"value": 3.952710932887806
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_image-classification_google/vit-base-patch16-224 | {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2574.761984,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.6610362229999964,
1.5238756629999557,
1.5143060410000544,
1.5217349879999915,
1.5177956429999995
],
"count": 5,
"total": 7.7387485579999975,
"mean": 1.5477497115999994,
"p50": 1.5217349879999915,
"p90": 1.60617199899998,
"p95": 1.6336041109999881,
"p99": 1.655549800599995,
"stdev": 0.05673851431719875,
"stdev_": 3.665871419129184
},
"throughput": {
"unit": "samples/s",
"value": 6.460992966144645
},
"energy": {
"unit": "kWh",
"cpu": 0.00031181906431666535,
"ram": 0.000013034458617008284,
"gpu": 0,
"total": 0.00032485352293367363
},
"efficiency": {
"unit": "samples/kWh",
"value": 30783.104673430713
}
} | {
"memory": {
"unit": "MB",
"max_ram": 2574.761984,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.6610362229999964,
1.5238756629999557
],
"count": 2,
"total": 3.184911885999952,
"mean": 1.592455942999976,
"p50": 1.592455942999976,
"p90": 1.6473201669999924,
"p95": 1.6541781949999943,
"p99": 1.659664617399996,
"stdev": 0.06858028000002037,
"stdev_": 4.306573145805478
},
"throughput": {
"unit": "samples/s",
"value": 2.5118434312628644
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2574.761984,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.5143060410000544,
1.5217349879999915,
1.5177956429999995
],
"count": 3,
"total": 4.553836672000045,
"mean": 1.5179455573333485,
"p50": 1.5177956429999995,
"p90": 1.520947118999993,
"p95": 1.5213410534999923,
"p99": 1.5216562010999917,
"stdev": 0.0030347069174772386,
"stdev_": 0.19992198684704224
},
"throughput": {
"unit": "samples/s",
"value": 3.952710932887806
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"model": "google/vit-base-patch16-224",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.2970974209999895,
"mean": 1.459419484199998,
"stdev": 0.05210139006345095,
"p50": 1.4401334369999859,
"p90": 1.521250764199999,
"p95": 1.5379663595999886,
"p99": 1.5513388359199802,
"values": [
1.5546819549999782,
1.4241662819999874,
1.4401334369999859,
1.4711039780000306,
1.4070117690000075
]
},
"throughput": {
"unit": "samples/s",
"value": 6.852039532336137
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.9788482369999656,
"mean": 1.4894241184999828,
"stdev": 0.06525783649999539,
"p50": 1.4894241184999828,
"p90": 1.541630387699979,
"p95": 1.5481561713499787,
"p99": 1.5533767982699782,
"values": [
1.5546819549999782,
1.4241662819999874
]
},
"throughput": {
"unit": "samples/s",
"value": 2.6856017371522416
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.318249184000024,
"mean": 1.4394163946666747,
"stdev": 0.02617044676610726,
"p50": 1.4401334369999859,
"p90": 1.4649098698000216,
"p95": 1.4680069239000262,
"p99": 1.4704845671800297,
"values": [
1.4401334369999859,
1.4711039780000306,
1.4070117690000075
]
},
"throughput": {
"unit": "samples/s",
"value": 4.1683560241714614
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2879.823872,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.7899294580000173,
0.6948390950000203,
0.6938305699999887,
0.6825559779999821,
0.6868560240000079
],
"count": 5,
"total": 3.5480111250000164,
"mean": 0.7096022250000033,
"p50": 0.6938305699999887,
"p90": 0.7518933128000185,
"p95": 0.7709113854000179,
"p99": 0.7861258434800175,
"stdev": 0.040417938114141294,
"stdev_": 5.695858424646443
},
"throughput": {
"unit": "samples/s",
"value": 14.092402260999046
},
"energy": {
"unit": "kWh",
"cpu": 0.00014494537540000026,
"ram": 0.000006058641250594069,
"gpu": 0,
"total": 0.00015100401665059433
},
"efficiency": {
"unit": "samples/kWh",
"value": 66223.40399817862
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2879.823872,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.7899294580000173,
0.6948390950000203
],
"count": 2,
"total": 1.4847685530000376,
"mean": 0.7423842765000188,
"p50": 0.7423842765000188,
"p90": 0.7804204217000177,
"p95": 0.7851749398500175,
"p99": 0.7889785543700174,
"stdev": 0.047545181499998534,
"stdev_": 6.40438961398145
},
"throughput": {
"unit": "samples/s",
"value": 5.388045149417842
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2879.823872,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6938305699999887,
0.6825559779999821,
0.6868560240000079
],
"count": 3,
"total": 2.0632425719999787,
"mean": 0.687747523999993,
"p50": 0.6868560240000079,
"p90": 0.6924356607999925,
"p95": 0.6931331153999907,
"p99": 0.6936910790799891,
"stdev": 0.004645799922448451,
"stdev_": 0.6755095089878504
},
"throughput": {
"unit": "samples/s",
"value": 8.724131735296602
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_multiple-choice_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2879.823872,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.7899294580000173,
0.6948390950000203,
0.6938305699999887,
0.6825559779999821,
0.6868560240000079
],
"count": 5,
"total": 3.5480111250000164,
"mean": 0.7096022250000033,
"p50": 0.6938305699999887,
"p90": 0.7518933128000185,
"p95": 0.7709113854000179,
"p99": 0.7861258434800175,
"stdev": 0.040417938114141294,
"stdev_": 5.695858424646443
},
"throughput": {
"unit": "samples/s",
"value": 14.092402260999046
},
"energy": {
"unit": "kWh",
"cpu": 0.00014494537540000026,
"ram": 0.000006058641250594069,
"gpu": 0,
"total": 0.00015100401665059433
},
"efficiency": {
"unit": "samples/kWh",
"value": 66223.40399817862
}
} | {
"memory": {
"unit": "MB",
"max_ram": 2879.823872,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.7899294580000173,
0.6948390950000203
],
"count": 2,
"total": 1.4847685530000376,
"mean": 0.7423842765000188,
"p50": 0.7423842765000188,
"p90": 0.7804204217000177,
"p95": 0.7851749398500175,
"p99": 0.7889785543700174,
"stdev": 0.047545181499998534,
"stdev_": 6.40438961398145
},
"throughput": {
"unit": "samples/s",
"value": 5.388045149417842
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2879.823872,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6938305699999887,
0.6825559779999821,
0.6868560240000079
],
"count": 3,
"total": 2.0632425719999787,
"mean": 0.687747523999993,
"p50": 0.6868560240000079,
"p90": 0.6924356607999925,
"p95": 0.6931331153999907,
"p99": 0.6936910790799891,
"stdev": 0.004645799922448451,
"stdev_": 0.6755095089878504
},
"throughput": {
"unit": "samples/s",
"value": 8.724131735296602
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.581090587999995,
"mean": 0.716218117599999,
"stdev": 0.043372798377969854,
"p50": 0.697155070000008,
"p90": 0.7645524997999928,
"p95": 0.7826806403999967,
"p99": 0.7971831528799999,
"values": [
0.8008087810000006,
0.710168077999981,
0.6928677590000234,
0.697155070000008,
0.680090899999982
]
},
"throughput": {
"unit": "samples/s",
"value": 13.962227084549829
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.5109768589999817,
"mean": 0.7554884294999908,
"stdev": 0.04532035150000979,
"p50": 0.7554884294999908,
"p90": 0.7917447106999986,
"p95": 0.7962767458499996,
"p99": 0.7999023739700004,
"values": [
0.8008087810000006,
0.710168077999981
]
},
"throughput": {
"unit": "samples/s",
"value": 5.294588035778844
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.0701137290000133,
"mean": 0.6900379096666711,
"stdev": 0.007248103654729414,
"p50": 0.6928677590000234,
"p90": 0.696297607800011,
"p95": 0.6967263389000096,
"p99": 0.6970693237800083,
"values": [
0.6928677590000234,
0.697155070000008,
0.680090899999982
]
},
"throughput": {
"unit": "samples/s",
"value": 8.695174447587021
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2862.915584,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.703325331999963,
0.5838446669999939,
0.5709311329999878,
0.5724210480000238,
0.575583230999996
],
"count": 5,
"total": 3.0061054109999645,
"mean": 0.6012210821999929,
"p50": 0.575583230999996,
"p90": 0.6555330659999754,
"p95": 0.6794291989999692,
"p99": 0.6985461053999643,
"stdev": 0.05124732195740364,
"stdev_": 8.523873076752238
},
"throughput": {
"unit": "samples/s",
"value": 16.632816606177418
},
"energy": {
"unit": "kWh",
"cpu": 0.00012416476648333325,
"ram": 0.000005189943672332714,
"gpu": 0,
"total": 0.00012935471015566597
},
"efficiency": {
"unit": "samples/kWh",
"value": 77306.80999529094
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2862.915584,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.703325331999963,
0.5838446669999939
],
"count": 2,
"total": 1.287169998999957,
"mean": 0.6435849994999785,
"p50": 0.6435849994999785,
"p90": 0.6913772654999661,
"p95": 0.6973512987499646,
"p99": 0.7021305253499633,
"stdev": 0.05974033249998456,
"stdev_": 9.282430843851039
},
"throughput": {
"unit": "samples/s",
"value": 6.2151852561941725
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2862.915584,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.5709311329999878,
0.5724210480000238,
0.575583230999996
],
"count": 3,
"total": 1.7189354120000075,
"mean": 0.5729784706666692,
"p50": 0.5724210480000238,
"p90": 0.5749507944000015,
"p95": 0.5752670126999988,
"p99": 0.5755199873399965,
"stdev": 0.0019396810686961967,
"stdev_": 0.3385259949539375
},
"throughput": {
"unit": "samples/s",
"value": 10.47159763789887
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_text-classification_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2862.915584,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.703325331999963,
0.5838446669999939,
0.5709311329999878,
0.5724210480000238,
0.575583230999996
],
"count": 5,
"total": 3.0061054109999645,
"mean": 0.6012210821999929,
"p50": 0.575583230999996,
"p90": 0.6555330659999754,
"p95": 0.6794291989999692,
"p99": 0.6985461053999643,
"stdev": 0.05124732195740364,
"stdev_": 8.523873076752238
},
"throughput": {
"unit": "samples/s",
"value": 16.632816606177418
},
"energy": {
"unit": "kWh",
"cpu": 0.00012416476648333325,
"ram": 0.000005189943672332714,
"gpu": 0,
"total": 0.00012935471015566597
},
"efficiency": {
"unit": "samples/kWh",
"value": 77306.80999529094
}
} | {
"memory": {
"unit": "MB",
"max_ram": 2862.915584,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.703325331999963,
0.5838446669999939
],
"count": 2,
"total": 1.287169998999957,
"mean": 0.6435849994999785,
"p50": 0.6435849994999785,
"p90": 0.6913772654999661,
"p95": 0.6973512987499646,
"p99": 0.7021305253499633,
"stdev": 0.05974033249998456,
"stdev_": 9.282430843851039
},
"throughput": {
"unit": "samples/s",
"value": 6.2151852561941725
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2862.915584,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.5709311329999878,
0.5724210480000238,
0.575583230999996
],
"count": 3,
"total": 1.7189354120000075,
"mean": 0.5729784706666692,
"p50": 0.5724210480000238,
"p90": 0.5749507944000015,
"p95": 0.5752670126999988,
"p99": 0.5755199873399965,
"stdev": 0.0019396810686961967,
"stdev_": 0.3385259949539375
},
"throughput": {
"unit": "samples/s",
"value": 10.47159763789887
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.882509665999976,
"mean": 0.5765019331999952,
"stdev": 0.04978939696949581,
"p50": 0.5569985249999831,
"p90": 0.6300386333999881,
"p95": 0.6525100941999881,
"p99": 0.670487262839988,
"values": [
0.674981554999988,
0.5569985249999831,
0.5424031590000027,
0.5626242509999884,
0.5455021760000136
]
},
"throughput": {
"unit": "samples/s",
"value": 17.34599560576128
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.2319800799999712,
"mean": 0.6159900399999856,
"stdev": 0.05899151500000244,
"p50": 0.6159900399999856,
"p90": 0.6631832519999875,
"p95": 0.6690824034999878,
"p99": 0.673801724699988,
"values": [
0.674981554999988,
0.5569985249999831
]
},
"throughput": {
"unit": "samples/s",
"value": 6.493611487614465
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.6505295860000047,
"mean": 0.5501765286666682,
"stdev": 0.00889233078021606,
"p50": 0.5455021760000136,
"p90": 0.5591998359999935,
"p95": 0.5609120434999909,
"p99": 0.5622818094999888,
"values": [
0.5424031590000027,
0.5626242509999884,
0.5455021760000136
]
},
"throughput": {
"unit": "samples/s",
"value": 10.905590637500968
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2868.744192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6962681229999816,
0.6368247039999915,
0.6177456400000096,
0.6186691440000232,
0.6274819639999691
],
"count": 5,
"total": 3.196989574999975,
"mean": 0.639397914999995,
"p50": 0.6274819639999691,
"p90": 0.6724907553999856,
"p95": 0.6843794391999836,
"p99": 0.693890386239982,
"stdev": 0.029262039644876316,
"stdev_": 4.576499071767624
},
"throughput": {
"unit": "samples/s",
"value": 15.639713182361689
},
"energy": {
"unit": "kWh",
"cpu": 0.00013133819058888733,
"ram": 0.000005489823598354257,
"gpu": 0,
"total": 0.00013682801418724158
},
"efficiency": {
"unit": "samples/kWh",
"value": 73084.44881992918
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2868.744192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6962681229999816,
0.6368247039999915
],
"count": 2,
"total": 1.3330928269999731,
"mean": 0.6665464134999866,
"p50": 0.6665464134999866,
"p90": 0.6903237810999826,
"p95": 0.6932959520499822,
"p99": 0.6956736888099817,
"stdev": 0.029721709499995086,
"stdev_": 4.45906074926254
},
"throughput": {
"unit": "samples/s",
"value": 6.001082473756466
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2868.744192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6177456400000096,
0.6186691440000232,
0.6274819639999691
],
"count": 3,
"total": 1.863896748000002,
"mean": 0.6212989160000006,
"p50": 0.6186691440000232,
"p90": 0.6257193999999799,
"p95": 0.6266006819999745,
"p99": 0.6273057075999702,
"stdev": 0.004388300873682663,
"stdev_": 0.7063107243024159
},
"throughput": {
"unit": "samples/s",
"value": 9.657187298231168
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_text-generation_openai-community/gpt2 | {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2868.744192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6962681229999816,
0.6368247039999915,
0.6177456400000096,
0.6186691440000232,
0.6274819639999691
],
"count": 5,
"total": 3.196989574999975,
"mean": 0.639397914999995,
"p50": 0.6274819639999691,
"p90": 0.6724907553999856,
"p95": 0.6843794391999836,
"p99": 0.693890386239982,
"stdev": 0.029262039644876316,
"stdev_": 4.576499071767624
},
"throughput": {
"unit": "samples/s",
"value": 15.639713182361689
},
"energy": {
"unit": "kWh",
"cpu": 0.00013133819058888733,
"ram": 0.000005489823598354257,
"gpu": 0,
"total": 0.00013682801418724158
},
"efficiency": {
"unit": "samples/kWh",
"value": 73084.44881992918
}
} | {
"memory": {
"unit": "MB",
"max_ram": 2868.744192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6962681229999816,
0.6368247039999915
],
"count": 2,
"total": 1.3330928269999731,
"mean": 0.6665464134999866,
"p50": 0.6665464134999866,
"p90": 0.6903237810999826,
"p95": 0.6932959520499822,
"p99": 0.6956736888099817,
"stdev": 0.029721709499995086,
"stdev_": 4.45906074926254
},
"throughput": {
"unit": "samples/s",
"value": 6.001082473756466
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2868.744192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.6177456400000096,
0.6186691440000232,
0.6274819639999691
],
"count": 3,
"total": 1.863896748000002,
"mean": 0.6212989160000006,
"p50": 0.6186691440000232,
"p90": 0.6257193999999799,
"p95": 0.6266006819999745,
"p99": 0.6273057075999702,
"stdev": 0.004388300873682663,
"stdev_": 0.7063107243024159
},
"throughput": {
"unit": "samples/s",
"value": 9.657187298231168
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"model": "openai-community/gpt2",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.1791685380000274,
"mean": 0.6358337076000055,
"stdev": 0.07846941233493662,
"p50": 0.596941285000014,
"p90": 0.7161873142000047,
"p95": 0.7544328206000045,
"p99": 0.7850292257200044,
"values": [
0.7926783270000044,
0.6014507950000052,
0.596941285000014,
0.593667699000008,
0.5944304319999958
]
},
"throughput": {
"unit": "samples/s",
"value": 15.727382616668173
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.3941291220000096,
"mean": 0.6970645610000048,
"stdev": 0.0956137659999996,
"p50": 0.6970645610000048,
"p90": 0.7735555738000045,
"p95": 0.7831169504000044,
"p99": 0.7907660516800044,
"values": [
0.7926783270000044,
0.6014507950000052
]
},
"throughput": {
"unit": "samples/s",
"value": 5.738349392288174
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7850394160000178,
"mean": 0.5950131386666726,
"stdev": 0.0013985114990352936,
"p50": 0.5944304319999958,
"p90": 0.5964391144000103,
"p95": 0.5966901997000121,
"p99": 0.5968910679400136,
"values": [
0.596941285000014,
0.593667699000008,
0.5944304319999958
]
},
"throughput": {
"unit": "samples/s",
"value": 10.083810944822195
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 4375.740416,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.3242777090000004,
1.1163182639999718,
1.1666107809999744,
1.1331125829999564,
1.1239917860000332
],
"count": 5,
"total": 5.864311122999936,
"mean": 1.1728622245999873,
"p50": 1.1331125829999564,
"p90": 1.26121093779999,
"p95": 1.2927443233999951,
"p99": 1.3179710318799993,
"stdev": 0.07762899293407541,
"stdev_": 6.618764873303965
},
"throughput": {
"unit": "samples/s",
"value": 8.526150634112687
},
"energy": {
"unit": "kWh",
"cpu": 0.00023568561274999998,
"ram": 0.000009851840918840805,
"gpu": 0,
"total": 0.00024553745366884077
},
"efficiency": {
"unit": "samples/kWh",
"value": 40726.9842159686
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 4375.740416,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.3242777090000004,
1.1163182639999718
],
"count": 2,
"total": 2.440595972999972,
"mean": 1.220297986499986,
"p50": 1.220297986499986,
"p90": 1.3034817644999976,
"p95": 1.313879736749999,
"p99": 1.3221981145500001,
"stdev": 0.1039797225000143,
"stdev_": 8.520846846453065
},
"throughput": {
"unit": "samples/s",
"value": 3.2778878964413054
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 4375.740416,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.1666107809999744,
1.1331125829999564,
1.1239917860000332
],
"count": 3,
"total": 3.423715149999964,
"mean": 1.1412383833333213,
"p50": 1.1331125829999564,
"p90": 1.1599111413999708,
"p95": 1.1632609611999727,
"p99": 1.165940817039974,
"stdev": 0.01832332133274492,
"stdev_": 1.6055647619585216
},
"throughput": {
"unit": "samples/s",
"value": 5.257446724211326
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_token-classification_microsoft/deberta-v3-base | {
"name": "pytorch",
"version": "2.5.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.342208,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": "abe76ffe762e9c1733304501bfa77d8641c7629d",
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 4375.740416,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.3242777090000004,
1.1163182639999718,
1.1666107809999744,
1.1331125829999564,
1.1239917860000332
],
"count": 5,
"total": 5.864311122999936,
"mean": 1.1728622245999873,
"p50": 1.1331125829999564,
"p90": 1.26121093779999,
"p95": 1.2927443233999951,
"p99": 1.3179710318799993,
"stdev": 0.07762899293407541,
"stdev_": 6.618764873303965
},
"throughput": {
"unit": "samples/s",
"value": 8.526150634112687
},
"energy": {
"unit": "kWh",
"cpu": 0.00023568561274999998,
"ram": 0.000009851840918840805,
"gpu": 0,
"total": 0.00024553745366884077
},
"efficiency": {
"unit": "samples/kWh",
"value": 40726.9842159686
}
} | {
"memory": {
"unit": "MB",
"max_ram": 4375.740416,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.3242777090000004,
1.1163182639999718
],
"count": 2,
"total": 2.440595972999972,
"mean": 1.220297986499986,
"p50": 1.220297986499986,
"p90": 1.3034817644999976,
"p95": 1.313879736749999,
"p99": 1.3221981145500001,
"stdev": 0.1039797225000143,
"stdev_": 8.520846846453065
},
"throughput": {
"unit": "samples/s",
"value": 3.2778878964413054
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 4375.740416,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.1666107809999744,
1.1331125829999564,
1.1239917860000332
],
"count": 3,
"total": 3.423715149999964,
"mean": 1.1412383833333213,
"p50": 1.1331125829999564,
"p90": 1.1599111413999708,
"p95": 1.1632609611999727,
"p99": 1.165940817039974,
"stdev": 0.01832332133274492,
"stdev_": 1.6055647619585216
},
"throughput": {
"unit": "samples/s",
"value": 5.257446724211326
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"model": "microsoft/deberta-v3-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.583659903000068,
"mean": 1.1167319806000138,
"stdev": 0.07853258776030796,
"p50": 1.0731834320000075,
"p90": 1.2025418994000006,
"p95": 1.2374836681999908,
"p99": 1.265437083239983,
"values": [
1.2724254369999812,
1.0731834320000075,
1.0977165930000297,
1.0711078369999996,
1.0692266040000504
]
},
"throughput": {
"unit": "samples/s",
"value": 8.954700119385008
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.3456088689999888,
"mean": 1.1728044344999944,
"stdev": 0.09962100249998684,
"p50": 1.1728044344999944,
"p90": 1.252501236499984,
"p95": 1.2624633367499825,
"p99": 1.2704330169499816,
"values": [
1.2724254369999812,
1.0731834320000075
]
},
"throughput": {
"unit": "samples/s",
"value": 3.4106283045436583
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.2380510340000797,
"mean": 1.0793503446666932,
"stdev": 0.01300958794585394,
"p50": 1.0711078369999996,
"p90": 1.0923948418000236,
"p95": 1.0950557174000266,
"p99": 1.097184417880029,
"values": [
1.0977165930000297,
1.0711078369999996,
1.0692266040000504
]
},
"throughput": {
"unit": "samples/s",
"value": 5.558899415419021
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |