{
    "config": {
        "name": "cpu_inference_transformers_text-generation_openai-community/gpt2",
        "backend": {
            "name": "pytorch",
            "version": "2.4.0+cpu",
            "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
            "task": "text-generation",
            "library": "transformers",
            "model_type": "gpt2",
            "model": "openai-community/gpt2",
            "processor": "openai-community/gpt2",
            "device": "cpu",
            "device_ids": null,
            "seed": 42,
            "inter_op_num_threads": null,
            "intra_op_num_threads": null,
            "model_kwargs": {},
            "processor_kwargs": {},
            "no_weights": true,
            "device_map": null,
            "torch_dtype": null,
            "eval_mode": true,
            "to_bettertransformer": false,
            "low_cpu_mem_usage": null,
            "attn_implementation": null,
            "cache_implementation": null,
            "autocast_enabled": false,
            "autocast_dtype": null,
            "torch_compile": false,
            "torch_compile_target": "forward",
            "torch_compile_config": {},
            "quantization_scheme": null,
            "quantization_config": {},
            "deepspeed_inference": false,
            "deepspeed_inference_config": {},
            "peft_type": null,
            "peft_config": {}
        },
        "scenario": {
            "name": "inference",
            "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
            "iterations": 1,
            "duration": 1,
            "warmup_runs": 1,
            "input_shapes": {
                "batch_size": 1,
                "num_choices": 2,
                "sequence_length": 2
            },
            "new_tokens": null,
            "memory": true,
            "latency": true,
            "energy": true,
            "forward_kwargs": {},
            "generate_kwargs": {
                "max_new_tokens": 2,
                "min_new_tokens": 2
            },
            "call_kwargs": {
                "num_inference_steps": 2
            }
        },
        "launcher": {
            "name": "process",
            "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
            "device_isolation": false,
            "device_isolation_action": "error",
            "numactl": false,
            "numactl_kwargs": {},
            "start_method": "spawn"
        },
        "environment": {
            "cpu": " AMD EPYC 7763 64-Core Processor",
            "cpu_count": 4,
            "cpu_ram_mb": 16757.338112,
            "system": "Linux",
            "machine": "x86_64",
            "platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
            "processor": "x86_64",
            "python_version": "3.10.14",
            "optimum_benchmark_version": "0.4.0",
            "optimum_benchmark_commit": "391e9edbc7753b839aaebb1598c0d4977058930d",
            "transformers_version": "4.44.0",
            "transformers_commit": null,
            "accelerate_version": "0.33.0",
            "accelerate_commit": null,
            "diffusers_version": "0.30.0",
            "diffusers_commit": null,
            "optimum_version": null,
            "optimum_commit": null,
            "timm_version": "1.0.8",
            "timm_commit": null,
            "peft_version": null,
            "peft_commit": null
        }
    },
    "report": {
        "load": {
            "memory": {
                "unit": "MB",
                "max_ram": 1120.755712,
                "max_global_vram": null,
                "max_process_vram": null,
                "max_reserved": null,
                "max_allocated": null
            },
            "latency": {
                "unit": "s",
                "count": 1,
                "total": 4.988093385999974,
                "mean": 4.988093385999974,
                "stdev": 0.0,
                "p50": 4.988093385999974,
                "p90": 4.988093385999974,
                "p95": 4.988093385999974,
                "p99": 4.988093385999974,
                "values": [
                    4.988093385999974
                ]
            },
            "throughput": null,
            "energy": {
                "unit": "kWh",
                "cpu": 6.896250234999933e-05,
                "ram": 2.882448459674828e-06,
                "gpu": 0,
                "total": 7.184495080967416e-05
            },
            "efficiency": null
        },
        "prefill": {
            "memory": {
                "unit": "MB",
                "max_ram": 974.487552,
                "max_global_vram": null,
                "max_process_vram": null,
                "max_reserved": null,
                "max_allocated": null
            },
            "latency": {
                "unit": "s",
                "count": 14,
                "total": 0.6467411189998984,
                "mean": 0.046195794214278454,
                "stdev": 0.0025435933646531838,
                "p50": 0.04748822799999175,
                "p90": 0.048816484099995706,
                "p95": 0.04899593979998542,
                "p99": 0.049061973559973354,
                "values": [
                    0.048143619999962084,
                    0.0482896940000046,
                    0.048951493999993545,
                    0.048501461000000745,
                    0.04844157999997378,
                    0.049078481999970336,
                    0.04740545800001428,
                    0.04757099799996922,
                    0.04520194500003072,
                    0.04278982900001438,
                    0.04419391999999789,
                    0.042846966000013254,
                    0.04272500700000137,
                    0.04260066499995219
                ]
            },
            "throughput": {
                "unit": "tokens/s",
                "value": 43.293984528613834
            },
            "energy": {
                "unit": "kWh",
                "cpu": 1.6749823027777158e-06,
                "ram": 7.000016760075193e-08,
                "gpu": 0.0,
                "total": 1.7449824703784676e-06
            },
            "efficiency": {
                "unit": "tokens/kWh",
                "value": 1146143.318887451
            }
        },
        "decode": {
            "memory": {
                "unit": "MB",
                "max_ram": 975.273984,
                "max_global_vram": null,
                "max_process_vram": null,
                "max_reserved": null,
                "max_allocated": null
            },
            "latency": {
                "unit": "s",
                "count": 14,
                "total": 0.39717634499993437,
                "mean": 0.02836973892856674,
                "stdev": 0.0018229862510433408,
                "p50": 0.02936533950000353,
                "p90": 0.030062515099990604,
                "p95": 0.030242611849990907,
                "p99": 0.030492581570002814,
                "values": [
                    0.030034870000008596,
                    0.030006625999988046,
                    0.03055507400000579,
                    0.029632674999959363,
                    0.0297962410000423,
                    0.030004201999986435,
                    0.030074362999982895,
                    0.029098004000047695,
                    0.02609267199994747,
                    0.027326764999997977,
                    0.02666740900002651,
                    0.026017310999975507,
                    0.02591605099996741,
                    0.02595408199999838
                ]
            },
            "throughput": {
                "unit": "tokens/s",
                "value": 35.24882631165337
            },
            "energy": {
                "unit": "kWh",
                "cpu": 1.0142550457408585e-06,
                "ram": 4.238829719193719e-08,
                "gpu": 0.0,
                "total": 1.0566433429327957e-06
            },
            "efficiency": {
                "unit": "tokens/kWh",
                "value": 946393.1294209665
            }
        },
        "per_token": {
            "memory": null,
            "latency": {
                "unit": "s",
                "count": 14,
                "total": 0.39077375800007985,
                "mean": 0.02791241128571999,
                "stdev": 0.0017354573375610474,
                "p50": 0.02884216400002515,
                "p90": 0.029528714799994304,
                "p95": 0.029700565799987543,
                "p99": 0.029935828359996322,
                "values": [
                    0.029497212000023865,
                    0.02946966000001794,
                    0.029994643999998516,
                    0.029089978000001793,
                    0.029289111000025514,
                    0.029452889000026516,
                    0.029542215999981636,
                    0.028594350000048507,
                    0.02573093399996651,
                    0.026969264999991083,
                    0.026318145000004733,
                    0.025642728999969222,
                    0.02558107300001211,
                    0.025601552000011907
                ]
            },
            "throughput": {
                "unit": "tokens/s",
                "value": 35.82635658967954
            },
            "energy": null,
            "efficiency": null
        }
    }
}