IlyasMoutawwakil HF staff commited on
Commit
a07e01a
1 Parent(s): 7f479d1

Upload cpu_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "load": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 946.405376,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
@@ -11,15 +11,15 @@
11
  "latency": {
12
  "unit": "s",
13
  "values": [
14
- 1.222943784999984
15
  ],
16
  "count": 1,
17
- "total": 1.222943784999984,
18
- "mean": 1.222943784999984,
19
- "p50": 1.222943784999984,
20
- "p90": 1.222943784999984,
21
- "p95": 1.222943784999984,
22
- "p99": 1.222943784999984,
23
  "stdev": 0,
24
  "stdev_": 0
25
  },
@@ -30,7 +30,7 @@
30
  "forward": {
31
  "memory": {
32
  "unit": "MB",
33
- "max_ram": 957.939712,
34
  "max_global_vram": null,
35
  "max_process_vram": null,
36
  "max_reserved": null,
@@ -39,51 +39,53 @@
39
  "latency": {
40
  "unit": "s",
41
  "values": [
42
- 0.048797159999992346,
43
- 0.048929998000005526,
44
- 0.04841551099997332,
45
- 0.04970983900000192,
46
- 0.04918165599997337,
47
- 0.04859177900004852,
48
- 0.04805119399998148,
49
- 0.04821739199996955,
50
- 0.05964398499997969,
51
- 0.05548912799997652,
52
- 0.054109772000003886,
53
- 0.05165780200002246,
54
- 0.050210258999982216,
55
- 0.05408685999998397,
56
- 0.04883214599999519,
57
- 0.048457989999974416,
58
- 0.04842392699998754,
59
- 0.047398719000000256,
60
- 0.049894021000000066,
61
- 0.04814494999999397
 
 
62
  ],
63
- "count": 20,
64
- "total": 1.0062440879998462,
65
- "mean": 0.05031220439999231,
66
- "p50": 0.04888107200000036,
67
- "p90": 0.05424770760000115,
68
- "p95": 0.05569687084997668,
69
- "p99": 0.058854562169979086,
70
- "stdev": 0.00307272137801659,
71
- "stdev_": 6.107308186275892
72
  },
73
  "throughput": {
74
  "unit": "samples/s",
75
- "value": 19.875893173946338
76
  },
77
  "energy": {
78
  "unit": "kWh",
79
- "cpu": 1.7864058616161237e-06,
80
- "ram": 7.469644648415405e-08,
81
  "gpu": 0.0,
82
- "total": 1.8611023081002777e-06
83
  },
84
  "efficiency": {
85
  "unit": "samples/kWh",
86
- "value": 537315.9743274678
87
  }
88
  }
89
  }
 
2
  "load": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 946.065408,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
 
11
  "latency": {
12
  "unit": "s",
13
  "values": [
14
+ 1.2297822710000332
15
  ],
16
  "count": 1,
17
+ "total": 1.2297822710000332,
18
+ "mean": 1.2297822710000332,
19
+ "p50": 1.2297822710000332,
20
+ "p90": 1.2297822710000332,
21
+ "p95": 1.2297822710000332,
22
+ "p99": 1.2297822710000332,
23
  "stdev": 0,
24
  "stdev_": 0
25
  },
 
30
  "forward": {
31
  "memory": {
32
  "unit": "MB",
33
+ "max_ram": 957.816832,
34
  "max_global_vram": null,
35
  "max_process_vram": null,
36
  "max_reserved": null,
 
39
  "latency": {
40
  "unit": "s",
41
  "values": [
42
+ 0.047197117999985494,
43
+ 0.04701091900000165,
44
+ 0.04414449500001183,
45
+ 0.04807227200001307,
46
+ 0.04629180799997812,
47
+ 0.04586062799995716,
48
+ 0.04595358300002772,
49
+ 0.045559904000015194,
50
+ 0.04723359700000174,
51
+ 0.04875624699997161,
52
+ 0.0464784879999911,
53
+ 0.047936456999991606,
54
+ 0.04665203400003293,
55
+ 0.049990254000022105,
56
+ 0.04643939599998248,
57
+ 0.04771427999997968,
58
+ 0.04619442500001014,
59
+ 0.04729299899997841,
60
+ 0.04600997799997231,
61
+ 0.046056226000018796,
62
+ 0.045464944999991985,
63
+ 0.04633350699998573
64
  ],
65
+ "count": 22,
66
+ "total": 1.0286435599999209,
67
+ "mean": 0.046756525454541856,
68
+ "p50": 0.04645894199998679,
69
+ "p90": 0.04805869050001092,
70
+ "p95": 0.048722048249973685,
71
+ "p99": 0.0497311125300115,
72
+ "stdev": 0.0012170828072060273,
73
+ "stdev_": 2.603022349018027
74
  },
75
  "throughput": {
76
  "unit": "samples/s",
77
+ "value": 21.387389038824775
78
  },
79
  "energy": {
80
  "unit": "kWh",
81
+ "cpu": 1.5597827668889425e-06,
82
+ "ram": 6.52226472597115e-08,
83
  "gpu": 0.0,
84
+ "total": 1.6250054141486537e-06
85
  },
86
  "efficiency": {
87
  "unit": "samples/kWh",
88
+ "value": 615382.5650629624
89
  }
90
  }
91
  }