IlyasMoutawwakil HF staff commited on
Commit
99beeb8
·
verified ·
1 Parent(s): 17c2837

Upload cpu_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.15",
82
  "optimum_benchmark_version": "0.5.0.dev0",
83
- "optimum_benchmark_commit": "1c20082e96a83cbb10a6021fdbbcd050ed6631b4",
84
  "transformers_version": "4.46.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "1.1.1",
@@ -101,7 +101,7 @@
101
  "load": {
102
  "memory": {
103
  "unit": "MB",
104
- "max_ram": 966.283264,
105
  "max_global_vram": null,
106
  "max_process_vram": null,
107
  "max_reserved": null,
@@ -110,15 +110,15 @@
110
  "latency": {
111
  "unit": "s",
112
  "values": [
113
- 1.2190277400000014
114
  ],
115
  "count": 1,
116
- "total": 1.2190277400000014,
117
- "mean": 1.2190277400000014,
118
- "p50": 1.2190277400000014,
119
- "p90": 1.2190277400000014,
120
- "p95": 1.2190277400000014,
121
- "p99": 1.2190277400000014,
122
  "stdev": 0,
123
  "stdev_": 0
124
  },
@@ -129,7 +129,7 @@
129
  "forward": {
130
  "memory": {
131
  "unit": "MB",
132
- "max_ram": 982.781952,
133
  "max_global_vram": null,
134
  "max_process_vram": null,
135
  "max_reserved": null,
@@ -138,38 +138,40 @@
138
  "latency": {
139
  "unit": "s",
140
  "values": [
141
- 0.18024533800002018,
142
- 0.18033672899997555,
143
- 0.17786964300000818,
144
- 0.16922332500001858,
145
- 0.09826937500002941,
146
- 0.09685704700001452,
147
- 0.10003447999997661
 
 
148
  ],
149
- "count": 7,
150
- "total": 1.002835937000043,
151
- "mean": 0.14326227671429187,
152
- "p50": 0.16922332500001858,
153
- "p90": 0.18028189440000233,
154
- "p95": 0.18030931169998893,
155
- "p99": 0.18033124553997823,
156
- "stdev": 0.03902445501029402,
157
- "stdev_": 27.23986795778804
158
  },
159
  "throughput": {
160
  "unit": "samples/s",
161
- "value": 13.960409159129883
162
  },
163
  "energy": {
164
  "unit": "kWh",
165
- "cpu": 3.8407534666667616e-06,
166
- "ram": 1.6050451571512105e-07,
167
  "gpu": 0.0,
168
- "total": 4.001257982381883e-06
169
  },
170
  "efficiency": {
171
  "unit": "samples/kWh",
172
- "value": 499842.80164045637
173
  }
174
  }
175
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.15",
82
  "optimum_benchmark_version": "0.5.0.dev0",
83
+ "optimum_benchmark_commit": "a0fa44f89e357c39216a81807cac957ae2b66da0",
84
  "transformers_version": "4.46.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "1.1.1",
 
101
  "load": {
102
  "memory": {
103
  "unit": "MB",
104
+ "max_ram": 968.72448,
105
  "max_global_vram": null,
106
  "max_process_vram": null,
107
  "max_reserved": null,
 
110
  "latency": {
111
  "unit": "s",
112
  "values": [
113
+ 1.216209875000004
114
  ],
115
  "count": 1,
116
+ "total": 1.216209875000004,
117
+ "mean": 1.216209875000004,
118
+ "p50": 1.216209875000004,
119
+ "p90": 1.216209875000004,
120
+ "p95": 1.216209875000004,
121
+ "p99": 1.216209875000004,
122
  "stdev": 0,
123
  "stdev_": 0
124
  },
 
129
  "forward": {
130
  "memory": {
131
  "unit": "MB",
132
+ "max_ram": 983.928832,
133
  "max_global_vram": null,
134
  "max_process_vram": null,
135
  "max_reserved": null,
 
138
  "latency": {
139
  "unit": "s",
140
  "values": [
141
+ 0.1759663729999943,
142
+ 0.17847407200002863,
143
+ 0.1264521909999985,
144
+ 0.10319367200003171,
145
+ 0.10670872400004328,
146
+ 0.09895815299995547,
147
+ 0.09160758799998803,
148
+ 0.09190421300002072,
149
+ 0.09167815099999643
150
  ],
151
+ "count": 9,
152
+ "total": 1.064943137000057,
153
+ "mean": 0.11832701522222856,
154
+ "p50": 0.10319367200003171,
155
+ "p90": 0.17646791280000115,
156
+ "p95": 0.1774709924000149,
157
+ "p99": 0.17827345608002587,
158
+ "stdev": 0.033111582917912316,
159
+ "stdev_": 27.98311345530507
160
  },
161
  "throughput": {
162
  "unit": "samples/s",
163
+ "value": 16.902310907140045
164
  },
165
  "energy": {
166
  "unit": "kWh",
167
+ "cpu": 3.665867004545505e-06,
168
+ "ram": 1.532027954026195e-07,
169
  "gpu": 0.0,
170
+ "total": 3.8190697999481244e-06
171
  },
172
  "efficiency": {
173
  "unit": "samples/kWh",
174
+ "value": 523687.7315065482
175
  }
176
  }
177
  }