IlyasMoutawwakil HF staff commited on
Commit
e805516
1 Parent(s): ba486a1

Upload cpu_inference_transformers_text-generation_openai-community/gpt2/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-generation_openai-community/gpt2/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cpu_inference_transformers_text-generation_openai-community/gpt2",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-generation",
9
  "library": "transformers",
@@ -73,23 +73,23 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
- "cpu_ram_mb": 16757.354496,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
- "optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
84
- "transformers_version": "4.41.1",
85
  "transformers_commit": null,
86
- "accelerate_version": "0.30.1",
87
  "accelerate_commit": null,
88
- "diffusers_version": "0.27.2",
89
  "diffusers_commit": null,
90
  "optimum_version": null,
91
  "optimum_commit": null,
92
- "timm_version": "1.0.3",
93
  "timm_commit": null,
94
  "peft_version": null,
95
  "peft_commit": null
@@ -99,7 +99,7 @@
99
  "prefill": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 955.838464,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -108,49 +108,49 @@
108
  "latency": {
109
  "unit": "s",
110
  "count": 13,
111
- "total": 0.6638927820000333,
112
- "mean": 0.0510686755384641,
113
- "stdev": 0.002045444067699494,
114
- "p50": 0.05120662400000242,
115
- "p90": 0.05376475900000628,
116
- "p95": 0.054256579000008284,
117
- "p99": 0.05438413660001174,
118
  "values": [
119
- 0.05207829200000447,
120
- 0.05441602600001261,
121
- 0.05415028100000541,
122
- 0.0515553809999858,
123
- 0.048570006000005606,
124
- 0.05222267100000977,
125
- 0.05146354100000394,
126
- 0.05041946099998995,
127
- 0.048943198000017674,
128
- 0.05120662400000242,
129
- 0.051157893000009835,
130
- 0.05108318399999234,
131
- 0.0466262239999935
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "tokens/s",
136
- "value": 39.16295026084302
137
  },
138
  "energy": {
139
  "unit": "kWh",
140
- "cpu": 1.766204891573404e-06,
141
- "ram": 7.381309496177772e-08,
142
  "gpu": 0.0,
143
- "total": 1.8400179865351817e-06
144
  },
145
  "efficiency": {
146
  "unit": "tokens/kWh",
147
- "value": 1086945.8965268433
148
  }
149
  },
150
  "decode": {
151
  "memory": {
152
  "unit": "MB",
153
- "max_ram": 955.838464,
154
  "max_global_vram": null,
155
  "max_process_vram": null,
156
  "max_reserved": null,
@@ -159,43 +159,43 @@
159
  "latency": {
160
  "unit": "s",
161
  "count": 13,
162
- "total": 0.40678622899997663,
163
- "mean": 0.031291248384613585,
164
- "stdev": 0.001711399813907373,
165
- "p50": 0.03150826099999904,
166
- "p90": 0.03236894979999079,
167
- "p95": 0.03312458519999381,
168
- "p99": 0.033997589040000095,
169
  "values": [
170
- 0.032397081999988586,
171
- 0.03421584000000166,
172
- 0.03225642099999959,
173
- 0.03081424100000163,
174
- 0.0313733799999909,
175
- 0.032005884999989576,
176
- 0.032052802000009706,
177
- 0.03150826099999904,
178
- 0.03156042799997749,
179
- 0.031312657000000854,
180
- 0.031474787999997034,
181
- 0.026685342000007495,
182
- 0.029129102000013063
183
  ]
184
  },
185
  "throughput": {
186
  "unit": "tokens/s",
187
- "value": 31.95781733309548
188
  },
189
  "energy": {
190
  "unit": "kWh",
191
- "cpu": 1.089805969293567e-06,
192
- "ram": 4.554324453157282e-08,
193
  "gpu": 0.0,
194
- "total": 1.1353492138251402e-06
195
  },
196
  "efficiency": {
197
  "unit": "tokens/kWh",
198
- "value": 880786.2707112546
199
  }
200
  },
201
  "per_token": {
@@ -203,31 +203,31 @@
203
  "latency": {
204
  "unit": "s",
205
  "count": 12,
206
- "total": 0.986620572000021,
207
- "mean": 0.08221838100000174,
208
- "stdev": 0.0033235099788530573,
209
- "p50": 0.08248187099998461,
210
- "p90": 0.08624422410000818,
211
- "p95": 0.08742922315001067,
212
- "p99": 0.08836903983001036,
213
  "values": [
214
- 0.08860399400001029,
215
- 0.08646804700001098,
216
- 0.08247610999998756,
217
- 0.07991193800000929,
218
- 0.08422981799998297,
219
- 0.08354884300001686,
220
- 0.08192640999999412,
221
- 0.08056282600000486,
222
- 0.08248763199998166,
223
- 0.08264841100000808,
224
- 0.07797228399999767,
225
- 0.07578425900001662
226
  ]
227
  },
228
  "throughput": {
229
  "unit": "tokens/s",
230
- "value": 12.16273037534002
231
  },
232
  "energy": null,
233
  "efficiency": null
 
3
  "name": "cpu_inference_transformers_text-generation_openai-community/gpt2",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-generation",
9
  "library": "transformers",
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
+ "cpu_ram_mb": 16757.338112,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-6.5.0-1022-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
+ "optimum_benchmark_commit": "d920fe9626db1e7915f6d3574b5b54b0159cd100",
84
+ "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
+ "accelerate_version": "0.31.0",
87
  "accelerate_commit": null,
88
+ "diffusers_version": "0.29.2",
89
  "diffusers_commit": null,
90
  "optimum_version": null,
91
  "optimum_commit": null,
92
+ "timm_version": "1.0.7",
93
  "timm_commit": null,
94
  "peft_version": null,
95
  "peft_commit": null
 
99
  "prefill": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 948.219904,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
108
  "latency": {
109
  "unit": "s",
110
  "count": 13,
111
+ "total": 0.6244059869999319,
112
+ "mean": 0.04803122976922553,
113
+ "stdev": 0.002659991941552128,
114
+ "p50": 0.04875405600000704,
115
+ "p90": 0.0509883479999985,
116
+ "p95": 0.0511371429999997,
117
+ "p99": 0.05124176619999503,
118
  "values": [
119
+ 0.050685386000026256,
120
+ 0.05104995700000359,
121
+ 0.05074191199997813,
122
+ 0.050600597999988395,
123
+ 0.049124685999998974,
124
+ 0.05126792199999386,
125
+ 0.04875405600000704,
126
+ 0.04520847499998126,
127
+ 0.046204921999986936,
128
+ 0.04494297899998401,
129
+ 0.047219673999990164,
130
+ 0.043867823999988786,
131
+ 0.04473759600000449
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "tokens/s",
136
+ "value": 41.63957511830013
137
  },
138
  "energy": {
139
  "unit": "kWh",
140
+ "cpu": 1.7411278065851922e-06,
141
+ "ram": 7.276344239316366e-08,
142
  "gpu": 0.0,
143
+ "total": 1.8138912489783558e-06
144
  },
145
  "efficiency": {
146
  "unit": "tokens/kWh",
147
+ "value": 1102601.9344469889
148
  }
149
  },
150
  "decode": {
151
  "memory": {
152
  "unit": "MB",
153
+ "max_ram": 948.219904,
154
  "max_global_vram": null,
155
  "max_process_vram": null,
156
  "max_reserved": null,
 
159
  "latency": {
160
  "unit": "s",
161
  "count": 13,
162
+ "total": 0.389256606000032,
163
+ "mean": 0.029942815846156306,
164
+ "stdev": 0.0020771328080415974,
165
+ "p50": 0.02917561000001001,
166
+ "p90": 0.032512111200014716,
167
+ "p95": 0.03256633820001298,
168
+ "p99": 0.03259535564000089,
169
  "values": [
170
+ 0.03260260999999787,
171
+ 0.03081702000000064,
172
+ 0.03254215700002305,
173
+ 0.03225978099999338,
174
+ 0.03239192799998136,
175
+ 0.031995488000006844,
176
+ 0.028002361999995173,
177
+ 0.02917561000001001,
178
+ 0.028242559999995365,
179
+ 0.027993085000019846,
180
+ 0.027426657999995996,
181
+ 0.02823492600001032,
182
+ 0.027572421000002123
183
  ]
184
  },
185
  "throughput": {
186
  "unit": "tokens/s",
187
+ "value": 33.39699262547373
188
  },
189
  "energy": {
190
  "unit": "kWh",
191
+ "cpu": 1.04916032385711e-06,
192
+ "ram": 4.3847111288990186e-08,
193
  "gpu": 0.0,
194
+ "total": 1.0930074351461e-06
195
  },
196
  "efficiency": {
197
  "unit": "tokens/kWh",
198
+ "value": 914906.8595917942
199
  }
200
  },
201
  "per_token": {
 
203
  "latency": {
204
  "unit": "s",
205
  "count": 12,
206
+ "total": 0.9307644740000001,
207
+ "mean": 0.07756370616666668,
208
+ "stdev": 0.0044240471243253265,
209
+ "p50": 0.07580525899999202,
210
+ "p90": 0.08325035319999471,
211
+ "p95": 0.08329273979999101,
212
+ "p99": 0.08329807435998844,
213
  "values": [
214
+ 0.08186743799998908,
215
+ 0.0832994079999878,
216
+ 0.0829179760000045,
217
+ 0.08149807000000919,
218
+ 0.08328728399999363,
219
+ 0.07693897900000479,
220
+ 0.07436066100001426,
221
+ 0.07451188199999592,
222
+ 0.07296351499999787,
223
+ 0.07467153899997925,
224
+ 0.07212025300000846,
225
+ 0.0723274690000153
226
  ]
227
  },
228
  "throughput": {
229
  "unit": "tokens/s",
230
+ "value": 12.892627872258045
231
  },
232
  "energy": null,
233
  "efficiency": null