IlyasMoutawwakil HF staff commited on
Commit
945cdce
1 Parent(s): 4554dd8

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cpu_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -45,7 +45,6 @@
45
  "warmup_runs": 1,
46
  "input_shapes": {
47
  "batch_size": 1,
48
- "num_choices": 2,
49
  "sequence_length": 2
50
  },
51
  "new_tokens": null,
@@ -73,23 +72,23 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
- "cpu_ram_mb": 16766.783488,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-6.8.0-1015-azure-x86_64-with-glibc2.39",
80
  "processor": "x86_64",
81
  "python_version": "3.10.15",
82
  "optimum_benchmark_version": "0.5.0.dev0",
83
- "optimum_benchmark_commit": "08c9f59440cf4e5a5d6711ec19e8329ab2de652d",
84
- "transformers_version": "4.45.2",
85
  "transformers_commit": null,
86
- "accelerate_version": "1.0.1",
87
  "accelerate_commit": null,
88
- "diffusers_version": "0.30.3",
89
  "diffusers_commit": null,
90
  "optimum_version": null,
91
  "optimum_commit": null,
92
- "timm_version": "1.0.9",
93
  "timm_commit": null,
94
  "peft_version": null,
95
  "peft_commit": null
@@ -101,7 +100,7 @@
101
  "load": {
102
  "memory": {
103
  "unit": "MB",
104
- "max_ram": 947.167232,
105
  "max_global_vram": null,
106
  "max_process_vram": null,
107
  "max_reserved": null,
@@ -110,15 +109,15 @@
110
  "latency": {
111
  "unit": "s",
112
  "values": [
113
- 1.2332180920000155
114
  ],
115
  "count": 1,
116
- "total": 1.2332180920000155,
117
- "mean": 1.2332180920000155,
118
- "p50": 1.2332180920000155,
119
- "p90": 1.2332180920000155,
120
- "p95": 1.2332180920000155,
121
- "p99": 1.2332180920000155,
122
  "stdev": 0,
123
  "stdev_": 0
124
  },
@@ -129,7 +128,7 @@
129
  "forward": {
130
  "memory": {
131
  "unit": "MB",
132
- "max_ram": 955.424768,
133
  "max_global_vram": null,
134
  "max_process_vram": null,
135
  "max_reserved": null,
@@ -138,60 +137,59 @@
138
  "latency": {
139
  "unit": "s",
140
  "values": [
141
- 0.03658645500001967,
142
- 0.0361032659999978,
143
- 0.03610821500001293,
144
- 0.036316623999994135,
145
- 0.035721485999999913,
146
- 0.03647771300001068,
147
- 0.03574233499995216,
148
- 0.0360932980000257,
149
- 0.03616041300000461,
150
- 0.03680985200003306,
151
- 0.03691355400002294,
152
- 0.036534498999969856,
153
- 0.036020262000022285,
154
- 0.03553399699995907,
155
- 0.035940604000018084,
156
- 0.036153218999970704,
157
- 0.03606949400000303,
158
- 0.03632082100000389,
159
- 0.0355914040000016,
160
- 0.035376623999979984,
161
- 0.03604786400001103,
162
- 0.036388205999969614,
163
- 0.035578268999984175,
164
- 0.032220391999999265,
165
- 0.03168374300003052,
166
- 0.03233158899996624,
167
- 0.03141736700001729,
168
- 0.0313566550000246,
169
- 0.03159928499997022
170
  ],
171
- "count": 29,
172
- "total": 1.021197504999975,
173
- "mean": 0.03521370706896466,
174
- "p50": 0.03604786400001103,
175
- "p90": 0.03654489019997982,
176
- "p95": 0.0367204932000277,
177
- "p99": 0.03688451744002577,
178
- "stdev": 0.0018028598173541696,
179
- "stdev_": 5.11976717993178
180
  },
181
  "throughput": {
182
  "unit": "samples/s",
183
- "value": 28.39803256276141
184
  },
185
  "energy": {
186
  "unit": "kWh",
187
- "cpu": 1.214893873737353e-06,
188
- "ram": 5.0801180209474435e-08,
189
  "gpu": 0.0,
190
- "total": 1.2656950539468274e-06
191
  },
192
  "efficiency": {
193
  "unit": "samples/kWh",
194
- "value": 790079.7248766136
195
  }
196
  }
197
  }
 
3
  "name": "cpu_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.5.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
45
  "warmup_runs": 1,
46
  "input_shapes": {
47
  "batch_size": 1,
 
48
  "sequence_length": 2
49
  },
50
  "new_tokens": null,
 
72
  "environment": {
73
  "cpu": " AMD EPYC 7763 64-Core Processor",
74
  "cpu_count": 4,
75
+ "cpu_ram_mb": 16757.342208,
76
  "system": "Linux",
77
  "machine": "x86_64",
78
+ "platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35",
79
  "processor": "x86_64",
80
  "python_version": "3.10.15",
81
  "optimum_benchmark_version": "0.5.0.dev0",
82
+ "optimum_benchmark_commit": "6807ba28334c8c98abf72a03d78f86133328d180",
83
+ "transformers_version": "4.46.3",
84
  "transformers_commit": null,
85
+ "accelerate_version": "1.1.1",
86
  "accelerate_commit": null,
87
+ "diffusers_version": "0.31.0",
88
  "diffusers_commit": null,
89
  "optimum_version": null,
90
  "optimum_commit": null,
91
+ "timm_version": "1.0.11",
92
  "timm_commit": null,
93
  "peft_version": null,
94
  "peft_commit": null
 
100
  "load": {
101
  "memory": {
102
  "unit": "MB",
103
+ "max_ram": 967.348224,
104
  "max_global_vram": null,
105
  "max_process_vram": null,
106
  "max_reserved": null,
 
109
  "latency": {
110
  "unit": "s",
111
  "values": [
112
+ 1.259143196000025
113
  ],
114
  "count": 1,
115
+ "total": 1.259143196000025,
116
+ "mean": 1.259143196000025,
117
+ "p50": 1.259143196000025,
118
+ "p90": 1.259143196000025,
119
+ "p95": 1.259143196000025,
120
+ "p99": 1.259143196000025,
121
  "stdev": 0,
122
  "stdev_": 0
123
  },
 
128
  "forward": {
129
  "memory": {
130
  "unit": "MB",
131
+ "max_ram": 975.60576,
132
  "max_global_vram": null,
133
  "max_process_vram": null,
134
  "max_reserved": null,
 
137
  "latency": {
138
  "unit": "s",
139
  "values": [
140
+ 0.04134855299997753,
141
+ 0.03980547300000126,
142
+ 0.036524000999975215,
143
+ 0.0363250100000414,
144
+ 0.03849340599998641,
145
+ 0.033670236999967074,
146
+ 0.03662184399996704,
147
+ 0.03400348899998562,
148
+ 0.03581257400003324,
149
+ 0.0367397139999639,
150
+ 0.03581842500000221,
151
+ 0.03429234700001871,
152
+ 0.03452939999999671,
153
+ 0.035058477999996285,
154
+ 0.03288451100002021,
155
+ 0.033002731999999924,
156
+ 0.03274014200002284,
157
+ 0.032771269999955166,
158
+ 0.033267905999991854,
159
+ 0.03382941400002437,
160
+ 0.034541262000004735,
161
+ 0.034603747999994994,
162
+ 0.040686757000003126,
163
+ 0.03903687999996919,
164
+ 0.038817409999978736,
165
+ 0.040607710000017505,
166
+ 0.04100459099998943,
167
+ 0.040230747000009615
 
168
  ],
169
+ "count": 28,
170
+ "total": 1.0170680309998943,
171
+ "mean": 0.036323858249996226,
172
+ "p50": 0.03581549950001772,
173
+ "p90": 0.04063142410001319,
174
+ "p95": 0.040893349099994224,
175
+ "p99": 0.04125568325998074,
176
+ "stdev": 0.0028184781986868956,
177
+ "stdev_": 7.759302933319834
178
  },
179
  "throughput": {
180
  "unit": "samples/s",
181
+ "value": 27.530115141337
182
  },
183
  "energy": {
184
  "unit": "kWh",
185
+ "cpu": 1.2973221294444778e-06,
186
+ "ram": 5.4216226129131595e-08,
187
  "gpu": 0.0,
188
+ "total": 1.3515383555736092e-06
189
  },
190
  "efficiency": {
191
  "unit": "samples/kWh",
192
+ "value": 739897.6106568488
193
  }
194
  }
195
  }