IlyasMoutawwakil HF staff commited on
Commit
6b20fe3
1 Parent(s): ff1efdb

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.0",
83
- "optimum_benchmark_commit": "b3690661eecec40c5905418f03e56991fd2bca89",
84
  "transformers_version": "4.40.2",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.30.1",
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 939.52,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,54 +107,64 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 23,
111
- "total": 1.0057511740000677,
112
- "mean": 0.04372831191304642,
113
- "stdev": 0.0023125447387848854,
114
- "p50": 0.04356103399999256,
115
- "p90": 0.04687751440000056,
116
- "p95": 0.047137526699992806,
117
- "p99": 0.047848402879983494,
118
  "values": [
119
- 0.04714847299999292,
120
- 0.045444992000000184,
121
- 0.042257148999965466,
122
- 0.04413896300002307,
123
- 0.04314970600000834,
124
- 0.04155089100004261,
125
- 0.041811617999997,
126
- 0.04024419999996098,
127
- 0.04144779800003562,
128
- 0.0421822499999962,
129
- 0.039129048000006605,
130
- 0.04177494900000056,
131
- 0.04356103399999256,
132
- 0.04267057100003058,
133
- 0.043619594000006146,
134
- 0.0422186780000402,
135
- 0.04703900999999178,
136
- 0.045515042999966226,
137
- 0.04602180999995653,
138
- 0.04804581899998084,
139
- 0.04623153200003571,
140
- 0.04513527400001749,
141
- 0.04541277200002014
 
 
 
 
 
 
 
 
 
 
142
  ]
143
  },
144
  "throughput": {
145
  "unit": "samples/s",
146
- "value": 22.868479395877348
147
  },
148
  "energy": {
149
  "unit": "kWh",
150
- "cpu": 1.5374729266533486e-06,
151
- "ram": 6.425178728138602e-08,
152
  "gpu": 0.0,
153
- "total": 1.6017247139347345e-06
154
  },
155
  "efficiency": {
156
  "unit": "samples/kWh",
157
- "value": 624327.009067269
158
  }
159
  }
160
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.0",
83
+ "optimum_benchmark_commit": "20967d442ecbde73b309c993163c266eac5ccef4",
84
  "transformers_version": "4.40.2",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.30.1",
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 939.29472,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 33,
111
+ "total": 1.012392686999931,
112
+ "mean": 0.03067856627272518,
113
+ "stdev": 0.00026836118750447755,
114
+ "p50": 0.030664161000004242,
115
+ "p90": 0.030970338600013747,
116
+ "p95": 0.03115793939998639,
117
+ "p99": 0.031353485119984724,
118
  "values": [
119
+ 0.03090960899999118,
120
+ 0.030685661000006803,
121
+ 0.030865768000012395,
122
+ 0.03092520899997453,
123
+ 0.030469896999989032,
124
+ 0.030420794999997725,
125
+ 0.030819160000021384,
126
+ 0.030832895999992616,
127
+ 0.030528617000015856,
128
+ 0.03062238199999001,
129
+ 0.03128507100001343,
130
+ 0.03138567999997122,
131
+ 0.03028177499999174,
132
+ 0.030441924999990988,
133
+ 0.03063069799998175,
134
+ 0.03098162100002355,
135
+ 0.030727678999994623,
136
+ 0.030743228999995154,
137
+ 0.031073184999968362,
138
+ 0.03060062200000857,
139
+ 0.030664161000004242,
140
+ 0.030665442999975312,
141
+ 0.030313345000024583,
142
+ 0.030358819000014137,
143
+ 0.030813851000004888,
144
+ 0.030789545000004637,
145
+ 0.030437928000026204,
146
+ 0.03046308500000805,
147
+ 0.030464998000013566,
148
+ 0.030275873999983105,
149
+ 0.030478983999955744,
150
+ 0.030920839999964755,
151
+ 0.030514335000020765
152
  ]
153
  },
154
  "throughput": {
155
  "unit": "samples/s",
156
+ "value": 32.59604738729434
157
  },
158
  "energy": {
159
  "unit": "kWh",
160
+ "cpu": 1.1620449084861606e-06,
161
+ "ram": 4.856383479275682e-08,
162
  "gpu": 0.0,
163
+ "total": 1.2106087432789175e-06
164
  },
165
  "efficiency": {
166
  "unit": "samples/kWh",
167
+ "value": 826030.7102123791
168
  }
169
  }
170
  }