IlyasMoutawwakil HF staff commited on
Commit
bf7de71
·
verified ·
1 Parent(s): 70e3854

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -81,7 +81,7 @@
81
  "processor": "x86_64",
82
  "python_version": "3.10.14",
83
  "optimum_benchmark_version": "0.3.1",
84
- "optimum_benchmark_commit": "73dd36ea04ff1cc8259f4039b6f9d56dfb9eecd1",
85
  "transformers_version": "4.42.4",
86
  "transformers_commit": null,
87
  "accelerate_version": "0.32.1",
@@ -100,7 +100,7 @@
100
  "forward": {
101
  "memory": {
102
  "unit": "MB",
103
- "max_ram": 938.590208,
104
  "max_global_vram": null,
105
  "max_process_vram": null,
106
  "max_reserved": null,
@@ -108,60 +108,58 @@
108
  },
109
  "latency": {
110
  "unit": "s",
111
- "count": 29,
112
- "total": 1.004014856999902,
113
- "mean": 0.03462120196551386,
114
- "stdev": 0.0025904509593886087,
115
- "p50": 0.03579219799996736,
116
- "p90": 0.037216727200006974,
117
- "p95": 0.03738750659998687,
118
- "p99": 0.037535513079976685,
119
  "values": [
120
- 0.03724305599996569,
121
- 0.036284351000006154,
122
- 0.03552998699996124,
123
- 0.03755562099996723,
124
- 0.03642006500001571,
125
- 0.03693139199998541,
126
- 0.03519646300003387,
127
- 0.03721014500001729,
128
- 0.03626918200001228,
129
- 0.03663399499998832,
130
- 0.03622485899995809,
131
- 0.03579219799996736,
132
- 0.036731337999981406,
133
- 0.036616682999977,
134
- 0.03575532999997222,
135
- 0.03661960900001304,
136
- 0.03748380700000098,
137
- 0.03658197800001517,
138
- 0.0356384219999768,
139
- 0.03160141999995858,
140
- 0.031102004000047145,
141
- 0.030923769999958495,
142
- 0.03119372500003692,
143
- 0.030934419999994134,
144
- 0.03117821700004697,
145
- 0.030818362000047728,
146
- 0.03134465800002317,
147
- 0.030933457999992697,
148
- 0.031266341999980796
149
  ]
150
  },
151
  "throughput": {
152
  "unit": "samples/s",
153
- "value": 28.884034730974935
154
  },
155
  "energy": {
156
  "unit": "kWh",
157
- "cpu": 1.2084031345868353e-06,
158
- "ram": 5.050091595452696e-08,
159
  "gpu": 0.0,
160
- "total": 1.2589040505413623e-06
161
  },
162
  "efficiency": {
163
  "unit": "samples/kWh",
164
- "value": 794341.7129923232
165
  }
166
  }
167
  }
 
81
  "processor": "x86_64",
82
  "python_version": "3.10.14",
83
  "optimum_benchmark_version": "0.3.1",
84
+ "optimum_benchmark_commit": "f837debaab7bc93d6490c10e065312a52e57b795",
85
  "transformers_version": "4.42.4",
86
  "transformers_commit": null,
87
  "accelerate_version": "0.32.1",
 
100
  "forward": {
101
  "memory": {
102
  "unit": "MB",
103
+ "max_ram": 939.061248,
104
  "max_global_vram": null,
105
  "max_process_vram": null,
106
  "max_reserved": null,
 
108
  },
109
  "latency": {
110
  "unit": "s",
111
+ "count": 27,
112
+ "total": 1.0046774289998552,
113
+ "mean": 0.037210275148142784,
114
+ "stdev": 0.002973652806825925,
115
+ "p50": 0.03872174299999642,
116
+ "p90": 0.03994491459997107,
117
+ "p95": 0.0401707419999866,
118
+ "p99": 0.040761405499987406,
119
  "values": [
120
+ 0.040256817999988925,
121
+ 0.03878219799997851,
122
+ 0.03714330799999743,
123
+ 0.03932511099998237,
124
+ 0.03824359199995797,
125
+ 0.039282150000019556,
126
+ 0.03872174299999642,
127
+ 0.039969897999981185,
128
+ 0.04093869299998687,
129
+ 0.039546208999979626,
130
+ 0.03858397400000513,
131
+ 0.03992825899996433,
132
+ 0.03932582199996659,
133
+ 0.038750848999995924,
134
+ 0.038841681000008066,
135
+ 0.03933889699999327,
136
+ 0.039409901999988506,
137
+ 0.03835199700000658,
138
+ 0.037515691000010065,
139
+ 0.03340292699999736,
140
+ 0.03348249800001213,
141
+ 0.033424458000013146,
142
+ 0.03309147100003429,
143
+ 0.03203313500000604,
144
+ 0.032101354000019455,
145
+ 0.032825498999955016,
146
+ 0.03205929500001048
 
 
147
  ]
148
  },
149
  "throughput": {
150
  "unit": "samples/s",
151
+ "value": 26.874297382074353
152
  },
153
  "energy": {
154
  "unit": "kWh",
155
+ "cpu": 1.3751970853842083e-06,
156
+ "ram": 5.7470542370216706e-08,
157
  "gpu": 0.0,
158
+ "total": 1.432667627754425e-06
159
  },
160
  "efficiency": {
161
  "unit": "samples/kWh",
162
+ "value": 697998.6010903368
163
  }
164
  }
165
  }