IlyasMoutawwakil HF staff commited on
Commit
8926a83
·
verified ·
1 Parent(s): c6db70b

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.0",
83
- "optimum_benchmark_commit": "ebd20fcf042acf5db8d8956e7057fa93c82e14ab",
84
  "transformers_version": "4.40.2",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.30.1",
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 938.184704,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,61 +107,58 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 30,
111
- "total": 1.0082519999999704,
112
- "mean": 0.03360839999999901,
113
- "stdev": 0.0011557893681127009,
114
- "p50": 0.03358253450002735,
115
- "p90": 0.03504524500000344,
116
- "p95": 0.03552685609996331,
117
- "p99": 0.03571159646001888,
118
  "values": [
119
- 0.03358627100004696,
120
- 0.03428198699998575,
121
- 0.03303920199999766,
122
- 0.0332638009999755,
123
- 0.03391645599998583,
124
- 0.03497853300001452,
125
- 0.03354319999999689,
126
- 0.03184948699998813,
127
- 0.03185056900002792,
128
- 0.03238422200001878,
129
- 0.03254428100001405,
130
- 0.032321064999962346,
131
- 0.03450406000001749,
132
- 0.03552011199997196,
133
- 0.03453324399998792,
134
- 0.03244097799995416,
135
- 0.03234579100001156,
136
- 0.03245402200002445,
137
- 0.03194622799998115,
138
- 0.03473418800001582,
139
- 0.03366266299997278,
140
- 0.03357879800000774,
141
- 0.033161209000013514,
142
- 0.035532373999956235,
143
- 0.03390343199998824,
144
- 0.03264342500000339,
145
- 0.03499248200000693,
146
- 0.03578480000004447,
147
- 0.034552249000000756,
148
- 0.034402870999997504
149
  ]
150
  },
151
  "throughput": {
152
  "unit": "samples/s",
153
- "value": 29.754466145369296
154
  },
155
  "energy": {
156
  "unit": "kWh",
157
- "cpu": 1.3016993028146252e-06,
158
- "ram": 5.440036534121381e-08,
159
  "gpu": 0.0,
160
- "total": 1.356099668155839e-06
161
  },
162
  "efficiency": {
163
  "unit": "samples/kWh",
164
- "value": 737408.9261152175
165
  }
166
  }
167
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.0",
83
+ "optimum_benchmark_commit": "2d3261cf9b992810a685ce736f41254a91cd091e",
84
  "transformers_version": "4.40.2",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.30.1",
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 938.47552,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 27,
111
+ "total": 1.0055812380001043,
112
+ "mean": 0.03724374955555942,
113
+ "stdev": 0.001135183103969783,
114
+ "p50": 0.037510698000005505,
115
+ "p90": 0.038273721600000955,
116
+ "p95": 0.03854815469999835,
117
+ "p99": 0.03917231396003558,
118
  "values": [
119
+ 0.03690168700001095,
120
+ 0.03562123800003292,
121
+ 0.03564668699999629,
122
+ 0.03647698100002117,
123
+ 0.03408332599997266,
124
+ 0.037355728000022737,
125
+ 0.037032312000008005,
126
+ 0.03689864199998283,
127
+ 0.037305964999973185,
128
+ 0.037539271000014196,
129
+ 0.035390135999989525,
130
+ 0.03557062400000177,
131
+ 0.03751674000000094,
132
+ 0.038046081000004506,
133
+ 0.0378514770000038,
134
+ 0.03802572299997564,
135
+ 0.037510698000005505,
136
+ 0.03811543099999426,
137
+ 0.03790593900004069,
138
+ 0.03746882999996615,
139
+ 0.03797440799996821,
140
+ 0.03935893000004853,
141
+ 0.03864117599999872,
142
+ 0.03823546600000327,
143
+ 0.037051768000026186,
144
+ 0.03772486900004424,
145
+ 0.03833110499999748
 
 
 
146
  ]
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
+ "value": 26.850142961793406
151
  },
152
  "energy": {
153
  "unit": "kWh",
154
+ "cpu": 1.2381615324152842e-06,
155
+ "ram": 5.1744711971067524e-08,
156
  "gpu": 0.0,
157
+ "total": 1.2899062443863518e-06
158
  },
159
  "efficiency": {
160
  "unit": "samples/kWh",
161
+ "value": 775250.1426766338
162
  }
163
  }
164
  }