IlyasMoutawwakil HF staff commited on
Commit
e8d9932
1 Parent(s): f05daa4

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
- "optimum_benchmark_commit": "88e73025c13ec1a964328bd3aa8fa3b1b8ab2582",
84
  "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.31.0",
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 936.771584,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,59 +107,61 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 28,
111
- "total": 1.0059070249998285,
112
- "mean": 0.03592525089285102,
113
- "stdev": 0.00196947294479433,
114
- "p50": 0.03658272849997957,
115
- "p90": 0.037491591399981414,
116
- "p95": 0.037970250849986086,
117
- "p99": 0.03827234601999862,
118
  "values": [
119
- 0.03729326000001265,
120
- 0.03649752399996942,
121
- 0.03624167599997463,
122
- 0.03545219199997973,
123
- 0.036897701000043526,
124
- 0.03634514999998828,
125
- 0.036348164999992605,
126
- 0.03833632900000339,
127
- 0.038099354999985735,
128
- 0.03773048599998674,
129
- 0.03722083399998155,
130
- 0.036848959999986164,
131
- 0.03651132999999618,
132
- 0.036654126999962955,
133
- 0.03649788500001705,
134
- 0.03691661700003124,
135
- 0.03706347999997206,
136
- 0.03670915899999727,
137
- 0.037389207999979135,
138
- 0.03733694199996762,
139
- 0.03695067000001018,
140
- 0.03601361099998712,
141
- 0.031191546999991715,
142
- 0.0319279310000411,
143
- 0.032116223999992144,
144
- 0.032740670000009686,
145
- 0.03434033600001385,
146
- 0.032235655999954815
 
 
147
  ]
148
  },
149
  "throughput": {
150
  "unit": "samples/s",
151
- "value": 27.835574565159014
152
  },
153
  "energy": {
154
  "unit": "kWh",
155
- "cpu": 1.2275965677367317e-06,
156
- "ram": 5.130300258727516e-08,
157
  "gpu": 0.0,
158
- "total": 1.278899570324007e-06
159
  },
160
  "efficiency": {
161
  "unit": "samples/kWh",
162
- "value": 781922.2268928058
163
  }
164
  }
165
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
+ "optimum_benchmark_commit": "dd02f26cb819965cbf86e16d9ce013cddc3b86af",
84
  "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.31.0",
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 937.426944,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 30,
111
+ "total": 1.024279958999898,
112
+ "mean": 0.03414266529999661,
113
+ "stdev": 0.002003813720873649,
114
+ "p50": 0.034858600000006845,
115
+ "p90": 0.03565418159996057,
116
+ "p95": 0.035895198800008646,
117
+ "p99": 0.03596436321000283,
118
  "values": [
119
+ 0.035967481000000134,
120
+ 0.03514002799994387,
121
+ 0.035576557000013054,
122
+ 0.03432157199995345,
123
+ 0.03563575799995533,
124
+ 0.03488703400000759,
125
+ 0.03473305500000379,
126
+ 0.03504126299998234,
127
+ 0.03468897199996945,
128
+ 0.034071383000025435,
129
+ 0.03581999400000768,
130
+ 0.03483442499998546,
131
+ 0.03480534999999918,
132
+ 0.03475111799997421,
133
+ 0.03510695500000338,
134
+ 0.03486445100003266,
135
+ 0.03482678099999248,
136
+ 0.035538115000008474,
137
+ 0.035070797999992465,
138
+ 0.03485274899998103,
139
+ 0.035258930999987115,
140
+ 0.0355690829999844,
141
+ 0.03595673000000943,
142
+ 0.03492708800001765,
143
+ 0.03258657999998604,
144
+ 0.030260984000051394,
145
+ 0.029760967000015626,
146
+ 0.029760896000027515,
147
+ 0.029487383000002865,
148
+ 0.030177477999984603
149
  ]
150
  },
151
  "throughput": {
152
  "unit": "samples/s",
153
+ "value": 29.288867497995223
154
  },
155
  "energy": {
156
  "unit": "kWh",
157
+ "cpu": 1.1727454226001415e-06,
158
+ "ram": 4.900842445615982e-08,
159
  "gpu": 0.0,
160
+ "total": 1.2217538470563013e-06
161
  },
162
  "efficiency": {
163
  "unit": "samples/kWh",
164
+ "value": 818495.47878192
165
  }
166
  }
167
  }