IlyasMoutawwakil HF staff commited on
Commit
b5057af
1 Parent(s): ce5d404

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -78,7 +78,7 @@
78
  "processor": "x86_64",
79
  "python_version": "3.10.14",
80
  "optimum_benchmark_version": "0.2.1",
81
- "optimum_benchmark_commit": "48414f58841d7ba7c7fd42d74fd524d1d23c3081",
82
  "transformers_version": "4.40.2",
83
  "transformers_commit": null,
84
  "accelerate_version": "0.30.1",
@@ -97,7 +97,7 @@
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
- "max_ram": 938.590208,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
@@ -105,59 +105,60 @@
105
  },
106
  "latency": {
107
  "unit": "s",
108
- "count": 28,
109
- "total": 1.021487527999966,
110
- "mean": 0.036481697428570214,
111
- "stdev": 0.0008459361313503326,
112
- "p50": 0.03652491449997797,
113
- "p90": 0.03750313130001359,
114
- "p95": 0.037567588950025765,
115
- "p99": 0.03799413227998514,
116
  "values": [
117
- 0.038149553999971886,
118
- 0.037332528000035836,
119
- 0.03757391800002097,
120
- 0.037480544000004556,
121
- 0.03436062599996603,
122
- 0.03651485099999263,
123
- 0.03687416100001428,
124
- 0.0365349779999633,
125
- 0.035414565000053244,
126
- 0.036579381999956695,
127
- 0.03607976799997914,
128
- 0.036629694999987805,
129
- 0.037048807000019224,
130
- 0.036248583000030976,
131
- 0.036300519999997505,
132
- 0.03644517000003589,
133
- 0.03697467000000643,
134
- 0.036477542000000085,
135
- 0.035670582999955514,
136
- 0.03463719299998047,
137
- 0.03616286300001548,
138
- 0.03679490399997576,
139
- 0.036220600999968156,
140
- 0.03625604699999485,
141
- 0.03521152500002245,
142
- 0.03680860899999061,
143
- 0.03714950599999156,
144
- 0.037555835000034676
 
145
  ]
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
- "value": 27.411005257032304
150
  },
151
  "energy": {
152
  "unit": "kWh",
153
- "cpu": 1.2576988093741906e-06,
154
- "ram": 5.2561183128230576e-08,
155
  "gpu": 0.0,
156
- "total": 1.3102599925024213e-06
157
  },
158
  "efficiency": {
159
  "unit": "samples/kWh",
160
- "value": 763207.306734699
161
  }
162
  }
163
  }
 
78
  "processor": "x86_64",
79
  "python_version": "3.10.14",
80
  "optimum_benchmark_version": "0.2.1",
81
+ "optimum_benchmark_commit": "2516ce57a5b64eefeb78dc75a171e0cdff88823e",
82
  "transformers_version": "4.40.2",
83
  "transformers_commit": null,
84
  "accelerate_version": "0.30.1",
 
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
+ "max_ram": 936.808448,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
 
105
  },
106
  "latency": {
107
  "unit": "s",
108
+ "count": 29,
109
+ "total": 1.0147147829998744,
110
+ "mean": 0.03499016493103015,
111
+ "stdev": 0.0007279547982049233,
112
+ "p50": 0.03503617699999495,
113
+ "p90": 0.03573674660002553,
114
+ "p95": 0.03618377639999153,
115
+ "p99": 0.03629199856001151,
116
  "values": [
117
+ 0.03631409000001895,
118
+ 0.035046595999972396,
119
+ 0.03451718899998468,
120
+ 0.036106652999990274,
121
+ 0.03358430999998063,
122
+ 0.03516179100000727,
123
+ 0.034629198000004635,
124
+ 0.03514327699997466,
125
+ 0.03498393999996097,
126
+ 0.03543017199996257,
127
+ 0.035018965999995544,
128
+ 0.03289524500002017,
129
+ 0.03562841100000469,
130
+ 0.034976516999961405,
131
+ 0.03501180199998544,
132
+ 0.03463143200002605,
133
+ 0.03503617699999495,
134
+ 0.03374574099996153,
135
+ 0.03519043500000407,
136
+ 0.034186072000011336,
137
+ 0.03514610100000937,
138
+ 0.03539599700002327,
139
+ 0.03501294300002655,
140
+ 0.035431072999983826,
141
+ 0.03514369700002362,
142
+ 0.03450180999999475,
143
+ 0.03564427000003434,
144
+ 0.03623519199999237,
145
+ 0.03496568599996408
146
  ]
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
+ "value": 28.579459455853407
151
  },
152
  "energy": {
153
  "unit": "kWh",
154
+ "cpu": 1.1748274946524428e-06,
155
+ "ram": 4.909816506872075e-08,
156
  "gpu": 0.0,
157
+ "total": 1.2239256597211635e-06
158
  },
159
  "efficiency": {
160
  "unit": "samples/kWh",
161
+ "value": 817043.0875906479
162
  }
163
  }
164
  }