IlyasMoutawwakil HF staff commited on
Commit
b1b6e8f
1 Parent(s): 7ce3fa0

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -78,7 +78,7 @@
78
  "processor": "x86_64",
79
  "python_version": "3.10.14",
80
  "optimum_benchmark_version": "0.2.0",
81
- "optimum_benchmark_commit": "94e541958c0ccf6b26acb5568dd64a080a1112b6",
82
  "transformers_version": "4.40.2",
83
  "transformers_commit": null,
84
  "accelerate_version": "0.30.1",
@@ -97,7 +97,7 @@
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
- "max_ram": 936.615936,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
@@ -105,60 +105,59 @@
105
  },
106
  "latency": {
107
  "unit": "s",
108
- "count": 29,
109
- "total": 1.016155399999775,
110
- "mean": 0.035039841379302585,
111
- "stdev": 0.0026555138290588897,
112
- "p50": 0.03376868799995236,
113
- "p90": 0.03870816119999745,
114
- "p95": 0.0389829794000093,
115
- "p99": 0.03914403499998343,
116
  "values": [
117
- 0.037724196999988635,
118
- 0.038665170999991005,
119
- 0.037768078999988575,
120
- 0.03794336500004647,
121
- 0.03905155100000002,
122
- 0.03758536799995227,
123
- 0.03888012200002322,
124
- 0.03810063800000307,
125
- 0.03800771599998143,
126
- 0.03918000099997698,
127
- 0.03813908000000765,
128
- 0.03485563899999988,
129
- 0.03261198699999568,
130
- 0.03357896899996149,
131
- 0.03251609799997368,
132
- 0.03276753000000099,
133
- 0.032610121999994135,
134
- 0.0321942460000173,
135
- 0.032280436999997164,
136
- 0.032438221000006706,
137
- 0.03347698999999693,
138
- 0.03238878899998099,
139
- 0.03376868799995236,
140
- 0.03243923599995924,
141
- 0.03250867599996354,
142
- 0.03353637200001458,
143
- 0.03409169699995118,
144
- 0.03526088900002833,
145
- 0.03178552600002149
146
  ]
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
- "value": 28.538941976794515
151
  },
152
  "energy": {
153
  "unit": "kWh",
154
- "cpu": 1.3160001790082015e-06,
155
- "ram": 5.4997153930277895e-08,
156
  "gpu": 0.0,
157
- "total": 1.3709973329384794e-06
158
  },
159
  "efficiency": {
160
  "unit": "samples/kWh",
161
- "value": 729396.0214034005
162
  }
163
  }
164
  }
 
78
  "processor": "x86_64",
79
  "python_version": "3.10.14",
80
  "optimum_benchmark_version": "0.2.0",
81
+ "optimum_benchmark_commit": "c78698d47f8aff03fc8162bfacd2946f5882e025",
82
  "transformers_version": "4.40.2",
83
  "transformers_commit": null,
84
  "accelerate_version": "0.30.1",
 
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
+ "max_ram": 936.767488,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
 
105
  },
106
  "latency": {
107
  "unit": "s",
108
+ "count": 28,
109
+ "total": 1.0101344830001722,
110
+ "mean": 0.036076231535720434,
111
+ "stdev": 0.000694139201075085,
112
+ "p50": 0.03610273799998254,
113
+ "p90": 0.0368631685000139,
114
+ "p95": 0.0370095093000117,
115
+ "p99": 0.03728659555002707,
116
  "values": [
117
+ 0.036092067999959454,
118
+ 0.03657427499996402,
119
+ 0.03627225300004966,
120
+ 0.037370143000032385,
121
+ 0.03492255499998009,
122
+ 0.036914426000009826,
123
+ 0.036810113000001365,
124
+ 0.03633397800001603,
125
+ 0.036814120000030925,
126
+ 0.03706070800001271,
127
+ 0.035931369000024915,
128
+ 0.03616614599997092,
129
+ 0.036841201000015644,
130
+ 0.035238772000013796,
131
+ 0.035587830999986636,
132
+ 0.035485891999996966,
133
+ 0.03598002900002939,
134
+ 0.035328329000037684,
135
+ 0.03523166899998387,
136
+ 0.03447857999998405,
137
+ 0.03579876300000251,
138
+ 0.03569773500004203,
139
+ 0.03595847999997659,
140
+ 0.03611340800000562,
141
+ 0.035329200000035144,
142
+ 0.0367212780000159,
143
+ 0.0365498379999849,
144
+ 0.036531324000009135
 
145
  ]
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
+ "value": 27.719081440362263
150
  },
151
  "energy": {
152
  "unit": "kWh",
153
+ "cpu": 1.2407072302367953e-06,
154
+ "ram": 5.18503560006689e-08,
155
  "gpu": 0.0,
156
+ "total": 1.2925575862374642e-06
157
  },
158
  "efficiency": {
159
  "unit": "samples/kWh",
160
+ "value": 773659.9209563445
161
  }
162
  }
163
  }