IlyasMoutawwakil HF staff commited on
Commit
60bd520
1 Parent(s): d9cf307

Upload cpu_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cpu_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
@@ -76,14 +76,14 @@
76
  "cpu_ram_mb": 16757.342208,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-6.5.0-1023-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.3.1",
83
- "optimum_benchmark_commit": "328b924ef0be0164f0dc86652abfd3746f634c6b",
84
  "transformers_version": "4.42.4",
85
  "transformers_commit": null,
86
- "accelerate_version": "0.32.1",
87
  "accelerate_commit": null,
88
  "diffusers_version": "0.29.2",
89
  "diffusers_commit": null,
@@ -99,7 +99,7 @@
99
  "load": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 1162.952704,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -108,31 +108,31 @@
108
  "latency": {
109
  "unit": "s",
110
  "count": 1,
111
- "total": 4.924514202000012,
112
- "mean": 4.924514202000012,
113
  "stdev": 0.0,
114
- "p50": 4.924514202000012,
115
- "p90": 4.924514202000012,
116
- "p95": 4.924514202000012,
117
- "p99": 4.924514202000012,
118
  "values": [
119
- 4.924514202000012
120
  ]
121
  },
122
  "throughput": null,
123
  "energy": {
124
  "unit": "kWh",
125
- "cpu": 6.925355990727743e-05,
126
- "ram": 2.8946190004342044e-06,
127
  "gpu": 0,
128
- "total": 7.214817890771163e-05
129
  },
130
  "efficiency": null
131
  },
132
  "forward": {
133
  "memory": {
134
  "unit": "MB",
135
- "max_ram": 1191.059456,
136
  "max_global_vram": null,
137
  "max_process_vram": null,
138
  "max_reserved": null,
@@ -141,37 +141,37 @@
141
  "latency": {
142
  "unit": "s",
143
  "count": 7,
144
- "total": 1.039402308999911,
145
- "mean": 0.14848604414284441,
146
- "stdev": 0.006116268189182509,
147
- "p50": 0.15128922699994973,
148
- "p90": 0.15310421939999513,
149
- "p95": 0.1531341311999995,
150
- "p99": 0.15315806064000298,
151
  "values": [
152
- 0.15306433699998934,
153
- 0.15093825100001368,
154
- 0.15128922699994973,
155
- 0.15316404300000386,
156
- 0.15304609300000038,
157
- 0.1378352639999889,
158
- 0.14006509399996503
159
  ]
160
  },
161
  "throughput": {
162
  "unit": "samples/s",
163
- "value": 6.73463964760213
164
  },
165
  "energy": {
166
  "unit": "kWh",
167
- "cpu": 5.338464511765375e-06,
168
- "ram": 2.231095761349877e-07,
169
  "gpu": 0.0,
170
- "total": 5.561574087900363e-06
171
  },
172
  "efficiency": {
173
  "unit": "samples/kWh",
174
- "value": 179805.21057439077
175
  }
176
  }
177
  }
 
3
  "name": "cpu_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
 
76
  "cpu_ram_mb": 16757.342208,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-6.5.0-1024-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.3.1",
83
+ "optimum_benchmark_commit": "b0801269b3611e452bb077a62163b08a99ceb2a9",
84
  "transformers_version": "4.42.4",
85
  "transformers_commit": null,
86
+ "accelerate_version": "0.33.0",
87
  "accelerate_commit": null,
88
  "diffusers_version": "0.29.2",
89
  "diffusers_commit": null,
 
99
  "load": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 1168.285696,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
108
  "latency": {
109
  "unit": "s",
110
  "count": 1,
111
+ "total": 5.0882778890000395,
112
+ "mean": 5.0882778890000395,
113
  "stdev": 0.0,
114
+ "p50": 5.0882778890000395,
115
+ "p90": 5.0882778890000395,
116
+ "p95": 5.0882778890000395,
117
+ "p99": 5.0882778890000395,
118
  "values": [
119
+ 5.0882778890000395
120
  ]
121
  },
122
  "throughput": null,
123
  "energy": {
124
  "unit": "kWh",
125
+ "cpu": 7.658259206347995e-05,
126
+ "ram": 3.2010127144644685e-06,
127
  "gpu": 0,
128
+ "total": 7.978360477794442e-05
129
  },
130
  "efficiency": null
131
  },
132
  "forward": {
133
  "memory": {
134
  "unit": "MB",
135
+ "max_ram": 1199.529984,
136
  "max_global_vram": null,
137
  "max_process_vram": null,
138
  "max_reserved": null,
 
141
  "latency": {
142
  "unit": "s",
143
  "count": 7,
144
+ "total": 1.0530784339999286,
145
+ "mean": 0.1504397762857041,
146
+ "stdev": 0.0055780564465626355,
147
+ "p50": 0.15176010499999393,
148
+ "p90": 0.15562274339998794,
149
+ "p95": 0.15702729870000098,
150
+ "p99": 0.15815094294001142,
151
  "values": [
152
+ 0.15291728300002205,
153
+ 0.15124883799995814,
154
+ 0.15176010499999393,
155
+ 0.15375000299997055,
156
+ 0.15843185400001403,
157
+ 0.14447072999996635,
158
+ 0.14049962100000357
159
  ]
160
  },
161
  "throughput": {
162
  "unit": "samples/s",
163
+ "value": 6.647178191098037
164
  },
165
  "energy": {
166
  "unit": "kWh",
167
+ "cpu": 4.731459970827457e-06,
168
+ "ram": 1.977376332126823e-07,
169
  "gpu": 0.0,
170
+ "total": 4.9291976040401396e-06
171
  },
172
  "efficiency": {
173
  "unit": "samples/kWh",
174
+ "value": 202872.7757191892
175
  }
176
  }
177
  }