Commit
•
5e6df89
1
Parent(s):
a107411
Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub
Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json
CHANGED
@@ -80,7 +80,7 @@
|
|
80 |
"processor": "x86_64",
|
81 |
"python_version": "3.10.14",
|
82 |
"optimum_benchmark_version": "0.2.0",
|
83 |
-
"optimum_benchmark_commit": "
|
84 |
"transformers_version": "4.40.2",
|
85 |
"transformers_commit": null,
|
86 |
"accelerate_version": "0.30.1",
|
@@ -99,7 +99,7 @@
|
|
99 |
"forward": {
|
100 |
"memory": {
|
101 |
"unit": "MB",
|
102 |
-
"max_ram":
|
103 |
"max_global_vram": null,
|
104 |
"max_process_vram": null,
|
105 |
"max_reserved": null,
|
@@ -107,65 +107,62 @@
|
|
107 |
},
|
108 |
"latency": {
|
109 |
"unit": "s",
|
110 |
-
"count":
|
111 |
-
"total": 1.
|
112 |
-
"mean": 0.
|
113 |
-
"stdev": 0.
|
114 |
-
"p50": 0.
|
115 |
-
"p90": 0.
|
116 |
-
"p95": 0.
|
117 |
-
"p99": 0.
|
118 |
"values": [
|
119 |
-
0.
|
120 |
-
0.
|
121 |
-
0.
|
122 |
-
0.
|
123 |
-
0.
|
124 |
-
0.
|
125 |
-
0.
|
126 |
-
0.
|
127 |
-
0.
|
128 |
-
0.
|
129 |
-
0.
|
130 |
-
0.
|
131 |
-
0.
|
132 |
-
0.
|
133 |
-
0.
|
134 |
-
0.
|
135 |
-
0.
|
136 |
-
0.
|
137 |
-
0.
|
138 |
-
0.
|
139 |
-
0.
|
140 |
-
0.
|
141 |
-
0.
|
142 |
-
0.
|
143 |
-
0.
|
144 |
-
0.
|
145 |
-
0.
|
146 |
-
0.
|
147 |
-
0.
|
148 |
-
0.
|
149 |
-
0.
|
150 |
-
0.03006020800000897,
|
151 |
-
0.02938689699999486,
|
152 |
-
0.02881693599999835
|
153 |
]
|
154 |
},
|
155 |
"throughput": {
|
156 |
"unit": "samples/s",
|
157 |
-
"value":
|
158 |
},
|
159 |
"energy": {
|
160 |
"unit": "kWh",
|
161 |
-
"cpu": 1.
|
162 |
-
"ram":
|
163 |
"gpu": 0.0,
|
164 |
-
"total": 1.
|
165 |
},
|
166 |
"efficiency": {
|
167 |
"unit": "samples/kWh",
|
168 |
-
"value":
|
169 |
}
|
170 |
}
|
171 |
}
|
|
|
80 |
"processor": "x86_64",
|
81 |
"python_version": "3.10.14",
|
82 |
"optimum_benchmark_version": "0.2.0",
|
83 |
+
"optimum_benchmark_commit": "8ef7b0420a1d5aa903d786290f7ab247769da879",
|
84 |
"transformers_version": "4.40.2",
|
85 |
"transformers_commit": null,
|
86 |
"accelerate_version": "0.30.1",
|
|
|
99 |
"forward": {
|
100 |
"memory": {
|
101 |
"unit": "MB",
|
102 |
+
"max_ram": 941.412352,
|
103 |
"max_global_vram": null,
|
104 |
"max_process_vram": null,
|
105 |
"max_reserved": null,
|
|
|
107 |
},
|
108 |
"latency": {
|
109 |
"unit": "s",
|
110 |
+
"count": 31,
|
111 |
+
"total": 1.0028979849999473,
|
112 |
+
"mean": 0.03235154790322411,
|
113 |
+
"stdev": 0.0007254913291455751,
|
114 |
+
"p50": 0.0321730819999857,
|
115 |
+
"p90": 0.032961761000024126,
|
116 |
+
"p95": 0.033839419999992515,
|
117 |
+
"p99": 0.0347247656999798,
|
118 |
"values": [
|
119 |
+
0.034753568999974505,
|
120 |
+
0.03257596299999932,
|
121 |
+
0.032491295999989234,
|
122 |
+
0.03238255299999082,
|
123 |
+
0.03199964000003774,
|
124 |
+
0.03161154600002192,
|
125 |
+
0.03302128199999288,
|
126 |
+
0.03189084699999967,
|
127 |
+
0.03237584999999399,
|
128 |
+
0.03226585499999146,
|
129 |
+
0.03179584000002933,
|
130 |
+
0.03183968199999754,
|
131 |
+
0.03179251399996019,
|
132 |
+
0.03208153200000652,
|
133 |
+
0.032130536000011034,
|
134 |
+
0.0321730819999857,
|
135 |
+
0.032154007000031015,
|
136 |
+
0.03189455399996177,
|
137 |
+
0.03170949000002565,
|
138 |
+
0.031932163999954355,
|
139 |
+
0.03212115499997026,
|
140 |
+
0.03159412400003703,
|
141 |
+
0.03465755799999215,
|
142 |
+
0.032474474000025566,
|
143 |
+
0.032671211999968364,
|
144 |
+
0.032961761000024126,
|
145 |
+
0.03241093399998363,
|
146 |
+
0.032696487999999135,
|
147 |
+
0.03245410599998877,
|
148 |
+
0.03152453500001684,
|
149 |
+
0.03245983599998681
|
|
|
|
|
|
|
150 |
]
|
151 |
},
|
152 |
"throughput": {
|
153 |
"unit": "samples/s",
|
154 |
+
"value": 30.910422060526553
|
155 |
},
|
156 |
"energy": {
|
157 |
"unit": "kWh",
|
158 |
+
"cpu": 1.221868023276329e-06,
|
159 |
+
"ram": 5.1063289788414087e-08,
|
160 |
"gpu": 0.0,
|
161 |
+
"total": 1.272931313064743e-06
|
162 |
},
|
163 |
"efficiency": {
|
164 |
"unit": "samples/kWh",
|
165 |
+
"value": 785588.342227495
|
166 |
}
|
167 |
}
|
168 |
}
|