IlyasMoutawwakil HF staff commited on
Commit
a32d598
1 Parent(s): 5034423

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -97,7 +97,7 @@
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
- "max_ram": 937.49248,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
@@ -106,57 +106,57 @@
106
  "latency": {
107
  "unit": "s",
108
  "count": 27,
109
- "total": 1.0255669960001228,
110
- "mean": 0.03798396281481936,
111
- "stdev": 0.0013657641120332562,
112
- "p50": 0.03817636399998037,
113
- "p90": 0.03896361340000567,
114
- "p95": 0.03951649189999103,
115
- "p99": 0.04031425351999132,
116
  "values": [
117
- 0.038789833999999246,
118
- 0.037985176000006504,
119
- 0.03855480300001091,
120
- 0.03869269200004055,
121
- 0.03647314000005508,
122
- 0.03890146300000197,
123
- 0.03817636399998037,
124
- 0.03971348599998237,
125
- 0.038459595000006175,
126
- 0.03883157199999232,
127
- 0.03855275899996968,
128
- 0.03571738400000868,
129
- 0.038099398999975165,
130
- 0.03821538799996915,
131
- 0.03765985500001534,
132
- 0.03812939599998799,
133
- 0.03905683900001122,
134
- 0.03849970999999641,
135
- 0.037342201000001296,
136
- 0.038257356000030995,
137
- 0.037666808000039964,
138
- 0.03800090600003614,
139
- 0.03811510900004578,
140
- 0.03757702899997639,
141
- 0.03666334799999049,
142
- 0.04052533399999447,
143
- 0.03291004999999814
144
  ]
145
  },
146
  "throughput": {
147
  "unit": "samples/s",
148
- "value": 26.326900246697065
149
  },
150
  "energy": {
151
  "unit": "kWh",
152
- "cpu": 1.2473264502154457e-06,
153
- "ram": 5.212426219828536e-08,
154
  "gpu": 0.0,
155
- "total": 1.299450712413731e-06
156
  },
157
  "efficiency": {
158
  "unit": "samples/kWh",
159
- "value": 769555.9288605098
160
  }
161
  }
162
  }
 
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
+ "max_ram": 936.808448,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
 
106
  "latency": {
107
  "unit": "s",
108
  "count": 27,
109
+ "total": 1.0110681230000296,
110
+ "mean": 0.037446967518519614,
111
+ "stdev": 0.000688873697815681,
112
+ "p50": 0.03757569100002911,
113
+ "p90": 0.03811274559999447,
114
+ "p95": 0.03818280149999964,
115
+ "p99": 0.038317894080009865,
116
  "values": [
117
+ 0.03812847099999317,
118
+ 0.036572600999988936,
119
+ 0.037022318999959225,
120
+ 0.0373983699999485,
121
+ 0.035084966999988865,
122
+ 0.038102261999995335,
123
+ 0.03739908200003583,
124
+ 0.03702988300000243,
125
+ 0.0368103839999776,
126
+ 0.03757569100002911,
127
+ 0.03631468000003224,
128
+ 0.03701594699998623,
129
+ 0.0376170079999838,
130
+ 0.03805969300003653,
131
+ 0.03794287499999882,
132
+ 0.03766147999999703,
133
+ 0.0376612800000089,
134
+ 0.0375501329999679,
135
+ 0.03729028800000833,
136
+ 0.03783137099998157,
137
+ 0.037796753000009176,
138
+ 0.03835717800001248,
139
+ 0.038087705000009464,
140
+ 0.0374619789999997,
141
+ 0.038206086000002415,
142
+ 0.038016283000047224,
143
+ 0.03707335400002876
144
  ]
145
  },
146
  "throughput": {
147
  "unit": "samples/s",
148
+ "value": 26.704432061299602
149
  },
150
  "energy": {
151
  "unit": "kWh",
152
+ "cpu": 1.2579312033977987e-06,
153
+ "ram": 5.257019776060419e-08,
154
  "gpu": 0.0,
155
+ "total": 1.310501401158403e-06
156
  },
157
  "efficiency": {
158
  "unit": "samples/kWh",
159
+ "value": 763066.7156220216
160
  }
161
  }
162
  }