IlyasMoutawwakil HF staff commited on
Commit
8c365e2
1 Parent(s): 890e2b1

Upload cpu_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 940.224512,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
@@ -10,55 +10,58 @@
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 24,
14
- "total": 1.002178846999982,
15
- "mean": 0.04175745195833258,
16
- "stdev": 0.0024072203492112466,
17
- "p50": 0.040595760999991626,
18
- "p90": 0.04619594190000385,
19
- "p95": 0.04656686894998643,
20
- "p99": 0.04693666874001593,
21
  "values": [
22
- 0.040718163999997614,
23
- 0.04075360000001638,
24
- 0.040562773999994306,
25
- 0.040545672999996896,
26
- 0.0401632989999996,
27
- 0.0400997899999993,
28
- 0.03984096899998235,
29
- 0.03989334600001371,
30
- 0.04009529199998951,
31
- 0.040628747999988946,
32
- 0.04232462999999598,
33
- 0.04016376000001287,
34
- 0.04129198400002565,
35
- 0.04133614699998134,
36
- 0.04056064900001388,
37
- 0.04471263399997838,
38
- 0.0459473950000131,
39
- 0.047033191000025454,
40
- 0.046613528999984055,
41
- 0.04630246199999988,
42
- 0.04331350399999678,
43
- 0.03951679400000785,
44
- 0.04007705799998007,
45
- 0.039683454999988044
 
 
 
46
  ]
47
  },
48
  "throughput": {
49
  "unit": "samples/s",
50
- "value": 23.94782136127089
51
  },
52
  "energy": {
53
  "unit": "kWh",
54
- "cpu": 1.5722363260057238e-06,
55
- "ram": 6.570593579440356e-08,
56
  "gpu": 0.0,
57
- "total": 1.6379422618001276e-06
58
  },
59
  "efficiency": {
60
  "unit": "samples/kWh",
61
- "value": 610522.1309211366
62
  }
63
  }
64
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 944.050176,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
 
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 27,
14
+ "total": 1.0370580829999199,
15
+ "mean": 0.03840955862962666,
16
+ "stdev": 0.0011961263415751992,
17
+ "p50": 0.03818323900000564,
18
+ "p90": 0.039068168799997235,
19
+ "p95": 0.04038487190000808,
20
+ "p99": 0.04267542652000031,
21
  "values": [
22
+ 0.038746060999983456,
23
+ 0.03839107200002445,
24
+ 0.03838606200000072,
25
+ 0.038361575999999786,
26
+ 0.03868244099999174,
27
+ 0.038506819999980735,
28
+ 0.03832324300000778,
29
+ 0.03866820499999335,
30
+ 0.038941749999992226,
31
+ 0.03925779700000476,
32
+ 0.03845991099998969,
33
+ 0.03818323900000564,
34
+ 0.03802958999997941,
35
+ 0.03794215400000667,
36
+ 0.043310501999997086,
37
+ 0.0408679040000095,
38
+ 0.037612151999979915,
39
+ 0.037402585999984694,
40
+ 0.03773981300000173,
41
+ 0.037985825999982126,
42
+ 0.0377236019999998,
43
+ 0.037744472000014184,
44
+ 0.03746149799999898,
45
+ 0.03760681200000704,
46
+ 0.03741033099998958,
47
+ 0.03770844699999998,
48
+ 0.03760421699999483
49
  ]
50
  },
51
  "throughput": {
52
  "unit": "samples/s",
53
+ "value": 26.0351859192848
54
  },
55
  "energy": {
56
  "unit": "kWh",
57
+ "cpu": 1.4719451405874496e-06,
58
+ "ram": 6.151547720944732e-08,
59
  "gpu": 0.0,
60
+ "total": 1.533460617796897e-06
61
  },
62
  "efficiency": {
63
  "unit": "samples/kWh",
64
+ "value": 652119.7795328366
65
  }
66
  }
67
  }