IlyasMoutawwakil HF staff commited on
Commit
c72a67d
1 Parent(s): d995ad0

Upload cpu_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "load": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 970.899456,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
@@ -11,15 +11,15 @@
11
  "latency": {
12
  "unit": "s",
13
  "values": [
14
- 1.317152555000007
15
  ],
16
  "count": 1,
17
- "total": 1.317152555000007,
18
- "mean": 1.317152555000007,
19
- "p50": 1.317152555000007,
20
- "p90": 1.317152555000007,
21
- "p95": 1.317152555000007,
22
- "p99": 1.317152555000007,
23
  "stdev": 0,
24
  "stdev_": 0
25
  },
@@ -30,7 +30,7 @@
30
  "forward": {
31
  "memory": {
32
  "unit": "MB",
33
- "max_ram": 884.785152,
34
  "max_global_vram": null,
35
  "max_process_vram": null,
36
  "max_reserved": null,
@@ -39,52 +39,53 @@
39
  "latency": {
40
  "unit": "s",
41
  "values": [
42
- 0.04893034999997781,
43
- 0.04933273299999996,
44
- 0.04940153199999031,
45
- 0.048061482999997907,
46
- 0.04845851599998241,
47
- 0.048120995000004996,
48
- 0.04769704099999217,
49
- 0.0480179509999914,
50
- 0.04878281299997411,
51
- 0.048490686000008054,
52
- 0.047670457999998916,
53
- 0.04906001599999854,
54
- 0.04761642400001165,
55
- 0.04878908899999601,
56
- 0.04808342800001242,
57
- 0.04883948299999474,
58
- 0.049956234000006816,
59
- 0.04861369400001081,
60
- 0.0438475379999943,
61
- 0.04385777799998891,
62
- 0.04458401700000536
 
63
  ],
64
- "count": 21,
65
- "total": 1.0062122589999376,
66
- "mean": 0.047914869476187505,
67
- "p50": 0.04845851599998241,
68
- "p90": 0.04933273299999996,
69
- "p95": 0.04940153199999031,
70
- "p99": 0.04984529360000352,
71
- "stdev": 0.0016706889424348956,
72
- "stdev_": 3.4867859616422128
73
  },
74
  "throughput": {
75
  "unit": "samples/s",
76
- "value": 20.87034799284959
77
  },
78
  "energy": {
79
  "unit": "kWh",
80
- "cpu": 1.7173942265700344e-06,
81
- "ram": 7.181181576903356e-08,
82
  "gpu": 0.0,
83
- "total": 1.789206042339068e-06
84
  },
85
  "efficiency": {
86
  "unit": "samples/kWh",
87
- "value": 558907.1221180755
88
  }
89
  }
90
  }
 
2
  "load": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 970.031104,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
 
11
  "latency": {
12
  "unit": "s",
13
  "values": [
14
+ 1.3080673270000034
15
  ],
16
  "count": 1,
17
+ "total": 1.3080673270000034,
18
+ "mean": 1.3080673270000034,
19
+ "p50": 1.3080673270000034,
20
+ "p90": 1.3080673270000034,
21
+ "p95": 1.3080673270000034,
22
+ "p99": 1.3080673270000034,
23
  "stdev": 0,
24
  "stdev_": 0
25
  },
 
30
  "forward": {
31
  "memory": {
32
  "unit": "MB",
33
+ "max_ram": 884.047872,
34
  "max_global_vram": null,
35
  "max_process_vram": null,
36
  "max_reserved": null,
 
39
  "latency": {
40
  "unit": "s",
41
  "values": [
42
+ 0.04683651999999938,
43
+ 0.04638292499998897,
44
+ 0.04696747399998458,
45
+ 0.04733514899999136,
46
+ 0.046675909999976284,
47
+ 0.046504872000014075,
48
+ 0.04784939699999313,
49
+ 0.04830640900001981,
50
+ 0.04814082000001463,
51
+ 0.046431405000021186,
52
+ 0.04835216300000411,
53
+ 0.05096218599999247,
54
+ 0.04859122000004845,
55
+ 0.04717217500001425,
56
+ 0.047541462999959094,
57
+ 0.04778462600000921,
58
+ 0.04831450400001813,
59
+ 0.04395642299999736,
60
+ 0.043303436000030615,
61
+ 0.04316184300000714,
62
+ 0.04432506999995667,
63
+ 0.04377436400000079
64
  ],
65
+ "count": 22,
66
+ "total": 1.0286703540000417,
67
+ "mean": 0.046757743363638256,
68
+ "p50": 0.04706982449999941,
69
+ "p90": 0.04834839710000551,
70
+ "p95": 0.048579267150046235,
71
+ "p99": 0.05046428314000422,
72
+ "stdev": 0.0019256390536110247,
73
+ "stdev_": 4.118331884914108
74
  },
75
  "throughput": {
76
  "unit": "samples/s",
77
+ "value": 21.386831956857492
78
  },
79
  "energy": {
80
  "unit": "kWh",
81
+ "cpu": 1.6670503564814387e-06,
82
+ "ram": 6.97067467201581e-08,
83
  "gpu": 0.0,
84
+ "total": 1.736757103201597e-06
85
  },
86
  "efficiency": {
87
  "unit": "samples/kWh",
88
+ "value": 575785.755046901
89
  }
90
  }
91
  }