IlyasMoutawwakil HF staff commited on
Commit
5dc1845
1 Parent(s): fe2209a

Upload cpu_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "load": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 969.756672,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
@@ -11,15 +11,15 @@
11
  "latency": {
12
  "unit": "s",
13
  "values": [
14
- 1.3268693419999522
15
  ],
16
  "count": 1,
17
- "total": 1.3268693419999522,
18
- "mean": 1.3268693419999522,
19
- "p50": 1.3268693419999522,
20
- "p90": 1.3268693419999522,
21
- "p95": 1.3268693419999522,
22
- "p99": 1.3268693419999522,
23
  "stdev": 0,
24
  "stdev_": 0
25
  },
@@ -30,7 +30,7 @@
30
  "forward": {
31
  "memory": {
32
  "unit": "MB",
33
- "max_ram": 883.642368,
34
  "max_global_vram": null,
35
  "max_process_vram": null,
36
  "max_reserved": null,
@@ -39,50 +39,54 @@
39
  "latency": {
40
  "unit": "s",
41
  "values": [
42
- 0.056869258000006084,
43
- 0.058061517999988155,
44
- 0.0522853390000364,
45
- 0.05243081999998367,
46
- 0.05326780699999745,
47
- 0.055420992999984264,
48
- 0.05466478699997879,
49
- 0.05310630600001787,
50
- 0.05267342199999803,
51
- 0.054737862000024506,
52
- 0.05671406999999817,
53
- 0.05285651100001587,
54
- 0.055796852000014496,
55
- 0.052001501000006556,
56
- 0.05440275799998062,
57
- 0.05993688600000269,
58
- 0.055317720999994435,
59
- 0.05281495399998448,
60
- 0.05182520199997498
 
 
 
 
61
  ],
62
- "count": 19,
63
- "total": 1.0351845669999875,
64
- "mean": 0.054483398263157234,
65
- "p50": 0.05440275799998062,
66
- "p90": 0.0571077100000025,
67
- "p95": 0.058249054799989604,
68
- "p99": 0.059599319760000075,
69
- "stdev": 0.00218809072210616,
70
- "stdev_": 4.016068732602884
71
  },
72
  "throughput": {
73
  "unit": "samples/s",
74
- "value": 18.3542148962507
75
  },
76
  "energy": {
77
  "unit": "kWh",
78
- "cpu": 2.3141269575164147e-06,
79
- "ram": 9.675788286311704e-08,
80
  "gpu": 0.0,
81
- "total": 2.410884840379532e-06
82
  },
83
  "efficiency": {
84
  "unit": "samples/kWh",
85
- "value": 414785.46932278015
86
  }
87
  }
88
  }
 
2
  "load": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 970.264576,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
 
11
  "latency": {
12
  "unit": "s",
13
  "values": [
14
+ 1.3057213939999883
15
  ],
16
  "count": 1,
17
+ "total": 1.3057213939999883,
18
+ "mean": 1.3057213939999883,
19
+ "p50": 1.3057213939999883,
20
+ "p90": 1.3057213939999883,
21
+ "p95": 1.3057213939999883,
22
+ "p99": 1.3057213939999883,
23
  "stdev": 0,
24
  "stdev_": 0
25
  },
 
30
  "forward": {
31
  "memory": {
32
  "unit": "MB",
33
+ "max_ram": 884.0192,
34
  "max_global_vram": null,
35
  "max_process_vram": null,
36
  "max_reserved": null,
 
39
  "latency": {
40
  "unit": "s",
41
  "values": [
42
+ 0.046211259000017435,
43
+ 0.04559323399996629,
44
+ 0.046232780000025286,
45
+ 0.04598622899999327,
46
+ 0.046600116000035996,
47
+ 0.04639245700002448,
48
+ 0.04765987700000096,
49
+ 0.04757826499997009,
50
+ 0.047311877000026925,
51
+ 0.047730449000027875,
52
+ 0.04630255000000716,
53
+ 0.049103245999958745,
54
+ 0.04881664999999202,
55
+ 0.04657648300002393,
56
+ 0.04700572500001954,
57
+ 0.04705417499997111,
58
+ 0.04776282000000265,
59
+ 0.04187629599999809,
60
+ 0.039299558999971396,
61
+ 0.03899048099998481,
62
+ 0.039586024000016096,
63
+ 0.0396858000000293,
64
+ 0.039652748999969845
65
  ],
66
+ "count": 23,
67
+ "total": 1.0390091010000333,
68
+ "mean": 0.045174308739131884,
69
+ "p50": 0.04639245700002448,
70
+ "p90": 0.04775634580000769,
71
+ "p95": 0.048711266999993084,
72
+ "p99": 0.04904019487996607,
73
+ "stdev": 0.003300905176473354,
74
+ "stdev_": 7.307040812810427
75
  },
76
  "throughput": {
77
  "unit": "samples/s",
78
+ "value": 22.136475972985018
79
  },
80
  "energy": {
81
  "unit": "kWh",
82
+ "cpu": 1.5387237100427616e-06,
83
+ "ram": 6.433947586866954e-08,
84
  "gpu": 0.0,
85
+ "total": 1.603063185911431e-06
86
  },
87
  "efficiency": {
88
  "unit": "samples/kWh",
89
+ "value": 623805.7294238493
90
  }
91
  }
92
  }