IlyasMoutawwakil HF staff commited on
Commit
de6721b
·
verified ·
1 Parent(s): b8bc718

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 938.590208,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
@@ -10,59 +10,60 @@
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 28,
14
- "total": 1.021487527999966,
15
- "mean": 0.036481697428570214,
16
- "stdev": 0.0008459361313503326,
17
- "p50": 0.03652491449997797,
18
- "p90": 0.03750313130001359,
19
- "p95": 0.037567588950025765,
20
- "p99": 0.03799413227998514,
21
  "values": [
22
- 0.038149553999971886,
23
- 0.037332528000035836,
24
- 0.03757391800002097,
25
- 0.037480544000004556,
26
- 0.03436062599996603,
27
- 0.03651485099999263,
28
- 0.03687416100001428,
29
- 0.0365349779999633,
30
- 0.035414565000053244,
31
- 0.036579381999956695,
32
- 0.03607976799997914,
33
- 0.036629694999987805,
34
- 0.037048807000019224,
35
- 0.036248583000030976,
36
- 0.036300519999997505,
37
- 0.03644517000003589,
38
- 0.03697467000000643,
39
- 0.036477542000000085,
40
- 0.035670582999955514,
41
- 0.03463719299998047,
42
- 0.03616286300001548,
43
- 0.03679490399997576,
44
- 0.036220600999968156,
45
- 0.03625604699999485,
46
- 0.03521152500002245,
47
- 0.03680860899999061,
48
- 0.03714950599999156,
49
- 0.037555835000034676
 
50
  ]
51
  },
52
  "throughput": {
53
  "unit": "samples/s",
54
- "value": 27.411005257032304
55
  },
56
  "energy": {
57
  "unit": "kWh",
58
- "cpu": 1.2576988093741906e-06,
59
- "ram": 5.2561183128230576e-08,
60
  "gpu": 0.0,
61
- "total": 1.3102599925024213e-06
62
  },
63
  "efficiency": {
64
  "unit": "samples/kWh",
65
- "value": 763207.306734699
66
  }
67
  }
68
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 936.808448,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
 
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 29,
14
+ "total": 1.0147147829998744,
15
+ "mean": 0.03499016493103015,
16
+ "stdev": 0.0007279547982049233,
17
+ "p50": 0.03503617699999495,
18
+ "p90": 0.03573674660002553,
19
+ "p95": 0.03618377639999153,
20
+ "p99": 0.03629199856001151,
21
  "values": [
22
+ 0.03631409000001895,
23
+ 0.035046595999972396,
24
+ 0.03451718899998468,
25
+ 0.036106652999990274,
26
+ 0.03358430999998063,
27
+ 0.03516179100000727,
28
+ 0.034629198000004635,
29
+ 0.03514327699997466,
30
+ 0.03498393999996097,
31
+ 0.03543017199996257,
32
+ 0.035018965999995544,
33
+ 0.03289524500002017,
34
+ 0.03562841100000469,
35
+ 0.034976516999961405,
36
+ 0.03501180199998544,
37
+ 0.03463143200002605,
38
+ 0.03503617699999495,
39
+ 0.03374574099996153,
40
+ 0.03519043500000407,
41
+ 0.034186072000011336,
42
+ 0.03514610100000937,
43
+ 0.03539599700002327,
44
+ 0.03501294300002655,
45
+ 0.035431072999983826,
46
+ 0.03514369700002362,
47
+ 0.03450180999999475,
48
+ 0.03564427000003434,
49
+ 0.03623519199999237,
50
+ 0.03496568599996408
51
  ]
52
  },
53
  "throughput": {
54
  "unit": "samples/s",
55
+ "value": 28.579459455853407
56
  },
57
  "energy": {
58
  "unit": "kWh",
59
+ "cpu": 1.1748274946524428e-06,
60
+ "ram": 4.909816506872075e-08,
61
  "gpu": 0.0,
62
+ "total": 1.2239256597211635e-06
63
  },
64
  "efficiency": {
65
  "unit": "samples/kWh",
66
+ "value": 817043.0875906479
67
  }
68
  }
69
  }