IlyasMoutawwakil HF staff commited on
Commit
f2032eb
1 Parent(s): f81a04c

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "forward": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 939.39712,
6
+ "max_global_vram": null,
7
+ "max_process_vram": null,
8
+ "max_reserved": null,
9
+ "max_allocated": null
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 33,
14
+ "total": 1.0082548800001518,
15
+ "mean": 0.030553178181822783,
16
+ "stdev": 0.00175309056604995,
17
+ "p50": 0.029733420000013666,
18
+ "p90": 0.03370184000001473,
19
+ "p95": 0.03401826339999161,
20
+ "p99": 0.03441965515998618,
21
+ "values": [
22
+ 0.034472210999979325,
23
+ 0.033724815000027775,
24
+ 0.033609939999962535,
25
+ 0.03430797400000074,
26
+ 0.033498162000000775,
27
+ 0.033825122999985524,
28
+ 0.031096044999969763,
29
+ 0.029733420000013666,
30
+ 0.02981750600002897,
31
+ 0.03249516899995797,
32
+ 0.030046884999990198,
33
+ 0.0291586160000179,
34
+ 0.029142075000038403,
35
+ 0.029352648000042336,
36
+ 0.030184923000035724,
37
+ 0.029891866000014033,
38
+ 0.02960390099997312,
39
+ 0.02963804099999834,
40
+ 0.029160449000016797,
41
+ 0.029296944000009262,
42
+ 0.02949319099997183,
43
+ 0.02924647999998342,
44
+ 0.029365111000004163,
45
+ 0.0298047030000248,
46
+ 0.02982126300003074,
47
+ 0.029433700000026874,
48
+ 0.029380349999996724,
49
+ 0.02941878199999337,
50
+ 0.029517145000056644,
51
+ 0.029141724999988128,
52
+ 0.02994289100001879,
53
+ 0.03215953199998012,
54
+ 0.029473294000013084
55
+ ]
56
+ },
57
+ "throughput": {
58
+ "unit": "samples/s",
59
+ "value": 32.729819269503594
60
+ },
61
+ "energy": {
62
+ "unit": "kWh",
63
+ "cpu": 1.1300574408637152e-06,
64
+ "ram": 4.722718188192435e-08,
65
+ "gpu": 0.0,
66
+ "total": 1.1772846227456398e-06
67
+ },
68
+ "efficiency": {
69
+ "unit": "samples/kWh",
70
+ "value": 849412.266736161
71
+ }
72
+ }
73
+ }