begumcig commited on
Commit
9c827e3
·
verified ·
1 Parent(s): e5c58ae

Upload folder using huggingface_hub (#2)

Browse files

- e38d3c813644106466dd76e3d4466d85601a9089753ae72fc89f6292a14c6f6e (94ef98c6e919848480f3d09f8d1e026b8d992bbc)

Files changed (3) hide show
  1. base_results.json +19 -0
  2. plots.png +0 -0
  3. smashed_results.json +19 -0
base_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "current_gpu_type": "Tesla T4",
3
+ "current_gpu_total_memory": 15095.0625,
4
+ "perplexity": 3.4586403369903564,
5
+ "memory_inference_first": 808.0,
6
+ "memory_inference": 808.0,
7
+ "token_generation_latency_sync": 38.27063522338867,
8
+ "token_generation_latency_async": 37.724124267697334,
9
+ "token_generation_throughput_sync": 0.026129694324719784,
10
+ "token_generation_throughput_async": 0.026508236292082377,
11
+ "token_generation_CO2_emissions": 1.899763620684379e-05,
12
+ "token_generation_energy_consumption": 0.0018430467100968824,
13
+ "inference_latency_sync": 118.64526290893555,
14
+ "inference_latency_async": 47.777557373046875,
15
+ "inference_throughput_sync": 0.008428486527671445,
16
+ "inference_throughput_async": 0.02093032911230698,
17
+ "inference_CO2_emissions": 1.8905640125185246e-05,
18
+ "inference_energy_consumption": 6.517715960578611e-05
19
+ }
plots.png ADDED
smashed_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "current_gpu_type": "Tesla T4",
3
+ "current_gpu_total_memory": 15095.0625,
4
+ "perplexity": 3.4594500064849854,
5
+ "memory_inference_first": 358.0,
6
+ "memory_inference": 358.0,
7
+ "token_generation_latency_sync": 73.98157806396485,
8
+ "token_generation_latency_async": 74.57008305937052,
9
+ "token_generation_throughput_sync": 0.01351687847392759,
10
+ "token_generation_throughput_async": 0.013410203649683872,
11
+ "token_generation_CO2_emissions": 9.466104547144268e-06,
12
+ "token_generation_energy_consumption": 0.003595659407598109,
13
+ "inference_latency_sync": 85.0370491027832,
14
+ "inference_latency_async": 66.51849746704102,
15
+ "inference_throughput_sync": 0.011759580212988254,
16
+ "inference_throughput_async": 0.015033412330088874,
17
+ "inference_CO2_emissions": 9.849718661233337e-06,
18
+ "inference_energy_consumption": 3.3456618429771876e-05
19
+ }