e38d3c813644106466dd76e3d4466d85601a9089753ae72fc89f6292a14c6f6e
Browse files- base_results.json +19 -0
- plots.png +0 -0
- smashed_results.json +19 -0
base_results.json
ADDED
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"current_gpu_type": "Tesla T4",
|
3 |
+
"current_gpu_total_memory": 15095.0625,
|
4 |
+
"perplexity": 3.4586403369903564,
|
5 |
+
"memory_inference_first": 808.0,
|
6 |
+
"memory_inference": 808.0,
|
7 |
+
"token_generation_latency_sync": 38.27063522338867,
|
8 |
+
"token_generation_latency_async": 37.724124267697334,
|
9 |
+
"token_generation_throughput_sync": 0.026129694324719784,
|
10 |
+
"token_generation_throughput_async": 0.026508236292082377,
|
11 |
+
"token_generation_CO2_emissions": 1.899763620684379e-05,
|
12 |
+
"token_generation_energy_consumption": 0.0018430467100968824,
|
13 |
+
"inference_latency_sync": 118.64526290893555,
|
14 |
+
"inference_latency_async": 47.777557373046875,
|
15 |
+
"inference_throughput_sync": 0.008428486527671445,
|
16 |
+
"inference_throughput_async": 0.02093032911230698,
|
17 |
+
"inference_CO2_emissions": 1.8905640125185246e-05,
|
18 |
+
"inference_energy_consumption": 6.517715960578611e-05
|
19 |
+
}
|
plots.png
ADDED
smashed_results.json
ADDED
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"current_gpu_type": "Tesla T4",
|
3 |
+
"current_gpu_total_memory": 15095.0625,
|
4 |
+
"perplexity": 3.4594500064849854,
|
5 |
+
"memory_inference_first": 358.0,
|
6 |
+
"memory_inference": 358.0,
|
7 |
+
"token_generation_latency_sync": 73.98157806396485,
|
8 |
+
"token_generation_latency_async": 74.57008305937052,
|
9 |
+
"token_generation_throughput_sync": 0.01351687847392759,
|
10 |
+
"token_generation_throughput_async": 0.013410203649683872,
|
11 |
+
"token_generation_CO2_emissions": 9.466104547144268e-06,
|
12 |
+
"token_generation_energy_consumption": 0.003595659407598109,
|
13 |
+
"inference_latency_sync": 85.0370491027832,
|
14 |
+
"inference_latency_async": 66.51849746704102,
|
15 |
+
"inference_throughput_sync": 0.011759580212988254,
|
16 |
+
"inference_throughput_async": 0.015033412330088874,
|
17 |
+
"inference_CO2_emissions": 9.849718661233337e-06,
|
18 |
+
"inference_energy_consumption": 3.3456618429771876e-05
|
19 |
+
}
|