AppleSwing
commited on
Commit
•
9ffef81
1
Parent(s):
f38163c
Fix some bugs
Browse files- backend-cli.py +26 -24
- src/backend/envs.py +1 -1
- src/display/utils.py +1 -0
- src/submission/check_validity.py +2 -1
- src/utils.py +102 -1
backend-cli.py
CHANGED
@@ -17,7 +17,7 @@ from src.backend.manage_requests import EvalRequest
|
|
17 |
from src.leaderboard.read_evals import EvalResult
|
18 |
|
19 |
from src.envs import QUEUE_REPO, RESULTS_REPO, API, DEBUG_QUEUE_REPO, DEBUG_RESULTS_REPO
|
20 |
-
from src.utils import my_snapshot_download, analyze_gpu_stats, parse_nvidia_smi, monitor_gpus
|
21 |
|
22 |
from src.leaderboard.read_evals import get_raw_eval_results
|
23 |
|
@@ -142,9 +142,6 @@ def request_to_result_name(request: EvalRequest) -> str:
|
|
142 |
def process_evaluation(task: Task, eval_request: EvalRequest, limit: Optional[int] = None) -> dict:
|
143 |
batch_size = 1
|
144 |
batch_size = eval_request.batch_size
|
145 |
-
|
146 |
-
if args.debug:
|
147 |
-
RESULTS_REPO = DEBUG_RESULTS_REPO
|
148 |
|
149 |
init_gpu_info = analyze_gpu_stats(parse_nvidia_smi())
|
150 |
# if init_gpu_info['Mem(M)'] > 500:
|
@@ -388,21 +385,7 @@ def maybe_refresh_results(thr: int, hard_task_lst: Optional[list[str]] = None) -
|
|
388 |
|
389 |
return False
|
390 |
|
391 |
-
|
392 |
-
def get_gpu_details():
|
393 |
-
gpus = GPUtil.getGPUs()
|
394 |
-
gpu = gpus[0]
|
395 |
-
name = gpu.name.replace(" ", "-")
|
396 |
-
# Convert memory from MB to GB and round to nearest whole number
|
397 |
-
memory_gb = round(gpu.memoryTotal / 1024)
|
398 |
-
memory = f"{memory_gb}GB"
|
399 |
-
formatted_name = f"{name}-{memory}"
|
400 |
-
return formatted_name
|
401 |
-
|
402 |
def process_pending_requests() -> bool:
|
403 |
-
if args.debug:
|
404 |
-
QUEUE_REPO = DEBUG_QUEUE_REPO
|
405 |
-
|
406 |
sanity_checks()
|
407 |
print("Processing pending requests")
|
408 |
current_pending_status = [PENDING_STATUS]
|
@@ -472,6 +455,7 @@ def get_args():
|
|
472 |
parser.add_argument("--limit", type=int, default=None, help="Limit for the number of samples")
|
473 |
parser.add_argument("--gpu-type", type=str, default="NVIDIA-A100-PCIe-80GB",
|
474 |
help="GPU type. NVIDIA-A100-PCIe-80GB; NVIDIA-RTX-A5000-24GB; NVIDIA-H100-PCIe-80GB")
|
|
|
475 |
return parser.parse_args()
|
476 |
|
477 |
|
@@ -479,7 +463,7 @@ if __name__ == "__main__":
|
|
479 |
args = get_args()
|
480 |
local_debug = args.debug
|
481 |
# debug specific task by ping
|
482 |
-
if local_debug:
|
483 |
# debug_model_names = [args.model] # Use model from arguments
|
484 |
# debug_task_name = [args.task] # Use task from arguments
|
485 |
debug_model_names = args.model.split(",")
|
@@ -510,25 +494,43 @@ if __name__ == "__main__":
|
|
510 |
results = process_evaluation(task, eval_request, limit=args.limit)
|
511 |
# except Exception as e:
|
512 |
# print(f"debug running error: {e}")
|
513 |
-
|
|
|
|
|
514 |
while True:
|
515 |
res = False
|
516 |
-
|
517 |
# if random.randint(0, 10) == 0:
|
518 |
res = process_pending_requests()
|
519 |
print(f"waiting for 60 seconds")
|
520 |
time.sleep(60)
|
521 |
-
|
522 |
# if res is False:
|
523 |
# if random.randint(0, 5) == 0:
|
524 |
# res = maybe_refresh_results(100)
|
525 |
# else:
|
526 |
# res = process_finished_requests(100)
|
527 |
-
|
528 |
# time.sleep(60)
|
529 |
-
|
530 |
# if res is False:
|
531 |
# if random.randint(0, 5) == 0:
|
532 |
# res = maybe_refresh_results(0)
|
533 |
# else:
|
534 |
# res = process_finished_requests(0)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
from src.leaderboard.read_evals import EvalResult
|
18 |
|
19 |
from src.envs import QUEUE_REPO, RESULTS_REPO, API, DEBUG_QUEUE_REPO, DEBUG_RESULTS_REPO
|
20 |
+
from src.utils import my_snapshot_download, analyze_gpu_stats, parse_nvidia_smi, monitor_gpus, get_gpu_details
|
21 |
|
22 |
from src.leaderboard.read_evals import get_raw_eval_results
|
23 |
|
|
|
142 |
def process_evaluation(task: Task, eval_request: EvalRequest, limit: Optional[int] = None) -> dict:
|
143 |
batch_size = 1
|
144 |
batch_size = eval_request.batch_size
|
|
|
|
|
|
|
145 |
|
146 |
init_gpu_info = analyze_gpu_stats(parse_nvidia_smi())
|
147 |
# if init_gpu_info['Mem(M)'] > 500:
|
|
|
385 |
|
386 |
return False
|
387 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
388 |
def process_pending_requests() -> bool:
|
|
|
|
|
|
|
389 |
sanity_checks()
|
390 |
print("Processing pending requests")
|
391 |
current_pending_status = [PENDING_STATUS]
|
|
|
455 |
parser.add_argument("--limit", type=int, default=None, help="Limit for the number of samples")
|
456 |
parser.add_argument("--gpu-type", type=str, default="NVIDIA-A100-PCIe-80GB",
|
457 |
help="GPU type. NVIDIA-A100-PCIe-80GB; NVIDIA-RTX-A5000-24GB; NVIDIA-H100-PCIe-80GB")
|
458 |
+
parser.add_argument("--debug_repo", action="store_true", help="Use debug repo")
|
459 |
return parser.parse_args()
|
460 |
|
461 |
|
|
|
463 |
args = get_args()
|
464 |
local_debug = args.debug
|
465 |
# debug specific task by ping
|
466 |
+
if local_debug and not args.debug_repo:
|
467 |
# debug_model_names = [args.model] # Use model from arguments
|
468 |
# debug_task_name = [args.task] # Use task from arguments
|
469 |
debug_model_names = args.model.split(",")
|
|
|
494 |
results = process_evaluation(task, eval_request, limit=args.limit)
|
495 |
# except Exception as e:
|
496 |
# print(f"debug running error: {e}")
|
497 |
+
elif local_debug and args.debug_repo:
|
498 |
+
QUEUE_REPO = DEBUG_QUEUE_REPO
|
499 |
+
RESULTS_REPO = DEBUG_RESULTS_REPO
|
500 |
while True:
|
501 |
res = False
|
|
|
502 |
# if random.randint(0, 10) == 0:
|
503 |
res = process_pending_requests()
|
504 |
print(f"waiting for 60 seconds")
|
505 |
time.sleep(60)
|
|
|
506 |
# if res is False:
|
507 |
# if random.randint(0, 5) == 0:
|
508 |
# res = maybe_refresh_results(100)
|
509 |
# else:
|
510 |
# res = process_finished_requests(100)
|
|
|
511 |
# time.sleep(60)
|
|
|
512 |
# if res is False:
|
513 |
# if random.randint(0, 5) == 0:
|
514 |
# res = maybe_refresh_results(0)
|
515 |
# else:
|
516 |
# res = process_finished_requests(0)
|
517 |
+
elif not local_debug and not args.debug_repo:
|
518 |
+
while True:
|
519 |
+
res = False
|
520 |
+
# if random.randint(0, 10) == 0:
|
521 |
+
res = process_pending_requests()
|
522 |
+
print(f"waiting for 60 seconds")
|
523 |
+
time.sleep(60)
|
524 |
+
# if res is False:
|
525 |
+
# if random.randint(0, 5) == 0:
|
526 |
+
# res = maybe_refresh_results(100)
|
527 |
+
# else:
|
528 |
+
# res = process_finished_requests(100)
|
529 |
+
# time.sleep(60)
|
530 |
+
# if res is False:
|
531 |
+
# if random.randint(0, 5) == 0:
|
532 |
+
# res = maybe_refresh_results(0)
|
533 |
+
# else:
|
534 |
+
# res = process_finished_requests(0)
|
535 |
+
else:
|
536 |
+
raise Exception("Cannot use debug_repo without local debug flag")
|
src/backend/envs.py
CHANGED
@@ -57,7 +57,7 @@ class Tasks(Enum):
|
|
57 |
|
58 |
# task20 = Task("race", "acc", "RACE", 0)
|
59 |
task21 = Task("mmlu", "acc", "MMLU", 5)
|
60 |
-
task22 = Task("gsm8k", "
|
61 |
|
62 |
|
63 |
EVAL_REQUESTS_PATH_BACKEND = os.path.join(CACHE_PATH, "eval-queue-bk")
|
|
|
57 |
|
58 |
# task20 = Task("race", "acc", "RACE", 0)
|
59 |
task21 = Task("mmlu", "acc", "MMLU", 5)
|
60 |
+
task22 = Task("gsm8k", "em", "GSM8K", 5)
|
61 |
|
62 |
|
63 |
EVAL_REQUESTS_PATH_BACKEND = os.path.join(CACHE_PATH, "eval-queue-bk")
|
src/display/utils.py
CHANGED
@@ -75,6 +75,7 @@ class Tasks(Enum):
|
|
75 |
# # XXX include me back at some point
|
76 |
selfcheck = Task("selfcheckgpt", "max-selfcheckgpt", "SelfCheckGPT")
|
77 |
mmlu = Task("mmlu", "acc", "MMLU") #MMLU/Acc (5-shot)
|
|
|
78 |
|
79 |
|
80 |
# These classes are for user facing column names,
|
|
|
75 |
# # XXX include me back at some point
|
76 |
selfcheck = Task("selfcheckgpt", "max-selfcheckgpt", "SelfCheckGPT")
|
77 |
mmlu = Task("mmlu", "acc", "MMLU") #MMLU/Acc (5-shot)
|
78 |
+
gsm8k = Task("gsm8k", "em", "GSM8K") #GSM8K/EM (5-shot)
|
79 |
|
80 |
|
81 |
# These classes are for user facing column names,
|
src/submission/check_validity.py
CHANGED
@@ -130,7 +130,8 @@ def already_submitted_models(requested_models_dir: str) -> set[str]:
|
|
130 |
continue
|
131 |
with open(os.path.join(root, file), "r") as f:
|
132 |
info = json.load(f)
|
133 |
-
|
|
|
134 |
|
135 |
# Select organisation
|
136 |
if info["model"].count("/") == 0 or "submitted_time" not in info:
|
|
|
130 |
continue
|
131 |
with open(os.path.join(root, file), "r") as f:
|
132 |
info = json.load(f)
|
133 |
+
if not info["status"] == "FINISHED" and not info["status"] == "RUNNING":
|
134 |
+
file_names.append(f"{info['model']}_{info['revision']}_{info['precision']}_{info['inference_framework']}_{info['gpu_type']}")
|
135 |
|
136 |
# Select organisation
|
137 |
if info["model"].count("/") == 0 or "submitted_time" not in info:
|
src/utils.py
CHANGED
@@ -3,12 +3,48 @@ from huggingface_hub import snapshot_download
|
|
3 |
import subprocess
|
4 |
import re
|
5 |
import os
|
|
|
6 |
|
7 |
try:
|
8 |
from src.display.utils import GPU_TEMP, GPU_Mem, GPU_Power, GPU_Util, GPU_Name
|
9 |
except:
|
10 |
print("local debug: from display.utils")
|
11 |
from display.utils import GPU_TEMP, GPU_Mem, GPU_Power, GPU_Util, GPU_Name
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
def my_snapshot_download(repo_id, revision, local_dir, repo_type, max_workers):
|
14 |
for i in range(10):
|
@@ -56,7 +92,7 @@ def parse_nvidia_smi():
|
|
56 |
gpu_stats = []
|
57 |
|
58 |
gpu_info_pattern = re.compile(r'(\d+)C\s+P\d+\s+(\d+)W / \d+W\s+\|\s+(\d+)MiB / \d+MiB\s+\|\s+(\d+)%')
|
59 |
-
gpu_name_pattern = re.compile(r'NVIDIA\s+([\w\s]
|
60 |
|
61 |
gpu_name = ""
|
62 |
for index in gpu_indices:
|
@@ -131,5 +167,70 @@ def analyze_gpu_stats(stats_list):
|
|
131 |
|
132 |
return avg_stats
|
133 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
134 |
if __name__ == "__main__":
|
135 |
print(analyze_gpu_stats(parse_nvidia_smi()))
|
|
|
3 |
import subprocess
|
4 |
import re
|
5 |
import os
|
6 |
+
import GPUtil
|
7 |
|
8 |
try:
|
9 |
from src.display.utils import GPU_TEMP, GPU_Mem, GPU_Power, GPU_Util, GPU_Name
|
10 |
except:
|
11 |
print("local debug: from display.utils")
|
12 |
from display.utils import GPU_TEMP, GPU_Mem, GPU_Power, GPU_Util, GPU_Name
|
13 |
+
|
14 |
+
MEM_BW_DICT ={
|
15 |
+
"NVIDIA-A100-PCIe-80GB": 1935,
|
16 |
+
"NVIDIA-A100-SXM-80GB": 2039,
|
17 |
+
"NVIDIA-H100-PCIe-80GB": 2039,
|
18 |
+
"NVIDIA-RTX-A5000-24GB": 768
|
19 |
+
}
|
20 |
+
|
21 |
+
PEAK_FLOPS_DICT = {
|
22 |
+
"float32":{
|
23 |
+
"NVIDIA-A100-PCIe-80GB": 312e12,
|
24 |
+
"NVIDIA-A100-SXM-80GB": 312e12,
|
25 |
+
"NVIDIA-H100-PCIe-80GB": 756e12,
|
26 |
+
"NVIDIA-RTX-A5000-24GB": 222.2e12
|
27 |
+
},
|
28 |
+
"float16":{
|
29 |
+
"NVIDIA-A100-PCIe-80GB": 624e12,
|
30 |
+
"NVIDIA-A100-SXM-80GB": 624e12,
|
31 |
+
"NVIDIA-H100-PCIe-80GB": 1513e12,
|
32 |
+
"NVIDIA-RTX-A5000-24GB": 444.4e12
|
33 |
+
},
|
34 |
+
"8bit":{
|
35 |
+
"NVIDIA-A100-PCIe-80GB": 1248e12,
|
36 |
+
"NVIDIA-A100-SXM-80GB": 1248e12,
|
37 |
+
"NVIDIA-H100-PCIe-80GB": 3026e12,
|
38 |
+
"NVIDIA-RTX-A5000-24GB": 889e12
|
39 |
+
},
|
40 |
+
"4bit": {
|
41 |
+
"NVIDIA-A100-PCIe-80GB": 2496e12,
|
42 |
+
"NVIDIA-A100-SXM-80GB": 2496e12,
|
43 |
+
"NVIDIA-H100-PCIe-80GB": 6052e12,
|
44 |
+
"NVIDIA-RTX-A5000-24GB": 1778e12
|
45 |
+
}
|
46 |
+
|
47 |
+
}
|
48 |
|
49 |
def my_snapshot_download(repo_id, revision, local_dir, repo_type, max_workers):
|
50 |
for i in range(10):
|
|
|
92 |
gpu_stats = []
|
93 |
|
94 |
gpu_info_pattern = re.compile(r'(\d+)C\s+P\d+\s+(\d+)W / \d+W\s+\|\s+(\d+)MiB / \d+MiB\s+\|\s+(\d+)%')
|
95 |
+
gpu_name_pattern = re.compile(r'NVIDIA\s+([\w\s]+\d+(?:\s*GB)?)')
|
96 |
|
97 |
gpu_name = ""
|
98 |
for index in gpu_indices:
|
|
|
167 |
|
168 |
return avg_stats
|
169 |
|
170 |
+
def get_gpu_number():
|
171 |
+
visible_devices = os.getenv('CUDA_VISIBLE_DEVICES', None)
|
172 |
+
if visible_devices is not None:
|
173 |
+
gpu_indices = visible_devices.split(',')
|
174 |
+
else:
|
175 |
+
# Query all GPU indices if CUDA_VISIBLE_DEVICES is not set
|
176 |
+
result = subprocess.run(['nvidia-smi', '--query-gpu=index', '--format=csv,noheader'], capture_output=True, text=True)
|
177 |
+
if result.returncode != 0:
|
178 |
+
print("Failed to query GPU indices.")
|
179 |
+
return []
|
180 |
+
gpu_indices = result.stdout.strip().split('\n')
|
181 |
+
# print(f"gpu_indices: {gpu_indices}")
|
182 |
+
gpu_stats = []
|
183 |
+
|
184 |
+
gpu_info_pattern = re.compile(r'(\d+)C\s+P\d+\s+(\d+)W / \d+W\s+\|\s+(\d+)MiB / \d+MiB\s+\|\s+(\d+)%')
|
185 |
+
|
186 |
+
for index in gpu_indices:
|
187 |
+
result = subprocess.run(['nvidia-smi', '-i', index], capture_output=True, text=True)
|
188 |
+
output = result.stdout.strip()
|
189 |
+
lines = output.split("\n")
|
190 |
+
for line in lines:
|
191 |
+
match = gpu_info_pattern.search(line)
|
192 |
+
gpu_info = {}
|
193 |
+
if match:
|
194 |
+
temp, power_usage, mem_usage, gpu_util = map(int, match.groups())
|
195 |
+
gpu_info.update({
|
196 |
+
GPU_TEMP: temp,
|
197 |
+
GPU_Power: power_usage,
|
198 |
+
GPU_Mem: round(mem_usage / 1024, 2),
|
199 |
+
GPU_Util: gpu_util
|
200 |
+
})
|
201 |
+
|
202 |
+
if len(gpu_info) >= 4:
|
203 |
+
gpu_stats.append(gpu_info)
|
204 |
+
|
205 |
+
return len(gpu_stats)
|
206 |
+
|
207 |
+
def get_gpu_details():
|
208 |
+
gpus = GPUtil.getGPUs()
|
209 |
+
gpu = gpus[0]
|
210 |
+
name = gpu.name.replace(" ", "-")
|
211 |
+
# Convert memory from MB to GB and round to nearest whole number
|
212 |
+
memory_gb = round(gpu.memoryTotal / 1024)
|
213 |
+
memory = f"{memory_gb}GB"
|
214 |
+
formatted_name = f"{name}-{memory}"
|
215 |
+
return formatted_name
|
216 |
+
|
217 |
+
def get_peak_bw(gpu_name):
|
218 |
+
return MEM_BW_DICT[gpu_name]
|
219 |
+
|
220 |
+
def get_peak_flops(gpu_name, precision):
|
221 |
+
return PEAK_FLOPS_DICT[precision][gpu_name]
|
222 |
+
|
223 |
+
def transfer_precision2bytes(precision):
|
224 |
+
if precision == "float32":
|
225 |
+
return 4
|
226 |
+
elif precision == "float16":
|
227 |
+
return 2
|
228 |
+
elif precision == "8bit":
|
229 |
+
return 1
|
230 |
+
elif precision == "4bit":
|
231 |
+
return 0.5
|
232 |
+
else:
|
233 |
+
raise ValueError(f"Unsupported precision: {precision}")
|
234 |
+
|
235 |
if __name__ == "__main__":
|
236 |
print(analyze_gpu_stats(parse_nvidia_smi()))
|