add gpu info
Browse files- backend-cli.py +2 -2
backend-cli.py
CHANGED
@@ -435,12 +435,12 @@ if __name__ == "__main__":
|
|
435 |
if local_debug:
|
436 |
# debug_model_names = [args.model] # Use model from arguments
|
437 |
# debug_task_name = [args.task] # Use task from arguments
|
438 |
-
debug_model_names = ["
|
439 |
"databricks/dbrx-instruct", "databricks/dbrx-base",
|
440 |
"mistralai/Mixtral-8x22B-v0.1", "mistralai/Mixtral-8x22B-Instruct-v0.1", "alpindale/WizardLM-2-8x22B",
|
441 |
"CohereForAI/c4ai-command-r-plus"] # Use model from arguments
|
442 |
debug_task_name = ['mmlu', 'selfcheckgpt'] # Use task from arguments
|
443 |
-
precisions = ['4bit', '
|
444 |
task_lst = TASKS_HARNESS.copy()
|
445 |
for precision in precisions:
|
446 |
for debug_model_name in debug_model_names:
|
|
|
435 |
if local_debug:
|
436 |
# debug_model_names = [args.model] # Use model from arguments
|
437 |
# debug_task_name = [args.task] # Use task from arguments
|
438 |
+
debug_model_names = ["mistralai/Mixtral-8x7B-Instruct-v0.1", "mistralai/Mixtral-8x7B-v0.1",
|
439 |
"databricks/dbrx-instruct", "databricks/dbrx-base",
|
440 |
"mistralai/Mixtral-8x22B-v0.1", "mistralai/Mixtral-8x22B-Instruct-v0.1", "alpindale/WizardLM-2-8x22B",
|
441 |
"CohereForAI/c4ai-command-r-plus"] # Use model from arguments
|
442 |
debug_task_name = ['mmlu', 'selfcheckgpt'] # Use task from arguments
|
443 |
+
precisions = ['4bit', 'float32', 'float16', '8bit']
|
444 |
task_lst = TASKS_HARNESS.copy()
|
445 |
for precision in precisions:
|
446 |
for debug_model_name in debug_model_names:
|