Spaces:
Sleeping
Sleeping
meg-huggingface
commited on
Commit
•
3b2722d
1
Parent(s):
971bce4
Adding CPU (float32) support
Browse files
scripts/create_request_file.py
CHANGED
@@ -12,7 +12,7 @@ from src.envs import TOKEN
|
|
12 |
EVAL_REQUESTS_PATH = "eval-queue"
|
13 |
QUEUE_REPO = "open-llm-leaderboard/requests"
|
14 |
|
15 |
-
precisions = ("float16", "bfloat16", "8bit (LLM.int8)", "4bit (QLoRA / FP4)", "GPTQ")
|
16 |
model_types = ("pretrained", "fine-tuned", "RL-tuned", "instruction-tuned")
|
17 |
weight_types = ("Original", "Delta", "Adapter")
|
18 |
|
|
|
12 |
EVAL_REQUESTS_PATH = "eval-queue"
|
13 |
QUEUE_REPO = "open-llm-leaderboard/requests"
|
14 |
|
15 |
+
precisions = ("float16", "bfloat16", "8bit (LLM.int8)", "4bit (QLoRA / FP4)", "GPTQ", "float32")
|
16 |
model_types = ("pretrained", "fine-tuned", "RL-tuned", "instruction-tuned")
|
17 |
weight_types = ("Original", "Delta", "Adapter")
|
18 |
|
src/backend/manage_requests.py
CHANGED
@@ -26,7 +26,7 @@ class EvalRequest:
|
|
26 |
def get_model_args(self):
|
27 |
model_args = f"pretrained={self.model},revision={self.revision}"
|
28 |
|
29 |
-
if self.precision in ["float16", "bfloat16"]:
|
30 |
model_args += f",dtype={self.precision}"
|
31 |
# Quantized models need some added config, the install of bits and bytes, etc
|
32 |
#elif self.precision == "8bit":
|
|
|
26 |
def get_model_args(self):
|
27 |
model_args = f"pretrained={self.model},revision={self.revision}"
|
28 |
|
29 |
+
if self.precision in ["float16", "bfloat16", "float32"]:
|
30 |
model_args += f",dtype={self.precision}"
|
31 |
# Quantized models need some added config, the install of bits and bytes, etc
|
32 |
#elif self.precision == "8bit":
|