meg-huggingface commited on
Commit
3b2722d
1 Parent(s): 971bce4

Adding CPU (float32) support

Browse files
scripts/create_request_file.py CHANGED
@@ -12,7 +12,7 @@ from src.envs import TOKEN
12
  EVAL_REQUESTS_PATH = "eval-queue"
13
  QUEUE_REPO = "open-llm-leaderboard/requests"
14
 
15
- precisions = ("float16", "bfloat16", "8bit (LLM.int8)", "4bit (QLoRA / FP4)", "GPTQ")
16
  model_types = ("pretrained", "fine-tuned", "RL-tuned", "instruction-tuned")
17
  weight_types = ("Original", "Delta", "Adapter")
18
 
 
12
  EVAL_REQUESTS_PATH = "eval-queue"
13
  QUEUE_REPO = "open-llm-leaderboard/requests"
14
 
15
+ precisions = ("float16", "bfloat16", "8bit (LLM.int8)", "4bit (QLoRA / FP4)", "GPTQ", "float32")
16
  model_types = ("pretrained", "fine-tuned", "RL-tuned", "instruction-tuned")
17
  weight_types = ("Original", "Delta", "Adapter")
18
 
src/backend/manage_requests.py CHANGED
@@ -26,7 +26,7 @@ class EvalRequest:
26
  def get_model_args(self):
27
  model_args = f"pretrained={self.model},revision={self.revision}"
28
 
29
- if self.precision in ["float16", "bfloat16"]:
30
  model_args += f",dtype={self.precision}"
31
  # Quantized models need some added config, the install of bits and bytes, etc
32
  #elif self.precision == "8bit":
 
26
  def get_model_args(self):
27
  model_args = f"pretrained={self.model},revision={self.revision}"
28
 
29
+ if self.precision in ["float16", "bfloat16", "float32"]:
30
  model_args += f",dtype={self.precision}"
31
  # Quantized models need some added config, the install of bits and bytes, etc
32
  #elif self.precision == "8bit":