la-llama-que-llama / modules /exllama /doc /model_compatibility.md
juanma1907's picture
Upload folder using huggingface_hub
462dacf

A newer version of the Gradio SDK is available: 5.6.0

Upgrade

Working models

As of 2023-07-19, the following GPTQ models on HuggingFace all appear to be working:

  • iambestfeed/open_llama_3b_4bit_128g
  • Neko-Institute-of-Science/LLaMA-7B-4bit-128g
  • Neko-Institute-of-Science/LLaMA-13B-4bit-128g
  • Neko-Institute-of-Science/LLaMA-30B-4bit-32g
  • Neko-Institute-of-Science/LLaMA-30B-4bit-128g
  • Neko-Institute-of-Science/LLaMA-65B-4bit-32g
  • Neko-Institute-of-Science/LLaMA-65B-4bit-128g
  • Panchovix/LLaMA-2-70B-GPTQ-transformers4.32.0.dev0
  • reeducator/bluemoonrp-13b
  • reeducator/bluemoonrp-30b
  • TehVenom/Metharme-13b-4bit-GPTQ
  • TheBloke/airoboros-13B-GPTQ
  • TheBloke/gpt4-x-vicuna-13B-GPTQ
  • TheBloke/GPT4All-13B-snoozy-GPTQ
  • TheBloke/guanaco-33B-GPTQ
  • TheBloke/guanaco-65B-GPTQ
  • TheBloke/h2ogpt-oasst1-512-30B-GPTQ
  • TheBloke/koala-13B-GPTQ-4bit-128g
  • TheBloke/Llama-2-13B-chat-GPTQ (128g)
  • TheBloke/Llama-2-13B-GPTQ (32g, 64g, 128g)
  • TheBloke/Llama-2-70B-GPTQ (32g, 128g)
  • TheBloke/Manticore-13B-GPTQ
  • TheBloke/medalpaca-13B-GPTQ-4bit
  • TheBloke/medalpaca-13B-GPTQ-4bit (compat version)
  • TheBloke/Nous-Hermes-13B-GPTQ
  • TheBloke/robin-65B-v2-GPTQ
  • TheBloke/tulu-7B-GPTQ
  • TheBloke/Tulu-13B-SuperHOT-8K-GPTQ
  • TheBloke/tulu-30B-GPTQ
  • TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g
  • TheBloke/VicUnlocked-30B-LoRA-GPTQ
  • TheBloke/wizard-mega-13B-GPTQ
  • TheBloke/Wizard-Vicuna-7B-Uncensored-GPTQ
  • TheBloke/Wizard-Vicuna-13B-Uncensored-GPTQ
  • TheBloke/WizardLM-7B-uncensored-GPTQ
  • TheBloke/WizardLM-30B-Uncensored-GPTQ
  • TheBloke/WizardLM-33B-V1.0-Uncensored-SuperHOT-8K-GPTQ
  • tmpupload/superhot-30b-8k-no-rlhf-test-128g-GPTQ
  • Yhyu13/chimera-inst-chat-13b-gptq-4bit
  • Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-gptq-4bit

Non-working models

None as of 2023-07-19.