gguf quantized version of t5xxl encoder with mochi (test pack revised)

screenshot

setup (once)

  • drag mochi_fp8_e4m3fn.safetensors (10GB) to > ./ComfyUI/models/diffusion_models
  • drag t5xxl_fp16-q4_0.gguf (2.9GB) to > ./ComfyUI/models/text_encoders
  • drag mochi_vae_fp8_e4m3fn.safetensors (460MB) to > ./ComfyUI/models/vae

run it straight (no installation needed way)

  • run the .bat file in the main directory (assuming you are using the gguf-node pack below)
  • drag the workflow json file (below) to > your browser

workflow

review

  • revised workflow to bypass oom issue and around 50% faster with the new fp8_e4m3fn file
  • t5xxl gguf works fine as text encoder
  • model gguf file might not work; if so, please wait for the code update

reference

prompt test#

screenshot prompt: "a fox moving quickly in a beautiful winter scenery nature trees sunset tracking camera"

Prompt
a fox moving quickly in a beautiful winter scenery nature trees sunset tracking camera
Prompt
same prompt as 1st one <metadata inside>
Prompt
same prompt as 1st one; but with new workflow to bypass oom <metadata inside>
Downloads last month
292
GGUF
Model size
10B params
Architecture
mochi

4-bit

8-bit

16-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for calcuis/mochi

Quantized
(3)
this model