This is a development model meant to help test the HunyuanVideoPipeline integration to diffusers. Please help out if you can on the PR.

pip install -qq git+https://github.com/huggingface/diffusers.git@hunyuan-video
import torch
pipe = HunyuanVideoPipeline.from_pretrained(
    "magespace/hyvideo-diffusers-dev",
    torch_dtype=torch.bfloat16,
)

Post-processing:

import PIL.Image
from diffusers.utils import export_to_video

export_to_video(result.frames[0], "output.mp4", fps=24)

For faster generation, you can optimize the transformer with torch.compile. Additionally, increasing shift in the scheduler can allow for lower step values as shown in the original paper.

Generation time is quadratic with the number of pixels, so reducing the height and width and decreasing the number of frames will drastically speed up generation at the price of video quality.

Downloads last month
3
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model’s pipeline type.