Edit model card

🎥 CogvideoX-5b LoRa to control camera movement

Usage

The LoRa was trained to control camera movement in 6 directions: left, right, up, down, zoom_in, zoom_out. Start prompt with text like this:

'Сamera moves to the {}...',
'Сamera is moving to the {}...',
'{} camera movement...',
'{} camera turn...',

Inference examples

ComfyUI example

image/png JSON File Example

Minimal code example

import torch
from diffusers import CogVideoXImageToVideoPipeline
from diffusers.utils import export_to_video, load_image


pipe = CogVideoXImageToVideoPipeline.from_pretrained(
    "THUDM/CogVideoX1.5-5B-I2V", torch_dtype=torch.bfloat16
)

pipe.load_lora_weights("NimVideo/cogvideox1.5-5b-prompt-camera-motion", adapter_name="cogvideox-lora")
pipe.set_adapters(["cogvideox-lora"], [1.0])

pipe.enable_sequential_cpu_offload()
pipe.vae.enable_slicing()
pipe.vae.enable_tiling()

height = 768 
width = 1360
image = load_image("resources/car.jpg").resize((width, height))
prompt = "Camera is moving to the left. A red sports car driving on a winding road."

video_generate = pipe(
    image=image,
    prompt=prompt,
    height=height, 
    width=width, 
    num_inference_steps=50,  
    num_frames=81,  
    guidance_scale=6.0,
    generator=torch.Generator().manual_seed(42), 
).frames[0]

export_to_video(video_generate, output_path, fps=8)

Inference with cli and jupyter-notebook examlple you can find on our Github

Acknowledgements

Original code and models CogVideoX.

Contacts

Issues should be raised directly in the repository.

Downloads last month
79
Inference API
Unable to determine this model’s pipeline type. Check the docs .