File size: 2,206 Bytes
8ea9d23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
import gradio as gr
import torch
from diffusers import AnimateDiffPipeline, MotionAdapter, DDIMScheduler
from diffusers.utils import export_to_gif
import random

def generate_gif(image, animation_type):
    # Load the motion adapter
    adapter = MotionAdapter.from_pretrained("guoyww/animatediff-motion-adapter-v1-5-2", torch_dtype=torch.float16)

    # Load SD 1.5 based finetuned model
    model_id = "SG161222/Realistic_Vision_V6.0_B1_noVAE"
    pipe = AnimateDiffPipeline.from_pretrained(model_id, motion_adapter=adapter, torch_dtype=torch.float16)

    # Scheduler setup
    scheduler = DDIMScheduler(
        clip_sample=False,
        beta_start=0.00085,
        beta_end=0.012,
        beta_schedule="linear",
        timestep_spacing="trailing",
        steps_offset=1
    )
    pipe.scheduler = scheduler

    # Enable memory savings
    pipe.enable_vae_slicing()
    pipe.enable_model_cpu_offload()

    # Load ip_adapter
    pipe.load_ip_adapter("h94/IP-Adapter", subfolder="models", weight_name="ip-adapter_sd15.bin")

    # Load the selected motion adapter
    pipe.load_lora_weights(f"guoyww/animatediff-motion-lora-{animation_type}", adapter_name=animation_type)

    # Generate a random seed
    seed = random.randint(0, 2**32 - 1)
    prompt = "best quality, high quality, trending on artstation"

    # Set adapter weights for the selected adapter
    adapter_weight = [0.75]

    pipe.set_adapters([animation_type], adapter_weights=adapter_weight)

    # Generate GIF
    output = pipe(
        prompt=prompt,
        num_frames=16,
        guidance_scale=7.5,
        num_inference_steps=30,
        ip_adapter_image=image,
        generator=torch.Generator("cpu").manual_seed(seed),
    )
    frames = output.frames[0]

    gif_path = "output_animation.gif"
    export_to_gif(frames, gif_path)
    return gif_path

# Gradio interface
iface = gr.Interface(
    fn=generate_gif,
    inputs=[gr.Image(type="pil"), gr.Radio(["zoom-out", "tilt-up", "pan-left"])],
    outputs=gr.Image(type="pil", label="Generated GIF"),
    title="AnimateDiff + IP Adapter Demo",
    description="Upload an image and select an motion module type to generate a GIF!"
)

iface.launch(debug=True,share=True)