--- license: apache-2.0 language: - en pipeline_tag: text-to-video tags: - art - code --- # RCNA MINI **RCNA MINI** is a compact **LoRA** (Low-Rank Adaptation) model designed for generating high-quality, 4-step text-to-video outputs. It can create video clips ranging from 4 to 16 seconds long, making it ideal for generating short animations with rich details and smooth transitions. ## Key Features: - **4-step Text-to-Video**: Generates videos from a text prompt in just 4 steps. - **Video Length**: Can generate videos from 4 seconds to 16 seconds long. - **High Quality**: Supports high-resolution and detailed outputs (up to 8K). - **Fast Sampling**: Leveraging decoupled consistency learning, the model is optimized for speed while maintaining quality. ## Example Outputs: - **Prompt**: "Astronaut in a jungle, cold color palette, muted colors, detailed, 8K" - Generates a high-quality video with rich details and smooth motion. ## How it Works: RCNA MINI is based on the LoRA architecture, which fine-tunes diffusion models using low-rank adaptations. This results in faster generation and less computational overhead compared to full model retraining. ## Applications: - Short-form animations for social media content - Video generation for creative projects - Artistic video generation based on textual descriptions ## Model Details: - **Architecture**: LoRA applied to diffusion models - **Inference Steps**: 4-step generation - **Output Length**: 4 to 16 seconds - ## Using AnimateLCM with Diffusers ```python import torch from diffusers import AnimateDiffPipeline, LCMScheduler, MotionAdapter, DiffusionPipeline from diffusers.utils import export_to_gif # Load AnimateLCM for video generation adapter = MotionAdapter.from_pretrained("Binarybardakshat/RCNA_MINI") pipe = AnimateDiffPipeline.from_pretrained("emilianJR/epiCRealism", motion_adapter=adapter, torch_dtype=torch.float16) pipe.scheduler = LCMScheduler.from_config(pipe.scheduler.config, beta_schedule="linear") pipe.load_lora_weights("Binarybardakshat/RCNA_MINI", weight_name="RCNA_LORA_MINI_1.safetensors", adapter_name="lcm-lora") pipe.set_adapters(["lcm-lora"], [0.8]) pipe.enable_vae_slicing() pipe.enable_model_cpu_offload() # Generate video using RCNA MINI output = pipe( prompt="A space rocket with trails of smoke behind it launching into space from the desert, 4k, high resolution", negative_prompt="bad quality, worse quality, low resolution", num_frames=16, guidance_scale=2.0, num_inference_steps=6, generator=torch.Generator("cpu").manual_seed(0), ) frames = output.frames[0] export_to_gif(frames, "animatelcm.gif") print("Video and image generation complete!") ``` ## License: This model is licensed under the [MIT License](LICENSE).