|
- sections: |
|
- local: index |
|
title: 🧨 Diffusers |
|
- local: quicktour |
|
title: Quicktour |
|
- local: stable_diffusion |
|
title: Effective and efficient diffusion |
|
- local: installation |
|
title: Installation |
|
title: Get started |
|
- sections: |
|
- local: tutorials/tutorial_overview |
|
title: Overview |
|
- local: using-diffusers/write_own_pipeline |
|
title: Understanding pipelines, models and schedulers |
|
- local: tutorials/autopipeline |
|
title: AutoPipeline |
|
- local: tutorials/basic_training |
|
title: Train a diffusion model |
|
- local: tutorials/using_peft_for_inference |
|
title: Load LoRAs for inference |
|
- local: tutorials/fast_diffusion |
|
title: Accelerate inference of text-to-image diffusion models |
|
title: Tutorials |
|
- sections: |
|
- local: using-diffusers/loading |
|
title: Load pipelines |
|
- local: using-diffusers/custom_pipeline_overview |
|
title: Load community pipelines and components |
|
- local: using-diffusers/schedulers |
|
title: Load schedulers and models |
|
- local: using-diffusers/other-formats |
|
title: Model files and layouts |
|
- local: using-diffusers/loading_adapters |
|
title: Load adapters |
|
- local: using-diffusers/push_to_hub |
|
title: Push files to the Hub |
|
title: Load pipelines and adapters |
|
- sections: |
|
- local: using-diffusers/unconditional_image_generation |
|
title: Unconditional image generation |
|
- local: using-diffusers/conditional_image_generation |
|
title: Text-to-image |
|
- local: using-diffusers/img2img |
|
title: Image-to-image |
|
- local: using-diffusers/inpaint |
|
title: Inpainting |
|
- local: using-diffusers/text-img2vid |
|
title: Text or image-to-video |
|
- local: using-diffusers/depth2img |
|
title: Depth-to-image |
|
title: Generative tasks |
|
- sections: |
|
- local: using-diffusers/overview_techniques |
|
title: Overview |
|
- local: training/distributed_inference |
|
title: Distributed inference with multiple GPUs |
|
- local: using-diffusers/merge_loras |
|
title: Merge LoRAs |
|
- local: using-diffusers/scheduler_features |
|
title: Scheduler features |
|
- local: using-diffusers/callback |
|
title: Pipeline callbacks |
|
- local: using-diffusers/reusing_seeds |
|
title: Reproducible pipelines |
|
- local: using-diffusers/image_quality |
|
title: Controlling image quality |
|
- local: using-diffusers/weighted_prompts |
|
title: Prompt techniques |
|
title: Inference techniques |
|
- sections: |
|
- local: advanced_inference/outpaint |
|
title: Outpainting |
|
title: Advanced inference |
|
- sections: |
|
- local: using-diffusers/sdxl |
|
title: Stable Diffusion XL |
|
- local: using-diffusers/sdxl_turbo |
|
title: SDXL Turbo |
|
- local: using-diffusers/kandinsky |
|
title: Kandinsky |
|
- local: using-diffusers/ip_adapter |
|
title: IP-Adapter |
|
- local: using-diffusers/controlnet |
|
title: ControlNet |
|
- local: using-diffusers/t2i_adapter |
|
title: T2I-Adapter |
|
- local: using-diffusers/inference_with_lcm |
|
title: Latent Consistency Model |
|
- local: using-diffusers/textual_inversion_inference |
|
title: Textual inversion |
|
- local: using-diffusers/shap-e |
|
title: Shap-E |
|
- local: using-diffusers/diffedit |
|
title: DiffEdit |
|
- local: using-diffusers/inference_with_tcd_lora |
|
title: Trajectory Consistency Distillation-LoRA |
|
- local: using-diffusers/svd |
|
title: Stable Video Diffusion |
|
- local: using-diffusers/marigold_usage |
|
title: Marigold Computer Vision |
|
title: Specific pipeline examples |
|
- sections: |
|
- local: training/overview |
|
title: Overview |
|
- local: training/create_dataset |
|
title: Create a dataset for training |
|
- local: training/adapt_a_model |
|
title: Adapt a model to a new task |
|
- isExpanded: false |
|
sections: |
|
- local: training/unconditional_training |
|
title: Unconditional image generation |
|
- local: training/text2image |
|
title: Text-to-image |
|
- local: training/sdxl |
|
title: Stable Diffusion XL |
|
- local: training/kandinsky |
|
title: Kandinsky 2.2 |
|
- local: training/wuerstchen |
|
title: Wuerstchen |
|
- local: training/controlnet |
|
title: ControlNet |
|
- local: training/t2i_adapters |
|
title: T2I-Adapters |
|
- local: training/instructpix2pix |
|
title: InstructPix2Pix |
|
title: Models |
|
- isExpanded: false |
|
sections: |
|
- local: training/text_inversion |
|
title: Textual Inversion |
|
- local: training/dreambooth |
|
title: DreamBooth |
|
- local: training/lora |
|
title: LoRA |
|
- local: training/custom_diffusion |
|
title: Custom Diffusion |
|
- local: training/lcm_distill |
|
title: Latent Consistency Distillation |
|
- local: training/ddpo |
|
title: Reinforcement learning training with DDPO |
|
title: Methods |
|
title: Training |
|
- sections: |
|
- local: optimization/fp16 |
|
title: Speed up inference |
|
- local: optimization/memory |
|
title: Reduce memory usage |
|
- local: optimization/torch2.0 |
|
title: PyTorch 2.0 |
|
- local: optimization/xformers |
|
title: xFormers |
|
- local: optimization/tome |
|
title: Token merging |
|
- local: optimization/deepcache |
|
title: DeepCache |
|
- local: optimization/tgate |
|
title: TGATE |
|
- sections: |
|
- local: using-diffusers/stable_diffusion_jax_how_to |
|
title: JAX/Flax |
|
- local: optimization/onnx |
|
title: ONNX |
|
- local: optimization/open_vino |
|
title: OpenVINO |
|
- local: optimization/coreml |
|
title: Core ML |
|
title: Optimized model formats |
|
- sections: |
|
- local: optimization/mps |
|
title: Metal Performance Shaders (MPS) |
|
- local: optimization/habana |
|
title: Habana Gaudi |
|
title: Optimized hardware |
|
title: Accelerate inference and reduce memory |
|
- sections: |
|
- local: conceptual/philosophy |
|
title: Philosophy |
|
- local: using-diffusers/controlling_generation |
|
title: Controlled generation |
|
- local: conceptual/contribution |
|
title: How to contribute? |
|
- local: conceptual/ethical_guidelines |
|
title: Diffusers' Ethical Guidelines |
|
- local: conceptual/evaluation |
|
title: Evaluating Diffusion Models |
|
title: Conceptual Guides |
|
- sections: |
|
- isExpanded: false |
|
sections: |
|
- local: api/configuration |
|
title: Configuration |
|
- local: api/logging |
|
title: Logging |
|
- local: api/outputs |
|
title: Outputs |
|
title: Main Classes |
|
- isExpanded: false |
|
sections: |
|
- local: api/loaders/ip_adapter |
|
title: IP-Adapter |
|
- local: api/loaders/lora |
|
title: LoRA |
|
- local: api/loaders/single_file |
|
title: Single files |
|
- local: api/loaders/textual_inversion |
|
title: Textual Inversion |
|
- local: api/loaders/unet |
|
title: UNet |
|
- local: api/loaders/peft |
|
title: PEFT |
|
title: Loaders |
|
- isExpanded: false |
|
sections: |
|
- local: api/models/overview |
|
title: Overview |
|
- local: api/models/unet |
|
title: UNet1DModel |
|
- local: api/models/unet2d |
|
title: UNet2DModel |
|
- local: api/models/unet2d-cond |
|
title: UNet2DConditionModel |
|
- local: api/models/unet3d-cond |
|
title: UNet3DConditionModel |
|
- local: api/models/unet-motion |
|
title: UNetMotionModel |
|
- local: api/models/uvit2d |
|
title: UViT2DModel |
|
- local: api/models/vq |
|
title: VQModel |
|
- local: api/models/autoencoderkl |
|
title: AutoencoderKL |
|
- local: api/models/asymmetricautoencoderkl |
|
title: AsymmetricAutoencoderKL |
|
- local: api/models/autoencoder_tiny |
|
title: Tiny AutoEncoder |
|
- local: api/models/consistency_decoder_vae |
|
title: ConsistencyDecoderVAE |
|
- local: api/models/transformer2d |
|
title: Transformer2DModel |
|
- local: api/models/pixart_transformer2d |
|
title: PixArtTransformer2DModel |
|
- local: api/models/dit_transformer2d |
|
title: DiTTransformer2DModel |
|
- local: api/models/hunyuan_transformer2d |
|
title: HunyuanDiT2DModel |
|
- local: api/models/transformer_temporal |
|
title: TransformerTemporalModel |
|
- local: api/models/sd3_transformer2d |
|
title: SD3Transformer2DModel |
|
- local: api/models/prior_transformer |
|
title: PriorTransformer |
|
- local: api/models/controlnet |
|
title: ControlNetModel |
|
- local: api/models/controlnet_sd3 |
|
title: SD3ControlNetModel |
|
title: Models |
|
- isExpanded: false |
|
sections: |
|
- local: api/pipelines/overview |
|
title: Overview |
|
- local: api/pipelines/amused |
|
title: aMUSEd |
|
- local: api/pipelines/animatediff |
|
title: AnimateDiff |
|
- local: api/pipelines/attend_and_excite |
|
title: Attend-and-Excite |
|
- local: api/pipelines/audioldm |
|
title: AudioLDM |
|
- local: api/pipelines/audioldm2 |
|
title: AudioLDM 2 |
|
- local: api/pipelines/auto_pipeline |
|
title: AutoPipeline |
|
- local: api/pipelines/blip_diffusion |
|
title: BLIP-Diffusion |
|
- local: api/pipelines/consistency_models |
|
title: Consistency Models |
|
- local: api/pipelines/controlnet |
|
title: ControlNet |
|
- local: api/pipelines/controlnet_sd3 |
|
title: ControlNet with Stable Diffusion 3 |
|
- local: api/pipelines/controlnet_sdxl |
|
title: ControlNet with Stable Diffusion XL |
|
- local: api/pipelines/controlnetxs |
|
title: ControlNet-XS |
|
- local: api/pipelines/controlnetxs_sdxl |
|
title: ControlNet-XS with Stable Diffusion XL |
|
- local: api/pipelines/dance_diffusion |
|
title: Dance Diffusion |
|
- local: api/pipelines/ddim |
|
title: DDIM |
|
- local: api/pipelines/ddpm |
|
title: DDPM |
|
- local: api/pipelines/deepfloyd_if |
|
title: DeepFloyd IF |
|
- local: api/pipelines/diffedit |
|
title: DiffEdit |
|
- local: api/pipelines/dit |
|
title: DiT |
|
- local: api/pipelines/hunyuandit |
|
title: Hunyuan-DiT |
|
- local: api/pipelines/i2vgenxl |
|
title: I2VGen-XL |
|
- local: api/pipelines/pix2pix |
|
title: InstructPix2Pix |
|
- local: api/pipelines/kandinsky |
|
title: Kandinsky 2.1 |
|
- local: api/pipelines/kandinsky_v22 |
|
title: Kandinsky 2.2 |
|
- local: api/pipelines/kandinsky3 |
|
title: Kandinsky 3 |
|
- local: api/pipelines/latent_consistency_models |
|
title: Latent Consistency Models |
|
- local: api/pipelines/latent_diffusion |
|
title: Latent Diffusion |
|
- local: api/pipelines/ledits_pp |
|
title: LEDITS++ |
|
- local: api/pipelines/marigold |
|
title: Marigold |
|
- local: api/pipelines/panorama |
|
title: MultiDiffusion |
|
- local: api/pipelines/musicldm |
|
title: MusicLDM |
|
- local: api/pipelines/paint_by_example |
|
title: Paint by Example |
|
- local: api/pipelines/pia |
|
title: Personalized Image Animator (PIA) |
|
- local: api/pipelines/pixart |
|
title: PixArt-α |
|
- local: api/pipelines/pixart_sigma |
|
title: PixArt-Σ |
|
- local: api/pipelines/self_attention_guidance |
|
title: Self-Attention Guidance |
|
- local: api/pipelines/semantic_stable_diffusion |
|
title: Semantic Guidance |
|
- local: api/pipelines/shap_e |
|
title: Shap-E |
|
- local: api/pipelines/stable_cascade |
|
title: Stable Cascade |
|
- sections: |
|
- local: api/pipelines/stable_diffusion/overview |
|
title: Overview |
|
- local: api/pipelines/stable_diffusion/text2img |
|
title: Text-to-image |
|
- local: api/pipelines/stable_diffusion/img2img |
|
title: Image-to-image |
|
- local: api/pipelines/stable_diffusion/svd |
|
title: Image-to-video |
|
- local: api/pipelines/stable_diffusion/inpaint |
|
title: Inpainting |
|
- local: api/pipelines/stable_diffusion/depth2img |
|
title: Depth-to-image |
|
- local: api/pipelines/stable_diffusion/image_variation |
|
title: Image variation |
|
- local: api/pipelines/stable_diffusion/stable_diffusion_safe |
|
title: Safe Stable Diffusion |
|
- local: api/pipelines/stable_diffusion/stable_diffusion_2 |
|
title: Stable Diffusion 2 |
|
- local: api/pipelines/stable_diffusion/stable_diffusion_3 |
|
title: Stable Diffusion 3 |
|
- local: api/pipelines/stable_diffusion/stable_diffusion_xl |
|
title: Stable Diffusion XL |
|
- local: api/pipelines/stable_diffusion/sdxl_turbo |
|
title: SDXL Turbo |
|
- local: api/pipelines/stable_diffusion/latent_upscale |
|
title: Latent upscaler |
|
- local: api/pipelines/stable_diffusion/upscale |
|
title: Super-resolution |
|
- local: api/pipelines/stable_diffusion/k_diffusion |
|
title: K-Diffusion |
|
- local: api/pipelines/stable_diffusion/ldm3d_diffusion |
|
title: LDM3D Text-to-(RGB, Depth), Text-to-(RGB-pano, Depth-pano), LDM3D Upscaler |
|
- local: api/pipelines/stable_diffusion/adapter |
|
title: T2I-Adapter |
|
- local: api/pipelines/stable_diffusion/gligen |
|
title: GLIGEN (Grounded Language-to-Image Generation) |
|
title: Stable Diffusion |
|
- local: api/pipelines/stable_unclip |
|
title: Stable unCLIP |
|
- local: api/pipelines/text_to_video |
|
title: Text-to-video |
|
- local: api/pipelines/text_to_video_zero |
|
title: Text2Video-Zero |
|
- local: api/pipelines/unclip |
|
title: unCLIP |
|
- local: api/pipelines/unidiffuser |
|
title: UniDiffuser |
|
- local: api/pipelines/value_guided_sampling |
|
title: Value-guided sampling |
|
- local: api/pipelines/wuerstchen |
|
title: Wuerstchen |
|
title: Pipelines |
|
- isExpanded: false |
|
sections: |
|
- local: api/schedulers/overview |
|
title: Overview |
|
- local: api/schedulers/cm_stochastic_iterative |
|
title: CMStochasticIterativeScheduler |
|
- local: api/schedulers/consistency_decoder |
|
title: ConsistencyDecoderScheduler |
|
- local: api/schedulers/ddim_inverse |
|
title: DDIMInverseScheduler |
|
- local: api/schedulers/ddim |
|
title: DDIMScheduler |
|
- local: api/schedulers/ddpm |
|
title: DDPMScheduler |
|
- local: api/schedulers/deis |
|
title: DEISMultistepScheduler |
|
- local: api/schedulers/multistep_dpm_solver_inverse |
|
title: DPMSolverMultistepInverse |
|
- local: api/schedulers/multistep_dpm_solver |
|
title: DPMSolverMultistepScheduler |
|
- local: api/schedulers/dpm_sde |
|
title: DPMSolverSDEScheduler |
|
- local: api/schedulers/singlestep_dpm_solver |
|
title: DPMSolverSinglestepScheduler |
|
- local: api/schedulers/edm_multistep_dpm_solver |
|
title: EDMDPMSolverMultistepScheduler |
|
- local: api/schedulers/edm_euler |
|
title: EDMEulerScheduler |
|
- local: api/schedulers/euler_ancestral |
|
title: EulerAncestralDiscreteScheduler |
|
- local: api/schedulers/euler |
|
title: EulerDiscreteScheduler |
|
- local: api/schedulers/flow_match_euler_discrete |
|
title: FlowMatchEulerDiscreteScheduler |
|
- local: api/schedulers/heun |
|
title: HeunDiscreteScheduler |
|
- local: api/schedulers/ipndm |
|
title: IPNDMScheduler |
|
- local: api/schedulers/stochastic_karras_ve |
|
title: KarrasVeScheduler |
|
- local: api/schedulers/dpm_discrete_ancestral |
|
title: KDPM2AncestralDiscreteScheduler |
|
- local: api/schedulers/dpm_discrete |
|
title: KDPM2DiscreteScheduler |
|
- local: api/schedulers/lcm |
|
title: LCMScheduler |
|
- local: api/schedulers/lms_discrete |
|
title: LMSDiscreteScheduler |
|
- local: api/schedulers/pndm |
|
title: PNDMScheduler |
|
- local: api/schedulers/repaint |
|
title: RePaintScheduler |
|
- local: api/schedulers/score_sde_ve |
|
title: ScoreSdeVeScheduler |
|
- local: api/schedulers/score_sde_vp |
|
title: ScoreSdeVpScheduler |
|
- local: api/schedulers/tcd |
|
title: TCDScheduler |
|
- local: api/schedulers/unipc |
|
title: UniPCMultistepScheduler |
|
- local: api/schedulers/vq_diffusion |
|
title: VQDiffusionScheduler |
|
title: Schedulers |
|
- isExpanded: false |
|
sections: |
|
- local: api/internal_classes_overview |
|
title: Overview |
|
- local: api/attnprocessor |
|
title: Attention Processor |
|
- local: api/activations |
|
title: Custom activation functions |
|
- local: api/normalization |
|
title: Custom normalization layers |
|
- local: api/utilities |
|
title: Utilities |
|
- local: api/image_processor |
|
title: VAE Image Processor |
|
- local: api/video_processor |
|
title: Video Processor |
|
title: Internal classes |
|
title: API |
|
|