TryOffDiff: Virtual-Try-Off via High-Fidelity Garment Reconstruction using Diffusion Models
Abstract
This paper introduces Virtual Try-Off (VTOFF), a novel task focused on generating standardized garment images from single photos of clothed individuals. Unlike traditional Virtual Try-On (VTON), which digitally dresses models, VTOFF aims to extract a canonical garment image, posing unique challenges in capturing garment shape, texture, and intricate patterns. This well-defined target makes VTOFF particularly effective for evaluating reconstruction fidelity in generative models. We present TryOffDiff, a model that adapts Stable Diffusion with SigLIP-based visual conditioning to ensure high fidelity and detail retention. Experiments on a modified VITON-HD dataset show that our approach outperforms baseline methods based on pose transfer and virtual try-on with fewer pre- and post-processing steps. Our analysis reveals that traditional image generation metrics inadequately assess reconstruction quality, prompting us to rely on DISTS for more accurate evaluation. Our results highlight the potential of VTOFF to enhance product imagery in e-commerce applications, advance generative model evaluation, and inspire future work on high-fidelity reconstruction. Demo, code, and models are available at: https://rizavelioglu.github.io/tryoffdiff/
Community
TL;DR: While current Virtual Try-On (VTON) technologies focus on digitally dressing models, our novel Virtual Try-Off (VTOFF) task extracts canonical garment images from single photos. Using TryOffDiff, a Stable Diffusion-based model with SigLIP visual conditioning, we achieve high-fidelity garment reconstruction that advances e-commerce product imagery and generative model evaluation.
Differences between VTON and VTOFF:
Model architecture:
Paper: https://arxiv.org/abs/2411.17190
Project Page: https://rizavelioglu.github.io/tryoffdiff
Code: https://github.com/rizavelioglu/tryoffdiff
Training code will be released soon...
Hi @neltherion , as stated in the previous comment, I am currently cleaning the repository and will release it next Friday, at the latest ๐ค
Code is not available (404 Error)... This is becoming a trend...
You should ask for a refund...
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- TED-VITON: Transformer-Empowered Diffusion Models for Virtual Try-On (2024)
- FitDiT: Advancing the Authentic Garment Details for High-fidelity Virtual Try-on (2024)
- Controllable Human Image Generation with Personalized Multi-Garments (2024)
- GS-VTON: Controllable 3D Virtual Try-on with Gaussian Splatting (2024)
- Try-On-Adapter: A Simple and Flexible Try-On Paradigm (2024)
- OSDFace: One-Step Diffusion Model for Face Restoration (2024)
- High-Fidelity Virtual Try-on with Large-Scale Unpaired Learning (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
The code for TryOffDiff is now officially released! ๐ค (
@neltherion
)
You can find all scripts for training, prediction, and evaluation included.
Check it out at: https://github.com/rizavelioglu/tryoffdiff/
PS: Scripts for ablation studies and baselines will also be available soon. Stay tuned!
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper