TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Improving Diffusion Models for Authentic Virtual Try-on in...

Improving Diffusion Models for Authentic Virtual Try-on in the Wild

Yisol Choi, Sangkyung Kwak, Kyungmin Lee, Hyungwon Choi, Jinwoo Shin

2024-03-08Virtual Try-on
PaperPDFCode(official)

Abstract

This paper considers image-based virtual try-on, which renders an image of a person wearing a curated garment, given a pair of images depicting the person and the garment, respectively. Previous works adapt existing exemplar-based inpainting diffusion models for virtual try-on to improve the naturalness of the generated visuals compared to other methods (e.g., GAN-based), but they fail to preserve the identity of the garments. To overcome this limitation, we propose a novel diffusion model that improves garment fidelity and generates authentic virtual try-on images. Our method, coined IDM-VTON, uses two different modules to encode the semantics of garment image; given the base UNet of the diffusion model, 1) the high-level semantics extracted from a visual encoder are fused to the cross-attention layer, and then 2) the low-level features extracted from parallel UNet are fused to the self-attention layer. In addition, we provide detailed textual prompts for both garment and person images to enhance the authenticity of the generated visuals. Finally, we present a customization method using a pair of person-garment images, which significantly improves fidelity and authenticity. Our experimental results show that our method outperforms previous approaches (both diffusion-based and GAN-based) in preserving garment details and generating authentic virtual try-on images, both qualitatively and quantitatively. Furthermore, the proposed customization method demonstrates its effectiveness in a real-world scenario. More visualizations are available in our project page: https://idm-vton.github.io

Results

TaskDatasetMetricValueModel
Virtual Try-onVITON-HDFID6.29IDM-VTON
1 Image, 2*2 StitchiVITON-HDFID6.29IDM-VTON

Related Papers

TalkFashion: Intelligent Virtual Try-On Assistant Based on Multimodal Large Language Model2025-07-08Video Virtual Try-on with Conditional Diffusion Transformer Inpainter2025-06-26Real-Time Per-Garment Virtual Try-On with Temporal Consistency for Loose-Fitting Garments2025-06-14Low-Barrier Dataset Collection with Real Human Body for Interactive Per-Garment Virtual Try-On2025-06-12VITON-DRR: Details Retention Virtual Try-on via Non-rigid Registration2025-05-29Inverse Virtual Try-On: Generating Multi-Category Product-Style Images from Clothed Individuals2025-05-27VTBench: Comprehensive Benchmark Suite Towards Real-World Virtual Try-on Models2025-05-26HF-VTON: High-Fidelity Virtual Try-On via Consistent Geometric and Semantic Alignment2025-05-26