Skip to content

gxl-groups/OF-VTON

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Recurrent Appearance Flow for Occlusion-Free Virtual Try-On (OF-VTON)

Introduction

Image-based virtual try-on aims at transferring a target in-shop garment onto a reference person, which has garnered significant attention from the research communities recently. However, previous methods have faced severe challenges in handling occlusion problems. To address this limitation, we classify occlusion problems into three types based on the reference person’s arm postures: single-arm occlusion, two-arm non-crossed occlusion, and two-arm crossed occlusion. Specifically, we propose a novel Occlusion-Free Virtual Try-On Network (OF-VTON) that effectively overcomes these occlusion challenges. The OF-VTON framework consists of two core components: i) a new Recurrent Appearance Flow based Deformation (RAFD) model that robustly aligns the in-shop garment to the reference person by adopting a multi-task learning strategy. This model jointly produces the dense appearance flow to warp the garment and predicts a human segmentation map to provide semantic guidance for the subsequent image synthesis model. ii) a powerful Multi-mask Image SynthesiS (MISS) model that generates photo-realistic try-on results by introducing a new mask generation and selection mechanism. Experimental results demonstrate that our proposed OF-VTON significantly outperforms existing state-of-theart methods by mitigating the impact of occlusion problems.

baseline

Environment

  • pytorch(1.10.0)
  • torchvision
  • scipy
  • Pillow
  • einops
  • opencv-python

Getting Started

Data Preperation

We provide our dataset files , extracted keypoints files and extracted parsing files for convience.

Data Preperation

Pretrained models

Download the models below and put it under checkpoint/

OFVTON-checkpoints

Test the model

python test.py --name test --warp_checkpoint ./checkpoints/warp.pth --gen_checkpoint ./checkpoints/gen.pth --dataroot ./datasets/

Citation

If any part of our paper and code is helpful to your work, please generously cite with:

@article{gu2024recurrent,
  title={Recurrent Appearance Flow for Occlusion-Free Virtual Try-On},
  author={Gu, Xiaoling and Zhu, Junkai and Wong, Yongkang and Wu, Zizhao and Yu, Jun and Fan, Jianping and Kankanhalli, Mohan S},
  journal={ACM Transactions on Multimedia Computing, Communications and Applications},
  year={2024},
  publisher={ACM New York, NY}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published