Vis enkel innførsel

dc.contributor.authorZhao, Fuwei
dc.contributor.authorXie, Zhenyu
dc.contributor.authorKampffmeyer, Michael
dc.contributor.authorDong, Haoye
dc.contributor.authorHan, Songfang
dc.contributor.authorZheng, Tianxiang
dc.contributor.authorZhang, Tao
dc.contributor.authorLiang, Xiaodan
dc.date.accessioned2022-03-28T12:32:11Z
dc.date.available2022-03-28T12:32:11Z
dc.date.issued2022-02-28
dc.description.abstractVirtual 3D try-on can provide an intuitive and realistic view for online shopping and has a huge potential commercial value. However, existing 3D virtual try-on methods mainly rely on annotated 3D human shapes and garment templates, which hinders their applications in practical scenarios. 2D virtual try-on approaches provide a faster alternative to manipulate clothed humans, but lack the rich and realistic 3D representation. In this paper, we propose a novel Monocular-to-3D Virtual Try-On Network (M3D-VTON) that builds on the merits of both 2D and 3D approaches. By integrating 2D information efficiently and learning a mapping that lifts the 2D representation to 3D, we make the first attempt to reconstruct a 3D try-on mesh only taking the target clothing and a person image as inputs. The proposed M3D-VTON includes three modules: 1) The Monocular Prediction Module (MPM) that estimates an initial full-body depth map and accomplishes 2D clothes-person alignment through a novel two-stage warping procedure; 2) The Depth Refinement Module (DRM) that refines the initial body depth to produce more detailed pleat and face characteristics; 3) The Texture Fusion Module (TFM) that fuses the warped clothing with the non-target body part to refine the results. We also construct a high-quality synthesized Monocular-to-3D virtual try-on dataset, in which each person image is associated with a front and a back depth map. Extensive experiments demonstrate that the proposed M3D-VTON can manipulate and reconstruct the 3D human body wearing the given clothing with compelling details and is more efficient than other 3D approaches.en_US
dc.description© 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.identifier.citationZhao, Xie, Kampffmeyer MC, Dong, Han, Zheng, Zhang T, Liang X. M3D-VTON: A Monocular-to-3D Virtual Try-On Network. IEEE International Conference on Computer Vision (ICCV). 2021en_US
dc.identifier.cristinIDFRIDAID 1941703
dc.identifier.doi10.1109/ICCV48922.2021.01299
dc.identifier.issn1550-5499
dc.identifier.issn2380-7504
dc.identifier.urihttps://hdl.handle.net/10037/24603
dc.language.isoengen_US
dc.publisherIEEEen_US
dc.relation.journalIEEE International Conference on Computer Vision (ICCV)
dc.relation.projectIDNorges forskningsråd: 315029en_US
dc.relation.projectIDNorges forskningsråd: 303514en_US
dc.relation.projectIDNorges forskningsråd: 309439en_US
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2021 The Author(s)en_US
dc.titleM3D-VTON: A Monocular-to-3D Virtual Try-On Networken_US
dc.type.versionacceptedVersionen_US
dc.typeJournal articleen_US
dc.typeTidsskriftartikkelen_US
dc.typePeer revieweden_US


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel