Vis enkel innførsel

dc.contributor.authorZhang, Xujie
dc.contributor.authorYang, Binbin
dc.contributor.authorKampffmeyer, Michael Christian
dc.contributor.authorZhang, Wenqing
dc.contributor.authorZhang, Shiyue
dc.contributor.authorLu, Guansong
dc.contributor.authorLin, Liang
dc.contributor.authorXu, Hang
dc.contributor.authorLiang, Xiaodan
dc.date.accessioned2024-02-16T13:36:48Z
dc.date.available2024-02-16T13:36:48Z
dc.date.issued2024-01-15
dc.description.abstractCross-modal garment synthesis and manipulation will significantly benefit the way fashion designers generate garments and modify their designs via flexible linguistic interfaces. However, despite the significant progress that has been made in generic image synthesis using diffusion models, producing garment images with garment part level semantics that are well aligned with input text prompts and then flexibly manipulating the generated results still remains a problem. Current approaches follow the general text-to-image paradigm and mine cross-modal relations via simple cross-attention modules, neglecting the structural correspondence between visual and textual representations in the fashion design domain. In this work, we instead introduce DiffCloth, a diffusion-based pipeline for cross-modal garment synthesis and manipulation, which empowers diffusion models with flexible compositionality in the fashion domain by structurally aligning the cross-modal semantics. Specifically, we formulate the part-level cross-modal alignment as a bipartite matching problem between the linguistic Attribute-Phrases (AP) and the visual garment parts which are obtained via constituency parsing and semantic segmentation, respectively. To mitigate the issue of attribute confusion, we further propose a semantic-bundled cross-attention to preserve the spatial structure similarities between the attention maps of attribute adjectives and part nouns in each AP. Moreover, DiffCloth allows for manipulation of the generated results by simply replacing APs in the text prompts. The manipulation-irrelevant regions are recognized by blended masks obtained from the bundled attention maps of the APs and kept unchanged. Extensive experiments on the CM-Fashion benchmark demonstrate that DiffCloth both yields state-of-the-art garment synthesis results by leveraging the inherent structural information and supports flexible manipulation with region consistency.en_US
dc.identifier.citationZhang X, Yang, Kampffmeyer MC, Zhang W, Zhang, Lu, Lin L, Xu H, Liang X. DiffCloth: Diffusion Based Garment Synthesis and Manipulation via Structural Cross-modal Semantic Alignment. IEEE International Conference on Computer Vision (ICCV). 2023en_US
dc.identifier.cristinIDFRIDAID 2185862
dc.identifier.doi10.1109/ICCV51070.2023.02116
dc.identifier.issn1550-5499
dc.identifier.issn2380-7504
dc.identifier.urihttps://hdl.handle.net/10037/32955
dc.language.isoengen_US
dc.publisherIEEEen_US
dc.relation.journalIEEE International Conference on Computer Vision (ICCV)
dc.relation.projectIDNorges forskningsråd: 309439en_US
dc.relation.projectIDNorges forskningsråd: 315029en_US
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2023 The Author(s)en_US
dc.titleDiffCloth: Diffusion Based Garment Synthesis and Manipulation via Structural Cross-modal Semantic Alignmenten_US
dc.type.versionacceptedVersionen_US
dc.typeJournal articleen_US
dc.typeTidsskriftartikkelen_US
dc.typePeer revieweden_US


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel