Vis enkel innførsel

dc.contributor.authorLin, Luoyang
dc.contributor.authorJiang, Zutao
dc.contributor.authorLiang, Xiaodan
dc.contributor.authorMa, Liqian
dc.contributor.authorKampffmeyer, Michael Christian
dc.contributor.authorCao, Xiaochun
dc.date.accessioned2025-03-17T10:00:41Z
dc.date.available2025-03-17T10:00:41Z
dc.date.issued2024-03-24
dc.description.abstractTalking upper-body synthesis is a promising task due to its versatile potential for video creation and consists of animating the body and face from a source image with the motion from a given driving video. However, prior synthesis approaches fall short in addressing this task and have been either limited to animating heads of a target person only, or have animated the upper body but neglected the synthesis of precise facial details. To tackle this task, we propose a Photo-realistic Talking Upper-body Synthesis method via 3D-aware motion decomposition warping, named PTUS, to both precisely synthesize the upper body as well as recover the details of the face such as blinking and lip synchronization. In particular, the motion decomposition mechanism consists of a face-body motion decomposition, which decouples the 3D motion estimation of the face and body, and a local-global motion decomposition, which decomposes the 3D face motion into global and local motions resulting in the transfer of facial expression. The 3D-aware warping module transfers the large-scale and subtle 3D motions to the extracted 3D depth-aware features in a coarse-tofine manner. Moreover, we present a new dataset, Talking-UB, which includes upper-body images with high-resolution faces, addressing the limitations of prior datasets that either consist of only facial images or upper-body images with blurry faces. Experimental results demonstrate that our proposed method can synthesize high-quality videos that preserve facial details, and achieves superior results compared to state-of-the-art cross-person motion transfer approaches. Code and collected dataset are released in https://github.com/cooluoluo/PTUS.en_US
dc.identifier.citationLin, Jiang, Liang, Ma, Kampffmeyer, Cao. PTUS: Photo-Realistic Talking Upper-Body Synthesis via 3D-Aware Motion Decomposition Warping. Proceedings of the AAAI Conference on Artificial Intelligence. 2024;38(4)en_US
dc.identifier.cristinIDFRIDAID 2296322
dc.identifier.doi10.1609/aaai.v38i4.28131
dc.identifier.issn2159-5399
dc.identifier.issn2374-3468
dc.identifier.urihttps://hdl.handle.net/10037/36705
dc.language.isoengen_US
dc.publisherAssociation for the Advancement of Artificial Intelligenceen_US
dc.relation.journalProceedings of the AAAI Conference on Artificial Intelligence
dc.relation.projectIDNorges forskningsråd: 309439en_US
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2024 The Author(s)en_US
dc.titlePTUS: Photo-Realistic Talking Upper-Body Synthesis via 3D-Aware Motion Decomposition Warpingen_US
dc.type.versionacceptedVersionen_US
dc.typeJournal articleen_US
dc.typeTidsskriftartikkelen_US
dc.typePeer revieweden_US


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel