File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: LEO: Generative Latent Image Animator for Human Video Synthesis

TitleLEO: Generative Latent Image Animator for Human Video Synthesis
Authors
KeywordsDeep generative models
Diffusion models
Human analysis
Video generation
Issue Date2024
Citation
International Journal of Computer Vision, 2024 How to Cite?
AbstractSpatio-temporal coherency is a major challenge in synthesizing high quality videos, particularly in synthesizing human videos that contain rich global and local deformations. To resolve this challenge, previous approaches have resorted to different features in the generation process aimed at representing appearance and motion. However, in the absence of strict mechanisms to guarantee such disentanglement, a separation of motion from appearance has remained challenging, resulting in spatial distortions and temporal jittering that break the spatio-temporal coherency. Motivated by this, we here propose LEO, a novel framework for human video synthesis, placing emphasis on spatio-temporal coherency. Our key idea is to represent motion as a sequence of flow maps in the generation process, which inherently isolate motion from appearance. We implement this idea via a flow-based image animator and a Latent Motion Diffusion Model (LMDM). The former bridges a space of motion codes with the space of flow maps, and synthesizes video frames in a warp-and-inpaint manner. LMDM learns to capture motion prior in the training data by synthesizing sequences of motion codes. Extensive quantitative and qualitative analysis suggests that LEO significantly improves coherent synthesis of human videos over previous methods on the datasets TaichiHD, FaceForensics and CelebV-HQ. In addition, the effective disentanglement of appearance and motion in LEO allows for two additional tasks, namely infinite-length human video synthesis, as well as content-preserving video editing. Project page: https://wyhsirius.github.io/LEO-project/.
Persistent Identifierhttp://hdl.handle.net/10722/352476
ISSN
2023 Impact Factor: 11.6
2023 SCImago Journal Rankings: 6.668

 

DC FieldValueLanguage
dc.contributor.authorWang, Yaohui-
dc.contributor.authorMa, Xin-
dc.contributor.authorChen, Xinyuan-
dc.contributor.authorChen, Cunjian-
dc.contributor.authorDantcheva, Antitza-
dc.contributor.authorDai, Bo-
dc.contributor.authorQiao, Yu-
dc.date.accessioned2024-12-16T03:59:18Z-
dc.date.available2024-12-16T03:59:18Z-
dc.date.issued2024-
dc.identifier.citationInternational Journal of Computer Vision, 2024-
dc.identifier.issn0920-5691-
dc.identifier.urihttp://hdl.handle.net/10722/352476-
dc.description.abstractSpatio-temporal coherency is a major challenge in synthesizing high quality videos, particularly in synthesizing human videos that contain rich global and local deformations. To resolve this challenge, previous approaches have resorted to different features in the generation process aimed at representing appearance and motion. However, in the absence of strict mechanisms to guarantee such disentanglement, a separation of motion from appearance has remained challenging, resulting in spatial distortions and temporal jittering that break the spatio-temporal coherency. Motivated by this, we here propose LEO, a novel framework for human video synthesis, placing emphasis on spatio-temporal coherency. Our key idea is to represent motion as a sequence of flow maps in the generation process, which inherently isolate motion from appearance. We implement this idea via a flow-based image animator and a Latent Motion Diffusion Model (LMDM). The former bridges a space of motion codes with the space of flow maps, and synthesizes video frames in a warp-and-inpaint manner. LMDM learns to capture motion prior in the training data by synthesizing sequences of motion codes. Extensive quantitative and qualitative analysis suggests that LEO significantly improves coherent synthesis of human videos over previous methods on the datasets TaichiHD, FaceForensics and CelebV-HQ. In addition, the effective disentanglement of appearance and motion in LEO allows for two additional tasks, namely infinite-length human video synthesis, as well as content-preserving video editing. Project page: https://wyhsirius.github.io/LEO-project/.-
dc.languageeng-
dc.relation.ispartofInternational Journal of Computer Vision-
dc.subjectDeep generative models-
dc.subjectDiffusion models-
dc.subjectHuman analysis-
dc.subjectVideo generation-
dc.titleLEO: Generative Latent Image Animator for Human Video Synthesis-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1007/s11263-024-02231-3-
dc.identifier.scopuseid_2-s2.0-85204903721-
dc.identifier.eissn1573-1405-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats