File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Taming Diffusion Models for Audio-Driven Co-Speech Gesture Generation

TitleTaming Diffusion Models for Audio-Driven Co-Speech Gesture Generation
Authors
Issue Date22-Aug-2023
Abstract

Animating virtual avatars to make co-speech gestures facilitates various applications in human-machine interaction. The existing methods mainly rely on generative adversarial networks (GANs), which typically suffer from notorious mode collapse and unstable training, thus making it difficult to learn accurate audio-gesture joint distributions. In this work, we propose a novel diffusion-based framework, named Diffusion Co-Speech Gesture (DiffGesture), to effectively capture the cross-modal audio-to-gesture associations and preserve temporal coherence for high-fidelity audio-driven co-speech gesture generation. Specifically, we first establish the diffusion-conditional generation process on clips of skeleton sequences and audio to enable the whole framework. Then, a novel Diffusion Audio-Gesture Transformer is devised to better attend to the information from multiple modalities and model the long-term temporal dependency. Moreover, to eliminate temporal inconsistency, we propose an effective Diffusion Gesture Stabilizer with an annealed noise sampling strategy. Benefiting from the architectural advantages of diffusion models, we further incorporate implicit classifier-free guidance to trade off between diversity and gesture quality. Extensive experiments demonstrate that DiffGesture achieves state-of-the-art performance, which renders coherent gestures with better mode coverage and stronger audio correlations. Code is available at https://github.com/Advocate99/DiffGesture.


Persistent Identifierhttp://hdl.handle.net/10722/333824

 

DC FieldValueLanguage
dc.contributor.authorZhu, L-
dc.contributor.authorLiu, X-
dc.contributor.authorLiu, X-
dc.contributor.authorQian, R-
dc.contributor.authorLiu, Z-
dc.contributor.authorYu, L-
dc.date.accessioned2023-10-06T08:39:23Z-
dc.date.available2023-10-06T08:39:23Z-
dc.date.issued2023-08-22-
dc.identifier.urihttp://hdl.handle.net/10722/333824-
dc.description.abstract<p>Animating virtual avatars to make co-speech gestures facilitates various applications in human-machine interaction. The existing methods mainly rely on generative adversarial networks (GANs), which typically suffer from notorious mode collapse and unstable training, thus making it difficult to learn accurate audio-gesture joint distributions. In this work, we propose a novel diffusion-based framework, named Diffusion Co-Speech Gesture (DiffGesture), to effectively capture the cross-modal audio-to-gesture associations and preserve temporal coherence for high-fidelity audio-driven co-speech gesture generation. Specifically, we first establish the diffusion-conditional generation process on clips of skeleton sequences and audio to enable the whole framework. Then, a novel Diffusion Audio-Gesture Transformer is devised to better attend to the information from multiple modalities and model the long-term temporal dependency. Moreover, to eliminate temporal inconsistency, we propose an effective Diffusion Gesture Stabilizer with an annealed noise sampling strategy. Benefiting from the architectural advantages of diffusion models, we further incorporate implicit classifier-free guidance to trade off between diversity and gesture quality. Extensive experiments demonstrate that DiffGesture achieves state-of-the-art performance, which renders coherent gestures with better mode coverage and stronger audio correlations. Code is available at https://github.com/Advocate99/DiffGesture.<br></p>-
dc.languageeng-
dc.relation.ispartof2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (17/06/2023-24/06/2023, Vancouver, BC, Canada)-
dc.titleTaming Diffusion Models for Audio-Driven Co-Speech Gesture Generation-
dc.typeConference_Paper-
dc.identifier.doi10.1109/CVPR52729.2023.01016-
dc.identifier.spage10544-
dc.identifier.epage10553-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats