Diffusion-based generative models have proven to be highly effective in various domains of synthesis. In this work, we propose a conditional paradigm utilizing the denoising diffusion probabilistic model (DDPM) to address the challenge of realistic and diverse action-conditioned 3D skeleton-based motion generation. The proposed method leverages bidirectional Markov chains to generate samples by inferring the reversed Markov chain based on the learned distribution mapping during the forward diffusion process. To the best of our knowledge, our work is the first to employ DDPM to synthesize a variable number of motion sequences conditioned on a categorical action. The proposed method is evaluated on the NTU RGB+D dataset and the NTU RGB+D two-person dataset, showing significant improvements over state-of-theart motion generation methods.
Denoising Diffusion Probabilistic Models for Action-Conditioned 3D Motion Generation / Zhao, Mengyi; Liu, Mengyuan; Ren, Bin; Dai, Shuling; Sebe, Nicu. - (2024), pp. 4225-4229. (Intervento presentato al convegno 49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 tenutosi a Seoul nel 14-19, April 2024) [10.1109/ICASSP48485.2024.10446185].
Denoising Diffusion Probabilistic Models for Action-Conditioned 3D Motion Generation
Ren, Bin;Sebe, Nicu
2024-01-01
Abstract
Diffusion-based generative models have proven to be highly effective in various domains of synthesis. In this work, we propose a conditional paradigm utilizing the denoising diffusion probabilistic model (DDPM) to address the challenge of realistic and diverse action-conditioned 3D skeleton-based motion generation. The proposed method leverages bidirectional Markov chains to generate samples by inferring the reversed Markov chain based on the learned distribution mapping during the forward diffusion process. To the best of our knowledge, our work is the first to employ DDPM to synthesize a variable number of motion sequences conditioned on a categorical action. The proposed method is evaluated on the NTU RGB+D dataset and the NTU RGB+D two-person dataset, showing significant improvements over state-of-theart motion generation methods.File | Dimensione | Formato | |
---|---|---|---|
Denoising_Diffusion_Probabilistic_Models_for_Action-Conditioned_3D_Motion_Generation.pdf
Solo gestori archivio
Tipologia:
Versione editoriale (Publisher’s layout)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
2 MB
Formato
Adobe PDF
|
2 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione