Diffusion-based generative models have proven to be highly effective in various domains of synthesis. In this work, we propose a conditional paradigm utilizing the denoising diffusion probabilistic model (DDPM) to address the challenge of realistic and diverse action-conditioned 3D skeleton-based motion generation. The proposed method leverages bidirectional Markov chains to generate samples by inferring the reversed Markov chain based on the learned distribution mapping during the forward diffusion process. To the best of our knowledge, our work is the first to employ DDPM to synthesize a variable number of motion sequences conditioned on a categorical action. The proposed method is evaluated on the NTU RGB+D dataset and the NTU RGB+D two-person dataset, showing significant improvements over state-of-theart motion generation methods.

Denoising Diffusion Probabilistic Models for Action-Conditioned 3D Motion Generation / Zhao, Mengyi; Liu, Mengyuan; Ren, Bin; Dai, Shuling; Sebe, Nicu. - (2024), pp. 4225-4229. (Intervento presentato al convegno 49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 tenutosi a Seoul nel 14-19, April 2024) [10.1109/ICASSP48485.2024.10446185].

Denoising Diffusion Probabilistic Models for Action-Conditioned 3D Motion Generation

Ren, Bin;Sebe, Nicu
2024-01-01

Abstract

Diffusion-based generative models have proven to be highly effective in various domains of synthesis. In this work, we propose a conditional paradigm utilizing the denoising diffusion probabilistic model (DDPM) to address the challenge of realistic and diverse action-conditioned 3D skeleton-based motion generation. The proposed method leverages bidirectional Markov chains to generate samples by inferring the reversed Markov chain based on the learned distribution mapping during the forward diffusion process. To the best of our knowledge, our work is the first to employ DDPM to synthesize a variable number of motion sequences conditioned on a categorical action. The proposed method is evaluated on the NTU RGB+D dataset and the NTU RGB+D two-person dataset, showing significant improvements over state-of-theart motion generation methods.
2024
49th IEEE International Conference on Acoustics, Speech, and Signal Processing
Piscataway, NJ USA
Institute of Electrical and Electronics Engineers Inc.
979-8-3503-4485-1
979-8-3503-4486-8
Zhao, Mengyi; Liu, Mengyuan; Ren, Bin; Dai, Shuling; Sebe, Nicu
Denoising Diffusion Probabilistic Models for Action-Conditioned 3D Motion Generation / Zhao, Mengyi; Liu, Mengyuan; Ren, Bin; Dai, Shuling; Sebe, Nicu. - (2024), pp. 4225-4229. (Intervento presentato al convegno 49th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2024 tenutosi a Seoul nel 14-19, April 2024) [10.1109/ICASSP48485.2024.10446185].
File in questo prodotto:
File Dimensione Formato  
Denoising_Diffusion_Probabilistic_Models_for_Action-Conditioned_3D_Motion_Generation.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 2 MB
Formato Adobe PDF
2 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/413711
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact