Not Like Transformers: Drop the Beat Representation for Dance Generation with Mamba-Based Diffusion Model
Keywords: 3D Dance Generation, Music-to-Dance Generation, Diffusion Models, Selective State Space Models, Beat Representation
TL;DR: A Mamba-based diffusion model for 3D dance generation with an explicit Gaussian-based beat representation
Abstract: Dance is a form of human motion characterized by emotional expression and communication, playing a role in various fields such as music, virtual reality, and content creation. Existing methods for dance generation often fail to adequately capture the inherently sequential, rhythmical, and music-synchronized characteristics of dance. In this paper, we propose a new dance generation approach that leverages a Mamba-based diffusion model. Mamba, specialized for handling long and autoregressive sequences, is integrated into our diffusion model as an alternative to the off-the-shelf Transformer. Additionally, considering the critical role of musical beats in dance choreography, we propose a Gaussian-based beat representation to explicitly guide the decoding of dance sequences. Experiments on AIST++ dataset show that our proposed method effectively reflects essential dance characteristics and advances performance compared to the state-of-the-art methods.
Supplementary Material: zip
Submission Number: 7
Loading