- Abstract: In this study, we employ skip connections into a deep recurrent neural network for modeling basic dance steps using audio as input. Our model consists of two blocks, one encodes the audio input sequences, and another generates the motion. The encoder uses a configuration called convolutional, long short-term memory deep neural network (CLDNN) which handle the power features of audio. Furthermore, we implement skip connections between the contexts of music encoder and motion decoder (i.e. delayed skip) for consistent motion generation. The experimental results show that the trained model generate predictive basic dance steps from a narrow dataset with low error and maintains similar motion beat fscore to the baseline dancer.
- TL;DR: Employing skip connections into a deep recurrent neural network for modeling basic dance steps using audio as input
- Keywords: Deep Learning, Skip Connectios, CLDRNN