Abstract: This paper proposes a method to control a music generation process from videos using users’ facial expressions, aligning the music with their emotions. Unlike previous works, this method considers users’ facial expressions, which reflect their emotions, to create more resonant music. We establish a functional relationship between facial expressions and music attributes and integrate these functions into the music generation process in a generation phase. Then zero-shot controllable music generation with facial expressions becomes feasible. To address the challenge of objectively verifying the generated music’s emotional correspondence, we introduce a novel evaluation metric that compares video, music, and emotions within a common latent space. The experimental results show the effectiveness of our method.
Loading