Abstract: We present two approaches to animate an android robot head according to audio speech input, both are adopted from recent machine learning based works in computer graphics animation. More concrete we implemented a viseme-based and a mesh-based approach on our robot. After a subjective comparison we conduct a speech-reading study to evaluate our preferred, the mesh-based, approach. The results show that on average the intelligibility is not increased by the visual cues provided through the robot head in comparison to noisy audio alone. This underlines the importance of carefully designing and controlling the facial co-speech movements of talking android heads.
Loading