Envision Human-AI Perceptual Alignment from a Multimodal Interaction Perspective

Published: 06 Mar 2025, Last Modified: 21 Apr 2025ICLR 2025 Bi-Align Workshop PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Multimodal Interaction, Perceptual Alignment, Touch, Olfaction
Abstract: Aligning AI with human intent has seen progress, yet perceptual alignment—how AI interprets what we see, hear, feel, or smell—remains underexplored. This paper advocates for expanding perceptual alignment efforts across multimodal sensory modalities, such as touch and olfaction, which are critical for how humans perceive and interpret their environment. We envision AI systems enabling natural, multimodal interactions in everyday contexts, such as selecting clothing that aligns with temperature and texture preferences or recreating rich sensory ambiances that evoke specific sights, sounds, and smells. By advancing multimodal representation learning and perceptual alignment, this work aims to inspire the computer science and human-computer interaction (HCI) communities to design inclusive, human-centered AI systems for everyday, multisensory experiences.
Submission Type: Short Paper (4 Pages)
Archival Option: This is an archival submission
Presentation Venue Preference: ICLR 2025
Submission Number: 100
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview