Sample Efficiency Matters: Training Multimodal Conversational Recommendation Systems in a Low Resource SettingDownload PDF

Anonymous

16 Feb 2024ACL ARR 2024 February Blind SubmissionReaders: Everyone
Abstract: Multi-modal conversational recommendation (multi-modal CRS) can potentially revolutionize how customers interact with e-commerce platforms. Yet conversational samples, as training data for such a system, are difficult to obtain in large quantities, particularly in new platforms. Motivated by this challenge, we consider multimodal CRS in a low resource setting. Specifically, assuming the availability of a small number of samples with dialog states, we devise an effective dialog state encoder to bridge the semantic gap between conversation and product representations for recommendation. To reduce the cost of dialog state annotation, a semi-supervised learning method is developed to effectively train the dialog state encoder with a smaller set of labeled conversations. In addition, we design a correlation regularisation that leverages knowledge in the multi-modal domain database to better align textual and visual modalities. Experiments on two datasets demonstrate the effectiveness of our method. Particularly, with only 5% of the MMD training set, our method (namely SeMANTIC) is comparable to the state-of-the-art model trained on the full dataset.
Paper Type: long
Research Area: Dialogue and Interactive Systems
Contribution Types: Approaches to low-resource settings
Languages Studied: English
0 Replies

Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview