Evaluating the role of ‘Constitutions’ for learning from AI feedback

Published: 30 Oct 2024, Last Modified: 13 Dec 2024LanGame PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: in-context learning, ai feedback, medical ai, llms, human-ai communication
TL;DR: We compare different constitutions as the basis for AI feedback and find that detailed constitutions lead to better outcomes, but also find limitations to AI feedback in certain areas.
Abstract: The growing capabilities of large language models (LLMs) have led to their use as substitutes for human feedback for training and assessing other LLMs. These methods often rely on `constitutions', written guidelines which a critic model uses to provide feedback and improve generations. We investigate how the choice of constitution affects feedback quality by using four different constitutions to improve patient-centered communication in medical interviews. In pairwise comparisons conducted by 215 human raters, we found that detailed constitutions led to better results regarding emotive qualities. However, none of the constitutions outperformed the baseline in learning more practically-oriented skills related to information gathering and provision. Our findings indicate that while detailed constitutions should be prioritised, there are possible limitations to the effectiveness AI feedback as a reward signal in certain areas.
Submission Number: 12
Loading