Teach Large Language Models the Concept of Meta-cognition to Reduce Hallucination Text Generation

20 Sept 2023 (modified: 11 Feb 2024)Submitted to ICLR 2024EveryoneRevisionsBibTeX
Primary Area: generative models
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Keywords: large language models, meta-learning, low rank adapter, ai hallucination, prompt engineering
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
TL;DR: In this paper, we construct an algorithm to train a language model to evaluate itself's ability to answer certain questions.
Abstract: We introduce an algorithm that endows language models with enduring meta-cognitive capabilities. Inspired by meta-learning, our approach involves fine-tuning models on diverse datasets, including the original base model. Throughout each training iteration, we randomly select various fine-tuned model versions, gauge their meta-cognitive capacities, and employ the meta-cognitive error average as the loss function for gradient updates. This empowers these models to assess their competence when interpreting human instructions, thereby averting the generation of responses beyond their abilities and mitigating hallucinatory text production. The meta-cognitive ability will be adapt to various fine-tuned versions of the main model, providing evaluations that align with the fine-tuned models' knowledge capacity.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Supplementary Material: zip
Submission Number: 2625
Loading