Reverse Thinking in Large Language Models

ACL ARR 2024 June Submission5277 Authors

16 Jun 2024 (modified: 02 Jul 2024)ACL ARR 2024 June SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Humans are accustomed to reading and writing in a forward manner, and this natural bias extends to text understanding in auto-regressive large language models (LLMs). This paper investigates whether LLMs, like humans, struggle with reverse thinking, specifically with reversed text inputs. We found that publicly available pre-trained LLMs cannot understand such inputs. However, LLMs trained from scratch with both forward and reverse texts can understand them equally well during inference. Our case study shows that different-content texts result in different losses if input (to LLMs) in different directions---some get lower losses for forward while some for reverse. This leads us to a simple and nice solution for data selection based on the loss differences between forward and reverse directions. Using our selected data in continued pretraining can boost LLMs' performance by a large margin for the task of Massive Multitask Language Understanding.
Paper Type: Short
Research Area: Interpretability and Analysis of Models for NLP
Research Area Keywords: pre-training,llm,large language models,large language model,reverse thinking,
Contribution Types: Model analysis & interpretability
Languages Studied: English,Korean,German,Arabic
Submission Number: 5277
Loading