Large Language Models Still Face Challenges in Multi-Hop Reasoning with External Knowledge

Published: 01 Jan 2024, Last Modified: 22 Sept 2025CoRR 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: We carry out a series of experiments to test large language models' multi-hop reasoning ability from three aspects: selecting and combining external knowledge, dealing with non-sequential reasoning tasks and generalising to data samples with larger numbers of hops. We test the GPT-3.5 model on four reasoning benchmarks with Chain-of-Thought prompting (and its variations). Our results reveal that despite the amazing performance achieved by large language models on various reasoning tasks, models still suffer from severe drawbacks which shows a large gap with humans.
Loading