Revisiting the Graph Reasoning Ability of Large Language Models: Case Studies in Translation, Connectivity and Shortest Path

ACL ARR 2024 December Submission1974 Authors

16 Dec 2024 (modified: 05 Feb 2025)ACL ARR 2024 December SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract:

Large Language Models (LLMs) have achieved great success in various reasoning tasks. However, their capacity for graph reasoning remains poorly understood. Although recent theoretical analyses suggest that LLMs can, in principle, perform complex graph tasks, empirical evaluations reveal numerous failures. To bridge this gap, we revisit the graph reasoning ability by introducing a new, balanced, and comprehensive benchmark. Through systematic experimentation, we identify key factors influencing performance, including node connectivity types, graph sizes, graph descriptions, and node naming methods. Moreover, we also demonstrate the impact of training data, model size and fine-tuning on graph reasoning. All the implementations and datasets are publicly available.

Paper Type: Long
Research Area: Interpretability and Analysis of Models for NLP
Research Area Keywords: graph reasoning ability
Contribution Types: Model analysis & interpretability
Languages Studied: English
Submission Number: 1974
Loading