Large Language Models Show Signs of Alignment with Human Neurocognition During Abstract Reasoning

20 Sept 2025 (modified: 12 Feb 2026)ICLR 2026 Conference Desk Rejected SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Abstract Reasoning; AI; Artificial Neural Networks; NeuroAI; Deep Learning; Large Language Models (LLMs); Electroencephalography (EEG); Fixation Related Potentials (FRPs); Repre- sentational Similarity Analysis (RSA)
Abstract: This study investigates whether large language models (LLMs) mirror human neurocognition during abstract reasoning. We compared the performance and neural representations of human participants with those of eight open-source LLMs on an abstract-pattern-completion task. We leveraged pattern type differences in task performance and in fixation-related potentials (FRPs) as recorded by electroen- cephalography (EEG) during the task. Our findings indicate that only the largest tested LLMs (∼70 billion parameters) achieve human-comparable accuracy, with Qwen-2.5-72B and DeepSeek-R1-70B also showing similarities with the human pattern-specific difficulty profile. Critically, every LLM tested forms representations that distinctly cluster the abstract pattern categories within their intermediate layers, although the strength of this clustering scales with their performance on the task. Moderate positive correlations were observed between the representational geometries of task-optimal LLM layers and human frontal FRPs. These results consistently diverged from comparisons with other EEG measures (response-locked ERPs and resting EEG), suggesting a potential shared representational space for abstract patterns. This indicates that LLMs might mirror human brain mechanisms in abstract reasoning, offering preliminary evidence of shared principles between biological and artificial intelligence.
Supplementary Material: pdf
Primary Area: foundation or frontier models, including LLMs
Submission Number: 22508
Loading