Beyond surprisal: GPT-derived attention metrics offer additional explanatory power in predicting the N400 during naturalistic reading

Published: 03 Oct 2025, Last Modified: 13 Nov 2025CPL 2025 SpotlightPosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: language models, surprisal, attention, N400, EEG, reading
TL;DR: GPT-derived attention metrics offer additional explanatory power in predicting the N400 during naturalistic reading
Abstract: The N400 component of the EEG signal is a well-established neural correlate of real-time language comprehension, sensitive to a range of lexical and contextual variables. While earlier studies have linked measures of word predictability (such as language model surprisal) to N400 amplitude, the N400 is known to reflect a broader array of cognitive processes than lexical expectation alone. For reading time prediction, prior work has found significant increases in language model psychometric predictive power based on metrics that go beyond surprisal, by capturing incremental changes in model attention patterns across timesteps (Oh & Schuler, 2022; Ryu & Lewis, 2025). In this study, we examine whether additional metrics derived from GPT language models, particularly attention-based measures, can provide complementary predictive value for N400 amplitude. We evaluate language model N400 predictivity on EEG data from the RaCCooNs dataset of Dutch naturalistic reading (Frank & Aumeistere, 2024), which contains data from 37 participants reading 200 Dutch sentences each. We fed the same sentences to four GPT-based language models (including monolingual and multilingual variants), and extracted surprisal, lexical entropy, and attention-derived metrics (the three metrics with highest explanatory value in a reading times study by Oh and Schuler, 2022). To compare the predictive power of these different metrics, we fit mixed-effect regression models with language model-based metrics as fixed-effect predictors and word- and participant-level random effects, including word frequency, word position and word length as covariates of no interest. Next we examined the quality of these model fits across metrics. We show that GPT surprisal robustly predicts N400 amplitude during naturalistic reading of Dutch (Figure 1). Crucially, we demonstrate for the first time that attention-based metrics, namely Normalized Attention Entropy (NAE), ΔNAE, and Manhattan Distance (MD), offer significant additional explanatory power beyond surprisal and lexical entropy in predicting the N400 (Figure 2), based on analyses of two representative GPT models. These findings demonstrate that attention metrics from Transformer models can potentially serve as powerful and cognitively informative predictors of neural language processing in a naturalistic context.
Submission Number: 30
Loading