Keywords: causal intervention, function vector, verb bias, structural priming, in-context learning, mechanistic interpretability, computational psycholinguistics
TL;DR: We show that verb bias, a continuous and context-dependent feature, is compactly represented in LLMs, and gradient manipulations of the associated subspaces play a predictable causal role in downstream structural choices.
Abstract: We investigate how large language models (LLMs) encode and causally use continuous, context-dependent properties in syntactic processing, focusing on verb bias and its influence on structural priming. Building on prior work that localized binary morphosyntactic features in non-basis-aligned subspaces, we introduce a simple and efficient method combining principal component analysis with beta regression to identify verb-bias subspaces in function vectors. Function vectors are compact task representations derived from in-context learning sequences. We show that function vectors can be employed to simulate structural priming in LLMs. Our method supports counterfactual continuous manipulation of the verb-bias subspace, and doing so yields the predicted shifts in priming magnitudes, confirming that the subspace is causally involved in syntactic choice. Our method thus extends causal interpretability methods to continuous linguistic variables, and our application of this method supports the proposal that the same mechanism is responsible for in-context learning and structural priming in LLMs.
Submission Number: 49
Loading