Can You See Me Think? Grounding LLM feedback in keystrokes and revision behaviour

Published: 24 Nov 2025, Last Modified: 24 Nov 20255th Muslims in ML Workshop co-located with NeurIPS 2025EveryoneRevisionsBibTeXCC BY 4.0
Keywords: Large Language Models (LLMs), Process-Aware Feedback, Second Language Writing, Ablation Study, Keystroke Logging, Snapshot Data, Cognitive Grounding, Human-AI Collaboration
TL;DR: This ablation study shows that adding keystroke and snapshot data to LLM feedback on L2 essays leads to more revision-aware and cognitively grounded suggestions.
Abstract: As large language models (LLMs) increasingly assist in evaluating student writing, researchers have begun to explore whether these systems can attend not just to final drafts, but to the writing process itself. We examine how LLM feedback can be anchored in student writing processes, using keystroke logs and revision snapshots as cognitive proxies. We compare two conditions: C1 (final essay only) and C2 (final essay + process data), using an ablation study on 52 student essays. While rubric scores changed little, but process-aware feedback (C2) offered more explicit recognition of revisions and organization changes. These findings suggest that cognitively-grounded feedback from LLMs is more pedagogically aligned and reflective of actual student effort.
Submission Number: 71
Loading