Psycholinguistic Diagnosis of Language Models’ Commonsense ReasoningDownload PDF

10 Mar 2022, 22:17 (modified: 11 Apr 2022, 01:09)ACL 2022 Workshop CSRRReaders: Everyone
Keywords: language models, commonsense pragmatics, cognitive plausibility, implicit meaning, implicatures, presuppositions
TL;DR: Analyze and fine-tune neural language models' commonsense pragmatics knowledge using human behavioral/neural data.
Abstract: Neural language models have attracted a lot of attention in the past few years. More and more researchers are getting intrigued by how language models encode commonsense, specifically what kind of commonsense they understand, and why they do. This paper analyzed neural language models’ understanding of commonsense pragmatics (i.e., implied meanings) through human behavioral and neurophysiological data. These psycholinguistic tests are designed to draw conclusions based on predictive responses in context, making them very well suited to test word-prediction models such as BERT in natural settings. They can provide the appropriate prompts and tasks to answer questions about linguistic mechanisms underlying predictive responses. This paper adopted psycholinguistic datasets to probe language models' commonsense reasoning. Findings suggest that GPT-3's performance was mostly at chance in the psycholinguistic tasks. We also showed that DistillBERT had some understanding of the (implied) intent that's shared among most people. Such intent is implicitly reflected in the usage of conversational implicatures and presuppositions. Whether or not fine-tuning improved its performance to human-level depends on the type of commonsense reasoning.
Published: No
Archival: Yes
4 Replies