DeCode: Decoupling Content and Delivery for Medical QA

ACL ARR 2026 January Submission2374 Authors

02 Jan 2026 (modified: 20 Mar 2026)ACL ARR 2026 January SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Medical Question Answering, Large Language Models, Contextualized Response Generation, Content–Delivery Decoupling, Modular Prompting, Patient-Centered Communication, HealthBench
Abstract: Large language models (LLMs) exhibit strong medical knowledge and can generate factually accurate responses. However, existing models often fail to account for individual patient contexts, producing answers that are clinically correct yet poorly aligned with patients’ needs. In this work, we introduce DeCode (Decoupling Content and Delivery), a training-free, model-agnostic framework that adapts existing LLMs to produce contextualized answers in clinical settings. We evaluate DeCode on OpenAI HealthBench, a comprehensive and challenging benchmark designed to assess clinical relevance and validity of LLM responses. DeCode improves the previous state-of-the-art from 28.4% to 49.8%, corresponding to a 75% relative improvement. Experimental results suggest the effectiveness of DeCode in improving clinical question answering of LLMs.
Paper Type: Long
Research Area: Clinical and Biomedical Applications
Research Area Keywords: biomedical QA, healthcare applications, clinical NLP
Contribution Types: NLP engineering experiment
Languages Studied: English
Submission Number: 2374
Loading