A Comparative Analysis of Retrieval-Augmented Generation and Crowdsourcing for Fact-Checking

Published: 2025, Last Modified: 12 Jan 2026ECIR (3) 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: In contrasting misinformation, both automated approaches using Large Language Models (LLMs) and human-based crowdsourcing have been explored. We enhance LLMs with a retrieval component to create a Retrieval-Augmented Generation (RAG) fact-checking system. Using state-of-the-art LLMs and a popular dataset, we empirically evaluate the system’s effectiveness and compare it with crowdsourced annotations. Our findings indicate that while RAG systems excel at detecting clear misinformation, they struggle with subtler distinctions where human judgment is more discerning. Interestingly, RAG systems show higher effectiveness on newer statements, suggesting adaptability to emerging misinformation.
Loading