SUMIE: A Synthetic Benchmark for Incremental Entity Summarization

ACL ARR 2024 August Submission177 Authors

15 Aug 2024 (modified: 22 Sept 2024)ACL ARR 2024 August SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: No existing dataset adequately tests how well language models can incrementally update entity summaries – a crucial ability as these models rapidly advance. The Incremental Entity Summarization (IES) task is vital for maintaining accurate, up-to-date knowledge. To address this, we introduce SUMIE, a fully synthetic dataset designed to expose real-world IES challenges. This dataset effectively highlights problems like incorrect entity association and incomplete information presentation. Unlike common synthetic datasets, ours captures the complexity and nuances found in real-world data. We generate informative and diverse attributes, summaries, and unstructured paragraphs in sequence, ensuring high quality. The alignment between generated summaries and paragraphs exceeds 99%, confirming the dataset's quality. Extensive experiments demonstrate the dataset's difficulty – state-of-the-art LLMs struggle to update summaries with an F1 higher than 80.4%. We will open source the benchmark and the evaluation metrics to help the community make progress on IES tasks.
Paper Type: Long
Research Area: Summarization
Research Area Keywords: Incremental entity summarization, synthetic dataset, large language model, knowledge update
Contribution Types: NLP engineering experiment, Data resources
Languages Studied: English
Submission Number: 177
Loading