ZeroDL: Zero-shot Distribution Learning for Text Clustering via Large Language Models

ACL ARR 2024 December Submission675 Authors

15 Dec 2024 (modified: 05 Feb 2025)ACL ARR 2024 December SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: The advancements in large language models (LLMs) have brought significant progress in NLP tasks. However, if a task cannot be fully described in prompts, the models could fail to carry out the task. In this paper, we propose a simple yet effective method to contextualize a task toward a LLM. The method utilizes (1) open-ended zero-shot inference from the entire dataset, (2) aggregating the inference results, and (3) finally incorporate the aggregated meta-information for the actual task. We show the effectiveness in text clustering tasks, empowering LLMs to perform text-to-text-based clustering and leading to improvements on several datasets. Lastly, we explore the generated class labels for clustering, showing how the LLM understands the task through data.
Paper Type: Short
Research Area: Language Modeling
Research Area Keywords: prompting
Contribution Types: NLP engineering experiment
Languages Studied: English
Submission Number: 675
Loading