Text Ranking and Classification using Data CompressionDownload PDF

Published: 18 Oct 2021, Last Modified: 22 Oct 2023ICBINB@NeurIPS2021 SpotlightReaders: Everyone
Keywords: data compression,text ranking,classification,machine learning,zstandard,text categorization,ranking,text classification,sentence embedding,compression
Abstract: A well-known but rarely used approach to text categorization uses conditional entropy estimates computed using data compression tools. Text affinity scores derived from compressed sizes can be used for classification and ranking tasks, but their success depends on the compression tools used. We use the Zstandard compressor and strengthen these ideas in several ways, calling the resulting language-agnostic technique Zest. In applications, this approach simplifies configuration, avoiding careful feature extraction and large ML models. Our ablation studies confirm the value of individual enhancements we introduce. We show that Zest complements and can compete with language-specific multidimensional content embeddings in production, but cannot outperform other counting methods on public datasets.
Category: Stuck paper: I hope to get ideas in this workshop that help me unstuck and improve this paper
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/arxiv:2109.11577/code)
1 Reply

Loading