Better entity matching with transformers through ensembles

Published: 01 Jan 2024, Last Modified: 27 Jul 2024Knowl. Based Syst. 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•Ensembles can be constructed with just one transformer architecture.•Single-architecture ensembles save on memory, making LLM ensembles viable.•Non-stochastic shuffling of LLM input tokens is sufficiently unique for ensembles.•More aggregator tokens equals more information preserved equals better EM performance.•Extra self-attention keeps model focused only on the informative aggregator tokens.
Loading