Addressing the Polysemy Problem in Language Modeling with Attentional Multi-Sense EmbeddingsDownload PDFOpen Website

Published: 01 Jan 2020, Last Modified: 09 Feb 2024ICASSP 2020Readers: Everyone
Abstract: Neural network language models have gained considerable popularity due to their promising performance. Distributed word embeddings are utilized to represent semantic information. However, each word is associated with a single vector in the embedding layer, disabling the model from capturing the meanings of polysemous words. In this work, we address this problem by assigning multiple fine-grained sense embeddings to each word in the embedding layers. The proposed model discriminates among different senses of a word with attention mechanism in an unsupervised manner. Experiments demonstrate the benefits of our approach in language modeling and ASR rescoring. Investigations are also made on standard word similarity tasks. The results indicate that our proposed method is efficient in modeling polysemy and therefore obtains better word representations.
0 Replies

Loading