Addressing The Polysemy Problem In Language Modeling With Attentional Multi-Sense Embeddings
Rao Ma, Lesheng Jin, Qi Liu, Kai Yu, Lu Chen
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 14:58
Neural network language models have gained considerable popularity due to their promising performance. Distributed word embeddings are utilized to represent semantic information. However, each word is associated with a single vector in the embedding layer, disabling the model from capturing the meanings of polysemous words. In this work, we address this problem by assigning multiple fine-grained sense embeddings to each word in the embedding layers. The proposed model discriminates among different senses of a word with attention mechanism in an unsupervised manner. Experiments demonstrate the benefits of our approach in language modeling and ASR rescoring. Investigations are also made on standard word similarity tasks. The results indicate that our proposed method is efficient in modeling polysemy and therefore obtains better word representations.