Entropy (Aug 2015)

A Gloss Composition and Context Clustering Based Distributed Word Sense Representation Model

  • Tao Chen,
  • Ruifeng Xu,
  • Yulan He,
  • Xuan Wang

DOI
https://doi.org/10.3390/e17096007
Journal volume & issue
Vol. 17, no. 9
pp. 6007 – 6024

Abstract

Read online

In recent years, there has been an increasing interest in learning a distributed representation of word sense. Traditional context clustering based models usually require careful tuning of model parameters, and typically perform worse on infrequent word senses. This paper presents a novel approach which addresses these limitations by first initializing the word sense embeddings through learning sentence-level embeddings from WordNet glosses using a convolutional neural networks. The initialized word sense embeddings are used by a context clustering based model to generate the distributed representations of word senses. Our learned representations outperform the publicly available embeddings on half of the metrics in the word similarity task, 6 out of 13 sub tasks in the analogical reasoning task, and gives the best overall accuracy in the word sense effect classification task, which shows the effectiveness of our proposed distributed distribution learning model.

Keywords