Information (Feb 2024)

Robust Chinese Short Text Entity Disambiguation Method Based on Feature Fusion and Contrastive Learning

  • Qishun Mei,
  • Xuhui Li

DOI
https://doi.org/10.3390/info15030139
Journal volume & issue
Vol. 15, no. 3
p. 139

Abstract

Read online

To address the limitations of existing methods of short-text entity disambiguation, specifically in terms of their insufficient feature extraction and reliance on massive training samples, we propose an entity disambiguation model called COLBERT, which fuses LDA-based topic features and BERT-based semantic features, as well as using contrastive learning, to enhance the disambiguation process. Experiments on a publicly available Chinese short-text entity disambiguation dataset show that the proposed model achieves an F1-score of 84.0%, which outperforms the benchmark method by 0.6%. Moreover, our model achieves an F1-score of 74.5% with a limited number of training samples, which is 2.8% higher than the benchmark method. These results demonstrate that our model achieves better effectiveness and robustness and can reduce the burden of data annotation as well as training costs.

Keywords