International Journal of Crowd Science (Sep 2024)

Converse Attention Knowledge Transfer for Low-Resource Named Entity Recognition

  • Shengfei Lyu,
  • Linghao Sun,
  • Huixiong Yi,
  • Yong Liu,
  • Huanhuan Chen,
  • Chunyan Miao

DOI
https://doi.org/10.26599/IJCS.2023.9100014
Journal volume & issue
Vol. 8, no. 3
pp. 140 – 148

Abstract

Read online

In recent years, great success has been achieved in many tasks of natural language processing (NLP), e.g., named entity recognition (NER), especially in the high-resource language, i.e., English, thanks in part to the considerable amount of labeled resources. More labeled resources, better word representations. However, most low-resource languages do not have such an abundance of labeled data as high-resource English, leading to poor performance of NER in these low-resource languages due to poor word representations. In the paper, we propose converse attention network (CAN) to augment word representations in low-resource languages from the high-resource language, improving the performance of NER in low-resource languages by transferring knowledge learned in the high-resource language. CAN first translates sentences in low-resource languages into high-resource English using an attention-based translation module. In the process of translation, CAN obtains the attention matrices that align word representations of high-resource language space and low-resource language space. Furthermore, CAN augments word representations learned in low-resource language space with word representations learned in high-resource language space using the attention matrices. Experiments on four low-resource NER datasets show that CAN achieves consistent and significant performance improvements, which indicates the effectiveness of CAN.

Keywords