Journal of Jilin University (Information Science Edition) ›› 2023, Vol. 41 ›› Issue (5): 866-875.

Previous Articles     Next Articles

Ancient Chinese Named Entity Recognition Based on SikuBERT Model and MHA

CHEN Xuesong a , ZHAN Ziyi a , WANG Haochang b   

  1. a. School of Electrical and Information Engineering; b. School of Computer and Information Technology, Northeast Petroleum University, Daqing 163318, China
  • Received:2022-09-28 Online:2023-10-09 Published:2023-10-10

Abstract:

Aiming at the problem that the traditional named entity recognition method can not fully learn the complex sentence structure information of ancient Chinese and it is easy to cause information loss in the process of long sequence feature extraction, an ancient Chinese fusion of SikuBERT ( Siku Bidirectional Encoder Representation from Transformers) model and MHA (Multi-Head Attention) is proposed. First, the SikuBERT model is used to pre-train the ancient Chinese corpus, the information vector obtained from the training into the BiLSTM (Bidirectional Long Short-Term Memory) network is input to extract features, and then the output features of the BiLSTM layer are assigned different weights through MHA to reduce the information loss problem of long sequences. And finally the predicted sequence labels are obtained through CRF (Conditional Random Field) decoding. Experiments show that compared with commonly used BiLSTM-CRF, BERT-BiLSTM-CRF and other models, the F1 value of this method has been significantly improved, which verifies that this method can effectively improve the effect of ancient Chinese named entity recognition.

Key words: ancient Chinese, named entity recognition, siku bidirectional encoder representation from transformers(SikuBERT) model, multi-head attention mechanism

CLC Number: 

  • TP391. 1