Journal of Jilin University (Information Science Edition) ›› 2023, Vol. 41 ›› Issue (6): 1048-1053.
Previous Articles Next Articles
CHEN Xuesong, ZOU Meng
Received:
Online:
Published:
Abstract: To address the problem that traditional language models can not solve the problem of deep bidirectional representation and the problem that classification models can not adequately capture salient features of text, a text classification model based on BERT-BiGRU-CNN ( Bidirectional Encoder Representation from Transformers-Bidirectional Gating Recurrent Unit-Convolutional Neural Networks) is proposed. Firstly, the BERT pre-training model is used for text representation; secondly, the output data of BERT is input into BiGRU to capture the global semantic information of text. The results of BiGRU layer again are input into CNN to capture the local semantic features of text. Finally, the feature vectors are input into Softmax layer to obtain the classification results. The Chinese news text headlines dataset is used, and the experimental results show that the BERT-BiGRU-CNN based text classification model achieves an F1 value of 0. 948 5 on the dataset, which is better than other baseline models, proving that the BERT-BiGRU-CNN model can improve theshort text classification performance.
Key words: text classification, bidirectional encoder representation from transformers(BERT)word embedding, bidirectional gating recurrent unit(BiGRU), convolutional neural networks(CNN)
CLC Number:
CHEN Xuesong, ZOU Meng . Research on Short Text Classification Based on BERT-BiGRU-CNN Model[J].Journal of Jilin University (Information Science Edition), 2023, 41(6): 1048-1053.
0 / / Recommend
Add to citation manager EndNote|Reference Manager|ProCite|BibTeX|RefWorks
URL: http://xuebao.jlu.edu.cn/xxb/EN/
http://xuebao.jlu.edu.cn/xxb/EN/Y2023/V41/I6/1048
Cited