extractive text summarization, bidirectional encoder representations from transformers ( BERT), transformer, deep Q-learning (DQN) ,"/> 基于 Deep Q-Learning 的抽取式摘要生成方法

吉林大学学报(信息科学版) ›› 2023, Vol. 41 ›› Issue (2): 306-314.

• • 上一篇    下一篇

基于 Deep Q-Learning 的抽取式摘要生成方法

王灿宇1 , 孙晓海1,2 , 吴叶辉1 , 季荣彪1 , 李亚东1 , 张少如3 , 杨士豪   

  1.  (1. 云南农业大学 大数据学院, 昆明 650201; 2. 吉林海诚科技有限公司, 长春 130033; 3. 东北师范大学 信息科学与技术学院, 长春 130117) 
  • 收稿日期:2022-10-09 出版日期:2023-04-13 发布日期:2023-04-16
  • 通讯作者: 杨士豪(1995— ), 男, 山东梁山人, 东北师范大学博士研究生, 主要从事深度学习、 自然语言 处理研究, (Tel)86-18443148644(E-mai)yangsh861@ nenu. edu. cn
  • 作者简介:王灿宇(1979), 男, 云南鹤庆人, 云南农业大学副教授, 主要从事神经网络研究, (Tel)86-13518712079(E-mai)736559039 @ qq. com;
  • 基金资助:
    吉林省科技厅基金资助项目(20220201140GX); 长春市科技局基金资助项目(21ZY31)

 Generation Method of Extractive Text Summarization Based on Deep Q-Learning 

WANG Canyu 1 , SUN Xiaohai 1,2 , WU Yehui 1 , JI Rongbiao 1 , LI Yadong 1 , ZHANG Shaoru 3 , YANG Shihao 3   

  1. (1. College of Big Dated, Yunnan Agriculture University, Kunming 650201, China; 2. Jilin Haicheng Technology Company Limited, Changchun 130033, China 3. College of Information Science and Technology, Northeast Normal University, Changchun 130117, China)
  • Received:2022-10-09 Online:2023-04-13 Published:2023-04-16

摘要: 为解决训练过程中需要句子级标签的问题, 提出一种基于深度强化学习的无标签抽取式摘要生成 方法, 将文本摘要转化为 Q-learning 问题, 并利用 DQN( Deep Q-Network) 学习 Q 函数。 为有效表示文档, 利用BERT(Bidirectional Encoder Representations from Transformers)作为句子编码器, Transformer 作为文档编码 器。 解码器充分考虑了句子的信息富集度、 显著性、 位置重要性以及其与当前摘要之间的冗余程度等重要性 等信息。 该方法在抽取摘要时不需要句子级标签, 可显著减少标注工作量。 实验结果表明, 该方法在 CNN (Cable News Network) / DailyMail 数据集上取得了最高的 Rouge-L(38. 35) 以及可比较的 Rouge-1 (42. 07) 和 Rouge-2 (18. 32) 。

关键词: 抽取式文本摘要, BERT 模型, 编码器, 深度强化学习

Abstract: Extractive text summarization is a method of extracting key text fragments from the input text to serve as the summary. In order to solve the problem of requiring sentence-level labels during training, extractive text summarization is modeled as a Q-Learning problem and DQN(Deep Q-Network) to learn the Q value function. The document representation method is crucial for the quality of the generated summarization. To effectively represent the document, we adopt a hierarchical document representation method, which uses Bidirectional Encoder Representations from Transformers to obtain sentence-level vector representation and uses Transformer to obtain document-level vector representation. The decoder considers the sentence information enrichment, saliency, position, and redundancy degree between a sentence and the current summarization. This method does not require sentence-level labels when extracting sentences, which significantly reduces workload. Experiments on CNN( Cable News Network) / DailyMail data sets show that, compared with other extraction models, this model achieves the best Rouge-L(38. 35) and comparable Rouge-1(42. 07) and Rouge-2(18. 32) performance.

Key words: extractive text summarization')">

extractive text summarization, bidirectional encoder representations from transformers ( BERT), transformer, deep Q-learning (DQN)

中图分类号: 

  • TP391