This paper presents a deep context-aware model with a copy mechanism based on reinforcement learning for abstractive text summarization. Our model is optimized using weighted ROUGEs as global prediction-based rewards and the self-critical policy gradient training algorithm, which can reduce the inconsistency between training and testing by directly optimizing the evaluation metrics. To alleviate the lexical diversity and component diversity problems caused by global prediction rewards, we improve the richness of the multi-head self-attention mechanism to capture context through global deep context representation with copy mechanism. We conduct experiments and demonstrate that our model outperforms many existing benchmarks over the Gigaword, LCSTS, and CNN/DM datasets. The experimental results demonstrate that our model has a significant effect on improving the quality of summarization.
Siyao LiDeren LeiPengda QinWilliam Yang Wang
Figen Beken FikriKemal OflazerBerrin Yanıkoğlu
Rishank TambeDisha ThaokarEshika PachgharePrachi SawanePranay MehendolePriti Kakde