José Ángel GonzálezEncarna SegarraFernando GarcíaEmilio SanchísLluís-F. Hurtado
In this paper, we present an extractive approach to document summarization, the Siamese Hierarchical Transformer Encoders system, that is based on the use of siamese neural networks and the transformer encoders which are extended in a hierarchical way. The system, trained for binary classification, is able to assign attention scores to each sentence in the document. These scores are used to select the most relevant sentences to build the summary. The main novelty of our proposal is the use of self-attention mechanisms at sentence level for document summarization, instead of using only attentions at word level. The experimentation carried out using the CNN/DailyMail summarization corpus shows promising results in-line with the state-of-the-art.
José Ángel GonzálezEncarna SegarraFernando GarcíaEmilio SanchísLlu rsquo ıs-F. Hurtado
K ArjunM. HariharanPooja AnandV. PradeepReshma RajAnuraj Mohan
Luca BaccoAndrea CiminoFelice Dell’Orletta⋄Mario Merone
Ella Hofmann-CoyleMayank KulkarniLingjue XieMounica MaddelaDaniel Preoţiuc-Pietro
Shihao YangShaoru ZhangMing FangFengqin YangShuhua Liu