Attentive Encoder-based Extractive Text Summarization
| Authors |
|
|---|---|
| Publication date | 2018 |
| Book title | CIKM'18 |
| Book subtitle | proceedings of the 2018 ACM International Conference on Information and Knowledge Management : October 22-26, 2018, Torino, Italy |
| ISBN (electronic) |
|
| Event | 27th ACM International Conference on Information and Knowledge Management, CIKM 2018 |
| Pages (from-to) | 1499-1502 |
| Number of pages | 4 |
| Publisher | New York, NY: The Association for Computing Machinery |
| Organisations |
|
| Abstract |
In previous work on text summarization, encoder-decoder architectures and attention mechanisms have both been widely used. Attention-based encoder-decoder approaches typically focus on taking the sentences preceding a given sentence in a document into account for document representation, failing to capture the relationships between a sentence and sentences that follow it in a document in the encoder. We propose an attentive encoder-based summarization (AES) model to generate article summaries. AES can generate a rich document representation by considering both the global information of a document and the relationships of sentences in the document. A unidirectional recurrent neural network (RNN) and a bidirectional RNN are considered to construct the encoders, giving rise to unidirectional attentive encoder-based summarization (Uni-AES) and bidirectional attentive encoder-based summarization (Bi-AES), respectively. Our experimental results show that Bi-AES outperforms Uni-AES. We obtain substantial improvements over a relevant start-of-the-art baseline. |
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1145/3269206.3269251 |
| Other links | https://www.scopus.com/pages/publications/85058024539 |
| Permalink to this page | |
