Sequence-to-Sequence Abstractive Text Summarization Model for Headline Generation with Attention

Authors

  • Shreya Regundwar, Radhika Bhagwat, Sakshi Bhosale, Rajlaxmi Chougale, Sakshi Abbu

Keywords:

Abstractive text summarization, Attention mechanism, Deep learning, Encoder-Decoder architecture, LSTM, Sequence-to-Sequence model, Single sentence summarization, Text preprocessing techniques

Abstract

Text summarization focuses on creating a brief and concise summary from source text while preserving the main idea and eliminating unnecessary details. Generating summaries through manual efforts by humans is a tedious, tiresome, and expensive process. Hence, this study’s objective is to build an automated abstractive text summarizer that can minimize manual efforts and generate concise summaries swiftly. The aim is to develop a text summarizer model using deep learning to form a single-line abstractive summary resembling a headline. It also explores the impact of adjusting the model's hyperparameters on the generated summary to achieve better results. A subset of instances from the Gigaword dataset is utilized to develop the model. The proposed summarizer is a sequence-to-sequence model with an LSTM-driven encoder-decoder architecture. It incorporates a Bahdanau attention mechanism and utilizes the Adam optimizer. Based on experimental analysis and the results obtained after adjusting hyperparameters and selecting the optimal values as final, the proposed architecture attained scores as 24.27, 8.57, and 23.13, for ROUGE-1, ROUGE-2, and ROUGE-L respectively.

Downloads

Download data is not yet available.

References

Suleiman D, Awajan A. Deep learning based abstractive text summarization: approaches, datasets, evaluation measures, and challenges. Mathematical problems in engineering. 2020 Aug 24;2020:1-29.

Dilawari A, Khan MU, Saleem S, Shaikh FS. Neural Attention Model for Abstractive Text Summarization Using Linguistic Feature Space. IEEE Access. 2023 Feb 27;11:23557-64.

Jobson E, Gutiérrez A. Abstractive text summarization using attentive sequence-to-sequence rnns.

Rush AM, Chopra S, Weston J. A neural attention model for abstractive sentence summarization. arXiv preprint arXiv:1509.00685. 2015 Sep 2.

Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473. 2014 Sep 1.

Parker, Robert, et al. English Gigaword Fifth Edition LDC2011T07. Web Download. Philadelphia: Linguistic Data Consortium, 2011.

Resnik P, Niv M, Nossal M, Schnitzer G, Stoner J, Kapit A, Toren R. Using intrinsic and extrinsic metrics to evaluate accuracy and facilitation in computer-assisted coding. InPerspectives in health information management computer assisted coding conference proceedings 2006 Aug (pp. 2006-2006).

Liu F, Liu Y. Exploring correlation between ROUGE and human evaluation on meeting summaries. IEEE Transactions on Audio, Speech, and Language Processing. 2009 Jun 10;18(1):187-96.

Jiang J, Zhang H, Dai C, Zhao Q, Feng H, Ji Z, Ganchev I. Enhancements of attention-based bidirectional lstm for hybrid automatic text summarization. IEEE Access. 2021 Sep 3;9:123660-71.

Li Z, Peng Z, Tang S, Zhang C, Ma H. Text summarization method based on double attention pointer network. IEEE Access. 2020 Jan 10;8:11279-88.

Hanunggul PM, Suyanto S. The impact of local attention in lstm for abstractive text summarization. In2019 International Seminar on Research of Information Technology and Intelligent Systems (ISRITI) 2019 Dec 5 (pp. 54-57). IEEE.

Nallapati R, Zhou B, Gulcehre C, Xiang B. Abstractive text summarization using sequence-to-sequence rnns and beyond. arXiv preprint arXiv:1602.06023. 2016 Feb 19.

Wong KF, Wu M, Li W. Extractive summarization using supervised and semi-supervised learning. InProceedings of the 22nd international conference on computational linguistics (Coling 2008) 2008 Aug (pp. 985-992).

Moratanch N, Chitrakala S. A survey on extractive text summarization. In2017 international conference on computer, communication and signal processing (ICCCSP) 2017 Jan 10 (pp. 1-6). IEEE.

Mridha MF, Lima AA, Nur K, Das SC, Hasan M, Kabir MM. A survey of automatic text summarization: Progress, process and challenges. IEEE Access. 2021 Nov 22;9:156043-70.

See A, Liu PJ, Manning CD. Get to the point: Summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368. 2017 Apr 14.

Shi T, Keneshloo Y, Ramakrishnan N, Reddy CK. Neural abstractive text summarization with sequence-to-sequence models. ACM Transactions on Data Science. 2021 Jan 3;2(1):1-37.

Sakhare DY. A Sequence-to-Sequence Text Summarization Using Long Short-Term Memory Based Neural Approach. International Journal of Intelligent Engineering & Systems. 2023 Mar 1;16(2).

Singh S, Singh JP, Deepak A. Deep Learning based Abstractive Summarization for English Language. InWorking Notes of FIRE 2022-Forum for Information Retrieval Evaluation, Kolkata, India 2022 Dec 9.

Kouris P, Alexandridis G, Stafylopatis A. Abstractive text summarization: Enhancing sequence-to-sequence models using word sense disambiguation and semantic content generalization. Computational Linguistics. 2021 Dec 23;47(4):813-59.

Wazery YM, Saleh ME, Alharbi A, Ali AA. Abstractive Arabic text summarization based on deep learning. Computational Intelligence and Neuroscience. 2022 Jan 11;2022.

Siddiqui T, Shamsi JA. Generating abstractive summaries using sequence to sequence attention model. In2018 International Conference on Frontiers of Information Technology (FIT) 2018 Dec 17 (pp. 212-217). IEEE.

Masum AK, Abujar S, Talukder MA, Rabby AS, Hossain SA. Abstractive method of text summarization with sequence to sequence RNNs. In2019 10th international conference on computing, communication and networking technologies (ICCCNT) 2019 Jul 6 (pp. 1-5). IEEE.

Rahman MM, Siddiqui FH. An optimized abstractive text summarization model using peephole convolutional LSTM. Symmetry. 2019 Oct 14;11(10):1290.

Anvitha Aravinda, Gururaja H S, Padmanabha J, Unique Combinations of LSTM for Text Summarization, International Journal of Engineering Research & Technology (IJERT) ICEI – 2022 (Volume 10 – Issue 11)

Sanjabi N. Abstractive text summarization with attention-based mechanism (Master's thesis, Universitat Politècnica de Catalunya).

Huang L, Wu H, Gao Q, Liu G. Attention Localness in Shared Encoder-Decoder Model For Text Summarization. InICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2023 Jun 4 (pp. 1-5). IEEE.

Parker, Robert, et al. English Gigaword Fifth Edition LDC2011T07. Web Download. Philadelphia: Linguistic Data Consortium, 2011.

Downloads

Published

16.03.2024

How to Cite

Sakshi Bhosale, Rajlaxmi Chougale, Sakshi Abbu, S. R. R. B. . (2024). Sequence-to-Sequence Abstractive Text Summarization Model for Headline Generation with Attention. International Journal of Intelligent Systems and Applications in Engineering, 12(3), 842–851. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/5363

Issue

Section

Research Article