Automatic Headline Generation for Hindi News using Fine-tuned Large Language Models

Authors

  • Jeetendra Kumar GLA University, Mathura, Uttar Pradesh, India
  • Shashi Shekhar GLA University, Mathura, Uttar Pradesh, India
  • Rashmi Gupta Atal Bihari Vajpayee University, Bilaspur, Chhattisgarh, India

Keywords:

Text Summarization, Fine-tuning, BART, indicBART, mT5

Abstract

Generating news headlines is one use case for automated text summarization. With only a few text lines, it creates a summary of the longer content, cutting down on reading time. Text summarization is a very challenging task and it is very difficult to generate summaries as a human being. Most summary tools available in the market primarily concentrate on summarizing English material, resulting in a scarcity of summarizers for other languages. In this study, we used two datasets gathered from Dainik Bhaskar, NavBharat Times, and one publically available dataset to fine-tune four pre-trained language models: Someman/bart-hindi, facebook/mbart-large-50, indicBART, and mT5. To conduct a full performance assessment, a variety of evaluation metrics are utilized. Our comprehensive examination consistently reveals that the facebook/mbart-large-50 model exhibits superior performance compared to other models in terms of these metrics. This highlights its potential to enhance automated summarization systems and facilitate enhanced content retrieval and comprehension within the Hindi-speaking community.

Downloads

Download data is not yet available.

References

A. P. Widyassari et al., “Review of automatic text summarization techniques & methods,” J. King Saud Univ. - Comput. Inf. Sci., vol. 34, no. 4, pp. 1029–1046, Apr. 2022, doi: 10.1016/J.JKSUCI.2020.05.006.

G. Sharma and D. Sharma, “Automatic Text Summarization Methods: A Comprehensive Review,” SN Comput. Sci., vol. 4, no. 1, pp. 1–18, Jan. 2023, doi: 10.1007/S42979-022-01446-W/METRICS.

Y. Kumar, K. Kaur, and S. Kaur, “Study of automatic text summarization approaches in different languages,” Artif. Intell. Rev., vol. 54, no. 8, pp. 5897–5929, Dec. 2021, doi: 10.1007/S10462-021-09964-4/METRICS.

A. V. Pradeepika Verma, “Accountability of NLP Tools in Text Summarization for Indian Languages,” J. Sci. Res., vol. 64, no. 1, pp. 358–363, 2020.

W. S. El-Kassas, C. R. Salama, A. A. Rafea, and H. K. Mohamed, “Automatic text summarization: A comprehensive survey,” Expert Syst. Appl., vol. 165, p. 113679, Mar. 2021, doi: 10.1016/J.ESWA.2020.113679.

A. K. Yadav et al., “Extractive text summarization using deep learning approach,” Int. J. Inf. Technol., vol. 14, no. 5, pp. 2407–2415, Aug. 2022, doi: 10.1007/S41870-022-00863-7/METRICS.

E. R. Mahalleh and F. S. Gharehchopogh, “An automatic text summarization based on valuable sentences selection,” Int. J. Inf. Technol., vol. 14, no. 6, pp. 2963–2969, Oct. 2022, doi: 10.1007/S41870-022-01049-X/METRICS.

S. Mandal, G. K. Singh, and A. Pal, “Single document text summarization technique using optimal combination of cuckoo search algorithm, sentence scoring and sentiment score,” Int. J. Inf. Technol., vol. 13, no. 5, pp. 1805–1813, Oct. 2021, doi: 10.1007/S41870-021-00739-2/METRICS.

P. J. Goutom, N. Baruah, and P. Sonowal, “An abstractive text summarization using deep learning in Assamese,” Int. J. Inf. Technol., vol. 15, no. 5, pp. 2365–2372, Jun. 2023, doi: 10.1007/S41870-023-01279-7/METRICS.

G. B. Mohan and R. P. Kumar, “Lattice abstraction-based content summarization using baseline abstractive lexical chaining progress,” Int. J. Inf. Technol., vol. 15, no. 1, pp. 369–378, Jan. 2023, doi: 10.1007/S41870-022-01080-Y/METRICS.

K. Rudra, S. Banerjee, N. Ganguly, P. Goyal, M. Imran, and P. Mitra, “Summarizing situational tweets in crisis scenario,” HT 2016 - Proc. 27th ACM Conf. Hypertext Soc. Media, pp. 137–147, Jul. 2016, doi: 10.1145/2914586.2914600.

I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to Sequence Learning with Neural Networks,” Adv. Neural Inf. Process. Syst., vol. 4, no. January, pp. 3104–3112, Sep. 2014, doi: 10.48550/arxiv.1409.3215.

V. Dalal and L. Malik, “Semantic graph based automatic text summarization for hindi documents using particle swarm optimization,” Smart Innov. Syst. Technol., vol. 84, pp. 284–289, 2018, doi: 10.1007/978-3-319-63645-0_31/COVER.

M. L. Joshi, N. Joshi, and N. Mittal, “SGATS: Semantic Graph-based Automatic Text Summarization from Hindi Text Documents,” Trans. Asian Low-Resource Lang. Inf. Process., vol. 20, no. 6, Sep. 2021, doi: 10.1145/3464381.

A. Jain, A. Arora, J. Morato, D. Yadav, and K. V. Kumar, “Automatic Text Summarization for Hindi Using Real Coded Genetic Algorithm,” Appl. Sci., vol. 12, no. 13, 2022, doi: 10.3390/app12136584.

S. Dhankhar and M. K. Gupta, “A statistically based sentence scoring method using mathematical combination for extractive Hindi text summarization,” J. Interdiscip. Math., vol. 25, no. 3, pp. 773–790, 2022, doi: 10.1080/09720502.2021.2015096.

S. S. Aote, A. Pimpalshende, A. Potnurwar, and S. Lohi, “Binary Particle Swarm Optimization with an improved genetic algorithm to solve multi-document text summarization problem of Hindi documents,” Eng. Appl. Artif. Intell., vol. 117, p. 105575, 2023, doi: https://doi.org/10.1016/j.engappai.2022.105575.

R. Bhargava, G. Sharma, and Y. Sharma, “Deep Text Summarization using Generative Adversarial Networks in Indian Languages,” Procedia Comput. Sci., vol. 167, pp. 147–153, Jan. 2020, doi: 10.1016/J.PROCS.2020.03.192.

M. Singh and V. Yadav, “Abstractive Text Summarization Using Attention-based Stacked LSTM,” pp. 236–241, Oct. 2022, doi: 10.1109/CCICT56684.2022.00052.

A. Shah, D. Zanzmera, and K. Mehta, “Deep Learning based Automatic Hindi Text Summarization,” Proc. - 6th Int. Conf. Comput. Methodol. Commun. ICCMC 2022, pp. 1455–1461, 2022, doi: 10.1109/ICCMC53470.2022.9753735.

S. S. Aote, A. Pimpalshende, A. Potnurwar, and S. Lohi, “Binary Particle Swarm Optimization with an improved genetic algorithm to solve multi-document text summarization problem of Hindi documents,” Eng. Appl. Artif. Intell., vol. 117, p. 105575, Jan. 2023, doi: 10.1016/J.ENGAPPAI.2022.105575.

S. Bandari and V. V. Bulusu, “Feature extraction based deep long short term memory for Hindi document summarization using political elephant herding optimization,” Int. J. Intell. Robot. Appl., vol. 7, no. 1, pp. 103–118, 2023, doi: 10.1007/s41315-022-00237-z.

R. Bhansali, A. Bhave, G. Bharat, V. Mahajan, and M. L. Dhore, “Abstractive Text Summarization of Hindi Corpus Using Transformer Encoder-Decoder Model,” Smart Innov. Syst. Technol., vol. 333, pp. 171–185, 2023, doi: 10.1007/978-981-19-8094-7_13/COVER.

D. Taunk and V. Varma, “Summarizing Indian Languages using Multilingual Transformers based Models,” CEUR Workshop Proc., vol. 3395, pp. 435–442, Mar. 2023, Accessed: Sep. 29, 2023. [Online]. Available: https://arxiv.org/abs/2303.16657v1

and S. S. Agarwal, Arjit, Soham Naik, “Abstractive Text Summarization for Hindi Language using IndicBART,” Work. Notes FIRE 2022-Forum Inf. Retr. Eval., 2022.

“Hindi Text Short Summarization Corpus | Kaggle.” https://www.kaggle.com/datasets/disisbig/hindi-text-short-summarization-corpus (accessed Feb. 02, 2023).

G. Arora, “iNLTK: Natural Language Toolkit for Indic Languages,” pp. 66–71, Sep. 2020, doi: 10.18653/v1/2020.nlposs-1.10.

R. Dabre, H. Shrotriya, A. Kunchukuttan, R. Puduppully, M. M. Khapra, and P. Kumar, “IndicBART: A Pre-trained Model for Indic Natural Language Generation,” Proc. Annu. Meet. Assoc. Comput. Linguist., vol. 2, pp. 1849–1863, 2022, doi: 10.18653/V1/2022.FINDINGS-ACL.145.

L. Xue et al., “mT5: A massively multilingual pre-trained text-to-text transformer,” NAACL-HLT 2021 - 2021 Conf. North Am. Chapter Assoc. Comput. Linguist. Hum. Lang. Technol. Proc. Conf., pp. 483–498, Oct. 2020, doi: 10.18653/v1/2021.naacl-main.41.

C.-Y. Lin, “ROUGE: A Package for Automatic Evaluation of Summaries,” in Text Summarization Branches Out, Barcelona, Spain: Association for Computational Linguistics, Jul. 2004, pp. 74–81. [Online]. Available: https://aclanthology.org/W04-1013

K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, “Bleu: a Method for Automatic Evaluation of Machine Translation,” Proc. 40th Annu. Meet. Assoc. Comput. Linguist. - ACL ’02, pp. 311–318, 2002, doi: 10.3115/1073083.1073135.

T. Zhang, V. Kishore, F. Wu, K. Q. Weinberger, and Y. Artzi, “Bertscore: Evaluating text generation with bert,” arXiv Prepr. arXiv1904.09675, 2019.

“GitHub - TechyNilesh/SSEM: SSEM is a semantic similarity-based evaluation library for natural language processing (NLP) text generation tasks. It supports various similarity metrics and evaluation levels, and is compatible with any Hugging Face pre-trained transformer model.” https://github.com/TechyNilesh/SSEM (accessed Sep. 25, 2023).

Downloads

Published

25.12.2023

How to Cite

Kumar, J. ., Shekhar, S. ., & Gupta, R. . (2023). Automatic Headline Generation for Hindi News using Fine-tuned Large Language Models. International Journal of Intelligent Systems and Applications in Engineering, 12(2), 391–399. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/4282

Issue

Section

Research Article