We apologize for a recent technical issue with our email system, which temporarily affected account activations. Accounts have now been activated. Authors may proceed with paper submissions. PhDFocusTM
CFP last date
20 December 2024
Reseach Article

Encoder Decoder based Nepali News Headline Generation

by Kaushal Raj Mishra, Jayshree Rathi, Janardan Banjara
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 175 - Number 20
Year of Publication: 2020
Authors: Kaushal Raj Mishra, Jayshree Rathi, Janardan Banjara
10.5120/ijca2020920735

Kaushal Raj Mishra, Jayshree Rathi, Janardan Banjara . Encoder Decoder based Nepali News Headline Generation. International Journal of Computer Applications. 175, 20 ( Sep 2020), 1-4. DOI=10.5120/ijca2020920735

@article{ 10.5120/ijca2020920735,
author = { Kaushal Raj Mishra, Jayshree Rathi, Janardan Banjara },
title = { Encoder Decoder based Nepali News Headline Generation },
journal = { International Journal of Computer Applications },
issue_date = { Sep 2020 },
volume = { 175 },
number = { 20 },
month = { Sep },
year = { 2020 },
issn = { 0975-8887 },
pages = { 1-4 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume175/number20/31565-2020920735/ },
doi = { 10.5120/ijca2020920735 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-02-07T00:25:32.776933+05:30
%A Kaushal Raj Mishra
%A Jayshree Rathi
%A Janardan Banjara
%T Encoder Decoder based Nepali News Headline Generation
%J International Journal of Computer Applications
%@ 0975-8887
%V 175
%N 20
%P 1-4
%D 2020
%I Foundation of Computer Science (FCS), NY, USA
Abstract

In this paper, a method for Nepali News Headline Generation is presented. The proposed method uses GRUs, in an encoder-decoder fashion, taking the news content as input and generating a headline as the output. The news is converted into word tokens, which are vectorized using FastText, trained on a corpus of Nepali news articles and headlines collected from several web portals. The headline generation model is also trained on the same corpus. A sequence to sequence model, with an encoder and a decoder GRU is used as the generation model. The model was able to attain a BLEU score of 22.1 on the test set.

References
  1. Text summarization: An extractive approach. In Soft Computing: Theories and Applications, pages 629–637, Singapore, 2020. Springer Singapore.
  2. P. B. Baxendale. Machine-made index for technical literaturean experiment. IBM Journal of Research and Development, 2(4):354–361, 1958.
  3. Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. Enriching word vectors with subword information. Transactions of the Association for Computational Linguistics, 5:135–146, 2017.
  4. Latesh G. Malik Chetana Thaokar. Test model for summarizing hindi text using extraction method. IEEE Conference on Information and Communication Technologies, pages 1138– 1143, 2013.
  5. Kyunghyun Cho, Bart Van Merrinboer, Caglar Gulcehre, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. 2014.
  6. Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv.org, 2014.
  7. Apurva Khude Dipali Telavane. Automatic summarization of hindi text documents using supervised learning method. International Journal for Research in Engineering Application Management, 4(10), 2019.
  8. Akshay Kishore Gattani. Automated natural language headline generation using discriminative machine learning models. PhD thesis, School of Computing Science-Simon Fraser University, 2007.
  9. D. Jain and G. Cardona. The indo-aryan languages. 2007.
  10. Konstantin Lopyrev. Generating news headlines with recurrent neural networks. 12 2015.
  11. H. P. Luhn. The automatic creation of literature abstracts. IBM J. Res. Dev., 2:159–165, 1958.
  12. R. Kathleen Mckeown and R. Dragomir Radev. Generating summaries of multiple news articles. SIGIR, pages 74–82, 1995.
  13. Joel Larocca Neto, Alex A. Freitas, and Celso A. A. Kaestner. Automatic text summarization using a machine learning approach. In Guilherme Bittencourt and Geber L. Ramalho, editors, Advances in Artificial Intelligence, pages 205–215, Berlin, Heidelberg, 2002. Springer Berlin Heidelberg.
  14. Kishore Papineni, Salim Roukos, Todd Ward, and Wei jing Zhu. Bleu: a method for automatic evaluation of machine translation. pages 311–318, 2002.
  15. M. Schuster and K. K. Paliwal. Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, 45(11):2673–2681, 1997.
  16. Ilya Sutskever, Oriol Vinyals, and Quoc V Le. Sequence to sequence learning with neural networks. In Advances in neural information processing systems, pages 3104–3112, 2014.
Index Terms

Computer Science
Information Sciences

Keywords

Recurrent Neural Network Gated Recurrent Unit FastText Bilingual Evaluation Understudy (BLEU) Encoder Decoder