Open Access Open Access  Restricted Access Subscription Access
Open Access Open Access Open Access  Restricted Access Restricted Access Subscription Access

Automated Summarization of Restaurant Reviews Using Hybrid Approaches


Affiliations
1 Department of Computer Science, Avinashilingam Institute for Home Science and Higher Education for Women, India
     

   Subscribe/Renew Journal


The arena of automatic text summarization incorporates the paramount and relevant information from a large document. This research paper attempts at representing two hybrid models for automatic text summarization. Extractive summarization followed by an abstractive summarization, is the strategy which is adopted in this paper to produce an informative and concise summary. The LexRank algorithm is used for extractive summarization, while BART (Bidirectional and Auto Regressive Transformers) and T5 (Text-To-Text Transfer Transformer) are used for abstractive summarization. BART and T5 are advanced per-trained models based on Transformer. The Transformer-based Per-trained models are causing a stir in the deep learning world. The first hybrid model is constructed using LexRank with BART (LRB) and the second hybrid model is constructed with LexRank with T5 (LRT). This specific approach will result in the generation of the extractive summary using the LexRank algorithm. The resulted output of LexRank is used as the input for BART and T5. The efficiency of two hybrid models is analyzed using qualitative and quantitative methods. The human-generated summary is used to evaluate the quality of the models, while the ROUGE score provides a quantitative assessment of their performance. Thus, this work may be concluded in the precision that, the LRT hybrid model is more effective than LRB hybrid model.

Keywords

BART, T5, LexRank, ROUGE Score, Extractive Summarization, Abstractive Summarization
Subscription Login to verify subscription
User
Notifications
Font Size

  • A. Nawaz, M. Bakhtyar and Abdul Basit, “Extractive Text Summarization Models for Urdu Language”, Information Processing and Management, Vol. 57, No. 6, pp. 1-13, 2020.
  • Vladislav Tretyak and Denis Stepanov, “Combination of Abstractive and Extractive Approaches for Summarization of Long Scientific Texts”, Proceedings of IEEE International Conference on Audio, Speech, and Language Processing, pp. 1-8, 2020.
  • X. Wang, “Summarization Based on Task-Oriented Discourse Parsing”, IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 23, No. 8, pp. 1358-1367, 2015.
  • R. Piryani, V. Gupta and V. Kumar Singh, “Generating Aspect-based Extractive Opinion Summary: Drawing Inferences from Social Media Texts”, Computacion Sistemas, Vol. 22, No. 1, pp. 83-91, 2020.
  • S.Chen, S. Liu, B. Chen and H. Wang, “An Information Distillation Framework for Extractive Summarization”, IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 26, No. 1, pp. 161-170, 2018.
  • H. Wang, L. Jing and H. Shao, “Research on Method of Sentence Similarity Based on Ontology”, Proceedings of Global Congress on Intelligent Systems, pp. 465-469, 2009.
  • D.K. Gaikwad and C. Namrata Mahender, “A Review Paper on Text Summarization”, International Journal of Advanced Research in Computer and Communication Engineering, Vol. 5, No. 3, pp. 1-13, 2016.
  • A. Chouigui, O. Ben Khiroun and B Elayeb, “An Arabic Multi-Source News Corpus: Experimenting on Single-Document Extractive Summarization”, Arabian Journal for Science and Engineering, Vol. 46, No. 4, pp. 3925-3938, 2021.
  • Divakar Yadav, Naman Lalit, Riya Kaushik, Yogendra Singh, Mohit, Dinesh, Arun Kr. Yadav, Kishor V. Bhadane, Adarsh Kumar and Baseem Khan, “Qualitative Analysis of Text Summarization Techniques and Its Applications in Health Domain”, Computational Intelligence and Neuroscience, Vol. 2022, pp. 1-14, 2022.
  • T. Wolf, L. Debut and J. Brew, “Hugging Face’s Transformers: State-of the-Art Natural Language Processing”, Clinical Orthopaedics and Related Research, Vol. 12, No. 3, pp. 1-13, 2019.
  • Y. Liu and V. Stoyanov, “RoBERTa: A robustly optimized BERT pretraining approach”, Proceedings of IEEE International Conference on Audio, Speech, and Language Processing, pp. 1-14, 2019.
  • M. Chang, K. Lee and K. Toutanova, “BERT: Pretraining of Deep Bidirectional Transformers for Language Understanding”, Proceedings of IEEE International Conference on Audio, Speech, and Language Processing, pp. 598-612, 2018.
  • C. Raffel, N. Shazeer and P.J. Liu, “Exploring the Limits of Transfer Learning with a Unified Text to-Text Transformer”, Clinical Orthopaedics and Related Research, Vol. 23, No. 1, pp. 1-12, 2019.
  • F. Zhuang, “A Comprehensive Survey on Transfer Learning”, Proceedings of the IEEE, Vol. 109, No. 1, pp. 43-76, 2021.
  • Y. Iwasaki, A. Yamashita, Y. Konno and K. Matsubayashi, “Japanese Abstractive Text Summarization using BERT”, Proceedings of International Conference on Technologies and Applications of Artificial Intelligence, pp. 1-5, 2019.
  • M. Farahani, M. Gharachorloo and M. Manthouri, “Leveraging ParsBERT and Pretrained mT5 for Persian Abstractive Text Summarization”, Proceedings of International Computer Conference, Computer Society of Iran, pp. 1-6, 2021
  • Yang Liu and Mirella Lapata, “Text Summarization with Pretrained Encoders”, Proceedings of International Conference on Technologies and Applications of Artificial Intelligence, pp. 34-45, 2019.
  • Mihir Kale and Abhinav Rastogi, “Text-to-Text Pre-Training for Data-to-Text Tasks”, Proceedings of International Conference on Technologies and Applications of Artificial Intelligence, pp. 143-154, 2020.
  • Ana Valdivia, M. Victoria Luzon and Francisco Herrera, “Sentiment Analysis in TripAdvisor”, Proceedings of International Conference on Affective Computing and Sentiment Analysis, pp. 1-13, 2019.
  • S. Neha, A Singh, I Raj and S Kumar, “An Evaluation for Various Text Summarization Algorithms on Blog Summarization Dataset”, International Journal of Scientific and Engineering Research, Vol 8, No. 11, pp.1-13, 2019.
  • S. Sidhpurwala, S. Jain and S Verma, “A Hybrid Approach for Text Summarization”, Topics in Intelligent Computing and Industry Design, Vol. 2, No. 2, pp. 142-146, 2020.
  • Varun Deokar and Kanishk Shah, “Automated Text Summarization of News Articles”, International Research Journal of Engineering and Technology, Vol. 8, No. 9, pp. 1-13, 2021.
  • Jiachen Xu, “Abstractive Summarization on COVID-19 Publications with BART”, Available at http://cs230.stanford.edu/projects_spring_2020/reports/38866168.pdf, Accessed at 2020
  • M. Lewis, Y. Liu, N. Goyal, M. Ghazvininejad, A. Mohamed, O. Levy, V.Stoyanov and L Zettlemoyer, “BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and
  • Comprehension”, Proceedings of IEEE International Conference on Natural Language, pp. 1-12, 2019.
  • J.L.E. Kedieng Fendji, D.M. Tair, M. Atemkeng and A Musa Ali, “WATS-SMS: A T5-based French Wikipedia Abstractive Text Summarizer for SMS”, Proceedings of IEEE International Conference on Natural Language, pp. 59-98, 2019.
  • C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li and P.J. Liu, “Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer”, Proceedings of IEEE International Conference on Natural Language, pp. 356-367, 2019.
  • K. Manju, S. David Peter and S. Mary Idicula, “A Framework for Generating Extractive Summary from Multiple Malayalam Documents”, Information, Vol. 12, No. 1, pp. 41-46, 2021.

Abstract Views: 72

PDF Views: 2




  • Automated Summarization of Restaurant Reviews Using Hybrid Approaches

Abstract Views: 72  |  PDF Views: 2

Authors

Shini George
Department of Computer Science, Avinashilingam Institute for Home Science and Higher Education for Women, India
V. Srividhya
Department of Computer Science, Avinashilingam Institute for Home Science and Higher Education for Women, India

Abstract


The arena of automatic text summarization incorporates the paramount and relevant information from a large document. This research paper attempts at representing two hybrid models for automatic text summarization. Extractive summarization followed by an abstractive summarization, is the strategy which is adopted in this paper to produce an informative and concise summary. The LexRank algorithm is used for extractive summarization, while BART (Bidirectional and Auto Regressive Transformers) and T5 (Text-To-Text Transfer Transformer) are used for abstractive summarization. BART and T5 are advanced per-trained models based on Transformer. The Transformer-based Per-trained models are causing a stir in the deep learning world. The first hybrid model is constructed using LexRank with BART (LRB) and the second hybrid model is constructed with LexRank with T5 (LRT). This specific approach will result in the generation of the extractive summary using the LexRank algorithm. The resulted output of LexRank is used as the input for BART and T5. The efficiency of two hybrid models is analyzed using qualitative and quantitative methods. The human-generated summary is used to evaluate the quality of the models, while the ROUGE score provides a quantitative assessment of their performance. Thus, this work may be concluded in the precision that, the LRT hybrid model is more effective than LRB hybrid model.

Keywords


BART, T5, LexRank, ROUGE Score, Extractive Summarization, Abstractive Summarization

References