skip to main content
research-article

A Framework for Extractive Text Summarization Based on Deep Learning Modified Neural Network Classifier

Authors Info & Claims
Published:14 July 2021Publication History
Skip Abstract Section

Abstract

There is an exponential growth of text data over the internet, and it is expected to gain significant growth and attention in the coming years. Extracting meaningful insights from text data is crucially important as it offers value-added solutions to business organizations and end-users. Automatic text summarization (ATS) automates text summarization by reducing the initial size of the text without the loss of key information elements. In this article, we propose a novel text summarization algorithm for documents using Deep Learning Modifier Neural Network (DLMNN) classifier. It generates an informative summary of the documents based on the entropy values. The proposed DLMNN framework comprises six phases. In the initial phase, the input document is pre-processed. Subsequently, the features are extracted using pre-processed data. Next, the most appropriate features are selected using the improved fruit fly optimization algorithm (IFFOA). The entropy value for every chosen feature is computed. These values are then classified into two classes, (a) highest entropy values and (b) lowest entropy values. Finally, the class that holds the highest entropy values is chosen, representing the informative sentences that form the last summary. The results observed from the experiment indicate that the DLMNN classifier gives 81.56, 91.21, and 83.53 of sensitivity, accuracy, specificity, precision, and f-measure. Whereas the existing schemes such as ANN relatively provide lesser value in contrast to DLMNN.

References

  1. D. Muhammad Noorul Mubarak. 2016. An overview of extractive-based automatic text summarization systems. Int. J. Comput. Sci. Info. Technol. 8, 5 (2016).Google ScholarGoogle Scholar
  2. B. Laxmi Rananavare and P. Venkata Subba Reddy. 2017. An overview of text summarization. Int. J. Comput. Appl. 171, 10 (2017), 1–17.Google ScholarGoogle Scholar
  3. Abdelkrime Aries and Walid Khaled Hidouci. 2019. Automatic text summarization: What has been done and what has to be done. Comput. Lang. 2019.Google ScholarGoogle Scholar
  4. Aziz Qaroush, Ibrahim Abu Farha, Wasel Ghanem, Mahdi Washaha, and Eman Maali. 2019. An efficient single document arabic text summarization using a combination of statistical and semantic features. J. King Saud Univ.-Comput. Info. Sci. 2019.Google ScholarGoogle Scholar
  5. Wan-Ting Hsu, Chieh-Kai Lin, Ming-Ying Lee, Kerui Min, Jing Tang, and Min Sun. 2018. A unified model for extractive and abstractive summarization using inconsistency loss. Comput. Language. Retrieved from https://arxiv.org/abs/1805.06266.Google ScholarGoogle Scholar
  6. Yangbin Chen, Yun Ma, Xudong Mao, and Qing Li. 2019. Multi-task learning for abstractive and extractive summarization. Data Sci. Eng. 4, 1 (2019), 14–23.Google ScholarGoogle ScholarCross RefCross Ref
  7. Vishal Gupta and Gurpreet Singh Lehal. 2010. A survey of text summarization extractive techniques. J. Emerg. Technol. Web Intell. 2, 3 (2010), 258–268.Google ScholarGoogle Scholar
  8. N. Moratanch and S. Chitrakala. 2016. A survey on abstractive text summarization. In Proceedings of the International Conference on Circuit, Power and Computing Technologies (ICCPCT’16), IEEE, 1–7.Google ScholarGoogle Scholar
  9. Nedunchelian Ramanujam1 and Manivannan Kaliappan. 2016. An automatic multidocument text summarization approach based on naive bayesian classifier using timestamp strategy. Scientific World J. 10 (2016).Google ScholarGoogle Scholar
  10. Folajimi Yetunde and Tijesuni Obereke. 2016. Towards efficient model for automatic text summarization. In Proceedings of the International Conference on Computing Research and Innovations (OcRI,’16) 53–57.Google ScholarGoogle Scholar
  11. Nikita Munot and Sharvari S. Govilkar. 2014. Comparative study of text summarization methods. Int. J. Comput. Appl. 102, 12 (2014).Google ScholarGoogle Scholar
  12. Neelima Bhatia and Arunima Jaiswal. 2016. Automatic text summarization and it's methods-a review. In Proceedings of the 6th International Conference-Cloud System and Big Data Engineering (Confluence’16), IEEE, 65–72.Google ScholarGoogle ScholarCross RefCross Ref
  13. Mehdi Allahyari, Seyedamin Pouriyeh, Mehdi Assefi, Saeid Safaei, Elizabeth D. Trippe, Juan B. Gutierrez, and Krys Kochut. 2017. Text summarization techniques: A brief survey Retrieved from https://arxiv.org/abs/1707.02268.Google ScholarGoogle Scholar
  14. A. Qasem Al-Radaideh and Dareen Q. Bataineh. 2018. A hybrid approach for arabic text summarization using domain knowledge and genetic algorithms. Cogn. Comput. 10, 4 (2018), 651–669.Google ScholarGoogle ScholarCross RefCross Ref
  15. Htet Myet Lynn, Chang Choi, and Pankoo Kim. 2018. An improved method of automatic text summarization for web contents using lexical chain with semantic-related terms. Soft Comput. 22, 12 (2018), 4013–4023. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. E. Akanksha Joshi, Fidalgo E. Alegre, and Laura Fernández-Robles. 2019. SummCoder: An unsupervised framework for extractive text summarization based on deep auto-encoders. Expert Syst. Appl. 129, 200–215.Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Mozhgan Nasr Azadani, Nasser Ghadiri, and Ensieh Davoodijam. 2018. Graph-based biomedical text summarization: An itemset mining and sentence clustering approach. J. Biomed. Info. 84, 42–58.Google ScholarGoogle ScholarCross RefCross Ref
  18. Cheng-Ying Liu, Ming-Syan Chen, and Chi-Yao Tseng. 2015. Incrests: Towards real-time incremental short text summarization on comment streams from social network services. IEEE Trans. Knowl. Data Eng. 27, 11 (2015), 2986–3000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Changjian Fang, Dejun Mu, Zhenghong Deng, and Zhiang Wu. 2017. Word-sentence co-ranking for automatic extractive text summarization. Expert Syst. Appl. 72, 189–195. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. V. Priya and K. Umamaheswari. 2019. Enhanced continuous and discrete multi objective particle swarm optimization for text summarization. Cluster Comput. 22, 1 (2019), 229–240.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. A Framework for Extractive Text Summarization Based on Deep Learning Modified Neural Network Classifier

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format
    About Cookies On This Site

    We use cookies to ensure that we give you the best experience on our website.

    Learn more

    Got it!