skip to main content
research-article

Persian Fake News Detection: Neural Representation and Classification at Word and Text Levels

Authors Info & Claims
Published:01 November 2021Publication History
Skip Abstract Section

Abstract

Nowadays, broadcasting news on social media and websites has grown at a swifter pace, which has had negative impacts on both the general public and governments; hence, this has urged us to build a fake news detection system. Contextualized word embeddings have achieved great success in recent years due to their power to embed both syntactic and semantic features of textual contents. In this article, we aim to address the problem of the lack of fake news datasets in Persian by introducing a new dataset crawled from different news agencies, and propose two deep models based on the Bidirectional Encoder Representations from Transformers model (BERT), which is a deep contextualized pre-trained model for extracting valuable features. In our proposed models, we benefit from two different settings of BERT, namely pool-based representation, which provides a representation for the whole document, and sequence representation, which provides a representation for each token of the document. In the former one, we connect a Single Layer Perceptron (SLP) to the BERT to use the embedding directly for detecting fake news. The latter one uses Convolutional Neural Network (CNN) after the BERT’s embedding layer to extract extra features based on the collocation of words in a corpus. Furthermore, we present the TAJ dataset, which is a new Persian fake news dataset crawled from news agencies’ websites. We evaluate our proposed models on the newly provided TAJ dataset as well as the two different Persian rumor datasets as baselines. The results indicate the effectiveness of using deep contextualized embedding approaches for the fake news detection task. We also show that both BERT-SLP and BERT-CNN models achieve superior performance to the previous baselines and traditional machine learning models, with 15.58% and 17.1% improvement compared to the reported results by Zamani et al. [30], and 11.29% and 11.18% improvement compared to the reported results by Jahanbakhsh-Nagadeh et al. [9].

REFERENCES

  1. [1] Ahmed Hadeer, Traore Issa, and Saad Sherif. 2017. Detection of online fake news using n-gram analysis and machine learning techniques. In International Conference on Intelligent, Secure, and Dependable Systems in Distributed and Cloud Environments. Springer, 127138.Google ScholarGoogle Scholar
  2. [2] Ahmed Hadeer, Traore Issa, and Saad Sherif. 2018. Detecting opinion spams and fake news using text classification. Security and Privacy 1, 1 (2018), e9.Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Chawla Nitesh V., Bowyer Kevin W., Hall Lawrence O., and Kegelmeyer W. Philip. 2002. SMOTE: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research 16 (2002), 321357.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. [4] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, MN, 41714186. DOI: DOI: https://doi.org/10.18653/v1/N19-1423Google ScholarGoogle Scholar
  5. [5] Goldani Mohammad, Momtazi Saeedeh, and Safabakhsh Reza. 2021. Detecting fake news with capsule neural networks. Applied Soft Computing 101 (2021), 106991. DOI: DOI: https://doi.org/10.1016/j.asoc.2020.106991Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Goldani Mohammad Hadi, Safabakhsh Reza, and Momtazi Saeedeh. 2021. Convolutional neural network with margin loss for fake news detection. Information Processing & Management 58, 1 (2021), 102418.Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Granik Mykhailo and Mesyura Volodymyr. 2017. Fake news detection using naive Bayes classifier. In 2017 IEEE 1st Ukraine Conference on Electrical and Computer Engineering (UKRCON’17). IEEE, 900903.Google ScholarGoogle Scholar
  8. [8] Isotani Ryosuke and Matsunaga Shoichi. 1994. Speech recognition using a stochastic language model integrating local and global constraints. In Human Language Technology: Proceedings of a Workshop held at Plainsboro, New Jersey. https://www.aclweb.org/anthology/H94-1015.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Jahanbakhsh-Nagadeh Zoleikha, Feizi-Derakhshi Mohammad-Reza, Ramezani Majid, Rahkar-Farshi Taymaz, Asgari-Chenaghlu Meysam, Nikzad-Khasmakhi Narjes, Feizi-Derakhshi Ali-Reza, Ranjbar-Khadivi Mehrdad, Zafarani-Moattar Elnaz, and Balafar Mohammad-Ali. 2020. A model to measure the spread power of rumors. arXiv preprint arXiv:2002.07563 (2020).Google ScholarGoogle Scholar
  10. [10] Khattar Dhruv, Goud Jaipal Singh, Gupta Manish, and Varma Vasudeva. 2019. Mvae: Multimodal variational autoencoder for fake news detection. In The World Wide Web Conference. 29152921.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. [11] Liu Yang and Lapata Mirella. 2019. Text summarization with pretrained encoders. arXiv preprint arXiv:1908.08345 (2019).Google ScholarGoogle Scholar
  12. [12] Liu Zhan, Shabani Shaban, Balet Nicole Glassey, and Sokhn Maria. 2019. Detection of satiric news on social media: Analysis of the phenomenon with a French dataset. In 2019 28th International Conference on Computer Communication and Networks (ICCCN’19). 16.Google ScholarGoogle Scholar
  13. [13] Mahmoodabad Sajjad Dehghani, Farzi Saeed, and Bakhtiarvand Danial Bidekani. 2018. Persian rumor detection on twitter. In 2018 9th International Symposium on Telecommunications (IST’18). IEEE, 597602.Google ScholarGoogle Scholar
  14. [14] Mikolov Tomas, Sutskever Ilya, Chen Kai, Corrado Greg S., and Dean Jeff. 2013. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems. 31113119.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. [15] Momtazi Saeedeh and Klakow Dietrich. 2010. Hierarchical Pitman-Yor language model for information retrieval. In Proceedings of the 33rd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’10). Association for Computing Machinery, New York, NY, 793794. DOI: DOI: https://doi.org/10.1145/1835449.1835619Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. [16] Munikar M., Shakya S., and Shrestha A.. 2019. Fine-grained sentiment classification using BERT. In 2019 Artificial Intelligence for Transforming Business and Society (AITB’19), Vol. 1. 15.Google ScholarGoogle Scholar
  17. [17] Pennington Jeffrey, Socher Richard, and Manning Christopher D.. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP’14). 15321543.Google ScholarGoogle ScholarCross RefCross Ref
  18. [18] Peters Matthew, Neumann Mark, Iyyer Mohit, Gardner Matt, Clark Christopher, Lee Kenton, and Zettlemoyer Luke. 2018. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). Association for Computational Linguistics, New Orleans, LA, 22272237. DOI: DOI: https://doi.org/10.18653/v1/N18-1202Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Ponte Jay M. and Croft W. Bruce. 1998. A language modeling approach to information retrieval. In Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’98). Association for Computing Machinery, New York, NY, 275281. DOI: DOI: https://doi.org/10.1145/290941.291008Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Reyes Jesús and Palafox Leon. 2019. Detection of fake news based on readability. In Proceedings of Reunión Internacional de Inteligencia Artificial y sus Aplicaciones.Google ScholarGoogle Scholar
  21. [21] Rubin Victoria L., Chen Yimin, and Conroy Nadia K.. 2015. Deception detection for news: Three types of fakes. Proceedings of the Association for Information Science and Technology 52, 1 (2015), 14.Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Silverman Craig, Strapagiel Lauren, Shaban Hamza, Hall Ellie, and Singer-Vine Jeremy. 2016. Hyperpartisan Facebook pages are publishing false and misleading information at an alarming rate. Buzzfeed News 20 (2016).Google ScholarGoogle Scholar
  23. [23] Taher Ehsan, Hoseini Seyed Abbas, and Shamsfard Mehrnoush. 2019. Beheshti-NER: Persian named entity recognition using BERT. In Proceedings of International Conference on Natural Language and Speech Processing, Trento, Italy.Google ScholarGoogle Scholar
  24. [24] Trnka Keith. 2008. Adaptive language modeling for word prediction. In Proceedings of the ACL-08: HLT Student Research Workshop. Association for Computational Linguistics, Columbus, OH, 6166. https://www.aclweb.org/anthology/P08-3011.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Vicario Michela Del, Quattrociocchi Walter, Scala Antonio, and Zollo Fabiana. 2019. Polarization and fake news: Early warning of potential misinformation targets. ACM Transactions on the Web (TWEB) 13, 2 (2019), 122.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Vogel Inna and Jiang Peter. 2019. Fake news detection with the new German dataset “GermanFakeNC.” In International Conference on Theory and Practice of Digital Libraries. Springer, 288295.Google ScholarGoogle Scholar
  27. [27] Wang William Yang. 2017. “Liar, liar pants on fire”: A new benchmark dataset for fake news detection. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Association for Computational Linguistics, Vancouver, Canada, 422426. DOI: DOI: https://doi.org/10.18653/v1/P17-2067Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Wang Zhiguo, Ng Patrick, Ma Xiaofei, Nallapati Ramesh, and Xiang Bing. 2019. Multi-passage BERT: A globally normalized BERT model for open-domain question answering. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP’19). Association for Computational Linguistics, Hong Kong, China, 58785882. DOI: DOI: https://doi.org/10.18653/v1/D19-1599Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Yang Jiacheng, Wang Mingxuan, Zhou Hao, Zhao Chengqi, Zhang Weinan, Yu Yong, and Li Lei. 2020. Towards making the most of BERT in neural machine translation. Proceedings of the AAAI Conference on Artificial Intelligence 34 (2020), 93789385. DOI: DOI: https://doi.org/10.1609/aaai.v34i05.6479Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Zamani Somayeh, Asadpour Masoud, and Moazzami Dara. 2017. Rumor detection for persian tweets. In 2017 Iranian Conference on Electrical Engineering (ICEE’17). IEEE, 15321536.Google ScholarGoogle Scholar
  31. [31] Zarharan Majid, Ahangar Samane, Rezvaninejad Fateme, Bidhendi Mahdi, Pilevar Mohammad Taher, Minaei Behrouz, and Eetemadi Sauleh. 2019. Persian stance classification data set. In Conference For Truth and Trust Online. DOI: DOI: https://doi.org/10.36370/tto.2019.30Google ScholarGoogle Scholar

Index Terms

  1. Persian Fake News Detection: Neural Representation and Classification at Word and Text Levels

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Asian and Low-Resource Language Information Processing
      ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 21, Issue 1
      January 2022
      442 pages
      ISSN:2375-4699
      EISSN:2375-4702
      DOI:10.1145/3494068
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 1 November 2021
      • Accepted: 1 June 2021
      • Revised: 1 March 2021
      • Received: 1 November 2020
      Published in tallip Volume 21, Issue 1

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format
    About Cookies On This Site

    We use cookies to ensure that we give you the best experience on our website.

    Learn more

    Got it!