Abstract
Fake news is risky, since it has been created to manipulate readers’ opinions and beliefs. In this work, we compared the language of false news to the real one of real news from an emotional perspective, considering a set of false information types (propaganda, hoax, clickbait, and satire) from social media and online news article sources. Our experiments showed that false information has different emotional patterns in each of its types, and emotions play a key role in deceiving the reader. Based on that, we proposed an LSTM neural network model that is emotionally infused to detect false news.
- Magda B. Arnold. 1960. Emotion and Personality. Columbia University Press.Google Scholar
- Gaurav Bhatt, Aman Sharma, Shivam Sharma, Ankush Nagpal, Balasubramanian Raman, and Ankush Mittal. 2018. Combining neural, statistical, and external features for fake news stance identification. In Proceedings of the Web Conference. International World Wide Web Conferences Steering Committee, 1353--1357.Google Scholar
Digital Library
- Carlos Castillo, Marcelo Mendoza, and Barbara Poblete. 2011. Information credibility on Twitter. In Proceedings of the 20th International Conference on World Wide Web. ACM, 675--684.Google Scholar
Digital Library
- Abhijnan Chakraborty, Bhargavi Paranjape, Sourya Kakarla, and Niloy Ganguly. 2016. Stop clickbait: Detecting and preventing clickbaits in online news media. In Proceedings of the IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining. IEEE Press, 9--16.Google Scholar
Cross Ref
- Jorge Carrillo de Albornoz, Laura Plaza, and Pablo Gervás. 2012. SentiSense: An easily scalable concept-based affective lexicon for sentiment analysis. In Proceedings of the International Conference on Language Resources and Evaluation (LREC’12). 3562--3567.Google Scholar
- Paul Ekman. 1992. An argument for basic emotions. Cog. Emot. 6, 3–4 (1992), 169--200.Google Scholar
Cross Ref
- Ethan Fast, Binbin Chen, and Michael S. Bernstein. 2016. Empath: Understanding topic signals in large-scale text. In Proceedings of the CHI Conference on Human Factors in Computing Systems. ACM, 4647--4657.Google Scholar
- Simona Frenda, Bilal Ghanem, and Manuel Montes-y Gòmez. 2018. Exploration of misogyny in Spanish and English Tweets. In Proceedings of the 3rd Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval’18), Vol. 2150. Ceur Workshop Proceedings, 260--267.Google Scholar
- Bilal Ghanem, Manuel Montes-y Gòmez, Francisco Rangel, and Paolo Rosso. 2018. UPV-INAOE-Autoritas—Check that: An approach based on external sources to detect claims credibility. In Proceedings of the Conference and Labs of the Evaluation Forum (CLEF ’18).Google Scholar
- Bilal Ghanem, Paolo Rosso, and Francisco Rangel. 2018. Stance detection in fake news a combined feature representation. In Proceedings of the 1st Workshop on Fact Extraction and VERification (FEVER’18). 66--71.Google Scholar
Cross Ref
- Andreas Hanselowski, Avinesh P. V. S., Benjamin Schiller, Felix Caspelherr, Debanjan Chaudhuri, Christian M. Meyer, and Iryna Gurevych. 2018. A retrospective analysis of the fake news challenge stance detection task. arXiv preprint arXiv:1806.05180 (2018).Google Scholar
- Delia Irazú Hernańdez Farías, Viviana Patti, and Paolo Rosso. 2016. Irony detection in Twitter: The role of affective content. ACM Trans. Internet Technol. 16, 3 (2016), 19.Google Scholar
- Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neur. Comput. 9, 8 (1997), 1735--1780.Google Scholar
Digital Library
- Georgi Karadzhov, Preslav Nakov, Lluís Màrquez, Alberto Barrón-Cedeño, and Ivan Koychev. 2017. Fully automated fact checking using external sources. In Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP’17). 344--353.Google Scholar
Cross Ref
- Alireza Karduni, Ryan Wesslen, Sashank Santhanam, Isaac Cho, Svitlana Volkova, Dustin Arendt, Samira Shaikh, and Wenwen Dou. 2018. Can you verifi this? Studying uncertainty and decision-making about misinformation using visual analytics. In Proceedings of the 12th International AAAI Conference on Web and Social Media.Google Scholar
- Elena Kochkina, Maria Liakata, and Arkaitz Zubiaga. 2018. All-in-one: Multi-task learning for rumour verification. arXiv preprint arXiv:1806.03713 (2018).Google Scholar
- Srijan Kumar, Robert West, and Jure Leskovec. 2016. Disinformation on the web: Impact, characteristics, and detection of Wikipedia hoaxes. In Proceedings of the 25th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, 591--602.Google Scholar
Digital Library
- Xian Li, Weiyi Meng, and Clement Yu. 2011. T-verifier: Verifying truthfulness of fact statements. In Proceedings of the IEEE 27th International Conference on Data Engineering (ICDE’11). IEEE, 63--74.Google Scholar
Digital Library
- Mario Livio. 2017. Why?: What Makes Us Curious. Simon and Schuster Publishing.Google Scholar
- Jing Ma, Wei Gao, Prasenjit Mitra, Sejeong Kwon, Bernard J. Jansen, Kam-Fai Wong, and Meeyoung Cha. 2016. Detecting rumors from microblogs with recurrent neural networks. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI’16). 3818--3824.Google Scholar
Digital Library
- Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. J. Mach. Learn. Res. 9, Nov. (2008), 2579--2605.Google Scholar
- Saif M. Mohammad and Peter D. Turney. 2010. Emotions evoked by common words and phrases: Using mechanical Turk to create an emotion lexicon. In Proceedings of the NAACL HLT Workshop on Computational Approaches to Analysis and Generation of Emotion in Text. Association for Computational Linguistics, 26--34.Google Scholar
- Brendan Nyhan and Jason Reifler. 2010. When corrections fail: The persistence of political misperceptions. Polit. Behav. 32, 2 (2010), 303--330.Google Scholar
Cross Ref
- W. Gerrod Parrott. 2001. Emotions in Social Psychology: Essential Readings. Psychology Press.Google Scholar
- Robert Plutchik. 2001. The nature of emotions: Human emotions have deep evolutionary roots, a fact that may explain their complexity and provide tools for clinical practice. Amer. Sci. 89, 4 (2001), 344--350.Google Scholar
Cross Ref
- Kashyap Popat, Subhabrata Mukherjee, Jannik Strötgen, and Gerhard Weikum. 2016. Credibility assessment of textual claims on the web. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management. ACM, 2173--2178.Google Scholar
Digital Library
- Kashyap Popat, Subhabrata Mukherjee, Andrew Yates, and Gerhard Weikum. 2018. DeClarE: Debunking fake news and false claims using evidence-aware deep learning. arXiv preprint arXiv:1809.06416 (2018).Google Scholar
- Soujanya Poria, Alexander Gelbukh, Amir Hussain, Newton Howard, Dipankar Das, and Sivaji Bandyopadhyay. 2013. Enhanced SenticNet with affective labels for concept-based opinion mining. IEEE Intel. Syst. 28, 2 (2013), 31--38.Google Scholar
Digital Library
- Vahed Qazvinian, Emily Rosengren, Dragomir R. Radev, and Qiaozhu Mei. 2011. Rumor has it: Identifying misinformation in microblogs. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 1589--1599.Google Scholar
Digital Library
- Colin Raffel and Daniel P. W. Ellis. 2015. Feed-forward networks with attention can solve some long-term memory problems. arXiv preprint arXiv:1512.08756 (2015).Google Scholar
- Francisco Rangel and Paolo Rosso. 2016. On the impact of emotions on author profiling. Inf. Proc. Manag. 52, 1 (2016), 73--92.Google Scholar
Digital Library
- Hannah Rashkin, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, and Yejin Choi. 2017. Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 2931--2937.Google Scholar
Cross Ref
- Victoria L. Rubin, Yimin Chen, and Niall J. Conroy. 2015. Deception detection for news: Three types of fakes. In Proceedings of the 78th ASIS8T Meeting: Information Science with Impact: Research in and for the Community. American Society for Information Science, 83.Google Scholar
- Natali Ruchansky, Sungyong Seo, and Yan Liu. 2017. CSI: A hybrid deep model for fake news detection. In Proceedings of the ACM Conference on Information and Knowledge Management. ACM, 797--806.Google Scholar
Digital Library
- Yla R. Tausczik and James W. Pennebaker. 2010. The psychological meaning of words: LIWC and computerized text analysis methods. J. Lang. Soc. Psychol. 29, 1 (2010), 24--54.Google Scholar
Cross Ref
- Svitlana Volkova, Kyle Shaffer, Jin Yea Jang, and Nathan Hodas. 2017. Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on Twitter. In Proceedings of the 55th Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 647--653.Google Scholar
Cross Ref
- Soroush Vosoughi, Deb Roy, and Sinan Aral. 2018. The spread of true and false news online. Science 359, 6380 (2018), 1146--1151.Google Scholar
- Po-Ya Angela Wang. 2013. # Irony or # Sarcasm—A quantitative and qualitative study based on Twitter. In Proceedings of the 27th Pacific Asia Conference on Language, Information, and Computation (PACLIC’13). 349--356.Google Scholar
- Savvas Zannettou, Michael Sirivianos, Jeremy Blackburn, and Nicolas Kourtellis. 2018. The web of false information: Rumors, fake news, hoaxes, clickbait, and various other shenanigans. arXiv preprint arXiv:1804.03461 (2018).Google Scholar
- Zhe Zhao, Paul Resnick, and Qiaozhu Mei. 2015. Enquiring minds: Early detection of rumors in social media from enquiry posts. In Proceedings of the 24th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, 1395--1405.Google Scholar
Digital Library
Index Terms
An Emotional Analysis of False Information in Social Media and News Articles
Recommendations
The Web of False Information: Rumors, Fake News, Hoaxes, Clickbait, and Various Other Shenanigans
Special Issue on Combating Digital Misinformation and Disinformation and On the HorizonA new era of Information Warfare has arrived. Various actors, including state-sponsored ones, are weaponizing information on Online Social Networks to run false-information campaigns with targeted manipulation of public opinion on specific topics. These ...
On the influence blocking maximization for minimizing the spreading of fake information in social media
SpringSim '20: Proceedings of the 2020 Spring Simulation ConferenceInfluence can be used to propagate the (fake or true) information in social media where a set of influential nodes (individuals) in social media can leverage their connections (e.g., followers in Tweeter) to impact others. Lately, most of the social ...
Predicting emotional reactions to news articles in social networks
Highlights- Novel method for predicting the distribution of emotions to be expected in a social network after the publication of a news article.
AbstractAfter reading a news article, some readers post their opinion to social networks, particularly as tweets. These opinions (responses) have an important emotional content. By analyzing users’ responses in context, it is possible to find ...






Comments