Abstract
The emerging uptake of AI chatbots for social support entails systematic comparisons between human and non-human entities as sources of support. In a between-subject experimental study, a human and two types of ostensible chatbots (using a wizard of oz design) had supportive conversations with college students who were experiencing stressful situations during the pandemic. We found that when compared with a less ideal chatbot (i.e., low-contingent chatbot), (1) the human support provider was perceived with more warmth, which directly reduced emotional distress among participants; (2) the ideal chatbot (i.e., high-contingent chatbot) was perceived to be more competent, which activated participants' cognitive reappraisal of their stressful situations and subsequently reduced emotional distress. The human provider and the ideal chatbot did not differ in users' perceived competence or warmth, although the human provider was more effective at activating participants' cognitive reappraisal. This study integrates human communication theories into human-computer interaction work and contributes by positioning and theorizing user perceptions of chatbots in a larger process from support sources with varying communication competence to users' cognitive and emotional responses, and ultimately to the stress outcome. Theoretical and design implications are discussed.
- Tamara D. Afifi, Anne F. Merrill, and Sharde Davis. 2016. The theory of resilience and relational load. Personal Relationships 23, 4 (2016), 663--683.Google Scholar
Cross Ref
- Tim Althoff, Kevin Clark, and Jure Leskovec. 2016. Large-scale analysis of counseling conversations: An application of natural language processing to mental health. Transactions of the Association for Computational Linguistics 4 (2016), 463--476.Google Scholar
Cross Ref
- Roy F. Baumeister. 1997. Esteem threat, self-regulatory breakdown, and emotional distress as factors in self-defeating behavior. Review of General Psychology 1, 2 (1997), 145--174.Google Scholar
Cross Ref
- Kirsten Bergmann, Friederike Eyssel, and Stefan Kopp. 2012. A second chance to make a first impression? How appearance and nonverbal behavior affect perceived warmth and competence of virtual agents over time. In Proceedings of the International Conference on Intelligent Virtual Agents. 126--138.Google Scholar
Digital Library
- Timothy Bickmore and Daniel Schulman. 2007. Practical approaches to comforting users with relational agents. In Proceedings of CHI'07 Extended Abstracts on Human Factors in Computing Systems. 2291--2296.Google Scholar
Digital Library
- Timothy Bickmore and Daniel Schulman. 2009. A virtual laboratory for studying long-term relationships between humans and virtual agents. In Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems. 297--304.Google Scholar
- Timothy W. Bickmore, Daniel Schulman, and Candace Sidner. 2013. Automated interventions for multiple health behaviors using conversational agents. Patient Education and Counseling 92, 2 (2013), 142--148.Google Scholar
Cross Ref
- Graham D. Bodie. 2011. The role of thinking in the comforting process: An empirical test of a dual-process framework. Communication Research 40, 4 (Dec. 2011), 533--558.Google Scholar
- Petter Bae Brandtzæg, Marita Skjuve, Kim Kristoffer Kristoffer Dysthe, and Asbjørn Følstad. 2021. When the social becomes non-human: young people's perception of social support in chatbots. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1--13.Google Scholar
Digital Library
- Sharon S. Brehm and Jack W. Brehm. 2013. Psychological Reactance: A Theory of Freedom and Control. Academic Press.Google Scholar
- Julia E.H. Brown and Jodi Halpern. 2021. AI chatbots cannot replace human interactions in the pursuit of more inclusive mental healthcare. SSM - Mental Health 1 (2021), 100017. DOI: https://doi.org/10.1016/j.ssmmh.2021.100017Google Scholar
Cross Ref
- Brant R. Burleson and Daena J.Goldsmith. 1998. How the comforting process works: Alleviating emotional distress through conversationally induced reappraisals. In P. A. Andersen & L. K. Guerrero (Eds.), Handbook of communication and emotion: Research, theory, applications, and contexts (pp. 245--280). Academic Press.Google Scholar
- Brant R. Burleson and Erina L. MacGeorge. 2002. Supportive communication. In M. L. Knapp & J. A. Daly (Eds.), Handbook of interpersonal communication (3rd ed., pp. 374--424). Sage, Thousand Oaks, CA.Google Scholar
- Brant R. Burleson. 2008. What counts as effective emotional support? Explorations of individual and situational differences. Studies in Applied Interpersonal Communication. 207--227.Google Scholar
- Heloisa Candello, Claudio Pinhanez, and Flavio Figueiredo. 2017. Typefaces and the perception of humanness in natural language chatbots. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. 3476--3487. https://doi.org/10.1145/3025453.3025919Google Scholar
Digital Library
- Astrid Carolus, Catharina Schmidt, Florian Schneider, Jule Mayr, and Ricardo Muench. 2018. Are people polite to smartphones? In M. Kurosu (Ed.), Human-Computer Interaction. Interaction in Context (pp. 500--511). Springer International Publishing. DOI: https://doi.org/10.1007/978--3--319--91244--8_39Google Scholar
Cross Ref
- Mauro de Gennaro, Eva G. Krumhuber, and Gale Lucas. 2020. Effectiveness of an empathic chatbot in combating adverse effects of social exclusion on mood. Frontiers in Psychology 10 (2020), 3061. DOI: https://doi.org/10.3389/fpsyg.2019.03061Google Scholar
Cross Ref
- Virginie Demeure, Radoslaw Niewiadomski, and Catherine Pelachaud. 2011. How Is believability of a virtual agent related to warmth, competence, personification, and embodiment? Presence 20, 5 (Oct. 2011), 431--448. DOI: https://doi.org/10.1162/PRES_a_00065Google Scholar
Digital Library
- Michelle Drouin, Susan Sprecher, Robert Nicola, and Taylor Perkins. 2021. Is chatting with a sophisticated chatbot as good as chatting online or FTF with a stranger?. Computers in Human Behavior 128 (2021), 107100. DOI: https://doi.org/10.1016/j.chb.2021.107100Google Scholar
Digital Library
- Nicholas Epley, Adam Waytz, and John T. Cacioppo. 2007. On seeing human: a three-factor theory of anthropomorphism. Psychological Review 114, 4 (2007), 864.Google Scholar
Cross Ref
- Friederike Eyssel and Frank Hegel. 2012. (s) he's got the look: Gender stereotyping of robots 1. Journal of Applied Social Psychology 42, 9 (2012), 2213--2230. DOI: https://doi.org/10.1111/j.1559--1816.2012.00937.xGoogle Scholar
Cross Ref
- Susan T. Fiske, Amy J.C. Cuddy, and Peter Glick. 2006. Universal dimensions of social cognition: Warmth and competence. Trends in Cognitive Sciences 11, 2 (2006). 77--83.Google Scholar
- Andrew J. Flanagin and Miriam J. Metzger. 2003. The perceived credibility of personal Web page information as influenced by the sex of the source. Computers in Human Behavior 19, 6 (2003), 683--701. DOI: https://doi.org/10.1016/S0747--5632(03)00021--9Google Scholar
Cross Ref
- Brian J. Fogg. 2002. Persuasive technology: using computers to change what we think and do. Ubiquity 2002, Article 5 (2002), 32 pages. DOI: https://doi.org/10.1145/764008.763957Google Scholar
Digital Library
- Keith Frankish and William M. Ramsey. 2014. The Cambridge Handbook of Artificial Intelligence (Eds.). Cambridge University Press. Cambridge, England.Google Scholar
- Andrew Gambino, Jesse Fox, and Rabindra A. Ratan. 2020. Building a stronger CASA: Extending the computers are social actors paradigm. Human-Machine Communication 1 1(2020), 71--85. DOI: https://doi.org/10.30658/hmc.1.5Google Scholar
Cross Ref
- Zohar Gilad, Ofra Amir, and Liat Levontin. 2021. The Effects of Warmth and Competence Perceptions on Users' Choice of an AI System. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1--13.Google Scholar
Digital Library
- Kurt Gray and Daniel M. Wegner. 2012. Feeling robots and human zombies: Mind perception and the uncanny valley. Cognition 125 (2012), 125--130.Google Scholar
Cross Ref
- Andrew F Hayes. 2017. Introduction to mediation, moderation, and conditional process analysis: A regression-based approach. Guilford Publications.Google Scholar
- Annabell Ho, Jeff Hancock, and Adam S. Miner. 2018. Psychological, relational, and emotional effects of self-disclosure after conversations with a chatbot. Journal of Communication 68, 4(2018), 712--733. DOI: https://doi.org/10.1093/joc/jqy026Google Scholar
Cross Ref
- Laura Hoffmann, Melanie Derksen, Stefan Kopp. 2020. What a pity, Pepper!: How warmth in robots' language impacts reactions to errors during a collaborative task. In Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction. 245--247.Google Scholar
Digital Library
- Maurice Jakesch, Megan French, Xiao Ma, Jeffrey T. Hancock, and Mor Naaman. 2019. AI-mediated communication: How the perception that profile text was written by AI affects trustworthiness. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1--13.Google Scholar
Digital Library
- Susanne M. Jones and John G. Wirtz. 2006. How does the comforting process work? An empirical test of an appraisal-based model of comforting. Human Communication Research 32, 3 (2006), 217--243. DOI: https://doi.org/10.1111/j.1468--2958.2006.00274.xGoogle Scholar
Cross Ref
- Takeshi Kamita, Atsuko Matsumoto, Boyu Sun, Tomoo Inoue. 2020. Promotion of continuous use of a self-guided mental healthcare system by a chatbot. In Conference Companion Publication of the 2020 on Computer Supported Cooperative Work and Social Computing. 293--298.Google Scholar
Digital Library
- Pranav Khadpe, Ranjay Krishna, Li Fei-Fei, Jeffrey T. Hancock, and Michael S. Bernstein. 2020. Conceptual metaphors impact perceptions of human-AI collaboration. In Proceedings of the ACM on Human-Computer Interaction 4, CSCW2 (2020), 1--26.Google Scholar
- Junhan Kim, Yoojung Kim, Byungjoon Kim, Sukyung Yun, Minjoon Kim, and Joongseek Lee. 2018. Can a machine tend to teenagers' emotional needs? A study with conversational agents. In Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems. 1--6.Google Scholar
Digital Library
- Seo Young Kim, Bernd H. Schmitt, and Nadia M. Thalmann. 2019. Eliza in the uncanny valley: Anthropomorphizing consumer robots increases their perceived warmth but decreases liking. Marketing Letters 30 (2019), 1--12.Google Scholar
Cross Ref
- Oswald D. Kothgassner, Andreas Goreis, Johanna X. Kafka, Marlene Kaufmann, Katharina Atteneder, Leon Beutl, Kristina Hennig-Fast, Helmut Hlavacs, and Anna Felnhofer. 2019. Virtual social support buffers stress response: An experimental comparison of real-life and virtual support prior to a social stressor. Journal of Behavior Therapy and Experimental Psychiatry 63 (2019), 57--65. DOI: https://doi.org/10.1016/j.jbtep.2018.11.003Google Scholar
Cross Ref
- Janice L. Krupnick, Stuart M. Sotsky, Irene Elkin, Sam Simmens, Janet Moyer, John Watkins, and Paul A. Pilkonis. 2006. The role of the therapeutic alliance in psychotherapy and pharmacotherapy outcome: Findings in the national institute of mental health treatment of depression collaborative research program. Focus 64, 2 (2006), 532--277. DOI: https://doi.org/10.1176/foc.4.2.269Google Scholar
Cross Ref
- Liliana Laranjo, Adam G Dunn, Huong Ly Tong, Ahmet Baki Kocaballi, Jessica Chen, Rabia Bashir, Didi Surian, Blanca Gallego, Farah Magrabi, Annie Y S Lau, Enrico Coiera. 2018. Conversational agents in healthcare: a systematic review. Journal of the American Medical Informatics Association 25, 9 (2018), 1248--1258. DOI: https://doi.org/10.1093/jamia/ocy072Google Scholar
Cross Ref
- Emily G. Lattie, Rachel Kornfield, Kathryn E. Ringland, Renwen Zhang, Nathan Winquist, Madhu Reddy. 2020. Designing mental health technologies that support the social ecosystem of college students. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1--15.Google Scholar
Digital Library
- Kristina O. Lavik, Helga Frøysa, Karen F. Brattebø, John McLeod, and Christian Moltu. 2018. The first sessions of psychotherapy: A qualitative meta-analysis of alliance formation processes. Journal of Psychotherapy Integration 28, 3 (2018), 348--366. DOI: https://doi.org/10.1037/int0000101Google Scholar
Cross Ref
- Richard S. Lazarus and Susan Folkman. 1984. Stress, Appraisal, and Coping. Springer, New York, NY.Google Scholar
- Seungcheol Austin Lee and Yuhua Liang. 2015. Reciprocity in computer--human interaction: Source-based, norm-based, and affect-based explanations. Cyberpsychology, Behavior, and Social Networking 18, 4(2015), 234--240. DOI: https://doi.org/10.1089/cyber.2014.0458Google Scholar
Cross Ref
- Seungcheol Austin Lee and Yuhua Liang. 2016. The role of reciprocity in verbally persuasive robots." Cyberpsychology, Behavior, and Social Networking 19, 8(2016), 524--527. DOI: https://doi.org/10.1089/cyber.2016.0124Google Scholar
Cross Ref
- Seungcheol Austin Lee and Yuhua Jake Liang. 2019. Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Computers in Human Behavior, 90 (2019), 351--356. DOI: https://doi.org/10.1016/j.chb.2018.08.026Google Scholar
Cross Ref
- Eun-Ju Lee. 2009. I like you, but I won't listen to you: Effects of rationality on affective and behavioral responses to computers that flatter. International Journal of Human-Computer Studies 67, 8(2009), 628--638. DOI: https://doi.org/10.1016/j.ijhcs.2009.03.003Google Scholar
Digital Library
- Eun-Ju Lee. 2010a. What triggers social responses to flattering computers? Experimental tests of anthropomorphism and mindlessness explanations. Communication Research 37, 2(2010), 191--214. DOI: https://doi.org/10.1177/0093650209356389Google Scholar
Cross Ref
- Eun-Ju Lee. 2010b. The more humanlike, the better? How speech type and users' cognitive style affect social responses to computers. Computers in Human Behavior 26, 4 (2010), 665--672. DOI: https://doi.org/10.1016/j.chb.2010.01.003Google Scholar
Digital Library
- Robert W. Lent and Frederick G. Lopez. 2002. Cognitive ties That bind: A tripartite view of efficacy beliefs in growth-promoting relationships. Journal of Social and Clinical Psychology 21, 3 (2002), 256--286. DOI: https://doi.org/10.1521/jscp.21.3.256.22535Google Scholar
Cross Ref
- Jiwei Li, Will Monroe, Alan Ritter, Michel Galley, Jianfeng Gao, and Dan Jurafsky. 2016. Deep reinforcement learning for dialogue generation. arXiv preprint arXiv:1606.01541.Google Scholar
- Q. Vera Liao, Yi-Chia Wang, Timothy Bickmore, Pascale Fung, Jonathan Grudin, Zhou Yu, and Michelle Zhou. 2019. Human-agent communication: Connecting research and development in HCI and AI. In Conference Companion Publication of the 2019 on Computer Supported Cooperative Work and Social Computing. 122--126.Google Scholar
Digital Library
- Bingjie Liu and S. Shyam Sundar. 2018. Should machines express sympathy and empathy? Experiments with a health advice chatbot. Cyberpsychology, Behavior, and Social Networking 21, 10 (2018), 625--636. DOI: https://doi.org/10.1089/cyber.2018.0110Google Scholar
Cross Ref
- Gale M. Lucas, Jonathan Gratch, Aisha King, and Louis-Philippe Morency. 2014. It's only a computer: Virtual humans increase willingness to disclose. Computers in Human Behavior 37 (2014), 94--100. DOI: https://doi.org/10.1016/j.chb.2014.04.043Google Scholar
Digital Library
- Erina L. MacGeorge, Bo Feng, and Brant R. Burleson. 2011. Supportive Communication. Vol.4: Handbook of interpersonal communication (Eds.). SAGE, Thousand Oaks, CA.Google Scholar
- James C. McCroskey and Jason J. Teven. 1999. Goodwill: A reexamination of the construct and its measurement. Communications Monographs 66, 1 (Jun. 2009), 90--103. DOI: https://doi.org/10.1080/03637759909376464Google Scholar
Cross Ref
- Rachel M. McLaren, Denise Haunani Solomon, Jennifer S. Priem. 2011. Explaining variation in contemporaneous responses to hurt in premarital romantic relationships: A relational turbulence model perspective. Communication Research 38, 4 (2011), 543--564. DOI: https://doi.org/10.1177/0093650210377896Google Scholar
Cross Ref
- Jingbo Meng and Yue Nancy Dai. 2021. Emotional support from AI chatbots: Should a supportive partner self-disclose or not?. Journal of Computer-Mediated Communication 26, 4 (2021), 207--222. DOI: https://doi.org/10.1093/jcmc/zmab005Google Scholar
Cross Ref
- Alexander Miloff, Per Carlbring, William Hamilton, Gerhard Andersson, Lena Reuterskiöld, and Philip Lindner. 2020. Measuring alliance toward embodied virtual therapists in the era of automated treatments with the virtual therapist alliance scale (VTAS): Development and psychometric evaluation. Journal of Medical Internet Research 22, 3 (2020), 1--13.Google Scholar
Cross Ref
- Adam S. Miner, Arnold Milstein, and Jefferey T. Hancock. 2017. Talking to machines about personal mental health problems. Jama 318, 13 (2017), 1217--1218.Google Scholar
Cross Ref
- Antonio Justiniano Moraes Neto and Márcia Aparecida Fernandes. 2019. Chatbot and conversational analysis to promote collaborative learning in distance education. In 2019 IEEE 19th International Conference on Advanced Learning Technologies. 324--326.Google Scholar
Cross Ref
- Robert R. Morris, Kareem Kouddous, Rohan Kshirsagar, and Stephen M Schueller. 2018. Towards an artificially empathic conversational agent for mental health applications: system design and user perceptions. Journal of Medical Internet Research 20, 6 (2018), e10148.Google Scholar
Cross Ref
- Nasim Motalebi, Eugene Cho, S. Shyam Sundar, and Saeed Abdullah. 2019. Can Alexa be your therapist? How back-channeling transforms smart-speakers to be active listeners. In Conference Companion Publication of the 2019 on Computer Supported Cooperative Work and Social Computing. 309--313.Google Scholar
- Jaya Narain, Tina Quach, Monique Davey, Hae Won Park, Cynthia Breazeal, and Rosalind Picard. 2020. Promoting wellbeing with Sunny, a chatbot that facilitates positive messages within social groups. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems. 1--8.Google Scholar
Digital Library
- Clifford Nass and Youngme Moon. 2000. Machines and mindlessness: social responses to computers. Journal of Social Issues 56, 1 (2000), 81--103.Google Scholar
Cross Ref
- Clifford Nass, B. J. Fogg, and Youngme Moon. 1996. Can computers be teammates?. International Journal of Human-Computer Studies 45, 6 (1996), 669--678. DOI: https://doi.org/10.1006/ijhc.1996.0073Google Scholar
Digital Library
- Clifford Nass, Youngme Moon, Brian J. Fogg, Byron Reeves, and D. Christopher Dryer. 1995. Can computer personalities be human personalities?. International Journal of Human-Computer Studies 43, 2 (1995), 223--239. DOI: https://doi.org/10.1006/ijhc.1995.1042Google Scholar
Digital Library
- Clifford Nass, Ing-Marie Jonsson, Helen Harris, Ben Reaves, Jack Endo, Scott Brave, and Leila Takayama. 2005. Improving automotive safety by pairing driver emotion and car voice emotion. In CHI'05 extended abstracts on Human factors in computing systems, 1973--1976. DOI: https://doi.org/10.1145/1056808.1057070Google Scholar
Digital Library
- Truong-Huy D. Nguyen, Elin Carstensdottir, Nhi Ngo, Magy Seif El-Nasr, Matt Gray, Derek Isaacowitz, and David Desteno. 2015. Modeling Warmth and Competence in Virtual Characters. In International Conference on Intelligent Virtual Agents, Springer International Publishing, Cham, 167--180.Google Scholar
- Yoo Jung Oh, Jingwen Zhang, Min-Lin Fang, and Yoshimi Fukuoka. 2021. A systematic review of artificial intelligence chatbots for promoting physical activity, healthy diet, and weight loss. International Journal of Behavioral Nutrition and Physical Activity 18, 1 (Dec. 2021), 1--25.Google Scholar
Cross Ref
- Chinedu Wilfred Okonkwo and Abejide Ade-Ibijola. 2021. Chatbots applications in education: A systematic review. Computers and Education: Artificial Intelligence 2 (2021), 10 pages. DOI: https://doi.org/10.1016/j.caeai.2021.100033Google Scholar
Cross Ref
- Laura M. Pfeifer and Timothy Bickmore. 2011. Is the media equation a flash in the pan? The durability and longevity of social responses to computers. In Proceedings of the SIGCHI conference on human factors in computing systems, 777--780. DOI: https://doi.org/10.1145/1978942.1979055Google Scholar
Digital Library
- Martin Porcheron, Leigh Clark, Matt Jones, Heloisa Candello, Benjamin R. Cowan, Christine Murad, Jaisie Sin, Matthew P. Aylett, Minha Lee, Cosmin Munteanu, Joel E. Fischer, Philip R. Doyle, and Jofish Kaye. 2020. CUI@ CSCW: Collaborating through Conversational User Interfaces. In Conference Companion Publication of the 2020 on Computer Supported Cooperative Work and Social Computing. 483--492.Google Scholar
Digital Library
- Pooja Prajod, Mohammed Al Owayyed, Tim Rietveld, Jaap Jan Van Der Steeg, and Joost Broekens. 2019. The Efect of virtual agent warmth on human-agent negotiation. In Proceedings of the 18th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2019), Montreal, Canada, 71--76.Google Scholar
Digital Library
- Sheizaf Rafaeli and Sudweeks. F .1997. Networked Interactivity. Journal of Computer-Mediated Communication 2, 4 (1997).Google Scholar
- Sheizaf Rafaeli. 1988. Interactivity: From new media to communication. In R. Hawkins, J. M. Wiemann, & S. Pingree (Eds.), Sage Annual Review of Communication Research: Advancing Communication Science (Vol. 16, pp. 110--134). Sage.Google Scholar
- Stephen A. Rains and Andrew C High. 2021. The effects of person-centered social support messages on recipient distress over time within a conversation. Journal of Communication 71, 3 (2021), 380--402. DOI: https://doi.org/10.1093/joc/jqab010Google Scholar
Cross Ref
- Stephen A. Rains, Corey A. Pavlich, Eric Tsetsi, and Chelsie Akers. 2019. Evaluating the scope and theoretical mechanisms of person-centered social support: The case of the comforting computer program. Communication Research 47, 4 (2019), 512--540. DOI: https://doi.org/10.1177/0093650219852554Google Scholar
Cross Ref
- Rachel M. Ranney, Emma Bruehlman-Senecal, and Ozlem Ayduk. 2016. Comparing the effects of three online cognitive reappraisal trainings on well-being. Journal of Happiness Studies 18, 5(2017), 1319--1338.Google Scholar
Cross Ref
- Byron Reeves and Clifford Nass. 1996. The media equation: How people treat computers, television, and new media like real people and places (pp. xiv, 305). Cambridge University Press.Google Scholar
- Lazlo Ring, Lin Shi, Kathleen Totzke, and Timothy Bickmore. 2014. Social support agents for older adults: longitudinal affective computing in the home. Journal on Multimodal User Interfaces 9, 1 (Jun. 2014), 79--88. DOI: https://doi.org/10.1007/s12193-014-0157-0Google Scholar
Cross Ref
- Rebecca B. Rubin and Matthew M. Martin. 1994. Development of a measure of interpersonal communication competence. Communication Research Reports 11, 1 (1994), 33--44.Google Scholar
Cross Ref
- Shruti Sannon, Brett Stoll, Dominic DiFranzo, Malte Jung, and Natalya N. Bazarova. 2018. How personification and interactivity influence stress-related disclosures to conversational agents. In Companion of the 2018 ACM Conference on Computer Supported Cooperative Work and Social Computing. 285--288.Google Scholar
- Jessica Schroeder, Chelsey Wilkes, Kael Rowan, Arturo Toledo, Ann Paradiso, Mary Czerwinski, Gloria Mark, and Marsha M. Linehan. 2018. Pocket skills: A conversational mobile web app to support dialectical behavioral therapy. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1--15.Google Scholar
- Chad E. Shenk and Alan E. Fruzzetti. 2011. The impact of validating and invalidating responses on emotional reactivity. Journal of Social and Clinical Psychology 30, 2 (2011), 163--183. DOI: https://doi.org/10.1521/jscp.2011.30.2.163Google Scholar
Cross Ref
- Weiyan Shi, Xuewei Wang, Yoo Jung Oh, Jingwen Zhang, Saurav Sahay, and Zhou Yu. 2020. Effects of persuasive dialogues: testing bot identities and inquiry strategies. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1--13.Google Scholar
Digital Library
- Marita Skjuve and Petter Bae Brandzaeg. 2018. Measuring user experience in chatbots: An approach to interpersonal communication competence. In International Conference on Internet Science. 113--120.Google Scholar
- Marita Skjuve, Asbjørn Følstad, Knut Inge Fostervold, and Petter Bae Brandtzaeg. 2021. My chatbot companion-a study of human-chatbot relationships. International Journal of Human-Computer Studies 149 (2021), 14 pages.Google Scholar
Cross Ref
- Magnus Söderlund. 2021. The robot-to-robot service encounter: an examination of the impact of inter-robot warmth. Journal of Services Marketing 35, 9 (2021), 15--27. DOI: https://doi.org/10.1108/JSM-01--2021-0006Google Scholar
Cross Ref
- Jan-Philipp Stein and Peter Ohler. 2017. Venturing into the uncanny valley of mind-The influence of mind attribution on the acceptance of human-like characters in a virtual reality setting. Cognition 160 (2017), 43--50. DOI: https://doi.org/10.1016/j.cognition.2016.12.010Google Scholar
Cross Ref
- Laura E. Stevens and Susan T. Fiske. 1995. Motivation and cognition in social life: A social survival perspective. Social Cognition 13, 3 (1995), 189--214.Google Scholar
Cross Ref
- S. Shyam Sundar and Jinyoung Kim. 2019. Machine heuristic: When we trust computers more than humans with our personal information. In Proceedings of the 2019 CHI Conference on human factors in computing systems. 1--9.Google Scholar
Digital Library
- S. Shyam Sundar, Saraswathi Bellur, Jeeyun Oh, Haiyan Jia, Hyang-Sook Kim. 2016. Theoretical importance of contingency in human-computer interaction: Effects of message interactivity on user engagement. Communication Research 43, 5 (2016), 595--625.Google Scholar
Cross Ref
- S. Shyam Sundar. 2020. Rise of machine agency: A framework for studying the psychology of human--AI interaction (HAII). Journal of Computer-Mediated Communication 25, 1 (2020), 74--88. DOI: https://doi.org/10.1093/jcmc/zmz026Google Scholar
Cross Ref
- Vivian Ta, Caroline Griffith, Carolynn Boatfield, Xinyu Wang, Maria Civitello, Haley Bader, Esther DeCero, and Alexia Loggarakis. 2020. User experiences of social support from companion chatbots in everyday contexts: Thematic analysis. Journal of Medical Internet Research 22, 3 (2020), 1--10. DOI: https://doi.org/10.2196/16235Google Scholar
Cross Ref
- Bert N. Uchino. 2004. Social support and physical health: Understanding the health consequences of relationships. Yale University Press, New Haven, CT. DOI: https://doi.org/10.12987/yale/9780300102185.001.0001Google Scholar
Cross Ref
- Astrid M. Von der Pütten, Nicole C. Krämer, Jonathan Gratch, and Sin-Hwa Kang. 2010. ?It doesn't matter what you are!" explaining social effects of agents and avatars. Computers in Human Behavior 26, 6 (2010), 1641--1650. DOI: https://doi.org/10.1016/j.chb.2010.06.012Google Scholar
Digital Library
- Liuping Wang, Dakuo Wang, Feng Tian, Zhenhui Peng, Xiangmin Fan, Zhan Zhang, Mo Yu, Xiaojuan Ma, and Hongan Wang. 2021. Cass: Towards building a social-support chatbot for online health community. In Proceedings of the ACM on Human-Computer Interaction. 1--31.Google Scholar
Digital Library
- Adam Waytz and Michael I. Norton. 2014. Botsourcing and outsourcing: Robot, British, Chinese, and German workers are for thinking-not feeling-jobs. Emotion 14, 2 (2014), 434--444.Google Scholar
Cross Ref
- Honglian Xiang, Jia Zhou, and Bingjun Xie. 2022. AI tools for debunking online spam reviews? Trust of younger and older adults in AI detection criteria. Behaviour & Information Technology (2022): 1--20. DOI: 10.1080/0144929X.2021.2024252Google Scholar
Cross Ref
- Anbang Xu, Zhe Liu, Yufan Guo, Vibha Sinha, and Rama Akkiraju. 2017. A new chatbot for customer service on social media. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. 3506--3510.Google Scholar
Digital Library
- Seounmi Youn and S. Venus Jin. 2021. "In AI we trust?" The effects of parasocial interaction and technopian versus luddite ideological views on chatbot-based customer relationship management in the emerging ?feeling economy". Computers in Human Behavior 119 (2021), 106721. DOI: https://doi.org/10.1016/j.chb.2021.106721Google Scholar
Digital Library
- Zhou Yu, Leah Nicolich-Henkin, Alan W. Black, and Alexander Rudnicky. 2016. A wizard-of-oz study on a non-task-oriented dialog systems that reacts to user engagement. In Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue. 55--63.Google Scholar
Cross Ref
Index Terms
- Mediated Social Support for Distress Reduction: AI Chatbots vs. Human
Recommendations
When the Social Becomes Non-Human: Young People's Perception of Social Support in Chatbots
CHI '21: Proceedings of the 2021 CHI Conference on Human Factors in Computing SystemsAlthough social support is important for health and well-being, many young people are hesitant to reach out for support. The emerging uptake of chatbots for social and emotional purposes entails opportunities and concerns regarding non-human agents as ...
Exploring how politeness impacts the user experience of chatbots for mental health support
AbstractPoliteness is important in human–human interaction when asking people to engage in sensitive conversations. If politeness manifests similarly in human–chatbot interaction, it may play an important role in the design of sensitive chatbot ...
Highlights- Politeness can both positively and negatively impact the chatbot user experience.
- The Personal politeness chatbot was experienced as caring and encouraging.
- The Passive politeness chatbot was experienced as too apologetic and ...
Interactivity in online chat: Conversational cues and visual cues in the service recovery process
Highlights- The paper investigates interactivity in online chat.
- Verbal cues lead to higher ...
AbstractUse of both verbal and nonverbal cues in computer-mediated communication can influence customers’ perceptions and their behavioral intentions. Drawing on the compensation effect theory, this study investigates how verbal and nonverbal ...





Comments