Computer Science > Human-Computer Interaction
[Submitted on 18 Jun 2021 (v1), last revised 22 Aug 2022 (this version, v2)]
Title:Influence of agent's self-disclosure on human empathy
View PDFAbstract:As AI technologies progress, social acceptance of AI agents, including intelligent virtual agents and robots, is becoming even more important for more applications of AI in human society. One way to improve the relationship between humans and anthropomorphic agents is to have humans empathize with the agents. By empathizing, humans act positively and kindly toward agents, which makes it easier for them to accept the agents. In this study, we focus on self-disclosure from agents to humans in order to increase empathy felt by humans toward anthropomorphic agents. We experimentally investigate the possibility that self-disclosure from an agent facilitates human empathy. We formulate hypotheses and experimentally analyze and discuss the conditions in which humans have more empathy toward agents. Experiments were conducted with a three-way mixed plan, and the factors were the agents' appearance (human, robot), self-disclosure (high-relevance self-disclosure, low-relevance self-disclosure, no self-disclosure), and empathy before/after a video stimulus. An analysis of variance (ANOVA) was performed using data from 918 participants. We found that the appearance factor did not have a main effect, and self-disclosure that was highly relevant to the scenario used facilitated more human empathy with a statistically significant difference. We also found that no self-disclosure suppressed empathy.These results support our hypotheses. This study reveals that self-disclosure represents an important characteristic of anthropomorphic agents which helps humans to accept them.
Submission history
From: Takahiro Tsumura [view email][v1] Fri, 18 Jun 2021 04:14:58 UTC (3,645 KB)
[v2] Mon, 22 Aug 2022 08:53:37 UTC (4,332 KB)
References & Citations
Bibliographic and Citation Tools
Bibliographic Explorer (What is the Explorer?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)
Code, Data and Media Associated with this Article
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)
Demos
Recommenders and Search Tools
Influence Flower (What are Influence Flowers?)
Connected Papers (What is Connected Papers?)
CORE Recommender (What is CORE?)
arXivLabs: experimental projects with community collaborators
arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.
Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.
Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.