当前位置: X-MOL 学术Cyberpsychology, Behavior, and Social Networking › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Content Moderation on Social Media: Does It Matter Who and Why Moderates Hate Speech?
Cyberpsychology, Behavior, and Social Networking ( IF 4.2 ) Pub Date : 2023-05-03 , DOI: 10.1089/cyber.2022.0158
Sai Wang 1 , Ki Joon Kim 2
Affiliation  

Artificial intelligence (AI) has been increasingly integrated into content moderation to detect and remove hate speech on social media. An online experiment (N = 478) was conducted to examine how moderation agents (AI vs. human vs. human-AI collaboration) and removal explanations (with vs. without) affect users' perceptions and acceptance of removal decisions for hate speech targeting social groups with certain characteristics, such as religion or sexual orientation. The results showed that individuals exhibit consistent levels of perceived trustworthiness and acceptance of removal decisions regardless of the type of moderation agent. When explanations for the content takedown were provided, removal decisions made jointly by humans and AI were perceived as more trustworthy than the same decisions made by humans alone, which increased users' willingness to accept the verdict. However, this moderated mediation effect was only significant when Muslims, not homosexuals, were the target of hate speech.

中文翻译:

社交媒体上的内容审核:由谁以及为何审核仇恨言论重要吗?

人工智能 (AI) 已越来越多地融入内容审核中,以检测和消除社交媒体上的仇恨言论。进行了一项在线实验(N = 478),以研究审核代理(人工智能与人类与人类与人工智能协作)和删除解释(有与无)如何影响用户对针对社交的仇恨言论的删除决定的看法和接受度具有某些特征的群体,例如宗教或性取向。结果表明,无论调节剂的类型如何,个人都表现出一致的感知可信度和接受移除决定的水平。当提供内容删除的解释时,人类和人工智能共同做出的删除决定被认为比人类单独做出的相同决定更值得信赖,这增加了用户接受裁决的意愿。然而,只有当穆斯林而不是同性恋者成为仇恨言论的目标时,这种缓和的中介效应才显着。
更新日期:2023-05-03
down
wechat
bug