当前位置: X-MOL 学术Mem. Cogn. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Unimodal and cross-modal identity judgements using an audio-visual sorting task: Evidence for independent processing of faces and voices
Memory & Cognition ( IF 2.482 ) Pub Date : 2021-07-12 , DOI: 10.3758/s13421-021-01198-7
Nadine Lavan 1, 2 , Harriet M J Smith 3 , Carolyn McGettigan 1
Affiliation  

Unimodal and cross-modal information provided by faces and voices contribute to identity percepts. To examine how these sources of information interact, we devised a novel audio-visual sorting task in which participants were required to group video-only and audio-only clips into two identities. In a series of three experiments, we show that unimodal face and voice sorting were more accurate than cross-modal sorting: While face sorting was consistently most accurate followed by voice sorting, cross-modal sorting was at chancel level or below. In Experiment 1, we compared performance in our novel audio-visual sorting task to a traditional identity matching task, showing that unimodal and cross-modal identity perception were overall moderately more accurate than the traditional identity matching task. In Experiment 2, separating unimodal from cross-modal sorting led to small improvements in accuracy for unimodal sorting, but no change in cross-modal sorting performance. In Experiment 3, we explored the effect of minimal audio-visual training: Participants were shown a clip of the two identities in conversation prior to completing the sorting task. This led to small, nonsignificant improvements in accuracy for unimodal and cross-modal sorting. Our results indicate that unfamiliar face and voice perception operate relatively independently with no evidence of mutual benefit, suggesting that extracting reliable cross-modal identity information is challenging.



中文翻译:

使用视听排序任务的单模态和跨模态身份判断:人脸和声音独立处理的证据

人脸和声音提供的单模态和跨模态信息有助于身份感知。为了检查这些信息源如何相互作用,我们设计了一种新颖的视听分类任务,其中要求参与者将纯视频和纯音频剪辑分组为两个身份。在一系列三个实验中,我们表明单模态人脸和语音排序比跨模态排序更准确:虽然人脸排序始终是最准确的,其次是语音排序,但跨模态排序处于通道级别或以下。在实验 1 中,我们将新颖的视听排序任务与传统身份匹配任务的性能进行了比较,表明单模态和跨模态身份感知总体上比传统的身份匹配任务更准确。在实验 2 中,将单模态排序与跨模态排序分开导致单模态排序的准确性略有提高,但跨模态排序性能没有变化。在实验 3 中,我们探索了最小视听训练的效果:在完成分类任务之前,向参与者展示了对话中两个身份的剪辑。这导致单模态和跨模态排序的准确性略有提高。我们的结果表明,陌生的面部和语音感知相对独立地运作,没有互惠互利的证据,这表明提取可靠的跨模式身份信息具有挑战性。在完成分类任务之前,参与者在对话中看到了两个身份的剪辑。这导致单模态和跨模态排序的准确性略有提高。我们的结果表明,陌生的面部和语音感知相对独立地运作,没有互惠互利的证据,这表明提取可靠的跨模式身份信息具有挑战性。在完成分类任务之前,参与者在对话中看到了两个身份的剪辑。这导致单模态和跨模态排序的准确性略有提高。我们的结果表明,陌生的面部和语音感知相对独立地运作,没有互惠互利的证据,这表明提取可靠的跨模式身份信息具有挑战性。

更新日期:2021-07-12
down
wechat
bug