当前位置: X-MOL 学术arXiv.cs.AI › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
A Bayesian Account of Measures of Interpretability in Human-AI Interaction
arXiv - CS - Artificial Intelligence Pub Date : 2020-11-22 , DOI: arxiv-2011.10920
Sarath Sreedharan, Anagha Kulkarni, Tathagata Chakraborti, David E. Smith, Subbarao Kambhampati

Existing approaches for the design of interpretable agent behavior consider different measures of interpretability in isolation. In this paper we posit that, in the design and deployment of human-aware agents in the real world, notions of interpretability are just some among many considerations; and the techniques developed in isolation lack two key properties to be useful when considered together: they need to be able to 1) deal with their mutually competing properties; and 2) an open world where the human is not just there to interpret behavior in one specific form. To this end, we consider three well-known instances of interpretable behavior studied in existing literature -- namely, explicability, legibility, and predictability -- and propose a revised model where all these behaviors can be meaningfully modeled together. We will highlight interesting consequences of this unified model and motivate, through results of a user study, why this revision is necessary.

中文翻译:

人机交互中可解释性度量的贝叶斯解释

现有的可解释代理行为设计方法考虑孤立地解释可解释性的不同方法。在本文中,我们假定,在现实世界中设计和部署人类感知代理时,可解释性的概念只是众多考虑因素中的一部分。孤立地开发的技术缺少两个关键特性,在一起考虑时它们是有用的:它们需要能够:1)处理它们相互竞争的特性;2)一个开放的世界,人类不仅在那里以一种特定的形式解释行为。为此,我们考虑了现有文献中研究的三个著名的可解释行为实例,即可解释性,易读性和可预测性,并提出了一个修订模型,可以将所有这些行为一起有意义地建模。
更新日期:2020-11-25
down
wechat
bug