论文标题

解释可以减少决策过程中对AI系统的过度依赖

Explanations Can Reduce Overreliance on AI Systems During Decision-Making

论文作者

Vasconcelos, Helena, Jörke, Matthew, Grunde-McLaughlin, Madeleine, Gerstenberg, Tobias, Bernstein, Michael, Krishna, Ranjay

论文摘要

先前的工作已经确定了一种有弹性的现象,威胁人类决策团队的表现:过度依赖,当人们同意AI时,即使是不正确的。令人惊讶的是,与仅提供预测相比,当AI对其预测产生解释时,过度依赖并不会减少。一些人认为,过度依赖是由认知偏见或未校准的信任导致的,将过分依赖归因于人类认知的必然性。相比之下,我们的论文认为,人们从战略上选择是否参与AI解释,从经验上证明,在某些情况下,AI解释会降低过度依赖。为了实现这一目标,我们将这种战略选择正式化在成本效益框架中,在这种框架中,参与任务的成本和收益与依靠AI的成本和收益权衡。我们操纵迷宫任务中的成本和收益,参与者与模拟的AI合作找到迷宫的出口。通过5项研究(n = 731),我们发现诸如任务难度(研究1),解释难度(研究2、3)和诸如货币补偿(研究4)诸如任务难度之类的成本(研究4)会影响过度依赖。最后,研究5调整了认知努力折现范式以量化不同解释的实用性,从而为我们的框架提供了进一步的支持。我们的结果表明,文献中发现的一些无效效应可能部分是由于解释不足以降低验证AI预测的成本。

Prior work has identified a resilient phenomenon that threatens the performance of human-AI decision-making teams: overreliance, when people agree with an AI, even when it is incorrect. Surprisingly, overreliance does not reduce when the AI produces explanations for its predictions, compared to only providing predictions. Some have argued that overreliance results from cognitive biases or uncalibrated trust, attributing overreliance to an inevitability of human cognition. By contrast, our paper argues that people strategically choose whether or not to engage with an AI explanation, demonstrating empirically that there are scenarios where AI explanations reduce overreliance. To achieve this, we formalize this strategic choice in a cost-benefit framework, where the costs and benefits of engaging with the task are weighed against the costs and benefits of relying on the AI. We manipulate the costs and benefits in a maze task, where participants collaborate with a simulated AI to find the exit of a maze. Through 5 studies (N = 731), we find that costs such as task difficulty (Study 1), explanation difficulty (Study 2, 3), and benefits such as monetary compensation (Study 4) affect overreliance. Finally, Study 5 adapts the Cognitive Effort Discounting paradigm to quantify the utility of different explanations, providing further support for our framework. Our results suggest that some of the null effects found in literature could be due in part to the explanation not sufficiently reducing the costs of verifying the AI's prediction.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源