论文标题
校准医疗保健AI:建立可靠且可解释的深度预测模型
Calibrating Healthcare AI: Towards Reliable and Interpretable Deep Predictive Models
论文作者
论文摘要
在临床决策中,广泛采用表示技术的广泛采用强烈强调了表征模型可靠性并实现模型行为的严格内省的必要性。尽管通常通过结合不确定性量化策略来解决前者的需求,但后者的挑战是使用广泛的可解释性技术来解决的。在本文中,我们认为这两个目标不一定是不同的,并建议利用预测校准来满足这两个目标。更具体地说,我们的方法由校准驱动的学习方法组成,该方法也用于设计基于反事实推理的可解释性技术。此外,我们介绍了\ textit {可靠性图},这是一种用于模型可靠性的整体评估机制。使用皮肤镜图像的病变分类问题,我们证明了方法的有效性,并推断出有关模型行为的有趣见解。
The wide-spread adoption of representation learning technologies in clinical decision making strongly emphasizes the need for characterizing model reliability and enabling rigorous introspection of model behavior. While the former need is often addressed by incorporating uncertainty quantification strategies, the latter challenge is addressed using a broad class of interpretability techniques. In this paper, we argue that these two objectives are not necessarily disparate and propose to utilize prediction calibration to meet both objectives. More specifically, our approach is comprised of a calibration-driven learning method, which is also used to design an interpretability technique based on counterfactual reasoning. Furthermore, we introduce \textit{reliability plots}, a holistic evaluation mechanism for model reliability. Using a lesion classification problem with dermoscopy images, we demonstrate the effectiveness of our approach and infer interesting insights about the model behavior.