论文标题
学习脚手架:优化教学模型解释
Learning to Scaffold: Optimizing Model Explanations for Teaching
论文作者
论文摘要
现代机器学习模型是不透明的,因此,关于解释这些模型行为的方法有一个新兴的学术子领域。但是,提供此类解释的确切目标是什么?我们如何证明解释实现这一目标?一些研究认为,解释应帮助教学学生(人类或机器)模拟所解释的模型,并且可以通过学生在无法解释的示例中的模拟准确性来衡量解释的质量。在这项工作中,利用元学习技术,我们扩展了这一想法,以提高解释本身的质量,特别是通过优化解释,使学生模型更有效地学习模拟原始模型。我们在三个自然语言处理和计算机视觉任务上培训模型,发现接受了使用我们框架的解释培训的学生能够比以前的方法更有效地模拟教师。通过人类注释和用户研究,我们进一步发现,这些学习的解释与人类如何解释这些任务中所需的决策更加一致。我们的代码可从https://github.com/coderpat/learning-scaffold获得
Modern machine learning models are opaque, and as a result there is a burgeoning academic subfield on methods that explain these models' behavior. However, what is the precise goal of providing such explanations, and how can we demonstrate that explanations achieve this goal? Some research argues that explanations should help teach a student (either human or machine) to simulate the model being explained, and that the quality of explanations can be measured by the simulation accuracy of students on unexplained examples. In this work, leveraging meta-learning techniques, we extend this idea to improve the quality of the explanations themselves, specifically by optimizing explanations such that student models more effectively learn to simulate the original model. We train models on three natural language processing and computer vision tasks, and find that students trained with explanations extracted with our framework are able to simulate the teacher significantly more effectively than ones produced with previous methods. Through human annotations and a user study, we further find that these learned explanations more closely align with how humans would explain the required decisions in these tasks. Our code is available at https://github.com/coderpat/learning-scaffold