Core Concepts
Interpretable machine learning can be viewed as a form of sensitivity analysis applied to machine learning systems, integrating recent advances in interpretability into a broader framework for explaining complex systems.
Abstract
This position paper argues that interpretable machine learning (IML) can be seen as a form of sensitivity analysis (SA) applied to machine learning (ML) systems. The authors aim to bridge the gap between the ML and SA communities by:
Formally describing the ML process as a system suitable for SA, with interconnected functions representing different components of the ML workflow (e.g., hyperparameter optimization, model training, model interpretation).
Highlighting how existing IML methods, such as partial dependence, FANOVA, and Shapley values, relate to this SA-based perspective on ML. These methods can be viewed as computing sensitivities within the ML system.
Discussing how other SA techniques, typically used in domains like environmental modeling or engineering, could be applied to ML to provide new interpretations and insights.
The authors argue that this unified SA-based view of IML can help better credit related work from other fields, identify potential research gaps, and establish a common framework and terminology for discussing and developing interpretability methods. They also discuss how ML can contribute to the advancement of SA, for example, through better metamodeling practices or handling of feature dependencies.
Overall, the paper aims to initiate a discussion on the connections between IML and SA, with the goal of fostering collaboration and cross-pollination between the two research communities.