Elsevier

Decision Support Systems

Volume 150, November 2021, 113556
Decision Support Systems

Simpler is better: Lifting interpretability-performance trade-off via automated feature engineering

https://doi.org/10.1016/j.dss.2021.113556Get rights and content
Under a Creative Commons license
open access

Highlights

  • Complex models are not always better.

  • SAFE extracts interpretable features from complex models.

  • Complex models can be automatically simplified without loss of performance.

Abstract

Machine learning has proved to generate useful predictive models that can and should support decision makers in many areas. The availability of tools for AutoML makes it possible to quickly create an effective but complex predictive model. However, the complexity of such models is often a major obstacle in applications, especially in terms of high-stake decisions. We are experiencing a growing number of examples where the use of black boxes leads to decisions that are harmful, unfair or simply wrong. In this paper, we show that very often we can simplify complex models without compromising their performance; however, with the benefit of much needed transparency.

We propose a framework that uses elastic black boxes as supervisor models to create simpler, less opaque, yet still accurate and interpretable glass box models. The new models were created using newly engineered features extracted with the help of a supervisor model. We supply the analysis using a large-scale benchmark on several tabular data sets from the OpenML database. There are tree main results of this paper: 1) we show that extracting information from complex models may improve the performance of simpler models, 2) we question a common myth that complex predictive models outperform simpler predictive models, 3) we present a real-life application of the proposed method.

Keywords

Interpretability
Machine learning
Feature engineering
Decision-making

Cited by (0)

Alicja Gosiewska is a Ph.D. student in Computer Science at Warsaw University of Technology and holds a Master's degree in Mathematics. She works on an explainable artificial intelligence methods for tabular and sequential data. Her interests also include the application of Deep Learning to model protein structure prediction.

Anna Kozak holds master degreee in Mathematical Statistics and Data Analysis at Warsaw University of Technology. Her interests are explainable machine learning and data visualization.

Przemyslaw Biecek works as associate professor at Warsaw University of Technology and University of Warsaw. His research interests include model exploration, explainable artificial intelligence and automated machine learning, as well as the development of statistical software. He is an active developer of the DALEX package and many other packages that support Explanatory Model Analysis (EMA).