Shapley values for feature selection: the good, the bad and the axioms

A secret-holder style payoff

Abstract

The Shapley value has become popular in the Explainable AI (XAI) literature, thanks, to a large extent, to a solid theoretical foundation, including four ``favourable and fair" axioms for attribution in transferable utility games. The Shapley value is provably the only solution concept satisfying these axioms. In this paper, we introduce the Shapley value and draw attention to its recent uses as a feature selection tool. We call into question this use of the Shapley value, using simple, abstract `toy' counterexamples to illustrate that the axioms may work against the goals of feature selection. From this, we develop a number of insights that are then investigated in concrete simulation settings, with a variety of Shapley value formulations, including SHapley Additive exPlanations (SHAP) and Shapley Additive Global importancE (SAGE).

Dr. Danny Fryer
Dr. Danny Fryer
Senior Data Scientist / Hon. Lecturer / AFHEA

Experienced in leading and innovating in data science, I specialise in data engineering, explainable AI/ML, and statistical analysis, focusing on simplification, uncertainty, decision-making and teaching.