Decision-support systems need user confidence to go into production. This is especially critical for hybrid #MachineLearning + #ORMS pipelines. So, how to explain data-driven decisions based on contextual information? We introduce new methodologies: https://arxiv.org/abs/2301.10074

We provide counterfactual explanations highlighting why a solution is better than another (or any other) due to context, e.g., in which context should we use the green path instead of the blue path (or any other path) today?

Explainable Data-Driven Optimization: From Context to Decision and Back Again

Data-driven optimization uses contextual information and machine learning algorithms to find solutions to decision problems with uncertain parameters. While a vast body of work is dedicated to interpreting machine learning models in the classification setting, explaining decision pipelines involving learning algorithms remains unaddressed. This lack of interpretability can block the adoption of data-driven solutions as practitioners may not understand or trust the recommended decisions. We bridge this gap by introducing a counterfactual explanation methodology tailored to explain solutions to data-driven problems. We introduce two classes of explanations and develop methods to find nearest explanations of random forest and nearest-neighbor predictors. We demonstrate our approach by explaining key problems in operations management such as inventory management and routing.

arXiv.org
More precisely, we ask what is the context "closest to the current known situation" that renders the green path optimal. In this specific case, a difference in a single feature (period of the day) explains this different decision. Such contrastive explanations are helpful in understanding why experts make certain decisions: they may have or ignore decisive information on the current context. In both cases, recognizing these differences will be instrumental in improving current practice.
If you are curious about how the methods work under the hood, what the paper does is akin to solving an inverse optimization problem formulated as an integer bilevel program. Our current approach applies to pipelines with random forests and nearest-neighbor predictors. The resulting explanation problems can be solved at scale for various applications. In particular, we report experiments on multi-item newsvendor, 2-stages shipment planning, and CVaR shortest paths.

Ultimately, we hope it contributes to bridging the gap between #ExplainableML, #XAI, and #ORMS. To facilitate future work and experimental reproducibility, all our algorithms and scripts are provided in #opensource at https://github.com/alexforel/Explainable-CSO.

Many thanks to my amazing co-authors Alexandre Forel and Axel Parmentier, as well as Polytechnique Montréal and SCALE AI for the financial support through the SCALE-AI Chair in Data-Driven Supply Chains!

GitHub - alexforel/Explainable-CSO: Code for "Explainable Data-Driven Optimization"

Code for "Explainable Data-Driven Optimization". Contribute to alexforel/Explainable-CSO development by creating an account on GitHub.

GitHub