pith. machine review for the scientific record. sign in

arxiv: 1805.10820 · v1 · submitted 2018-05-28 · 💻 cs.AI

Recognition: unknown

Local Rule-Based Explanations of Black Box Decision Systems

Authors on Pith no claims yet
classification 💻 cs.AI
keywords decisionexplanationsblackinterpretablelocallorepredictorreasons
0
0 comments X
read the original abstract

The recent years have witnessed the rise of accurate but obscure decision systems which hide the logic of their internal decision processes to the users. The lack of explanations for the decisions of black box systems is a key ethical issue, and a limitation to the adoption of machine learning components in socially sensitive and safety-critical contexts. %Therefore, we need explanations that reveals the reasons why a predictor takes a certain decision. In this paper we focus on the problem of black box outcome explanation, i.e., explaining the reasons of the decision taken on a specific instance. We propose LORE, an agnostic method able to provide interpretable and faithful explanations. LORE first leans a local interpretable predictor on a synthetic neighborhood generated by a genetic algorithm. Then it derives from the logic of the local interpretable predictor a meaningful explanation consisting of: a decision rule, which explains the reasons of the decision; and a set of counterfactual rules, suggesting the changes in the instance's features that lead to a different outcome. Wide experiments show that LORE outperforms existing methods and baselines both in the quality of explanations and in the accuracy in mimicking the black box.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 3 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. What Should Explanations Contain? A Human-Centered Explanation Content Model for Local, Post-Hoc Explanations

    cs.HC 2026-05 accept novelty 7.0

    A 14-code content model for local post-hoc AI explanations, derived from 325 user statements and validated by experts with high reliability scores.

  2. PREF-XAI: Preference-Based Personalized Rule Explanations of Black-Box Machine Learning Models

    cs.LG 2026-04 unverdicted novelty 7.0

    PREF-XAI treats explanations as ranked alternatives and learns additive utility functions from limited user feedback to select and discover personalized rule explanations for black-box models.

  3. Interpretable and Explainable Surrogate Modeling for Simulations: A State-of-the-Art Survey and Perspectives on Explainable AI for Decision-Making

    cs.AI 2026-04 unverdicted novelty 5.0

    This survey synthesizes XAI methods with surrogate modeling workflows for simulations and outlines a research agenda to embed explainability into simulation-driven design and decision-making.