Concisely Explaining the Doubt: Minimum-Size Abductive Explanations for Linear Models with a Reject Option
Trustworthiness in artificial intelligence depends not only on what a model decides, but also on how it handles and explains cases in which a reliable decision cannot be made. In critical domains such as healthcare and finance, a reject option allows the model to abstain when evidence is insufficient, making it essential to explain why an instance is rejected in order to support informed human intervention. In these settings, explanations must not only be interpretable, but also faithful to the underlying model and computationally efficient enough to support real-time decision making. Abductive explanations guarantee fidelity, but their exact computation is known to be NP-hard for many classes of models, limiting their practical applicability. Computing \textbf{minimum-size} abductive explanations is an even more challenging problem, as it requires reasoning not only about fidelity but also about optimality. Prior work has addressed this challenge in restricted settings, including log-linear-time algorithms for computing minimum-size abductive explanations in linear models without rejection, as well as a polynomial-time method based on linear programming for computing abductive explanations, without guarantees of minimum size, for linear models with a reject option. In this work, we bridge these lines of research by computing minimum-size abductive explanations for linear models with a reject option. For accepted instances, we adapt the log-linear algorithm to efficiently compute optimal explanations. For rejected instances, we formulate a 0-1 integer linear programming problem that characterizes minimum-size abductive explanations of rejection. Although this formulation is NP-hard in theory, our experimental results show that it is consistently more efficient in practice than the linear-programming-based approach that does not guarantee minimum-size explanations.
💡 Research Summary
**
The paper introduces MINABRO (Minimum‑size ABductive explanations for linear models with a Reject Option), a framework that produces faithful, concise, and computationally efficient explanations for linear classifiers equipped with a reject option. In high‑stakes domains such as healthcare and finance, models often abstain from making a prediction when the evidence is insufficient; explaining why a particular instance is rejected is crucial for informed human intervention. Existing abductive explanation methods either lack guarantees of minimal size or are limited to models without rejection, and computing minimum‑size abductive explanations is generally NP‑hard.
For accepted predictions (i.e., the model outputs +1 or –1), MINABRO adapts the log‑linear‑time greedy algorithm previously proposed for plain linear classifiers. The algorithm sorts features by their contribution to the linear score, then incrementally adds the most influential features while checking whether the current subset still forces the original prediction given the reject thresholds. This procedure retains O(n log n) complexity, where n is the number of features, and yields provably minimum‑size explanations.
For rejected predictions (output 0), the authors formulate a 0‑1 integer linear programming (ILP) problem that captures the requirement: a subset of features is an explanation if, when those features are fixed to their observed values, the linear score remains within the reject interval
Comments & Academic Discussion
Loading comments...
Leave a Comment