Rule Extraction with Reject Option
2025 (English)In: Machine Learning and Soft Computing: 9th International Conference, ICMLSC 2025, Tokyo, Japan, January 24–26, 2025, Revised Selected Papers, Part I / [ed] L. Huang, Springer, 2025, Vol. 2487, p. 278-300Conference paper, Published paper (Refereed)
Abstract [en]
In many applications, it is vital to simultaneously obtain the best possible predictive performance and be able to understand the underlying logic. Rule extraction can then be employed, by having predictions made by an opaque model and an extracted transparent model to provide explanations and allowing inspection and analysis of relationships found. Using this setup, the quality of both global and local explanations depend on the fidelity of the extracted model. In this paper, a novel approach to rule extraction, which includes a reject option based on well-calibrated fidelity estimates, is presented and empirically evaluated. With the proposed method, the user can balance the required fidelity against the number of instances explained, by rejecting explanations in parts of feature space where the extracted model is generally weak in approximating the opaque model, or when the two models don’t agree. The transparent model can be visualized, using different rejection levels, thus identifying the parts of feature space where it is a good approximation of the opaque model. Empirical investigation, using a number of benchmark data sets, shows that extracted models are highly faithful, while often being small enough to be comprehensible. Most importantly, it is demonstrated how both local and global explanations can be provided through rule extraction with a reject option by leveraging well-calibrated fidelity estimates.
Place, publisher, year, edition, pages
Springer, 2025. Vol. 2487, p. 278-300
Series
Communications in Computer and Information Science, ISSN 1865-0929, E-ISSN 1865-0937 ; 2487
Keywords [en]
calibration, explanations, reject option, rule extraction, Benchmark data, Data set, Empirical investigation, Explanation, Feature space, Predictive performance, Rejection levels, Rules extraction
National Category
Information Systems
Identifiers
URN: urn:nbn:se:hj:diva-68365DOI: 10.1007/978-981-96-6400-9_21Scopus ID: 2-s2.0-105006899968ISBN: 978-981-96-6399-6 (print)ISBN: 978-981-96-6400-9 (electronic)OAI: oai:DiVA.org:hj-68365DiVA, id: diva2:1967279
Conference
9th International Conference, ICMLSC, 2025, Tokyo, Japan, January 24–26, 2025
Funder
Knowledge Foundation2025-06-112025-06-112026-01-19Bibliographically approved