Abstract: Adopting opaque machine learning predictors, which achieve very high predictive performance, often necessitates incorporating symbolic knowledge-extraction techniques. These techniques aim to explain the opaque predictions, thus making them applicable in high-stakes scenarios. The development of symbolic knowledge-extraction procedures is evolving alongside the dynamic machine learning landscape. However, there are recurring drawbacks that tend to be overlooked or addressed in a suboptimum way. Common examples include the non-exhaustiveness of the global explanations generated for a black-box predictor or the unwanted discretisation introduced in the prediction of continuous variables. To tackle these challenges, in this work, we introduce the HEx algorithm, its formalisation and its properties. This algorithm aims to obtain a symbolic, hierarchical representation of the knowledge acquired by opaque machine learning classifiers and regressors, always ensuring knowledge exhaustiveness and avoiding any output discretisation. Experiments demonstrating the superior capabilities of HEx compared to state-of-the-art competitors in terms of predictive performance, completeness, and human readability are presented.
Loading