IEEE Access (Jan 2020)

Learning Explainable Decision Rules via Maximum Satisfiability

  • Henrik E. C. Cao,
  • Riku Sarlin,
  • Alexander Jung

DOI
https://doi.org/10.1109/access.2020.3041040
Journal volume & issue
Vol. 8
pp. 218180 – 218185

Abstract

Read online

Decision trees are a popular choice for providing explainable machine learning, since they make explicit how different features contribute towards the prediction. We apply tools from constraint satisfaction to learn optimal decision trees in the form of sparse k-CNF (Conjunctive Normal Form) rules. We develop two methods offering different trade-offs between accuracy and computational complexity: one offline method that learns decision trees using the entire training dataset and one online method that learns decision trees over a local subset of the training dataset. This subset is obtained from training examples near a query point. The developed methods are applied on a number of datasets both in an online and an offline setting. We found that our methods learn decision trees which are significantly more accurate than those learned by existing heuristic approaches. However, the global decision tree model tends to be computationally more expensive compared to heuristic approaches. The online method is faster to train and finds smaller decision trees with an accuracy comparable to that of the k-nearest-neighbour method.

Keywords