Small additive ensembles of symbolic rules that offer interpretable prediction models traditionally use rule conditions based on threshold propositions, resulting in axis-parallel polytopes as decision regions.
A new approach introduces logical propositions with learnable sparse linear transformations of input variables, enabling decision regions as general polytopes with oblique faces.
The proposed learning method utilizes a sequential greedy optimization based on logistic regression to efficiently construct rule ensembles with reduced model complexity across benchmark datasets.
Experimental results show that the new method achieves the same test risk as state-of-the-art methods while decreasing model complexity.