Recently, we have proposed a new decision tree family called soft decision trees where a node chooses both its left and right children with different probabilities as given by a gating function, different from a hard decision node which chooses one of the two. In this paper, we extend the original algorithm by introducing local dimension reduction via L1 and L2 regularization for feature selection and smoother fitting. We compare our novel approach with the standard decision tree algorithms over 27 classification data sets. We see that both regularized versions have similar generalization ability with less complexity in terms of number of nodes, where seems to work slightly better than L1. © 2013 Springer International Publishing.
CITATION STYLE
Yildiz, O. T., & Alpaydin, E. (2014). Regularizing soft decision trees. In Lecture Notes in Electrical Engineering (Vol. 264 LNEE, pp. 15–21). Springer Verlag. https://doi.org/10.1007/978-3-319-01604-7_2
Mendeley helps you to discover research relevant for your work.