R. Blanquero, E. Carrizosa, C. Molero-Río, D. Romero Morales

Classic regression trees are defined by a set of orthogonal cuts, i.e., the branching rules are of the form variable X not lower than threshold c. Oblique cuts, with at least two variables, have also been proposed, leading to generally more accurate and smaller-sized trees. The variables and thresholds are selected by a greedy procedure. The use of a greedy strategy yields low computational cost, but may lead to myopic decisions. The latest advances in Optimization techniques have motivated further research on procedures to build optimal regression trees. In this talk, we propose a continuous optimization approach to tackle this problem. This is achieved by including a cumulative density function that will indicate the path to be followed inside the tree, yielding a random implementation of the cuts. In contrast to classic regression trees, our formulation is flexible enough since sparsity or performance guarantee in a subsample could be easily included.

Keywords: Classification and Regression Trees; Optimal Decision Trees; Nonlinear Programming

Scheduled

AM-2 Multivariate Analysis
September 4, 2019  2:45 PM
I3L8. Georgina Blanes building


Other papers in the same session


Cookie policy

We use cookies in order to be able to identify and authenticate you on the website. They are necessary for the correct functioning of it, and therefore they can not be disabled. If you continue browsing the website, you are agreeing with their acceptance, as well as our Privacy Policy.

Additionally, we use Google Analytics in order to analyze the website traffic. They also use cookies and you can accept or refuse them with the buttons below.

You can read more details about our Cookie Policy and our Privacy Policy.