The effect of splitting on random forests
Gespeichert in:
Verfasser / Beitragende:
[Hemant Ishwaran]
Ort, Verlag, Jahr:
2015
Enthalten in:
Machine Learning, 99/1(2015-04-01), 75-118
Format:
Artikel (online)
Online Zugang:
| LEADER | caa a22 4500 | ||
|---|---|---|---|
| 001 | 605478465 | ||
| 003 | CHVBK | ||
| 005 | 20210128100405.0 | ||
| 007 | cr unu---uuuuu | ||
| 008 | 210128e20150401xx s 000 0 eng | ||
| 024 | 7 | 0 | |a 10.1007/s10994-014-5451-2 |2 doi |
| 035 | |a (NATIONALLICENCE)springer-10.1007/s10994-014-5451-2 | ||
| 100 | 1 | |a Ishwaran |D Hemant |u Division of Biostatistics, University of Miami, 1120 NW 14th Street, 33136, Miami, FL, USA |4 aut | |
| 245 | 1 | 4 | |a The effect of splitting on random forests |h [Elektronische Daten] |c [Hemant Ishwaran] |
| 520 | 3 | |a The effect of a splitting rule on random forests (RF) is systematically studied for regression and classification problems. A class of weighted splitting rules, which includes as special cases CART weighted variance splitting and Gini index splitting, are studied in detail and shown to possess a unique adaptive property to signal and noise. We show for noisy variables that weighted splitting favors end-cut splits. While end-cut splits have traditionally been viewed as undesirable for single trees, we argue for deeply grown trees (a trademark of RF) end-cut splitting is useful because: (a) it maximizes the sample size making it possible for a tree to recover from a bad split, and (b) if a branch repeatedly splits on noise, the tree minimal node size will be reached which promotes termination of the bad branch. For strong variables, weighted variance splitting is shown to possess the desirable property of splitting at points of curvature of the underlying target function. This adaptivity to both noise and signal does not hold for unweighted and heavy weighted splitting rules. These latter rules are either too greedy, making them poor at recognizing noisy scenarios, or they are overly ECP aggressive, making them poor at recognizing signal. These results also shed light on pure random splitting and show that such rules are the least effective. On the other hand, because randomized rules are desirable because of their computational efficiency, we introduce a hybrid method employing random split-point selection which retains the adaptive property of weighted splitting rules while remaining computational efficient. | |
| 540 | |a The Author(s), 2014 | ||
| 690 | 7 | |a CART |2 nationallicence | |
| 690 | 7 | |a End-cut preference |2 nationallicence | |
| 690 | 7 | |a Law of the iterated logarithm |2 nationallicence | |
| 690 | 7 | |a Splitting rule |2 nationallicence | |
| 690 | 7 | |a Split-point |2 nationallicence | |
| 773 | 0 | |t Machine Learning |d Springer US; http://www.springer-ny.com |g 99/1(2015-04-01), 75-118 |x 0885-6125 |q 99:1<75 |1 2015 |2 99 |o 10994 | |
| 856 | 4 | 0 | |u https://doi.org/10.1007/s10994-014-5451-2 |q text/html |z Onlinezugriff via DOI |
| 898 | |a BK010053 |b XK010053 |c XK010000 | ||
| 900 | 7 | |a Metadata rights reserved |b Springer special CC-BY-NC licence |2 nationallicence | |
| 908 | |D 1 |a research-article |2 jats | ||
| 949 | |B NATIONALLICENCE |F NATIONALLICENCE |b NL-springer | ||
| 950 | |B NATIONALLICENCE |P 856 |E 40 |u https://doi.org/10.1007/s10994-014-5451-2 |q text/html |z Onlinezugriff via DOI | ||
| 950 | |B NATIONALLICENCE |P 100 |E 1- |a Ishwaran |D Hemant |u Division of Biostatistics, University of Miami, 1120 NW 14th Street, 33136, Miami, FL, USA |4 aut | ||
| 950 | |B NATIONALLICENCE |P 773 |E 0- |t Machine Learning |d Springer US; http://www.springer-ny.com |g 99/1(2015-04-01), 75-118 |x 0885-6125 |q 99:1<75 |1 2015 |2 99 |o 10994 | ||