Format

Send to

Choose Destination
J Mach Learn Res. 2014 Jan;15(1):1625-1651.

Confidence Intervals for Random Forests: The Jackknife and the Infinitesimal Jackknife.

Author information

1
Department of Statistics, Stanford University, Stanford, CA 94305, USA.

Abstract

We study the variability of predictions made by bagged learners and random forests, and show how to estimate standard errors for these methods. Our work builds on variance estimates for bagging proposed by Efron (1992, 2013) that are based on the jackknife and the infinitesimal jackknife (IJ). In practice, bagged predictors are computed using a finite number B of bootstrap replicates, and working with a large B can be computationally expensive. Direct applications of jackknife and IJ estimators to bagging require B = Θ(n1.5) bootstrap replicates to converge, where n is the size of the training set. We propose improved versions that only require B = Θ(n) replicates. Moreover, we show that the IJ estimator requires 1.7 times less bootstrap replicates than the jackknife to achieve a given accuracy. Finally, we study the sampling distributions of the jackknife and IJ variance estimates themselves. We illustrate our findings with multiple experiments and simulation studies.

KEYWORDS:

Monte Carlo noise; bagging; jackknife methods; variance estimation

PMID:
25580094
PMCID:
PMC4286302

Supplemental Content

Full text links

Icon for PubMed Central
Loading ...
Support Center