fbpx
Wikipedia

Bootstrapping (statistics)

Bootstrapping is any test or metric that uses random sampling with replacement (e.g. mimicking the sampling process), and falls under the broader class of resampling methods. Bootstrapping assigns measures of accuracy (bias, variance, confidence intervals, prediction error, etc.) to sample estimates.[1][2] This technique allows estimation of the sampling distribution of almost any statistic using random sampling methods.[3][4]

Bootstrapping estimates the properties of an estimand (such as its variance) by measuring those properties when sampling from an approximating distribution. One standard choice for an approximating distribution is the empirical distribution function of the observed data. In the case where a set of observations can be assumed to be from an independent and identically distributed population, this can be implemented by constructing a number of resamples with replacement, of the observed data set (and of equal size to the observed data set).

It may also be used for constructing hypothesis tests.[5] It is often used as an alternative to statistical inference based on the assumption of a parametric model when that assumption is in doubt, or where parametric inference is impossible or requires complicated formulas for the calculation of standard errors.

History edit

The bootstrap was published by Bradley Efron in "Bootstrap methods: another look at the jackknife" (1979),[6][7][8] inspired by earlier work on the jackknife.[9][10][11] Improved estimates of the variance were developed later.[12][13] A Bayesian extension was developed in 1981.[14] The bias-corrected and accelerated ( ) bootstrap was developed by Efron in 1987,[15] and the approximate bootstrap confidence interval (ABC, or approximate  ) procedure in 1992.[16] Other names that Efron's colleagues suggested for the "bootstrap" method were: Swiss Army Knife, Meat Axe, Swan-Dive, Jack-Rabbit, and Shotgun.[8]

Approach edit

 
A sample is drawn from a population. From this sample, resamples are generated by drawing with replacement (orange). Data points that were drawn more than once (which happens for approx. 26.4% of data points) are shown in red and slightly offsetted. From the resamples, the statistic   is calculated and, therefore, a histogram can be calculated to estimate the distribution of  .

The basic idea of bootstrapping is that inference about a population from sample data (sample → population) can be modeled by resampling the sample data and performing inference about a sample from resampled data (resampled → sample). [17] As the population is unknown, the true error in a sample statistic against its population value is unknown. In bootstrap-resamples, the 'population' is in fact the sample, and this is known; hence the quality of inference of the 'true' sample from resampled data (resampled → sample) is measurable.

More formally, the bootstrap works by treating inference of the true probability distribution J, given the original data, as being analogous to an inference of the empirical distribution Ĵ, given the resampled data. The accuracy of inferences regarding Ĵ using the resampled data can be assessed because we know Ĵ. If Ĵ is a reasonable approximation to J, then the quality of inference on J can in turn be inferred.

As an example, assume we are interested in the average (or mean) height of people worldwide. We cannot measure all the people in the global population, so instead, we sample only a tiny part of it, and measure that. Assume the sample is of size N; that is, we measure the heights of N individuals. From that single sample, only one estimate of the mean can be obtained. In order to reason about the population, we need some sense of the variability of the mean that we have computed. The simplest bootstrap method involves taking the original data set of heights, and, using a computer, sampling from it to form a new sample (called a 'resample' or bootstrap sample) that is also of size N. The bootstrap sample is taken from the original by using sampling with replacement (e.g. we might 'resample' 5 times from [1,2,3,4,5] and get [2,5,4,4,1]), so, assuming N is sufficiently large, for all practical purposes there is virtually zero probability that it will be identical to the original "real" sample. This process is repeated a large number of times (typically 1,000 or 10,000 times), and for each of these bootstrap samples, we compute its mean (each of these is called a "bootstrap estimate"). We now can create a histogram of bootstrap means. This histogram provides an estimate of the shape of the distribution of the sample mean from which we can answer questions about how much the mean varies across samples. (The method here, described for the mean, can be applied to almost any other statistic or estimator.)

Discussion edit

Advantages edit

A great advantage of bootstrap is its simplicity. It is a straightforward way to derive estimates of standard errors and confidence intervals for complex estimators of the distribution, such as percentile points, proportions, Odds ratio, and correlation coefficients. However, despite its simplicity, bootstrapping can be applied to complex sampling designs (e.g. for population divided into s strata with ns observations per strata, bootstrapping can be applied for each stratum).[18] Bootstrap is also an appropriate way to control and check the stability of the results. Although for most problems it is impossible to know the true confidence interval, bootstrap is asymptotically more accurate than the standard intervals obtained using sample variance and assumptions of normality.[19] Bootstrapping is also a convenient method that avoids the cost of repeating the experiment to get other groups of sample data.

Disadvantages edit

Bootstrapping depends heavily on the estimator used and, though simple, naive use of bootstrapping will not always yield asymptotically valid results and can lead to inconsistency.[20] Although bootstrapping is (under some conditions) asymptotically consistent, it does not provide general finite-sample guarantees. The result may depend on the representative sample. The apparent simplicity may conceal the fact that important assumptions are being made when undertaking the bootstrap analysis (e.g. independence of samples or large enough of a sample size) where these would be more formally stated in other approaches. Also, bootstrapping can be time-consuming and there are not many available software for bootstrapping as it is difficult to automate using traditional statistical computer packages.[18]

Recommendations edit

Scholars have recommended more bootstrap samples as available computing power has increased. If the results may have substantial real-world consequences, then one should use as many samples as is reasonable, given available computing power and time. Increasing the number of samples cannot increase the amount of information in the original data; it can only reduce the effects of random sampling errors which can arise from a bootstrap procedure itself. Moreover, there is evidence that numbers of samples greater than 100 lead to negligible improvements in the estimation of standard errors.[21] In fact, according to the original developer of the bootstrapping method, even setting the number of samples at 50 is likely to lead to fairly good standard error estimates.[22]

Adèr et al. recommend the bootstrap procedure for the following situations:[23]

  • When the theoretical distribution of a statistic of interest is complicated or unknown. Since the bootstrapping procedure is distribution-independent it provides an indirect method to assess the properties of the distribution underlying the sample and the parameters of interest that are derived from this distribution.
  • When the sample size is insufficient for straightforward statistical inference. If the underlying distribution is well-known, bootstrapping provides a way to account for the distortions caused by the specific sample that may not be fully representative of the population.
  • When power calculations have to be performed, and a small pilot sample is available. Most power and sample size calculations are heavily dependent on the standard deviation of the statistic of interest. If the estimate used is incorrect, the required sample size will also be wrong. One method to get an impression of the variation of the statistic is to use a small pilot sample and perform bootstrapping on it to get impression of the variance.

However, Athreya has shown[24] that if one performs a naive bootstrap on the sample mean when the underlying population lacks a finite variance (for example, a power law distribution), then the bootstrap distribution will not converge to the same limit as the sample mean. As a result, confidence intervals on the basis of a Monte Carlo simulation of the bootstrap could be misleading. Athreya states that "Unless one is reasonably sure that the underlying distribution is not heavy tailed, one should hesitate to use the naive bootstrap".

Types of bootstrap scheme edit

In univariate problems, it is usually acceptable to resample the individual observations with replacement ("case resampling" below) unlike subsampling, in which resampling is without replacement and is valid under much weaker conditions compared to the bootstrap. In small samples, a parametric bootstrap approach might be preferred. For other problems, a smooth bootstrap will likely be preferred.

For regression problems, various other alternatives are available.[1]

Case resampling edit

The bootstrap is generally useful for estimating the distribution of a statistic (e.g. mean, variance) without using normality assumptions (as required, e.g., for a z-statistic or a t-statistic). In particular, the bootstrap is useful when there is no analytical form or an asymptotic theory (e.g., an applicable central limit theorem) to help estimate the distribution of the statistics of interest. This is because bootstrap methods can apply to most random quantities, e.g., the ratio of variance and mean. There are at least two ways of performing case resampling.

  1. The Monte Carlo algorithm for case resampling is quite simple. First, we resample the data with replacement, and the size of the resample must be equal to the size of the original data set. Then the statistic of interest is computed from the resample from the first step. We repeat this routine many times to get a more precise estimate of the Bootstrap distribution of the statistic.[1]
  2. The 'exact' version for case resampling is similar, but we exhaustively enumerate every possible resample of the data set. This can be computationally expensive as there are a total of   different resamples, where n is the size of the data set. Thus for n = 5, 10, 20, 30 there are 126, 92378, 6.89 × 1010 and 5.91 × 1016 different resamples respectively.[25]

Estimating the distribution of sample mean edit

Consider a coin-flipping experiment. We flip the coin and record whether it lands heads or tails. Let X = x1, x2, …, x10 be 10 observations from the experiment. xi = 1 if the i th flip lands heads, and 0 otherwise. By invoking the assumption that the average of the coin flips is normally distributed, we can use the t-statistic to estimate the distribution of the sample mean,

 

Such a normality assumption can be justified either as an approximation of the distribution of each individual coin flip or as an approximation of the distribution of the average of a large number of coin flips. The former is a poor approximation because the true distribution of the coin flips is Bernoulli instead of normal. The latter is a valid approximation in infinitely large samples due to the central limit theorem.

However, if we are not ready to make such a justification, then we can use the bootstrap instead. Using case resampling, we can derive the distribution of  . We first resample the data to obtain a bootstrap resample. An example of the first resample might look like this X1* = x2, x1, x10, x10, x3, x4, x6, x7, x1, x9. There are some duplicates since a bootstrap resample comes from sampling with replacement from the data. Also the number of data points in a bootstrap resample is equal to the number of data points in our original observations. Then we compute the mean of this resample and obtain the first bootstrap mean: μ1*. We repeat this process to obtain the second resample X2* and compute the second bootstrap mean μ2*. If we repeat this 100 times, then we have μ1*, μ2*, ..., μ100*. This represents an empirical bootstrap distribution of sample mean. From this empirical distribution, one can derive a bootstrap confidence interval for the purpose of hypothesis testing.

Regression edit

In regression problems, case resampling refers to the simple scheme of resampling individual cases – often rows of a data set. For regression problems, as long as the data set is fairly large, this simple scheme is often acceptable.[26][27][28] However, the method is open to criticism[citation needed].[18]

In regression problems, the explanatory variables are often fixed, or at least observed with more control than the response variable. Also, the range of the explanatory variables defines the information available from them. Therefore, to resample cases means that each bootstrap sample will lose some information. As such, alternative bootstrap procedures should be considered.

Bayesian bootstrap edit

Bootstrapping can be interpreted in a Bayesian framework using a scheme that creates new data sets through reweighting the initial data. Given a set of   data points, the weighting assigned to data point   in a new data set   is  , where   is a low-to-high ordered list of   uniformly distributed random numbers on  , preceded by 0 and succeeded by 1. The distributions of a parameter inferred from considering many such data sets   are then interpretable as posterior distributions on that parameter.[29]

Smooth bootstrap edit

Under this scheme, a small amount of (usually normally distributed) zero-centered random noise is added onto each resampled observation. This is equivalent to sampling from a kernel density estimate of the data. Assume K to be a symmetric kernel density function with unit variance. The standard kernel estimator   of   is

  [30]

where   is the smoothing parameter. And the corresponding distribution function estimator   is

  [30]

Parametric bootstrap edit

Based on the assumption that the original data set is a realization of a random sample from a distribution of a specific parametric type, in this case a parametric model is fitted by parameter θ, often by maximum likelihood, and samples of random numbers are drawn from this fitted model. Usually the sample drawn has the same sample size as the original data. Then the estimate of original function F can be written as  . This sampling process is repeated many times as for other bootstrap methods. Considering the centered sample mean in this case, the random sample original distribution function   is replaced by a bootstrap random sample with function  , and the probability distribution of   is approximated by that of  , where  , which is the expectation corresponding to  .[31] The use of a parametric model at the sampling stage of the bootstrap methodology leads to procedures which are different from those obtained by applying basic statistical theory to inference for the same model.

Resampling residuals edit

Another approach to bootstrapping in regression problems is to resample residuals. The method proceeds as follows.

  1. Fit the model and retain the fitted values   and the residuals  .
  2. For each pair, (xi, yi), in which xi is the (possibly multivariate) explanatory variable, add a randomly resampled residual,  , to the fitted value  . In other words, create synthetic response variables   where j is selected randomly from the list (1, ..., n) for every i.
  3. Refit the model using the fictitious response variables  , and retain the quantities of interest (often the parameters,  , estimated from the synthetic  ).
  4. Repeat steps 2 and 3 a large number of times.

This scheme has the advantage that it retains the information in the explanatory variables. However, a question arises as to which residuals to resample. Raw residuals are one option; another is studentized residuals (in linear regression). Although there are arguments in favor of using studentized residuals; in practice, it often makes little difference, and it is easy to compare the results of both schemes.

Gaussian process regression bootstrap edit

When data are temporally correlated, straightforward bootstrapping destroys the inherent correlations. This method uses Gaussian process regression (GPR) to fit a probabilistic model from which replicates may then be drawn. GPR is a Bayesian non-linear regression method. A Gaussian process (GP) is a collection of random variables, any finite number of which have a joint Gaussian (normal) distribution. A GP is defined by a mean function and a covariance function, which specify the mean vectors and covariance matrices for each finite collection of the random variables.[32]

Regression model:

    is a noise term.

Gaussian process prior:

For any finite collection of variables, x1, ..., xn, the function outputs   are jointly distributed according to a multivariate Gaussian with mean   and covariance matrix  

Assume   Then  ,

where  , and   is the standard Kronecker delta function.[32]

Gaussian process posterior:

According to GP prior, we can get

 ,

where   and  

Let x1*,...,xs* be another finite collection of variables, it's obvious that

 ,

where  ,  ,  

According to the equations above, the outputs y are also jointly distributed according to a multivariate Gaussian. Thus,

 

where  ,  ,  , and   is   identity matrix.[32]

Wild bootstrap edit

The wild bootstrap, proposed originally by Wu (1986),[33] is suited when the model exhibits heteroskedasticity. The idea is, as the residual bootstrap, to leave the regressors at their sample value, but to resample the response variable based on the residuals values. That is, for each replicate, one computes a new   based on

 

so the residuals are randomly multiplied by a random variable   with mean 0 and variance 1. For most distributions of   (but not Mammen's), this method assumes that the 'true' residual distribution is symmetric and can offer advantages over simple residual sampling for smaller sample sizes. Different forms are used for the random variable  , such as

  • A distribution suggested by Mammen (1993).[34]
 
Approximately, Mammen's distribution is:
 
 

Block bootstrap edit

The block bootstrap is used when the data, or the errors in a model, are correlated. In this case, a simple case or residual resampling will fail, as it is not able to replicate the correlation in the data. The block bootstrap tries to replicate the correlation by resampling inside blocks of data (see Blocking (statistics)). The block bootstrap has been used mainly with data correlated in time (i.e. time series) but can also be used with data correlated in space, or among groups (so-called cluster data).

Time series: Simple block bootstrap edit

In the (simple) block bootstrap, the variable of interest is split into non-overlapping blocks.

Time series: Moving block bootstrap edit

In the moving block bootstrap, introduced by Künsch (1989),[35] data is split into n − b + 1 overlapping blocks of length b: Observation 1 to b will be block 1, observation 2 to b + 1 will be block 2, etc. Then from these n − b + 1 blocks, n/b blocks will be drawn at random with replacement. Then aligning these n/b blocks in the order they were picked, will give the bootstrap observations.

This bootstrap works with dependent data, however, the bootstrapped observations will not be stationary anymore by construction. But, it was shown that varying randomly the block length can avoid this problem.[36] This method is known as the stationary bootstrap. Other related modifications of the moving block bootstrap are the Markovian bootstrap and a stationary bootstrap method that matches subsequent blocks based on standard deviation matching.

Time series: Maximum entropy bootstrap edit

Vinod (2006),[37] presents a method that bootstraps time series data using maximum entropy principles satisfying the Ergodic theorem with mean-preserving and mass-preserving constraints. There is an R package, meboot,[38] that utilizes the method, which has applications in econometrics and computer science.

Cluster data: block bootstrap edit

Cluster data describes data where many observations per unit are observed. This could be observing many firms in many states or observing students in many classes. In such cases, the correlation structure is simplified, and one does usually make the assumption that data is correlated within a group/cluster, but independent between groups/clusters. The structure of the block bootstrap is easily obtained (where the block just corresponds to the group), and usually only the groups are resampled, while the observations within the groups are left unchanged. Cameron et al. (2008) discusses this for clustered errors in linear regression.[39]

Methods for improving computational efficiency edit

The bootstrap is a powerful technique although may require substantial computing resources in both time and memory. Some techniques have been developed to reduce this burden. They can generally be combined with many of the different types of Bootstrap schemes and various choices of statistics.

Poisson bootstrap edit

 
Graph showing the convergence of the Binomial Distribution to the Poisson as the Binomial parameters are n*p=1 and n grows

The ordinary bootstrap requires the random selection of n elements from a list, which is equivalent to drawing from a binomial distribution. This may require a large number of passes over the data and is challenging to run these computations in parallel. For large values of n, the Poisson bootstrap is an efficient method of generating bootstrapped data sets.[40] When generating a single bootstrap sample, instead of randomly drawing from the sample data with replacement, each data point is assigned a random weight distributed according to the Poisson distribution with  . For large sample data, this will approximate random sampling with replacement. This is due to the following approximation:

 

This method also lends itself well to streaming data and growing data sets, since the total number of samples does not need to be known in advance of beginning to take bootstrap samples.

For large enough n, the results are relatively similar to the original bootstrap estimations.[41]

A way to improve on the poisson bootstrap, termed "sequential bootstrap", is by taking the first samples so that the proportion of unique values is ≈0.632 of the original sample size n. This provides a distribution with main empirical characteristics being within a distance of  .[42] Empirical investigation has shown this method can yield good results.[43] This is related to the reduced bootstrap method.[44]

Bag of Little Bootstraps edit

For massive data sets, it is often computationally prohibitive to hold all the sample data in memory and resample from the sample data. The Bag of Little Bootstraps (BLB)[45] provides a method of pre-aggregating data before bootstrapping to reduce computational constraints. This works by partitioning the data set into   equal-sized buckets and aggregating the data within each bucket. This pre-aggregated data set becomes the new sample data over which to draw samples with replacement. This method is similar to the Block Bootstrap, but the motivations and definitions of the blocks are very different. Under certain assumptions, the sample distribution should approximate the full bootstrapped scenario. One constraint is the number of buckets  where   and the authors recommend usage of   as a general solution.

Choice of statistic edit

The bootstrap distribution of a point estimator of a population parameter has been used to produce a bootstrapped confidence interval for the parameter's true value if the parameter can be written as a function of the population's distribution.

Population parameters are estimated with many point estimators. Popular families of point-estimators include mean-unbiased minimum-variance estimators, median-unbiased estimators, Bayesian estimators (for example, the posterior distribution's mode, median, mean), and maximum-likelihood estimators.

A Bayesian point estimator and a maximum-likelihood estimator have good performance when the sample size is infinite, according to asymptotic theory. For practical problems with finite samples, other estimators may be preferable. Asymptotic theory suggests techniques that often improve the performance of bootstrapped estimators; the bootstrapping of a maximum-likelihood estimator may often be improved using transformations related to pivotal quantities.[46]

Deriving confidence intervals from the bootstrap distribution edit

The bootstrap distribution of a parameter-estimator has been used to calculate confidence intervals for its population-parameter.[1]

Bias, asymmetry, and confidence intervals edit

  • Bias: The bootstrap distribution and the sample may disagree systematically, in which case bias may occur.
    If the bootstrap distribution of an estimator is symmetric, then percentile confidence-interval are often used; such intervals are appropriate especially for median-unbiased estimators of minimum risk (with respect to an absolute loss function). Bias in the bootstrap distribution will lead to bias in the confidence interval.
    Otherwise, if the bootstrap distribution is non-symmetric, then percentile confidence intervals are often inappropriate.

Methods for bootstrap confidence intervals edit

There are several methods for constructing confidence intervals from the bootstrap distribution of a real parameter:

  • Basic bootstrap,[46] also known as the Reverse Percentile Interval.[47] The basic bootstrap is a simple scheme to construct the confidence interval: one simply takes the empirical quantiles from the bootstrap distribution of the parameter (see Davison and Hinkley 1997, equ. 5.6 p. 194):
  where   denotes the   percentile of the bootstrapped coefficients  .
  • Percentile bootstrap. The percentile bootstrap proceeds in a similar way to the basic bootstrap, using percentiles of the bootstrap distribution, but with a different formula (note the inversion of the left and right quantiles):
  where   denotes the   percentile of the bootstrapped coefficients  .
See Davison and Hinkley (1997, equ. 5.18 p. 203) and Efron and Tibshirani (1993, equ 13.5 p. 171).
This method can be applied to any statistic. It will work well in cases where the bootstrap distribution is symmetrical and centered on the observed statistic[48] and where the sample statistic is median-unbiased and has maximum concentration (or minimum risk with respect to an absolute value loss function). When working with small sample sizes (i.e., less than 50), the basic / reversed percentile and percentile confidence intervals for (for example) the variance statistic will be too narrow. So that with a sample of 20 points, 90% confidence interval will include the true variance only 78% of the time.[49] The basic / reverse percentile confidence intervals are easier to justify mathematically[50][47] but they are less accurate in general than percentile confidence intervals, and some authors discourage their use.[47]
  • Studentized bootstrap. The studentized bootstrap, also called bootstrap-t, is computed analogously to the standard confidence interval, but replaces the quantiles from the normal or student approximation by the quantiles from the bootstrap distribution of the Student's t-test (see Davison and Hinkley 1997, equ. 5.7 p. 194 and Efron and Tibshirani 1993 equ 12.22, p. 160):
  where   denotes the   percentile of the bootstrapped Student's t-test  , and   is the estimated standard error of the coefficient in the original model.
The studentized test enjoys optimal properties as the statistic that is bootstrapped is pivotal (i.e. it does not depend on nuisance parameters as the t-test follows asymptotically a N(0,1) distribution), unlike the percentile bootstrap.
  • Bias-corrected bootstrap – adjusts for bias in the bootstrap distribution.
  • Accelerated bootstrap – The bias-corrected and accelerated (BCa) bootstrap, by Efron (1987),[15] adjusts for both bias and skewness in the bootstrap distribution. This approach is accurate in a wide variety of settings, has reasonable computation requirements, and produces reasonably narrow intervals.[15]

Bootstrap hypothesis testing edit

Efron and Tibshirani[1] suggest the following algorithm for comparing the means of two independent samples: Let   be a random sample from distribution F with sample mean   and sample variance  . Let   be another, independent random sample from distribution G with mean   and variance  

  1. Calculate the test statistic  
  2. Create two new data sets whose values are   and   where   is the mean of the combined sample.
  3. Draw a random sample ( ) of size   with replacement from   and another random sample ( ) of size   with replacement from  .
  4. Calculate the test statistic  
  5. Repeat 3 and 4   times (e.g.  ) to collect   values of the test statistic.
  6. Estimate the p-value as   where   when condition is true and 0 otherwise.

Example applications edit

Smoothed bootstrap edit

In 1878, Simon Newcomb took observations on the speed of light.[51] The data set contains two outliers, which greatly influence the sample mean. (The sample mean need not be a consistent estimator for any population mean, because no mean needs to exist for a heavy-tailed distribution.) A well-defined and robust statistic for the central tendency is the sample median, which is consistent and median-unbiased for the population median.

The bootstrap distribution for Newcomb's data appears below. We can reduce the discreteness of the bootstrap distribution by adding a small amount of random noise to each bootstrap sample. A conventional choice is to add noise with a standard deviation of   for a sample size n; this noise is often drawn from a Student-t distribution with n-1 degrees of freedom.[52] This results in an approximately-unbiased estimator for the variance of the sample mean. [53] This means that samples taken from the bootstrap distribution will have a variance which is, on average, equal to the variance of the total population.

Histograms of the bootstrap distribution and the smooth bootstrap distribution appear below. The bootstrap distribution of the sample-median has only a small number of values. The smoothed bootstrap distribution has a richer support. However, note that whether the smoothed or standard bootstrap procedure is favorable is case-by-case and is shown to depend on both the underlying distribution function and on the quantity being estimated.[54]

 

In this example, the bootstrapped 95% (percentile) confidence-interval for the population median is (26, 28.5), which is close to the interval for (25.98, 28.46) for the smoothed bootstrap.

Relation to other approaches to inference edit

Relationship to other resampling methods edit

The bootstrap is distinguished from:

  • the jackknife procedure, used to estimate biases of sample statistics and to estimate variances, and
  • cross-validation, in which the parameters (e.g., regression weights, factor loadings) that are estimated in one subsample are applied to another subsample.

For more details see resampling.

Bootstrap aggregating (bagging) is a meta-algorithm based on averaging model predictions obtained from models trained on multiple bootstrap samples.

U-statistics edit

In situations where an obvious statistic can be devised to measure a required characteristic using only a small number, r, of data items, a corresponding statistic based on the entire sample can be formulated. Given an r-sample statistic, one can create an n-sample statistic by something similar to bootstrapping (taking the average of the statistic over all subsamples of size r). This procedure is known to have certain good properties and the result is a U-statistic. The sample mean and sample variance are of this form, for r = 1 and r = 2.

Methods for proving consistency of bootstrap estimates edit

It is possible to use the central limit theorem to show the consistency of the bootstrap procedure for estimating the distribution of the sample mean.

Specifically, let us consider   independent identically distributed random variables with   and   for each  . Let  . In addition, for each  , conditional on  , let   be independent random variables with distribution equal to the empirical distribution of  . This is the sequence of bootstrap samples.

Then it can be shown that

 
where   represents probability conditional on  ,  ,  , and  .

To see this, note that   satisfies the Lindeberg condition, so the CLT holds.[55]

See also edit

References edit

  1. ^ a b c d e Efron, B.; Tibshirani, R. (1993). An Introduction to the Bootstrap. Boca Raton, FL: Chapman & Hall/CRC. ISBN 0-412-04231-2. software Archived 2012-07-12 at archive.today
  2. ^ Second Thoughts on the Bootstrap – Bradley Efron, 2003
  3. ^ Varian, H.(2005). "Bootstrap Tutorial". Mathematica Journal, 9, 768–775.
  4. ^ Weisstein, Eric W. "Bootstrap Methods." From MathWorld--A Wolfram Web Resource. http://mathworld.wolfram.com/BootstrapMethods.html
  5. ^ Lehmann E.L. (1992) "Introduction to Neyman and Pearson (1933) On the Problem of the Most Efficient Tests of Statistical Hypotheses". In: Breakthroughs in Statistics, Volume 1, (Eds Kotz, S., Johnson, N.L.), Springer-Verlag. ISBN 0-387-94037-5 (followed by reprinting of the paper).
  6. ^ Notes for Earliest Known Uses of Some of the Words of Mathematics: Bootstrap (John Aldrich)
  7. ^ Earliest Known Uses of Some of the Words of Mathematics (B) (Jeff Miller)
  8. ^ a b Efron, B. (1979). "Bootstrap methods: Another look at the jackknife". The Annals of Statistics. 7 (1): 1–26. doi:10.1214/aos/1176344552.
  9. ^ Quenouille M (1949) Approximate tests of correlation in time-series. J Roy Statist Soc Ser B 11 68–84
  10. ^ Tukey J (1958) Bias and confidence in not-quite large samples (abstract). Ann Math Statist 29 614
  11. ^ Jaeckel L (1972) The infinitesimal jackknife. Memorandum MM72-1215-11, Bell Lab
  12. ^ Bickel P, Freeman D (1981) Some asymptotic theory for the bootstrap. Ann Statist 9 1196–1217
  13. ^ Singh K (1981) On the asymptotic accuracy of Efron’s bootstrap. Ann Statist 9 1187–1195
  14. ^ Rubin D (1981). The Bayesian bootstrap. Ann Statist 9 130–134
  15. ^ a b c Efron, B. (1987). "Better Bootstrap Confidence Intervals". Journal of the American Statistical Association. 82 (397). Journal of the American Statistical Association, Vol. 82, No. 397: 171–185. doi:10.2307/2289144. JSTOR 2289144.
  16. ^ Dicicio, Thomas; Efron, Bradley (1992-06-01). "More accurate confidence intervals in exponential families". Biometrika. pp. 231–245. doi:10.2307/2336835. ISSN 0006-3444. OCLC 5545447518. Retrieved 2024-01-31.
  17. ^ Good, P. (2006) Resampling Methods. 3rd Ed. Birkhauser.
  18. ^ a b c "21 Bootstrapping Regression Models" (PDF). (PDF) from the original on 2015-07-24.
  19. ^ DiCiccio TJ, Efron B (1996) Bootstrap confidence intervals (with Discussion). Statistical Science 11: 189–228
  20. ^ Hinkley, David (1994-08-01). "[Bootstrap: More than a Stab in the Dark?]: Comment". Statistical Science. 9 (3). doi:10.1214/ss/1177010387. ISSN 0883-4237.
  21. ^ Goodhue, D.L., Lewis, W., & Thompson, R. (2012). Does PLS have advantages for small sample size or non-normal data? MIS Quarterly, 36(3), 981–1001.
  22. ^ Efron, B., Rogosa, D., & Tibshirani, R. (2004). Resampling methods of estimation. In N.J. Smelser, & P.B. Baltes (Eds.). International Encyclopedia of the Social & Behavioral Sciences (pp. 13216–13220). New York, NY: Elsevier.
  23. ^ Adèr, H. J., Mellenbergh G. J., & Hand, D. J. (2008). Advising on research methods: A consultant's companion. Huizen, The Netherlands: Johannes van Kessel Publishing. ISBN 978-90-79418-01-5.
  24. ^ Bootstrap of the mean in the infinite variance case Athreya, K.B. Ann Stats vol 15 (2) 1987 724–731
  25. ^ . Archived from the original on 2019-09-14. Retrieved 2019-12-09.
  26. ^ Jenkins, David G.; Quintana-Ascencio, Pedro F. (2020-02-21). "A solution to minimum sample size for regressions". PLOS ONE. 15 (2): e0229345. Bibcode:2020PLoSO..1529345J. doi:10.1371/journal.pone.0229345. ISSN 1932-6203. PMC 7034864. PMID 32084211.
  27. ^ Lumley, Thomas (2002). "The Importance of the Normality Assumption in Large Public Health Data Sets". Annual Review of Public Health. 23: 151–169. doi:10.1146/annurev.publhealth.23.100901.140546. PMID 11910059.
  28. ^ Li, Xiang; Wong, Wanling; Lamoureux, Ecosse L.; Wong, Tien Y. (2012-05-01). "Are Linear Regression Techniques Appropriate for Analysis When the Dependent (Outcome) Variable Is Not Normally Distributed?". Investigative Ophthalmology & Visual Science. 53 (6): 3082–3083. doi:10.1167/iovs.12-9967. ISSN 1552-5783. PMID 22618757.
  29. ^ Rubin, D. B. (1981). "The Bayesian bootstrap". Annals of Statistics, 9, 130.
  30. ^ a b WANG, SUOJIN (1995). "Optimizing the smoothed bootstrap". Ann. Inst. Statist. Math. 47: 65–80. doi:10.1007/BF00773412. S2CID 122041565.
  31. ^ A modern introduction to probability and statistics : understanding why and how. Dekking, Michel, 1946-. London: Springer. 2005. ISBN 978-1-85233-896-1. OCLC 262680588.{{cite book}}: CS1 maint: others (link)
  32. ^ a b c Kirk, Paul (2009). "Gaussian process regression bootstrapping: exploring the effects of uncertainty in time course data". Bioinformatics. 25 (10): 1300–1306. doi:10.1093/bioinformatics/btp139. PMC 2677737. PMID 19289448.
  33. ^ Wu, C.F.J. (1986). "Jackknife, bootstrap and other resampling methods in regression analysis (with discussions)" (PDF). Annals of Statistics. 14: 1261–1350. doi:10.1214/aos/1176350142.
  34. ^ Mammen, E. (Mar 1993). "Bootstrap and wild bootstrap for high dimensional linear models". Annals of Statistics. 21 (1): 255–285. doi:10.1214/aos/1176349025.
  35. ^ Künsch, H. R. (1989). "The Jackknife and the Bootstrap for General Stationary Observations". Annals of Statistics. 17 (3): 1217–1241. doi:10.1214/aos/1176347265.
  36. ^ Politis, D. N.; Romano, J. P. (1994). "The Stationary Bootstrap". Journal of the American Statistical Association. 89 (428): 1303–1313. doi:10.1080/01621459.1994.10476870. hdl:10983/25607.
  37. ^ Vinod, HD (2006). "Maximum entropy ensembles for time series inference in economics". Journal of Asian Economics. 17 (6): 955–978. doi:10.1016/j.asieco.2006.09.001.
  38. ^ Vinod, Hrishikesh; López-de-Lacalle, Javier (2009). "Maximum entropy bootstrap for time series: The meboot R package". Journal of Statistical Software. 29 (5): 1–19. doi:10.18637/jss.v029.i05.
  39. ^ Cameron, A. C.; Gelbach, J. B.; Miller, D. L. (2008). "Bootstrap-based improvements for inference with clustered errors" (PDF). Review of Economics and Statistics. 90 (3): 414–427. doi:10.1162/rest.90.3.414.
  40. ^ Chamandy, N; Muralidharan, O; Najmi, A; Naidu, S (2012). "Estimating Uncertainty for Massive Data Streams".
  41. ^ Hanley, James A., and Brenda MacGibbon. "Creating non-parametric bootstrap samples using Poisson frequencies." computer methods and programs in biomedicine 83.1 (2006): 57-62. pdf
  42. ^ Babu, G. Jogesh, P. K. Pathak, and C. R. Rao. "Second-order correctness of the Poisson bootstrap." The Annals of Statistics 27.5 (1999): 1666-1683. link
  43. ^ Shoemaker, Owen J., and P. K. Pathak. "The sequential bootstrap: a comparison with regular bootstrap." Communications in Statistics-Theory and Methods 30.8-9 (2001): 1661-1674. link
  44. ^ Jiménez-Gamero, María Dolores, Joaquín Muñoz-García, and Rafael Pino-Mejías. "Reduced bootstrap for the median." Statistica Sinica (2004): 1179-1198. link
  45. ^ Kleiner, A; Talwalkar, A; Sarkar, P; Jordan, M. I. (2014). "A scalable bootstrap for massive data". Journal of the Royal Statistical Society, Series B (Statistical Methodology). 76 (4): 795–816. arXiv:1112.5016. doi:10.1111/rssb.12050. ISSN 1369-7412. S2CID 3064206.
  46. ^ a b Davison, A. C.; Hinkley, D. V. (1997). Bootstrap methods and their application. Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press. ISBN 0-521-57391-2. software.
  47. ^ a b c Hesterberg, Tim C (2014). "What Teachers Should Know about the Bootstrap: Resampling in the Undergraduate Statistics Curriculum". arXiv:1411.5279 [stat.OT].
  48. ^ Efron, B. (1982). The jackknife, the bootstrap, and other resampling plans. Vol. 38. Society of Industrial and Applied Mathematics CBMS-NSF Monographs. ISBN 0-89871-179-7.
  49. ^ Scheiner, S. (1998). Design and Analysis of Ecological Experiments. CRC Press. ISBN 0412035618. Ch13, p300
  50. ^ Rice, John. Mathematical Statistics and Data Analysis (2 ed.). p. 272. "Although this direct equation of quantiles of the bootstrap sampling distribution with confidence limits may seem initially appealing, it’s rationale is somewhat obscure."
  51. ^ Data from examples in Bayesian Data Analysis
  52. ^ Chihara, Laura; Hesterberg, Tim (3 August 2018). Mathematical Statistics with Resampling and R (2nd ed.). John Wiley & Sons, Inc. doi:10.1002/9781119505969. ISBN 9781119416548. S2CID 60138121.
  53. ^ Voinov, Vassily [G.]; Nikulin, Mikhail [S.] (1993). Unbiased estimators and their applications. Vol. 1: Univariate case. Dordrect: Kluwer Academic Publishers. ISBN 0-7923-2382-3.
  54. ^ Young, G. A. (July 1990). "Alternative Smoothed Bootstraps". Journal of the Royal Statistical Society, Series B (Methodological). 52 (3): 477–484. doi:10.1111/j.2517-6161.1990.tb01801.x. ISSN 0035-9246.
  55. ^ Gregory, Karl (29 Dec 2023). "Some results based on the Lindeberg central limit theorem" (PDF). Retrieved 29 Dec 2023.

Further reading edit

  • Diaconis, P.; Efron, B. (May 1983). (PDF). Scientific American. 248 (5): 116–130. Bibcode:1983SciAm.248e.116D. doi:10.1038/scientificamerican0583-116. Archived from the original (PDF) on 2016-03-13. Retrieved 2016-01-19. popular-science
  • Efron, B. (1981). "Nonparametric estimates of standard error: The jackknife, the bootstrap and other methods". Biometrika. 68 (3): 589–599. doi:10.1093/biomet/68.3.589.
  • Hesterberg, T. C.; D. S. Moore; S. Monaghan; A. Clipson & R. Epstein (2005). (PDF). In David S. Moore & George McCabe (eds.). Introduction to the Practice of Statistics. software. Archived from the original (PDF) on 2006-02-15. Retrieved 2007-03-23.
  • Efron, Bradley (1979). "Bootstrap methods: Another look at the jackknife". The Annals of Statistics. 7: 1–26. doi:10.1214/aos/1176344552.
  • Efron, Bradley (1981). "Nonparametric estimates of standard error: The jackknife, the bootstrap and other methods". Biometrika. 68 (3): 589–599. doi:10.2307/2335441. JSTOR 2335441.
  • Efron, Bradley (1982). The jackknife, the bootstrap, and other resampling plans, In Society of Industrial and Applied Mathematics CBMS-NSF Monographs, 38.
  • Diaconis, P.; Efron, Bradley (1983), "Computer-intensive methods in statistics," Scientific American, May, 116–130.
  • Efron, Bradley; Tibshirani, Robert J. (1993). An introduction to the bootstrap, New York: Chapman & Hall, software.
  • Davison, A. C. and Hinkley, D. V. (1997): Bootstrap Methods and their Application, software.
  • Mooney, C Z & Duval, R D (1993). Bootstrapping. A Nonparametric Approach to Statistical Inference. Sage University Paper series on Quantitative Applications in the Social Sciences, 07-095. Newbury Park, CA: Sage.
  • Simon, J. L. (1997): .
  • Wright, D.B., London, K., Field, A.P. Using Bootstrap Estimation and the Plug-in Principle for Clinical Psychology Data. 2011 Textrum Ltd. Online: https://www.researchgate.net/publication/236647074_Using_Bootstrap_Estimation_and_the_Plug-in_Principle_for_Clinical_Psychology_Data. Retrieved on 25/04/2016.
  • An Introduction to the Bootstrap. Monographs on Statistics and applied probability 57. Chapman&Hall/CHC. 1998. Online https://books.google.com/books?id=gLlpIUxRntoC&q=plug+in+principle.&pg=PA35 Retrieved on 25 04 2016.
  • Gail Gong (1986) Cross-Validation, the Jackknife, and the Bootstrap: Excess Error Estimation in Forward Logistic Regression, Journal of the American Statistical Association, 81:393, 108–113, DOI: 10.1080/01621459.1986.10478245

External links edit

  • Bootstrap sampling tutorial using MS Excel
  • bootstrapping tutorial
  • What is the bootstrap?

Software edit

  • Statistics101: Resampling, Bootstrap, Monte Carlo Simulation program. Free program written in Java to run on any operating system.

bootstrapping, statistics, other, uses, bootstrapping, disambiguation, bootstrapping, test, metric, that, uses, random, sampling, with, replacement, mimicking, sampling, process, falls, under, broader, class, resampling, methods, bootstrapping, assigns, measur. For other uses see Bootstrapping disambiguation Bootstrapping is any test or metric that uses random sampling with replacement e g mimicking the sampling process and falls under the broader class of resampling methods Bootstrapping assigns measures of accuracy bias variance confidence intervals prediction error etc to sample estimates 1 2 This technique allows estimation of the sampling distribution of almost any statistic using random sampling methods 3 4 Bootstrapping estimates the properties of an estimand such as its variance by measuring those properties when sampling from an approximating distribution One standard choice for an approximating distribution is the empirical distribution function of the observed data In the case where a set of observations can be assumed to be from an independent and identically distributed population this can be implemented by constructing a number of resamples with replacement of the observed data set and of equal size to the observed data set It may also be used for constructing hypothesis tests 5 It is often used as an alternative to statistical inference based on the assumption of a parametric model when that assumption is in doubt or where parametric inference is impossible or requires complicated formulas for the calculation of standard errors Contents 1 History 2 Approach 3 Discussion 3 1 Advantages 3 2 Disadvantages 3 3 Recommendations 4 Types of bootstrap scheme 4 1 Case resampling 4 1 1 Estimating the distribution of sample mean 4 1 2 Regression 4 2 Bayesian bootstrap 4 3 Smooth bootstrap 4 4 Parametric bootstrap 4 5 Resampling residuals 4 6 Gaussian process regression bootstrap 4 7 Wild bootstrap 4 8 Block bootstrap 4 8 1 Time series Simple block bootstrap 4 8 2 Time series Moving block bootstrap 4 8 3 Time series Maximum entropy bootstrap 4 8 4 Cluster data block bootstrap 5 Methods for improving computational efficiency 5 1 Poisson bootstrap 5 2 Bag of Little Bootstraps 6 Choice of statistic 7 Deriving confidence intervals from the bootstrap distribution 7 1 Bias asymmetry and confidence intervals 7 2 Methods for bootstrap confidence intervals 8 Bootstrap hypothesis testing 9 Example applications 9 1 Smoothed bootstrap 10 Relation to other approaches to inference 10 1 Relationship to other resampling methods 10 2 U statistics 11 Methods for proving consistency of bootstrap estimates 12 See also 13 References 14 Further reading 15 External links 15 1 SoftwareHistory editThe bootstrap was published by Bradley Efron in Bootstrap methods another look at the jackknife 1979 6 7 8 inspired by earlier work on the jackknife 9 10 11 Improved estimates of the variance were developed later 12 13 A Bayesian extension was developed in 1981 14 The bias corrected and accelerated B C a displaystyle BC a nbsp bootstrap was developed by Efron in 1987 15 and the approximate bootstrap confidence interval ABC or approximate B C a displaystyle BC a nbsp procedure in 1992 16 Other names that Efron s colleagues suggested for the bootstrap method were Swiss Army Knife Meat Axe Swan Dive Jack Rabbit and Shotgun 8 Approach edit nbsp A sample is drawn from a population From this sample resamples are generated by drawing with replacement orange Data points that were drawn more than once which happens for approx 26 4 of data points are shown in red and slightly offsetted From the resamples the statistic x displaystyle x nbsp is calculated and therefore a histogram can be calculated to estimate the distribution of x displaystyle x nbsp The basic idea of bootstrapping is that inference about a population from sample data sample population can be modeled by resampling the sample data and performing inference about a sample from resampled data resampled sample 17 As the population is unknown the true error in a sample statistic against its population value is unknown In bootstrap resamples the population is in fact the sample and this is known hence the quality of inference of the true sample from resampled data resampled sample is measurable More formally the bootstrap works by treating inference of the true probability distribution J given the original data as being analogous to an inference of the empirical distribution Ĵ given the resampled data The accuracy of inferences regarding Ĵ using the resampled data can be assessed because we know Ĵ If Ĵ is a reasonable approximation to J then the quality of inference on J can in turn be inferred As an example assume we are interested in the average or mean height of people worldwide We cannot measure all the people in the global population so instead we sample only a tiny part of it and measure that Assume the sample is of size N that is we measure the heights of N individuals From that single sample only one estimate of the mean can be obtained In order to reason about the population we need some sense of the variability of the mean that we have computed The simplest bootstrap method involves taking the original data set of heights and using a computer sampling from it to form a new sample called a resample or bootstrap sample that is also of size N The bootstrap sample is taken from the original by using sampling with replacement e g we might resample 5 times from 1 2 3 4 5 and get 2 5 4 4 1 so assuming N is sufficiently large for all practical purposes there is virtually zero probability that it will be identical to the original real sample This process is repeated a large number of times typically 1 000 or 10 000 times and for each of these bootstrap samples we compute its mean each of these is called a bootstrap estimate We now can create a histogram of bootstrap means This histogram provides an estimate of the shape of the distribution of the sample mean from which we can answer questions about how much the mean varies across samples The method here described for the mean can be applied to almost any other statistic or estimator Discussion editThis section includes a list of references related reading or external links but its sources remain unclear because it lacks inline citations Please help improve this section by introducing more precise citations June 2012 Learn how and when to remove this message Advantages edit A great advantage of bootstrap is its simplicity It is a straightforward way to derive estimates of standard errors and confidence intervals for complex estimators of the distribution such as percentile points proportions Odds ratio and correlation coefficients However despite its simplicity bootstrapping can be applied to complex sampling designs e g for population divided into s strata with ns observations per strata bootstrapping can be applied for each stratum 18 Bootstrap is also an appropriate way to control and check the stability of the results Although for most problems it is impossible to know the true confidence interval bootstrap is asymptotically more accurate than the standard intervals obtained using sample variance and assumptions of normality 19 Bootstrapping is also a convenient method that avoids the cost of repeating the experiment to get other groups of sample data Disadvantages edit Bootstrapping depends heavily on the estimator used and though simple naive use of bootstrapping will not always yield asymptotically valid results and can lead to inconsistency 20 Although bootstrapping is under some conditions asymptotically consistent it does not provide general finite sample guarantees The result may depend on the representative sample The apparent simplicity may conceal the fact that important assumptions are being made when undertaking the bootstrap analysis e g independence of samples or large enough of a sample size where these would be more formally stated in other approaches Also bootstrapping can be time consuming and there are not many available software for bootstrapping as it is difficult to automate using traditional statistical computer packages 18 Recommendations edit Scholars have recommended more bootstrap samples as available computing power has increased If the results may have substantial real world consequences then one should use as many samples as is reasonable given available computing power and time Increasing the number of samples cannot increase the amount of information in the original data it can only reduce the effects of random sampling errors which can arise from a bootstrap procedure itself Moreover there is evidence that numbers of samples greater than 100 lead to negligible improvements in the estimation of standard errors 21 In fact according to the original developer of the bootstrapping method even setting the number of samples at 50 is likely to lead to fairly good standard error estimates 22 Ader et al recommend the bootstrap procedure for the following situations 23 When the theoretical distribution of a statistic of interest is complicated or unknown Since the bootstrapping procedure is distribution independent it provides an indirect method to assess the properties of the distribution underlying the sample and the parameters of interest that are derived from this distribution When the sample size is insufficient for straightforward statistical inference If the underlying distribution is well known bootstrapping provides a way to account for the distortions caused by the specific sample that may not be fully representative of the population When power calculations have to be performed and a small pilot sample is available Most power and sample size calculations are heavily dependent on the standard deviation of the statistic of interest If the estimate used is incorrect the required sample size will also be wrong One method to get an impression of the variation of the statistic is to use a small pilot sample and perform bootstrapping on it to get impression of the variance However Athreya has shown 24 that if one performs a naive bootstrap on the sample mean when the underlying population lacks a finite variance for example a power law distribution then the bootstrap distribution will not converge to the same limit as the sample mean As a result confidence intervals on the basis of a Monte Carlo simulation of the bootstrap could be misleading Athreya states that Unless one is reasonably sure that the underlying distribution is not heavy tailed one should hesitate to use the naive bootstrap Types of bootstrap scheme editThis section includes a list of references related reading or external links but its sources remain unclear because it lacks inline citations Please help improve this section by introducing more precise citations June 2012 Learn how and when to remove this message In univariate problems it is usually acceptable to resample the individual observations with replacement case resampling below unlike subsampling in which resampling is without replacement and is valid under much weaker conditions compared to the bootstrap In small samples a parametric bootstrap approach might be preferred For other problems a smooth bootstrap will likely be preferred For regression problems various other alternatives are available 1 Case resampling edit The bootstrap is generally useful for estimating the distribution of a statistic e g mean variance without using normality assumptions as required e g for a z statistic or a t statistic In particular the bootstrap is useful when there is no analytical form or an asymptotic theory e g an applicable central limit theorem to help estimate the distribution of the statistics of interest This is because bootstrap methods can apply to most random quantities e g the ratio of variance and mean There are at least two ways of performing case resampling The Monte Carlo algorithm for case resampling is quite simple First we resample the data with replacement and the size of the resample must be equal to the size of the original data set Then the statistic of interest is computed from the resample from the first step We repeat this routine many times to get a more precise estimate of the Bootstrap distribution of the statistic 1 The exact version for case resampling is similar but we exhaustively enumerate every possible resample of the data set This can be computationally expensive as there are a total of 2 n 1 n 2 n 1 n n 1 displaystyle binom 2n 1 n frac 2n 1 n n 1 nbsp different resamples where n is the size of the data set Thus for n 5 10 20 30 there are 126 92378 6 89 1010 and 5 91 1016 different resamples respectively 25 Estimating the distribution of sample mean edit Consider a coin flipping experiment We flip the coin and record whether it lands heads or tails Let X x1 x2 x10 be 10 observations from the experiment xi 1 if the i th flip lands heads and 0 otherwise By invoking the assumption that the average of the coin flips is normally distributed we can use the t statistic to estimate the distribution of the sample mean x 1 10 x 1 x 2 x 10 displaystyle bar x frac 1 10 x 1 x 2 cdots x 10 nbsp Such a normality assumption can be justified either as an approximation of the distribution of each individual coin flip or as an approximation of the distribution of the average of a large number of coin flips The former is a poor approximation because the true distribution of the coin flips is Bernoulli instead of normal The latter is a valid approximation in infinitely large samples due to the central limit theorem However if we are not ready to make such a justification then we can use the bootstrap instead Using case resampling we can derive the distribution of x displaystyle bar x nbsp We first resample the data to obtain a bootstrap resample An example of the first resample might look like this X1 x2 x1 x10 x10 x3 x4 x6 x7 x1 x9 There are some duplicates since a bootstrap resample comes from sampling with replacement from the data Also the number of data points in a bootstrap resample is equal to the number of data points in our original observations Then we compute the mean of this resample and obtain the first bootstrap mean m1 We repeat this process to obtain the second resample X2 and compute the second bootstrap mean m2 If we repeat this 100 times then we have m1 m2 m100 This represents an empirical bootstrap distribution of sample mean From this empirical distribution one can derive a bootstrap confidence interval for the purpose of hypothesis testing Regression edit In regression problems case resampling refers to the simple scheme of resampling individual cases often rows of a data set For regression problems as long as the data set is fairly large this simple scheme is often acceptable 26 27 28 However the method is open to criticism citation needed 18 In regression problems the explanatory variables are often fixed or at least observed with more control than the response variable Also the range of the explanatory variables defines the information available from them Therefore to resample cases means that each bootstrap sample will lose some information As such alternative bootstrap procedures should be considered Bayesian bootstrap edit Bootstrapping can be interpreted in a Bayesian framework using a scheme that creates new data sets through reweighting the initial data Given a set of N displaystyle N nbsp data points the weighting assigned to data point i displaystyle i nbsp in a new data set D J displaystyle mathcal D J nbsp is w i J x i J x i 1 J displaystyle w i J x i J x i 1 J nbsp where x J displaystyle mathbf x J nbsp is a low to high ordered list of N 1 displaystyle N 1 nbsp uniformly distributed random numbers on 0 1 displaystyle 0 1 nbsp preceded by 0 and succeeded by 1 The distributions of a parameter inferred from considering many such data sets D J displaystyle mathcal D J nbsp are then interpretable as posterior distributions on that parameter 29 Smooth bootstrap edit Under this scheme a small amount of usually normally distributed zero centered random noise is added onto each resampled observation This is equivalent to sampling from a kernel density estimate of the data Assume K to be a symmetric kernel density function with unit variance The standard kernel estimator f h x displaystyle hat f h x nbsp of f x displaystyle f x nbsp is f h x 1 n h i 1 n K x X i h displaystyle hat f h x 1 over nh sum i 1 n K left x X i over h right nbsp 30 where h displaystyle h nbsp is the smoothing parameter And the corresponding distribution function estimator F h x displaystyle hat F h x nbsp is F h x x f h t d t displaystyle hat F h x int infty x hat f h t dt nbsp 30 Parametric bootstrap edit Based on the assumption that the original data set is a realization of a random sample from a distribution of a specific parametric type in this case a parametric model is fitted by parameter 8 often by maximum likelihood and samples of random numbers are drawn from this fitted model Usually the sample drawn has the same sample size as the original data Then the estimate of original function F can be written as F F 8 displaystyle hat F F hat theta nbsp This sampling process is repeated many times as for other bootstrap methods Considering the centered sample mean in this case the random sample original distribution function F 8 displaystyle F theta nbsp is replaced by a bootstrap random sample with function F 8 displaystyle F hat theta nbsp and the probability distribution of X n m 8 displaystyle bar X n mu theta nbsp is approximated by that of X n m displaystyle bar X n mu nbsp where m m 8 displaystyle mu mu hat theta nbsp which is the expectation corresponding to F 8 displaystyle F hat theta nbsp 31 The use of a parametric model at the sampling stage of the bootstrap methodology leads to procedures which are different from those obtained by applying basic statistical theory to inference for the same model Resampling residuals edit Another approach to bootstrapping in regression problems is to resample residuals The method proceeds as follows Fit the model and retain the fitted values y i displaystyle widehat y i nbsp and the residuals e i y i y i i 1 n displaystyle widehat varepsilon i y i widehat y i i 1 dots n nbsp For each pair xi yi in which xi is the possibly multivariate explanatory variable add a randomly resampled residual e j displaystyle widehat varepsilon j nbsp to the fitted value y i displaystyle widehat y i nbsp In other words create synthetic response variables y i y i e j displaystyle y i widehat y i widehat varepsilon j nbsp where j is selected randomly from the list 1 n for every i Refit the model using the fictitious response variables y i displaystyle y i nbsp and retain the quantities of interest often the parameters m i displaystyle widehat mu i nbsp estimated from the synthetic y i displaystyle y i nbsp Repeat steps 2 and 3 a large number of times This scheme has the advantage that it retains the information in the explanatory variables However a question arises as to which residuals to resample Raw residuals are one option another is studentized residuals in linear regression Although there are arguments in favor of using studentized residuals in practice it often makes little difference and it is easy to compare the results of both schemes Gaussian process regression bootstrap edit When data are temporally correlated straightforward bootstrapping destroys the inherent correlations This method uses Gaussian process regression GPR to fit a probabilistic model from which replicates may then be drawn GPR is a Bayesian non linear regression method A Gaussian process GP is a collection of random variables any finite number of which have a joint Gaussian normal distribution A GP is defined by a mean function and a covariance function which specify the mean vectors and covariance matrices for each finite collection of the random variables 32 Regression model y x f x e e N 0 s 2 displaystyle y x f x varepsilon varepsilon sim mathcal N 0 sigma 2 nbsp e displaystyle varepsilon nbsp is a noise term Gaussian process prior For any finite collection of variables x1 xn the function outputs f x 1 f x n displaystyle f x 1 ldots f x n nbsp are jointly distributed according to a multivariate Gaussian with mean m m x 1 m x n displaystyle m m x 1 ldots m x n intercal nbsp and covariance matrix K i j k x i x j displaystyle K ij k x i x j nbsp Assume f x G P m k displaystyle f x sim mathcal GP m k nbsp Then y x G P m l displaystyle y x sim mathcal GP m l nbsp where l x i x j k x i x j s 2 d x i x j displaystyle l x i x j k x i x j sigma 2 delta x i x j nbsp and d x i x j displaystyle delta x i x j nbsp is the standard Kronecker delta function 32 Gaussian process posterior According to GP prior we can get y x 1 y x r N m 0 K 0 displaystyle y x 1 ldots y x r sim mathcal N m 0 K 0 nbsp where m 0 m x 1 m x r displaystyle m 0 m x 1 ldots m x r intercal nbsp and K 0 i j k x i x j s 2 d x i x j displaystyle K 0 ij k x i x j sigma 2 delta x i x j nbsp Let x1 xs be another finite collection of variables it s obvious that y x 1 y x r f x 1 f x s N m 0 m K 0 K K K displaystyle y x 1 ldots y x r f x 1 ldots f x s intercal sim mathcal N binom m 0 m begin pmatrix K 0 amp K K intercal amp K end pmatrix nbsp where m m x 1 m x s displaystyle m m x 1 ldots m x s intercal nbsp K i j k x i x j displaystyle K ij k x i x j nbsp K i j k x i x j displaystyle K ij k x i x j nbsp According to the equations above the outputs y are also jointly distributed according to a multivariate Gaussian Thus f x 1 f x s y x y N m post K post displaystyle f x 1 ldots f x s intercal mid y x intercal y sim mathcal N m text post K text post nbsp where y y 1 y r displaystyle y y 1 y r intercal nbsp m post m K K O s 2 I r 1 y m 0 displaystyle m text post m K intercal K O sigma 2 I r 1 y m 0 nbsp K post K K K O s 2 I r 1 K displaystyle K text post K K intercal K O sigma 2 I r 1 K nbsp and I r displaystyle I r nbsp is r r displaystyle r times r nbsp identity matrix 32 Wild bootstrap edit The wild bootstrap proposed originally by Wu 1986 33 is suited when the model exhibits heteroskedasticity The idea is as the residual bootstrap to leave the regressors at their sample value but to resample the response variable based on the residuals values That is for each replicate one computes a new y displaystyle y nbsp based on y i y i e i v i displaystyle y i widehat y i widehat varepsilon i v i nbsp so the residuals are randomly multiplied by a random variable v i displaystyle v i nbsp with mean 0 and variance 1 For most distributions of v i displaystyle v i nbsp but not Mammen s this method assumes that the true residual distribution is symmetric and can offer advantages over simple residual sampling for smaller sample sizes Different forms are used for the random variable v i displaystyle v i nbsp such as The standard normal distribution A distribution suggested by Mammen 1993 34 v i 5 1 2 with probability 5 1 2 5 5 1 2 with probability 5 1 2 5 displaystyle v i begin cases sqrt 5 1 2 amp text with probability sqrt 5 1 2 sqrt 5 sqrt 5 1 2 amp text with probability sqrt 5 1 2 sqrt 5 end cases nbsp dd dd Approximately Mammen s distribution is dd v i 0 6180 with a 0 in the units place with probability 0 7236 1 6180 with a 1 in the units place with probability 0 2764 displaystyle v i begin cases 0 6180 quad text with a 0 in the units place amp text with probability 0 7236 1 6180 quad text with a 1 in the units place amp text with probability 0 2764 end cases nbsp dd dd Or the simpler distribution linked to the Rademacher distribution v i 1 with probability 1 2 1 with probability 1 2 displaystyle v i begin cases 1 amp text with probability 1 2 1 amp text with probability 1 2 end cases nbsp dd dd Block bootstrap edit The block bootstrap is used when the data or the errors in a model are correlated In this case a simple case or residual resampling will fail as it is not able to replicate the correlation in the data The block bootstrap tries to replicate the correlation by resampling inside blocks of data see Blocking statistics The block bootstrap has been used mainly with data correlated in time i e time series but can also be used with data correlated in space or among groups so called cluster data Time series Simple block bootstrap edit In the simple block bootstrap the variable of interest is split into non overlapping blocks Time series Moving block bootstrap edit In the moving block bootstrap introduced by Kunsch 1989 35 data is split into n b 1 overlapping blocks of length b Observation 1 to b will be block 1 observation 2 to b 1 will be block 2 etc Then from these n b 1 blocks n b blocks will be drawn at random with replacement Then aligning these n b blocks in the order they were picked will give the bootstrap observations This bootstrap works with dependent data however the bootstrapped observations will not be stationary anymore by construction But it was shown that varying randomly the block length can avoid this problem 36 This method is known as the stationary bootstrap Other related modifications of the moving block bootstrap are the Markovian bootstrap and a stationary bootstrap method that matches subsequent blocks based on standard deviation matching Time series Maximum entropy bootstrap edit Vinod 2006 37 presents a method that bootstraps time series data using maximum entropy principles satisfying the Ergodic theorem with mean preserving and mass preserving constraints There is an R package meboot 38 that utilizes the method which has applications in econometrics and computer science Cluster data block bootstrap edit Cluster data describes data where many observations per unit are observed This could be observing many firms in many states or observing students in many classes In such cases the correlation structure is simplified and one does usually make the assumption that data is correlated within a group cluster but independent between groups clusters The structure of the block bootstrap is easily obtained where the block just corresponds to the group and usually only the groups are resampled while the observations within the groups are left unchanged Cameron et al 2008 discusses this for clustered errors in linear regression 39 Methods for improving computational efficiency editThe bootstrap is a powerful technique although may require substantial computing resources in both time and memory Some techniques have been developed to reduce this burden They can generally be combined with many of the different types of Bootstrap schemes and various choices of statistics Poisson bootstrap edit nbsp Graph showing the convergence of the Binomial Distribution to the Poisson as the Binomial parameters are n p 1 and n grows The ordinary bootstrap requires the random selection of n elements from a list which is equivalent to drawing from a binomial distribution This may require a large number of passes over the data and is challenging to run these computations in parallel For large values of n the Poisson bootstrap is an efficient method of generating bootstrapped data sets 40 When generating a single bootstrap sample instead of randomly drawing from the sample data with replacement each data point is assigned a random weight distributed according to the Poisson distribution with l 1 displaystyle lambda 1 nbsp For large sample data this will approximate random sampling with replacement This is due to the following approximation lim n Binomial n 1 n Poisson 1 displaystyle lim n to infty operatorname Binomial n 1 n operatorname Poisson 1 nbsp This method also lends itself well to streaming data and growing data sets since the total number of samples does not need to be known in advance of beginning to take bootstrap samples For large enough n the results are relatively similar to the original bootstrap estimations 41 A way to improve on the poisson bootstrap termed sequential bootstrap is by taking the first samples so that the proportion of unique values is 0 632 of the original sample size n This provides a distribution with main empirical characteristics being within a distance of O n 3 4 displaystyle O n 3 4 nbsp 42 Empirical investigation has shown this method can yield good results 43 This is related to the reduced bootstrap method 44 Bag of Little Bootstraps edit For massive data sets it is often computationally prohibitive to hold all the sample data in memory and resample from the sample data The Bag of Little Bootstraps BLB 45 provides a method of pre aggregating data before bootstrapping to reduce computational constraints This works by partitioning the data set into b displaystyle b nbsp equal sized buckets and aggregating the data within each bucket This pre aggregated data set becomes the new sample data over which to draw samples with replacement This method is similar to the Block Bootstrap but the motivations and definitions of the blocks are very different Under certain assumptions the sample distribution should approximate the full bootstrapped scenario One constraint is the number of buckets b n g displaystyle b n gamma nbsp where g 0 5 1 displaystyle gamma in 0 5 1 nbsp and the authors recommend usage of b n 0 7 displaystyle b n 0 7 nbsp as a general solution Choice of statistic editThe bootstrap distribution of a point estimator of a population parameter has been used to produce a bootstrapped confidence interval for the parameter s true value if the parameter can be written as a function of the population s distribution Population parameters are estimated with many point estimators Popular families of point estimators include mean unbiased minimum variance estimators median unbiased estimators Bayesian estimators for example the posterior distribution s mode median mean and maximum likelihood estimators A Bayesian point estimator and a maximum likelihood estimator have good performance when the sample size is infinite according to asymptotic theory For practical problems with finite samples other estimators may be preferable Asymptotic theory suggests techniques that often improve the performance of bootstrapped estimators the bootstrapping of a maximum likelihood estimator may often be improved using transformations related to pivotal quantities 46 Deriving confidence intervals from the bootstrap distribution editThe bootstrap distribution of a parameter estimator has been used to calculate confidence intervals for its population parameter 1 Bias asymmetry and confidence intervals edit Bias The bootstrap distribution and the sample may disagree systematically in which case bias may occur If the bootstrap distribution of an estimator is symmetric then percentile confidence interval are often used such intervals are appropriate especially for median unbiased estimators of minimum risk with respect to an absolute loss function Bias in the bootstrap distribution will lead to bias in the confidence interval Otherwise if the bootstrap distribution is non symmetric then percentile confidence intervals are often inappropriate Methods for bootstrap confidence intervals edit There are several methods for constructing confidence intervals from the bootstrap distribution of a real parameter Basic bootstrap 46 also known as the Reverse Percentile Interval 47 The basic bootstrap is a simple scheme to construct the confidence interval one simply takes the empirical quantiles from the bootstrap distribution of the parameter see Davison and Hinkley 1997 equ 5 6 p 194 2 8 8 1 a 2 2 8 8 a 2 displaystyle 2 widehat theta theta 1 alpha 2 2 widehat theta theta alpha 2 nbsp where 8 1 a 2 displaystyle theta 1 alpha 2 nbsp denotes the 1 a 2 displaystyle 1 alpha 2 nbsp percentile of the bootstrapped coefficients 8 displaystyle theta nbsp dd Percentile bootstrap The percentile bootstrap proceeds in a similar way to the basic bootstrap using percentiles of the bootstrap distribution but with a different formula note the inversion of the left and right quantiles 8 a 2 8 1 a 2 displaystyle theta alpha 2 theta 1 alpha 2 nbsp where 8 1 a 2 displaystyle theta 1 alpha 2 nbsp denotes the 1 a 2 displaystyle 1 alpha 2 nbsp percentile of the bootstrapped coefficients 8 displaystyle theta nbsp dd See Davison and Hinkley 1997 equ 5 18 p 203 and Efron and Tibshirani 1993 equ 13 5 p 171 This method can be applied to any statistic It will work well in cases where the bootstrap distribution is symmetrical and centered on the observed statistic 48 and where the sample statistic is median unbiased and has maximum concentration or minimum risk with respect to an absolute value loss function When working with small sample sizes i e less than 50 the basic reversed percentile and percentile confidence intervals for for example the variance statistic will be too narrow So that with a sample of 20 points 90 confidence interval will include the true variance only 78 of the time 49 The basic reverse percentile confidence intervals are easier to justify mathematically 50 47 but they are less accurate in general than percentile confidence intervals and some authors discourage their use 47 Studentized bootstrap The studentized bootstrap also called bootstrap t is computed analogously to the standard confidence interval but replaces the quantiles from the normal or student approximation by the quantiles from the bootstrap distribution of the Student s t test see Davison and Hinkley 1997 equ 5 7 p 194 and Efron and Tibshirani 1993 equ 12 22 p 160 8 t 1 a 2 se 8 8 t a 2 se 8 displaystyle widehat theta t 1 alpha 2 cdot widehat text se theta widehat theta t alpha 2 cdot widehat text se theta nbsp where t 1 a 2 displaystyle t 1 alpha 2 nbsp denotes the 1 a 2 displaystyle 1 alpha 2 nbsp percentile of the bootstrapped Student s t test t 8 8 se 8 displaystyle t widehat theta widehat theta widehat text se widehat theta nbsp and se 8 displaystyle widehat text se theta nbsp is the estimated standard error of the coefficient in the original model dd The studentized test enjoys optimal properties as the statistic that is bootstrapped is pivotal i e it does not depend on nuisance parameters as the t test follows asymptotically a N 0 1 distribution unlike the percentile bootstrap Bias corrected bootstrap adjusts for bias in the bootstrap distribution Accelerated bootstrap The bias corrected and accelerated BCa bootstrap by Efron 1987 15 adjusts for both bias and skewness in the bootstrap distribution This approach is accurate in a wide variety of settings has reasonable computation requirements and produces reasonably narrow intervals 15 Bootstrap hypothesis testing editThis article may require cleanup to meet Wikipedia s quality standards The specific problem is there are other bootstrap tests Please help improve this article if you can July 2023 Learn how and when to remove this message Efron and Tibshirani 1 suggest the following algorithm for comparing the means of two independent samples Let x 1 x n displaystyle x 1 ldots x n nbsp be a random sample from distribution F with sample mean x displaystyle bar x nbsp and sample variance s x 2 displaystyle sigma x 2 nbsp Let y 1 y m displaystyle y 1 ldots y m nbsp be another independent random sample from distribution G with mean y displaystyle bar y nbsp and variance s y 2 displaystyle sigma y 2 nbsp Calculate the test statistic t x y s x 2 n s y 2 m displaystyle t frac bar x bar y sqrt sigma x 2 n sigma y 2 m nbsp Create two new data sets whose values are x i x i x z displaystyle x i x i bar x bar z nbsp and y i y i y z displaystyle y i y i bar y bar z nbsp where z displaystyle bar z nbsp is the mean of the combined sample Draw a random sample x i displaystyle x i nbsp of size n displaystyle n nbsp with replacement from x i displaystyle x i nbsp and another random sample y i displaystyle y i nbsp of size m displaystyle m nbsp with replacement from y i displaystyle y i nbsp Calculate the test statistic t x y s x 2 n s y 2 m displaystyle t frac bar x bar y sqrt sigma x 2 n sigma y 2 m nbsp Repeat 3 and 4 B displaystyle B nbsp times e g B 1000 displaystyle B 1000 nbsp to collect B displaystyle B nbsp values of the test statistic Estimate the p value as p i 1 B I t i t B displaystyle p frac sum i 1 B I t i geq t B nbsp where I condition 1 displaystyle I text condition 1 nbsp when condition is true and 0 otherwise Example applications editThis section includes a list of references related reading or external links but its sources remain unclear because it lacks inline citations Please help improve this section by introducing more precise citations June 2012 Learn how and when to remove this message Smoothed bootstrap edit In 1878 Simon Newcomb took observations on the speed of light 51 The data set contains two outliers which greatly influence the sample mean The sample mean need not be a consistent estimator for any population mean because no mean needs to exist for a heavy tailed distribution A well defined and robust statistic for the central tendency is the sample median which is consistent and median unbiased for the population median The bootstrap distribution for Newcomb s data appears below We can reduce the discreteness of the bootstrap distribution by adding a small amount of random noise to each bootstrap sample A conventional choice is to add noise with a standard deviation of s n displaystyle sigma sqrt n nbsp for a sample size n this noise is often drawn from a Student t distribution with n 1 degrees of freedom 52 This results in an approximately unbiased estimator for the variance of the sample mean 53 This means that samples taken from the bootstrap distribution will have a variance which is on average equal to the variance of the total population Histograms of the bootstrap distribution and the smooth bootstrap distribution appear below The bootstrap distribution of the sample median has only a small number of values The smoothed bootstrap distribution has a richer support However note that whether the smoothed or standard bootstrap procedure is favorable is case by case and is shown to depend on both the underlying distribution function and on the quantity being estimated 54 nbsp In this example the bootstrapped 95 percentile confidence interval for the population median is 26 28 5 which is close to the interval for 25 98 28 46 for the smoothed bootstrap Relation to other approaches to inference editRelationship to other resampling methods edit The bootstrap is distinguished from the jackknife procedure used to estimate biases of sample statistics and to estimate variances and cross validation in which the parameters e g regression weights factor loadings that are estimated in one subsample are applied to another subsample For more details see resampling Bootstrap aggregating bagging is a meta algorithm based on averaging model predictions obtained from models trained on multiple bootstrap samples U statistics edit Main article U statistic In situations where an obvious statistic can be devised to measure a required characteristic using only a small number r of data items a corresponding statistic based on the entire sample can be formulated Given an r sample statistic one can create an n sample statistic by something similar to bootstrapping taking the average of the statistic over all subsamples of size r This procedure is known to have certain good properties and the result is a U statistic The sample mean and sample variance are of this form for r 1 and r 2 Methods for proving consistency of bootstrap estimates editIt is possible to use the central limit theorem to show the consistency of the bootstrap procedure for estimating the distribution of the sample mean Specifically let us consider X n 1 X n n displaystyle X n1 ldots X nn nbsp independent identically distributed random variables with E X n 1 m displaystyle mathbb E X n1 mu nbsp and Var X n 1 s 2 lt displaystyle text Var X n1 sigma 2 lt infty nbsp for each n 1 displaystyle n geq 1 nbsp Let X n n 1 X n 1 X n n displaystyle bar X n n 1 X n1 cdots X nn nbsp In addition for each n 1 displaystyle n geq 1 nbsp conditional on X n 1 X n n displaystyle X n1 ldots X nn nbsp let X n 1 X n n displaystyle X n1 ldots X nn nbsp be independent random variables with distribution equal to the empirical distribution of X n 1 X n n displaystyle X n1 ldots X nn nbsp This is the sequence of bootstrap samples Then it can be shown thatsup x R P n X n X n s n x P n X n m s x 0 in probability as n displaystyle sup x in mathbb R left P left frac sqrt n bar X n bar X n hat sigma n leq x right P left frac sqrt n bar X n mu sigma leq x right right to 0 text in probability as n to infty nbsp where P displaystyle P nbsp represents probability conditional on X n 1 X n n displaystyle X n1 ldots X nn nbsp n 1 displaystyle n geq 1 nbsp X n n 1 X n 1 X n n displaystyle bar X n n 1 X n1 cdots X nn nbsp and s n 2 n 1 i 1 n X n i X n 2 displaystyle hat sigma n 2 n 1 sum i 1 n X ni bar X n 2 nbsp To see this note that X n i X n n s n displaystyle X ni bar X n sqrt n hat sigma n nbsp satisfies the Lindeberg condition so the CLT holds 55 See also editAccuracy and precision Bootstrap aggregating Bootstrapping Empirical likelihood Imputation statistics Reliability statistics Reproducibility ResamplingReferences edit a b c d e Efron B Tibshirani R 1993 An Introduction to the Bootstrap Boca Raton FL Chapman amp Hall CRC ISBN 0 412 04231 2 software Archived 2012 07 12 at archive today Second Thoughts on the Bootstrap Bradley Efron 2003 Varian H 2005 Bootstrap Tutorial Mathematica Journal 9 768 775 Weisstein Eric W Bootstrap Methods From MathWorld A Wolfram Web Resource http mathworld wolfram com BootstrapMethods html Lehmann E L 1992 Introduction to Neyman and Pearson 1933 On the Problem of the Most Efficient Tests of Statistical Hypotheses In Breakthroughs in Statistics Volume 1 Eds Kotz S Johnson N L Springer Verlag ISBN 0 387 94037 5 followed by reprinting of the paper Notes for Earliest Known Uses of Some of the Words of Mathematics Bootstrap John Aldrich Earliest Known Uses of Some of the Words of Mathematics B Jeff Miller a b Efron B 1979 Bootstrap methods Another look at the jackknife The Annals of Statistics 7 1 1 26 doi 10 1214 aos 1176344552 Quenouille M 1949 Approximate tests of correlation in time series J Roy Statist Soc Ser B 11 68 84 Tukey J 1958 Bias and confidence in not quite large samples abstract Ann Math Statist 29 614 Jaeckel L 1972 The infinitesimal jackknife Memorandum MM72 1215 11 Bell Lab Bickel P Freeman D 1981 Some asymptotic theory for the bootstrap Ann Statist 9 1196 1217 Singh K 1981 On the asymptotic accuracy of Efron s bootstrap Ann Statist 9 1187 1195 Rubin D 1981 The Bayesian bootstrap Ann Statist 9 130 134 a b c Efron B 1987 Better Bootstrap Confidence Intervals Journal of the American Statistical Association 82 397 Journal of the American Statistical Association Vol 82 No 397 171 185 doi 10 2307 2289144 JSTOR 2289144 Dicicio Thomas Efron Bradley 1992 06 01 More accurate confidence intervals in exponential families Biometrika pp 231 245 doi 10 2307 2336835 ISSN 0006 3444 OCLC 5545447518 Retrieved 2024 01 31 Good P 2006 Resampling Methods 3rd Ed Birkhauser a b c 21 Bootstrapping Regression Models PDF Archived PDF from the original on 2015 07 24 DiCiccio TJ Efron B 1996 Bootstrap confidence intervals with Discussion Statistical Science 11 189 228 Hinkley David 1994 08 01 Bootstrap More than a Stab in the Dark Comment Statistical Science 9 3 doi 10 1214 ss 1177010387 ISSN 0883 4237 Goodhue D L Lewis W amp Thompson R 2012 Does PLS have advantages for small sample size or non normal data MIS Quarterly 36 3 981 1001 Efron B Rogosa D amp Tibshirani R 2004 Resampling methods of estimation In N J Smelser amp P B Baltes Eds International Encyclopedia of the Social amp Behavioral Sciences pp 13216 13220 New York NY Elsevier Ader H J Mellenbergh G J amp Hand D J 2008 Advising on research methods A consultant s companion Huizen The Netherlands Johannes van Kessel Publishing ISBN 978 90 79418 01 5 Bootstrap of the mean in the infinite variance case Athreya K B Ann Stats vol 15 2 1987 724 731 How many different bootstrap samples are there Statweb stanford edu Archived from the original on 2019 09 14 Retrieved 2019 12 09 Jenkins David G Quintana Ascencio Pedro F 2020 02 21 A solution to minimum sample size for regressions PLOS ONE 15 2 e0229345 Bibcode 2020PLoSO 1529345J doi 10 1371 journal pone 0229345 ISSN 1932 6203 PMC 7034864 PMID 32084211 Lumley Thomas 2002 The Importance of the Normality Assumption in Large Public Health Data Sets Annual Review of Public Health 23 151 169 doi 10 1146 annurev publhealth 23 100901 140546 PMID 11910059 Li Xiang Wong Wanling Lamoureux Ecosse L Wong Tien Y 2012 05 01 Are Linear Regression Techniques Appropriate for Analysis When the Dependent Outcome Variable Is Not Normally Distributed Investigative Ophthalmology amp Visual Science 53 6 3082 3083 doi 10 1167 iovs 12 9967 ISSN 1552 5783 PMID 22618757 Rubin D B 1981 The Bayesian bootstrap Annals of Statistics 9 130 a b WANG SUOJIN 1995 Optimizing the smoothed bootstrap Ann Inst Statist Math 47 65 80 doi 10 1007 BF00773412 S2CID 122041565 A modern introduction to probability and statistics understanding why and how Dekking Michel 1946 London Springer 2005 ISBN 978 1 85233 896 1 OCLC 262680588 a href Template Cite book html title Template Cite book cite book a CS1 maint others link a b c Kirk Paul 2009 Gaussian process regression bootstrapping exploring the effects of uncertainty in time course data Bioinformatics 25 10 1300 1306 doi 10 1093 bioinformatics btp139 PMC 2677737 PMID 19289448 Wu C F J 1986 Jackknife bootstrap and other resampling methods in regression analysis with discussions PDF Annals of Statistics 14 1261 1350 doi 10 1214 aos 1176350142 Mammen E Mar 1993 Bootstrap and wild bootstrap for high dimensional linear models Annals of Statistics 21 1 255 285 doi 10 1214 aos 1176349025 Kunsch H R 1989 The Jackknife and the Bootstrap for General Stationary Observations Annals of Statistics 17 3 1217 1241 doi 10 1214 aos 1176347265 Politis D N Romano J P 1994 The Stationary Bootstrap Journal of the American Statistical Association 89 428 1303 1313 doi 10 1080 01621459 1994 10476870 hdl 10983 25607 Vinod HD 2006 Maximum entropy ensembles for time series inference in economics Journal of Asian Economics 17 6 955 978 doi 10 1016 j asieco 2006 09 001 Vinod Hrishikesh Lopez de Lacalle Javier 2009 Maximum entropy bootstrap for time series The meboot R package Journal of Statistical Software 29 5 1 19 doi 10 18637 jss v029 i05 Cameron A C Gelbach J B Miller D L 2008 Bootstrap based improvements for inference with clustered errors PDF Review of Economics and Statistics 90 3 414 427 doi 10 1162 rest 90 3 414 Chamandy N Muralidharan O Najmi A Naidu S 2012 Estimating Uncertainty for Massive Data Streams Hanley James A and Brenda MacGibbon Creating non parametric bootstrap samples using Poisson frequencies computer methods and programs in biomedicine 83 1 2006 57 62 pdf Babu G Jogesh P K Pathak and C R Rao Second order correctness of the Poisson bootstrap The Annals of Statistics 27 5 1999 1666 1683 link Shoemaker Owen J and P K Pathak The sequential bootstrap a comparison with regular bootstrap Communications in Statistics Theory and Methods 30 8 9 2001 1661 1674 link Jimenez Gamero Maria Dolores Joaquin Munoz Garcia and Rafael Pino Mejias Reduced bootstrap for the median Statistica Sinica 2004 1179 1198 link Kleiner A Talwalkar A Sarkar P Jordan M I 2014 A scalable bootstrap for massive data Journal of the Royal Statistical Society Series B Statistical Methodology 76 4 795 816 arXiv 1112 5016 doi 10 1111 rssb 12050 ISSN 1369 7412 S2CID 3064206 a b Davison A C Hinkley D V 1997 Bootstrap methods and their application Cambridge Series in Statistical and Probabilistic Mathematics Cambridge University Press ISBN 0 521 57391 2 software a b c Hesterberg Tim C 2014 What Teachers Should Know about the Bootstrap Resampling in the Undergraduate Statistics Curriculum arXiv 1411 5279 stat OT Efron B 1982 The jackknife the bootstrap and other resampling plans Vol 38 Society of Industrial and Applied Mathematics CBMS NSF Monographs ISBN 0 89871 179 7 Scheiner S 1998 Design and Analysis of Ecological Experiments CRC Press ISBN 0412035618 Ch13 p300 Rice John Mathematical Statistics and Data Analysis 2 ed p 272 Although this direct equation of quantiles of the bootstrap sampling distribution with confidence limits may seem initially appealing it s rationale is somewhat obscure Data from examples in Bayesian Data Analysis Chihara Laura Hesterberg Tim 3 August 2018 Mathematical Statistics with Resampling and R 2nd ed John Wiley amp Sons Inc doi 10 1002 9781119505969 ISBN 9781119416548 S2CID 60138121 Voinov Vassily G Nikulin Mikhail S 1993 Unbiased estimators and their applications Vol 1 Univariate case Dordrect Kluwer Academic Publishers ISBN 0 7923 2382 3 Young G A July 1990 Alternative Smoothed Bootstraps Journal of the Royal Statistical Society Series B Methodological 52 3 477 484 doi 10 1111 j 2517 6161 1990 tb01801 x ISSN 0035 9246 Gregory Karl 29 Dec 2023 Some results based on the Lindeberg central limit theorem PDF Retrieved 29 Dec 2023 Further reading editDiaconis P Efron B May 1983 Computer intensive methods in statistics PDF Scientific American 248 5 116 130 Bibcode 1983SciAm 248e 116D doi 10 1038 scientificamerican0583 116 Archived from the original PDF on 2016 03 13 Retrieved 2016 01 19 popular science Efron B 1981 Nonparametric estimates of standard error The jackknife the bootstrap and other methods Biometrika 68 3 589 599 doi 10 1093 biomet 68 3 589 Hesterberg T C D S Moore S Monaghan A Clipson amp R Epstein 2005 Bootstrap methods and permutation tests PDF In David S Moore amp George McCabe eds Introduction to the Practice of Statistics software Archived from the original PDF on 2006 02 15 Retrieved 2007 03 23 Efron Bradley 1979 Bootstrap methods Another look at the jackknife The Annals of Statistics 7 1 26 doi 10 1214 aos 1176344552 Efron Bradley 1981 Nonparametric estimates of standard error The jackknife the bootstrap and other methods Biometrika 68 3 589 599 doi 10 2307 2335441 JSTOR 2335441 Efron Bradley 1982 The jackknife the bootstrap and other resampling plans In Society of Industrial and Applied Mathematics CBMS NSF Monographs 38 Diaconis P Efron Bradley 1983 Computer intensive methods in statistics Scientific American May 116 130 Efron Bradley Tibshirani Robert J 1993 An introduction to the bootstrap New York Chapman amp Hall software Davison A C and Hinkley D V 1997 Bootstrap Methods and their Application software Mooney C Z amp Duval R D 1993 Bootstrapping A Nonparametric Approach to Statistical Inference Sage University Paper series on Quantitative Applications in the Social Sciences 07 095 Newbury Park CA Sage Simon J L 1997 Resampling The New Statistics Wright D B London K Field A P Using Bootstrap Estimation and the Plug in Principle for Clinical Psychology Data 2011 Textrum Ltd Online https www researchgate net publication 236647074 Using Bootstrap Estimation and the Plug in Principle for Clinical Psychology Data Retrieved on 25 04 2016 An Introduction to the Bootstrap Monographs on Statistics and applied probability 57 Chapman amp Hall CHC 1998 Online https books google com books id gLlpIUxRntoC amp q plug in principle amp pg PA35 Retrieved on 25 04 2016 Gail Gong 1986 Cross Validation the Jackknife and the Bootstrap Excess Error Estimation in Forward Logistic Regression Journal of the American Statistical Association 81 393 108 113 DOI 10 1080 01621459 1986 10478245External links editBootstrap sampling tutorial using MS Excel Bootstrap example to simulate stock prices using MS Excel bootstrapping tutorial What is the bootstrap Software edit Statistics101 Resampling Bootstrap Monte Carlo Simulation program Free program written in Java to run on any operating system Retrieved from https en wikipedia org w index php title Bootstrapping statistics amp oldid 1208968780, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.