Abstrakt: |
Let X, X, X, ... be a sequence of i.i.d. random variables with mean µ = EX. Let { v, ..., v} be vectors of nonnegative random variables (weights), independent of the data sequence { X, ..., X}, and put m = Σ v. Consider v X, ..., v X, a bootstrap sample, resulting from re-sampling or stochastically re-weighing the random sample X, ..., X, n ≥ 1. Put $$\bar X_n = \sum\nolimits_{i = 1}^n {X_i } /n$$, the original sample mean, and define $$\bar X_{m_n }^* = \sum\nolimits_{i = 1}^n {v_i^{(n)} X_i /m_n }$$, where m:= Σ v, the bootstrap sample mean. Thus, $$\bar X_{m_n }^* - \bar X_n = \sum\nolimits_{i = 1}^n {\left( {v_i^{(n)} /m_n - 1/n} \right)X_i }$$. Put V = Σ( v/ m − 1/ n) and let S, $$S_{m_n }^{*2}$$ respectively be the original sample variance and the bootstrap sample variance. The main aim of this exposition is to study the asymptotic behavior of the bootstrapped t-statistics $$T_{m_n }^* : = (\bar X_{m_n }^* - \bar X_n )/(S_n V_n )$$ and $$T_{m_n }^{**} : = \sqrt {m_n } (\bar X_{m_n }^* - \bar X_n )/S_{m_n }^*$$ in terms of conditioning on the weights via assuming that, as n → ∞, max( v/ m − 1/ n)/ V = o(1) almost surely or in probability on the probability space of the weights. In consequence of these maximum negligibility conditions on the weights, a characterization of the validity of this approach to the bootstrap is obtained as a direct consequence of the Lindeberg-Feller central limit theorem (CLT). This view of justifying the validity of bootstrapping i.i.d. observables is believed to be new. The need for it arises naturally in practice when exploring the nature of information contained in a random sample via re-sampling, for example. Conditioning on the data is also revisited for Efron's bootstrap weights under conditions on n, m as n → ∞ that differ from requiring m/ n to be in the interval [ λ, λ] with 0 < λ < λ < ∞ as in Mason and Shao (2001). The validity of the bootstrapped t-intervals is established for both approaches to conditioning. [ABSTRACT FROM AUTHOR] |