Abstract
The multiplier bootstrap is a fast and easy-to-implement alternative to the standard bootstrap; it has been used successfully in many statistical contexts. In this paper, resampling methods based on multipliers are proposed in a general framework where one investigates the stochastic behavior of a random vector \(\mathbf {Y}\in \mathbb {R}^d\) conditional on a covariate \(X \in \mathbb {R}\). Specifically, two versions of the multiplier bootstrap adapted to empirical conditional distributions are introduced as alternatives to the conditional bootstrap and their asymptotic validity is formally established. As the method walks hand-in-hand with the functional delta method, theory around the estimation of statistical functionals is developed accordingly; this includes the interval estimation of conditional mean and variance, conditional correlation coefficient, Kendall’s dependence measure and copula. Composite inference about univariate and joint conditional distributions is also considered. The sample behavior of the new bootstrap schemes and related estimation methods are investigated via simulations and an illustration on real data is provided.
Similar content being viewed by others
References
Aerts, M., Janssen, P., Veraverbeke, N.: Bootstrapping regression quantiles. J. Nonparametr. Stat. 4(1), 1–20 (1994). doi:10.1080/10485259408832597
Benedetti, J.K.: On the nonparametric estimation of regression functions. J. R. Stat. Soc. Ser B (Methodol.) 39(2), 248–253 (1977). URL http://www.jstor.org/stable/2984802
Bücher, A., Volgushev, S.: Empirical and sequential empirical copula processes under serial dependence. J. Multivar. Anal. 119, 61–70 (2013). doi:10.1016/j.jmva.2013.04.003
Calonico, S., Cattaneo, M.D., Titiunik, R.: Robust nonparametric confidence intervals for regression-discontinuity designs. Econometrica 82(6), 2295–2326 (2014). doi:10.3982/ECTA11757
Gasser, T., Müller, H.G.: Kernel estimation of regression functions. In: Smoothing Techniques for Curve Estimation (Proceedings of a Workshop, Heidelberg, 1979). Lecture Notes in Mathematics, vol. 757, pp. 23–68. Springer, Berlin (1979)
Gijbels, I., Veraverbeke, N., Omelka, M.: Conditional copulas, association measures and their applications. Comput. Stat. Data Anal. 55(5), 1919–1932 (2011). doi:10.1016/j.csda.2010.11.010
Hall, P.: Efficient Bootstrap Simulation. Exploring the Limits of Bootstrap. Wiley, New York (1992)
Härdle, W., Bowman, A.W.: Bootstrapping in nonparametric regression: local adaptive smoothing and confidence bands. J. Am. Stat. Assoc. 83(401), 102–110 (1988). URL http://links.jstor.org/sici?sici=0162-1459(198803)83:401<102:BINRLA>2.0.CO;2-T&origin=MSN
Hardle, W., Marron, J.S.: Bootstrap simultaneous error bars for nonparametric regression. Ann. Stat. 19(2), 778–796 (1991). doi:10.1214/aos/1176348120
Hoeffding, W.: The Collected Works of Wassily Hoeffding. Springer Series in Statistics: Perspectives in Statistics. In: Fisher, N.I., Sen, P.K. (eds.) Springer, New York (1994)
Hušková, M., Meintanis, S.G.: Goodness-of-fit tests for parametric regression models based on empirical characteristic functions. Kybernetika 45(6), 960–971 (2009)
Kosorok, M.: Introduction to Empirical Processes and Semiparametric Inference. Springer, New York (2008)
Lehmann, E.L.: Some concepts of dependence. Ann. Math. Stat. 37, 1137–1153 (1966)
McMurry, T.L., Politis, D.N.: Bootstrap confidence intervals in nonparametric regression with built-in bias correction. Stat. Probab. Lett. 78(15), 2463–2469 (2008). doi:10.1016/j.spl.2008.02.032. URL http://www.sciencedirect.com/science/article/pii/S0167715208001508
Mesfioui, M., Quessy, J.F.: Dependence structure of conditional Archimedean copulas. J. Multivar. Anal. 99(3), 372–385 (2008). doi:10.1016/j.jmva.2006.10.007
Nelsen, R.B.: An Introduction to Copulas. Springer Series in Statistics, 2nd edn. Springer, New York (2006)
Omelka, M., Veraverbeke, N., Gijbels, I.: Bootstrapping the conditional copula. J. Stat. Plan. Inference 143(1), 1–23 (2013). doi:10.1016/j.jspi.2012.06.001
Priestley, M.B., Chao, M.T.: Non-parametric function fitting. J. R. Stat. Soc. Ser. B 34, 385–392 (1972)
van der Vaart, A.W., Wellner, J.A.: Weak Convergence and Empirical Processes: With Applications to Statistics. Springer Series in Statistics. Springer, New York (1996)
Veraverbeke, N., Omelka, M., Gijbels, I.: Estimation of a conditional copula and association measures. Scand. J. Stat. 38(4), 766–780 (2011). doi:10.1111/j.1467-9469.2011.00744.x
Watson, G.S.: Smooth regression analysis. Sankhyā Ser. A 26, 359–372 (1964)
Author information
Authors and Affiliations
Corresponding author
Electronic supplementary material
Below is the link to the electronic supplementary material.
Appendices
Appendix 1: Assumptions on the weights
The following notation will be adopted in the sequel. For a sequence of random variables \((Z_n)_{n\in \mathbb {N}}\) and a sequence of real numbers \((a_n)_{n\in \mathbb {N}}\), one writes \(Z_n = O_e(a_n)\) if there exist finite constants \(C_1, C_2\) and \(\eta > 0\) such that \(\mathrm{P}( Z_n/a_n \ge C_1 ) \le C_2 \exp ( - n^\eta / C_1 )\). Also, let \(J_{nx}= [ \, \min _{i \in I_{nx}} \, x_i ,\max _{i \in I_{nx}} \, x_i \, ]\), where \(I_{nx} = \{ j \in \{ 1, \ldots , n \} : w_{nj}(x,h_n) > 0 \}\). The assumptions on the weights listed below are for fixed designs; the adaptation to random designs generally consists in replacing o by \(o_{{\text {a.s.}}}\) and O by \(O_{\text {a.s.}}\), except for \(W_4\), \(W_6\), \(W'_2\) and \(W''_2\), where O has to be replaced by \(O_e\). This stronger requirement ensures the validity of Lemma 3 and Lemma 4 in Omelka et al. (2013) in the context of a random design. These lemmas will prove useful in the proofs presented in the online resource.
Assumptions \(W_1\)–\(W_5\) below are needed to establish the asymptotic behavior of \(\mathbb {F}_{xh}\) stated in Sect. 2.1; they are also needed, as well as \(W_6\), in the proof of Proposition 1.
- \(W_1\).:
-
\(\displaystyle \sqrt{n h_n} \max _{1 \le i \le n} \left| w_{ni}(x,h_n) \right| = o(1)\);
- \(W_2\).:
-
\(\displaystyle \sqrt{n h_n} \left| \sum _{i=1}^n w_{ni}(x,h_n)(x_i-x) - h_n^2 \, K_2 \right| = o(1)\) for some \(K_2 = K_2(x) \in (0,\infty )\);
- \(W_3\).:
-
\(\displaystyle \sqrt{n h_n} \left| \sum _{i=1}^n w_{ni}(x,h_n) (x_i-x)^2 / 2 - h_n^2 \, K_3 \right| = o(1)\) for some \(K_3 = K_3(x) \in (0,\infty )\);
- \(W_4\).:
-
\(\displaystyle n h_n \sum _{i=1}^n \{ w_{ni}(x,h_n) \}^2 - K_4 = O(1)\) for some \(K_4 = K_4(x) \in (0,\infty )\);
- \(W_5\).:
-
\(\displaystyle \max _{i \in I_{nx}} x_i - \min _{i \in I_{nx}} x_i = o(1)\);
- \(W_6\).:
-
\(\displaystyle (n h_n)^2 \sum _{i=1}^n \{ w_{ni}(x,h_n) \}^4 = O(n^{-\delta })\) for some \(\delta > 0\).
The next assumptions are needed in order to establish the consistency of \({\widehat{B_x}}\) and the validity of the two multiplier bootstrap methods.
- \(W'_1\).:
-
\(\displaystyle \left| \sum _{i=1}^n w'_{ni}(x,g_n) \right| = o(1)\);
- \(W'_2\).:
-
\(\displaystyle n^\delta \sum _{i=1}^n \{ w'_{ni}(x,g_n) \}^2 = O(1)\) for some \(\delta >0\);
- \(W'_3\).:
-
\(\displaystyle \left| \sum _{i=1}^n w'_{ni}(x,g_n)(x_i - x) - 1 \right| = o(1)\);
- \(W''_1\).:
-
\(\displaystyle \sup _{z \in J_{nx}} \left| \sum _{i=1}^n w''_{ni}(z,g_n) \right| = o(1)\);
- \(W''_2\).:
-
\(\displaystyle n^\delta \sup _{z \in J_{nx}} \sum _{i=1}^n \{ w''_{ni}(x,g_n) \}^2 = O(1)\) for some \(\delta >0\);
- \(W''_3\).:
-
\(\displaystyle \sup _{z \in J_{nx}} \left| \sum _{i=1}^n w''_{ni}(z,g_n)(x_i-x) \right| = o(1)\);
- \(W''_4\).:
-
\(\displaystyle \sup _{z \in J_{nx}} \left| \sum _{i=1}^n \{ w''_{ni}(x,g_n)(x_i-x) \}^2 - 1 \right| = o(1)\);
- \(W''_5\).:
-
\(\displaystyle g_n^2 \sup _{z \in J_{nx}} \sum _{i=1}^n \left| w''_{ni}(z,g_n) \right| = O(1)\);
- \(W''_6\).:
-
One can find constants \(C_1, C_2 < \infty \) and \(\alpha >0\) such that for all \(z_1,z_2 \in J_{nx}\),
$$\begin{aligned} \max _{i \in I_{nx}} \left| w''_{ni}(z_1,g_n) - w''_{ni}(z_2,g_n) \right| \le C_1 \, g_n^{-C_2} |z_1-z_2|^\alpha . \end{aligned}$$
Appendix 2: Complementary computations
1.1 Hadamard derivative of the variance functional
Let \(\mathcal {D}\) be the space of distribution functions. Then, for any \(\delta \in \mathcal {D}\), define \(\delta _t = \delta + t {\varDelta }_t \in \ell ^\infty (\mathbb {R})\) such that \({\varDelta }_t \rightarrow {\varDelta }\in \mathbb {D}_0\) uniformly as \(t\rightarrow 0\). Then,
\(\{ \delta _t(y_1\wedge y_2) - \delta _t(y_1)\delta _t(y_2) \} - \{ \delta (y_1\wedge y_2) - \delta (y_1)\delta (y_2) \}\)
It then follows that as \(t\rightarrow 0\),
1.2 Conditional variance and correlation
First introduce the notation \(\mathbf{A}^{(j)} = ( Y_{j1} - \theta _{xjh}^\mathrm{m}, \ldots , Y_{jn} - \theta _{xjh}^\mathrm{m} )\) and \(\mathbf{B}^{(jj')} = \mathrm{diag} ( (\mathbf{A}^{(j)})^\top \mathbf{A}^{(j')})\) for \(j,j' \in \{1,2\}\). For the variance functional, one has
\(\displaystyle {\widehat{({\varLambda }^\mathrm{v})'_{F_x}}}({\varDelta }) = ({\varLambda }^\mathrm{v})_{F_{xh}}'({\varDelta }) =\)
and then long but straightforward computations yield \({\widehat{{\varTheta }}}_x^{(1)} = {\widehat{{\varTheta }}}_x^{(2)} = \sqrt{n h_n} \, \mathbf{B}^{(11)} \, \mathbf{L}_x\). For the correlation functional, one can show that \({\varLambda }^\mathrm{Cov}(F_{xh}) = \mathbf{B}^{(12)} \, \mathbf{w}_x^\top \), \({\varLambda }^\mathrm{v}(F_{1xh}) = \mathbf{B}^{(11)} \, \mathbf{w}_x^\top \) and \({\varLambda }^\mathrm{v}(F_{2xh}) = \mathbf{B}^{(22)} \, \mathbf{w}_x^\top \), so that one can write \(\rho _{xh} = \mathbf{B}^{(12)} \, \mathbf{w}_x^\top / \{ \mathbf{B}^{(11)} \, \mathbf{w}_x^\top \}^{1/2} \{ \mathbf{B}^{(22)} \, \mathbf{w}_x^\top \}^{1/2}\). By similar computations, one can show that
and
1.3 Kendall’s tau
For \(\mathbf {Y}_i = (Y_{1i},Y_{2i})\), \(i \in \{ 1, \ldots , n \}\), let \((\mathcal {K}_{ii'}) = \mathbb {I}(\mathbf {Y}_i \le \mathbf {Y}_{i'})\). Then, one obtains from a direct computation that \(\tau _{xh} = {\varLambda }^\tau (F_{xh}) = -1 + 4 \, \mathbf{w}_x \, \mathcal {K} \, \mathbf{w}_x^\top \). Next, since \(\mathrm{d}F_{xh}\) puts mass \(w_{ni}(x,h_n)\) at \(\mathbf {Y}_i\), one obtains
\(\displaystyle ({\varLambda }^\tau )_{F_{xh}}'({\varDelta })\)
If \(\lim _{\gamma \rightarrow \infty } {\varDelta }(\gamma ,y) = \lim _{\gamma \rightarrow \infty } {\varDelta }(y,\gamma ) = 0\), the last equality reduces to
Hence,
Also, one can show that \({\varLambda }^\tau ({\widehat{F}}_x) = - 1 + 4 ( \mathbf{w}_x + \mathbf{L}_x ) \, \mathcal {K} \, ( \mathbf{w}_x + \mathbf{L}_x )^\top \), and then
1.4 Copula
The empirical conditional copula is \(C_{xh}(\mathbf {u}) = {\varLambda }^\mathrm{C}(F_{xh}) = \mathcal {I}^\mathrm{C}(\mathbf {u}) \, \mathbf{w}_x^\top \), where \((\mathcal {I}^\mathrm{C})_i = \mathbb {I}\{ \mathbf {Y}_i \le \mathbf{F}_{xh}^{-1}(\cdot )\}\). One also obtains \({\widehat{{\varTheta }}}_x^{(1)} = {\widehat{{\varTheta }}}_x^{(1)}(\mathbf {u}) = {\widetilde{\mathcal {I}}}^\mathrm{C}(\mathbf {u}) \, \mathbf{L}_x^\top \), where
The formula for \({\widehat{{\varTheta }}}_x^{(2)}\) is more involved.
1.5 Testing a covariate’s influence
Let \((\mathcal {Y}_{ii'})_{i,i'=1}^n = Y_i \wedge Y_{i'}\). Then by straightforward computations, one can show that
and
Rights and permissions
About this article
Cite this article
Lemyre, F.C., Quessy, JF. Multiplier bootstrap methods for conditional distributions. Stat Comput 27, 805–821 (2017). https://doi.org/10.1007/s11222-016-9655-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11222-016-9655-0