Abstract
In sample surveys, there is often insufficient sample size to obtain reliable direct estimates for parameters of interest for certain domains. Precision can be increased by introducing small area models which ‘borrow strength’ by connecting different areas through use of explicit linking models, area-specific random effects, and auxiliary covariate information. One consequence of the use of small area models is that small area estimates at a lower (for example, county) geographic level typically will not aggregate to the estimate at the corresponding higher (for example, state) geographic level. Benchmarking is the statistical procedure for reconciling these differences. This paper provides new perspectives for the benchmarking problem, especially for complex Bayesian small area models which require Markov Chain Monte Carlo estimation. Two new approaches to Bayesian benchmarking are introduced: one procedure based on minimum discrimination information, and another procedure for fully Bayesian self-consistent conditional benchmarking. Notably the proposed procedures construct adjusted posterior distributions whose first and higher order moments are consistent with the benchmarking constraints. It is shown that certain existing benchmarked estimators are special cases of the proposed methodology under normality, giving a distributional justification for the use of benchmarked estimates. Additionally, a ‘flexible’ benchmarking constraint is introduced, where the higher geographic level estimate is not considered fixed, and is simultaneously adjusted, along with lower level estimates.

Similar content being viewed by others
Notes
SAHIE estimates can be found at www.census.gov/did/www/sahie/data/index.html.
References
Azzalini A (1985) A class of distributions which includes the normal ones. Scand J Stat 12:171–178
Battese GE, Harter RH, Fuller WA (1988) An error-components model for prediction of county crop areas using survey and satellite data. J Am Stat Assoc 83:28–36
Bell WR, Datta GS, Ghosh M (2013) Benchmarking small area estimators. Biometrika 100:189–202
Berger JO (1985) Statistical decision theory and Bayesian analysis, 2nd edn. Springer, New York
Datta GS, Ghosh M, Steorts R, Maples J (2011) Bayesian benchmarking with applications to small area estimation. TEST 20:574–588
Fay RE, Herriot RA (1979) Estimates of income from small places: an application of James–Stein procedures to census data. J Am Stat Assoc 74:269–277
Gelfand AE, Smith AFM (1990) Sampling-based approaches to calculating marginal densities. J Am Stat Assoc 85:398–409
Ghosh M, Steorts R (2013) Two-stage Bayesian benchmarking as applied to small area estimation. TEST 22(4):670–687
Ghosh M, Kubokawa T, Kawakubo Y (2015) Benchmarked empirical bayes methods in multiplicative area-level models with risk evaluation. Biometrika 102:647–659
Isaki CT, Tsay JH, Fuller WA (2000) Estimation of census adjustment factors. Surv Methodol 26:31–42
Jaynes ET (1957) Information theory and statistical mechanics. Phys Rev 106:620–630
Knottnerus P (2003) Sample survey theory: some pythagorean perspectives. Springer, New York
Kullback S (1959) Information theory and statistics. Wiley, New York
Kullback S, Liebler RA (1951) On information and sufficiency. Ann Math Stat 22:79–86
Nandram B, Toto MCS, Choi JW (2011) A Bayesian benchmarking of the Scott–Smith model for small areas. J Stat Comput Simul 81:1593–1608
Pfeffermann D, Barnard CH (1991) New estimators for small-area means with applications to the assessment of farmland values. J Bus Econ Stat 9:73–84
R Development Core Team (2011) R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. ISBN 3-900051-07-0. http://www.R-project.org/
Rao JNK, Molina I (2015) Small area estimation, 2nd edn. Wiley, New York
Seber GAF (2008) A matrix handbook for statisticians. Wiley, Hoboken
Sostra K, Traat I (2009) Optimal domain estimation under summation restriction. J Stat Plann Inference 139:3928–3941
Toto MCS, Nandram B (2010) A Bayesian predictive inference for small area means incorporating covariates and sampling weights. J Stat Plann Inference 140:2963–2979
Wang J, Fuller WA, Qu Y (2008) Small area estimation under a restriction. Surv Methodol 34:29–36
You Y, Rao JNK (2002) A pseudo-empirical best linear unbiased prediction approach to small area estimation using survey weights. Can J Stat 30:431–439
You Y, Rao JNK, Dick P (2004) Benchmarking hierarchical Bayes small area estimators in the Canadian census undercoverage estimation. Stat Transit 6:631–640
Acknowledgements
This report is released to inform interested parties of ongoing research and to encourage discussion of work in progress. The views expressed are those of the authors and not necessarily those of the U. S. Census Bureau. The authors wish to thank the SAHIE group at the U. S. Census Bureau for many discussions about benchmarking problems related to health insurance estimation. The authors are also grateful to the associate editor and referee for their careful and detailed reviews, which greatly improved the paper.
Author information
Authors and Affiliations
Corresponding author
Appendix 1: Minimum discrimination information
Appendix 1: Minimum discrimination information
Using the same notation as in the main text, assume the posterior distribution, \(\pi \), conditional on the data and auxiliary information, is \(\varvec{\theta } \sim N_{m + 1} \left( \tilde{\varvec{\theta }}, \varvec{\varSigma } \right) \). The goal of this section is to compute the distribution \(\pi ^*\) which minimizes the K–L divergence (7) from the posterior distribution \(\pi \), subject to the benchmarking constraints (9). This constrained minimization problem can be solved in a straightforward way using Lagrange multipliers; however, the calculations are long and tedious. The MDI distribution can be found in a more direct way due to the structure of the constraints and properties of the normal distribution.
Kullback (1959) shows that the solution \(\pi ^*\) is a member of an exponential family, and if \(\pi \) is Gaussian, so is \(\pi ^*\). Hence, \(\pi ^* \sim N_{m + 1} \left( \varvec{\mu ^*}, \varvec{\varSigma }^* \right) \), where \(\varvec{\mu ^*}\) and \(\varvec{\varSigma }^*\) are chosen to satisfy equation (9). Using the properties of the multivariate normal distribution, it is easy to show that
Let \(E^* \left( \cdot \right) \) and \(Var^* \left( \cdot \right) \) denote the expectation and variance operators, respectively, with respect to \(\pi ^*\). Then the constraints (9) can be written
and, using (24),
so that the first constraint (24) is a function only of \(\varvec{\mu }^*\) while the second constraint (25) is a function only of \(\varvec{\varSigma }^*\). Since (23) can be written as the sum of two terms, one which is a function of \(\tilde{\varvec{\theta }}\) and \(\varvec{\mu }^*\), and the other which is a function of \(\varvec{\varSigma }^*\), the optimization problem can be simplified by minimizing (23) over \(\varvec{\mu }^*\) subject to constraint (24), and then minimizing (23) over \(\varvec{\varSigma }^*\) subject to constraint (25). Furthermore, since the terms in (23) involving \(\varvec{\varSigma }^*\) do not involve \(\tilde{\varvec{\theta }}\), the covariance of \(\pi ^*\) will not be a function of \(\tilde{\varvec{\theta }}\), so we may set \(\tilde{\varvec{\theta }} = \varvec{0}\) when solving for \(\varvec{\varSigma }^*\), without loss of generality. We can then use a result of Kullback (1959) which states that for a general restriction T, the solution to the minimization problem
is given by \(\pi ^* \left( \varvec{\theta }\right) \propto e^{\tau ^* T \left( \varvec{\theta }\right) } \pi \left( \varvec{\theta }\right) \), where \(\tau ^*\) is the solution to \(\frac{d}{d \tau } \log M_2(\tau ^*) = 0\) and \(M_2(\tau ) = \int {e^{\tau T ( \varvec{\theta })} \pi ( \varvec{\theta }) d \varvec{\theta }}\).
1.1 Appendix 1.1: Flexible benchmarking constraint
The first moment condition, \(T_1\) in (8), can be used to calculate \(\varvec{\mu ^*}\), which is the mean of \(\pi ^*\), the solution to
Based on the moment generating function of the multivariate normal distribution,
The solution to the equation \(\frac{\partial }{\partial \tau } \log M_2 \left( \tau \right) = 0\) is \(\tau ^* = - \tilde{\varvec{\theta }}^T {\varvec{R}}( {\varvec{R}}^T \varvec{\varSigma } {\varvec{R}})^{-1}\); the MDI distribution, \(\pi ^*\), is then
so that the multivariate normal MDI distribution satisfying the first moment benchmarking constraint has mean
The second moment condition, \(T_2\), can be used to calculate \(\varvec{\varSigma }^*\), which is the covariance of \(\pi ^*\), the solution to
Without loss of generality, assume \(\tilde{\varvec{\theta }} = {\mathbf {0}}\). Under this assumption,
where \(\sigma ^2_s = {\varvec{C}}^T \varvec{\varSigma }_s {\varvec{C}}\). Thus, we have the following independent distributions:
Using the moment generating function of the Chi-Squared distribution,
Solving \( \partial \log M_2 \left( \tau \right) / \partial \tau = 0 \) for \(\tau \) gives \(\tau ^* = \left( \sigma ^2_s - \sigma ^2 \right) / \left( 4 \sigma ^2_s \sigma ^2 \right) \), hence the MDI distribution, \(\pi ^*\), is
This factorization shows that the multivariate normal MDI distribution satisfying the second moment benchmarking constraint has covariance matrix
Combining Eqs. (26) and (27), the MDI distribution satisfying the first and second moment benchmarking restrictions is \(\varvec{\theta }^* \sim N_{m + 1} \left( \varvec{\mu ^*}, \varvec{\varSigma }^* \right) \).
1.2 Appendix 1.2: Fixed benchmarking constraint
If the higher level parameters are assumed fixed and known, the moment restrictions are
Using the same procedures as above, the first and second moments of the MDI distribution can be computed separately, and it can be shown that the MDI distribution is \(\varvec{\theta }^* \sim N_m \left( \varvec{\mu }^*, \varvec{\varSigma }^* \right) \), where
and
Rights and permissions
About this article
Cite this article
Janicki, R., Vesper, A. Benchmarking techniques for reconciling Bayesian small area models at distinct geographic levels. Stat Methods Appl 26, 557–581 (2017). https://doi.org/10.1007/s10260-017-0379-x
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10260-017-0379-x