首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
For small area estimation of area‐level data, the Fay–Herriot model is extensively used as a model‐based method. In the Fay–Herriot model, it is conventionally assumed that the sampling variances are known, whereas estimators of sampling variances are used in practice. Thus, the settings of knowing sampling variances are unrealistic, and several methods are proposed to overcome this problem. In this paper, we assume the situation where the direct estimators of the sampling variances are available as well as the sample means. Using this information, we propose a Bayesian yet objective method producing shrinkage estimation of both means and variances in the Fay–Herriot model. We consider the hierarchical structure for the sampling variances, and we set uniform prior on model parameters to keep objectivity of the proposed model. For validity of the posterior inference, we show under mild conditions that the posterior distribution is proper and has finite variances. We investigate the numerical performance through simulation and empirical studies.  相似文献   

2.
A nonasymptotic Bayesian approach is developed for analysis of data from threshold autoregressive processes with two regimes. Using the conditional likelihood function, the marginal posterior distribution for each of the parameters is derived along with posterior means and variances. A test for linear functions of the autoregressive coefficients is presented. The approach presented uses a posterior p-value averaged over the values of the threshold. The one-step ahead predictive distribution is derived along with the predictive mean and variance. In addition, equivalent results are derived conditional upon a value of the threshold. A numerical example is presented to illustrate the approach.  相似文献   

3.
4.
In the health and social sciences, researchers often encounter categorical data for which complexities come from a nested hierarchy and/or cross-classification for the sampling structure. A common feature of these studies is a non-standard data structure with repeated measurements which may have some degree of clustering. In this paper, methodology is presented for the joint estimation of quantities of interest in the context of a stratified two-stage sample with bivariate dichotomous data. These quantities are the mean value π of an observed dichotomous response for a certain condition or time-point and a set of correlation coefficients for intra-cluster association for each condition or time period and for inter-condition correlation within and among clusters. The methodology uses the cluster means and pairwise joint probability parameters from each cluster. They together provide appropriate information across clusters for the estimation of the correlation coefficients.  相似文献   

5.
Ori Davidov  Chang Yu 《Statistics》2013,47(2):163-173
We provide a method for estimating the sample mean of a continuous outcome in a stratified population using a double sampling scheme. The stratified sample mean is a weighted average of stratum specific means. It is assumed that the fallible and true outcome data are related by a simple linear regression model in each stratum. The optimal stratified double sampling plan, i.e. , the double sampling plan that minimizes the cost of sampling for fixed variances, or alternatively, minimizes the variance for fixed costs, is found and compared to a standard sampling plan. The design parameters are the total sample size and the number of doubly sampled units in each stratum. We show that the optimal double sampling plan is a function of the between-strata and within-strata cost and variance ratios. The efficiency gains, relative to standard sampling plans, under broad set of conditions, are considerable.  相似文献   

6.
Although Cox proportional hazards regression is the default analysis for time to event data, there is typically uncertainty about whether the effects of a predictor are more appropriately characterized by a multiplicative or additive model. To accommodate this uncertainty, we place a model selection prior on the coefficients in an additive-multiplicative hazards model. This prior assigns positive probability, not only to the model that has both additive and multiplicative effects for each predictor, but also to sub-models corresponding to no association, to only additive effects, and to only proportional effects. The additive component of the model is constrained to ensure non-negative hazards, a condition often violated by current methods. After augmenting the data with Poisson latent variables, the prior is conditionally conjugate, and posterior computation can proceed via an efficient Gibbs sampling algorithm. Simulation study results are presented, and the methodology is illustrated using data from the Framingham heart study.  相似文献   

7.
For a linear regression model over m populations with separate regression coefficients but a common error variance, a Bayesian model is employed to obtain regression coefficient estimates which are shrunk toward an overall value. The formulation uses Normal priors on the coefficients and diffuse priors on the grand mean vectors, the error variance, and the between-to-error variance ratios. The posterior density of the parameters which were given diffuse priors is obtained. From this the posterior means and variances of regression coefficients and the predictive mean and variance of a future observation are obtained directly by numerical integration in the balanced case, and with the aid of series expansions in the approximately balanced case. An example is presented and worked out for the case of one predictor variable. The method is an extension of Box & Tiao's Bayesian estimation of means in the balanced one-way random effects model.  相似文献   

8.
Ghosh and Lahiri (1987a,b) considered simultaneous estimation of several strata means and variances where each stratum contains a finite number of elements, under the assumption that the posterior expectation of any stratum mean is a linear function of the sample observations - the so called“posterior linearity” property. In this paper we extend their result by retaining the “posterior linearity“ property of each stratum mean but allowing the superpopulation model whose mean as well as the variance-covariance structure changes from stratum to stratum. The performance of the proposed empirical Bayes estimators are found to be satisfactory both in terms of “asymptotic optimality” (Robbins (1955)) and “relative savings loss” (Efron and Morris (1973)).  相似文献   

9.
Empirical Bayes methods are used to estimate the extent of the undercount at the local level in the 1980 U.S. census. "Grouping of like subareas from areas such as states, counties, and so on into strata is a useful way of reducing the variance of undercount estimators. By modeling the subareas within a stratum to have a common mean and variances inversely proportional to their census counts, and by taking into account sampling of the areas (e.g., by dual-system estimation), empirical Bayes estimators that compromise between the (weighted) stratum average and the sample value can be constructed. The amount of compromise is shown to depend on the relative importance of stratum variance to sampling variance. These estimators are evaluated at the state level (51 states, including Washington, D.C.) and stratified on race/ethnicity (3 strata) using data from the 1980 postenumeration survey (PEP 3-8, for the noninstitutional population)."  相似文献   

10.
A Bayesian analysis is presented for the K-group Behrens-Fisher problem. Both exact posterior distributions and approximations were developed for both a general linear contrast of the K means and the K variances, given either proper diffuse or informative conjugate priors. The contrast of variances is a unique feature of the heterogeneous variance model that enables investigators to test specific effects of experimental manipulations on variance. Finally, important-differences were observed between the heterogeneous variance model and the homogeneous model.  相似文献   

11.
Analysis of categorical data by linear models is extended to data obtained by stratified random sampling. It is shown that, asymptotically, proportional allocation reduces the variances of estimators from those obtained hy simple random sampling. The difference between the asymptotic covariance matrices of estimated parameters obtained by simple random sampling and stratified random sampling with proportional allocation is shown to be positive definite vinder fairly non-restrictive conditions, when an asymptotically efficient method of estimation is used. Data from a major community study of mental health are used to illustrate application of the technique.  相似文献   

12.
The implications of including autoregressive disturbances in linear logit models of demand systems are explored. It is argued that the normality assumption of the error terms is more appropriate in the linear logit model than in a share equation model with additive disturbances (commonly found in the literature). Autoregressive disturbances and their implications for model estimation are discussed in that context. Both theoretical arguments and empirical evidence are presented in favor of the logit specification given the presence of serial correlation.  相似文献   

13.
A Bayesian model consists of two elements: a sampling model and a prior density. The problem of selecting a prior density is nothing but the problem of selecting a Bayesian model where the sampling model is fixed. A predictive approach is used through a decision problem where the loss function is the squared L 2 distance between the sampling density and the posterior predictive density, because the aim of the method is to choose the prior that provides a posterior predictive density as good as possible. An algorithm is developed for solving the problem; this algorithm is based on Lavine's linearization technique.  相似文献   

14.
MODEL-BASED VARIANCE ESTIMATION IN SURVEYS WITH STRATIFIED CLUSTERED DESIGN   总被引:1,自引:0,他引:1  
A model-based method for estimating the sampling variances of estimators of (sub-)population means, proportions, quantiles, and regression parameters in surveys with stratified clustered design is described and applied to a survey of US secondary education. The method is compared with the jackknife by a simulation study. The model-based estimators of the sampling variances have much smaller mean squared errors than their jackknife counterparts. In addition, they can be improved by incorporating information about the unknown parameters (variances) from external sources. A regression-based smoothing method for estimating the sampling variances of the estimators for a large number of subpopulation means is proposed. Such smoothing may be invaluable when subpopulations are represented in the sample by only few subjects.  相似文献   

15.
In this paper we present a two-stage sampling procedure for testing the equality of normal means against ordered alternatives in one-way analysis of variance with unequal unknown variances. A table of approximated percentiles needed for implementation is provided. Some Monte Carlo results for estimating the power of the proposed test statistic are presented.  相似文献   

16.
Usual stratified sampling design assume that one is able to draw units directly from given strata. If this is not possible, one can use the following double sampling procedure: First take a large simple random sample out of the whole population and find out, to which stratum each sample unit belongs. Out of these chosen units take a second stratified sample. In this paper unbiased estimators for this procedure in the cases of known (part I) and unknown (part II) stratum weights are proposed for sampling with replacement and sampling without replacement and their variances are evaluated.  相似文献   

17.
The multivariate regression model is considered with p regressors. A latent vector with p binary entries serves to identify one of two types of regression coefficients: those close to 0 and those not. Specializing our general distributional setting to the linear model with Gaussian errors and using natural conjugate prior distributions, we derive the marginal posterior distribution of the binary latent vector. Fast algorithms aid its direct computation, and in high dimensions these are supplemented by a Markov chain Monte Carlo approach to sampling from the known posterior distribution. Problems with hundreds of regressor variables become quite feasible. We give a simple method of assigning the hyperparameters of the prior distribution. The posterior predictive distribution is derived and the approach illustrated on compositional analysis of data involving three sugars with 160 near infrared absorbances as regressors.  相似文献   

18.
Numerous optimization problems arise in survey designs. The problem of obtaining an optimal (or near optimal) sampling design can be formulated and solved as a mathematical programming problem. In multivariate stratified sample surveys usually it is not possible to use the individual optimum allocations for sample sizes to various strata for one reason or another. In such situations some criterion is needed to work out an allocation which is optimum for all characteristics in some sense. Such an allocation may be called an optimum compromise allocation. This paper examines the problem of determining an optimum compromise allocation in multivariate stratified random sampling, when the population means of several characteristics are to be estimated. Formulating the problem of allocation as an all integer nonlinear programming problem, the paper develops a solution procedure using a dynamic programming technique. The compromise allocation discussed is optimal in the sense that it minimizes a weighted sum of the sampling variances of the estimates of the population means of various characteristics under study. A numerical example illustrates the solution procedure and shows how it compares with Cochran's average allocation and proportional allocation.  相似文献   

19.
Although bootstrapping has become widely used in statistical analysis, there has been little reported concerning bootstrapped Bayesian analyses, especially when there is proper prior informa-tion concerning the parameter of interest. In this paper, we first propose an operationally implementable definition of a Bayesian bootstrap. Thereafter, in simulated studies of the estimation of means and variances, this Bayesian bootstrap is compared to various parametric procedures. It turns out that little information is lost in using the Bayesian bootstrap even when the sampling distribution is known. On the other hand, the parametric procedures are at times very sensitive to incorrectly specified sampling distributions, implying that the Bayesian bootstrap is a very robust procedure for determining the posterior distribution of the parameter.  相似文献   

20.
Empirical researchers face a trade-off between the lower resource costs associated with smaller samples and the increased confidence in the results gained from larger samples. Choice of sampling strategy is one tool researchers can use to reduce costs yet still attain desired confidence levels. This study uses Monte Carlo simulation to examine the impact of nine sampling strategies on the finite sample performance of the maximum likelihood logit estimator. The results show stratified random sampling with balanced strata sizes and a bias correction for choice-based sampling outperforms all other sampling strategies with respect to four small-sample performance measures.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号