首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 287 毫秒
1.
Ghoudi, Khoudraji & Rivest [The Canadian Journal of Statistics 1998;26:187–197] showed how to test whether the dependence structure of a pair of continuous random variables is characterized by an extreme‐value copula. The test is based on a U‐statistic whose finite‐ and large‐sample variance are determined by the present authors. They propose estimates of this variance which they compare to the jackknife estimate of Ghoudi, Khoudraji & Rivest ( 1998 ) through simulations. They study the finite‐sample and asymptotic power of the test under various alternatives. They illustrate their approach using financial and geological data. The Canadian Journal of Statistics © 2009 Statistical Society of Canada  相似文献   

2.
Kernel density estimation has been used with great success with data that may be assumed to be generated from independent and identically distributed (iid) random variables. The methods and theoretical results for iid data, however, do not directly apply to data from stratified multistage samples. We present finite-sample and asymptotic properties of a modified density estimator introduced in Buskirk (Proceedings of the Survey Research Methods Section, American Statistical Association (1998), pp. 799–801) and Bellhouse and Stafford (Statist. Sin. 9 (1999) 407–424); this estimator incorporates both the sampling weights and the kernel weights. We present regularity conditions which lead the sample estimator to be consistent and asymptotically normal under various modes of inference used with sample survey data. We also introduce a superpopulation structure for model-based inference that allows the population model to reflect naturally occurring clustering. The estimator, and confidence bands derived from the sampling design, are illustrated using data from the US National Crime Victimization Survey and the US National Health and Nutrition Examination Survey.  相似文献   

3.
The authors consider a double robust estimation of the regression parameter defined by an estimating equation in a surrogate outcome set‐up. Under a correct specification of the propensity score, the proposed estimator has smallest trace of asymptotic covariance matrix whether the “working outcome regression model” involved is specified correct or not, and it is particularly meaningful when it is incorrectly specified. Simulations are conducted to examine the finite sample performance of the proposed procedure. Data on obesity and high blood pressure are analyzed for illustration. The Canadian Journal of Statistics 38: 633–646; 2010 © 2010 Statistical Society of Canada  相似文献   

4.
A new general method of combining estimators is proposed in order to obtain an estimator with “improved” small sample properties. It is based on a specification test statistic and incorporates some well-known methods like preliminary testing. It is used to derive an alternative estimator for the slope in the simple errors-in-variables model, combining OLS and the modified instrumental variable estimator by Fuller. Small sample properties of the new estimator are investigated by means of a Monte Carlo study.  相似文献   

5.
We present a multi-level rotation sampling design which includes most of the existing rotation designs as special cases. When an estimator is defined under this sampling design, its variance and bias remain the same over survey months, but it is not so under other existing rotation designs. Using the properties of this multi-level rotation design, we derive the mean squared error (MSE) of the generalized composite estimator (GCE), incorporating the two types of correlations arising from rotating sample units. We show that the MSEs of other existing composite estimators currently used can be expressed as special cases of the GCE. Furthermore, since the coefficients of the GCE are unknown and difficult to determine, we present the minimum risk window estimator (MRWE) as an alternative estimator. This MRWE has the smallest MSE under this rotation design and yet, it is easy to calculate. The MRWE is unbiased for monthly and yearly changes and preserves the internal consistency in total. Our numerical study shows that the MRWE is as efficient as GCE and more efficient than the existing composite estimators and does not suffer from the drift problem [Fuller W.A., Rao J.N.K., 2001. A regression composite estimator with application to the Canadian Labour Force Survey. Surv. Methodol. 27 (2001) 45–51] unlike the regression composite estimators.  相似文献   

6.
In many applications, a finite population contains a large proportion of zero values that make the population distribution severely skewed. An unequal‐probability sampling plan compounds the problem, and as a result the normal approximation to the distribution of various estimators has poor precision. The central‐limit‐theorem‐based confidence intervals for the population mean are hence unsatisfactory. Complex designs also make it hard to pin down useful likelihood functions, hence a direct likelihood approach is not an option. In this paper, we propose a pseudo‐likelihood approach. The proposed pseudo‐log‐likelihood function is an unbiased estimator of the log‐likelihood function when the entire population is sampled. Simulations have been carried out. When the inclusion probabilities are related to the unit values, the pseudo‐likelihood intervals are superior to existing methods in terms of the coverage probability, the balance of non‐coverage rates on the lower and upper sides, and the interval length. An application with a data set from the Canadian Labour Force Survey‐2000 also shows that the pseudo‐likelihood method performs more appropriately than other methods. The Canadian Journal of Statistics 38: 582–597; 2010 © 2010 Statistical Society of Canada  相似文献   

7.
Horvitz and Thompson's (HT) [1952. A generalization of sampling without replacement from a finite universe. J. Amer. Statist. Assoc. 47, 663–685] well-known unbiased estimator for a finite population total admits an unbiased estimator for its variance as given by [Yates and Grundy, 1953. Selection without replacement from within strata with probability proportional to size. J. Roy. Statist. Soc. B 15, 253–261], provided the parent sampling design involves a constant number of distinct units in every sample to be chosen. If the design, in addition, ensures uniform non-negativity of this variance estimator, Rao and Wu [1988. Resampling inference with complex survey data. J. Amer. Statist. Assoc. 83, 231–241] have given their re-scaling bootstrap technique to construct confidence interval and to estimate mean square error for non-linear functions of finite population totals of several real variables. Horvitz and Thompson's estimators (HTE) are used to estimate the finite population totals. Since they need to equate the bootstrap variance of the bootstrap estimator to the Yates and Grundy's estimator (YGE) for the variance of the HTE in case of a single variable, i.e., in the linear case the YG variance estimator is required to be positive for the sample usually drawn.  相似文献   

8.
The authors propose to estimate nonlinear small area population parameters by using the empirical Bayes (best) method, based on a nested error model. They focus on poverty indicators as particular nonlinear parameters of interest, but the proposed methodology is applicable to general nonlinear parameters. They use a parametric bootstrap method to estimate the mean squared error of the empirical best estimators. They also study small sample properties of these estimators by model‐based and design‐based simulation studies. Results show large reductions in mean squared error relative to direct area‐specific estimators and other estimators obtained by “simulated” censuses. The authors also apply the proposed method to estimate poverty incidences and poverty gaps in Spanish provinces by gender with mean squared errors estimated by the mentioned parametric bootstrap method. For the Spanish data, results show a significant reduction in coefficient of variation of the proposed empirical best estimators over direct estimators for practically all domains. The Canadian Journal of Statistics 38: 369–385; 2010 © 2010 Statistical Society of Canada  相似文献   

9.
For binomial data analysis, many methods based on empirical Bayes interpretations have been developed, in which a variance‐stabilizing transformation and a normality assumption are usually required. To achieve the greatest model flexibility, we conduct nonparametric Bayesian inference for binomial data and employ a special nonparametric Bayesian prior—the Bernstein–Dirichlet process (BDP)—in the hierarchical Bayes model for the data. The BDP is a special Dirichlet process (DP) mixture based on beta distributions, and the posterior distribution resulting from it has a smooth density defined on [0, 1]. We examine two Markov chain Monte Carlo procedures for simulating from the resulting posterior distribution, and compare their convergence rates and computational efficiency. In contrast to existing results for posterior consistency based on direct observations, the posterior consistency of the BDP, given indirect binomial data, is established. We study shrinkage effects and the robustness of the BDP‐based posterior estimators in comparison with several other empirical and hierarchical Bayes estimators, and we illustrate through examples that the BDP‐based nonparametric Bayesian estimate is more robust to the sample variation and tends to have a smaller estimation error than those based on the DP prior. In certain settings, the new estimator can also beat Stein's estimator, Efron and Morris's limited‐translation estimator, and many other existing empirical Bayes estimators. The Canadian Journal of Statistics 40: 328–344; 2012 © 2012 Statistical Society of Canada  相似文献   

10.
Donor imputation is frequently used in surveys. However, very few variance estimation methods that take into account donor imputation have been developed in the literature. This is particularly true for surveys with high sampling fractions using nearest donor imputation, often called nearest‐neighbour imputation. In this paper, the authors develop a variance estimator for donor imputation based on the assumption that the imputed estimator of a domain total is approximately unbiased under an imputation model; that is, a model for the variable requiring imputation. Their variance estimator is valid, irrespective of the magnitude of the sampling fractions and the complexity of the donor imputation method, provided that the imputation model mean and variance are accurately estimated. They evaluate its performance in a simulation study and show that nonparametric estimation of the model mean and variance via smoothing splines brings robustness with respect to imputation model misspecifications. They also apply their variance estimator to real survey data when nearest‐neighbour imputation has been used to fill in the missing values. The Canadian Journal of Statistics 37: 400–416; 2009 © 2009 Statistical Society of Canada  相似文献   

11.
Using survey weights, You & Rao [You and Rao, The Canadian Journal of Statistics 2002; 30, 431–439] proposed a pseudo‐empirical best linear unbiased prediction (pseudo‐EBLUP) estimator of a small area mean under a nested error linear regression model. This estimator borrows strength across areas through a linking model, and makes use of survey weights to ensure design consistency and preserve benchmarking property in the sense that the estimators add up to a reliable direct estimator of the mean of a large area covering the small areas. In this article, a second‐order approximation to the mean squared error (MSE) of the pseudo‐EBLUP estimator of a small area mean is derived. Using this approximation, an estimator of MSE that is nearly unbiased is derived; the MSE estimator of You & Rao [You and Rao, The Canadian Journal of Statistics 2002; 30, 431–439] ignored cross‐product terms in the MSE and hence it is biased. Empirical results on the performance of the proposed MSE estimator are also presented. The Canadian Journal of Statistics 38: 598–608; 2010 © 2010 Statistical Society of Canada  相似文献   

12.
The penalized spline is a popular method for function estimation when the assumption of “smoothness” is valid. In this paper, methods for estimation and inference are proposed using penalized splines under additional constraints of shape, such as monotonicity or convexity. The constrained penalized spline estimator is shown to have the same convergence rates as the corresponding unconstrained penalized spline, although in practice the squared error loss is typically smaller for the constrained versions. The penalty parameter may be chosen with generalized cross‐validation, which also provides a method for determining if the shape restrictions hold. The method is not a formal hypothesis test, but is shown to have nice large‐sample properties, and simulations show that it compares well with existing tests for monotonicity. Extensions to the partial linear model, the generalized regression model, and the varying coefficient model are given, and examples demonstrate the utility of the methods. The Canadian Journal of Statistics 40: 190–206; 2012 © 2012 Statistical Society of Canada  相似文献   

13.
This paper considers an improvement of the customary estimator of a finite population mean under a single stage sampling design when paired data, are available on each unit of the sample. Guided by the well known problem of “corninon mean”, a mixture i.e. a weighted combination of the mean of the principal characteristic and that of the auxiliary (possibly transformed) characteristic is proposed. It is shown that, under some conditions, improveinent (with respect to MSE) over the traditional estimator is possible for a broad range of the values of the mixing constant. An estimator of the MSE of the proposed estimator is also provided.  相似文献   

14.
A multivariate “errors in variables” regression model is proposed which generalizes a model previously considered by Gleser and Watson (1973). Maximum likelihood estimators [MLE's] for the parameters of this model are obtained, and the consistency properties of these estimators are investigated. Distribution of the MLE of the “error” variance is obtained in a simple case while the mean and the variance of the estimator are obtained in this case without appealing to the exact distribution.  相似文献   

15.
This article focuses on two‐phase sampling designs for a population with unknown number of rare objects. The first phase is used to estimate the number of rare or potentially rare objects in a population, and the second phase to design sampling plans to capture a certain number or a certain proportion of such type of objects. A hypergeometric‐binomial model is applied to infer the number of rare or potentially rare objects and Monte Carlo simulation based approaches are developed to calculate needed sample sizes. Simulations and real data applications are discussed. The Canadian Journal of Statistics 37: 417–434; 2009 © 2009 Statistical Society of Canada  相似文献   

16.
In this paper the interest is in testing the null hypothesis of positive quadrant dependence (PQD) between two random variables. Such a testing problem is important since prior knowledge of PQD is a qualitative restriction that should be taken into account in further statistical analysis, for example, when choosing an appropriate copula function to model the dependence structure. The key methodology of the proposed testing procedures consists of evaluating a “distance” between a nonparametric estimator of a copula and the independence copula, which serves as a reference case in the whole set of copulas having the PQD property. Choices of appropriate distances and nonparametric estimators of copula are discussed, and the proposed methods are compared with testing procedures based on bootstrap and multiplier techniques. The consistency of the testing procedures is established. In a simulation study the authors investigate the finite sample size and power performances of three types of test statistics, Kolmogorov–Smirnov, Cramér–von‐Mises, and Anderson–Darling statistics, together with several nonparametric estimators of a copula, including recently developed kernel type estimators. Finally, they apply the testing procedures on some real data. The Canadian Journal of Statistics 38: 555–581; 2010 © 2010 Statistical Society of Canada  相似文献   

17.
In this paper, we consider a regression analysis for a missing data problem in which the variables of primary interest are unobserved under a general biased sampling scheme, an outcome‐dependent sampling (ODS) design. We propose a semiparametric empirical likelihood method for accessing the association between a continuous outcome response and unobservable interesting factors. Simulation study results show that ODS design can produce more efficient estimators than the simple random design of the same sample size. We demonstrate the proposed approach with a data set from an environmental study for the genetic effects on human lung function in COPD smokers. The Canadian Journal of Statistics 40: 282–303; 2012 © 2012 Statistical Society of Canada  相似文献   

18.
Although “choose all that apply” questions are common in modern surveys, methods for analyzing associations among responses to such questions have only recently been developed. These methods are generally valid only for simple random sampling, but these types of questions often appear in surveys conducted under more complex sampling plans. The purpose of this article is to provide statistical analysis methods that can be applied to “choose all that apply” questions in complex survey sampling situations. Loglinear models are developed to incorporate the multiple responses inherent in these types of questions. Statistics to compare models and to measure association are proposed and their asymptotic distributions are derived. Monte Carlo simulations show that tests based on adjusted Pearson statistics generally hold their correct size when comparing models. These simulations also show that confidence intervals for odds ratios estimated from loglinear models have good coverage properties, while being shorter than those constructed using empirical estimates. Furthermore, the methods are shown to be applicable to more general problems of modeling associations between elements of two or more binary vectors. The proposed analysis methods are applied to data from the National Health and Nutrition Examination Survey. The Canadian Journal of Statistics © 2009 Statistical Society of Canada  相似文献   

19.
In stratified sampling, methods for the allocation of effort among strata usually rely on some measure of within-stratum variance. If we do not have enough information about these variances, adaptive allocation can be used. In adaptive allocation designs, surveys are conducted in two phases. Information from the first phase is used to allocate the remaining units among the strata in the second phase. Brown et al. [Adaptive two-stage sequential sampling, Popul. Ecol. 50 (2008), pp. 239–245] introduced an adaptive allocation sampling design – where the final sample size was random – and an unbiased estimator. Here, we derive an unbiased variance estimator for the design, and consider a related design where the final sample size is fixed. Having a fixed final sample size can make survey-planning easier. We introduce a biased Horvitz–Thompson type estimator and a biased sample mean type estimator for the sampling designs. We conduct two simulation studies on honey producers in Kurdistan and synthetic zirconium distribution in a region on the moon. Results show that the introduced estimators are more efficient than the available estimators for both variable and fixed sample size designs, and the conventional unbiased estimator of stratified simple random sampling design. In order to evaluate efficiencies of the introduced designs and their estimator furthermore, we first review some well-known adaptive allocation designs and compare their estimator with the introduced estimators. Simulation results show that the introduced estimators are more efficient than available estimators of these well-known adaptive allocation designs.  相似文献   

20.
When sampling from a continuous population (or distribution), we often want a rather small sample due to some cost attached to processing the sample or to collecting information in the field. Moreover, a probability sample that allows for design‐based statistical inference is often desired. Given these requirements, we want to reduce the sampling variance of the Horvitz–Thompson estimator as much as possible. To achieve this, we introduce different approaches to using the local pivotal method for selecting well‐spread samples from multidimensional continuous populations. The results of a simulation study clearly indicate that we succeed in selecting spatially balanced samples and improve the efficiency of the Horvitz–Thompson estimator.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号