首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
We consider the classic problem of interval estimation of a proportion p based on binomial sampling. The ‘exact’ Clopper–Pearson confidence interval for p is known to be unnecessarily conservative. We propose coverage adjustments of the Clopper–Pearson interval that incorporate prior or posterior beliefs into the interval. Using heatmap‐type plots for comparing confidence intervals, we show that the coverage‐adjusted intervals have satisfying coverage and shorter expected lengths than competing intervals found in the literature.  相似文献   

2.
Yu et al. [An improved score interval with a modified midpoint for a binomial proportion. J Stat Comput Simul. 2014;84:1022–1038] propose a novel confidence interval (CI) for a binomial proportion by modifying the midpoint of the score interval. This CI is competitive with the various commonly used methods. At the same time, Martín and Álvarez [Two-tailed asymptotic inferences for a proportion. J Appl Stat. 2014;41:1516–1529] analyse the performance of 29 asymptotic two-tailed CI for a proportion. The CI they selected is based on the arcsin transformation (when this is applied to the data increased by 0.5), although they also refer to the good behaviour of the classical methods of score and Agresti and Coull (which may be preferred in certain circumstances). The aim of this commentary is to compare the four methods referred to previously. The conclusion (for the classic error α of 5%) is that with a small sample size (≤80) the method that should be used is that of Yu et al.; for a large sample size (n?≥?100), the four methods perform in a similar way, with a slight advantage for the Agresti and Coull method. In any case the Agresti and Coull method does not perform badly and tends to be conservative. The program which determines these four intervals are available from the address http://www.ugr.es/local/bioest/Z_LINEAR_K.EXEhttp://www.ugr.es/local/bioest/Z_LINEAR_K.EXE.  相似文献   

3.
Estimation of the parameters of an exponential distribution based on record data has been treated by Samaniego and Whitaker [On estimating population characteristics from record-breaking observations, I. Parametric results, Naval Res. Logist. Q. 33 (1986), pp. 531–543] and Doostparast [A note on estimation based on record data, Metrika 69 (2009), pp. 69–80]. Recently, Doostparast and Balakrishnan [Optimal record-based statistical procedures for the two-parameter exponential distribution, J. Statist. Comput. Simul. 81(12) (2011), pp. 2003–2019] obtained optimal confidence intervals as well as uniformly most powerful tests for one- and two-sided hypotheses concerning location and scale parameters based on record data from a two-parameter exponential model. In this paper, we derive optimal statistical procedures including point and interval estimation as well as most powerful tests based on record data from a two-parameter Pareto model. For illustrative purpose, a data set on annual wages of a sample of production-line workers in a large industrial firm is analysed using the proposed procedures.  相似文献   

4.
Rasul A. Khan 《Statistics》2015,49(3):705-710
Let X1, X2, …, Xn be iid N(μ, aμ2) (a>0) random variables with an unknown mean μ>0 and known coefficient of variation (CV) √a. The estimation of μ is revisited and it is shown that a modified version of an unbiased estimator of μ [cf. Khan RA. A note on estimating the mean of a normal distribution with known CV. J Am Stat Assoc. 1968;63:1039–1041] is more efficient. A certain linear minimum mean square estimator of Gleser and Healy [Estimating the mean of a normal distribution with known CV. J Am Stat Assoc. 1976;71:977–981] is also modified and improved. These improved estimators are being compared with the maximum likelihood estimator under squared-error loss function. Based on asymptotic consideration, a large sample confidence interval is also mentioned.  相似文献   

5.
Inference concerning the negative binomial dispersion parameter, denoted by c, is important in many biological and biomedical investigations. Properties of the maximum-likelihood estimator of c and its bias-corrected version have been studied extensively, mainly, in terms of bias and efficiency [W.W. Piegorsch, Maximum likelihood estimation for the negative binomial dispersion parameter, Biometrics 46 (1990), pp. 863–867; S.J. Clark and J.N. Perry, Estimation of the negative binomial parameter κ by maximum quasi-likelihood, Biometrics 45 (1989), pp. 309–316; K.K. Saha and S.R. Paul, Bias corrected maximum likelihood estimator of the negative binomial dispersion parameter, Biometrics 61 (2005), pp. 179–185]. However, not much work has been done on the construction of confidence intervals (C.I.s) for c. The purpose of this paper is to study the behaviour of some C.I. procedures for c. We study, by simulations, three Wald type C.I. procedures based on the asymptotic distribution of the method of moments estimate (mme), the maximum-likelihood estimate (mle) and the bias-corrected mle (bcmle) [K.K. Saha and S.R. Paul, Bias corrected maximum likelihood estimator of the negative binomial dispersion parameter, Biometrics 61 (2005), pp. 179–185] of c. All three methods show serious under-coverage. We further study parametric bootstrap procedures based on these estimates of c, which significantly improve the coverage probabilities. The bootstrap C.I.s based on the mle (Boot-MLE method) and the bcmle (Boot-BCM method) have coverages that are significantly better (empirical coverage close to the nominal coverage) than the corresponding bootstrap C.I. based on the mme, especially for small sample size and highly over-dispersed data. However, simulation results on lengths of the C.I.s show evidence that all three bootstrap procedures have larger average coverage lengths. Therefore, for practical data analysis, the bootstrap C.I. Boot-MLE or Boot-BCM should be used, although Boot-MLE method seems to be preferable over the Boot-BCM method in terms of both coverage and length. Furthermore, Boot-MLE needs less computation than Boot-BCM.  相似文献   

6.
This paper discusses the classic but still current problem of interval estimation of a binomial proportion. Bootstrap methods are presented for constructing such confidence intervals in a routine, automatic way. Three confidence intervals for a binomial proportion are compared and studied by means of a simulation study, namely: the Wald confidence interval, the Agresti–Coull interval and the bootstrap-t interval. A new confidence interval, the Agresti–Coull interval with bootstrap critical values, is also introduced and its good behaviour related to the average coverage probability is established by means of simulations.  相似文献   

7.
Small sample properties of seven confidence intervals for the binomial parameterp (based on various normal approximations) and of the Clopper-Pearson interval are compared. Coverage probabilities and expected lower and upper limits of the intervals are graphically displayed as functions of the binomial parameterp for various sample sizes.  相似文献   

8.
The paper deals with generalized confidence intervals for the between-group variance in one-way heteroscedastic (unbalanced) ANOVA with random effects. The approach used mimics the standard one applied in mixed linear models with two variance components, where interval estimators are based on a minimal sufficient statistic derived after an initial reduction by the principle of invariance. A minimal sufficient statistic under heteroscedasticity is found to resemble its homoscedastic counterpart and further analogies between heteroscedastic and homoscedastic cases lead us to two classes of fiducial generalized pivots for the between-group variance. The procedures suggested formerly by Wimmer and Witkovský [Between group variance component interval estimation for the unbalanced heteroscedastic one-way random effects model, J. Stat. Comput. Simul. 73 (2003), pp. 333–346] and Li [Comparison of confidence intervals on between group variance in unbalanced heteroscedastic one-way random models, Comm. Statist. Simulation Comput. 36 (2007), pp. 381–390] are found to belong to these two classes. We comment briefly on some of their properties that were not mentioned in the original papers. In addition, properties of another particular generalized pivot are considered.  相似文献   

9.
ABSTRACT

For interval estimation of a binomial proportion and a Poisson mean, matching pseudocounts are derived, which give the one-sided Wald confidence intervals with second-order accuracy. The confidence intervals remove the bias of coverage probabilities given by the score confidence intervals. Partial poor behavior of the confidence intervals by the matching pseudocounts is corrected by hybrid methods using the score confidence interval depending on sample values.  相似文献   

10.
Guogen Shan 《Statistics》2018,52(5):1086-1095
In addition to point estimate for the probability of response in a two-stage design (e.g. Simon's two-stage design for binary endpoints), confidence limits should be computed and reported. The current method of inverting the p-value function to compute the confidence interval does not guarantee coverage probability in a two-stage setting. The existing exact approach to calculate one-sided limits is based on the overall number of responses to order the sample space. This approach could be conservative because many sample points have the same limits. We propose a new exact one-sided interval based on p-value for the sample space ordering. Exact intervals are computed by using binomial distributions directly, instead of a normal approximation. Both exact intervals preserve the nominal confidence level. The proposed exact interval based on the p-value generally performs better than the other exact interval with regard to expected length and simple average length of confidence intervals.  相似文献   

11.
Doubly robust (DR) estimators of the mean with missing data are compared. An estimator is DR if either the regression of the missing variable on the observed variables or the missing data mechanism is correctly specified. One method is to include the inverse of the propensity score as a linear term in the imputation model [D. Firth and K.E. Bennett, Robust models in probability sampling, J. R. Statist. Soc. Ser. B. 60 (1998), pp. 3–21; D.O. Scharfstein, A. Rotnitzky, and J.M. Robins, Adjusting for nonignorable drop-out using semiparametric nonresponse models (with discussion), J. Am. Statist. Assoc. 94 (1999), pp. 1096–1146; H. Bang and J.M. Robins, Doubly robust estimation in missing data and causal inference models, Biometrics 61 (2005), pp. 962–972]. Another method is to calibrate the predictions from a parametric model by adding a mean of the weighted residuals [J.M Robins, A. Rotnitzky, and L.P. Zhao, Estimation of regression coefficients when some regressors are not always observed, J. Am. Statist. Assoc. 89 (1994), pp. 846–866; D.O. Scharfstein, A. Rotnitzky, and J.M. Robins, Adjusting for nonignorable drop-out using semiparametric nonresponse models (with discussion), J. Am. Statist. Assoc. 94 (1999), pp. 1096–1146]. The penalized spline propensity prediction (PSPP) model includes the propensity score into the model non-parametrically [R.J.A. Little and H. An, Robust likelihood-based analysis of multivariate data with missing values, Statist. Sin. 14 (2004), pp. 949–968; G. Zhang and R.J. Little, Extensions of the penalized spline propensity prediction method of imputation, Biometrics, 65(3) (2008), pp. 911–918]. All these methods have consistency properties under misspecification of regression models, but their comparative efficiency and confidence coverage in finite samples have received little attention. In this paper, we compare the root mean square error (RMSE), width of confidence interval and non-coverage rate of these methods under various mean and response propensity functions. We study the effects of sample size and robustness to model misspecification. The PSPP method yields estimates with smaller RMSE and width of confidence interval compared with other methods under most situations. It also yields estimates with confidence coverage close to the 95% nominal level, provided the sample size is not too small.  相似文献   

12.
Tests for the equality of variances are of interest in many areas such as quality control, agricultural production systems, experimental education, pharmacology, biology, as well as a preliminary to the analysis of variance, dose–response modelling or discriminant analysis. The literature is vast. Traditional non-parametric tests are due to Mood, Miller and Ansari–Bradley. A test which usually stands out in terms of power and robustness against non-normality is the W50 Brown and Forsythe [Robust tests for the equality of variances, J. Am. Stat. Assoc. 69 (1974), pp. 364–367] modification of the Levene test [Robust tests for equality of variances, in Contributions to Probability and Statistics, I. Olkin, ed., Stanford University Press, Stanford, 1960, pp. 278–292]. This paper deals with the two-sample scale problem and in particular with Levene type tests. We consider 10 Levene type tests: the W50, the M50 and L50 tests [G. Pan, On a Levene type test for equality of two variances, J. Stat. Comput. Simul. 63 (1999), pp. 59–71], the R-test [R.G. O'Brien, A general ANOVA method for robust tests of additive models for variances, J. Am. Stat. Assoc. 74 (1979), pp. 877–880], as well as the bootstrap and permutation versions of the W50, L50 and R tests. We consider also the F-test, the modified Fligner and Killeen [Distribution-free two-sample tests for scale, J. Am. Stat. Assoc. 71 (1976), pp. 210–213] test, an adaptive test due to Hall and Padmanabhan [Adaptive inference for the two-sample scale problem, Technometrics 23 (1997), pp. 351–361] and the two tests due to Shoemaker [Tests for differences in dispersion based on quantiles, Am. Stat. 49(2) (1995), pp. 179–182; Interquantile tests for dispersion in skewed distributions, Commun. Stat. Simul. Comput. 28 (1999), pp. 189–205]. The aim is to identify the effective methods for detecting scale differences. Our study is different with respect to the other ones since it is focused on resampling versions of the Levene type tests, and many tests considered here have not ever been proposed and/or compared. The computationally simplest test found robust is W50. Higher power, while preserving robustness, is achieved by considering the resampling version of Levene type tests like the permutation R-test (recommended for normal- and light-tailed distributions) and the bootstrap L50 test (recommended for heavy-tailed and skewed distributions). Among non-Levene type tests, the best one is the adaptive test due to Hall and Padmanabhan.  相似文献   

13.
In this paper, we suggest three new ratio estimators of the population mean using quartiles of the auxiliary variable when there are missing data from the sample units. The suggested estimators are investigated under the simple random sampling method. We obtain the mean square errors equations for these estimators. The suggested estimators are compared with the sample mean and ratio estimators in the case of missing data. Also, they are compared with estimators in Singh and Horn [Compromised imputation in survey sampling, Metrika 51 (2000), pp. 267–276], Singh and Deo [Imputation by power transformation, Statist. Papers 45 (2003), pp. 555–579], and Kadilar and Cingi [Estimators for the population mean in the case of missing data, Commun. Stat.-Theory Methods, 37 (2008), pp. 2226–2236] and present under which conditions the proposed estimators are more efficient than other estimators. In terms of accuracy and of the coverage of the bootstrap confidence intervals, the suggested estimators performed better than other estimators.  相似文献   

14.
In the context of discrete data, a sequential fixed-width confidence interval for an unknown parameter in a parametric model is constructed using a minimum Hellinger distance estimator (MHD) as the center of the interval. It is shown that our sequential procedure is asymptotically consistent and efficient, when the assumed parametric model is correct. These results, in addition to being exactly same as those obtained by Khan [1969, A general method of determining fixed-width confidence intervals. Ann. Math. Statist. 40, 704–709] and Yu [1989, On fixed-width confidence intervals associated with maximum likelihood estimation. J. Theoret. Probab. 2, 193–199] using a maximum likelihood estimator (MLE), offer an alternative which has several in-built robustness properties. Monte Carlo simulations show that the performance of our sequential procedure based on MHD, measured in terms of average sample size and the coverage probability, are as good as those based on MLE, when the assumed Poisson model is correct. However, when the samples come from a gross-error contaminated Poisson model, our numerical results show that the deviation from the Poisson model assumption severely affects the performance of the sequential procedure based on MLE, while the procedure based on MHD continues to perform well, thus exhibiting robustness of MHD against gross-error contaminations even for random sample sizes.  相似文献   

15.
Record scheme is a method to reduce the total time on test of an experiment. In this scheme, items are sequentially observed and only values smaller than all previous ones are recorded. In some situations, when the experiments are time-consuming and sometimes the items are lost during the experiment, the record scheme dominates the usual random sample scheme [M. Doostparast and N. Balakrishnan, Optimal sample size for record data and associated cost analysis for exponential distribution, J. Statist. Comput. Simul. 80(12) (2010), pp. 1389–1401]. Estimation of the mean of an exponential distribution based on record data has been treated by Samaniego and Whitaker [On estimating population characteristics from record breaking observations I. Parametric results, Naval Res. Logist. Q. 33 (1986), pp. 531–543] and Doostparast [A note on estimation based on record data, Metrika 69 (2009), pp. 69–80]. The lognormal distribution is used in a wide range of applications when the multiplicative scale is appropriate and the log-transformation removes the skew and brings about symmetry of the data distribution [N.T. Longford, Inference with the lognormal distribution, J. Statist. Plann. Inference 139 (2009), pp. 2329–2340]. In this paper, point estimates as well as confidence intervals for the unknown parameters are obtained. This will also be addressed by the Bayesian point of view. To carry out the performance of the estimators obtained, a simulation study is conducted. For illustration proposes, a real data set, due to Lawless [Statistical Models and Methods for Lifetime Data, 2nd ed., John Wiley & Sons, New York, 2003], is analysed using the procedures obtained.  相似文献   

16.
This paper demonstrates that cross-validation (CV) and Bayesian adaptive bandwidth selection can be applied in the estimation of associated kernel discrete functions. This idea is originally proposed by Brewer [A Bayesian model for local smoothing in kernel density estimation, Stat. Comput. 10 (2000), pp. 299–309] to derive variable bandwidths in adaptive kernel density estimation. Our approach considers the adaptive binomial kernel estimator and treats the variable bandwidths as parameters with beta prior distribution. The best variable bandwidth selector is estimated by the posterior mean in the Bayesian sense under squared error loss. Monte Carlo simulations are conducted to examine the performance of the proposed Bayesian adaptive approach in comparison with the performance of the Asymptotic mean integrated squared error estimator and CV technique for selecting a global (fixed) bandwidth proposed in Kokonendji and Senga Kiessé [Discrete associated kernels method and extensions, Stat. Methodol. 8 (2011), pp. 497–516]. The Bayesian adaptive bandwidth estimator performs better than the global bandwidth, in particular for small and moderate sample sizes.  相似文献   

17.
For the complete sample and the right Type II censored sample, Chen [Joint confidence region for the parameters of Pareto distribution. Metrika 44 (1996), pp. 191–197] proposed the interval estimation of the parameter θ and the joint confidence region of the two parameters of Pareto distribution. This paper proposed two methods to construct the confidence region of the two parameters of the Pareto distribution for the progressive Type II censored sample. A simulation study comparing the performance of the two methods is done and concludes that Method 1 is superior to Method 2 by obtaining a smaller confidence area. The interval estimation of parameter ν is also given under progressive Type II censoring. In addition, the predictive intervals of the future observation and the ratio of the two future consecutive failure times based on the progressive Type II censored sample are also proposed. Finally, one example is given to illustrate all interval estimations in this paper.  相似文献   

18.
The demand for reliable statistics in subpopulations, when only reduced sample sizes are available, has promoted the development of small area estimation methods. In particular, an approach that is now widely used is based on the seminal work by Battese et al. [An error-components model for prediction of county crop areas using survey and satellite data, J. Am. Statist. Assoc. 83 (1988), pp. 28–36] that uses linear mixed models (MM). We investigate alternatives when a linear MM does not hold because, on one side, linearity may not be assumed and/or, on the other, normality of the random effects may not be assumed. In particular, Opsomer et al. [Nonparametric small area estimation using penalized spline regression, J. R. Statist. Soc. Ser. B 70 (2008), pp. 265–283] propose an estimator that extends the linear MM approach to the case in which a linear relationship may not be assumed using penalized splines regression. From a very different perspective, Chambers and Tzavidis [M-quantile models for small area estimation, Biometrika 93 (2006), pp. 255–268] have recently proposed an approach for small-area estimation that is based on M-quantile (MQ) regression. This allows for models robust to outliers and to distributional assumptions on the errors and the area effects. However, when the functional form of the relationship between the qth MQ and the covariates is not linear, it can lead to biased estimates of the small area parameters. Pratesi et al. [Semiparametric M-quantile regression for estimating the proportion of acidic lakes in 8-digit HUCs of the Northeastern US, Environmetrics 19(7) (2008), pp. 687–701] apply an extended version of this approach for the estimation of the small area distribution function using a non-parametric specification of the conditional MQ of the response variable given the covariates [M. Pratesi, M.G. Ranalli, and N. Salvati, Nonparametric m-quantile regression using penalized splines, J. Nonparametric Stat. 21 (2009), pp. 287–304]. We will derive the small area estimator of the mean under this model, together with its mean-squared error estimator and compare its performance to the other estimators via simulations on both real and simulated data.  相似文献   

19.
Epstein [Truncated life tests in the exponential case, Ann. Math. Statist. 25 (1954), pp. 555–564] introduced a hybrid censoring scheme (called Type-I hybrid censoring) and Chen and Bhattacharyya [Exact confidence bounds for an exponential parameter under hybrid censoring, Comm. Statist. Theory Methods 17 (1988), pp. 1857–1870] derived the exact distribution of the maximum-likelihood estimator (MLE) of the mean of a scaled exponential distribution based on a Type-I hybrid censored sample. Childs et al. [Exact likelihood inference based on Type-I and Type-II hybrid censored samples from the exponential distribution, Ann. Inst. Statist. Math. 55 (2003), pp. 319–330] provided an alternate simpler expression for this distribution, and also developed analogous results for another hybrid censoring scheme (called Type-II hybrid censoring). The purpose of this paper is to derive the exact bivariate distribution of the MLE of the parameter vector of a two-parameter exponential model based on hybrid censored samples. The marginal distributions are derived and exact confidence bounds for the parameters are obtained. The results are also used to derive the exact distribution of the MLE of the pth quantile, as well as the corresponding confidence bounds. These exact confidence intervals are then compared with parametric bootstrap confidence intervals in terms of coverage probabilities. Finally, we present some numerical examples to illustrate the methods of inference developed here.  相似文献   

20.
The well-known Wilson and Agresti–Coull confidence intervals for a binomial proportion p are centered around a Bayesian estimator. Using this as a starting point, similarities between frequentist confidence intervals for proportions and Bayesian credible intervals based on low-informative priors are studied using asymptotic expansions. A Bayesian motivation for a large class of frequentist confidence intervals is provided. It is shown that the likelihood ratio interval for p approximates a Bayesian credible interval based on Kerman’s neutral noninformative conjugate prior up to O(n? 1) in the confidence bounds. For the significance level α ? 0.317, the Bayesian interval based on the Jeffreys’ prior is then shown to be a compromise between the likelihood ratio and Wilson intervals. Supplementary materials for this article are available online.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号