首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 78 毫秒
1.
In this article, we present the EM-algorithm for performing maximum likelihood estimation of an asymmetric linear calibration model with the assumption of skew-normally distributed error. A simulation study is conducted for evaluating the performance of the calibration estimator with interpolation and extrapolation situations. As one application in a real data set, we fitted the model studied in a dimensional measurement method used for calculating the testicular volume through a caliper and its calibration by using ultrasonography as the standard method. By applying this methodology, we do not need to transform the variables to have symmetrical errors. Another interesting aspect of the approach is that the developed transformation to make the information matrix nonsingular, when the skewness parameter is near zero, leaves the parameter of interest unchanged. Model fitting is implemented and the best choice between the usual calibration model and the model proposed in this article was evaluated by developing the Akaike information criterion, Schwarz’s Bayesian information criterion and Hannan–Quinn criterion.  相似文献   

2.
Measurement error, the difference between a measured (observed) value of quantity and its true value, is perceived as a possible source of estimation bias in many surveys. To correct for such bias, a validation sample can be used in addition to the original sample for adjustment of measurement error. Depending on the type of validation sample, we can either use the internal calibration approach or the external calibration approach. Motivated by Korean Longitudinal Study of Aging (KLoSA), we propose a novel application of fractional imputation to correct for measurement error in the analysis of survey data. The proposed method is to create imputed values of the unobserved true variables, which are mis-measured in the main study, by using validation subsample. Furthermore, the proposed method can be directly applicable when the measurement error model is a mixture distribution. Variance estimation using Taylor linearization is developed. Results from a limited simulation study are also presented.  相似文献   

3.
In this paper, me shall investigate a bootstrap method hasd on a martingale representation of the relevant statistic for inference to a class of functionals of the survival distribution. The method is similar in spirit to Efron's (1981) bootstrap, and thus in the present paper will be referred to as “martingale-based bootstrap” The method was derived from Lin,Wei and Ying (1993), who appiied the method in checking the Cox model with cumulative sums of martingale-based residuals. It is shown that this martingale-based bootstrap gives a correct first-order asymptotic approximation to the distribution function of the corresponding functional of the Kaplan-Meier estimator. As a consequence, confidence intervals constructed by the martingale-based bootstrap have asymptotially correct coverage probability. Our simulation study indicats that the martingale-based bootst strap method for a small and moderate sample sizes can be uniformly better than the usual bootstrap method in estimating the sampling distribution for a mean function and a point probability in survival analysis.  相似文献   

4.
Bayesian calibration of computer models   总被引:5,自引:0,他引:5  
We consider prediction and uncertainty analysis for systems which are approximated using complex mathematical models. Such models, implemented as computer codes, are often generic in the sense that by a suitable choice of some of the model's input parameters the code can be used to predict the behaviour of the system in a variety of specific applications. However, in any specific application the values of necessary parameters may be unknown. In this case, physical observations of the system in the specific context are used to learn about the unknown parameters. The process of fitting the model to the observed data by adjusting the parameters is known as calibration. Calibration is typically effected by ad hoc fitting, and after calibration the model is used, with the fitted input values, to predict the future behaviour of the system. We present a Bayesian calibration technique which improves on this traditional approach in two respects. First, the predictions allow for all sources of uncertainty, including the remaining uncertainty over the fitted parameters. Second, they attempt to correct for any inadequacy of the model which is revealed by a discrepancy between the observed data and the model predictions from even the best-fitting parameter values. The method is illustrated by using data from a nuclear radiation release at Tomsk, and from a more complex simulated nuclear accident exercise.  相似文献   

5.
Counting by weighing is widely used in industry and often more efficient than counting manually which is time consuming and prone to human errors especially when the number of items is large. Lower confidence bounds on the numbers of items in infinitely many future bags based on the weights of the bags have been proposed recently in Liu et al. [Counting by weighing: Know your numbers with confidence, J. Roy. Statist. Soc. Ser. C 65(4) (2016), pp. 641–648]. These confidence bounds are constructed using the data from one calibration experiment and for different parameters (or numbers), but have the frequency interpretation similar to a usual confidence set for one parameter only. In this paper, the more challenging problem of constructing two-sided confidence intervals is studied. A simulation-based method for computing the critical constant is proposed. This method is proven to give the required critical constant when the number of simulations goes to infinity, and shown to be easily implemented on an ordinary computer to compute the critical constant accurately and quickly. The methodology is illustrated with a real data example.  相似文献   

6.
In the field of education, it is often of great interest to estimate the percentage of students who start out in the top test quantile at time 1 and who remain there at time 2, which is termed as “persistence rate,” to measure the students’ academic growth. One common difficulty is that students’ performance may be subject to measurement errors. We therefore considered a correlation calibration method and the simulation–extrapolation (SIMEX) method for correcting the measurement errors. Simulation studies are presented to compare various measurement error correction methods in estimating the persistence rate.  相似文献   

7.
Abstract

In the present article, an effort has been made to develop calibration estimators of the population mean under two-stage stratified random sampling design when auxiliary information is available at primary stage unit (psu) level. The properties of the developed estimators are derived in-terms of design based approximate variance and approximate consistent design based estimator of the variance. Some simulation studies have been conducted to investigate the relative performance of calibration estimator over the usual estimator of the population mean without using auxiliary information in two-stage stratified random sampling. Proposed calibration estimators have outperformed the usual estimator without using auxiliary information.  相似文献   

8.
In mixture experiments the properties of mixtures are usually studied by mixing the amounts of the mixture components that are required to obtain the necessary proportions. This paper considers the impact of inaccuracies in discharging the required amounts of the mixture components on the statistical analysis of the data. It shows how the regression calibration approach can be used to minimize the resulting bias in the model and in the estimates of the model parameters, as well as to find correct estimates of the corresponding variances. Its application is made difficult by the complex structure of these errors. We also show how knowledge of the form of the model bias allows for choosing a manufacturing setting for a mixture product that is not biased and has smaller signal to noise ratio.  相似文献   

9.
The usual (global) breakdown point describes the worst effect that a given number of gross errors can have. In a two-way layout, without interaction, one is frustrated by the small number of gross errors such a design can tolerate. However, neither the whole fit nor all parameter estimates need to be affected by such a breakdown. An example from molecular spectroscopy serves to illustrate such partial breakdown in a large, “sparse” two-factor model. Because the global finite sample breakdown point is zero for all usual estimators in this example, this concept does not make sense in such problems. The more appropriate concept of partial breakdown point is discussed in this paper. It also provides a crude quantification of the robustness properties of an estimator, yet for any linear combination of the estimated parameters. The maximum number of gross errors to which the linear combination of the estimated parameters can resist is related to the minimum number of observations that must be omitted to make the linear function a non-estimable function. In the example, we are mainly interested in differences of parameters. Then the maximal partial breakdown point for regression equivariant estimators is one half, and Huber-type regression M-estimators with bounded ψ-function reach this limit.  相似文献   

10.
Many of the existing methods of finding calibration intervals in simple linear regression rely on the inversion of prediction limits. In this article, we propose an alternative procedure which involves two stages. In the first stage, we find a confidence interval for the value of the explanatory variable which corresponds to the given future value of the response. In the second stage, we enlarge the confidence interval found in the first stage to form a confidence interval called, calibration interval, for the value of the explanatory variable which corresponds to the theoretical mean value of the future observation. In finding the confidence interval in the first stage, we have used the method based on hypothesis testing and percentile bootstrap. When the errors are normally distributed, the coverage probability of resulting calibration interval based on hypothesis testing is comparable to that of the classical calibration interval. In the case of non normal errors, the coverage probability of the calibration interval based on hypothesis testing is much closer to the target value than that of the calibration interval based on percentile bootstrap.  相似文献   

11.
In a linear model with missing observations, one can substitute algebraic quantities and then minimize the error sum of squares for the augmented model. This gives the correct error sum of squares. But this method does not produce the correct hypothesis sum of squares for testing a linear hypothesis about the parameters. The sum of squares obtained is biased but practitioners still use it. The distribution of this biased sum of squares is derived in this paper and the consequences of using this biased sum of squares on the type I and II errors is examined.  相似文献   

12.
The ranked set sampling (RSS) method as suggested by McIntyre (1952) may be modified to come up with new sampling methods that can be made more efficient than the usual RSS method. Two such modifications, namely extreme and median ranked set sampling methods, are considered in this study. These two methods are generally easier to use in the field and less prone to problems resulting from errors in ranking. Two regression-type estimators based on extreme ranked set sampling (ERSS) and median ranked set sampling (MRSS) for estimating the population mean of the variable of interest are considered in this study and compared with the regression-type estimators based on RSS suggested by Yu & Lam (1997). It turned out that when the variable of interest and the concomitant variable jointly followed a bivariate normal distribution, the regression-type estimator of the population mean based on ERSS dominates all other estimators considered.  相似文献   

13.
The problem of error estimation of parameters b in a linear model,Y = Xb+ e, is considered when the elements of the design matrix X are functions of an unknown ‘design’ parameter vector c. An estimated value c is substituted in X to obtain a derived design matrix [Xtilde]. Even though the usual linear model conditions are not satisfied with [Xtilde], there are situations in physical applications where the least squares solution to the parameters is used without concern for the magnitude of the resulting error. Such a solution can suffer from serious errors.

This paper examines bias and covariance errors of such estimators. Using a first-order Taylor series expansion, we derive approximations to the bias and covariance matrix of the estimated parameters. The bias approximation is a sum of two terms:One is due to the dependence between ? and Y; the other is due to the estimation errors of ? and is proportional to b, the parameter being estimated. The covariance matrix approximation, on the other hand, is composed of three omponents:One component is due to the dependence between ? and Y; the second is the covariance matrix ∑b corresponding to the minimum variance unbiased b, as if the design parameters were known without error; and the third is an additional component due to the errors in the design parameters. It is shown that the third error component is directly proportional to bb'. Thus, estimation of large parameters with wrong design matrix [Xtilde] will have larger errors of estimation. The results are illustrated with a simple linear example.  相似文献   

14.
The use of Monte Carlo methods to generate exam datasets is nowadays a well-established practice among econometrics and statistics examiners all over the world. Its advantages are well known: providing each student a different data set ensures that estimates are actually computed individually, rather than copied from someone sitting nearby. The method however has a major fault: initial “random errors,” such as mistakes in downloading the assigned dataset, might generate downward bias in student evaluation. We propose a set of calibration algorithms, typical of indirect estimation methods, that solve the issue of initial “random errors” and reduce evaluation bias. Ensuring round initial estimates of the parameters for each individual dataset, our calibration procedures allow the students to determine if they have started the exam correctly. When initial estimates are not round numbers, this random error in the initial stage of the exam can be corrected for immediately, thus reducing evaluation bias. The procedure offers the further advantage of rounding markers’ life by allowing them to check round numbers answers only, rather than lists of numbers with many decimal digits1.  相似文献   

15.
ABSTRACT

The broken-stick (BS) is a popular stopping rule in ecology to determine the number of meaningful components of principal component analysis. However, its properties have not been systematically investigated. The purpose of the current study is to evaluate its ability to detect the correct dimensionality in a data set and whether it tends to over- or underestimate it. A Monte Carlo protocol was carried out. Two main correlation matrices deemed usual in practice were used with three levels of correlation (0, 0.10 and 0.30) between components (generating oblique structure) and with different sample sizes. Analyses of the population correlation matrices indicated that, for extremely large sample sizes, the BS method could be correct for only one of the six simulated structure. It actually failed to identify the correct dimensionality half the time with orthogonal structures and did even worse with some oblique ones. In harder conditions, results show that the power of the BS decreases as sample size increases: weakening its usefulness in practice. Since the BS method seems unlikely to identify the underlying dimensionality of the data, and given that better stopping rules exist it appears as a poor choice when carrying principal component analysis.  相似文献   

16.
Physical activity measurements derived from self-report surveys are prone to measurement errors. Monitoring devices like accelerometers offer more objective measurements of physical activity, but are impractical for use in large-scale surveys. A model capable of predicting objective measurements of physical activity from self-reports would offer a practical alternative to obtaining measurements directly from monitoring devices. Using data from National Health and Nutrition Examination Survey 2003–2006, we developed and validated models for predicting objective physical activity from self-report variables and other demographic characteristics. The prediction intervals produced by the models were large, suggesting that the ability to predict objective physical activity for individuals from self-reports is limited.  相似文献   

17.
We describe novel, analytical, data-analysis, and Monte-Carlo-simulation studies of strongly heteroscedastic data of both small and wide range.Many different types of heteroscedasticity and fixed or variable weighting are incorporated through error-variance models.Attention is given to parameter bias determinations, evaluations of their significances, and to new ways to correct for bias.The error-variance models allow for both additive and independent power-law errors, and the power exponent is shown to be able to be well determined for typical physicalsciences data by the rapidly-converging, general-purpose, extended-least-squares program we use.The fitting and error-variance models are applied to both low-and high-heteroscedasticity situations, including single-response data from radioactive decay.Monte-Carlo simulations of data with similar parameters are used to evaluate the analytical models developed and the various minimization methods em-ployed, such as extended and generalized least squares.Logarithmic and inversion transformations are investigated in detail, and it is shown analytically and by simulations that exponential data with constant percentage errors can be logarithmically transformed to allow a simple parameter-bias-removal procedure.A more-general bias-reduction approach combining direct and inversion fitting is also developed.Distributions of fitting-model and error-variance-model parameters are shown to be typically non-normal, thus invalidating the usual estimates of parameter bias and precision.Errors in conventional confidence-interval estimates are quantified by comparison with accurate simulation results.  相似文献   

18.
We propose to use a general mixing distribution in modeling the heterogeneity of the fecundability of couples. We introduce a sequence of parameters called canonical moments, which is in one to one correspondence with the moments, to characterize the mixing distribution. By using the bootstrap method, we can estimate the standard errors of our estimates. Our method modifies the usual moment estimates so that the resulting mixing distribution is always supported on [0, 1]. Moreover, the downward bias of the moment estimate of the number of support points would be reduced. Our approach can be used for censored data. The application of our technique in finding the sterile subpopulation is also discussed. The theory is illustrated with several data examples and simulations.  相似文献   

19.
Inverse probability weighting (IPW) and multiple imputation are two widely adopted approaches dealing with missing data. The former models the selection probability, and the latter models data distribution. Consistent estimation requires correct specification of corresponding models. Although the augmented IPW method provides an extra layer of protection on consistency, it is usually not sufficient in practice as the true data‐generating process is unknown. This paper proposes a method combining the two approaches in the same spirit of calibration in sampling survey literature. Multiple models for both the selection probability and data distribution can be simultaneously accounted for, and the resulting estimator is consistent if any model is correctly specified. The proposed method is within the framework of estimating equations and is general enough to cover regression analysis with missing outcomes and/or missing covariates. Results on both theoretical and numerical investigation are provided.  相似文献   

20.
This paper considers the classical and inverse calibration estimators and discusses the consequences of departure from normality of errors on their bias and mean squared error properties when the errors in calibration process are small.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号