首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Structural equation modeling (SEM) typically utilizes first- and second-order moment structures. This limits its applicability since many unidentified models and many equivalent models that researchers would like to distinguish are created. In this paper, we relax this restriction and assume non-normal distributions on exogenous variables. We shall provide a solution to the problems of underidentifiability and equivalence of SEM models by making use of non-normality (higher-order moment structures). The non-normal SEM is applied to finding the possible direction of a path in simple regression models. The method of (generalized) least squares is employed to estimate model parameters. A test statistic for examining a fit of a model is proposed. A simulation result and a real data example are reported to study how the non-normal SEM approach works empirically.  相似文献   

2.
Results of a simulation study of the fit of data to an estimated parametric model are reported. Three particular models including the two-parameter normal and exponential distributions, and the simple linear regression model are considered. A number of scaled versions of the least squares residuals from the regression model and quantities that we call residuals from the other two models arc seen follow the parent distribution form loo well. i.e., to be supernormal and superexponential. A point of particular interest is that this tendency does not appear to decrease with increasing sample size, at least for the sample sizes considered here.  相似文献   

3.
Abstract

This paper investigates the statistical analysis of grouped accelerated temperature cycling test data when the product lifetime follows a Weibull distribution. A log-linear acceleration equation is derived from the Coffin-Manson model. The problem is transformed to a constant-stress accelerated life test with grouped data and multiple acceleration variables. The Jeffreys prior and reference priors are derived. Maximum likelihood estimation and Bayesian estimation with objective priors are obtained by applying the technique of data augmentation. A simulation study shows that both of these two methods perform well when sample size is large, and the Bayesian method gives better performance under small sample sizes.  相似文献   

4.
In longitudinal clinical trials, a common objective is to compare the rates of changes in an outcome variable between two treatment groups. Generalized estimating equation (GEE) has been widely used to examine if the rates of changes are significantly different between treatment groups due to its robustness to misspecification of the true correlation structure and randomly missing data. The sample size formula for repeated outcomes is based on the assumption of missing completely at random and a large sample approximation. A simulation study is conducted to investigate the performance of GEE sample size formula with small sample sizes, damped exponential family of correlation structure and non‐ignorable missing data. Copyright © 2008 John Wiley & Sons, Ltd.  相似文献   

5.
Multilevel latent class analysis is conducive to providing more effective information on both individual and group typologies. However, model selection issues still need further investigation. Current study probed into issue of high-level class numeration for a more complex model using AIC, AIC3, BIC, and BIC*. Data simulation was conducted and its result was verified by empirical data. The result demonstrated that these criteria have a certain inclination relative to sample sizes. Sample size per group plays an evident role in improving accuracy of AIC3 and BIC. The complex model requires more sample size per group to ensure accurate class numeration.  相似文献   

6.
In this paper, we suggest a technique to quantify model risk, particularly model misspecification for binary response regression problems found in financial risk management, such as in credit risk modelling. We choose the probability of default model as one instance of many other credit risk models that may be misspecified in a financial institution. By way of illustrating the model misspecification for probability of default, we carry out quantification of two specific statistical predictive response techniques, namely the binary logistic regression and complementary log–log. The maximum likelihood estimation technique is employed for parameter estimation. The statistical inference, precisely the goodness of fit and model performance measurements, are assessed. Using the simulation dataset and Taiwan credit card default dataset, our finding reveals that with the same sample size and very small simulation iterations, the two techniques produce similar goodness-of-fit results but completely different performance measures. However, when the iterations increase, the binary logistic regression technique for balanced dataset reveals prominent goodness of fit and performance measures as opposed to the complementary log–log technique for both simulated and real datasets.  相似文献   

7.
We consider in this article the problem of numerically approximating the quantiles of a sample statistic for a given population, a problem of interest in many applications, such as bootstrap confidence intervals. The proposed Monte Carlo method can be routinely applied to handle complex problems that lack analytical results. Furthermore, the method yields estimates of the quantiles of a sample statistic of any sample size though Monte Carlo simulations for only two optimally selected sample sizes are needed. An analysis of the Monte Carlo design is performed to obtain the optimal choices of these two sample sizes and the number of simulated samples required for each sample size. Theoretical results are presented for the bias and variance of the numerical method proposed. The results developed are illustrated via simulation studies for the classical problem of estimating a bivariate linear structural relationship. It is seen that the size of the simulated samples used in the Monte Carlo method does not have to be very large and the method provides a better approximation to quantiles than those based on an asymptotic normal theory for skewed sampling distributions.  相似文献   

8.
A method based on pseudo-observations has been proposed for direct regression modeling of functionals of interest with right-censored data, including the survival function, the restricted mean and the cumulative incidence function in competing risks. The models, once the pseudo-observations have been computed, can be fitted using standard generalized estimating equation software. Regression models can however yield problematic results if the number of covariates is large in relation to the number of events observed. Guidelines of events per variable are often used in practice. These rules of thumb for the number of events per variable have primarily been established based on simulation studies for the logistic regression model and Cox regression model. In this paper we conduct a simulation study to examine the small sample behavior of the pseudo-observation method to estimate risk differences and relative risks for right-censored data. We investigate how coverage probabilities and relative bias of the pseudo-observation estimator interact with sample size, number of variables and average number of events per variable.  相似文献   

9.
In this article, we propose some tests of fit based on sample entropy for the composite Gumbel (Extreme Value) hypothesis. The proposed test statistics are constructed using different entropy estimates. Through a Monte Carlo simulation, critical values of the test statistics for various sample sizes are obtained. Since the tests based on the empirical distribution function (EDF) are commonly used in practice, the power values of the entropy-based tests with those of the EDF tests are compared against various alternatives and different sample sizes. Finally, two real data sets are modeled by the Gumbel distribution.KEYWORDS: Entropy estimator, Gumbel distribution, Monte Carlo simulation, test power  相似文献   

10.
For the linear hypothesis in a strucural equation model, the properties of test statistics based on the two stage least squares estimator (2SLSE) have been examined since these test statistics are easily derived in the instrumental variable estimation framework. Savin (1976) has shown that inequalities exist among the test statistics for the linear hypothesis, but it is well known that there is no systematic inequality among these statistics based on 2SLSE for the linear hypothesis in a structural equation model. Morimune and Oya (1994) derived the constrained limited information maximum likelihood estimator (LIMLE) subject to general linear constraints on the coefficients of the structural equation, as well as Wald, LM and Lr Test statistics for the adequacy of the linear constraints.

In this paper, we derive the inequalities among these three test statistics based on LIMLE and the local power functions based on Limle and 2SLSE to show that there is no test statistic which is uniformly most powerful, and the LR test statistic based on LIMLE is locally unbised and the other test statistics are not. Monte Carlo simulations are used to examine the actual sizes of these test statistics and some numerical examples of the power differences among these test statistics are given. It is found that the actual sizes of these test statistics are greater than the nominal sizes, the differences between the actual and nominal sizes of Wald test statistics are generally the greatest, those of LM test statistics are the smallest, and the power functions depend on the correlations between the endogenous explanatory variables and the error term of the structural equation, the asymptotic variance of estimator of coefficients of the structural equation and the number of restrictions imposed on the coefficients.  相似文献   

11.
Despite tremendous effort on different designs with cross-sectional data, little research has been conducted for sample size calculation and power analyses under repeated measures design. In addition to time-averaged difference, changes in mean response over time (CIMROT) is the primary interest in repeated measures analysis. We generalized sample size calculation and power analysis equations for CIMROT to allow unequal sample size between groups for both continuous and binary measures, through simulation, evaluated the performance of proposed methods, and compared our approach to that of a two-stage model formulization. We also created a software procedure to implement the proposed methods.  相似文献   

12.
For the linear hypothesis in a strucural equation model, the properties of test statistics based on the two stage least squares estimator (2SLSE) have been examined since these test statistics are easily derived in the instrumental variable estimation framework. Savin (1976) has shown that inequalities exist among the test statistics for the linear hypothesis, but it is well known that there is no systematic inequality among these statistics based on 2SLSE for the linear hypothesis in a structural equation model. Morimune and Oya (1994) derived the constrained limited information maximum likelihood estimator (LIMLE) subject to general linear constraints on the coefficients of the structural equation, as well as Wald, LM and Lr Test statistics for the adequacy of the linear constraints.

In this paper, we derive the inequalities among these three test statistics based on LIMLE and the local power functions based on Limle and 2SLSE to show that there is no test statistic which is uniformly most powerful, and the LR test statistic based on LIMLE is locally unbised and the other test statistics are not. Monte Carlo simulations are used to examine the actual sizes of these test statistics and some numerical examples of the power differences among these test statistics are given. It is found that the actual sizes of these test statistics are greater than the nominal sizes, the differences between the actual and nominal sizes of Wald test statistics are generally the greatest, those of LM test statistics are the smallest, and the power functions depend on the correlations between the endogenous explanatory variables and the error term of the structural equation, the asymptotic variance of estimator of coefficients of the structural equation and the number of restrictions imposed on the coefficients.  相似文献   

13.
Alternating logistic regressions (ALRs) seem to offer some of the advantages of marginal models estimated via generalized estimating equations (GEE) and generalized linear mixed models (GLMMs). Via simulation study we compared ALRs to marginal models estimated via GEE and subject-specific models estimated via GLMMs, with a focus on estimation of the correlation structure in three-level data sets (e.g. students in classes in schools). Data set size and structure, and amount of correlation in the data sets were varied. For simple correlation structures, ALRs performed well. For three-level correlation structures, all approaches, but especially ALRs, had difficulty assigning the correlation to the correct level, though sample sizes used were small. In addition, ALRs and GEEs had trouble attaching correct inference to the mean effects, though this improved as overall sample size improved. ALRs are a valuable addition to the data analyst's toolkit, though care should be taken when modelling data with three-level structures.  相似文献   

14.
Summary.  A Bayesian non-parametric methodology has been recently proposed to deal with the issue of prediction within species sampling problems. Such problems concern the evaluation, conditional on a sample of size n , of the species variety featured by an additional sample of size m . Genomic applications pose the additional challenge of having to deal with large values of both n and m . In such a case the computation of the Bayesian non-parametric estimators is cumbersome and prevents their implementation. We focus on the two-parameter Poisson–Dirichlet model and provide completely explicit expressions for the corresponding estimators, which can be easily evaluated for any sizes of n and m . We also study the asymptotic behaviour of the number of new species conditionally on the observed sample: such an asymptotic result, combined with a suitable simulation scheme, allows us to derive asymptotic highest posterior density intervals for the estimates of interest. Finally, we illustrate the implementation of the proposed methodology by the analysis of five expressed sequence tags data sets.  相似文献   

15.
We consider optimal sample designs for observing classes of objects. Suppose we will take a simple random sample of equal-sized sectors from a study population and observe the classes existing on these sectors. The classes might be many different things, for example, herbaceous plant species (in sampling quadrats), microinvertebrate species (in sampling cores), and side effects from a drug (in conducting medical trials). Under a nonparametric mixture model and data from a previous related study, we first estimate the optimal number of sample sectors of a given size. Then for negative binomial dispersions of individuals with a common aggregation parameter k, we consider the optimal size as well as number of sample sectors. A simple test exists to check our common k assumption and our optimal size method requires far less data than would be required by a grid method or other method which utilizes data from sample sectors of several different sizes.  相似文献   

16.
The model chi-square that is used in linear structural equation modeling compares the fitted covariance matrix of a target model to an unstructured covariance matrix to assess global fit. For models with nonlinear terms, i.e., interaction or quadratic terms, this comparison is very problematic because these models are not nested within the saturated model that is represented by the unstructured covariance matrix. We propose a novel measure that quantifies the heteroscedasticity of residuals in structural equation models. It is based on a comparison of the likelihood for the residuals under the assumption of heteroscedasticity with the likelihood under the assumption of homoscedasticity. The measure is designed to respond to omitted nonlinear terms in the structural part of the model that result in heteroscedastic residual scores. In a small Monte Carlo study, we demonstrate that the measure appears to detect omitted nonlinear terms reliably when falsely a linear model is analyzed and the omitted nonlinear terms account for substantial nonlinear effects. The results also indicate that the measure did not respond when the correct model or an overparameterized model were used.  相似文献   

17.
In this paper we propose a sequential procedure to design optimum experiments for discriminating between two binary data models. For the problem to be fully specified, not only the mode1link functions should be provided but also their associated linear predictor structures. Further, we suppose that one of the models is true, albeit it is not known which of them. Under these assumptions the procedure consists of making sequential choices of single experimental units to discriminate between the rival models as efficiently as possible. Depending on whether the models are nested or not, alternative methods are proposed.

To illustrate the procedure, a simulation study for the classical case of pro bit versus logit model is presented. It enables us to estimate the total sample sizes required to gain a certain power of discrimination and compare them to sample sizes for methods that were previously suggested in the literature.  相似文献   

18.
A Monte Carlo study was used to examine the Type I error and power values of five multivariate tests for the single-factor repeated measures model The performance of Hotelling's T2 and four nonparametric tests, including a chi-square and an F-test version of a rank-transform procedure, were investigated for different distributions, sample sizes, and numbers of repeated measures. The results indicated that both Hotellings T* and the F-test version of the rank-transform performed well, producing Type I error rates which were close to the nominal value. The chi-square version of the rank-transform test, on the other hand, produced inflated Type I error rates for every condition studied. The Hotelling and F-test version of the rank-transform procedure showed similar power for moderately-skewed distributions, but for strongly skewed distributions the F-test showed much better power. The performance of the other nonparametric tests depended heavily on sample size. Based on these results, the F-test version of the rank-transform procedure is recommended for the single-factor repeated measures model.  相似文献   

19.
We consider tied survival data based on Cox proportional regression model. The standard approaches are the Breslow and Efron approximations and various so called exact methods. All these methods lead to biased estimates when the true underlying model is in fact a Cox model. In this paper we review the methods and suggest a new method based on the missing-data principle using EM-algorithm that leads to a score equation that can be solved directly. This score has mean zero. We also show that all the considered methods have the same asymptotic properties and that there is no loss of asymptotic efficiency when the tie sizes are bounded or even converge to infinity at a given rate. A simulation study is conducted to compare the finite sample properties of the methods.  相似文献   

20.
The extreme value distribution has been extensively used to model natural phenomena such as rainfall and floods, and also in modeling lifetimes and material strengths. Maximum likelihood estimation (MLE) for the parameters of the extreme value distribution leads to likelihood equations that have to be solved numerically, even when the complete sample is available. In this paper, we discuss point and interval estimation based on progressively Type-II censored samples. Through an approximation in the likelihood equations, we obtain explicit estimators which are approximations to the MLEs. Using these approximate estimators as starting values, we obtain the MLEs using an iterative method and examine numerically their bias and mean squared error. The approximate estimators compare quite favorably to the MLEs in terms of both bias and efficiency. Results of the simulation study, however, show that the probability coverages of the pivotal quantities (for location and scale parameters) based on asymptotic normality are unsatisfactory for both these estimators and particularly so when the effective sample size is small. We, therefore, suggest the use of unconditional simulated percentage points of these pivotal quantities for the construction of confidence intervals. The results are presented for a wide range of sample sizes and different progressive censoring schemes. We conclude with an illustrative example.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号