首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 93 毫秒
1.
Monte Carlo methods are used to compare the methods of maximum likelihood and least squares to estimate a cumulative distribution function. When the probabilistic model used is correct or nearly correct, the two methods produce similar results with the MLE usually slightly superior When an incorrect model is used, or when the data is contaminated, the least squares technique often gives substantially superior results.  相似文献   

2.
The number of variables in a regression model is often too large and a more parsimonious model may be preferred. Selection strategies (e.g. all-subset selection with various penalties for model complexity, or stepwise procedures) are widely used, but there are few analytical results about their properties. The problems of replication stability, model complexity, selection bias and an over-optimistic estimate of the predictive value of a model are discussed together with several proposals based on resampling methods. The methods are applied to data from a case–control study on atopic dermatitis and a clinical trial to compare two chemotherapy regimes by using a logistic regression and a Cox model. A recent proposal to use shrinkage factors to reduce the bias of parameter estimates caused by model building is extended to parameterwise shrinkage factors and is discussed as a further possibility to illustrate problems of models which are too complex. The results from the resampling approaches favour greater simplicity of the final regression model.  相似文献   

3.
Nowadays, Bayesian methods are routinely used for estimating parameters of item response theory (IRT) models. However, the marginal likelihoods are still rarely used for comparing IRT models due to their complexity and a relatively high dimension of the model parameters. In this paper, we review Monte Carlo (MC) methods developed in the literature in recent years and provide a detailed development of how these methods are applied to the IRT models. In particular, we focus on the “best possible” implementation of these MC methods for the IRT models. These MC methods are used to compute the marginal likelihoods under the one-parameter IRT model with the logistic link (1PL model) and the two-parameter logistic IRT model (2PL model) for a real English Examination dataset. We further use the widely applicable information criterion (WAIC) and deviance information criterion (DIC) to compare the 1PL model and the 2PL model. The 2PL model is favored by all of these three Bayesian model comparison criteria for the English Examination data.  相似文献   

4.
阎军  顾岚 《统计研究》1996,13(2):48-54
The paper presents the application of Threshold Auto-regression ( TAR ) model in dealing with the nonlinearity in economy. With SETAR model, forecasting of 31 China’s macro-economic series is performed and the accuracy of different preprocessing methods are compared. With TARSC model, China’s Adjacent National Income Index and Accumulation rate are used to analyze economic fluctuation, empirical in terpretation of the fluctuation and leading and lagging are also presented. The results of the application are satisfactory.  相似文献   

5.
The present study investigates the performance of fice discrimination methods for data consisting of a mixture of continuous and binary variables. The methods are Fisher’s linear discrimination, logistic discrimination, quadratic discrimination, a kernal model and an independence model. Six-dimensional data, consisting of three binary and three continuous variables, are simulated according to a location model. The results show an almost identical performance for Fisher’s linear discrimination and logistic discrimination. Only in situations with independently distributed variables the independence model does have a reasonable discriminatory ability for the dimensionality considered. If the log likelihood ratio is non-linear ratio is non-linear with respect to its continuous and binary part, the quadratic discrimination method is substantial better than linear and logistic discrimination, followed by the kernel method. A very good performance is obtained when in every situation the better one of linear and quardratic discrimination is used.  相似文献   

6.
Some quality characteristics are well defined when treated as the response variables and their relationships are identified to some independent variables. This relationship is called a profile. The parametric models, such as linear models, may be used to model the profiles. However, due to the complexity of many processes in practical applications, it is inappropriate to model the process using parametric models. In these cases non parametric methods are used to model the processes. One of the most applicable non parametric methods used to model complicated profiles is the wavelet. Many authors considered the use of the wavelet transformation only for monitoring the processes in phase II. The problem of estimating the in-control profile in phase I using wavelet transformation is not deeply addressed. Usually classical estimators are used in phase I to estimate the in-control profiles, even when the wavelet transformation is used. These estimators are suitable if the data do not contain outliers. However, when the outliers exist, these estimators cannot estimate the in-control profile properly. In this research, a robust method of estimating the in-control profiles is proposed, which is insensitive to the presence of outliers and could be applied when the wavelet transformation is used. The proposed estimator is the combination of the robust clustering and the S-estimator. This estimator is compared with the classical estimator of the in-control profile in the presence of outliers. The results from a large simulation study show that using the proposed method, one can estimate the in-control profile precisely when the data are contaminated either locally or globally.  相似文献   

7.
Estimating parameters in a stochastic volatility (SV) model is a challenging task. Among other estimation methods and approaches, efficient simulation methods based on importance sampling have been developed for the Monte Carlo maximum likelihood estimation of univariate SV models. This paper shows that importance sampling methods can be used in a general multivariate SV setting. The sampling methods are computationally efficient. To illustrate the versatility of this approach, three different multivariate stochastic volatility models are estimated for a standard data set. The empirical results are compared to those from earlier studies in the literature. Monte Carlo simulation experiments, based on parameter estimates from the standard data set, are used to show the effectiveness of the importance sampling methods.  相似文献   

8.
Estimating parameters in a stochastic volatility (SV) model is a challenging task. Among other estimation methods and approaches, efficient simulation methods based on importance sampling have been developed for the Monte Carlo maximum likelihood estimation of univariate SV models. This paper shows that importance sampling methods can be used in a general multivariate SV setting. The sampling methods are computationally efficient. To illustrate the versatility of this approach, three different multivariate stochastic volatility models are estimated for a standard data set. The empirical results are compared to those from earlier studies in the literature. Monte Carlo simulation experiments, based on parameter estimates from the standard data set, are used to show the effectiveness of the importance sampling methods.  相似文献   

9.
We examine three media exposure distribution (e.d.) simulation methods. The first is based on the maximum likelihood estimate of an individual's exposure, the second on ‘personal probability’ (Greene 1970) and the third on a dependent Bernoulli trials model (Klotz 1973). The last method uses population exposure probabilities rather than individual exposure probabilities, thereby markedly reducing computation time. Magazine exposure data are used to compare the accuracy and computation times of the simulation methods with a log–linear e.d. model (Danaher 1988b) and the popular Metheringham (1964) model based on the beta–binomial distribution (BBD). The results show that the simulation methods are not as accurate as the log– linear model but are more accurate than Metheringham's model, However, all the simulation methods take less computation time than the log–linear model for schedules with more than six magazines, making them viable competitors for large schedule sizes  相似文献   

10.
Dynamic regression models are widely used because they express and model the behaviour of a system over time. In this article, two dynamic regression models, the distributed lag (DL) model and the autoregressive distributed lag model, are evaluated focusing on their lag lengths. From a classical statistics point of view, there are various methods to determine the number of lags, but none of them are the best in all situations. This is a serious issue since wrong choices will provide bad estimates for the effects of the regressors on the response variable. We present an alternative for the aforementioned problems by considering a Bayesian approach. The posterior distributions of the numbers of lags are derived under an improper prior for the model parameters. The fractional Bayes factor technique [A. O'Hagan, Fractional Bayes factors for model comparison (with discussion), J. R. Statist. Soc. B 57 (1995), pp. 99–138] is used to handle the indeterminacy in the likelihood function caused by the improper prior. The zero-one loss function is used to penalize wrong decisions. A naive method using the specified maximum number of DLs is also presented. The proposed and the naive methods are verified using simulation data. The results are promising for the method we proposed. An illustrative example with a real data set is provided.  相似文献   

11.
Compliance with one specified dosing strategy of assigned treatments is a common problem in randomized drug clinical trials. Recently, there has been much interest in methods used for analysing treatment effects in randomized clinical trials that are subject to non-compliance. In this paper, we estimate and compare treatment effects based on the Grizzle model (GM) (ignorable non-compliance) as the custom model and the generalized Grizzle model (GGM) (non-ignorable non-compliance) as the new model. A real data set based on the treatment of knee osteoarthritis is used to compare these models. The results based on the likelihood ratio statistics and simulation study show the advantage of the proposed model (GGM) over the custom model (GGM).  相似文献   

12.
A fully parametric multistate model is explored for the analysis of animal carcinogenicity experiments in which the time of tumour onset is not known. This model does not require assumptions about tumour lethality or cause of death judgements and can be fitted in the absence of sacrifice data. The model is constructed as a three-state model with simple parametric forms for the transition rates. Maximum likelihood methods are used to estimate the transition rates and different treatment groups are compared using likelihood ratio tests. Selection of an appropriate model and methods to assess the fit of the model are illustrated with data from animal experiments. Comparisons with standard methods are made.  相似文献   

13.
The generalized Pareto distribution (GPD) has been widely used to model exceedances over a threshold. This article generalizes the method of generalized probability weighted moments, and applies this method to estimate the parameters of GPD. The estimator is computationally easy. Some asymptotic results of this method are provided. Two simulations are carried out to investigate the behavior of this method and to compare them with other methods suggested in the literature. The simulation results show that the performance of the proposed method is better than some other methods. Finally, this method is applied to analyze a real-life data.  相似文献   

14.
In this article, we formulate a semiparametric model for counting processes in which the effect of covariates is to transform the time scale for a baseline rate function. We assume an arbitrary dependence structure for the counting process and propose a class of estimating equations for the regression parameters. Asymptotic results for these estimators are derived. In addition, goodness of fit methods for assessing the adequacy of the accelerated rates model are proposed. The finite-sample behavior of the proposed methods is examined in simulation studies, and data from a chronic granulomatous disease study are used to illustrate the methodology.  相似文献   

15.
In clinical studies, pairwise comparisons are frequently performed to examine differences in efficacy between treatments. The statistical methods of pairwise comparisons are available when treatment responses are measured on an ordinal scale. The Wilcoxon–Mann–Whitney test and the latent normal model are popular examples. However, these procedures cannot be used to compare treatments in parallel groups (a two-way design) when overall type I error must be controlled. In this paper, we explore statistical approaches to the pairwise testing of treatments that satisfy the requirements of a two-way layout. The results of our simulation indicate that the latent normal approach is superior to the Wilcoxon–Mann–Whitney test. Clinical examples are used to illustrate our suggested testing methods.  相似文献   

16.
This article describes three approximation methods I used to solve the growth model (Model 1) studied by the National Bureau of Economic Research's nonlinear rational-expectations-modeling group project, the results of which were summarized by Taylor and Uhlig (1990). The methods involve computing exact solutions to models that approximate Model 1 in different ways. The first two methods approximate Model 1 about its nonstochastic steady state. The third method works with a version of the model in which the state space has been discretized. A value function iteration method is used to solve that model.  相似文献   

17.
Cluster analysis is one of the most widely used method in statistical analyses, in which homogeneous subgroups are identified in a heterogeneous population. Due to the existence of the continuous and discrete mixed data in many applications, so far, some ordinary clustering methods such as, hierarchical methods, k-means and model-based methods have been extended for analysis of mixed data. However, in the available model-based clustering methods, by increasing the number of continuous variables, the number of parameters increases and identifying as well as fitting an appropriate model may be difficult. In this paper, to reduce the number of the parameters, for the model-based clustering mixed data of continuous (normal) and nominal data, a set of parsimonious models is introduced. Models in this set are extended, using the general location model approach, for modeling distribution of mixed variables and applying factor analyzer structure for covariance matrices. The ECM algorithm is used for estimating the parameters of these models. In order to show the performance of the proposed models for clustering, results from some simulation studies and analyzing two real data sets are presented.  相似文献   

18.
Bayesian hierarchical models are developed to estimate the frequencies of the alleles at the HLA-C locus in the presence of non-identifiable alleles and possible spatial correlations in a large but sparse, spatially defined database from Papua New Guinea. Bayesian model selection methods are applied to investigate the effects of altitude and language on the genetic diversity of HLA-C alleles. The general model includes fixed altitudinal effects, random language effects and random spatially structured location effects. Conditional autoregressive priors are used to incorporate the geographical structure of the map, and Markov chain Monte Carlo simulation methods are applied for estimation and inference. The results show that HLA-C allele frequencies are explained more by linguistic than altitudinal differences, indicating that genetic diversity at this locus in Papua New Guinea probably tracks population movements and is less influenced by natural selection than is variation at HLA-A and HLA-B.  相似文献   

19.
Detecting excessive similarity in answers on multiple choice exams   总被引:2,自引:0,他引:2  
This paper provides a simple and robust method for detecting cheating. Unlike some methods, non-cheating behaviour and not cheating behaviour is modelled because this requires the fewest assumptions. The main concern is the prevention of false accusations. The model is suitable for screening large classes and the results are simple to interpret. Simulation and the Bonferroni inequality are used to prevent false accusation due to 'data dredging'. The model has received considerable application in practice and has been verified through the adjacent seating method.  相似文献   

20.
We explore the construction of new symplectic numerical integration schemes to be used in Hamiltonian Monte Carlo and study their efficiency. Integration schemes from Blanes et al., and a new scheme are considered as candidates to the commonly used leapfrog method. All integration schemes are tested within the framework of the No-U-Turn sampler (NUTS), both for a logistic regression model and a student t-model. The results show that the leapfrog method is inferior to all the new methods both in terms of asymptotic expected acceptance probability for a model problem and the efficient sample size per computing time for the realistic models.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号