首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 171 毫秒
1.
This paper introduces the Dogit ordered generalized extreme value (DOGEV) model, for handling discrete variables that are ordered and heterogeneous. In particular, the DOGEV model can be applied to questionnaire responses on questions allowing a discrete set of ordered possible responses, where there is a preference for particular responses and possibly multiple modes in the data. The DOGEV model combines a model for choice set generation with the ordered generalized extreme value model. The paper illustrates the model using two empirical examples: a model of inflationary expectations and a model for students' evaluations of teaching.  相似文献   

2.
A general stochastic model for the spread of an epidemic developing in a closed population is introduced. Each model consisting of a discrete-time Markov chain involves a deterministic counterpart represented by an ordinary differential equation. Our framework involves various epidemic models such as a stochastic version of the Kermack and McKendrick model and the SIS epidemic model. We prove the asymptotic consistency of the stochastic model regarding a deterministic model; this means that for a large population both modelings are similar. Moreover, a Central Limit Theorem for the fluctuations of the stochastic modeling regarding the deterministic model is also proved.  相似文献   

3.
Crossover designs are popular in early phases of clinical trials and in bioavailability and bioequivalence studies. Assessment of carryover effects, in addition to the treatment effects, is a critical issue in crossover trails. The observed data from a crossover trial can be incomplete because of potential dropouts. A joint model for analyzing incomplete data from crossover trials is proposed in this article; the model includes a measurement model and an outcome dependent informative model for the dropout process. The informative-dropout model is compared with the ignorable-dropout model as specific cases of the latter are nested subcases of the proposed joint model. Markov chain sampling methods are used for Bayesian analysis of this model. The joint model is used to analyze depression score data from a clinical trial in women with late luteal phase dysphoric disorder. Interestingly, carryover effect is found to have a strong effect in the informative dropout model, but it is less significant when dropout is considered ignorable.  相似文献   

4.
For the analysis of binary data, various deterministic models have been proposed, which are generally simpler to fit and easier to understand than probabilistic models. We claim that corresponding to any deterministic model is an implicit stochastic model in which the deterministic model fits imperfectly, with errors occurring at random. In the context of binary data, we consider a model in which the probability of error depends on the model prediction. We show how to fit this model using a stochastic modification of deterministic optimization schemes.The advantages of fitting the stochastic model explicitly (rather than implicitly, by simply fitting a deterministic model and accepting the occurrence of errors) include quantification of uncertainty in the deterministic model’s parameter estimates, better estimation of the true model error rate, and the ability to check the fit of the model nontrivially. We illustrate this with a simple theoretical example of item response data and with empirical examples from archeology and the psychology of choice.  相似文献   

5.
The Cox proportional hazards model has become the standard model for survival analysis. It is often seen as the null model in that "... explicit excuses are now needed to use different models" (Keiding, Proceedings of the XIXth International Biometric Conference, Cape Town, 1998). However, converging hazards also occur frequently in survival analysis. The Burr model, which may be derived as the marginal from a gamma frailty model, is one commonly used tool to model converging hazards. We outline this approach and introduce a mixed model which extends the Burr model and allows for both proportional and converging hazards. Although a semi-parametric model in its own right, we demonstrate how the mixed model can be derived via a gamma frailty interpretation, suggesting an E-M fitting procedure. We illustrate the modelling techniques using data on survival of hospice patients.  相似文献   

6.
In this article, the normal inverse Gaussian stochastic volatility model of Barndorff-Nielsen is extended. The resulting model has a more flexible lag structure than the original one. In addition, the second-and fourth-order moments, important properties of a volatility model, are derived. The model can be considered either as a generalized autoregressive conditional heteroscedasticity model with nonnormal errors or as a stochastic volatility model with an inverse Gaussian distributed conditional variance. A simulation study is made to investigate the performance of the maximum likelihood estimator of the model. Finally, the model is applied to stock returns and exchange-rate movements. Its fit to two stylized facts and its forecasting performance is compared with two other volatility models.  相似文献   

7.
针对GM(1,1)幂模型灰微分方程与白化方程无法匹配的缺陷,以灰微分方程的重构为基础,建立无偏GM(1,1)幂模型。该方法使得差分方程的参数与其在微分方程中对应的参数具有更好的一致性。将无偏GM(1,1)幂模型应用到旅游客源预测中,实例应用结果显示无偏GM(1,1)幂模型预测精度高于GM(1,1)模型。  相似文献   

8.
We propose a new cure model for survival data with a surviving or cure fraction. The new model is a mixture cure model where the covariate effects on the proportion of cure and the distribution of the failure time of uncured patients are separately modeled. Unlike the existing mixture cure models, the new model allows covariate effects on the failure time distribution of uncured patients to be negligible at time zero and to increase as time goes by. Such a model is particularly useful in some cancer treatments when the treat effect increases gradually from zero, and the existing models usually cannot handle this situation properly. We develop a rank based semiparametric estimation method to obtain the maximum likelihood estimates of the parameters in the model. We compare it with existing models and methods via a simulation study, and apply the model to a breast cancer data set. The numerical studies show that the new model provides a useful addition to the cure model literature.  相似文献   

9.
Summary.  In process characterization the quality of information that is obtained depends directly on the quality of process model. The current quality revolution is now providing a strong stimulus for rethinking and re-evaluating many statistical ideas. Among these are the role of theoretic knowledge and data in statistical inference and some issues in theoretic–empirical modelling. With this concern the paper takes a broad, pragmatic view of statistical inference to include all aspects of model formulation. The estimation of model parameters traditionally assumes that a model has a prespecified known form and takes no account of possible uncertainty regarding model structure. But in practice model structural uncertainty is a fact of life and is likely to be more serious than other sources of uncertainty which have received far more attention. This is true whether the model is specified on subject-matter grounds or when a model is formulated, fitted and checked on the same data set in an iterative interactive way. For that reason novel modelling techniques have been fashioned for reducing model uncertainty. Using available knowledge for theoretic model elaboration the techniques that have been created approximate the exact unknown process model concurrently by accessible theoretic and polynomial empirical functions. The paper examines the effects of uncertainty for hybrid theoretic–empirical models and, for reducing uncertainty, additive and multiplicative methods of model formulation are fashioned. Such modelling techniques have been successfully applied to perfect a steady flow model for an air gauge sensor. Validation of the models elaborated has revealed that the multiplicative modelling approach allows us to attain a satisfactory model with small discrepancy from empirical evidence.  相似文献   

10.
The purpose of this paper is to develop a new linear regression model for count data, namely generalized-Poisson Lindley (GPL) linear model. The GPL linear model is performed by applying generalized linear model to GPL distribution. The model parameters are estimated by the maximum likelihood estimation. We utilize the GPL linear model to fit two real data sets and compare it with the Poisson, negative binomial (NB) and Poisson-weighted exponential (P-WE) models for count data. It is found that the GPL linear model can fit over-dispersed count data, and it shows the highest log-likelihood, the smallest AIC and BIC values. As a consequence, the linear regression model from the GPL distribution is a valuable alternative model to the Poisson, NB, and P-WE models.  相似文献   

11.
Social network data represent the interactions between a group of social actors. Interactions between colleagues and friendship networks are typical examples of such data.The latent space model for social network data locates each actor in a network in a latent (social) space and models the probability of an interaction between two actors as a function of their locations. The latent position cluster model extends the latent space model to deal with network data in which clusters of actors exist — actor locations are drawn from a finite mixture model, each component of which represents a cluster of actors.A mixture of experts model builds on the structure of a mixture model by taking account of both observations and associated covariates when modeling a heterogeneous population. Herein, a mixture of experts extension of the latent position cluster model is developed. The mixture of experts framework allows covariates to enter the latent position cluster model in a number of ways, yielding different model interpretations.Estimates of the model parameters are derived in a Bayesian framework using a Markov Chain Monte Carlo algorithm. The algorithm is generally computationally expensive — surrogate proposal distributions which shadow the target distributions are derived, reducing the computational burden.The methodology is demonstrated through an illustrative example detailing relationships between a group of lawyers in the USA.  相似文献   

12.
A partially linear model is a semiparametric regression model that consists of parametric and nonparametric regression components in an additive form. In this article, we propose a partially linear model using a Gaussian process regression approach and consider statistical inference of the proposed model. Based on the proposed model, the estimation procedure is described by posterior distributions of the unknown parameters and model comparisons between parametric representation and semi- and nonparametric representation are explored. Empirical analysis of the proposed model is performed with synthetic data and real data applications.  相似文献   

13.
We consider the problem of selecting a regression model from a large class of possible models in the case where no true model is believed to exist. In practice few statisticians, or scientists who employ statistical methods, believe that a "true" model exists, but nonetheless they seek to select a model as a proxy from which they want to predict. Unlike much of the recent work in this area we address this problem explicitly. We develop Bayesian predictive model selection techniques when proper conjugate priors are used and obtain an easily computed expression for the model selection criterion. We also derive expressions for updating the value of the statistic when a predictor is dropped from the model and apply this approach to a large well-known data set.  相似文献   

14.
Model selection methods are important to identify the best approximating model. To identify the best meaningful model, purpose of the model should be clearly pre-stated. The focus of this paper is model selection when the modelling purpose is classification. We propose a new model selection approach designed for logistic regression model selection where main modelling purpose is classification. The method is based on the distance between the two clustering trees. We also question and evaluate the performances of conventional model selection methods based on information theory concepts in determining best logistic regression classifier. An extensive simulation study is used to assess the finite sample performances of the cluster tree based and the information theoretic model selection methods. Simulations are adjusted for whether the true model is in the candidate set or not. Results show that the new approach is highly promising. Finally, they are applied to a real data set to select a binary model as a means of classifying the subjects with respect to their risk of breast cancer.  相似文献   

15.
A statistical model assuming a preferential attachment network, which is generated by adding nodes sequentially according to a few simple rules, usually describes real-life networks better than a model assuming, for example, a Bernoulli random graph, in which any two nodes have the same probability of being connected, does. Therefore, to study the propagation of “infection” across a social network, we propose a network epidemic model by combining a stochastic epidemic model and a preferential attachment model. A simulation study based on the subsequent Markov Chain Monte Carlo algorithm reveals an identifiability issue with the model parameters. Finally, the network epidemic model is applied to a set of online commissioning data.  相似文献   

16.
The two-part model and Heckman's sample selection model are often used in economic studies which involve analyzing the demand for limited variables. This study proposed a simultaneous equation model (SEM) and used the expectation-maximization algorithm to obtain the maximum likelihood estimate. We then constructed a simulation to compare the performance of estimates of price elasticity using SEM with those estimates from the two-part model and the sample selection model. The simulation shows that the estimates of price elasticity by SEM are more precise than those by the sample selection model and the two-part model when the model includes limited independent variables. Finally, we analyzed a real example of cigarette consumption as an application. We found an increase in cigarette price associated with a decrease in both the propensity to consume cigarettes and the amount actually consumed.  相似文献   

17.
We wish to model pulse wave velocity (PWV) as a function of longitudinal measurements of pulse pressure (PP) at the same and prior visits at which the PWV is measured. A number of approaches are compared. First, we use the PP at the same visit as the PWV in a linear regression model. In addition, we use the average of all available PPs as the explanatory variable in a linear regression model. Next, a two-stage process is applied. The longitudinal PP is modeled using a linear mixed-effects model. This modeled PP is used in the regression model to describe PWV. An approach for using the longitudinal PP data is to obtain a measure of the cumulative burden, the area under the PP curve. This area under the curve is used as an explanatory variable to model PWV. Finally, a joint Bayesian model is constructed similar to the two-stage model.  相似文献   

18.
The number of variables in a regression model is often too large and a more parsimonious model may be preferred. Selection strategies (e.g. all-subset selection with various penalties for model complexity, or stepwise procedures) are widely used, but there are few analytical results about their properties. The problems of replication stability, model complexity, selection bias and an over-optimistic estimate of the predictive value of a model are discussed together with several proposals based on resampling methods. The methods are applied to data from a case–control study on atopic dermatitis and a clinical trial to compare two chemotherapy regimes by using a logistic regression and a Cox model. A recent proposal to use shrinkage factors to reduce the bias of parameter estimates caused by model building is extended to parameterwise shrinkage factors and is discussed as a further possibility to illustrate problems of models which are too complex. The results from the resampling approaches favour greater simplicity of the final regression model.  相似文献   

19.
A combined simple linear and Haar-wavelet regression model is the combination of a simple linear model and a Haar-wavelet regression model. In this article we show how to construct D-optimal designs for a combined simple linear and Haar-wavelet regression model. An example is also given.  相似文献   

20.
Statistical space–time modelling has traditionally been concerned with separable covariance functions, meaning that the covariance function is a product of a purely temporal function and a purely spatial function. We draw attention to a physical dispersion model which could model phenomena such as the spread of an air pollutant. We show that this model has a non-separable covariance function. The model is well suited to a wide range of realistic problems which will be poorly fitted by separable models. The model operates successively in time: the spatial field at time t +1 is obtained by 'blurring' the field at time t and adding a spatial random field. The model is first introduced at discrete time steps, and the limit is taken as the length of the time steps goes to 0. This gives a consistent continuous model with parameters that are interpretable in continuous space and independent of sampling intervals. Under certain conditions the blurring must be a Gaussian smoothing kernel. We also show that the model is generated by a stochastic differential equation which has been studied by several researchers previously.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号