首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 906 毫秒
1.
In this work we study robustness in Bayesian models through a generalization of the Normal distribution. We show new appropriate techniques in order to deal with this distribution in Bayesian inference. Then we propose two approaches to decide, in some applications, if we should replace the usual Normal model by this generalization. First, we pose this dilemma as a model rejection problem, using diagnostic measures. In the second approach we evaluate the model's predictive efficiency. We illustrate those perspectives with a simulation study, a non linear model and a longitudinal data model.  相似文献   

2.
In a calibration of near-infrared (NIR) instrument, we regress some chemical compositions of interest as a function of their NIR spectra. In this process, we have two immediate challenges: first, the number of variables exceeds the number of observations and, second, the multicollinearity between variables are extremely high. To deal with the challenges, prediction models that produce sparse solutions have recently been proposed. The term ‘sparse’ means that some model parameters are zero estimated and the other parameters are estimated naturally away from zero. In effect, a variable selection is embedded in the model to potentially achieve a better prediction. Many studies have investigated sparse solutions for latent variable models, such as partial least squares and principal component regression, and for direct regression models such as ridge regression (RR). However, in the latter, it mainly involves an L1 norm penalty to the objective function such as lasso regression. In this study, we investigate new sparse alternative models for RR within a random effects model framework, where we consider Cauchy and mixture-of-normals distributions on the random effects. The results indicate that the mixture-of-normals model produces a sparse solution with good prediction and better interpretation. We illustrate the methods using NIR spectra datasets from milk and corn specimens.  相似文献   

3.
The classification of a random variable based on a mixture can be meaningfully discussed only if the class of all finite mixtures is identifiable. In this paper, we find the maximum-likelihood estimates of the parameters of the mixture of two inverse Weibull distributions by using classified and unclassified observations. Next, we estimate the nonlinear discriminant function of the underlying model. Also, we calculate the total probabilities of misclassification as well as the percentage bias. In addition, we investigate the performance of all results through a series of simulation experiments by means of relative efficiencies. Finally, we analyse some simulated and real data sets through the findings of the paper.  相似文献   

4.
In this paper we consider from maximum likelihood and Bayesian points of view the generalized growth curve model when the covariance matrix has a Toeplitz structure. This covariance is a generalization of the AR(1) dependence structure. Inferences on the parameters as well as the future values are included. The results are illustrated with several real data sets.  相似文献   

5.
In this paper we deal with a Bayesian analysis for right-censored survival data suitable for populations with a cure rate. We consider a cure rate model based on the negative binomial distribution, encompassing as a special case the promotion time cure model. Bayesian analysis is based on Markov chain Monte Carlo (MCMC) methods. We also present some discussion on model selection and an illustration with a real data set.  相似文献   

6.
Abstract

In choice experiments the process of decision-making can be more complex than the proposed by the Multinomial Logit Model (MNL). In these scenarios, models such as the Nested Multinomial Logit Model (NMNL) are often employed to model a more complex decision-making. Understanding the decision-making process is important in some fields such as marketing. Achieving a precise estimation of the models is crucial to the understanding of this process. To do this, optimal experimental designs are required. To construct an optimal design, information matrix is key. A previous research by others has developed the expression for the information matrix of the two-level NMNL model with two nests: Alternatives nest (J alternatives) and No-Choice nest (1 alternative). In this paper, we developed the likelihood function for a two-stage NMNL model for M nests and we present the expression for the information matrix for 2 nests with any amount of alternatives in them. We also show alternative D-optimal designs for No-Choice scenarios with similar relative efficiency but with less complex alternatives which can help to obtain more reliable answers and one application of these designs.  相似文献   

7.
Abstract.  In a recent paper we extended and refined some tools introduced by O'Hagan for criticism of Bayesian hierarchical models. Especially, avoiding double use of data by a data-splitting approach was a main concern. Such tools can be applied at each node of the model, with a view to diagnosing problems of model fit at any point in the model structure. As O'Hagan, we investigated a Gaussian model of one-way analysis of variance. Through extensive Markov chain Monte Carlo simulations it was shown that our method detects model misspecification about as well as the one of O'Hagan, when this is properly calibrated, while retaining the desired false warning probability for data generated from the assumed model. In the present paper, we suggest some new measures of conflict based on tail probabilities of the so-called integrated posterior distributions introduced in our recent paper. These new measures are equivalent to the measure applied in the latter paper in simple Gaussian models, but seem more appropriately adjusted to deviations from normality and to conflicts not concerning location parameters. A general linear normal model with known covariance matrices is considered in detail.  相似文献   

8.
We wish to model pulse wave velocity (PWV) as a function of longitudinal measurements of pulse pressure (PP) at the same and prior visits at which the PWV is measured. A number of approaches are compared. First, we use the PP at the same visit as the PWV in a linear regression model. In addition, we use the average of all available PPs as the explanatory variable in a linear regression model. Next, a two-stage process is applied. The longitudinal PP is modeled using a linear mixed-effects model. This modeled PP is used in the regression model to describe PWV. An approach for using the longitudinal PP data is to obtain a measure of the cumulative burden, the area under the PP curve. This area under the curve is used as an explanatory variable to model PWV. Finally, a joint Bayesian model is constructed similar to the two-stage model.  相似文献   

9.
The main focus of our paper is to compare the performance of different model selection criteria used for multivariate reduced rank time series. We consider one of the most commonly used reduced rank model, that is, the reduced rank vector autoregression (RRVAR (p, r)) introduced by Velu et al. [Reduced rank models for multiple time series. Biometrika. 1986;7(31):105–118]. In our study, the most popular model selection criteria are included. The criteria are divided into two groups, that is, simultaneous selection and two-step selection criteria, accordingly. Methods from the former group select both an autoregressive order p and a rank r simultaneously, while in the case of two-step criteria, first an optimal order p is chosen (using model selection criteria intended for the unrestricted VAR model) and then an optimal rank r of coefficient matrices is selected (e.g. by means of sequential testing). Considered model selection criteria include well-known information criteria (such as Akaike information criterion, Schwarz criterion, Hannan–Quinn criterion, etc.) as well as widely used sequential tests (e.g. the Bartlett test) and the bootstrap method. An extensive simulation study is carried out in order to investigate the efficiency of all model selection criteria included in our study. The analysis takes into account 34 methods, including 6 simultaneous methods and 28 two-step approaches, accordingly. In order to carefully analyse how different factors affect performance of model selection criteria, we consider over 150 simulation settings. In particular, we investigate the influence of the following factors: time series dimension, different covariance structure, different level of correlation among components and different level of noise (variance). Moreover, we analyse the prediction accuracy concerned with the application of the RRVAR model and compare it with results obtained for the unrestricted vector autoregression. In this paper, we also present a real data application of model selection criteria for the RRVAR model using the Polish macroeconomic time series data observed in the period 1997–2007.  相似文献   

10.
In the present paper, certain random damage models are examined, such as the generalized MARKOV-POLY A (GMP), the Quasi-Binomial, and the Quasi-Hypergeo-metric, in which an integer random variable N is reduced to B. Following JANAEDAN (1973 b) who has characterized the Multivariate Hypergeometric distribution in terms of the Multinomial, we have shown that under the GMP damage model, the distributions of N and B both belong to the family of the generalised POLYA-EGGENBERGER (GPE) distributions. We have also shown that the damage model can be uniquely identified as the GMPD given that B and N belong to the same GPE family. A physical interpretation of the result is given  相似文献   

11.
In this paper, we consider an estimation for the unknown parameters of a conditional Gaussian MA(1) model. In the majority of cases, a maximum-likelihood estimator is chosen because the estimator is consistent. However, for small sample sizes the error is large, because the estimator has a bias of O(n? 1). Therefore, we provide a bias of O(n? 1) for the maximum-likelihood estimator for the conditional Gaussian MA(1) model. Moreover, we propose new estimators for the unknown parameters of the conditional Gaussian MA(1) model based on the bias of O(n? 1). We investigate the properties of the bias, as well as the asymptotical variance of the maximum-likelihood estimators for the unknown parameters, by performing some simulations. Finally, we demonstrate the validity of the new estimators through this simulation study.  相似文献   

12.
13.
Internet traffic data is characterized by some unusual statistical properties, in particular, the presence of heavy-tailed variables. A typical model for heavy-tailed distributions is the Pareto distribution although this is not adequate in many cases. In this article, we consider a mixture of two-parameter Pareto distributions as a model for heavy-tailed data and use a Bayesian approach based on the birth-death Markov chain Monte Carlo algorithm to fit this model. We estimate some measures of interest related to the queueing system k-Par/M/1 where k-Par denotes a mixture of k Pareto distributions. Heavy-tailed variables are difficult to model in such queueing systems because of the lack of a simple expression for the Laplace Transform (LT). We use a procedure based on recent LT approximating results for the Pareto/M/1 system. We illustrate our approach with both simulated and real data.  相似文献   

14.
《随机性模型》2013,29(3):299-324
In this paper we consider a bottleneck link and buffer used by one or two fluid sources that are subject to feedback. The feedback is such that the model captures essential aspects of the behavior of the Transmission Control Protocol as used in the Internet. During overflow, the buffer sends negative feedback signals to the sources to indicate that the sending rate should be reduced. Otherwise the buffer sends positive signals so as to increase the rate. In this context we find closed form expressions for the solution of the one-source case. The two-source case extends the single-source model considerably: we can control the behavior and parameters of each source individually. This enables us to study the impact of these parameters on the sharing of links and buffers. For the two-source case we solve the related two-point boundary value problem in the stationary case. We also establish a numerically efficient procedure to compute the coefficients of the solution of the differential equations. The numerical results of this model are presented in an accompanying paper.  相似文献   

15.
In recent years, a number of statistical models have been proposed for the purposes of high-level image analysis tasks such as object recognition. However, in general, these models remain hard to use in practice, partly as a result of their complexity, partly through lack of software. In this paper we concentrate on a particular deformable template model which has proved potentially useful for locating and labelling cells in microscope slides Rue and Hurn (1999). This model requires the specification of a number of rather non-intuitive parameters which control the shape variability of the deformed templates. Our goal is to arrange the estimation of these parameters in such a way that the microscope user's expertise is exploited to provide the necessary training data graphically by identifying a number of cells displayed on a computer screen, but that no additional statistical input is required. In this paper we use maximum likelihood estimation incorporating the error structure in the generation of our training data.  相似文献   

16.
In this paper, we consider a special finite mixture model named Combination of Uniform and shifted Binomial (CUB), recently introduced in the statistical literature to analyse ordinal data expressing the preferences of raters with regards to items or services. Our aim is to develop a variable selection procedure for this model using a Bayesian approach. Bayesian methods for variable selection and model choice have become increasingly popular in recent years, due to advances in Markov chain Monte Carlo computational algorithms. Several methods have been proposed in the case of linear and generalized linear models (GLM). In this paper, we adapt to the CUB model some of these algorithms: the Kuo–Mallick method together with its ‘metropolized’ version and the Stochastic Search Variable Selection method. Several simulated examples are used to illustrate the algorithms and to compare their performance. Finally, an application to real data is introduced.  相似文献   

17.
Functional regression models that relate functional covariates to a scalar response are becoming more common due to the availability of functional data and computational advances. We introduce a functional nonlinear model with a scalar response where the true parameter curve is monotone. Using the Newton-Raphson method within a backfitting procedure, we discuss a penalized least squares criterion for fitting the functional nonlinear model with the smoothing parameter selected using generalized cross validation. Connections between a nonlinear mixed effects model and our functional nonlinear model are discussed, thereby providing an additional model fitting procedure using restricted maximum likelihood for smoothing parameter selection. Simulated relative efficiency gains provided by a monotone parameter curve estimator relative to an unconstrained parameter curve estimator are presented. In addition, we provide an application of our model with data from ozonesonde measurements of stratospheric ozone in which the measurements are biased as a function of altitude.  相似文献   

18.
Item response theory (IRT) models provide an important contribution in the analysis of polytomous items, such as Likert scale items in survey data. We propose a bifactor generalized partial credit model (bifac-GPC model) with flexible link functions - probit, logit and complementary log-log - for use in analysis of ordered polytomous item scale data. In order to estimate the parameters of the proposed model, we use a Bayesian approach through the NUTS algorithm and show the advantages of implementing IRT models through the Stan language. We present an application to marketing scale data. Specifically, we apply the model to a dataset of non-users of a mobile banking service in order to highlight the advantages of this model. The results show important managerial implications resulting from consumer perceptions. We provide a discussion of the methodology for this type of data and extensions. Codes are available for practitioners and researchers to replicate the application.  相似文献   

19.
In this article, we extend a previously formulated threshold dose-response model with random litter effects that was applied to a data set from a developmental toxicity study. The dose-response pattern of the data indicates that a threshold dose level may exist. Additionally, there is noticeable variation between the responses across the dose levels. With threshold estimation being critical, the assumed variability structure should adequately model the variation while not taking away from the estimation of the threshold as well as the other parameters directly involved in the dose-response relationship. In the prior formulation, the random effect was modeled assuming identical variation in the interlitter response probabilities across all dose levels, that is, the model had a single parameter to account for the interlitter variability. In this new model, the random effect is modeled as having different response variability across dose levels, that is, multiple interlitter variability parameters. We performed the likelihood ratio test (LRT) to compare our extended model to the previous model. We conducted a simulation study to compare the bias of each model when fit to data generated with the underlying parametric structure of the opposing model. The extended threshold dose-response model with multiple response variation was less biased.  相似文献   

20.
The logistic regression model has become a standard tool to investigate the relationship between a binary outcome and a set of potential predictors. When analyzing binary data, it often arises that the observed proportion of zeros is greater than expected under the postulated logistic model. Zero-inflated binomial (ZIB) models have been developed to fit binary data that contain too many zeros. Maximum likelihood estimators in these models have been proposed and their asymptotic properties established. Several aspects of ZIB models still deserve attention however, such as the estimation of odds-ratios and event probabilities. In this article, we propose estimators of these quantities and we investigate their properties both theoretically and via simulations. Based on these results, we provide recommendations about the range of conditions (minimum sample size, maximum proportion of zeros in excess) under which a reliable statistical inference on the odds-ratios and event probabilities can be obtained in a ZIB regression model. A real-data example illustrates the proposed estimators.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号