首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.

In this paper the use of the empirical Fisher information matrix as an estimator of the information matrix is considered in the context of response models and incomplete data problems. The introduction of an additional stochastic component into such models is shown to greatly increase the range of situations in which the estimator can be employed. In particular the conditions for its use in incomplete data problems are shown to be the same as those needed to justify the use of the EM algorithm.  相似文献   

2.
Abstract

The assumption of underlying return distribution plays an important role in asset pricing models. While the return distribution used in the traditional theories of asset pricing is the unimodal distribution, numerous studies which have investigated the empirical behavior of asset returns in financial markets use multi-modal distribution. We introduce a new parsimonious multi-modal distribution, referred to as the multi-modal tempered stable (MMTS) distribution. In this article we also generate the exponential Lévy market models and derive the value-at-risk (VaR) induced from them. To demonstrate the advantages, we will present the results of the parameter estimation and the VaRs for financial data.  相似文献   

3.
The issue of modelling non-Gaussian time series data is one that has been examined by several authors in recent years. Zeger (1988) introduced a parameter-driven model for a time series of counts as well as a more general observation-driven model for non-Gaussian data (Zeger & Qaqish, 1988). This paper examines the application of the added variable plot to these two models. This plot is useful for determining the strength of relationships and the detection of influential or outlying observations.  相似文献   

4.
Abstract

Variable selection in finite mixture of regression (FMR) models is frequently used in statistical modeling. The majority of applications of variable selection in FMR models use a normal distribution for regression error. Such assumptions are unsuitable for a set of data containing a group or groups of observations with heavy tails and outliers. In this paper, we introduce a robust variable selection procedure for FMR models using the t distribution. With appropriate selection of the tuning parameters, the consistency and the oracle property of the regularized estimators are established. To estimate the parameters of the model, we develop an EM algorithm for numerical computations and a method for selecting tuning parameters adaptively. The parameter estimation performance of the proposed model is evaluated through simulation studies. The application of the proposed model is illustrated by analyzing a real data set.  相似文献   

5.
This paper proposes a working estimating equation which is computationally easy to use for spatial count data. The proposed estimating equation is a modification of quasi-likelihood estimating equations without the need of correctly specifying the covariance matrix. Under some regularity conditions, we show that the proposed estimator has consistency and asymptotic normality. A simulation comparison also indicates that the proposed method has competitive performance in dealing with over-dispersion data from a parameter-driven model.  相似文献   

6.
ABSTRACT

In this paper we propose a class of skewed t link models for analyzing binary response data with covariates. It is a class of asymmetric link models designed to improve the overall fit when commonly used symmetric links, such as the logit and probit links, do not provide the best fit available for a given binary response dataset. Introducing a skewed t distribution for the underlying latent variable, we develop the class of models. For the analysis of the models, a Bayesian and non-Bayesian methods are pursued using a Markov chain Monte Carlo (MCMC) sampling based approach. Necessary theories involved in modelling and computation are provided. Finally, a simulation study and a real data example are used to illustrate the proposed methodology.  相似文献   

7.
Useful models for time series of counts or simply wrong ones?   总被引:1,自引:0,他引:1  
There has been a considerable and growing interest in low integer-valued time series data leading to a diversification of modelling approaches. In addition to static regression models, both observation-driven and parameter-driven models are considered here. We compare and contrast a variety of time series models for counts using two very different data sets as a testbed. A range of diagnostic devices is employed to help inform model adequacy. Special attention is paid to dynamic structure and underlying distributional assumptions including associated dispersion properties. Competing models show attractive features, but overall no one modelling approach is seen to dominate.  相似文献   

8.
ABSTRACT

Standard prior elicitation procedures require experts to explicitly quantify their beliefs about parameters in the form of multiple summaries. In this article, we draw on recent advances in the statistical graphics and information visualization communities to propose a novel elicitation scheme that implicitly learns an expert’s opinions through their sequential selection of graphics of carefully constructed hypothetical future samples. While the scheme can be applied to a broad array of models, we use it to construct procedures for elicitation in data models commonly used in practice: Bernoulli, Poisson, and Normal. We also provide open-source, web-based Shiny implementations of the procedures.  相似文献   

9.
Abstract

Although stochastic volatility and GARCH (generalized autoregressive conditional heteroscedasticity) models have successfully described the volatility dynamics of univariate asset returns, extending them to the multivariate models with dynamic correlations has been difficult due to several major problems. First, there are too many parameters to estimate if available data are only daily returns, which results in unstable estimates. One solution to this problem is to incorporate additional observations based on intraday asset returns, such as realized covariances. Second, since multivariate asset returns are not synchronously traded, we have to use the largest time intervals such that all asset returns are observed to compute the realized covariance matrices. However, in this study, we fail to make full use of the available intraday informations when there are less frequently traded assets. Third, it is not straightforward to guarantee that the estimated (and the realized) covariance matrices are positive definite.

Our contributions are the following: (1) we obtain the stable parameter estimates for the dynamic correlation models using the realized measures, (2) we make full use of intraday informations by using pairwise realized correlations, (3) the covariance matrices are guaranteed to be positive definite, (4) we avoid the arbitrariness of the ordering of asset returns, (5) we propose the flexible correlation structure model (e.g., such as setting some correlations to be zero if necessary), and (6) the parsimonious specification for the leverage effect is proposed. Our proposed models are applied to the daily returns of nine U.S. stocks with their realized volatilities and pairwise realized correlations and are shown to outperform the existing models with respect to portfolio performances.  相似文献   

10.
Abstract

In this paper, we establish that the usual stochastic, hazard rate, reversed hazard rate, likelihood ratio, dispersive and star orders are all preserved for parallel systems under exponentiated models for lifetimes of components. We then use the multiple-outlier exponentiated gamma models to illustrate this result. Finally, we consider the dual family with exponentiated survival function and establish similar results for series systems. The results established here extend some well-known results for series and parallel systems arising from different exponentiated distributions such as generalized exponential and exponentiated Weibull, established previously in the literature.  相似文献   

11.
In partly linear models, the dependence of the response y on (x T, t) is modeled through the relationship y=x T β+g(t)+?, where ? is independent of (x T, t). We are interested in developing an estimation procedure that allows us to combine the flexibility of the partly linear models, studied by several authors, but including some variables that belong to a non-Euclidean space. The motivating application of this paper deals with the explanation of the atmospheric SO2 pollution incidents using these models when some of the predictive variables belong in a cylinder. In this paper, the estimators of β and g are constructed when the explanatory variables t take values on a Riemannian manifold and the asymptotic properties of the proposed estimators are obtained under suitable conditions. We illustrate the use of this estimation approach using an environmental data set and we explore the performance of the estimators through a simulation study.  相似文献   

12.
Abstract

In this article, we have considered three different shared frailty models under the assumption of generalized Pareto Distribution as baseline distribution. Frailty models have been used in the survival analysis to account for the unobserved heterogeneity in an individual risks to disease and death. These three frailty models are with gamma frailty, inverse Gaussian frailty and positive stable frailty. Then we introduce the Bayesian estimation procedure using Markov chain Monte Carlo (MCMC) technique to estimate the parameters. We applied these three models to a kidney infection data and find the best fitted model for kidney infection data. We present a simulation study to compare true value of the parameters with the estimated values. Model comparison is made using Bayesian model selection criterion and a well-fitted model is suggested for the kidney infection data.  相似文献   

13.
Time series of counts occur in many different contexts, the counts being usually of certain events or objects in specified time intervals. In this paper we introduce a model called parameter-driven state-space model to analyse integer-valued time series data. A key property of such model is that the distribution of the observed count data is independent, conditional on the latent process, although the observations are correlated marginally. Our simulation shows that the Monte Carlo Expectation Maximization (MCEM) algorithm and the particle method are useful for the parameter estimation of the proposed model. In the application to Malaysia dengue data, our model fits better when compared with several other models including that of Yang et al. (2015)  相似文献   

14.
ABSTRACT

Inflated data are prevalent in many situations and a variety of inflated models with extensions have been derived to fit data with excessive counts of some particular responses. The family of information criteria (IC) has been used to compare the fit of models for selection purposes. Yet despite the common use in statistical applications, there are not too many studies evaluating the performance of IC in inflated models. In this study, we studied the performance of IC for data with dual-inflated data. The new zero- and K-inflated Poisson (ZKIP) regression model and conventional inflated models including Poisson regression and zero-inflated Poisson (ZIP) regression were fitted for dual-inflated data and the performance of IC were compared. The effect of sample sizes and the proportions of inflated observations towards selection performance were also examined. The results suggest that the Bayesian information criterion (BIC) and consistent Akaike information criterion (CAIC) are more accurate than the Akaike information criterion (AIC) in terms of model selection when the true model is simple (i.e. Poisson regression (POI)). For more complex models, such as ZIP and ZKIP, the AIC was consistently better than the BIC and CAIC, although it did not reach high levels of accuracy when sample size and the proportion of zero observations were small. The AIC tended to over-fit the data for the POI, whereas the BIC and CAIC tended to under-parameterize the data for ZIP and ZKIP. Therefore, it is desirable to study other model selection criteria for dual-inflated data with small sample size.  相似文献   

15.
ABSTRACT

Online consumer product ratings data are increasing rapidly. While most of the current graphical displays mainly represent the average ratings, Ho and Quinn proposed an easily interpretable graphical display based on an ordinal item response theory (IRT) model, which successfully accounts for systematic interrater differences. Conventionally, the discrimination parameters in IRT models are constrained to be positive, particularly in the modeling of scored data from educational tests. In this article, we use real-world ratings data to demonstrate that such a constraint can have a great impact on the parameter estimation. This impact on estimation was explained through rater behavior. We also discuss correlation among raters and assess the prediction accuracy for both the constrained and the unconstrained models. The results show that the unconstrained model performs better when a larger fraction of rater pairs exhibit negative correlations in ratings.  相似文献   

16.
17.
This paper presents a methodology for model fitting and inference in the context of Bayesian models of the type f(Y | X,θ)f(X|θ)f(θ), where Y is the (set of) observed data, θ is a set of model parameters and X is an unobserved (latent) stationary stochastic process induced by the first order transition model f(X (t+1)|X (t),θ), where X (t) denotes the state of the process at time (or generation) t. The crucial feature of the above type of model is that, given θ, the transition model f(X (t+1)|X (t),θ) is known but the distribution of the stochastic process in equilibrium, that is f(X|θ), is, except in very special cases, intractable, hence unknown. A further point to note is that the data Y has been assumed to be observed when the underlying process is in equilibrium. In other words, the data is not collected dynamically over time. We refer to such specification as a latent equilibrium process (LEP) model. It is motivated by problems in population genetics (though other applications are discussed), where it is of interest to learn about parameters such as mutation and migration rates and population sizes, given a sample of allele frequencies at one or more loci. In such problems it is natural to assume that the distribution of the observed allele frequencies depends on the true (unobserved) population allele frequencies, whereas the distribution of the true allele frequencies is only indirectly specified through a transition model. As a hierarchical specification, it is natural to fit the LEP within a Bayesian framework. Fitting such models is usually done via Markov chain Monte Carlo (MCMC). However, we demonstrate that, in the case of LEP models, implementation of MCMC is far from straightforward. The main contribution of this paper is to provide a methodology to implement MCMC for LEP models. We demonstrate our approach in population genetics problems with both simulated and real data sets. The resultant model fitting is computationally intensive and thus, we also discuss parallel implementation of the procedure in special cases.  相似文献   

18.
ABSTRACT

The living hours data of individuals' time spent on daily activities are compositional and include many zeros because individuals do not pursue all activities every day. Thus, we should exercise caution in using such data for empirical analyses. The Bayesian method offers several advantages in analyzing compositional data. In this study, we analyze the time allocation of Japanese married couples using the Bayesian model. Based on the Bayes factors, we compare models that consider and do not consider the correlations between married couples' time use data. The model that considers the correlation shows superior performance. We show that the Bayesian method can adequately take into account the correlations of wives' and husbands' living hours, facilitating the calculation of partial effects that their activities' variables have on living hours. The partial effects of the model that considers the correlations between the couples' time use are easily calculated from the posterior results.  相似文献   

19.
Abstract

The multivariate elliptically contoured distributions provide a viable framework for modeling time-series data. It includes the multivariate normal, power exponential, t, and Cauchy distributions as special cases. For multivariate elliptically contoured autoregressive models, we derive the exact likelihood equations for the model parameters. They are closely related to the Yule-Walker equations and involve simple function of the data. The maximum likelihood estimators are obtained by alternately solving two linear systems and illustrated using the simulation data.  相似文献   

20.
ABSTRACT

The Lindley distribution is an important distribution for analysing the stress–strength reliability models and lifetime data. In many ways, the Lindley distribution is a better model than that based on the exponential distribution. Order statistics arise naturally in many of such applications. In this paper, we derive the exact explicit expressions for the single, double (product), triple and quadruple moments of order statistics from the Lindley distribution. Then, we use these moments to obtain the best linear unbiased estimates (BLUEs) of the location and scale parameters based on Type-II right-censored samples. Next, we use these results to determine the mean, variance, and coefficients of skewness and kurtosis of some certain linear functions of order statistics to develop Edgeworth approximate confidence intervals of the location and scale Lindley parameters. In addition, we carry out some numerical illustrations through Monte Carlo simulations to show the usefulness of the findings. Finally, we apply the findings of the paper to some real data set.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号