首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Abstract

Augmented mixed beta regression models are suitable choices for modeling continuous response variables on the closed interval [0, 1]. The random eeceeects in these models are typically assumed to be normally distributed, but this assumption is frequently violated in some applied studies. In this paper, an augmented mixed beta regression model with skew-normal independent distribution for random effects are used. Next, we adopt a Bayesian approach for parameter estimation using the MCMC algorithm. The methods are then evaluated using some intensive simulation studies. Finally, the proposed models have applied to analyze a dataset from an Iranian Labor Force Survey.  相似文献   

2.
In this article, a generalized linear mixed model (GLMM) based on a frequentist approach is employed to examine spatial trend of asthma data. However, the frequentist analysis of GLMM is computationally difficult. On the other hand, the Bayesian analysis of GLMM has been computationally convenient due to the advent of Markov chain Monte Carlo algorithms. Recently developed data cloning (DC) method, which yields to maximum likelihood estimate, provides frequentist approach to complex mixed models and equally computationally convenient method. We use DC to conduct frequentist analysis of spatial models. The advantages of the DC approach are that the answers are independent of the choice of the priors, non-estimable parameters are flagged automatically, and the possibility of improper posterior distributions is completely avoided. We illustrate this approach using a real dataset of asthma visits to hospital in the province of Manitoba, Canada, during 2000–2010. The performance of the DC approach in our application is also studied through a simulation study.  相似文献   

3.
ABSTRACT

In this article, we propose a new distribution by mixing normal and Pareto distributions, and the new distribution provides an unusual hazard function. We model the mean and the variance with covariates for heterogeneity. Estimation of the parameters is obtained by the Bayesian method using Markov Chain Monte Carlo (MCMC) algorithms. Proposal distribution in MCMC is proposed with a defined working variable related to the observations. Through the simulation, the method shows a dependable performance of the model. We demonstrate through establishing model under a real dataset that the proposed model and method can be more suitable than the previous report.  相似文献   

4.
Abstract

A procedure based on the sum of N Shiryayev–Roberts processes is proposed to detect common changes in panel data and shown to perform better for small portions of changed panels. The change-point for each panel is estimated by using the CUSUM process calculated in parallel. The changed panels are isolated by using the scores formed by the post-change parameter estimations and the common change point is then estimated from the isolated changed panels. A real example is used for illustration. An adaptive detection procedures is also proposed when the unknown post-change parameters are estimated adaptively in each panel.  相似文献   

5.
Abstract

It is one of the important issues in survival analysis to compare two hazard rate functions to evaluate treatment effect. It is quite common that the two hazard rate functions cross each other at one or more unknown time points, representing temporal changes of the treatment effect. In certain applications, besides survival data, we also have related longitudinal data available regarding some time-dependent covariates. In such cases, a joint model that accommodates both types of data can allow us to infer the association between the survival and longitudinal data and to assess the treatment effect better. In this paper, we propose a modelling approach for comparing two crossing hazard rate functions by joint modelling survival and longitudinal data. Maximum likelihood estimation is used in estimating the parameters of the proposed joint model using the EM algorithm. Asymptotic properties of the maximum likelihood estimators are studied. To illustrate the virtues of the proposed method, we compare the performance of the proposed method with several existing methods in a simulation study. Our proposed method is also demonstrated using a real dataset obtained from an HIV clinical trial.  相似文献   

6.
In this article, a semiparametric approach is proposed for the regression analysis of panel count data. Panel count data commonly arise in clinical trials and demographical studies where the response variable is the number of multiple recurrences of the event of interest and observation times are not fixed, varying from subject to subject. It is assumed that two processes exist in this data: the first is for a recurrent event and the second is for observation time. Many studies have been done to estimate mean function and regression parameters under the independency between recurrent event process and observation time process. In this article, the same statistical inference is studied, but the situation where these two processes may be related is also considered. The mixed Poisson process is applied for the recurrent event processes, and a frailty intensity function for the observation time is also used, respectively. Simulation studies are conducted to study the performance of the suggested methods. The bladder tumor data are applied to compare previous studie' results.  相似文献   

7.
This article studies computation problem in the context of estimating parameters of linear mixed model for massive data. Our algorithms combine the factored spectrally transformed linear mixed model method with a sequential singular value decomposition calculation algorithm. This combination solves the operation limitation of the method and also makes this algorithm feasible to big dataset, especially when the data has a tall and thin design matrix. Our simulation studies show that our algorithms make the calculation of linear mixed model feasible for massive data on ordinary desktop and have same estimating accuracy with the method based on the whole data.  相似文献   

8.
Abstract

The locally weighted censored quantile regression approach is proposed for panel data models with fixed effects, which allows for random censoring. The resulting estimators are obtained by employing the fixed effects quantile regression method. The weights are selected either parametrically, semi-parametrically or non-parametrically. The large panel data asymptotics are used in an attempt to cope with the incidental parameter problem. The consistency and limiting distribution of the proposed estimator are also derived. The finite sample performance of the proposed estimators are examined via Monte Carlo simulations.  相似文献   

9.
Abstract

A nonparametric procedure is proposed to estimate multiple change-points of location changes in a univariate data sequence by using ranks instead of the raw data. While existing rank-based multiple change-point detection methods are mostly based on sequential tests, we treat it as a model selection problem. We derive the corresponding Schwarz’s information criterion for rank-statistics, theoretically prove the consistency of the change-point estimator and use a pruned dynamic programing algorithm to achieve the change-point estimator. Simulation studies show our method’s robustness, effectiveness and efficiency in detecting mean-changes. We also apply the method to a gene dataset as an illustration.  相似文献   

10.
Spatial modeling is widely used in environmental sciences, biology, and epidemiology. Generalized linear mixed models are employed to account for spatial variations of point-referenced data called spatial generalized linear mixed models (SGLMMs). Frequentist analysis of these type of data is computationally difficult. On the other hand, the advent of the Markov chain Monte Carlo algorithm has made the Bayesian analysis of SGLMM computationally convenient. Recent introduction of the method of data cloning, which leads to maximum likelihood estimate, has made frequentist analysis of mixed models also equally computationally convenient. Recently, the data cloning was employed to estimate model parameters in SGLMMs, however, the prediction of spatial random effects and kriging are also very important. In this article, we propose a frequentist approach based on data cloning to predict (and provide prediction intervals) spatial random effects and kriging. We illustrate this approach using a real dataset and also by a simulation study.  相似文献   

11.
Summary.  The paper proposes an estimation approach for panel models with mixed continuous and ordered categorical outcomes based on generalized estimating equations for the mean and pseudoscore equations for the covariance parameters. A numerical study suggests that efficiency can be gained in the mean parameter estimators by using individual covariance matrices in the estimating equations for the mean parameters. The approach is applied to estimate the returns to occupational qualification in terms of income and perceived job security in a 9-year period based on the German Socio-Economic Panel. To compensate for missing data, a combined multiple imputation–weighting approach is adopted.  相似文献   

12.
Abstract

We propose a new class of two-stage parameter estimation methods for semiparametric ordinary differential equation (ODE) models. In the first stage, state variables are estimated using a penalized spline approach; In the second stage, form of numerical discretization algorithms for an ODE solver is used to formulate estimating equations. Estimated state variables from the first stage are used to obtain more data points for the second stage. Asymptotic properties for the proposed estimators are established. Simulation studies show that the method performs well, especially for small sample. Real life use of the method is illustrated using Influenza specific cell-trafficking study.  相似文献   

13.
ABSTRACT

In this paper we propose a class of skewed t link models for analyzing binary response data with covariates. It is a class of asymmetric link models designed to improve the overall fit when commonly used symmetric links, such as the logit and probit links, do not provide the best fit available for a given binary response dataset. Introducing a skewed t distribution for the underlying latent variable, we develop the class of models. For the analysis of the models, a Bayesian and non-Bayesian methods are pursued using a Markov chain Monte Carlo (MCMC) sampling based approach. Necessary theories involved in modelling and computation are provided. Finally, a simulation study and a real data example are used to illustrate the proposed methodology.  相似文献   

14.
Abstract

Non-normality is a common phenomenon in data from agricultural and biological research, especially in molecular data (for example; -omics, RNAseq, flow cytometric data, etc.). For over half a century, the leading paradigm called for using analysis of variance (ANOVA) after applying a data transformation. The introduction of generalized linear mixed models (GLMM) provides a new way of analyzing non-normal data. Selecting an apt link function in GLMM can be quite influential, however, and is as critical as selecting an appropriate transformation for ANOVA. In this paper, we assess the performance of different parametric link families available in literature. Then, we propose a new estimation method for selecting an appropriate link function with a suitable variance function in a quasi-likelihood framework. We apply these methods to a proteomics data set, showing that GLMMs provide a very flexible framework for analyzing these kinds of data.  相似文献   

15.
For micro-datasets considered for release as scientific or public use files, statistical agencies have to face the dilemma of guaranteeing the confidentiality of survey respondents on the one hand and offering sufficiently detailed data on the other hand. For that reason, a variety of methods to guarantee disclosure control is discussed in the literature. In this paper, we present an application of Rubin’s (J. Off. Stat. 9, 462–468, 1993) idea to generate synthetic datasets from existing confidential survey data for public release.We use a set of variables from the 1997 wave of the German IAB Establishment Panel and evaluate the quality of the approach by comparing results from an analysis by Zwick (Ger. Econ. Rev. 6(2), 155–184, 2005) with the original data with the results we achieve for the same analysis run on the dataset after the imputation procedure. The comparison shows that valid inferences can be obtained using the synthetic datasets in this context, while confidentiality is guaranteed for the survey participants.  相似文献   

16.
ABSTRACT

Restricted canonical correlation analysis and the lasso shrinkage method were paired together for canonical correlation analysis with non-negativity restrictions on datasets, where a sample size is much smaller than the number of variables. The method was implemented in an alternating least-squares algorithm and applied to cross-language information retrieval on a dataset with aligned documents in eight languages. A set of experiments was ran to evaluate the method and compare it to other methods in the field.  相似文献   

17.
Abstract

Solar radiation is a global ecological phenomenon that affects life everywhere. In this study, a new statistical method, called the Quartiles-Moment's method, is proposed to estimate the scale and shape parameters of the exponentiated Gumbel maximum distribution (EGMD). The Kolomogorov–Smirnov test and the percentiles of the dataset are thus used to fit the dataset of the daily global solar radiation and the corresponding daily maximum temperature with EGMD. Thence, multiple nonlinear regression of the daily global solar radiation and the corresponding daily maximum temperature are produced and compared with the real dataset accordingly.  相似文献   

18.
Abstract

Cluster analysis is the distribution of objects into different groups or more precisely the partitioning of a data set into subsets (clusters) so that the data in subsets share some common trait according to some distance measure. Unlike classification, in clustering one has to first decide the optimum number of clusters and then assign the objects into different clusters. Solution of such problems for a large number of high dimensional data points is quite complicated and most of the existing algorithms will not perform properly. In the present work a new clustering technique applicable to large data set has been used to cluster the spectra of 702248 galaxies and quasars having 1,540 points in wavelength range imposed by the instrument. The proposed technique has successfully discovered five clusters from this 702,248X1,540 data matrix.  相似文献   

19.
A researcher using complex longitudinal survey data for event history analysis has to make several choices that affect the analysis results. These choices include the following: whether a design-based or a model-based approach for the analysis is taken, which subset of data to use and, if a design-based approach is chosen, which weights to use. We discuss different choices and illustrate their effects using longitudinal register data linked at person-level with the Finnish subset of the European Community Household Panel data. The use of register data enables us to construct an event history data set without nonresponse and attrition. Design-based estimates from these data are used as benchmarks against design-based and model-based estimates from subsets of data usually available for a survey data analyst. Our illustration suggests that the often recommended way to use panel data for longitudinal analyses, data from total respondents and weights from the last wave analysed may not be the best way to go. Instead, using all available data and weights from the first survey wave appears to be a safe choice for longitudinal analyses based on multipurpose survey data.  相似文献   

20.
《随机性模型》2013,29(1):149-184
Abstract

We consider several known algorithms and introduce some new algorithms that can be used to calculate the probability of return to the initial level in the Markov stochastic fluid flow model. We give the physical interpretations of these algorithms within the fluid flow environment. The rates of convergence are explained in terms of the physical properties of the fluid flow processes. We compare these algorithms with respect to the number of iterations required and their complexity. The performance of the algorithms depends on the nature of the process considered in the analysis. We illustrate this with examples and give appropriate recommendations.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号