首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In survey sampling, policy decisions regarding the allocation of resources to sub‐groups of a population depend on reliable predictors of their underlying parameters. However, in some sub‐groups, called small areas due to small sample sizes relative to the population, the information needed for reliable estimation is typically not available. Consequently, data on a coarser scale are used to predict the characteristics of small areas. Mixed models are the primary tools in small area estimation (SAE) and also borrow information from alternative sources (e.g., previous surveys and administrative and census data sets). In many circumstances, small area predictors are associated with location. For instance, in the case of chronic disease or cancer, it is important for policy makers to understand spatial patterns of disease in order to determine small areas with high risk of disease and establish prevention strategies. The literature considering SAE with spatial random effects is sparse and mostly in the context of spatial linear mixed models. In this article, small area models are proposed for the class of spatial generalized linear mixed models to obtain small area predictors and corresponding second‐order unbiased mean squared prediction errors via Taylor expansion and a parametric bootstrap approach. The performance of the proposed approach is evaluated through simulation studies and application of the models to a real esophageal cancer data set from Minnesota, U.S.A. The Canadian Journal of Statistics 47: 426–437; 2019 © 2019 Statistical Society of Canada  相似文献   

2.
A simulation experiment compares the accuracy and precision of three alternate estimation techniques for the parameters of the STARMA model. Maximum likelihood estimation, in most ways the "best" estimation procedure, involves a large amount of computational effort so that two approximate techniques, exact least squares and conditional maximum likelihood, are often proposed for series of moderate lengths. This simulation experiment compares the accuracy of these three estimation procedures for simulated series of various lengths, and discusses the appropriateness of the three procedures as a function of the length of the observed series.  相似文献   

3.
Hailin Sang 《Statistics》2015,49(1):187-208
We propose a sparse coefficient estimation and automated model selection procedure for autoregressive processes with heavy-tailed innovations based on penalized conditional maximum likelihood. Under mild moment conditions on the innovation processes, the penalized conditional maximum likelihood estimator satisfies a strong consistency, OP(N?1/2) consistency, and the oracle properties, where N is the sample size. We have the freedom in choosing penalty functions based on the weak conditions on them. Two penalty functions, least absolute shrinkage and selection operator and smoothly clipped average deviation, are compared. The proposed method provides a distribution-based penalized inference to AR models, which is especially useful when the other estimation methods fail or under perform for AR processes with heavy-tailed innovations [Feigin, Resnick. Pitfalls of fitting autoregressive models for heavy-tailed time series. Extremes. 1999;1:391–422]. A simulation study confirms our theoretical results. At the end, we apply our method to a historical price data of the US Industrial Production Index for consumer goods, and obtain very promising results.  相似文献   

4.
Ordinary differential equations (ODEs) are normally used to model dynamic processes in applied sciences such as biology, engineering, physics, and many other areas. In these models, the parameters are usually unknown, and thus they are often specified artificially or empirically. Alternatively, a feasible method is to estimate the parameters based on observed data. In this study, we propose a Bayesian penalized B-spline approach to estimate the parameters and initial values for ODEs used in epidemiology. We evaluated the efficiency of the proposed method based on simulations using the Markov chain Monte Carlo algorithm for the Kermack–McKendrick model. The proposed approach is also illustrated based on a real application to the transmission dynamics of hepatitis C virus in mainland China.  相似文献   

5.
The skew-normal and the skew-t distributions are parametric families which are currently under intense investigation since they provide a more flexible formulation compared to the classical normal and t distributions by introducing a parameter which regulates their skewness. While these families enjoy attractive formal properties from the probability viewpoint, a practical problem with their usage in applications is the possibility that the maximum likelihood estimate of the parameter which regulates skewness diverges. This situation has vanishing probability for increasing sample size, but for finite samples it occurs with non-negligible probability, and its occurrence has unpleasant effects on the inferential process. Methods for overcoming this problem have been put forward both in the classical and in the Bayesian formulation, but their applicability is restricted to simple situations. We formulate a proposal based on the idea of penalized likelihood, which has connections with some of the existing methods, but it applies more generally, including the multivariate case.  相似文献   

6.
In survey sampling, policymaking regarding the allocation of resources to subgroups (called small areas) or the determination of subgroups with specific properties in a population should be based on reliable estimates. Information, however, is often collected at a different scale than that of these subgroups; hence, the estimation can only be obtained on finer scale data. Parametric mixed models are commonly used in small‐area estimation. The relationship between predictors and response, however, may not be linear in some real situations. Recently, small‐area estimation using a generalised linear mixed model (GLMM) with a penalised spline (P‐spline) regression model, for the fixed part of the model, has been proposed to analyse cross‐sectional responses, both normal and non‐normal. However, there are many situations in which the responses in small areas are serially dependent over time. Such a situation is exemplified by a data set on the annual number of visits to physicians by patients seeking treatment for asthma, in different areas of Manitoba, Canada. In cases where covariates that can possibly predict physician visits by asthma patients (e.g. age and genetic and environmental factors) may not have a linear relationship with the response, new models for analysing such data sets are required. In the current work, using both time‐series and cross‐sectional data methods, we propose P‐spline regression models for small‐area estimation under GLMMs. Our proposed model covers both normal and non‐normal responses. In particular, the empirical best predictors of small‐area parameters and their corresponding prediction intervals are studied with the maximum likelihood estimation approach being used to estimate the model parameters. The performance of the proposed approach is evaluated using some simulations and also by analysing two real data sets (precipitation and asthma).  相似文献   

7.
This article introduces a novel non parametric penalized likelihood hazard estimation when the censoring time is dependent on the failure time for each subject under observation. More specifically, we model this dependence using a copula, and the method of maximum penalized likelihood (MPL) is adopted to estimate the hazard function. We do not consider covariates in this article. The non negatively constrained MPL hazard estimation is obtained using a multiplicative iterative algorithm. The consistency results and the asymptotic properties of the proposed hazard estimator are derived. The simulation studies show that our MPL estimator under dependent censoring with an assumed copula model provides a better accuracy than the MPL estimator under independent censoring if the sign of dependence is correctly specified in the copula function. The proposed method is applied to a real dataset, with a sensitivity analysis performed over various values of correlation between failure and censoring times.  相似文献   

8.
In this article, small area estimation under a multivariate linear model for repeated measures data is considered. The proposed model aims to get a model which borrows strength both across small areas and over time. The model accounts for repeated surveys, grouped response units, and random effects variations. Estimation of model parameters is discussed within a likelihood based approach. Prediction of random effects, small area means across time points, and per group units are derived. A parametric bootstrap method is proposed for estimating the mean squared error of the predicted small area means. Results are supported by a simulation study.  相似文献   

9.
Scheike and Zhang [An additive-multiplicative Cox-Aalen regression model. Scand J Stat. 2002;29:75–88] proposed a flexible additive-multiplicative hazard model, called the Cox-Aalen model, by replacing the baseline hazard function in the well-known Cox model with a covariate-dependent Aalen model, which allows for both fixed and dynamic covariate effects. In this paper, based on left-truncated and mixed interval-censored (LT-MIC) data, we consider maximum likelihood estimation for the Cox-Aalen model with fixed covariates. We propose expectation-maximization (EM) algorithms for obtaining the conditional maximum likelihood estimators (cMLE) of the regression coefficients for the Cox-Aalen model. We establish the consistency of the cMLE. Numerical studies show that estimation via the EM algorithms performs well.  相似文献   

10.
We consider a multinomial distribution in which the cell probabilities are known arbitrary functions of a vector parameter θ. It is desired to estimate θ by least squares. Three variations of the least squares approach are investigated, and each is found to be equivalent, in the very strong sense of being algebraically identical, to one of the following estimation procedures: maximum likelihood, minimum χ2 and minimum modified χ2. Two of these results also apply to the multiple hypergeometric distribution.  相似文献   

11.
In this paper considering an appropriate transformation on the Lindley distribution, we propose the unit-Lindley distribution and investigate some of its statistical properties. An important fact associated with this new distribution is that it is possible to obtain the analytical expression for bias correction of the maximum likelihood estimator. Moreover, it belongs to the exponential family. This distribution allows us to incorporate covariates directly in the mean and consequently to quantify their influences on the average of the response variable. Finally, a practical application is presented to show that our model fits much better than the Beta regression.  相似文献   

12.
13.
A log-linear model is defined for multiway contingency tables with negative multinomial frequency counts. The maximum likelihood estimator of the model parameters and the estimator covariance matrix is given. The likelihood ratio test for the general log-linear hypothesis also is presented.  相似文献   

14.
We study a group lasso estimator for the multivariate linear regression model that accounts for correlated error terms. A block coordinate descent algorithm is used to compute this estimator. We perform a simulation study with categorical data and multivariate time series data, typical settings with a natural grouping among the predictor variables. Our simulation studies show the good performance of the proposed group lasso estimator compared to alternative estimators. We illustrate the method on a time series data set of gene expressions.  相似文献   

15.
A nested-error regression model having both fixed and random effects is introduced to estimate linear parameters of small areas. The model is applicable to data having a proportion of domains where the variable of interest cannot be described by a standard linear mixed model. Algorithms and formulas to fit the model, to calculate EBLUP and to estimate mean-squared errors are given. A Monte Carlo simulation experiment is presented to illustrate the gain of precision obtained by using the proposed model and to obtain some practical conclusions. A motivating application to Spanish Labour Force Survey data is also given.  相似文献   

16.
This article discusses regression analysis of mixed interval-censored failure time data. Such data frequently occur across a variety of settings, including clinical trials, epidemiologic investigations, and many other biomedical studies with a follow-up component. For example, mixed failure times are commonly found in the two largest studies of long-term survivorship after childhood cancer, the datasets that motivated this work. However, most existing methods for failure time data consider only right-censored or only interval-censored failure times, not the more general case where times may be mixed. Additionally, among regression models developed for mixed interval-censored failure times, the proportional hazards formulation is generally assumed. It is well-known that the proportional hazards model may be inappropriate in certain situations, and alternatives are needed to analyze mixed failure time data in such cases. To fill this need, we develop a maximum likelihood estimation procedure for the proportional odds regression model with mixed interval-censored data. We show that the resulting estimators are consistent and asymptotically Gaussian. An extensive simulation study is performed to assess the finite-sample properties of the method, and this investigation indicates that the proposed method works well for many practical situations. We then apply our approach to examine the impact of age at cranial radiation therapy on risk of growth hormone deficiency in long-term survivors of childhood cancer.  相似文献   

17.
In this paper, we propose a new iterative sparse algorithm (ISA) to compute the maximum likelihood estimator (MLE) or penalized MLE of the mixed effects model. The sparse approximation based on the arrow-head (A-H) matrix is one solution which is popularly used in practice. The A-H method provides an easy computation of the inverse of the Hessian matrix and is computationally efficient. However, it often has non-negligible error in approximating the inverse of the Hessian matrix and in the estimation. Unlike the A-H method, in the ISA, the sparse approximation is applied “iteratively” to reduce the approximation error at each Newton Raphson step. The advantages of the ISA over the exact and A-H method are illustrated using several synthetic and real examples.  相似文献   

18.
The proportional hazards model is the most commonly used model in regression analysis of failure time data and has been discussed by many authors under various situations (Kalbfleisch & Prentice, 2002. The Statistical Analysis of Failure Time Data, Wiley, New York). This paper considers the fitting of the model to current status data when there exist competing risks, which often occurs in, for example, medical studies. The maximum likelihood estimates of the unknown parameters are derived and their consistency and convergence rate are established. Also we show that the estimates of regression coefficients are efficient and have asymptotically normal distributions. Simulation studies are conducted to assess the finite sample properties of the estimates and an illustrative example is provided. The Canadian Journal of Statistics © 2009 Statistical Society of Canada  相似文献   

19.
We develop estimates for the parameters of the Dirichlet-multinomial distribution (DMD) when there is insufficient data to obtain maximum likelihood or method of moment estimates known in the literature. We do, however, have supplemetary beta-binomial data pertaining to the marginals of the DMD, and use these data when estimating the DMD parameters. A real situation and data set are given where our estimates are applicable.  相似文献   

20.
A model for the lifetime of a system is considered in which the system is susceptible to simultaneous failures of two or more components, the failures having a common external cause. Three sets of discrete failure data from the US nuclear industry are examined to motivate and illustrate the model derivation: they are for motor-operated valves, cooling fans and emergency diesel generators. To achieve target reliabilities, these components must be placed in systems that have built-in redundancy. Consequently, multiple failures due to a common cause are critical in the risk of core meltdown. Vesely has offered a simple methodology for inference, called the binomial failure rate model: external events are assumed to be governed by a Poisson shock model in which resulting shocks kill X out of m system components, X having a binomial distribution with parameters ( m , p ), 0< p <1. In many applications the binomial failure rate model fits failure data poorly, and the model has not typically been applied to probabilistic risk assessments in the nuclear industry. We introduce a realistic generalization of the binomial failure rate model by assigning a mixing distribution to the unknown parameter p . The distribution is generally identifiable, and its unique nonparametric maximum likelihood estimator can be obtained by using a simple iterative scheme.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号