首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 115 毫秒
1.
Ruiqin Tian 《Statistics》2017,51(5):988-1005
In this paper, empirical likelihood inference for longitudinal data within the framework of partial linear regression models are investigated. The proposed procedures take into consideration the correlation within groups without involving direct estimation of nuisance parameters in the correlation matrix. The empirical likelihood method is used to estimate the regression coefficients and the baseline function, and to construct confidence intervals. A nonparametric version of Wilk's theorem for the limiting distribution of the empirical likelihood ratio is derived. Compared with methods based on normal approximations, the empirical likelihood does not require consistent estimators for the asymptotic variance and bias. The finite sample behaviour of the proposed method is evaluated with simulation and illustrated with an AIDS clinical trial data set.  相似文献   

2.
Abstract.  Comparison of two samples can sometimes be conducted on the basis of analysis of receiver operating characteristic (ROC) curves. A variety of methods of point estimation and confidence intervals for ROC curves have been proposed and well studied. We develop smoothed empirical likelihood-based confidence intervals for ROC curves when the samples are censored and generated from semiparametric models. The resulting empirical log-likelihood function is shown to be asymptotically chi-squared. Simulation studies illustrate that the proposed empirical likelihood confidence interval is advantageous over the normal approximation-based confidence interval. A real data set is analysed using the proposed method.  相似文献   

3.
A smoothed bootstrap method is presented for the purpose of bandwidth selection in nonparametric hazard rate estimation for iid data. In this context, two new bootstrap bandwidth selectors are established based on the exact expression of the bootstrap version of the mean integrated squared error of some approximations of the kernel hazard rate estimator. This is very useful since Monte Carlo approximation is no longer needed for the implementation of the two bootstrap selectors. A simulation study is carried out in order to show the empirical performance of the new bootstrap bandwidths and to compare them with other existing selectors. The methods are illustrated by applying them to a diabetes data set.  相似文献   

4.
In this article, empirical likelihood inferences for semiparametric varying-coefficient partially linear models with longitudinal data are investigated. We propose a groupwise empirical likelihood procedure to handle the inter-series dependence of the longitudinal data. By using residual-adjustment, an empirical likelihood ratio function for the nonparametric component is constructed, and a nonparametric version Wilks' phenomenons is proved. Compared with methods based on normal approximations, the empirical likelihood does not require consistent estimators for the asymptotic variance and bias. A simulation study is undertaken to assess the finite sample performance of the proposed confidence regions.  相似文献   

5.
This paper is concerned with the estimation of a general class of nonlinear panel data models in which the conditional distribution of the dependent variable and the distribution of the heterogeneity factors are arbitrary. In general, exact analytical results for this problem do not exist. Here, Laplace and small-sigma appriximations for the marginal likelihood are presented. The computation of the MLE from both approximations is straightforward. It is shown that the accuracy of the Laplace approximation depends on both the sample size and the variance of the individual effects, whereas the accuracy of the small-sigma approximation is 0(1) with respect to the sample size. The results are applied to count, duration and probit panel data models. The accuracy of the approximations is evaluated through a Monte Carlo simulation experiment. The approximations are also applied in an analysis of youth unemployment in Australia.  相似文献   

6.
The paper shows that many estimation methods, including ML, moments, even-points, empirical c.f. and minimum chi-square, can be regarded as scoring procedures using weighted sums of the discrepancies between observed and expected frequencies The nature of the weights is investigated for many classes of distributions; the study of approximations to the weights clarifies the relationships between estimation methods, and also leads to useful formulae for initial values for ML iteration.  相似文献   

7.
Two approximations to the F-distribution are evaluated in the context of testing for intraclass correlation in the analysis of family data. The evaluation is based on a computation of empirical significance levels and a comparison between p-values associated with these approximations and the corresponding exact p-values. It is found that the approximate methods may give very unsatisfactory results, and exact methods are therefore recommended for general use.  相似文献   

8.
Given two independent samples of size n and m drawn from univariate distributions with unknown densities f and g, respectively, we are interested in identifying subintervals where the two empirical densities deviate significantly from each other. The solution is built by turning the nonparametric density comparison problem into a comparison of two regression curves. Each regression curve is created by binning the original observations into many small size bins, followed by a suitable form of root transformation to the binned data counts. Turned as a regression comparison problem, several nonparametric regression procedures for detection of sparse signals can be applied. Both multiple testing and model selection methods are explored. Furthermore, an approach for estimating larger connected regions where the two empirical densities are significantly different is also derived, based on a scale-space representation. The proposed methods are applied on simulated examples as well as real-life data from biology.  相似文献   

9.
Two-sample comparison of survival times with “cured patients” is of major interest and a challenging issue in many areas, particularly in cancer clinical research. Recently, several authors have proposed various procedures of comparison, including tests of no overall, no short-term and no long-term differences between two samples. In clinical practice, it is often of interest to detect the difference in treatment effects among noncured patients regardless of the difference between cure fractions. In this paper, we propose a statistical test to compare two samples with cured patients and possibly heterogeneous treatment effects based on a class of semi-parametric transformation models, and our main focus is on the survival times of noncured patients. The empirical and quantile processes are used to construct strong approximations for the empirical curves. The two-sample test is then constructed from general least squares estimators derived from these processes. Simulation results show that the proposed test perform well. As an example of application, a set of bladder cancer data is analyzed to illustrate the proposed methods.  相似文献   

10.
The authors propose a bootstrap procedure which estimates the distribution of an estimating function by resampling its terms using bootstrap techniques. Studentized versions of this so‐called estimating function (EF) bootstrap yield methods which are invariant under reparametrizations. This approach often has substantial advantage, both in computation and accuracy, over more traditional bootstrap methods and it applies to a wide class of practical problems where the data are independent but not necessarily identically distributed. The methods allow for simultaneous estimation of vector parameters and their components. The authors use simulations to compare the EF bootstrap with competing methods in several examples including the common means problem and nonlinear regression. They also prove symptotic results showing that the studentized EF bootstrap yields higher order approximations for the whole vector parameter in a wide class of problems.  相似文献   

11.
This paper proposes two methods of estimation for the parameters in a Poisson-exponential model. The proposed methods combine the method of moments with a regression method based on the empirical moment generating function. One of the methods is an adaptation of the mixed-moments procedure of Koutrouvelis & Canavos (1999). The asymptotic distribution of the estimator obtained with this method is derived. Finite-sample comparisons are made with the maximum likelihood estimator and the method of moments. The paper concludes with an exploratory-type analysis of real data based on the empirical moment generating function.  相似文献   

12.
Maximum likelihood (ML) estimation with spatial econometric models is a long-standing problem that finds application in several areas of economic importance. The problem is particularly challenging in the presence of missing data, since there is an implied dependence between all units, irrespective of whether they are observed or not. Out of the several approaches adopted for ML estimation in this context, that of LeSage and Pace [Models for spatially dependent missing data. J Real Estate Financ Econ. 2004;29(2):233–254] stands out as one of the most commonly used with spatial econometric models due to its ability to scale with the number of units. Here, we review their algorithm, and consider several similar alternatives that are also suitable for large datasets. We compare the methods through an extensive empirical study and conclude that, while the approximate approaches are suitable for large sampling ratios, for small sampling ratios the only reliable algorithms are those that yield exact ML or restricted ML estimates.  相似文献   

13.
基于区域间投入产出表的空间静态特征和人口投入产出表的跨期动态特征,本文研究了省区间人口投入产出表的编制,研究的主要内容从编制思路创新、数据估计方法创新及实证检验分析等方面展开。研究结论认为:人口流动跨期特征和空间静态分布特征的结合合理解决了省区间人口投入产出表的编制思路;分类估计省区间人口投入产出表中间流量和非中间流量数据,有效解决了省区间人口投入产出表数据的估计;实证检验及经验数据检验结果验证了本文所编制中国省区间人口投入产出表是可行的,GM-RAS方法估计的省区间人口投入产出表数据是可靠的;实证应用分析发现中国人口迁移特征符合“胡焕庸线”理论,且表现出“胡焕庸线”向东南移动的特征。  相似文献   

14.
The analysis of survival endpoints subject to right-censoring is an important research area in statistics, particularly among econometricians and biostatisticians. The two most popular semiparametric models are the proportional hazards model and the accelerated failure time (AFT) model. Rank-based estimation in the AFT model is computationally challenging due to optimization of a non-smooth loss function. Previous work has shown that rank-based estimators may be written as solutions to linear programming (LP) problems. However, the size of the LP problem is O(n 2+p) subject to n 2 linear constraints, where n denotes sample size and p denotes the dimension of parameters. As n and/or p increases, the feasibility of such solution in practice becomes questionable. Among data mining and statistical learning enthusiasts, there is interest in extending ordinary regression coefficient estimators for low-dimensions into high-dimensional data mining tools through regularization. Applying this recipe to rank-based coefficient estimators leads to formidable optimization problems which may be avoided through smooth approximations to non-smooth functions. We review smooth approximations and quasi-Newton methods for rank-based estimation in AFT models. The computational cost of our method is substantially smaller than the corresponding LP problem and can be applied to small- or large-scale problems similarly. The algorithm described here allows one to couple rank-based estimation for censored data with virtually any regularization and is exemplified through four case studies.  相似文献   

15.
Summary.  Problems of the analysis of data with incomplete observations are all too familiar in statistics. They are doubly difficult if we are also uncertain about the choice of model. We propose a general formulation for the discussion of such problems and develop approximations to the resulting bias of maximum likelihood estimates on the assumption that model departures are small. Loss of efficiency in parameter estimation due to incompleteness in the data has a dual interpretation: the increase in variance when an assumed model is correct; the bias in estimation when the model is incorrect. Examples include non-ignorable missing data, hidden confounders in observational studies and publication bias in meta-analysis. Doubling variances before calculating confidence intervals or test statistics is suggested as a crude way of addressing the possibility of undetectably small departures from the model. The problem of assessing the risk of lung cancer from passive smoking is used as a motivating example.  相似文献   

16.
Motivated by the national evaluation of readmission rates among kidney dialysis facilities in the United States, we evaluate the impact of including discharging hospitals on the estimation of facility-level standardized readmission ratios (SRRs). The estimation of SRRs consists of two steps. First, we model the dependence of readmission events on facilities and patient-level characteristics, with or without an adjustment for discharging hospitals. Second, using results from the models, standardization is achieved by computing the ratio of the number of observed events to the number of expected events assuming a population norm and given the case-mix in that facility. A challenging aspect of our motivating example is that the number of parameters is very large and estimation of high-dimensional parameters is troublesome. To solve this problem, we propose a structured Newton-Raphson algorithm for a logistic fixed effects model and an approximate EM algorithm for the logistic mixed effects model. We consider a re-sampling and simulation technique to obtain p-values for the proposed measures. Finally, our method of identifying outlier facilities involves converting the observed p-values to Z-statistics and using the empirical null distribution, which accounts for overdispersion in the data. The finite-sample properties of proposed measures are examined through simulation studies. The methods developed are applied to national dialysis data. It is our great pleasure to present this paper in honor of Ross Prentice, who has been instrumental in the development of modern methods of modeling and analyzing life history and failure time data, and in the inventive applications of these methods to important national data problem.  相似文献   

17.
In ridge regression, the estimation of ridge parameter k is an important problem. There are several methods available in the literature to do this job some what efficiently. However, no attempts were made to suggest a confidence interval for the ridge parameter using the knwoledge from the data. In this article, we propose a data dependent confidence interval for the ridge parameter k. The method of obtaining the confidence interval is illustrated with the help of a data set. A simulation study indicates that the empirical coverage probability of the suggested confidence intervals are quite high.  相似文献   

18.
In this paper, we consider the problem of estimating the number of components of a superimposed nonlinear sinusoids model of a signal in the presence of additive noise. We propose and provide a detailed empirical comparison of robust methods for estimation of the number of components. The proposed methods, which are robust modifications of the commonly used information theoretic criteria, are based on various M-estimator approaches and are robust with respect to outliers present in the data and heavy-tailed noise. The proposed methods are compared with the usual non-robust methods through extensive simulations under varied model scenarios. We also present real signal analysis of two speech signals to show the usefulness of the proposed methodology.  相似文献   

19.
This paper deals with estimation problems under an extended growth curve model with two hierarchical within-individuals design matrices. The model in cludes the one whose mean structure consists of polynomial growth curves with two different degrees. First we propose certain simple estimators of the mean and covariance parameters which are closely related to the MEE's. Some basic properties of the estimators are given. Simultaneous confidence intervals are constructed, based on the estimators, for each and both of two growth curves. We give asymptotic approximations for the corresponding critical points. A numerical example is also given.  相似文献   

20.
Covariance matrices, or in general matrices of sums of squares and cross-products, are used as input in many multivariate analyses techniques. The eigenvalues of these matrices play an important role in the statistical analysis of data including estimation and hypotheses testing. It has been recognized that one or few observations can exert an undue influence on the eigenvalues of a covariance matrix. The relationship between the eigenvalues of the covariance matrix computed from all data and the eigenvalues of the perturbed covariance matrix (a covariance matrix computed after a small subset of the observations has been deleted) cannot in general be written in closed-form. Two methods for approximating the eigenvalues of a perturbed covariance matrix have been suggested by Hadi (1988) and Wang and Nyquist (1991) for the case of a perturbation by a single observation. In this paper we improve on these two methods and give some additional theoretical results that may give further insight into the problem. We also compare the two improved approximations in terms of their accuracies.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号