首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
We investigate empirical likelihood for the additive hazards model with current status data. An empirical log-likelihood ratio for a vector or subvector of regression parameters is defined and its limiting distribution is shown to be a standard chi-squared distribution. The proposed inference procedure enables us to make empirical likelihood-based inference for the regression parameters. Finite sample performance of the proposed method is assessed in simulation studies to compare with that of a normal approximation method, it shows that the empirical likelihood method provides more accurate inference than the normal approximation method. A real data example is used for illustration.  相似文献   

2.
Hierarchical models are widely-used to characterize the performance of individual healthcare providers. However, little attention has been devoted to system-wide performance evaluations, the goals of which include identifying extreme (e.g., top 10%) provider performance and developing statistical benchmarks to define high-quality care. Obtaining optimal estimates of these quantities requires estimating the empirical distribution function (EDF) of provider-specific parameters that generate the dataset under consideration. However, the difficulty of obtaining uncertainty bounds for a square-error loss minimizing EDF estimate has hindered its use in system-wide performance evaluations. We therefore develop and study a percentile-based EDF estimate for univariate provider-specific parameters. We compute order statistics of samples drawn from the posterior distribution of provider-specific parameters to obtain relevant uncertainty assessments of an EDF estimate and its features, such as thresholds and percentiles. We apply our method to data from the Medicare End Stage Renal Disease (ESRD) Program, a health insurance program for people with irreversible kidney failure. We highlight the risk of misclassifying providers as exceptionally good or poor performers when uncertainty in statistical benchmark estimates is ignored. Given the high stakes of performance evaluations, statistical benchmarks should be accompanied by precision estimates.  相似文献   

3.
We extend univariate regression quantile splines to problems with several covariates. We adopt an ANOVA-type decomposition approach with main effects captured by linear splines and second-order ‘interactions’ modeled by bi-linear tensor-product splines. Both univariate linear splines and bi-linear tensor-product splines are optimal when fidelity to data are balanced by a roughness penalty on the fitted function. The problem of sub-model selection and asymptotic justification for using a smaller sub-space of the spline functions in the approximation are discussed. Two examples are considered to illustrate the empirical performance of the proposed methods.  相似文献   

4.
In this paper, we address the problem of simulating from a data-generating process for which the observed data do not follow a regular probability distribution. One existing method for doing this is bootstrapping, but it is incapable of interpolating between observed data. For univariate or bivariate data, in which a mixture structure can easily be identified, we could instead simulate from a Gaussian mixture model. In general, though, we would have the problem of identifying and estimating the mixture model. Instead of these, we introduce a non-parametric method for simulating datasets like this: Kernel Carlo Simulation. Our algorithm begins by using kernel density estimation to build a target probability distribution. Then, an envelope function that is guaranteed to be higher than the target distribution is created. We then use simple accept–reject sampling. Our approach is more flexible than others, can simulate intelligently across gaps in the data, and requires no subjective modelling decisions. With several univariate and multivariate examples, we show that our method returns simulated datasets that, compared with the observed data, retain the covariance structures and have distributional characteristics that are remarkably similar.  相似文献   

5.
In the presence of univariate censoring, a class of nonparametric estimators is proposed for linear functionals of a bivariate distribution of paired failure times. The estimators are shown to be root-n consistent and asymptotically normal. An adjusted empirical log-likelihood ratio statistic is developed and proved to follow a chi-square distribution asymptotically. Two types of confidence intervals, based on the normal approximation method and the empirical likelihood method, respectively, are constructed to make inference about the linear functionals. Their performance is evaluated in several simulation studies and a real example.  相似文献   

6.
A goodness-of-fit test for multivariate normality is proposed which is based on Shapiro–Wilk's statistic for univariate normality and on an empirical standardization of the observations. The critical values can be approximated by using a transformation of the univariate standard normal distribution. A Monte Carlo study reveals that this test has a better power performance than some of the best known tests for multinormality against a wide range of alternatives.  相似文献   

7.
The currently existing estimation methods and goodness-of-fit tests for the Cox model mainly deal with right censored data, but they do not have direct extension to other complicated types of censored data, such as doubly censored data, interval censored data, partly interval-censored data, bivariate right censored data, etc. In this article, we apply the empirical likelihood approach to the Cox model with complete sample, derive the semiparametric maximum likelihood estimators (SPMLE) for the Cox regression parameter and the baseline distribution function, and establish the asymptotic consistency of the SPMLE. Via the functional plug-in method, these results are extended in a unified approach to doubly censored data, partly interval-censored data, and bivariate data under univariate or bivariate right censoring. For these types of censored data mentioned, the estimation procedures developed here naturally lead to Kolmogorov-Smirnov goodness-of-fit tests for the Cox model. Some simulation results are presented.  相似文献   

8.
In this paper, we study the multi-class differential gene expression detection for microarray data. We propose a likelihood-based approach to estimating an empirical null distribution to incorporate gene interactions and provide a more accurate false-positive control than the commonly used permutation or theoretical null distribution-based approach. We propose to rank important genes by p-values or local false discovery rate based on the estimated empirical null distribution. Through simulations and application to lung transplant microarray data, we illustrate the competitive performance of the proposed method.  相似文献   

9.
Within the context of mixture modeling, the normal distribution is typically used as the components distribution. However, if a cluster is skewed or heavy tailed, then the normal distribution will be inefficient and many may be needed to model a single cluster. In this paper, we present an attempt to solve this problem. We define a cluster, in the absence of further information, to be a group of data which can be modeled by a unimodal density function. Hence, our intention is to use a family of univariate distribution functions, to replace the normal, for which the only constraint is unimodality. With this aim, we devise a new family of nonparametric unimodal distributions, which has large support over the space of univariate unimodal distributions. The difficult aspect of the Bayesian model is to construct a suitable MCMC algorithm to sample from the correct posterior distribution. The key will be the introduction of strategic latent variables and the use of the Product Space view of Reversible Jump methodology.  相似文献   

10.
There are situations in the analysis of failure time or lifetime data where the censoring times of unfailed units are missing. The non-parametric estimator of the lifetime distribution for such data is available in literature. In this paper we consider an extension of this situation to the univariate and bivariate competing risk setups. The maximum likelihood and simple moment estimators of cause specific distribution functions in both univariate and bivariate situations are developed. A simulation study is carried out to assess the performance of the estimators. Finally, we illustrate the method with real data set.  相似文献   

11.
On some study of skew-t distributions   总被引:1,自引:0,他引:1  
Abstract

In this note, through ratio of independent random variables, new families of univariate and bivariate skew-t distributions are introduced. Probability density function for each skew-t distribution will be given. We also derive explicit forms of moments of the univariate skew-t distribution and recurrence relations for its cumulative distribution function. Finally we illustrate the flexibility of this class of distributions with applications to a simulated data and the volcanos heights data.  相似文献   

12.
In practice, a financial or actuarial data set may be a skewed or heavy-tailed and this motivates us to study a class of distribution functions in risk management theory that provide more information about these characteristics resulting in a more accurate risk analysis. In this paper, we consider a multivariate tail conditional expectation (MTCE) for multivariate scale mixtures of skew-normal (SMSN) distributions. This class of distributions contains skewed distributions and some members of this class can be used to analyse heavy-tailed data sets. We also provide a closed form for TCE in a univariate skew-normal distribution framework. Numerical examples are also provided for illustration.  相似文献   

13.
We consider here a univariate skew-elliptical distribution, which is a special case of the unified multivariate skew-elliptical distribution studied recently by Arellano-Valle and Azzalini (2006) [1]. We then derive the exact distribution of a linear combination of a variable and order statistics from the other two variables in the case of a trivariate elliptical distribution. We show that the cumulative distribution function (cdf) of this linear combination is a mixture of the univariate skew-elliptical distribution functions.  相似文献   

14.
Small area estimation plays a prominent role in survey sampling due to a growing demand for reliable small area estimates from both public and private sectors. Popularity of model-based inference is increasing in survey sampling, particularly, in small area estimation. The estimates of the small area parameters can profitably ‘borrow strength’ from data on related multiple characteristics and/or auxiliary variables from other neighboring areas through appropriate models. Fay (1987, Small Area Statistics, Wiley, New York, pp. 91–102) proposed multivariate regression for small area estimation of multiple characteristics. The success of this modeling rests essentially on the strength of correlation of these dependent variables. To estimate small area mean vectors of multiple characteristics, multivariate modeling has been proposed in the literature via a multivariate variance components model. We use this approach to empirical best linear unbiased and empirical Bayes prediction of small area mean vectors. We use data from Battese et al. (1988, J. Amer. Statist. Assoc. 83, 28 –36) to conduct a simulation which shows that the multivariate approach may achieve substantial improvement over the usual univariate approach.  相似文献   

15.
A rule of thumb for testing symmetry of an unknown univariate continuous distribution against the alternative of a long right tail is proposed. Our proposed test is based on the concept of exceedance statistic and is ad hoc in nature. Exact performances of the proposed rule are investigated in detail. Some results from an asymptotic point of view are also provided. We compare our proposed test with several classical tests which are practically applicable and are known to be exact or nearly distribution free. We see that the proposed rule is better than most of the existing tests for symmetry and can be applied with ease. An illustration with real data is provided.  相似文献   

16.
The rootogram is a graphical tool associated with the work of J. W. Tukey that was originally used for assessing goodness of fit of univariate distributions. Here, we extend the rootogram to regression models and show that this is particularly useful for diagnosing and treating issues such as overdispersion and/or excess zeros in count data models. We also introduce a weighted version of the rootogram that can be applied out of sample or to (weighted) subsets of the data, for example, in finite mixture models. An empirical illustration revisiting a well-known dataset from ethology is included, for which a negative binomial hurdle model is employed. Supplementary materials providing two further illustrations are available online: the first, using data from public health, employs a two-component finite mixture of negative binomial models; the second, using data from finance, involves underdispersion. An R implementation of our tools is available in the R package countreg. It also contains the data and replication code.  相似文献   

17.
Heterogeneity in lifetime data may be modelled by multiplying an individual's hazard by an unobserved frailty. We test for the presence of frailty of this kind in univariate and bivariate data with Weibull distributed lifetimes, using statistics based on the ordered Cox–Snell residuals from the null model of no frailty. The form of the statistics is suggested by outlier testing in the gamma distribution. We find through simulation that the sum of the k largest or k smallest order statistics, for suitably chosen k, provides a powerful test when the frailty distribution is assumed to be gamma or positive stable, respectively. We provide recommended values of k for sample sizes up to 100 and simple formulae for estimated critical values for tests at the 5% level.  相似文献   

18.
In financial analysis it is useful to study the dependence between two or more time series as well as the temporal dependence in a univariate time series. This article is concerned with the statistical modeling of the dependence structure in a univariate financial time series using the concept of copula. We treat the series of financial returns as a first order Markov process. The Archimedean two-parameter BB7 copula is adopted to describe the underlying dependence structure between two consecutive returns, while the log-Dagum distribution is employed to model the margins marked by skewness and kurtosis. A simulation study is carried out to evaluate the performance of the maximum likelihood estimates. Furthermore, we apply the model to the daily returns of four stocks and, finally, we illustrate how its fitting to data can be improved when the dependence between consecutive returns is described through a copula function.  相似文献   

19.
We derive an identity for nonparametric maximum likelihood estimators (NPMLE) and regularized MLEs in censored data models which expresses the standardized maximum likelihood estimator in terms of the standardized empirical process. This identity provides an effective starting point in proving both consistency and efficiency of NPMLE and regularized MLE. The identity and corresponding method for proving efficiency is illustrated for the NPMLE in the univariate right-censored data model, the regularized MLE in the current status data model and for an implicit NPMLE based on a mixture of right-censored and current status data. Furthermore, a general algorithm for estimation of the limiting variance of the NPMLE is provided. This revised version was published online in July 2006 with corrections to the Cover Date.  相似文献   

20.
Assuming a statistical model in which the joint distribution of the unobservable errors is drawn from independent univariate Student t's that are identically and symmetrically distributed, the sampling performance of traditional robust estimators and a family of Stein-like estimators are compared and evaluated. These results suggest that under thick-tailed distributions, the relative sampling performances and risk characteristics for a range of nonconventional Stein estimators remains approximately the same as in the case of their normal counterparts. The empirical risk implications of misspecifying the error distribution are investigated.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号