首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
ABSTRACT

The log-logistic distribution is commonly used to model lifetime data. We propose a wider distribution, named the exponentiated log-logistic geometric distribution, based on a double activation approach. We obtain the quantile function, ordinary moments, and generating function. The method of maximum likelihood is used to estimate the model parameters. We propose a new extended regression model based on the logarithm of the exponentiated log-logistic geometric distribution. This regression model can be very useful in the analysis of real data and could provide better fits than other special regression models. The potentiality of the new models is illustrated by means of two applications to real lifetime data sets.  相似文献   

2.
This article describes two bivariate geometric distributions. We investigate characterizations of bivariate geometric distributions using conditional failure rates and study properties of the bivariate geometric distributions. The bivariate models are fitted to real-life data using the Method of Moments, Maximum Likelihood, and Bayes Estimators. Two methods of moments estimators, in each bivariate geometric model, are compared and evaluated for their performance in terms of bias vector and covariance matrix. This comparison is done through a Monte Carlo simulation. Chi-square goodness-of-fit tests are used to evaluate model performance.  相似文献   

3.
We apply geometric programming, developed by Duffin, Peterson and Zener (1967), to the optimal allocation of stratified samples with several variance constraints arising from several estimates of deficiency rates in the quality control of administrative decisions. We develop also a method for imposing constraints on sample sizes to equalize workloads over time, as required by the practicalities of clerical work for quality control.

We allocate samples by an extension of the work of Neyman (1934), following the exposition of Cochran (1977). Davis and Schwartz (1987) developed methods for multiconstraint Neyman allocation by geometric programming for integrated sampling. They also applied geometric programming to Neyman allocation of a sample for estimating college enrollments by Cornell (1947) and Cochran (1977). This paper continues the application of geometric programming to Neyman allocation with multiple constraints on variances and workloads and minimpal sampling costs.  相似文献   

4.
In this article, we consider clustering based on principal component analysis (PCA) for high-dimensional mixture models. We present theoretical reasons why PCA is effective for clustering high-dimensional data. First, we derive a geometric representation of high-dimension, low-sample-size (HDLSS) data taken from a two-class mixture model. With the help of the geometric representation, we give geometric consistency properties of sample principal component scores in the HDLSS context. We develop ideas of the geometric representation and provide geometric consistency properties for multiclass mixture models. We show that PCA can cluster HDLSS data under certain conditions in a surprisingly explicit way. Finally, we demonstrate the performance of the clustering using gene expression datasets.  相似文献   

5.
Correspondence analysis (CA) and nonsymmetric correspondence analysis are based on generalized singular value decomposition, and, in general, they are not equivalent. Taxicab correspondence analysis (TCA) is a \(\hbox {L}_{1}\) variant of CA, and it is based on the generalized taxicab singular value decomposition (GTSVD). Our aim is to study the taxicab variant of nonsymmetric correspondence analysis. We find that for diagonal metric matrices GTSVDs of a given data set are equivalent; from which we deduce the equivalence of TCA and taxicab nonsymmetric correspondence analysis. We also attempt to show that TCA stays as close as possible to the original correspondence matrix without calculating a dissimilarity (or similarity) measure between rows or columns. Further, we discuss some new geometric and distance aspects of TCA.  相似文献   

6.
This paper reviews two types of geometric methods proposed in recent years for defining statistical decision rules based on 2-dimensional parameters that characterize treatment effect in a medical setting. A common example is that of making decisions, such as comparing treatments or selecting a best dose, based on both the probability of efficacy and the probability toxicity. In most applications, the 2-dimensional parameter is defined in terms of a model parameter of higher dimension including effects of treatment and possibly covariates. Each method uses a geometric construct in the 2-dimensional parameter space based on a set of elicited parameter pairs as a basis for defining decision rules. The first construct is a family of contours that partitions the parameter space, with the contours constructed so that all parameter pairs on a given contour are equally desirable. The partition is used to define statistical decision rules that discriminate between parameter pairs in term of their desirabilities. The second construct is a convex 2-dimensional set of desirable parameter pairs, with decisions based on posterior probabilities of this set for given combinations of treatments and covariates under a Bayesian formulation. A general framework for all of these methods is provided, and each method is illustrated by one or more applications.  相似文献   

7.
In multivariate statistics, estimation of the covariance or correlation matrix is of crucial importance. Computational and other arguments often lead to the use of coordinate-dependent estimators, yielding matrices that are symmetric but not positive semidefinite. We briefly discuss existing methods, based on shrinking, for transforming such matrices into positive semidefinite matrices, A simple method based on eigenvalues is also considered. Taking into account the geometric structure of correlation matrices, a new method is proposed which uses techniques similar to those of multidimensional scaling.  相似文献   

8.
通过逆抽样过程获得的分布又称为负二项分布,在流行病学研究和二分类变量分布的研究中应用极为广泛。因此,提出两种基于梯度统计量的逆抽样下风险差的置信区间的构建方法,分别依据风险差的极大似然估计(MLE)和方差最小无偏一致估计量(UMVUE)。与现有的WALD方法和得分方法相比,该方法所构建置信区间的优点在于:置信区间构建方法既不需要计算Fisher信息阵也不需要计算其逆矩阵,可使计算得以大大简化;对所提出的基于梯度统计量的置信区间构建方法进行蒙特卡洛模拟研究,模拟结果表明提出的构建方法可以得到很好的覆盖概率和较短的区间宽度。  相似文献   

9.
A semicompeting risks problem involves two-types of events: a nonterminal and a terminal event (death). Typically, the nonterminal event is the focus of the study, but the terminal event can preclude the occurrence of the nonterminal event. Semicompeting risks are ubiquitous in studies of aging. Examples of semicompeting risk dyads include: dementia and death, frailty syndrome and death, disability and death, and nursing home placement and death. Semicompeting risk models can be divided into two broad classes: models based only on observables quantities (class \(\mathcal {O}\) ) and those based on potential (latent) failure times (class \(\mathcal {L}\) ). The classical illness-death model belongs to class \(\mathcal {O}\) . This model is a special case of the multistate models, which has been an active area of methodology development. During the past decade and a half, there has also been a flurry of methodological activity on semicompeting risks based on latent failure times ( \(\mathcal {L}\) models). These advances notwithstanding, the semicompeting risks methodology has not penetrated biomedical research, in general, and gerontological research, in particular. Some possible reasons for this lack of uptake are: the methods are relatively new and sophisticated, conceptual problems associated with potential failure time models are difficult to overcome, paucity of expository articles aimed at educating practitioners, and non-availability of readily usable software. The main goals of this review article are: (i) to describe the major types of semicompeting risks problems arising in aging research, (ii) to provide a brief survey of the semicompeting risks methods, (iii) to suggest appropriate methods for addressing the problems in aging research, (iv) to highlight areas where more work is needed, and (v) to suggest ways to facilitate the uptake of the semicompeting risks methodology by the broader biomedical research community.  相似文献   

10.
Generalized exponential, geometric extreme exponential and Weibull distributions are three non-negative skewed distributions that are suitable for analysing lifetime data. We present diagnostic tools based on the likelihood ratio test (LRT) and the minimum Kolmogorov distance (KD) method to discriminate between these models. Probability of correct selection has been calculated for each model and for several combinations of shape parameters and sample sizes using Monte Carlo simulation. Application of LRT and KD discrimination methods to some real data sets has also been studied.  相似文献   

11.
We present a new statistical framework for landmark ?>curve-based image registration and surface reconstruction. The proposed method first elastically aligns geometric features (continuous, parameterized curves) to compute local deformations, and then uses a Gaussian random field model to estimate the full deformation vector field as a spatial stochastic process on the entire surface or image domain. The statistical estimation is performed using two different methods: maximum likelihood and Bayesian inference via Markov Chain Monte Carlo sampling. The resulting deformations accurately match corresponding curve regions while also being sufficiently smooth over the entire domain. We present several qualitative and quantitative evaluations of the proposed method on both synthetic and real data. We apply our approach to two different tasks on real data: (1) multimodal medical image registration, and (2) anatomical and pottery surface reconstruction.  相似文献   

12.
Model selection methods are important to identify the best approximating model. To identify the best meaningful model, purpose of the model should be clearly pre-stated. The focus of this paper is model selection when the modelling purpose is classification. We propose a new model selection approach designed for logistic regression model selection where main modelling purpose is classification. The method is based on the distance between the two clustering trees. We also question and evaluate the performances of conventional model selection methods based on information theory concepts in determining best logistic regression classifier. An extensive simulation study is used to assess the finite sample performances of the cluster tree based and the information theoretic model selection methods. Simulations are adjusted for whether the true model is in the candidate set or not. Results show that the new approach is highly promising. Finally, they are applied to a real data set to select a binary model as a means of classifying the subjects with respect to their risk of breast cancer.  相似文献   

13.
We apply geometric programming, developed by Duffin, Peterson Zener (1967), to the optimal allocation of stratified samples. As an introduction, we show how geometric programming is used to allocate samples according to Neyman (1934), using the data of Cornell (1947) and following the exposition of Cochran (1953).

Then we use geometric programming to allocate an integrated sample introduced by Schwartz (1978) for more efficient sampling of three U. S. Federal welfare quality control systems, Aid to Families with Dependent Children, Food Stamps and Medicaid.

We develop methods for setting up the allocation problem, interpreting it as a geometric programming primal problem, transforming it to the corresponding dual problem, solving that, and finding the sample sizes required in the allocation problem. We show that the integrated sample saves sampling costs.  相似文献   

14.
永续盘存法核算资本存量的两种途径及其比较   总被引:1,自引:0,他引:1  
资本存量核算最常用的方法是永续盘存法。永续盘存法的实质是通过对过去购置的并估算出使用年限的资产进行累加来完成的,核算中需要考虑资产的相对效率与重置需求、资本的租赁价格与资本折旧。为此探讨永续盘存法下的资本存量核算的传统途径与新途径,针对传统途径和几何模式相对效率与双曲线相对效率模式下的新途径对中国资本存量净额进行估算和比较,认为中国的资本核算可以采用新途径的思路来进行。  相似文献   

15.
We address the issue of recovering the structure of large sparse directed acyclic graphs from noisy observations of the system. We propose a novel procedure based on a specific formulation of the \(\ell _1\)-norm regularized maximum likelihood, which decomposes the graph estimation into two optimization sub-problems: topological structure and node order learning. We provide convergence inequalities for the graph estimator, as well as an algorithm to solve the induced optimization problem, in the form of a convex program embedded in a genetic algorithm. We apply our method to various data sets (including data from the DREAM4 challenge) and show that it compares favorably to state-of-the-art methods. This algorithm is available on CRAN as the R package GADAG.  相似文献   

16.
The study of count data time series has been active in the past decade, mainly in theory and model construction. There are different ways to construct time series models with a geometric autocorrelation function, and a given univariate margin such as negative binomial. In this paper, we investigate negative binomial time series models based on the binomial thinning and two other expectation thinning operators, and show how they differ in conditional variance or heteroscedasticity. Since the model construction is in terms of probability generating functions, typically, the relevant conditional probability mass functions do not have explicit forms. In order to do simulations, likelihood inference, graphical diagnostics and prediction, we use a numerical method for inversion of characteristic functions. We illustrate the numerical methods and compare the various negative binomial time series models for a real data example.  相似文献   

17.
Summary. We use a multipath (multistate) model to describe data with multiple end points. Statistical inference based on the intermediate end point is challenging because of the problems of nonidentifiability and dependent censoring. We study nonparametric estimation for the path probability and the sojourn time distributions between the states. The methodology proposed can be applied to analyse cure models which account for the competing risk of death. Asymptotic properties of the estimators proposed are derived. Simulation shows that the methods proposed have good finite sample performance. The methodology is applied to two data sets.  相似文献   

18.
We consider the statistical evaluation and estimation of vaccine efficacy when the protective effect wanes with time. We reanalyse data from a 5-year trial of two oral cholera vaccines in Matlab, Bangladesh. In this field trial, one vaccine appears to confer better initial protection than the other, but neither appears to offer protection for a period longer than about 3 years. Time-dependent vaccine effects are estimated by obtaining smooth estimates of a time-varying relative risk RR( t ) using survival analysis. We compare two approaches based on the Cox model in terms of their strategies for detecting time-varying vaccine effects, and their estimation techniques for obtaining a time-dependent RR( t ) estimate. These methods allow an exploration of time-varying vaccine effects while making minimal parametric assumptions about the functional form of RR( t ) for vaccinated compared wit unvaccinated subjects.  相似文献   

19.
Many characterization results of the bivariate exponential distribution and the bivariate geometric distribution have been proved in the literature. Recently Nair and Nair (1988b, Ann. Inst. Statist. Math. 40 (2), 267–271) obtained a characterization result of the Gumbel bivariate exponential distribution and a bivariate geometric distribution based on truncated moments. In this note, we extend the results of Nair and Nair (1988b) to obtain a general result, characterizing these two bivariate distributions based on the truncated expectation of a function h, satisfying some mild conditions.  相似文献   

20.
Summary.  As a part of the EUREDIT project new methods to detect multivariate outliers in incomplete survey data have been developed. These methods are the first to work with sampling weights and to be able to cope with missing values. Two of these methods are presented here. The epidemic algorithm simulates the propagation of a disease through a population and uses extreme infection times to find outlying observations. Transformed rank correlations are robust estimates of the centre and the scatter of the data. They use a geometric transformation that is based on the rank correlation matrix. The estimates are used to define a Mahalanobis distance that reveals outliers. The two methods are applied to a small data set and to one of the evaluation data sets of the EUREDIT project.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号