首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
The use of the correlation coefficient is suggested as a technique for summarizing and objectively evaluating the information contained in probability plots. Goodness-of-fit tests are constructed using this technique for several commonly used plotting positions for the normal distribution. Empirical sampling methods are used to construct the null distribution for these tests, which are then compared on the basis of power against certain nonnormal alternatives. Commonly used regression tests of fit are also included in the comparisons. The results indicate that use of the plotting position pi = (i - .375)/(n + .25) yields a competitive regression test of fit for normality.  相似文献   

2.
Most biomedical research is carried out using longitudinal studies. The method of generalized estimating equations (GEEs) introduced by Liang and Zeger [Longitudinal data analysis using generalized linear models, Biometrika 73 (1986), pp. 13–22] and Zeger and Liang [Longitudinal data analysis for discrete and continuous outcomes, Biometrics 42 (1986), pp. 121–130] has become a standard method for analyzing non-normal longitudinal data. Since then, a large variety of GEEs have been proposed. However, the model diagnostic problem has not been explored intensively. Oh et al. [Modeldiagnostic plots for repeated measures data using the generalized estimating equations approach, Comput. Statist. Data Anal. 53 (2008), pp. 222–232] proposed residual plots based on the quantile–quantile (Q–Q) plots of the χ2-distribution for repeated-measures data using the GEE methodology. They considered the Pearson, Anscombe and deviance residuals. In this work, we propose to extend this graphical diagnostic using a generalized residual. A simulation study is presented as well as two examples illustrating the proposed generalized Q–Q plots.  相似文献   

3.
In this article, it is explicitly demonstrated that the probability of non exceedance of the mth value in n order ranked events equals m/(n + 1). Consequently, the plotting position in the extreme value analysis should be considered not as an estimate, but to be equal to m/(n + 1), regardless of the parent distribution and the application. The many other suggested plotting formulas and numerical methods to determine them should thus be abandoned. The article is intended to mark the end of the century-long controversial discussion on the plotting positions.  相似文献   

4.
Formulas for plotting probability and techniques for subjectively drawing lines on probability plots are reviewed. A method is presented for plotting data and drawing an objective line on the probability plot to obtain a test of the distributional assumption.  相似文献   

5.
The quantile–quantile plot is widely used to check normality. The plot depends on the plotting positions. Many commonly used plotting positions do not depend on the sample values. We propose an adaptive plotting position that depends on the relative distances of the two neighbouring sample values. The correlation coefficient obtained from the adaptive plotting position is used to test normality. The test using the adaptive plotting position is better than the Shapiro–Wilk W test for small samples and has larger power than Hazen's and Blom's plotting positions for symmetric alternatives with shorter tail than normal and skewed alternatives when n is 20 or larger. The Brown–Hettmansperger T* test is designed for detecting bad tail behaviour, so it does not have power for symmetric alternatives with shorter tail than normal, but it is generally better than the other tests when β2 is greater than 3.25.  相似文献   

6.
Probability paper was used as early as 1896, and was mentioned in the literature more than 30 times before 1950, mainly by hydrologists, most of whom used the plotting position (i-0.5)/n proposed by Hazen (1914). Gumbel (1942a) considered the modal position (i-1)/(n-1) and the mean position i/(n+1) [the latter proposed by Weibull (1939a,b)], and chose the latter. Lebedev (1952) and others proposed the use of (i-0.3)/(n+0.4), which is approximately the median position advocated by Johnson (1951). Blom (1958) sug-gested (i-α)/(n-2α+1), where a is a constant (usually 0 ≤ α ≤ 1), which includes all of the above plotting positions as special cases. Moreover, by proper choice of α, one can approximate F[E(xi)], the position proposed by Kimball (1946), for any distri-bution of interest. Gumbel (1954) stated five postulates which plotting positions should satisfy. Chernoff & Lieberman (1954) discussed the optimum choice of plotting positions in various situ-ations. It is clear that the optimum plotting position depends on the use that is to be made of the results and may also depend on the underlying distribution. The author endeavors to formulate recommendations as to the best choice in various situations.  相似文献   

7.
This article presents a universal quantile-quantile (QQ) plot that may be used to assess the fit of a family of absolutely continuous distribution functions in a possibly non-homogeneous population. This plot is more general than probability plotting papers because it may be used for distributions having more than two parameters. It is also more general than standard quantile-quantile plots because it may be used for families of not-necessarily identical distributions. In particular, the universal QQ plot may be used in the context of non-homogeneous Poisson processes, generalized linear models, and other general models.  相似文献   

8.
In reliability studies, one typically would assume a lifetime distribution for the units under study and then carry out the required analysis. One popular choice for the lifetime distribution is the family of two-parameter Weibull distributions (with scale and shape parameters) which, through a logarithmic transformation, can be transformed to the family of two-parameter extreme value distributions (with location and scale parameters). In carrying out a parametric analysis of this type, it is highly desirable to be able to test the validity of such a model assumption. A basic tool that is useful for this purpose is a quantile–quantile (QQ) plot, but in its use, the issue of the choice of plotting position arises. Here, by adopting the optimal plotting points based on Pitman closeness criterion proposed recently by Balakrishnan et al. (2010b Balakrishnan , N. , Davies , K. F. , Keating , J. P. , Mason , R. L. ( 2010b ). Computation of optimal plotting points based on Pitman Closeness with an application to goodness of fit for location-scale families. Submitted to Computational Statistics & Data Analysis.  [Google Scholar]), and referred to as simultaneous closeness probability (SCP) plotting points, we propose a correlation-type goodness of fit test for the extreme value distribution. We compute the SCP plotting points for various sample sizes and use them to determine the mean, standard deviation and critical values for the proposed correlation-type test statistic. Using these critical values, we carry out a power study, similar to the one carried out by Kinnison (1989 Kinnison , R. ( 1989 ). Correlation coefficient goodness of fit test for extreme value distribution . The American Statistician 43 : 98100 .[Taylor & Francis Online], [Web of Science ®] [Google Scholar]), through which we demonstrate that the use of SCP plotting points results in better power than with the use of mean ranks as plotting points and nearly the same power as with the use of median ranks. We then demonstrate the use of the SCP plotting points and the associated correlation-type test for Weibull analysis with an illustrative example. Finally, for the sake of comparison, we also adapt two statistics proposed by Gan and Koehler (1990 Gan , F. F. , Koehler , K. J. ( 1990 ). Goodness of fit based on P-P probability plots . Technometrics 32 : 289303 .[Taylor & Francis Online], [Web of Science ®] [Google Scholar]), in the context of probability–probability (PP) plots, based on SCP plotting points and compare their performance to those based on mean ranks. The empirical study also reveals that the tests from the QQ plot have better power than those from the PP plot.  相似文献   

9.
《随机性模型》2013,29(4):425-447
Abstract

In this paper, we define a birth–death‐modulated Markovian arrival process (BDMMAP) as a Markovian arrival process (MAP) with an underlying birth–death process. It is proved that the zeros of det(zI ? A(z)) in the unit disk are real and simple. In order to analyze a BDMMAP/G/1 queue, two spectral methods are proposed. The first one is a bisection method for calculation of the zeros from which the boundary vector is derived. The second one is the Fourier inversion transform of the probability generating function for the calculation of the stationary probability distribution of the queue length. Eigenvalues required in this calculation are obtained by the Duran–Kerner–Aberth (DKA) method. For numerical examples, the stationary probability distribution of the queue length is calculated by using the spectral methods. Comparisons of the spectral methods with the currently best methods available are discussed.  相似文献   

10.
The problem of goodness of fit of a lognormal distribution is usually reduced to testing goodness of fit of the logarithmic data to a normal distribution. In this paper, new goodness-of-fit tests for a lognormal distribution are proposed. The new procedures make use of a characterization property of the lognormal distribution which states that the Kullback–Leibler measure of divergence between a probability density function (p.d.f) and its r-size weighted p.d.f is symmetric only for the lognormal distribution [Tzavelas G, Economou P. Characterization properties of the log-normal distribution obtained with the help of divergence measures. Stat Probab Lett. 2012;82(10):1837–1840]. A simulation study examines the performance of the new procedures in comparison with existing goodness-of-fit tests for the lognormal distribution. Finally, two well-known data sets are used to illustrate the methods developed.  相似文献   

11.
Multinomial logit (also termed multi-logit) models permit the analysis of the statistical relation between a categorical response variable and a set of explicative variables (called covariates or regressors). Although multinomial logit is widely used in both the social and economic sciences, the interpretation of regression coefficients may be tricky, as the effect of covariates on the probability distribution of the response variable is nonconstant and difficult to quantify. The ternary plots illustrated in this article aim at facilitating the interpretation of regression coefficients and permit the effect of covariates (either singularly or jointly considered) on the probability distribution of the dependent variable to be quantified. Ternary plots can be drawn both for ordered and for unordered categorical dependent variables, when the number of possible outcomes equals three (trinomial response variable); these plots allow not only to represent the covariate effects over the whole parameter space of the dependent variable but also to compare the covariate effects of any given individual profile. The method is illustrated and discussed through analysis of a dataset concerning the transition of master’s graduates of the University of Trento (Italy) from university to employment.  相似文献   

12.
ABSTRACT

We propose a new generalized geometric distribution which permits inflation/deflation of the zero count probability and study some of its properties. We also present an actuarial application of this distribution and fit it to three datasets used by other researchers. It is observed that the proposed distribution fits reasonably well to these data. Further, in a regression setup, the performance of this distribution is studied vis–a–vis other competing distributions used for explaining variability in a response variable.  相似文献   

13.
The speed of convergence of the distribution of the normalized maximum, of a sample of independent and identically distributed random variables, to its asymptotic distribution is considered in this article. Assuming that the cumulative distribution function of the random variables is known, the error committed replacing the actual distribution of the normalized maximum by its asympotic distribution is studied. Instead of using the arithmetical scale of probabilities, we measure the difference between the actual and asympotic distribution in terms of the double-log scale used for building the probability plotting paper for the the latter. We demonstrate that the difference between the double-log values corresponding to two probabilities in the upper tail is almost exactly equal to the logarithm of the distribution may not be uniform in this double-log scale and that the difference between the actual and asymptotic distributions, on the probebility plotting paper, may be a logarithmic, power, or even exponential function in the upper tail when the latter distribution is of Fisher-Tippett type I, but that difference is at most logarithmic in the upper tail for type II and III distributions. This fact is exploited to obtain transformed variables that converge tothe asymptotic distribution faster than the original variable on the probabilites plotting paper  相似文献   

14.
This paper introduces a modified one-sample test of goodness-of-fit based on the cumulative distribution function. Damico [A new one-sample test for goodness-of-fit. Commun Stat – Theory Methods. 2004;33:181–193] proposed a test for testing goodness-of-fit of univariate distribution that uses the concept of partitioning the probability range into n intervals of equal probability mass 1/n and verifies that the hypothesized distribution evaluated at the observed data would place one case into each interval. The present paper extends this notion by allowing for m intervals of probability mass r/n, where r≥1 and n=m×r. A simulation study for small and moderate sample sizes demonstrates that the proposed test for two observations per interval under various alternatives is more powerful than the test proposed by Damico (2004).  相似文献   

15.
In this paper, a probability plots class of tests for multivariate normality is introduced. Based on independent standardized principal components of a d-variate normal data set, we obtained the sum of squared differences between corresponding observations of an ordered set of each principal component observations and the set of the population pth quantiles of the standard normal distribution. We proposed the sum of these d-sums of squared differences as an appropriate statistic for testing multivariate normality. We evaluated empirical critical values of the statistic and compared its power with those of some highly regarded techniques with a wonderful result.  相似文献   

16.
A life distribution is said to have a weak memoryless property if its conditional probability of survival beyond a fixed time point is equal to its (unconditional) survival probability at that point. Goodness‐of‐fit testing of this notion is proposed in the current investigation, both when the fixed time point is known and when it is unknown but estimable from the data. The limiting behaviour of the proposed test statistic is obtained and the null variance is explicitly given. The empirical power of the test is evaluated for a commonly known alternative using Monte Carlo methods, showing that the test performs well. The case when the fixed time point t0 equals a quantile of the distribution F gives a distribution‐free test procedure. The procedure works even if t0 is unknown but is estimable.  相似文献   

17.
In this paper, we proposed a new family of distributions namely exponentiated exponential–geometric (E2G) distribution. The E2G distribution is a straightforwardly generalization of the exponential–geometric (EG) distribution proposed by Adamidis and Loukas [A lifetime distribution with decreasing failure rate, Statist. Probab. Lett. 39 (1998), pp. 35–42], which accommodates increasing, decreasing and unimodal hazard functions. It arises on a latent competing risk scenarios, where the lifetime associated with a particular risk is not observable but only the minimum lifetime value among all risks. The properties of the proposed distribution are discussed, including a formal proof of its probability density function and explicit algebraic formulas for its survival and hazard functions, moments, rth moment of the ith order statistic, mean residual lifetime and modal value. Maximum-likelihood inference is implemented straightforwardly. From a mis-specification simulation study performed in order to assess the extent of the mis-specification errors when testing the EG distribution against the E2G, and we observed that it is usually possible to discriminate between both distributions even for moderate samples with presence of censoring. The practical importance of the new distribution was demonstrated in three applications where we compare the E2G distribution with several lifetime distributions.  相似文献   

18.
In this paper, the three-decision procedures to classify p treatments as better than or worse than one control, proposed for normal/symmetric probability models [Bohrer, Multiple three-decision rules for parametric signs. J. Amer. Statist. Assoc. 74 (1979), pp. 432–437; Bohrer et al., Multiple three-decision rules for factorial simple effects: Bonferroni wins again!, J. Amer. Statist. Assoc. 76 (1981), pp. 119–124; Liu, A multiple three-decision procedure for comparing several treatments with a control, Austral. J. Statist. 39 (1997), pp. 79–92 and Singh and Mishra, Classifying logistic populations using sample medians, J. Statist. Plann. Inference 137 (2007), pp. 1647–1657]; in the literature, have been extended to asymmetric two-parameter exponential probability models to classify p(p≥1) treatments as better than or worse than the best of q(q≥1) control treatments in terms of location parameters. Critical constants required for the implementation of the proposed procedure are tabulated for some pre-specified values of probability of no misclassification. Power function of the proposed procedure is also defined and a common sample size necessary to guarantee various pre-specified power levels are tabulated. Optimal allocation scheme is also discussed. Finally, the implementation of the proposed methodology is demonstrated through numerical example.  相似文献   

19.
The Benjamini–Hochberg procedure is widely used in multiple comparisons. Previous power results for this procedure have been based on simulations. This article produces theoretical expressions for expected power. To derive them, we make assumptions about the number of hypotheses being tested, which null hypotheses are true, which are false, and the distributions of the test statistics under each null and alternative. We use these assumptions to derive bounds for multiple dimensional rejection regions. With these bounds and a permanent based representation of the joint density function of the largest p-values, we use the law of total probability to derive the distribution of the total number of rejections. We derive the joint distribution of the total number of rejections and the number of rejections when the null hypothesis is true. We give an analytic expression for the expected power for a false discovery rate procedure that assumes the hypotheses are independent.  相似文献   

20.
ABSTRACT

In this paper, we consider the problem of constructing non parametric confidence intervals for the mean of a positively skewed distribution. We suggest calibrated, smoothed bootstrap upper and lower percentile confidence intervals. For the theoretical properties, we show that the proposed one-sided confidence intervals have coverage probability α + O(n? 3/2). This is an improvement upon the traditional bootstrap confidence intervals in terms of coverage probability. A version smoothed approach is also considered for constructing a two-sided confidence interval and its theoretical properties are also studied. A simulation study is performed to illustrate the performance of our confidence interval methods. We then apply the methods to a real data set.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号