首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
We propose a multivariate extension of the univariate chi-squared normality test. Using a known result for the distribution of quadratic forms in normal variables, we show that the proposed test statistic has an approximated chi-squared distribution under the null hypothesis of multivariate normality. As in the univariate case, the new test statistic is based on a comparison of observed and expected frequencies for specified events in sample space. In the univariate case, these events are the standard class intervals, but in the multivariate extension we propose these become hyper-ellipsoidal annuli in multivariate sample space. We assess the performance of the new test using Monte Carlo simulation. Keeping the type I error rate fixed, we show that the new test has power that compares favourably with other standard normality tests, though no uniformly most powerful test has been found. We recommend the new test due to its competitive advantages.  相似文献   

2.
A. Roy  D. Klein 《Statistics》2018,52(2):393-408
Testing hypotheses about the structure of a covariance matrix for doubly multivariate data is often considered in the literature. In this paper the Rao's score test (RST) is derived to test the block exchangeable covariance matrix or block compound symmetry (BCS) covariance structure under the assumption of multivariate normality. It is shown that the empirical distribution of the RST statistic under the null hypothesis is independent of the true values of the mean and the matrix components of a BCS structure. A significant advantage of the RST is that it can be performed for small samples, even smaller than the dimension of the data, where the likelihood ratio test (LRT) cannot be used, and it outperforms the standard LRT in a number of contexts. Simulation studies are performed for the sample size consideration, and for the estimation of the empirical quantiles of the null distribution of the test statistic. The RST procedure is illustrated on a real data set from the medical studies.  相似文献   

3.
Various nonparametric procedures are known for the goodness-of-fit test in the univariate case. The distribution-free nature of these procedures does not extend to the multivariate case. In this paper, we consider an application of the theory of statistically equivalent blocks(SEB)to obtain distribution-free procedures for the multivariate case. The sample values are transformed to random variables which are distributed as sample spacings from a uniform distribution on [0, 1], under the null hypothesis. Various test statistics are known, based on the spacings, which are used for testing uniformity in the univariate case. Any of these statistics can be used in the multivariate situation, based on the spacings generated from the SEB. This paper gives an expository development of the theory of SEB and a review of tests for goodness-of-fit, based on sample spacings. To show an application of the SEB, we consider a test of bivariate normality.  相似文献   

4.
Tests on multivariate means that are hypothesized to be in a specified direction have received attention from both theoretical and applied points of view. One of the most common procedures used to test this cone alternative is the likelihood ratio test (LRT) assuming a multivariate normal model for the data. However, the resulting test for an ordered alternative is biased in that the only usable critical values are bounds on the null distribution. The present paper provides empirical evidence that bootstrapping the null distribution of the likelihood ratio statistic results in a bootstrap test (BT) with comparable power properties without the additional burden of assuming multivariate normality. Additionally, the tests based on the LRT statistic can reject the null hypothesis in favor of the alternative even though the true means are far from the alternative region. The BT also has similar properties for normal and nonnormal data. This anomalous behavior is due to the formulation of the null hypothesis and a possible remedy is to reformulate the null to be the complement of the alternative hypothesis. We discuss properties of a BT for the modified set of hypotheses (MBT) based on a simulation study. The resulting test is conservative in general and in some specific cases has power estimates comparable to those for existing methods. The BT has higher sensitivity but relatively lower specificity, whereas the MBT has higher specificity but relatively lower sensitivity.  相似文献   

5.
ABSTRACT

This article argues that researchers do not need to completely abandon the p-value, the best-known significance index, but should instead stop using significance levels that do not depend on sample sizes. A testing procedure is developed using a mixture of frequentist and Bayesian tools, with a significance level that is a function of sample size, obtained from a generalized form of the Neyman–Pearson Lemma that minimizes a linear combination of α, the probability of rejecting a true null hypothesis, and β, the probability of failing to reject a false null, instead of fixing α and minimizing β. The resulting hypothesis tests do not violate the Likelihood Principle and do not require any constraints on the dimensionalities of the sample space and parameter space. The procedure includes an ordering of the entire sample space and uses predictive probability (density) functions, allowing for testing of both simple and compound hypotheses. Accessible examples are presented to highlight specific characteristics of the new tests.  相似文献   

6.
ABSTRACT

A statistical test can be seen as a procedure to produce a decision based on observed data, where some decisions consist of rejecting a hypothesis (yielding a significant result) and some do not, and where one controls the probability to make a wrong rejection at some prespecified significance level. Whereas traditional hypothesis testing involves only two possible decisions (to reject or not a null hypothesis), Kaiser’s directional two-sided test as well as the more recently introduced testing procedure of Jones and Tukey, each equivalent to running two one-sided tests, involve three possible decisions to infer the value of a unidimensional parameter. The latter procedure assumes that a point null hypothesis is impossible (e.g., that two treatments cannot have exactly the same effect), allowing a gain of statistical power. There are, however, situations where a point hypothesis is indeed plausible, for example, when considering hypotheses derived from Einstein’s theories. In this article, we introduce a five-decision rule testing procedure, equivalent to running a traditional two-sided test in addition to two one-sided tests, which combines the advantages of the testing procedures of Kaiser (no assumption on a point hypothesis being impossible) and Jones and Tukey (higher power), allowing for a nonnegligible (typically 20%) reduction of the sample size needed to reach a given statistical power to get a significant result, compared to the traditional approach.  相似文献   

7.
ABSTRACT

We derive the influence function of the likelihood ratio test statistic for multivariate normal sample. The derived influence function does not depend on the influence functions of the parameters under the null hypothesis. So we can obtain directly the empirical influence function with only the maximum likelihood estimators under the null hypothesis. Since the derived formula is a general form, it can be applied to influence analysis on many statistical testing problems.  相似文献   

8.
This paper introduces W-tests for assessing homogeneity in mixtures of discrete probability distributions. A W-test statistic depends on the data solely through parameter estimators and, if a penalized maximum likelihood estimation framework is used, has a tractable asymptotic distribution under the null hypothesis of homogeneity. The large-sample critical values are quantiles of a chi-square distribution multiplied by an estimable constant for which we provide an explicit formula. In particular, the estimation of large-sample critical values does not involve simulation experiments or random field theory. We demonstrate that W-tests are generally competitive with a benchmark test in terms of power to detect heterogeneity. Moreover, in many situations, the large-sample critical values can be used even with small to moderate sample sizes. The main implementation issue (selection of an underlying measure) is thoroughly addressed, and we explain why W-tests are well-suited to problems involving large and online data sets. Application of a W-test is illustrated with an epidemiological data set.  相似文献   

9.
A new statistical procedure for testing normality is proposed. The Q statistic is derived as the ratio of two linear combinations of the ordered random observations. The coefficients of the linear combinations are utilizing the expected values of the order statistics from the standard normal distribution. This test is omnibus to detect the deviations from normality that result from either skewness or kurtosis. The statistic is independent of the origin and the scale under the null hypothesis of normality, and the null distribution of Q can be very well approximated by the Cornish-Fisher expansion. The powers for various alternative distributions were compared with several other test statistics by simulations.  相似文献   

10.
In this article, we propose a new goodness-of-fit test for Type I or Type II censored samples from a completely specified distribution. This test is a generalization of Michael's test for censored data, which is based on the empirical distribution and a variance stabilizing transformation. Using Monte Carlo methods, the distributions of the test statistics are analyzed under the null hypothesis. Tables of quantiles of these statistics are also provided. The power of the proposed test is studied and compared to that of other well-known tests also using simulation. The proposed test is more powerful in most of the considered cases. Acceptance regions for the PP, QQ, and Michael's stabilized probability plots are derived, which enable one to visualize which data contribute to the decision of rejecting the null hypothesis. Finally, an application in quality control is presented as illustration.  相似文献   

11.
A probability property that connects the skew normal (SN) distribution with the normal distribution is used for proposing a goodness-of-fit test for the composite null hypothesis that a random sample follows an SN distribution with unknown parameters. The random sample is transformed to approximately normal random variables, and then the Shapiro–Wilk test is used for testing normality. The implementation of this test does not require neither parametric bootstrap nor the use of tables for different values of the slant parameter. An additional test for the same problem, based on a property that relates the gamma and SN distributions, is also introduced. The results of a power study conducted by the Monte Carlo simulation show some good properties of the proposed tests in comparison to existing tests for the same problem.  相似文献   

12.
The density of the multiple correlation coefficient is derived by direct integration when the sample covariance matrix has a linear non-central distribution. Using the density, we deduce the null and non-null distribution of the multiple correlation coefficient when sampling from a mixture of two multivariate normal populations with the same covariance matrix. We also compute actual significance levels of the test of the hypothesis Ho : ρ1·2…p = 0 versus Ha1·2…p > 0, given the mixture model.  相似文献   

13.
Abstract

In this paper we present several goodness-of-fit tests for the centralized Wishart process, a popular matrix-variate time series model used to capture the stochastic properties of realized covariance matrices. The new test procedures are based on the extended Bartlett decomposition derived from the properties of the Wishart distribution and allows to obtain sets of independently and standard normally distributed random variables under the null hypothesis. Several tests for normality and independence are then applied to these variables in order to support or to reject the underlying assumption of a centralized Wishart process. In order to investigate the influence of estimated parameters on the suggested testing procedures in the finite-sample case, a simulation study is conducted. Finally, the new test methods are applied to real data consisting of realized covariance matrices computed for the returns on six assets traded on the New York Stock Exchange.  相似文献   

14.
Characterization theorems in probability and statistics are widely appreciated for their role in clarifying the structure of the families of probability distributions. Less well known is the role characterization theorems have as a natural, logical and effective starting point for constructing goodness-of-fit tests. The characteristic independence of the mean and variance and of the mean and the third central moment of a normal sample were used, respectively, by Lin and Mudholkar [1980. A simple test for normality against asymmetric alternatives. Biometrika 67, 455–461] and by Mudholkar et al. [2002a. Independence characterizations and testing normality against skewness-kurtosis alternatives. J. Statist. Plann. Inference 104, 485–501] for developing tests of normality. The characteristic independence of the maximum likelihood estimates of the population parameters was similarly used by Mudholkar et al. [2002b. Independence characterization and inverse Gaussian goodness-of-fit. Sankhya A 63, 362–374] to develop a test of the composite inverse Gaussian hypothesis. The gamma models are extensively used for applied research in the areas of econometrics, engineering and biomedical sciences; but there are few goodness-of-fit tests available to test if the data indeed come from a gamma population. In this paper we employ Hwang and Hu's [1999. On a characterization of the gamma distribution: the independence of the sample mean and the sample coefficient of variation. Ann. Inst. Statist. Math. 51, 749–753] characterization of the gamma population in terms of the independence of sample mean and coefficient of variation for developing such a test. The asymptotic null distribution of the proposed test statistic is obtained and empirically refined for use with samples of moderate size.  相似文献   

15.
To detect the dependence on the covariates in the lower and upper tails of the response distribution, regression quantiles are very useful tools in linear model problems with univariate response. We consider here a notion of regression quantiles for problems with multivariate responses. The approach is based on minimizing a loss function equivalent to that in the case of univariate response. To construct an affine equivariant notion of multivariate regression quantiles, we have considered a transformation retransformation procedure based on ‘data-driven coordinate systems’. We indicate some algorithm to compute the proposed estimates and establish asymptotic normality for them. We also, suggest an adaptive procedure to select the optimal data-driven coordinate system. We discuss the performance of our estimates with the help of a finite sample simulation study and to illustrate our methodology, we analyzed an interesting data-set on blood pressures of a group of women and another one on the dependence of sales performances on creative test scores.  相似文献   

16.
The quantile–quantile plot is widely used to check normality. The plot depends on the plotting positions. Many commonly used plotting positions do not depend on the sample values. We propose an adaptive plotting position that depends on the relative distances of the two neighbouring sample values. The correlation coefficient obtained from the adaptive plotting position is used to test normality. The test using the adaptive plotting position is better than the Shapiro–Wilk W test for small samples and has larger power than Hazen's and Blom's plotting positions for symmetric alternatives with shorter tail than normal and skewed alternatives when n is 20 or larger. The Brown–Hettmansperger T* test is designed for detecting bad tail behaviour, so it does not have power for symmetric alternatives with shorter tail than normal, but it is generally better than the other tests when β2 is greater than 3.25.  相似文献   

17.
For a multivariate linear model, Wilk's likelihood ratio test (LRT) constitutes one of the cornerstone tools. However, the computation of its quantiles under the null or the alternative hypothesis requires complex analytic approximations, and more importantly, these distributional approximations are feasible only for moderate dimension of the dependent variable, say p≤20. On the other hand, assuming that the data dimension p as well as the number q of regression variables are fixed while the sample size n grows, several asymptotic approximations are proposed in the literature for Wilk's Λ including the widely used chi-square approximation. In this paper, we consider necessary modifications to Wilk's test in a high-dimensional context, specifically assuming a high data dimension p and a large sample size n. Based on recent random matrix theory, the correction we propose to Wilk's test is asymptotically Gaussian under the null hypothesis and simulations demonstrate that the corrected LRT has very satisfactory size and power, surely in the large p and large n context, but also for moderately large data dimensions such as p=30 or p=50. As a byproduct, we give a reason explaining why the standard chi-square approximation fails for high-dimensional data. We also introduce a new procedure for the classical multiple sample significance test in multivariate analysis of variance which is valid for high-dimensional data.  相似文献   

18.
In this paper, a probability plots class of tests for multivariate normality is introduced. Based on independent standardized principal components of a d-variate normal data set, we obtained the sum of squared differences between corresponding observations of an ordered set of each principal component observations and the set of the population pth quantiles of the standard normal distribution. We proposed the sum of these d-sums of squared differences as an appropriate statistic for testing multivariate normality. We evaluated empirical critical values of the statistic and compared its power with those of some highly regarded techniques with a wonderful result.  相似文献   

19.
We propose a new goodness-of-fit test for normal and lognormal distributions with unknown parameters and type-II censored data. This test is a generalization of Michael's test for censored samples, which is based on the empirical distribution and a variance stabilizing transformation. We estimate the parameters of the model by using maximum likelihood and Gupta's methods. The quantiles of the distribution of the test statistic under the null hypothesis are obtained through Monte Carlo simulations. The power of the proposed test is estimated and compared to that of the Kolmogorov–Smirnov test also using simulations. The new test is more powerful than the Kolmogorov–Smirnov test in most of the studied cases. Acceptance regions for the PP, QQ and Michael's stabilized probability plots are derived, making it possible to visualize which data contribute to the decision of rejecting the null hypothesis. Finally, an illustrative example is presented.  相似文献   

20.
In many situations, we want to verify the existence of a relationship between multivariate time series. In this paper, we generalize the procedure developed by Haugh (1976) for univariate time series in order to test the hypothesis of noncorrelation between two multivariate stationary ARMA series. The test statistics are based on residual cross-correlation matrices. Under the null hypothesis of noncorrelation, we show that an arbitrary vector of residual cross-correlations asymptotically follows the same distribution as the corresponding vector of cross-correlations between the two innovation series. From this result, it follows that the test statistics considered are asymptotically distributed as chi-square random variables. Two test procedures are described. The first one is based on the residual cross-correlation matrix at a particular lag, whilst the second one is based on a portmanteau type statistic that generalizes Haugh's statistic. We also discuss how the procedures for testing noncorrelation can be adapted to determine the directions of causality in the sense of Granger (1969) between the two series. An advantage of the proposed procedures is that their application does not require the estimation of a global model for the two series. The finite-sample properties of the statistics introduced were studied by simulation under the null hypothesis. It led to modified statistics whose upper quantiles are much better approximated by those of the corresponding chi-square distribution. Finally, the procedures developed are applied to two different sets of economic data.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号