首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Summary. The evaluation of the cumulative distribution function of a multivariate normal distribution is considered. The multivariate normal distribution can have any positive definite correlation matrix and any mean vector. The approach taken has two stages. In the first stage, it is shown how non-centred orthoscheme probabilities can be evaluated by using a recursive integration method. In the second stage, some ideas of Schläfli and Abrahamson are extended to show that any non-centred orthant probability can be expressed as differences between at most ( m −1)! non-centred orthoscheme probabilities. This approach allows an accurate evaluation of many multivariate normal probabilities which have important applications in statistical practice.  相似文献   

2.
We consider estimation of the number of cells in a multinomial distribution. This is one version of the species problem: there are many applications, such as the estimation of the number of unobserved species of animals; estimation of vocabulary size, etc. We describe the results of a simulation comparison of three principal frequent-ist' procedures for estimating the number of cells (or species). The first procedure postulates a functional form for the cell probabilities; the second procedure approxi mates the distribution of the probabilities by a parametric probability density function; and the third procedure is based on an estimate of the sample coverage, i.e. the sum of the probabilities of the observed cells. Among the procedures studied, we find that the third (non-parametric) method is globally preferable; the second (functional parametric) method cannot be recommended; and that, when based on the inverse Gaussian density, the first method is competitive in some cases with the third method. We also discuss Sichel's recent generalized inverse Gaussian-based procedure which, with some refine ment, promises to perform at least as well as the non-parametric method in all cases.  相似文献   

3.
Cook (1977) proposed a diagnostic to quantify the impact of deleting an observation on the estimated regression coefficients of a General Linear Univariate Model (GLUM). Simulations of models with Gaussian response and predictors demonstrate that his suggestion of comparing the diagnostic to the median of the F for overall regression captures an erratically varying proportion of the values.

We describe the exact distribution of Cook's statistic for a GLUM with Gaussian predictors and response. We also present computational forms, simple approximations, and asymptotic results. A simulation supports the accuracy of the results. The methods allow accurate evaluation of a single value or the maximum value from a regression analysis. The approximations work well for a single value, but less well for the maximum. In contrast, the cut-point suggested by Cook provides widely varying tail probabilities. As with all diagnostics, the data analyst must use scientific judgment in deciding how to treat highlighted observations.  相似文献   

4.
Inverse Gaussian distribution has been used in a wide range of applications in modeling duration and failure phenomena. In these applications, one-sided lower tolerance limits are employed, for instance, for designing safety limits of medical devices. Tang and Chang (1994) proposed lowersided tolerance limits via Bonferroni inequality when parameters in the inverse Gaussian distribution are unknown. However, their simulation results showed conservative coverage probabilities, and consequently larger interval width. In their paper, they also proposed an alternative to construct lesser conservative limits. But simulation results yielded unsatisfactory coverage probabilities in many cases. In this article, the exact lower-sided tolerance limit is proposed. The proposed limit has a similar form to that of the confidence interval for mean under inverse Gaussian. The comparison between the proposed limit and Tang and Chang's method is compared via extensive Monte Carlo simulations. Simulation results suggest that the proposed limit is superior to Tang and Chang's method in terms of narrower interval width and approximate to nominal level of coverage probability. Similar argument can be applied to the formulation of two-sided tolerance limits. A summary and conclusion of the proposed limits is included.  相似文献   

5.
We consider the problem of evaluation of the probability that all elements of a multivariate normally distributed vector have non-negative coordinates; this probability is called the non-centred orthant probability. The necessity for the evaluation of this probability arises frequently in statistics. The probability is defined by the integral of the probability density function. However, direct numerical integration is not practical. In this article, a method is proposed for the computation of the probability. The method involves the evaluation of a measure on a unit sphere surface in p-dimensional space that satisfies conditions derived from a covariance matrix. The required computational time for the p-dimensional problem is proportional to p2·2p?1, and it increases at a rate that is lower than that in the case of the existing method.  相似文献   

6.
A method of regularized discriminant analysis for discrete data, denoted DRDA, is proposed. This method is related to the regularized discriminant analysis conceived by Friedman (1989) in a Gaussian framework for continuous data. Here, we are concerned with discrete data and consider the classification problem using the multionomial distribution. DRDA has been conceived in the small-sample, high-dimensional setting. This method has a median position between multinomial discrimination, the first-order independence model and kernel discrimination. DRDA is characterized by two parameters, the values of which are calculated by minimizing a sample-based estimate of future misclassification risk by cross-validation. The first parameter is acomplexity parameter which provides class-conditional probabilities as a convex combination of those derived from the full multinomial model and the first-order independence model. The second parameter is asmoothing parameter associated with the discrete kernel of Aitchison and Aitken (1976). The optimal complexity parameter is calculated first, then, holding this parameter fixed, the optimal smoothing parameter is determined. A modified approach, in which the smoothing parameter is chosen first, is discussed. The efficiency of the method is examined with other classical methods through application to data.  相似文献   

7.
In this paper, we introduce the p-generalized polar methods for the simulation of the p-generalized Gaussian distribution. On the basis of geometric measure representations, the well-known Box–Muller method and the Marsaglia–Bray rejecting polar method for the simulation of the Gaussian distribution are generalized to simulate the p-generalized Gaussian distribution, which fits much more flexibly to data than the Gaussian distribution and has already been applied in various fields of modern sciences. To prove the correctness of the p-generalized polar methods, we give stochastic representations, and to demonstrate their adequacy, we perform a comparison of six simulation techniques w.r.t. the goodness of fit and the complexity. The competing methods include adapted general methods and another special method. Furthermore, we prove stochastic representations for all the adapted methods.  相似文献   

8.
Statistics and Computing - A new procedure is proposed for evaluating non-centred orthant probabilities of elliptical distributed vectors, which is the probabilities that all elements of a vector...  相似文献   

9.
In this paper, we propose a method for testing absolutely regular and possibly nonstationary nonlinear time-series, with application to general AR-ARCH models. Our test statistic is based on a marked empirical process of residuals which is shown to converge to a Gaussian process with respect to the Skohorod topology. This testing procedure was first introduced by Stute [Nonparametric model checks for regression, Ann. Statist. 25 (1997), pp. 613–641] and then widely developed by Ngatchou-Wandji [Weak convergence of some marked empirical processes: Application to testing heteroscedasticity, J. Nonparametr. Stat. 14 (2002), pp. 325–339; Checking nonlinear heteroscedastic time series models, J. Statist. Plann. Inference 133 (2005), pp. 33–68; Local power of a Cramer-von Mises type test for parametric autoregressive models of order one, Compt. Math. Appl. 56(4) (2008), pp. 918–929] under more general conditions. Applications to general AR-ARCH models are given.  相似文献   

10.
We propose a new model for regression and dependence analysis when addressing spatial data with possibly heavy tails and an asymmetric marginal distribution. We first propose a stationary process with t marginals obtained through scale mixing of a Gaussian process with an inverse square root process with Gamma marginals. We then generalize this construction by considering a skew‐Gaussian process, thus obtaining a process with skew‐t marginal distributions. For the proposed (skew) t process, we study the second‐order and geometrical properties and in the t case, we provide analytic expressions for the bivariate distribution. In an extensive simulation study, we investigate the use of the weighted pairwise likelihood as a method of estimation for the t process. Moreover we compare the performance of the optimal linear predictor of the t process versus the optimal Gaussian predictor. Finally, the effectiveness of our methodology is illustrated by analyzing a georeferenced dataset on maximum temperatures in Australia.  相似文献   

11.
Abstract

This article mainly considers the uniform asymptotics for the finite-time ruin probabilities of a two-dimensional renewal risk model with heavy-tailed claims. In this model, the two claim-number processes are arbitrarily dependent and each of them is generated by widely orthant dependent claim inter-arrival times. Two types of ruin are studied and for each type of ruin, an asymptotic formula for the finite-time ruin probability is established. These formulae possess a certain uniformity feature in the time horizon.  相似文献   

12.
We study the computation of Gaussian orthant probabilities, i.e. the probability that a Gaussian variable falls inside a quadrant. The Geweke–Hajivassiliou–Keane (GHK) algorithm (Geweke, Comput Sci Stat 23:571–578 1991, Keane, Simulation estimation for panel data models with limited dependent variables, 1993, Hajivassiliou, J Econom 72:85–134, 1996, Genz, J Comput Graph Stat 1:141–149, 1992) is currently used for integrals of dimension greater than 10. In this paper, we show that for Markovian covariances GHK can be interpreted as the estimator of the normalizing constant of a state-space model using sequential importance sampling. We show for an AR(1) the variance of the GHK, properly normalized, diverges exponentially fast with the dimension. As an improvement we propose using a particle filter. We then generalize this idea to arbitrary covariance matrices using Sequential Monte Carlo with properly tailored MCMC moves. We show empirically that this can lead to drastic improvements on currently used algorithms. We also extend the framework to orthants of mixture of Gaussians (Student, Cauchy, etc.), and to the simulation of truncated Gaussians.  相似文献   

13.
A simulation study was done to compare seven confidence interval methods, based on the normal approximation, for the difference of two binomial probabilities. Cases considered included minimum expected cell sizes ranging from 2 to 15 and smallest group sizes (NMIN) ranging from 6 to 100. Our recommendation is to use a continuity correction of 1/(2 NMIN) combined with the use of (N ? 1) rather than N in the estimate of the standard error. For all of the cases considered with minimum expected cell size of at least 3, this method gave coverage probabilities close to or greater than the nominal 90% and 95%. The Yates method is also acceptable, but it is slightly more conservative. At the other extreme, the usual method (with no continuity correction) does not provide adequate coverage even at the larger sample sizes. For the 99% intervals, our recommended method and the Yates correction performed equally well and are reasonable for minimum expected cell sizes of at least 5. None of the methods performed consistently well for a minimum expected cell size of 2.  相似文献   

14.
Generalized additive models represented using low rank penalized regression splines, estimated by penalized likelihood maximisation and with smoothness selected by generalized cross validation or similar criteria, provide a computationally efficient general framework for practical smooth modelling. Various authors have proposed approximate Bayesian interval estimates for such models, based on extensions of the work of Wahba, G. (1983) [Bayesian confidence intervals for the cross validated smoothing spline. J. R. Statist. Soc. B 45 , 133–150] and Silverman, B.W. (1985) [Some aspects of the spline smoothing approach to nonparametric regression curve fitting. J. R. Statist. Soc. B 47 , 1–52] on smoothing spline models of Gaussian data, but testing of such intervals has been rather limited and there is little supporting theory for the approximations used in the generalized case. This paper aims to improve this situation by providing simulation tests and obtaining asymptotic results supporting the approximations employed for the generalized case. The simulation results suggest that while across‐the‐model performance is good, component‐wise coverage probabilities are not as reliable. Since this is likely to result from the neglect of smoothing parameter variability, a simple and efficient simulation method is proposed to account for smoothing parameter uncertainty: this is demonstrated to substantially improve the performance of component‐wise intervals.  相似文献   

15.
This paper considers the evaluation of probabilities which are defined by a set of linear inequalities of a trivariate normal distribution. It is shown that these probabilities can be evaluated by a one-dimensional numerical integration. The trivariate normal distribution can have any covariance matrix and any mean vector, and the probability can be defined by any number of one-sided and two-sided linear inequalities. This affords a practical and efficient method for the calculation of these probabilities which is superior to basic simulation methods. An application of this method to the analysis of pairwise comparisons of four treatment effects is discussed.  相似文献   

16.
The problem of calculating orthant probabilities for sets of variables (X1,., Xn) is considered in the case where they are jointly normally distributed with zero means and a correlation matrix such that the correlation between Xi and Xi is zero if |i-j|> 1. An effective method is given which works for quite large n when the correlations between Xi and Xi+1 have the values 1/2, 2/5, 3/10, 4/17, 5/26,. and more approximate methods are given for other values. The accuracy is investigated numerically.  相似文献   

17.
A general saddlepoint/Monte Carlo method to approximate (conditional) multivariate probabilities is presented. This method requires a tractable joint moment generating function (m.g.f.), but does not require a tractable distribution or density. The method is easy to program and has a third-order accuracy with respect to increasing sample size in contrast to standard asymptotic approximations which are typically only accurate to the first order.

The method is most easily described in the context of a continuous regular exponential family. Here, inferences can be formulated as probabilities with respect to the joint density of the sufficient statistics or the conditional density of some sufficient statistics given the others. Analytical expressions for these densities are not generally available, and it is often not possible to simulate exactly from the conditional distributions to obtain a direct Monte Carlo approximation of the required integral. A solution to the first of these problems is to replace the intractable density by a highly accurate saddlepoint approximation. The second problem can be addressed via importance sampling, that is, an indirect Monte Carlo approximation involving simulation from a crude approximation to the true density. Asymptotic normality of the sufficient statistics suggests an obvious candidate for an importance distribution.

The more general problem considers the computation of a joint probability for a subvector of random T, given its complementary subvector, when its distribution is intractable, but its joint m.g.f. is computable. For such settings, the distribution may be tilted, maintaining T as the sufficient statistic. Within this tilted family, the computation of such multivariate probabilities proceeds as described for the exponential family setting.  相似文献   

18.
While standard techniques are available for the analysis of time-series (longitudinal) data, and for ordinal (rating) data, not much is available for the combination of the two, at least in a readily-usable form. However, this data type is common place in the natural and health sciences where repeated ratings are recorded on the same subject. To analyse these data, this paper considers a transition (Markov) model where the rating of a subject at one time depends explicitly on the observed rating at the previous point of time by incorporating the previous rating as a predictor variable. Complications arise with adequate handling of data at the first observation (t=1), as there is no prior observation to use as a predictor. To overcome this, it is postulated the existence of a rating at time t=0; however it is treated as ‘missing data’ and the expectation–maximisation algorithm used to accommodate this. The particular benefits of this method are shown for shorter time series.  相似文献   

19.
In this article, we investigated the bootstrap calibrated generalized confidence limits for process capability indices C pk for the one-way random effect model. Also, we derived Bissell's approximation formula for the lower confidence limit using Satterthwaite's method and calculated its coverage probabilities and expected values. Then we compared it with standard bootstrap (SB) method and generalized confidence interval method. The simulation results indicate that the confidence limit obtained offers satisfactory coverage probabilities. The proposed method is illustrated with the help of simulation studies and data sets.  相似文献   

20.
We consider a nonparametric autoregression model under conditional heteroscedasticity with the aim to test whether the innovation distribution changes in time. To this end, we develop an asymptotic expansion for the sequential empirical process of nonparametrically estimated innovations (residuals). We suggest a Kolmogorov–Smirnov statistic based on the difference of the estimated innovation distributions built from the first ?ns?and the last n ? ?ns? residuals, respectively (0 ≤ s ≤ 1). Weak convergence of the underlying stochastic process to a Gaussian process is proved under the null hypothesis of no change point. The result implies that the test is asymptotically distribution‐free. Consistency against fixed alternatives is shown. The small sample performance of the proposed test is investigated in a simulation study and the test is applied to a data example.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号