首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 24 毫秒
1.
On the Tukey depth of an atomic measure   总被引:1,自引:0,他引:1  
This paper gives a relation between the convex Tukey trimmed region (see [J.C. Massé, R. Theodorescu, Halfplane trimming for bivariate distributions, J. Multivariate Anal. 48(2) (1994) 188–202]) of an atomic measure and the support of the measure. It is shown that an atomic measure is concentrated on the extreme points of its Tukey trimmed region. A property concerning the extreme points which have 0 mass is given. As a corollary, we give a new method of proof of the Koshevoy characterization result.  相似文献   

2.
The Tukey depth (Proceedings of the International Congress of Mathematicians, vol. 2, pp. 523–531, 1975) of a point p with respect to a finite set S of points is the minimum number of elements of S contained in any closed halfspace that contains p. Algorithms for computing the Tukey depth of a point in various dimensions are considered. The running times of these algorithms depend on the value of the output, making them suited to situations, such as outlier removal, where the value of the output is typically small. This research was partly funded by the NSERC Canada.  相似文献   

3.
The concept of location depth was introduced as a way to extend the univariate notion of ranking to a bivariate configuration of data points. It has been used successfully for robust estimation, hypothesis testing, and graphical display. The depth contours form a collection of nested polygons, and the center of the deepest contour is called the Tukey median. The only available implemented algorithms for the depth contours and the Tukey median are slow, which limits their usefulness. In this paper we describe an optimal algorithm which computes all bivariate depth contours in O(n 2) time and space, using topological sweep of the dual arrangement of lines. Once these contours are known, the location depth of any point can be computed in O(log2 n) time with no additional preprocessing or in O(log n) time after O(n 2) preprocessing. We provide fast implementations of these algorithms to allow their use in everyday statistical practice.  相似文献   

4.
In univariate statistics, the trimmed mean has long been regarded as a robust and efficient alternative to the sample mean. A multivariate analogue calls for a notion of trimmed region around the center of the sample. Using Tukey's depth to achieve this goal, this paper investigates two types of multivariate trimmed means obtained by averaging over the trimmed region in two different ways. For both trimmed means, conditions ensuring asymptotic normality are obtained; in this respect, one of the main features of the paper is the systematic use of Hadamard derivatives and empirical processes methods to derive the central limit theorems. Asymptotic efficiency relative to the sample mean as well as breakdown point are also studied. The results provide convincing evidence that these location estimators have nice asymptotic behavior and possess highly desirable finite-sample robustness properties; furthermore, relative to the sample mean, both of them can in some situations be highly efficient for dimensions between 2 and 10.  相似文献   

5.
6.
Modern methods for detecting changes in the scale or covariance of multivariate distributions rely primarily on testing for the constancy of the covariance matrix. These depend on higher-order moment conditions, and also do not work well when the dimension of the data is large or even moderate relative to the sample size. In this paper, we propose a nonparametric change point test for multivariate data using rankings obtained from data depth measures. As the data depth of an observation measures its centrality relative to the sample, changes in data depth may signify a change of scale of the underlying distribution, and the proposed test is particularly responsive to detecting such changes. We provide a full asymptotic theory for the proposed test statistic under the null hypothesis that the observations are stable, and natural conditions under which the test is consistent. The finite sample properties are investigated by means of a Monte Carlo simulation, and these along with the theoretical results confirm that the test is robust to heavy tails, skewness and high dimensionality. The proposed methods are demonstrated with an application to structural break detection in the rate of change of pollutants linked to acid rain measured in Turkey lake, a lake in central Ontario, Canada. Our test suggests a change in the rate of acid rain in the late 1980s/early 1990s, which coincides with clean air legislation in Canada and the US. The Canadian Journal of Statistics 48: 417–446; 2020 © 2020 Statistical Society of Canada  相似文献   

7.
This paper extends the univariate time series smoothing approach provided by penalized least squares to a multivariate setting, thus allowing for joint estimation of several time series trends. The theoretical results are valid for the general multivariate case, but particular emphasis is placed on the bivariate situation from an applied point of view. The proposal is based on a vector signal-plus-noise representation of the observed data that requires the first two sample moments and specifying only one smoothing constant. A measure of the amount of smoothness of an estimated trend is introduced so that an analyst can set in advance a desired percentage of smoothness to be achieved by the trend estimate. The required smoothing constant is determined by the chosen percentage of smoothness. Closed form expressions for the smoothed estimated vector and its variance-covariance matrix are derived from a straightforward application of generalized least squares, thus providing best linear unbiased estimates for the trends. A detailed algorithm applicable for estimating bivariate time series trends is also presented and justified. The theoretical results are supported by a simulation study and two real applications. One corresponds to Mexican and US macroeconomic data within the context of business cycle analysis, and the other one to environmental data pertaining to a monitored site in Scotland.  相似文献   

8.
Methods for assessing the variability of an estimated contour of a density are discussed. A new method called the coverage plot is proposed. Techniques including sectioning and bootstrap techniques are compared for a particular problem which arises in Monte Carlo simulation approaches to estimating the spatial distribution of risk in the operation of weapons firing ranges. It is found that, for computational reasons, the sectioning procedure outperforms the bootstrap for this problem. The roles of bias and sample size are also seen in the examples shown.  相似文献   

9.
The estimator for the coefficients of a distributed lag model with smoothness priors proposed by Shi Her (1973) is examined and an alternative ridge-type estimator is obtained utilizing restricted least squares approach, A Monte Carlo experiment is carried out to compare the behavior of these estimators.  相似文献   

10.
Contours may be viewed as the 2D outline of the image of an object. This type of data arises in medical imaging as well as in computer vision and can be modeled as data on a manifold and can be studied using statistical shape analysis. Practically speaking, each observed contour, while theoretically infinite dimensional, must be discretized for computations. As such, the coordinates for each contour as obtained at k sampling times, resulting in the contour being represented as a k-dimensional complex vector. While choosing large values of k will result in closer approximations to the original contour, this will also result in higher computational costs in the subsequent analysis. The goal of this study is to determine reasonable values for k so as to keep the computational cost low while maintaining accuracy. To do this, we consider two methods for selecting sample points and determine lower bounds for k for obtaining a desired level of approximation error using two different criteria. Because this process is computationally inefficient to perform on a large scale, we then develop models for predicting the lower bounds for k based on simple characteristics of the contours.  相似文献   

11.
Forecasting of future snow depths is useful for many applications like road safety, winter sport activities, avalanche risk assessment and hydrology. Motivated by the lack of statistical forecasts models for snow depth, in this paper we present a set of models to fill this gap. First, we present a model to do short-term forecasts when we assume that reliable weather forecasts of air temperature and precipitation are available. The covariates are included nonlinearly into the model following basic physical principles of snowfall, snow aging and melting. Due to the large set of observations with snow depth equal to zero, we use a zero-inflated gamma regression model, which is commonly used to similar applications like precipitation. We also do long-term forecasts of snow depth and much further than traditional weather forecasts for temperature and precipitation. The long-term forecasts are based on fitting models to historic time series of precipitation, temperature and snow depth. We fit the models to data from six locations in Norway with different climatic and vegetation properties. Forecasting five days into the future, the results showed that, given reliable weather forecasts of temperature and precipitation, the forecast errors in absolute value was between 3 and 7?cm for different locations in Norway. Forecasting three weeks into the future, the forecast errors were between 7 and 16?cm.  相似文献   

12.
This paper presents a method to estimate mortality trends of two-dimensional mortality tables. Comparability of mortality trends for two or more of such tables is enhanced by applying penalized least squares and imposing a desired percentage of smoothness to be attained by the trends. The smoothing procedure is basically determined by the smoothing parameters that are related to the percentage of smoothness. To quantify smoothness, we employ an index defined first for the one-dimensional case and then generalized to the two-dimensional one. The proposed method is applied to data from member countries of the OECD. We establish as goal the smoothed mortality surface for one of those countries and compare it with some other mortality surfaces smoothed with the same percentage of two-dimensional smoothness. Our aim is to be able to see whether convergence exists in the mortality trends of the countries under study, in both year and age dimensions.  相似文献   

13.
Based on data depth, three types of nonparametric goodness-of-fit tests for multivariate distribution are proposed in this paper. They are Pearson’s chi-square test, tests based on EDF and tests based on spacings, respectively. The Anderson–Darling (AD) test and the Greenwood test for bivariate normal distribution and uniform distribution are simulated. The results of simulation show that these two tests have low type I error rates and become more efficient with the increase in sample size. The AD-type test performs more powerfully than the Greenwood type test.  相似文献   

14.
In this article, the local convergence rate of the mean square error (MSE) corresponding to a delta sequence-based density estimators is investigated by using second-order modulus of continuity type majorants. We look at the rate of convergency of the MSE of estimator for densities belonging to the class of functions which are defined by the second-order finite differences. The main contribution of this study is to obtain stronger convergence rate of a MSE by relaxing the second-order differentiation condition when compared with the class of density functions defined by the first-order finite differences.  相似文献   

15.
In this paper it is shown that data depth does not only provide consistent and robust estimators but also consistent and robust tests. Thereby, consistency of a test means that the Type I (αα) error and the Type II (ββ) error converge to zero with growing sample size in the interior of the nullhypothesis and the alternative, respectively. Robustness is measured by the breakdown point which depends here on a so-called concentration parameter. The consistency and robustness properties are shown for cases where the parameter of maximum depth is a biased estimator and has to be corrected. This bias is a disadvantage for estimation but an advantage for testing. It causes that the corresponding simplicial depth is not a degenerated U-statistic so that tests can be derived easily. However, the straightforward tests have a very poor power although they are asymptotic α-levelα-level tests. To improve the power, a new method is presented to modify these tests so that even consistency of the modified tests is achieved. Examples of two-dimensional copulas and the Weibull distribution show the applicability of the new method.  相似文献   

16.
This article introduces a fast cross-validation algorithm that performs wavelet shrinkage on data sets of arbitrary size and irregular design and also simultaneously selects good values of the primary resolution and number of vanishing moments.We demonstrate the utility of our method by suggesting alternative estimates of the conditional mean of the well-known Ethanol data set. Our alternative estimates outperform the Kovac-Silverman method with a global variance estimate by 25% because of the careful selection of number of vanishing moments and primary resolution. Our alternative estimates are simpler than, and competitive with, results based on the Kovac-Silverman algorithm equipped with a local variance estimate.We include a detailed simulation study that illustrates how our cross-validation method successfully picks good values of the primary resolution and number of vanishing moments for unknown functions based on Walsh functions (to test the response to changing primary resolution) and piecewise polynomials with zero or one derivative (to test the response to function smoothness).  相似文献   

17.
Let D be a saturated fractional factorial design of the general K1 x K2 ...x Kt factorial such that it consists of m distinct treatment combinations and it is capable of providing an unbiased estimator of a subvector of m factorial parameters under the assumption that the remaining k-m,t (k = H it ) factorial parameters are negligible. Such a design will not provide an unbiased estimator of the varianceσ2 Suppose that D is an optimal design with respect to some optimality criterion (e.g. d-optimality, a-optimality or e-optimality) and it is desirable to augment D with c treatmentcombinations with the aim to estimate 2 Suppose that D is an optimal design with respect to some optimality criterion (e.g. d-optimality, a-optimality or e-optimality) and it is desirable to augment D with c treatment combinations with the aim to estimate σ2 unbiasedly. The problem then is how to select the c treatment combinations such that the augmented design D retains its optimality property. This problem, in all its generality is extremely complex. The objective of this paper is to provide some insight in the problem by providing a partial answer in the case of the 2tfactorial, using the d-optimality criterion.  相似文献   

18.
19.
This paper investigates two estimators under the non-parametric neighbourhoods of an exponential scale parametric family. It uses the relative efficiency approach and shows that the tighter lower bounds on the relative efficiency of the upper trimmed mean to mean can be obtained under a sufficient condition. This condition gives the relationship between the possible positive lower bound and the degree of asymmetry of some related distributions. Similar arguments can be applied to the comparison of dispersion estimators under the neighbourhoods of a normal distribution.  相似文献   

20.
We suggest several constructions suitable to define the depth of set-valued observations with respect to a sample of convex sets or with respect to the distribution of a random closed convex set. With the concept of a depth, it is possible to determine if a given convex set should be regarded an outlier with respect to a sample of convex closed sets. Some of our constructions are motivated by the known concepts of half-space depth and band depth for function-valued data. A novel construction derives the depth from a family of non-linear expectations of random sets. Furthermore, we address the role of positions of sets for evaluation of their depth. Two case studies concern interval regression for Greek wine data and detection of outliers in a sample of particles.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号