首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The paper compares several methods for computing robust 1-α confidence intervals for σ 1 2-σ 2 2, or σ 1 2/σ 2 2, where σ 1 2 and σ 2 2 are the population variances corresponding to two independent treatment groups. The emphasis is on a Box-Scheffe approach when distributions have different shapes, and so the results reported here have implications about comparing means. The main result is that for unequal sample sizes, a Box-Scheffe approach can be considerably less robust than indicated by past investigations. Several other procedures for comparing variances, not based on a Box-Scheffe approach, were also examined and found to be highly unsatisfactory although previously published papers found them to be robust when the distributions have identical shapes. Included is a new result on why the procedures examined here are not robust, and an illustration that increasing σ 1 2-σ 2 2 can reduce power in certain situations. Constants needed to apply Dunnett’s robust comparison of means are included.  相似文献   

2.
Fitting multiplicative models by robust alternating regressions   总被引:1,自引:0,他引:1  
In this paper a robust approach for fitting multiplicative models is presented. Focus is on the factor analysis model, where we will estimate factor loadings and scores by a robust alternating regression algorithm. The approach is highly robust, and also works well when there are more variables than observations. The technique yields a robust biplot, depicting the interaction structure between individuals and variables. This biplot is not predetermined by outliers, which can be retrieved from the residual plot. Also provided is an accompanying robust R 2-plot to determine the appropriate number of factors. The approach is illustrated by real and artificial examples and compared with factor analysis based on robust covariance matrix estimators. The same estimation technique can fit models with both additive and multiplicative effects (FANOVA models) to two-way tables, thereby extending the median polish technique.  相似文献   

3.
Multivariate mixture regression models can be used to investigate the relationships between two or more response variables and a set of predictor variables by taking into consideration unobserved population heterogeneity. It is common to take multivariate normal distributions as mixing components, but this mixing model is sensitive to heavy-tailed errors and outliers. Although normal mixture models can approximate any distribution in principle, the number of components needed to account for heavy-tailed distributions can be very large. Mixture regression models based on the multivariate t distributions can be considered as a robust alternative approach. Missing data are inevitable in many situations and parameter estimates could be biased if the missing values are not handled properly. In this paper, we propose a multivariate t mixture regression model with missing information to model heterogeneity in regression function in the presence of outliers and missing values. Along with the robust parameter estimation, our proposed method can be used for (i) visualization of the partial correlation between response variables across latent classes and heterogeneous regressions, and (ii) outlier detection and robust clustering even under the presence of missing values. We also propose a multivariate t mixture regression model using MM-estimation with missing information that is robust to high-leverage outliers. The proposed methodologies are illustrated through simulation studies and real data analysis.  相似文献   

4.
Trimming principles play an important role in robust statistics. However, their use for clustering typically requires some preliminary information about the contamination rate and the number of groups. We suggest a fresh approach to trimming that does not rely on this knowledge and that proves to be particularly suited for solving problems in robust cluster analysis. Our approach replaces the original K‐population (robust) estimation problem with K distinct one‐population steps, which take advantage of the good breakdown properties of trimmed estimators when the trimming level exceeds the usual bound of 0.5. In this setting, we prove that exact affine equivariance is lost on one hand but, on the other hand, an arbitrarily high breakdown point can be achieved by “anchoring” the robust estimator. We also support the use of adaptive trimming schemes, in order to infer the contamination rate from the data. A further bonus of our methodology is its ability to provide a reliable choice of the usually unknown number of groups.  相似文献   

5.
This contribution deals with the statistical problem of evaluating the stress–strength reliability parameter R = P(X < Y), when both stress and strength data are prone to contamination. Standard likelihood inference can be badly affected by mild data inadequacies, that often occur in the form of several outliers. Then, robust tools are recommended. Here, inference relies on the weighted likelihood methodology. This approach has the advantage to lead to robust estimators, tests, and confidence intervals that share the main asymptotic properties of their classical counterparts. The accuracy of the proposed methodology is illustrated both by numerical studies and real-data applications.  相似文献   

6.
In this paper, we consider the problem of robust estimation of the fractional parameter, d, in long memory autoregressive fractionally integrated moving average processes, when two types of outliers, i.e. additive and innovation, are taken into account without knowing their number, position or intensity. The proposed method is a weighted likelihood estimation (WLE) approach for which needed definitions and algorithm are given. By an extensive Monte Carlo simulation study, we compare the performance of the WLE method with the performance of both the approximated maximum likelihood estimation (MLE) and the robust M-estimator proposed by Beran (Statistics for Long-Memory Processes, Chapman & Hall, London, 1994). We find that robustness against the two types of considered outliers can be achieved without loss of efficiency. Moreover, as a byproduct of the procedure, we can classify the suspicious observations in different kinds of outliers. Finally, we apply the proposed methodology to the Nile River annual minima time series.  相似文献   

7.
In the past decade, different robust estimators have been proposed by several researchers to improve the ability to detect non-random patterns such as trend, process mean shift, and outliers in multivariate control charts. However, the use of the sample mean vector and the mean square successive difference matrix in the T 2 control chart is sensitive in detecting process mean shift or trend but less sensitive in detecting outliers. On the other hand, the minimum volume ellipsoid (MVE) estimators in the T 2 control chart are sensitive in detecting multiple outliers but less sensitive in detecting trend or process mean shift. Therefore, new robust estimators using both merits of the mean square successive difference matrix and the MVE estimators are developed to modify Hotelling's T 2 control chart. To compare the detection performance among various control charts, a simulation approach for establishing control limits and calculating signal probabilities is provided as well. Our simulation results show that a multivariate control chart using the new robust estimators can achieve a well-balanced sensitivity in detecting the above-mentioned non-random patterns. Finally, three numerical examples further demonstrate the usefulness of our new robust estimators.  相似文献   

8.
Abstract

This article introduces a parametric robust way of comparing two population means and two population variances. With large samples the comparison of two means, under model misspecification, is lesser a problem, for, the validity of inference is protected by the central limit theorem. However, the assumption of normality is generally required, so that the inference for the ratio of two variances can be carried out by the familiar F statistic. A parametric robust approach that is insensitive to the distributional assumption will be proposed here. More specifically, it will be demonstrated that the normal likelihood function can be adjusted for asymptotically valid inferences for all underlying distributions with finite fourth moments. The normal likelihood function, on the other hand, is itself robust for the comparison of two means so that no adjustment is needed.  相似文献   

9.
The trend test is often used for the analysis of 2×K ordered categorical data, in which K pre-specified increasing scores are used. There have been discussions on how to assign these scores and the impact of the outcomes on different scores. The scores are often assigned based on the data-generating model. When this model is unknown, using the trend test is not robust. We discuss the weighted average of a trend test over all scientifically plausible choices of scores or models. This approach is more computationally efficient than a commonly used robust test MAX when K is large. Our discussion is for any ordered 2×K table, but simulation and applications to real data are focused on case-control genetic association studies. Although there is no single test optimal for all choices of scores, our numerical results show that some score averaging tests can achieve the performance of MAX.  相似文献   

10.
When process data follow a particular curve in quality control, profile monitoring is suitable and appropriate for assessing process stability. Previous research in profile monitoring focusing on nonlinear parametric (P) modeling, involving both fixed and random-effects, was made under the assumption of an accurate nonlinear model specification. Lately, nonparametric (NP) methods have been used in the profile monitoring context in the absence of an obvious linear P model. This study introduces a novel technique in profile monitoring for any nonlinear and auto-correlated data. Referred to as the nonlinear mixed robust profile monitoring (NMRPM) method, it proposes a semiparametric (SP) approach that combines nonlinear P and NP profile fits for scenarios in which a nonlinear P model is adequate over part of the data but inadequate of the rest. These three methods (P, NP, and NMRPM) account for the auto-correlation within profiles and treats the collection of profiles as a random sample with a common population. During Phase I analysis, a version of Hotelling’s T2 statistic is proposed for each approach to identify abnormal profiles based on the estimated random effects and obtain the corresponding control limits. The performance of the NMRPM method is then evaluated using a real data set. Results reveal that the NMRPM method is robust to model misspecification and performs adequately against a correctly specified nonlinear P model. Control charts with the NMRPM method have excellent capability of detecting changes in Phase I data with control limits that are easily computable.  相似文献   

11.
Although the t-type estimator is a kind of M-estimator with scale optimization, it has some advantages over the M-estimator. In this article, we first propose a t-type joint generalized linear model as a robust extension to the classical joint generalized linear models for modeling data containing extreme or outlying observations. Next, we develop a t-type pseudo-likelihood (TPL) approach, which can be viewed as a robust version to the existing pseudo-likelihood (PL) approach. To determine which variables significantly affect the variance of the response variable, we then propose a unified penalized maximum TPL method to simultaneously select significant variables for the mean and dispersion models in t-type joint generalized linear models. Thus, the proposed variable selection method can simultaneously perform parameter estimation and variable selection in the mean and dispersion models. With appropriate selection of the tuning parameters, we establish the consistency and the oracle property of the regularized estimators. Simulation studies are conducted to illustrate the proposed methods.  相似文献   

12.
We investigate and develop methods for structural break detection, considering time series from thermal spraying process monitoring. Since engineers induce technical malfunctions during the processes, the time series exhibit structural breaks at known time points, giving us valuable information to conduct the investigations. First, we consider a recently developed robust online (also real-time) filtering (i.e. smoothing) procedure that comprises a test for local linearity. This test rejects when jumps and trend changes are present, so that it can also be useful to detect such structural breaks online. Second, based on the filtering procedure we develop a robust method for the online detection of ongoing trends. We investigate these two methods as to the online detection of structural breaks by simulations and applications to the time series from the manipulated spraying processes. Third, we consider a recently developed fluctuation test for constant variances that can be applied offline, i.e. after the whole time series has been observed, to control the spraying results. Since this test is not reliable when jumps are present in the time series, we suggest data transformation based on filtering and demonstrate that this transformation makes the test applicable.  相似文献   

13.
The autoregressive Cauchy estimator uses the sign of the first lag as instrumental variable (IV); under independent and identically distributed (i.i.d.) errors, the resulting IV t-type statistic is known to have a standard normal limiting distribution in the unit root case. With unconditional heteroskedasticity, the ordinary least squares (OLS) t statistic is affected in the unit root case; but the paper shows that, by using some nonlinear transformation behaving asymptotically like the sign as instrument, limiting normality of the IV t-type statistic is maintained when the series to be tested has no deterministic trends. Neither estimation of the so-called variance profile nor bootstrap procedures are required to this end. The Cauchy unit root test has power in the same 1/T neighborhoods as the usual unit root tests, also for a wide range of magnitudes for the initial value. It is furthermore shown to be competitive with other, bootstrap-based, robust tests. When the series exhibit a linear trend, however, the null distribution of the Cauchy test for a unit root becomes nonstandard, reminiscent of the Dickey-Fuller distribution. In this case, inference robust to nonstationary volatility is obtained via the wild bootstrap.  相似文献   

14.
ABSTRACT

In many statistical applications estimation of population quantiles is desired. In this study, a log–flip–robust (LFR) approach is proposed to estimate, specifically, lower-end quantiles (those below the median) from a continuous, positive, right-skewed distribution. Characteristics of common right-skewed distributions suggest that a logarithm transformation (L) followed by flipping the lower half of the sample (F) allows for the estimation of the lower-end quantile using robust methods (R) based on symmetric populations. Simulations show that this approach is superior in many cases to current methods, while not suffering from the sample size restrictions of other approaches.  相似文献   

15.
ABSTRACT

In this article, we study the recursive kernel estimator of the conditional quantile of a scalar response variable Y given a random variable (rv) X taking values in a semi-metric space. Two estimators are considered. While the first one is given by inverting the double-kernel estimate of the conditional distribution function, the second estimator is obtained by using the robust approach. We establish the almost complete consistency of these estimates when the observations are sampled from a functional ergodic process. Finally, a simulation study is carried out to illustrate the finite sample performance of these estimators.  相似文献   

16.
Despite the long and frustrating history of struggling with the wrong signs or other types of implausible estimates under multicollinearity, it turns out that the problem can be solved in a surprisingly easy way. This paper presents a simple approach that ensures both statistically sound and theoretically consistent estimates under multicollinearity. The approach is simple in the sense that it requires nothing but basic statistical methods plus a piece of a priori knowledge. In addition, the approach is robust even to the extreme case when the a priori knowledge is wrong. A simulation test shows astonishingly superior performance of the method in repeated samples comparing to the OLS, the Ridge Regression and the Dropping-Variable approach.  相似文献   

17.
ABSTRACT

This paper deals with Bayes, robust Bayes, and minimax predictions in a subfamily of scale parameters under an asymmetric precautionary loss function. In Bayesian statistical inference, the goal is to obtain optimal rules under a specified loss function and an explicit prior distribution over the parameter space. However, in practice, we are not able to specify the prior totally or when a problem must be solved by two statisticians, they may agree on the choice of the prior but not the values of the hyperparameters. A common approach to the prior uncertainty in Bayesian analysis is to choose a class of prior distributions and compute some functional quantity. This is known as Robust Bayesian analysis which provides a way to consider the prior knowledge in terms of a class of priors Γ for global prevention against bad choices of hyperparameters. Under a scale invariant precautionary loss function, we deal with robust Bayes predictions of Y based on X. We carried out a simulation study and a real data analysis to illustrate the practical utility of the prediction procedure.  相似文献   

18.
A robust biplot     
This paper introduces a robust biplot which is related to multivariate M-estimates. The n × p data matrix is first considered as a sample of size n from some p-variate population, and robust M-estimates of the population location vector and scatter matrix are calculated. In the construction of the biplot, each row of the data matrix is assigned a weight determined in the preliminary robust estimation. In a robust biplot, one can plot the variables in order to represent characteristics of the robust variance-covariance matrix: the length of the vector representing a variable is proportional to its robust standard deviation, while the cosine of the angle between two variables is approximately equal to their robust correlation. The proposed biplot also permits a meaningful representation of the variables in a robust principal-component analysis. The discrepancies between least-squares and robust biplots are illustrated in an example.  相似文献   

19.
Consider the situation where measurements are taken at two different times and let Mj(x) be some conditional robust measure of location associated with the random variable Y at time j, given that some covariate X=x. The goal is to test H0: M1(x)=M2(x) for each xx1,?…?, xK such that the probability of one or more Type I errors is less than α, where x1,?…?, xK are K specified values of the covariate. The paper reports simulation results comparing two methods aimed at accomplishing this goal without specifying some parametric form for the regression line. The first method is based on a simple modification of the method in Wilcox [Introduction to robust estimation and hypothesis testing. 3rd ed. San Diego, CA: Academic Press; 2012, Section 11.11.1]. The main result here is that the second method, which has never been studied, can have higher power, sometimes substantially so. Data from the Well Elderly 2 study, which motivated this paper, are used to illustrate that the alternative approach can make a practical difference. Here, the estimate of Mj(x) is based in part on either a 20% trimmed mean or the Harrell–Davis quantile estimator, but in principle the more successful method can be used with any robust location estimator.  相似文献   

20.
Under some nonstochastic linear restrictions based on either additional information or prior knowledge in a semiparametric regression model, a family of feasible generalized robust estimators for the regression parameter is proposed. The least trimmed squares (LTS) method proposed by Rousseeuw as a highly robust regression estimator is a statistical technique for fitting a regression model based on the subset of h observations (out of n) whose least-square fit possesses the smallest sum of squared residuals. The coverage h may be set between n/2 and n. The LTS estimator involves computing the hyperplane that minimizes the sum of the smallest h squared residuals. For practical purpose, it is assumed that the covariance matrix of the error term is unknown and thus feasible estimators are replaced. Then, we develop an algorithm for the LTS estimator based on feasible methods. Through the Monte Carlo simulation studies and a real data example, performance of the feasible type of robust estimators is compared with the classical ones in restricted semiparametric regression models.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号