共查询到20条相似文献,搜索用时 15 毫秒
1.
This paper considers the problem of selecting a robust threshold of wavelet shrinkage. Previous approaches reported in literature
to handle the presence of outliers mainly focus on developing a robust procedure for a given threshold; this is related to
solving a nontrivial optimization problem. The drawback of this approach is that the selection of a robust threshold, which
is crucial for the resulting fit is ignored. This paper points out that the best fit can be achieved by a robust wavelet shrinkage
with a robust threshold. We propose data-driven selection methods for a robust threshold. These approaches are based on a
coupling of classical wavelet thresholding rules with pseudo data. The concept of pseudo data has influenced the implementation
of the proposed methods, and provides a fast and efficient algorithm. Results from a simulation study and a real example demonstrate
the promising empirical properties of the proposed approaches. 相似文献
2.
Gabriel Huerta 《Journal of applied statistics》2005,32(5):529-542
In recent years, wavelet shrinkage has become a very appealing method for data de-noising and density function estimation. In particular, Bayesian modelling via hierarchical priors has introduced novel approaches for Wavelet analysis that had become very popular, and are very competitive with standard hard or soft thresholding rules. In this sense, this paper proposes a hierarchical prior that is elicited on the model parameters describing the wavelet coefficients after applying a Discrete Wavelet Transformation (DWT). In difference to other approaches, the prior proposes a multivariate Normal distribution with a covariance matrix that allows for correlations among Wavelet coefficients corresponding to the same level of detail. In addition, an extra scale parameter is incorporated that permits an additional shrinkage level over the coefficients. The posterior distribution for this shrinkage procedure is not available in closed form but it is easily sampled through Markov chain Monte Carlo (MCMC) methods. Applications on a set of test signals and two noisy signals are presented. 相似文献
3.
Statistical inference in the wavelet domain remains a vibrant area of contemporary statistical research because of desirable properties of wavelet representations and the need of scientific community to process, explore, and summarize massive data sets. Prime examples are biomedical, geophysical, and internet related data. We propose two new approaches to wavelet shrinkage/thresholding. In the spirit of Efron and Tibshirani's recent work on local false discovery rate, we propose Bayesian Local False Discovery Rate (BLFDR), where the underlying model on wavelet coefficients does not assume known variances. This approach to wavelet shrinkage is shown to be connected with shrinkage based on Bayes factors. The second proposal, Bayesian False Discovery Rate (BaFDR), is based on ordering of posterior probabilities of hypotheses on true wavelets coefficients being null, in Bayesian testing of multiple hypotheses. We demonstrate that both approaches result in competitive shrinkage methods by contrasting them to some popular shrinkage techniques. 相似文献
4.
Taewoon Kong 《统计学通讯:模拟与计算》2017,46(6):4399-4418
We can use wavelet shrinkage to estimate a possibly multivariate regression function g under the general regression setup, y = g + ε. We propose an enhanced wavelet-based denoising methodology based on Bayesian adaptive multiresolution shrinkage, an effective Bayesian shrinkage rule in addition to the semi-supervised learning mechanism. The Bayesian shrinkage rule is advanced by utilizing the semi-supervised learning method in which the neighboring structure of a wavelet coefficient is adopted and an appropriate decision function is derived. According to decision function, wavelet coefficients follow one of two prespecified Bayesian rules obtained using varying related parameters. The decision of a wavelet coefficient depends not only on its magnitude, but also on the neighboring structure on which the coefficient is located. We discuss the theoretical properties of the suggested method and provide recommended parameter settings. We show that the proposed method is often superior to several existing wavelet denoising methods through extensive experimentation. 相似文献
5.
《Journal of statistical planning and inference》1996,49(3):305-317
We introduce the concept of inferential distributions corresponding to inference rules. Fiducial and posterior distributions are special cases. Inferential distributions are essentially unique. They correspond to or represent inference rules and are defined on the parameter space. Not all inference rules can be represented by an inferential distribution. A constructive method is given to investigate its existence for any given inference rule. 相似文献
6.
《Journal of Statistical Computation and Simulation》2012,82(10):741-764
We consider an empirical Bayes approach to standard nonparametric regression estimation using a nonlinear wavelet methodology. Instead of specifying a single prior distribution on the parameter space of wavelet coefficients, which is usually the case in the existing literature, we elicit the ?-contamination class of prior distributions that is particularly attractive to work with when one seeks robust priors in Bayesian analysis. The type II maximum likelihood approach to prior selection is used by maximizing the predictive distribution for the data in the wavelet domain over a suitable subclass of the ?-contamination class of prior distributions. For the prior selected, the posterior mean yields a thresholding procedure which depends on one free prior parameter and it is level- and amplitude-dependent, thus allowing better adaptation in function estimation. We consider an automatic choice of the free prior parameter, guided by considerations on an exact risk analysis and on the shape of the thresholding rule, enabling the resulting estimator to be fully automated in practice. We also compute pointwise Bayesian credible intervals for the resulting function estimate using a simulation-based approach. We use several simulated examples to illustrate the performance of the proposed empirical Bayes term-by-term wavelet scheme, and we make comparisons with other classical and empirical Bayes term-by-term wavelet schemes. As a practical illustration, we present an application to a real-life data set that was collected in an atomic force microscopy study. 相似文献
7.
We suggest a procedure to improve the overall performances of several existing methods for determining the number of factors in factor analysis by using alternative measures of correlation: Pearson's, Spearman's, Gini's, and a robust estimator of the covariance matrix (MCD). We examine the effect of the choice of the covariance used on the number of factors chosen by the KG rule of one, the 80% rule, the Minimum average partial (MAP), and the Parallel Analysis Methodology (PAM). Extensive simulations show that when the entire (or part) of the data come from heavy-tail (lognormal) distributions, ranking the variables which come from non symmetric distributions improves the performances of the methods. In this case, Gini is slightly better than Spearman. The PAM and MAP procedures are qualitatively superior to the KG and the 80% rules in determining the true number of factors. A real example involving data on document authorship is analyzed. 相似文献
8.
Paul H. Garthwaite 《统计学通讯:理论与方法》2013,42(12):3601-3619
A scoring rule for evaluating the usefulness of an assessed prior distribution should reflect the purpose for which the distribution is to be used. In this paper we suppose that sample data is to become available and that the posterior distribution will be used to estimate some quantity under a quadratic loss function. The utility of a prior distribution is consequently determined by its preposterior expected quadratic loss. It is shown that this loss function has properties desirable in a scoring rule and formulae are derived for calculating the scores it gives in some common problems. Many scoring rules give a very poor score to any improper prior distribution but, in contrast, the scoring rule proposed here provides a meaningful measure for comparing the usefulness of assessed prior distributions and non-informative (improper) prior distributions. Results for making this comparison in various situations are also given. 相似文献
9.
In the present work, whenever the response variables are binary, we frame an adaptive allocation rule for a two-treatment two-period crossover design in the presence of possible carry-over effects. The proposed rule is a combination of the play-the-winner and randomized play-the-winner rules. We study various properties of the proposed rule through asymptotics and simulations. Some related inferential problems are also considered. The proposed procedure is compared with some possible competitor. 相似文献
10.
Ying-Ying Zhang 《统计学通讯:理论与方法》2017,46(14):7125-7133
For the variance parameter of the hierarchical normal and inverse gamma model, we analytically calculate the Bayes rule (estimator) with respect to a prior distribution IG (alpha, beta) under Stein's loss function. This estimator minimizes the posterior expected Stein's loss (PESL). We also analytically calculate the Bayes rule and the PESL under the squared error loss. Finally, the numerical simulations exemplify that the PESLs depend only on alpha and the number of observations. The Bayes rules and PESLs under Stein's loss are unanimously smaller than those under the squared error loss. 相似文献
11.
A Bayesian discovery procedure 总被引:1,自引:0,他引:1
Michele Guindani Peter Müller Song Zhang 《Journal of the Royal Statistical Society. Series B, Statistical methodology》2009,71(5):905-925
Summary. We discuss a Bayesian discovery procedure for multiple-comparison problems. We show that, under a coherent decision theoretic framework, a loss function combining true positive and false positive counts leads to a decision rule that is based on a threshold of the posterior probability of the alternative. Under a semiparametric model for the data, we show that the Bayes rule can be approximated by the optimal discovery procedure, which was recently introduced by Storey. Improving the approximation leads us to a Bayesian discovery procedure, which exploits the multiple shrinkage in clusters that are implied by the assumed non-parametric model. We compare the Bayesian discovery procedure and the optimal discovery procedure estimates in a simple simulation study and in an assessment of differential gene expression based on microarray data from tumour samples. We extend the setting of the optimal discovery procedure by discussing modifications of the loss function that lead to different single-thresholding statistics. Finally, we provide an application of the previous arguments to dependent (spatial) data. 相似文献
12.
The problem of classification into two univariate normal populations with a common mean is considered. Several classification rules are proposed based on efficient estimators of the common mean. Detailed numerical comparisons of probabilities of misclassifications using these rules have been carried out. It is shown that the classification rule based on the Graybill-Deal estimator of the common mean performs the best. Classification rules are also proposed for the case when variances are assumed to be ordered. Comparison of these rules with the rule based on the Graybill-Deal estimator has been done with respect to individual probabilities of misclassification. 相似文献
13.
This paper is devoted to robust Bayes sample size determination under the quadratic loss function. The idea behind the proposed approach is that the smaller a chosen posterior functional, the more robust the posterior inference. Such desired posterior functional has been taken, in the literature, as the range of posterior mean over a class of priors but we show that dealing with the posterior mean is not the only method leading to an optimal sample size. To provide an alternative approach, we propose implementing most stable rules into the context of sample size determination. We discuss properties of the desired most stable estimate and provide some examples in the normal model. We then compare the proposed approach with that of a recent global robustness study from both numerical and theoretical aspects. We illustrate the practical utility of our proposed method by analyzing a real data set. 相似文献
14.
We consider a heteroscedastic convolution density model under the “ordinary smooth assumption.” We introduce a new adaptive wavelet estimator based on term-by-term hard thresholding rule. Its asymptotic properties are explored via the minimax approach under the mean integrated squared error over Besov balls. We prove that our estimator attains near optimal rates of convergence (lower bounds are determined). Simulation results are reported to support our theoretical findings. 相似文献
15.
It is often the case that high-dimensional data consist of only a few informative components. Standard statistical modeling and estimation in such a situation is prone to inaccuracies due to overfitting, unless regularization methods are practiced. In the context of classification, we propose a class of regularization methods through shrinkage estimators. The shrinkage is based on variable selection coupled with conditional maximum likelihood. Using Stein's unbiased estimator of the risk, we derive an estimator for the optimal shrinkage method within a certain class. A comparison of the optimal shrinkage methods in a classification context, with the optimal shrinkage method when estimating a mean vector under a squared loss, is given. The latter problem is extensively studied, but it seems that the results of those studies are not completely relevant for classification. We demonstrate and examine our method on simulated data and compare it to feature annealed independence rule and Fisher's rule. 相似文献
16.
This paper concerns wavelet regression using a block thresholding procedure. Block thresholding methods utilize neighboring wavelet coefficients information to increase estimation accuracy. We propose to construct a data-driven block thresholding procedure using the smoothly clipped absolute deviation (SCAD) penalty. A simulation study demonstrates competitive finite sample performance of the proposed estimator compared to existing methods. We also show that the proposed estimator achieves optimal convergence rates in Besov spaces. 相似文献
17.
18.
In this paper, we propose a design that uses a short‐term endpoint for accelerated approval at interim analysis and a long‐term endpoint for full approval at final analysis with sample size adaptation based on the long‐term endpoint. Two sample size adaptation rules are compared: an adaptation rule to maintain the conditional power at a prespecified level and a step function type adaptation rule to better address the bias issue. Three testing procedures are proposed: alpha splitting between the two endpoints; alpha exhaustive between the endpoints; and alpha exhaustive with improved critical value based on correlation. Family‐wise error rate is proved to be strongly controlled for the two endpoints, sample size adaptation, and two analysis time points with the proposed designs. We show that using alpha exhaustive designs greatly improve the power when both endpoints are effective, and the power difference between the two adaptation rules is minimal. The proposed design can be extended to more general settings. Copyright © 2015 John Wiley & Sons, Ltd. 相似文献
19.
Ahmed S. E 《统计学通讯:理论与方法》2013,42(5-6):1273-1291
The improved large sample estimation theory for the probabilities of multi¬nomial distribution is developed under uncertain prior information (UPI) that the true proportion is a known quantity. Several estimators based on pretest and the Stein-type shrinkage rules are constructed. The expressions for the bias and risk of the proposed estimators are derived and compared with the maximum likelihood (ml) estimators. It is demonstrated that the shrinkage estimators are superior to the ml estimators. It is also shown that none of the preliminary test and shrinkage estimators dominate each other, though they perform y/ell relative to the ml estimators. The relative dominance picture of the estimators is presented. A simulation study is carried out to assess the performance of the estimators numerically in small samples. 相似文献
20.
When calculating independently the false alarm rate of the eight usual runs rules used in SPC control chart, it appears that the proposed rule designed to detect mixture patterns corresponds to a Type-I error strongly lower than the seven other rules. This discrepancy is underlined and the mixture rule is showed to be useless both for in-control and out-of-control processes. Thus a modification of the mixture detection rule is proposed and the impact of this new mixture rule is then illustrated and discussed using Monte Carlo calculations. 相似文献