首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
In this note we propose a new and novel kernel density estimator for directly estimating the probability and cumulative distribution function of an L-estimate from a single population based on utilizing the theory in Knight (1985) in conjunction with classic inversion theory. This idea is further developed for a kernel density estimator for the difference of L-estimates from two independent populations. The methodology is developed via a “plug-in” approach, but it is distinct from the classic bootstrap methodology in that it is analytically and computationally feasible to provide an exact estimate of the distribution function and thus eliminates the resampling related error. The asymptotic and finite sample properties of our estimators are examined. The procedure is illustrated via generating the kernel density estimate for the Tukey's trimean from a small data set.  相似文献   

2.
Large Deviations Limit Theorems for the Kernel Density Estimator   总被引:2,自引:0,他引:2  
We establish pointwise and uniform large deviations limit theorems of Chernoff-type for the non-parametric kernel density estimator based on a sequence of independent and identically distributed random variables. The limits are well-identified and depend upon the underlying kernel and density function. We derive then some implications of our results in the study of asymptotic efficiency of the goodness-of-fit test based on the maximal deviation of the kernel density estimator as well as the inaccuracy rate of this estimate  相似文献   

3.
In this article, we develop a new and novel kernel density estimator for a sum of weighted averages from a single population based on utilizing the well defined kernel density estimator in conjunction with classic inversion theory. This idea is further developed for a kernel density estimator for the difference of weighed averages from two independent populations. The resulting estimator is “bootstrap-like” in terms of its properties with respect to the derivation of approximate confidence intervals via a “plug-in” approach. This new approach is distinct from the bootstrap methodology in that it is analytically and computationally feasible to provide an exact estimate of the distribution function through direct calculation. Thus, our approach eliminates the error due to Monte Carlo resampling that arises within the context of simulation based approaches that are oftentimes necessary in order to derive bootstrap-based confidence intervals for statistics involving weighted averages of i.i.d. random variables. We provide several examples and carry forth a simulation study to show that our kernel density estimator performs better than the standard central limit theorem based approximation in term of coverage probability.  相似文献   

4.
The purpose of this paper is to prove, through the analysis of the behaviour of a standard kernel density estimator, that the notion of weak dependence defined in a previous paper (cf. Doukhan & Louhichi, 1999) has sufficiently sharp properties to be used in various situations. More precisely we investigate the asymptotics of high order losses, asymptotic distributions and uniform almost sure behaviour of kernel density estimates. We prove that they are the same as for independent samples (with some restrictions for a.s. behaviours). Recall finally that this weak dependence condition extends on the previously defined ones such as mixing, association and it allows considerations of new classes such as weak shifts processes based on independent sequences as well as some non-mixing Markov processes.  相似文献   

5.
Abstract

Multistage sampling is a common sampling technique employed in many studies. In this setting, observations are identically distributed but not independent, thus many traditional kernel smoothing techniques, which assume that the data are independent and identically distributed (i.i.d.), may not produce reasonable density estimates. In this paper, we sample repeatedly with replacement from each cluster, create multiple i.i.d. samples containing one observation from each cluster, and then create a kernel density estimate from each i.i.d. sample. These estimates will then be combined to form an estimate of the marginal probability density function of the population.  相似文献   

6.
A simple random sample on a random variable A allows its density to be consistently estimated, by a histogram or preferably a kernel density estimate. When the sampling is biased towards certain x-values these methods instead estimate a weighted version of the density function. This article proposes a method for estimating both the density and the sampling bias simultaneously. The technique requires two independent samples and utilises ideas from mark-recapture experiments. An estimator of the size of the sampled population also follows simply from this density estimate.  相似文献   

7.
Abstract. Several old and new density estimators may have good theoretical performance, but are hampered by not being bona fide densities; they may be negative in certain regions or may not integrate to 1. One can therefore not simulate from them, for example. This paper develops general modification methods that turn any density estimator into one which is a bona fide density, and which is always better in performance under one set of conditions and arbitrarily close in performance under a complementary set of conditions. This improvement-for-free procedure can, in particular, be applied for higher-order kernel estimators, classes of modern h 4 bias kernel type estimators, superkernel estimators, the sinc kernel estimator, the k -NN estimator, orthogonal expansion estimators, and for various recently developed semi-parametric density estimators.  相似文献   

8.
Kernel density classification and boosting: an L2 analysis   总被引:1,自引:0,他引:1  
Kernel density estimation is a commonly used approach to classification. However, most of the theoretical results for kernel methods apply to estimation per se and not necessarily to classification. In this paper we show that when estimating the difference between two densities, the optimal smoothing parameters are increasing functions of the sample size of the complementary group, and we provide a small simluation study which examines the relative performance of kernel density methods when the final goal is classification.A relative newcomer to the classification portfolio is boosting, and this paper proposes an algorithm for boosting kernel density classifiers. We note that boosting is closely linked to a previously proposed method of bias reduction in kernel density estimation and indicate how it will enjoy similar properties for classification. We show that boosting kernel classifiers reduces the bias whilst only slightly increasing the variance, with an overall reduction in error. Numerical examples and simulations are used to illustrate the findings, and we also suggest further areas of research.  相似文献   

9.
Nonparametric density estimation in the presence of measurement error is considered. The usual kernel deconvolution estimator seeks to account for the contamination in the data by employing a modified kernel. In this paper a new approach based on a weighted kernel density estimator is proposed. Theoretical motivation is provided by the existence of a weight vector that perfectly counteracts the bias in density estimation without generating an excessive increase in variance. In practice a data driven method of weight selection is required. Our strategy is to minimize the discrepancy between a standard kernel estimate from the contaminated data on the one hand, and the convolution of the weighted deconvolution estimate with the measurement error density on the other hand. We consider a direct implementation of this approach, in which the weights are optimized subject to sum and non-negativity constraints, and a regularized version in which the objective function includes a ridge-type penalty. Numerical tests suggest that the weighted kernel estimation can lead to tangible improvements in performance over the usual kernel deconvolution estimator. Furthermore, weighted kernel estimates are free from the problem of negative estimation in the tails that can occur when using modified kernels. The weighted kernel approach generalizes to the case of multivariate deconvolution density estimation in a very straightforward manner.  相似文献   

10.
The author considers the estimation of the common probability density of independent and identically distributed random variables observed with added white noise. She assumes that the unknown density belongs to some class of supersmooth functions, and that the error distribution is ordinarily smooth, meaning that its characteristic function decays polynomially asymptotically. In this context, the author evaluates the minimax rate of convergence of the pointwise risk and describes a kernel estimator having this rate. She computes upper bounds for the L2 risk of this estimator.  相似文献   

11.
12.
The standard approach to non-parametric bivariate density estimation is to use a kernel density estimator. Practical performance of this estimator is hindered by the fact that the estimator is not adaptive (in the sense that the level of smoothing is not sensitive to local properties of the density). In this paper a simple, automatic and adaptive bivariate density estimator is proposed based on the estimation of marginal and conditional densities. Asymptotic properties of the estimator are examined, and guidance to practical application of the method is given. Application to two examples illustrates the usefulness of the estimator as an exploratory tool, particularly in situations where the local behaviour of the density varies widely. The proposed estimator is also appropriate for use as a pilot estimate for an adaptive kernel estimate, since it is relatively inexpensive to calculate.  相似文献   

13.
This paper considers linear and nonlinear regression with a response variable that is allowed to be “missing at random”. The only structural assumptions on the distribution of the variables are that the errors have mean zero and are independent of the covariates. The independence assumption is important. It enables us to construct an estimator for the response density that uses all the observed data, in contrast to the usual local smoothing techniques, and which therefore permits a faster rate of convergence. The idea is to write the response density as a convolution integral which can be estimated by an empirical version, with a weighted residual-based kernel estimator plugged in for the error density. For an appropriate class of regression functions, and a suitably chosen bandwidth, this estimator is consistent and converges with the optimal parametric rate n1/2. Moreover, the estimator is proved to be efficient (in the sense of Hájek and Le Cam) if an efficient estimator is used for the regression parameter.  相似文献   

14.
Problems with truncated data arise frequently in survival analyses and reliability applications. The estimation of the density function of the lifetimes is often of interest. In this article, the estimation of density function by the kernel method is considered, when truncated data are showing some kind of dependence. We apply the strong Gaussian approximation technique to study the strong uniform consistency for kernel estimators of the density function under a truncated dependent model. We also apply the strong approximation results to study the integrated square error properties of the kernel density estimators under the truncated dependent scheme.  相似文献   

15.
In some experiments, such as destructive stress testing and industrial quality control experiments, only values smaller than all previous ones are observed. Here, for such record-breaking data, kernel estimation of the cumulative distribution function and smooth density estimation is considered. For a single record-breaking sample, consistent estimation is not possible, and replication is required for global results. For m independent record-breaking samples, the proposed distribution function and density estimators are shown to be strongly consistent and asymptotically normal as m → ∞. Also, for small m, the mean squared errors and biases of the estimators and their smoothing parameters are investigated through computer simulations.  相似文献   

16.
In this paper, the kernel density estimator for negatively superadditive dependent random variables is studied. The exponential inequalities and the exponential rate for the kernel estimator of density function with a uniform version, over compact sets are investigated. Also, the optimal bandwidth rate of the estimator is obtained using mean integrated squared error. The results are generalized and used to improve the ones obtained for the case of associated sequences. As an application, FGM sequences that fulfil our assumptions are investigated. Also, the convergence rate of the kernel density estimator is illustrated via a simulation study. Moreover, a real data analysis is presented.  相似文献   

17.
This paper presents a method for Bayesian inference for the regression parameters in a linear model with independent and identically distributed errors that does not require the specification of a parametric family of densities for the error distribution. This method first selects a nonparametric kernel density estimate of the error distribution which is unimodal and based on the least-squares residuals. Once the error distribution is selected, the Metropolis algorithm is used to obtain the marginal posterior distribution of the regression parameters. The methodology is illustrated with data sets, and its performance relative to standard Bayesian techniques is evaluated using simulation results.  相似文献   

18.
The commonly used survey technique of clustering introduces dependence into sample data. Such data is frequently used in economic analysis, though the dependence induced by the sample structure of the data is often ignored. In this paper, the effect of clustering on the non-parametric, kernel estimate of the density, f(x), is examined. The window width commonly used for density estimation for the case of i.i.d. data is shown to no longer be optimal. A new optimal bandwidth using a higher-order kernel is proposed and is shown to give a smaller integrated mean squared error than two window widths which are widely used for the case of i.i.d. data. Several illustrations from simulation are provided.  相似文献   

19.
The normal/independent family of distributions is an attractive class of symmetric heavy-tailed density functions. They have a nice hierarchical representation to make inferences easily. We propose the Sinh-normal/independent distribution which extends the Sinh-normal (SN) distribution [23]. We discuss some of its properties and propose the Sinh-normal/independent nonlinear regression model based on a similar setup of Lemonte and Cordeiro [18], who applied the Birnbaum–Saunders distribution. We develop an EM-algorithm for maximum likelihood estimation of the model parameters. In order to examine the robustness of this flexible class against outlying observations, we perform a simulation study and analyze a real data set to illustrate the usefulness of the new model.  相似文献   

20.
The performance of kernel density estimation, in terms of mean integrated squared error, is investigated in the opposite of the usual situation, namely when the bandwidth is large. This affords noteworthy insights including the special role taken by the normal density function as kernel and a tie-in with ‘semiparametric’ approaches to density estimation.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号