首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
In this article, we propose a class of partial deconvolution kernel estimators for the nonparametric regression function when some covariates are measured with error and some are not. The estimation procedure combines the classical kernel methodology and the deconvolution kernel technique. According to whether the measurement error is ordinarily smooth or supersmooth, we establish the optimal local and global convergence rates for these proposed estimators, and the optimal bandwidths are also identified. Furthermore, lower bounds for the convergence rates of all possible estimators for the nonparametric regression functions are developed. It is shown that, in both the super and ordinarily smooth cases, the convergence rates of the proposed partial deconvolution kernel estimators attain the lower bound. The Canadian Journal of Statistics 48: 535–560; 2020 © 2020 Statistical Society of Canada  相似文献   

2.
We consider nonparametric estimation problems in the presence of dependent data, notably nonparametric regression with random design and nonparametric density estimation. The proposed estimation procedure is based on a dimension reduction. The minimax optimal rate of convergence of the estimator is derived assuming a sufficiently weak dependence characterised by fast decreasing mixing coefficients. We illustrate these results by considering classical smoothness assumptions. However, the proposed estimator requires an optimal choice of a dimension parameter depending on certain characteristics of the function of interest, which are not known in practice. The main issue addressed in our work is an adaptive choice of this dimension parameter combining model selection and Lepski's method. It is inspired by the recent work of Goldenshluger and Lepski [(2011), ‘Bandwidth Selection in Kernel Density Estimation: Oracle Inequalities and Adaptive Minimax Optimality’, The Annals of Statistics, 39, 1608–1632]. We show that this data-driven estimator can attain the lower risk bound up to a constant provided a fast decay of the mixing coefficients.  相似文献   

3.
We propose a modification of local polynomial estimation which improves the efficiency of the conventional method when the observation errors are correlated. The procedure is based on a pre-transformation of the data as a generalization of the pre-whitening procedure introduced by Xiao et al. [(2003), ‘More Efficient Local Polynomial Estimation in Nonparametric Regression with Autocorrelated Errors’, Journal of the American Statistical Association, 98, 980–992]. While these authors assumed a linear process representation for the error process, we avoid any structural assumption. We further allow the regressors and the errors to be dependent. More importantly, we show that the inclusion of both leading and lagged variables in the approximation of the error terms outperforms the best approximation based on lagged variables only. Establishing its asymptotic distribution, we show that the proposed estimator is more efficient than the standard local polynomial estimator. As a by-product we prove a suitable version of a central limit theorem which allows us to improve the asymptotic normality result for local polynomial estimators by Masry and Fan [(1997), ‘Local Polynomial Estimation of Regression Functions for Mixing Processes’, Scandinavian Journal of Statistics, 24, 165–179]. A simulation study confirms the efficiency of our estimator on finite samples. An application to climate data also shows that our new method leads to an estimator with decreased variability.  相似文献   

4.
This paper deals with the problem of estimating the multivariate version of the Conditional-Tail-Expectation, proposed by Di Bernardino et al. [(2013), ‘Plug-in Estimation of Level Sets in a Non-Compact Setting with Applications in Multivariable Risk Theory’, ESAIM: Probability and Statistics, (17), 236–256]. We propose a new nonparametric estimator for this multivariate risk-measure, which is essentially based on Kendall's process [Genest and Rivest, (1993), ‘Statistical Inference Procedures for Bivariate Archimedean Copulas’, Journal of American Statistical Association, 88(423), 1034–1043]. Using the central limit theorem for Kendall's process, proved by Barbe et al. [(1996), ‘On Kendall's Process’, Journal of Multivariate Analysis, 58(2), 197–229], we provide a functional central limit theorem for our estimator. We illustrate the practical properties of our nonparametric estimator on simulations and on two real test cases. We also propose a comparison study with the level sets-based estimator introduced in Di Bernardino et al. [(2013), ‘Plug-In Estimation of Level Sets in A Non-Compact Setting with Applications in Multivariable Risk Theory’, ESAIM: Probability and Statistics, (17), 236–256] and with (semi-)parametric approaches.  相似文献   

5.
Pretest–posttest studies are an important and popular method for assessing the effectiveness of a treatment or an intervention in many scientific fields. While the treatment effect, measured as the difference between the two mean responses, is of primary interest, testing the difference of the two distribution functions for the treatment and the control groups is also an important problem. The Mann–Whitney test has been a standard tool for testing the difference of distribution functions with two independent samples. We develop empirical likelihood-based (EL) methods for the Mann–Whitney test to incorporate the two unique features of pretest–posttest studies: (i) the availability of baseline information for both groups; and (ii) the structure of the data with missing by design. Our proposed methods combine the standard Mann–Whitney test with the EL method of Huang, Qin and Follmann [(2008), ‘Empirical Likelihood-Based Estimation of the Treatment Effect in a Pretest–Posttest Study’, Journal of the American Statistical Association, 103(483), 1270–1280], the imputation-based empirical likelihood method of Chen, Wu and Thompson [(2015), ‘An Imputation-Based Empirical Likelihood Approach to Pretest–Posttest Studies’, The Canadian Journal of Statistics accepted for publication], and the jackknife empirical likelihood method of Jing, Yuan and Zhou [(2009), ‘Jackknife Empirical Likelihood’, Journal of the American Statistical Association, 104, 1224–1232]. Theoretical results are presented and finite sample performances of proposed methods are evaluated through simulation studies.  相似文献   

6.
We study the class of bivariate penalised splines that use tensor product splines and a smoothness penalty. Similar to Claeskens, G., Krivobokova, T., and Opsomer, J.D. [(2009), ‘Asymptotic Properties of Penalised Spline Estimators’, Biometrika, 96(3), 529–544] for the univariate penalised splines, we show that, depending on the number of knots and penalty, the global asymptotic convergence rate of bivariate penalised splines is either similar to that of tensor product regression splines or to that of thin plate splines. In each scenario, the bivariate penalised splines are found rate optimal in the sense of Stone, C.J. [(12, 1982), ‘Optimal Global Rates of Convergence for Nonparametric Regression’, The Annals of Statistics, 10(4), 1040–1053] for a corresponding class of functions with appropriate smoothness. For the scenario where a small number of knots is used, we obtain expressions for the local asymptotic bias and variance and derive the point-wise and uniform asymptotic normality. The theoretical results are applicable to tensor product regression splines.  相似文献   

7.
Recently, Lad, Sanfilippo, and Agro [(2015), ‘Extropy: Complementary Dual of Entropy’, Statistical Science, 30, 40–58.] showed the measure of entropy has a complementary dual, which is termed extropy. The present article introduces some estimators of the extropy of a continuous random variable. Properties of the proposed estimators are stated, and comparisons are made with Qiu and Jia’s estimators [(2018a), ‘Extropy Estimators with Applications in Testing uniformity’, Journal of Nonparametric Statistics, 30, 182–196]. The results indicate that the proposed estimators have a smaller mean squared error than competing estimators. A real example is presented and analysed.  相似文献   

8.
Semi-competing risks data arise when two types of events, non-terminal and terminal, may be observed. When the terminal event occurs first, it censors the non-terminal event. Otherwise the terminal event is observable after the occurrence of the non-terminal event. In practice, it can be hard to ascertain all terminal event information after the non-terminal event. Yu and Yiannoutsos [(2015), ‘Marginal and Conditional Distribution Estimation from Double-Sampled Semi-Competing Risks Data’, Scandinavian Journal of Statistics, 42, 87–103] considered a setting when the terminal event is ascertained via double sampling from only a subset of patients who experienced the non-terminal event. They discussed estimation for marginal and conditional distributions under this double sampled semi-competing risk data framework. We propose a more efficient estimation method in the same setting by fully utilising the non-terminal event information. The efficiency gain can be substantial as observed in our simulation study.  相似文献   

9.
A finite mixture model is considered in which the mixing probabilities vary from observation to observation. A parametric model is assumed for one mixture component distribution, while the others are nonparametric nuisance parameters. Generalized estimating equations (GEE) are proposed for the semi‐parametric estimation. Asymptotic normality of the GEE estimates is demonstrated and the lower bound for their dispersion (asymptotic covariance) matrix is derived. An adaptive technique is developed to derive estimates with nearly optimal small dispersion. An application to the sociological analysis of voting results is discussed. The Canadian Journal of Statistics 41: 217–236; 2013 © 2013 Statistical Society of Canada  相似文献   

10.
Abstract.  A new semiparametric method for density deconvolution is proposed, based on a model in which only the ratio of the unconvoluted to convoluted densities is specified parametrically. Deconvolution results from reweighting the terms in a standard kernel density estimator, where the weights are defined by the parametric density ratio. We propose that in practice, the density ratio be modelled on the log-scale as a cubic spline with a fixed number of knots. Parameter estimation is based on maximization of a type of semiparametric likelihood. The resulting asymptotic properties for our deconvolution estimator mirror the convergence rates in standard density estimation without measurement error when attention is restricted to our semiparametric class of densities. Furthermore, numerical studies indicate that for practical sample sizes our weighted kernel estimator can provide better results than the classical non-parametric kernel estimator for a range of densities outside the specified semiparametric class.  相似文献   

11.
In this note, we consider the problem of the existence of adaptive confidence bands in the fixed design regression model, adapting ideas in Hoffmann and Nickl [(2011), ‘On Adaptive Inference and Confidence Bands’, Annals of Statistics, 39, 2383–2409] to the present case. In the course of the proof, we show that sup-norm adaptive estimators exist as well in the regression setting.  相似文献   

12.
Abstract.  We derive the asymptotic distribution of the integrated square error of a deconvolution kernel density estimator in supersmooth deconvolution problems. Surprisingly, in contrast to direct density estimation as well as ordinary smooth deconvolution density estimation, the asymptotic distribution is no longer a normal distribution but is given by a normalized chi-squared distribution with 2 d.f. A simulation study shows that the speed of convergence to the asymptotic law is reasonably fast.  相似文献   

13.
The study of differences among groups is an interesting statistical topic in many applied fields. It is very common in this context to have data that are subject to mechanisms of loss of information, such as censoring and truncation. In the setting of a two‐sample problem with data subject to left truncation and right censoring, we develop an empirical likelihood method to do inference for the relative distribution. We obtain a nonparametric generalization of Wilks' theorem and construct nonparametric pointwise confidence intervals for the relative distribution. Finally, we analyse the coverage probability and length of these confidence intervals through a simulation study and illustrate their use with a real data set on gastric cancer. The Canadian Journal of Statistics 38: 453–473; 2010 © 2010 Statistical Society of Canada  相似文献   

14.
The performance of nonparametric function estimates often depends on the choice of design points. Based on the mean integrated squared error criterion, we propose a sequential design procedure that updates the model knowledge and optimal design density sequentially. The methodology is developed under a general framework covering a wide range of nonparametric inference problems, such as conditional mean and variance functions, the conditional distribution function, the conditional quantile function in quantile regression, functional coefficients in varying coefficient models and semiparametric inferences. Based on our empirical studies, nonparametric inference based on the proposed sequential design is more efficient than the uniform design and its performance is close to the true but unknown optimal design. The Canadian Journal of Statistics 40: 362–377; 2012 © 2012 Statistical Society of Canada  相似文献   

15.
This article addresses issues in creating public-use data files in the presence of missing ordinal responses and subsequent statistical analyses of the dataset by users. The authors propose a fully efficient fractional imputation (FI) procedure for ordinal responses with missing observations. The proposed imputation strategy retrieves the missing values through the full conditional distribution of the response given the covariates and results in a single imputed data file that can be analyzed by different data users with different scientific objectives. Two most critical aspects of statistical analyses based on the imputed data set,  validity  and  efficiency, are examined through regression analysis involving the ordinal response and a selected set of covariates. It is shown through both theoretical development and simulation studies that, when the ordinal responses are missing at random, the proposed FI procedure leads to valid and highly efficient inferences as compared to existing methods. Variance estimation using the fractionally imputed data set is also discussed. The Canadian Journal of Statistics 48: 138–151; 2020 © 2019 Statistical Society of Canada  相似文献   

16.
Priors are introduced into goodness‐of‐fit tests, both for unknown parameters in the tested distribution and on the alternative density. Neyman–Pearson theory leads to the test with the highest expected power. To make the test practical, we seek priors that make it likely a priori that the power will be larger than the level of the test but not too close to one. As a result, priors are sample size dependent. We explore this procedure in particular for priors that are defined via a Gaussian process approximation for the logarithm of the alternative density. In the case of testing for the uniform distribution, we show that the optimal test is of the U‐statistic type and establish limiting distributions for the optimal test statistic, both under the null hypothesis and averaged over the alternative hypotheses. The optimal test statistic is shown to be of the Cramér–von Mises type for specific choices of the Gaussian process involved. The methodology when parameters in the tested distribution are unknown is discussed and illustrated in the case of testing for the von Mises distribution. The Canadian Journal of Statistics 47: 560–579; 2019 © 2019 Statistical Society of Canada  相似文献   

17.
We focus on the nonparametric regression of a scalar response on a functional explanatory variable. As an alternative to the well-known Nadaraya-Watson estimator for regression function in this framework, the locally modelled regression estimator performs very well [cf. [Barrientos-Marin, J., Ferraty, F., and Vieu, P. (2010), ‘Locally Modelled Regression and Functional Data’, Journal of Nonparametric Statistics, 22, 617–632]. In this paper, the asymptotic properties of locally modelled regression estimator for functional data are considered. The mean-squared convergence as well as asymptotic normality for the estimator are established. We also adapt the empirical likelihood method to construct the point-wise confidence intervals for the regression function and derive the Wilk's phenomenon for the empirical likelihood inference. Furthermore, a simulation study is presented to illustrate our theoretical results.  相似文献   

18.
Summary.  Deconvolution problems are naturally represented in the Fourier domain, whereas thresholding in wavelet bases is known to have broad adaptivity properties. We study a method which combines both fast Fourier and fast wavelet transforms and can recover a blurred function observed in white noise with O { n    log ( n )2} steps. In the periodic setting, the method applies to most deconvolution problems, including certain 'boxcar' kernels, which are important as a model of motion blur, but having poor Fourier characteristics. Asymptotic theory informs the choice of tuning parameters and yields adaptivity properties for the method over a wide class of measures of error and classes of function. The method is tested on simulated light detection and ranging data suggested by underwater remote sensing. Both visual and numerical results show an improvement over competing approaches. Finally, the theory behind our estimation paradigm gives a complete characterization of the 'maxiset' of the method: the set of functions where the method attains a near optimal rate of convergence for a variety of L p loss functions.  相似文献   

19.
We study the benefit of exploiting the gene–environment independence (GEI) assumption for inferring the joint effect of genotype and environmental exposure on disease risk in a case–control study. By transforming the problem into a constrained maximum likelihood estimation problem we derive the asymptotic distribution of the maximum likelihood estimator (MLE) under the GEI assumption (MLE‐GEI) in a closed form. Our approach uncovers a transparent explanation of the efficiency gained by exploiting the GEI assumption in more general settings, thus bridging an important gap in the existing literature. Moreover, we propose an easy‐to‐implement numerical algorithm for estimating the model parameters in practice. Finally, we conduct simulation studies to compare the proposed method with the traditional prospective logistic regression method and the case‐only estimator. The Canadian Journal of Statistics 47: 473–486; 2019 © 2019 Statistical Society of Canada  相似文献   

20.
The two-parameter generalized exponential (GE) distribution was introduced by Gupta and Kundu [Gupta, R.D. and Kundu, D., 1999, Generalized exponential distribution. Australian and New Zealand Journal of Statistics, 41(2), 173–188.]. It was observed that the GE can be used in situations where a skewed distribution for a nonnegative random variable is needed. In this article, the Bayesian estimation and prediction for the GE distribution, using informative priors, have been considered. Importance sampling is used to estimate the parameters, as well as the reliability function, and the Gibbs and Metropolis samplers data sets are used to predict the behavior of further observations from the distribution. Two data sets are used to illustrate the Bayesian procedure.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号