首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 21 毫秒
1.
Variable selection problem is one of the most important tasks in regression analysis, especially in a high-dimensional setting. In this paper, we study this problem in the context of scalar response functional regression model, which is a linear model with scalar response and functional regressors. The functional model can be represented by certain multiple linear regression model via basis expansions of functional variables. Based on this model and random subspace method of Mielniczuk and Teisseyre (Comput Stat Data Anal 71:725–742, 2014), two simple variable selection procedures for scalar response functional regression model are proposed. The final functional model is selected by using generalized information criteria. Monte Carlo simulation studies conducted and a real data example show very satisfactory performance of new variable selection methods under finite samples. Moreover, they suggest that considered procedures outperform solutions found in the literature in terms of correctly selected model, false discovery rate control and prediction error.  相似文献   

2.
Classification of gene expression microarray data is important in the diagnosis of diseases such as cancer, but often the analysis of microarray data presents difficult challenges because the gene expression dimension is typically much larger than the sample size. Consequently, classification methods for microarray data often rely on regularization techniques to stabilize the classifier for improved classification performance. In particular, numerous regularization techniques, such as covariance-matrix regularization, are available, which, in practice, lead to a difficult choice of regularization methods. In this paper, we compare the classification performance of five covariance-matrix regularization methods applied to the linear discriminant function using two simulated high-dimensional data sets and five well-known, high-dimensional microarray data sets. In our simulation study, we found the minimum distance empirical Bayes method reported in Srivastava and Kubokawa [Comparison of discrimination methods for high dimensional data, J. Japan Statist. Soc. 37(1) (2007), pp. 123–134], and the new linear discriminant analysis reported in Thomaz, Kitani, and Gillies [A Maximum Uncertainty LDA-based approach for Limited Sample Size problems – with application to Face Recognition, J. Braz. Comput. Soc. 12(1) (2006), pp. 1–12], to perform consistently well and often outperform three other prominent regularization methods. Finally, we conclude with some recommendations for practitioners.  相似文献   

3.
ABSTRACT

In this paper, we first consider the entropy estimators introduced by Vasicek [A test for normality based on sample entropy. J R Statist Soc, Ser B. 1976;38:54–59], Ebrahimi et al. [Two measures of sample entropy. Stat Probab Lett. 1994;20:225–234], Yousefzadeh and Arghami [Testing exponentiality based on type II censored data and a new cdf estimator. Commun Stat – Simul Comput. 2008;37:1479–1499], Alizadeh Noughabi and Arghami [A new estimator of entropy. J Iran Statist Soc. 2010;9:53–64], and Zamanzade and Arghami [Goodness-of-fit test based on correcting moments of modified entropy estimator. J Statist Comput Simul. 2011;81:2077–2093], and the nonparametric distribution functions corresponding to them. We next introduce goodness-of-fit test statistics for the Laplace distribution based on the moments of nonparametric distribution functions of the aforementioned estimators. We obtain power estimates of the proposed test statistics with Monte Carlo simulation and compare them with the competing test statistics against various alternatives. Performance of the proposed new test statistics is illustrated in real cases.  相似文献   

4.
Jae Keun Yoo 《Statistics》2018,52(2):409-425
In this paper, a model-based approach to reduce the dimension of response variables in multivariate regression is newly proposed, following the existing context of the response dimension reduction developed by Yoo and Cook [Response dimension reduction for the conditional mean in multivariate regression. Comput Statist Data Anal. 2008;53:334–343]. The related dimension reduction subspace is estimated by maximum likelihood, assuming an additive error. In the new approach, the linearity condition, which is assumed for the methodological development in Yoo and Cook (2008), is understood through the covariance matrix of the random error. Numerical studies show potential advantages of the proposed approach over Yoo and Cook (2008). A real data example is presented for illustration.  相似文献   

5.
We deal with a general class of extreme-value regression models introduced by Barreto-Souza and Vasconcellos [Bias and skewness in a general extreme-value regression model, Comput. Statist. Data Anal. 55 (2011), pp. 1379–1393]. Our goal is to derive an adjusted likelihood ratio statistic that is approximately distributed as χ2 with a high degree of accuracy. Although the adjusted statistic requires more computational effort than its unadjusted counterpart, it is shown that the adjustment term has a simple compact form that can be easily implemented in standard statistical software. Further, we compare the finite-sample performance of the three classical tests (likelihood ratio, Wald, and score), the gradient test that has been recently proposed by Terrell [The gradient statistic, Comput. Sci. Stat. 34 (2002), pp. 206–215], and the adjusted likelihood ratio test obtained in this article. Our simulations favour the latter. Applications of our results are presented.  相似文献   

6.
We propose a new distribution, the so-called beta-Weibull geometric distribution, whose failure rate function can be decreasing, increasing or an upside-down bathtub. This distribution contains special sub-models the exponential geometric [K. Adamidis and S. Loukas, A lifetime distribution with decreasing failure rate, Statist. Probab. Lett. 39 (1998), pp. 35–42], beta exponential [S. Nadarajah and S. Kotz, The exponentiated type distributions, Acta Appl. Math. 92 (2006), pp. 97–111; The beta exponential distribution, Reliab. Eng. Syst. Saf. 91 (2006), pp. 689–697], Weibull geometric [W. Barreto-Souza, A.L. de Morais, and G.M. Cordeiro, The Weibull-geometric distribution, J. Stat. Comput. Simul. 81 (2011), pp. 645–657], generalized exponential geometric [R.B. Silva, W. Barreto-Souza, and G.M. Cordeiro, A new distribution with decreasing, increasing and upside-down bathtub failure rate, Comput. Statist. Data Anal. 54 (2010), pp. 935–944; G.O. Silva, E.M.M. Ortega, and G.M. Cordeiro, The beta modified Weibull distribution, Lifetime Data Anal. 16 (2010), pp. 409–430] and beta Weibull [S. Nadarajah, G.M. Cordeiro, and E.M.M. Ortega, General results for the Kumaraswamy-G distribution, J. Stat. Comput. Simul. (2011). DOI: 10.1080/00949655.2011.562504] distributions, among others. The density function can be expressed as a mixture of Weibull density functions. We derive expansions for the moments, generating function, mean deviations and Rénvy entropy. The parameters of the proposed model are estimated by maximum likelihood. The model fitting using envelops was conducted. The proposed distribution gives a good fit to the ozone level data in New York.  相似文献   

7.
We point out and comment on the confusions, deficiencies and errors of Wang [Life prediction under random censorship, J. Stat. Comput. Simul. 78 (2008), pp. 1033–1044].  相似文献   

8.
We consider the first-order Poisson autoregressive model proposed by McKenzie [Some simple models for discrete variate time series. Water Resour Bull. 1985;21:645–650] and Al-Osh and Alzaid [First-order integer valued autoregressive (INAR(1)) process. J Time Ser Anal. 1987;8:261–275], which may be suitable in situations where the time series data are non-negative and integer valued. We derive the second-order bias of the squared difference estimator [Weiß. Process capability analysis for serially dependent processes of Poisson counts. J Stat Comput Simul. 2012;82:383–404] for one of the parameters and show that this bias can be used to define a bias-reduced estimator. The behaviour of a modified conditional least-squares estimator is also studied. Furthermore, we access the asymptotic properties of the estimators here discussed. We present numerical evidence, based upon Monte Carlo simulation studies, showing that the here proposed bias-adjusted estimator outperforms the other estimators in small samples. We also present an application to a real data set.  相似文献   

9.
Quantile regression methods have been used to estimate upper and lower quantile reference curves as the function of several covariates. In this article, it is demonstrated that the estimating equation of Zhou [A weighted quantile regression for randomly truncated data, Comput. Stat. Data Anal. 55 (2011), pp. 554–566.] can be extended to analyse left-truncated and right-censored data. We evaluate the finite sample performance of the proposed estimators through simulation studies. The proposed estimator β?(q) is applied to the Veteran's Administration lung cancer data reported by Prentice [Exponential survival with censoring and explanatory variables, Biometrika 60 (1973), pp. 279–288].  相似文献   

10.
In this paper, based on a jointly type-II censored sample from two exponential populations, the Bayesian inference for the two unknown parameters are developed with the use of squared-error, linear-exponential and general entropy loss functions. The problem of predicting the future failure times, both point and interval prediction, based on the observed joint type-II censored data, is also addressed from a Bayesian viewpoint. A Monte Carlo simulation study is conducted to compare the Bayesian estimators with the maximum likelihood estimator developed by Balakrishnan and Rasouli [Exact likelihood inference for two exponential populations under joint type-II censoring. Comput Stat Data Anal. 2008;52:2725–2738]. Finally, a numerical example is utilized for the purpose of illustration.  相似文献   

11.
The Rayleigh distribution has been used to model right skewed data. Rayleigh [On the resultant of a large number of vibrations of the some pitch and of arbitrary phase. Philos Mag. 1880;10:73–78] derived it from the amplitude of sound resulting from many important sources. In this paper, a new goodness-of-fit test for the Rayleigh distribution is proposed. This test is based on the empirical likelihood ratio methodology proposed by Vexler and Gurevich [Empirical likelihood ratios applied to goodness-of-fit tests based on sample entropy. Comput Stat Data Anal. 2010;54:531–545]. Consistency of the proposed test is derived. It is shown that the distribution of the proposed test does not depend on scale parameter. Critical values of the test statistic are computed, through a simulation study. A Monte Carlo study for the power of the proposed test is carried out under various alternatives. The performance of the test is compared with some well-known competing tests. Finally, an illustrative example is presented and analysed.  相似文献   

12.
Jin Zhang 《Statistics》2018,52(4):874-884
In this article, we establish the minimum-volume confidence sets for normal linear regression models, extending the results in Zhang [Minimum volume confidence sets for parameters of normal distributions. Adv Stat Anal. 2017;101:309–320] on building the minimum-volume confidence sets for parameters of normal distributions. Compared with classical confidence sets, the proposed optimal confidence set is proved to have the smallest volume, for whatever confidence level, sample size and sample data.  相似文献   

13.
We review a few unusual aspects of Bootstrap and some of the recent theoretical as well as methodological advances. We discuss the handling of non-linearity by Bootstrap through a numerical example in Section  2. Application to the estimation of high-dimensional inverse covariance matrix is presented in Section  3 with emphasis on the Augmented Bootstrap and a Bayesian version of it. Another high dimensional example, namely, Random Forest and its offshoot random survival forest (Ishwaran et al. (2008)  [32]) are discussed in Section  4. Bootstrap for massive data, introduced by Kleiner et al. (2011) [35], is discussed in Section  4. In Section  5, we discuss some aspects of Bootstrap in the context of hypothesis testing in high-dimension.  相似文献   

14.
In this paper, we investigate some stochastic comparisons in terms of likelihood ratio ordering between spacings from independent random variables exponentially distributed with different scale parameters. We partially solve some open problems in [Wen S, Lu Q, Hu T. Likelihood ratio orderings of spacings of heterogeneous exponential random variables. J Multivariate Anal. 2007;98:743–756] for a one-sample problem and in [Hu T, Lu Q, Wen S. Stochastic comparisons and dependence of spacings from two samples of exponential random variables. Commun Stat – Theory Methods 2006;35:979–988] for a two-sample problem. Specifically, we prove that the second spacing is always smaller than the third spacing in terms of the likelihood ratio order and we provide the ordering among all spacings in the case n=4. In the two-sample case, we establish comparisons between the second spacings related to each sample under certain conditions.  相似文献   

15.
The hybrid censoring scheme, which is a mixture of Type-I and Type-II censoring schemes, has been extended to the case of progressive censoring schemes by Kundu and Joarder [Analysis of Type-II progressively hybrid censored data, Comput. Stat. Data Anal. 50 (2006), pp. 2509–2528] and Childs et al. [Exact likelihood inference for an exponential parameter under progressive hybrid censoring schemes, in Statistical Models and Methods for Biomedical and Technical Systems, F. Vonta, M. Nikulin, N. Limnios, and C. Huber-Carol, eds., Birkhäuser, Boston, MA, 2007, pp. 323–334]. In this paper, we derive a simple expression for the Fisher information contained in Type-I and Type-II progressively hybrid censored data. An illustrative example is provided applicable to a scaled-exponential distribution to demonstrate our methodologies.  相似文献   

16.
Kadilar and Cingi [Ratio estimators in simple random sampling, Appl. Math. Comput. 151 (3) (2004), pp. 893–902] introduced some ratio-type estimators of finite population mean under simple random sampling. Recently, Kadilar and Cingi [New ratio estimators using correlation coefficient, Interstat 4 (2006), pp. 1–11] have suggested another form of ratio-type estimators by modifying the estimator developed by Singh and Tailor [Use of known correlation coefficient in estimating the finite population mean, Stat. Transit. 6 (2003), pp. 655–560]. Kadilar and Cingi [Improvement in estimating the population mean in simple random sampling, Appl. Math. Lett. 19 (1) (2006), pp. 75–79] have suggested yet another class of ratio-type estimators by taking a weighted average of the two known classes of estimators referenced above. In this article, we propose an alternative form of ratio-type estimators which are better than the competing ratio, regression, and other ratio-type estimators considered here. The results are also supported by the analysis of three real data sets that were considered by Kadilar and Cingi.  相似文献   

17.
Grubbs’s model (Grubbs, Encycl Stat Sci 3:42–549, 1983) is used for comparing several measuring devices, and it is common to assume that the random terms have a normal (or symmetric) distribution. In this paper, we discuss the extension of this model to the class of scale mixtures of skew-normal distributions. Our results provide a useful generalization of the symmetric Grubbs’s model (Osorio et al., Comput Stat Data Anal, 53:1249–1263, 2009) and the asymmetric skew-normal model (Montenegro et al., Stat Pap 51:701–715, 2010). We discuss the EM algorithm for parameter estimation and the local influence method (Cook, J Royal Stat Soc Ser B, 48:133–169, 1986) for assessing the robustness of these parameter estimates under some usual perturbation schemes. The results and methods developed in this paper are illustrated with a numerical example.  相似文献   

18.
In this paper, we develop a new forecasting algorithm for value-at-risk (VaR) based on ARMA–GARCH (autoregressive moving average–generalized autoregressive conditional heteroskedastic) models whose innovations follow a Gaussian mixture distribution. For the parameter estimation, we employ the conditional least squares and quasi-maximum-likelihood estimator (QMLE) for ARMA and GARCH parameters, respectively. In particular, Gaussian mixture parameters are estimated based on the residuals obtained from the QMLE of GARCH parameters. Our algorithm provides a handy methodology, spending much less time in calculation than the existing resampling and bias-correction method developed in Hartz et al. [Accurate value-at-risk forecasting based on the normal-GARCH model, Comput. Stat. Data Anal. 50 (2006), pp. 3032–3052]. Through a simulation study and a real-data analysis, it is shown that our method provides an accurate VaR prediction.  相似文献   

19.
A new five-parameter distribution called the beta Weibull-geometric (BWG) distribution is proposed. The new distribution is generated from the logit of a beta random variable and includes the Weibull-geometric distribution of Barreto-Souza et al. [The Weibull-geometric distribution, J. Stat. Comput. Simul. 81 (2011), pp. 645–657], beta Weibull (BW), beta exponential, exponentiated Weibull, and some other lifetime distributions as special cases. A comprehensive mathematical treatment of this distribution is provided. The density function can be expressed as an infinite mixture of BW densities and then we derive some mathematical properties of the new distribution from the corresponding properties of the BW distribution. The density function of the order statistics and also estimation of the stress–strength parameter are obtained using two general expressions. To estimate the model parameters, we use the maximum likelihood method and the asymptotic distribution of the estimators is also discussed. The capacity of the new distribution are examined by various tools, using two real data sets.  相似文献   

20.
Lin et al. [Exact Bayesian variable sampling plans for the exponential distribution with progressive hybrid censoring, J. Stat. Comput. Simul. 81 (2011), pp. 873–882] claimed to have derived exact Bayesian sampling plans for exponential distributions with progressive hybrid censoring. We comment on the accuracy of the design parameters of their proposed sampling plans and the associated Bayes risks given in tables of Lin et al. [Exact Bayesian variable sampling plans for the exponential distribution with progressive hybrid censoring, J. Stat. Comput. Simul. 81 (2011), pp. 873–882]. Counter-examples to their claim are provided.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号