首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 187 毫秒
1.
The multivariate skew-t distribution (J Multivar Anal 79:93–113, 2001; J R Stat Soc, Ser B 65:367–389, 2003; Statistics 37:359–363, 2003) includes the Student t, skew-Cauchy and Cauchy distributions as special cases and the normal and skew–normal ones as limiting cases. In this paper, we explore the use of Markov Chain Monte Carlo (MCMC) methods to develop a Bayesian analysis of repeated measures, pretest/post-test data, under multivariate null intercept measurement error model (J Biopharm Stat 13(4):763–771, 2003) where the random errors and the unobserved value of the covariate (latent variable) follows a Student t and skew-t distribution, respectively. The results and methods are numerically illustrated with an example in the field of dentistry.  相似文献   

2.
Lin et al. [Exact Bayesian variable sampling plans for the exponential distribution with progressive hybrid censoring, J. Stat. Comput. Simul. 81 (2011), pp. 873–882] claimed to have derived exact Bayesian sampling plans for exponential distributions with progressive hybrid censoring. We comment on the accuracy of the design parameters of their proposed sampling plans and the associated Bayes risks given in tables of Lin et al. [Exact Bayesian variable sampling plans for the exponential distribution with progressive hybrid censoring, J. Stat. Comput. Simul. 81 (2011), pp. 873–882]. Counter-examples to their claim are provided.  相似文献   

3.
In the first part of the paper, we introduce the matrix-variate generalized hyperbolic distribution by mixing the matrix normal distribution with the matrix generalized inverse Gaussian density. The p-dimensional generalized hyperbolic distribution of [Barndorff-Nielsen, O. (1978). Hyperbolic distributions and distributions on hyperbolae. Scand. J. Stat., 5, 151–157], the matrix-T distribution and many well-known distributions are shown to be special cases of the new distribution. Some properties of the distribution are also studied. The second part of the paper deals with the application of the distribution in the Bayesian analysis of the normal multivariate linear model.  相似文献   

4.
Sinh-normal/independent distributions are a class of symmetric heavy-tailed distributions that include the sinh-normal distribution as a special case, which has been used extensively in Birnbaum–Saunders regression models. Here, we explore the use of Markov Chain Monte Carlo methods to develop a Bayesian analysis in nonlinear regression models when Sinh-normal/independent distributions are assumed for the random errors term, and it provides a robust alternative to the sinh-normal nonlinear regression model. Bayesian mechanisms for parameter estimation, residual analysis and influence diagnostics are then developed, which extend the results of Farias and Lemonte [Bayesian inference for the Birnbaum-Saunders nonlinear regression model, Stat. Methods Appl. 20 (2011), pp. 423-438] who used the Sinh-normal/independent distributions with known scale parameter. Some special cases, based on the sinh-Student-t (sinh-St), sinh-slash (sinh-SL) and sinh-contaminated normal (sinh-CN) distributions are discussed in detail. Two real datasets are finally analyzed to illustrate the developed procedures.  相似文献   

5.
Composite quantile regression models have been shown to be effective techniques in improving the prediction accuracy [H. Zou and M. Yuan, Composite quantile regression and the oracle model selection theory, Ann. Statist. 36 (2008), pp. 1108–1126; J. Bradic, J. Fan, and W. Wang, Penalized composite quasi-likelihood for ultrahighdimensional variable selection, J. R. Stat. Soc. Ser. B 73 (2011), pp. 325–349; Z. Zhao and Z. Xiao, Efficient regressions via optimally combining quantile information, Econometric Theory 30(06) (2014), pp. 1272–1314]. This paper studies composite Tobit quantile regression (TQReg) from a Bayesian perspective. A simple and efficient MCMC-based computation method is derived for posterior inference using a mixture of an exponential and a scaled normal distribution of the skewed Laplace distribution. The approach is illustrated via simulation studies and a real data set. Results show that combine information across different quantiles can provide a useful method in efficient statistical estimation. This is the first work to discuss composite TQReg from a Bayesian perspective.  相似文献   

6.
V. Nekoukhou  H. Bidram 《Statistics》2013,47(4):876-887
In this paper, we shall attempt to introduce another discrete analogue of the generalized exponential distribution of Gupta and Kundu [Generalized exponential distributions, Aust. N. Z. J. Stat. 41(2) (1999), pp. 173–188], different to that of Nekoukhou et al. [A discrete analogue of the generalized exponential distribution, Comm. Stat. Theory Methods, to appear (2011)]. This new discrete distribution, which we shall call a discrete generalized exponential distribution of the second type (DGE2(α, p)), can be viewed as another generalization of the geometric distribution. We shall first study some basic distributional and moment properties, as well as order statistics distributions of this family of new distributions. Certain compounded DGE2(α, p) distributions are also discussed as the results of which some previous lifetime distributions such as that of Adamidis and Loukas [A lifetime distribution with decreasing failure rate, Statist. Probab. Lett. 39 (1998), pp. 35–42] follow as corollaries. Then, we will investigate estimation of the parameters involved. Finally, we will examine the model with a real data set.  相似文献   

7.
Tests for the equality of variances are of interest in many areas such as quality control, agricultural production systems, experimental education, pharmacology, biology, as well as a preliminary to the analysis of variance, dose–response modelling or discriminant analysis. The literature is vast. Traditional non-parametric tests are due to Mood, Miller and Ansari–Bradley. A test which usually stands out in terms of power and robustness against non-normality is the W50 Brown and Forsythe [Robust tests for the equality of variances, J. Am. Stat. Assoc. 69 (1974), pp. 364–367] modification of the Levene test [Robust tests for equality of variances, in Contributions to Probability and Statistics, I. Olkin, ed., Stanford University Press, Stanford, 1960, pp. 278–292]. This paper deals with the two-sample scale problem and in particular with Levene type tests. We consider 10 Levene type tests: the W50, the M50 and L50 tests [G. Pan, On a Levene type test for equality of two variances, J. Stat. Comput. Simul. 63 (1999), pp. 59–71], the R-test [R.G. O'Brien, A general ANOVA method for robust tests of additive models for variances, J. Am. Stat. Assoc. 74 (1979), pp. 877–880], as well as the bootstrap and permutation versions of the W50, L50 and R tests. We consider also the F-test, the modified Fligner and Killeen [Distribution-free two-sample tests for scale, J. Am. Stat. Assoc. 71 (1976), pp. 210–213] test, an adaptive test due to Hall and Padmanabhan [Adaptive inference for the two-sample scale problem, Technometrics 23 (1997), pp. 351–361] and the two tests due to Shoemaker [Tests for differences in dispersion based on quantiles, Am. Stat. 49(2) (1995), pp. 179–182; Interquantile tests for dispersion in skewed distributions, Commun. Stat. Simul. Comput. 28 (1999), pp. 189–205]. The aim is to identify the effective methods for detecting scale differences. Our study is different with respect to the other ones since it is focused on resampling versions of the Levene type tests, and many tests considered here have not ever been proposed and/or compared. The computationally simplest test found robust is W50. Higher power, while preserving robustness, is achieved by considering the resampling version of Levene type tests like the permutation R-test (recommended for normal- and light-tailed distributions) and the bootstrap L50 test (recommended for heavy-tailed and skewed distributions). Among non-Levene type tests, the best one is the adaptive test due to Hall and Padmanabhan.  相似文献   

8.
Cross-validation has been widely used in the context of statistical linear models and multivariate data analysis. Recently, technological advancements give possibility of collecting new types of data that are in the form of curves. Statistical procedures for analysing these data, which are of infinite dimension, have been provided by functional data analysis. In functional linear regression, using statistical smoothing, estimation of slope and intercept parameters is generally based on functional principal components analysis (FPCA), that allows for finite-dimensional analysis of the problem. The estimators of the slope and intercept parameters in this context, proposed by Hall and Hosseini-Nasab [On properties of functional principal components analysis, J. R. Stat. Soc. Ser. B: Stat. Methodol. 68 (2006), pp. 109–126], are based on FPCA, and depend on a smoothing parameter that can be chosen by cross-validation. The cross-validation criterion, given there, is time-consuming and hard to compute. In this work, we approximate this cross-validation criterion by such another criterion so that we can turn to a multivariate data analysis tool in some sense. Then, we evaluate its performance numerically. We also treat a real dataset, consisting of two variables; temperature and the amount of precipitation, and estimate the regression coefficients for the former variable in a model predicting the latter one.  相似文献   

9.
This paper demonstrates that cross-validation (CV) and Bayesian adaptive bandwidth selection can be applied in the estimation of associated kernel discrete functions. This idea is originally proposed by Brewer [A Bayesian model for local smoothing in kernel density estimation, Stat. Comput. 10 (2000), pp. 299–309] to derive variable bandwidths in adaptive kernel density estimation. Our approach considers the adaptive binomial kernel estimator and treats the variable bandwidths as parameters with beta prior distribution. The best variable bandwidth selector is estimated by the posterior mean in the Bayesian sense under squared error loss. Monte Carlo simulations are conducted to examine the performance of the proposed Bayesian adaptive approach in comparison with the performance of the Asymptotic mean integrated squared error estimator and CV technique for selecting a global (fixed) bandwidth proposed in Kokonendji and Senga Kiessé [Discrete associated kernels method and extensions, Stat. Methodol. 8 (2011), pp. 497–516]. The Bayesian adaptive bandwidth estimator performs better than the global bandwidth, in particular for small and moderate sample sizes.  相似文献   

10.
Partially linear models (PLMs) are an important tool in modelling economic and biometric data and are considered as a flexible generalization of the linear model by including a nonparametric component of some covariate into the linear predictor. Usually, the error component is assumed to follow a normal distribution. However, the theory and application (through simulation or experimentation) often generate a great amount of data sets that are skewed. The objective of this paper is to extend the PLMs allowing the errors to follow a skew-normal distribution [A. Azzalini, A class of distributions which includes the normal ones, Scand. J. Statist. 12 (1985), pp. 171–178], increasing the flexibility of the model. In particular, we develop the expectation-maximization (EM) algorithm for linear regression models and diagnostic analysis via local influence as well as generalized leverage, following [H. Zhu and S. Lee, Local influence for incomplete-data models, J. R. Stat. Soc. Ser. B 63 (2001), pp. 111–126]. A simulation study is also conducted to evaluate the efficiency of the EM algorithm. Finally, a suitable transformation is applied in a data set on ragweed pollen concentration in order to fit PLMs under asymmetric distributions. An illustrative comparison is performed between normal and skew-normal errors.  相似文献   

11.
We propose a new distribution, the so-called beta-Weibull geometric distribution, whose failure rate function can be decreasing, increasing or an upside-down bathtub. This distribution contains special sub-models the exponential geometric [K. Adamidis and S. Loukas, A lifetime distribution with decreasing failure rate, Statist. Probab. Lett. 39 (1998), pp. 35–42], beta exponential [S. Nadarajah and S. Kotz, The exponentiated type distributions, Acta Appl. Math. 92 (2006), pp. 97–111; The beta exponential distribution, Reliab. Eng. Syst. Saf. 91 (2006), pp. 689–697], Weibull geometric [W. Barreto-Souza, A.L. de Morais, and G.M. Cordeiro, The Weibull-geometric distribution, J. Stat. Comput. Simul. 81 (2011), pp. 645–657], generalized exponential geometric [R.B. Silva, W. Barreto-Souza, and G.M. Cordeiro, A new distribution with decreasing, increasing and upside-down bathtub failure rate, Comput. Statist. Data Anal. 54 (2010), pp. 935–944; G.O. Silva, E.M.M. Ortega, and G.M. Cordeiro, The beta modified Weibull distribution, Lifetime Data Anal. 16 (2010), pp. 409–430] and beta Weibull [S. Nadarajah, G.M. Cordeiro, and E.M.M. Ortega, General results for the Kumaraswamy-G distribution, J. Stat. Comput. Simul. (2011). DOI: 10.1080/00949655.2011.562504] distributions, among others. The density function can be expressed as a mixture of Weibull density functions. We derive expansions for the moments, generating function, mean deviations and Rénvy entropy. The parameters of the proposed model are estimated by maximum likelihood. The model fitting using envelops was conducted. The proposed distribution gives a good fit to the ozone level data in New York.  相似文献   

12.
This paper proposes various double unit root tests for cross-sectionally dependent panel data. The cross-sectional correlation is handled by the projection method [P.C.B. Phillips and D. Sul, Dynamic panel estimation and homogeneity testing under cross section dependence, Econom. J. 6 (2003), pp. 217–259; H.R. Moon and B. Perron, Testing for a unit root in panels with dynamic factors, J. Econom. 122 (2004), pp. 81–126] or the subtraction method [J. Bai and S. Ng, A PANIC attack on unit roots and cointegration, Econometrica 72 (2004), pp. 1127–1177]. Pooling or averaging is applied to combine results from different panel units. Also, to estimate autoregressive parameters the ordinary least squares estimation [D.P. Hasza and W.A. Fuller, Estimation for autoregressive processes with unit roots, Ann. Stat. 7 (1979), pp. 1106–1120] or the symmetric estimation [D.L. Sen and D.A. Dickey, Symmetric test for second differencing in univariate time series, J. Bus. Econ. Stat. 5 (1987), pp. 463–473] are used, and to adjust mean functions the ordinary mean adjustment or the recursive mean adjustment are used. Combinations of different methods in defactoring to eliminate the cross-sectional dependency, integrating results from panel units, estimating the parameters, and adjusting mean functions yields various available tests for double unit roots in panel data. Simple asymptotic distributions of the proposed test statistics are derived, which can be used to find critical values of the test statistics.

We perform a Monte Carlo experiment to compare the performance of these tests and to suggest optimal tests for a given panel data. Application of the proposed tests to a real data, the yearly export panel data sets of several Latin–American countries for the past 50 years, illustrates the usefulness of the proposed tests for panel data, in that they reveal stronger evidence of double unit roots than the componentwise double unit root tests of Hasza and Fuller [Estimation for autoregressive processes with unit roots, Ann. Stat. 7 (1979), pp. 1106–1120] or Sen and Dickey [Symmetric test for second differencing in univariate time series, J. Bus. Econ. Stat. 5 (1987), pp. 463–473].  相似文献   


13.
Dealing with incomplete data is a pervasive problem in statistical surveys. Bayesian networks have been recently used in missing data imputation. In this research, we propose a new methodology for the multivariate imputation of missing data using discrete Bayesian networks and conditional Gaussian Bayesian networks. Results from imputing missing values in coronary artery disease data set and milk composition data set as well as a simulation study from cancer-neapolitan network are presented to demonstrate and compare the performance of three Bayesian network-based imputation methods with those of multivariate imputation by chained equations (MICE) and the classical hot-deck imputation method. To assess the effect of the structure learning algorithm on the performance of the Bayesian network-based methods, two methods called Peter-Clark algorithm and greedy search-and-score have been applied. Bayesian network-based methods are: first, the method introduced by Di Zio et al. [Bayesian networks for imputation, J. R. Stat. Soc. Ser. A 167 (2004), 309–322] in which, each missing item of a variable is imputed using the information given in the parents of that variable; second, the method of Di Zio et al. [Multivariate techniques for imputation based on Bayesian networks, Neural Netw. World 15 (2005), 303–310] which uses the information in the Markov blanket set of the variable to be imputed and finally, our new proposed method which applies the whole available knowledge of all variables of interest, consisting the Markov blanket and so the parent set, to impute a missing item. Results indicate the high quality of our new proposed method especially in the presence of high missingness percentages and more connected networks. Also the new method have shown to be more efficient than the MICE method for small sample sizes with high missing rates.  相似文献   

14.
In this article, the restricted rk class estimator and restricted rd class estimator are introduced, which are general estimators of the rk class estimator by Baye and Parker [Combining ridge and principal component regression: A money demand illustration, Commun. Stat. Theory Methods 13(2) (1984), pp. 197–205] and the rd class estimator by Kaç?ranlar and Sakall?o?lu [Combining the Liu estimator and the principal component regression estimator, Commun. Stat. Theory Methods 30(12) (2001), pp. 2699–2705], respectively. For the two cases when the restrictions are true and not true, the superiority of the restricted rk class estimator and rd class estimator over the restricted ridge regression estimator by Sarkar [A new estimator combining the ridge regression and the restricted least squares methods of estimation, Commun. Stat. Theory Methods 21 (1992), pp. 1987–2000] and the restricted Liu estimator by Kaç?ranlar et al. [A new biased estimator in linear regression and a detailed analysis of the widely analysed dataset on Portland cement, Sankhya - Indian J. Stat. 61B(3) (1999), pp. 443–459] are discussed with respect to the mean squared error matrix criterion. Furthermore, a Monte Carlo evaluation of the estimators is given to illustrate some of the theoretical results.  相似文献   

15.
Data envelopment analysis (DEA) is the most commonly used approach for evaluating healthcare efficiency [B. Hollingsworth, The measurement of efficiency and productivity of health care delivery. Health Economics 17(10) (2008), pp. 1107–1128], but a long-standing concern is that DEA assumes that data are measured without error. This is quite unlikely, and DEA and other efficiency analysis techniques may yield biased efficiency estimates if it is not realized [B.J. Gajewski, R. Lee, M. Bott, U. Piamjariyakul, and R.L. Taunton, On estimating the distribution of data envelopment analysis efficiency scores: an application to nursing homes’ care planning process. Journal of Applied Statistics 36(9) (2009), pp. 933–944; J. Ruggiero, Data envelopment analysis with stochastic data. Journal of the Operational Research Society 55 (2004), pp. 1008–1012]. We propose to address measurement error systematically using a Bayesian method (Bayesian DEA). We will apply Bayesian DEA to data from the National Database of Nursing Quality Indicators® to estimate nursing units’ efficiency. Several external reliability studies inform the posterior distribution of the measurement error on the DEA variables. We will discuss the case of generalizing the approach to situations where an external reliability study is not feasible.  相似文献   

16.
We derive a generalization of the exponential distribution by making log transformation of the standard two-sided power distribution. We show that this new generalization is in fact a mixture of a truncated exponential distribution and truncated generalized exponential distribution introduced by Gupta and Kundu [Generalized exponential distributions. Aust. N. Z. J. Stat. 41(1999):173–188]. The newly defined distribution is more flexible for modeling data than the ordinary exponential distribution. We study its properties, estimate the parameters, and demonstrate it on some well-known real data sets comparing other existing methods.  相似文献   

17.
In this paper, we first propose a new estimator of entropy for continuous random variables. Our estimator is obtained by correcting the coefficients of Vasicek's [A test for normality based on sample entropy, J. R. Statist. Soc. Ser. B 38 (1976), pp. 54–59] entropy estimator. We prove the consistency of our estimator. Monte Carlo studies show that our estimator is better than the entropy estimators proposed by Vasicek, Ebrahimi et al. [Two measures of sample entropy, Stat. Probab. Lett. 20 (1994), pp. 225–234] and Correa [A new estimator of entropy, Commun. Stat. Theory Methods 24 (1995), pp. 2439–2449] in terms of root mean square error. We then derive the non-parametric distribution function corresponding to our proposed entropy estimator as a piece-wise uniform distribution. We also introduce goodness-of-fit tests for testing exponentiality and normality based on the said distribution and compare its performance with their leading competitors.  相似文献   

18.
In a recent issue of this journal, Holgersson et al. [Dummy variables vs. category-wise models, J. Appl. Stat. 41(2) (2014), pp. 233–241, doi:10.1080/02664763.2013.838665] compared the use of dummy coding in regression analysis to the use of category-wise models (i.e. estimating separate regression models for each group) with respect to estimating and testing group differences in intercept and in slope. They presented three objections against the use of dummy variables in a single regression equation, which could be overcome by the category-wise approach. In this note, I first comment on each of these three objections and next draw attention to some other issues in comparing these two approaches. This commentary further clarifies the differences and similarities between dummy variable and category-wise approaches.  相似文献   

19.
Meta-analysis refers to a quantitative method for combining results from independent studies in order to draw overall conclusions. We consider hierarchical models including selection models under a skewed heavy tailed error distribution proposed originally by Chen, Dey, and Shao [M. H. Chen, D. K. Dey, Q. M. Shao, A new skewed link model for dichotomous quantal response data, J. Amer. Statist. Assoc. 94 (1983), pp. 1172–1186.] and Branco and Dey [D. Branco and D.K. Dey, A general class of multivariate skew-elliptical distributions, J. Multivariate Anal. 79, pp. 99–113.]. These rich classes of models combine the information of independent studies, allowing investigation of variability both between and within studies and incorporating weight functions. We constructed a detailed computational scheme under skewed normal and skewed Student's t distribution using the MCMC method. Bayesian model selection was conducted by Bayes factor under a different skewed error. Finally, we illustrated our methodology using a real data example taken from Johnson [M.F. Johnson, Comparative efficacy of Naf and SMFP dentifrices in caries prevention: a meta-analysis overview, J Eur. Organ. Caries Res. 27 (1993), pp. 328–336.].  相似文献   

20.
In a recent volume of this journal, Holden [Testing the normality assumption in the Tobit Model, J. Appl. Stat. 31 (2004) pp. 521–532] presents Monte Carlo evidence comparing several tests for departures from normality in the Tobit Model. This study adds to the work of Holden by considering another test, and several information criteria, for detecting departures from normality in the Tobit Model. The test given here is a modified likelihood ratio statistic based on a partially adaptive estimator of the Censored Regression Model using the approach of Caudill [A partially adaptive estimator for the Censored Regression Model based on a mixture of normal distributions, Working Paper, Department of Economics, Auburn University, 2007]. The information criteria examined include the Akaike’s Information Criterion (AIC), the Consistent AIC (CAIC), the Bayesian information criterion (BIC), and the Akaike’s BIC (ABIC). In terms of fewest ‘rejections’ of a true null, the best performance is exhibited by the CAIC and the BIC, although, like some of the statistics examined by Holden, there are computational difficulties with each.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号