首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 187 毫秒
1.
An inequality for the sum of squares of rank differences associated with Spearman’s rank correlation coefficient, when ties and missing data are present in both rankings, was established numerically in Loukas and Papaioannou (1991 Loukas, S., Papaioannou, T. (1991). Rank correlation inequalities with ties and missing data. Stat. Probab. Lett. 11:5356.[Crossref], [Web of Science ®] [Google Scholar]). That inequality is improved and generalized.  相似文献   

2.
Vangeneugden et al. [15 Vangeneugden, T., Molenberghs, G., Laenen, A., Geys, H., Beunckens, C. and Sotto, C. 2007. Marginal correlation in longitudinal binary data based on generalized linear mixed models, Tech. Rep., Hasselt University. submitted for publication [Google Scholar]] derived approximate correlation functions for longitudinal sequences of general data type, Gaussian and non-Gaussian, based on generalized linear mixed-effects models (GLMM). Their focus was on binary sequences, as well as on a combination of binary and Gaussian sequences. Here, we focus on the specific case of repeated count data, important in two respects. First, we employ the model proposed by Molenberghs et al. [13 Molenberghs, G., Verbeke, G. and Demétrio, C. G.B. 2007. An extended random-effects approach to modeling repeated, overdispersed count data. Lifetime Data Anal., 13: 513531. [Crossref], [PubMed], [Web of Science ®] [Google Scholar]], which generalizes at the same time the Poisson-normal GLMM and the conventional overdispersion models, in particular the negative-binomial model. The model flexibly accommodates data hierarchies, intra-sequence correlation, and overdispersion. Second, means, variances, and joint probabilities can be expressed in closed form, allowing for exact intra-sequence correlation expressions. Next to the general situation, some important special cases such as exchangeable clustered outcomes are considered, producing insightful expressions. The closed-form expressions are contrasted with the generic approximate expressions of Vangeneugden et al. [15 Vangeneugden, T., Molenberghs, G., Laenen, A., Geys, H., Beunckens, C. and Sotto, C. 2007. Marginal correlation in longitudinal binary data based on generalized linear mixed models, Tech. Rep., Hasselt University. submitted for publication [Google Scholar]]. Data from an epileptic-seizures trial are analyzed and correlation functions derived. It is shown that the proposed extension strongly outperforms the classical GLMM.  相似文献   

3.
In hierarchical data settings, be it of a longitudinal, spatial, multi-level, clustered, or otherwise repeated nature, often the association between repeated measurements attracts at least part of the scientific interest. Quantifying the association frequently takes the form of a correlation function, including but not limited to intraclass correlation. Vangeneugden et al. (2010 Vangeneugden, T., Molenberghs, G., Laenen, A., Geys, H., Beunckens, C., Sotto, C. (2010). Marginal correlation in longitudinal binary data based on generalized linear mixed models. Communi. Stati. Theory &; Methods. 39:35423557. [Google Scholar]) derived approximate correlation functions for longitudinal sequences of general data type, Gaussian and non-Gaussian, based on generalized linear mixed-effects models. Here, we consider the extended model family proposed by Molenberghs et al. (2010 Molenberghs, G., Verbeke, G., Demétrio, C., Vieira, A. (2010). A family of generalized linear models for repeated measures with normal and conjugate random effects. Stat. Sci. 25:325347.[Crossref], [Web of Science ®] [Google Scholar]). This family flexibly accommodates data hierarchies, intra-sequence correlation, and overdispersion. The family allows for closed-form means, variance functions, and correlation function, for a variety of outcome types and link functions. Unfortunately, for binary data with logit link, closed forms cannot be obtained. This is in contrast with the probit link, for which such closed forms can be derived. It is therefore that we concentrate on the probit case. It is of interest, not only in its own right, but also as an instrument to approximate the logit case, thanks to the well-known probit-logit ‘conversion.’ Next to the general situation, some important special cases such as exchangeable clustered outcomes receive attention because they produce insightful expressions. The closed-form expressions are contrasted with the generic approximate expressions of Vangeneugden et al. (2010 Vangeneugden, T., Molenberghs, G., Laenen, A., Geys, H., Beunckens, C., Sotto, C. (2010). Marginal correlation in longitudinal binary data based on generalized linear mixed models. Communi. Stati. Theory &; Methods. 39:35423557. [Google Scholar]) and with approximations derived for the so-called logistic-beta-normal combined model. A simulation study explores performance of the method proposed. Data from a schizophrenia trial are analyzed and correlation functions derived.  相似文献   

4.
Some extensions of Shannon entropy to the survival function have been recently proposed. Misagh et al. (2011 Misagh, F., Panahi, Y., Yari, G.H., Shahi, R. (2011, September). Weighted cumulative entropy and its estimation. In: Quality and Reliability (ICQR), 2011, IEEE International conference (pp. 477480), IEEE.[Crossref] [Google Scholar]) introduced weighted cumulative residual entropy (WCRE) that was studied more by Mirali et al. (2015 Mirali, M., Baratpour, S., Fakoor, V. (2015). On weighted cumulative residual entropy. Commun. Stat. Theory Methods. doi:10.1080103610926.2015.1053932.[Web of Science ®] [Google Scholar]). In this article, the dynamic version of WCRE is proposed. Some relationships of this measure with well-known reliability measures and ageing classes are studied and some characterization results for exponential and Rayleigh distributions are provided. Also, a non parametric estimation of dynamic version of WCRE is introduced and its asymptotic behavior is investigated.  相似文献   

5.
There are many situations where n objects are ranked by b>2 independent sources or observers and in which the interest is focused on agreement on the top rankings. Kendall's coefficient of concordance [10 M. Kendall and B. Smith, The problem of m rankings, Ann. Math. Stat. 10 (1939), pp. 275287. doi: 10.1214/aoms/1177732186[Crossref] [Google Scholar]] assigns equal weights to all rankings. In this paper, a new coefficient of concordance is introduced which is more sensitive to agreement on the top rankings. The limiting distribution of the new concordance coefficient under the null hypothesis of no association among the rankings is presented, and a summary of the exact and approximate quantiles for this coefficient is provided. A simulation study is carried out to compare the performance of Kendall's, the top-down and the new concordance coefficients in detecting the agreement on the top rankings. Finally, examples are given for illustration purposes, including a real data set from financial market indices.  相似文献   

6.
Abstract

Takahasi and Wakimoto (1968 Takahasi, K., and K. Wakimoto. 1968. On unbiased estimates of the population mean based on the sample stratified by means of ordering. Annals of the Institute of Statistical Mathematics 20:131.[Crossref], [Web of Science ®] [Google Scholar]) derived a sharp upper bound on the efficiency of the balanced ranked-set sampling (RSS) sample mean relative to the simple random sampling (SRS) sample mean under perfect rankings. The bound depends on the set size and is achieved for uniform distributions. Here we generalize the Takahasi and Wakimoto (1968 Takahasi, K., and K. Wakimoto. 1968. On unbiased estimates of the population mean based on the sample stratified by means of ordering. Annals of the Institute of Statistical Mathematics 20:131.[Crossref], [Web of Science ®] [Google Scholar]) result by finding a sharp upper bound in the case of unbalanced RSS. The bound depends on the particular unbalanced design, and the distributions where the bound is achieved can be highly nonuniform. The bound under perfect rankings can be exceeded under imperfect rankings.  相似文献   

7.
This article suggests random and fixed effects spatial two-stage least squares estimators for the generalized mixed regressive spatial autoregressive panel data model. This extends the generalized spatial panel model of Baltagi et al. (2013 Baltagi, B. H., Egger, P., Pfaffermayr, M. (2013). A generalized spatial panel data model with random effects. Econometric Reviews 32:650685.[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]) by the inclusion of a spatial lag term. The estimation method utilizes the Generalized Moments method suggested by Kapoor et al. (2007 Kapoor, M., Kelejian, H. H., Prucha, I. R. (2007). Panel data models with spatially correlated error components. Journal of Econometrics 127(1):97130.[Crossref], [Web of Science ®] [Google Scholar]) for a spatial autoregressive panel data model. We derive the asymptotic distributions of these estimators and suggest a Hausman test a la Mutl and Pfaffermayr (2011 Mutl, J., Pfaffermayr, M. (2011). The Hausman test in a Cliff and Ord panel model. Econometrics Journal 14:4876.[Crossref], [Web of Science ®] [Google Scholar]) based on the difference between these estimators. Monte Carlo experiments are performed to investigate the performance of these estimators as well as the corresponding Hausman test.  相似文献   

8.
The Hosmer–Lemeshow test is a widely used method for evaluating the goodness of fit of logistic regression models. But its power is much influenced by the sample size, like other chi-square tests. Paul, Pennell, and Lemeshow (2013 Paul, P., M. L. Pennell, and S. Lemeshow. 2013. Standardizing the power of the Hosmer–Lemeshow goodness of fit test in large data sets. Statistics in Medicine 32:6780.[Crossref], [PubMed], [Web of Science ®] [Google Scholar]) considered using a large number of groups for large data sets to standardize the power. But simulations show that their method performs poorly for some models. In addition, it does not work when the sample size is larger than 25,000. In the present paper, we propose a modified Hosmer–Lemeshow test that is based on estimation and standardization of the distribution parameter of the Hosmer–Lemeshow statistic. We provide a mathematical derivation for obtaining the critical value and power of our test. Through simulations, we can see that our method satisfactorily standardizes the power of the Hosmer–Lemeshow test. It is especially recommendable for enough large data sets, as the power is rather stable. A bank marketing data set is also analyzed for comparison with existing methods.  相似文献   

9.
In recent articles, Fajardo et al. (2009 Fajardo Molinares, F., Reisen, V.A., Cribari-Neto, F. (2009). Robust estimation in long-memory processes under additive outliers. Journal of Statistical Planning and Inference 139:25112525.[Crossref], [Web of Science ®] [Google Scholar]) and Reisen and Fajardo (2012) propose an alternative semiparametric estimator of the fractional parameter in ARFIMA models which is robust to the presence of additive outliers. The results are very interesting, however, they use samples of 300 or 800 observations which are rarely found in macroeconomics. In order to perform a comparison, I estimate the fractional parameter using the procedure of Geweke and Porter-Hudak (1983 Geweke, J., Porter-Hudak, S. (1983). The estimation and application of long memory time series model. Journal of Time Series Analysis 4:221238.[Crossref] [Google Scholar]) augmented with dummy variables associated with the (previously) detected outliers using the statistic τd suggested by Perron and Rodríguez (2003 Perron, P., Rodríguez, G. (2003). Searching for additive outliers in nonstationary time series. Journal of Time Series Analysis 24(2):193220.[Crossref], [Web of Science ®] [Google Scholar]). Comparing with Fajardo et al. (2009 Fajardo Molinares, F., Reisen, V.A., Cribari-Neto, F. (2009). Robust estimation in long-memory processes under additive outliers. Journal of Statistical Planning and Inference 139:25112525.[Crossref], [Web of Science ®] [Google Scholar]) and Reisen and Fajardo (2012), I found better results for the mean and bias of the fractional parameter when T = 100 and the results in terms of the standard deviation and the MSE are very similar. However, for higher sample sizes such as 300 or 800, the robust procedure performs better. Empirical applications for seven monthly Latin-American inflation series with very small sample sizes contaminated by additive outliers are discussed.  相似文献   

10.
The objective of this paper is to study U-type designs for Bayesian non parametric response surface prediction under correlated errors. The asymptotic Bayes criterion is developed in terms of the asymptotic approach of Mitchell et al. (1994 Mitchell, T., Sacks, J., Ylvisaker, D. (1994). Asymptotic Bayes criteria for nonparametric response surface design. Ann. Stat. 22:634651.[Crossref], [Web of Science ®] [Google Scholar]) for a more general covariance kernel proposed by Chatterjee and Qin (2011 Chatterjee, K., Qin, H. (2011). Generalized discrete discrepancy and its applications in experimental designs. J. Stat. Plann. Inference 141:951960.[Crossref], [Web of Science ®] [Google Scholar]). A relationship between the asymptotic Bayes criterion and other criteria, such as orthogonality and aberration, is then developed. A lower bound for the criterion is also obtained, and numerical results show that this lower bound is tight. The established results generalize those of Yue et al. (2011 Yue, R.X., Qin, H., Chatterjee, K. (2011). Optimal U-type design for Bayesian nonparametric multiresponse prediction. J. Stat. Plann. Inference 141:24722479.[Crossref], [Web of Science ®] [Google Scholar]) from symmetrical case to asymmetrical U-type designs.  相似文献   

11.
The inverse Gaussian distribution is often suited for modeling positive and/or positively skewed data (see Chhikara and Folks, 1989 Chhikara , R. S. , Folks , J. L. ( 1989 ). The Inverse Gaussian Distribution . New York : Marcel Dekker . [Google Scholar]) and presents an interesting alternative to the Gaussian model in such cases. We note here that overlap coefficients and their variants are widely studied in the literature for Gaussian populations (see Mulekar and Mishra, 1994 Mulekar , M. , Mishra , S. N. ( 1994 ). Overlap coefficients of two normal densities: equal means case . J. Japan. Statist. Soc. 24 : 169180 . [Google Scholar], 2000 Mulekar , M. , Mishra , S. N. ( 2000 ). Confidence interval estimation of overlap: equal means case . Computat. Statist. Data Anal. 34 : 121137 .[Crossref], [Web of Science ®] [Google Scholar], and references therein for further details). This article studies the properties and addresses point estimation for large samples of commonly used measures of overlap when the populations are described by inverse Gaussian distributions. The bias and mean square error properties of the estimators are studied through a simulation study.  相似文献   

12.
In this article, we investigate the relationships among intraday serial correlation, jump-robust volatility, positive and negative jumps based on Shanghai composite index high frequency data. We implement variance ratio test to quantify intraday serial correlation. We also measure the continuous part of realized volatility using jump-robust MedRV estimator and disentangle positive and negative jumps using Realized Downside Risk Measure and Realized Upside Potential Measure proposed by Bi et al., (2013 Bi, T., Zhang, B., Wu, H. (2013). Measuring downside risk using high frequency data–realized downside risk measure. Communications in Statistics–Simulation and Computation 42(4):741754.[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]). We find that intraday serial correlation are positively correlated with jump-robust volatility and negatively correlated with negative jumps which confirm the LeBaron effect.  相似文献   

13.
In this paper, the focus is on sequential analysis of multivariate financial time series with heavy tails. The mean vector and the covariance matrix of multivariate non linear models are simultaneously monitored by modifying conventional control charts to identify structural changes in the data. The considered target process is a constant conditional correlation model (cf. Bollerslev, 1990 Bollerslev, T. (1990). Modeling the coherence in short-run nominal exchange rates: A multivariate generalized ARCH model. Rev. Econ. Stat. 72:498505.[Crossref], [Web of Science ®] [Google Scholar]), an extended constant conditional correlation model (cf. He and Teräsvirta, 2004 He, C., Teräsvirta, T. (2004). An extended constant conditional correlation GARCH model and its fourth-moment structure. Economet. Theory 20:904926.[Crossref], [Web of Science ®] [Google Scholar]), a dynamic conditional correlation model (cf. Engle, 2002 Engle, R.F. (2002). Dynamic conditional correlation: A simple class of multivariate GARCH models. J. Bus. Econ. Stat. 20(3):339350.[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]), or a generalized dynamic conditional correlation model (cf. Capiello et al., 2006 Capiello, L., Engle, R., Sheppard, K. (2006). Asymmetric correlations in the dynamics of global equity and bond returns. J. Financial Economet. 4(4):537572.[Crossref] [Google Scholar]). For statistical surveillance we use control charts based on residuals. Further, the procedures are constructed for t-distribution. The detection speed of these charts is compared via Monte Carlo simulation. In the empirical study, the procedure with the best performance is applied to log-returns of the stock market indices FTSE and CAC.  相似文献   

14.
We show that the rearrangement algorithm (RA) introduced in Puccetti and Rüschendorf (2012 Puccetti, G., Rüschendorf, L. (2012). Computation of sharp bounds on the distribution of a function of dependent risks. Journal of Computational and Applied Mathematics 236(7):18331840.[Crossref], [Web of Science ®] [Google Scholar]) to compute distributional bounds can be used also to compute sharp lower and upper bounds on the expected value of a supermodular function of d random variables having fixed marginal distributions. Compared to the analytical methods existing in the literature the algorithm is widely applicable, more easily obtained and gives insight into the dependence structures attaining the bounds.  相似文献   

15.
Recently, Abbasnejad et al. (2010 Abbasnejad, M., Arghami, N.R., Morgenthaler, S., Mohtashami Borzadaran, G.R. (2010). On the dynamic survival entropy. Stat. Probab. Lett. 80:19621971.[Crossref], [Web of Science ®] [Google Scholar]) proposed a measure of uncertainty based on survival function, called the survival entropy of order α. A dynamic form of the survival entropy of order α is also proposed by them. In this paper, we derive the weighted form of these measures. The properties of the new measures are also discussed.  相似文献   

16.
This paper presents a new variable weight method, called the singular value decomposition (SVD) approach, for Kohonen competitive learning (KCL) algorithms based on the concept of Varshavsky et al. [18 R. Varshavsky, A. Gottlieb, M. Linial, and D. Horn, Novel unsupervised feature filtering of bilogical data, Bioinformatics 22 (2006), pp. 507513.[Crossref], [PubMed], [Web of Science ®] [Google Scholar]]. Integrating the weighted fuzzy c-means (FCM) algorithm with KCL, in this paper, we propose a weighted fuzzy KCL (WFKCL) algorithm. The goal of the proposed WFKCL algorithm is to reduce the clustering error rate when data contain some noise variables. Compared with the k-means, FCM and KCL with existing variable-weight methods, the proposed WFKCL algorithm with the proposed SVD's weight method provides a better clustering performance based on the error rate criterion. Furthermore, the complexity of the proposed SVD's approach is less than Pal et al. [17 S.K. Pal, R.K. De, and J. Basak, Unsupervised feature evaluation: a neuro-fuzzy approach, IEEE. Trans. Neural Netw. 11 (2000), pp. 366376.[Crossref], [PubMed], [Web of Science ®] [Google Scholar]], Wang et al. [19 X.Z. Wang, Y.D. Wang, and L.J. Wang, Improving fuzzy c-means clustering based on feature-weight learning, Pattern Recognit. Lett. 25 (2004), pp. 11231132.[Crossref], [Web of Science ®] [Google Scholar]] and Hung et al. [9 W. -L. Hung, M. -S. Yang, and D. -H. Chen, Bootstrapping approach to feature-weight selection in fuzzy c-means algorithms with an application in color image segmentation, Pattern Recognit. Lett. 29 (2008), pp. 13171325.[Crossref], [Web of Science ®] [Google Scholar]].  相似文献   

17.
The logistic distribution and the S-shaped pattern of its cumulative distribution and quantile functions have been extensively used in many different spheres affecting human life. By far, the most well-known application of logistic distribution is in the logistic regression that is used for modeling categorical response variables. The exponentiated-exponential logistic distribution, a generalization of the logistic distribution, is obtained using the technique proposed by Alzaatreh et al. (2013 Alzaatreh, A., C. Lee, and F. Famoye. 2013. A new method for generating families of continuous distribution. Metron. 71:6379.[Crossref] [Google Scholar]) of mixing two distributions, hereafter called the EEL distribution. This distribution subsumes various types of logistic distribution. The structural analysis of the distribution in this paper includes limiting behavior, quantiles, moments, mode, skewness, kurtosis, order statistics, the large sample distributions of the sample maximum and the sample minimum, and the distribution of the sample median. For illustrative purposes, a real-life data set is considered as an application of the EEL distribution.  相似文献   

18.
The hypothesis of structural stability that the regression coefficients do not change over time is central to all applications of linear regression models. It is rather surprising that existing theory as well as practice focus on testing for structural change under homoskedasticity – that is, regression coefficients may change, but the variances remain the same. Since structural change can, and often does, involve changes in variances, this is a puzzling gap in the literature. Our main focus in this paper is to utilize a newly developed test (MZ) by Maasoumi et al. (2010 Maasoumi, E., Zaman, A., Ahmed, M. (2010). Tests for structural change, aggregation, and homogeneity. Econ Model. 27(6):13821391.[Crossref], [Web of Science ®] [Google Scholar]) that tests simultaneously for break in regression coefficients as well as in variance. Currently, the sup F test is most widely used for structural change. This has certain optimality properties shown by Andrews (1993 Andrews, D.W.K. (1993). Test for parameter instability and structural change with unknown change point. Econometrica. 61(4):821856.[Crossref], [Web of Science ®] [Google Scholar]). However, this test assumes homoskedasticity across the structural change. We introduce the sup MZ test which caters to unknown breakpoints, and also compare it to the sup F. Our Monte Carlo results show that sup MZ test incurs only a low cost in case of homoskedasticity while having hugely better performance in case of heteroskedasticity. The simulation results are further supported by providing a real-world application. In real-world datasets, we find that structural change often involves heteroskedasticity. In such cases, the sup F test can fail to detect structural breaks and give misleading results, while the sup MZ test works well. We conclude that the sup MZ test is superior to current methodology for detecting structural change.  相似文献   

19.
By using the medical data analyzed by Kang et al. (2007 Kang, C.W., Lee, M.S., Seong, Y.J., Hawkins, D.M. (2007). A control chart for the coefficient of variation. J. Qual. Technol. 39(2):151158.[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]), a Bayesian procedure is applied to obtain control limits for the coefficient of variation. Reference and probability matching priors are derived for a common coefficient of variation across the range of sample values. By simulating the posterior predictive density function of a future coefficient of variation, it is shown that the control limits are effectively identical to those obtained by Kang et al. (2007 Kang, C.W., Lee, M.S., Seong, Y.J., Hawkins, D.M. (2007). A control chart for the coefficient of variation. J. Qual. Technol. 39(2):151158.[Taylor &; Francis Online], [Web of Science ®] [Google Scholar]) for the specific dataset they used. This article illustrates the flexibility and unique features of the Bayesian simulation method for obtaining posterior distributions, predictive intervals, and run-lengths in the case of the coefficient of variation. A simulation study shows that the 95% Bayesian confidence intervals for the coefficient of variation have the correct frequentist coverage.  相似文献   

20.
In this note, we introduce a new class of dependent random variables (henceforth rvs), together with some its basic properties. This class includes independent rvs and pairwise negatively dependent rvs. Some extensions of Ranjbar et al. (2008) are discussed. The complete convergence for the new class of rvs is also proved, and some results of Beak and Park (2010 Beak, J.-II., and S. T. Park. 2010. Convergence of weighted sums for arrays of negatively dependent random variables and its applications. J. Stat. Plann. Inference 140:24612469.[Crossref], [Web of Science ®] [Google Scholar]) are extended to this class conveniently.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号