首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 437 毫秒
1.
Various matrix-trace Cauchy-Schwarz and related inequalities involving positive semidefinite matrices are obtained. Applications of some of these results to canonical correlation analysis are presented.  相似文献   

2.
The Gaussian rank correlation equals the usual correlation coefficient computed from the normal scores of the data. Although its influence function is unbounded, it still has attractive robustness properties. In particular, its breakdown point is above 12%. Moreover, the estimator is consistent and asymptotically efficient at the normal distribution. The correlation matrix obtained from pairwise Gaussian rank correlations is always positive semidefinite, and very easy to compute, also in high dimensions. We compare the properties of the Gaussian rank correlation with the popular Kendall and Spearman correlation measures. A simulation study confirms the good efficiency and robustness properties of the Gaussian rank correlation. In the empirical application, we show how it can be used for multivariate outlier detection based on robust principal component analysis.  相似文献   

3.
Statistical inference for the diffusion coefficients of multivariate diffusion processes has been well established in recent years; however, it is not the case for the drift coefficients. Furthermore, most existing estimation methods for the drift coefficients are proposed under the assumption that the diffusion matrix is positive definite and time homogeneous. In this article, we put forward two estimation approaches for estimating the drift coefficients of the multivariate diffusion models with the time inhomogeneously positive semidefinite diffusion matrix. They are maximum likelihood estimation methods based on both the martingale representation theorem and conditional characteristic functions and the generalized method of moments based on conditional characteristic functions, respectively. Consistency and asymptotic normality of the generalized method of moments estimation are also proved in this article. Simulation results demonstrate that these methods work well.  相似文献   

4.
ABSTRACT

Stress testing correlation matrix is a challenging exercise for portfolio risk management. Most existing methods directly modify the estimated correlation matrix to satisfy stress conditions while maintaining positive semidefiniteness. The focus lies on technical optimization issues but the resultant stressed correlation matrices usually lack statistical interpretations. In this article, we suggest a novel approach using Empirical Likelihood method to modify the probability weights of sample observations to construct a stressed correlation matrix. The resultant correlations correspond to a stress scenario that is nearest to the observed scenario in a Kullback–Leibler divergence sense. Besides providing a clearer statistical interpretation, the proposed method is non-parametric in distribution, simple in computation and free from subjective tunings. We illustrate the method through an application to a portfolio of international assets.  相似文献   

5.
Suppose estimates are available for correlations between pairs of variables but that the matrix of correlation estimates is not positive definite. In various applications, having a valid correlation matrix is important in connection with follow‐up analyses that might, for example, involve sampling from a valid distribution. We present new methods for adjusting the initial estimates to form a proper, that is, nonnegative definite, correlation matrix. These are based on constructing certain pseudo‐likelihood functions, formed by multiplying together exact or approximate likelihood contributions associated with the individual correlations. Such pseudo‐likelihoods may then be maximized over the range of proper correlation matrices. They may also be utilized to form pseudo‐posterior distributions for the unknown correlation matrix, by factoring in relevant prior information for the separate correlations. We illustrate our methods on two examples from a financial time series and genomic pathway analysis.  相似文献   

6.
One of the important goals of regression diagnostics is the detection of cases or groups of cases which have an inordinate impact on the regression results. Such observations are generally described as influential. A number of influence measures have been proposed, each focusing on a different aspect of the regression. For single cases, these measures are relatively simple and inexpensive to calculate. However, the detection of multiple-case or joint influence is more difficult on two counts. First, calculation of influence for a single subset is more involved than for an individual case, and second, the sheer number of subsets of cases makes the computation overwhelming for all but the smallest data sets.Barrett and Gray (1992) described methods for efficiently examining subset influence for those measures that can be expressed as the trace of a product of positive semidefinite (psd) matrices. There are, however, other popular measures that do not take this form, but rather are expressible as the ratio of determinants of psd matrices. This article focuses on reducing the computation for the determinantal ratio measures by making use of upper and lower bounds on the influence to limit the number of subsets for which the actual influence must be explicitly determined.  相似文献   

7.
We consider a vector conditional heteroscedastic autoregressive nonlinear (CHARN) model in which both the conditional mean and the conditional variance (volatility) matrix are unknown functions of the past. Nonparametric estimators of these functions are constructed based on local polynomial fitting. We examine the rates of convergence of these estimators and give a result on their asymptotic normality. These results are applied to estimation of volatility matrices in foreign exchange markets. Estimation of the conditional covariance surface for the Deutsche Mark/US Dollar (DEM/USD) and Deutsche Mark/British Pound (DEM/GBP) daily returns show negative correlation when the two series have opposite lagged values and positive correlation elsewhere. The relation of our findings to the capital asset pricing model is discussed.  相似文献   

8.
Block-structured correlation matrices are correlation matrices in which the p variables are subdivided into homogeneous groups, with equal correlations for variables within each group, and equal correlations between any given pair of variables from different groups. Block-structured correlation matrices arise as approximations for certain data sets’ true correlation matrices. A block structure in a correlation matrix entails a certain number of properties regarding its eigendecomposition and, therefore, a principal component analysis of the underlying data. This paper explores these properties, both from an algebraic and a geometric perspective, and discusses their robustness. Suggestions are also made regarding the choice of variables to be subjected to a principal component analysis, when in the presence of (approximately) block-structured variables.  相似文献   

9.
In semidefinite programming (SDP), we minimize a linear objective function subject to a linear matrix being positive semidefinite. A powerful program, SeDuMi, has been developed in MATLAB to solve SDP problems. In this article, we show in detail how to formulate A-optimal and E-optimal design problems as SDP problems and solve them by SeDuMi. This technique can be used to construct approximate A-optimal and E-optimal designs for all linear and nonlinear regression models with discrete design spaces. In addition, the results on discrete design spaces provide useful guidance for finding optimal designs on any continuous design space, and a convergence result is derived. Moreover, restrictions in the designs can be easily incorporated in the SDP problems and solved by SeDuMi. Several representative examples and one MATLAB program are given.  相似文献   

10.
Abstract

Although stochastic volatility and GARCH (generalized autoregressive conditional heteroscedasticity) models have successfully described the volatility dynamics of univariate asset returns, extending them to the multivariate models with dynamic correlations has been difficult due to several major problems. First, there are too many parameters to estimate if available data are only daily returns, which results in unstable estimates. One solution to this problem is to incorporate additional observations based on intraday asset returns, such as realized covariances. Second, since multivariate asset returns are not synchronously traded, we have to use the largest time intervals such that all asset returns are observed to compute the realized covariance matrices. However, in this study, we fail to make full use of the available intraday informations when there are less frequently traded assets. Third, it is not straightforward to guarantee that the estimated (and the realized) covariance matrices are positive definite.

Our contributions are the following: (1) we obtain the stable parameter estimates for the dynamic correlation models using the realized measures, (2) we make full use of intraday informations by using pairwise realized correlations, (3) the covariance matrices are guaranteed to be positive definite, (4) we avoid the arbitrariness of the ordering of asset returns, (5) we propose the flexible correlation structure model (e.g., such as setting some correlations to be zero if necessary), and (6) the parsimonious specification for the leverage effect is proposed. Our proposed models are applied to the daily returns of nine U.S. stocks with their realized volatilities and pairwise realized correlations and are shown to outperform the existing models with respect to portfolio performances.  相似文献   

11.
Regression methods for common data types such as measured, count and categorical variables are well understood but increasingly statisticians need ways to model relationships between variable types such as shapes, curves, trees, correlation matrices and images that do not fit into the standard framework. Data types that lie in metric spaces but not in vector spaces are difficult to use within the usual regression setting, either as the response and/or a predictor. We represent the information in these variables using distance matrices which requires only the specification of a distance function. A low-dimensional representation of such distance matrices can be obtained using methods such as multidimensional scaling. Once these variables have been represented as scores, an internal model linking the predictors and the responses can be developed using standard methods. We call scoring as the transformation from a new observation to a score, whereas backscoring is a method to represent a score as an observation in the data space. Both methods are essential for prediction and explanation. We illustrate the methodology for shape data, unregistered curve data and correlation matrices using motion capture data from an experiment to study the motion of children with cleft lip.  相似文献   

12.
The estimation of variance-covariance matrices through optimization of an objective function, such as a log-likelihood function, is usually a difficult numerical problem. Since the estimates should be positive semi-definite matrices, we must use constrained optimization, or employ a parametrization that enforces this condition. We describe here five different parametrizations for variance-covariance matrices that ensure positive definiteness, thus leaving the estimation problem unconstrained. We compare the parametrizations based on their computational efficiency and statistical interpretability. The results described here are particularly useful in maximum likelihood and restricted maximum likelihood estimation in linear and non-linear mixed-effects models, but are also applicable to other areas of statistics.  相似文献   

13.
基于小波协方差的中国股市波动序列相关性的实证分析   总被引:1,自引:0,他引:1  
在介绍概率变化协调的相关性度量方法的同时,证明了该方法是传统方法的推广。又依据小波协方差在不同尺度下的分解理论,提出了基于小波协方差的相关性度量方法,并对沪深股市波动序列之间的相关性进行了实证分析。结果表明:沪深股市波动序列在整体上具有正相关性,但在不同尺度下沪深股市波动序列之间的相关性不同,小尺度下相关性小。对投资者而言,最好以小尺度为基准选择分散投资策略。  相似文献   

14.
Incomplete growth curve data often result from missing or mistimed observations in a repeated measures design. Virtually all methods of analysis rely on the dispersion matrix estimates. A Monte Carlo simulation was used to compare three methods of estimation of dispersion matrices for incomplete growth curve data. The three methods were: 1) maximum likelihood estimation with a smoothing algorithm, which finds the closest positive semidefinite estimate of the pairwise estimated dispersion matrix; 2) a mixed effects model using the EM (estimation maximization) algorithm; and 3) a mixed effects model with the scoring algorithm. The simulation included 5 dispersion structures, 20 or 40 subjects with 4 or 8 observations per subject and 10 or 30% missing data. In all the simulations, the smoothing algorithm was the poorest estimator of the dispersion matrix. In most cases, there were no significant differences between the scoring and EM algorithms. The EM algorithm tended to be better than the scoring algorithm when the variances of the random effects were close to zero, especially for the simulations with 4 observations per subject and two random effects.  相似文献   

15.
Multivariate associated kernel estimators, which depend on both target point and bandwidth matrix, are appropriate for distributions with partially or totally bounded supports and generalize the classical ones such as the Gaussian. Previous studies on multivariate associated kernels have been restricted to products of univariate associated kernels, also considered having diagonal bandwidth matrices. However, it has been shown in classical cases that, for certain forms of target density such as multimodal ones, the use of full bandwidth matrices offers the potential for significantly improved density estimation. In this paper, general associated kernel estimators with correlation structure are introduced. Asymptotic properties of these estimators are presented; in particular, the boundary bias is investigated. Generalized bivariate beta kernels are handled in more details. The associated kernel with a correlation structure is built with a variant of the mode-dispersion method and two families of bandwidth matrices are discussed using the least squared cross validation method. Simulation studies are done. In the particular situation of bivariate beta kernels, a very good performance of associated kernel estimators with correlation structure is observed compared to the diagonal case. Finally, an illustration on a real dataset of paired rates in a framework of political elections is presented.  相似文献   

16.
ABSTRACT

This article describes a new algorithm for generating correlation matrices with specified eigenvalues. The algorithm uses the method of alternating projections (MAP) that was first described by Neumann. The MAP algorithm for generating correlation matrices is both easy to understand and to program in higher-level computer languages, making this method accessible to applied researchers with no formal training in advanced mathematics. Simulations indicate that the new algorithm has excellent convergence properties. Correlation matrices with specified eigenvalues can be profitably used in Monte Carlo research in statistics, psychometrics, computer science, and related disciplines. To encourage such use, R code (R Core Team) for implementing the algorithm is provided in the supplementary material.  相似文献   

17.
A NEW PROCEDURE FOR ASSESSING LARGE SETS OF CORRELATIONS   总被引:1,自引:0,他引:1  
In this paper, a new test of the hypothesis that all the correlations between a set of variables are zero is proposed. It is based on the asymptotic behaviour of the largest of the observed correlation coefficients. Here “asymptotic” refers to the size of the correlation matrix considered. Simulations show that the critical levels, calculated using the asymptotic theory, are conservative but quite accurate, even for small correlation matrices.  相似文献   

18.
This study, based on systematic review and meta-analysis, aimed to collect and analyze evidence on correlation between perception-based journal rankings and the most popular citation-based measure, Thomson Reuters Journal Impact Factor (JIF). A search was conducted in the Web of Science, Scopus, and Google Scholar databases. After the screening of titles, abstracts, and full text, 18 articles were selected as eligible for review and analysis. The included studies belonged to various subject areas in social sciences, science, and technology. The correlation coefficients found in most of the studies were statistically significant in a positive direction. The heterogeneity test was positive. Therefore, the random-effect method of meta-analysis was applied. The value of pooled correlation coefficient indicated a moderate positive relationship between two methods of assessing the quality of academic journals. The absence of a high correlation makes decision making based on a single ranking method dangerous. Therefore, a hybrid approach for journal assessment is recommended.  相似文献   

19.
The well-known INDSCAL model for simultaneous metric multidimensional scaling (MDS) of three-way data analyzes doubly centered matrices of squared dissimilarities. An alternative approach, called for short DINDSCAL, is proposed for analyzing directly the input matrices of squared dissimilarities. An important consequence is that missing values can be easily handled. The DINDSCAL problem is solved by means of the projected gradient approach. First, the problem is transformed into a gradient dynamical system on a product matrix manifold (of Stiefel sub-manifold of zero-sum matrices and non-negative diagonal matrices). The constructed dynamical system can be numerically integrated which gives a globally convergent algorithm for solving the DINDSCAL. The DINDSCAL problem and its solution are illustrated by well-known data routinely used in metric MDS and INDSCAL. Alternatively, the problem can also be solved by iterative algorithm based on the conjugate (projected) gradient method, which MATLAB implementation is enclosed as an appendix.  相似文献   

20.
We present geodesic Lagrangian Monte Carlo, an extension of Hamiltonian Monte Carlo for sampling from posterior distributions defined on general Riemannian manifolds. We apply this new algorithm to Bayesian inference on symmetric or Hermitian positive definite (PD) matrices. To do so, we exploit the Riemannian structure induced by Cartan's canonical metric. The geodesics that correspond to this metric are available in closed-form and – within the context of Lagrangian Monte Carlo – provide a principled way to travel around the space of PD matrices. Our method improves Bayesian inference on such matrices by allowing for a broad range of priors, so we are not limited to conjugate priors only. In the context of spectral density estimation, we use the (non-conjugate) complex reference prior as an example modelling option made available by the algorithm. Results based on simulated and real-world multivariate time series are presented in this context, and future directions are outlined.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号