首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 375 毫秒
1.
The most popular method for trying to detect an association between two random variables is to test H 0 ?:?ρ=0, the hypothesis that Pearson's correlation is equal to zero. It is well known, however, that Pearson's correlation is not robust, roughly meaning that small changes in any distribution, including any bivariate normal distribution as a special case, can alter its value. Moreover, the usual estimate of ρ, r, is sensitive to only a few outliers which can mask a true association. A simple alternative to testing H 0 ?:?ρ =0 is to switch to a measure of association that guards against outliers among the marginal distributions such as Kendall's tau, Spearman's rho, a Winsorized correlation, or a so-called percentage bend correlation. But it is known that these methods fail to take into account the overall structure of the data. Many measures of association that do take into account the overall structure of the data have been proposed, but it seems that nothing is known about how they might be used to detect dependence. One such measure of association is selected, which is designed so that under bivariate normality, its estimator gives a reasonably accurate estimate of ρ. Then methods for testing the hypothesis of a zero correlation are studied.  相似文献   

2.
Abstract

In this paper, we introduce Liu estimator for the vector of parameters in linear measurement error models and discuss its asymptotic properties. Based on the Liu estimator, diagnostic measures are developed to identify influential observations. Additionally, the analogs of Cook’s distance and likelihood distance are proposed to determine influential observations using case deletion approach. A parametric bootstrap procedure is used to obtain empirical distributions of the test statistics. Finally, the performance of the influence measures have been illustrated through simulation study and analyzing a real data set.  相似文献   

3.
ABSTRACT

In this work, we deal with a bivariate time series of wind speed and direction. Our observed data have peculiar features, such as informative missing values, non-reliable measures under a specific condition and interval-censored data, that we take into account in the model specification. We analyse the time series with a non-parametric Bayesian hidden Markov model, introducing a new emission distribution, suitable to model our data, based on the invariant wrapped Poisson, the Poisson and the hurdle density. The model is estimated on simulated datasets and on the real data example that motivated this work.  相似文献   

4.
ABSTRACT

In this paper we discuss the identification of influential observations in a growth curve model with Rao's simple covariance structure. Based on the generalized Cook-type distance and the volume of a confidence ellipsoid, a variety of influence measures are proposed in terms of the case-deletion technique. Also, the influence of observations on a linear combination of regression coefficients is considered. For illustration, a practical example is analyzed using the proposed approach.  相似文献   

5.
ABSTRACT

There is a growing interest to get a fully MR based radiotherapy. The most important development needed is to obtain improved bone tissue estimation. The existing model-based methods perform poorly on bone tissues. This paper was aimed at obtaining improved bone tissue estimation. Skew-Gaussian mixture model and Gaussian mixture model were proposed to investigate CT image estimation from MR images by partitioning the data into two major tissue types. The performance of the proposed models was evaluated using the leave-one-out cross-validation method on real data. In comparison with the existing model-based approaches, the model-based partitioning approach outperformed in bone tissue estimation, especially in dense bone tissue estimation.  相似文献   

6.
ABSTRACT

In this article, we look into the properties and characterizations of the New Zenga curve. The relationship of the curve with other measures of inequality as well as some reliability concepts are examined. Classification of lifetime distributions using the Zenga curve and an illustration for the behaviour of the curve using a survival data are also provided.  相似文献   

7.
ABSTRACT

This paper extends the classical methods of analysis of a two-way contingency table to the fuzzy environment for two cases: (1) when the available sample of observations is reported as imprecise data, and (2) the case in which we prefer to categorize the variables based on linguistic terms rather than as crisp quantities. For this purpose, the α-cuts approach is used to extend the usual concepts of the test statistic and p-value to the fuzzy test statistic and fuzzy p-value. In addition, some measures of association are extended to the fuzzy version in order to evaluate the dependence in such contingency tables. Some practical examples are provided to explain the applicability of the proposed methods in real-world problems.  相似文献   

8.
Abstract

In this paper, we show that Y can be introduced into data sharpening to produce non-parametric regression estimators that enjoy high orders of bias reduction. Compared with those in existing literature, the proposed data-sharpening estimator has advantages including simplicity of the estimators, good performance of expectation and variance, and mild assumptions. We generalize this estimator to dependent errors. Finally, we conduct a limited simulation to illustrate that the proposed estimator performs better than existing ones.  相似文献   

9.
ABSTRACT

Discrepancies are measures which are defined as the deviation between the empirical and the theoretical uniform distribution. In this way, discrepancy is a measure of uniformity which provides a way of construction a special kind of space filling designs, namely uniform designs. Several discrepancies have been proposed in recent literature. A brief, selective review of these measures including some construction algorithms are given in this paper. Furthermore, a critical discussion along with some comparisons is provided, as well.  相似文献   

10.
ABSTRACT

We consider semiparametric inference on the partially linearsingle-index model (PLSIM). The generalized likelihood ratio (GLR) test is proposed to examine whether or not a family of new semiparametric models fits adequately our given data in the PLSIM. A new GLR statistic is established to deal with the testing of the index parameter α0 in the PLSIM. The newly proposed statistic is shown to asymptotically follow a χ2-distribution with the scale constant and the degrees of freedom being independent of the nuisance parameters or function. Some finite sample simulations and a real example are used to illustrate our proposed methodology.  相似文献   

11.
ABSTRACT

This article studies the outlier detection problem in mixed regressive-spatial autoregressive model. The formulae for testing outliers and their approximate distributions are derived under the mean-shift model and the variance-weight model, respectively. The simulation studies are conducted for examining the power and size of the test, as well as for the detection of outliers when a simulated data contains several outliers. A real data is analyzed to illustrate the proposed method, and modified models based on mean-shift and variance-weight models in which detected outliers are taken into account are suggested to deal with the outliers and confirm theconclusions.  相似文献   

12.
Abstract

ROC curve is a fundamental evaluation tool in medical researches and survival analysis. The estimation of ROC curve has been studied extensively with complete data and right-censored survival data. However, these methods are not suitable to analyze the length-biased and right-censored data. Since this kind of data includes the auxiliary information that truncation time and residual time share the same distribution, the two new estimators for the ROC curve are proposed by taking into account this auxiliary information to improve estimation efficiency. Numerical simulation studies with different assumed cases and real data analysis are conducted.  相似文献   

13.
Abstract

Cluster analysis is the distribution of objects into different groups or more precisely the partitioning of a data set into subsets (clusters) so that the data in subsets share some common trait according to some distance measure. Unlike classification, in clustering one has to first decide the optimum number of clusters and then assign the objects into different clusters. Solution of such problems for a large number of high dimensional data points is quite complicated and most of the existing algorithms will not perform properly. In the present work a new clustering technique applicable to large data set has been used to cluster the spectra of 702248 galaxies and quasars having 1,540 points in wavelength range imposed by the instrument. The proposed technique has successfully discovered five clusters from this 702,248X1,540 data matrix.  相似文献   

14.
ABSTRACT

In this paper we introduce the exponentiated Fréchet regression for modelling positive responses having a long-tailed distribution in a regression model, which are common in actuarial statistics. We propose two parameterizations each of which links the regression parameters with the explanatory variables. We then discuss the maximum likelihood estimation of the parameters both theoretically and empirically. In order to meet the needs of an actuary, closed-form expressions for certain risk measures for the exponentiated Fréchet distribution are also derived. We employ the proposed model to a motorcycle claim size data set.  相似文献   

15.
Abstract

Failure time data occur in many areas and also in various forms and in particular, many authors have discussed regression analysis of failure time data in the presence of interval censoring, a cured subgroup or mismeasured covariates. However, it does not seem to exist an established procedure that can deal with all three issues together. Corresponding to this, we propose a sieve maximum likelihood estimation procedure that takes into account all three issues with the use of the SIMEX algorithm. The asymptotic properties of the proposed estimators are established, and an extensive simulation study is also conducted and suggests that the proposed method works well for practical situations.  相似文献   

16.
In this article, we study Bayesian estimation for the covariance matrix Σ and the precision matrix Ω (the inverse of the covariance matrix) in the star-shaped model with missing data. Based on a Cholesky-type decomposition of the precision matrix Ω = ΨΨ, where Ψ is a lower triangular matrix with positive diagonal elements, we develop the Jeffreys prior and a reference prior for Ψ. We then introduce a class of priors for Ψ, which includes the invariant Haar measures, Jeffreys prior, and reference prior. The posterior properties are discussed and the closed-form expressions for Bayesian estimators for the covariance matrix Σ and the precision matrix Ω are derived under the Stein loss, entropy loss, and symmetric loss. Some simulation results are given for illustration.  相似文献   

17.
ABSTRACT

Joint models are statistical tools for estimating the association between time-to-event and longitudinal outcomes. One challenge to the application of joint models is its computational complexity. Common estimation methods for joint models include a two-stage method, Bayesian and maximum-likelihood methods. In this work, we consider joint models of a time-to-event outcome and multiple longitudinal processes and develop a maximum-likelihood estimation method using the expectation–maximization algorithm. We assess the performance of the proposed method via simulations and apply the methodology to a data set to determine the association between longitudinal systolic and diastolic blood pressure measures and time to coronary artery disease.  相似文献   

18.
In this article we consider a control chart based on the sample variances of two quality characteristics. The points plotted on the chart correspond to the maximum value of these two statistics. The main reason to consider the proposed chart instead of the generalized variance | S | chart is its better diagnostic feature, that is, with the new chart it is easier to relate an out-of-control signal to the variables whose parameters have moved away from their in-control values. We study the control chart efficiency considering different shifts in the covariance matrix. In this way, we obtain the average run length (ARL) that measures the effectiveness of a control chart in detecting process shifts. The proposed chart always detects process disturbances faster than the generalized variance | S | chart. The same is observed when the size of the samples is variable, except in a few cases in which the size of the samples switches between small size and very large size.  相似文献   

19.
ABSTRACT

Physical phenomena are commonly modelled by time consuming numerical simulators, function of many uncertain parameters whose influences can be measured via a global sensitivity analysis. The usual variance-based indices require too many simulations, especially as the inputs are numerous. To address this limitation, we consider recent advances in dependence measures, focusing on the distance correlation and the Hilbert–Schmidt independence criterion. We study and use these indices for a screening purpose. Numerical tests reveal differences between variance-based indices and dependence measures. Then, two approaches are proposed to use the latter for a screening purpose. The first approach uses independence tests, with existing asymptotic versions and spectral extensions; bootstrap versions are also proposed. The second considers a linear model with dependence measures, coupled to a bootstrap selection method or a Lasso penalization. Numerical experiments show their potential in the presence of many non-influential inputs and give successful results for a nuclear reliability application.  相似文献   

20.
Abstract

The homogeneity hypothesis is investigated in a location family of distributions. A moment-based test is introduced based on data collected from a ranked set sampling scheme. The asymptotic distribution of the proposed test statistic is determined and the performance of the test is studied via simulation. Furthermore, for small sample sizes, the bootstrap procedure is used to distinguish the homogeneity of data. An illustrative example is also presented to explain the proposed procedures in this paper.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号