共查询到20条相似文献,搜索用时 31 毫秒
1.
The cost and time consumption of many industrial experimentations can be reduced using the class of supersaturated designs since this can be used for screening out the important factors from a large set of potentially active variables. A supersaturated design is a design for which there are fewer runs than effects to be estimated. Although there exists a wide study of construction methods for supersaturated designs, their analysis methods are yet in an early research stage. In this article, we propose a method for analyzing data using a correlation-based measure, named as symmetrical uncertainty. This method combines measures from the information theory field and is used as the main idea of variable selection algorithms developed in data mining. In this work, the symmetrical uncertainty is used from another viewpoint in order to determine more directly the important factors. The specific method enables us to use supersaturated designs for analyzing data of generalized linear models for a Bernoulli response. We evaluate our method by using some of the existing supersaturated designs, obtained according to methods proposed by Tang and Wu (1997) as well as by Koukouvinos et al. (2008). The comparison is performed by some simulating experiments and the Type I and Type II error rates are calculated. Additionally, Receiver Operating Characteristics (ROC) curves methodology is applied as an additional statistical tool for performance evaluation. 相似文献
2.
Jean-François Quessy 《统计学通讯:理论与方法》2013,42(19):3510-3531
Population and sample versions of Kendall and Spearman measures of association suitable for multivariate ordinal data are defined. The latter generalize the indices of dependence of Ruymgaart and van Zuijlen (1978), Joe (1990), and Schmid and Schmidt (2007) by allowing atoms in the underlying distribution. The representation of the proposed empirical measures as U-statistics enables to establish their asymptotic normality under general distributions. A special attention is given to tests of independence for multivariate ordinal data, where the power of the new methodologies are investigated under fixed and contiguous alternatives. 相似文献
3.
Feng-Shou Ko 《统计学通讯:理论与方法》2013,42(15):2681-2698
A proposed method based on frailty models is used to identify longitudinal biomarkers or surrogates for a multivariate survival. This method is an extention of earlier models by Wulfsohn and Tsiatis (1997) and Song et al. (2002). In this article, similar to Henderson et al. (2002), a joint likelihood function combines the likelihood functions of the longitudinal biomarkers and the multivariate survival times. We use simulations to explore how the number of individuals, the number of time points per individual and the functional form of the random effects from the longitudianl biomarkers influence the power to detect the association of a longitudinal biomarker and the multivariate survival time. The proposed method is illustrate by using the gastric cancer data. 相似文献
4.
In this article, we find designs insensitive to the presence of an outlier in a diallel cross design setup for estimating a complete set of orthonormal contrasts among the effects of the general combining abilities of a set of parental lines. The criterion of robustness, suggested by Mandal (1989) in block design setup and used by Biswas (2012) in treatment-control setup, is adapted here. Complete diallel cross designs, suggested by Gupta and Kageyama (1994), and partial diallel cross designs, suggested by Gupta et al. (1995) and Mukerjee (1997), are found to be robust under certain conditions. 相似文献
5.
Olivier Darné 《统计学通讯:模拟与计算》2013,42(5):1037-1050
Unit root tests with structural break developed by Zivot and Andrews (1992) and Perron and Rodriguez (2003) in the presence of additive outliers and breaks are studied by simulation experiments. The results show that the Zivot–Andrews test appears to have size distortions due to the additive outliers whereas the Perron–Rodriguez test exhibits good properties of size and power. However, the two tests are biased when a second break is present but not taken into account. Furthermore, these endogenous break unit root tests tend to determine the break point incorrectly at one period behind the true break point, leading to spurious rejections of the unit root null hypothesis. 相似文献
6.
This article considers the problem of estimating the population mean using information on two auxiliary variables in the presence of non response under two-phase sampling. Some improved ratio-in-regression type estimators have been proposed in four different situations of non response along with their properties under large sample approximation. Efficiency comparisons of the proposed estimators with the usual unbiased estimator by Hansen and Hurwitz (1946), conventional ratio and regression estimators using single auxiliary variable and Singh and Kumar (2010b) estimators using two auxiliary variables have been made. Finally, these theoretical findings are illustrated by a numerical example. 相似文献
7.
This article deals with the adaptive estimation of a periodic autoregressive model, with unspecified innovation density satisfying only some general technical assumptions. We first establish, while verifying the adapted sufficient conditions of Swensen (1985) to our model, the Local Asymptotic Normality (LAN), the Local Asymptotic Quadratic (LAQ), and the Local Asymptotic properties satisfied by its central sequence. Secondly, the Locally Asymptotically Minimax (LAM) estimators are constructed. Using these results, we construct the adaptive estimators of the unknown autoregressive parameters. The performances of the established estimators are shown, via simulation studies. 相似文献
8.
Chung-Ho Chen 《统计学通讯:理论与方法》2013,42(10):1767-1778
Economic selection of process parameters has been an important topic in modern statistical process control. The optimum process parameters setting have a major effect on the expected profit/cost per item. There are some concerns on the problem of setting process parameters. Boucher and Jafari (1991) first considered the attribute single sampling plan applied in the selection of process target. Pulak and Al-Sultan (1996) extended Boucher and Jafari's model and presented the rectifying inspection plan for determining the optimum process mean. In this article, we further propose a modified Pulak and Al-Sultan model for determining the optimum process mean and standard deviation under the rectifying inspection plan with the average outgoing quality limit (AOQL) protection. Taguchi's (1986) symmetric quadratic quality loss function is adopted for evaluating the product quality. By solving the modified model, we can obtain the optimum process parameters with the maximum expected profit per item and the specified quality level can be reached. 相似文献
9.
In a recent article, Pedeli and Karlis (2010) examined the extension of the classical Integer–valued Autoregressive (INAR) model to the bivariate case. In the present article, we examine estimation methods for the case of bivariate Poisson innovations. This is a simple extension of the classical INAR model allowing for two discrete valued time series to be correlated. Properties of different estimators are given. We also compare their properties via a small simulation experiment. Extensions to incorporate covariate information is discussed. A real data application is also provided. 相似文献
10.
Several extensions of the popular E(s 2) criterion of Booth and Cox (1962) to multilevel supersaturated designs have been advanced in literature. These extensions are not unique due to different ways they measure overall nonorthogonality between all pairs of the columns of the model matrix. We exploit the connection of the E(s 2) criterion with A- and D-optimality that naturally lends itself to a generalized criterion for the multilevel situation in a unified way. The extensions provided in literature follow as special cases of the generalized criterion. A lower bound to the generalized criterion is derived for a wide class of designs, and a method of construction for the symmetrical case is discussed. 相似文献
11.
Suppose that some information is available for the particular factor. The experimenter may apply the technique of foldover to isolate the factor and the two-factor interactions involving this factor. In fact, under some situations this can be done by the method of semi-folding. We will discuss this property in detail in this article. Furthermore, we use the computer to search the corresponding optimal semi-folding design for the given 2 k?p designs that are tabulated in Chen et al. (1993) research. 相似文献
12.
The Significance Analysis of Microarrays (SAM; Tusher et al., 2001) method is widely used in analyzing gene expression data while controlling the FDR by using resampling-based procedure in the microarray setting. One of the main components of the SAM procedure is the adjustment of the test statistic. The introduction of the fudge factor to the test statistic aims at deflating the large value of test statistics due to the small standard error of gene-expression. Lin et al. (2008) pointed out that the fudge factor does not effectively improve the power and the control of the FDR as compared to the SAM procedure without the fudge factor in the presence of small variance genes. Motivated by the simulation results presented in Lin et al. (2008), in this article, we extend our study to compare several methods for choosing the fudge factor in the modified t-type test statistics and use simulation studies to investigate the power and the control of the FDR of the considered methods. 相似文献
13.
Luigi Greco 《统计学通讯:理论与方法》2013,42(5):1039-1048
In some real situations the population of interest is divided into two groups, of which one contains only a few units. In other cases, the population may be considered as subdivided into two group', for example, if only a few units display a value of the variable of interest which is highly different from zero, while all the other units show a value equal to or near zero. In both cases, inverse sampling is more efficient than classical fixed sample-size designs to obtain the parameter estimators for the whole population as well as for its groups (e.g., Salehi and Seber, 2004). In fact, in this design the procedure selection continues until a prefixed number of units with the characteristic of interest is sampled. Since it is not known a priori to which group the population units belong, the sample size is a random variable. Christman and Lan (2001) and Salehi and Seber (2001 2004) considered inverse sampling designs when all the population units have equal selection probabilities. In this article, we consider the general case in which the units may have unequal probabilities of being included in the sample. In fact, in many real situations different units may have different selection probabilities because of some inherent features of the sampling procedure, or in order to obtain better estimates. We derive unbiased estimators of the totals of the two groups, their variance and the corresponding unbiased variance estimators in inverse sampling with replacement. Finally, we derive similar results for more complex designs, where the selection procedure stops before observing the prefixed number of units from the rare group. 相似文献
14.
Magda (1980) and Hedayat (1981) first considered the construction of circular strongly balanced repeated measurements designs. Sen and Mukerjee (1987) and Roy (1988) considered the optimality and existence of circular strongly balanced repeated measurements designs based on the method of differences and Hamiltonian decomposition of lexicographic product of two graphs. In this article, we consider the construction of circular strongly balanced repeated measurements designs using the newly proposed method called cyclic shifts, and propose some new designs for p < v. 相似文献
15.
We consider non-parametric estimation of a continuous cdf of a random vector (X 1, X 2). With bivariate RC data, it is stated in van der Laan (1996, p. 59810, Ann. Statist.), Quale et al. (2006, JASA) etc. that “it is well known that the NPMLE for continuous data is inconsistent (Tsai et al. (1986)).” The claim is based on a result in Tsai et al. (1986, p.1352, Ann. Statist.) that if X 1 is right censored but not X 2, then common ways for defining one NPMLE lead to inconsistency. If X 1 is right censored and X 2 is type I right-censored (which includes the case in Tsai et al.), we present a consistent NPMLE. The result corrects a common misinterpretation of Tsai's example (Tsai et al., 1986, Ann. Statist.). 相似文献
16.
We present an efficient new approach to estimate global sensitivity indices based on ANOVA high-dimensional model representation. The method makes use of the properties of orthogonal arrays and extend the results of Wang et al. (2010) to a more general situation. A theorem is given to illustrate that both the non influential and significant sensitivity indices are well estimated. This new method offers higher sampling efficiency than those of Sobol and Saltelli. We test its performance on two different models, which are widely used in engineering and statistical areas. 相似文献
17.
《统计学通讯:理论与方法》2012,41(13-14):2445-2455
In this article, the problem of estimation of the individual weights of three objects using a chemical balance weighing design is considered. We use the criterion of D-optimality. We assume that the covariance matrix of errors is the matrix of first-order autoregressive process. Such problems were discussed in Li and Yang (2005) and also in Yeh and Lo Huang (2005). We present some results of D-optimal designs in certain class of designs with the design matrix X ∈ M n×3(±1) such that each column of matrix X has at least one 1 and one ?1. 相似文献
18.
Huang (2010) proposed an optional randomized response model using a linear combination scrambling which is a generalization of the multiplicative scrambling of Eichhorn and Hayre (1983) and the additive scrambling of Gupta et al. (2006, 2010). In this article, we discuss two main issues. (1) Can the Huang (2010) model be improved further by using a two-stage approach?; (2) Does the linear combination scrambling provide any benefit over the additive scrambling of Gupta et al. (2010)? We will note that the answer to the first question is “yes” but the answer to the second question is “no.” 相似文献
19.
Zheng Su 《统计学通讯:模拟与计算》2013,42(8):1163-1170
Johns (1988), Davison (1988), and Do and Hall (1991) used importance sampling for calculating bootstrap distributions of one-dimensional statistics. Realizing that their methods can not be extended easily to multi-dimensional statistics, Fuh and Hu (2004) proposed an exponential tilting formula for statistics of multi-dimension, which is optimal in the sense that the asymptotic variance is minimized for estimating tail probabilities of asymptotically normal statistics. For one-dimensional statistics, Hu and Su (2008) proposed a multi-step variance minimization approach that can be viewed as a generalization of the two-step variance minimization approach proposed by Do and Hall (1991). In this article, we generalize the approach of Hu and Su (2008) to multi-dimensional statistics, which applies to general statistics and does not resort to asymptotics. Empirical results on a real survival data set show that the proposed algorithm provides significant computational efficiency gains. 相似文献
20.
Hyun Jip Choi 《统计学通讯:模拟与计算》2013,42(7):1377-1384
We suggest a method for constructing a multidimensional distribution of correlated categorical data with fixed marginal distributions and specified degrees of association based on the log-linear models. A convex combination approach by Lee (1997) is applied to get a joint distribution with fixed Pearson chi-square coefficient. By using the suggested method, we can generate three-dimensional distributions which have a fixed association among three variables. Therefore, the suggested method could be extended to higher dimensions. 相似文献