首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
This article introduces a non parametric warping model for functional data. When the outcome of an experiment is a sample of curves, data can be seen as realizations of a stochastic process, which takes into account the variations between the different observed curves. The aim of this work is to define a mean pattern which represents the main behaviour of the set of all the realizations. So, we define the structural expectation of the underlying stochastic function. Then, we provide empirical estimators of this structural expectation and of each individual warping function. Consistency and asymptotic normality for such estimators are proved.  相似文献   

2.
Segmentation of the mean of heteroscedastic data via cross-validation   总被引:1,自引:0,他引:1  
This paper tackles the problem of detecting abrupt changes in the mean of a heteroscedastic signal by model selection, without knowledge on the variations of the noise. A new family of change-point detection procedures is proposed, showing that cross-validation methods can be successful in the heteroscedastic framework, whereas most existing procedures are not robust to heteroscedasticity. The robustness to heteroscedasticity of the proposed procedures is supported by an extensive simulation study, together with recent partial theoretical results. An application to Comparative Genomic Hybridization (CGH) data is provided, showing that robustness to heteroscedasticity can indeed be required for their analysis.  相似文献   

3.
We review some issues related to the implications of different missing data mechanisms on statistical inference for contingency tables and consider simulation studies to compare the results obtained under such models to those where the units with missing data are disregarded. We confirm that although, in general, analyses under the correct missing at random and missing completely at random models are more efficient even for small sample sizes, there are exceptions where they may not improve the results obtained by ignoring the partially classified data. We show that under the missing not at random (MNAR) model, estimates on the boundary of the parameter space as well as lack of identifiability of the parameters of saturated models may be associated with undesirable asymptotic properties of maximum likelihood estimators and likelihood ratio tests; even in standard cases the bias of the estimators may be low only for very large samples. We also show that the probability of a boundary solution obtained under the correct MNAR model may be large even for large samples and that, consequently, we may not always conclude that a MNAR model is misspecified because the estimate is on the boundary of the parameter space.  相似文献   

4.
Computational expressions for the exact CDF of Roy’s test statistic in MANOVA and the largest eigenvalue of a Wishart matrix are derived based upon their Pfaffian representations given in Gupta and Richards (SIAM J. Math. Anal. 16:852–858, 1985). These expressions allow computations to proceed until a prespecified degree of accuracy is achieved. For both distributions, convergence acceleration methods are used to compute CDF values which achieve reasonably fast run times for dimensions up to 50 and error degrees of freedom as large as 100. Software that implements these computations is described and has been made available on the Web.  相似文献   

5.
When the experimenter suspects that there might be a quadratic relation between the response variable and the explanatory parameters, a design with at least three points must be employed to establish and explore this relation (second-order design). Orthogonal arrays (OAs) with three levels are often used as second-order response surface designs. Generally, we assume that the data are independent observations; however, there are many situations where this assumption may not be sustainable. In this paper, we want to compare three-level OAs with 18, 27, and 36 runs under the presence of three specific forms of correlation in observations. The aim is to derive the best designs that can be efficiently used for response surface modeling.  相似文献   

6.
Abstract

The problem of obtaining the maximum probability 2 × c contingency table with fixed marginal sums, R  = (R 1R 2) and C  = (C 1, … , C c ), and row and column independence is equivalent to the problem of obtaining the maximum probability points (mode) of the multivariate hypergeometric distribution MH(R 1; C 1, … , C c ). The most simple and general method for these problems is Joe's (Joe, H. (1988 Joe, H. 1988. Extreme probabilities for contingency tables under row and column independence with application to Fisher's exact test. Commun. Statist. Theory Meth., 17(11): 36773685. [Taylor & Francis Online], [Web of Science ®] [Google Scholar]). Extreme probabilities for contingency tables under row and column independence with application to Fisher's exact test. Commun. Statist. Theory Meth. 17(11):3677–3685.) In this article we study a family of MH's in which a connection relationship is defined between its elements. Based on this family and on a characterization of the mode described in Requena and Martín (Requena, F., Martín, N. (2000 Requena, F. and Martín, N. 2000. Characterization of maximum probability points in the multivariate hypergeometric distribution. Statist. Probab. Lett., 50: 3947.  [Google Scholar]). Characterization of maximum probability points in the multivariate hypergeometric distribution. Statist. Probab. Lett. 50:39–47.), we develop a new method for the above problems, which is completely general, non recursive, very simple in practice and more efficient than the Joe's method. Also, under weak conditions (which almost always hold), the proposed method provides a simple explicit solution to these problems. In addition, the well-known expression for the mode of a hypergeometric distribution is just a particular case of the method in this article.  相似文献   

7.
The correct and efficient estimation of memory parameters in a stationary Gaussian processes is an important issue, since otherwise, forecasts based on the resulting time series would be misleading. On the other hand, if the memory parameters are suspected to fall in a smaller subspace through some hypothesis restrictions, it becomes a hard decision whether to use estimators based on the restricted spaces or to use unrestricted estimators over the full parameter space. In this article, we propose James-Stein-type estimators of the memory parameters of a stationary Gaussian times series process, which can efficiently incorporate the hypothetical restrictions. We show theoretically that the proposed estimators are more efficient than the usual unrestricted maximum likelihood estimators over the entire parameter space.  相似文献   

8.
This article considers the estimation of R = P(Y < X) when X and Y are distributed as two independent three-parameter generalized exponential (GE) random variables with different shape parameters but having the same location and scale parameters. A modified maximum likelihood method and a Bayesian technique are used to estimate R on the basis of independent complete samples. The Bayes estimator cannot be obtained in explicit form, and therefore it has been determined using an importance sampling procedure. An analysis of a real life data set is presented for illustrative purposes.  相似文献   

9.
In early January 2015, the multi-event national cyclo-cross bicycle races were set to take place in Austin, Texas. Cyclo-cross has a rich history in this country, and throughout the world, attracting huge crowds and competitors. Being primarily a winter sport, these athletes often compete in harsh conditions, which include rain, snow, mud, and revel in the excitement that comes with such elements. Unfortunately, the competition was postponed mid-event when a local arborist group protested to the parks department. The issue: there was too much mud, in an event where many spectators and racers alike hope for such conditions. For many competitors, the postponement generated additional expenses, such as flights, hotels, and car rentals. Although people on opposite sides of the debate may greatly disagree, we instead focus on the competitors themselves. We analyze the financial impact of the disagreement using a hierarchical Bayesian mixed model which accounts for heterogeneity within the costs endured by the event's participants.  相似文献   

10.
Dynamic programming (DP) is a fast, elegant method for solving many one-dimensional optimisation problems but, unfortunately, most problems in image analysis, such as restoration and warping, are two-dimensional. We consider three generalisations of DP. The first is iterated dynamic programming (IDP), where DP is used to recursively solve each of a sequence of one-dimensional problems in turn, to find a local optimum. A second algorithm is an empirical, stochastic optimiser, which is implemented by adding progressively less noise to IDP. The final approach replaces DP by a more computationally intensive Forward-Backward Gibbs Sampler, and uses a simulated annealing cooling schedule. Results are compared with existing pixel-by-pixel methods of iterated conditional modes (ICM) and simulated annealing in two applications: to restore a synthetic aperture radar (SAR) image, and to warp a pulsed-field electrophoresis gel into alignment with a reference image. We find that IDP and its stochastic variant outperform the remaining algorithms.  相似文献   

11.
This article develops a new and stable estimator for information matrix when the EM algorithm is used in maximum likelihood estimation. This estimator is constructed using the smoothed individual complete-data scores that are readily available from running the EM algorithm. The method works for dependent data sets and when the expectation step is an irregular function of the conditioning parameters. In comparison to the approach of Louis (J. R. Stat. Soc., Ser. B 44:226–233, 1982), this new estimator is more stable and easier to implement. Both real and simulated data are used to demonstrate the use of this new estimator.  相似文献   

12.
We consider the problem of estimating the stress-strength reliability when the available data is in the form of record values. The one parameter and two parameters exponential distribution are considered. In the case of two parameters exponential distributions we considered the case where the location parameter is common and the case where the scale parameter is common. The maximum likelihood estimators and the associated confidence intervals are derived.  相似文献   

13.
Two symmetrical fractional factorial designs are said to be combinatorially equivalent if one design can be obtained from another by reordering the runs, relabeling the factors and relabeling the levels of one or more factors. This article presents concepts of ordered distance frequency matrix, distance frequency vector, and reduced distance frequency vector for a design. Necessary conditions for two designs to be combinatorial equivalent based on these concepts are presented. A new algorithm based on the results obtained is proposed to check combinatorial non-equivalence of two factorial designs and several illustrating examples are provided.  相似文献   

14.
The 2 × 2 crossover is commonly used to establish average bioequivalence of two treatments. In practice, the sample size for this design is often calculated under a supposition that the true average bioavailabilities of the two treatments are almost identical. However, the average bioequivalence analysis that is subsequently carried out does not reflect this prior belief and this leads to a loss in efficiency. We propose an alternate average bioequivalence analysis that avoids this inefficiency. The validity and substantial power advantages of our proposed method are illustrated by simulations, and two numerical examples with real data are provided.  相似文献   

15.
It is customary to use two groups of indices to evaluate a diagnostic method with a binary outcome: validity indices with a standard rater (sensitivity, specificity, and positive or negative predictive values) and reliability indices (positive, negative and overall agreements) without a standard rater. However neither of these classic indices is chance-corrected, and this may distort the analysis of the problem (especially in comparative studies). One way of chance-correcting these indices is by using the Delta model (an alternative to the Kappa model), but this means having to use a computer program to work out the calculations. This paper gives an asymptotic version of the Delta model, thus allowing simple expressions to be obtained for the estimator of each of the above-mentioned chance-corrected indices (as well as for its standard error).  相似文献   

16.
Data envelopment analysis (DEA) and free disposal hull (FDH) estimators are widely used to estimate efficiency of production. Practitioners use DEA estimators far more frequently than FDH estimators, implicitly assuming that production sets are convex. Moreover, use of the constant returns to scale (CRS) version of the DEA estimator requires an assumption of CRS. Although bootstrap methods have been developed for making inference about the efficiencies of individual units, until now no methods exist for making consistent inference about differences in mean efficiency across groups of producers or for testing hypotheses about model structure such as returns to scale or convexity of the production set. We use central limit theorem results from our previous work to develop additional theoretical results permitting consistent tests of model structure and provide Monte Carlo evidence on the performance of the tests in terms of size and power. In addition, the variable returns to scale version of the DEA estimator is proved to attain the faster convergence rate of the CRS-DEA estimator under CRS. Using a sample of U.S. commercial banks, we test and reject convexity of the production set, calling into question results from numerous banking studies that have imposed convexity assumptions. Supplementary materials for this article are available online.  相似文献   

17.
Some alternative estimators to the maximum likelihood estimators of the two parameters of the Birnbaum–Saunders distribution are proposed. Most have high efficiencies as measured by root mean square error and are robust to departure from the model as well as to outliers. In addition, the proposed estimators are easy to compute. Both complete and right-censored data are discussed. Simulation studies are provided to compare the performance of the estimators.  相似文献   

18.
This article considers K pairs of incomplete correlated 2 × 2 tables in which the interesting measurement is the risk difference between marginal and conditional probabilities. A Wald-type statistic and a score-type statistic are presented to test the homogeneity hypothesis about risk differences across strata. Powers and sample size formulae based on the above two statistics are deduced. Figures about sample size against risk difference (or marginal probability) are given. A real example is used to illustrate the proposed methods.  相似文献   

19.
We consider the fitting of a Bayesian model to grouped data in which observations are assumed normally distributed around group means that are themselves normally distributed, and consider several alternatives for accommodating the possibility of heteroscedasticity within the data. We consider the case where the underlying distribution of the variances is unknown, and investigate several candidate prior distributions for those variances. In each case, the parameters of the candidate priors (the hyperparameters) are themselves given uninformative priors (hyperpriors). The most mathematically convenient model for the group variances is to assign them inverse gamma distributed priors, the inverse gamma distribution being the conjugate prior distribution for the unknown variance of a normal population. We demonstrate that for a wide class of underlying distributions of the group variances, a model that assigns the variances an inverse gamma-distributed prior displays favorable goodness-of-fit properties relative to other candidate priors, and hence may be used as standard for modeling such data. This allows us to take advantage of the elegant mathematical property of prior conjugacy in a wide variety of contexts without compromising model fitness. We test our findings on nine real world publicly available datasets from different domains, and on a wide range of artificially generated datasets.  相似文献   

20.
Electronic resources are becoming the backbone of many library collections. Electronic journals in particular have become the default format for most periodical literature in all fields of scholarship. The mission of librarians, the publishing industry, and nonprofit organizations is to ensure that electronic scholarship remains accessible to future generations in an affordable and sustainable manner. The focus for the first edition of “In Layman's Terms” is on LOCKSS, CLOCKSS, CHORUS, Portico, and the Keepers Registry, five initiatives dedicated to the preservation and to the perpetual access of electronic scholarship. It includes a brief history of each initiative. It looks at their missions and their different roles in preserving and providing access to electronic resources and the benefits and costs in joining them and being involved with them. It identifies the relationships among these initiatives.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号