首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Classical regression analysis is usually performed in two steps. In the first step, an appropriate model is identified to describe the data generating process and in the second step, statistical inference is performed in the identified model. An intuitively appealing approach to the design of experiment for these different purposes are sequential strategies, which use parts of the sample for model identification and adapt the design according to the outcome of the identification steps. In this article, we investigate the finite sample properties of two sequential design strategies, which were recently proposed in the literature. A detailed comparison of sequential designs for model discrimination in several regression models is given by means of a simulation study. Some non-sequential designs are also included in the study.  相似文献   

2.
This paper deals with the problem of local sensitivity analysis in ordered parameter models. In addition to order restrictions, some constraints imposed on the parameters by the model and/or the data are considered. Measures for assessing how much a change in the data modifies the results and conclusions of a statistical analysis of these models are presented. The sensitivity measures are derived using recent results in mathematical programming. The estimation problem is formulated as a primal nonlinear programming problem, and the sensitivities of the parameter estimates as well as the objective function sensitivities with respect to data are obtained. They are very effective in revealing the influential observations in this type of models and in evaluating the changes due to changes in data values. The methods are illustrated by their application to a wide variety of examples of order-restricted models including ordered exponential family parameters, ordered multinomial parameters, ordered linear model parameters, ordered and data constrained parameters, and ordered functions of parameters.  相似文献   

3.
The evaluation of hazards from complex, large scale, technologically advanced systems often requires the construction of computer implemented mathematical models. These models are used to evaluate the safety of the systems and to evaluate the consequences of modifications to the systems. These evaluations, however, are normally surrounded by significant uncertainties related to the uncertainty inherent in natural phenomena such as the weather and those related to uncertainties in the parameters and models used in the evaluation.

Another use of these models is to evaluate strategies for improving information used in the modeling process itself. While sensitivity analysis is useful in defining variables in the model that are important, uncertainty analysis provides a tool for assessing the importance of uncertainty about these variables. A third complementary technique, is decision analysis. It provides a methodology for explicitly evaluating and ranking potential improvements to the model. Its use in the development of information gathering strategies for a nuclear waste repository are discussed in this paper.  相似文献   

4.
《统计学通讯:理论与方法》2012,41(13-14):2367-2385
Orthogonal regression is a proper tool to analyze relations between two variables when three-part compositional data, i.e., three-part observations carrying relative information (like proportions or percentages), are under examination. When linear statistical models with type-II constraints (constraints involving other parameters besides the ones of the unknown model) are employed for estimating the parameters of the regression line, approximate variances and covariances of the estimated line coefficients can be determined. Moreover, the additional assumption of normality enables to construct confidence domains and perform hypotheses testing. The theoretical results are applied to a real-world example.  相似文献   

5.
Calculations of local influence curvatures and leverage have been well developed when the parameters are unrestricted. In this article, we discuss the assessment of local influence and leverage under linear equality parameter constraints with extensions to inequality constraints. Using a penalized quadratic function we express the normal curvature of local influence for arbitrary perturbation schemes and the generalized leverage matrix in interpretable forms, which depend on restricted and unrestricted components. The results are quite general and can be applied in various statistical models. In particular, we derive the normal curvature under three useful perturbation schemes for generalized linear models. Four illustrative examples are analyzed by the methodology developed in the article.  相似文献   

6.
The growth curve model introduced by potthoff and Roy 1964 is a general statistical model which includes as special cases regression models and both univariate and multivariate analysis of variance models. The methods currently available for estimating the parameters of this model assume an underlying multivariate normal distribution of errors. In this paper, we discuss tw robst estimators of the growth curve loction and scatter parameters based upon M-estimation techniques and the work done by maronna 1976. The asymptotic distribution of these robust estimators are discussed and a numerical example given.  相似文献   

7.
The general linear test approach for testing hypotheses concerning parameters of a full-rank multivariate regression model is often not sufficiently discussed in applied multivariate analysis textbooks. This note makes mention of this approach and suggests that major emphasis of this topic in an applied multivariate analysis course provides students with a unifying structure for the analysis of statistical models.  相似文献   

8.
Summary Nonsymmetric correspondence analysis is a model meant for the analysis of the dependence in a two-way continengy table, and is an alternative to correspondence analysis. Correspondence analysis is based on the decomposition of Pearson's Ф2-index Nonsymmetric correspondence analysis is based on the decomposition of Goodman-Kruskal's τ-index for predicatablity. In this paper, we approach nonsymmetric correspondence analysis as a statistical model based on a probability distribution. We provide algorithms for the maximum likelihood and the least-squares estimation with linear constraints upon model parameters. The nonsymmetric correspondence analysis model has many properties that can be useful for prediction analysis in contingency tables. Predictability measures are introduced to identify the categories of the response variable that can be best predicted, as well as the categories of the explanatory variable having the highest predictability power. We describe the interpretation of model parameters in two examples. In the end, we discuss the relations of nonsymmetric correspondence analysis with other reduced-rank models.  相似文献   

9.
Motivated by the Singapore Longitudinal Aging Study (SLAS), we propose a Bayesian approach for the estimation of semiparametric varying-coefficient models for longitudinal continuous and cross-sectional binary responses. These models have proved to be more flexible than simple parametric regression models. Our development is a new contribution towards their Bayesian solution, which eases computational complexity. We also consider adapting all kinds of familiar statistical strategies to address the missing data issue in the SLAS. Our simulation results indicate that a Bayesian imputation (BI) approach performs better than complete-case (CC) and available-case (AC) approaches, especially under small sample designs, and may provide more useful results in practice. In the real data analysis for the SLAS, the results for longitudinal outcomes from BI are similar to AC analysis, differing from those with CC analysis.  相似文献   

10.
In the framework of model-based cluster analysis, finite mixtures of Gaussian components represent an important class of statistical models widely employed for dealing with quantitative variables. Within this class, we propose novel models in which constraints on the component-specific variance matrices allow us to define Gaussian parsimonious clustering models. Specifically, the proposed models are obtained by assuming that the variables can be partitioned into groups resulting to be conditionally independent within components, thus producing component-specific variance matrices with a block diagonal structure. This approach allows us to extend the methods for model-based cluster analysis and to make them more flexible and versatile. In this paper, Gaussian mixture models are studied under the above mentioned assumption. Identifiability conditions are proved and the model parameters are estimated through the maximum likelihood method by using the Expectation-Maximization algorithm. The Bayesian information criterion is proposed for selecting the partition of the variables into conditionally independent groups. The consistency of the use of this criterion is proved under regularity conditions. In order to examine and compare models with different partitions of the set of variables a hierarchical algorithm is suggested. A wide class of parsimonious Gaussian models is also presented by parameterizing the component-variance matrices according to their spectral decomposition. The effectiveness and usefulness of the proposed methodology are illustrated with two examples based on real datasets.  相似文献   

11.
Generalized linear mixed models (GLMMs) are often used for analyzing cluster correlated data, including longitudinal data and repeated measurements. Full unrestricted maximum likelihood (ML) approaches for inference on both fixed‐and random‐effects parameters in GLMMs have been extensively studied in the literature. However, parameter orderings or constraints may occur naturally in practice, and in such cases, the efficiency of a statistical method is improved by incorporating the parameter constraints into the ML estimation and hypothesis testing. In this paper, inference for GLMMs under linear inequality constraints is considered. The asymptotic properties of the constrained ML estimators and constrained likelihood ratio tests for GLMMs have been studied. Simulations investigated the empirical properties of the constrained ML estimators, compared to their unrestricted counterparts. An application to a recent survey on Canadian youth smoking patterns is also presented. As these survey data exhibit natural parameter orderings, a constrained GLMM has been considered for data analysis. The Canadian Journal of Statistics 40: 243–258; 2012 © 2012 Crown in the right of Canada  相似文献   

12.
On identifiability of parametric statistical models   总被引:1,自引:0,他引:1  
Summary This is a review article on statistical identifiability. Besides the definition of the main concepts, we deal with several questions relevant to the statistician: parallelism between parametric identifiability and sample sufficiency; relationship of identifiability with measures of sample information and with the inferential concept of estimability; several strategies of making inferences in unidentifiable models with emphasis on the distinct behaviour of the classical and Bayesian approaches. The concepts, ideas and methods discussed are illustrated with simple examples of statistical models. Centro de Análise e Processamento de Sinais da UTL  相似文献   

13.
Measuring the efficiency of public services: the limits of analysis   总被引:2,自引:0,他引:2  
Summary.  Policy makers are increasingly seeking to develop overall measures of the effi-ciency of public service organizations. For that, the use of 'off-the-shelf' statistical tools such as data envelopment analysis and stochastic frontier analysis have been advocated as tools to measure organizational efficiency. The analytical sophistication of such methods has reached an advanced stage of development. We discuss the context within which such models are deployed, their underlying assumptions and their usefulness for a regulator of public services. Four specific model building issues are discussed: the weights that are attached to public service outputs; the specification of the statistical model; the treatment of environmental influences on performance; the treatment of dynamic effects. The paper concludes with recommendations for policy makers and researchers on the development and use of efficiency measurement techniques.  相似文献   

14.
Log Gaussian Cox processes as introduced in Moller et al. (1998) are extended to space-time models called log Gaussian Cox birth processes. These processes allow modelling of spatial and temporal heterogeneity in time series of increasing point processes consisting of different types of points. The models are shown to be easy to analyse yet flexible enough for a detailed statistical analysis of a particular agricultural experiment concerning the development of two weed species on an organic barley field. Particularly, the aspects of estimation, model validation and intensity surface prediction are discussed.  相似文献   

15.
Mixed effect models, which contain both fixed effects and random effects, are frequently used in dealing with correlated data arising from repeated measurements (made on the same statistical units). In mixed effect models, the distributions of the random effects need to be specified and they are often assumed to be normal. The analysis of correlated data from repeated measurements can also be done with GEE by assuming any type of correlation as initial input. Both mixed effect models and GEE are approaches requiring distribution specifications (likelihood, score function). In this article, we consider a distribution-free least square approach under a general setting with missing value allowed. This approach does not require the specifications of the distributions and initial correlation input. Consistency and asymptotic normality of the estimation are discussed.  相似文献   

16.
We develope an M-estimator for partially linear models in which the nonparametric component is subject to various shape constraints. Bernstein polynomials are used to approximate the unknown nonparametric function, and shape constraints are imposed on the coefficients. Asymptotic normality of regression parameters and the optimal rate of convergence of the shape-restricted nonparametric function estimator are established under very mild conditions. Some simulation studies and a real data analysis are conducted to evaluate the finite sample performance of the proposed method.  相似文献   

17.
This article performs a sensitivity analyses of the synthetic T2 chart using fractional factorial design, which integrates the interaction effects. We are interested in the effects of the input parameters on the optimal cost, chart's parameters, and average run lengths. We also look at the input parameters responsible for the increase in cost and improvement in statistical performance under statistical constraints, and investigate how the input parameters influence the binding effect of the statistical constraints. The sensitivity analyses of the synthetic T2 chart are compared with that of the Hotelling's T2 chart, and parameters responsible for the cost advantage of the synthetic T2 chart are identified.  相似文献   

18.
This paper presents information theory and statistical analysis as two fundamental conceptual tools for data mining. A data mining technique based on these two conceptual tools consists of three steps. The first step is a statistical approach for discovering data patterns. The second step is an information-theoretic approach for identifying models that encapsulate the statistical behavior of the data patterns. The last step is a probabilistic approach for pattern-based inference that uncovers unknown significant event patterns.  相似文献   

19.
Missing data are common in many experiments, including surveys, clinical trials, epidemiological studies, and environmental studies. Unconstrained likelihood inferences for generalized linear models (GLMs) with nonignorable missing covariates have been studied extensively in the literature. However, parameter orderings or constraints may occur naturally in practice, and thus the efficiency of a statistical method may be improved by incorporating parameter constraints into the likelihood function. In this paper, we consider constrained inference for analysing GLMs with nonignorable missing covariates under linear inequality constraints on the model parameters. Specifically, constrained maximum likelihood (ML) estimation is based on the gradient projection expectation maximization approach. Further, we investigate the asymptotic null distribution of the constrained likelihood ratio test (LRT). Simulations study the empirical properties of the constrained ML estimators and LRTs, which demonstrate improved precision of these constrained techniques. An application to contaminant levels in an environmental study is also presented.  相似文献   

20.
A new method for analyzing high-dimensional categorical data, Linear Latent Structure (LLS) analysis, is presented. LLS models belong to the family of latent structure models, which are mixture distribution models constrained to satisfy the local independence assumption. LLS analysis explicitly considers a family of mixed distributions as a linear space, and LLS models are obtained by imposing linear constraints on the mixing distribution.LLS models are identifiable under modest conditions and are consistently estimable. A remarkable feature of LLS analysis is the existence of a high-performance numerical algorithm, which reduces parameter estimation to a sequence of linear algebra problems. Simulation experiments with a prototype of the algorithm demonstrated a good quality of restoration of model parameters.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号