首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The United States is experiencing a major public health problem relating to increasing levels of excess body fat. This paper is about the relationship in the United States between trends in the distribution of body mass index (BMI), including trends in overweight and obesity, and demographic change. We provide estimates of the counterfactual distribution of BMI that would have been observed in 2003–2008 had demographics remained fixed at 1980 values, roughly the beginning of the period of increasing overweight and obesity. We find that changes in demographics are partly responsible for the changes in the population distribution of BMI and are capable of explaining about 8.6% of the increase in the combined rate of overweight and obesity among women and about 7.2% of the increase among men. We also use demographic projections to predict a BMI distribution and corresponding rates of overweight and obesity for 2050.  相似文献   

2.
Despite tremendous effort on different designs with cross-sectional data, little research has been conducted for sample size calculation and power analyses under repeated measures design. In addition to time-averaged difference, changes in mean response over time (CIMROT) is the primary interest in repeated measures analysis. We generalized sample size calculation and power analysis equations for CIMROT to allow unequal sample size between groups for both continuous and binary measures, through simulation, evaluated the performance of proposed methods, and compared our approach to that of a two-stage model formulization. We also created a software procedure to implement the proposed methods.  相似文献   

3.
We tackle an important although rarely addressed question of accounting for a variety of asymmetries frequently observed in stochastic temporal/spatial records. First, we review some measures intending to capture such asymmetries that have been introduced on various occasions in the past and then propose a family of measures that is motivated by Rice's formula for crossing level distributions of the slope. We utilize those asymmetry measures to demonstrate how a class of second‐order models built on the skewed Laplace distributions can account for sample path asymmetries. It is shown that these models are capable of mimicking not only distributional skewness but also more complex geometrical asymmetries in the sample path such as tilting, front‐back slope asymmetry and time irreversibility. Simple moment‐based estimation techniques are briefly discussed to allow direct application to modelling and fitting actual records.  相似文献   

4.
ABSTRACT

Environmental data is typically indexed in space and time. This work deals with modelling spatio-temporal air quality data, when multiple measurements are available for each space-time point. Typically this situation arises when different measurements referring to several response variables are observed in each space-time point, for example, different pollutants or size resolved data on particular matter. Nonetheless, such a kind of data also arises when using a mobile monitoring station moving along a path for a certain period of time. In this case, each spatio-temporal point has a number of measurements referring to the response variable observed several times over different locations in a close neighbourhood of the space-time point. We deal with this type of data within a hierarchical Bayesian framework, in which observed measurements are modelled in the first stage of the hierarchy, while the unobserved spatio-temporal process is considered in the following stages. The final model is very flexible and includes autoregressive terms in time, different structures for the variance-covariance matrix of the errors, and can manage covariates available at different space-time resolutions. This approach is motivated by the availability of data on urban pollution dynamics: fast measures of gases and size resolved particulate matter have been collected using an Optical Particle Counter located on a cabin of a public conveyance that moves on a monorail on a line transect of a town. Urban microclimate information is also available and included in the model. Simulation studies are conducted to evaluate the performance of the proposed model over existing alternatives that do not model data over the first stage of the hierarchy.  相似文献   

5.
We consider cross-sectional aggregation of time series with long-range dependence. This question arises for instance from the statistical analysis of networks where aggregation is defined via routing matrices. Asymptotically, aggregation turns out to increase dependence substantially, transforming a hyperbolic decay of autocorrelations to a slowly varying rate. This effect has direct consequences for statistical inference. For instance, unusually slow rates of convergence for nonparametric trend estimators and nonstandard formulas for optimal bandwidths are obtained. The situation changes, when time-dependent aggregation is applied. Suitably chosen time-dependent aggregation schemes can preserve a hyperbolic rate or even eliminate autocorrelations completely.  相似文献   

6.
We formulate a prior distribution for the energy function of stationary binary Markov random fields (MRFs) defined on a rectangular lattice. In the prior we assign distributions to all parts of the energy function. In particular we define priors for the neighbourhood structure of the MRF, what interactions to include in the model, and for potential values. We define a reversible jump Markov chain Monte Carlo (RJMCMC) procedure to simulate from the corresponding posterior distribution when conditioned to an observed scene. Thereby we are able to learn both the neighbourhood structure and the parametric form of the MRF from the observed scene. We circumvent evaluations of the intractable normalising constant of the MRF when running the RJMCMC algorithm by adopting a previously defined approximate auxiliary variable algorithm. We demonstrate the usefulness of our prior in two simulation examples and one real data example.  相似文献   

7.
We consider the problem of minimax-variance, robust estimation of a location parameter, through the use of L- and R-estimators. We derive an easily checked necessary condition for L-estimation to be minimax, and a related sufficient condition for R-estimation to be minimax. Those cases in the literature in which L-estimation is known not to be minimax, and those in which R-estimation is minimax, are derived as consequences of these conditions. New classes of examples are given in each case. As well, we answer a question of Scholz (1974), who showed essentially that the asymptotic variance of an R-estimator never exceeds that of an L-estimator, if both are efficient at the same strongly unimodal distribution. Scholz raised the question of whether or not the assumption of strong unimodality could be dropped. We answer this question in the negative, theoretically and by examples. In the examples, the minimax property fails both for L-estimation and for R-estimation, but the variance of the L-estimator, as the distribution of the observation varies over the given neighbourhood, remains unbounded. That of the R-estimator is unbounded.  相似文献   

8.
We propose a new approach for outlier detection, based on a ranking measure that focuses on the question of whether a point is ‘central’ for its nearest neighbours. Using our notations, a low cumulative rank implies that the point is central. For instance, a point centrally located in a cluster has a relatively low cumulative sum of ranks because it is among the nearest neighbours of its own nearest neighbours, but a point at the periphery of a cluster has a high cumulative sum of ranks because its nearest neighbours are closer to each other than the point. Use of ranks eliminates the problem of density calculation in the neighbourhood of the point and this improves the performance. Our method performs better than several density-based methods on some synthetic data sets as well as on some real data sets.  相似文献   

9.
Summary. Nearest neighbour algorithms are among the most popular methods used in statistical pattern recognition. The models are conceptually simple and empirical studies have shown that their performance is highly competitive against other techniques. However, the lack of a formal framework for choosing the size of the neighbourhood k is problematic. Furthermore, the method can only make discrete predictions by reporting the relative frequency of the classes in the neighbourhood of the prediction point. We present a probabilistic framework for the k -nearest-neighbour method that largely overcomes these difficulties. Uncertainty is accommodated via a prior distribution on k as well as in the strength of the interaction between neighbours. These prior distributions propagate uncertainty through to proper probabilistic predictions that have continuous support on (0, 1). The method makes no assumptions about the distribution of the predictor variables. The method is also fully automatic with no user-set parameters and empirically it proves to be highly accurate on many bench-mark data sets.  相似文献   

10.
Missing data are often problematic in social network analysis since what is missing may potentially alter the conclusions about what we have observed as tie-variables need to be interpreted in relation to their local neighbourhood and the global structure. Some ad hoc methods for dealing with missing data in social networks have been proposed but here we consider a model-based approach. We discuss various aspects of fitting exponential family random graph (or p-star) models (ERGMs) to networks with missing data and present a Bayesian data augmentation algorithm for the purpose of estimation. This involves drawing from the full conditional posterior distribution of the parameters, something which is made possible by recently developed algorithms. With ERGMs already having complicated interdependencies, it is particularly important to provide inference that adequately describes the uncertainty, something that the Bayesian approach provides. To the extent that we wish to explore the missing parts of the network, the posterior predictive distributions, immediately available at the termination of the algorithm, are at our disposal, which allows us to explore the distribution of what is missing unconditionally on any particular parameter values. Some important features of treating missing data and of the implementation of the algorithm are illustrated using a well-known collaboration network and a variety of missing data scenarios.  相似文献   

11.
Over the past decades, various principles for causal effect estimation have been proposed, all differing in terms of how they adjust for measured confounders: either via traditional regression adjustment, by adjusting for the expected exposure given those confounders (e.g., the propensity score), or by inversely weighting each subject's data by the likelihood of the observed exposure, given those confounders. When the exposure is measured with error, this raises the question whether these different estimation strategies might be differently affected and whether one of them is to be preferred for that reason. In this article, we investigate this by comparing inverse probability of treatment weighted (IPTW) estimators and doubly robust estimators for the exposure effect in linear marginal structural mean models (MSM) with G-estimators, propensity score (PS) adjusted estimators and ordinary least squares (OLS) estimators for the exposure effect in linear regression models. We find analytically that these estimators are equally affected when exposure misclassification is independent of the confounders, but not otherwise. Simulation studies reveal similar results for time-varying exposures and when the model of interest includes a logistic link.  相似文献   

12.
Using kernel density estimation, this paper describes the real income distribution and how it evolved over time in Italy. Data are cross-sectional samples from the population of Italian households during the period 1987–1998. A non parametric test is applied to asses whether the observed changes in the distribution are statistically significant, while the presence of more than one mode in the distributions is investigated by a bootstrap test. Empirical results show that the Italian income distribution significantly changed over time, accompanied by a decreasing inequality pattern until 1991 and a subsequent rise. No marked income gains were perceived, while the real “losers” of the decade seem to be households in the middle-upper income range. Supported by the MURST project 98-13-45. We would like to thank Nicholas Longford for his precious support and encouragement, two anonymous referees, the participants of the seminar at CEPS/INSTEAD in Luxembarg, and of the 40 th SIS Conference in Florence, for helpful comments and suggestions. The usual disclaimers apply.  相似文献   

13.
Much effort in life-history theory has been addressed to the dependence of life-history traits on age, especially the phenomenon of senescence and its evolution. Although senescent declines in survival are well documented in humans and in domestic and laboratory animals, evidence for their occurrence and importance in wild animal species remains limited and equivocal. Several recent papers have suggested that methodological issues may contribute to this problem, and have encouraged investigators to improve sampling designs and to analyse their data using recently developed approaches to modelling of capture-mark-recapture data. Here we report on a three-year, two-site, mark-recapture study of known-aged common terns (Sterna hirundo) in the north-eastern USA. The study was nested within a long-term ecological study in which large numbers of chicks had been banded in each year for > 25 years. We used a range of models to test the hypothesis of an influence of age on survival probability. We also tested for a possible influence of sex on survival. The cross-sectional design of the study (one year's parameter estimates) avoided the possible confounding of effects of age and time. The study was conducted at a time when one of the study sites was being colonized and numbers were increasing rapidly. We detected two-way movements between the sites and estimated movement probabilities in the year for which they could be modelled. We also obtained limited data on emigration from our study area to more distant sites. We found no evidence that survival depended on either sex or age, except that survival was lower among the youngest birds (ages 2-3 years). Despite the large number of birds included in the study (1599 known-aged birds, 2367 total), confidence limits on estimates of survival probability were wide, especially for the oldest age-classes, so that a slight decline in survival late in life could not have been detected. In addition, the cross-sectional design of this study meant that a decline in survival probability within individuals (actuarial senescence) could have been masked by heterogeneity in survival probability among individuals (mortality selection). This emphasizes the need for the development of modelling tools permitting separation of these two phenomena, valid under field conditions in which the recapture probabilities are less than one.  相似文献   

14.
Repeated neuropsychological measurements, such as mini-mental state examination (MMSE) scores, are frequently used in Alzheimer’s disease (AD) research to study change in cognitive function of AD patients. A question of interest among dementia researchers is whether some AD patients exhibit transient “plateaus” of cognitive function in the course of the disease. We consider a statistical approach to this question, based on irregularly spaced repeated MMSE scores. We propose an algorithm that formalizes the measurement of an apparent cognitive plateau, and a procedure to evaluate the evidence of plateaus in AD using this algorithm based on applying the algorithm to the observed data and to data sets simulated from a linear mixed model. We apply these methods to repeated MMSE data from the Michigan Alzheimer’s Disease Research Center, finding a high rate of apparent plateaus and also a high rate of false discovery. Simulation studies are also conducted to assess the performance of the algorithm. In general, the false discovery rate of the algorithm is high unless the rate of decline is high compared with the measurement error of the cognitive test. It is argued that the results are not a problem of the specific algorithm chosen, but reflect a lack of information concerning the presence of plateaus in the data.  相似文献   

15.
In a likelihood-ratio test for a two-component Normal location mixture, the natural parametrisation degenerates to non-uniqueness under the null hypothesis. One consequence of this ambiguity is that the limiting distribution of the likelihood-ratio statistic is quite irregular, being of extreme-value type rather than chi-squared. Another irregular feature is that the likelihood-ratio statistic diverges to infinity, and so limit theory is nonstandard in this respect as well. These results, in a form applying directly to the likelihood-ratio statistic rather than to an approximating stochastic process, have recently been established by Liu and Shao (2004). While they address only properties under the null hypothesis, they hint that the power of the likelihood-ratio test may be less than in more conventional settings. In this paper we show that this is indeed the case. Using a system of local alternative hypotheses we quantify the extent to which power is reduced. We show that, in a large class of circumstances, the reduction in power can be appreciated in terms of inflation (by a log–log factor) of the displacement of the closest local alternative that can just be distinguished from the null hypothesis. However, in important respects the properties of power under local alternatives are significantly more complex than this, and exhibit two types of singularity. In particular, in two quite different respects, small changes in the local alternative, in the neighbourhood of a threshold, can dramatically alter power.  相似文献   

16.
We consider minimax-bias M-estimation of a location parameter in a Kolmogorov neighbourhood K() of a normal distribution. The maximum asymptotic bias of M-estimators for the Kolmogorov normal neighbourhood is derived, and its relation with the gross-error sensitivity of the estimator at the nominal model (the Gaussian case) is found. In addition, efficient bias-robust M-estimators Ti are constructed. Numerical results are also obtained to show the percentage of increase in maximum asymptotic bias and the efficiency we can achieve for some well-known -functions.  相似文献   

17.
It i s well known that even if the sample observations are correlated and not normal, the sample mean is normal in 1arge samples. But how large is large? This question i s investigated in this paper. In particular , the relation between the rate of convergence and the correlation property of the observations i s explored. It i s observed that the correlation, in general, retards the rate of convergence.  相似文献   

18.
In the first part of this article, we briefly review the history of seasonal adjustment and statistical time series analysis in order to understand why seasonal adjustment methods have evolved into their present form. This review provides insight into some of the problems that must be addressed by seasonal adjustment procedures and points out that advances in modem time series analysis raise the question of whether seasonal adjustment should be performed at all. This in turn leads to a discussion in the second part of issues involved in seasonal adjustment. We state our opinions about the issues raised and review some of the work of other authors. First, we comment on reasons that have been given for doing seasonal adjustment and suggest a new possible justification. We then emphasize the need to define precisely the seasonal and nonseasonal components and offer our definitions. Finally, we discuss criteria for evaluating seasonal adjustments. We contend that proposed criteria based on empirical comparisons of estimated components are of little value and suggest that seasonal adjustment methods should be evaluated based on whether they are consistent with the information in the observed data. This idea is illustrated with an example.  相似文献   

19.
In the first part of this article, we briefly review the history of seasonal adjustment and statistical time series analysis in order to understand why seasonal adjustment methods have evolved into their present form. This review provides insight into some of the problems that must be addressed by seasonal adjustment procedures and points out that advances in modern time series analysis raise the question of whether seasonal adjustment should be performed at all. This in turn leads to a discussion in the second part of issues invloved in seasonal adjustment. We state our opinions about the issues raised and renew some of the work of our authors. First, we comment on reasons that have been given for doing seasonal adjustment and suggest a new possible justification. We then emphasize the need to define precisely the seasonal and nonseasonal components and offer our definitions. Finally, we discuss our criteria for evaluating seasonal adjustments. We contend that proposed criteria based on empirical comparisons of estimated components are of little value and suggest that seasonal adjustment methods should be evaluated based on whether they are consistent with the information in the observed data. This idea is illustrated with an example.  相似文献   

20.
ABSTRACT

Panel datasets have been increasingly used in economics to analyze complex economic phenomena. Panel data is a two-dimensional array that combines cross-sectional and time series data. Through constructing a panel data matrix, the clustering method is applied to panel data analysis. This method solves the heterogeneity question of the dependent variable, which belongs to panel data, before the analysis. Clustering is a widely used statistical tool in determining subsets in a given dataset. In this article, we present that the mixed panel dataset is clustered by agglomerative hierarchical algorithms based on Gower's distance and by k-prototypes. The performance of these algorithms has been studied on panel data with mixed numerical and categorical features. The effectiveness of these algorithms is compared by using cluster accuracy. An experimental analysis is illustrated on a real dataset using Stata and R package software.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号