首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
Individual-level models (ILMs) for infectious disease can be used to model disease spread between individuals while taking into account important covariates. One important covariate in determining the risk of infection transfer can be spatial location. At the same time, measurement error is a concern in many areas of statistical analysis, and infectious disease modelling is no exception. In this paper, we are concerned with the issue of measurement error in the recorded location of individuals when using a simple spatial ILM to model the spread of disease within a population. An ILM that incorporates spatial location random effects is introduced within a hierarchical Bayesian framework. This model is tested upon both simulated data and data from the UK 2001 foot-and-mouth disease epidemic. The ability of the model to successfully identify both the spatial infection kernel and the basic reproduction number (R 0) of the disease is tested.  相似文献   

2.
Population-level proportions of individuals that fall at different points in the spectrum [of disease severity], from asymptomatic infection to severe disease, are often difficult to observe, but estimating these quantities can provide information about the nature and severity of the disease in a particular population. Logistic and multinomial regression techniques are often applied to infectious disease modeling of large populations and are suited to identifying variables associated with a particular disease or disease state. However, they are less appropriate for estimating infection state prevalence over time because they do not naturally accommodate known disease dynamics like duration of time an individual is infectious, heterogeneity in the risk of acquiring infection, and patterns of seasonality. We propose a Bayesian compartmental model to estimate latent infection state prevalence over time that easily incorporates known disease dynamics. We demonstrate how and why a stochastic compartmental model is a better approach for determining infection state proportions than multinomial regression is by using a novel method for estimating Bayes factors for models with high-dimensional parameter spaces. We provide an example using visceral leishmaniasis in Brazil and present an empirically-adjusted reproductive number for the infection.  相似文献   

3.
We present a sequential Monte Carlo algorithm for Markov chain trajectories with proposals constructed in reverse time, which is advantageous when paths are conditioned to end in a rare set. The reverse time proposal distribution is constructed by approximating the ratio of Green’s functions in Nagasawa’s formula. Conditioning arguments can be used to interpret these ratios as low-dimensional conditional sampling distributions of some coordinates of the process given the others. Hence, the difficulty in designing SMC proposals in high dimension is greatly reduced. Empirically, our method outperforms an adaptive multilevel splitting algorithm in three examples: estimating an overflow probability in a queueing model, the probability that a diffusion follows a narrowing corridor, and the initial location of an infection in an epidemic model on a network.  相似文献   

4.
Generalized additive models for location, scale and shape   总被引:10,自引:0,他引:10  
Summary.  A general class of statistical models for a univariate response variable is presented which we call the generalized additive model for location, scale and shape (GAMLSS). The model assumes independent observations of the response variable y given the parameters, the explanatory variables and the values of the random effects. The distribution for the response variable in the GAMLSS can be selected from a very general family of distributions including highly skew or kurtotic continuous and discrete distributions. The systematic part of the model is expanded to allow modelling not only of the mean (or location) but also of the other parameters of the distribution of y , as parametric and/or additive nonparametric (smooth) functions of explanatory variables and/or random-effects terms. Maximum (penalized) likelihood estimation is used to fit the (non)parametric models. A Newton–Raphson or Fisher scoring algorithm is used to maximize the (penalized) likelihood. The additive terms in the model are fitted by using a backfitting algorithm. Censored data are easily incorporated into the framework. Five data sets from different fields of application are analysed to emphasize the generality of the GAMLSS class of models.  相似文献   

5.
In this paper, we propose a hidden Markov model for the analysis of the time series of bivariate circular observations, by assuming that the data are sampled from bivariate circular densities, whose parameters are driven by the evolution of a latent Markov chain. The model segments the data by accounting for redundancies due to correlations along time and across variables. A computationally feasible expectation maximization (EM) algorithm is provided for the maximum likelihood estimation of the model from incomplete data, by treating the missing values and the states of the latent chain as two different sources of incomplete information. Importance-sampling methods facilitate the computation of bootstrap standard errors of the estimates. The methodology is illustrated on a bivariate time series of wind and wave directions and compared with popular segmentation models for bivariate circular data, which ignore correlations across variables and/or along time.  相似文献   

6.
As new technologies permit the generation of hitherto unprecedented volumes of data (e.g. genome-wide association study data), researchers struggle to keep up with the added complexity and time commitment required for its analysis. For this reason, model selection commonly relies on machine learning and data-reduction techniques, which tend to afford models with obscure interpretations. Even in cases with straightforward explanatory variables, the so-called ‘best’ model produced by a given model-selection technique may fail to capture information of vital importance to the domain-specific questions at hand. Herein we propose a new concept for model selection, feasibility, for use in identifying multiple models that are in some sense optimal and may unite to provide a wider range of information relevant to the topic of interest, including (but not limited to) interaction terms. We further provide an R package and associated Shiny Applications for use in identifying or validating feasible models, the performance of which we demonstrate on both simulated and real-life data.  相似文献   

7.
When analyzing data on subjective expectations of continuous outcomes, researchers have access to a limited number of reported probabilities for each respondent from which to construct complete distribution functions. Moreover, reported probabilities may be rounded and thus not equal to true beliefs. Using survival expectations elicited from a representative sample from the Netherlands, we investigate what can be learned if we take these two sources of missing information into account and expectations are therefore only partially identified. We find novel evidence for rounding by checking whether reported expectations are consistent with a hazard of death that increases weakly with age. Only 39% of reported beliefs are consistent with this under the assumption that all probabilities are reported precisely, while 92% are if we allow for rounding. Using the available information to construct bounds on subjective life expectancy, we show that the data alone are not sufficiently informative to allow for useful inference in partially identified linear models, even in the absence of rounding. We propose to improve precision by interpolation between rounded probabilities. Interpolation in combination with a limited amount of rounding does yield informative intervals.  相似文献   

8.
Identification of the type of disease pattern and spread in a field is critical in epidemiological investigations of plant diseases. For example, an aggregation pattern of infected plants suggests that, at the time of observation, the pathogen is spreading from a proximal source. Conversely, a random pattern suggests a lack of spread from a proximal source. Most of the existing methods of spatial pattern analysis work with only one variety of plant at each location and with uniform genetic disease susceptibility across the field. Pecan orchards, used in this study, and other orchard crops are usually composed of different varieties with different levels of susceptibility to disease. A new measure is suggested to characterize the spatio-temporal transmission patterns of disease; a Monte Carlo test procedure is proposed to test whether the transmission of disease is random or aggregated. In addition, we propose a mixed-transmission model, which allows us to quantify the degree of aggregation effect.  相似文献   

9.
An important marker for identifying the progression of human immunodeficiency virus (HIV) infection in an individual is the CD4 cell count. Antiretroviral therapy (ART) is a treatment for HIV/AIDS (AIDS, acquired immune-deficiency syndrome) which prolongs and improves the lives of patients by improving the CD4 cell count and strengthen the immune system. This strengthening of the immune system in terms of CD4 count, not only depends on various biological factors, but also other behavioral factors. Previous studies have shown the effect of CD4 count on the mortality, but nobody has attempted to study the factors which are likely to influence the improvement in CD4 count of patients diagnosed of AIDS and undergoing ART. In this paper, we use Poisson regression model (GPR) for exploring the effect of various socio-demographic covariates such as age, gender, geographical location, and drug usage on the improvement in the CD4 count of AIDS patients. However, if the CD4 count data suffers from under or overdispersion, we use GPR model and compare it with negative binomial distribution. Finally, the model is applied for the analysis of data on patients undergoing the ART in the Ram Manohar Lohia Hospital, Delhi, India. The data exhibited overdispersion and hence, GPR model provided the best fit.  相似文献   

10.
Summary.  We propose an adaptive varying-coefficient spatiotemporal model for data that are observed irregularly over space and regularly in time. The model is capable of catching possible non-linearity (both in space and in time) and non-stationarity (in space) by allowing the auto-regressive coefficients to vary with both spatial location and an unknown index variable. We suggest a two-step procedure to estimate both the coefficient functions and the index variable, which is readily implemented and can be computed even for large spatiotemporal data sets. Our theoretical results indicate that, in the presence of the so-called nugget effect, the errors in the estimation may be reduced via the spatial smoothing—the second step in the estimation procedure proposed. The simulation results reinforce this finding. As an illustration, we apply the methodology to a data set of sea level pressure in the North Sea.  相似文献   

11.
We propose a new class of time dependent random probability measures and show how this can be used for Bayesian nonparametric inference in continuous time. By means of a nonparametric hierarchical model we define a random process with geometric stick-breaking representation and dependence structure induced via a one dimensional diffusion process of Wright-Fisher type. The sequence is shown to be a strongly stationary measure-valued process with continuous sample paths which, despite the simplicity of the weights structure, can be used for inferential purposes on the trajectory of a discretely observed continuous-time phenomenon. A simple estimation procedure is presented and illustrated with simulated and real financial data.  相似文献   

12.
As reported by Kalbfleisch and Prentice (1980), the generalized Wilcoxon test fails to detect a difference between the lifetime distributions of the male and female mice died from Thymic Leukemia. This failure is a result of the test's inability to detect a distributional difference when a location shift and a scale change exist simultaneously. In this article, we propose an estimator based on the minimization of an average distance between two independent quantile processes under a location-scale model. Large sample inference on the proposed estimator, with possible right-censorship, is discussed. The mouse leukemia data are used as an example for illustration purpose.  相似文献   

13.
In modeling disease transmission, contacts are assumed to have different infection rates. A proper simulation must model the heterogeneity in the transmission rates. In this article, we present a computationally efficient algorithm that can be applied to a population with heterogeneous transmission rates. We conducted a simulation study to show that the algorithm is more efficient than other algorithms for sampling the disease transmission in a subset of the heterogeneous population. We use a valid stochastic model of pandemic influenza to illustrate the algorithm and to estimate the overall infection attack rates of influenza A (H1N1) in a Canadian city.  相似文献   

14.
We propose a hidden Markov model for longitudinal count data where sources of unobserved heterogeneity arise, making data overdispersed. The observed process, conditionally on the hidden states, is assumed to follow an inhomogeneous Poisson kernel, where the unobserved heterogeneity is modeled in a generalized linear model (GLM) framework by adding individual-specific random effects in the link function. Due to the complexity of the likelihood within the GLM framework, model parameters may be estimated by numerical maximization of the log-likelihood function or by simulation methods; we propose a more flexible approach based on the Expectation Maximization (EM) algorithm. Parameter estimation is carried out using a non-parametric maximum likelihood (NPML) approach in a finite mixture context. Simulation results and two empirical examples are provided.  相似文献   

15.
Extreme Value Theory (EVT) aims to study the tails of probability distributions in order to measure and quantify extreme events of maximum and minimum. In river flow data, an extreme level of a river may be related to the level of a neighboring river that flows into it. In this type of data, it is very common for flooding of a location to have been caused by a very large flow from an affluent river that is tens or hundreds of kilometers from this location. In this sense, an interesting approach is to consider a conditional model for the estimation of a multivariate model. Inspired by this idea, we propose a Bayesian model to describe the dependence of exceedance between rivers, where we considered a conditionally independent structure. In this model, the dependence between rivers is captured by modeling the excess marginally of one river as a consequence of linear functions of the other rivers. The results showed that there is a strong and positive connection between excesses in one river caused by the excesses of the other rivers.  相似文献   

16.
投入产出方法是产业经济研究的基本分析工具,研究产业之间价格波动传导又是产业经济调控的重要内容。运用离散状态方程,引入价格传导时滞因素对传统投入产出分析方法进行拓展,建立一个含价格传导时滞的投入产出拓展模型。该模型给出了时滞状态下产业部门之间价格传导的解析公式,可应用于产业部门之间价格传导的分析与计算。  相似文献   

17.
We study the most basic Bayesian forecasting model for exponential family time series, the power steady model (PSM) of Smith, in terms of observable properties of one-step forecast distributions and sample paths. The PSM implies a constraint between location and spread of the forecast distribution. Including a scale parameter in the models does not always give an exact solution free of this problem, but it does suggest how to define related models free of the constraint. We define such a class of models which contains the PSM. We concentrate on the case where observations are non-negative. Probability theory and simulation show that under very mild conditions almost all sample paths of these models converge to some constant, making them unsuitable for modelling in many situations. The results apply more generally to non-negative models defined in terms of exponentially weighted moving averages. We use these and related results to motivate, define and apply very simple models based on directly specifying the forecast distributions.  相似文献   

18.
Abstract. The modelling process in Bayesian Statistics constitutes the fundamental stage of the analysis, since depending on the chosen probability laws the inferences may vary considerably. This is particularly true when conflicts arise between two or more sources of information. For instance, inference in the presence of an outlier (which conflicts with the information provided by the other observations) can be highly dependent on the assumed sampling distribution. When heavy‐tailed (e.g. t) distributions are used, outliers may be rejected whereas this kind of robust inference is not available when we use light‐tailed (e.g. normal) distributions. A long literature has established sufficient conditions on location‐parameter models to resolve conflict in various ways. In this work, we consider a location–scale parameter structure, which is more complex than the single parameter cases because conflicts can arise between three sources of information, namely the likelihood, the prior distribution for the location parameter and the prior for the scale parameter. We establish sufficient conditions on the distributions in a location–scale model to resolve conflicts in different ways as a single observation tends to infinity. In addition, for each case, we explicitly give the limiting posterior distributions as the conflict becomes more extreme.  相似文献   

19.
The main purpose of this paper is the longitudinal analysis of the poverty phenomenon. By interpreting poverty as a latent variable, we are able to resort to the statistical methodology developed for latent structure analysis. In particular, we propose to use the mixture latent Markov model which allows us to achieve two goals: (i) a time-invariant classification of households into homogenous groups, representing different levels of poverty; (ii) the dynamic analysis of the poverty phenomenon which highlights the distinction between transitory and permanent poverty situations. Furthermore, we exploit the flexibility provided by the model in order to achieve the measurement of poverty in a multidisciplinary framework, using several socio-economic indicators as covariates and identifying the main relevant factors which influence permanent and transitory poverty. The analysis of the longitudinal data of the Survey on Households Income and Wealth of the Bank of Italy provides the identification of two groups of households which are characterized by different dynamic features. Moreover, the inclusion of socio-economic covariates such as level of education, employment status, geographical area and residence size of the household head shows a direct association with permanent poverty.  相似文献   

20.
We describe the use of perfect sampling algorithms for Bayesian variable selection in a linear regression model. Starting with a basic case solved by Huang and Djurić (EURASIP J. Appl. Si. Pr. 1 (2002) 38), where the model coefficients and noise variance are assumed to be known, we generalize the model step by step to allow for other sources of randomness. We specify perfect simulation algorithms that solve these cases by incorporating various techniques including Gibbs sampling, the perfect independent Metropolis–Hastings (IMH) algorithm, and recently developed “slice coupling” algorithms. Applications to simulated data sets suggest that our algorithms perform well in identifying relevant predictor variables.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号