首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Summary. The human immunodeficiency virus–acquired immune deficiency syndrome (HIV–AIDS) epidemic in Hong Kong has been under surveillance in the form of voluntary reporting since 1984. However, there has been little discussion or research on the reconstruction of the HIV incidence curve. This paper is the first to use a modified back-projection method to estimate the incidence of HIV in Hong Kong on the basis of the number of positive HIV tests only. The model proposed has several advantages over the original back-projection method based on AIDS data only. First, not all HIV-infected individuals will develop AIDS by the time of analysis, but some of them may undertake an HIV test; therefore, the HIV data set contains more information than the AIDS data set. Second, the HIV diagnosis curve usually has a smoother pattern than the AIDS diagnosis curve, as it is not affected by redefinition of AIDS. Third, the time to positive HIV diagnosis is unlikely to be affected by treatment effects, as it is unlikely that an individual receives medication before the diagnosis of HIV. Fourth, the induction period from HIV infection to the first HIV positive test is usually shorter than the incubation period which is from HIV infection to diagnosis of AIDS. With a shorter induction period, more information becomes available for estimating the HIV incidence curve. Finally, this method requires the number of positive HIV diagnoses only, which is readily available from HIV–AIDS surveillance systems in many countries. It is estimated that, in Hong Kong, the cumulative number of HIV infections during the period 1979–2000 is about 2600, whereas an estimate based only on AIDS data seems to give an underestimate.  相似文献   

2.
In multi-category response models, categories are often ordered. In the case of ordinal response models, the usual likelihood approach becomes unstable with ill-conditioned predictor space or when the number of parameters to be estimated is large relative to the sample size. The likelihood estimates do not exist when the number of observations is less than the number of parameters. The same problem arises if constraint on the order of intercept values is not met during the iterative procedure. Proportional odds models (POMs) are most commonly used for ordinal responses. In this paper, penalized likelihood with quadratic penalty is used to address these issues with a special focus on POMs. To avoid large differences between two parameter values corresponding to the consecutive categories of an ordinal predictor, the differences between the parameters of two adjacent categories should be penalized. The considered penalized-likelihood function penalizes the parameter estimates or differences between the parameter estimates according to the type of predictors. Mean-squared error for parameter estimates, deviance of fitted probabilities and prediction error for ridge regression are compared with usual likelihood estimates in a simulation study and an application.  相似文献   

3.
4.
Summary.  We present models for the combined analysis of evidence from randomized controlled trials categorized as being at either low or high risk of bias due to a flaw in their conduct. We formulate a bias model that incorporates between-study and between-meta-analysis heterogeneity in bias, and uncertainty in overall mean bias. We obtain algebraic expressions for the posterior distribution of the bias-adjusted treatment effect, which provide limiting values for the information that can be obtained from studies at high risk of bias. The parameters of the bias model can be estimated from collections of previously published meta-analyses. We explore alternative models for such data, and alternative methods for introducing prior information on the bias parameters into a new meta-analysis. Results from an illustrative example show that the bias-adjusted treatment effect estimates are sensitive to the way in which the meta-epidemiological data are modelled, but that using point estimates for bias parameters provides an adequate approximation to using a full joint prior distribution. A sensitivity analysis shows that the gain in precision from including studies at high risk of bias is likely to be low, however numerous or large their size, and that little is gained by incorporating such studies, unless the information from studies at low risk of bias is limited. We discuss approaches that might increase the value of including studies at high risk of bias, and the acceptability of the methods in the evaluation of health care interventions.  相似文献   

5.
Optimal group definitions, maximizing retained information, are well known for grouping normal distributions with known parameters. When these group definitions are used with sample estimates substituted for unknown parameters, there is a shrinkage of retained information. If there are k groups, and if the size of the sample yielding parameter estimates is n, then this shrinkage is less than 5% for n <= 30 and minimization of shrinkage yields less than 1 % improvement for n <= 5, when k >= 12.  相似文献   

6.
Exponential regression model is important in analyzing data from heterogeneous populations. In this paper we propose a simple method to estimate the regression parameters using binary data. Under certain design distributions, including ellipticaily symmetric distributions, for the explanatory variables, the estimators are shown to be consistent and asymptotically normal when sample size is large. For finite samples, the new estimates were shown to behave reasonably well. They are competitive with the maximum likelihood estimates and more importantly, according to our simulation results, the cost of CPU time for computing new estimates is only 1/7 of that required for computing the usual maximum likelihood estimates. We expect the savings in CPU time would be more dramatic with larger dimension of the regression parameter space.  相似文献   

7.
Consider a large number of econometric investigations using different estimation techniques and/or different subsets of all available data to estimate a fixed set of parameters. The resulting empirical distribution of point estimates can be shown - under suitable conditions - to coincide with a Bayesian posterior measure on the parameter space induced by a minimum information procedure. This Bayesian interpretation makes it easier to combine the results of various empirical exercises for statistical decision making. The collection of estimators may be generated by one investigator to ensure the satisfaction of our conditions, or they may be collected from published works, where behavioral assumptions need to be made regarding the dependence structure of econometric studies.  相似文献   

8.
Data from a surveillance system can be used to estimate the size of a disease population. For certain surveillance systems, a binomial mixture model arises as a natural choice. The Chao estimator estimates a lower bound of the population size. The Zelterman estimator estimates a parameter that is neither a lower bound nor an upper bound. By comparing the Chao estimator and the Zelterman estimator both theoretically and numerically, we conclude that the Chao estimator is better.  相似文献   

9.
This study presents statistical techniques to obtain local approximate query answers for aggregate multivariate materialized views thus eliminating the need for repetitive scanning of the source data. In widely distributed management information systems, detailed data do not necessarily reside in the same physical location as the decision-maker; thus, requiring scanning of the source data as needed by the query demand. Decision-making, business intelligence and data analysis could involve multiple data sources, data diversity, aggregates and large amounts of data. Management often confronts delays in information acquisition from remote sites. Management decisions usually involve analyses that require the most precise summary data available. These summaries are readily available from data warehouses and can be used to estimate or approximate data in exchange for a quicker response. An approach to supporting aggregate materialized view management is proposed that reconstructs data sets locally using posterior parameter estimates based on sufficient statistics in a log-linear model with a multinomial likelihood.  相似文献   

10.
The power function distribution is often used to study the electrical component reliability. In this paper, we model a heterogeneous population using the two-component mixture of the power function distribution. A comprehensive simulation scheme including a large number of parameter points is followed to highlight the properties and behavior of the estimates in terms of sample size, censoring rate, parameters size and the proportion of the components of the mixture. The parameters of the power function mixture are estimated and compared using the Bayes estimates. A simulated mixture data with censored observations is generated by probabilistic mixing for the computational purposes. Elegant closed form expressions for the Bayes estimators and their variances are derived for the censored sample as well as for the complete sample. Some interesting comparison and properties of the estimates are observed and presented. The system of three non-linear equations, required to be solved iteratively for the computations of maximum likelihood (ML) estimates, is derived. The complete sample expressions for the ML estimates and for their variances are also given. The components of the information matrix are constructed as well. Uninformative as well as informative priors are assumed for the derivation of the Bayes estimators. A real-life mixture data example has also been discussed. The posterior predictive distribution with the informative Gamma prior is derived, and the equations required to find the lower and upper limits of the predictive intervals are constructed. The Bayes estimates are evaluated under the squared error loss function.  相似文献   

11.
In testing product reliability, there is often a critical cutoff level that determines whether a specimen is classified as failed. One consequence is that the number of degradation data collected varies from specimen to specimen. The information of random sample size should be included in the model, and our study shows that it can be influential in estimating model parameters. Two-stage least squares (LS) and maximum modified likelihood (MML) estimation, which both assume fixed sample sizes, are commonly used for estimating parameters in the repeated measurements models typically applied to degradation data. However, the LS estimate is not consistent in the case of random sample sizes. This article derives the likelihood for the random sample size model and suggests using maximum likelihood (ML) for parameter estimation. Our simulation studies show that ML estimates have smaller biases and variances compared to the LS and MML estimates. All estimation methods can be greatly improved if the number of specimens increases from 5 to 10. A data set from a semiconductor application is used to illustrate our methods.  相似文献   

12.
殷崔红等 《统计研究》2019,36(3):100-112
本文建立了索赔次数的多风险类别混合泊松模型。首先,考虑索赔次数的零膨胀、厚尾性和异质性等特征,建立风险类别待定的开放式混合泊松模型,开放式结构使该模型对实际数据的多样特征和风险类别具有良好的自适应性;其次,定义混合权重参数的iSCAD惩罚函数,实现对权重参数的筛选;最后,借助EM算法求得模型参数,实现对各风险类别下索赔次数的估计。借助iSCAD惩罚函数,给出最优混合数,避免传统混合模型中主观选择的弊端,克服传统混合模型中结构复杂、参数估计没有显式表达式、估计结果不便于解释等问题。基于三组风险特征多样数据的实证分析,本文发现该模型可以显著改进现有模型的拟合效果。  相似文献   

13.
14.
Summary.  We present a Bayesian evidence synthesis model combining data on seroprevalence, seroconversion and tests of recent infection, to produce estimates of current incidence of toxoplasmosis in the UK. The motivation for the study was the need for an estimate of current average incidence in the UK, with a realistic assessment of its uncertainty, to inform a decision model for a national screening programme to prevent congenital toxoplasmosis. The model has a hierarchical structure over geographic region, a random-walk model for temporal effects and a fixed age effect, with one or more types of data informing the regional estimates of incidence. Inference is obtained by using Markov chain Monte Carlo simulations. A key issue in the synthesis of evidence from multiple sources is model selection and the consistency of different types of evidence. Alternative models of incidence are compared by using the deviance information criterion, and we find that temporal effects are region specific. We assess the consistency of the various forms of evidence by using cross-validation where practical, and posterior and mixed prediction otherwise, and we discuss how these measures can be used to assess different aspects of consistency in a complex evidence structure. We discuss the contribution of the various forms of evidence to estimated current average incidence.  相似文献   

15.
In this article, we use a latent class model (LCM) with prevalence modeled as a function of covariates to assess diagnostic test accuracy in situations where the true disease status is not observed, but observations on three or more conditionally independent diagnostic tests are available. A fast Monte Carlo expectation–maximization (MCEM) algorithm with binary (disease) diagnostic data is implemented to estimate parameters of interest; namely, sensitivity, specificity, and prevalence of the disease as a function of covariates. To obtain standard errors for confidence interval construction of estimated parameters, the missing information principle is applied to adjust information matrix estimates. We compare the adjusted information matrix-based standard error estimates with the bootstrap standard error estimates both obtained using the fast MCEM algorithm through an extensive Monte Carlo study. Simulation demonstrates that the adjusted information matrix approach estimates the standard error similarly with the bootstrap methods under certain scenarios. The bootstrap percentile intervals have satisfactory coverage probabilities. We then apply the LCM analysis to a real data set of 122 subjects from a Gynecologic Oncology Group study of significant cervical lesion diagnosis in women with atypical glandular cells of undetermined significance to compare the diagnostic accuracy of a histology-based evaluation, a carbonic anhydrase-IX biomarker-based test and a human papillomavirus DNA test.  相似文献   

16.
This paper addresses the problem of obtaining maximum likelihood estimates for the parameters of the Pearson Type I distribution (beta distribution with unknown end points and shape parameters). Since they do not seem to have appeared in the literature, the likelihood equations and the information matrix are derived. The regularity conditions which ensure asymptotic normality and efficiency are examined, and some apparent conflicts in the literature are noted. To ensure regularity, the shape parameters must be greater than two, giving an (assymmetrical) bell-shaped distribution with high contact in the tails. A numerical investigation was carried out to explore the bias and variance of the maximum likelihood estimates and their dependence on sample size. The numerical study indicated that only for large samples (n ≥ 1000) does the bias in the estimates become small and does the Cramér-Rao bound give a good approximation for their variance. The likelihood function has a global maximum which corresponds to parameter estimates that are inadmissable. Useful parameter estimates can be obtained at a local maximum, which is sometimes difficult to locate when the sample size is small.  相似文献   

17.
Back-projection is a commonly used method in reconstructing HIV incidence. Instead of using AIDS incidence data in back-projection, this paper uses HIV positive tests data. Both multinomial and Poisson settings are used. The two settings give similar results when a parametric form or step function is assumed for the infection curve. However, this may not be true when the HIV infection in each year is characterized by a different parameter. This paper attempts to use simulation studies to compare these two settings by constructing various scenarios for the infection curve. Results show that both methods give approximately the same estimates of the number of HIV infections in the past, whilst the estimates for HIV infections in the recent past differ a lot. The multinomial setting always gives a levelling-off pattern for the recent past, while the Poisson setting is more sensitive to the change in the shape of the HIV infection curve. Nonetheless, the multinomial setting gives a relatively narrower point-wise probability interval. When the size of the epidemic is large, the narrow probability interval may be under-estimating the true underlying variation.  相似文献   

18.
The maximum likelihood and Bayesian approaches for parameter estimations and prediction of future record values have been considered for the two-parameter Burr Type XII distribution based on record values with the number of trials following the record values (inter-record times). Firstly, the Bayes estimates are obtained based on a joint bivariate prior for the shape parameters. In this case, the Bayes estimates of the parameters have been developed by using Lindley's approximation and the Markov Chain Monte Carlo (MCMC) method due to the lack of explicit forms under the squared error and the linear-exponential loss functions. The MCMC method has been also used to construct the highest posterior density credible intervals. Secondly, the Bayes estimates are obtained with respect to a discrete prior for the first shape parameter and a conjugate prior for other shape parameter. The Bayes and the maximum likelihood estimates are compared in terms of the estimated risk by the Monte Carlo simulations. We further consider the non-Bayesian and Bayesian prediction for future lower record arising from the Burr Type XII distribution based on record data. The comparison of the derived predictors is carried out by using Monte Carlo simulations. A real data are analysed for illustration purposes.  相似文献   

19.
This paper discusses maximum likelihood parameter estimation in the Pareto distribution for multicensored samples. In particu-

lar, the modality of the associated conditional log-likelihood function is investigated in order to resolve questions concerninc

the existence and uniqurneas of the lnarimum likelihood estimates.For the cases with one parameter known, the maximum likelihood

estimates of the remaining unknown parameters are shown to exist and to be unique. When both parameters are unknown, the maximum likelihood estimates may or may not exist and be unique. That is, their existence and uniqueness would seem to depend solely upon the information inherent in the sample data. In viav of the possible nonexistence and/or non-uniqueness of the maximum likelihood estimates when both parameters are unknown, alternatives to standard iterative numerical methods are explored.  相似文献   

20.
Non-central chi-squared distribution plays a vital role in statistical testing procedures. Estimation of the non-centrality parameter provides valuable information for the power calculation of the associated test. We are interested in the statistical inference property of the non-centrality parameter estimate based on one observation (usually a summary statistic) from a truncated chi-squared distribution. This work is motivated by the application of the flexible two-stage design in case–control studies, where the sample size needed for the second stage of a two-stage study can be determined adaptively by the results of the first stage. We first study the moment estimate for the truncated distribution and prove its existence, uniqueness, and inadmissibility and convergence properties. We then define a new class of estimates that includes the moment estimate as a special case. Among this class of estimates, we recommend to use one member that outperforms the moment estimate in a wide range of scenarios. We also present two methods for constructing confidence intervals. Simulation studies are conducted to evaluate the performance of the proposed point and interval estimates.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号