首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 250 毫秒
1.
ABSTRACT

A bivariate distribution, whose marginal distributions are truncated Poisson distributions, is developed as a product of truncated Poisson distributions and a multiplicative factor. The multiplicative factor takes into account the correlation, either positive or negative, between the two random variables. The distributional properties of this model are studied and the model is fitted to a real life bivariate data.  相似文献   

2.
3.
This paper deals with the asymptotics of a class of tests for association in 2-way contingency tables based on square forms in cell frequencies, given the total number of observations (multinomial sampling) or one set of marginal totals (stratified sampling). The case when both row and column marginal totals are fixed (hypergeometric sampling) was studied in Kulinskaya (1994), The class of tests under consideration includes a number of classical measures for association, Its two subclasses are the tests based on statistics using centralized cell frequencies (asymptotically distributed as weighted sums of central chi-squares) and those using the non-centralized cell frequencies (asymptotically normal). The parameters of asymptotic distributions depend on the sampling model and on true marginal probabilities. Maximum efficiency for asymptotically normal statistics is achieved under hypergeometric sampling, If the cell frequencies or the statistic as a whole are centralized using marginal proportions as estimates for marginal probabilities, the asymptotic distribution does not differ much between models and it is equivalent to that under hypergeometric sampling. These findings give an extra justification for the use of permutation tests for association (which are based on hypergeometric sampling). As an application, several well known measures of association are analysed.  相似文献   

4.
This paper sets out to identify the abilities that a person needs to be able to successfully use an experimental device, such as a probability wheel or balls in an urn, for the elicitation of subjective probabilities. It is assumed that the successful use of the device requires that the person elicits unique probability values that obey the standard probability laws. This leads to a definition of probability based on the idea of the similarity between the likeliness of events and this concept is naturally extended to the idea that probabilities have strengths, which relates to information about the likeliness of an event that lies beyond a simple probability value. The latter notion is applied to the problem of explaining the Ellsberg paradox. To avoid the definition of probability being circular, probabilities are defined such that they depend on the choice of a reference set of events R which, in simple cases, corresponds to the raw outcomes produced by using an experimental device. However, it is shown that even when the events in R are considered as having an “equal chance” of occurring, the values and/or strengths of probabilities can still be affected by the choice of the set R.  相似文献   

5.
In models using categorical data, one may use adjacency relations to justify smoothing to improve upon simple histogram approximations of the probabilities. This is particularly convenient for sparsely observed or rather peaked distributions. Moreover, in a few models, prior knowledge of a marginal distribution is available. We adapt local polynomial estimators to include this partial information about the underlying distribution and give explicit representations for the proposed estimators. An application to a set of anthropological data is included.  相似文献   

6.
Disease prediction based on longitudinal data can be done using various modeling approaches. Alternative approaches are compared using data from a longitudinal study to predict the onset of disease. The data are modeled using linear mixed-effects models. Posterior probabilities of group membership are computed starting with the first observation and sequentially adding observations until the subject is classified as developing the disease or until the last measurement is used. Individuals are classified by computing posterior probabilities using the marginal distributions of the mixed-effects models, the conditional distributions (conditional on the group-specific random effects), and the distributions of the random effects.  相似文献   

7.
In this article, we highlight some interesting facts about Bayesian variable selection methods for linear regression models in settings where the design matrix exhibits strong collinearity. We first demonstrate via real data analysis and simulation studies that summaries of the posterior distribution based on marginal and joint distributions may give conflicting results for assessing the importance of strongly correlated covariates. The natural question is which one should be used in practice. The simulation studies suggest that posterior inclusion probabilities and Bayes factors that evaluate the importance of correlated covariates jointly are more appropriate, and some priors may be more adversely affected in such a setting. To obtain a better understanding behind the phenomenon, we study some toy examples with Zellner’s g-prior. The results show that strong collinearity may lead to a multimodal posterior distribution over models, in which joint summaries are more appropriate than marginal summaries. Thus, we recommend a routine examination of the correlation matrix and calculation of the joint inclusion probabilities for correlated covariates, in addition to marginal inclusion probabilities, for assessing the importance of covariates in Bayesian variable selection.  相似文献   

8.
Mutual information (also known as Kullback–Leibler divergence) can be viewed as a measure of multivariate association in a random vector. The definition incorporates the joint density as well as the marginal densities. We will focus on a representation of mutual information in terms of copula densities that is thus independent of the marginal distributions. This representation yields a different approach to estimating mutual information than the original definition does, as only the copula density has to be estimated. We review analytical properties and examples for selected distributions and discuss methods of nonparametric estimation of copula densities and hence of the mutual information from a sample. Based on a simulation study, we compare the performance of these estimators with respect to bias, standard deviation, and the root mean squared error. The Gauss and the Frank copula are considered as examples.  相似文献   

9.
Tests are proposed for the equality of two unknown distributions. For empirical probability measures that are defined for samples from the two distributions, the proposed tests are based on the supremum of the absolute differences between the corresponding empirical probabilities, the supremum being taken over all possible events (Borel sets). In contrast, competing EDF tests compare only empirical probabilities of a subclass of Borel sets. The proposed tests are compared for simulated samples to the Kolmogorov-Smirnov, Cramér-von Mises, Kuiper, and Mann-Whitney-Wilcoxon tests  相似文献   

10.
Asymptotic Normality in Mixtures of Power Series Distributions   总被引:1,自引:0,他引:1  
Abstract.  The problem of estimating the individual probabilities of a discrete distribution is considered. The true distribution of the independent observations is a mixture of a family of power series distributions. First, we ensure identifiability of the mixing distribution assuming mild conditions. Next, the mixing distribution is estimated by non-parametric maximum likelihood and an estimator for individual probabilities is obtained from the corresponding marginal mixture density. We establish asymptotic normality for the estimator of individual probabilities by showing that, under certain conditions, the difference between this estimator and the empirical proportions is asymptotically negligible. Our framework includes Poisson, negative binomial and logarithmic series as well as binomial mixture models. Simulations highlight the benefit in achieving normality when using the proposed marginal mixture density approach instead of the empirical one, especially for small sample sizes and/or when interest is in the tail areas. A real data example is given to illustrate the use of the methodology.  相似文献   

11.
Many probability distributions can be represented as compound distributions. Consider some parameter vector as random. The compound distribution is the expected distribution of the variable of interest given the random parameters. Our idea is to define a partition of the domain of definition of the random parameters, so that we can represent the expected density of the variable of interest as a finite mixture of conditional densities. We then model the mixture probabilities of the conditional densities using information on population categories, thus modifying the original overall model. We thus obtain specific models for sub-populations that stem from the overall model. The distribution of a sub-population of interest is thus completely specified in terms of mixing probabilities. All characteristics of interest can be derived from this distribution and the comparison between sub-populations easily proceeds from the comparison of the mixing probabilities. A real example based on EU-SILC data is given. Then the methodology is investigated through simulation.  相似文献   

12.
A method for efficiently calculating exact marginal, conditional and joint distributions for change points defined by general finite state Hidden Markov Models is proposed. The distributions are not subject to any approximation or sampling error once parameters of the model have been estimated. It is shown that, in contrast to sampling methods, very little computation is needed. The method provides probabilities associated with change points within an interval, as well as at specific points.  相似文献   

13.
Summary We consider the analysis of discrete serially correlated data in the presence of time dependent covariates. If the interest is to relate the covariates to the marginal distribution of the data, Markov chains are an obvious tool to consider, but their use is complicated by the fact that they are expressed in terms of transitional rather than marginal probabilities. We show how to parametrize the transition matrix in a suitable way so that interpretation is as desired. The focus is on binary and Poisson data, but the methodology can be adopted also with other discrete data distributions.  相似文献   

14.
The problem of comparing several experimental treatments to a standard arises frequently in medical research. Various multi-stage randomized phase II/III designs have been proposed that select one or more promising experimental treatments and compare them to the standard while controlling overall Type I and Type II error rates. This paper addresses phase II/III settings where the joint goals are to increase the average time to treatment failure and control the probability of toxicity while accounting for patient heterogeneity. We are motivated by the desire to construct a feasible design for a trial of four chemotherapy combinations for treating a family of rare pediatric brain tumors. We present a hybrid two-stage design based on two-dimensional treatment effect parameters. A targeted parameter set is constructed from elicited parameter pairs considered to be equally desirable. Bayesian regression models for failure time and the probability of toxicity as functions of treatment and prognostic covariates are used to define two-dimensional covariate-adjusted treatment effect parameter sets. Decisions at each stage of the trial are based on the ratio of posterior probabilities of the alternative and null covariate-adjusted parameter sets. Design parameters are chosen to minimize expected sample size subject to frequentist error constraints. The design is illustrated by application to the brain tumor trial.  相似文献   

15.
Predictive distributions are developed and illustrated for prediction in some Poisson errors in variables models. Two different situations in which multiplicative treatment effects are appropriate are considered within the context of predicting counts of road accidents. Hierarchical prior structures are investigated, and numerical integration and Gibbs sampling routines are used to derive the predictive and posterior probabilities. Examples of analyses are provided with data from road accidents in Sweden.  相似文献   

16.
The aim of this paper is to develop some bivariate generalizations of the Hofmann distribution. The Hofmann distribution is known to give nice fits for overdispersed data sets. Two bivariate models are proposed. Recursive formulae are given for the evaluation of the probability function. Moments, conditional distributions and marginal distributions are studied. Two data sets are fitted based on the proposed models. Parameters are estimated by maximum likelihood.  相似文献   

17.
Kontkanen  P.  Myllymäki  P.  Silander  T.  Tirri  H.  Grünwald  P. 《Statistics and Computing》2000,10(1):39-54
In this paper we are interested in discrete prediction problems for a decision-theoretic setting, where the task is to compute the predictive distribution for a finite set of possible alternatives. This question is first addressed in a general Bayesian framework, where we consider a set of probability distributions defined by some parametric model class. Given a prior distribution on the model parameters and a set of sample data, one possible approach for determining a predictive distribution is to fix the parameters to the instantiation with the maximum a posteriori probability. A more accurate predictive distribution can be obtained by computing the evidence (marginal likelihood), i.e., the integral over all the individual parameter instantiations. As an alternative to these two approaches, we demonstrate how to use Rissanen's new definition of stochastic complexity for determining predictive distributions, and show how the evidence predictive distribution with Jeffrey's prior approaches the new stochastic complexity predictive distribution in the limit with increasing amount of sample data. To compare the alternative approaches in practice, each of the predictive distributions discussed is instantiated in the Bayesian network model family case. In particular, to determine Jeffrey's prior for this model family, we show how to compute the (expected) Fisher information matrix for a fixed but arbitrary Bayesian network structure. In the empirical part of the paper the predictive distributions are compared by using the simple tree-structured Naive Bayes model, which is used in the experiments for computational reasons. The experimentation with several public domain classification datasets suggest that the evidence approach produces the most accurate predictions in the log-score sense. The evidence-based methods are also quite robust in the sense that they predict surprisingly well even when only a small fraction of the full training set is used.  相似文献   

18.
Summary. Consider a pair of random variables, both subject to random right censoring. New estimators for the bivariate and marginal distributions of these variables are proposed. The estimators of the marginal distributions are not the marginals of the corresponding estimator of the bivariate distribution. Both estimators require estimation of the conditional distribution when the conditioning variable is subject to censoring. Such a method of estimation is proposed. The weak convergence of the estimators proposed is obtained. A small simulation study suggests that the estimators of the marginal and bivariate distributions perform well relatively to respectively the Kaplan–Meier estimator for the marginal distribution and the estimators of Pruitt and van der Laan for the bivariate distribution. The use of the estimators in practice is illustrated by the analysis of a data set.  相似文献   

19.
The polyhazard model with dependent causes, first introduced to fit lifetime data, generalized the traditional polyhazard model by allowing the latent causes of failure to be dependent by using copula functions. When modeling lifetime data, marginal distributions are supported on the positive reals. Dropping this restriction, the method generates a rich family of univariate distributions with asymmetries and multiple modes. We show that this new family of distributions is able to approximate other distributions proposed in the literature, such as the generalized beta-generated distributions. These distributions are fitted to three real data sets.  相似文献   

20.
We study the performance of six proposed bivariate survival curve estimators on simulated right censored data. The performance of the estimators is compared for data generated by three bivariate models with exponential marginal distributions. The estimators are compared in their ability to estimate correlations and survival functions probabilities. Simulated data results are presented so that the proposed estimators in this relatively new area of analysis can be explicitly compared to the known distribution of the data and the parameters of the underlying model. The results show clear differences in the performance of the estimators.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号