首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 109 毫秒
1.
2.
Although applications of Bayesian analysis for numerical quadrature problems have been considered before, it is only very recently that statisticians have focused on the connections between statistics and numerical analysis of differential equations. In line with this very recent trend, we show how certain commonly used finite difference schemes for numerical solutions of ordinary and partial differential equations can be considered in a regression setting. Focusing on this regression framework, we apply a simple Bayesian strategy to obtain confidence intervals for the finite difference solutions. We apply this framework on several examples to show how the confidence intervals are related to truncation error and illustrate the utility of the confidence intervals for the examples considered.  相似文献   

3.
Summary. We present a technique for extending generalized linear models to the situation where some of the predictor variables are observations from a curve or function. The technique is particularly useful when only fragments of each curve have been observed. We demonstrate, on both simulated and real data sets, how this approach can be used to perform linear, logistic and censored regression with functional predictors. In addition, we show how functional principal components can be used to gain insight into the relationship between the response and functional predictors. Finally, we extend the methodology to apply generalized linear models and principal components to standard missing data problems.  相似文献   

4.
5.
Summary.  Mastitis resistance data on dairy cattle are modelled as first-passage times of stochastic processes. Population heterogeneity is included by expressing process parameters as functions of shared random variables. We show how dependences between individuals, e.g. genetic relationships, can be exploited in the analyses. The method can be extended to handle situations with multiple hidden causes of failure. Markov chain Monte Carlo methods are used for estimation.  相似文献   

6.
We show how a simple modification of the splitting method based on Gibbs sampler can be efficiently used for decision making in the sense that one can efficiently decide whether or not a given set of integer program constraints has at least one feasible solution. We also show how to incorporate the classic capture-recapture method into the splitting algorithm in order to obtain a low variance estimator for the counting quantity representing, say the number of feasible solutions on the set of the constraints of an integer program. We finally present numerical with with both, the decision making and the capture-recapture estimators and show their superiority as compared to the conventional one, while solving quite general decision making and counting ones, like the satisfiability problems.  相似文献   

7.
We discuss how lower previsions induced by multi-valued mappings fit into the framework of the behavioural theory of imprecise probabilities, and show how the notions of coherence and natural extension from that theory can be used to prove and generalise existing results in an elegant and straightforward manner. This provides a clear example for their explanatory and unifying power.  相似文献   

8.
Summary.  Key ecological studies involve the regular censusing of populations of wild animals, resulting in individual case history data which record when marked individuals are seen alive and/or found dead. We show how current conditional methods of analysing case history data may be biased. We then show how a correction can be applied, making use of results from a mark–recovery–recapture analysis. This allows a simple investigation of the effect of time-varying individual covariates such as weight that often contain missing values. The work is motivated and illustrated by the study of Soay sheep in the St Kilda archipelago.  相似文献   

9.
Logistic regression is frequently used for classifying observations into two groups. Unfortunately there are often outlying observations in a data set and these might affect the estimated model and the associated classification error rate. In this paper, the authors study the effect of observations in the training sample on the error rate by deriving influence functions. They obtain a general expression for the influence function of the error rate, and they compute it for the maximum likelihood estimator as well as for several robust logistic discrimination procedures. Besides being of interest in their own right, the influence functions are also used to derive asymptotic classification efficiencies of different logistic discrimination rules. The authors also show how influential points can be detected by means of a diagnostic plot based on the values of the influence function  相似文献   

10.
The analysis of non stationary data streams requires a continuous adaption of the model to the relevant most recent data. This requires that changes in the data stream must be distinguished from noise. Many approaches are based on heuristic adaptation schemes. We analyze simple regression models to understand the joint effects of noise and concept drift and derive the optimal sliding window size for the regression models. Our theoretical analysis and simulations show that a near optimal window size can be crucial. Our models can be used as benchmarks for other models to see how they cope with noise and drift.  相似文献   

11.
The Dirichlet process has been used extensively in Bayesian non parametric modeling, and has proven to be very useful. In particular, mixed models with Dirichlet process random effects have been used in modeling many types of data and can often outperform their normal random effect counterparts. Here we examine the linear mixed model with Dirichlet process random effects from a classical view, and derive the best linear unbiased estimator (BLUE) of the fixed effects. We are also able to calculate the resulting covariance matrix and find that the covariance is directly related to the precision parameter of the Dirichlet process, giving a new interpretation of this parameter. We also characterize the relationship between the BLUE and the ordinary least-squares (OLS) estimator and show how confidence intervals can be approximated.  相似文献   

12.
ABSTRACT

In this article we propose indices for technological breadth and depth of patent documents. We conceptually compare and contrast several indices that have been used in the existing literature, explaining main limitations. Motivated by the drawbacks of those we demonstrate how Rao’s Quadratic Entropy, a statistical index used in ecology for measuring biodiversity, can be decomposed to separately measure technological breadth and depth. The properties of breadth and depth are then investigated using patents for business data processing. For the technological domains with the highest patenting activity, we show how the novel measures of technological breadth and depth can be used to rank the patenting entities by average breadth, and identify ones with inventions that on average exceed the domain-specific average depth. Practical implications of the proposed indices are also exemplified and discussed in the context of competitor analysis for entrepreneurial ventures in the area of network security for business data processing.  相似文献   

13.
Summary.  The forward–backward algorithm is an exact filtering algorithm which can efficiently calculate likelihoods, and which can be used to simulate from posterior distributions. Using a simple result which relates gamma random variables with different rates, we show how the forward–backward algorithm can be used to calculate the distribution of a sum of gamma random variables, and to simulate from their joint distribution given their sum. One application is to calculating the density of the time of a specific event in a Markov process, as this time is the sum of exponentially distributed interevent times. This enables us to apply the forward–backward algorithm to a range of new problems. We demonstrate our method on three problems: calculating likelihoods and simulating allele frequencies under a non-neutral population genetic model, analysing a stochastic epidemic model and simulating speciation times in phylogenetics.  相似文献   

14.
We describe some developments in the P OPAN system for the analysis of mark-recapture data from Jolly-Seber (JS) type experiments. The latest version, P OPAN-6, adopts the Design Matrix approach for specifying constraints and then uses it in the constrained maximization of the likelihood. We describe how this is done and the difference it makes to convergence and parameter identifiability over the constraint contrast-equation methods used in P OPAN-5. Then we show how the SIMULATE capabilities of P OPAN can be used to explore the properties of estimates, including their identifiability, precision, and robustness to model misspecification or capture heterogeneity.  相似文献   

15.
Data envelopment analysis models are used for measuring composite indicators in various areas. Although there are many models for measuring composite indicators in the literature, surprisingly, there is no methodology that clearly shows how composite indicators improvement could be performed. This article proposes a slack analysis framework for improving the composite indicator of inefficient entities. For doing so, two dual problems originated from two data envelopment analysis models in the literature are proposed, which can guide decision makers on how to adjust the subindicators of inefficient entities to improve their composite indicators through identifying which subindicators must be improved and how much they should be augmented. The proposed methodology for improving composite indicators is inspired from data envelopment analysis and slack analysis approaches. Applicability of the proposed methodology is investigated for improving two well-known composite indicators, i.e., Sustainable Energy Index and Human Development Index. The results show that 12 out of 18 economies are inefficient in the context of sustainable energy index, for which the proposed slack analysis models provide the suggested adjustments in terms of their respected subindicators. Furthermore, the proposed methodology suggests how to adjust the life expectancy, the education, and the gross domestic product (GDP) as the three socioeconomic indicators to improve the human development index of 24 countries which are identified as inefficient entities among 27 countries.  相似文献   

16.
Meta-analytical approaches have been extensively used to analyze medical data. In most cases, the data come from different studies or independent trials with similar characteristics. However, these methods can be applied in a broader sense. In this paper, we show how existing meta-analytic techniques can also be used as well when dealing with parameters estimated from individual hierarchical data. Specifically, we propose to apply statistical methods that account for the variances (and possibly covariances) of such measures. The estimated parameters together with their estimated variances can be incorporated into a general linear mixed model framework. We illustrate the methodology by using data from a first-in-man study and a simulated data set. The analysis was implemented with the SAS procedure MIXED and example code is offered.  相似文献   

17.
For a trial with primary endpoint overall survival for a molecule with curative potential, statistical methods that rely on the proportional hazards assumption may underestimate the power and the time to final analysis. We show how a cure proportion model can be used to get the necessary number of events and appropriate timing via simulation. If phase 1 results for the new drug are exceptional and/or the medical need in the target population is high, a phase 3 trial might be initiated after phase 1. Building in a futility interim analysis into such a pivotal trial may mitigate the uncertainty of moving directly to phase 3. However, if cure is possible, overall survival might not be mature enough at the interim to support a futility decision. We propose to base this decision on an intermediate endpoint that is sufficiently associated with survival. Planning for such an interim can be interpreted as making a randomized phase 2 trial a part of the pivotal trial: If stopped at the interim, the trial data would be analyzed, and a decision on a subsequent phase 3 trial would be made. If the trial continues at the interim, then the phase 3 trial is already underway. To select a futility boundary, a mechanistic simulation model that connects the intermediate endpoint and survival is proposed. We illustrate how this approach was used to design a pivotal randomized trial in acute myeloid leukemia and discuss historical data that informed the simulation model and operational challenges when implementing it.  相似文献   

18.
The Joy of Copulas: Bivariate Distributions with Uniform Marginals   总被引:1,自引:0,他引:1  
We describe a class of bivariate distributions whose marginals are uniform on the unit interval. Such distributions are often called “copulas.” The particular copulas we present are especially well suited for use in undergraduate mathematical statistics courses, as many of their basic properties can be derived using elementary calculus. In particular, we show how these copulas can be used to illustrate the existence of distributions with singular components and to give a geometric interpretation to Kendall's tau.  相似文献   

19.
We present a methodology for rating in real-time the creditworthiness of public companies in the U.S. from the prices of traded assets. Our approach uses asset pricing data to impute a term structure of risk neutral survival functions or default probabilities. Firms are then clustered into ratings categories based on their survival functions using a functional clustering algorithm. This allows all public firms whose assets are traded to be directly rated by market participants. For firms whose assets are not traded, we show how they can be indirectly rated by matching them to firms that are traded based on observable characteristics. We also show how the resulting ratings can be used to construct loss distributions for portfolios of bonds. Finally, we compare our ratings to Standard & Poors and find that, over the period 2005 to 2011, our ratings lead theirs for firms that ultimately default.  相似文献   

20.
In the present paper we consider collective models in risk theory and their thinning and decomposition. We give three applications to reinsurance, multiplicative tariffs, and loss reserving. For each of these applications we show how maximum-likelihood and marginal-sum estimation can be used to estimate the parameters.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号