首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 265 毫秒
1.
Given the random walk model, we show, for the traditional unrestricted regression used in testing stationarity, that no matter what the initial value of the random walk is or its drift or its error standard deviation, the sampling distributions of certain statistics remain unchanged. Using Monte Carlo simulations, we estimate, for different finite samples, the sampling distributions of these statistics. After smoothing the percentiles of the empirical sampling distributions, we come up with a new set of critical values for testing the existence of a random walk, if each statistic is being used on an individual base. Combining the new sets of critical values, we finally suggest a general methodology for testing for a random walk model.  相似文献   

2.
We propose a Monte Carlo sampling algorithm for estimating guananteed-coverage tolerance factors for non-normal continuous distributions with known shape but u n p w n location and scale. The algorithm is based on reformulating this root-finding problem as a quantile-estimation problem. The reformulation leads to a geometrical interpretation of the tolerance-interval factor. For arbitrary distribution shapes, we analytically and empirically investigate various relationships among tolerance- interval coverage, confidence, and sample size.  相似文献   

3.
Two new implementations of the EM algorithm are proposed for maximum likelihood fitting of generalized linear mixed models. Both methods use random (independent and identically distributed) sampling to construct Monte Carlo approximations at the E-step. One approach involves generating random samples from the exact conditional distribution of the random effects (given the data) by rejection sampling, using the marginal distribution as a candidate. The second method uses a multivariate t importance sampling approximation. In many applications the two methods are complementary. Rejection sampling is more efficient when sample sizes are small, whereas importance sampling is better with larger sample sizes. Monte Carlo approximation using random samples allows the Monte Carlo error at each iteration to be assessed by using standard central limit theory combined with Taylor series methods. Specifically, we construct a sandwich variance estimate for the maximizer at each approximate E-step. This suggests a rule for automatically increasing the Monte Carlo sample size after iterations in which the true EM step is swamped by Monte Carlo error. In contrast, techniques for assessing Monte Carlo error have not been developed for use with alternative implementations of Monte Carlo EM algorithms utilizing Markov chain Monte Carlo E-step approximations. Three different data sets, including the infamous salamander data of McCullagh and Nelder, are used to illustrate the techniques and to compare them with the alternatives. The results show that the methods proposed can be considerably more efficient than those based on Markov chain Monte Carlo algorithms. However, the methods proposed may break down when the intractable integrals in the likelihood function are of high dimension.  相似文献   

4.
We give an upper bound for the expected value of the largest order statistic of a simple random sample of size n from a discrete distribution on N points. We also characterize the distributions that attain such bound. In the particular case n=2, we obtain a characterization of the discrete uniform distribution. © 1998 Elsevier Science B.V. All rights reserved.  相似文献   

5.
We give algorithms for sampling from non-exchangeable Archimedean copulas created by the nesting of Archimedean copula generators, where in the most general algorithm the generators may be nested to an arbitrary depth. These algorithms are based on mixture representations of these copulas using Laplace transforms. While in principle the approach applies to all nested Archimedean copulas, in practice the approach is restricted to certain cases where we are able to sample distributions with given Laplace transforms. Precise instructions are given for the case when all generators are taken from the Gumbel parametric family or the Clayton family; the Gumbel case in particular proves very easy to simulate.  相似文献   

6.
A general framework is proposed for modelling clustered mixed outcomes. A mixture of generalized linear models is used to describe the joint distribution of a set of underlying variables, and an arbitrary function relates the underlying variables to be observed outcomes. The model accommodates multilevel data structures, general covariate effects and distinct link functions and error distributions for each underlying variable. Within the framework proposed, novel models are developed for clustered multiple binary, unordered categorical and joint discrete and continuous outcomes. A Markov chain Monte Carlo sampling algorithm is described for estimating the posterior distributions of the parameters and latent variables. Because of the flexibility of the modelling framework and estimation procedure, extensions to ordered categorical outcomes and more complex data structures are straightforward. The methods are illustrated by using data from a reproductive toxicity study.  相似文献   

7.
Block and Basu bivariate exponential distribution is one of the most popular absolute continuous bivariate distributions. Recently, Kundu and Gupta [A class of absolute continuous bivariate distributions. Statist Methodol. 2010;7:464–477] introduced Block and Basu bivariate Weibull (BBBW) distribution, which is a generalization of the Block and Basu bivariate exponential distribution, and provided the maximum likelihood estimators using EM algorithm. In this paper, we consider the Bayesian inference of the unknown parameters of the BBBW distribution. The Bayes estimators are obtained with respect to the squared error loss function, and the prior distributions allow for prior dependence among the unknown parameters. Prior independence also can be obtained as a special case. It is observed that the Bayes estimators of the unknown parameters cannot be obtained in explicit forms. We propose to use the importance sampling technique to compute the Bayes estimates and also to construct the associated highest posterior density credible intervals. The analysis of two data sets has been performed for illustrative purposes. The performances of the proposed estimators are quite satisfactory. Finally, we generalize the results for the multivariate case.  相似文献   

8.
In this article, we consider the multiple step stress model based on the cumulative exposure model assumption. Here, it is assumed that for a given stress level, the lifetime of the experimental units follows exponential distribution and the expected lifetime decreases as the stress level increases. We mainly focus on the order restricted inference of the unknown parameters of the lifetime distributions. First we consider the order restricted maximum likelihood estimators (MLEs) of the model parameters. It is well known that the order restricted MLEs cannot be obtained in explicit forms. We propose an algorithm that stops in finite number of steps and it provides the MLEs. We further consider the Bayes estimates and the associated credible intervals under the squared error loss function. Due to the absence of explicit form of the Bayes estimates, we propose to use the importance sampling technique to compute Bayes estimates. We provide an extensive simulation study in case of three stress levels mainly to see the performance of the proposed methods. Finally the analysis of one real data set has been provided for illustrative purposes.  相似文献   

9.
This article describes a method for computing approximate statistics for large data sets, when exact computations may not be feasible. Such situations arise in applications such as climatology, data mining, and information retrieval (search engines). The key to our approach is a modular approximation to the cumulative distribution function (cdf) of the data. Approximate percentiles (as well as many other statistics) can be computed from this approximate cdf. This enables the reduction of a potentially overwhelming computational exercise into smaller, manageable modules. We illustrate the properties of this algorithm using a simulated data set. We also examine the approximation characteristics of the approximate percentiles, using a von Mises functional type approach. In particular, it is shown that the maximum error between the approximate cdf and the actual cdf of the data is never more than 1% (or any other preset level). We also show that under assumptions of underlying smoothness of the cdf, the approximation error is much lower in an expected sense. Finally, we derive bounds for the approximation error of the percentiles themselves. Simulation experiments show that these bounds can be quite tight in certain circumstances.  相似文献   

10.
New recursive algorithms for fast computation of the normalizing constant for the autologistic model on the lattice make feasible a sample-based maximum likelihood estimation (MLE) of the autologistic parameters. We demonstrate by sampling from 12 simulated 420×420 binary lattices with square lattice plots of size 4×4, …, 7×7 and sample sizes between 20 and 600. Sample-based results are compared with ‘benchmark’ MCMC estimates derived from all binary observations on a lattice. Sample-based estimates are, on average, biased systematically by 3%–7%, a bias that can be reduced by more than half by a set of calibrating equations. MLE estimates of sampling variances are large and usually conservative. The variance of the parameter of spatial association is about 2–10 times higher than the variance of the parameter of abundance. Sample distributions of estimates were mostly non-normal. We conclude that sample-based MLE estimation of the autologistic parameters with an appropriate sample size and post-estimation calibration will furnish fully acceptable estimates. Equations for predicting the expected sampling variance are given.  相似文献   

11.
Multistage ranked-set sampling (MRSS) is a generalization of ranked-set sampling in which multiple stages of ranking are used. It is known that for a fixed distribution under perfect rankings, each additional stage provides a gain in efficiency when estimating the population mean. However, the maximum possible efficiency for the MRSS sample mean relative to the simple random sampling sample mean has not previously been determined. In this paper, we provide a method for computing this maximum possible efficiency under perfect rankings for any choice of the set size and the number of stages. The maximum efficiency tends to infinity as the number of stages increases, and, for large numbers of stages, the efficiency-maximizing distributions are symmetric multi-modal distributions where the number of modes matches the set size. The results in this paper correct earlier assertions in the literature that the maximum efficiency is bounded and that it is achieved when the distribution is uniform.  相似文献   

12.
Simulation of multivariate distributions is important in many applications but remains computationally challenging in practice. In this article, we introduce three classes of multivariate distributions from which simulation can be conducted by means of their stochastic representations related to the Dirichlet random vector. More emphasis is made to simulation from the class of uniform distributions over a polyhedron, which is useful for solving some constrained optimization problems and ha`s many applications in sampling and Monte Carlo simulations. Numerical evidences show that, by utilizing state-of-the-art Dirichlet generation algorithms, the introduced methods become superior to other approaches in terms of computational efficiency.  相似文献   

13.
In this paper, we consider the problem of making statistical inference for a truncated normal distribution under progressive type I interval censoring. We obtain maximum likelihood estimators of unknown parameters using the expectation-maximization algorithm and in sequel, we also compute corresponding midpoint estimates of parameters. Estimation based on the probability plot method is also considered. Asymptotic confidence intervals of unknown parameters are constructed based on the observed Fisher information matrix. We obtain Bayes estimators of parameters with respect to informative and non-informative prior distributions under squared error and linex loss functions. We compute these estimates using the importance sampling procedure. The highest posterior density intervals of unknown parameters are constructed as well. We present a Monte Carlo simulation study to compare the performance of proposed point and interval estimators. Analysis of a real data set is also performed for illustration purposes. Finally, inspection times and optimal censoring plans based on the expected Fisher information matrix are discussed.  相似文献   

14.
Convergence of Heavy-tailed Monte Carlo Markov Chain Algorithms   总被引:1,自引:0,他引:1  
Abstract.  In this paper, we use recent results of Jarner & Roberts ( Ann. Appl. Probab., 12, 2002, 224) to show polynomial convergence rates of Monte Carlo Markov Chain algorithms with polynomial target distributions, in particular random-walk Metropolis algorithms, Langevin algorithms and independence samplers. We also use similar methodology to consider polynomial convergence of the Gibbs sampler on a constrained state space. The main result for the random-walk Metropolis algorithm is that heavy-tailed proposal distributions lead to higher rates of convergence and thus to qualitatively better algorithms as measured, for instance, by the existence of central limit theorems for higher moments. Thus, the paper gives for the first time a theoretical justification for the common belief that heavy-tailed proposal distributions improve convergence in the context of random-walk Metropolis algorithms. Similar results are shown to hold for Langevin algorithms and the independence sampler, while results for the mixing of Gibbs samplers on uniform distributions on constrained spaces are rather different in character.  相似文献   

15.
In this paper we describe a sequential importance sampling (SIS) procedure for counting the number of vertex covers in general graphs. The optimal SIS proposal distribution is the uniform over a suitably restricted set, but is not implementable. We will consider two proposal distributions as approximations to the optimal. Both proposals are based on randomization techniques. The first randomization is the classic probability model of random graphs, and in fact, the resulting SIS algorithm shows polynomial complexity for random graphs. The second randomization introduces a probabilistic relaxation technique that uses Dynamic Programming. The numerical experiments show that the resulting SIS algorithm enjoys excellent practical performance in comparison with existing methods. In particular the method is compared with cachet—an exact model counter, and the state of the art SampleSearch, which is based on Belief Networks and importance sampling.  相似文献   

16.
In this paper, we study the statistical inference based on the Bayesian approach for regression models with the assumption that independent additive errors follow normal, Student-t, slash, contaminated normal, Laplace or symmetric hyperbolic distribution, where both location and dispersion parameters of the response variable distribution include nonparametric additive components approximated by B-splines. This class of models provides a rich set of symmetric distributions for the model error. Some of these distributions have heavier or lighter tails than the normal as well as different levels of kurtosis. In order to draw samples of the posterior distribution of the interest parameters, we propose an efficient Markov Chain Monte Carlo (MCMC) algorithm, which combines Gibbs sampler and Metropolis–Hastings algorithms. The performance of the proposed MCMC algorithm is assessed through simulation experiments. We apply the proposed methodology to a real data set. The proposed methodology is implemented in the R package BayesGESM using the function gesm().  相似文献   

17.
In this paper, we consider the finite mixture of quantile regression model from a Bayesian perspective by assuming the errors have the asymmetric Laplace distribution (ALD), and develop the Gibbs sampling algorithm to estimate various quantile conditional on covariate in different groups using the Normal-Exponential representation of the ALD. We conduct several simulations under different error distributions to demonstrate the performance of the algorithm, and finally apply it to analyse a real data set, finding that the procedure has good performance.  相似文献   

18.
The hybrid censoring scheme is a mixture of Type-I and Type-II censoring schemes. Based on hybrid censored samples, we first derive the maximum likelihood estimators of the unknown parameters and the expected Fisher’s information matrix of the generalized inverted exponential distribution (GIED). Monte Carlo simulations are performed to study the performance of the maximum likelihood estimators. Next we consider Bayes estimation under the squared error loss function. These Bayes estimates are evaluated by applying Lindley’s approximation method, the importance sampling procedure and Metropolis–Hastings algorithm. The importance sampling technique is used to compute the highest posterior density credible intervals. Two data sets are analyzed for illustrative purposes. Finally, we discuss a method of obtaining the optimum hybrid censoring scheme.  相似文献   

19.
Ensemble methods using the same underlying algorithm trained on different subsets of observations have recently received increased attention as practical prediction tools for massive data sets. We propose Subsemble: a general subset ensemble prediction method, which can be used for small, moderate, or large data sets. Subsemble partitions the full data set into subsets of observations, fits a specified underlying algorithm on each subset, and uses a clever form of V-fold cross-validation to output a prediction function that combines the subset-specific fits. We give an oracle result that provides a theoretical performance guarantee for Subsemble. Through simulations, we demonstrate that Subsemble can be a beneficial tool for small- to moderate-sized data sets, and often has better prediction performance than the underlying algorithm fit just once on the full data set. We also describe how to include Subsemble as a candidate in a SuperLearner library, providing a practical way to evaluate the performance of Subsemble relative to the underlying algorithm fit just once on the full data set.  相似文献   

20.
Inverse Weibull (IW) distribution is one of the widely used probability distributions for nonnegative data modelling, specifically, for describing degradation phenomena of mechanical components. In this paper, by compounding IW and power series distributions we introduce a new lifetime distribution. The compounding procedure follows the same set-up carried out by Adamidis and Loukas [A lifetime distribution with decreasing failure rate. Stat Probab Lett. 1998;39:35–42]. We provide mathematical properties of this new distribution such as moments, estimation by maximum likelihood with censored data, inference for a large sample and the EM algorithm to determine the maximum likelihood estimates of the parameters. Furthermore, we characterize the proposed distributions using a simple relationship between two truncated moments and maximum entropy principle under suitable constraints. Finally, to show the flexibility of this type of distributions, we demonstrate applications of two real data sets.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号