首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
One of the two independent stochastic processes (or ‘arms’) is selected and observed sequentially at each of n(≤ ∝) stages. Arm 1 yields observations identically distributed with unknown probability measure P with a Dirichlet process prior whereas observations from arm 2 have known probability measure Q. Future observations are discounted and at stage m, the payoff is a m(≥0) times the observation Z m at that stage. The objective is to maximize the total expected payoff. Clayton and Berry (1985) consider this problem when a m equals 1 for mn and 0 for m > n(< ∝) In this paper, the Clayton and Berry (1985) results are extended to the case of regular discount sequences of horizon n, which may also be infinite. The results are illustrated with numerical examples. In case of geometric discounting, the results apply to a bandit with many independent unknown Dirichlet arms.  相似文献   

2.
We consider the situation where there is a known regression model that can be used to predict an outcome, Y, from a set of predictor variables X . A new variable B is expected to enhance the prediction of Y. A dataset of size n containing Y, X and B is available, and the challenge is to build an improved model for Y| X ,B that uses both the available individual level data and some summary information obtained from the known model for Y| X . We propose a synthetic data approach, which consists of creating m additional synthetic data observations, and then analyzing the combined dataset of size n + m to estimate the parameters of the Y| X ,B model. This combined dataset of size n + m now has missing values of B for m of the observations, and is analyzed using methods that can handle missing data (e.g., multiple imputation). We present simulation studies and illustrate the method using data from the Prostate Cancer Prevention Trial. Though the synthetic data method is applicable to a general regression context, to provide some justification, we show in two special cases that the asymptotic variances of the parameter estimates in the Y| X ,B model are identical to those from an alternative constrained maximum likelihood estimation approach. This correspondence in special cases and the method's broad applicability makes it appealing for use across diverse scenarios. The Canadian Journal of Statistics 47: 580–603; 2019 © 2019 Statistical Society of Canada  相似文献   

3.
We investigate the construction of a BCa-type bootstrap procedure for setting approximate prediction intervals for an efficient estimator θm of a scalar parameter θ, based on a future sample of size m. The results are also extended to nonparametric situations, which can be used to form bootstrap prediction intervals for a large class of statistics. These intervals are transformation-respecting and range-preserving. The asymptotic performance of our procedure is assessed by allowing both the past and future sample sizes to tend to infinity. The resulting intervals are then shown to be second-order correct and second-order accurate. These second-order properties are established in terms of min(m, n), and not the past sample size n alone.  相似文献   

4.
The traditional non-parametric bootstrap (referred to as the n-out-of-n bootstrap) is a widely applicable and powerful tool for statistical inference, but in important situations it can fail. It is well known that by using a bootstrap sample of size m, different from n, the resulting m-out-of-n bootstrap provides a method for rectifying the traditional bootstrap inconsistency. Moreover, recent studies have shown that interesting cases exist where it is better to use the m-out-of-n bootstrap in spite of the fact that the n-out-of-n bootstrap works. In this paper, we discuss another case by considering its application to hypothesis testing. Two new data-based choices of m are proposed in this set-up. The results of simulation studies are presented to provide empirical comparisons between the performance of the traditional bootstrap and the m-out-of-n bootstrap, based on the two data-dependent choices of m, as well as on an existing method in the literature for choosing m. These results show that the m-out-of-n bootstrap, based on our choice of m, generally outperforms the traditional bootstrap procedure as well as the procedure based on the choice of m proposed in the literature.  相似文献   

5.
This paper focuses on bivariate kernel density estimation that bridges the gap between univariate and multivariate applications. We propose a subsampling-extrapolation bandwidth matrix selector that improves the reliability of the conventional cross-validation method. The proposed procedure combines a U-statistic expression of the mean integrated squared error and asymptotic theory, and can be used in both cases of diagonal bandwidth matrix and unconstrained bandwidth matrix. In the subsampling stage, one takes advantage of the reduced variability of estimating the bandwidth matrix at a smaller subsample size m (m < n); in the extrapolation stage, a simple linear extrapolation is used to remove the incurred bias. Simulation studies reveal that the proposed method reduces the variability of the cross-validation method by about 50% and achieves an expected integrated squared error that is up to 30% smaller than that of the benchmark cross-validation. It shows comparable or improved performance compared to other competitors across six distributions in terms of the expected integrated squared error. We prove that the components of the selected bivariate bandwidth matrix have an asymptotic multivariate normal distribution, and also present the relative rate of convergence of the proposed bandwidth selector.  相似文献   

6.
Point process models are a natural approach for modelling data that arise as point events. In the case of Poisson counts, these may be fitted easily as a weighted Poisson regression. Point processes lack the notion of sample size. This is problematic for model selection, because various classical criteria such as the Bayesian information criterion (BIC) are a function of the sample size, n, and are derived in an asymptotic framework where n tends to infinity. In this paper, we develop an asymptotic result for Poisson point process models in which the observed number of point events, m, plays the role that sample size does in the classical regression context. Following from this result, we derive a version of BIC for point process models, and when fitted via penalised likelihood, conditions for the LASSO penalty that ensure consistency in estimation and the oracle property. We discuss challenges extending these results to the wider class of Gibbs models, of which the Poisson point process model is a special case.  相似文献   

7.
8.
An exact filter is an algorithm for calculating the a-posteriori distribution of the state ξ n of a process, given observations ηt, …,ηnup to time n. We describe a method to determine an appropriate algorithm for processes, where the distributions involved are members of exponential families, The resulting algorithm consists essen tially of a prediction term, combined with an affine transformation depending on the chosen model.  相似文献   

9.
With the advent of modern technology, manufacturing processes have become very sophisticated; a single quality characteristic can no longer reflect a product's quality. In order to establish performance measures for evaluating the capability of a multivariate manufacturing process, several new multivariate capability (NMC) indices, such as NMC p and NMC pm , have been developed over the past few years. However, the sample size determination for multivariate process capability indices has not been thoroughly considered in previous studies. Generally, the larger the sample size, the more accurate an estimation will be. However, too large a sample size may result in excessive costs. Hence, the trade-off between sample size and precision in estimation is a critical issue. In this paper, the lower confidence limits of NMC p and NMC pm indices are used to determine the appropriate sample size. Moreover, a procedure for conducting the multivariate process capability study is provided. Finally, two numerical examples are given to demonstrate that the proper determination of sample size for multivariate process indices can achieve a good balance between sampling costs and estimation precision.  相似文献   

10.
A control procedure is presented for monitoring changes in variation for a multivariate normal process in a Phase II operation where the subgroup size, m, is less than p, the number of variates. The methodology is based on a form of Wilk' statistic, which can be expressed as a function of the ratio of the determinants of two separate estimates of the covariance matrix. One estimate is based on the historical data set from Phase I and the other is based on an augmented data set including new data obtained in Phase II. The proposed statistic is shown to be distributed as the product of independent beta distributions that can be approximated using either a chi-square or F-distribution. An ARL study of the statistic is presented for a range of conditions for the population covariance matrix. Cases are considered where a p-variate process is being monitored using a sample of m observations per subgroup and m < p. Data from an industrial multivariate process is used to illustrate the proposed technique.  相似文献   

11.
Making predictions of future realized values of random variables based on currently available data is a frequent task in statistical applications. In some applications, the interest is to obtain a two-sided simultaneous prediction interval (SPI) to contain at least k out of m future observations with a certain confidence level based on n previous observations from the same distribution. A closely related problem is to obtain a one-sided upper (or lower) simultaneous prediction bound (SPB) to exceed (or be exceeded) by at least k out of m future observations. In this paper, we provide a general approach for computing SPIs and SPBs based on data from a particular member of the (log)-location-scale family of distributions with complete or right censored data. The proposed simulation-based procedure can provide exact coverage probability for complete and Type II censored data. For Type I censored data, our simulation results show that our procedure provides satisfactory results in small samples. We use three applications to illustrate the proposed simultaneous prediction intervals and bounds.  相似文献   

12.
Many multivariate quality control techniques are used for multivariate variable processes, but few work for multivariate attribute processes. To monitor multivariate attributes, controlling the false alarms (type I errors) and considering the correlation between attributes are two important issues. By taking into account these two issues, a new control chart is presented to monitor a bivariate binomial process. An example is illustrated for the proposed method. To evaluate the performance of the proposed method, a simulation study is conducted to compare the results with those using both the multivariate np chart and skewness reduction approaches. The results show that the correlation is taken into account in the designed chart and the overall false alarm is controlled at the nominal value. Moreover, the process shift can be quickly detected and the variable that is responsible for a signal can be determined.  相似文献   

13.
Statistical process control tools have been used routinely to improve process capabilities through reliable on-line monitoring and diagnostic processes. In the present paper, we propose a novel multivariate control chart that integrates a support vector machine (SVM) algorithm, a bootstrap method, and a control chart technique to improve multivariate process monitoring. The proposed chart uses as the monitoring statistic the predicted probability of class (PoC) values from an SVM algorithm. The control limits of SVM-PoC charts are obtained by a bootstrap approach. A simulation study was conducted to evaluate the performance of the proposed SVM–PoC chart and to compare it with other data mining-based control charts and Hotelling's T 2 control charts under various scenarios. The results showed that the proposed SVM–PoC charts outperformed other multivariate control charts in nonnormal situations. Further, we developed an exponential weighed moving average version of the SVM–PoC charts for increasing sensitivity to small shifts.  相似文献   

14.
A discrete distribution associated with a pure birth process starting with no individuals, with birth rates λ n =λ forn=0, 2, …,m−1 and λ n forn≥m is considered in this paper. The probability mass function is expressed in terms of an integral that is very convenient for computing probabilities, moments, generating functions and others. Using this representation, the mean and the k-th factorial moments of the distribution are obtained. Some nice characterizations of this distribution are also given.  相似文献   

15.
In recent years, statistical process control (SPC) of multivariate and autocorrelated processes has received a great deal of attention. Modern manufacturing/service systems with more advanced technology and higher production rates can generate complex processes in which consecutive observations are dependent and each variable is correlated. These processes obviously violate the assumption of the independence of each observation that underlies traditional SPC and thus deteriorate the performance of its traditional tools. The popular way to address this issue is to monitor the residuals—the difference between the actual value and the fitted value—with the traditional SPC approach. However, this residuals-based approach requires two steps: (1) finding the residuals; and (2) monitoring the process. Also, an accurate prediction model is necessary to obtain the uncorrelated residuals. Furthermore, these residuals are not the original values of the observations and consequently may have lost some useful information about the targeted process. The main purpose of this article is to examine the feasibility of using one-class classification-based control charts to handle multivariate and autocorrelated processes. The article uses simulated data to present an analysis and comparison of one-class classification-based control charts and the traditional Hotelling's T 2 chart.  相似文献   

16.
A Galton-Watson process in varying environments (Zn), with essentially constant offspring means, i.e. E(Zn)/mn→α∈(0, ∞), and exactly two rates of growth is constructed. The underlying sample space Ω can be decomposed into parts A and B such that (Zn)n grows like 2non A and like mnon B (m > 4).  相似文献   

17.
!n this paper we consider the predicf an problem of the future nth record value based an the first m (m < n) observed record values from one-parameter exponential distribution. We introduce four procedures for obtaining prediction intervals for the nth record value. The performance of the so obtained intervals is assessed through numerical and simulation studies. In these studies, we provide the means and standard errors of lower limits. upper limits and lengths of prediction intervals. Further, we check the validation of these intervals based on some point predictors.  相似文献   

18.
The shape features of run chart patterns of the most recent m observations arising from stable and unstable processes are different. Using this fact, a new monitoring statistic is defined whose value for given m depends on the pattern parameters but not on the process parameters. A control chart for this statistic for given m, therefore, will be globally applicable to normal processes. The simulation study reveals that the proposed statistic approximately follows normal distribution. The performances of the globally applicable control chart in terms of average run lengths (ARLs) are evaluated and compared with the X chart. Both in-control ARL and out-of-control ARLs with respect to different abnormal process conditions are found to be larger than the X chart. However, the proposed concept is promising because it can eliminate the burden of designing separate control charts for different quality characteristics or processes in a manufacturing set-up.  相似文献   

19.
Traditional resampling methods for estimating sampling distributions sometimes fail, and alternative approaches are then needed. For example, if the classical central limit theorem does not hold and the naïve bootstrap fails, the m/n bootstrap, based on smaller-sized resamples, may be used as an alternative. An alternative to the naïve bootstrap, the sufficient bootstrap, which uses only the distinct observations in a bootstrap sample, is another recently proposed bootstrap approach that has been suggested to reduce the computational burden associated with bootstrapping. It works as long as naïve bootstrap does. However, if the naïve bootstrap fails, so will the sufficient bootstrap. In this paper, we propose combining the sufficient bootstrap with the m/n bootstrap in order to both regain consistent estimation of sampling distributions and to reduce the computational burden of the bootstrap. We obtain necessary and sufficient conditions for asymptotic normality of the proposed method, and propose new values for the resample size m. We compare the proposed method with the naïve bootstrap, the sufficient bootstrap, and the m/n bootstrap by simulation.  相似文献   

20.
ABSTRACT

This article deals with a distribution associated with a pure birth process starting with no individuals, with birth rates λ n  = λ for n = 0, 2,…, m ? 1 and λ n  = μ for n ≥ m. The probability mass function is expressed in terms of an integral that is very convenient for computing probabilities, moments, generating functions, and others. Using this representation, the kth factorial moments of the distribution is obtained. Some other forms of this distribution are also given.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号