首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
This paper deals with the asymptotics of a class of tests for association in 2-way contingency tables based on square forms in cell frequencies, given the total number of observations (multinomial sampling) or one set of marginal totals (stratified sampling). The case when both row and column marginal totals are fixed (hypergeometric sampling) was studied in Kulinskaya (1994), The class of tests under consideration includes a number of classical measures for association, Its two subclasses are the tests based on statistics using centralized cell frequencies (asymptotically distributed as weighted sums of central chi-squares) and those using the non-centralized cell frequencies (asymptotically normal). The parameters of asymptotic distributions depend on the sampling model and on true marginal probabilities. Maximum efficiency for asymptotically normal statistics is achieved under hypergeometric sampling, If the cell frequencies or the statistic as a whole are centralized using marginal proportions as estimates for marginal probabilities, the asymptotic distribution does not differ much between models and it is equivalent to that under hypergeometric sampling. These findings give an extra justification for the use of permutation tests for association (which are based on hypergeometric sampling). As an application, several well known measures of association are analysed.  相似文献   

2.
Order sampling with fixed distribution shape is a class of sampling schemes with inclusion probabilities approximately proportional to given size measures. In a recent article, methods were provided to compute the exact first and second order inclusion probabilities numerically when the distribution shape is of the Pareto type. In the same article, procedures were also provided for this case to adjust the parameters to get predetermined inclusion probabilities. In this paper we prove the existence and uniqueness of a solution for the latter problem, in general for any order sampling of fixed distribution shape.  相似文献   

3.
A method for constructing confidence limits for a distribution function is proposed. This method is a simple modification of the common method based on a normal approximation to the distribution of the estimated distribution function. The methods differ in how the estimated standard errors are used. The coverage properties of the two methods are compared in a simulation study. Coverage probabilities for the proposed method are found to be much closer to the nominal levels, particularly in the tails of the population distribution.  相似文献   

4.
This article develops a new generalized formula to compute the inclusion probabilities of a median-ranked set sample in a finite population setting. The use of this formula is illustrated in a numerical example. Furthermore, the inclusion probabilities of a median-ranked set sample is compared with the inclusion probabilities of ranked set and simple random samples.  相似文献   

5.
The purpose of acceptance sampling is to develop decision rules to accept or reject production lots based on sample data. When testing is destructive or expensive, dependent sampling procedures cumulate results from several preceding lots. This chaining of past lot results reduces the required size of the samples. A large part of these procedures only chain past lot results when defects are found in the current sample. However, such selective use of past lot results only achieves a limited reduction of sample sizes. In this article, a modified approach for chaining past lot results is proposed that is less selective in its use of quality history and, as a result, requires a smaller sample size than the one required for commonly used dependent sampling procedures, such as multiple dependent sampling plans and chain sampling plans of Dodge. The proposed plans are applicable for inspection by attributes and inspection by variables. Several properties of their operating characteristic-curves are derived, and search procedures are given to select such modified chain sampling plans by using the two-point method.  相似文献   

6.
Consideration of coverage yields a new class of estimators of population size for the standard mark-recapture model which permits heterogeneity of capture probabilities. Real data and simulation studies are used to assess these coverage-adjusted estimators. The simulations highlight the need for estimators that perform well for a wide range of values of the mean and coefficient of variation of the capture probabilities. When judged for this type of robustness, the simulations provide good grounds for preferring the new estimators to earlier ones for this model, except when the number of sampling occasions is large. A bootstrapping approach is used to estimate the standard errors of the new estimators, and to obtain confidence intervals for the population size.  相似文献   

7.
In proteomics, identification of proteins from complex mixtures of proteins extracted from biological samples is an important problem. Among the experimental technologies, mass spectrometry (MS) is the most popular one. Protein identification from MS data typically relies on a ‘two-step’ procedure of identifying the peptide first followed by the separate protein identification procedure next. In this setup, the interdependence of peptides and proteins is neglected resulting in relatively inaccurate protein identification. In this article, we propose a Markov chain Monte Carlo based Bayesian hierarchical model, a first of its kind in protein identification, which integrates the two steps and performs joint analysis of proteins and peptides using posterior probabilities. We remove the assumption of independence of proteins by using clustering group priors to the proteins based on the assumption that proteins sharing the same biological pathway are likely to be present or absent together and are correlated. The complete conditionals of the proposed joint model being tractable, we propose and implement a Gibbs sampling scheme for full posterior inference that provides the estimation and statistical uncertainties of all relevant parameters. The model has better operational characteristics compared to two existing ‘one-step’ procedures on a range of simulation settings as well as on two well-studied datasets.  相似文献   

8.
Poisson sampling is a method for unequal probabilities sampling with random sample size. There exist several implementations of the Poisson sampling design, with fixed sample size, which almost all are rejective methods, that is, the sample is not always accepted. Thus, the existing methods can be time-consuming or even infeasible in some situations. In this paper, a fast and non-rejective method, which is efficient even for large populations, is proposed and studied. The method is a new design for selecting a sample of fixed size with unequal inclusion probabilities. For the population of large size, the proposed design is very close to the strict πps sampling which is similar to the conditional Poisson (CP) sampling design, but the implementation of the design is much more efficient than the CP sampling. And the inclusion probabilities can be calculated recursively.  相似文献   

9.
In this work we propose Bayesian measures to quantify the influence of observations on the structural parameters of the simple measurement error model (MEM). Different influence measures, like those based on q-divergence between posterior distributions and Bayes risk, are studied to evaluate the influence. A strategy based on the perturbation function and MCMC samples is used to compute these measures. The samples from the posterior distributions are obtained by using the Metropolis-Hastings algorithm and assuming specific proper prior distributions. The results are illustrated with an application to a real example modeled with MEM in the literature.  相似文献   

10.
In outcome‐dependent sampling, the continuous or binary outcome variable in a regression model is available in advance to guide selection of a sample on which explanatory variables are then measured. Selection probabilities may either be a smooth function of the outcome variable or be based on a stratification of the outcome. In many cases, only data from the final sample is accessible to the analyst. A maximum likelihood approach for this data configuration is developed here for the first time. The likelihood for fully general outcome‐dependent designs is stated, then the special case of Poisson sampling is examined in more detail. The maximum likelihood estimator differs from the well‐known maximum sample likelihood estimator, and an information bound result shows that the former is asymptotically more efficient. A simulation study suggests that the efficiency difference is generally small. Maximum sample likelihood estimation is therefore recommended in practice when only sample data is available. Some new smooth sample designs show considerable promise.  相似文献   

11.
Item response theory (IRT) models are commonly used in educational and psychological testing to assess the (latent) ability of examinees and the effectiveness of the test items in measuring this underlying trait. The focus of this paper is on the assessment of item fit for unidimensional IRT models for dichotomous items using a Bayesian method. This paper will illustrate and compare the effectiveness of several discrepancy measures, used within the posterior predictive model check procedure, in detecting misfitted items. The effectiveness of the different discrepancy measures are illustrated in a simulation study using artificially altered simulated data. Using the best discrepancy measure among those studied, this method was applied to real data coming from a mathematics placement exam.  相似文献   

12.
This paper considers the multiple comparisons problem for normal variances. We propose a solution based on a Bayesian model selection procedure to this problem in which no subjective input is considered. We construct the intrinsic and fractional priors for which the Bayes factors and model selection probabilities are well defined. The posterior probability of each model is used as a model selection tool. The behaviour of these Bayes factors is compared with the Bayesian information criterion of Schwarz and some frequentist tests.  相似文献   

13.
Abstract

HYGARCH model is basically used to model long-range dependence in volatility. We propose Markov switch smooth-transition HYGARCH model, where the volatility in each state is a time-dependent convex combination of GARCH and FIGARCH. This model provides a flexible structure to capture different levels of volatilities and also short and long memory effects. The necessary and sufficient condition for the asymptotic stability is derived. Forecast of conditional variance is studied by using all past information through a parsimonious way. Bayesian estimations based on Gibbs sampling are provided. A simulation study has been given to evaluate the estimations and model stability. The competitive performance of the proposed model is shown by comparing it with the HYGARCH and smooth-transition HYGARCH models for some period of the S&P500 and Dow Jones industrial average indices based on volatility and value-at-risk forecasts.  相似文献   

14.
The purpose of this paper is to build a model for aggregate losses which constitutes a crucial step in evaluating premiums for health insurance systems. It aims at obtaining the predictive distribution of the aggregate loss within each age class of insured persons over the time horizon involved in planning employing the Bayesian methodology. The model proposed using the Bayesian approach is a generalization of the collective risk model, a commonly used model for analysing risk of an insurance system. Aggregate loss prediction is based on past information on size of loss, number of losses and size of population at risk. In modelling the frequency and severity of losses, the number of losses is assumed to follow a negative binomial distribution, individual loss sizes are independent and identically distributed exponential random variables, while the number of insured persons in a finite number of possible age groups is assumed to follow the multinomial distribution. Prediction of aggregate losses is based on the Gibbs sampling algorithm which incorporates the missing data approach.  相似文献   

15.
This article utilizes stochastic ideas for reasoning about association rule mining, and provides a formal statistical view of this discipline. A simple stochastic model is proposed, based on which support and confidence are reasonable estimates for certain probabilities of the model. Statistical properties of the corresponding estimators, like moments and confidence intervals, are derived, and items and itemsets are observed for correlations. After a brief review of measures of interest of association rules, with the main focus on interestingness measures motivated by statistical principles, two new measures are described. These measures, called α- and σ-precision, respectively, rely on statistical properties of the estimators discussed before. Experimental results demonstrate the effectivity of both measures.  相似文献   

16.
Abstract.  Typically, regression analysis for multistate models has been based on regression models for the transition intensities. These models lead to highly nonlinear and very complex models for the effects of covariates on state occupation probabilities. We present a technique that models the state occupation or transition probabilities in a multistate model directly. The method is based on the pseudo-values from a jackknife statistic constructed from non-parametric estimators for the probability in question. These pseudo-values are used as outcome variables in a generalized estimating equation to obtain estimates of model parameters. We examine this approach and its properties in detail for two special multistate model probabilities, the cumulative incidence function in competing risks and the current leukaemia-free survival used in bone marrow transplants. The latter is the probability a patient is alive and in either a first or second post-transplant remission. The techniques are illustrated on a dataset of leukaemia patients given a marrow transplant. We also discuss extensions of the model that are of current research interest.  相似文献   

17.
This article considers a k level step-stress accelerated life testing (ALT) on series system products, where independent Weibull-distributed lifetimes are assumed for the components. Due to cost considerations or environmental restrictions, causes of system failures are masked and type-I censored observations might occur in the collected data. Bayesian approach combined with auxiliary variables is developed for estimating the parameters of the model. Further, the reliability and hazard rate functions of the system and components are estimated at a specified time at use stress level. The proposed method is illustrated through a numerical example based on two priors and various masking probabilities.  相似文献   

18.
A finite mixture model is considered in which the mixing probabilities vary from observation to observation. A parametric model is assumed for one mixture component distribution, while the others are nonparametric nuisance parameters. Generalized estimating equations (GEE) are proposed for the semi‐parametric estimation. Asymptotic normality of the GEE estimates is demonstrated and the lower bound for their dispersion (asymptotic covariance) matrix is derived. An adaptive technique is developed to derive estimates with nearly optimal small dispersion. An application to the sociological analysis of voting results is discussed. The Canadian Journal of Statistics 41: 217–236; 2013 © 2013 Statistical Society of Canada  相似文献   

19.
Recursive computation of inclusion probabilities in ranked-set sampling   总被引:1,自引:0,他引:1  
We derive recursive algorithms for computing first-order and second-order inclusion probabilities for ranked-set sampling from a finite population. These algorithms make it practical to compute inclusion probabilities even for relatively large sample and population sizes. As an application, we use the inclusion probabilities to examine the performance of Horvitz-Thompson estimators under different varieties of balanced ranked-set sampling. We find that it is only for balanced Level 2 sampling that the Horvitz-Thompson estimator can be relied upon to outperform the simple random sampling mean estimator.  相似文献   

20.
This paper develops a new class of option price models and applies it to options on the Australian S&P200 Index. The class of models generalizes the traditional Black‐Scholes framework by accommodating time‐varying conditional volatility, skewness and excess kurtosis in the underlying returns process. An important property of these more general pricing models is that the computational requirements are essentially the same as those associated with the Black‐Scholes model, with both methods being based on one‐dimensional integrals. Bayesian inferential methods are used to evaluate a range of models nested in the general framework, using observed market option prices. The evaluation is based on posterior parameter distributions, as well as posterior model probabilities. Various fit and predictive measures, plus implied volatility graphs, are also used to rank the alternative models. The empirical results provide evidence that time‐varying volatility, leptokurtosis and a small degree of negative skewness are priced in Australian stock market options.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号