首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
We restrict attention to a class of Bernoulli subset selection procedures which take observations one-at-a-time and can be compared directly to the Gupta-Sobel single-stage procedure. For the criterion of minimizing the expected total number of observations required to terminate experimentation, we show that optimal sampling rules within this class are not of practical interest. We thus turn to procedures which, although not optimal, exhibit desirable behavior with regard to this criterion. A procedure which employs a modification of the so-called least-failures sampling rule is proposed, and is shown to possess many desirable properties among a restricted class of Bernoulli subset selection procedures. Within this class, it is optimal for minimizing the number of observations taken from populations excluded from consideration following a subset selection experiment, and asymptotically optimal for minimizing the expected total number of observations required. In addition, it can result in substantial savings in the expected total num¬ber of observations required as compared to a single-stage procedure, thus it may be de¬sirable to a practitioner if sampling is costly or the sample size is limited.  相似文献   

2.
Wald and Wolfowitz (1948) have shown that the Sequential Probability Ratio Test (SPRT) for deciding between two simple hypotheses is, under very restrictive conditions, optimal in three attractive senses. First, it can be a Bayes-optimal rule. Second, of all level α tests having the same power, the test with the smallest joint-expected number of observations is the SPRT, where this expectation is taken jointly with respect to both data and prior over the two hypotheses. Third, the level α test needing the fewest conditional-expected number of observat ions is the SPRT, where this expectation is now taken with respect to the data conditional on either hypothesis being true. Principal among the strong restrictions is that sampling can proceed only in a one-at-a-time manner. In this paper, we relax some of the conditions and show that there are sequential procedures that strictly dominate the SPRT in all three senses. We conclude that the third type of optimality occurs rarely and that decision-makers are better served by looking for sequential procedures that possess the first two types of optimality. By relaxing the one-at-a-time sampling restriction, we obtain optimal (in the first two senses) variable-s ample-size- sequential probability ratio tests.  相似文献   

3.
Nonparametric control chart are presented for the problem of detecting changes in the process median (or mean), or changes in the process variability when samples are taken at regular time intervals. The proposed procedures are based on sign-test statistics computed for each sample, and are used in Shewhart and cumulative sum control charts. When the process is in control the run length distributions for the proposed nonparametric control charts do not depend on the distribution of the observations. An additional advantage of the non-parametric control charts is that the variance of the process does not need to be established in order to set up a control chart for the mean. Comparisons with the corresponding parametric control charts are presented. It is also shown that curtailed sampling plans can considerably reduce the expected number of observations used in the Shewhart control schemes based on the sign statistic.  相似文献   

4.
Statistics and Computing - We present an importance sampling algorithm that can produce realisations of Markovian epidemic models that exactly match observations, taken to be the number of a single...  相似文献   

5.
The sampling designs dependent on sample moments of auxiliary variables are well known. Lahiri (Bull Int Stat Inst 33:133–140, 1951) considered a sampling design proportionate to a sample mean of an auxiliary variable. Sing and Srivastava (Biometrika 67(1):205–209, 1980) proposed the sampling design proportionate to a sample variance while Wywiał (J Indian Stat Assoc 37:73–87, 1999) a sampling design proportionate to a sample generalized variance of auxiliary variables. Some other sampling designs dependent on moments of an auxiliary variable were considered e.g. in Wywiał (Some contributions to multivariate methods in, survey sampling. Katowice University of Economics, Katowice, 2003a); Stat Transit 4(5):779–798, 2000) where accuracy of some sampling strategies were compared, too.These sampling designs cannot be useful in the case when there are some censored observations of the auxiliary variable. Moreover, they can be much too sensitive to outliers observations. In these cases the sampling design proportionate to the order statistic of an auxiliary variable can be more useful. That is why such an unequal probability sampling design is proposed here. Its particular cases as well as its conditional version are considered, too. The sampling scheme implementing this sampling design is proposed. The inclusion probabilities of the first and second orders were evaluated. The well known Horvitz–Thompson estimator is taken into account. A ratio estimator dependent on an order statistic is constructed. It is similar to the well known ratio estimator based on the population and sample means. Moreover, it is an unbiased estimator of the population mean when the sample is drawn according to the proposed sampling design dependent on the appropriate order statistic.  相似文献   

6.
Transductive methods are useful in prediction problems when the training dataset is composed of a large number of unlabeled observations and a smaller number of labeled observations. In this paper, we propose an approach for developing transductive prediction procedures that are able to take advantage of the sparsity in the high dimensional linear regression. More precisely, we define transductive versions of the LASSO (Tibshirani, 1996) and the Dantzig Selector (Candès and Tao, 2007). These procedures combine labeled and unlabeled observations of the training dataset to produce a prediction for the unlabeled observations. We propose an experimental study of the transductive estimators that shows that they improve the LASSO and Dantzig Selector in many situations, and particularly in high dimensional problems when the predictors are correlated. We then provide non-asymptotic theoretical guarantees for these estimation methods. Interestingly, our theoretical results show that the Transductive LASSO and Dantzig Selector satisfy sparsity inequalities under weaker assumptions than those required for the “original” LASSO.  相似文献   

7.
Some properties of control procedures with variable sampling intervals (VSI) have been investigated in recent years by Amin, Renolds et al, and others. Such procedures have been shown to be more efficient when compared to the corresponding fixed sampling interval (FSI) charts with respect to the Average Time to Signal (ATS) when the Average Run Length (ARL) values for both types of procedures are held equal. Frequent switching between the different sampling intervals can be a complicating factor in the application of control charts with variable sampling intervals (VSI). This problem is being addressed in this article, and improved switching rules are presented and evaluated for Shewhart, CUSUM, and EWMA control procedures. The proposed rules considerably reduce the average number of switches between the sampling intervals and also improve the ATS properties of the control procedures when compared to the conventional variable sampling interval procedures  相似文献   

8.
In a two-sample testing problem, sometimes one of the sample observations are difficult and/or costlier to collect compared to the other one. Also, it may be the situation that sample observations from one of the populations have been previously collected and for operational advantages we do not wish to collect any more observations from the second population that are necessary for reaching a decision. Partially sequential technique is found to be very useful in such situations. The technique gained its popularity in statistics literature due to its very nature of capitalizing the best aspects of both fixed and sequential procedures. The literature is enriched with various types of partially sequential techniques useable under different types of data set-up. Nonetheless, there is no mention of multivariate data framework in this context, although very common in practice. The present paper aims at developing a class of partially sequential nonparametric test procedures for two-sample multivariate continuous data. For this we suggest a suitable stopping rule adopting inverse sampling technique and propose a class of test statistics based on the samples drawn using the suggested sampling scheme. Various asymptotic properties of the proposed tests are explored. An extensive simulation study is also performed to study the asymptotic performance of the tests. Finally the benefit of the proposed test procedure is demonstrated with an application to a real-life data on liver disease.  相似文献   

9.
Two-stage procedures are introduced to control the width and coverage (validity) of confidence intervals for the estimation of the mean, the between groups variance component and certain ratios of the variance components in one-way random effects models. The procedures use the pilot sample data to estimate an “optimal” group size and then proceed to determine the number of groups by a stopping rule. Such sampling plans give rise to unbalanced data, which are consequently analyzed by the harmonic mean method. Several asymptotic results concerning the proposed procedures are given along with simulation results to assess their performance in moderate sample size situations. The proposed procedures were found to effectively control the width and probability of coverage of the resulting confidence intervals in all cases and were also found to be robust in the presence of missing observations. From a practical point of view, the procedures are illustrated using a real data set and it is shown that the resulting unbalanced designs tend to require smaller sample sizes than is needed in a corresponding balanced design where the group size is arbitrarily pre-specified.  相似文献   

10.
Hypothesis Testing in Two-Stage Cluster Sampling   总被引:1,自引:0,他引:1  
Correlated observations often arise in complex sampling schemes such as two-stage cluster sampling. The resulting observations from this sampling scheme usually exhibit certain positive intracluster correlation, as a result of which the standard statistical procedures for testing hypotheses concerning linear combinations of the parameters may lack some of the optimal properties that these possess when the data are uncorrelated. The aim of this paper is to present exact methods for testing these hypotheses by combining within and between cluster information much as in Zhou & Mathew (1993).  相似文献   

11.
12.
Allocation of samples in stratified and/or multistage sampling is one of the central issues of sampling theory. In a survey of a population often the constraints for precision of estimators of subpopulations parameters have to be taken care of during the allocation of the sample. Such issues are often solved with mathematical programming procedures. In many situations it is desirable to allocate the sample, in a way which forces the precision of estimates at the subpopulations level to be both: optimal and identical, while the constraints of the total (expected) size of the sample (or samples, in two-stage sampling) are imposed. Here our main concern is related to two-stage sampling schemes. We show that such problem in a wide class of sampling plans has an elegant mathematical and computational solution. This is done due to a suitable definition of the optimization problem, which enables to solve it through a linear algebra setting involving eigenvalues and eigenvectors of matrices defined in terms of some population quantities. As a final result, we obtain a very simple and relatively universal method for calculating the subpopulation optimal and equal-precision allocation which is based on one of the most standard algorithms of linear algebra (available, e.g., in R software). Theoretical solutions are illustrated through a numerical example based on the Labour Force Survey. Finally, we would like to stress that the method we describe allows to accommodate quite automatically for different levels of precision priority for subpopulations.  相似文献   

13.
Single sampling plans are widely used for appraising incoming product quality. However, for situations where a continuous product flow exists, lot-by-lot demarcations may not exist, and it may be necessary to use alternate procedures, such as CSP-1, for continuous processes. In this case, one would like to be able to understand how average performance of the continuous sampling procedures compares to the more commonly used single sampling plans.

In this study, a model is devised which can be used to relate plan performance between single sample lot acceptance procedures and Dodge's(1943) CSP-1 continuous sampling plan. It is shown that it is generally not possible to match up performance based upon operating characteristic curve expressions for the two plans. Instead, the plans are matched by equating expressions for π(p), the long run proportion of product which is accepted, under both procedures. This is shown to be equivalent to matching up properties on an average outgoing quality basis. The methodology may be extended for any derivative plan under MIL-STD-1235B (1982), the military standard for continuous acceptance sampling.  相似文献   

14.
The main interest of prediction intervals lies in the results of a future sample from a previously sampled population. In this article, we develop procedures for the prediction intervals which contain all of a fixed number of future observations for general balanced linear random models. Two methods based on the concept of a generalized pivotal quantity (GPQ) and one based on ANOVA estimators are presented. A simulation study using the balanced one-way random model is conducted to evaluate the proposed methods. It is shown that one of the two GPQ-based and the ANOVA-based methods are computationally more efficient and they also successfully maintain the simulated coverage probabilities close to the nominal confidence level. Hence, they are recommended for practical use. In addition, one example is given to illustrate the applicability of the recommended methods.  相似文献   

15.
Shewhart, cumulative sum (CUSUM), and exponentially weighted moving average (EWMA) control procedures with variable sampling intervals (VSI) have been investigated in recent years for detecting shifts in the process mean. Such procedures have been shown to be more efficient when compared with the corresponding fixed sampling interval (FSI) charts with respect to the average time to signal (ATS) when the average run length (ARL) values of both types of procedures are held equal. Frequent switching between the different sampling intervals can be a complicating factor in the application of control charts with variable sampling intervals. In this article, we propose using a double exponentially weighted moving average control procedure with variable sampling intervals (VSI-DEWMA) for detecting shifts in the process mean. It is shown that the proposed VSI-DEWMA control procedure is more efficient when compared with the corresponding fixed sampling interval FSI-DEWMA chart with respect to the average time to signal (ATS) when the average run length (ARL) values of both types of procedures are held equal. It is also shown that the VSI-DEWMA procedure reduces the average number of switches between the sampling intervals and has similar ATS properties as compared to the VSI-EMTMA control procedure  相似文献   

16.
SUMMARY Ranked-set sampling is a widely used sampling procedure when sample observations are expensive or difficult to obtain. It departs from simple random sampling by seeking to spread the observations in the sample widely over the distribution or population. This is achieved by ranking methods which may need to employ concomitant information. The ranked-set sample mean is known to be more efficient than the corresponding simple random sample mean. Instead of the ranked-set sample mean, this paper considers the corresponding optimal estimator: the ranked-set best linear unbiased estimator. This is shown to be more efficient, even for normal data, but particularly for skew data, such as from an exponential distribution. The corresponding forms of the estimators are quite distinct from the ranked-set sample mean. Improvement holds where the ordering is perfect or imperfect, with this prospect of improper ordering being explored through the use of concomitants. In addition, the corresponding optimal linear estimator of a scale parameter is also discussed. The results are applied to a biological problem that involves the estimation of root weights for experimental plants, where the expense of measurement implies the need to minimize the number of observations taken.  相似文献   

17.
In longitudinal surveys where a number of observations have to be made on the same sampling unit at specified time intervals, it is not uncommon that observations for some of the time stages for some of the sampled units are found missing. In the present investigation an estimation procedure for estimating the population total based on such incomplete data from multiple observations is suggested which makes use of all the available information and is seen to be more efficient than the one based on only completely observed units. Estimators are also proposed for two other situations; firstly when data is collected only for a sample of time stages and secondly when data is observed for only one time stage per sampled unit.  相似文献   

18.
Many sampling problems from multiple populations can be considered under the semiparametric framework of the biased, or weighted, sampling model. Included under this framework is logistic regression under case–control sampling. For any model, atypical observations can greatly influence the maximum likelihood estimate of the parameters. Several robust alternatives have been proposed for the special case of logistic regression. However, some current techniques can exhibit poor behavior in many common situations. In this paper a new family of procedures are constructed to estimate the parameters in the semiparametric biased sampling model. The procedures incorporate a minimum distance approach, but are instead based on characteristic functions. The estimators can also be represented as the minimizers of quadratic forms in simple residuals, thus yielding straightforward computation. For the case of logistic regression, the resulting estimators are shown to be competitive with the existing robust approaches in terms of both robustness and efficiency, while maintaining affine equivariance. The approach is developed under the case–control sampling scheme, yet is shown to be applicable under prospective sampling logistic regression as well.  相似文献   

19.
Summary.  Traditional lot quality assurance sampling (LQAS) methods require simple random sampling to guarantee valid results. However, cluster sampling has been proposed to reduce the number of random starting points. This study uses simulations to examine the classification error of two such designs, a 67×3 (67 clusters of three observations) and a 33×6 (33 clusters of six observations) sampling scheme to assess the prevalence of global acute malnutrition (GAM). Further, we explore the use of a 67×3 sequential sampling scheme for LQAS classification of GAM prevalence. Results indicate that, for independent clusters with moderate intracluster correlation for the GAM outcome, the three sampling designs maintain approximate validity for LQAS analysis. Sequential sampling can substantially reduce the average sample size that is required for data collection. The presence of intercluster correlation can impact dramatically the classification error that is associated with LQAS analysis.  相似文献   

20.
This paper presents a set of new tables and procedures for the selection of the following three types of Quick Switching (QS) systems for a given Acceptable Quality Level (AQL), Limiting Quality Level (LQL), producer's risk and consumer's risk.

(1) A Single sampling QS system with equal sample sizes but with different acceptance numbers

(2) A Single sampling QS system with two different sample sizes but with same acceptance number and

(3) A QS systen with double sampling normal inspection and single sampling tightened inspection

The third type of QS systen is the one newly presented in this paper. The tables provide unique plans for a given set of conditions as well as providing a smaller sample size or a smaller sum of Average Sample Numbers(ASN) at the AQL and LQL  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号