首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Systematic sampling is the simplest and easiest of the most common sampling methods. However, when the population size N cannot be evenly divided by the sampling size n, systematic sampling cannot be performed. Not only is it difficult to determine the sampling interval k equivalent to the sampling probability of the sampling unit, but also the sample size will be inconstant and the sample mean will be a biased estimator of the population mean. To solve this problem, this paper introduces an improved method for systematic sampling: the remainder Markov systematic sampling method. This new method involves separately finding the first-order and second-order inclusion probabilities. This approach uses the Horvitz-Thompson estimator as an unbiased estimator of the population mean to find the variance of the estimator. This study examines the effectiveness of the proposed method for different super-populations.  相似文献   

2.
Samples of size n are drawn from a finite population on each of two occasions. On the first occasion a variate x is measured, and on the second a variate y. In estimating the population mean of y, the variance of the best linear unbiased combination of means for matched and unmatched samples is itself minimized, with respect to the sampling design on the second occasion, by a certain degree of matching. This optimal allocation depends on the population correlation coefficient, which previous authors have assumed known. We estimate the correlation from an initial matched sample, then an approximately optimal allocation is completed and an estimator formed which, under a bivariate normal superpopulation model, has model expected mean square error equal, apart from an error of order n-2, to the minimum enjoyed by any linear, unbiased estimator.  相似文献   

3.
The problem of estimating the sample size for a phase III trial on the basis of existing phase II data is considered, where data from phase II cannot be combined with those of the new phase III trial. Focus is on the test for comparing the means of two independent samples. A launching criterion is adopted in order to evaluate the relevance of phase II results: phase III is run if the effect size estimate is higher than a threshold of clinical importance. The variability in sample size estimation is taken into consideration. Then, the frequentist conservative strategies with a fixed amount of conservativeness and Bayesian strategies are compared. A new conservative strategy is introduced and is based on the calibration of the optimal amount of conservativeness – calibrated optimal strategy (COS). To evaluate the results we compute the Overall Power (OP) of the different strategies, as well as the mean and the MSE of sample size estimators. Bayesian strategies have poor characteristics since they show a very high mean and/or MSE of sample size estimators. COS clearly performs better than the other conservative strategies. Indeed, the OP of COS is, on average, the closest to the desired level; it is also the highest. COS sample size is also the closest to the ideal phase III sample size MI, showing averages and MSEs lower than those of the other strategies. Costs and experimental times are therefore considerably reduced and standardized. However, if the ideal sample size MI is to be estimated the phase II sample size n should be around the ideal phase III sample size, i.e. n?2MI/3. Copyright © 2010 John Wiley & Sons, Ltd.  相似文献   

4.
This paper is concerned with ranked set sampling theory which is useful to estimate the population mean when the order of a sample of small size can be found without measurements or with rough methods. Consider n sets of elements each set having size m. All elements of each set are ranked but only one is selected and quantified. The average of the quantified elements is adopted as the estimator. In this paper we introduce the notion of selective probability which is a generalization of a notion from Yanagawa and Shirahata (1976). Uniformly optimal unbiased procedures are found for some (n,m). Furthermore, procedures which are unbiased for all distributions and are good for symmetric distributions are studied for (n,m) which do not allow uniformly optimal unbiased procedures.  相似文献   

5.
Previous work has been carried out on the use of double sampling schemes for inference from binomial data which are subject to misclassification. The double sampling scheme utilizes a sample of n units which are classified by both a fallible and a true device and another sample of n2 units which are classified only by a fallible device. A triple sampljng scheme incorporates an additional sample of nl units which are classified only by the true device. In this paper we apply this triple sampling to estimation from binomialdata. First estimation of a binomial proportion is discussed under different misclassification structures. Then, the problem of optimal allocation of sample sizes is discussed.  相似文献   

6.
A fast general extension algorithm of Latin hypercube sampling (LHS) is proposed, which reduces the time consumption of basic general extension and preserves the most original sampling points. The extension algorithm starts with an original LHS of size m and constructs a new LHS of size m?+?n that remains the original points. This algorithm is the further research of basic general extension, which cost too much time to get the new LHS. During selecting the original sampling points to preserve, time consumption is cut from three aspects. The first measure of the proposed algorithm is to select isolated vertices and divide the adjacent matrix into blocks. Secondly, the relationship of original LHS structure and new LHS structure is discussed. Thirdly, the upper and lower bounds help reduce the time consumption. The proposed algorithm is applied for two functions to demonstrate the effectiveness.  相似文献   

7.
ABSTRACT

Recently, distance sampling emerged as an advantageous technique to estimate the abundance of many animal populations, including ungulates. Its basic design involves the random selection of several samplers (transects or points) within the population range, and a Horvitz–Thompson-like estimator is then applied to estimate the population abundance while correcting for animal detectability. Ensuring even coverage probability is essential for subsequent inference on the population size, but it may not be achievable because of limited access to parts of the population range. Moreover, in several environmental conditions, a random selection of samplers may induce very high survey costs because it does not minimize the displacement time of the observer(s) between successive samplers. We thus tested whether two-stage designs – based on the random selection of points and then of nearby samplers – could be more cost-effective, for a given population size and when even area coverage cannot be guaranteed. Here, we further extend our analyses to assess the performance of two-stage designs under varying animal densities.  相似文献   

8.
In order to reduce the effect of autocorrelation on the X¯ monitoring scheme, a new sampling strategy is proposed to form rational subgroup samples of size n. It requires sampling to be done such that: (i) observations from two consecutive samples are merged, and (ii) some consecutive observations are skipped before sampling. This technique which is a generalized version of the mixed samples strategy is shown to yield a better reduction of the negative effect of autocorrelation when monitoring the mean of processes with and without measurement errors. For processes subjected to a combined effect of autocorrelation and measurement errors, the proposed sampling technique, together with multiple measurement strategy, yields an uniformly better zero-state run-length performance than its two main existing competitors for any autocorrelation level. However, in steady-state mode, it yields the best performance only when the monitoring process is subject to a high level of autocorrelation, for any given level of measurement errors. A real life example is used to illustrate the implementation of the proposed sampling strategy.KEYWORDS: Autocorrelation, measurement errors, mixed samples strategy, multiple measurements, skipping sampling strategy, steady-state, zero-state  相似文献   

9.
A sample (X1 …, Xn) is drawn from a population of size N. Karlin (1974) conjectured that for any function ? in a certain class of real-valued functions on the sample space, ? is at least as large for sampling with replacement as for any other random replacement sampling plan. This conjecture is proved under the assumption that ?  相似文献   

10.
In this paper we examine the failure-censored sampling plans for the two–parameter exponential distri- bution based on m random samples, each of size n. The suggested procedure is based on exact results and only the first failure time of each sample is needed. The values of the acceptability constant are also tabulated for selected values of p α 1 p β 1, α and β. Further, a comparison of the proposed sampling plans with ordinary sampling plans using a sample of size mn is made. When compared to ordinary sampling plans, the proposed plan has an advantage in terms of shorter test-time and a saving of resources.  相似文献   

11.
Results in five areas of survey sampling dealing with the choice of the sampling design are reviewed. In Section 2, the results and discussions surrounding the purposive selection methods suggested by linear regression superpopulation models are reviewed. In Section 3, similar models to those in the previous section are considered; however, random sampling designs are considered and attention is focused on the optimal choice of πj. Then in Section 4, systematic sampling methods obtained under autocorrelated superpopulation models are reviewed. The next section examines minimax sampling designs. The work in the final section is based solely on the randomization. In Section 6 methods of sample selection which yield inclusion probabilities πj = n/N and πij = n(n - 1)/N(N - 1), but for which there are fewer than NCn possible samples, are mentioned briefly.  相似文献   

12.
In a number of experiments, such as destructive stress testings, sampling is conducted sequentially. In such experiments, in which destruction of sample units may be expensive, one may wonder if it is more economical to observe n lower record values than to observe n iid observations from the original distribution. In this paper, we establish some general results concerning the comparison of the amount of the Fisher information contained in n record values and inter-record times with that contained in n iid observations from the original distribution. Some specific common distributions are classified according to this criterion.  相似文献   

13.
We consider the adjustment, based upon a sample of size n, of collections of vectors drawn from either an infinite or finite population. The vectors may be judged to be either normally distributed or, more generally, second-order exchangeable. We develop the work of Goldstein and Wooff (1998) to show how the familiar univariate finite population corrections (FPCs) naturally generalise to individual quantities in the multivariate population. The types of information we gain by sampling are identified with the orthogonal canonical variable directions derived from a generalised eigenvalue problem. These canonical directions share the same co-ordinate representation for all sample sizes and, for equally defined individuals, all population sizes enabling simple comparisons between both the effects of different sample sizes and of different population sizes. We conclude by considering how the FPC is modified for multivariate cluster sampling with exchangeable clusters. In univariate two-stage cluster sampling, we may decompose the variance of the population mean into the sum of the variance of cluster means and the variance of the cluster members within clusters. The first term has a FPC relating to the sampling fraction of clusters, the second term has a FPC relating to the sampling fraction of cluster size. We illustrate how this generalises in the multivariate case. We decompose the variance into two terms: the first relating to multivariate finite population sampling of clusters and the second to multivariate finite population sampling within clusters. We solve two generalised eigenvalue problems to show how to generalise the univariate to the multivariate: each of the two FPCs attaches to one, and only one, of the two eigenbases.  相似文献   

14.
The present article deals with some methods for estimation of finite populations means in the presence of linear trend among the population values. As a result, we provided a strategy for the selection of sampling interval k for the case of circular systematic sampling, which ensures better estimator for the population mean compared to other choices of the sampling interval. This has been established based on empirical studies. Further we more, applied multiple random starts methods for selecting random samples for the case of linear systematic sampling and diagonal systematic sampling schemes. We also derived the explicit expressions for the variances and their estimates. The relative performances of simple random sampling, linear systematic sampling and diagonal systematic sampling schemes with single and multiple random starts are also assessed based on numerical examples.  相似文献   

15.
This study investigates the statistical properties of the adaptive Hotelling's T 2 charts with run rules in which the sample size and sampling interval are allowed to vary according on the current and past sampling points. The adaptive charts include variable sample size (VSS), variable sampling interval (VSI), and variable sample size and sampling interval (VSSI) charts. The adaptive Hotelling's T 2 charts with run rules are compared with the fixed sampling rate Hotelling's T 2 chart with run rules. The numerical results show that the VSS, VSI, and VSSI features improve the performance of the Hotelling's T 2 chart with run rules.  相似文献   

16.
We present a surprising though obvious result that seems to have been unnoticed until now. In particular, we demonstrate the equivalence of two well-known problems—the optimal allocation of the fixed overall sample size n among L strata under stratified random sampling and the optimal allocation of the H = 435 seats among the 50 states for apportionment of the U.S. House of Representatives following each decennial census. In spite of the strong similarity manifest in the statements of the two problems, they have not been linked and they have well-known but different solutions; one solution is not explicitly exact (Neyman allocation), and the other (equal proportions) is exact. We give explicit exact solutions for both and note that the solutions are equivalent. In fact, we conclude by showing that both problems are special cases of a general problem. The result is significant for stratified random sampling in that it explicitly shows how to minimize sampling error when estimating a total TY while keeping the final overall sample size fixed at n; this is usually not the case in practice with Neyman allocation where the resulting final overall sample size might be near n + L after rounding. An example reveals that controlled rounding with Neyman allocation does not always lead to the optimum allocation, that is, an allocation that minimizes variance.  相似文献   

17.
This paper deals with techniques for obtaining random point samples from spatial databases. We seek random points from a continuous domain (usually 2) which satisfy a spatial predicate that is represented in the database as a collection of polygons. Several applications of spatial sampling (e.g. environmental monitoring, agronomy, forestry, etc) are described. Sampling problems are characterized in terms of two key parameters: coverage (selectivity), and expected stabbing number (overlap). We discuss two fundamental approaches to sampling with spatial predicates, depending on whether we sample first or evaluate the predicate first. The approaches are described in the context of both quadtrees and R-trees, detailing the sample first, acceptance/rejection tree, and partial area tree algorithms. A sequential algorithm, the one-pass spatial reservoir algorithm is also described. The relative performance of the various sampling algorithms is compared and choice of preferred algorithms is suggested. We conclude with a short discussion of possible extensions.  相似文献   

18.
The use of a range estimator of the population standard deviation, sigma (σ), for determining sample sizes is discussed in this study. Standardized mean ranges (dn's), when divided into the ranges of sampling frames, provide estimates of the standard deviation of the population. These estimates can be used for determining sample sizes. The dn's are provided for seven different distributions for sampling frame sizes that range from 2 to 2000, For each of the seven distributions, functional relationships are developed such that dn = f(nSF) where nSF is the size of the sample frame. From these functions, dn's can be estimated for sampling frame sizes which are not presented in the study.  相似文献   

19.
Efficient stochastic algorithms are presented in order to simulate allele configurations distributed according to a family π A , 0<A<∞, of exchangeable sampling distributions arising in population genetics. Each distribution π A has two parameters n and k, the sample size and the number of alleles, respectively. For A→0, the distribution π A is induced from neutral sampling, whereas for A→∞, it is induced from Maxwell–Boltzmann sampling. Three different Monte Carlo methods (independent sampling procedures) are provided, based on conditioning, sequential methods and a generalization of Pitmans ‘Chinese restaurant process’. Moreover, an efficient Markov chain Monte Carlo method is provided. The algorithms are applied to the homozygosity test and to the Ewens–Watterson–Slatkin test in order to test the hypothesis of selective neutrality.  相似文献   

20.
Suppose particles are randomly distributed in a certain medium, powder or liquid, which is conceptually divided into N cells. Let pi denote the probability that a particle falls in the ith cell and Yi denote the number of particles in the ith cell. Assume that the joint probability function of the Yi follows a multinomial distribution with cell probabilities pi respectively. Take n (≤N) cells at random without replacement and put each of the cells separately through a mixing mechanism of dilution and swirl. These n cells constitute the first stage samples and the number of particles in these cells are not observable. Now conceptually divide each of n cells into M subcells of equal size and let Xij denote the number of particles in the jth subcell of the ith cell selected in the first stage; i=1,2,…,N and j=1,2,…,M. Consequently assume that the conditional joint probability function of the Xij given Yi=yi follows a multinomial distribution with equal cell probabilities. Now take m (≤M) subcells at random from each of the cells selected in the first stage sample. Assume that the numbers of particles in M×N subcells are observable. The properties of the estimator of the particle density per sample unit are investigated under the modified two-stage cluster sampling method. A laboratory experiment for Xanthan Gum Products is analyzed in order to examine the appropriateness of the model assumed in this paper.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号