首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 750 毫秒
1.
Classes of higher-order kernels for estimation of a probability density are constructed by iterating the twicing procedure. Given a kernel K of order l, we build a family of kernels Km of orders l(m + 1) with the attractive property that their Fourier transforms are simply 1 — {1 —$(.)}m+1, where ? is the Fourier transform of K. These families of higher-order kernels are well suited when the fast Fourier transform is used to speed up the calculation of the kernel estimate or the least-squares cross-validation procedure for selection of the window width. We also compare the theoretical performance of the optimal polynomial-based kernels with that of the iterative twicing kernels constructed from some popular second-order kernels.  相似文献   

2.
Fantasy sports, particularly the daily variety in which new lineups are selected each day, are a rapidly growing industry. The two largest companies in the daily fantasy business, DraftKings and Fanduel, have been valued as high as $2 billion. This research focuses on the development of a complete system for daily fantasy basketball, including both the prediction of player performance and the construction of a team. First, a Bayesian random effects model is used to predict an aggregate measure of daily NBA player performance. The predictions are then used to construct teams under the constraints of the game, typically related to a fictional salary cap and player positions. Permutation based and K-nearest neighbors approaches are compared in terms of the identification of “successful” teams—those who would be competitive more often than not based on historical data. We demonstrate the efficacy of our system by comparing our predictions to those from a well-known analytics website, and by simulating daily competitions over the course of the 2015–2016 season. Our results show an expected profit of approximately $9,000 on an initial $500 investment using the K-nearest neighbors approach, a 36% increase relative to using the permutation-based approach alone. Supplementary materials for this article are available online.  相似文献   

3.
In this article the problem of the optimal selection and allocation of time points in repeated measures experiments is considered. D‐ optimal designs for linear regression models with a random intercept and first order auto‐regressive serial correlations are computed numerically and compared with designs having equally spaced time points. When the order of the polynomial is known and the serial correlations are not too small, the comparison shows that for any fixed number of repeated measures, a design with equally spaced time points is almost as efficient as the D‐ optimal design. When, however, there is no prior knowledge about the order of the underlying polynomial, the best choice in terms of efficiency is a D‐ optimal design for the highest possible relevant order of the polynomial. A design with equally‐spaced time points is the second best choice  相似文献   

4.
The purpose of this article is to present the optimal designs based on D-, G-, A-, I-, and D β-optimality criteria for random coefficient regression (RCR) models with heteroscedastic errors. A sufficient condition for the heteroscedastic structure is given to make sure that the search of optimal designs can be confined at extreme settings of the design region when the criteria satisfy the assumption of the real valued monotone design criteria. Analytical solutions of D-, G-, A-, I-, and D β-optimal designs for the RCR models are derived. Two examples are presented for random slope models with specific heteroscedastic errors.  相似文献   

5.
Abstract

When we consider the improvement of the functional performances that are released by the new updates of the products, it is an interesting problem to revisit the existing replacement policies. For such a viewpoint, four replacement models with product update announcements, i.e., PUA for abbreviation, are given in this paper: Model 1, the unit is replaced at time T or at PUA over time T. Model 2, the unit is replaced at the Kth failure or at PUA over the Kth failure. By considering both time T and failure K, Models 3 and 4 are obtained based on the approaches of replacement first and last. We obtain the expected cost rates for four models and discuss analytically their optimal replacement policies Further, numerical examples are given when the time for PUA has an exponential distribution.  相似文献   

6.
Abstract

K-means inverse regression was developed as an easy-to-use dimension reduction procedure for multivariate regression. This approach is similar to the original sliced inverse regression method, with the exception that the slices are explicitly produced by a K-means clustering of the response vectors. In this article, we propose K-medoids clustering as an alternative clustering approach for slicing and compare its performance to K-means in a simulation study. Although the two methods often produce comparable results, K-medoids tends to yield better performance in the presence of outliers. In addition to isolation of outliers, K-medoids clustering also has the advantage of accommodating a broader range of dissimilarity measures, which could prove useful in other graphical regression applications where slicing is required.  相似文献   

7.
The trend test is often used for the analysis of 2×K ordered categorical data, in which K pre-specified increasing scores are used. There have been discussions on how to assign these scores and the impact of the outcomes on different scores. The scores are often assigned based on the data-generating model. When this model is unknown, using the trend test is not robust. We discuss the weighted average of a trend test over all scientifically plausible choices of scores or models. This approach is more computationally efficient than a commonly used robust test MAX when K is large. Our discussion is for any ordered 2×K table, but simulation and applications to real data are focused on case-control genetic association studies. Although there is no single test optimal for all choices of scores, our numerical results show that some score averaging tests can achieve the performance of MAX.  相似文献   

8.
Suppose it is desired to obtain a large number Ns of items for which individual counting is impractical, but one can demand a batch to weigh at least w units so that the number of items N in the batch may be close to the desired number Ns. If the items have mean weight ωTH, it is reasonable to have w equal to ωTHNs when ωTH is known. When ωTH is unknown, one can take a sample of size n, not bigger than Ns, estimate ωTH by a good estimator ωn, and set w equal to ωnNs. Let Rn = Kp2N2s/n + Ksn be a measure of loss, where Ke and Ks are the coefficients representing the cost of the error in estimation and the cost of the sampling respectively, and p is the coefficient of variation for the weight of the items. If one determines the sample size to be the integer closest to pCNs when p is known, where C is (Ke/Ks)1/2, then Rn will be minimized. If p is unknown, a simple sequential procedure is proposed for which the average sample number is shown to be asymptotically equal to the optimal fixed sample size. When the weights are assumed to have a gamma distribution given ω and ω has a prior inverted gamma distribution, the optimal sample size can be found to be the nonnegative integer closest to pCNs + p2A(pC – 1), where A is a known constant given in the prior distribution.  相似文献   

9.
Approximations to the distribution of a discrete random variable originating from the classical occupancy problem are explored. The random variable X of interest is defined to be how many of N elements selected by or assigned to K individuals when each of the N elements is equally likely to be chosen by or assigned to any of the K individuals. Assuming that N represents the number of cells and each of the K individuals is placed in exactly one of the cells, X can also be defined as the number of cells occupied by the Kindividuals. In the literature, various asymptotic results for the distributions of X and (N ? X) are given; however, no guidelines are specified with respect to their utilization. In this article, these approximations are explored for various values of K and N, and rules of thumb are given for their appropriate use.  相似文献   

10.
In this article, we investigate the precise large deviations for a sum of independent but not identical distributed random variables. {X n , n ≥ 1} are independent non-negative random variables with distribution functions {F n , n ≥ 1}. We assume that the average of right tails of distribution functions F n is equivalent to some distribution function F with consistently varying tails. In applications, we apply our main results to a realistic example (Pareto-type distribution) and obtain a specific result.  相似文献   

11.
From the economical viewpoint in reliability theory, this paper addresses a scheduling replacement problem for a single operating system which works at random times for multiple jobs. The system is subject to stochastic failure which results the imperfect maintenance activity based on some random failure mechanism: minimal repair due to type-I (repairable) failure, or corrective replacement due to type-II (non-repairable) failure. Three scheduling models for the system with multiple jobs are considered: a single work, N tandem works, and N parallel works. To control the deterioration process, the preventive replacement is planned to undergo at a scheduling time T or the job's completion time of for each model. The objective is to determine the optimal scheduling parameters (T* or N*) that minimizes the mean cost rate function in a finite time horizon for each model. A numerical example is provided to illustrate the proposed analytical model. Because the framework and analysis are general, the proposed models extend several existing results.  相似文献   

12.
This paper is concerned with the pricing of American options by simulation methods. In the traditional methods, in order to determine when to exercise, we have to store the simulated asset prices at all time steps on all paths. If N time steps and M paths are used, then the storage requirement is O(MN). In this paper, we present a simulation method for pricing American options where the number of storage required only grows like O(M). The only additional computational cost is that we have to generate each random number twice instead of once. For machines with limited memory, we can now use a larger N to improve the accuracy in pricing the options.  相似文献   

13.
In this article, we introduce a new class of estimators called the sK type principal components estimators to combat multicollinearity, which include the principal components regression (PCR) estimator, the rk estimator and the sK estimator as special cases. Necessary and sufficient conditions for the superiority of the new estimator over the PCR estimator, the rk estimator and the sK estimator are derived in the sense of the mean squared error matrix criterion. A Monte Carlo simulation study and a numerical example are given to illustrate the performance of the proposed estimator.  相似文献   

14.
The Fisher exact test has been unjustly dismissed by some as ‘only conditional,’ whereas it is unconditionally the uniform most powerful test among all unbiased tests, tests of size α and with power greater than its nominal level of significance α. The problem with this truly optimal test is that it requires randomization at the critical value(s) to be of size α. Obviously, in practice, one does not want to conclude that ‘with probability x the we have a statistical significant result.’ Usually, the hypothesis is rejected only if the test statistic's outcome is more extreme than the critical value, reducing the actual size considerably.

The randomized unconditional Fisher exact is constructed (using Neyman–structure arguments) by deriving a conditional randomized test randomizing at critical values c(t) by probabilities γ(t), that both depend on the total number of successes T (the complete-sufficient statistic for the nuisance parameter—the common success probability) conditioned upon.

In this paper, the Fisher exact is approximated by deriving nonrandomized conditional tests with critical region including the critical value only if γ (t) > γ0, for a fixed threshold value γ0, such that the size of the unconditional modified test is for all value of the nuisance parameter—the common success probability—smaller, but as close as possible to α. It will be seen that this greatly improves the size of the test as compared with the conservative nonrandomized Fisher exact test.

Size, power, and p value comparison with the (virtual) randomized Fisher exact test, and the conservative nonrandomized Fisher exact, Pearson's chi-square test, with the more competitive mid-p value, the McDonald's modification, and Boschloo's modifications are performed under the assumption of two binomial samples.  相似文献   

15.
In this paper, we propose two SUR type estimators based on combining the SUR ridge regression and the restricted least squares methods. In the sequel these estimators are designated as the restricted ridge Liu estimator and the restricted ridge HK estimator (see Liu in Commun Statist Thoery Methods 22(2):393–402, 1993; Sarkar in Commun Statist A 21:1987–2000, 1992). The study has been made using Monte Carlo techniques, (1,000 replications), under certain conditions where a number of factors that may effect their performance have been varied. The performance of the proposed and some of the existing estimators are evaluated by means of the TMSE and the PR criteria. Our results indicate that the proposed SUR restricted ridge estimators based on K SUR, K Sratio, K Mratio and [(K)\ddot]{\ddot{K}} produced smaller TMSE and/or PR values than the remaining estimators. In contrast with other ridge estimators, components of [(K)\ddot]{\ddot{K}} are defined in terms of the eigenvalues of X* X*{X^{{\ast^{\prime}}} X^{\rm \ast}} and all lie in the open interval (0, 1).  相似文献   

16.
Consider observations (representing lifelengths) taken on a random field indexed by lattice points. Estimating the distribution function F(x) = P(X i  ≤ x) is an important problem in survival analysis. We propose to estimate F(x) by kernel estimators, which take into account the smoothness of the distribution function. Under some general mixing conditions, our estimators are shown to be asymptotically unbiased and consistent. In addition, the proposed estimator is shown to be strongly consistent and sharp rates of convergence are obtained.  相似文献   

17.
Multivariate combination-based permutation tests have been widely used in many complex problems. In this paper we focus on the equipower property, derived directly from the finite-sample consistency property, and we analyze the impact of the dependency structure on the combined tests. At first, we consider the finite-sample consistency property which assumes that sample sizes are fixed (and possibly small) and considers on each subject a large number of informative variables. Moreover, since permutation test statistics do not require to be standardized, we need not assume that data are homoscedastic in the alternative. The equipower property is then derived from these two notions: consider the unconditional permutation power of a test statistic T for fixed sample sizes, with V ? 2 independent and identically distributed variables and fixed effect δ, calculated in two ways: (i) by considering two V-dimensional samples sized m1 and m2, respectively; (ii) by considering two unidimensional samples sized n1 = Vm1 and n2 = Vm2, respectively. Since the unconditional power essentially depends on the non centrality induced by T, and two ways are provided with exactly the same likelihood and the same non centrality, we show that they are provided with the same power function, at least approximately. As regards both investigating the equipower property and the power behavior in presence of correlation we performed an extensive simulation study.  相似文献   

18.
19.
We present a surprising though obvious result that seems to have been unnoticed until now. In particular, we demonstrate the equivalence of two well-known problems—the optimal allocation of the fixed overall sample size n among L strata under stratified random sampling and the optimal allocation of the H = 435 seats among the 50 states for apportionment of the U.S. House of Representatives following each decennial census. In spite of the strong similarity manifest in the statements of the two problems, they have not been linked and they have well-known but different solutions; one solution is not explicitly exact (Neyman allocation), and the other (equal proportions) is exact. We give explicit exact solutions for both and note that the solutions are equivalent. In fact, we conclude by showing that both problems are special cases of a general problem. The result is significant for stratified random sampling in that it explicitly shows how to minimize sampling error when estimating a total TY while keeping the final overall sample size fixed at n; this is usually not the case in practice with Neyman allocation where the resulting final overall sample size might be near n + L after rounding. An example reveals that controlled rounding with Neyman allocation does not always lead to the optimum allocation, that is, an allocation that minimizes variance.  相似文献   

20.
Let X1, X2,… be an independently and identically distributed sequence with ξX1 = 0, ξ exp (tX1 < ∞ (t ≧ 0) and partial sums Sn = X1 + … + Xn. Consider the maximum increment D1 (N, K) = max0≤nN - K (Sn + K - Sn)of the sequence (Sn) in (0, N) over a time K = KN, 1 ≦ KN. Under appropriate conditions on (KN) it is shown that in the case KN/log N → 0, but KN/(log N)1/2 → ∞, there exists a sequence (αN) such that K-1/2 D1 (N, K) - αN converges to 0 w. p. 1. This result provides a small increment analogue to the improved Erd?s-Rényi-type laws stated by Csörg? and Steinebach (1981).  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号