首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 609 毫秒
1.
ABSTRACT

In many real-world applications, the traditional theory of analysis of covariance (ANCOVA) leads to inadequate and unreliable results because of violation of the response variable observations from the essential Gaussian assumption that may be due to the heterogeneity of population, the presence of outlier or both of them. In this paper, we develop a Gaussian mixture ANCOVA model for modelling heterogeneous populations with a finite number of subpopulation. We provide the maximum likelihood estimates of the model parameters via an EM algorithm. We also drive the adjusted effects estimators for treatments and covariates. The Fisher information matrix of the model and asymptotic confidence intervals for the parameter are also discussed. We performed a simulation study to assess the performance of the proposed model. A real-world example is also worked out to explained the methodology.  相似文献   

2.
In this paper, a new nonparametric methodology is developed for testing whether the changing pattern of a response variable over multiple ordered sub-populations from one treatment group differs with the one from another treatment group. The question is formalized into a nonparametric two-sample comparison problem for the stochastic order among subsamples, through U-statistics with accommodations for zero-inflated distributions. A novel bootstrap procedure is proposed to obtain the critical values with given type I error. Following the procedure, bootstrapped p-values are obtained through simulated samples. It is proven that the distribution of the test statistics is independent from the underlying distributions of the subsamples, when certain sufficient statistics provided. Furthermore, this study also develops a feasible framework for power studies to determine sample sizes, which is necessary in real-world applications. Simulation results suggest that the test is consistent. The methodology is illustrated using a biological experiment with a split-plot design, and significant differences in changing patterns of seed weight between treatments are found with relative small subsample sizes.  相似文献   

3.
Abstract

In this article we examine the functional central limit theorem for the first passage time of reward processes defined over a finite state space semi-Markov process. In order to apply this process for a wider range of real-world applications, the reward functions, considered in this work, are assumed to have general forms instead of the constant rates reported in the other studies. We benefit from the martingale theory and Poisson equations to prove and establish the convergence of the first passage time of reward processes to a zero mean Brownian motion. Necessary conditions to derive the results presented in this article are the existence of variances for sojourn times in each state and second order integrability of reward functions with respect to the distribution of sojourn times. We finally verify the presented methodology through a numerical illustration.  相似文献   

4.
ABSTRACT

Seasonal autoregressive (SAR) models have been modified and extended to model high frequency time series characterized by exhibiting double seasonal patterns. Some researchers have introduced Bayesian inference for double seasonal autoregressive (DSAR) models; however, none has tackled the problem of Bayesian identification of DSAR models. Therefore, in order to fill this gap, we present a Bayesian methodology to identify the order of DSAR models. Assuming the model errors are normally distributed and using three priors, i.e. natural conjugate, g, and Jeffreys’ priors, on the model parameters, we derive the joint posterior mass function of the model order in a closed-form. Accordingly, the posterior mass function can be investigated and the best order of DSAR model is chosen as a value with the highest posterior probability for the time series being analyzed. We evaluate the proposed Bayesian methodology using simulation study, and we then apply it to real-world hourly internet amount of traffic dataset.  相似文献   

5.
ABSTRACT

Large sample properties of Life-Table estimator are discussed for interval censored bivariate survival data. We restrict our attention to the situation where response times within pairs are not distinguishable, and the univariate survival distribution is the same for any individual within any pair. The large sample properties are applied to test for equality of two distributions with correlated response times where treatments are applied to different independent sets of cohorts. Data, which can be separated into two independent sets, from an angioplasty study where more than one procedure is performed on some patients are used to illustrate this methodology.  相似文献   

6.
ABSTRACT

We consider the variance estimation in a general nonparametric regression model with multiple covariates. We extend difference methods to the multivariate setting by introducing an algorithm that orders the design points in higher dimensions. We also consider an adaptive difference estimator which requires much less strict assumptions on the covariate design and can significantly reduce mean squared error for small sample sizes.  相似文献   

7.
ABSTRACT

Online consumer product ratings data are increasing rapidly. While most of the current graphical displays mainly represent the average ratings, Ho and Quinn proposed an easily interpretable graphical display based on an ordinal item response theory (IRT) model, which successfully accounts for systematic interrater differences. Conventionally, the discrimination parameters in IRT models are constrained to be positive, particularly in the modeling of scored data from educational tests. In this article, we use real-world ratings data to demonstrate that such a constraint can have a great impact on the parameter estimation. This impact on estimation was explained through rater behavior. We also discuss correlation among raters and assess the prediction accuracy for both the constrained and the unconstrained models. The results show that the unconstrained model performs better when a larger fraction of rater pairs exhibit negative correlations in ratings.  相似文献   

8.

This work is motivated by the need to find experimental designs which are robust under different model assumptions. We measure robustness by calculating a measure of design efficiency with respect to a design optimality criterion and say that a design is robust if it is reasonably efficient under different model scenarios. We discuss two design criteria and an algorithm which can be used to obtain robust designs. The first criterion employs a Bayesian-type approach by putting a prior or weight on each candidate model and possibly priors on the corresponding model parameters. We define the first criterion as the expected value of the design efficiency over the priors. The second design criterion we study is the minimax design which minimizes the worst value of a design criterion over all candidate models. We establish conditions when these two criteria are equivalent when there are two candidate models. We apply our findings to the area of accelerated life testing and perform sensitivity analysis of designs with respect to priors and misspecification of planning values.  相似文献   

9.
We present a methodology for screening predictors that, given the response, follow a one-parameter exponential family distributions. Screening predictors can be an important step in regressions when the number of predictors p is excessively large or larger than n the number of observations. We consider instances where a large number of predictors are suspected irrelevant for having no information about the response. The proposed methodology helps remove these irrelevant predictors while capturing those linearly or nonlinearly related to the response.  相似文献   

10.
Classical time-series theory assumes values of the response variable to be ‘crisp’ or ‘precise’, which is quite often violated in reality. However, forecasting of such data can be carried out through fuzzy time-series analysis. This article presents an improved method of forecasting based on LR fuzzy sets as membership functions. As an illustration, the methodology is employed for forecasting India's total foodgrain production. For the data under consideration, superiority of proposed method over other competing methods is demonstrated in respect of modelling and forecasting on the basis of mean square error and average relative error criteria. Finally, out-of-sample forecasts are also obtained.  相似文献   

11.
Abstract

Designs for the first order trigonometric regression model over an interval on the real line are considered for the situation where estimation of the slope of the response surface at various points in the factor space is of primary interest. Minimization of the variance of the estimated slope at a point maximized over all points in the region of interest is taken as the design criterion. Optimal designs under the minimax criterion are derived for the situation where the design region and the region of interest are identical and a symmetric “partial cycle”. Some comparisons of the minimax designs with the traditional D- and A-optimal designs are provided. Efficiencies of some exact designs under the minimax criterion are also investigated.  相似文献   

12.
The purpose of this paper is to discuss response surface designs for multivariate generalized linear models (GLMs). Such models are considered whenever several response variables can be measured for each setting of a group of control variables, and the response variables are adequately represented by GLMs. The mean-squared error of prediction (MSEP) matrix is used to assess the quality of prediction associated with a given design. The MSEP incorporates both the prediction variance and the prediction bias, which results from using maximum likelihood estimates of the parameters of the fitted linear predictor. For a given design, quantiles of a scalar-valued function of the MSEP are obtained within a certain region of interest. The quantiles depend on the unknown parameters of the linear predictor. The dispersion of these quantiles over the space of the unknown parameters is determined and then depicted by the so-called quantile dispersion graphs. An application of the proposed methodology is presented using the special case of the bivariate binary distribution.  相似文献   

13.
Abstract

Robust parameter design (RPD) is an effective tool, which involves experimental design and strategic modeling to determine the optimal operating conditions of a system. The usual assumptions of RPD are that normally distributed experimental data and no contamination due to outliers. And generally the parameter uncertainties in response models are neglected. However, using normal theory modeling methods for a skewed data and ignoring parameter uncertainties can create a chain of degradation in optimization and production phases such that misleading fit, poor estimated optimal operating conditions, and poor quality products. This article presents a new approach based on confidence interval (CI) response modeling for the process mean. The proposed interval robust design makes the system median unbiased for the mean and uses midpoint of the interval as a measure of location performance response. As an alternative robust estimator for the process variance response modeling, using biweight midvariance is proposed which is both resistant and robust of efficiency where normality is not met. The results further show that the proposed interval robust design gives a robust solution to the skewed structure of the data and to contaminated data. The procedure and its advantages are illustrated using two experimental design studies.  相似文献   

14.
ABSTRACT

Motivated by some recent improvements for mean estimation in finite sampling theory, we propose, in a design-based approach, a new class of ratio-type estimators. The class is initially discussed on the assumption that the study variable has a nonsensitive nature, meaning that it deals with topics that do not generate embarrassment when respondents are directly questioned about them. Under this standard setting, some estimators belonging to the class are shown and the bias, mean square error and minimum mean square error are determined up to the first-order of approximation. The class is subsequently extended to the case where the study variable refers to sensitive issues which produce measurement errors due to nonresponses and/or untruthful reporting. These errors may be reduced by enhancing respondent cooperation through scrambled response methods that mask the true value of the sensitive variable. Hence, four methods (say the additive, multiplicative, mixed and combined additive-multiplicative methods) are discussed for the purposes of the article. Finally, a simulation study is carried out to assess the performance of the proposed class by comparing a number of competing estimators, both in the sensitive and the nonsensitive setting.  相似文献   

15.
The construction of a balanced incomplete block design (BIBD) is formulated in terms of combinatorial optimization by defining a cost function that reaches its lower bound on all and only those configurations corresponding to a BIBD. This cost function is a linear combination of distribution measures for each of the properties of a block design (number of plots, uniformity of rows, uniformity of columns, and balance). The approach generalizes naturally to a super-class BIBDs, which we call maximally balanced maximally uniform designs (MBMUDs), that allow two consecutive values for their design parameters [r,r+1;k,k+1;λ,λ+1]. In terms of combinatorial balance, MBMUDs are the closest possible approximation to BIBDs for all experimental settings where no set of admissible parameters exists. Thus, other design classes previously proposed with the same approximation aim—such as RDGs, SRDGs and NBIBDs of type I—can be viewed as particular cases of MBMUDs. Interestingly, experimental results show that the proposed combinatorial cost function has a monotonic relation with A- and D-statistical optimality in the space of designs with uniform rows and columns, while its computational cost is much lower.  相似文献   

16.
Arguments about using computer facilities in classroom teaching have received a lot of attention over time. Using the computer facilities will be helpful to demonstrate real-world applications, while poor data or inappropriate case studies might hinder the applications of the computer programs in classroom teaching. In this paper, we examine the impacts that using computer programs to teach business statistics have on students in the Krannert School of Management at Purdue University. The results show that students are attracted to the interactive computer programs designed for the business statistics course, and students are more motivated to attend classes when computer programs are applied in teaching. Furthermore, computer programs help students to understand confusing topics, and students feel that teaching them to use computer facilities really improves their own abilities to apply similar programs in analyzing real-world problems.  相似文献   

17.

A computer program that performs ridge analysis on quadratic response surfaces is presented in this paper, the primary goal of which is to seek the estimated optimum operating conditions inside a spherical region of experimentation during the stage of process optimization. The computational algorithm is developed based upon the trust-region methods in nonlinear optimization and guarantees the resulting operating conditions to be globally optimal without any priori assumption on the structure of response functions. Under a particular condition termed the "hard case" arising from the trust region literature, the conventional ridge analysis procedure fails to provide a set of acceptable optimum operating settings, yet the proposed algorithm has the capability of locating a pair of non-unique global solutions achieved on an identical estimated response value. Two illustrative examples taken from the response surface methodology (RSM) literature are given to demonstrate the effectiveness and efficiency of the method addressed in the paper.  相似文献   

18.
ABSRTACT

Since errors in factor levels affect the traditional statistical properties of response surface designs, an important question to consider is robustness of design to errors. However, when the actual design could be observed in the experimental settings, its optimality and prediction are of interest. Various numerical and graphical methods are useful tools for understanding the behavior of the designs. The D- and G-efficiencies and the fraction of design space plot are adapted to assess second-order response surface designs where the predictor variables are disturbed by a random error. Our study shows that the D-efficiencies of the competing designs are considerably low for big variance of the error, while the G-efficiencies are quite good. Fraction of design space plots display the distribution of the scaled prediction variance through the design space with and without errors in factor levels. The robustness of experimental designs against factor errors is explored through comparative study. The construction and use of the D- and G-efficiencies and the fraction of design space plots are demonstrated with several examples of different designs with errors.  相似文献   

19.
Abstract

The gambler's ruin problem is one of the most important problems in the emergence of probability. The problem has been long considered “solved” from a probabilistic viewpoint. However, we do not find the solution satisfactory. In this paper, the problem is recast as a statistical problem. Bounds of the estimate are derived over wide classes of priors. Interestingly, the probabilistic estimates ω(1/2) are identified as the most conservative solutions while the plug-in estimates are found to be out of range of the bounds. It implies that, although conservative, the probabilistic estimates ω(1/2) are justified by our analysis while the plug-in estimates are too extreme for estimating the ruin probability of gambler.  相似文献   

20.
Abstract

A number of tests have been proposed for assessing the location-scale assumption that is often invoked by practitioners. Existing approaches include Kolmogorov–Smirnov and Cramer–von Mises statistics that each involve measures of divergence between unknown joint distribution functions and products of marginal distributions. In practice, the unknown distribution functions embedded in these statistics are typically approximated using nonsmooth empirical distribution functions (EDFs). In a recent article, Li, Li, and Racine establish the benefits of smoothing the EDF for inference, though their theoretical results are limited to the case where the covariates are observed and the distributions unobserved, while in the current setting some covariates and their distributions are unobserved (i.e., the test relies on population error terms from a location-scale model) which necessarily involves a separate theoretical approach. We demonstrate how replacing the nonsmooth distributions of unobservables with their kernel-smoothed sample counterparts can lead to substantial power improvements, and extend existing approaches to the smooth multivariate and mixed continuous and discrete data setting in the presence of unobservables. Theoretical underpinnings are provided, Monte Carlo simulations are undertaken to assess finite-sample performance, and illustrative applications are provided.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号