首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 531 毫秒
1.
The purpose of this article is to develop a goodness-of-fit test based on score test statistics for cumulative logit models with extra variation of random effects. Two main theorems for the proposed score test statistics are derived. In simulation studies, the powers of the proposed tests are discussed and the power curve against a variety of dispersion parameters and bandwidths is depicted. The proposed method is illustrated by an ordinal data set from Mosteller and Tukey [23].  相似文献   

2.
Ranking objects by a panel of judges is commonly used in situations where objective attributes cannot easily be measured or interpreted. Under the assumption that the judge independently arrive at their rankings by making pair wise comparisons among the objects in an attempt to reproduce a common baseline ranking w0, we develop and explore confidence regions and Bayesian highest posterior density credible regions for w0 with emphasis on very small sample sizes.  相似文献   

3.
Abstract

This article focuses on the problem of estimating the shelf life of food products by modeling the results coming from sensory evaluations. In such studies, trained panelists are asked to judge food attributes by reference to a scale of numbers (scores varying often from 0 to 6). The usual statistical approach for data analysis is to fit a regression line relating the scores and the time of evaluation. The estimate of the shelf life is obtained by solving the regression equation and replacing the score by a cut-off point (which indicates product “failure”) previously chosen by the food company. The procedure used in these sensory evaluations is such that one never knows the exact “time to failure”. Consequently, data arising from these studies are either right or left censored. We propose a model which incorporates these informations and assumes a Weibull for the underlying distribution of the failure time. Simulation studies were implemented. The approach was used in a real data set coming from sensory evaluations of a dehydrated food product.  相似文献   

4.
In market research and some other areas, it is common that a sample of n judges (consumers, evaluators, etc.) are asked to independently rank a series of k objects or candidates. It is usually difficult to obtain the judges' full cooperation to completely rank all k objects. A practical way to overcome this difficulty is to give each judge the freedom to choose the number of top candidates he is willing to rank. A frequently encountered question in this type of survey is how to select the best object or candidate from the incompletely ranked data. This paper proposes a subset selection procedure which constructs a random subset of all the k objects involved in the survey such that the best object is included in the subset with a prespecified confidence. It is shown that the proposed subset selection procedure is distribution-free over a very broad class of underlying distributions. An example from a market research study is used to illustrate the proposed procedure.  相似文献   

5.
In this article, we address the similarity structure between pairs of order statistics of an identically distributed independent random variables X1,…, Xn. The overlapping coefficient (Δ) of Weitzman (1970 Weitzman , M. S. ( 1970 ). Measures of overlap of income distributions of white and Negro families in the United States . Technical Paper No. 22 , Washington , D.C. , U.S.A : Department of Commerce, Bureau of Census . [Google Scholar]), is used to assess the degree of similarity or closeness between pairs of order statistics. It appears that the degree of the similarity between any of such pairs is independent of the parent distribution. Using this notion, it is shown that for i < j, the degree of similarity between distributions of the ith and the jth order statistics decreases as i and j sunder. Some possible biometric applicability of the value of Δ are explored. In particular, the use of this measure in estimation of the number of possible strata, subgroups or natural subdivisions in a population are suggested.  相似文献   

6.
Dyadic matrices are natural data representations in a wide range of domains. A dyadic matrix often involves two types of abstract objects and is based on observations of pairs of elements with one element from each object. Owing to the increasing needs from practical applications, dyadic data analysis has recently attracted increasing attention and many techniques have been developed. However, most existing approaches, such as co-clustering and relational reasoning, only handle a single dyadic table and lack flexibility to perform prediction using multiple dyadic matrices. In this article, we propose a general nonparametric Bayesian framework with a cascaded structure to model multiple dyadic matrices and then describe an efficient hybrid Gibbs sampling algorithm for posterior inference and analysis. Empirical evaluations using both synthetic data and real data show that the proposed model captures the hidden structure of data and generalizes the predictive inference in a unique way.  相似文献   

7.
Outlier detection plays an important role in the pre-treatment of sequential datasets to obtain pure valuable data. This paper proposes an outlier detection scheme for dynamical sequential datasets. First, the conception of forward outlier factor(FOF) and backward outlier factor(BOF) are employed to measure an object’s similarity shared with its sequentially adjacent objects. The object that shows no similarity with its sequential neighbors is labeled as suspicious outliers, which will be treated subsequently to judge whether it is really an outlier in the dataset. Second, the sequentially adjacent suspicious outliers are defined as suspicious outlier series(SOS), then the expected path representing the ideal transition path through the suspicious outliers in the SOS and the measured path representing the real path through all the objects in the SOS are employed, and the ratio of the length of the expected path to that of the measured path indicates whether there exist outliers in the SOS. Third, in the case that there exist outliers in the SOS, if there are N suspicious outliers in the SOS, then 2N ? 2 remaining path will be generated by removing k(0 < k < N) suspicious outliers and sequentially connecting the remaining ones. The dynamical sequential outlier factor(DSOF) is employed to represent the ratio of the length of measured path of the considered remaining path to the that of the the expected path of the corresponding SOS, and the degree of the objects removed in a remaining path being outliers is indicated by the DSOF. The proposed outlier detection scheme is conducted from a dynamical perspective, and breaks the tight relation between being an outlier and being not similar with adjacent objects. Experiments are conducted to evaluate the effectiveness of the proposed scheme, and the experimental results verify that the proposed scheme has higher detection quality for sequential dataset. In addition, the proposed outlier detection scheme is not dependent on the size of dataset and needs no prior information about the distribution of the data.  相似文献   

8.
This article derives score tests for extra-Poisson variation in the positive or truncated-at-zero Poisson regression model against truncated-at-zero negative binomial family alternatives. It also develops size-corrected tests of overdispersion that are expected to improve their small-sample properties. Further, small-sample performance of the tests is investigated by means of Monte Carlo experiments. As an illustration, the proposed tests are applied to a model of strikes in U.S. manufacturing. The proposed tests have an interpretation as conditional moment tests and require only the positive Poisson model to be estimated. It is shown that most of the tests for overdispersion in the regular Poisson model given in the econometric and statistical literature can be obtained as special cases of the tests developed in this article. Monte Carlo experiments indicate that the size correction, based on the asymptotic expansions of the score function, is effective in improving the accuracy of the size and power of the tests in small samples.  相似文献   

9.
The rule of thumb given by Mosteller and Tukey for obtaining the degrees of freedom which are appropriate for the approximate t-statistic associated with the jackknife is discussed.Examples of possible misapplication of this rule are given.  相似文献   

10.
By using prior knowledge it may be possible to deduce pieces of individual information from a frequency distribution of a population. If the prior information is described by a stochastic model, an information-theoretic approach can be applied in order to judge the possibilities for disclosure. By specifying the stochastic model in various ways it is shown how the decrease in entropy caused by the publication of a frequency distribution can be determined and interpreted. The stochastic models are also used to derive formulae for disclosure risks and expected numbers of disclosures.  相似文献   

11.
An important aspect of paired comparison experiments is the decision of how to form pairs in advance of collecting data. A weakness of typical paired comparison experimental designs is the difficulty in incorporating prior information, which can be particularly relevant for the design of tournament schedules for players of games and sports. Pairing methods that make use of prior information are often ad hoc algorithms with little or no formal basis. The problem of pairing objects can be formalized as a Bayesian optimal design. Assuming a linear paired comparison model for outcomes, we develop a pairing method that maximizes the expected gain in Kullback–Leibler information from the prior to the posterior distribution. The optimal pairing is determined using a combinatorial optimization method commonly used in graph-theoretic contexts. We discuss the properties of our optimal pairing criterion, and demonstrate our method as an adaptive procedure for pairing objects multiple times. We compare the performance of our method on simulated data against random pairings, and against a system that is currently in use in tournament chess.  相似文献   

12.
The main object of this paper is to discuss properties of the score statistics for testing the null hypothesis of no association in Weibull model with measurement errors. Three different score statistics are considered. The efficient score statistics, a naive score statistics obtained by replacing the unobserved true covariate with the observed one and a score statistics based on the corrected score statistics. It is shown that corrected and naive score statistics are equivalent and the asymptotic relative efficiency between naive and efficient score statistics is derived.  相似文献   

13.
SUMMARY We compare properties of parameter estimators under Akaike information criterion (AIC) and 'consistent' AIC (CAIC) model selection in a nested sequence of open population capture-recapture models. These models consist of product multinomials, where the cell probabilities are parameterized in terms of survival ( ) and capture ( p ) i i probabilities for each time interval i . The sequence of models is derived from 'treatment' effects that might be (1) absent, model H ; (2) only acute, model H ; or (3) acute and 0 2 p chronic, lasting several time intervals, model H . Using a 35 factorial design, 1000 3 repetitions were simulated for each of 243 cases. The true number of parameters ranged from 7 to 42, and the sample size ranged from approximately 470 to 55 000 per case. We focus on the quality of the inference about the model parameters and model structure that results from the two selection criteria. We use achieved confidence interval coverage as an integrating metric to judge what constitutes a 'properly parsimonious' model, and contrast the performance of these two model selection criteria for a wide range of models, sample sizes, parameter values and study interval lengths. AIC selection resulted in models in which the parameters were estimated with relatively little bias. However, these models exhibited asymptotic sampling variances that were somewhat too small, and achieved confidence interval coverage that was somewhat below the nominal level. In contrast, CAIC-selected models were too simple, the parameter estimators were often substantially biased, the asymptotic sampling variances were substantially too small and the achieved coverage was often substantially below the nominal level. An example case illustrates a pattern: with 20 capture occasions, 300 previously unmarked animals are released at each occasion, and the survival and capture probabilities in the control group on each occasion were 0.9 and 0.8 respectively using model H . There was a strong acute treatment effect 3 on the first survival ( ) and first capture probability ( p ), and smaller, chronic effects 1 2 on the second and third survival probabilities ( and ) as well as on the second capture 2 3 probability ( p ); the sample size for each repetition was approximately 55 000. CAIC 3 selection led to a model with exactly these effects in only nine of the 1000 repetitions, compared with 467 times under AIC selection. Under CAIC selection, even the two acute effects were detected only 555 times, compared with 998 for AIC selection. AIC selection exhibited a balance between underfitted and overfitted models (270 versus 263), while CAIC tended strongly to select underfitted models. CAIC-selected models were overly parsimonious and poor as a basis for statistical inferences about important model parameters or structure. We recommend the use of the AIC and not the CAIC for analysis and inference from capture-recapture data sets.  相似文献   

14.
3D object recognition: Representation and matching   总被引:1,自引:0,他引:1  
Three-dimensional object recognition entails a number of fundamental problems in computer vision: representation of a 3D object, identification of the object from its image, estimation of its position and orientation, and registration of multiple views of the object for automatic model construction. This paper surveys three of those topics, namely representation, matching, and pose estimation. It also presents an overview of the free-form surface matching problem, and describes COSMOS, our framework for representing and recognizing free-form objects. The COSMOS system recognizes arbitrarily curved 3D rigid objects from a single view using dense surface data. We present both the theoretical aspects and the experimental results of a prototype recognition system based on COSMOS.  相似文献   

15.
The Bradley–Terry model is widely and often beneficially used to rank objects from paired comparisons. The underlying assumption that makes ranking possible is the existence of a latent linear scale of merit or equivalently of a kind of transitiveness of the preference. However, in some situations such as sensory comparisons of products, this assumption can be unrealistic. In these contexts, although the Bradley–Terry model appears to be significantly interesting, the linear ranking does not make sense. Our aim is to propose a 2-dimensional extension of the Bradley–Terry model that accounts for interactions between the compared objects. From a methodological point of view, this proposition can be seen as a multidimensional scaling approach in the context of a logistic model for binomial data. Maximum likelihood is investigated and asymptotic properties are derived in order to construct confidence ellipses on the diagram of the 2-dimensional scores. It is shown by an illustrative example based on real sensory data on how to use the 2-dimensional model to inspect the lack-of-fit of the Bradley–Terry model.  相似文献   

16.
Summary.  On-line auctions pose many challenges for the empirical researcher, one of which is the effective and reliable modelling of price paths. We propose a novel way of modelling price paths in eBay's on-line auctions by using functional data analysis. One of the practical challenges is that the functional objects are sampled only very sparsely and unevenly. Most approaches rely on smoothing to recover the underlying functional object from the data, which can be difficult if the data are irregularly distributed. We present a new approach that can overcome this challenge. The approach is based on the ideas of mixed models. Specifically, we propose a semiparametric mixed model with boosting to recover the functional object. As well as being able to handle sparse and unevenly distributed data, the model also results in conceptually more meaningful functional objects. In particular, we motivate our method within the framework of eBay's on-line auctions. On-line auctions produce monotonic increasing price curves that are often correlated across auctions. The semiparametric mixed model accounts for this correlation in a parsimonious way. It also manages to capture the underlying monotonic trend in the data without imposing model constraints. Our application shows that the resulting functional objects are conceptually more appealing. Moreover, when used to forecast the outcome of an on-line auction, our approach also results in more accurate price predictions compared with standard approaches. We illustrate our model on a set of 183 closed auctions for Palm M515 personal digital assistants.  相似文献   

17.
ABSTRACT

Cordeiro and Ferrari[1] Cordeiro, G.M. and Ferrari, S.L.P. 1991. A Modified Score Test Statistic Having Chi-Squared Distribution to Order n?1. Biometrika, 78: 573582. [Web of Science ®] [Google Scholar] obtained a Bartlett-type correction to the score statistic that is given by a polynomial of second degree in the statistic itself with coefficients that depend on cumulants of log–likelihood derivatives. Although the corrected statistic has good size properties, it is not always a monotone transformation of the original statistic. Recently, some monotone transformations of the score statistic have been proposed as alternatives to the polynomial transformation. In this paper we derive simple formulae for various modified score statistics for testing a scalar parameter in two-parameter exponential models which do not require knowledge of cumulants. The formulae are readily applicable to cover many important and commonly used distributions and involve only trivial operations on certain functions and their derivatives.  相似文献   

18.
The order of experimental runs in a fractional factorial experiment is essential when the cost of level changes in factors is considered. The generalized foldover scheme given by [1] Coster, D. C. and Cheng, C. S. 1988. Minimum cost trend free run orders of fractional factorial designs. The Annals of Statistics, 16: 11881205. [Crossref], [Web of Science ®] [Google Scholar]gives an optimal order to experimental runs in an experiment with specified defining contrasts. An experiment can be specified by a design requirement such as resolution or estimation of some interactions. To meet such a requirement, we can find several sets of defining contrasts. Applying the generalized foldover scheme to these sets of defining contrasts, we obtain designs with different numbers of level changes and then the design with minimum number of level changes. The difficulty is to find all the sets of defining contrasts. An alternative approach is investigated by [2] Cheng, C. S., Martin, R. J. and Tang, B. 1998. Two-level factorial designs with extreme numbers of level changes. The Annals of Statistics, 26: 15221539. [Crossref], [Web of Science ®] [Google Scholar]for two-level fractional factorial experiments. In this paper, we investigate experiments with all factors in slevels.  相似文献   

19.
In this paper, we introduce a compound size-dependent renewal risk model driven by two sequences of random sources. The individual claim sizes and their inter-arrival times form a sequence of independent and identically distributed random pairs with each pair obeying a specific dependence structure. The numbers of claims caused by individual events form another sequence of independent and identically distributed positive integer-valued random variables, independent of the random pairs above. Precise large deviations of aggregate claims for the compound size-dependent renewal risk model are investigated in the case of dominatedly varying claim sizes.  相似文献   

20.
In measurement error problems, two major and consistent estimation methods are the conditional score and the corrected score. They are functional methods that require no parametric assumptions on mismeasured covariates. The conditional score requires that a suitable sufficient statistic for the mismeasured covariate can be found, while the corrected score requires that the object score function can be estimated without bias. These assumptions limit their ranges of applications. The extensively corrected score proposed here is an extension of the corrected score. It yields consistent estimations in many cases when neither the conditional score nor the corrected score is feasible. We demonstrate its constructions in generalized linear models and the Cox proportional hazards model, assess its performances by simulation studies and illustrate its implementations by two real examples.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号