共查询到20条相似文献,搜索用时 15 毫秒
1.
This article develops a statistical test for the presence of a jump in an otherwise smooth transition process. In this testing, the null model is a threshold regression and the alternative model is a smooth transition model. We propose a quasi-Gaussian likelihood ratio statistic and provide its asymptotic distribution, which is defined as the maximum of a two parameter Gaussian process with a nonzero bias term. Asymptotic critical values can be tabulated and depend on the transition function employed. A simulation method to compute empirical critical values is also developed. Finite-sample performance of the test is assessed via Monte Carlo simulations. The test is applied to investigate the dynamics of racial segregation within cities across the United States. 相似文献
2.
3.
Ian W. Renner David I. Warton Francis K.C. Hui 《Australian & New Zealand Journal of Statistics》2021,63(1):144-158
Point process models are a natural approach for modelling data that arise as point events. In the case of Poisson counts, these may be fitted easily as a weighted Poisson regression. Point processes lack the notion of sample size. This is problematic for model selection, because various classical criteria such as the Bayesian information criterion (BIC) are a function of the sample size, n, and are derived in an asymptotic framework where n tends to infinity. In this paper, we develop an asymptotic result for Poisson point process models in which the observed number of point events, m, plays the role that sample size does in the classical regression context. Following from this result, we derive a version of BIC for point process models, and when fitted via penalised likelihood, conditions for the LASSO penalty that ensure consistency in estimation and the oracle property. We discuss challenges extending these results to the wider class of Gibbs models, of which the Poisson point process model is a special case. 相似文献
4.
A Bayesian approach to modelling binary data on a regular lattice is introduced. The method uses a hierarchical model where the observed data is the sign of a hidden conditional autoregressive Gaussian process. This approach essentially extends the familiar probit model to dependent data. Markov chain Monte Carlo simulations are used on real and simulated data to estimate the posterior distribution of the spatial dependency parameters and the method is shown to work well. The method can be straightforwardly extended to regression models. 相似文献
5.
Jörg Drechsler Agnes Dundler Stefan Bender Susanne Rässler Thomas Zwick 《AStA Advances in Statistical Analysis》2008,92(4):439-458
For micro-datasets considered for release as scientific or public use files, statistical agencies have to face the dilemma of guaranteeing the confidentiality of survey respondents on the one hand and offering sufficiently detailed data on the other hand. For that reason, a variety of methods to guarantee disclosure control is discussed in the literature. In this paper, we present an application of Rubin’s (J. Off. Stat. 9, 462–468, 1993) idea to generate synthetic datasets from existing confidential survey data for public release.We use a set of variables from the 1997 wave of the German IAB Establishment Panel and evaluate the quality of the approach by comparing results from an analysis by Zwick (Ger. Econ. Rev. 6(2), 155–184, 2005) with the original data with the results we achieve for the same analysis run on the dataset after the imputation procedure. The comparison shows that valid inferences can be obtained using the synthetic datasets in this context, while confidentiality is guaranteed for the survey participants. 相似文献
6.
7.
Peter A. Wright 《Journal of applied statistics》2000,27(8):1045-1049
Chen (1999) proposed an economic design, using Taguchi's quality loss function, for choosing a producer's lower specification limit eta for a product with a quality characteristic that has an exponential distribution with mean θ and 'the larger the better' tolerance. Chen (1999) developed an approximate solution that is applicable when 0.5 r m /θ r 0.7 and that requires numerical minimization. We derive a simple, exact solution that is applicable for all values of m /θ and does not require numerical minimization. 相似文献
8.
Smolinski Z 《Wiadomo?ci statystyczne (Warsaw, Poland : 1956)》1985,30(5):1-4
Global population trends are reviewed. The author concludes that a level of overpopulation is inevitable, not primarily because of problems of food supply but because of the environmental degradation that will result from population increases. The author suggests that these environmental changes will lead to increases in mortality and declines in fertility. 相似文献
9.
《Journal of Statistical Computation and Simulation》2012,82(1-4):127-138
In a former study (Chatillon, Gelinas, Martin and Laurencelle, 1987), the authors arrived at the conclusion that for small to moderate sample sizes (n≦90), and for population distributions that are not too skewed nor heavy tailed, the percentiles computed from a set of 9 classes are at least as precise as the corresponding percentiles computed with raw data. Their proof was based essentially on Monte Carlo simulations. The present paper gives a different and complementary proof, based on an exact evaluation of the mean squared error. The method of proof uses the trinomial distribution in an interesting way. 相似文献
10.
Most data used to study the durations of unemployment spells come from the Current Population Survey (CPS), which is a point-in-time survey and gives an incomplete picture of the underlying duration distribution. We introduce a new sample of completed unemployment spells obtained from panel data and apply CPS sampling and reporting techniques to replicate the type of data used by other researchers. Predicted duration distributions derived from this CPS-like data are then compared to the actual distribution. We conclude that the best inferences that can be made about unemployment durations by using CPS-like data are seriously biased. 相似文献
11.
Guangchao C. Feng 《统计学通讯:模拟与计算》2017,46(7):5085-5104
Researchers from different fields have been intrigued but also confused with regard to the ways of conceptualizing and operationalizing difference. This article critically reviewed a number of existing means of operationalizing difference and mathematically explicated and empirically tested why difference scores were problematic and why the diamond model that aimed to replace difference scores was equally troublesome. Ultimately, this article recommended polynomial regression as an alternative to difference scores. 相似文献
12.
13.
14.
15.
We consider non-parametric estimation of a continuous cdf of a random vector (X 1, X 2). With bivariate RC data, it is stated in van der Laan (1996, p. 59810, Ann. Statist.), Quale et al. (2006, JASA) etc. that “it is well known that the NPMLE for continuous data is inconsistent (Tsai et al. (1986)).” The claim is based on a result in Tsai et al. (1986, p.1352, Ann. Statist.) that if X 1 is right censored but not X 2, then common ways for defining one NPMLE lead to inconsistency. If X 1 is right censored and X 2 is type I right-censored (which includes the case in Tsai et al.), we present a consistent NPMLE. The result corrects a common misinterpretation of Tsai's example (Tsai et al., 1986, Ann. Statist.). 相似文献
16.
17.
The integration of different data sources is a widely discussed topic among both the researchers and the Official Statistics. Integrating data helps to contain costs and time required by new data collections. The non-parametric micro Statistical Matching (SM) allows to integrate ‘live’ data resorting only to the observed information, potentially avoiding the misspecification bias and speeding the computational effort. Despite these pros, the assessment of the integration goodness when we use this method is not robust. Moreover, several applications comply with some commonly accepted practices which recommend e.g. to use the biggest data set as donor. We propose a validation strategy to assess the integration goodness. We apply it to investigate these practices and to explore how different combinations of the SM techniques and distance functions perform in terms of the reliability of the synthetic (complete) data set generated. The validation strategy takes advantage of the relation existing among the variables pre-and-post the integration. The results show that ‘the biggest, the best’ rule must not be considered mandatory anymore. Indeed, the integration goodness increases in relation to the variability of the matching variables rather than with respect to the dimensionality ratio between the recipient and the donor data set. 相似文献
18.
Jan Lasek Zoltán Szlávik Marek Gagolewski Sandjai Bhulai 《Journal of applied statistics》2016,43(7):1349-1368
In this paper, we study the efficacy of the official ranking for international football teams compiled by FIFA, the body governing football competition around the globe. We present strategies for improving a team's position in the ranking. By combining several statistical techniques, we derive an objective function in a decision problem of optimal scheduling of future matches. The presented results display how a team's position can be improved. Along the way, we compare the official procedure to the famous Elo rating system. Although it originates from chess, it has been successfully tailored to ranking football teams as well. 相似文献
19.
20.
Orasa Anan Dankmar Böhning Antonello Maruotti 《Statistical Methods and Applications》2017,26(1):49-79
The purpose of the study is to estimate the population size under a truncated count model that accounts for heterogeneity. The proposed estimator is based on the Conway–Maxwell–Poisson distribution. The benefit of using the Conway–Maxwell–Poisson distribution is that it includes the Bernoulli, the Geometric and the Poisson distributions as special cases and, furthermore, allows for heterogeneity. Parameter estimates can be obtained by exploiting the ratios of successive frequency counts in a weighted linear regression framework. The results of the comparisons with Turing’s, the maximum likelihood Poisson, Zelterman’s and Chao’s estimators reveal that our proposal can be beneficially used. Furthermore, our proposal outperforms its competitors under all heterogeneous settings. The empirical examples consider the homogeneous case and several heterogeneous cases, each with its own features, and provide interesting insights on the behavior of the estimators. 相似文献