首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
吴翌琳  南金伶 《统计研究》2020,37(5):94-103
神经网络模型对大样本时间序列的拟合效果优于传统时间序列模型,但对于年度、月度、日度等低频时间序列的预测则难以发挥其优势。鉴于此,本文应用传统时间序列模型和神经网络模型,建立Holtwinters-BP组合模型,利用Holtwinters模型分别拟合各解释变量序列,利用BP模型拟合解释变量和自变量的非线性关系,基于某社交新闻类APP的日广告收入数据进行互联网企业广告收入预测研究。通过与循环神经网络(RNN)模型、长短期记忆神经网络(LSTM)模型等预测结果的对比发现:Holtwinters-BP组合模型的预测精度和稳定性更高;证明多维变量对于广告收入的显著影响,多变量模型的预测准确性高于单变量模型;构建的Holtwinters-BP组合模型对于低频数据预测有较好的有效性和适用性。  相似文献   

2.
It is quite common that raters may need to classify a sample of subjects on a categorical scale. Perfect agreement can rarely be observed partly because of different perceptions about the meanings of the category labels between raters and partly because of factors such as intrarater variability. Usually, category indistinguishability occurs between adjacent categories. In this article, we propose a simple log-linear model combining ordinal scale information and category distinguishability between ordinal categories for modelling agreement between two raters. For the proposed model, no score assignment is required to the ordinal categories. An algorithm and statistical properties will be provided.  相似文献   

3.
Summary.  Controversy has intensified regarding the death-rate from cancer that is induced by a dose of radiation. In the models that are usually considered the hazard function is an increasing function of the dose of radiation. Such models can mask local variations. We consider the models of excess relative risk and of absolute risk and propose a nonparametric estimation of the effect of the dose by using a model selection procedure. This estimation deals with stratified data. We approximate the function of the dose by a collection of splines and select the best one according to the Akaike information criterion. In the same way between the models of excess relative risk or excess absolute risk, we choose the model that best fits the data. We propose a bootstrap method for calculating a pointwise confidence interval of the dose function. We apply our method for estimating the solid cancer and leukaemia death hazard functions to Hiroshima.  相似文献   

4.
Parameter design or robust parameter design (RPD) is an engineering methodology intended as a cost-effective approach for improving the quality of products and processes. The goal of parameter design is to choose the levels of the control variables that optimize a defined quality characteristic. An essential component of RPD involves the assumption of well estimated models for the process mean and variance. Traditionally, the modeling of the mean and variance has been done parametrically. It is often the case, particularly when modeling the variance, that nonparametric techniques are more appropriate due to the nature of the curvature in the underlying function. Most response surface experiments involve sparse data. In sparse data situations with unusual curvature in the underlying function, nonparametric techniques often result in estimates with problematic variation whereas their parametric counterparts may result in estimates with problematic bias. We propose the use of semi-parametric modeling within the robust design setting, combining parametric and nonparametric functions to improve the quality of both mean and variance model estimation. The proposed method will be illustrated with an example and simulations.  相似文献   

5.
基于辅助回归模型的空间Hausman检验   总被引:1,自引:0,他引:1  
 基于面板数据空间误差分量模型,提出空间Hausman检验,并通过数理推导,构造辅助回归模型的空间Hausman检验,进而通过Monte Carlo模拟实验,研究空间Hausman检验,以及辅助回归空间Hausman检验的有限样本性质。研究结果表明,空间Hausman检验能有效矫正空间面板数据下经典Hausman检验的水平扭曲,但随着空间相关性和样本量增大,其水平扭曲偏离理想值;辅助回归空间Hausman检验始终保持理想的水平扭曲。此外,二者均具有优越的检验功效。  相似文献   

6.
We propose a new procedure for combining multiple tests in samples of right-censored observations. The new method is based on multiple constrained censored empirical likelihood where the constraints are formulated as linear functionals of the cumulative hazard functions. We prove a version of Wilks’ theorem for the multiple constrained censored empirical likelihood ratio, which provides a simple reference distribution for the test statistic of our proposed method. A useful application of the proposed method is, for example, examining the survival experience of different populations by combining different weighted log-rank tests. Real data examples are given using the log-rank and Gehan-Wilcoxon tests. In a simulation study of two sample survival data, we compare the proposed method of combining tests to previously developed procedures. The results demonstrate that, in addition to its computational simplicity, the combined test performs comparably to, and in some situations more reliably than previously developed procedures. Statistical software is available in the R package ‘emplik’.  相似文献   

7.
We consider the problem of model selection based on quantile analysis and with unknown parameters estimated using quantile leasts squares. We propose a model selection test for the null hypothesis that the competing models are equivalent against the alternative hypothesis that one model is closer to the true model. We follow with two applications of the proposed model selection test. The first application is in model selection for time series with non-normal innovations. The second application is in model selection in the NoVas method, short for normalizing and variance stabilizing transformation, forecast. A set of simulation results also lends strong support to the results presented in the paper.  相似文献   

8.
This article introduces a parsimonious structure for mixture of autoregressive models, where the weighting coefficients are determined through latent random variables, as functions of all past observations. These latent variables follow a Markov model. We propose a dynamic programming algorithm for forecasting, which reduces the volume of calculations. We also derive limiting behavior of unconditional first moment of the process and an appropriate upper bound for the limiting value of the variance. Further more, we show convergence and stability of the second moment. Finally, we illustrate the efficacy of the proposed model by simulation.  相似文献   

9.
Model selection methods are important to identify the best approximating model. To identify the best meaningful model, purpose of the model should be clearly pre-stated. The focus of this paper is model selection when the modelling purpose is classification. We propose a new model selection approach designed for logistic regression model selection where main modelling purpose is classification. The method is based on the distance between the two clustering trees. We also question and evaluate the performances of conventional model selection methods based on information theory concepts in determining best logistic regression classifier. An extensive simulation study is used to assess the finite sample performances of the cluster tree based and the information theoretic model selection methods. Simulations are adjusted for whether the true model is in the candidate set or not. Results show that the new approach is highly promising. Finally, they are applied to a real data set to select a binary model as a means of classifying the subjects with respect to their risk of breast cancer.  相似文献   

10.
The GARCH and stochastic volatility (SV) models are two competing, well-known and often used models to explain the volatility of financial series. In this paper, we consider a closed form estimator for a stochastic volatility model and derive its asymptotic properties. We confirm our theoretical results by a simulation study. In addition, we propose a set of simple, strongly consistent decision rules to compare the ability of the GARCH and the SV model to fit the characteristic features observed in high frequency financial data such as high kurtosis and slowly decaying autocorrelation function of the squared observations. These rules are based on a number of moment conditions that is allowed to increase with sample size. We show that our selection procedure leads to choosing the model that fits best, or the simplest model under equivalence, with probability one as the sample size increases. The finite sample size behavior of our procedure is analyzed via simulations. Finally, we provide an application to stocks in the Dow Jones industrial average index.  相似文献   

11.
We propose a Bayesian stochastic search approach to selecting restrictions on multivariate regression models where the errors exhibit deterministic or stochastic conditional volatilities. We develop a Markov chain Monte Carlo (MCMC) algorithm that generates posterior restrictions on the regression coefficients and Cholesky decompositions of the covariance matrix of the errors. Numerical simulations with artificially generated data show that the proposed method is effective in selecting the data-generating model restrictions and improving the forecasting performance of the model. Applying the method to daily foreign exchange rate data, we conduct stochastic search on a VAR model with stochastic conditional volatilities.  相似文献   

12.
We address the issue of model selection in beta regressions with varying dispersion. The model consists of two submodels, namely: for the mean and for the dispersion. Our focus is on the selection of the covariates for each submodel. Our Monte Carlo evidence reveals that the joint selection of covariates for the two submodels is not accurate in finite samples. We introduce two new model selection criteria that explicitly account for varying dispersion and propose a fast two step model selection scheme which is considerably more accurate and is computationally less costly than usual joint model selection. Monte Carlo evidence is presented and discussed. We also present the results of an empirical application.  相似文献   

13.
In practice, it often happens that we have a number of base methods of classification. We are not able to clearly determine which method is optimal in the sense of the smallest error rate. Then we have a combined method that allows us to consolidate information from multiple sources in a better classifier. I propose a different approach, a sequential approach. Sequentiality is understood here in the sense of adding posterior probabilities to the original data set and so created data are used during classification process. We combine posterior probabilities obtained from base classifiers using all combining methods. Finally, we combine these probabilities using a mean combining method. To the original data set we add obtained posterior probabilities as additional features. In each step we change our additional probabilities to achieve the minimum error rate for base methods. Experimental results on different data sets demonstrate that the method is efficient and that this approach outperforms base methods providing a reduction in the mean classification error rate.  相似文献   

14.
We propose a new criterion for model selection in prediction problems. The covariance inflation criterion adjusts the training error by the average covariance of the predictions and responses, when the prediction rule is applied to permuted versions of the data set. This criterion can be applied to general prediction problems (e.g. regression or classification) and to general prediction rules (e.g. stepwise regression, tree-based models and neural nets). As a by-product we obtain a measure of the effective number of parameters used by an adaptive procedure. We relate the covariance inflation criterion to other model selection procedures and illustrate its use in some regression and classification problems. We also revisit the conditional bootstrap approach to model selection.  相似文献   

15.
In this article we propose a new method to select a discrete model f(x; θ), based on the conditional density of a sample given the value of a sufficient statistic for θ. The main idea is to work with a broad family of discrete distributions, called the family of power series distribution, for which there is a common sufficient statistic for the parameter of interest. The proposed method uses the maximum conditional density in order to select the best model.

We compare our proposal with the usual methodology based on Bayes factors. We provide several examples that show that our proposal works fine in most instances. Bayes factors are strongly dependent on the prior information about the parameters. Since our method does not require the specification of a prior distribution, it provides a useful alternative to Bayes factors.  相似文献   

16.

We propose a semiparametric version of the EM algorithm under the semiparametric mixture model introduced by Anderson (1979, Biometrika , 66 , 17-26). It is shown that the sequence of proposed EM iterates, irrespective of the starting value, converges to the maximum semiparametric likelihood estimator of the vector of parameters in the semiparametric mixture model. The proposed EM algorithm preserves the appealing monotone convergence property of the standard EM algorithm and can be implemented by employing the standard logistic regression program. We present one example to demonstrate the performance of the proposed EM algorithm.  相似文献   

17.
Systemic risk analysis reveals the interdependencies of risk factors especially in tail event situations. In applications the focus of interest is on capturing joint tail behavior rather than a variation around the mean. Quantile and expectile regression are used here as tools of data analysis. When it comes to characterizing tail event curves one faces a dimensionality problem, which is important for CoVaR (Conditional Value at Risk) determination. A projection-based single-index model specification may come to the rescue but for ultrahigh-dimensional regressors one faces yet another dimensionality problem and needs to balance precision versus dimension. Such a balance is achieved by combining semiparametric ideas with variable selection techniques. In particular, we propose a projection-based single-index model specification for very high-dimensional regressors. This model is used for practical CoVaR estimates with a systemically chosen indicator. In simulations we demonstrate the practical side of the semiparametric CoVaR method. The application to the U.S. financial sector shows good backtesting results and indicate market coagulation before the crisis period. Supplementary materials for this article are available online.  相似文献   

18.
In this paper we consider the linear compartment model and consider the estimation procedures of the different parameters. We discuss a method to obtain the initial estimators, which can be used for any iterative procedures to obtain the least-squares estimators. Four different types of confidence intervals have been discussed and they have been compared by computer simulations. We propose different methods to estimate the number of components of the linear compartment model. One data set has been used to see how the different methods work in practice.  相似文献   

19.
The equity premium, return on equity minus return on risk-free asset, is expected to be positive. We consider imposing such positivity constraint in local historical average (LHA) in nonparametric kernel regression framework. It is also extended to the semiparametric single index model when multiple predictors are used. We construct the constrained LHA estimator via an indicator function which operates as “model-selection” between the unconstrained LHA and the bound of the constraint (zero for the positivity constraint). We smooth the indicator function by bagging, which operates as “model-averaging” and yields a combined forecast of unconstrained LHA forecasts and the bound of the constraint. The local combining weights are determined by the probability that the constraint is binding. Asymptotic properties of the constrained LHA estimators without and with bagging are established, which show how the positive constraint and bagging can help reduce the asymptotic variance and mean squared errors. Monte Carlo simulations are conducted to show the finite sample behavior of the asymptotic properties. In predicting U.S. equity premium, we show that substantial nonlinearity can be captured by LHA and that the local positivity constraint can improve out-of-sample prediction of the equity premium.  相似文献   

20.
顾嘉等 《统计研究》2021,38(9):114-127
不同于传统( Susceptible-Exposed-Infected-Removed)SEIR流行病传播动力学模型,本文在近期研究的Varying Coefficient Susceptible-Exposed-Infected-Diagnosed-Removed (vSEIdR)模型基础上加上人口迁徙(Migration) 模块,设计开发了vSEIdRm模型,该模型考虑了跨区域人口迁徙对疫情传播的影响,并允许流行病传播参数随时间变化。本文首先对人口迁移数据进行统计分析,建立其与各省新冠肺炎疫情发展的联系。之后,基于vSEIdRm模型估计了疫情初期各省份来自武汉的输入病例数,并定量刻画了离汉交通管控的效果。研究结果显示,离汉交通管控措施有效地减少了各省份的疫情规模。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号