首页 | 本学科首页   官方微博 | 高级检索  
     检索      


Some results concerning off-training-set and IID error for the Gibbs and the Bayes optimal generalizers
Authors:DAVID H WOLPERT  EMANUEL KNILL  TAL GROSSMAN
Institution:(1) NASA Ames Research Centre, Caelum Research, MS 269-2, Moffet Field, CA, 94035, USA email;(2) CIC-3 Computer Research and Applications, MSB265, LANL, Los Alamos, NM, 87545, USA;(3) Theoretical Division and, CNLS, MS B213, LANL, Los Alamos, NM 87545, USA
Abstract:In this paper we analyse the average behaviour of the Bayes-optimal and Gibbs learning algorithms. We do this both for off-training-set error and conventional IID (independent identically distributed) error (for which test sets overlap with training sets). For the IID case we provide a major extension to one of the better known results. We also show that expected IID test set error is a non-increasing function of training set size for either algorithm. On the other hand, as we show, the expected off-training-set error for both learning algorithms can increase with training set size, for non-uniform sampling distributions. We characterize the relationship the sampling distribution must have with the prior for such an increase. We show in particular that for uniform sampling distributions and either algorithm, the expected off-training-set error is a non-increasing function of training set size. For uniform sampling distributions, we also characterize the priors for which the expected error of the Bayes-optimal algorithm stays constant. In addition we show that for the Bayes-optimal algorithm, expected off-training-set error can increase with training set size when the target function is fixed, but if and only if the expected error averaged over all targets decreases with training set size. Our results hold for arbitrary noise and arbitrary loss functions.
Keywords:Supervised learning  learning curves  off-training-set  Bayes-optimal  Gibbs  IID
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号