【24h】

Model Selection and Error Estimation

机译:模型选择和误差估计

获取原文
获取原文并翻译 | 示例

摘要

We study model selection strategies based on penalized empirical loss minimization. We point out a tight relationship between error estimation and data-based complexity penalization: any good error estimate may be converted into a data-based penalty function and the performance of the estimate is governed by the quality of the error estimate. We consider several penalty functions, involving error estimates on independent test data, empirical vc dimension, empirical VC entropy, and margin-based quantities. We also consider the maximal difference between the error on the first half of the training data and the second half, and the expected maximal discrepancy, a closely related capacity estimate that can be calculated by Monte Carlo integration. Maximal discrepancy penalty functions are appealing for pattern classification problems, since their computation is equivalent to empirical risk minimization over the training data with some labels flipped.
机译:我们研究基于经验损失最小化的模型选择策略。我们指出了错误估计和基于数据的复杂度惩罚之间的紧密关系:任何好的错误估计都可以转换为基于数据的惩罚函数,并且估计的性能由错误估计的质量决定。我们考虑几种惩罚函数,包括对独立测试数据的误差估计,经验vc维数,经验VC熵和基于余量的数量。我们还考虑了训练数据上半部分和下半部分的误差之间的最大差,以及预期的最大差异,这是可以通过蒙特卡洛积分计算的密切相关的能力估计。最大差异惩罚函数对于模式分类问题很有吸引力,因为它们的计算等效于在翻转某些标签的情况下对训练数据进行经验风险最小化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号