【24h】

Generalisation Error Bounds for Sparse Linear Classifiers

机译:稀疏线性分类器的泛化误差界

获取原文
获取原文并翻译 | 示例

摘要

We provide small sample size bounds on the generalisation error of linear classifiers that are sparse in their dual representation given by the expansion coefficients of the weight vector in terms of the training data. These results theoretically justify algorithms like the Support Vector Machine, the Relevance Vector Machine and K-nearest-neighbour. The bounds are a-posteriori bounds to be evaluated after learning when the attained level of sparsity is known. In a PAC-Bayesian style prior knowledge about the expected sparsity is incorporated into the bounds. The proofs avoid the use of double sample arguments by taking into account the sparsity that leaves unused training points for the evaluation of classifiers. We furthermore give a PAC-Bayesian bound on the average generalisation error over subsets of parameter space that may pave the way combining sparsity in the expansion coefficients and margin in a single bound. Finally, reinterpreting a mistake bound for the classical perceptron algorithm due to Novikoff we demonstrate that our new results put classifiers found by this algorithm on a firm theoretical basis.
机译:我们提供了线性分类器泛化误差的小样本范围,线性分类器的双重表示是稀疏的,这是根据训练数据权重向量的扩展系数给出的。这些结果从理论上证明了算法如支持向量机,相关向量机和K最近邻算法。边界是已知的稀疏度达到后要在学习后评估的后边界。在PAC-贝叶斯风格中,关于预期稀疏性的先验知识被并入边界。通过考虑稀疏性,这些稀疏性为评估分类器留下了未使用的训练点,从而避免了使用双重样本参数。我们还给出了参数空间子集上的平均泛化误差的PAC-贝叶斯界,这可能为在单个界中组合扩展系数的稀疏性和边距的方式铺平了道路。最后,重新解释了由于Novikoff而导致的经典感知器算法的错误界限,我们证明了我们的新结果使该算法找到的分类器具有牢固的理论基础。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号