首页> 美国政府科技报告 >Bayesian Classification Using Noninformative Dirichlet Priors
【24h】

Bayesian Classification Using Noninformative Dirichlet Priors

机译:使用非信息Dirichlet priors的贝叶斯分类

获取原文

摘要

In this dissertation, the Combined Bayes Test (CBT) and its average probability of error, P(s), are developed. The CBT combines training and test data to infer symbol probabilities where a Dirichlet (completely noninformative) prior is assumed for all classes. Using P(s), several results are shown based on the best quantization complexity, M*(which is related to the Hughes Phenomenon). For example, it is shown that M* increases with the training and test data. Also, it is demonstrated that the OST outperforms a more conventional Maximum Likelihood (ML) based test, and the Kolmogorov-Smimov Test (KST). With this, the Bayesian Data Reduction Algorithm (BDRA) is developed. The BDRA uses P(s) (conditioned on the training data) and a greedy approach for reducing irrelevant features from each class, and its performance is shown to be superior to that of a neural network From here, the CBT is extended to demonstrate performance when the training data of each class are mislabeled. Performance is shown to degrade when mislabeling exists in the training data, being dependent on the mislabeling probabilities. However, it is also shown that the BRDA can be used to diminish the effect of mislabeling. Further, the BDRA is modified, using two different approaches, to classify test observations when the training data of each class contain missing feature values. In the first approach, each missing feature is assumed to be uniformly distributed over its range of values; in the second approach, the number of discrete levels for each feature is increased by one. Both methods of modeling missing features are shown to perform similarly, and both also outperform a neural network.

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号