【24h】

Learning Compositional Categorization Models

机译:学习组成分类模型

获取原文
获取原文并翻译 | 示例

摘要

This contribution proposes a compositional approach to visual object categorization of scenes. Compositions are learned from the Caltech 101 database1 and form intermediate abstractions of images that are semantically situated between low-level representations and the high-level categorization. Salient regions, which are described by localized feature histograms, are detected as image parts. Subsequently compositions are formed as bags of parts with a locality constraint. After performing a spatial binding of compositions by means of a shape model, coupled probabilistic kernel classifiers are applied thereupon to establish the final image categorization. In contrast to the discriminative training of the cat-egorizer, intermediate compositions are learned in a generative manner yielding relevant part agglomerations, i.e. groupings which are frequently appearing in the dataset while simultaneously supporting the discrimination between sets of categories. Consequently, compositionality simplifies the learning of a complex categorization model for complete scenes by splitting it up into simpler, sharable compositions. The architecture is evaluated on the highly challenging Caltech 101 database which exhibits large intra-category variations. Our compositional approach shows competitive retrieval rates in the range of 53.6 ±0.88% or, with a multi-scale feature set, rates of 57.8 ± 0.79%.
机译:该贡献提出了一种对场景的视觉对象分类的合成方法。构图可从Caltech 101数据库中学习,并形成图像的中间抽象,这些图像在语义上位于低级表示和高级分类之间。通过局部特征直方图描述的显着区域被检测为图像部分。随后,将组合物形成为具有局部约束的零件袋。在通过形状模型执行构图的空间绑定之后,在其上应用耦合的概率核分类器以建立最终的图像分类。与区分猫训练器的区分训练相反,以生成方式学习中间成分,产生相关的部分集聚,即在数据集中经常出现的分组,同时支持类别集之间的区分。因此,合成将其分为更简单,可共享的合成,从而简化了针对完整场景的复杂分类模型的学习。该体系结构在极富挑战性的Caltech 101数据库中进行了评估,该数据库显示出较大的类别内差异。我们的合成方法显示出竞争性检索率在53.6±0.88%的范围内,或者在具有多尺度功能集的情况下,其检索率在57.8±0.79%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号