首页> 外文会议>Rough Sets and Current Trends in Computing >It is a new application area that integrating rough set theory and ORDBMS to implement data mining in ORDBMS. A suitable rough set algebra must be designed to implement the tight coupling between rough set and ORDBMS. Equivalence matrices algebra doesn't
【24h】

It is a new application area that integrating rough set theory and ORDBMS to implement data mining in ORDBMS. A suitable rough set algebra must be designed to implement the tight coupling between rough set and ORDBMS. Equivalence matrices algebra doesn't

机译:它是将粗糙集理论与ORDBMS集成在一起以在ORDBMS中实现数据挖掘的一个新的应用领域。必须设计合适的粗糙集代数以实现粗糙集和ORDBMS之间的紧密耦合。等价矩阵代数不

获取原文

摘要

Feature subset selection is an importent component of knowledge discovery and data mining systems to help reduce the data dimensionality. Rough sets theory provides a mechanism of selecting feature subsets. In the rough set community, most feature subset selection algorithms are attributes reduct-oriented; that is, finding minimum reducts of the conditional attributes of a decision table. Two main approaches to finding attribute reducts are categorized as discernibility functions-based and attribute dependency-based. These algorithms, however, suffer from intensive computations of either discernibility functions for the former or positive regions for the latter. In this paper, we propose a new concept, called relative attribute dependency, and present a sufficient and necessary condition of the minimum conditional attributes reduct of a decision table represented with the relative attribute dependency. The relative attribute dependency can be calculated by counting the distinct rows of the sub-decision table, instead of generating discernibility functions or positive regions. Thus the computation efficiency of minimum reducts are highly improved. We develop two algorithms for finding minimum reducts of the conditional attributes, one brute-force algorithm and the other heuristic algorithm using attribute entropy as the heuristic function. We also show the results of these algorithms by an illustrative example.
机译:特征子集选择是知识发现和数据挖掘系统的重要组成部分,有助于降低数据维数。粗糙集理论提供了一种选择特征子集的机制。在粗糙集社区中,大多数特征子集选择算法都是面向属性约简的。也就是说,找到决策表的条件属性的最小化约简。查找属性归约的两种主要方法分为基于区分性函数和基于属性依赖项。但是,这些算法需要大量计算前者的辨别函数或后者的正区域。在本文中,我们提出了一个新概念,称为相对属性依赖关系,并提出了用相对属性依赖关系表示的决策表的最小条件属性约简的充要条件。可以通过对子决策表的不同行进行计数来计算相对属性依存关系,而不用生成区分函数或正区域。因此极大地减少了最小减法的计算效率。我们开发了两种用于查找条件属性的最小约简的算法,一种是蛮力算法,另一种是使用属性熵作为启发式函数的启发式算法。我们还将通过一个示例说明这些算法的结果。

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号