首页> 外文会议>Annual neural information processing systems conference >Feature Selection in Mixture-Based Clustering
【24h】

Feature Selection in Mixture-Based Clustering

机译:基于混合的聚类功能选择

获取原文

摘要

There exist many approaches to clustering, but the important issue of feature selection, i.e., selecting the data attributes that are relevant for clustering, is rarely addressed. Feature selection for clustering is difficult due to the absence of class labels. We propose two approaches to feature selection in the context of Gaussian mixture-based clustering. In the first one, instead of making hard selections, we estimate feature saliencies. An expectation-maximization (EM) algorithm is derived for this task. The second approach extends Koller and Sahami's mutual-information-based feature relevance criterion to the unsupervised case. Feature selection is then carried out by a backward search scheme. This scheme can be classified as a "wrapper", since it wraps mixture estimation in an outer layer that performs feature selection. Experimental results on synthetic and real data show that both methods have promising performance.
机译:群集存在许多方法,但是要素选择的重要问题,即选择与群集相关的数据属性,则很少寻址。由于没有类标签,群集的特征选择很难。我们提出了在高斯混合基于聚类的背景下进行了两种特征选择的方法。在第一个,而不是制造艰难的选择,我们估计了特征纠纷。为此任务导出了期望 - 最大化(EM)算法。第二种方法将Koller和Sahami基于互信的特征相关性标准扩展到无监督的情况。然后通过向后搜索方案执行特征选择。该方案可以被归类为“包装器”,因为它包裹在执行特征选择的外层中的混合估计。合成和实际数据的实验结果表明,两种方法都具有很有希望的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号