4.5 Article

An efficient rough feature selection algorithm with a multi-granulation view

期刊

出版社

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ijar.2012.02.004

关键词

Feature selection; Multi-granulation view; Rough set theory; Large-scale data sets

资金

  1. National Natural Science Fund of China [71031006, 60903110, 70971080]
  2. National Key Basic Research and Development Program of China (973) [2011CB311805]
  3. Research Fund for the Doctoral Program of Higher Education [20101401110002]
  4. Key Technologies R&D Program of Shanxi province [20110321027-01]

向作者/读者索取更多资源

Feature selection is a challenging problem in many areas such as pattern recognition, machine learning and data mining. Rough set theory, as a valid soft computing tool to analyze various types of data, has been widely applied to select helpful features (also called attribute reduction). In rough set theory, many feature selection algorithms have been developed in the literatures, however, they are very time-consuming when data sets are in a large scale. To overcome this limitation, we propose in this paper an efficient rough feature selection algorithm for large-scale data sets, which is stimulated from multi-granulation. A sub-table of a data set can be considered as a small granularity. Given a large-scale data set, the algorithm first selects different small granularities and then estimate on each small granularity the reduct of the original data set. Fusing all of the estimates on small granularities together, the algorithm can get an approximate reduct. Because of that the total time spent on computing reducts for sub-tables is much less than that for the original large-scale one, the algorithm yields in a much less amount of time a feature subset (the approximate reduct). According to several decision performance measures, experimental results show that the proposed algorithm is feasible and efficient for large-scale data sets. (c) 2012 Elsevier Inc. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据