4.5 Article

Instance reduction for one-class classification

Journal

KNOWLEDGE AND INFORMATION SYSTEMS
Volume 59, Issue 3, Pages 601-628

Publisher

SPRINGER LONDON LTD
DOI: 10.1007/s10115-018-1220-z

Keywords

Machine learning; One-class classification; Instance reduction; Training set selection; Evolutionary computing

Funding

  1. Polish National Science Center [UMO-2015/19/B/ST6/01597]
  2. Spanish National Research Project [TIN2014-57251-P]
  3. Andalusian Research Plan [P11-TIC-7765]

Ask authors/readers for more resources

Instance reduction techniques are data preprocessing methods originally developed to enhance the nearest neighbor rule for standard classification. They reduce the training data by selecting or generating representative examples of a given problem. These algorithms have been designed and widely analyzed in multi-class problems providing very competitive results. However, this issue was rarely addressed in the context of one-class classification. In this specific domain a reduction of the training set may not only decrease the classification time and classifier's complexity, but also allows us to handle internal noisy data and simplify the data description boundary. We propose two methods for achieving this goal. The first one is a flexible framework that adjusts any instance reduction method to one-class scenario by introduction of meaningful artificial outliers. The second one is a novel modification of evolutionary instance reduction technique that is based on differential evolution and uses consistency measure for model evaluation in filter or wrapper modes. It is a powerful native one-class solution that does not require an access to counterexamples. Both of the proposed algorithms can be applied to any type of one-class classifier. On the basis of extensive computational experiments, we show that the proposed methods are highly efficient techniques to reduce the complexity and improve the classification performance in one-class scenarios.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available