4.7 Article

A generalized mean distance-based k-nearest neighbor classifier

Journal

EXPERT SYSTEMS WITH APPLICATIONS
Volume 115, Issue -, Pages 356-372

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2018.08.021

Keywords

K-nearest neighbor rule; Local mean vector; Generalized mean distance; Pattern recognition

Funding

  1. National Natural Science Foundation of China [61502208, 61762021, 61402122]
  2. Natural Science Foundation of Jiangsu Province of China [BK20150522, BK20170558]
  3. Research Foundation for Talented Scholars of Jiangsu University [14JDG037]
  4. China Postdoctoral Science Foundation [2015M570411]
  5. Open Foundation of Artificial Intelligence Key Laboratory of Sichuan Province [2017RYJ04]
  6. Natural Science Foundation of Guizhou Province [[2017]1130]
  7. Research Foundation of North Minzu University [2018XYZDX10]

Ask authors/readers for more resources

K-nearest neighbor (KNN) rule is a well-known non-parametric classifier that is widely used in pattern recognition. However, the sensitivity of the neighborhood size k always seriously degrades the KNN-based classification performance, especially in the case of the small sample size with the existing outliers. To overcome this issue, in this article we propose a generalized mean distance-based k-nearest neighbor classifier (GMDKNN) by introducing multi-generalized mean distances and the nested generalized mean distance that are based on the characteristic of the generalized mean. In the proposed method, multi-local mean vectors of the given query sample in each class are calculated by adopting its class-specific k nearest neighbors. Using the achieved k local mean vectors per class, the corresponding k generalized mean distances are calculated and then used to design the categorical nested generalized mean distance. In the classification phase, the categorical nested generalized mean distance is used as the classification decision rule and the query sample is classified into the class with the minimum nested generalized mean distance among all the classes. Extensive experiments on the UCI and KEEL data sets, synthetic data sets, the KEEL noise data sets and the UCR time series data sets are conducted by comparing the proposed method to the state-of-art KNN-based methods. The experimental results demonstrate that the proposed GMDKNN performs better and has the less sensitiveness to k. Thus, our proposed GMDKNN with the robust and effective classification performanrP could-be a promising method for-pattern-recognition-in some expert and intelligence systems. (C) 2018 Elsevier Ltd. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available