4.6 Article

Tri-partition cost-sensitive active learning through kNN

Journal

SOFT COMPUTING
Volume 23, Issue 5, Pages 1557-1572

Publisher

SPRINGER
DOI: 10.1007/s00500-017-2879-x

Keywords

Active learning; Classification; Cost; k-Nearest neighbors; Tri-partition

Funding

  1. National Natural Science Foundation of China [61379089]
  2. Natural Science Foundation of Department of Education of Sichuan Province [16ZA0060]

Ask authors/readers for more resources

Active learning differs from the training-testing scenario in that class labels can be obtained upon request. It is widely employed in applications where the labeling of instances incurs a heavy manual cost. In this paper, we propose a new algorithm called tri-partition active learning through k-nearest neighbors (TALK). The optimization objective is to minimize the total teacher and misclassification costs. First, a k-nearest neighbors classifier is employed to divide unlabeled instances into three disjoint regions. Region I contains instances for which the expected misclassification cost is lower than the teacher cost, Region II contains instances to be labeled by human experts, and Region III contains the remaining instances. Various strategies are designed to determine which instances are in Region II. Second, instances in Regions I and II are labeled and added to the training set, and the tri-partition process is repeated until all instances have been labeled. Experiments are undertaken on eight University of California, Irvine, datasets using different cost settings. Compared with the state-of-the-art cost-sensitive classification and active learning algorithms, our new algorithm generally exhibits a lower total cost.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available