4.8 Article

Scalable Active Learning for Multiclass Image Classification

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2012.21

Keywords

Active learning; scalable machine learning; multiclass classification; object recognition

Funding

  1. US National Science Foundation (NSF) [IIP-0443945, IIP-0726109, CNS-0708344, CNS-0821474, IIP-0934327, CNS-1039741, IIS-1017344, IIP-1032018, SMA-1028076]
  2. Minnesota Department of Transportation
  3. ITS Institute at the University of Minnesota
  4. Direct For Social, Behav & Economic Scie
  5. SBE Off Of Multidisciplinary Activities [1028076] Funding Source: National Science Foundation
  6. Division Of Computer and Network Systems
  7. Direct For Computer & Info Scie & Enginr [1039741] Funding Source: National Science Foundation

Ask authors/readers for more resources

Machine learning techniques for computer vision applications like object recognition, scene classification, etc., require a large number of training samples for satisfactory performance. Especially when classification is to be performed over many categories, providing enough training samples for each category is infeasible. This paper describes new ideas in multiclass active learning to deal with the training bottleneck, making it easier to train large multiclass image classification systems. First, we propose a new interaction modality for training which requires only yes-no type binary feedback instead of a precise category label. The modality is especially powerful in the presence of hundreds of categories. For the proposed modality, we develop a Value-of-Information (VOI) algorithm that chooses informative queries while also considering user annotation cost. Second, we propose an active selection measure that works with many categories and is extremely fast to compute. This measure is employed to perform a fast seed search before computing VOI, resulting in an algorithm that scales linearly with dataset size. Third, we use locality sensitive hashing to provide a very fast approximation to active learning, which gives sublinear time scaling, allowing application to very large datasets. The approximation provides up to two orders of magnitude speedups with little loss in accuracy. Thorough empirical evaluation of classification accuracy, noise sensitivity, imbalanced data, and computational performance on a diverse set of image datasets demonstrates the strengths of the proposed algorithms.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available