4.7 Article

Large-Margin Label-Calibrated Support Vector Machines for Positive and Unlabeled Learning

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2019.2892403

Keywords

Support vector machines; Calibration; Training; Training data; Data models; Learning systems; Intserv networks; Label calibration; large margin; positive and unlabeled learning (PU Learning); support vector machines (SVMs)

Funding

  1. NSF of China [61602246, U1713208]
  2. NSF of Jiangsu Province [BK20171430]
  3. Fundamental Research Funds for the Central Universities [30918011319]
  4. State Key Laboratory of Integrated Services Networks, Xidian University [ISN1903]
  5. 111 Program [AH92005]
  6. Summit of the Six Top Talents Program [DZXX-027]
  7. Program for Changjiang Scholars
  8. Lift Program for Young Talents of Jiangsu Province
  9. CAST Lift Program for Young Talents
  10. Australian Research Council [FL-170100117, DE-1901014738, DP-180103424]

Ask authors/readers for more resources

Positive and unlabeled learning (PU learning) aims to train a binary classifier based on only PU data. Existing methods usually cast PU learning as a label noise learning problem or a cost-sensitive learning problem. However, none of them fully take the data distribution information into consideration when designing the model, which hinders them from acquiring more encouraging performance. In this paper, we argue that the clusters formed by positive examples and potential negative examples in the feature space should be critically utilized to establish the PU learning model, especially when the negative data are not explicitly available. To this end, we introduce a hat loss to discover the margin between data clusters, a label calibration regularizer to amend the biased decision boundary to the potentially correct one, and propose a novel discriminative PU classifier termed Large-margin Label-calibrated Support Vector Machines (LLSVM). Our LLSVM classifier can work properly in the absence of negative training examples and effectively achieve the max-margin effect between positive and negative classes. Theoretically, we derived the generalization error bound of LLSVM which reveals that the introduction of PU data does help to enhance the algorithm performance. Empirically, we compared LLSVM with state-of-the-art PU methods on various synthetic and practical data sets, and the results confirm that the proposed LLSVM is more effective than other compared methods on dealing with PU learning tasks.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available