3.8 Article

Asymptotic properties of distance-weighted discrimination and its bias correction for high-dimension, low-sample-size data

Journal

JAPANESE JOURNAL OF STATISTICS AND DATA SCIENCE
Volume 4, Issue 2, Pages 821-840

Publisher

SPRINGERNATURE
DOI: 10.1007/s42081-021-00135-x

Keywords

Bias-corrected DWD; Discriminant analysis; HDLSS; Large p small n; Weighted DWD

Funding

  1. JSPS [20H00576, 18H05290, 19K22837, 18K03409]
  2. Grants-in-Aid for Scientific Research [20H00576, 19K22837, 18H05290, 18K03409] Funding Source: KAKEN

Ask authors/readers for more resources

The DWD is sensitive to imbalanced sample sizes in high-dimensional, low-sample-size settings. The proposed BC-DWD corrects for this bias and shows consistency in misclassification rates. Optimal weights are also proposed for the WDWD to improve its performance.
While distance-weighted discrimination (DWD) was proposed to improve the support vector machine in high-dimensional settings, it is known that the DWD is quite sensitive to the imbalanced ratio of sample sizes. In this paper, we study asymptotic properties of the DWD in high-dimension, low-sample-size (HDLSS) settings. We show that the DWD includes a huge bias caused by a heterogeneity of covariance matrices as well as sample imbalance. We propose a bias-corrected DWD (BC-DWD) and show that the BC-DWD can enjoy consistency properties about misclassification rates. We also consider the weighted DWD (WDWD) and propose an optimal choice of weights in the WDWD. Finally, we discuss performances of the BC-DWD and the WDWD with the optimal weights in numerical simulations and actual data analyses.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available