期刊
DATA MINING AND KNOWLEDGE DISCOVERY
卷 34, 期 6, 页码 1676-1712出版社
SPRINGER
DOI: 10.1007/s10618-020-00692-x
关键词
Dimensionality reduction; Kernel methods; Diffusion maps; Classification
资金
- Israel Science Foundation [ISF 1556/17]
- US-Israel Binational Science Foundation [BSF 2012282]
- Blavatnik Computer Science Research Fund
- Blavatink ICRC Funds
- Pazy Foundation
Kernel methods play a critical role in many machine learning algorithms. They are useful in manifold learning, classification, clustering and other data analysis tasks. Setting the kernel's scale parameter, also referred to as the kernel's bandwidth, highly affects the performance of the task in hand. We propose to set a scale parameter that is tailored to one of two types of tasks: classification and manifold learning. For manifold learning, we seek a scale which is best at capturing the manifold's intrinsic dimension. For classification, we propose three methods for estimating the scale, which optimize the classification results in different senses. The proposed frameworks are simulated on artificial and on real datasets. The results show a high correlation between optimal classification rates and the estimated scales. Finally, we demonstrate the approach on a seismic event classification task.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据