期刊
IEEE TRANSACTIONS ON IMAGE PROCESSING
卷 22, 期 7, 页码 2911-2920出版社
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2013.2253485
关键词
Higher-order tensor; support tensor machine (STM); support vector machine (SVM); tensor classification; tensor rank-one decomposition
资金
- National Science Foundation of China [61273295, 61070033]
- National Social Science Foundation of China [11ZD156]
- Key Laboratory of Symbolic Computation and Knowledge Engineering of the Chinese Ministry of Education [93K-17-2009-K04]
- China Scholarship Council
There has been growing interest in developing more effective learning machines for tensor classification. At present, most of the existing learning machines, such as support tensor machine (STM), involve nonconvex optimization problems and need to resort to iterative techniques. Obviously, it is very time-consuming and may suffer from local minima. In order to overcome these two shortcomings, in this paper, we present a novel linear support higher-order tensor machine (SHTM) which integrates the merits of linear C-support vector machine (C-SVM) and tensor rank-one decomposition. Theoretically, SHTM is an extension of the linear C-SVM to tensor patterns. When the input patterns are vectors, SHTM degenerates into the standard C-SVM. A set of experiments is conducted on nine second-order face recognition datasets and three third-order gait recognition datasets to illustrate the performance of the proposed SHTM. The statistic test shows that compared with STM and C-SVM with the RBF kernel, SHTM provides significant performance gain in terms of test accuracy and training speed, especially in the case of higher-order tensors.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据