Journal
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING
Volume 9, Issue 9, Pages 4400-4406Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JSTARS.2016.2574876
Keywords
Compute unified device architecture (CUDA); dimensionality reduction; graphics processing units (GPUs); hyperspectral imaging; HySime; subspace identification
Categories
Funding
- Natural Science Foundation of China [61301290]
Ask authors/readers for more resources
Signal subspace identification provides a performance improvement in hyperspectral applications, such as target detection, spectral unmixing, and classification. The HySime method is a well-known unsupervised approach for hyperspectral signal subspace identification. It computes the estimated noise and signal correlation matrices from which a subset of eigenvectors is selected to best represent the signal subspace in the least square sense. Depending on the complexity and dimensionality of the hyperspectral scene, the HySime algorithm may be computationally expensive. In this paper, we propose a massively parallel design of the HySime method for acceleration on NVIDIA's graphics processing units (GPUs). Our pure GPU-based implementation includes the optimal use of the page-locked hostmemory, block size, and the number of registers per thread. The proposed implementation was validated in terms of accuracy and performance using the NASA AVIRIS hyperspectral data. The benchmark with the NVIDIA GeForce GTX 580 and Tesla K20 GPUs shows significant speedups with regards to the optimized CPU-based serial counterpart. This new fast implementation of the HySime method demonstrates good potential for real-time hyperspectral applications.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available