4.6 Article

Coupled Deep Autoencoder for Single Image Super-Resolution

Journal

IEEE TRANSACTIONS ON CYBERNETICS
Volume 47, Issue 1, Pages 27-37

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCYB.2015.2501373

Keywords

Autoencoder; deep learning; single image super-resolution (SR)

Funding

  1. National Natural Science Foundation of China [61472110, 61373077]
  2. Zhejiang Provincial Natural Science Foundation of China [LR15F020002]
  3. Specialized Research Fund for the Doctoral Program of Higher Education of China [20110121110020]
  4. National Defense Basic Scientific Research Program of China [B-0110155]
  5. National Defense Science and Technology Key Laboratory Foundation [9140C-30211ZS-8]

Ask authors/readers for more resources

Sparse coding has been widely applied to learning-based single image super-resolution (SR) and has obtained promising performance by jointly learning effective representations for low-resolution (LR) and high-resolution (HR) image patch pairs. However, the resulting HR images often suffer from ringing, jaggy, and blurring artifacts due to the strong yet ad hoc assumptions that the LR image patch representation is equal to, is linear with, lies on a manifold similar to, or has the same support set as the corresponding HR image patch representation. Motivated by the success of deep learning, we develop a data-driven model coupled deep autoencoder (CDA) for single image SR. CDA is based on a new deep architecture and has high representational capability. CDA simultaneously learns the intrinsic representations of LR and HR image patches and a big-data-driven function that precisely maps these LR representations to their corresponding HR representations. Extensive experimentation demonstrates the superior effectiveness and efficiency of CDA for single image SR compared to other state-of-the-art methods on Set5 and Set14 datasets.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available