4.7 Review

Low Rank Regularization: A review

Journal

NEURAL NETWORKS
Volume 136, Issue -, Pages 218-232

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.neunet.2020.09.021

Keywords

Low rank; Regularization; Optimization

Funding

  1. National Key Research and Development Program of China [2018YFB1403501]
  2. National Natural Science Foundation of China [61936014, 61772427, 61751202]
  3. Fundamental Research Funds for the Central Universities, PR China [G2019KY0501]

Ask authors/readers for more resources

This paper provides a comprehensive survey on Low Rank Regularization (LRR), focusing on recent advances in rank-norm relaxation and model optimization. It emphasizes the superiority of non-convex relaxations over convex relaxations in improving the performance of existing LRR models.
Low Rank Regularization (LRR), in essence, involves introducing a low rank or approximately low rank assumption to target we aim to learn, which has achieved great success in many data analysis tasks. Over the last decade, much progress has been made in theories and applications. Nevertheless, the intersection between these two lines is rare. In order to construct a bridge between practical applications and theoretical studies, in this paper we provide a comprehensive survey for LRR. Specifically, we first review the recent advances in two issues that all LRR models are faced with: (1) rank-norm relaxation, which seeks to find a relaxation to replace the rank minimization problem; (2) model optimization, which seeks to use an efficient optimization algorithm to solve the relaxed LRR models. For the first issue, we provide a detailed summarization for various relaxation functions and conclude that the non-convex relaxations can alleviate the punishment bias problem compared with the convex relaxations. For the second issue, we summarize the representative optimization algorithms used in previous studies, and analyze their advantages and disadvantages. As the main goal of this paper is to promote the application of non-convex relaxations, we conduct extensive experiments to compare different relaxation functions. The experimental results demonstrate that the non-convex relaxations generally provide a large advantage over the convex relaxations. Such a result is inspiring for further improving the performance of existing LRR models. (c) 2020 Elseiver Ltd. All Rights Reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available