4.8 Article

Small Data Challenges in Big Data Era: A Survey of Recent Progress on Unsupervised and Semi-Supervised Methods

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2020.3031898

Keywords

Data models; Task analysis; Training; Adaptation models; Gallium nitride; Big Data; Supervised learning; Unsupervised methods; semi-supervised methods; domain adaptation; transformation equivariance and invariance; disentangled representations; generative models; auto-encoders; generative adversarial networks; auto-regressive models; flow-based generative models; transformers; self-supervised methods; teach-student models; instance discrimination and equivariance

Ask authors/readers for more resources

This paper reviews the recent progresses in unsupervised and semi-supervised learning, presenting the principles and applications of various models. By exploring the distribution of unlabeled data, these models expand the domain of existing autoencoders, generative adversarial networks, and other deep networks, providing more powerful representations.
Representation learning with small labeled data have emerged in many problems, since the success of deep neural networks often relies on the availability of a huge amount of labeled data that is expensive to collect. To address it, many efforts have been made on training sophisticated models with few labeled data in an unsupervised and semi-supervised fashion. In this paper, we will review the recent progresses on these two major categories of methods. A wide spectrum of models will be categorized in a big picture, where we will show how they interplay with each other to motivate explorations of new ideas. We will review the principles of learning the transformation equivariant, disentangled, self-supervised and semi-supervised representations, all of which underpin the foundation of recent progresses. Many implementations of unsupervised and semi-supervised generative models have been developed on the basis of these criteria, greatly expanding the territory of existing autoencoders, generative adversarial nets (GANs) and other deep networks by exploring the distribution of unlabeled data for more powerful representations. We will discuss emerging topics by revealing the intrinsic connections between unsupervised and semi-supervised learning, and propose in future directions to bridge the algorithmic and theoretical gap between transformation equivariance for unsupervised learning and supervised invariance for supervised learning, and unify unsupervised pretraining and supervised finetuning. We will also provide a broader outlook of future directions to unify transformation and instance equivariances for representation learning, connect unsupervised and semi-supervised augmentations, and explore the role of the self-supervised regularization for many learning problems.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available