Internet publication
A critical analysis of self-supervision, or what we can learn from a single image
- Abstract:
- We look critically at popular self-supervision techniques for learning deep convolutional neural networks without manual labels. We show that three different and representative methods, BiGAN, RotNet and DeepCluster, can learn the first few layers of a convolutional network from a single image as well as using millions of images and manual labels, provided that strong data augmentation is used. However, for deeper layers the gap with manual supervision cannot be closed even if millions of unlabelled images are used for training. We conclude that: (1) the weights of the early layers of deep networks contain limited information about the statistics of natural images, that (2) such low-level statistics can be learned through self-supervision just as well as through strong supervision, and that (3) the low-level statistics can be captured via synthetic transformations instead of using a large image dataset.
- Publication status:
- Published
- Peer review status:
- Not peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Version of record, pdf, 8.7MB, Terms of use)
-
- Publisher copy:
- 10.48550/arxiv.1904.13132
Authors
- Host title:
- arXiv
- Publication date:
- 2019-04-30
- DOI:
- Language:
-
English
- Pubs id:
-
1212131
- Local pid:
-
pubs:1212131
- Deposit date:
-
2024-11-25
Terms of use
- Copyright holder:
- Asano et al
- Copyright date:
- 2019
- Rights statement:
- ©2019 The Authors
- Licence:
- Other
If you are the owner of this record, you can report an update to it here: Report update to this record