Conference item icon

Conference item : Poster

Finding an unsupervised image segmenter in each of your deep generative models

Abstract:
Recent research has shown that numerous human-interpretable directions exist in the latent space of GANs. In this paper, we develop an automatic procedure for finding directions that lead to foreground-background image separation, and we use these directions to train an image segmentation model without human supervision. Our method is generator-agnostic, producing strong segmentation results with a wide range of different GAN architectures. Furthermore, by leveraging GANs pretrained on large datasets such as ImageNet, we are able to segment images from a range of domains without further training or finetuning. Evaluating our method on image segmentation benchmarks, we compare favorably to prior work while using neither human supervision nor access to the training data. Broadly, our results demonstrate that automatically extracting foreground-background structure from pretrained deep generative models can serve as a remarkably effective substitute for human supervision.
Publication status:
Published
Peer review status:
Peer reviewed

Actions


Access Document


Publication website:
https://openreview.net/forum?id=Ug-bgjgSlKV

Authors


More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Oxford college:
New College
Role:
Author
ORCID:
0000-0003-1374-2858


Publisher:
OpenReview
Host title:
International Conference on Learning Representations
Article number:
1643
Publication date:
2021-09-29
Acceptance date:
2022-01-24
Event title:
Tenth International Conference on Learning Representations (ICLR 2022)
Event location:
Virtual event
Event website:
https://iclr.cc/Conferences/2022/
Event start date:
2022-04-25
Event end date:
2022-04-29


Language:
English
Keywords:
Subtype:
Poster
Pubs id:
1281495
Local pid:
pubs:1281495
Deposit date:
2022-10-05

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP