Internet publication icon

Internet publication

Self-supervised learning of interpretable keypoints from unlabelled videos

Abstract:
We propose KeypointGAN, a new method for recognizing the pose of objects from a single image that for learning uses only unlabelled videos and a weak empirical prior on the object poses. Video frames differ primarily in the pose of the objects they contain, so our method distils the pose information by analyzing the differences between frames. The distillation uses a new dual representation of the geometry of objects as a set of 2D keypoints, and as a pictorial representation, i.e. a skeleton image. This has three benefits: (1) it provides a tight `geometric bottleneck' which disentangles pose from appearance, (2) it can leverage powerful image-to-image translation networks to map between photometry and geometry, and (3) it allows to incorporate empirical pose priors in the learning process. The pose priors are obtained from unpaired data, such as from a different dataset or modality such as mocap, such that no annotated image is ever used in learning the pose recognition network. In standard benchmarks for pose recognition for humans and faces, our method achieves state-of-the-art performance among methods that do not require any labelled images for training.
Publication status:
Published
Peer review status:
Not peer reviewed

Actions


Access Document


Files:
Publisher copy:
10.48550/arxiv.1907.02055

Authors


More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Research group:
Visual Geometry Group
Role:
Author
More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Research group:
Visual Geometry Group
Oxford college:
New College
Role:
Author
ORCID:
0000-0003-1374-2858


Host title:
arXiv
Publication date:
2019-07-03
DOI:


Language:
English
Pubs id:
1212129
Local pid:
pubs:1212129
Deposit date:
2024-11-22

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP