Journal article icon

Journal article

Discovering fully semantic representations via centroid- and orientation-aware feature learning

Abstract:
Learning meaningful representations of images in scientific domains that are robust to variations in centroids and orientations remains an important challenge. Here we introduce centroid- and orientation-aware disentangling autoencoder (CODAE), an encoder–decoder-based neural network that learns meaningful content of objects in a latent space. Specifically, a combination of a translation- and rotation-equivariant encoder, Euler encoding and an image moment loss enables CODAE to extract features invariant to positions and orientations of objects of interest from randomly translated and rotated images. We evaluate this approach on several publicly available scientific datasets, including protein images from life sciences, four-dimensional scanning transmission electron microscopy data from material science and galaxy images from astronomy. The evaluation shows that CODAE learns centroids, orientations and their invariant features and outputs, as well as aligned reconstructions and the exact view reconstructions of the input images with high quality.
Publication status:
Published
Peer review status:
Peer reviewed

Actions


Access Document


Publisher copy:
10.1038/s42256-024-00978-5

Authors


More by this author
Role:
Author
ORCID:
0000-0002-2498-4214
More by this author
Institution:
University of Oxford
Division:
SSD
Department:
International Development
Sub department:
Refugee Studies Centre
Role:
Author


Publisher:
Nature Research
Journal:
Nature Machine Intelligence More from this journal
Volume:
7
Issue:
2
Pages:
307-314
Publication date:
2025-02-06
Acceptance date:
2024-12-20
DOI:
EISSN:
2522-5839


Language:
English
Source identifiers:
2711564
Deposit date:
2025-02-24
This ORA record was generated from metadata provided by an external service. It has not been edited by the ORA Team.

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP