Conference item
Set transformer: A framework for attention-based permutation-invariant neural networks
- Abstract:
- Many machine learning tasks such as multiple instance learning, 3D shape recognition, and few-shot image classification are defined on sets of instances. Since solutions to such problems do not depend on the order of elements of the set, models used to address them should be permutation invariant. We present an attention-based neural network module, the Set Transformer, specifically designed to model interactions among elements in the input set. The model consists of an encoder and a decoder, both of which rely on attention mechanisms. In an effort to reduce computational complexity, we introduce an attention scheme inspired by inducing point methods from sparse Gaussian process literature. It reduces the computation time of self-attention from quadratic to linear in the number of elements in the set. We show that our model is theoretically attractive and we evaluate it on a range of tasks, demonstrating the state-of-the-art performance compared to recent methods for set-structured data.
- Publication status:
- Published
- Peer review status:
- Peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Version of record, 4.7MB, Terms of use)
-
(Preview, Supplementary materials, 157.4KB, Terms of use)
-
- Publication website:
- http://proceedings.mlr.press/v97/lee19d.html
Authors
- Publisher:
- Proceedings of Machine Learning Research
- Volume:
- 97
- Pages:
- 3744-3753
- Publication date:
- 2019-05-24
- Acceptance date:
- 2019-04-24
- Event title:
- 36th International Conference on Machine Learning (ICML 2019)
- Event location:
- Long Beach, California, USA
- Event website:
- https://icml.cc/Conferences/2019
- Event start date:
- 2019-06-09
- Event end date:
- 2019-06-15
- ISSN:
-
2640-3498
- Language:
-
English
- Keywords:
- Pubs id:
-
1128467
- Local pid:
-
pubs:1128467
- Deposit date:
-
2020-08-27
Terms of use
- Copyright holder:
- Lee et al.
- Copyright date:
- 2019
- Rights statement:
- Copyright © The authors and PMLR 2020. This paper is Open Access under a Creative Commons CC-BY licence.
- Notes:
- This paper was presented at the 36th International Conference on Machine Learning (ICML 2019), Long Beach, California, USA, June 2019. This is the publisher's version of the paper. The final version is available online from Proceedings of Machine Learning Research at: http://proceedings.mlr.press/v97/lee19d.html
- Licence:
- CC Attribution (CC BY)
If you are the owner of this record, you can report an update to it here: Report update to this record