Conference item
Self-supervised learning of audio-visual objects from video
- Abstract:
- Our objective is to transform a video into a set of discrete audio-visual objects using self-supervised learning. To this end, we introduce a model that uses attention to localize and group sound sources, and optical flow to aggregate information over time. We demonstrate the effectiveness of the audio-visual object embeddings that our model learns by using them for four downstream speech-oriented tasks: (a) multi-speaker sound source separation, (b) localizing and tracking speakers, (c) correcting misaligned audio-visual data, and (d) active speaker detection. Using our representation, these tasks can be solved entirely by training on unlabeled video, without the aid of object detectors. We also demonstrate the generality of our method by applying it to non-human speakers, including cartoons and puppets. Our model significantly outperforms other self-supervised approaches, and obtains performance competitive with methods that use supervised face detection.
- Publication status:
- Published
- Peer review status:
- Peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Accepted manuscript, 9.5MB, Terms of use)
-
- Publisher copy:
- 10.1007/978-3-030-58523-5_13
Authors
- Publisher:
- Springer
- Series:
- Lecture Notes in Computer Science
- Series number:
- 12363
- Publication date:
- 2020-12-04
- Acceptance date:
- 2020-07-02
- Event title:
- 16th European Conference on Computer Vision (ECCV 2020)
- Event website:
- https://eccv2020.eu/
- Event start date:
- 2020-08-23
- Event end date:
- 2020-09-28
- DOI:
- EISBN:
- 9783030585235
- ISBN:
- 97830305852
- Language:
-
English
- Keywords:
- Pubs id:
-
1131225
- Local pid:
-
pubs:1131225
- Deposit date:
-
2020-09-09
Terms of use
- Copyright holder:
- Springer Nature
- Copyright date:
- 2020
- Rights statement:
- © Springer Nature Switzerland AG 2020.
- Notes:
- This paper was presented at the 16th European Conference on Computer Vision (ECCV 2020), August 2020. This is the accepted manuscript version of the paper. The final version is available online from Springer at: https://doi.org/10.1007/978-3-030-58523-5_13
If you are the owner of this record, you can report an update to it here: Report update to this record