Thesis
Combine and conquer: representation learning from multmodal data
- Abstract:
-
Supervised learning, which involves training a model using a labelled dataset, is becoming less popular due to its high cost and various issues with generalization and robustness. This is unsurprising, as data such as images and language are complex and cannot be accurately represented by a single label. When models are trained using this method, they often learn features that spuriously correlate with the label, resulting in poor performance when deployed in the real world.
This thesis explores representation learning using multiple sources of data, such as images and language or photos and sketches. We demonstrate through both generative and discriminative models that extracting common abstract concepts between multiple modalities or domains can lead to more accurate and generalisable representations. In addition, we investigate ways to improve the data efficiency of these models, including using fewer multimodal pairs through contrastivestyle objectives and generating multimodal pairs through masked image modeling. Finally, we systematically evaluate the robustness of different learning objectives on distribution shift tasks in order to understand their usefulness in the real world.
Actions
Authors
Contributors
- Institution:
- University of Oxford
- Division:
- MPLS
- Department:
- Engineering Science
- Role:
- Supervisor
- Institution:
- University of Oxford
- Division:
- MPLS
- Department:
- Engineering Science
- Role:
- Supervisor
- ORCID:
- 0000-0003-4911-7333
- Institution:
- University of Oxford
- Division:
- MPLS
- Department:
- Engineering Science
- Role:
- Examiner
- Role:
- Examiner
- DOI:
- Type of award:
- DPhil
- Level of award:
- Doctoral
- Awarding institution:
- University of Oxford
- Language:
-
English
- Keywords:
- Subjects:
- Deposit date:
-
2024-07-23
Terms of use
- Copyright holder:
- Shi, Y
- Copyright date:
- 2023
- Notes:
-
Open Access (no embargo / immediate file release)
If you are the owner of this record, you can report an update to it here: Report update to this record