Thesis icon

Thesis

Combine and conquer: representation learning from multmodal data

Abstract:

Supervised learning, which involves training a model using a labelled dataset, is becoming less popular due to its high cost and various issues with generalization and robustness. This is unsurprising, as data such as images and language are complex and cannot be accurately represented by a single label. When models are trained using this method, they often learn features that spuriously correlate with the label, resulting in poor performance when deployed in the real world.

This thesis explores representation learning using multiple sources of data, such as images and language or photos and sketches. We demonstrate through both generative and discriminative models that extracting common abstract concepts between multiple modalities or domains can lead to more accurate and generalisable representations. In addition, we investigate ways to improve the data efficiency of these models, including using fewer multimodal pairs through contrastivestyle objectives and generating multimodal pairs through masked image modeling. Finally, we systematically evaluate the robustness of different learning objectives on distribution shift tasks in order to understand their usefulness in the real world.

Actions


Access Document


Files:

Authors


More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Role:
Author

Contributors

Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Role:
Supervisor
Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Role:
Supervisor
ORCID:
0000-0003-4911-7333
Institution:
University of Oxford
Division:
MPLS
Department:
Engineering Science
Role:
Examiner
Role:
Examiner


More from this funder
Grant:
DFR04650
Programme:
Phd Scholarship Programme


DOI:
Type of award:
DPhil
Level of award:
Doctoral
Awarding institution:
University of Oxford


Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP