Thesis icon

Thesis

Identifying and exploiting structures for reliable deep learning

Abstract:

Deep learning research has recently witnessed an impressively fast-paced progress in a wide range of tasks including computer vision, natural language processing, and reinforcement learning. The extraordinary performance of these systems often gives the impression that they can be used to revolutionise our lives for the better. However, as recent works point out, these systems suffer from several issues that make them unreliable for use in the real world, including vulnerability to adversarial attacks (Szegedy et al. [243]), tendency to memorise noise (Zhang et al. [286]), being over-confident on incorrect predictions (miscalibration) (Guo et al. [99]), and unsuitability for handling private data (Gilad-Bachrach et al. [88]). In this the- sis, we look at each of these issues in detail, investigate their causes, and propose computationally cheap algorithms for mitigating them in practice.

To do this, we identify structures in deep neural networks that can be exploited to mitigate the above causes of unreliability of deep learning algorithms. In Chapter 4, we show that minimising a property of matrices, called stable rank, for individual weight matrix in a neural network reduces the tendency of the network to memorise noise without sacrificing its performance on noiseless data.

In Chapter 5, we prove that memorising label noise or doing improper representation learning makes achieving adversarial robustness impossible. Chapter 6 shows that a low-rank prior on the representation space of neural networks increases the robustness of neural networks to adversarial perturbations without inducing any tradeoff with accuracy in practice.

In Chapter 7, we highlight the use of focal loss, which weights loss components from individual samples differentially by how well the neural network classifies each of them, as an alternative loss function to cross-entropy for minimising miscalibration in neural networks.

In Chapter 8, we first define a new framework called Encrypted Prediction As A Service (EPAAS) along with a set of computational and privacy constraints. Then we propose the use of a Fully Homomorphic Encryption [84] scheme which can be used with a Binary neural network [61], along with a set of algebraic and computational tricks, to satisfy all our conditions for EPAAS while being computationally efficient.

Actions


Access Document


Files:

Authors


More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Computer Science
Sub department:
Computer Science
Research group:
The Alan Turing Institute
Oxford college:
St Hugh's College
Role:
Author
ORCID:
https://orcid.org/0000-0002-4190-0449

Contributors

Division:
MPLS
Department:
Computer Science
Sub department:
Computer Science
Oxford college:
Lady Margaret Hall
Role:
Supervisor
ORCID:
https://orcid.org/0000-0002-2300-4819
Division:
MPLS
Department:
Engineering Science
Sub department:
Engineering Science
Oxford college:
St Catherine's College
Role:
Supervisor


More from this funder
Funding agency for:
Sanyal, A
Grant:
TU/C/000023
Programme:
The Turing doctoral studentship


DOI:
Type of award:
DPhil
Level of award:
Doctoral
Awarding institution:
University of Oxford

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP