Thesis icon

Thesis

Compression of models and data in deep learning

Abstract:
We face many challenges in deploying high-performance neural networks in practice. These challenges are predominantly due to the size of neural networks and apply to both training and inference. Compressing neural networks to make them train and run more efficiently is therefore crucial and has been a parallel line of research from the early days of neural networks development. The two main compression techniques in deep learning, which are the focus of this thesis, are pruning and quantization. This thesis explores how the information from higher-order gradients (meta-gradients) be used to improve deep learning compression. We start by identifying a fundamental limitation in the formulation of pruning: Although many methods, such as saliency-based pruning, follow pruning by a training or fine-tuning stage, parameter saliencies only look at a snapshot of parameters without taking into account the "trainability" of the parameters. We show how meta-gradients can be used as a more informative signal to find better trainable subnetworks at initialization. We then look at quantized neural networks and show how meta-gradients can be used in a regularization scheme to "learn" models with inherent robustness against post-training quantization. Finally, we look at the dual compression problem, i.e. using neural networks to compress data sources. We start with images and propose a simple autoencoder-free architecture where we store weights of a neural network instead of RGB values of image pixels. We then use meta-gradients to meta-learn a base network to amortize the cost of training one network per input. A significant advantage of our learning compression is that it becomes agnostic to the data type, and we present results on various data types beyond 2D images. Importantly, we evaluate the usefulness of standard DNN compression techniques, e.g., quantization, for this new type of neural network.

Actions


Access Document


Files:

Authors


More by this author
Division:
MPLS
Department:
Computer Science
Sub department:
Computer Science
Oxford college:
Linacre College
Role:
Author

Contributors

Institution:
University of Cambridge
Role:
Supervisor
Institution:
University of Oxford
Division:
MPLS
Department:
Computer Science
Sub department:
Computer Science
Research group:
Oxford Applied and Theoretical Machine Learning Group (OATML)
Oxford college:
Christ Church
Role:
Supervisor
ORCID:
0000-0002-2733-2078
Institution:
University of Oxford
Division:
MPLS
Department:
Computer Science
Sub department:
Computer Science
Oxford college:
Kellogg College
Role:
Examiner
Institution:
Massachusetts Institute of Technology
Role:
Examiner


More from this funder
Funder identifier:
http://dx.doi.org/10.13039/501100000266
Funding agency for:
Lane, N
Alizadeh, M
Grant:
EP/R512333/1
Programme:
NPIF EPSRC Doctoral - University of Oxford 2017


Type of award:
DPhil
Level of award:
Doctoral
Awarding institution:
University of Oxford

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP