Thesis
Compression of models and data in deep learning
- Abstract:
- We face many challenges in deploying high-performance neural networks in practice. These challenges are predominantly due to the size of neural networks and apply to both training and inference. Compressing neural networks to make them train and run more efficiently is therefore crucial and has been a parallel line of research from the early days of neural networks development. The two main compression techniques in deep learning, which are the focus of this thesis, are pruning and quantization. This thesis explores how the information from higher-order gradients (meta-gradients) be used to improve deep learning compression. We start by identifying a fundamental limitation in the formulation of pruning: Although many methods, such as saliency-based pruning, follow pruning by a training or fine-tuning stage, parameter saliencies only look at a snapshot of parameters without taking into account the "trainability" of the parameters. We show how meta-gradients can be used as a more informative signal to find better trainable subnetworks at initialization. We then look at quantized neural networks and show how meta-gradients can be used in a regularization scheme to "learn" models with inherent robustness against post-training quantization. Finally, we look at the dual compression problem, i.e. using neural networks to compress data sources. We start with images and propose a simple autoencoder-free architecture where we store weights of a neural network instead of RGB values of image pixels. We then use meta-gradients to meta-learn a base network to amortize the cost of training one network per input. A significant advantage of our learning compression is that it becomes agnostic to the data type, and we present results on various data types beyond 2D images. Importantly, we evaluate the usefulness of standard DNN compression techniques, e.g., quantization, for this new type of neural network.
Actions
Authors
Contributors
+ Lane, N
- Institution:
- University of Cambridge
- Role:
- Supervisor
+ Gal, Y
- Institution:
- University of Oxford
- Division:
- MPLS
- Department:
- Computer Science
- Sub department:
- Computer Science
- Research group:
- Oxford Applied and Theoretical Machine Learning Group (OATML)
- Oxford college:
- Christ Church
- Role:
- Supervisor
- ORCID:
- 0000-0002-2733-2078
+ Markham, A
- Institution:
- University of Oxford
- Division:
- MPLS
- Department:
- Computer Science
- Sub department:
- Computer Science
- Oxford college:
- Kellogg College
- Role:
- Examiner
+ Han, S
- Institution:
- Massachusetts Institute of Technology
- Role:
- Examiner
+ Engineering and Physical Sciences Research Council, ARM Inc.
More from this funder
- Funder identifier:
- http://dx.doi.org/10.13039/501100000266
- Funding agency for:
- Lane, N
- Alizadeh, M
- Grant:
- EP/R512333/1
- Programme:
- NPIF EPSRC Doctoral - University of Oxford 2017
- Type of award:
- DPhil
- Level of award:
- Doctoral
- Awarding institution:
- University of Oxford
- Language:
-
English
- Keywords:
- Subjects:
- Deposit date:
-
2022-12-25
Terms of use
- Copyright holder:
- Alizadeh, M
- Copyright date:
- 2022
If you are the owner of this record, you can report an update to it here: Report update to this record