Conference item
AdaGeo: Adaptive geometric learning for optimization and sampling
- Abstract:
- Gradient-based optimization and Markov Chain Monte Carlo sampling can be found at the heart of a multitude of machine learning methods. In high-dimensional settings, well-known issues such as slow-mixing, non-convexity and correlations can hinder the algorithms’ efficiency. In order to overcome these difficulties, we propose AdaGeo, a preconditioning framework for adaptively learning the geometry of parameter space during optimization or sampling. We use the Gaussian Process latent variable model (GP-LVM) to represent a lower-dimensional embedding of the parameters, identifying the underlying Riemannian manifold on which the optimization or sampling are taking place. Samples or optimization steps are consequently proposed based on the geometry of the manifold. We apply our framework to stochastic gradient descent and stochastic gradient Langevin dynamics and show performance improvements for both optimization and sampling.
- Publication status:
- Published
- Peer review status:
- Peer reviewed
Actions
Authors
- Publisher:
- Proceedings of Machine Learning Research
- Host title:
- Twenty-First International Conference on Artificial Intelligence and Statistics, Apr 9, 2018 - Apr 11, 2018, Playa Blanca, Spain
- Journal:
- International Conference on Artificial Intelligence and Statistics More from this journal
- Volume:
- 84
- Pages:
- 226-234
- Publication date:
- 2018-04-09
- Acceptance date:
- 2017-12-22
- ISSN:
-
1938-7228
- Pubs id:
-
pubs:845063
- UUID:
-
uuid:5e0c3640-183b-409d-a2df-d56fb8e3b8c1
- Local pid:
-
pubs:845063
- Source identifiers:
-
845063
- Deposit date:
-
2018-05-02
Terms of use
- Copyright holder:
- 2018 Abbati, et al
- Copyright date:
- 2018
- Notes:
- This is the publisher's version of the article. The final version is available online from Proceedings of Machine Learning Research at: http://proceedings.mlr.press/v84/abbati18a.html
If you are the owner of this record, you can report an update to it here: Report update to this record