Conference item
Reinforcement learning for bandits with continuous actions and large context spaces
- Abstract:
- We consider the challenging scenario of contextual bandits with continuous actions and large context spaces. This is an increasingly important application area in personalised healthcare where an agent is requested to make dosing decisions based on a patient’s single image scan. In this paper, we first adapt a reinforcement learning (RL) algorithm for continuous control to outperform contextual bandit algorithms specifically hand-crafted for continuous action spaces. We empirically demonstrate this on a suite of standard benchmark datasets for vector contexts. Secondly, we demonstrate that our RL agent can generalise problems with continuous actions to large context spaces, providing results that outperform previous methods on image contexts. Thirdly, we introduce a new contextual bandits test domain with multi-dimensional continuous action space and image contexts which existing tree-based methods cannot handle. We provide initial results with our RL agent.
- Publication status:
- Published
- Peer review status:
- Peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Version of record, pdf, 862.1KB, Terms of use)
-
- Publisher copy:
- 10.3233/FAIA230320
Authors
- Publisher:
- IOS Press
- Host title:
- ECAI 2023
- Pages:
- 590-597
- Series:
- Frontiers in Artificial Intelligence and Applications
- Series number:
- 372
- Publication date:
- 2023-09-28
- Event title:
- 26th European Conference on Artificial Intelligence (ECAI 2023)
- Event location:
- Kraków, Poland
- Event website:
- https://ecai2023.eu/
- Event start date:
- 2023-09-30
- Event end date:
- 2023-10-04
- DOI:
- EISSN:
-
1879-8314
- ISSN:
-
0922-6389
- EISBN:
- 9781643684376
- ISBN:
- 9781643684369
- Language:
-
English
- Pubs id:
-
1562129
- Local pid:
-
pubs:1562129
- Deposit date:
-
2023-12-01
Terms of use
- Copyright holder:
- Duckworth et al.
- Copyright date:
- 2023
- Rights statement:
- © 2023 The Authors. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution Non-Commercial License 4.0 (CC BY-NC 4.0).
If you are the owner of this record, you can report an update to it here: Report update to this record