TY - JOUR
T1 - Multi-fidelity Gaussian process bandit optimisation
AU - Kandasamy, Kirthevasan
AU - Dasarathy, Gautam
AU - Oliva, Junier
AU - Schneider, Jeff
AU - Póczos, Barnabás
N1 - Funding Information:
We wish to thank Bharath Sriperumbudur for the helpful email discussions. This research is partly funded by DOE grant DESC0011114, NSF grant IIS1563887, and the Darpa D3M program. KK was supported by a Facebook fellowship and a Siebel scholarship. This work was done when KK, GD, and JO were at Carnegie Mellon University.
Publisher Copyright:
© 2019 AI Access Foundation. All rights reserved.
PY - 2019/9/1
Y1 - 2019/9/1
N2 - In many scientific and engineering applications, we are tasked with the maximisation of an expensive to evaluate black box function f. Traditional settings for this problem assume just the availability of this single function. However, in many cases, cheap approximations to f may be obtainable. For example, the expensive real world behaviour of a robot can be approximated by a cheap computer simulation. We can use these approximations to eliminate low function value regions cheaply and use the expensive evaluations of f in a small but promising region and speedily identify the optimum. We formalise this task as a multi-fidelity bandit problem where the target function and its approximations are sampled from a Gaussian process. We develop MF-GP-UCB, a novel method based on upper confidence bound techniques. In our theoretical analysis we demonstrate that it exhibits precisely the above behaviour and achieves better bounds on the regret than strategies which ignore multi-fidelity information. Empirically, MF-GP-UCB outperforms such naive strategies and other multi-fidelity methods on several synthetic and real experiments.
AB - In many scientific and engineering applications, we are tasked with the maximisation of an expensive to evaluate black box function f. Traditional settings for this problem assume just the availability of this single function. However, in many cases, cheap approximations to f may be obtainable. For example, the expensive real world behaviour of a robot can be approximated by a cheap computer simulation. We can use these approximations to eliminate low function value regions cheaply and use the expensive evaluations of f in a small but promising region and speedily identify the optimum. We formalise this task as a multi-fidelity bandit problem where the target function and its approximations are sampled from a Gaussian process. We develop MF-GP-UCB, a novel method based on upper confidence bound techniques. In our theoretical analysis we demonstrate that it exhibits precisely the above behaviour and achieves better bounds on the regret than strategies which ignore multi-fidelity information. Empirically, MF-GP-UCB outperforms such naive strategies and other multi-fidelity methods on several synthetic and real experiments.
UR - http://www.scopus.com/inward/record.url?scp=85075474780&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85075474780&partnerID=8YFLogxK
U2 - 10.1613/jair.1.11288
DO - 10.1613/jair.1.11288
M3 - Article
AN - SCOPUS:85075474780
SN - 1076-9757
VL - 66
SP - 151
EP - 196
JO - Journal of Artificial Intelligence Research
JF - Journal of Artificial Intelligence Research
ER -