Abstract
We extend the concept of the correlated knowledge-gradient policy for the ranking and selection of a finite set of alternatives to the case of continuous decision variables. We propose an approximate knowledge gradient for problems with continuous decision variables in the context of a Gaussian process regression model in a Bayesian setting, along with an algorithm to maximize the approximate knowledge gradient. In the problem class considered, we use the knowledge gradient for continuous parameters to sequentially choose where to sample an expensive noisy function in order to find the maximum quickly. We show that the knowledge gradient for continuous decisions is a generalization of the efficient global optimization algorithm proposed in [D. R. Jones, M. Schonlau, and W. J. Welch, J. Global Optim., 13 (1998), pp. 455-492].
Original language | English (US) |
---|---|
Pages (from-to) | 996-1026 |
Number of pages | 31 |
Journal | SIAM Journal on Optimization |
Volume | 21 |
Issue number | 3 |
DOIs | |
State | Published - 2011 |
All Science Journal Classification (ASJC) codes
- Software
- Theoretical Computer Science
- Applied Mathematics
Keywords
- Bayesian global optimization
- Expected improvement
- Gaussian process regression
- Knowledge gradient
- Model calibration