no code implementations • 3 Nov 2023 • Poompol Buathong, Jiayue Wan, Samuel Daulton, Raul Astudillo, Maximilian Balandat, Peter I. Frazier
Recent work has considered Bayesian optimization of function networks (BOFN), where the objective function is computed via a network of functions, each taking as input the output of previous nodes in the network and additional parameters.
1 code implementation • 28 Mar 2023 • Raul Astudillo, Zhiyuan Jerry Lin, Eytan Bakshy, Peter I. Frazier
Preferential Bayesian optimization (PBO) is a framework for optimizing a decision maker's latent utility function using preference feedback.
1 code implementation • 21 Mar 2022 • Zhiyuan Jerry Lin, Raul Astudillo, Peter I. Frazier, Eytan Bakshy
We consider Bayesian optimization of expensive-to-evaluate experiments that generate vector-valued outcomes over which a decision-maker (DM) has preferences.
no code implementations • 2 Jan 2022 • Raul Astudillo, Peter I. Frazier
However, internal information about objective function computation is often available.
1 code implementation • NeurIPS 2021 • Raul Astudillo, Peter I. Frazier
We consider Bayesian optimization of the output of a network of functions, where each function takes as input the output of its parent nodes, and where the network takes significant time to evaluate.
1 code implementation • NeurIPS 2021 • Raul Astudillo, Daniel R. Jiang, Maximilian Balandat, Eytan Bakshy, Peter I. Frazier
To overcome the shortcomings of existing approaches, we propose the budgeted multi-step expected improvement, a non-myopic acquisition function that generalizes classical expected improvement to the setting of heterogeneous and unknown evaluation costs.
2 code implementations • NeurIPS 2020 • Sait Cakmak, Raul Astudillo, Peter Frazier, Enlu Zhou
We consider Bayesian optimization of objective functions of the form $\rho[ F(x, W) ]$, where $F$ is a black-box expensive-to-evaluate function and $\rho$ denotes either the VaR or CVaR risk measure, computed with respect to the randomness induced by the environmental random variable $W$.
no code implementations • 14 Nov 2019 • Raul Astudillo, Peter I. Frazier
The outcome of our approach is a menu of designs and evaluated attributes from which the DM makes a final selection.
no code implementations • 4 Jun 2019 • Raul Astudillo, Peter I. Frazier
We consider optimization of composite objective functions, i. e., of the form $f(x)=g(h(x))$, where $h$ is a black-box derivative-free expensive-to-evaluate function with vector-valued outputs, and $g$ is a cheap-to-evaluate real-valued function.