This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

BOIS: Bayesian Optimization of Interconnected Systems

Leonardo D. González and Victor M. Zavala Department of Chemical and Biological Engineering, University of Wisconsin-Madison Madison, WI 53706 USA
Abstract

Bayesian optimization (BO) has proven to be an effective paradigm for the global optimization of expensive-to-sample systems. One of the main advantages of BO is its use of Gaussian processes (GPs) to characterize model uncertainty which can be leveraged to guide the learning and search processes. However, BO typically treats systems as black-boxes and this limits the ability to exploit structural knowledge (e.g., physics and sparse interconnections). Composite functions of the form f(x,y(x))f(x,y(x)), wherein GP modeling is shifted from the performance function ff to an intermediate function yy, offer an avenue for exploiting structural knowledge. However, the use of composite functions in a BO framework is complicated by the need to generate a probability density for ff from the Gaussian density of yy calculated by the GP (e.g., when ff is nonlinear it is not possible to obtain a closed-form expression). Previous work has handled this issue using sampling techniques; these are easy to implement and flexible but are computationally intensive. In this work, we introduce a new paradigm which allows for the efficient use of composite functions in BO; this uses adaptive linearizations of ff to obtain closed-form expressions for the statistical moments of the composite function. We show that this simple approach (which we call BOIS) enables the exploitation of structural knowledge, such as that arising in interconnected systems as well as systems that embed multiple GP models and combinations of physics and GP models. Using a chemical process optimization case study, we benchmark the effectiveness of BOIS against standard BO and sampling approaches. Our results indicate that BOIS achieves performance gains and accurately captures the statistics of composite functions.

keywords:
Bayesian optimization, grey-box modeling, composite functions
thanks: We acknowledge support from NSF-EFRI #2132036, UW-Madison GERS program, and the PPG Fellowship.

1 Introduction

Optimization of natural and engineered systems (e.g., chemical processes, biological systems, materials) is often challenging due to incomplete physical knowledge or the high complexity of experiments and available physical models. As a result, it is necessary to device optimization procedures that effectively combine experimental and model data while mitigating complexity, see Conn et al. (2009). These procedures, often referred to as black-box optimization algorithms, treat the system as a black-box, f(x)f(x), and use input/output observation data to direct their search for a solution. One of the most popular paradigms to have emerged in this setting is Bayesian optimization (BO), see Mockus (2012). BO uses input/output data to generate a Gaussian process (GP) model that estimates system performance as well as model uncertainty. These estimates are used to construct an acquisition function (AF) that allows for the search to be tuned to emphasize sampling from either high performance (exploitation) or high uncertainty (exploration) regions; this is a key feature that makes BO an especially powerful global optimizer, see Shahriari et al. (2016).

While the black-box assumption substantially facilitates the implementation of BO (no prior knowledge about the system is needed), it ignores the fact that there is often some form of structural knowledge available (e.g., physics or sparse interconnectivity). For example, when dealing with physical systems, several components might be well-modeled and understood, while others might not. For those that are not, the fundamental principles governing their behavior (e.g., conservation laws, equilibrium, value constraints) are, at least qualitatively, understood. Additionally, sparse connectivity, which provides information on how different components affect each other, is also often known. As a result, the system of interest is usually not an unobservable black-box but rather a “grey-box” that is partially observable with a known structure, see Astudillo and Frazier (2021). Several methods exist for performing grey-box BO; studies have shown that they are able to improve algorithm performance by effectively constraining the search space, resulting in lower sampling requirements and better solutions, see Lu et al. (2021); Raissi et al. (2019); Kandasamy et al. (2017). Of these approaches, the use of composite functions has proven to be one of the most widely used methods for incorporating preexisting knowledge into BO, see Astudillo and Frazier (2019).

A composite function expresses the system performance as f(x,y(x))f(x,y(x)) where xx are the system inputs, ff is a known scalar function, and yy is an unknown vector-valued function that describes the behavior of internal system components. This decomposition shifts the modeling task from estimating the performance function directly to estimating the values of yy which serve as inputs to f(x,y(x))f(x,y(x)). Additionally, because ff is now a known function, derivative information might be available to understand the effects of xx and yy, see Uhrenholt and Jensen (2019). This approach also readily lends itself to the inclusion of constraints as these are often dependent on internal variables which can be captured by yy Paulson and Lu (2022). Framing an optimization problem in this manner is therefore a fairly intuitive approach, especially in chemical engineering where the performance metric is usually an economic cost. For example, the cost equations for equipment, material streams, and utilities are often readily available and it is the parameters these equations rely on (compositions, flowrates, duties) that are unknown. Furthermore, traditional unit operations (heat exchangers, distillation columns, compressors) have significantly better mechanistic models available than those that tend to be more niche (bioreactors, non-equilibrium separators, solids-handling). It then makes sense to construct a composite function where the outer function, ff, returns the price of the system based on the known cost equations while its inputs, yy, are the mass and energy flows through the system and are estimated via either mechanistic or data-driven models. Constraints can then be incorporated using values estimated for yy to ensure that data-driven models obey fundamental physical laws as well as to enforce more traditional requirements such as product specifications, waste generation, utility consumption, and equipment sizing which are often important in process design.

While setting up a composite function optimization problem seems fairly straightforward, implementing it in a BO setting is not a trivial task. As previously stated, one of the main advantages of BO is the inclusion of uncertainty estimates in the surrogate model, which allows for greater exploration of the design space when compared to a deterministic model, see Boukouvala and Floudas (2017). However, when using a composite function, the GP models generated are of yy not ff. Given that ff is the performance measure being optimized, it is necessary propagate the predicted uncertainty from y(x)y(x) to f(x,y(x))f(x,y(x)) (i.e, the density of ff or desired summarizing statistics must be determined). A Gaussian density for y(x)y(x) is directly obtained from the GP model; as a result, when ff is a linear model, we can make use of the closure of Gaussian random variables under linear operations to generate the density of f(x,y(x))f(x,y(x)) (which is also a Gaussian). When ff is nonlinear, however, a closed-form solution is not readily available, and this operation is usually carried out numerically using sampling methods like Monte Carlo (MC), see Balandat et al. (2020).

Given the increased functionality composite functions can provide to BO and the computational intensity of sampling methods, we propose the Bayesian Optimization of Interconnected Systems (BOIS) framework. BOIS provides a novel method for estimating the distribution of ff in a significantly more efficient manner than sampling by linearizing the performance function in the neighborhood of a y(x)y(x) of interest. This allows us to construct a local Laplace approximation which we can use to generate closed-form expressions for the mean and uncertainty of f(x,y(x))f(x,y(x)). We test the performance of this approach by conducting extensive numerical experiments on a chemical process optimization case study and compare its performance to standard BO as well as MC-driven composite function BO. Our results illustrate that BOIS is able to outperform standard BO while also providing accurate estimates for the distribution of ff at a significantly lower computational cost than MC.

11footnotetext: Corresponding author: Victor M. Zavala (e-mail: [email protected])

2 Background

2.1 Bayesian optimization

Refer to caption
Figure 1: Block-flow diagram of the BO framework

The problems we are interested in solving are of the form

minxf(x)\displaystyle\min_{x}~{}~{}f(x) (1a)
s.t.xX\displaystyle\textrm{s.t.}~{}~{}x\in X (1b)

where f:Xf:X\to\mathbb{R} is a scalar performance function, XdxX\subseteq\mathbb{R}^{d_{x}} is the design or search space, and xx is a set of design inputs within XX. Generally, solving this problem is made difficult by the fact that derivatives cannot readily be calculated as sampling ff is expensive and the generated data can be noisy. Bayesian optimization manages these challenges by using input/output data to generate a surrogate model of the performance function that it uses to systematically explore the design space. The general framework for the algorithm is as follows: using an initial dataset of size {\ell}, 𝒟={x𝒦,f𝒦}\mathcal{D}^{\ell}=\{x_{\mathcal{K}},f_{\mathcal{K}}\}, where 𝒦={1,,}\mathcal{K}=\{1,...,\ell\}, train a Gaussian process surrogate model. The GP assumes that the output data have a prior of the form f(x𝒦)𝒩(m(x),K(x,x))f(x_{\mathcal{K}})\sim\mathcal{N}\left(\textbf{m}(x),\textbf{K}(x,x^{\prime})\right) where m(x)dx\textbf{m}(x)\in\mathbb{R}^{d_{x}} is the mean function and K(x,x)dx×dx\textbf{K}(x,x^{\prime})\in\mathbb{R}^{d_{x}\times d_{x}} is the covariance matrix. While m(x)\textbf{m}(x) is usually set equal to 0, K(x,x)\textbf{K}(x,x^{\prime}) is calculated using a kernel function, k(x,x)k(x,x^{\prime}), such that Kij=k(xi,xj)\textbf{K}_{ij}=k(x_{i},x_{j}). In our work, we have opted to use the Mátern class of kernels to construct the covariance matrix. Once the GP has been conditioned on 𝒟\mathcal{D}^{\ell}, it calculates the posterior distribution of ff, f^\hat{f}^{\ell}, at a set of nn new points 𝒳\mathcal{X}:

f^(𝒳)𝒩(mf(𝒳),Σf(𝒳))\hat{f}^{\ell}(\mathcal{X})\sim\mathcal{N}\left(m^{\ell}_{f}(\mathcal{X}),\Sigma^{\ell}_{f}(\mathcal{X})\right) (2)

where

mf(𝒳)\displaystyle m_{f}^{\ell}(\mathcal{X}) =K(𝒳,x𝒦)TK(x𝒦,x𝒦)1f𝒦\displaystyle=\textbf{K}(\mathcal{X},x_{\mathcal{K}})^{T}\textbf{K}(x_{\mathcal{K}},x_{\mathcal{K}})^{-1}f_{\mathcal{K}} (3a)
Σf(𝒳)\displaystyle\Sigma^{\ell}_{f}(\mathcal{X}) =K(𝒳,𝒳)K(𝒳,x𝒦)TK(x𝒦,x𝒦)1K(x𝒦,𝒳)\displaystyle=\textbf{K}(\mathcal{X},\mathcal{X})-\textbf{K}(\mathcal{X},x_{\mathcal{K}})^{T}\textbf{K}(x_{\mathcal{K}},x_{\mathcal{K}})^{-1}\textbf{K}(x_{\mathcal{K}},\mathcal{X}) (3b)

Equation (3) is used to construct an acquisition function (AF) of the form

𝒜(x)=mf(x)κσf(x)\mathcal{AF}(x)=m_{f}^{\ell}(x)-\kappa\cdot\sigma_{f}^{\ell}(x) (4)

that is then optimized to select a new sample point x+1x^{\ell+1}. Note that the parameter κ+\kappa\in\mathbb{R}_{+}, known as the exploration weight, determines the importance placed on the model uncertainty and can be modified to make the algorithm either more exploitative or explorative. After taking a sample at x+1x^{\ell+1}, the dataset is updated and the model can be retrained. This process is repeated until a satisfactory solution is found or the data collection budget is exhausted. For a more complete treatment of BO we refer reader to Garnett (2023).

2.2 Monte Carlo approach for composite functions

Optimization of a composite objective using BO is introduced in Astudillo and Frazier (2019). In this context, the performance function is now a known composition of the form f(x,y(x))f(x,y(x)) with f:X×Yf:X\times Y\to\mathbb{R}. The inner or intermediate function y:Xdyy:X\to\mathbb{R}^{d_{y}} is a vector-valued function with range YdyY\subseteq\mathbb{R}^{d_{y}} and is now the unknown, expensive-to-evaluate function. Note that in this approach, the GP model no longer estimates the system’s performance as in standard BO but is instead generating estimates for y(x)y(x), which serve as inputs to the performance function. The system is then no longer a black-box but rather a composition of interconnected black-boxes whose relation to each other and contributions to the system’s performance are known via f(x,y(x))f(x,y(x)). Additionally, because ff is a known function that can be easily evaluated, its derivatives are also available enabling the use of derivative-based techniques (gradient descent, Newton’s method, etc.) to optimize the function. Thus, in this context, it would be be more precise to consider the system a partially observable ”grey-box” rather than a black-box as shown in Figure 2.

In order to apply this paradigm in a BO setting, we must be able to provide the acquisition function with the mean and uncertainty estimates of ff. However, because the performance function is no longer being approximated by the GP model, these are no longer as readily available as they were in the standard setting. Therefore, it is necessary to translate the mean and uncertainty estimates provided for yy, my(x)m_{y}^{\ell}(x) and Σy(x)\Sigma_{y}^{\ell}(x) respectively, into the appropriate values for mf(x)m_{f}^{\ell}(x) and σf(x)\sigma_{f}^{\ell}(x). In the case where ff is a linear transformation of yy of the form f=aTy+bf=a^{T}y+b, then, given that the GP model assumes a normal distribution for yy, ff is normally distributed with

mf(x)\displaystyle m_{f}^{\ell}(x) =aTmy(x)+b\displaystyle=a^{T}m_{y}^{\ell}(x)+b (5a)
σf(x)\displaystyle\sigma^{\ell}_{f}(x) =(aTΣy(x)a)12\displaystyle=\left(a^{T}\Sigma^{\ell}_{y}(x)a\right)^{\frac{1}{2}} (5b)

However, in the more general case where ff is a nonlinear transformation, this property can no longer be used, and closed-form expressions for mf(x)m_{f}^{\ell}(x) and σf(x)\sigma_{f}^{\ell}(x) are not readily available. Various methods have proposed using Monte Carlo to address this problem, see Astudillo and Frazier (2019); Balandat et al. (2020); Paulson and Lu (2022). Using this approach, the statistical moments of ff are estimated by passing samples from the GP models for y(x)y(x) into f(x,y(x))f(x,y(x)) and then numerically estimating the mean and variance of the performance function as follows

m^f\displaystyle\hat{m}_{f}^{\ell} =1Ss=1Sf(x,my(x)+Ay(x)zs)\displaystyle=\frac{1}{S}\sum_{s=1}^{S}f(x,m_{y}^{\ell}(x)+A_{y}^{\ell}(x)z_{s}) (6a)
σ^f\displaystyle\hat{\sigma}^{\ell}_{f} =1S1s=1S(f(x,my(x)+Ay(x)zs)m^f)2\displaystyle=\frac{1}{S-1}\sqrt{\sum_{s=1}^{S}\left(f(x,m^{\ell}_{y}(x)+A_{y}^{\ell}(x)z_{s})-\hat{m}^{\ell}_{f}\right)^{2}} (6b)

where Ay(x)dy×dyA_{y}^{\ell}(x)\in\mathbb{R}^{d_{y}\times d_{y}} is the Cholesky factor of the GP covariance (Ay(Ay)T=Σy)\left(A_{y}^{\ell}(A_{y}^{\ell})^{T}=\Sigma_{y}^{\ell}\right) and zdyz\in\mathbb{R}^{d_{y}} is a random vector drawn from 𝒩(0,I)\mathcal{N}(\textbf{0},\textbf{I}). These estimates are then passed into the the AF presented in (4).

While MC provides a convenient manner for determining the distribution of ff, it is a computationally intensive method for doing so. The number of samples required to accurately estimate mf(x)m_{f}^{\ell}(x) and σf(x)\sigma_{f}^{\ell}(x) can be quite large (on the order of 10310^{3} or more) in regions of the design space with high model uncertainty. As a result, drawing the number of samples, SS, needed from a GP, which scales as 𝒪(S3)\mathcal{O}(S^{3}), see Shahriari et al. (2016), can require a significant amount of computational time. Additionally, while ff is a known function and is significantly cheaper to evaluate than the system, at large values of SS the cost of repeatedly calculating the value of f(x,y(x))f(x,y(x)) can also become nontrivial. This issue is compounded by the fact that (6) must be recalculated at every point of interest.

Refer to caption
Figure 2: Grey-box representation of a composite function system with black-box intermediate functions

3 The BOIS Approach

Refer to caption
Figure 3: Representation of the BOIS framework, note that bb represents the deterministic terms in the linearized expression

The fundamental challenge of composite function BO algorithm is the lack of closed-form expressions for mf(x)m_{f}^{\ell}(x) and σf(x)\sigma_{f}^{\ell}(x), which are needed to build the AF. As previously mentioned, obtaining these when ff is not a linear mapping of yy is generally not possible. However, when ff is a once-differentiable mapping, it is possible to conduct a linearization at the current iterate (as is done in standard optimization algorithms such as Newton’s method). If we choose to represent ff as

f(x,y)=g(x)+h(x,y)f(x,y)=g(x)+h(x,y) (7)

such that g(x)g(x) includes the terms that do not depend on yy and h(x,y)h(x,y) includes those that do, we can use a first-order Taylor series expansion to linearize ff with respect to yy around some reference point y0y_{0}:

f(x,y)g(x)+h(x,y0)+JT(yy0)f(x,y)\approx g(x)+h(x,y_{0})+J^{T}(y-y_{0}) (8)

where

J\displaystyle J =yh(x,y0)\displaystyle=\nabla_{y}h(x,y_{0}) (9a)
=yf(x,y0)\displaystyle=\nabla_{y}f(x,y_{0}) (9b)

Using this approach, if we then select some point of interest x+1x^{\ell+1} where the mean and covariance of y(x)y(x) are given by the GP model as y^=my(x+1)\hat{y}^{\ell}=m_{y}^{\ell}(x^{\ell+1}) and Σ^=Σy(x+1)\hat{\Sigma}^{\ell}=\Sigma_{y}^{\ell}(x^{\ell+1}) respectively and some reference point in the neighborhood of y^\hat{y}^{\ell}, which we denote as y^0\hat{y}_{0}^{\ell}, we can obtain the following estimate for ff at x+1x^{\ell+1}

f(x+1,y(x+1))\displaystyle f(x^{\ell+1},y(x^{\ell+1})) g(x+1)+h(x+1,y0^)\displaystyle\approx g(x^{\ell+1})+h(x^{\ell+1},\hat{y_{0}}^{\ell})
+JT(y(x+1)y^0)\displaystyle+J^{T}(y(x^{\ell+1})-\hat{y}_{0}^{\ell}) (10a)

Note that we make the assumption that g(x)g(x) is also a known, easy-to-evaluate function and, therefore, g(x+1)g(x^{\ell+1}) is a deterministic variable. Thus, we are now able to derive at a set of closed-form expressions for the mean and uncertainty of the performance function by making use of the closure of normal distributions under linear transformations.

mf(x+1)\displaystyle m_{f}^{\ell}(x^{\ell+1}) =JTy^+g(x+1)+h(x+1,y^0)JTy^0\displaystyle=J^{T}\hat{y}^{\ell}+g(x^{\ell+1})+h(x^{\ell+1},\hat{y}_{0}^{\ell})-J^{T}\hat{y}_{0}^{\ell} (11a)
σf(x+1)\displaystyle\sigma^{\ell}_{f}(x^{\ell+1}) =(JTΣ^J)12.\displaystyle=\left(J^{T}\hat{\Sigma}^{\ell}J\right)^{\frac{1}{2}}. (11b)

The proposed framework (which we call BOIS and is shown in Figure 3) is built on the expressions derived in (11). The reason we are able to generate these explicit formulations is due to the manner in which linearizing constructs the density of ff. While the MC approach is agnostic to the nature of the density of ff, BOIS approximates it using a Gaussian model around the neighborhood of the iterate y^0\hat{y}_{0}^{\ell}. In other words, BOIS generates a local Laplace approximation of the performance function by passing the mean and uncertainty estimates of y(x)y(x) given by the GP model into (10). This allows us to obtain closed-form expressions for the statistical moments of ff, such as mfm_{f}^{\ell} and σf\sigma_{f}^{\ell}, which can be used for constructing AFs. Given that this approximation of the density of ff is Gaussian, it is also possible to obtain expressions for the probabilities and quantiles (to construct different types of AFs). This assumption about the shape of ff also means that as the distance between y^\hat{y}^{\ell} and y^0\hat{y}_{0}^{\ell} grows, the Laplace approximation will result in a worse fit, similar to how the linearization itself becomes less accurate. However, it is important to note that the linearization is updated in an adaptive manner (by linearizing around the current iterate). If we compare BOIS to MC-driven approaches, we can see that at any point x+1x^{\ell+1} of interest, BOIS only samples from the GP once to determine y^\hat{y}^{\ell} and Σ^\hat{\Sigma}^{\ell} and evaluates ff once to calculate f(x+1,y^0)f(x^{\ell+1},\hat{y}_{0}); recall that this is done tens to thousands of times in MC. While BOIS also has to compute (9), this is also done only once and has been shown to have a computational cost similar to that of evaluating ff when methods like automatic differentiation are used, see Griewank and Walther (2008); Baur and Strassen (1983). As a result, we are able calculate values for mfm_{f}^{\ell} and σf\sigma_{f}^{\ell} at a significantly lower computational cost than when using MC.

4 Numerical Experiments

4.1 Optimization of a chemical process

Refer to caption
Figure 4: Iteration number vs average operating cost for the tested algorithms (left) and their distribution profiles across all runs with average performance shown in color (right)

Consider the following chemical process: two reagents, AA and BB, are compressed and heated and then fed into a reactor where they react to form product CC. The reactor effluent is sent to a separator where the product is recovered as a liquid. A fraction of the vapor stream exiting the separator, which contains mostly unreacted AA and BB, is recycled and fed back to the reactor after being heated and compressed, and the remainder is purged. Our aim is to determine the operating temperatures and pressures of the reactor and separator as well as the recycle fraction that will minimize the operating cost of the process. With this goal in mind, we formulate our cost function as

f1(y(x))\displaystyle f_{1}(y(x)) =j{A,B}wj0Fj+Fpi{A,B,C}wi1ψi\displaystyle=\sum_{j\in\{A,B\}}w_{j0}F_{j}+F_{p}\sum_{i\in\{A,B,C\}}w_{i1}\psi_{i}
+Foi{A,B,C}wi2ξi+w3(ψCFpF¯F¯)2\displaystyle+F_{o}\sum_{i\in\{A,B,C\}}w_{i2}\xi_{i}+w_{3}\left(\frac{\psi_{C}F_{p}-\bar{F}}{\bar{F}}\right)^{2} (12a)
f2(y(x))\displaystyle f_{2}(y(x)) =h=15whQ˙h,+wek=13W˙k\displaystyle=\sum_{h=1}^{5}w_{h}\dot{Q}_{h},+w_{e}\sum_{k=1}^{3}\dot{W}_{k} (12b)
f(y(x))\displaystyle f(y(x)) =f1(y(x))+f2(y(x))\displaystyle=f_{1}(y(x))+f_{2}(y(x)) (12c)

Here, FjF_{j} are the flowrates of reagents into the process. The product and purge streams exit the process at rates FpF_{p} with composition ψi\psi_{i} and FoF_{o} with composition ξi\xi_{i} respectively. The heat and power requirements of the process units are denoted as Q˙h\dot{Q}_{h} and W˙k\dot{W}_{k}. The costs of reagents and heat and power utilities are wjw_{j}, whw_{h} and wew_{e} respectively while wi1w_{i1} and wi2w_{i2} refer to the values of species ii in the product and purge streams. We are also targeting a certain production rate for CC, F¯\bar{F}, which we choose to enforce by incurring an additional cost, w3w_{3}, when the process operates at a different value of FpF_{p}. We define our design space as the box domain X=[673,250,288,140,0.5]×[973,450,338,170,0.9]X=[673,250,288,140,0.5]\times[973,450,338,170,0.9]; the optimal solution is at x=(844,346,288,170,0.9)x=(844,346,288,170,0.9). For the composite function BO algorithms, the intermediate black-box functions, y(x)y(x), are set to model the flowrates and compositions of the product and purge streams as well as the heat and power loads of the process. All algorithms tested, standard BO (S-BO), MC-driven composite function BO (MC-BO), and BOIS, were initialized at the same point; we conducted 243 trials, each at a different starting point on a 353^{5} grid of XX.

The left side of Figure 4 illustrates the average performance of the algorithms across all of the runs. We observe that, on average, BOIS is able to return a better solution than both S-BO and MC-BO. Additionally, we also see that during the first 10 trials BOIS samples from regions with significantly higher costs than the other two algorithms. We attribute this to the fact that, due to how σf(x)\sigma_{f}^{\ell}(x) is calculated in (11), large values in JJ and Σy(x)\Sigma_{y}^{\ell}(x) can result in the uncertainty estimate for f(y(x))f(y(x)) being high. This pushes the algorithm to be more explorative, especially at the beginning of the run when the model uncertainty of y(x)y(x) is at its highest. From the sharpness of this peak we can also determine that BOIS quickly moves away from these highly non-optimal areas; after 15 iterations it is exploring regions with similar values as BO and MC-BO.

The performance distributions shown on the right side of Figure 4 clearly illustrate the benefit of using a composite representation of the performance function. We observe that while the sampling behavior of BOIS and MC-BO is relatively stable after approximately 60 iterations across all runs, several instances of S-BO are still sampling from high-cost regions even at the end of their runs. The cause of this is the flow penalty term included in (12). By providing the composite function algorithms with the form of ff, we enable them to easily determine values of FpF_{p} that result in a small penalty. S-BO does not have access to this information and, as a result, has no way of knowing that the penalty is there and has to rely on sampling to learn it. From the results illustrated in the figure, we can surmise that this is a difficult task for the algorithm.

4.2 Statistical consistency of BOIS

Refer to caption
Figure 5: Parity plots of the values of mfm_{f}^{\ell} and σf\sigma_{f}^{\ell} at various points in XX calculated by BOIS with y^y^0=y^×103\hat{y}-\hat{y}_{0}=\hat{y}\times 10^{-3} and MC-BO using samples sizes S=10S=10, 10210^{2}, and 10410^{4}; the same trained GP model of y(x)y(x) was used by both approaches

We decided to use the accuracy of the statistical moments of ff calculated by BOIS as a metric for comparing its efficacy with that of MC-BO. We know that as we increase the number of samples (6) will return values closer to the true moments of ff. Using a trained GP model of y(x)y(x) we generated estimates for mfm_{f}^{\ell} and σf\sigma_{f}^{\ell} at various points in the design space using MC-BO with 10, 100, and 10,000 samples. We then used this same GP model and set y^y^0=y^×103\hat{y}-\hat{y}_{0}=\hat{y}\times 10^{-3} in the linearization step to obtain the corresponding estimates from BOIS. If the values given by BOIS are accurate, then we should expect that the difference between its estimates and those of MC-BO should decrease as SS increases. The results presented in Figure 5 demonstrate that this precisely the case. The shift in the difference between the estimates is especially clear when looking at the values of σf\sigma_{f}^{\ell} calculated by the two algorithms. The large discrepancies seen when using 10 samples shrink significantly when SS is increased to 100 and are virtually gone when S=S= 10,000. If we consider the time it took to generate these estimates, approximately 10 seconds for BOIS and 1 hour for MC-BO when using 10,000 samples, we can conclude that, not only is BOIS faster than MC-BO, it can also be just as accurate. This further reinforces our claim that BOIS is an efficient method for using composite functions in a BO setting.

5 Conclusions

We presented a framework, which we refer to as BOIS, to enable the use of composite functions f(x,y(x))f(x,y(x)) in BO to allow for the exploitation of structural knowledge (in the form of physics or sparse interconnections). The key contribution of this work is the derivation of a set of closed-form expressions for the moments of the composite, mf(x)m_{f}^{\ell}(x) and σf(x)\sigma_{f}^{\ell}(x), based on adaptive linearizations; this overcomes the tractability challenges of sampling based approaches. These expressions are obtained by linearizing f(x,y(x))f(x,y(x)) in order to generate a Laplace approximation around the current iterate. We demonstrate that BOIS outperforms standard BO by making use of the structure conveyed by f(x,y(x))f(x,y(x)). We also show that the statistical moments calculated by BOIS accurately represent the statistical moments of ff and that these estimates can be obtained in significantly less time than sampling-based approaches. As part of our future work, we plan to scale up BOIS and deploy it on high-dimensional systems where BO has traditionally not been applied and to obtain alternative types of AFs. Finally, while the GP is the most popular surrogate model choice in BO, the algorithm is not limited to using only GPs, any probabilistic model can be used. Therefore, we would like to explore the use of alternative models such as warped GPs, see Snelson et al. (2004), RNNs, see Thompson et al. (2023), and reference models, see Lu et al. (2021).

References

  • Astudillo and Frazier (2019) Astudillo, R. and Frazier, P. (2019). Bayesian optimization of composite functions. In K. Chaudhuri and R. Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, 354–363. PMLR.
  • Astudillo and Frazier (2021) Astudillo, R. and Frazier, P. (2021). Thinking inside the box: A tutorial on grey-box Bayesian optimization. In Proceedings of the 2021 Winter Simulation Conference.
  • Balandat et al. (2020) Balandat, M., Karrer, B., Jiang, D., Daulton, S., Letham, B., Wislon, A., and Bashky, E. (2020). BOTORCH: A framework for efficient Monte-Carlo Bayesian optimization. In Proceedings of the 34th Conference International Conference on Neural Information Processing Systems, NIPS ’20, 21524–21538. Curran Associates Inc.
  • Baur and Strassen (1983) Baur, W. and Strassen, V. (1983). The complexity of partial derivatives. Theoretical Computer Science, 22(3), 317–330.
  • Boukouvala and Floudas (2017) Boukouvala, F. and Floudas, C. (2017). ARGONAUT: AlgoRithms for Global Optimization of coNstrAined grey-box compUTational problems. Optimization Letters, 11(5), 895–913.
  • Conn et al. (2009) Conn, A., Scheinberg, K., and Vicente, L. (2009). Introduction to Derivative-free Optimization. SIAM.
  • Garnett (2023) Garnett, R. (2023). Baysian Optimization. Cambridge University Press.
  • Griewank and Walther (2008) Griewank, A. and Walther, A. (2008). Evaluating Derivatives: Principles and Techniques of Algorithmic Differentiation. SIAM, Philadelphia.
  • Kandasamy et al. (2017) Kandasamy, K., Dasarathy, G., Schnieder, J., and Pózcos, B. (2017). Multi-fidelity Bayesian optimisation with continuous approximations. In D. Precup and Y. Teh (eds.), Uncertainty in Artificial Intelligence, volume 70 of Proceedings of Machine Learning Research, 1799–1808. PMLR.
  • Lu et al. (2021) Lu, Q., González, L., Kumar, R., and Zavala, V. (2021). Bayesian optimization with reference models: A case study in MPC for HVAC central plants. Computers & Chemical Engineering, 154, 107491.
  • Mockus (2012) Mockus, J. (2012). Bayesian Approach to Global Optimization: Theory and Applications. Springer Science & Business Media.
  • Paulson and Lu (2022) Paulson, J. and Lu, C. (2022). COBALT: COnstrained Bayesian optimizAtion of computaionaLly expensive grey-box models exploiting derivaTive information. Computers & Chemical Engineering, 160, 107700.
  • Raissi et al. (2019) Raissi, M., Perdikaris, P., and Karniadakis, G. (2019). A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational Physics, 378, 686–707.
  • Shahriari et al. (2016) Shahriari, B., Swersky, K., Wang, Z., Adams, R., and de Freitas, N. (2016). Taking the human out of the loop: A review of Bayesian optimization. Proceedings of the IEEE, 104, 148–175.
  • Snelson et al. (2004) Snelson, E., Ghahramani, Z., and Rasmussen, C. (2004). Warped Gaussian processes. In S. Thrun, L. Saul, and B. Schölkopf (eds.), Advances in Neural Information Processing Systems, volume 16, 337–334. MIT Press.
  • Thompson et al. (2023) Thompson, J., Zavala, V., and Venturelli, O. (2023). Integrating a tailored recurrent neural network with Bayesian experimental design to optimize microbial community functions. PLOS Computational Biology, 19(9), 1–25.
  • Uhrenholt and Jensen (2019) Uhrenholt, A. and Jensen, B. (2019). Efficient Bayesian optimization for target vector estimation. In K. Chaudhuri and M. Sugiyama (eds.), Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89 of Proceedings of Machine Learning Research, 2661–2670. PMLR.