25
Journal of Machine Learning Research 21 (2020) 1-48 Submitted 4/20; Published / Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization Jingfan Chen [email protected] Guanghui Zhu * [email protected] Chunfeng Yuan [email protected] Yihua Huang * [email protected] State Key Laboratory for Novel Software Technology Nanjing University Nanjing 210023, China Editor: Abstract Bayesian optimization is a broadly applied methodology to optimize the expensive black- box function. Despite its success, it still faces the challenge from the high-dimensional search space. To alleviate this problem, we propose a novel Bayesian optimization frame- work, which finds a low-dimensional space to perform Bayesian optimization through a semi-supervised, iterative, and embedding learning-based method (SILBO). SILBO in- corporates both labeled and unlabeled points acquired from the acquisition function of Bayesian optimization to guide the learning of the embedding space. To accelerate the learning procedure, we present a randomized method for generating the projection ma- trix. Furthermore, to map from the low-dimensional space to the high-dimensional original space, we propose two mapping strategies: SILBO-BU and SILBO-TD according to the evaluation overhead of the objective function. Experimental results on both synthetic function and hyperparameter optimization tasks demonstrate that SILBO outperforms the existing state-of-the-art high-dimensional Bayesian optimization methods. Meanwhile, the semi-supervised dimensional reduction and iterative learning in SILBO are effective for high-dimensional Bayesian optimization. Keywords: Bayesian Optimization, High-dimensional Optimization, Semi-supervised Dimension Reduction, Embedding Learning 1. Introduction As a well-established approach for sample-efficient global optimization of black-box func- tions that are expensive to evaluate, Bayesian optimization (BO) is used in many tasks such as hyperparameter tuning (Hutter et al., 2011; Bergstra et al., 2011; Snoek et al., 2012), neural architecture search (Kandasamy et al., 2018), and chemical structure search (G´ omez-Bombarelli et al., 2018). BO provides a principled method for finding the global optimum of black-box function: using the cheap probability surrogate model of black-box function as the input to the acquisition function, repeatedly considering the trade-off be- tween exploitation and exploration to select the promising points. The surrogate model is *. Corresponding authors with equal contribution ©2020 Jingfan Chen, Guanghui Zhu, Chunfeng Yuan, and Yihua Huang. arXiv:2005.14601v3 [cs.LG] 19 Oct 2020

arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Journal of Machine Learning Research 21 (2020) 1-48 Submitted 4/20; Published /

Semi-supervised Embedding Learning for High-dimensionalBayesian Optimization

Jingfan Chen [email protected]

Guanghui Zhu∗ [email protected]

Chunfeng Yuan [email protected]

Yihua Huang∗ [email protected]

State Key Laboratory for Novel Software Technology

Nanjing University

Nanjing 210023, China

Editor:

Abstract

Bayesian optimization is a broadly applied methodology to optimize the expensive black-box function. Despite its success, it still faces the challenge from the high-dimensionalsearch space. To alleviate this problem, we propose a novel Bayesian optimization frame-work, which finds a low-dimensional space to perform Bayesian optimization through asemi-supervised, iterative, and embedding learning-based method (SILBO). SILBO in-corporates both labeled and unlabeled points acquired from the acquisition function ofBayesian optimization to guide the learning of the embedding space. To accelerate thelearning procedure, we present a randomized method for generating the projection ma-trix. Furthermore, to map from the low-dimensional space to the high-dimensional originalspace, we propose two mapping strategies: SILBO-BU and SILBO-TD according to theevaluation overhead of the objective function. Experimental results on both syntheticfunction and hyperparameter optimization tasks demonstrate that SILBO outperforms theexisting state-of-the-art high-dimensional Bayesian optimization methods. Meanwhile, thesemi-supervised dimensional reduction and iterative learning in SILBO are effective forhigh-dimensional Bayesian optimization.

Keywords: Bayesian Optimization, High-dimensional Optimization, Semi-supervisedDimension Reduction, Embedding Learning

1. Introduction

As a well-established approach for sample-efficient global optimization of black-box func-tions that are expensive to evaluate, Bayesian optimization (BO) is used in many taskssuch as hyperparameter tuning (Hutter et al., 2011; Bergstra et al., 2011; Snoek et al.,2012), neural architecture search (Kandasamy et al., 2018), and chemical structure search(Gomez-Bombarelli et al., 2018). BO provides a principled method for finding the globaloptimum of black-box function: using the cheap probability surrogate model of black-boxfunction as the input to the acquisition function, repeatedly considering the trade-off be-tween exploitation and exploration to select the promising points. The surrogate model is

∗. Corresponding authors with equal contribution

©2020 Jingfan Chen, Guanghui Zhu, Chunfeng Yuan, and Yihua Huang.

arX

iv:2

005.

1460

1v3

[cs

.LG

] 1

9 O

ct 2

020

Page 2: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

constructed based on the evaluation values observed so far. A widely-used surrogate modelis Gaussian Process regression, which provides the uncertainty quantification of the functionvalue by imposing a Gaussian Process prior.

While BO provides such an automated procedure, it still faces a huge challenge in high-dimensional scenarios. To ensure the converge for learning the function response value, thesample complexity depends exponentially on the number of dimensions (Shahriari et al.,2016). In practice, BO is limited to the optimization problem with around 20 dimensionswhen using Gaussian Process regression as a surrogate model (Frazier, 2018).

To handle high-dimensional Bayesian optimization, many methods have been proposed.Based on the assumption that only a small number of dimensions influence the responsevalue of the objective function, the embedding methods perform BO on a low-dimensionalspace. The corresponding projection matrix can be constructed randomly (Wang et al.,2013; Nayebi et al., 2019), or learned actively (Djolonga et al., 2013; Zhang et al., 2019).Some methods impose an additive structure on the objective function (Gardner et al.,2017; Kandasamy et al., 2015). Besides, many methods start from the way of learninglow-dimensional embedding and find an effective subspace through nonlinear dimensionreduction (Lu et al., 2018). However, there are two limitations in the existing methods.First, the projection matrix is immutable. Once the generated low-dimensional embeddingcannot represent the intrinsic structure of the objective function, finding the global optimumthrough Bayesian optimization will become very difficult. Second, the low-dimensional spaceis learned in a supervised way. The label of each point in the learning-based dimensionalreduction method indicates the response value of the black-box function. To learn aneffective low-dimensional space, a large number of labeled points are required, which leadsto huge computation costs especially when the evaluation overhead of the objective functionis expensive.

In this paper, we propose a novel framework called SILBO1 to mitigate the problemof the curse of dimensionality by learning the effective low-dimensional space iterativelythrough the semi-supervised dimension reduction method. After a low-dimensional spaceis identified, Bayesian optimization is performed in the low-dimensional space, leading to astable and reliable estimation of the global optimum. Specifically, the contribution of thispaper is as follows:

• We propose an iterative method in SILBO to update the projection matrix dynam-ically. During each iteration of BO, a semi-supervised low-dimensional embeddinglearning method is proposed to construct the projection matrix by utilizing both la-beled and unlabeled points acquired from the acquisition function of BO.

• To accelerate the semi-supervised dimension reduction, we further propose a ran-domized method to compute the high-dimensional generalized eigenvalue problemefficiently. We also analyze its time complexity in detail.

• Furthermore, to map from the low-dimensional space to the high-dimensional originalspace, we propose two mapping strategies: SILBO-BU and SILBO-TD according tothe evaluation overhead of the objective function.

1. SILBO stands for Semi-supervised, Iterative, and Learning-based Bayesian Optimization. The code isavailable at https://github.com/cjfcsjt/SILBO.

2

Page 3: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

• Experimental results on both synthetic function and neural network hyperparam-eter optimization tasks reveal that SILBO outperforms the existing state-of-the-arthigh-dimensional Bayesian optimization methods. Meanwhile, the semi-supervised di-mensional reduction and iterative learning in SILBO are effective for high-dimensionalBayesian optimization.

The rest of this paper is organized as follows. Section 2 gives an overview of relatedwork. Section 3 states the problem and lists relevant background materials. The SILBOalgorithm is proposed in Section 4. The experimental evaluation is presented in Section 5and the conclusion is given in Section 6.

2. Related Work

Bayesian optimization has achieved great success in many applications with low dimen-sions (Kandasamy et al., 2017; Klein et al., 2017; Swersky et al., 2013; Wu et al., 2019,2017; Hernandez-Lobato et al., 2016). However, extending BO to high dimensions is still achallenge. Recently, the high-dimensional BO has received increasing attention and a largebody of literature has been devoted to addressing this issue.

Given the assumption that only a few dimensions play a decisive role, the linear low-dimensional embedding method achieves dimension reduction using a projection matrix.In REMBO (Wang et al., 2013), the projection matrix is randomly generated accordingto Gaussian distribution. The promising points are searched in the low-dimensional spaceby performing Gaussian Process regression and then mapped back to the high-dimensionalspace for evaluation. It has been proven that REMBO has a great probability to find theglobal optimum by convex projection, although the high probability is not guaranteed whenthe box bound exists. Another problem of REMBO is the over-exploration of the boundary.To address the over exploration, a carefully-selected bound in the low-dimensional embed-ding space was proposed (Binois et al., 2019), which finds the corresponding points in thehigh-dimensional space by solving a quadratic programming problem. The BOCK algorithm(Oh et al., 2018) scales to the high-dimensional space using a cylindrical transformation ofthe search space. HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviatethe over-exploration of the boundary and use the hash technique to improve computationalefficiency. HeSBO also shows that the mean and variance function of Gaussian Process donot deviate greatly under certain condition. However, the above-mentioned methods onlyuse the prior information to generate the projection matrix randomly and do not employthe information of the actual initial points to learn a low-dimensional embedding actively.

Different from the previous methods, the learning-based methods have been proposed.SIRBO (Zhang et al., 2019) uses the supervised dimension reduction method to learn a low-dimensional embedding, while SI-BO (Djolonga et al., 2013) employs the low-rank matrixrecovery to learn the embedding. However, the low-dimensional embedding learned by thesemethods is immutable. Once the projection matrix is generated according to the initialpoints, it will not be updated. In some scenarios, because of the small number of initialpoints that have been evaluated, the low-dimensional embedding space cannot accuratelyreflect the information of the objective function.

Another way for handling the high-dimensional BO is to assume an additive structure(Gardner et al., 2017) of the objective function. Typically, ADD-BO (Kandasamy et al.,

3

Page 4: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

2015) optimizes the objective function on a disjoint subspace decomposed from the high-dimensional space. Unfortunately, the additive assumption does not hold in most practicalapplications. Besides, the non-linear embedding method is also attractive (Eissman et al.,2018; Lu et al., 2018; Moriconi et al., 2019). These non-linear methods use the VariationalAutoencoder (VAE) to learn a low-dimensional embedding space. The advantage of non-linear learning methods is that points in the original space can be easily reconstructedthrough the non-linear mapping. However, training VAE requires a large number of points.When the evaluation cost of the objective function is expensive, the non-linear embeddingmethod is almost impractical.

In this paper, we focus on the linear low-dimensional embedding method and proposean iterative, semi-supervised method to learn the embedding.

3. Preliminary

In this section, we give the problem setup and introduce Bayesian optimization (BO), semi-supervised discriminant analysis (SDA), and slice inverse regression (SIR).

3.1 Problem Setup

Consider the black-box function f : D → [0, 1], defined on a high-dimensional d and continu-ous domain D = [−1, 1]d ⊂ Rd. f(x) is computationally expensive and may be non-convex.Given x ∈ D, we can only access the noisy response value y extracted from f(x) withnoise ε ∼ N (0, σ2). Also, we assume that the objective function contains r 6 d intrinsicdimensions. In other words, given an embedding matrix B ∈ Rr×d with orthogonal rowsand a function g : Rr → [0, 1], f(x) = g(Bx). Our goal is to find the global optimum.

x∗ = arg maxx∈D

f(x) (1)

3.2 Bayesian Optimization

Bayesian optimization is an iterative framework, which combines a surrogate model of theblack-box function with a search strategy that tries to find possible points with large re-sponse values. Given t observation points x1, ...,xt ∈ D and their corresponding evaluationvalues y1, ..., yt, the surrogate model is usually Gaussian Process regression that imposes aprior, f(x1:t) ∼ GP(µ(x1:t), k(x1:t,x1:t)), to the objective function with mean function µat each xi and covariance function or kernel k at each pair of points (xi,xj). The kernelfunction describes the similarity between inputs. One of the widely-used kernel functionsis the Matern kernel. Then, given a new point x∗, the prediction of the response value canbe calibrated by the posterior probability distribution (noise-free).

f(x∗)|f(x1:t) ∼ N (µt(x∗), σt(x

∗)) (2)

µt(x∗) = k(x∗,x1:t)k(x1:t,x1:t)

−1(f(x1:t)− µ(x1:t)) + µ(x∗) (3)

σt(x∗) = k(x∗,x∗)− k(x∗,x1:t)k(x1:t,x1:t)

−1k(x1:t,x∗) (4)

At each iteration of Bayesian optimization, the predictive mean and variance are re-garded as uncertainty quantification, supporting the subsequent acquisition function op-

4

Page 5: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

timization. The acquisition function tries to balance between exploration (high variance)and exploitation (high mean value). The commonly-used acquisition function for searchingpromising points is UCB (Upper Confidence Bound) (Srinivas et al., 2010). UCB tries toselect the next point with the largest plausible response value according to Equation 5.

xt+1 = arg maxx∈D

µt(x) + β1/2t σt(x) (5)

where βt is a parameter set used to achieve the trade-off between exploration and exploita-tion. In this work, we also experiment with EI (Expected Improvement) (Snoek et al.,2012), which is another popular acquisition function.

3.3 Semi-supervised Discriminant Analysis

Semi-supervised discriminant analysis (SDA) (Cai et al., 2007) is a semi-supervised lineardimension reduction algorithm that leverages both labeled and unlabeled points. SDAaims to find a projection that reflects the discriminant structure inferred from the labeleddata points, as well as the intrinsic geometrical structure inferred from both labeled andunlabeled points. SDA is an extension of linear discriminant analysis (LDA). The originalLDA aims to find a projection β such that the ratio of the between-class scatter matrixSb to the total within-class scatter matrix St is maximized. When the number of trainingdata is small, it can easily lead to overfitting. SDA solves this problem by introducing aregularizer J(β) combined with unlabeled information.

β∗ = arg maxβ

β>Sbβ

β>Stβ + αJ(β)(6)

where α is a coefficient used to balance between model complexity and experience loss.

J(β) is constructed by considering a graph S incorporating neighbor information ofthe labeled points and the unlabeled points, where Sij indicates whether xi and xj areneighbors. Motivated from spectral dimension reduction (Belkin and Niyogi, 2001), theregularizer can be defined as J(β) = Σij(β

>xi − β>xj)2Sij for any two points xi and xj .Then, given the dataset X, J(β) can be written as:

J(β) = 2β>X(D − S)X>β = 2β>XLX>β (7)

where D is a diagonal matrix, Dii =∑

j Sij , and L is a Laplacian matrix (Chung, 1997).Finally, SDA can be reformulated as solving the following generalized eigenvalue problem.

Sbβ = λ(St + αXLX>)β (8)

3.4 Slice Inverse Regression

Sliced inverse regression (SIR) (Li, 1991) is a supervised dimension reduction method forcontinuous response values. SIR aims to find an effective low-dimensional space. Thedimension reduction model is:

Y = g(β>1 x, ..., β>r x, ε) (9)

5

Page 6: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

… …𝑋𝑙

points in the high-dimensional

space 𝑫evaluation

value

…𝑋𝑢

Semi-supervised

dimension reduction

Low-dimensional

space 𝑍

Bayesian

Optimization 𝑧𝑙

maximize

the acquisition

function

𝑧𝑢

close to the

maximum value

of the acquisition

function

𝑧𝑢…

map 𝒁 → 𝑫 and evaluate on the objective function

add to 𝑿𝒍

map 𝒁 → 𝑫

add to 𝑿𝒖

Figure 1: An overview of SILBO.Here, Y is the response variable, x ∈ Rd is an input vector, g is an unknown functionwith r + 1 arguments. β1, · · · , βr denotes orthogonal projection vectors, r denotes thedimensions of the e.d.r. (effective dimension reducing) space and ε is noise. The core idea ofSIR is to swap the positions of x and Y. The algorithm cuts response value Y into H slicesand consider the H-dimensional inverse regression curve E(x|Y ) = (E(x1|Y ), ..., E(xH |Y ))rather than regressing Y on x directly. SIR assumes the existence of e.d.r. directions, andthe curve that just falls into an r-dimensional space. SIR finds the e.d.r. directions byminimizing the total within-slice scatter Σx and maximize the between-slice scatter Ση.Similar to LDA, the problem can be reformulated as a generalized eigenvalue problem.

Σηβ = λΣxβ (10)

Give the assumption that only a small number of dimensions influence the response valueof the objective function, SIR can find an effective low-dimensional space without losingthe essential information to predict response values, regardless of whether the points areindependent and identically distributed (i.i.d.). SIR can be easily extended to support semi-supervised dimension reduction. For example, Semi-SIR (Wu et al., 2010) and Semi-KSIR(Huang and Su, 2014) take the unlabeled data into consideration.

4. The SILBO Algorithm

4.1 Overview

In this section, we propose a framework that addresses the high-dimensional optimizationchallenge by learning a low-dimensional embedding space Z associated with a projectionmatrix B. To learn the intrinsic structure of the objective function effectively, we iterativelyupdate B through semi-supervised dimension reduction. Moreover, we propose a random-ized method to accelerate the computation of the embedding matrix B. By performingBO on the learned low-dimensional space Z, the algorithm can approach the z∗ ∈ Z thatcorresponds to the optimum x∗ ∈ D as close as possible.

As shown in Figure 1, given the labeled points Xl and unlabeled points Xu where thelabel represents the evaluation value of the corresponding point, SILBO consists of threetightly-connected phases:

• semi-supervised embedding learning, which tries to find an effective low-dimensionalspace Z of the objective function by utilizing both Xl and Xu;

6

Page 7: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

• performing Gaussian Process regression on the learned low-dimensional embeddingspace and selecting candidate points according to the acquisition function of BO;

• evaluating points by the objective function f , then updating the Gaussian Processsurrogate model and the projection matrix B.

Algorithm 1: SILBO

Input: Objective function f , acquisition function α, high dimension d, effectivedimension r

Output: The optimum x∗

1 Initialize samples X0u, X

0l with size nu and nl;

2 Construct B0 with Algorithm 2;3 Let D0 = B0X

0l , f(X0

l );4 Construct Gaussian Process regression model based on D0;5 Z0

l = ∅,Y0 = ∅;6 for t = 1 to N do7 Ztu = ∅;8 Generate random point set C from the low-dimensional space;9 zl = arg maxz∈C α(z);

10 Ztl = Zt−1l ∪ zl;11 C = C − zl;12 for n = 1 to nu do13 zu = arg maxz∈C α(z);14 Ztu ∪ zu;15 C = C − zu;

16 end17 Construct Xt

l ,Xtu,Dt based on Ztl ,Z

tu;

18 Update Gaussian process regression model based on Dt;19 Update Bt with Algorithm 2 based on Xt

u , Xtl and Yt;

20 end

These phases are also summarized in Algorithm 1. The first step is to construct theprojection matrix B and find an effective low-dimensional space that can keep the infor-mation of the original space as much as possible. Both Xl and Xu are selected accordingto the acquisition function of BO. Xl is the labeled points that have been evaluated on theobjective function. Xu is the unlabeled points whose evaluation values are unknown.

The second step is to find the possible low-dimensional optimum zl ∈ Z which maximizesthe acquisition function α and select several promising unlabeled points zu ∈ Z that canbe used to update B in the next iteration. Then, zl and zu are mapped back to thehigh-dimensional space D through a specific mapping h : Z → D to get xl and xu.

Finally, we compute y by evaluating the objective function f on xl and update the GPmodel by (zl, y). The xl and xu will be added to Xl and Xu respectively for updating theembedding in the next iteration.

The low-dimensional embedding is learned through SIR combined with the semi-supervisedtechnique. The between-slice scatter matrix Ση and total within-slice scatter matrix Σx are

7

Page 8: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

constructed by utilizing the local information as well as the unlabeled points. Then, Bis obtained through solving the generalized eigenvalue problem. Based on the random-ized SVD method, Algorithm 2 is proposed to speed up the solution to this problem.Moreover, we carefully analyze the mapping h between the low-dimensional space and thehigh-dimensional space and further propose two strategies for the scenarios with differentevaluation costs.

4.2 Semi-supervised Dimension Reduction

The assumption is that there is a low-dimensional space that preserves the informationof the objective function f(x) defined in the high-dimensional space. In other words, thedimensionality of D can be reduced without losing the essential information to predictresponse values Y . If there are enough evaluated points for the initialization of Bayesianoptimization, we may be able to explore the intrinsic structure of f(x) through these points.However, for optimization problems with high computational cost, only a few evaluatedpoints are available. Thus, it is difficult to learn proper embedding only through them. Insuch a case, reasonable and effective use of unevaluated points acquired from the acquisitionfunction of BO will be helpful for embedding learning.

Although SDA provides an effective strategy to incorporate the unlabeled data, it isonly suitable for classification problems. Therefore, we need to extend it to the scenarioswhere the response value is continuous. In fact, SIR is equivalent to LDA, and SDA is anextension of LDA as a generalized eigenvalue problem (Wu et al., 2010). Thus, SDA can beapplied to regression setting by taking advantage of the discreteness brought by the slicingtechnique in SIR. We call such extension semi-SIR. Next, we introduce how to constructsemi-SIR to learn a low-dimensional embedding.

First, the evaluated points are sorted and cut into H slices according to their responsevalues. Then, similar to SDA (Cai et al., 2007), our problem can be reformulated as follows.Given labeled points Xl ∈ Rnl×d and unlabeled points Xu ∈ Rnu×d, nl and nu denote thenumber of the labeled and unlabeled points respectively, we aim to find the embeddingmatrix B through solving:

X>WXβ = λX>(I + αL)Xβ (11)

where X ∈ Rn×d denotes a centered data matrix whose rows represent n samples in Rd,n = nu + nl. W is a n× n weight matrix. I can be expressed as:

I =

[I 00 0

](12)

where I is a nl × nl identity matrix.Note that the between-slice scatter matrix X>WX can be written in an elegant linear

algebraic formulation.

X>WX = X>ΩΩ>X = X>[ΩlΩ

>l 0

0 0

]X (13)

where Ωl ∈ Rnl×H denotes the rescaled slice membership matrix for those evaluated sampleswith Ωlij = 1/

√nj if the i-th sample of Xl is a member of the j-th slice. Otherwise, Ωlij = 0.

nj denotes the number of the samples in the j-th slice.

8

Page 9: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

For the labeled points in the same slice, their response values are very close, but thiscloseness may not be retained in D. There could be a small number of points that arefar away from others in each slice. Although these outliers may indicate that there existother areas with large response values of the objective function, they are likely to interferewith the embedding we have learned. Thus, to reveal the main information of the objectivefunction as much as possible, we employ the localization technique. By strengthening thelocal information of each slice, the degeneracy issue in the original SIR can be mitigated(Wu et al., 2010). Next, we illustrate how to construct W in Equation 13 with the localinformation.

We note that Ωl in Equation 13 is a block matrix, and each block corresponds to a slice.

Ωl =

Ω1l

. . .

ΩHl

(14)

For the original between-slice scatter X>WX, Ωl can only indicate whether a samplepoint belongs to a slice. Here, we will strengthen the localization information by introducinga localized weight. For those evaluated samples, we let Ωh

ij = 1/kh if the i-th sample belongsto the k-nearest neighbors of the j-th sample (j can equal to i) in the h-th slice. Otherwise,Ωhij = 0. H denotes the number of slices, k is a parameter for kNN, and kh is the number

of neighbor pairs in the h-th slice.

In summary, we aim to estimate the intrinsic structure of the objective function andfind an effective subspace that reflects the large response value. Then, we are in a betterposition to identify the possible global optimum. Therefore, in combination with Bayesianoptimization whose acquisition function will provide us candidate points with potentiallylarge response values, semi-supervised dimension reduction can alleviate overfitting causedby the small size of labeled points. Thus, by utilizing both the labeled and unlabeled points,we can learn an r-dimensional space that preserves the intrinsic structure of the objectivefunction better.

4.3 Randomized Semi-supervised Dimension reduction

In the high-dimensional scenario, learning a low-dimensional embedding space still requiresa lot of time. The main bottleneck lies in the solution of generalized eigenvalue problemin Equation 11 and the computation of the neighbor information in each slice. The ma-jor computation cost of the traditional method such as the two-stage SVD algorithm isthe decomposition of the large-scale matrix. Fortunately, the randomized SVD (R-SVD)technique (Halko et al., 2011) shed a light on this problem.

The contribution of R-SVD here is that when the effective dimension r n < d,using randomized SVD to approximate a n × d matrix only requires O(ndr), rather thanO(n2d). Moreover, the empirical results find that when solving the generalized eigenvalueproblem, the performance of the randomized algorithm is superior to that of the traditionaldeterministic method in the high-dimensional case (Georgiev and Mukherjee, 2012). Thus,we first replace the two-stage SVD with R-SVD to accelerate the solution of Equation 11.

We note that the decomposition overhead of the XTΩ ∈ Rd×n in Equation 11 is hugewhen d is large. Thus, we decompose X>Ω = U1S1V

>1 using the R-SVD algorithm. Then,

9

Page 10: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

the between-slice scatter matrix can be expressed as:

X>ΩΩ>X = U1S21U>1 (15)

Similarly, due to the symmetry of matrix I + αL, the right hand side of Equation 11can be decomposed as:

X>(I + αL)X = X>MM>X (16)

where M can be constructed through Cholesky decomposition. Thus, the generalized eigen-value problem in Equation 11 can be reformulated as:

1

λe = S−11 U>1 X

>MM>XU1S−11 e (17)

where e = S1UT1 β.

Next, if we let A = S−11 U>1 X>M , then we can see that this is a traditional eigenvalue

problem. A is an r × n matrix, r is small enough and thus we can decompose it throughoriginal SVD naturally: A = U2S2V

>2 . Finally, the embedding matrix can be computed as:

B = (U1S−11 e1, · · · , U1S

−11 er) = U1S

−11 U2 (18)

The original time complexity of the construction of XTΩ ∈ Rd×n and XTM ∈ Rd×n isO(d2nnh) and O(d2nk) respectively, including the computation of the exact kNN for eachpoint, where nh is the number of samples in each slice. In addition, the time complexity ofdecomposition of XTΩ is O(n2d) and the time complexity of construction and computationof Equation 17 is O(ndr). Therefore, the overall time complexity is O(d2nmax(nh, k) +n2d+ ndr), which is prohibitive in the high-dimensional case.

To alleviate the expensive overhead in high-dimensional case, we use the fast kNN(Achlioptas, 2003; Ailon and Chazelle, 2009) to compute the neighbor information, wefind that constructing XTΩ and XTM only requires O(tdnnh) and O(tdnk) respectively.Moreover, using R-SVD to factorize the matrix XTΩ only requires O(ndr). Therefore, Theoverall time complexity reduces to O(tdnmax(nh, k) + ndr). t denotes a fixed parameterabout the logarithm of n.

Algorithm 2 summarizes the above steps.

Algorithm 2: Randomized semi-supervised dimension reduction

Input: Xl ∈ Rnl×d, Xu ∈ Rnu×d, number of slice H; number of nearest neighbor k,effective dimension r

Output: B ∈ Rd×r1 compute X>Ω, X>M from Equation 13 and Equation 16;

2 estimate [U1, S1, V1] = Randomized SVD(X>Ω,r);

3 A = S−11 U1X>M ;

4 factorize [U2, S2, V2]= SVD(A);5 compute B from Equation 18

10

Page 11: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

4.4 Mapping From Low to High

After semi-supervised dimensional reduction, we perform Bayesian optimization with GPon the low-dimensional space Z. But we need to further consider three issues. The first ishow to select the search domain S in the low-dimensional embedding space. The secondis how to specify a mapping h associate with B to map the point z from the embeddingspace Z to the high-dimensional space D where the response value y should be evaluated.The third is how to maintain consistency of the Gaussian process regression model afterupdating B. Only by addressing these issues, the low-dimensional Bayesian optimizationcan be updated by the data pair (z, y).

Before elaborating the selection of S in the first issue, we introduce the definition ofzonotope, which is a convex symmetric polytope.

Definition 1 Zonotope (Le et al., 2013) Given a linear mapping H ∈ Rr×d and vectorp ∈ Rr, an r-zonotope is defined as

Z = z ∈ Rr : z = p +Hx;x ∈ [−1, 1]d

where p is the center vector of the zonotope.

Without loss of generality, let p = 0, we note that H is exactly B computed by Equation18. Thus, the zonotope ZB associate with B is the low-dimensional space where Bayesianoptimization is performed. Then, we introduce the smallest box containing the zonotope.

Lemma 2 Given zonotope Z = z ∈ Rr : z = p + Hx;x ∈ [−1, 1]d, p = 0, the smallestbox containing this zonotope can be computed as

S =

− d∑j=1

|H1j | ,d∑j=1

|H1j |

× · · · ×− d∑

j=1

|Hrj | ,d∑j=1

|Hrj |

Although the boundary of zonotope is difficult to solve, we use the box S as an alternative(Binois et al., 2019).

Next, we focus on the second and third issues. As mentioned before, a mapping hshould connect the high-dimensional space and the low-dimensional space. As a result, forany point in the low-dimensional space, a corresponding point can be found in the high-dimensional space. In this way, the low-dimensional point z and the response value y of thecorresponding high-dimensional point x can be used as the training set to build a Gaussianprocess regression model. Figure 2 shows the relationship between them. We note that thereproduction of the correlation between z and y is the goal of the low-dimensional Gaussianprocess regression model, and the two are closely connected through x. Thus a reasonablechoice of h plays a great influence on such a middle layer. Meanwhile, due to the iterativenature of SILBO, the mapping h will change after updating B, which directly makes thecorrelation between z and y inconsistent before and after the update. Therefore, we needto develop a strategy to maintain consistency. Next, we introduce two different strategiesto address these two issues.

As shown in Algorithm 3, we can naturally get the response value y by evaluating thepoint x = B†z after finding the candidate point z ∈ S through BO. When the projection

11

Page 12: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

Algorithm 3: The bottom-up strategy

Input: Labeled low-dimensional points Ztl , unlabeled low-dimensional points ZtuOutput: Labeled high-dimensional points Xt

l , unlabeled high-dimensional pointsXtu, the training set Dt

1 Xtl = B†tZ

tl ;

2 Xtu = B†tZ

tu;

3 Yt = f(Xtl ), Dt = (Ztl , Yt);

(a) Bottom-up strategy (b) Top-down strategy

Figure 2: The relationship between response value y, high-dimensional point x, and low-dimensional point z. The solid line shows the process of mapping and evaluation withdifferent strategies. The left hierarchy denotes the bottom-up strategy. The dash line showsthe process of re-mapping and re-evaluation. z at the bottom is fixed after updating B.Then, re-map z to find and re-evaluate the corresponding x. The right hierarchy showsthe top-down strategy. The dash line shows the re-mapping process. x and y at the topare fixed after updating B. Then, x is re-mapped to find the corresponding z and a newGaussian Process regression model will be constructed.

matrix B changed, we fix z at the bottom of the hierarchy in Figure 2a and then update yat the top to maintain consistency (bottom-up for short).

Since we use B† to connect the points between the low-dimensional space Z and thehigh-dimensional space D, the actual evaluation space φ is only a part of D (i.e., φ ⊂ D)when B is fixed. While a great probability of containing the optimum in φ has been proved(Wang et al., 2013), one of the preconditions is that there is no limitation to the high-dimensional space of the objective function. In practice, the restriction often exists (e.g.,[−1, 1]d), making the high probability no longer guaranteed. In contrast, due to the iterativenature of the effective low-dimensional space learning mechanism in SILBO (Algorithm 1),we can alleviate this problem. Specifically, more and more φ associate with B will belearned, which gradually reveal the location of the optimum.

However, the bottom-up strategy may bring more evaluation overhead. When we getthe training set Dt = zi, yiTi=1 for BO through T iterations in Algorithm 1, we updateB to Bnew and get zT+1 through the acquisition function α. Due to the updating of B,

12

Page 13: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

the mapping relationship between the high-dimensional and low-dimensional spaces haschanged. As shown in Figure 2a, we need to find the corresponding x again for z in thetraining set and then evaluate x to maintain consistency. When the evaluation of theobjective function is expensive, the computational cost is huge.

To eliminate the re-evaluation overhead, we propose a new strategy. Specifically, wefirst obtain z in the low-dimensional space through the acquisition function α, and thensolve the following equation to find the corresponding x.

arg minx

||Bx− z||2 (19)

When the projection matrix B is changed, we fix x and y at the top of the hierarchy inFigure 2 and then update z in the bottom to maintain consistency (top-down for short).

Algorithm 4: The top-down strategy

Input: Labeled low-dimensional points Ztl , unlabeled low-dimensional points Ztu,labeled high-dimensional points Xt−1

l , the training set Dt−1Output: labeled high-dimensional points Xt

l , unlabeled high-dimensional pointsXtu, the training set Dt

1 Get the last point zl from Ztl ;2 Compute xl,X

tu from Equation 19 using zl,Z

tu;

3 Xtl = Xt−1

l ∪ xl;4 y = f(xl), Dt = Dt−1 ∪ (zl, y);

Different from the bottom-up strategy, the top-down strategy shown in Figure 2b updatesz directly, which enables us to reconstruct a new BO model efficiently only by replacing thetraining set with Bnewxi, yiTi=1 after updating B, instead of relocating x and re-evaluatingthem. The top-down strategy can be found in Algorithm 4. Next, we analyze the rationalityof this strategy theoretically.

Theorem 3 Given a matrix B ∈ Rr×d with orthogonal rows and its corresponding zonotopeZB. Given z ∈ ZB, for any x1,x2 ∈ UB = x|x = B>z + y,x ∈ Rd, y ∈ N(B),Bx1 = Bx2 = z.

Proof Let x = xV + x⊥V ∈ Rd be the orthogonal decomposition with respect to V =Row(B), then xV ∈ V . Let z ∈ ZB with B>z = xV . Due to the orthogonal decompos-tion, we have x − xV = x − B>z lies in V ⊥ = N(B). For any x1,x2 ∈ UB, we havex1 − xV = x1 − B>z and x2 − xV = x2 − B>z, then B(x1 − B>z) = B(x2 − B>z) = 0.Thus, Bx1 = Bx2 = BB>z = z.

According to Theorem 3, if we assume that B has orthogonal rows, it is clear that givenB and z, any point x ∈ UB will be the solution to Equation 19. However, our purpose is notx itself, but y = f(x), because (z, y) is the data pair used to update BO. Fortunately, if weuse SILBO to learn an ideal effective low-dimensional subspace B∗, then for any x1,x2 in thesolution set UB∗ , f(x1) = f(B∗>z+y1) = f(B∗>z), f(x2) = f(B∗>z+y2) = f(B∗>z) (i.e.,f(x1) = f(x2)). Thus, in each iteration T , we can obtain the unique data pair (zT , yT )

13

Page 14: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

under B∗ to update the low-dimensional Bayesian optimization model. Therefore, thediversity of solutions in the set UB does not affect the construction of BO and we canexplore in the original space D which liberates the shackles of φ.

In summary, both the two strategies aim to generate a consistent training set for thesubsequent GP construction when updating B. To maintain such consistency, the bottom-up strategy acquires more observations y from the objective function according to z in thetraining set while the top-down strategy directly changes z according to y. Note that themore response values, the more information about the objective function we can get. Thus,the bottom-up strategy can obtain more clues about the global optimum than the top-downstrategy.

5. Numerical Results

In this section, we evaluated our proposed algorithm SILBO on a set of high-dimensionalbenchmarks. First, we compared SILBO with other state-of-the-art algorithms on syntheticfunction2 and neural network hyperparameter optimization tasks. Second, we demonstratedthe effectiveness of semi-supervised embedding learning and iterative learning. Finally, weevaluated and analyzed the scalability of SILBO.

5.1 Experiment Setting

For comparison, we also evaluated other state-of-the-art algorithms: REMBO (Wang et al.,2013), which performs BO on a random generated linear embedding space, REMBO-KX ,which computes the kernel using the distance in the high-dimensional space, REMBO-Kψ

(Binois et al., 2015), which uses a warping kernel and finds the evaluation point throughsophisticated mapping, and HeSBO (Nayebi et al., 2019), which uses the count-sketch tech-nique to generate subspace and compute the projection matrix. Additional methods includeADD-BO (Kandasamy et al., 2015), which assumes that the objective function has an addi-tive structure, and recently-proposed SIR-BO (Zhang et al., 2019), which uses SIR to learna low-dimensional space purely. SILBO-BU and SILBO-TD represent the proposed SILBOalgorithm that employs the bottom-up and top-down mapping strategies respectively.

We employed the package GPy3 as the Gaussian Process framework and selected theMatern kernel for the GP model. We adopted the package ristretto4 to perform the R-SVD. We employed CMA-ES (Hansen, 2016) to compute Equation 19 for SILBO-TD. Theembedding matrix B is updated every 20 iterations. The size of the unlabeled points isset to 50 in each iteration. The number of neighbors k is set to 7 and the number ofunlabeled points is set to 50 for semi-supervised dimension reduction. To obtain errorbars, we performed 100 independent runs for each algorithm on synthetic functions and5 runs for the neural network hyperparameter search. We plotted the mean value alongwith one standard deviation. For ADD-BO, we used the authors’ implementation throughMATLAB, so we did not compare its scalability since other algorithms were implementedin Python. Also, we evaluated SILBO using two acquisition functions: UCB and EI. Thenumber of initial random points defaults to 50 and the response values of initial points are

2. The synthetic function can be found at https://www.sfu.ca/~ssurjano/optimization.html.3. https://github.com/SheffieldML/GPy4. https://github.com/erichson/ristretto

14

Page 15: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

(a) 100d Branin (b) 1000d Branin

Figure 3: Performance on Branin under dimension 100(a) and dimension 1000(b) withembedding dimension r = 2. We plotted mean and one standard deviation across 100independent runs.

(a) 100d hartmann6 (b) 1000d hartmann6

Figure 4: Performance on Hartmann6 under dimension 100(a) and dimension 1000(b) withembedding dimension r = 6. We plotted mean and one standard deviation across 100independent runs.

not considered when observing the experimental performance except in Section 5.5, whichuses different number of initial points, so the response values of these points are shown inthe performance curve.

5.2 Performance on Synthetic Functions

Similar to (Nayebi et al., 2019), these algorithms were under comparison upon the followingwidely-used test functions: (1) Branin (2) Hartmann-6 (3) Colville. The active dimen-sions for Branin, Hartmann-6, Colville are 2, 6, and 4 respectively. We studied the caseswith different input dimensions d ∈ 100, 1000. The experimental results are shown inFigure 3-5. The proposed SILBO-BU, outperforms all benchmarks especially in high di-mensions (d = 1000). SILBO-TD is also very competitive, surpassing most traditionalalgorithms. In addition, the experimental results under different acquisition functions are

15

Page 16: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

(a) 100d Colville (b) 100d Colville (enlarged)

(c) 1000d Colville (d) 1000d Colville (enlarged)

Figure 5: Performance on Colville under dimension 100(a)(b) and dimension 1000(c)(d)with embedding dimension r = 4. To see performance discrepancy clearly among differentalgorithms, we enlarged (a)(c) to get (b)(d) respectively. We plotted mean and one standarddeviation across 100 independent runs.

similar. ADD-BO performs excellently in Hartmann-6 due to the additive structure of theobjective function itself, but it performs poorly in the 1000-dimension case (Figure 4b)and in functions without the additive structure. HeSBO performs well in Colville andHartmann-6, but poorly in Branin. Moreover, we can find that HeSBO does not convergein most cases. The performance of SIRBO is similar to our proposed SILBO-TD since it alsotries to learn an embedding space actively. In summary, the proposed method SILBO-BUnearly beats all baselines due to effective semi-supervised dimension reduction and iterativeembedding learning.

5.3 Hyperparameter Optimization on Neural Network

Following (Oh et al., 2018), we also evaluated SILBO in a hyperparameter optimizationtask for a neural network on the MNIST dataset. Specifically, the neural network is amulti-Layer perceptron (MLP) with one hidden layer of size 50 and one output layer of size10. These 500 initial weights were viewed as hyperparameters and were optimized using theAdam (Kingma and Ba, 2015) optimizer.

16

Page 17: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

Figure 6: Performances on the MLP with embedding dimension r= 12. We plotted meanand one standard deviation across 5 independent runs.

Figure 6 shows that SILBO-BU converges faster than all other algorithms. The perfor-mance of SILBO-TD and SIRBO is close. These two methods perform better than othertraditional methods such as REMBO and show competitive performance. Similar to the per-formance on synthetic functions, SILBO-BU beats all other high-dimensional BO methodsand can obtain better response values in the neural network hyperparameter optimizationtasks. Next, we evaluated the effectiveness of semi-supervised dimension reduction anditerative embedding learning respectively.

5.4 Effectiveness of Semi-supervised Dimension Reduction

For the function with effective dimensions, the low-dimensional subspace can reflect its in-trinsic structure information. Thus, if we sample points from the function and project theminto the effective subspace, the response values of these points will be dispersed along thedirection of each subspace basis. Here, we compared the embedding learning performanceof SILBO and SIR on Branin and Camel functions. We first evaluated 50 high-dimensional(i.e., 1000 dimensions) points to get their response values and then projected these pointsinto the low-dimensional space generated by SILBO and SIR respectively. The goal of em-bedding learning is to find the e.d.r directions that preserve the information of the objectivefunction as much as possible. Figure 7 and Figure 8 summarize the observed performance.

In Figure 7b, Figure 8b, and Figure 8d, the information extracted by SIR stacks together.A large number of points are distributed in a very small interval, but correspond to manydifferent response values with large variance. The consequence is that a lot of informationwill be lost if we train a Gaussian process regression model using these low-dimensionalrepresentations. In contrast, due to the semi-supervised dimension reduction, the informa-tion extracted by SILBO is more complete without losing the intrinsic information of theobjective function.

To further demonstrate the effectiveness of semi-supervised dimension reduction, wealso evaluated SILBO only using the labeled points in each iteration. We call this methodSILBO-only-labeled. Figure 9 shows the comparison results between SILBO-semi andSILBO-only-labeled with the bottom-up mapping strategy. From Figure 9, we can see

17

Page 18: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

(a) (b)

Figure 7: Point distribution in the low-dimensional space generated by SILBO and SIRrespectively on Camel. The number of e.d.r directions is 1.

(a) (b)

(c) (d)

Figure 8: Point distribution in the low-dimensional space generated by SILBO and SIRrespectively on Branin. The number of e.d.r directions is 2. (a)(b) plot the points projectedto one estimated direction. (c)(d) plot the points projected to the other direction.

that the semi-supervised dimension reduction technique can significantly improve the per-formance of high-dimensional Bayesian optimization.

5.5 Effectiveness of Iterative Learning

In the iterative process of SILBO, the projection matrix is updated based on random initialpoints and the points acquired from the acquisition function. To verify the effectiveness ofiterative learning, we need to show that the points acquired from the acquisition functionindeed contribute to generating a better embedding. Figure 10 shows the performancecomparison results with and without the iterative learning process. SILBO-only-random

18

Page 19: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

(a) 1000d Branin (b) 1000d Hartmann6 (c) 1000d Colville

Figure 9: Performance comparison between SILBO-semi and SILBO-only-labeled on threesynthetic functions. All of the methods update the projection matrix every 20 iteration.

means that the low-dimensional embedding is learned only using the random initial points,while SILBO-random+acquisition contains the random initial points and the points fromthe acquisition function.

We chose 20 random initial points for SILBO-only-random and chose 10 random initialpoints accompanied with 10 points acquired from the acquisition function iteratively forSILBO-random+acquisition. We compared the optimization performance of the two algo-rithms after these 20 points. From Figure 10, we can see that SILBO-random+acquisitioncan find a better solution at most cases, which indicates the effectiveness of iterative learn-ing. The main reason why SILBO-random+acquisition performs slightly worse on Braninis due to less initial points. Moreover, the performance of SILBO-random+acquisition inFigure 10 is worse than other experiments for two reasons. One is that the initial points aremuch less, we only use 20 initial points. The other is that the low-dimensional embeddingspace is only updated once.

Furthermore, we compare the optimization performance of SILBO with bottom-up strat-egy under different embedding updating frequency. Figure 11 shows that the optimizationperformance is improved when using low updating frequency.

(a) 1000d Branin (b) 1000d Hartmann6 (c) 1000d Colville

Figure 10: The optimization performance of SILBO-only-random and SILBO-random+acquisition using 20 points to find the low-dimensional embedding space. Thetotal number of points to construct the projection matrix is the same.

19

Page 20: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

(a) 1000d Branin (b) 1000d Hartmann6 (c) 1000d Colville

Figure 11: The optimization performance of SILBO under different iterative frequency. Thenumber represents the frequency of updating the projection matrix.

(a) 1000d Branin (b) 500d MLP

Figure 12: Comparison of the cumulative time on Branin with dimension 1000 and on neuralnetwork (MLP) with dimension 500.

5.6 Scalability Analysis

We analyzed the scalability by comparing the cumulative time of each algorithm under thesame number of iterations. As shown in Figure 12, for low-cost objective functions such asBranin, SILBO-BU is fast, while SILBO-TD is relatively slow. This is because SILBO-TDtakes a lot of time to solve the Equation 19. For expensive objective functions such asneural network, SILBO-TD takes approximately the same time as most algorithms, whileSILBO-BU takes more time due to its re-evaluation procedure.

6. Conclusion and Future Work

High-dimensional Bayesian optimization is a very challenging task. To address the problem, we proposed a novel iterative embedding learning framework SILBO for high-dimensionalBayesian optimization through semi-supervised dimensional reduction. We also proposeda randomized fast algorithm for solving the embedding matrix efficiently. Moreover, ac-cording to the cost of the objective function, two different mapping strategies are proposed.Experimental results on both synthetic function and hyperparameter optimization tasksreveal that SILBO outperforms the existing state-of-the-art high-dimensional Bayesian op-timization methods.

20

Page 21: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

In the future, we plan to address the problem of how to determine the true dimension-ality during the dimension reduction process. Moreover, we also plan to use the unlabeledinformation to determine the kernel hyperparameters of Gaussian process and other hy-perparameters of Bayesian optimization. Also, we further plan to apply SILBO to moreAutoML (Automatic Machine Learning) tasks.

Acknowledgments

This work was supported by the National Natural Science Foundation of China(U1811461, 61702254), National Key R&D Program of China (2019YFC1711000), JiangsuProvince Science and Technology Program (BE2017155), National Natural Science Foun-dation of Jiangsu Province (BK20170651), and Collaborative Innovation Center of NovelSoftware Technology and Industrialization.

21

Page 22: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

References

Dimitris Achlioptas. Database-friendly random projections: Johnson-lindenstrauss withbinary coins. Journal of Computer and System Sciences, 66(4):671 – 687, 2003.

Nir. Ailon and Bernard. Chazelle. The fast johnson–lindenstrauss transform and approxi-mate nearest neighbors. SIAM Journal on Computing, 39(1):302–322, 2009.

Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps and spectral techniques for em-bedding and clustering. In Proceedings of the 14th International Conference on NeuralInformation Processing Systems: Natural and Synthetic, NIPS’01, page 585–591, Cam-bridge, MA, USA, 2001.

James Bergstra, Remi Bardenet, Yoshua Bengio, and Balazs Kegl. Algorithms for hyper-parameter optimization. In Proceedings of the 24th International Conference on NeuralInformation Processing Systems, NIPS’11, page 2546–2554, Red Hook, NY, USA, 2011.

Mickael Binois, David Ginsbourger, and Olivier Roustant. A warped kernel improvingrobustness in bayesian optimization via random embeddings. In Learning and IntelligentOptimization, pages 281–286, Cham, 2015.

Mickael Binois, David Ginsbourger, and Olivier Roustant. On the choice of the low-dimensional domain for global optimization via random embeddings. Journal of GlobalOptimization, 2019.

D. Cai, X. He, and J. Han. Semi-supervised discriminant analysis. In 2007 IEEE 11thInternational Conference on Computer Vision, pages 1–7, Oct 2007.

F. R. K. Chung. Spectral Graph Theory. American Mathematical Society, 1997.

Josip Djolonga, Andreas Krause, and Volkan Cevher. High-dimensional gaussian processbandits. In Advances in Neural Information Processing Systems 26, pages 1025–1033.2013.

Stephan Eissman, Daniel Levy, Rui Shu, Stefan Bartzsch, and Stefano Ermon. Bayesianoptimization and attribute adjustment. In Proc. 34th Conference on Uncertainty inArtificial Intelligence, 2018.

Peter I. Frazier. A Tutorial on Bayesian Optimization. art. arXiv:1807.02811, 2018.

Jacob Gardner, Chuan Guo, Kilian Weinberger, Roman Garnett, and Roger Grosse. Dis-covering and Exploiting Additive Structure for Bayesian Optimization. In Proceedings ofthe 20th International Conference on Artificial Intelligence and Statistics, volume 54 ofProceedings of Machine Learning Research, pages 1311–1319, Fort Lauderdale, FL, USA,20–22 Apr 2017.

Stoyan Georgiev and Sayan Mukherjee. Randomized Dimension Reduction on MassiveData. art. arXiv:1211.1642, 2012.

22

Page 23: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

Rafael Gomez-Bombarelli, Jennifer N. Wei, David Duvenaud, Jose Miguel Hernandez-Lobato, Benjamın Sanchez-Lengeling, Dennis Sheberla, Jorge Aguilera-Iparraguirre, Tim-othy D. Hirzel, Ryan P. Adams, and Alan Aspuru-Guzik. Automatic chemical designusing a data-driven continuous representation of molecules. ACS Central Science, 4(2):268–276, 2018.

N. Halko, P. G. Martinsson, and J. A. Tropp. Finding structure with randomness: Prob-abilistic algorithms for constructing approximate matrix decompositions. SIAM Review,53(2):217–288, 2011.

Nikolaus Hansen. The CMA Evolution Strategy: A Tutorial. arXiv e-prints, art.arXiv:1604.00772, April 2016.

Daniel Hernandez-Lobato, Jose Miguel Hernandez-Lobato, Amar Shah, and Ryan P. Adams.Predictive entropy search for multi-objective bayesian optimization. In Proceedings ofthe 33rd International Conference on International Conference on Machine Learning -Volume 48, ICML’16, page 1492–1501. JMLR.org, 2016.

Chiao-Ching Huang and Kuan-Ying Su. Semi-supervised dimension reduction with kernelsliced inverse regression. In Shin-Ming Cheng and Min-Yuh Day, editors, Technologies andApplications of Artificial Intelligence, pages 178–187, Cham, 2014. Springer InternationalPublishing.

Frank Hutter, Holger H. Hoos, and Kevin Leyton-Brown. Sequential model-based optimiza-tion for general algorithm configuration. In Learning and Intelligent Optimization, pages507–523, Berlin, Heidelberg, 2011.

Kirthevasan Kandasamy, Jeff Schneider, and Barnabas Poczos. High dimensional bayesianoptimisation and bandits via additive models. In Proceedings of the 32nd InternationalConference on International Conference on Machine Learning - Volume 37, ICML’15,page 295–304, 2015.

Kirthevasan Kandasamy, Gautam Dasarathy, Jeff Schneider, and Barnabas Poczos. Multi-fidelity bayesian optimisation with continuous approximations. In Proceedings of the 34thInternational Conference on Machine Learning - Volume 70, ICML’17, page 1799–1808,2017.

Kirthevasan Kandasamy, Willie Neiswanger, Jeff Schneider, Barnabas Poczos, and Eric P.Xing. Neural architecture search with bayesian optimisation and optimal transport. InProceedings of the 32nd International Conference on Neural Information Processing Sys-tems, NIPS’18, page 2020–2029, Red Hook, NY, USA, 2018.

Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA,USA, May 7-9, 2015, Conference Track Proceedings, 2015.

Aaron Klein, Stefan Falkner, Simon Bartels, Philipp Hennig, and Frank Hutter. FastBayesian Optimization of Machine Learning Hyperparameters on Large Datasets. InAarti Singh and Jerry Zhu, editors, Proceedings of the 20th International Conference

23

Page 24: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Chen, Zhu, Yuan, and Huang

on Artificial Intelligence and Statistics, volume 54 of Proceedings of Machine LearningResearch, pages 528–536, Fort Lauderdale, FL, USA, 20–22 Apr 2017. PMLR.

Vu Tuan Hieu Le, Cristina Stoica, T. Alamo, E.F. Camacho, and Didier Dumur. Zonotopes:From Guaranteed State Estimation to Control. Wiley-ISTE, October 2013.

Ker-Chau Li. Sliced inverse regression for dimension reduction. Journal of the AmericanStatistical Association, 86(414):316–327, 1991.

Xiaoyu Lu, Javier Gonzalez, Zhenwen Dai, and Neil Lawrence. Structured variationallyauto-encoded optimization. In Proceedings of the 35th International Conference on Ma-chine Learning, volume 80 of Proceedings of Machine Learning Research, pages 3267–3275,Stockholmsmassan, Stockholm Sweden, 10–15 Jul 2018.

Riccardo Moriconi, Marc P. Deisenroth, and K. S. Sesh Kumar. High-dimensional Bayesianoptimization using low-dimensional feature spaces. arXiv e-prints, art. arXiv:1902.10675,February 2019.

Amin Nayebi, Alexander Munteanu, and Matthias Poloczek. A framework for Bayesianoptimization in embedded subspaces. In Kamalika Chaudhuri and Ruslan Salakhutdinov,editors, Proceedings of the 36th International Conference on Machine Learning, volume 97of Proceedings of Machine Learning Research, pages 4752–4761, Long Beach, California,USA, 09–15 Jun 2019.

ChangYong Oh, Efstratios Gavves, and Max Welling. Bock : Bayesian optimization withcylindrical kernels. ICML, pages 3865–3874, 2018.

B. Shahriari, K. Swersky, Z. Wang, R. P. Adams, and N. de Freitas. Taking the human outof the loop: A review of bayesian optimization. Proceedings of the IEEE, 104(1):148–175,2016.

Jasper Snoek, Hugo Larochelle, and Ryan P. Adams. Practical bayesian optimization ofmachine learning algorithms. In Proceedings of the 25th International Conference onNeural Information Processing Systems - Volume 2, NIPS’12, page 2951–2959, Red Hook,NY, USA, 2012.

Niranjan Srinivas, Andreas Krause, Sham Kakade, and Matthias Seeger. Gaussian processoptimization in the bandit setting: No regret and experimental design. In Proceedingsof the 27th International Conference on International Conference on Machine Learning,ICML’10, page 1015–1022, Madison, WI, USA, 2010.

Kevin Swersky, Jasper Snoek, and Ryan P. Adams. Multi-task bayesian optimization.In Proceedings of the 26th International Conference on Neural Information ProcessingSystems - Volume 2, NIPS’13, page 2004–2012, Red Hook, NY, USA, 2013.

Ziyu Wang, Masrour Zoghi, Frank Hutter, David Matheson, and Nando De Freitas.Bayesian optimization in high dimensions via random embeddings. In Proceedings ofthe Twenty-Third International Joint Conference on Artificial Intelligence, IJCAI ’13,page 1778–1784, 2013.

24

Page 25: arXiv:2005.14601v1 [cs.LG] 29 May 2020 · 2020. 6. 1. · HeSBO (Nayebi et al., 2019) employs the count sketch method to alleviate the over-exploration of the boundary and use the

Semi-supervised Embedding Learning for High-dimensional Bayesian Optimization

Jian Wu, Matthias Poloczek, Andrew Gordon Wilson, and Peter I. Frazier. Bayesian opti-mization with gradients. In Proceedings of the 31st International Conference on NeuralInformation Processing Systems, NIPS’17, page 5273–5284, Red Hook, NY, USA, 2017.

Jian Wu, Saul Toscano-Palmerin, I. Peter Frazier, and Gordon Andrew Wilson. Practicalmulti-fidelity bayesian optimization for hyperparameter tuning. UAI, 2019.

Qiang Wu, Feng Liang, and Sayan Mukherjee. Localized sliced inverse regression. Journalof Computational and Graphical Statistics, 19(4):843–860, 2010.

Miao Zhang, Huiqi Li, and Steven Su. High dimensional bayesian optimization via su-pervised dimension reduction. In Proceedings of the Twenty-Eighth International JointConference on Artificial Intelligence, IJCAI-19, pages 4292–4298, 7 2019.

25