104
Meta-Model Assisted (Evolutionary) Optimization Tutorial at PPSN 2016 - 18.09.2016 Boris Naujoks, Jörg Stork, Martin Zaefferer, Thomas Bartz-Beielstein This project has received funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement No 692286.

Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

  • Upload
    lamcong

  • View
    219

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Meta-Model Assisted (Evolutionary) OptimizationTutorial at PPSN 2016 - 18.09.2016

Boris Naujoks, Jörg Stork, Martin Zaefferer, Thomas Bartz-Beielstein

This project has received funding from the European Union’s Horizon 2020research and innovation programme under grant agreement No 692286.

Page 2: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Intro and Motivation

• Synonyms• Metamodels• Surrogates• Response surface models• Approximation models• Simulation models• Data-driven models• Emulators

• From Latin surrogatus– a replacement for something, a substitute or alternativePerfect passive participle of surrogare

• Variant of subrogare, from• Sub (under) + rogare (ask)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 2 / 74

Page 3: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Most common applications• Engineering design• Long, expensive fitness function evaluations

• Finite elements models• Computational fluid dynamics models

• Examples• Airfoil design• Ship propulsion systems• etc.

Additional variable geometry parameters for the linear jetin contrast to the more simple propeller blade optimizationare:

• the hub’s diameter and length• the nozzle’s length and profile angle.

This results in an optimization problem featuring 14 deci-sion parameters in contrast to the nine decision parametersof the pure propeller blade optimization problem describedabove. The stator is not included into the optimization yet.

Again, like in the task presented before, a geometry todeliver more or less thrust compared to the desired valueis “punished” with a higher target value. By a similar”punishment” of a geometry generating cavitation and a“reward” for efficiency we calculate a single target valuewhich is returned to the optimization program after everyhydrodynamic simulation.

Fig. 2. Visualization of the linear jet propulsion systems, within the two-dimensional view, the three different parts blades, hub and nozzle can beidentified.

III. META-MODEL-ASSISTED EVOLUTIONARY

OPTIMIZATION

The idea to assist direct search algorithms by meta-modelshas first been explored by Torczon et al. [8], [9] for patternsearch algorithms. A similar approach can be employedin evolutionary algorithms (EA) by incorporating a pre-screening procedure before the offspring population is eval-uated with the time consuming evaluation tool. Algorithm1 gives an outline of MAEA which is, in fact, a modifiedversion of the basic (µ+λ)-EA described by Back, Hammel,and Schwefel [10] or Beyer and Schwefel [11]. Two featuresdistinguish MAEA from standard EA.

1) All exactly evaluated individuals are recorded andstored in a database. Up to 15 nearest neighbors areconsidered to set up the meta-model for each of the λindividuals per generation.

Fig. 3. Visualization of a more complex propulsion systems featuring rotor,hub, and nozzle. Within the three-dimensional figure of the geometry, thenozzle had to be hidden except for the corresponding grid to see the othercomponents. Nevertheless, this view enables to see the composition of huband blades in more detail.

2) During the pre-screening phase, the objective functionvalues for new solutions are predicted by the meta-model, before deciding whether they need to be re-evaluated by the exact and costly tool.

Thereby, at generation t, the set of offspring solutions Gt

is reduced to the subset of offspring solutions Qt, which willbe evaluated exactly and will also be considered in the finalselection procedure (cf. [1]).

Algorithm 1 (µ + ν < λ)-MAEAt← 0Pt ← init() /* Pt: Set of solutions */evaluate Pt preciselyinitialize database Dwhile t < tmax do

Gt ← generate(Pt) /* λ new offsprings */evaluate Gt with meta-modelQt ← select(Gt) /* |Qt| = ν */evaluate Qt preciselyupdate databasePt+1 ← select(Qt ∪ Pt) /* Select µ best */t← t + 1

end while

A. Pre-screening procedures

A ranking algorithm, applied over the offspring populationGt, identifies the most promising individuals in the newgeneration. In the general case, this algorithm is based on thevalues y(x) (predictions for f(x)) and s(x) (correspondingstandard deviations) obtained for each individual x ∈ Gt

through the meta-model. Comparisons with objective func-tion values for the parent population Pt are necessary. Vari-ous criteria for identifying promising solutions are discussedby Emmerich et al. [1]. Once the promising subset Qt of Gt

has been found, its members undergo exact evaluations.

Image takenfrom [Naujoks et al., 2007]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 3 / 74

Page 4: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Most common applications• Engineering design• Long, expensive fitness function evaluations

• Finite elements models• Computational fluid dynamics models

• Examples• Airfoil design• Ship propulsion systems• etc.

Additional variable geometry parameters for the linear jetin contrast to the more simple propeller blade optimizationare:

• the hub’s diameter and length• the nozzle’s length and profile angle.

This results in an optimization problem featuring 14 deci-sion parameters in contrast to the nine decision parametersof the pure propeller blade optimization problem describedabove. The stator is not included into the optimization yet.

Again, like in the task presented before, a geometry todeliver more or less thrust compared to the desired valueis “punished” with a higher target value. By a similar”punishment” of a geometry generating cavitation and a“reward” for efficiency we calculate a single target valuewhich is returned to the optimization program after everyhydrodynamic simulation.

Fig. 2. Visualization of the linear jet propulsion systems, within the two-dimensional view, the three different parts blades, hub and nozzle can beidentified.

III. META-MODEL-ASSISTED EVOLUTIONARY

OPTIMIZATION

The idea to assist direct search algorithms by meta-modelshas first been explored by Torczon et al. [8], [9] for patternsearch algorithms. A similar approach can be employedin evolutionary algorithms (EA) by incorporating a pre-screening procedure before the offspring population is eval-uated with the time consuming evaluation tool. Algorithm1 gives an outline of MAEA which is, in fact, a modifiedversion of the basic (µ+λ)-EA described by Back, Hammel,and Schwefel [10] or Beyer and Schwefel [11]. Two featuresdistinguish MAEA from standard EA.

1) All exactly evaluated individuals are recorded andstored in a database. Up to 15 nearest neighbors areconsidered to set up the meta-model for each of the λindividuals per generation.

Fig. 3. Visualization of a more complex propulsion systems featuring rotor,hub, and nozzle. Within the three-dimensional figure of the geometry, thenozzle had to be hidden except for the corresponding grid to see the othercomponents. Nevertheless, this view enables to see the composition of huband blades in more detail.

2) During the pre-screening phase, the objective functionvalues for new solutions are predicted by the meta-model, before deciding whether they need to be re-evaluated by the exact and costly tool.

Thereby, at generation t, the set of offspring solutions Gt

is reduced to the subset of offspring solutions Qt, which willbe evaluated exactly and will also be considered in the finalselection procedure (cf. [1]).

Algorithm 1 (µ + ν < λ)-MAEAt← 0Pt ← init() /* Pt: Set of solutions */evaluate Pt preciselyinitialize database Dwhile t < tmax do

Gt ← generate(Pt) /* λ new offsprings */evaluate Gt with meta-modelQt ← select(Gt) /* |Qt| = ν */evaluate Qt preciselyupdate databasePt+1 ← select(Qt ∪ Pt) /* Select µ best */t← t + 1

end while

A. Pre-screening procedures

A ranking algorithm, applied over the offspring populationGt, identifies the most promising individuals in the newgeneration. In the general case, this algorithm is based on thevalues y(x) (predictions for f(x)) and s(x) (correspondingstandard deviations) obtained for each individual x ∈ Gt

through the meta-model. Comparisons with objective func-tion values for the parent population Pt are necessary. Vari-ous criteria for identifying promising solutions are discussedby Emmerich et al. [1]. Once the promising subset Qt of Gt

has been found, its members undergo exact evaluations.

Image takenfrom [Naujoks et al., 2007]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 3 / 74

Page 5: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Most common applications• Engineering design• Long, expensive fitness function evaluations

• Finite elements models• Computational fluid dynamics models

• Examples• Airfoil design• Ship propulsion systems• etc.

Additional variable geometry parameters for the linear jetin contrast to the more simple propeller blade optimizationare:

• the hub’s diameter and length• the nozzle’s length and profile angle.

This results in an optimization problem featuring 14 deci-sion parameters in contrast to the nine decision parametersof the pure propeller blade optimization problem describedabove. The stator is not included into the optimization yet.

Again, like in the task presented before, a geometry todeliver more or less thrust compared to the desired valueis “punished” with a higher target value. By a similar”punishment” of a geometry generating cavitation and a“reward” for efficiency we calculate a single target valuewhich is returned to the optimization program after everyhydrodynamic simulation.

Fig. 2. Visualization of the linear jet propulsion systems, within the two-dimensional view, the three different parts blades, hub and nozzle can beidentified.

III. META-MODEL-ASSISTED EVOLUTIONARY

OPTIMIZATION

The idea to assist direct search algorithms by meta-modelshas first been explored by Torczon et al. [8], [9] for patternsearch algorithms. A similar approach can be employedin evolutionary algorithms (EA) by incorporating a pre-screening procedure before the offspring population is eval-uated with the time consuming evaluation tool. Algorithm1 gives an outline of MAEA which is, in fact, a modifiedversion of the basic (µ+λ)-EA described by Back, Hammel,and Schwefel [10] or Beyer and Schwefel [11]. Two featuresdistinguish MAEA from standard EA.

1) All exactly evaluated individuals are recorded andstored in a database. Up to 15 nearest neighbors areconsidered to set up the meta-model for each of the λindividuals per generation.

Fig. 3. Visualization of a more complex propulsion systems featuring rotor,hub, and nozzle. Within the three-dimensional figure of the geometry, thenozzle had to be hidden except for the corresponding grid to see the othercomponents. Nevertheless, this view enables to see the composition of huband blades in more detail.

2) During the pre-screening phase, the objective functionvalues for new solutions are predicted by the meta-model, before deciding whether they need to be re-evaluated by the exact and costly tool.

Thereby, at generation t, the set of offspring solutions Gt

is reduced to the subset of offspring solutions Qt, which willbe evaluated exactly and will also be considered in the finalselection procedure (cf. [1]).

Algorithm 1 (µ + ν < λ)-MAEAt← 0Pt ← init() /* Pt: Set of solutions */evaluate Pt preciselyinitialize database Dwhile t < tmax do

Gt ← generate(Pt) /* λ new offsprings */evaluate Gt with meta-modelQt ← select(Gt) /* |Qt| = ν */evaluate Qt preciselyupdate databasePt+1 ← select(Qt ∪ Pt) /* Select µ best */t← t + 1

end while

A. Pre-screening procedures

A ranking algorithm, applied over the offspring populationGt, identifies the most promising individuals in the newgeneration. In the general case, this algorithm is based on thevalues y(x) (predictions for f(x)) and s(x) (correspondingstandard deviations) obtained for each individual x ∈ Gt

through the meta-model. Comparisons with objective func-tion values for the parent population Pt are necessary. Vari-ous criteria for identifying promising solutions are discussedby Emmerich et al. [1]. Once the promising subset Qt of Gt

has been found, its members undergo exact evaluations.

Image takenfrom [Naujoks et al., 2007]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 3 / 74

Page 6: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Other areas

. . . where surrogates are applied, involved, used . . .

• No explicit fitness function available• Fitness depending on external factors, e.g. human interactions• Music and arts

• Uncertain environments• Noisy environments• Robustness wrt. design variables• Dynamic fitness landscapes

• Smoothing multi-modal fitness landscapes

Naujoks, Stork, Zaefferer, Bartz-Beielstein 4 / 74

Page 7: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 5 / 74

Page 8: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 6 / 74

Page 9: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Surrogate Modeling - Concepts and Methods

Questions to Answer:

1 What is the core concept of surrogate modeling?2 How does a typical surrogate optimization cycle work?3 Which models are common for surrogate optimization?4 Example method: Efficient Global Optimization

Naujoks, Stork, Zaefferer, Bartz-Beielstein 7 / 74

Page 10: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Costly real world (blackbox) problems• Real-world applications: commonly blackbox problems (machines, complexprocesses)

• Available information is very sparse, properties of the objective function aredifficult or impossible to determine,

• No a priori information about modality, convexity, gradients, or the minimalfunction value f(x∗) is known

• Most complex problems arise if physical experiments are involved. Aside frombeing costly in terms of needed resources (manpower, material, time)

• Wrong values can lead to hazardous effects, e.g., damaging or destroyingexperimental material.

• Instead of physical experiments, simulations are used, e.g., from the field ofComputational Fluid Dynamics (CFD).

• Require a lot of computational power and are very time demanding

Inevitable need to evaluate candidate solutions in the search space to retrieve anyinformation and high demand on resources for each of these evaluations.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 8 / 74

Page 11: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Surrogate Modeling - Application Layers

L1 The Real-World Application• Direct optimization is very costly or impossible as incorrectly chosen decisionvariable values

• Evaluations involve resource demanding prototype building or even hazardousexperiments

L2 The Simulation Model• Complex computational model from fluid- or structural dynamics• Single simulation process may take minutes, hours, or even weeks to compute• Available computational power limits the amount of available evaluations

L3 The Surrogate Model• Data-driven regression model• The accuracy heavily depends on the underlying surrogate type and number ofavailable information

• Typically cheapL4 The Optimization Process

• Any suitable optimization algorithm (deterministic, stochastic, metaheuristic...)• Can be tuned

Naujoks, Stork, Zaefferer, Bartz-Beielstein 9 / 74

Page 12: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Surrogate Modeling - Core Concept

tuning procedure

(f4f4) optimization algorithm

(f3f3) surrogate model

(f1f1) real world application, physical model

algorithm control parameters

optimized algorithm control parameters

decisionvariables

optimized variables

process parameters, estimated output

simulated input

input output

candidate solutions predicted fitness

(f2f2) simulation model

Naujoks, Stork, Zaefferer, Bartz-Beielstein 10 / 74

Page 13: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Surrogate Modeling - Costs and Benefits

• Each layer L1 to L4 imposes different evaluation costs and solutionaccuracies:

• Most expensive: L1 real world• Commonly cheapest: L2 Surrogate Model• Modeling process itself requires computational resources for evaluations,construction or validation of the surrogate.

The main benefit of using surrogates is the reduction of needed fitness evaluationson the objective function during the optimization.

• An other advantage is the availability of a surrogate itself, which can beutilized to gain further problem insight.-> This is particularly valuable for blackbox problems.

• The initial sampling design plan has a major impact on the optimizationperformance and should be carefully selected.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 11 / 74

Page 14: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Surrogate Modeling - Optimization Cycle

A common optimization process using surrogates is outlined by the following steps:

1 Sampling the objective function to generate a set of evaluated points2 Selecting a suitable surrogate3 Constructing the surrogate using the evaluated points4 Utilizing the surrogate to predict new promising locations5 Evaluating the objective function on one (or more) of the identified locations6 Updating the surrogate and repeating the optimization cycle

Naujoks, Stork, Zaefferer, Bartz-Beielstein 12 / 74

Page 15: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Surrogate Modeling - Important Publications

Important publications featuring overviews or surveys on surrogate modeling andsurrogate optimization:

• Design and analysis of computer experiments, [Sacks et al., 1989]• A taxonomy of global optimization methods based on response surfaces,[Jones, 2001]

• Surrogate-based analysis and optimization, [Queipo et al., 2005]• Recent advances in surrogate-based optimization, [Forrester and Keane, 2009]• Surrogate-assisted evolutionary computation: Recent advances and futurechallenges, [Jin, 2011]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 13 / 74

Page 16: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Linear Models

• Combination of linear predictor functions of each input to model the output• Basic LM: y = β0 + β1x1 + β2x2 + · · ·+ βnxn + ε, where ε is the error term• Extensions: Interactions between inputs, quadratic terms, response surfacemodels, polynomial regression

• Polynomial model takes the formy = a0 + a1x+ a2x

2 + a3x3 + · · ·+ anx

n + ε

Pro: white box, easy to interpret / analyse, simple and fast

Con: not suitable for complex functions, overfitting (by using too many terms)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 14 / 74

Page 17: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Decision Trees and Random Forests• Decision Trees [Breiman et al., 1984] model the objective function by usingtree-based approximations.

• At each node of the tree a split is made on the basis of an decision variablevalue

• The prediction of a new point is given by the mean value of associated points• Random Forests Regression [Breiman, 2001] a large number of decision treesis combined to an ensemble predictor

• Usually, each tree in the ensemble is fitted using a subset of the evaluatedpoints to avoid overfitting (bagging)

• Predictions of new individuals are then given by a cumulated mean of allpredictors in the ensemble

Pro: easy to interpret white box model (decision trees), fast, binary+integer+realvariables

Con: complex to interpret (RF), bad fit for complex functions (decision tree), nosmooth surface, overfitting (too large tree)Naujoks, Stork, Zaefferer, Bartz-Beielstein 15 / 74

Page 18: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Artificial Neural Networks and Deep Learning• Neural Networks [Haykin, 2004; Hornik et al., 1989] are inspired by thebiological brain

• They utilize so-called connected neurons to learn and approximate thebehavior of a function

• Neurons are weighted transform functions• Several layers of neurons: input, output and hidden layers• Layers consist of neurons with different forward an/or backward connections• Deep learning [Deng and Yu, 2014; Hinton et al., 2006]:• Complex structured networks with multiple processing layers and/or multiplenon-linear transformations and stacked model approaches

• Excellent results in approximation and specially classification tasks• Highly computational complex, lot of resources needed

Pro: very accurate (deep learning), universal approximator

Con: high computational effort, difficult to interpret, very complex (DeepLearning)Naujoks, Stork, Zaefferer, Bartz-Beielstein 16 / 74

Page 19: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Symbolic Regression

• Symbolic Regression [Flasch et al., 2010] is a high level method to fit ahuman-readable mathematical model

• Based on Genetic Programming (GP)• Mathematical expressions building blocks (+,−, sin, cos, exp ...)• Model is evolved using an evolutionary population-based approach

Pro: easy to interpret, fast prediction

Con: high computational complexity (building process)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 17 / 74

Page 20: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Kriging

• Kriging or Gaussian Process Regression [Sacks et al., 1989] is used to modelthe error term of the model instead of the linear coefficients

• Simplest form: β0 + ε, where β0 is the mean• The ε is then expressed by an gaussian stochastic process.• Modeling of the error term ε with help of a covariance distance matrix• The correlation between errors is related to the distance between thecorresponding points

• The covariance matrix is utilized to predict unknown candidates.• Outstanding feature of Kriging models: uncertainty measure for theprediction and Expected Improvement (EI):

Pro: suitable for complex functions, uncertainty measurment and EI

Con: not suitable for high dimensional data, high computational effort

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18 / 74

Page 21: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Expected ImprovementImprovement:

• Current best point for minimization: x∗ with function value f(x∗)• For a new point x’, the improvement in our objective function is

[f(x∗)− f(x′)]+

0

f(x*) - f(x')

Impr

ovem

ent

Naujoks, Stork, Zaefferer, Bartz-Beielstein 19 / 74

Page 22: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Expected Improvement2

found in [7]. Recent publications show, that metamodels arebeneficial to speed up the evolutionary search in constrainedand multi-objective optimization [15], [16], [17], [18], thoughthere are still open questions.

Recently, screening methods also consider the confidence ofthe predicted output have been suggested [3], [9], [13], [19].This information can be obtained through Gaussian RandomField models which predict the unknown evaluation result bymeans of a Gaussian distribution. The use of confidence in-formation increases the prediction accuracy of the metamodeland helps guiding the search towards less explored regions inthe search space. This also prevents premature convergence.

In this paper, the term pre-screening will denote the use ofmetamodels for the selection of promising members which arenot evaluated so far. Criteria that can be used to support pre-screening procedures by incorporating confidence informationare introduced; their concept is discussed and statistical studiesof their performance are presented. These criteria figure outimprovements in a set of new (offspring) solutions, assumingthat the unknown response is described by a Gaussian distri-bution.

In order to extend the application domain of the proposedmethods, pre-screening criteria used in single-objective prob-lems will be generalized to constrained and multi-objectiveproblems. After scrutinizing a number of mathematical opti-mization problems, a challenging aerodynamic design problemwith 3 objectives and 6 constraints will be solved. The resultspresented below indicate that it is very beneficial to considerthe confidence information within a MAEA, in order toimprove its robustness.

The structure of this paper is as follows: In section II,GRFM are presented and discussed. In section III, the single-and multi-objective EA used are presented. In section IV,the integration of GRFM in single- and multi-objective EAis outlined. Finally, using a number of academic test cases(section V) and a real-world test problem (section VI), theefficiency of the proposed MAEA is investigated.

II. GAUSSIAN RANDOM FIELD METAMODELS

The Gaussian Random Field (GRF) theory constitutes apowerful framework for building metamodels based on dataobtained through computer experiments. Gaussian RandomField Models (GRFM) will be defined below, by first puttingemphasis to the information required by the model as wellas its responses after training. Later, statistical assumptions,limitations and practicalities related to the model itself and itsuse will be discussed.

In the literature, GRFM are also known under differentnames, such as Kriging, Gaussian processes and Gaussianrandom functions methods. The term Kriging points directlyto the origin of these prediction methods dating back to thesixties, when the mining engineer Krige used GRFM-likemodels to predict the concentration of ore in gold- and uraniummines [20]. Today, Kriging includes a wide class of spatialprediction methods which do not necessarily assume Gaussianfields.

Note that the latter assumption is essential in our algorithmsand the term GRFM is used herein in the standard (strict)

�����������������������������������������������������������������������������

�����������������������������������������������������������������������������

������������������������������������������������������

������������������������������������������������������

����������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

����������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

�������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

�����������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

����������������������������������������������������������������������������������������������������������������������������������������������������������������

����������������������������������������������������������������������������������������������������������������������������������������������������������������

����������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

����������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������������

x x x(1) (2) (3)

y

y

(1) (3)

(2)^

y

Confidence Range

^

y

y(x’)

y(x’)+s(x’)

x’

^

^

y(x’)−s(x’)^

x

Predicted Function

Fig. 1. Outputs of Gaussian Random Field Metamodels using a R → Rmapping example.

sense. On the other hand, the term Gaussian random functions[21] might be misleading, because a random function is oftenassociated with a single random variable instead of a setof them. However, in the present paper, the term GaussianRandom Field seems to be more appropriate than GaussianProcess [22] since this paper is dealing with a multidimen-sional – spatial – rather than a one-dimensional – temporal –input space [23].

Apart from the predicted objective function value, anotherinformation provided by a GRFM is a measure of confidencefor its prediction. It is reasonable that the confidence isexpected to be higher if the training point density in theneighborhood of a newly proposed point is higher. Anotherimportant output of the metamodel is the variance of the outputvalues and the average correlation between responses at neigh-boring points. A GRFM interpolates data values and estimatestheir prediction accuracy. It provides the mean value and thestandard deviation for a one-dimensional Gaussian distributionwhich represents the likelihood for different realizations ofoutcomes to represent a precise function evaluation. Figure 1illustrates the use of GRFM in an example mapping R→ R.

The user of modern optimization methods desires to operatethe metamodel in the most efficient manner, i.e. to maximizeits prediction capabilities and minimize the CPU cost forits training. For this purpose, a better understanding of thestatistical assumptions, limitations and practicalities related tothe model itself and its use are needed.

Let y : Rd → R be the output of a computationallyexpensive computer experiment and X = {x(1), . . . ,x(m)}be a set of m input configurations which are available alongwith the corresponding responses y(1) = y(x(1)), y(2) =y(x(2)), . . . , y(m) = y(x(m)). No assumption on the regularityof the distribution of x(1), . . . , x(m) in S is made.

In GRF theory, the aim is to build a non-time-consumingtool capable of predicting the output corresponding to a newpoint x′ ∈ S, according to an approximated Rd → R mapping.With x′ ∈ X , the precise objective function value is returned.This corresponds to the well known exact interpolation prob-lem for which a large variety of methods, ranging from splines[24] to radial basis networks [25] and Shepard polynomials[26], are available.

The basic assumption in modeling with GRFM is that theoutput function is a realization (sample-path) of a Gaussian

Image taken from [Emmerich et al., 2006]

Intuition: Expected Improvement is every possible improvement value weighted byits probabilityNaujoks, Stork, Zaefferer, Bartz-Beielstein 20 / 74

Page 23: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Method Example: Efficient Global Optimization

• Efficient Global Optimization (EGO) by [Jones et al., 1998a] is a surrogateoptimization framework specialized on utilizing Kriging and expectedimprovement

• Focus on optimization of expensive blackbox functions• The original version of EGO starts by sampling the objective function by aspace-filling experimental design

• Example: LHD with approximate k = 10n points: convenient, finite-decimalvalue for the inter-point spacing, e.g., 21 design points for 2-dimensions

• Kriging surrogate is fit using maximum likelihood estimation on the selectedpoints

• The surrogate is then manually analyzed by applying different diagnostic tests• If it is satisfactory the iterative optimization process is started, if not, theobjective function is tried to be transformed (by log or inversetransformation) to acquire a better fit.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 21 / 74

Page 24: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Method Example: Efficient Global Optimization

The optimization cycle has following steps:

1 Calculate and maximize expected improvement on surrogate by a exactbranch-and-bound algorithm

2 Sample the objective function where expected improvement is maximized3 Re-estimate the Kriging surrogate including the new candidate by maximumlikelihood estimation

The authors introduce a stopping criterion, which is reached if the expectedimprovement is less than one percent of the current best candidate

Naujoks, Stork, Zaefferer, Bartz-Beielstein 22 / 74

Page 25: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

EGO Pseudo-Code Phase I: Building

Algorithm 1.1: EGO

beginphase 1, initial surrogate building:initialize population X of size k based on a space-filling DOEevaluate X on f(x)xc = best candidate in f(X)fit Kriging surrogate model fm with X by maximum likelihood estimationmanually verify fm by diagnostic testsif verify(fm)=false then

transform f(x) by log or inverse and repeat fitting processend

end

Naujoks, Stork, Zaefferer, Bartz-Beielstein 23 / 74

Page 26: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

EGO Pseudo-Code Phase II: Optimization

Algorithm 1.2: EGO

beginphase 2, use and refine surrogate:while not termination-condition do

xnew = calculate and maximize EI on surrogate model bybranch-and-bound optimizationif EI(xnew)/|f(xc)| < 0.01 then

stop algorithmendevaluate f(xnew)add xnew to Xxc = best candidate in f(X)re-estimate fm with X by maximum likelihood estimation

endend

Naujoks, Stork, Zaefferer, Bartz-Beielstein 24 / 74

Page 27: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 25 / 74

Page 28: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

A practical example

• language: R• installation of R and more: https://cran.r-project.org/• optional / recommended IDE: RStudio https://www.rstudio.com/• R tutorial:

https://cran.r-project.org/doc/manuals/r-release/R-intro.html• Uses a 1-dim benchmark function from Forrester et al. [2008]• See following code

Naujoks, Stork, Zaefferer, Bartz-Beielstein 26 / 74

Page 29: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

## To install the required packages, uncomment the following lines:# install.packages("SPOT")library("SPOT") #load required package: SPOT

## Initialize random number generator seed. Reproducibility.set.seed(1)

## Define objective functionobjectFun <- function(x){

(6*x-2)^2 * sin(12*x-4)}

## Plot the function:curve(objectFun(x),0,1)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 30: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 31: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

## Now, let us assume objectFun is expensive.## First, we start with making some initial## design of experiment, which in this case## is simply a regular grid:x <- seq(from=0, by=0.3,to=1)

## Evaluate with objective:y <- sapply(x,objectFun)

## Add to plot:curve(objectFun(x),0,1)points(x,y)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 32: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 33: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

## Build a model (here: Kriging, with the SPOT package.## But plenty of alternatives available)fit <- forrBuilder(as.matrix(x),as.matrix(y),

control=list(uselambda=FALSE #do not use nugget effect (regularization)))

## Evaluate prediction based on model fitxtest <- seq(from=0, by=0.001,to=1)pred <- predict(fit,as.matrix(xtest),predictAll=T)ypred <- pred$fspred <- pred$s

## Plot the prediction of the model:curve(objectFun(x),0,1)points(x,y)lines(xtest,ypred,lty=2,lwd=2)

## Plot suggested candidate solutionpoints(xtest[which.min(ypred)],ypred[which.min(ypred)],col="black",pch=20,cex=2)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 34: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 35: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

## Calculate expected improvement (EI)ei <- 10^(-spotInfillExpImp(ypred,spred,min(y)))## note: the function used above returns negative## log. of EI, for optimization purposes.

## Plot EIcurve(objectFun(x),0,1)points(x,y)lines(xtest,ypred,lty=2,lwd=2)par(new = T)plot(xtest,ei,lty=3,lwd=2, type="l", axes=F, xlab=NA, ylab=NA,

ylim=rev(range(ei)))axis(side = 4); mtext(side = 4, line = 1.4, 'EI')

## Determine solution that maximizes EInewx <- xtest[which.max(ei)]

## Plot suggested candidate solution, based on EIpoints(newx,max(ei),col="red",pch=20,cex=2)

## Add datax <- c(x,newx)y <- c(y,objectFun(newx))

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 36: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.4

0.3

0.2

0.1

0.0

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 37: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

## Now repeat the same as often as necessary:repeatThis <- expression({

curve(objectFun(x),0,1)points(x,y)fit <- forrBuilder(as.matrix(x),as.matrix(y),

control=list(uselambda=FALSE))

xtest <- seq(from=0, by=0.001,to=1)pred <- predict(fit,as.matrix(xtest),predictAll=T)ypred <- pred$fspred <- pred$slines(xtest,ypred,lty=2,lwd=2)points(xtest[which.min(ypred)],ypred[which.min(ypred)],col="black",pch=20,cex=2)ei <- 10^(-spotInfillExpImp(ypred,spred,min(y)))par(new = T)plot(xtest,ei,lty=3,lwd=2, type="l", axes=F, xlab=NA, ylab=NA,

ylim=rev(range(ei)))axis(side = 4); mtext(side = 4, line = 1.4, 'EI')points(xtest[which.max(ei)],max(ei),col="red",pch=20,cex=2)newx <- xtest[which.max(ei)]x <- c(x,newx)y <- c(y,objectFun(newx))

})

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 38: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.35

0.30

0.25

0.20

0.15

0.10

0.05

0.00

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 39: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.4

0.3

0.2

0.1

0.0

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 40: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.4

0.3

0.2

0.1

0.0

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 41: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.20

0.15

0.10

0.05

0.00

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 42: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.00

80.

006

0.00

40.

002

0.00

0E

I

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 43: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.01

50.

010

0.00

50.

000

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 44: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

3e+

072e

+07

1e+

070e

+00

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 45: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

1e+

098e

+08

6e+

084e

+08

2e+

080e

+00

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 46: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

• EI looks noisy, strange• Predicted mean is completely off target• Why?

• Common practical problem• Numerical issue• Closeness of solutions• Problem for Kriging model

• Near identical rows in correlation matrix• Badly conditioned

• Often, Kriging implementations will crash• Here, nonsensical predictions

• Potential remedy: use nugget (regularization) + reinterpolation[Forrester et al. 2008]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 47: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

## repeat as often as necessary (but now with regularization):repeatThis <- expression({

curve(objectFun(x),0,1)points(x,y)fit <- forrBuilder(as.matrix(x),as.matrix(y),

control=list(uselambda=TRUE, # Use nugget (parameter lambda)reinterpolate=T # Reinterpolation, to fix uncertainty estimates

))xtest <- seq(from=0, by=0.001,to=1)pred <- predict(fit,as.matrix(xtest),predictAll=T)ypred <- pred$fspred <- pred$slines(xtest,ypred,lty=2,lwd=2)points(xtest[which.min(ypred)],ypred[which.min(ypred)],col="black",pch=20,cex=2)ei <- 10^(-spotInfillExpImp(ypred,spred,min(y)))par(new = T)plot(xtest,ei,lty=3,lwd=2, type="l", axes=F, xlab=NA, ylab=NA,

ylim=rev(range(ei)))axis(side = 4); mtext(side = 4, line = 1.4, 'EI')points(xtest[which.max(ei)],max(ei),col="red",pch=20,cex=2)newx <- xtest[which.max(ei)]x <- c(x,newx)y <- c(y,objectFun(newx))

})

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 48: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.00

150.

0010

0.00

050.

0000

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 49: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

eval(repeatThis)

0.0 0.2 0.4 0.6 0.8 1.0

−5

05

1015

x

obje

ctF

un(x

)

0.00

120.

0008

0.00

040.

0000

EI

Naujoks, Stork, Zaefferer, Bartz-Beielstein 18.09.2016

Page 50: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 27 / 74

Page 51: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice

• Previous slides: numerical issues (Kriging)

• Other, more general issues:• Problem definition

• What is the objective• What variables impact the objective• ...

• Algorithm design, selection of:• Model• Optimizer• Parameters

Naujoks, Stork, Zaefferer, Bartz-Beielstein 28 / 74

Page 52: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Problem definition

• Very important, crucial to success• Often underestimated• Information based on

• Discussions with application experts, practitioners• Literature• Experience

Naujoks, Stork, Zaefferer, Bartz-Beielstein 29 / 74

Page 53: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Problem definition• Consider the following:

• Aims and goals• What are they?• Can they be clearly defined?• Can they be evaluated

(measured, computed)?• Cost of evaluation?• Budget?• Desired accuracy?

• Variables affecting theobjective(s)

• How many?• Independent variables?• Disturbance variables?• Data types?

• Constraints?• Noise?• Interfacing, data exchange

• Repeat the aforementioned, e.g.,after first results

Surrogate Model Optimization

CostlyExperiment or

Simulation

ComputeObjective(s)

Measurements & Results

Distu

rban

ce P

arameters

Objectives

Independent Parameters

Ob

jectives

OptimizerModel

Ind

epen

den

t P

arameters

No

ise

Naujoks, Stork, Zaefferer, Bartz-Beielstein 30 / 74

Page 54: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Model Selection

Spline models

Linear regression

Kriging

Support Vector Machines

Neural

Networks

Symbolic Regression

RBFNs

Random Forest

Regression

Trees

• Large variety of models available• Which to choose?

• Potential solutions:• Use the "default" (e.g., Kriging / EGO)• Exploit problem knowledge• Select performance-based or combine ->

Ensembles (open issue)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 31 / 74

Page 55: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Model Selection

• No problem is truly black-box• Use what you know, e.g.:

• Number of parameters• 20 or more: Kriging and related looseperformance

• Data types• Continuous: Kriging, SVMs, RBFNs• Integer, binary, categorical parameters:e.g., Random Forest

• Mixed: Treed Gaussian Processes (TGP)• Structured / combinatorial (e.g.:permutations, trees): see later slides

• Data set sizes (budget)• Large: Kriging may become slow• Small: Take care to use models that avoidoverfitting

Naujoks, Stork, Zaefferer, Bartz-Beielstein 32 / 74

Page 56: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Model Selection

• Structure of the fitness landscape:• Highly multi-modal: do not use simple linear models• Smooth: Kriging or related• Large plateaus or discontinuities: Kriging variants may perform poorly• Known trend: Use Kriging with trend function.

• Cost of the objective function• Rather high: Complex, powerful models (Kriging, SVMs)• Rather low: Less complex, cheaper models (linear regression,tree-based,k-Nearest Neighbor)

• Requirements of understandability / learning from the model• Variable importance: most models• Rule extraction: regression trees• Human readable formulas: linear models, genetic programming(symbolic regression)

• Availability of derivatives• e.g., Gradient Enhanced Kriging ?

$$$

?

Naujoks, Stork, Zaefferer, Bartz-Beielstein 33 / 74

Page 57: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Model Selection

• Other considerations• Customer / Practitioner preferences and knowledge

• Do they understand the models• Do they trust results from the models

• Your own preferences & experience• e.g., with regards to parameterization• or implementation

• Note, various model types often quite similar & related, interchangeable• e.g.: Spline models - Kriging - SVM - RBFN

Naujoks, Stork, Zaefferer, Bartz-Beielstein 34 / 74

Page 58: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Implementation

• Once models are selected -> Implementation• Can have significant impact• Options

• Frequently employed packages/libraries• Quality• Community support• Examples, documentation• Continuity of development

• Less frequently used work• For special tasks?• Because of specific features?

• Do it yourself• None other available (or too slow, buggy)• Specific features not available• More fun , but also more work /• You know what the model really does

Naujoks, Stork, Zaefferer, Bartz-Beielstein 35 / 74

Page 59: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Optimizer Selection

• Similar considerations as for models

• Optimizer also depends on model type (and vice versa)

• Smooth, differentiable models like Kriging: gradient-based optimizers are fine• Non-smooth (tree-based): GA, DE, PSO• Multimodality (of prediction or infill criterion, e.g., EI):Population based, restarts, niching, etc.

• Simple linear regression: analytical

Naujoks, Stork, Zaefferer, Bartz-Beielstein 36 / 74

Page 60: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Typical Problems in Practice: Parameter Selection

• Similar to model selection / optimizer selection ...• ... but with more attention to details

• Use expert / literature suggestions• Exploit problem knowledge

• Parameters affect:• complexity,• cost of modeling,• cost of model optimization,• noise handling,• robustness,• smoothness,• ...

• Tuning, benchmarks (open issue)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 37 / 74

Page 61: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 38 / 74

Page 62: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Research perspectives• Fields of Interest

• Multi-objective:SAMCO - Surrogate Assisted Multi-Criteria Optimisation

• Combinatorial surrogates models (optimisation)

• ... both handled in more detail later!

Naujoks, Stork, Zaefferer, Bartz-Beielstein 39 / 74

Page 63: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Research perspectives• Fields of Interest

• Multi-objective:SAMCO - Surrogate Assisted Multi-Criteria Optimisation

• Combinatorial surrogates models (optimisation)• ... both handled in more detail later!

Naujoks, Stork, Zaefferer, Bartz-Beielstein 39 / 74

Page 64: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Meaningful benchmarking and testing of algorithms• Noise handling• Complex resource limitations• High-dimensional / large scale data• Constraint handling• Aggregation: Model ensembles, Multi-fidelity models• Dynamic optimization problems

Naujoks, Stork, Zaefferer, Bartz-Beielstein 40 / 74

Page 65: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Meaningful benchmarking and testing of algorithms• Some benchmark sets available• (Almost?) not considered for evaluation• No standard implemented• Depending on people who apply?

• Noise handling• Surrogates considered for noisy problems• What about noise in models?

• Complex resource limitations• Resources like computation times may not be available constantly• Server availability, different calculation times per job . . .• Problem handled separately• Integration of resources handling in algorithm needed

Naujoks, Stork, Zaefferer, Bartz-Beielstein 41 / 74

Page 66: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Meaningful benchmarking and testing of algorithms• Some benchmark sets available• (Almost?) not considered for evaluation• No standard implemented• Depending on people who apply?

• Noise handling• Surrogates considered for noisy problems• What about noise in models?

• Complex resource limitations• Resources like computation times may not be available constantly• Server availability, different calculation times per job . . .• Problem handled separately• Integration of resources handling in algorithm needed

Naujoks, Stork, Zaefferer, Bartz-Beielstein 41 / 74

Page 67: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Meaningful benchmarking and testing of algorithms• Some benchmark sets available• (Almost?) not considered for evaluation• No standard implemented• Depending on people who apply?

• Noise handling• Surrogates considered for noisy problems• What about noise in models?

• Complex resource limitations• Resources like computation times may not be available constantly• Server availability, different calculation times per job . . .• Problem handled separately• Integration of resources handling in algorithm needed

Naujoks, Stork, Zaefferer, Bartz-Beielstein 41 / 74

Page 68: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• High-dimensional / large scale data• Models may fail / not be applicable• New models might need to be considered• New integration schemes needed as well?

• Constraint handling• Different scenarios possible• Most common: infeasible offspring of feasible ancestor

• Easy strategy: just omit . . . optimal?• Constraints to be considered by models as well?• Integration in algorithms?• Optimal strategy?

Naujoks, Stork, Zaefferer, Bartz-Beielstein 42 / 74

Page 69: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• High-dimensional / large scale data• Models may fail / not be applicable• New models might need to be considered• New integration schemes needed as well?

• Constraint handling• Different scenarios possible• Most common: infeasible offspring of feasible ancestor

• Easy strategy: just omit

. . . optimal?• Constraints to be considered by models as well?• Integration in algorithms?• Optimal strategy?

Naujoks, Stork, Zaefferer, Bartz-Beielstein 42 / 74

Page 70: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• High-dimensional / large scale data• Models may fail / not be applicable• New models might need to be considered• New integration schemes needed as well?

• Constraint handling• Different scenarios possible• Most common: infeasible offspring of feasible ancestor

• Easy strategy: just omit . . . optimal?• Constraints to be considered by models as well?• Integration in algorithms?• Optimal strategy?

Naujoks, Stork, Zaefferer, Bartz-Beielstein 42 / 74

Page 71: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Aggregation: Model ensembles, Multi-fidelity models• Which model in which situation?

- Again depending on many parameters- Some results available . . .

• How to aggregate ensembles best?• Setting may vary over time . . .

• Dynamic optimization problems• In general: time-varying fitness function• Surrogates used for forecasting, predicting future values• Other settings possible . . . see above

Naujoks, Stork, Zaefferer, Bartz-Beielstein 43 / 74

Page 72: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Open Issues

• Aggregation: Model ensembles, Multi-fidelity models• Which model in which situation?

- Again depending on many parameters- Some results available . . .

• How to aggregate ensembles best?• Setting may vary over time . . .

• Dynamic optimization problems• In general: time-varying fitness function• Surrogates used for forecasting, predicting future values• Other settings possible . . . see above

Naujoks, Stork, Zaefferer, Bartz-Beielstein 43 / 74

Page 73: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 44 / 74

Page 74: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO

Surrogate-Assisted Multi-Criteria Optimisation

• Intersection of• Multi-criteria optimisation• Surrogate-assisted optimisation

• MCO – Multi-Criteria Optimisation• EMO – Evolutionary Multi-objective Optimisation• EMOA – Evolutionary Multi-objective Optimisation Algorithm• MOEA – Multi-Objective Evolutionary Optimisation

Naujoks, Stork, Zaefferer, Bartz-Beielstein 45 / 74

Page 75: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Basics of Multi-criteria optimisation

• Multiple objective functions considered• Minimize

f : IRn −→ IRm, f(x) = (f1(x), . . . , fm(x))

Pareto Dominance• Solution x dominates solution y

x <p y :⇔ ∀i : fi(x) ≤ fi(y) (i = 1, . . .m)∃j : fj(x) < fj(y) (j = 1, . . .m)

• Pareto-Set: Set of all non-dominated solutions in the search space

{x | @z : z <p x}

• Pareto-Front: Image of Pareto-set in objective space

Naujoks, Stork, Zaefferer, Bartz-Beielstein 46 / 74

Page 76: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO

• Available budget: 100 to 10 000 evaluations• Different strategies

• Stochastic variation of EAs assisted (e.g. filtering solutions)• Completely replaced (e.g. optimizing figure of merit)

• Many algorithms already developed• However

• Very heterogeneous research fields• Different sciences / faculties involved

- Engineering- Statistics- Computer Science

- Mathematics- Aeronautics- Agriculture

• Thus: different backgrounds, also different languages to be considered

Naujoks, Stork, Zaefferer, Bartz-Beielstein 47 / 74

Page 77: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO

• Application driven• Proposed algorithms tested respective application tasks mainly• Comparison of different approaches hard to accomplish• Lacks existence of accepted benchmarks

• Theoretical aspects almost neglected due to focus on practical applications

• Methodological research areas• Choice of the surrogate model• Respective figure of merit (or infill criterion)

Naujoks, Stork, Zaefferer, Bartz-Beielstein 48 / 74

Page 78: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO• Easiest approach: one model per objective• Kriging and expected improvement used commonly

10

on the old Pareto front

0

5

10

15

20

0

5

10

15

20

0

0.2

f1

f2

Probability Density

Precise Evaluations

Mean values of approximations

Lower confidence bounds

x1

x2

x3

Fig. 8. Interval boxes for approximations in a solution space with twoobjectives.

Pareto−front

PSfrag replacements

Vd(P )

Vnd(P )

f1

f2

f(x1)

f(x2)

f(x3)

f(x4)

fmax

fmin

Fig. 9. Illustration of the hypervolume measure.

V. METHOD APPLICATION – RESULTS AND DISCUSSION

A. Mathematical Test Problems and Performance Measures

At first, experiments have been conducted on selected math-ematical test problems in order to compare the performanceof different MAES variants. These variants were tested ondifferent objective function landscapes, featuring minimizationof a simple convex function (sphere function, appendix A), anon-isotropic function (ellipsoid function, appendix B), a dis-continuous function (step function, appendix C) and - finally -a highly multimodal function (Ackley function, appendix D).

MAES testing was carried out using population sizes ofµ = 5 and λ = 100 individuals, among which ν = 20individuals at most were pre-selected for exact evaluation.Each run was repeated 20 times, each of which with a differentrandom number seed.

The median of the best results found after t evaluations(t ≤ 1000) was plotted. In order to get a reliability measure,the 16th worst function value, i. e. the 80%-quantile of thedistribution of the obtained function values, was recorded andpresented. For similar studies on 20–dimensional test-casesand with different population sizes the reader should refer to[9], [19], [45] and recently [3].

B. Prediction Accuracy Measures

It is well known that EA are rank-based strategies thatare invariant to monotonic transformations of the objectivefunction. Hence, for a metamodel used in conjunction withan EA to be successful, it suffices this to predict the subsetof Gt that would be selected by the recombination if allevaluations were precise improvements with respect to theparent population Pt. The so–called retrieval quality of any

pre-screening tool (metamodel) can be measured through therecall and precision measures defined below.

Let Mµ(A) denote the subset of the µ best solutions inA. Pre-screening aims at identifying the members of Gt ∩Mµ(Gt ∪Pt) which will enter the next generation. Thus, it isdesirable that

Qt ≈ Gt ∩Mµ(Gt ∪ Pt). (47)

It is reasonable that none of the metamodels could alwaysretrieve the ensemble of relevant individuals out of Gt. Anon-satisfactory metamodel is one which: (a) fails capturinga considerable part of Gt ∩Mµ(Gt ∪ Pt) or (b) in order tocapture as many as possible of them, it additionally selects toomany irrelevant individuals.

The retrieval accuracy, practically in relation to the first ofthe two unpleasant situations just mentioned, i. e. the ratiothe relevant solutions retrieved from Gt to the number of allrelevant solutions in Gt, is quantified as follows:

recall(t) =|Mµ(Gt ∪ Pt) ∩Qt||Mµ(Gt ∪ Pt) ∩Gt|

, (48)

where the optimal values is recall(t) = 1.On the other hand, precision(t) is a measure for controlling

the second unpleasant metamodel behavior. This is expressedby the ratio of the number of correctly retrieved solutions tothe total number of retrieved solutions, namely:

precision(t) =|Mµ(Gt ∪ Pt) ∩Qt|

|Qt|(49)

The optimal value for this criterion is precision(t) = 1.Unfortunately, in contrast to quantitative measures such as

y − y plots, specificity measures cannot be evaluated withoutperforming extra evaluations with the costly evaluation tool.Hence, these are useful for statistics on simple academic casesbut not for real–world problems.

C. Implementation details

The basic evolution strategy corresponds to the one de-scribed previously in section IV. The initial step-size wasset to 0.05% of the search space width. The database isformed only by exact evaluations. The metamodel is usedfrom the first generation on. As soon as there are more than2d solutions in the database, the algorithm switches to thelocal metamodeling strategy as described in section II. For allstrategies, the maximal number of pre-selected individuals wasset to µ.

D. Results – Discussion on the performance

The first comparison was conducted on the 20-dimensionalsphere model (cf. appendix A). The median of the best foundsolution is shown in Figure 10. All metamodel-based strategiesoutperformed conventional strategies (i.e. (5+20)-ES, (5+35)-ES4, (5+100)-ES) since they ask considerably less function

4This strategy has been added in order to be comparable with Ulmer et al.[9].

Image taken from [Emmerich et al., 2006]Naujoks, Stork, Zaefferer, Bartz-Beielstein 49 / 74

Page 79: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Alternative approaches

ParEGO – Pareto Efficient Global Optimisation [Knowles, 2006]

• Converts different cost values into single one• Parameterized scalarizing weight vector (augmented Tchebycheff function)• Using augmented Tchebycheff function• Different weight vector at each iteration• Weight vector is drawn uniformly at random• Allows for gradually building an approximation to whole Pareto front

• Learns a Gaussian processes model of search landscape• Scalar costs of all previously visited solutions is computed• DACE model of landscape is constructed by maximum-likelihood• Solution that maximizes expected improvement becomes next point• Evaluation on real, expensive cost function• Update after every function evaluation

• Ensures that weakly dominated solutions are rewarded less than Paretooptimal ones

Naujoks, Stork, Zaefferer, Bartz-Beielstein 50 / 74

Page 80: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Alternative approaches

RASM – Rank-based aggregated surrogate models [Loshchilov et al.,2010]

• Mono-surrogate approach again• Single surrogate model to reflect Pareto dominance in EMO framework• Locally approximates Pareto dominance relation

• Ranking neighbor points within the objective space• Offspring filter estimating whether they improve on their parents in terms ofapproximated Pareto-dominance

• Used for offspring generation in standard EMOA

• Modeling Pareto dominance within the rank-SVM framework

Naujoks, Stork, Zaefferer, Bartz-Beielstein 51 / 74

Page 81: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Existing libraries and approaches

• Many libraries already existing, e.g.- mlrMBO- DiceKriging- SUMO- parEGO

- GPareto- SPOT- Shark- QstatLab

• Overview on SAMCO homepage:http://samco.gforge.inria.fr/doku.php?id=surr_mco

• However: up-to-date overview is missing• List algorithms contained• Compare strengths and weaknesses

Naujoks, Stork, Zaefferer, Bartz-Beielstein 52 / 74

Page 82: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO Promising research areas

• Multiple objectives with different response surfaces+ specific requirements of set- and indicator- based optimization

• New variants of models• New infill criteria

• Approaches beyond one model per objective function• Model dominance relations• Model performance indicator landscapes

• Ensembles of Surrogates• Multiple surrogates simultaneously or successively

- To improve overall quality of prediction of each objective- Model evolves over time from a coarse grained to finer one- Different parts of search space with significantly different behavior

Naujoks, Stork, Zaefferer, Bartz-Beielstein 53 / 74

Page 83: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO Promising research areas

• Multiple objectives with different response surfaces+ specific requirements of set- and indicator- based optimization

• New variants of models• New infill criteria• Approaches beyond one model per objective function

• Model dominance relations• Model performance indicator landscapes

• Ensembles of Surrogates• Multiple surrogates simultaneously or successively

- To improve overall quality of prediction of each objective- Model evolves over time from a coarse grained to finer one- Different parts of search space with significantly different behavior

Naujoks, Stork, Zaefferer, Bartz-Beielstein 53 / 74

Page 84: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

SAMCO Promising research areas

• Collect existing approaches and libraries

• Benchmarking Surrogate-Assisted Optimizers lacks rigorously• Review of common test functions (academic vs. real-world)• Understand weaknesses and strengths of each algorithm• Algorithm recommendations for practice

• Overview on SAMCO homepage:http://samco.gforge.inria.fr/doku.php?id=benchmarking

Naujoks, Stork, Zaefferer, Bartz-Beielstein 54 / 74

Page 85: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Overview

• Motivation

• Concepts and methods

• Practical approach: instructive application

• Typical problems in application

• Open Issues / Research perspectives / Fields of Interest• Multi-criteria optimization• Combinatorial optimization

• Discussion• Typical problems and their solutions

Naujoks, Stork, Zaefferer, Bartz-Beielstein 55 / 74

Page 86: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Discrete / combinatorial / structured search spaces

• Well established in expensive, continuous optimization

What about combinatorial / discrete optimization problems?

• Let’s get an overview

Naujoks, Stork, Zaefferer, Bartz-Beielstein 56 / 74

Page 87: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Survey: combinatorial surrogatesMixed variablesmodel optimizer cost budget dimension remarks / topics reference

RBFN ES cheap /∼ expensive

560 /280

15 /23

benchmark /real-world:medical image analysis

Li et al. [2008]

Random Forest,Kriging NSGA2 ∼expensive - 4-76 algorithm tuning Hutter et al. [2010]

RBFN +cluster GA cheap 2,000 12

benchmark,real-world:chemical industry

Bajer and Holeňa [2010]

RBFN +GLM GA cheap several

thousand 4-13benchmark,real-world:chemical industry

Bajer and Holeňa [2013]

SVR NSGA2 ? 2,000 10 finite element,multi criteria Herrera et al. [2014]

Binary stringsmodel optimizer cost budget dimension remarks / topics reference

ANN SA expensive ? 16 real world,pump positioning Rao and Manju [2007]

RBFN GA cheap dimension2 10-25 NK-Landscape Moraglio and Kattan [2011a]

RBFN GA expensive 100 10-40 benchmark,package deal negotiation Fatima and Kattan [2011]

Kriging GA cheap dimension2 10-25 NK-Landscape Zaefferer et al. [2014b]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 57 / 74

Page 88: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Survey: combinatorial surrogates

Permutations

model optimizer cost budget dimension remarks / topics reference

custom brute force expensive 28 6signed permutation,real world:weld sequence

Voutchkov et al. [2005]

RBFN GA cheap 100 30 - 32 benchmark Moraglio et al. [2011]Kriging GA cheap 100 12 - 32 benchmark Zaefferer et al. [2014b]Kriging GA cheap 200 10 - 50 distance selection Zaefferer et al. [2014a]

Kriging ACO cheap 100 -1,000 50 - 100 benchmark,

tuning Pérez Cáceres et al. [2015]

RBFN GA* instancedependent 1,000 50 - 1,928

numerical stability,real world:cell suppression

Smith et al. [2016]

Kriging brute force,GA cheap 100 5 - 10 kernel definiteness Zaefferer and Bartz-Beielstein [2016]

*Different integration: GA produces random solutions, which are filtered by the model in each iteration

Naujoks, Stork, Zaefferer, Bartz-Beielstein 58 / 74

Page 89: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Survey: combinatorial surrogatesTreesmodel optimizer cost budget remarks / topics referenceRBFN GA cheap 100 symbolic regression Moraglio and Kattan [2011b]

kNN GA expensive 30,000 Phenotypic similarity,genetic programming Hildebrandt and Branke [2014]

RBFN* GA cheap 100 symbolic regression,parity Kattan and Ong [2015]

Random Forest GA cheap 15,000 benchmark,genetic programming Pilát and Neruda [2016]

*two models: semantic and fitness

Othermodel optimizer cost budget dimension remarks / topics reference

k-NN GA rathercheap

20000 -200000 161 - 259

real-valued+structure,real-world,protein structure

Custódio et al. [2010]

Kriging GA expensive fewhundreds

graph-based,real-world,protein structure

Romero et al. [2013]

ANN DE cheap severalhundreds 40 - 500 assignment problem,

dynamic Hao et al. [2016]

Naujoks, Stork, Zaefferer, Bartz-Beielstein 59 / 74

Page 90: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Summary: Strategies

• Strategies of dealing with discrete / combinatorial searchspaces

• Inherently discrete models (e.g., regression trees)• simple, but may not be efficient/feasible for any

representation

• Dummy variables• Only for linear regression, vector-based

• Feature based• Extract real-valued features from genotype / phenotype• Requires good features

• (Dis)similarity measure based (distance, kernel)• Requires good measure

Naujoks, Stork, Zaefferer, Bartz-Beielstein 60 / 74

Page 91: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Summary: Types of Models

• As varied as in the continuous case:

• Custom, application specific models (expertknowledge, physics)

• Artificial Neural Networks (ANN)• Markov Random Fields [Allmendinger et al., 2015]• Random Forest (Integer, Mixed Integer Problems)• (Probabilistic models - in Estimation of DistributionAlgorithms)

• (Pheromone trails - in Ant Colony Optimization)• ”Classical” kernel-based (similarity-based) models:

• k-Nearest Neighbour (k-NN)• Radial Basis Function Networks (RBFN)• Support Vector Regression (SVR)• Kriging (Gaussian Processes)

Linear regression

Markov Random FieldsKriging

Support Vector Machines

Neural

Networks

k-NN

RBFNs

Random Forest

custom

Proba-

bilistic

Models

Naujoks, Stork, Zaefferer, Bartz-Beielstein 61 / 74

Page 92: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Why kernel based approach, Kriging?

• Conceptually simple:• Replace kernel or distance function• e.g., with Gaussian kernel and arbitrary distance:

k(x, x′) = exp(−θd(x, x′))

• Transfer of popular method from continuous domain• Powerful predictor• Elegant parameter fitting (maximum likelihood estimation)• Uncertainty estimate, Expected Improvement

→ Efficient Global Optimization EGO [Jones et al., 1998b]

• Note:• None of these features exclusive to Kriging• Closely related to other model types

Naujoks, Stork, Zaefferer, Bartz-Beielstein 62 / 74

Page 93: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Combinatorial Surrogates: Research Questions

• Which kernel/distance works best and why?

• How to choose a suitable kernel/distance?

• Or else, combine?

• Genotypic vs phenotypic distances? [Hildebrandt and Branke, 2014]

• Definiteness?

• Dimensionality issues? Dimensionality reduction? See e.g., the very highdimensional problems in [Smith et al., 2016]

• Comparison of model types?1

• And again: benchmarking / testing?

1If you want to compare your approach to our methods: R Package for CombinatorialEfficient Global Optimization CEGO - https://cran.r-project.org/package=CEGO.Naujoks, Stork, Zaefferer, Bartz-Beielstein 63 / 74

Page 94: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Research Question: Choosing a Distance / Kernel[Zaefferer et al., 2014a]

* Choice crucial for success* Use prior knowledge (if available?)* Cross-validation• Fitness Distance Correlation (FDC)(potentially misleading)

FDC

Ham.Swa.Int.

Lev.R

Pos.Posq.Adj.

LCSeqLCStr

Euc.Man.Che.Lee

tho3

0kr

a32

nug3

0nu

g12

reC

05re

C13

reC

19re

C31

bayg

29fr

i26

gr24

atsp

10at

sp20

atsp

30w

t40a

wt4

0bw

t40c

wt4

0d

0.00.10.20.30.4

value

note: larger FDC values are better

• Maximum Lilkelihood Estimation(MLE) (seems to work well)

Performance reC19

● ●●

● ●

●●

Ham.All

Lev.Pos.

LCSeqLee

Man.GAInt.

Swa.Euc.

Posq.Adj.

RLCStr

2200 2300 2400

All (MLE) - GA (model-free) - Posq (squared position distance)

note: smaller performance values are better

Naujoks, Stork, Zaefferer, Bartz-Beielstein 64 / 74

Page 95: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Research Question: Definiteness[Zaefferer and Bartz-Beielstein, 2016]

So we can just replace the distance or kernel function with something appropriateand everything is fine, right?

Common requirement for kernels (distances): Definiteness2

• Definiteness may be unknown / lacking• Designing definite kernels may be hard / infeasible• Required: correction procedure• Some results from SVM field [Ong et al., 2004; Chen et al., 2009; Loosliet al., 2015] Survey: [Schleif and Tino, 2015]

• Can be transfered to Kriging, with some tweaks2Positive semi-definite kernel matrix: all eigenvalues are positive or zero

Naujoks, Stork, Zaefferer, Bartz-Beielstein 65 / 74

Page 96: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

That’s all Folks. Thanks for hanging on.

• Any questions?

• Discussion:

• Problems you encountered in practice?• New directions, challenges?• What is missing in the field?• Interesting applications?

Naujoks, Stork, Zaefferer, Bartz-Beielstein 66 / 74

Page 97: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Allmendinger, R., Coello, C. A. C., Emmerich, M. T. M., Hakanen, J., Jin, Y.,and Rigoni, E. (2015). Surrogate-assisted multicriteria optimization (wg6). InGreco, S., Klamroth, K., Knowles, J. D., and Rudolph, G., editors,Understanding Complexity in Multiobjective Optimization (Dagstuhl Seminar15031) - Dagstuhl Reports, volume 5, pages 96–163, Dagstuhl, Germany.Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik.

Bajer, L. and Holeňa, M. (2010). Surrogate model for continuous and discretegenetic optimization based on rbf networks. In Intelligent Data Engineering andAutomated Learning – IDEAL 2010, volume 6283 LNCS, pages 251–258.

Bajer, L. and Holeňa, M. (2013). Surrogate model for mixed-variablesevolutionary optimization based on glm and rbf networks. In SOFSEM 2013:Theory and Practice of Computer Science, volume 7741 LNCS, pages 481–490.

Breiman, L. (2001). Random forests. Machine learning, 45(1):5–32.Breiman, L., Friedman, J., Stone, C. J., and Olshen, R. A. (1984). Classificationand regression trees. CRC press.

Chen, Y., Gupta, M. R., and Recht, B. (2009). Learning kernels from indefinitesimilarities. In Proceedings of the 26th Annual International Conference onMachine Learning, ICML ’09, pages 145–152, New York, NY, USA. ACM.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 67 / 74

Page 98: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Custódio, F. L., Barbosa, H. J., and Dardenne, L. E. (2010). Full-atom ab initioprotein structure prediction with a genetic algorithm using a similarity-basedsurrogate model. In Proceedings of the Congress on Evolutionary Computation(CEC’10), pages 1–8, New York, NY, USA. IEEE.

Deng, L. and Yu, D. (2014). Deep learning: methods and applications.Foundations and Trends in Signal Processing, 7(3–4):197–387.

Emmerich, M. T. M., Giannakoglou, K. C., and Naujoks, B. (2006). Single- andmultiobjective evolutionary optimization assisted by gaussian random fieldmetamodels. IEEE Transactions on Evolutionary Computation, 10(4):421–439.

Fatima, S. and Kattan, A. (2011). Evolving optimal agendas for package dealnegotiation. In Proceedings of the 13th Annual Conference on Genetic andEvolutionary Computation, GECCO ’11, pages 505–512, New York, NY, USA.ACM.

Flasch, O., Mersmann, O., and Bartz-Beielstein, T. (2010). Rgp: An open sourcegenetic programming system for the r environment. In Proceedings of the 12thAnnual Conference Companion on Genetic and Evolutionary Computation,GECCO ’10, pages 2071–2072, New York, NY, USA. ACM.

Forrester, A., Sobester, A., and Keane, A. (2008). Engineering Design viaSurrogate Modelling. Wiley.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 68 / 74

Page 99: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Forrester, A. I. and Keane, A. J. (2009). Recent advances in surrogate-basedoptimization. Progress in Aerospace Sciences, 45(1):50–79.

Hao, J., Liu, M., Lin, J., and Wu, C. (2016). A hybrid differential evolutionapproach based on surrogate modelling for scheduling bottleneck stages.Computers & Operations Research, 66:215–224.

Haykin, S. (2004). A comprehensive foundation. Neural Networks, 2(2004).Herrera, M., Guglielmetti, A., Xiao, M., and Filomeno Coelho, R. (2014).Metamodel-assisted optimization based on multiple kernel regression for mixedvariables. Structural and Multidisciplinary Optimization, 49(6):979–991.

Hildebrandt, T. and Branke, J. (2014). On using surrogates with geneticprogramming. Evolutionary Computation, pages 1–25.

Hinton, G. E., Osindero, S., and Teh, Y.-W. (2006). A fast learning algorithm fordeep belief nets. Neural computation, 18(7):1527–1554.

Hornik, K., Stinchcombe, M., and White, H. (1989). Multilayer feedforwardnetworks are universal approximators. Neural networks, 2(5):359–366.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 69 / 74

Page 100: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Hutter, F., Hoos, H. H., and Leyton-Brown, K. (2010). Sequential model-basedoptimization for general algorithm configuration (extended version). TechnicalReport TR-2010-10, University of British Columbia, Department of ComputerScience. Available online:http://www.cs.ubc.ca/˜hutter/papers/10-TR-SMAC.pdf.

Jin, Y. (2011). Surrogate-assisted evolutionary computation: Recent advancesand future challenges. Swarm and Evolutionary Computation, 1(2):61–70.

Jones, D. R. (2001). A taxonomy of global optimization methods based onresponse surfaces. Journal of global optimization, 21(4):345–383.

Jones, D. R., Schonlau, M., and Welch, W. J. (1998a). Efficient globaloptimization of expensive black-box functions. Journal of Global optimization,13(4):455–492.

Jones, D. R., Schonlau, M., and Welch, W. J. (1998b). Efficient globaloptimization of expensive black-box functions. Journal of Global Optimization,13(4):455–492.

Kattan, A. and Ong, Y.-S. (2015). Surrogate genetic programming: A semanticaware evolutionary search. Information Sciences, 296:345–359.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 70 / 74

Page 101: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Knowles, J. (2006). Parego: A hybrid algorithm with on-line landscapeapproximation for expensive multiobjective optimization problems. IEEETransactions on Evolutionary Computation, 10(1):50–66.

Li, R., Emmerich, M. T. M., Eggermont, J., Bovenkamp, E. G. P., Bäck, T.,Dijkstra, J., and Reiber, J. (2008). Metamodel-assisted mixed integer evolutionstrategies and their application to intravascular ultrasound image analysis. InProceedings of the Congress on Evolutionary Computation (CEC’08), pages2764–2771, New York, NY, USA. IEEE.

Loosli, G., Canu, S., and Ong, C. (2015). Learning svm in krein spaces. IEEETransactions on Pattern Analysis and Machine Intelligence, 38(6):1204–1216.

Loshchilov, I., Schoenauer, M., and Sebag, M. (2010). Dominance-BasedPareto-Surrogate for Multi-Objective Optimization. In Simulated Evolution andLearning (SEAL 2010), volume 6457 of LNCS, pages 230–239. Springer.

Moraglio, A. and Kattan, A. (2011a). Geometric generalisation of surrogate modelbased optimisation to combinatorial spaces. In Proceedings of the 11thEuropean Conference on Evolutionary Computation in CombinatorialOptimization, EvoCOP’11, pages 142–154, Berlin, Heidelberg, Germany.Springer.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 71 / 74

Page 102: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Moraglio, A. and Kattan, A. (2011b). Geometric surrogate model basedoptimisation for genetic programming: Initial experiments. Technical report,University of Birmingham.

Moraglio, A., Kim, Y.-H., and Yoon, Y. (2011). Geometric surrogate-basedoptimisation for permutation-based problems. In Proceedings of the 13thAnnual Conference Companion on Genetic and Evolutionary Computation,GECCO ’11, pages 133–134, New York, NY, USA. ACM.

Naujoks, B., Steden, M., Muller, S. B., and Hundemer, J. (2007). Evolutionaryoptimization of ship propulsion systems. In 2007 IEEE Congress on EvolutionaryComputation, pages 2809–2816.

Ong, C. S., Mary, X., Canu, S., and Smola, A. J. (2004). Learning withnon-positive kernels. In Proceedings of the Twenty-first International Conferenceon Machine Learning, ICML ’04, pages 81–88, New York, NY, USA. ACM.

Pilát, M. and Neruda, R. (2016). Feature extraction for surrogate models ingenetic programming. In Parallel Problem Solving from Nature – PPSN XIV,pages 335–344. Springer Nature.

Pérez Cáceres, L., López-Ibáñez, M., and Stützle, T. (2015). Ant colonyoptimization on a limited budget of evaluations. Swarm Intelligence, pages1–22.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 72 / 74

Page 103: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Queipo, N. V., Haftka, R. T., Shyy, W., Goel, T., Vaidyanathan, R., and Tucker,P. K. (2005). Surrogate-based analysis and optimization. Progress in aerospacesciences, 41(1):1–28.

Rao, S. V. N. and Manju, S. (2007). Optimal pumping locations of skimmingwells. Hydrological Sciences Journal, 52(2):352–361.

Romero, P. A., Krause, A., and Arnold, F. H. (2013). Navigating the proteinfitness landscape with Gaussian processes. Proceedings of the NationalAcademy of Sciences, 110(3):E193–E201.

Sacks, J., Welch, W. J., Mitchell, T. J., and Wynn, H. P. (1989). Design andanalysis of computer experiments. Statistical science, pages 409–423.

Schleif, F.-M. and Tino, P. (2015). Indefinite proximity learning: A review. NeuralComputation, 27(10):2039–2096.

Smith, J., Stone, C., and Serpell, M. (2016). Exploiting diverse distance metricsfor surrogate-based optimisation of ordering problems: A case study. InProceedings of the 2016 on Genetic and Evolutionary Computation Conference,GECCO ’16, pages 701–708, New York, NY, USA. ACM.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 73 / 74

Page 104: Meta-Model Assisted (Evolutionary) Optimization - Tutorial at PPSN

Voutchkov, I., Keane, A., Bhaskar, A., and Olsen, T. M. (2005). Weld sequenceoptimization: The use of surrogate models for solving sequential combinatorialproblems. Computer Methods in Applied Mechanics and Engineering,194(30-33):3535–3551.

Zaefferer, M. and Bartz-Beielstein, T. (2016). Efficient global optimization withindefinite kernels. In Parallel Problem Solving from Nature–PPSN XIV, pages69–79. Springer.

Zaefferer, M., Stork, J., and Bartz-Beielstein, T. (2014a). Distance measures forpermutations in combinatorial efficient global optimization. In Bartz-Beielstein,T., Branke, J., Filipic, B., and Smith, J., editors, Parallel Problem Solving fromNature–PPSN XIII, pages 373–383, Cham, Switzerland. Springer.

Zaefferer, M., Stork, J., Friese, M., Fischbach, A., Naujoks, B., andBartz-Beielstein, T. (2014b). Efficient global optimization for combinatorialproblems. In Proceedings of the 2014 Conference on Genetic and EvolutionaryComputation, GECCO ’14, pages 871–878, New York, NY, USA. ACM.

Naujoks, Stork, Zaefferer, Bartz-Beielstein 74 / 74