38
Optimal Execution and Stochastic Approximation: Learning by Trading Sophie LARUELLE Laboratoire d’Analyse et de Math´ ematiques Appliqu´ ees Universit´ e Paris-Est Cr´ eteil The Mathematics of High Frequency Markets April 14, 2015

Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

  • Upload
    others

  • View
    23

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Optimal Execution and StochasticApproximation:

Learning by Trading

Sophie LARUELLE

Laboratoire d’Analyse et de Mathematiques AppliqueesUniversite Paris-Est Creteil

The Mathematics of High Frequency Markets

April 14, 2015

Page 2: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Outline

1 Introduction to Stochastic Approximation Algorithms

2 Applications to Optimal ExecutionOptimal split of orders across liquidity poolsOptimal posting price of limit orders: learning by tradingOptimal split and posting price of limit orders across lit pools

Page 3: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Outline

1 Introduction to Stochastic Approximation Algorithms

2 Applications to Optimal ExecutionOptimal split of orders across liquidity poolsOptimal posting price of limit orders: learning by tradingOptimal split and posting price of limit orders across lit pools

Page 4: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Motivations

In many fields, we often are faced with optimization or zero searchproblems.

Examples: In Finance,

extraction of implicit parameter (volatility),

calibration, optimization of an exogenous parameter for variancereduction (regression, importance sampling, etc).

Common point: all the concerned functions have a representation as anexpectation, i .e.

h(θ) = E [H(θ,Y )], where Y is a q-dimensional random vector.

The stochastic approximation is a tool based on simulation to solveoptimization or zero search problems.

Page 5: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Measurement : Robbins-Monro procedure

A dose θ of a chemical product creates a random effect measured byF (θ,Y ), Y being a random variable with distribution µ and F : R2 7→ R.

We assume that the mean effect

f (θ) = E [F (θ,Y )] is non-decreasing.

We want to determine the dose θ∗ which creates a mean effect of a giventhreshold α, i .e. to solve

f (θ∗) = α.

Naive idea: f (θn) ≈ 1

Nn

Nn∑k=1

F (θn,Yk), Yk i.i.d. with distribution µ,

Nn −→n→∞

∞.

Other idea: use F (θn,Yn+1) instead of f (θn) with γn 0 (“local”randomization).

Page 6: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

The Robbins-Monro procedure is the following

choose arbitrarily θ0 and administer it to a subject which reacts withthe effect F (θ0,Y1).

Recurrence: at instant n, choose a dose θn administered to a subject(independent of the previous ones), the effect is F (θn,Yn+1).

As (Yn)n≥1 is a sequence of i.i.d. random variables with distribution µ,then

f (θn) = E [F (θn,Yn+1) | F (θ0,Y1), . . . ,F (θn−1,Yn)].

The Robbins-Monro algorithm for the choice of θn then reads

θn+1 = θn − γn (F (θn,Yn+1)− α), (γn) nonnegative tending to 0.

By setting H(θ, y) := F (θ, y)− α, this procedure is then a zero search ofthe function

h(θ) := f (θ)− α = E [F (θ,Y )− α] = E [H(θ,Y )].

Page 7: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

General Framework

Consider the following recursive procedure defined on a filtered probabilityspace (Ω,A, (Fn)n≥0,P)

∀ n ≥ 0, θn+1 = θn − γn+1h(θn) + γn+1 (∆Mn+1 + rn+1), (1)

where

h : Rd → Rd is a continuous locally Lipschitz function,

θ0 is a F0-measurable finite random vector,

for every n ≥ 0, ∆Mn+1 is a Fn-martingale increment,

rn is a Fn-adapted reminder term.

Page 8: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Convergence: Martingale approach or ODE method

Martingale approach: Existence of a Lyapunov function and controlof both martingale increment and remainder term.

ODE method: Study of the SA as a discretization of

ODEh ≡ θ = −h(θ).

Assume that (γn)n≥1 is a nonnegative sequence such that∑n≥1

γn = +∞ and∑n≥1

γ2n < +∞.

Then,θn

a.s.−→n→∞

θ∗.

Page 9: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Rate of convergence: CLT (SDE method)

Let θ∗ be an equilibrium point of h = 0. Assume that the function h isdifferentiable at θ∗ and that all the eigenvalues of Dh(θ∗) have anonnegative real part. Specify the step sequence as follows

∀n ≥ 1, γn =α

n, α >

1

2Re(λmin)(2)

where λmin denotes the eigenvalue of Dh(θ∗) with the lowest real part.

Then, the previous convergence a.s. is ruled on θna.s.−→ θ∗ by the

following CLT √n (θn − θ∗)

L−→n→∞

N (0, αΣ)

with

Σ :=

∫ +∞

0

(e−(Dh(θ∗)− Id

)u)t

Γe−(Dh(θ∗)− Id

)udu.

Page 10: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Outline

1 Introduction to Stochastic Approximation Algorithms

2 Applications to Optimal ExecutionOptimal split of orders across liquidity poolsOptimal posting price of limit orders: learning by tradingOptimal split and posting price of limit orders across lit pools

Page 11: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Outline

1 Introduction to Stochastic Approximation Algorithms

2 Applications to Optimal ExecutionOptimal split of orders across liquidity poolsOptimal posting price of limit orders: learning by tradingOptimal split and posting price of limit orders across lit pools

Page 12: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Static modelling

The principle of a Dark pool is the following:

It proposes a bid price with no guarantee of executed quantity at theoccasion of an OTC transaction.

Usually this price is lower than the bid price offered on the regularmarket.

So one can model the impact of the existence of N dark pools (N ≥ 2) ona given transaction as follows:

Let V > 0 be the random volume to be executed,

Let θi ∈ (0, 1) be the discount factor proposed by the dark pool i .

Let ri denote the percentage of V sent to the dark pool i forexecution.

Let Di ≥ 0 be the quantity of securities that can be delivered (ormase available) by the dark pool i at price θiS .

Page 13: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Cost of the executed order

The reminder of the order is to be executed on the regular market, at priceS .

Then the cost C of the whole executed order is given by

C = SN∑i=1

θi min (riV ,Di ) + S

(V −

N∑i=1

min (riV ,Di )

)

= S

(V −

N∑i=1

ρi min (riV ,Di )

)

whereρi = 1− θi ∈ (0, 1), i = 1, . . . ,N.

Page 14: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Mean Execution Cost and dynamical aspect

Minimizing the mean execution cost, given the price S , amounts tosolving the following maximization problem

max

N∑i=1

ρiE (S min (riV ,Di )) , r ∈ PN

(3)

where PN :=r = (ri )1≤i≤N ∈ RN

+ |∑N

i=1 ri = 1

.

We consider the sequence Y n := (V n,Dn1 , . . . ,D

nN)n≥1. We will make

stationary assumptions on the sequence

either (Y n)n≥1 is i.i.d.,

or (Y n)n≥1 is ergodic with some rate (e.g . α-mixing).

Page 15: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Design of the stochastic algorithmWe us a Lagrangian approach to solve this maximisation problem underconstraints. We obtain that

r∗ ∈ arg maxPNΦ

m∀i ∈ IN , E

[V(ρi1r∗i V<Di −

1N

∑Nj=1 ρj1r∗j V<Dj

)]= 0.

Consequently, this leads to the following recursive zero search procedure

rn+1i = rni + γn+1Hi (r

n,Y n+1), r0 ∈ PN , i ∈ IN , (4)

where for i ∈ IN , every r ∈ PN , every V > 0 and every D1, . . . ,DN ≥ 0,

Hi (r ,Y ) = V

ρi1riV<Di −1

N

N∑j=1

ρj1rjV<Dj

where (Y n)n≥1 is a sequence of random vectors with non negative

components such that, for every n ≥ 1 and i ∈ IN , (V n,Dni )

d= (V ,Di ).

Page 16: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

The pseudo-real data setting

V is the traded volume of a very liquid security – namely the asset BNP –during an 11 day period.

Then we selected the N most correlated assets (in terms of tradedvolumes) with the original asset, denoted Si , i = 1, . . . ,N.

The available volumes of each dark pool i have been modeled as follows

∀1 ≤ i ≤ N, Di := βi

((1− αi )V + αiSi

EV

ESi

)where

αi ∈ (0, 1), i = 1, . . . ,N are the recombining coefficients,

βi , i = 1, . . . ,N are some scaling factors,

EV and ESi stand for the empirical mean of the data sets of V andSi .

Page 17: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Parameters

The shortage situation corresponds to∑N

i=1 βi < 1 since it implies

E

[N∑i=1

Di

]< EV .

We will use the following parameters

N = 4, β = (0.1, 0.2, 0.3, 0.2)′ and α = (0.4, 0.6, 0.8, 0.2)′

and the dark pool trading (rebate) parameters are set to

ρ = (0.01, 0.02, 0.04, 0.06)′.

Page 18: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Long-term optimization

0 1 2 3 4 5 6 7 8 9 10 11

5

10

15

20

25

Days

% o

f sat

isfa

ctio

n

Relative Cost Reductions

OracleOptimizationReinforcement

0 1 2 3 4 5 6 7 8 9 10 11

20

30

40

50

60

70

80

90

100

110

Days

Rat

ios

(%)

Performances

OneOptimization/OracleReinforcement/Oracle

Figure: Long term optimization: Case N = 4,∑N

i=1 βi < 1, 0.2 ≤ αi ≤ 0.8 andr0i = 1/N, 1 ≤ i ≤ N.

Page 19: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Daily resetting of the procedureWe reset the step γn at the beginning of each day and the satisfactionparameters and we keep the allocation coefficients of the preceding day.

0 1 2 3 4 5 6 7 8 9 10 11

1

1.5

2

2.5

3

3.5

4

4.5

Days

% o

f sat

isfa

ctio

n

Relative Cost Reductions

OracleOptimizationReinforcement

0 1 2 3 4 5 6 7 8 9 10 1140

50

60

70

80

90

100

110

Days

Rat

ios(

%)

Performances

OracleOptimization/OracleReinforcement/Oracle

Figure: Daily resetting of the algorithms parameters: Case N = 4,∑N

i=1 βi < 1,0.2 ≤ αi ≤ 0.8 and r0

i = 1/N 1 ≤ i ≤ N.

Page 20: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Outline

1 Introduction to Stochastic Approximation Algorithms

2 Applications to Optimal ExecutionOptimal split of orders across liquidity poolsOptimal posting price of limit orders: learning by tradingOptimal split and posting price of limit orders across lit pools

Page 21: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Modeling and design of the algorithm

We consider on a short period T a Poisson process of “execution” ofbuy orders(

N(δ)t

)0≤t≤T

with intensity ΛT (δ, S) :=

∫ T

0λ(St − (S0 − δ))dt (5)

where

• 0 ≤ δ ≤ δmax with δmax ∈ (0,S0) denotes the depth of the limit orderbook,

• (St)t≥0 is a stochastic process modeling the dynamic of the fair priceof a security stock,

• the function λ is defined on the whole real line as a finite nonincreasing convex function.

Page 22: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Optimization Problem

Then we introduce a market impact penalization function Φ : R 7→ R+,nondecreasing and convex, with Φ(0) = 0 to model the additional cost ofthe execution of the remaining quantity.

Then the resulting cost of execution on a period [0,T ] reads

C (δ) := E[

(S0 − δ)(QT ∧ N

(δ)T

)+ κSTΦ

((QT − N

(δ)T

)+

)](6)

where κ > 0.

Our aim is then to minimize this cost, namely to solve the followingoptimization problem

min0≤δ≤δmax

C (δ). (7)

To solve this optimization problem, we will devise a stochastic algorithmconstrained to stay in [0, δmax].

Page 23: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Strategy

We will devise a stochastic algorithm constrained to stay in [0, δmax].To this end we have to

• prove that C , C ′ and C ′′ admit representations as expectations, i .e.in particular, there exists a Borel functional

H : [0, δmax]× D ([0,T ],R) −→ R such that

∀δ ∈ [0, δmax], C ′(δ) = E[H(δ, (St)t∈[0,T ]

)]• find natural assumptions on QT and κ such that C is twice

differentiable, strictly convex on [0, δmax] with C ′(0) < 0.Consequently

argminδ∈[0,δmax]C (δ) = δ∗, δ∗ ∈ (0, δmax]

andδ∗ = δmax iff C is non increasing on [0, δmax].

Page 24: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Design of the algorithmOnce the two points are checked, we can devise the algorithm following thestandard stochastic approximation with projection (see [2]), namely

δn+1 = Proj[0,δmax]

(δn − γn+1H

(δn,(S

(n+1)ti

)0≤i≤m

)), δ0 ∈ [0, δmax],

(8)where

• Proj[0,δmax] denotes the projection on [0, δmax],

• the positive step sequence (γn)n≥1 satisfies at least the minimaldecreasing step assumption∑

n≥1

γn = +∞ and γn → 0, (9)

•(

S(n)ti

)0≤i≤m, n ≥ 0

is either a sequence of i.i.d. copies of the true

underlying dynamics of (Sti )0≤i≤m or at least of its Euler scheme or asequence sharing some averaging properties in the sense of [4] (e.g .stationary α-mixing).

Page 25: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Numerical experiments: Market data

As market data, we use the bid prices of Accor SA (ACCP.PA) of11/11/2010 for the fair price process (St)t∈[0,T ]. We divide the day intoperiods of 15 trades which will denote steps of the stochastic procedure.Let Ncycles be the number of these periods. For every n ∈ Ncycles, we have

a sequence of bid prices (S(n)ti )1≤i≤15 and we approximate the jump

intensity of the Poisson process ΛT n(δ, S), where T n =∑15

i=1 ti , by

∀n ∈ Ncycles, ΛT n(δ, S) = A15∑i=2

e−k(S(n)ti−St1 +δ)(ti − ti−1).

The empirical mean of the intensity function

Λ(δ, S) =1

Ncycles

Ncycles∑n=1

ΛT n(δ, S)

is plotted on the following figure.

Page 26: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Intensity function

0 0.02 0.04 0.06 0.08 0.10

200

400

600

800

1000

1200

δ

Intensity

Figure: Fit of the exponential model on real data (Accor SA (ACCP.PA)11/11/2010): A = 1/50, k = 50 and Ncycles = 220.

The penalization function is of the following form

Φ(x) = (1 + η(x))x with η(x) = A′ek′x .

Page 27: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Cost function and its derivative

0 0.02 0.04 0.06 0.08 0.13098.21

3098.215

3098.22

3098.225

3098.23

3098.235

3098.24

3098.245

3098.25

δ

Cost function

0 0.02 0.04 0.06 0.08 0.1−4

−3.5

−3

−2.5

−2

−1.5

−1

−0.5

0

0.5

1

δ

Derivative

Figure: η 6≡ 0: A = 1/50, k = 50, Q = 100, κ = 1, A′ = 0.001, k ′ = 0.0005 andNcycles = 220.

Page 28: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

δ and posting price obtained by SA

0 0.5 1 1.5 2 2.5

x 104

0

0.005

0.01

0.015

0.02

0.025

time (s)

Stochastic Approximation

0 0.5 1 1.5 2 2.5

x 104

30.75

30.8

30.85

30.9

30.95

31

31.05

31.1

31.15

time (s)

Fair and Posting Prices

Fair pricePosting price

Figure: η 6≡ 0: A = 1/50, k = 50, Q = 100, κ = 1, A′ = 0.001, k ′ = 0.0005 andNcycles = 220. Crude algorithm with γn = 1

550n .

Page 29: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Stochastic algorithm with RP averaging

0 0.5 1 1.5 2 2.5

x 104

0

0.005

0.01

0.015

0.02

0.025

time (s)

Stochastic Approximation

0 0.5 1 1.5 2 2.5

x 104

30.75

30.8

30.85

30.9

30.95

31

31.05

31.1

31.15

time (s)

Fair and Posting Prices

Fair pricePosting price

Figure: η 6≡ 0: A = 1/50, k = 50, Q = 100, κ = 1, A′ = 0.001, k ′ = 0.0005 andNcycles = 220. Averaging algorithm (Ruppert and Poliak) with γn = 1

550n0.95 .

Page 30: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Outline

1 Introduction to Stochastic Approximation Algorithms

2 Applications to Optimal ExecutionOptimal split of orders across liquidity poolsOptimal posting price of limit orders: learning by tradingOptimal split and posting price of limit orders across lit pools

Page 31: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Framework and Modelling

Assume that a trader wants to buy a volume V of an asset across Nlit pools with limit orders.

She has to determinate the proportions r = (r i )1≤i≤N to sent to eachlit pools and the posting prices (S − δ) = (S − δi )1≤i≤N ∈ [0, δmax]N .

The execution flow at the distance δi of the reference price S ismodelled by a random variable Q i (δi ) = Q ie−k

iδi where Q i is apositive random variable modeling the executed quantity at the firstlimit an k i > 0.

If she is not fully executed, she sends a market order of the remainingquantity.

Page 32: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Mean Execution Cost

The mean resulting cost of execution is the sum of each meanexecution costs on the lit pools, namely it reads

C (r , δ) :=N∑i=1

E[(S − δi )(r iV ∧ Q i (δi )) + κS(r iV − Q i (δi ))+

], (10)

where κ > 0 is a free tuning parameter.

Our aim is then to minimize this cost by choosing the proportions andthe distances to post at, namely to solve the following optimizationproblem

minr∈PN ,δ∈[0,δmax]N

C (r , δ). (11)

We take advantage of the representation of C and its first two derivativesas expectations to devise a recursive stochastic algorithm.

Page 33: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Design of the stochastic algorithmBased on a Lagrangian approach for the optimal proportions and on therepresentations as expectations for C ′ and C ′′, we can formally devise arecursive stochastic gradient descent

rn+1 = ProjPN

(rn − γn+1H(rn, δn, Qne

−kδn)), n ≥ 0,

δn+1 = Proj[0,δmax]N

(δn − γn+1G (rn, δn, Qne

−kδn)), n ≥ 0,

where, for every i ∈ 1, . . . ,N,

H(r in, δin, Q

ine−k iδin) = V

((S − δin)1r inV≤Q i

ne−ki δin + κS1r inV≥Q i

ne−ki δin

)− 1

N

∑Nj=1 V

((S − δjn)1

r jnV≤Q jne−kj δ

jn

+ κS1r jnV≥Q j

ne−kj δjn

),

and

G (r in, δin, Q

ine−k iδin) = −r inV ∧ Q i

ne−k iδin + k i (κS − (S − δin))1r inV≤Q i

ne−ki δin.

Page 34: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Convergence of the SA procedure

0 2000 4000 6000 8000 100000.475

0.48

0.485

0.49

0.495

0.5

0.505

0.51

0.515

0.52

0.525 Stochastic algorithm for the proportions

Periods

r1

r2

0 2000 4000 6000 8000 100000

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08 Stochastic algorithm for the posting prices

Periods

δ

1

δ2

Figure: Convergence of the stochastic algorithm with N = 2, V = 150, S = 100,m

Q= (8090)t , v

Q= (11)t , k = (2025)t , κ = 1, r i0 = 1/N, δi0 = 0.05 and

n = 10000.

Page 35: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Convergence of the SA procedure with RP averaging

0 200 400 600 800 10000

0.1

0.2

0.3

0.4

0.5

0.6

0.7 Stochastic algorithm for the proportions

Periods

r1

r2

0 200 400 600 800 10000

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.045

0.05 Stochastic algorithm for the posting prices

Periods

δ

1

δ2

Figure: Convergence of the stochastic algorithm with N = 2, V = 150, S = 100,m

Q= (8090)t , v

Q= (11)t , k = (2025)t , κ = 1, r i0 = 1/N, δi0 = 0.05 and

n = 1000.

Page 36: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Perspectives

Optimal split of large volumes across both dark and lit pools (withintegrated or mid-point book for dark pools and limit orders for litpools).

Optimal posting price of limit order with other kind of matchingmechanism : pro-rata or a mix between FIFO and pro-rata. Thiscould be generalized whith order split too.

Optimal execution for portfolios.

Page 37: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Some References

M. Duflo, Algorihtmes Stochastiques, volume 23 of Math. and Applications (Berlin),Springer-Verlag, Berlin, 1996.

H. J. Kushner, D. S. Clark, Stochastic approximation methods for constrained andunconstrained systems, volume 26 of Applied Mathematical Sciences. Springer-Verlag, NewYork, second edition, 1978.

S. Laruelle, C.-A. Lehalle, G. Pages, Optimal split of orders across liquidity pools: astochastic algorithm approach, SIAM Journal on Financial Mathematics, 2(1):1042–1076,2011.

S. Laruelle, G. Pages, Stochastic Approximation with Averaging Innovation Applied toFinance, Monte Carlo Methods and Applications, 18 (1):1–51, 2012.

S. Laruelle, C.-A. Lehalle, G. Pages, Optimal posting price of limit orders: learning bytrading, Mathematics and Financial Economics, 7(3):359–403, 2013.

S. Laruelle, Faisabilite de l’apprentissage des parametres d’un algorithme de trading sur desdonnees reelles, Cahier de la Chaire Finance and Sustainable Development, Hors-SerieMicrostructure des marches, juin 2013.

S. Laruelle, C.-A. Lehalle, Market Microstructure in Practice, World Scientific, 2013.

S. Laruelle, Optimal split and posting price of limit orders across lit pools using stochasticapproximation, Initiative de Recherche “Microstructure des marches” de Kepler-Cheuvreux,sous la tutelle de l’Institut Europlace de Finance, avril 2014.

Page 38: Optimal Execution and Stochastic Approximation: Learning ...helper.ipam.ucla.edu/publications/fmws2/fmws2_12690.pdfOptimal Execution and Stochastic Approximation: Learning by Trading

Thank you for your attention