369
Contents Preface ix 1 Dynamical Systems, Ensembles, and Transfer Operators 1 1.1 Ergodic Preamble ............................ 1 1.2 The Ensemble Perspective ........................ 2 1.3 Evolution of Ensembles ......................... 8 1.4 Various Useful Representations and Invariant Density of a Differential Equation ................................. 14 2 Dynamical Systems Terminology and Definitions 31 2.1 The Form of a Dynamical System .................... 32 2.2 Linearization ............................... 34 2.3 Hyperbolicity .............................. 36 2.4 Hyperbolicity: Nonautonomous vector fields .............. 40 3 Frobenius-Perron Operator and Infinitesimal Generator 45 3.1 Frobenius-Perron Operator ....................... 45 3.2 Infinitesimal Operators ......................... 47 3.3 Frobenius-Perron Operator of Discrete Stochastic Systems ....... 51 3.4 Invariant Density is a “Fixed Point" of the Frobenius-Perron Operator . 54 3.5 Invariant Sets and Ergodic Measure ................... 55 3.6 Relation between the Frobenius-Perron and Koopman Operators ... 66 4 Graph Theoretic Methods and Markov Models of Dynamical Transport 69 4.1 Finite-rank approximation of the Frobenius-Perron operator ...... 71 4.2 The Markov Partition: How it Relates to the Frobenius-Perron Operator 73 4.3 The Approximate Action of Dynamical System on Density looks like a Directed Graph: Ulam’s Method is a form of Galerkin’s Method . . . 78 4.4 Exact Representations are Dense and the Ulam-Galerkin’s Method .. 96 5 Graph Partition Methods and Their Relationship to Transport in Dynamical Systems 105 5.1 Graphs and Partitions .......................... 105 5.2 Weakly Transitive ............................ 105 5.3 Partition by Signs of The Second Eigenvector ............. 111 5.4 Graph Laplacian and Almost invariance ................. 117 v

Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Contents

Preface ix

1 Dynamical Systems, Ensembles, and Transfer Operators 11.1 Ergodic Preamble . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.2 The Ensemble Perspective . . . . . . . . . . . . . . . . . . . . . . . . 21.3 Evolution of Ensembles . . . . . . . . . . . . . . . . . . . . . . . . . 81.4 Various Useful Representations and Invariant Density of a Differential

Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

2 Dynamical Systems Terminology and Definitions 312.1 The Form of a Dynamical System . . . . . . . . . . . . . . . . . . . . 322.2 Linearization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342.3 Hyperbolicity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 362.4 Hyperbolicity: Nonautonomous vector fields . . . . . . . . . . . . . . 40

3 Frobenius-Perron Operator and Infinitesimal Generator 453.1 Frobenius-Perron Operator . . . . . . . . . . . . . . . . . . . . . . . 453.2 Infinitesimal Operators . . . . . . . . . . . . . . . . . . . . . . . . . 473.3 Frobenius-Perron Operator of Discrete Stochastic Systems . . . . . . . 513.4 Invariant Density is a “Fixed Point" of the Frobenius-Perron Operator . 543.5 Invariant Sets and Ergodic Measure . . . . . . . . . . . . . . . . . . . 553.6 Relation between the Frobenius-Perron and Koopman Operators . . . 66

4 Graph Theoretic Methods and Markov Models of Dynamical Transport 694.1 Finite-rank approximation of the Frobenius-Perron operator . . . . . . 714.2 The Markov Partition: How it Relates to the Frobenius-Perron Operator 734.3 The Approximate Action of Dynamical System on Density looks like

a Directed Graph: Ulam’s Method is a form of Galerkin’s Method . . . 784.4 Exact Representations are Dense and the Ulam-Galerkin’s Method . . 96

5 Graph Partition Methods and Their Relationship to Transport in DynamicalSystems 1055.1 Graphs and Partitions . . . . . . . . . . . . . . . . . . . . . . . . . . 1055.2 Weakly Transitive . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1055.3 Partition by Signs of The Second Eigenvector . . . . . . . . . . . . . 1115.4 Graph Laplacian and Almost invariance . . . . . . . . . . . . . . . . . 117

v

Page 2: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

vi

5.5 Finite-time coherent sets . . . . . . . . . . . . . . . . . . . . . . . . . 1245.6 Spectral partitioning for the coherent pair . . . . . . . . . . . . . . . . 1285.7 The SVD Connection . . . . . . . . . . . . . . . . . . . . . . . . . . 1305.8 Example 1: Idealized Stratospheric flow . . . . . . . . . . . . . . . . 1315.9 Example 2: Stratospheric polar vortex as coherent sets . . . . . . . . . 1325.10 Community Methods . . . . . . . . . . . . . . . . . . . . . . . . . . 1365.11 Open Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1455.12 Relative Measure and Finite Time Relative Coherence . . . . . . . . . 149

6 The Topological Dynamics Perspective of Symbol Dynamics 1536.1 Symbolization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1536.2 Chaos . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1706.3 Horseshoe Chaos by Melnikov Function Analysis . . . . . . . . . . . 1806.4 Learning Symbolic Grammar in Practice . . . . . . . . . . . . . . . . 1846.5 Stochasticity, Symbolic Dynamics and Finest Scale . . . . . . . . . . 197

7 Transport Mechanism, Lobe Dynamics, Flux Rates and Escape 2017.1 Transport Mechanism . . . . . . . . . . . . . . . . . . . . . . . . . . 2017.2 Markov Model Dynamics for Lobe Dynamics, A Henon Map Example 2157.3 On Lobe Dynamics of Resonance Overlap . . . . . . . . . . . . . . . 2217.4 Transport Rates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 222

8 Finite Time Lyapunov Exponents: FTLE 2418.1 Lyapunov exponents: One-dimensional Maps . . . . . . . . . . . . . 2418.2 Lyapunov exponents: Diffeomorphism and flow . . . . . . . . . . . . 2438.3 Finite-time Lyapunov exponents (FTLE) and Lagrangian coherent struc-

ture (LCS) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246

9 Information Theory in Dynamical Systems 2699.1 A Little Shannon Information on Coding by Example . . . . . . . . . 2699.2 A Little More Shannon Information on Coding . . . . . . . . . . . . . 2739.3 Many Random Variables and Taxonomy of the Entropy Zoo . . . . . . 2759.4 Information theory in Dynamical Systems . . . . . . . . . . . . . . . 2809.5 Formally Interpreting a Deterministic Dynamical System in the Lan-

guage of Information TheoryChapter 9. Information Theory in Dy-namical Systems9.5. Formally Interpreting a Deterministic Dynami-cal System in the Language of Information TheoryChapter 9. Infor-mation Theory in Dynamical Systems9.5. Deterministic DynamicalSystems in Information Theory TermsChapter 9. Information The-ory in Dynamical Systems9.5. Deterministic Dynamical Systems inInformation Theory Terms9.5. Deterministic Dynamical Systems inInformation Theory Terms . . . . . . . . . . . . . . . . . . . . . . . . 285

9.6 Computational Estimates of Topological Entropy and Symbolic Dy-namics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 289

9.7 Lyapunov Exponents, and Metric Entropy and the Ulam’s MethodConnection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300

9.8 Information Flow and Transfer Entropy . . . . . . . . . . . . . . . . . 302

Page 3: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

vii

9.9 Examples of Transfer Entropy and Mutual Information in DynamicalSystems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 306

A Computation, Codes, and Computational Complexity 315A.1 Matlab Codes and Implementations of Ulam-Galerkin’s Matrix and

Ulam’s Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 315A.2 Ulam-Galerkin Code by Rectangles . . . . . . . . . . . . . . . . . . . 319A.3 Delauney Triangulation in a Three-Dimensional Phase Space . . . . . 324A.4 Delauney Triangulation and Refinement . . . . . . . . . . . . . . . . 326A.5 Analysis of Refinement . . . . . . . . . . . . . . . . . . . . . . . . . 330

Bibliography 339

Page 4: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure
Page 5: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Preface

Measurable dynamics has traditionally referred to ergodic theory, which is in somesense a sister topic to dynamical systems and chaos theory. However, the topic has untilrecently been a highly theoretically mathematical topic which is generally less obviousto those practitioners in applied areas may not find obvious links to practical real worldproblems. During the recent decade, facilitated by the advent of high speed computers, ithas become practical to represent the notion of a transfer operator discretely but to highresolution thanks to rapidly developing algorithms and new numerical methods designedfor the purpose, but with an early book on this general topic beginning with, “Cell-to-cell mapping: a method of global analysis for nonlinear systems," [177] from 1987.1 Atremendous amount of progress and sophistication has come to the empirical perspectivesince then.

Rather than discussing the behaviors of complex dynamical systems in terms of fol-lowing the fate of single trajectories, it is now possible to empirically discuss global ques-tions in terms of evolution of density. Now complementary to the traditional geometricmethods of dynamical systems transport study particularly by stable and unstable mani-folds structure and bifurcation analysis, we can analyze transport activity and evolutionby matrix representation of the Frobenius-Perron transfer operator. While the traditionalmethods allow for a analytic approach, when they work, the new and fast developing com-putational tools discussed here allow for detailed analysis of real world problems that aresimply beyond the reach of traditional methods. Here we will draw connections betweenthe new methods of transport analysis based on transfer operators and the more traditionalmethods. The goal of this book is not to become a presentation of the general topic ofdynamical systems, as there are already several excellent textbooks that achieve this goalin a manner better than we can hope. We will bring together several areas, as we will drawconnections between topological dynamics, symbolic dynamics and information theory toshow that they are also highly relevant to the Ulam-Galerkin representations. In these partsof the discussion, we will compare and contrast notions from topological dynamics to mea-surable dynamics, the latter is more so the starting topic of this book. That is if measurabledynamics means a discussion of a dynamical system in consideration of how much, howbig and other notions that require measure structure to discuss transport rates, topologi-cal dynamics can be considered as a parallel topic of study that asks similar questions inabsent of a measure that begets scale. As such, mechanism and geometry of transport ismore so the focus. Therefore, including discussion of topological dynamics in our primarydiscussion here on measurable dynamics should be considered complementary.

1Recent terminology has come to call these “set oriented" methods.

ix

Page 6: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

x Preface

There are several excellent previous related texts on mathematical aspects of trans-fer operators which wish to recommend as possible supplements. In particular, Lasotaand Mackay [208] gives a highly regarded discussion of the theoretical perspective ofFrobenius-Perron operators in dynamical systems, of which material we overlap in as faras we need these elements for the computational discussion here. Gora and Boyarsky [53]also give a sharp presentation of an ensembles density perspective in dynamical system,but more so specialized for one dimensional maps, and some of the material and proofstherein are difficult to find elsewhere. Of course the book by Baladi [11] is important inthat it gives a thoroughly rigorous presentation of transfer operators including a unique per-spective. We recommend highly the book by Ding and Zhou, [337], which covers a greatdeal of theoretical information complementary to the work discussed in this book, includ-ing Ulam’s method and piecewise constant approximations of invariant density, piecewiselinear Markov models, and especially analysis of convergence. Also an in depth study canbe found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure theory and functionalanalysis. The book by McCauley, [225], includes a useful perspective regarding what isbecoming a modern perspective on computational insight into behaviors of dynamical sys-tems, especially experimentally observed dynamical systems. That is that finite realizationsof chaotic data can give a great deal of insight. This is a major theme which we also de-velop here toward the perspective that a finite time sample of a dynamical system is not justan estimate of the long time behavior, as suggested perhaps by traditional perspective, butin fact finite time samples are most useful in their own right toward understanding finitetime behavior of a dynamical system. After all, any practical real world observation of adynamical system can be argued to only exist during a time window which cannot possiblybe infinite in duration.

There are many excellent textbooks on the general theory of dynamical systems,clearly including Robinson [279], Guckenheimer and Holmes [156], Devaney [97], Alli-good, Sauer and Yorke [2], Strogatz [313], Perko [262], Meiss [228], Ott [255], Arnold [4],Wiggins [329], and Melo and van Strein [91] to name a few. Each of these has been verypopular and successful and each is particularly strong in special aspects of dynamical sys-tems as well as broad presentation. We cannot and should not hope to repeat these worksin this presentation, but we do give what we hope is enough background of the generaldynamical systems theory in order that this work can be somewhat self contained for thenonspecialist. Therefore there is some overlap with other texts in so far as background in-formation on the general theory is given and we encourage the reader to investigate furtherin some of the other cited texts for more depth and other perspectives. More to the pointof the central theme of this textbook, the review article by Dellnitz and Junge [89] andthen later the PhD thesis by Padberg [258], advised also by Dellnitz, both give excellentpresentations of a more computationally based perspective of measurable dynamical sys-tems in common with the present text and we highly recommend them. A summary of theGerman school’s approach can be found in the book [118] to empirical study of dynamicalsystems, particularly in the book chapter [84]. Also, we recommend the review in the bookchapter by Froyland [129]. Finally, we highly recommend the 1980 book by Hsu [177],and see also [176], which is an early and less often cited work in the current literature,as we rarely see “cell-to-cell mappings" cited lately. While lacking the transfer orientedformalism behind the analysis, this cell-to-cell mapping paradigm is clearly a precursor tothe computational methods which are now commonly called “set-oriented methods." Also,

Page 7: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Preface xi

we do include discussion and contrast to the early ideas by Ulam [319] called the “Ulam’s-method." Here we hope to give a useful broad presentation in a manner that also includessome necessary background to allow a sophisticated but otherwise not specialized studentor researcher to dive into this topic.

Page 8: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure
Page 9: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 1

Dynamical Systems,Ensembles, and TransferOperators

1.1 Ergodic PreambleIn this chapter, we present the heuristic arguments leading to the Frobenius-Perron operator,which we will restate with more mathematical rigor in the next chapter. This chapter ismeant to serve as a motivating preamble, leading to the technical language in the nextchapter. As such, this material is meant to serve as a quick start guide so that either the moredetailed discussion can be followed with more motivation, or even as enough backgroundso that the techniques in subsequent chapters can be understood without necessarily readingall of the mathematical theory in the middle chapters.

In terms of practical application, the field of measurable dynamics has been hidden ina forest of formal language of pure mathematics that may seem impenetrable to the appliedscientist. This language may be quite necessary for mathematical proof of the methodswithin the field of ergodic theory. However, the proofs often require restricting the rangeof problems quite dramatically, whereas, the utility may extend quite further. In reality, thebasic tools one needs to begin practice of measurable dynamics by transfer operator meth-ods are surprisingly simple, while still allowing useful studies of transport mechanisms ina wide array of real world dynamical systems. It is our primary goal in this writing to bringout the simplicity of the field for practicioners. We will attempt to highlight the languagenecessary to speak properly in terms necessary to prove convergence, invariance, steadystate, and several of the other issues rooted in the language of ergodic theory, but above all,we wish to leave a spine of simple techniques available to practicioners from outside thefield of mathematics. We hope this book will be useful to those experimentalists with realquestions coming from real data, and any students interested in such issues.

Our discussion here may be described as a contrast between the Lagrangian perspec-tive of following orbits of single initial conditions to the Eulerian perspective associatedwith the corresponding dynamical system of the transfer operator which describes the evo-lution of measurable of ensembles of initial conditions while focusing at a location. Thisleads to issues traditionally affiliated with ergodic theory, a field which has important prac-tical implications in the applied problems of transport study of interest here. Thus we hopethe reader will agree that both perspectives each allow important information to be derivedfrom a dynamical system. In particular, the transfer operator approach will allow us to

1

Page 10: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

discuss:

• Exploring global dynamics, and characterization of the global attractors.

• Estimating invariant manifolds

• Partitioning the phase space into invariant, almost invariant regions and coherent sets.

• Rates of transport between these partitioned regions.

• Decay of correlation.

• Associated information theoretic descriptions.

As we will discuss throughout this book, the question of transport can be boileddown to a question of walks in graphs, stochastic matrices, Markov chains, graph parti-tioning questions, and matrix analysis, together with Galerkin’s methods for discussingthe approximation. We leave this section with a picture in Fig. 1.1 which in some sense,highlights so many of the techniques in the book. We will refer back to this figure oftenthroughout this text. For now, we just note that the figure is an approximation of the ac-tion on the phase space of a Henon mapping as the action of a directed graph. The Henonmapping,

xn+1 = yn+1−ax2n ,

yn+1 = bxn, (1.1)

for parameter values a = 1.4, b = 0.3, is frequently used as a research tool and also as apedogical example of a smooth chaotic mapping in the plane. It is a diffeomorphism thathighlights many issues of chaos and chaotic attractors in more than one dimension. Suchmappings are not only interesting in their own right, but they offer allow a step towardunderstanding differential equations by Poincare’ section mappings.

1.2 The Ensemble PerspectiveThe dynamical systems point of view is generally Lagrangian, meaning, we focus on fol-lowing the fate of trajectories corresponding to the evolution of a single initial condition.Such is the perspective of an ODE, Eq. (2.1), as well as a map, Eq.(2.7). Here we contrastthe Lagrangian perspective of following single initial conditions to the Eulerian perspectiverooted in following measurable ensembles of initial conditions, based on the associated dy-namical system of transfer operators and leading to ergodic theory. We are most interestedhere in the transfer operator approach in that it may shed light on certain applied problemsto which we have already alluded and we will detail.

Example 1.1. (Following Initial Conditions, the Logistic Map)The logistic map,

xn+1 = L(xn)= 4xn(1− xn), (1.2)

is a model of resource limited growth in a population system. The logistic map is anextremely popular model of chaos, partly for pedogogical reasons of simplicity of analysis,and partly for historical reasons. In Fig. 1.2 we see the mapping and the time series it

Page 11: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.2. The Ensemble Perspective 3

Figure 1.1. Approximate action of a dynamical system on density looks like adirected graph: UlamÕs method is a form of GalerkinÕs method. In a sense, this couldbe the mantra of this book. (Above) we see an attractor of the Henon map, Eq. (1.1)partitioned by an abitrary graph, with the grid laid out according to a natural order of theplane phase space. (Below) The action of the dynamical system which moves (ensemblesof) initial conditions is better represented as a directed graph. The action shown here isfaithful (match the numbered boxes), but approximate since a Markov partition was notused, and a refinement would apparently be beneficial. [27]

Page 12: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

Figure 1.2. The Logistic map (Left) produces a time series shown (Right) for agiven initial condition x0 = 0.4.

produces for a specific initial condition, x0 = 0.4, where a time series is simply the functionof the output values with respect to time. An orbit is a sequence starting at a single initialcondition,

{x0, x1, x2, x3...} = {x0, L(x0), L2(x0), L3(x0), ...}, (1.3)

where,

xi = Li (x0)≡ L ◦ L ◦ ...◦ L(x0), which denotes the i th− composition. (1.4)

In this case, the orbit from x0 = 0.4 is the sequence {x0, x1, x2, ...} = {0.4,0.96,0.1536, ...}.The time series perspective illustrates the trajectory of a single initial condition in time,which as an orbit, runs “forever" and we are simply inspecting a finite segment. In thisperspective, we ask, how does a single initial state evolve? Perhaps there is a limit point?Perhaps there is a stable periodic orbit? Perhaps the orbit is unbounded? Or perhaps theorbit is chaotic?

At this stage it is useful to give some definition of a dynamical system. A math-ematically detailed definition of a dynamical system is given in Chapter 2 in Definitions2.1-2.3. Said plainly for now, a dynamical system is,

1. A state space (phase space), usually a manifold, together with

2. A notion of time, and

3. An evolution rule (often a continuous evolution rule) that brings forward states tonew states as time progresses.

Generally dynamical systems can be considered of two general types, continuous time asin flows (or semi flows) usually from differential equations, and discrete time mappings.For instance, the mapping xn+1 = L(xn) in Example 1.1, Eq. (1.2) is a discrete time map,L : [0,1]→ [0,1]. In this case, 1) the state space is the unit interval, [0,1], 2) time is

Page 13: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.2. The Ensemble Perspective 5

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

200

400

600

800

1000

1200

x

coun

t

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.5

1

1.5

2

2.5

3

3.5x 10

4

x

coun

t

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.5

1

1.5

2

2.5x 10

4

coun

t

x

Figure 1.3. Histograms depicting the evolution of many (N = 106) initial condi-tions under the logistic map. (Left) Initially {xi

0}i=1,...,N , are chosen uniformly by U (0,1) inthis experiment. (Middle), After one iterate, each initial conditions xi

0 moves to its iterate,x i

1 = L(xi0), and the full histogram, {xi

1}i=1,...,N is shown. (Right) The histogram of thesecond iterate, {xi

1}i=1,...,N is shown.

taken to be iteration number and it is discrete, 3) the mapping L(xn) = 4xn(1− xn) is theevolution rule which assigns new values to the old values. The phrase dynamical system isusually reserved to mean that the evolution rule is deterministic, meaning the same inputwill always yield the same output in a function mapping type relationship, where as thephrase stochastic dynamical system can be used to denote those systems with some kindof randomness in the behavior. Each of these will be discussed in the subsequent chapters.

Another perspective we pursue will be be to ask what happens to the evolution ofmany different initial conditions, the so-called ensemble of initial conditions. To illustratethis idea:

Example 1.2. (Following an Ensemble of Initial Conditions in the Logistic Map) Imag-ine that instead of following one initial condition, we choose N initial conditions, {xi

0}i=1,...,N ,(let N = 106, a million, for sake of specificity). Choosing those initial conditions by a ran-dom number generator, approximating uniform, U (0,1), we follow all of them - each andevery one. Now it would not be reasonable to plot a time series for all million states.The corresponding plot to Fig. 1.2(Right) would be too busy. We would only see a solidband. Instead, we accumulate the information as a histogram, an empirical representationof the probability density function. A histogram of N uniformly chosen initial conditionsis shown in Fig. 1.4(Left). Iterating each and every one of the initial conditions under thelogistic map yields, {xi

1}i=1,...,N = {L(xi0)}i=1,...,N , and so forth, through each iteration. Due

to the very large number of data points, we can only reasonably view the data statistically,as histograms, the profile of each evolve upon each successive iteration as shown in thesuccessive panels of Fig. 1.3.

There are central tenants of ergodic theory to be found in this example. The propertyof ergodic is defined explicitly in Secs. 3.5-3.5.1, but a main tenant is highlighted by theBirkhoff theorem describing coincidence of time averages and space averages Eq. 1.5. Twomajor questions that may be asked of this example are,

1. Will the profile of the histogram settle down to some form, or will it change forever?

2. Does the initial condition play a role, and in particular how does the specific initialcondition play a role in the answer to question #1?

Page 14: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.5

1

1.5

2

2.5x 10

4

coun

t

x0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

0

0.5

1

1.5

2

2.5x 10

4

coun

t

x0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

0

0.5

1

1.5

2

2.5x 10

4

coun

t

x

Figure 1.4. Following Fig. 1.3, histograms of {xi10}i=1,...,N , (Left), and

{xi25}i=1,...,N , (Middle), are shown. Arguably, a limit is apparent in the profile of the his-

togram. (Right) The histogram of the orbit of a single initial condition gives apparently thesame long term limit density. This is empirical evidence suggesting the ergodic hypothesis.

It is not always true that a dynamical system will have a long term steady state distribution,as approximated by the histogram; the specific dynamical system is relevant and for manydynamical systems, but not all, the initial condition may be relevant. When there is asteady state distribution, loosely said, we will discuss the issue of natural measure which isa sort of stable ergodic invariant measure, [181]. By invariant measure, we mean that theensemble of orbits may each move individually, but in such a way that their distributionnonetheless remains the same.

More generally, there is the notion of an invariant measure, (See Definition 3.3),where invariant measure and ergodic invariant measure are discussed further in Secs. 3.5-3.5.1. A transformation which has an invariant measure μ need not be ergodic, which is away of saying it favors just part of the phase space, or it may even be supported on just partof the phase space. By contrast, the density2 as illustrated here by the histogram showncovers the whole of [0,1], suggesting at least by empirical inspection,3 that the invariantdensity is absolutely continuous4.

Perhaps the greatest application of an ergodic invariant measure follows Birkhoff’sergodic theorem. Stated roughly, with respect to an ergodic T-invariant measure μ on ameasurable space (X ,A), where A is the sigma algebra of measurable sets, time averagesand spatial averages may be exchanged,

limn→∞

1

n

n∑i=1

f ◦ T i (x0)=∫

Xf (x)dμ(x), (1.5)

for μ-almost every initial condition. This is evidenced that a long orbit segment of a singleinitial condition {xj }106

j=1 yields essentially the same result as the long term ensemble, as

2We will often speak of measure and density interchangably, but in fact they are dual, but this is bestunderstood when there is a Radon-Nikodym derivative [200], in the case of a positive absolutely continuousmeasure-μ, dμ(x)= g(x)dx and g is the density function when it exists, which expression denotes, μ(B)=∫

B dμ(x) = ∫B g(x)dx and in the case the measurable functions are cells of a histogram’s partition, thisis descriptive of the histogram. In the case of continuous functions g this reminds us of the fundamentaltheorem of calculus.

3The result does in fact hold by arguments that the invariant measure is absolutely continuous to Lebesguemeasure, which will not present here, [53]

4A positive measure μ(x) is called absolutely continuous when it has a Radon-Nikodyn derivative preim-age [200] to Lebesgue measure. dμ(x)= g(x)dx .

Page 15: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.2. The Ensemble Perspective 7

seen in Figure 1.3.

Example 1.3. (Birkhoff Ergodic Theorem and Histograms)The statement that a histogram reveals the invariant measure for almost all initial

conditions can be sharpened by choosing the measurable function f in Eq. (1.5), as thecharacteristic (indicator) functions,

f (x)= χBi (x)=(

1 if x ∈ Bi0 else

)(1.6)

A histogram in these terms is an occupancy count of a data set “sprinkled" in a topologicalpartition, X =∪i (Bi ). Then in these terms, considering how many points of a sample orbit{xj }nj=1 occupy a cell Bi as part of building a histogram, in Eq. (1.5),

n∑j=1

f ◦ T j (x)=n∑

j=1

χBi (xj ). (1.7)

And Eq. (1.5) promises that we will almost never choose a bad initial condition but stillconverge toward the same occupancy for cell Bj . Likewise, repeating for each cell in thepartition produces a histogram such as in 1.3.

Example 1.4. (What Can Birkhoff Ergodic Theorem Say About Lyapunov Expo-nents?)

In Chapter 8, we will discuss Finite Time Lyapunov Exponents, which in brief arerelated to derivatives averaged along finite orbit segments, but multaplicatively, and howthe results vary depending on where in the phase space the initial condition is chosen, thetime length of the orbit segment, and how this information relates to transport. This is indramatic contrast to the traditional definition of a Lyapunov exponents, which are almostthe same quantity, but averaged along an infinite orbit. In other words, if we choose ameasuring function

f (x)= ln |T ′(x)| 5, (1.8)

then μ-almost every initial condition again will give the same result. Perhaps this “usual"way of thinking of orbits as infinitely long, and Lyapunov exponents as limit averages withthe Birkhoff theorem stating that almost every starting point is the same prevented the dis-covery of the brilliant-for-its-simplicity but powerful idea of FTLEs which are intrinsicallyspatially dependent due to the finite time aspect.

To state more clearly the question of how important is the initial condition, in brief,the answer is almost not at all. In this sense, almost every initial condition stated in themeasure theoretic sense is “typical". This means that with probability one, we will wechoose an initial condition which will behave as the ergodic case. To emphasize the state-ment of this rarity in perspective, in the same sense we may say that if we choose a numberrandomly from the unit interval, with probability zero, the number will be rational, and

5Specializing to a one dimensional setting so that we do not need to discuss issues related to the Jacobianderivative matrices and diagonalizability at this early part of the book. In this case the Birkhoff theoremdescribes the Lyapunov exponents as discussed here. The more general scenario in more than one dimensionrequires the Oseledet’s Multiplicative Ergodic Theorem to handle products along orbits, as discussed inSec.!8.2 as contrasted to the one-dimensional scenario in Sec. 8.1.

Page 16: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

likewise with probability one, the number will be irrational. Of course this does not meanit is impossible to choose a rational number, just that the Lebesgue measure of the rationalsis zero. By contrast, the situation is opposite when performing the random selection on acomputer. The number will always be rational because: 1) The random number generatoris just a model of the uniform random variable, as must be all algorithms descriptive of apseudo random number generator [322], 2) The computer can only represent rational num-bers, and in fact, a finite number of those. Nonetheless, when selecting a pseudo randomnumber, it will be ergodic-typical in the sense above for a “typical" dynamical system.

1.3 Evolution of EnsemblesPerhaps a paradoxical fact, but a fact central to the analysis of this book, is that whilea chaotic dynamical system6 may be nonlinear causing particular difficulty to predict thefate of individual orbits, the evolution of density is an associated linear7 dynamical systemwhich turns out to be especially straight forward to predict. That is, the dynamical system,

f : X → X , (1.9)

moves initial conditions, whereas, there is an associated linear dynamical system,

Pf : L1(X)→ L1(X), (1.10)

which is descriptive of the evolution of densities of ensembles of initial conditions. Theoperator, Pf , is called the Frobenius-Perron operator. Initially, we will specialize for sim-plicity of presentation to the logistic map as follows. The general theory will be saved forthe next Chaper 2.

The evolution of density follows a principle of mass conservation: emsembles ofinitial conditions evolve forward in time, and no individual orbits are lost. In terms of den-sities, it must be assumed that the transformation is nonsingular in that this will guaranteethat densities map to densities.8 If there are N initial conditions {x0,i}Ni=1, then in general,they may be distributed in X according to some initial distribution ρ0(x), for which wewrite, x0 ∼ ρ0(x). The question of evolution of density is as follows. After one iteration byf , each x0,i moves to x1,i = f (x0,i ) for each i . Generally, if we investigate the distributionof the points in their new positions, we must allow that the distribution of them may be dif-ferent than their initial configuration. If the actual new configuration of {x1,i}Ni=1 distributesaccording to some new density ρ1(x), then the question becomes to find ρ1(x) given ρ0(x).And likewise we can look for the orbit of distributions,

{ρ0(x),ρ1(x),ρ2(x), ...}. (1.11)

From the principle conservation of initial conditions follows a discrete continuityequation. ∫

Bρ1(x)dx =

∫f −1(B)

ρ0(x)dx ,∀B ∈A, (1.12)

6A dynamical system is defined to be chaotic if it displays sensitive dependence to initial conditions anda dense orbit [97, 12, 326, 223], or according to [2] an orbit is chaotic if it is bounded, not asymptoticallyperiodic, and has a positive Lyapunov exponent.

7A dynamical system T : X → X is linear if for any x1, x2 ∈ X , and a,b ∈ R, T (ax1+bx2)= aT (x1)+bT (x2), and otherwise the dynamical system is nonlinear.

8More precisely we wish that absolutely continuous densities map to absolutely continuous densities.

Page 17: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.3. Evolution of Ensembles 9

from which will follow the dynamical system,

Pf : L1(X)→ L1(X),

ρ0(x) �→ ρ1(x)= Pf [ρ0](x). (1.13)

and the assignment of a new density by the operator Pf is interpreted at each point x .This continuity equation may be interpreted as follows. Formally, over a measure space(X ,A,μ), B ∈A is any one of the measurable subsets. For simplicity of discussion, we mayinterpret the B’s to be any one or collection of the cells used in describing the histogramssuch as shown in Figures 1.3, or 1.7. Then ρ0(x) is an initial density descriptive of anensemble, such as the approximation depicted in Figure 1.4(Left). Asking the question,where do each and every one of the initial conditions go under the action of f , we arebetter suited to ask where orbits distributed by ρ1 came from after one iteration of f . Thepreimage version Eq. (1.12) and,∫

f (B)ρ1(x ′)dx ′ =

∫Bρ0(x ′)dx ′, (1.14)

would give the same result as Eq. (1.13) if the mapping were piecewise smooth and one-one, but many examples including the logistic map are not one-one. See as shown inFigure 1.5.

The continuity equation Eq. (1.12) is well stated for any measure space, (X ,A,μ),but assuming that X is an interval X = [a,b] and μ is Lebesgue measure, then we maywrite, ∫ x

aPf ρ(x ′)dx ′ =

∫f −1([a,x])

ρ(x ′)dx ′,∀x ∈ [a,b]. (1.15)

thus representing those B which are intervals [a, x]. For a noninvertible f , f −1 denotesthe union of all the preimages. Differentiating both sides of the equation, and assumingdifferentiability, then the fundamental theorem of calculus gives,

Pf ρ(x)= d

dx

∫f −1([a,x])

ρ(x ′)dx ′,∀x ∈ [a,b]. (1.16)

Further assuming that f is invertible and differentiable allows application of the fundamen-tal theorem of calculus and the chain rule to the right hand side of the equation,

Pf ρ(x)= d

dx

∫ f −1(x)

f −1(a)ρ(x ′)dx ′ = ρ( f −1(x))

d

dx( f −1(x))= ρ( f −1(x))

| f ′( f −1(x))| ,∀x ∈ [a,b].

(1.17)However, generally since f may not be invertible, the integral derivation is applied overeach preimage, resulting in a commonly presented form of the Frobenius-Perron operatorfor deterministic evolution of density in maps,

Pf [ρ](x)=∑

y:x= f (y)

ρ(y)

| f ′(y)| . (1.18)

The nature of this expression is a functional equation for the unknown density func-tion ρ(x). Questions of existence, and uniqueness of solutions are related to the funda-mental unique ergodicity question in ergodic theory. That is, can one find a special distri-bution function, which should be stated as a centrally important principle in the theory ofFrobenius-Perron operators,

Page 18: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

10 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

An invariant density is a fixed “point" of the Frobenius-Perron operator: Functionallythis can be stated,

ρ∗(x)= Pf [ρ∗](x). (1.19)

One note, is that we say, “an" invariant density rather than the invariant density asthere can be many and even infinitely many, but usually we are interested in the “domi-nant" invariant measure, or other information related to dominant behaviors such as almostinvariant sets. Further, is this fixed density (globally) stable? This is the critical question:will general (nearby or all?) ensemble distributions settle to some unique density profile?ρi (x)→ ρ∗(x) as n→∞? In the following example, we offer a geometric interpretation ofthe form of the Frobenius-Perron operator and its relationship to unique ergodicity. Moreon this principle can be found discussed in Sec. 3.4.

Example 1.5. (Frobenius-Perron Operator of the Logistic Map) The (usually two)preimage(s) of the logistic map Eq. (1.2) at each point, may be written,

L−1± (x)= 1±√1− x

2. (1.20)

Therefore, the Frobenius-Perron operator Eq. (1.18) specializes to,

Pf [ρ](x)= 1

4√

1− x(ρ(

1−√1− x

2)+ρ(

1+√1− x

2)). (1.21)

This functional equation can be interpreted pictorially as in Figure 1.5; the collective en-semble at cell B comes from those initial conditions at the two preimages f −1± (B) shown.The pre-image of the set may be seen as the cobweb diagram shows, scaled roughly as theinverse of the derivative at the preimages. Roughly, the scaling occurs almost as if we werewatching ray optics where the preimages f −1(B) focus on B through mirrors by the actionof the map, placed at B by the map, but scaled as if focused by the inverse of the derivativebecause the ensemble of initial conditions at f −1(B) shuttle into B .

It is a simple matter of substitution, and therefore application of trigonometric iden-tities, to check that the function,

ρ(x)= 1

π√

x(1− x), (1.22)

is a fixed point of the operator, Eq. (1.21). This guess and check method is a valid way forvalidating an invariant density. However, how does one make the guess? Comparison of theexperimental numerical histograms validate this density, comparing Eq. (1.22) to Figures1.3 and 1.4. By comparison to a simpler system, where the invariant density is easy toguess, the invariant density of this Logistic map is straight forward to derive.

Example 1.6. (Frobenius-Perron Operator of the Tent Map)The Tent map serves as a simple example to derive invariant density and for compar-

ison to the Logistic map.

xn+1 = T (xn)= 2(1−2|x− 1

2|), (1.23)

may also be viewed as a dynamical system on the unit interval, T : [0,1]→ [0,1], shown inFigure 1.6(Left), and a “typical" time series is shown in Figure 1.6(Middle). Repeating the

Page 19: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.3. Evolution of Ensembles 11

experiment of the evolution of an ensemble of initial conditions as was done for the Logisticmap, Figs. 1.3 and 1.4, yields the histogram Fig. 1.6(Right). Apparently from the empiricalexperiment, the uniform density, U (0,1), is invariant and this is straight forward to validateanalytically by checking that the Frobenius-Perron operator. Eq. (1.18) specializes to,

PT [ν](x)= 1

4(ν(

x

4)+ ν(1− x

4)). (1.24)

Further, invariance Eq. (1.19) has a solution,

ν(x)= 1. (1.25)

The well-known change of variables between the dynamics of the fully developedchaotic9 tent map (slope a = 2) and the fully developed logistic map (r = 4) is through thechange of variables,

h(x)= 1

2(1− cos(πx)), (1.26)

which is formally an example of a conjugacy in dynamical systems.The fundamental equivalence relationship in the field of dynamical system, compar-

ing two dynamical systems, g1 : X → X and g2 : Y → Y is a conjugacy,

Definition 1.1. (Conjugacy) Two dynamical systems,

g1 : X → X , and g2 : Y → Y , (1.27)

are a conjugate if there exists a function (a change of variables),

h : X → Y , (1.28)

such that h commutes (a pointwise functional requirement),

h ◦ g1(x)= g2 ◦ h(x). (1.29)

often written as a commuting diagram,

Xg1−−−−→ X

h

⏐⏐� ⏐⏐�h .

Yg2−−−−→ Y

(1.30)

and h is a homeomorphism between the two spaces X and Y . The function h is a homeo-morphism if,

• h is one-one,

• h is onto,

9Fully developed chaos as used here refers to the fact that as the parameter (a or r for the tent map orlogistic map for example) is varied, the corresponding symbol dynamics becomes complete in the sense thatthe corresponding symbol dynamics is a full shift, meaning the corresponding grammar has no restrictions.The symbol dynamics theory will be discussed in detail in Chapter 6. A different definition of can be foundin [225], which differs from our use here largely by including the notion that the chaotic set should denselyfill the interval.

Page 20: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

12 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

• h is continuous,

• h−1 is continuous.

Change of variables is a basic method in mathematical sciences since it is fair gameto change from a coordinate system where the problem may be hard (say Cartesian coordi-nates) to a coordinate system (say spherical coordinates) where the problem may be easierin some sense, often the goal being to decouple variables. The most basic requirementbeing that in the new coordinate system, solutions are neither created nor destroyed, as theabove definition allows. The principle behind defining a good coordinate transformation tobe a homeomorphism is that the two dynamical systems should take place in topologicallyequivalent phase spaces. Further, solutions should correspond to solutions with the samebehavior in a continuous manner, and this is further covered by the pointwise commutingprinciple h ◦g1(x)= g2◦h(x). By contrast for example, without requiring that h is one-one,two solutions may come from one, and so forth.

Returning to comparing the Logistic and the tent maps, it can be checked that thefunction,

h(x)= 1

2(1− cos(πx)), (1.31)

shown in Fig. 1.8(Upper Right) is a conjugacy between g1 as the Logistic map of Eq. (1.2),(with the parameter value 4) and g2 as the Tent map, Eq. (3.56), (with parameter value 2),with X = Y = [0,1].

A graphical way to represent that commuter function (a function simply satisfyingEq. (1.30) whether or not that function may be a homeomorphism [296]) is by what wecall a quadweb diagram, as illustrated in Figure 1.8). A quadweb is a direct and point-wise graphical representation of the commuting diagram. In [296], we discuss further howeven representing the commuting equation even when two systems may not be conjugate(and therefore the commuter function is not a homeomorphism) has interesting relevanceto relating dynamical systems. Here we will simply note that the quadweb illustrates that aconjugacy is a pointwise relationship. Of course, we named a quadweb as such since it isa direct play on the better known term “cobweb" diagram. When further h is a homeomor-phism, then the two maps compared are conjugate.

Inspecting Eq. (1.31), we see that the function is not simply continuous, it is differ-entiable. As it turns out, this most popular example of a conjugacy is atypical in the sensethat it is stronger than required. In fact, it is a diffeomorphism.

Definition 1.2. (Diffeomorphism) A diffeomorphism is a homeomorphism h which is bi-differentiable (meaning h and h−1 are differentiable), and when stated that two dynamicalsystems are diffeomorphic, there is is a conjugacy which is bi-differentiable.

Conjugacy is an equivalence relationship with many conserved quantities betweendynamical systems, including notably, topological entropy. Diffeomorphism is a strongerequivalence relationship which conserves quantities such as metric entropy and also Lya-punov exponents. Interestingly despite the aytpical nature of diffeomorphism, in the senseof genericity implying that most systems if conjugate have nondifferentiable conjugacies,the sole explicit example used for introduction in most textbooks is a diffeomorphism,Eq. (1.31). A nondifferentiable conjugacy of two maps in the interval will be a Lebesguesingular function [296] meaning it will be differentiable almost everywhere, but wherever

Page 21: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.3. Evolution of Ensembles 13

Figure 1.5. Cobweb of density. Note how infinitesimal density segments of Bgrow or shrink inversely proportionally to the derivative at the pre-image, as prescribed byEq. (1.18).

Figure 1.6. The Tent map (Left), Eq. (3.56), a sample time series (Middle) anda histogram of a sample ensemble (Right). This figure mirrors Figure 1.3 shown for theLogistic map. Apparently here, the tent map suggests an invariant density which is uniform,U (0,1).

it is differentiable, the derivative is zero - but nonetheless the function is monotone nonde-creasing in order to be one-one. These are topologically exotic in the sense that they are abit more like a devil’s staircase function [283, 339] than they are like a cosine function.

Most relevant for our problem here, being comparison between invariant densitiesof the logistic map, and the tent map, for which we require the differentiability of theconjugacy. Thus h must further be a diffeomorphism to execute the change of density. We

Page 22: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

14 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

require the infinitesimal comparison,10

ρ(x)dx = ν(y)dy. (1.32)

from which follows

ρ(x)= dy

dx= 1

π√

x(1− x), (1.33)

This result is in fact the fixed density already noted, in Eq. (1.22) and which agrees withFigures 1.3 and 1.4.

Finally in this section, we illustrate the ensemble perpective of invariant density foran example of a mapping whose phase space is more than an interval. The Henon mappingfrom Eq. (1.1). This is a diffeomorphism of the plane,

H : R2→ R2. (1.34)

As such, a density is a positive function over the phase space,

ρ : R2→ R+. (1.35)

In Fig. 1.1 we illustrated both the chaotic attractor as well as the action of this mappingwhich is approximately a directed graph. The resulting invariant density, of a long timesettling an ensemble of initial conditions, or alternatively of a long time behavior of onetypical orbit, is illustrated in Fig. 1.7. As we will describe further in the next chapter, thisinvariant density derived here by a histogram of a long orbit, may also be found as thedominant eigenvector of transition matrix of the graph shown in Fig. 1.1; this is the Ulamconjecture [319].

1.4 Various Useful Representations and Invariant Densityof a Differential Equation

An extremely popular differential equation considered often and early in the presentationof chaos in nonlinear differential equations, and historically central in the development ofthe theory, is the Duffing equation,

x+ax− x+ x3 = b sinωt . (1.36)

This equation in its most basic physical realization describes the situation of a massless ballbearing rolling in a double-welled potential,

P(x)=−x2+ 1

4x4, (1.37)

which is then sinusoidally forced, as depicted in Figure 1.9.11 This is a standard differentialequation in the pedagogy of dynamical systems. We take this problem as example to presentthe various presentations in representing the dynamics of a flow, including:

10This is equation in the simplest problems is simply called “u-change of variables" in elementary calculusbooks, but is a form of the Radon-Nikodym derivative theorem in more general settings, [200].

11The gradient system case, where the autonomous part can be written − ∂P∂x occurs when the viscuous

friction part is zero, a = 0.

Page 23: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.4. Useful Representations Density of an ODE 15

Figure 1.7. Henon Map histogram approximating the invariant density. Notice theirregular nature typical of the densities of such chaotic attractors which are often suspectedto not be absolutely continuous.

• Time-series, Fig. 1.10,

• Phase portrait, Fig. 1.11

• Poincare’ map, Figs. 1.12-1.13,

• Attractor, also seen in Fig. 1.12

• Invariant Density, Fig. 1.14.

Written in a convenient form as a system of first order equations, with the substitu-tion,

y ≡ x , (1.38)

gives a nonautonomous12 two dimensional equation,

x = y

y =−ay− x− x3+b cosωt . (1.39)

12A autonomous differential equation can be written x = F(x) without explicitly including t in the righthand side of the equation, and otherwise the differential equation is nonautonomous when it must be writtenx = f (x , t).

Page 24: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

16 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

Figure 1.8. A quadweb is a graphical way to pointwise represent the commutingdiagram eqcommdiagramapp. When further h is a homeomorphism, then the two mapscompared are conjugate. Here shown is the conjugacy h(x) = 1

2 (1− cos(πx)), changingvariables between the full tent map and the full logistic map, Eq. (1.31).

Figure 1.9. Duffing double well potential, Eq. (1.37) corresponding to the Duffingoscillator, (a = 0 case). Unforced, the gradient flow can be illustrated as a massless ballbearing in the double well as shown. Further forcing with a sinusoidal term can be thoughof as a ball bearing moving in the well, but the floor is oscillating, causing the ball tosometimes jump from one of the two wells to the other.

Page 25: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.4. Useful Representations Density of an ODE 17

Figure 1.10. A Duffing oscillator can give rise to a chaotic time series, here shownfor both x(t) and y(t) solutions from Eq. (1.39), with a = 0.02, b = 3 and ω = 1.

As a time-series of measured position x(t) and velocity, y(t) of these equations,with a = 0.02, b = 3 and ω = 1, we observe a signature chaotic oscillation as seen inFigure 1.10. This time-series of an apparently erratic oscillation nonetheless comes fromthe deterministic evolution of the ODE Eq. (1.36). This is simply a plot of the variables xor y as a function of time t .

A phase portrait in phase space however suppresses the time variable. Instead thet serves as a parameter which for representation of a solution curve in parametric form,(x(t), y(t))R2 as seen in Figure 1.11.

Augmenting with an extra time variable, τ (t) = t , from which dτdt = τ = 1 gives the

autonomous three dimensional equations of this flow,

x = y

y =−ay− x− x3+b sinωτ

τ = 1. (1.40)

This form of the dynamical system allows us to represent solutions in a phase space,(x(t), y(t),τ (t)) ∈ R3 for each t . In this representation, the time variable is not suppressedas we view the solution curves, (x(t), y(t),τ (t)). Thus generally one can represent a nonau-tonomous differential equation as an autonomous differential equation by embedding inlarger phase space.

Page 26: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

18 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

Figure 1.11. The Duffing equations, Eq. (1.39) nonautonomous phase space is(x(t), y(t)) ∈ R2, with a = 0.02, b = 3 and ω = 1.

A convenient way to study topological and also measurable properties of the dynam-ical system presented by a flow is to produce a discrete time mapping by the Poincare’section method to produce a Poincare’ mapping. That is, a codimension-1 “surface" isplaced transverse to the flow so that (almost every) solution will pierce it, and then ratherthan recording every point on the flow, it is sufficient to record the values at the instants ofpiercing.

In the case of the Duffing oscillator, a suitable Poincare surface is a special casecalled a “stroboscopic" section, by ωτ = 2πk for k ∈Z. The brilliance of Poincare’s “trick"allows the ordered discrete values (x(tk), y(tk), tk = 2πk

ω, or rather simply we write (xk , yk),

to represent the flow on its attractor. In this manner, Fig. 1.12 replaces Fig. 1.11, and inmany ways this representation as a discrete time mapping,

(xk+1, yk+1)= F(xk , yk), (1.41)

is easier to analyze, or at least there exists a great deal of new tools otherwise not availableto the ODEs perspective alone. For the sake of classification, when the right hand side ofthe differential equation is in the form of an autonomous vector field, as we represented in

Page 27: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.4. Useful Representations Density of an ODE 19

the case of Eq. (1.40) we write specifically,

G : R3→R3,

G(x , y,τ ) = < y,−ay− x− x3+b,1 > . (1.42)

Then simply, let,z = (x , y,τ ) and z = G(z), (1.43)

which is a general form for Eq. (1.40).If the vector field G is Lipschitz,13,14, then it is known that there is continuous de-

pendence both with respect to initial conditions and also with respect to parameters asproven through Gronwall’s inequality [174, 262]. It follows that the Poincare’ mapping Fin Eq. (1.41) must be a continuous function in two dimensions, F :R2→R2 correspondingto a two-dimensional dynamical system in its own right. If further, G ∈ C2(R3), then F isa diffeomorphism which brings with it a great deal of tools from the field of differentiabledynamical systems such as transport study by stable and unstable manifold analysis.

In fact, the Duffing oscillator is an excellent example for presentation of the Poincare’mapping method. There exists a two dimensional Duffing mapping - in this case a diffeo-morphism. Such is common with differential equations arising from physical and especiallymechanical problems. All this said the common scenario is that we cannot explicitly rep-resent the function F : R2 → R2. In Figure 1.12 we show the attractor corresponding tothe Duffing oscillator on the left, and a charicature of the stroboscopic method whose flightproduces F on the right. In practice, a computer is required for all examples we have expe-rience to numerically integrate chaotic differential equations, and thus further to estimatethe mapping F and a finite number of sample points.

Just as was presented in the case of the logistic map, where a histogram as in Fig-ures 1.3, 1.4 and 1.6 gives further information regarding the long term fate of ensemblesof initial conditions, we can make the same study in the case of differential equations byusing the Poincare’ mapping representation. The question is the same, but posed in termsof the Poincare’ mapping. How do ensembles of initial conditions evolve under the discretemapping, (xk+1, yk+1)= F(xk , yk), as represented by a histogram over R2? See Fig. 1.14.The result of an experiment of a numerical simulation of one initial condition is expectedto represent the same as the fate of many samples, for almost all initial conditions. That istrue if one believes the system is ergodic, and thus follows the Birkhoff ergodic theoremEq. (1.5). See also the discussion regarding natural measure near Eq. (3.78). The idea isthat the same long term averages sampled in the histogram boxes are almost always the

13G : Rn→Rn is Lipschitz in a region ⊂R

n if there exists a constant L > 0 ‖G(z)−G(z)‖ ≤ L‖z− z‖for all z, z ∈ ; the Lipschitz property can be considered as a form of stronger continuity (often calledLipschitz continuity) but not quite as strong as differentiability which allows for the difference quotient limitz→ z to maintain the constant L.

14Perhaps the most standard existence and uniqueness theorem used in ordinary differential equationstheory is the Picard-Lindelof theorem: an initial value problem z = G(t , z), z(t0)= z0 has a unique solutionz(t) at least for time t ∈ [t0−ε, t0+ε] for some time range ε > 0 if G is Lipschitz in z and continuous in t inan open neighborhood containing (t0, z(t0)). The standard proof relies on Picard iteration of an integral formof the ODE, z(t)= z(t0)= ∫ t

t0G(s, z(s))ds which with the Lipschitz condition can be proven to converge in

a Banach space by the contraction mapping theorem, [262]. Existence and uniqueness is a critical startingcondition to discuss an ODE as a dynamical system, meaning one initial condition does indeed lead to oneoutcome which continues (at least for awhile), and correspondingly often the analysis herein may be as adiscrete time mapping by Poincare’ section.

Page 28: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

20 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

Figure 1.12. A Poincare-stroboscopic mapping representation of the Duffing os-cillator. The discrete time mapping in R2 is derived by recording (x , y) each time that(x(t), y(t),τ (t))∈�, the Poincare’ surface in this case, � = {(x , y,τ ) : (x ,τ ) ∈R2,τ ∈ 2πk

ωas charicatured in Fig. 1.13. Eq. (1.39), with a = 0.02, b = 3 and ω = 1. [30]

same with respect to choosing initial conditions. Making these statements of ergodicityinto mathematically rigorous statements turn out to be notoriously difficult even for themost famous chaotic attractors from the most favorite differential equations from physicsand mathematics. This mathematical intricacy is certainly beyond the scope of this bookand we refer to Lai Sai Young for a good starting point, [336]. This is true despite theapparent ease by which we can simulate and seemingly confirm ergodicity of a map or dif-ferential equation through computer simulations. Related questions include existence of anatural measure, presence of uniform hyperbolicity, representation by symbolic dynamicsto name the few are all associated questions.

In subsequent chapters, we will present the theory of transfer operator methods tointerpret invariant density, mechanism and almost invariant sets leading to steady states,almost steady states, and coherent structures partitioning the phase space. Further, we willshow how the action of the mapping by a transfer operator may be approximated by agraph action generated by a stochastic matrix, through the now classic Ulam’s method, andfurther, graph partitioning methods approximate and can be pulled back to present relevantstructures in the phase space of the dynamical system.

Page 29: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.4. Useful Representations Density of an ODE 21

Figure 1.13. The Poincare’ mapping shown in Fig. 1.12 are surfaces � ={(x , y,τ ) : (x ,τ ) ∈ R2,τ ∈ 2πk

ωcharicatured as the flow from Eq. (1.39), with a = 0.02,

b = 3 and ω = 1, pierces the surfaces.

Finally in this section for sake of contrast, we may consider the the histogram result-ing directly from following a single orbit of the flow as shown in the phase space but withoutresorting to the Poincare’ mapping. That is, it is the approximation of relative density fromthe invariant measure of the attractor of the flow in the phase space. See Figure 1.15. Thisis in contrast to the density of the more commonly used and perhaps more useful Poincare’mapping as shown in Fig. 1.14. As was seen for the Henon map in Fig. 1.1, consider-ing the action of the mapping on a discrete grid leads to a directed graph approximationof the action of the map. We will see that this action becomes a discrete approximationof the Frobenius-Perron operator and as such it will serve as a useful computational tool.

Page 30: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

22 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

Figure 1.14. Duffing Density of the Poincare’-stroboscopic mapping method, es-timated by simulation of a single initial condition evolved over 100,000 mapping periods,and density approximated by a histogram The density is shown both as block heights above,and as a color intensity map below. Compare to the attractor shown in Fig. 1.12.

Discussion of convergence with respect to refinement we call the Ulam-Galerkin methodto be discussed in subsequent chapters. Also a major topic of this book will be the manyalgorithmic uses for this presentation as a method for transport analysis. There are a greatnumber of computational methods that we will see become available when considering thisdirected graph structures. We will be discussing these methods, as well as the correspond-ing questions of convergence and representation in subsequent chapters.

A major strength of this computational perspective for global analysis is the possi-bility to analyze systems known empirically only through data. As a case study and animportant aplication [37], consider the spreading of oil following the 2010 Gulf of MexicoDeep Water Horizon oil spill disaster. On April 20, 2010, an oil well cap explosion belowthe Deepwater Horizon, an off-shore oil rig in the Gulf of Mexico, started the worst human-caused submarine oil spill ever. Though an historic tragedy for the marine ecosystem, theunprecedented monitoring of the spill in real time by satellites and increased modeling ofthe natural oceanic flows has provided a wealth of data, allowing analysis of the flow dy-namics governing the spread of the oil. In [37] we studied two computational analysesdescribing the mixing, mass transport, and flow dynamics related to oil dispersion in theGulf of Mexico over the first 100 days of the spill. Both Transfer operator methods wereused to determine the spatial partitioning of regions of homogeneous dynamics into almost-invariant sets, and Finite Time Lyapunov Exponents were used to compute pseudobarriers

Page 31: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.4. Useful Representations Density of an ODE 23

Figure 1.15. The attractor of the Duffing oscillator flow in its phase space(x(t), y(t)) has relative density approximated by a histogram. Contrasts to the Poincare’mapping presentation shown of the same orbit segments in Figure 1.14.

to the mixing of the oil between these regions. The two methods give complementaryresults as we will see in subsequent chapters. As we will present from several different per-spectives, this data makes a useful presentation for generating a comprehensive descriptionof the oil flow dynamics over time, and as such, for discussion of the utility of many of themethods described herein.

Basic questions in oceanic systems concern large-scale and local flow dynamicswhich naturally partition the seascape into distinct regions. Following the initial explo-sion beneath the Deepwater Horizon drilling rig on April 20, 2010, oil continued to spillinto the Gulf of Mexico from the resulting fissure in the well head on the sea foor. Spillrates have been estimated at 53,000 barrels per day by the time the leak was controlled bythe“cap"-fix three months later. It is estimated that approximately 4.9 million barrels, or185 million gallons of crude oil flowed into the Gulf of Mexico, making it the largest-eversubmarine oil spill. The regional damage to marine ecology was extensive, but impactswere seen on much larger scales as well, as some oil seeped into the Gulf Stream, whichtransported the oil around Florida and into the Atlantic Ocean. Initially, the amount of oilthat would disperse into the Atlantic was overestimated, because a prominent dynamicalstructure arose in the gulf early in the summer preventing oil from entering the Gulf Stream.The importance of computational tools for analyzing the transport mechanisms governing

Page 32: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

24 Chapter 1. Dynamical Systems, Ensembles, and Transfer Operators

Figure 1.16. Toward the Ulam-Galerkin Method in the Duffing Oscillator in aPoincare’ mapping representation as explained in Chapter 4. Covering the attractor withrectangles, three are highlighted by colors magenta, red, and green. Under the Poincare’mapping in Eq. (1.41), F(Rectangle) yields the distorted images of the rectangles shown.Each rectangle is mapped correspondingly to the same colored regions. Considering therelative measures of how much of each of these rectangles maps across other rectanglesleads to a discrete approximation of the Frobenius-Perron operator, akin to the graph pre-sentation of the Henon map’s directed graph shown in Fig. 1.1. For generality, compare thisfigure to a similar presentation in the Gulf of Mexico in Fig. 4.3 allowing global analysisof a practical system known only through data.

the advective spread of the oil may therefore be considered self evident in this problem.Fig. 1.17. shows a satellite image of the Gulf of Mexico off the coast of Louisiana on May24, 2010. The oil is clearly visible in white in the center of the image, and the spread of theoil can already be seen, just over a month after the initial explosion. During the early daysof the spill the Gulf Stream was draining oil out of the gulf and, eventually, into the At-lantic. This spread was substantially tempered later in the summer, due to the developmentof a natural eddy in the central Gulf of Mexico, which acted as a barrier to transport.

The form of the data is an empirical nonautonomous vector field f (x , t), x ∈R2, herederived from an ocean modeling source called the HYCOM model [182]. One time shot

Page 33: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

1.4. Useful Representations Density of an ODE 25

from July 27, 2010 is shown in Fig. 1.18. Toward transfer operator methods, in Fig. 1.19 weillustrate time evolution of several rectangle boxes suggesting the Ulam-Galerkin methodto come in Chapter 4, and analogous to what was already shown in this chapter in Figs. 1.1and 1.16, and in practice a finer grid covering would be used rather than this coarse cover-ing used for illustrative purposes. The kind of partition result we may expect using thesedirected graph representations of the Frobenius-Perron transfer operator can be seen inFig. 1.20. Discussion of almost-invariant sets, coherent sets and issues related to transportand measure based partitions in dynamical systems from Markov models are discussed indetail in Chapter 5. For now we can say that the prime direction leading to this computa-tional avenue are simple questions of asking,

• Where does the product (oil) go, at least in relatively short time scales,

• Where does the product not go, at least in relatively short time scales,

• Are there regions which stay together and barriers to transport between regions.

These are questions of transport and of partition of the space relative to which transportcan be discussed. Also related to partition is the boundary of partition for which there isa complementary method that has become useful. The theory of Finite-Time LyapunovExponents will be discussed in Chapter 8, along with highlighting both interpretations asbarriers to transport and also shortcomings for such interpretations. See for example anFTLE computation for the Fig. 1.18 Gulf of Mexico data in Fig. 1.21. In the sequel chaptersthese computations and supporting theory will be discussed.

Page 34: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 1.17. Satellite view of the Gulf of Mexico near Louisiana during the oilspill disaster, May 24, 2010. The oil slick spread is clearly visible and large. The image,taken by NASA’s Terra satellite, is in the public domain.

Page 35: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 1.18. Vector field describing surface flow in the Gulf of Mexico on May24, 2010, computed using the HYCOM model [182]. Note the coherence of the Gulf Streamat this time. Oil spilling from south of Louisiana could flow directly into the Gulf Streamand out towards the Atlantic. Horizontal and vertical units are degrees longitude (negativeindicates west longitude) and degrees latitude (positive indicates north latitude), respec-tively.

Page 36: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 1.19. Evolution of rectangles of initial conditions illustrates the actionof the Frobenius-Perron transfer operator in the Gulf as estimated on a coarse grid byUlam-Galerkin’s method. Further discussion of such methods can be found in Chapter 4.Compare to Figs. 1.1 and 1.16.

Page 37: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 1.20. Partition of the Gulf of Mexico using transfer operator approach tobe discussed in Chapter 5. Regions in red correspond to coherent sets, i.e. areas into andout of which little transport occurs. [37]

Page 38: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 1.21. Finite-Time Lyapunov Exponents in the Gulf of Mexico help withunderstanding of transport mechanisms, as discussed in Chapter 8 including both inter-pretations and limitations. Roughly stated the redder regions represent slow to almost notransport occurs across these ridges. [37]

Page 39: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 2

Dynamical SystemsTerminology and Definitions

Some standard dynamical systems terminology and concepts will be useful as reviewed inthis chapter, which we review here those elements for use in the sequel. Some general andpopular references for the following materials include [156, 279, 329, 228]. The materialin this chapter uses a bit more formal presentation than the quick start guide of the previouschapter, but will be brief in that only necessary background will be given to support thecentral topic of the book, in an attempt to not overly repeat several of the many excellentgeneral dynamical systems textbooks.

In its most general form, dynamical systems can be described as the study of groupaction on manifolds, and perhaps including differentiable structure. Throughout the major-ity of this work, we will not require this most general perspective, but some initial descrip-tion can be helpful. Two major threads in the field of dynamical systems involve the studyof,

• Topological properties

• Measurable properties

of either groups or semi-groups, [228, 281] therein. Discussion of topological propertieswill be addressed in the Chapter 6 regarding symbolic dynamics and related to informa-tion theoretic aspects in Chapter 9. Discussion of measurable properties is more closelyallied with the specific theme here which involves transport mechanism of the fate of en-sembles of initial conditions. In a basic sense, these two perspectives are closely relatedas can be roughly understood by inspecting the Henon map example depicted in Fig.1.1;the figure depicts the action of the map on the phase space as approximated by a directedgraph. In topological dynamics, we are not concerned with relative scale of the sets beingmapped. As such, the approximations by directed graphs have no weights. Thus followsunweighted graphs and adjacency matrices which generate them. On the other hand, mea-surable dynamics is concerned with relative weights of the sets, and so the directed graphapproximation must be a weighted graph with the weights along the edges describing eitherprobability or relative ratios of transitions. Correspondingly, the graphs are generated bystochastic matrices rather than adjacency matrices. As we will see, both of these perspec-tives have their place in the algorithmic study of applied measurable dynamical systems.

31

Page 40: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

32 Chapter 2. Dynamical Systems Terminology and Definitions

2.1 The Form of a Dynamical SystemHere, we shall write,

x = f (x , t), x(t0)= x0, x(t) ∈ M , t ∈ R, (2.1)

to denote a nonautonomous continuous time dynamical system, an ODE (ordinary differen-tial equation) with initial condition. We shall assume sufficient regularity to allow existenceand uniqueness [262] to suggest a dynamical system. By this, we mean that (semi-)groupaction leads to a (semi)dynamical system with solutions that are (semi-)flows.

Definition 2.1. (Flow [228, 262]) A flow is a two-parameter family of differentiable mapson a manifold15,

1. Two-parameter mapping: x(t; t0, x0) : R×R×M → M which we interpret as map-ping phase space M , through the time parameters t0 and t , i.e., x0 �→ x(t ; t0, x0).

2. Identity property: for each “initial" x0 ∈M at the initial time t0, the identity evolutionis, x(t0; t0, x0)= x0.

3. Group addition property: For each time t and s in R, x(s; t , x(t ; t0, x0)) = x(t +s; t0, x0).

4. The function x(·; ·, ·) is differentiable with respect to all arguments.

Definition 2.2. (Semiflow) A semiflow is identical to that of a flow, weakened only by thelack of time reversibility. That is, property (3) is changed so that the parameters t and smust come from the positive reals, �+.

See also Definition 3.1. Whereas a flow is a group isomorphic to addition on thereals, a semiflow is not reversible. Hence group action is now a semigroup since we cannotexpect the property that each element has an inverse. The concatenation of an element andits inverse is expressed by,

x(−t+ t0; t , x(t; t0, x0))= x(−t+ t0+ t ; t0, x0)= x(t0; t0, x0)= x0, (2.2)

requiring use of a negative time −t , meaning prehistory.The concept of semiflows arises naturally in certain physical systems, such as a heat

equation ut = kuxx to cite a simple PDE example, or the leaky bucket problem to cite anODE example. Such are problems that “forget" their history, and the forgetting process wewill eventually discuss alternatively as dissipation or a friction.

Example 2.1. (The Leaky Bucket - no reversibility [313]). The initial value problem

x =−√|x |, x(0)= x0, (2.3)

describes the height of a column of water in a bucket, which leaks out due to a hole at thebottom, assuming that the rate of water loss due to the leak depends on the pressure above

15If f (x , t)≡ f (x), the ode is called autonomous and the flow associated with an autonomous equation isreduced to the one-parameter family of maps, i.e., x0 �→ x(t; x0).

Page 41: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2.1. The Form of a Dynamical System 33

it, that in turn is proportional to the volume of the column of water above the hole. It canbe shown by substitution that this problem when x0 = 0 allows solutions of the form,

x(t)=[

14 (t− c)2, t < c

0 t ≥ c,

](2.4)

for any c, as well as the constant zero solution, x(t)= 0. Analytically, it should be noted thatthe right hand side of Eq. (2.3) fails to be Lipschitz at t = 0, and hence it is not Lipschitzin any open set containing the initial condition. Therefore, the usual Picard-uniquenesstheorem [262, 200] fails to hold, suggesting that nonuniqueness is at least a possibility; asdemonstrated by multiple solutions nonuniqueness does in fact occur. Nonuniqueness inthis example quite simply corresponds to a physically natural observation, that an emptybucket cannot “remember" when it used to be full, or even if it was ever full. �

The function on the right hand side of Eq. (2.1),

f : M×R→ M , (2.5)

denotes a vector field in the phase space manifold, M . However, the same ansatz formcan be taken to denote a wider class of problems, even semidynamical systems (formaldefinition in 3.1) including certain PDEs (partial differential equations) such as reactiondiffusion equations when M is taken to be a Banach space [280]. In such case, throughGalerkin’s method when M is a Hilbert space, the PDE corresponds to an infinite set ofODEs describing energy oscillating between the time varying “Fourier" modes. While theidea is straight forward for our purposes, to make this statement rigorous it is necessaryto properly understand regularity and convergence issues by methods between functionalanalysis and partial differential equations theory [280, 67].

Here, we will be most interested in the ODE case. In particular, we contrast Eq. (2.1)to the autonomous case,

x = f (x), x(t0)= x0 ∈ M , (2.6)

where the right hand side does not explicitly incorporate time. Note that a nonautonomousdynamical system can be written as an autonomous system in a phase space of one moredimension, by augmenting the phase space to incorporate the time.16

A flow incorporates continuous time, whereas maps are a widely studied class ofdynamical systems descriptive of discrete time,

xn+1 = F(xn), given x0. (2.7)

For convenience, we will denote both the mapping x(t ;x0) : R×M→ M descriptiveof a flow, and those of a mapping Eq. (2.7), by the notation, φt (x0); the domain of theindependent t variable will determine the kind of dynamical system:

Definition 2.3. Dynamical systems including maps and semiflows can be classified withinthe language of the flow in Definition 2.1,

• If φt (·) denotes a flow, then require the time domain is R.

• If φt (·) denotes a semiflow, then require the time domain is R+.

16Given Eq. (2.1), then let τ = t and hence, τ = dτ/dt = 1, and x = f (x , t) is autonomous in M×R.

Page 42: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

34 Chapter 2. Dynamical Systems Terminology and Definitions

• If φt (·) denotes an invertible discrete time mapping, then require the time domainis Z.

• If φt (·) denotes a noninvertible discrete time mapping, then require the time is Z+.

We will refer to dynamical systems which are either semiflows, or noninvertible map-pings together as semidynamical systems (formal definition in 3.1) acknowledging thesemigroup nature.

Both discrete time and continuous time systems will be discussed here, as there arephysically relevant systems which are naturally cast in each category. A stereotypical dis-crete time system is a model descriptive of compounding money interest at the bank, whereinterest is awarded at each time epoch. On the other hand, discrete time sampling of con-tinuous systems, as well as numerical methods to compute estimates of solutions both takethe form of discrete time systems. As seen in the previous chapter, a flow gives rise to adiscrete time map through the method of Poincare’ mapping as in Fig. 1.12.

2.2 LinearizationIn this section, we summarize the information to be gained from the linearization of a map,which we will use to study and classify the local behavior “near" the fixed or periodicpoints of nonlinear differential equations. Loosely speaking, the Hartman-Grobman theo-rem [156, 262], shows when the local behavior near the hyperbolic fixed point is “similar"to that of the associated linearized system.

Consider maps f : U ⊂Rk →Rn , where U is an open subset of Rm . The first partialderivative of f at a point p can be expressed in a n× k matrix form called a Jacobianderivative matrix,

D fp =[∂ fi∂xj

]. (2.8)

We consider this matrix as a linear map from Rm to Rn , that is D fp ∈ L(Rm ,Rn), andrecall that L(Rm ,Rn) is isomorphic to Rmn . With this in mind we may define the (Fréchet)derivative in the following way.

Definition 2.4. A map f : U ⊂ Rm → Rn is said to be Fréchet differentiable at x ∈ Rm

if and only if there exists D fx ∈ L(Rm ,Rn), called the Fréchet derivative of f at x , forwhich

f (x+h)= f (x)+ Dfxh+o(‖h‖) as ‖h‖→ 0. (2.9)

The derivative is called continuous provided the map D f : U → L(Rm ,Rn) is con-tinuous with respect to the Euclidean norm on the domain and the operator norm onL(Rm ,Rn). If the first partial derivatives at all points in U exist and are continuous, thederivative is also continuous and the map f is called continuously differentiable or f ∈C1.Higher order derivatives can be defined recursively. E.g.,

D f (x+ k)= D f (x)+ D2 fx k+o(‖k‖) (2.10)

defines the second derivative, D2 fx ∈ L(Rm , L(Rm ,Rn)). We again note that L(Rm ,Rn)is isomorphic to bilinear maps in L(Rm ×Rm ,Rn). The mapping (h,k) �→ (D2 fx k)h is

Page 43: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2.2. Linearization 35

bounded and bilinear from Rm ×Rn into Rn . If v,w ∈ Rm are expressed in terms of thestandard basis as v =∑i vi ei and w =∑i wi ei , then we have

(D2 fxv)w =∑i, j

(∂2 f

∂xi∂xj

)xviwj (2.11)

Since the mixed cross derivatives are equal,17 D2 fx is a symmetric bilinear form. Gen-erally, if all the derivatives of order 1 ≤ j ≤ r exist and are continuous, then f is said tobe r−continuously differentiable, or f ∈Cr . If f :Rm→Rm is Cr , D fx is a linear isomor-phism at each point x ∈Rk , and f is 1-to-1 and onto, then f is called a Cr−diffeomorphism.

Now, consider a dynamical system as the flow from an ODE,

x = f (x , t) (2.12)

and assume that f : M → M is at least C2 in the domain M of x and C1 in time, t . Then,it is possible to linearize a dynamical system about a point x ∈ M . Suppose an initial pointx is advected to x(t) = x + δx(t) after time t by a vector field. Then, the linearization ofx = f (x , t) is

dδx

dt= Dfx · δx+o(‖δx‖). (2.13)

Note that D fx (t ) is generally time-dependent along the trajectory x(t) and it is also non-singular at each time t by the existence and uniqueness of the solution. If γ (t) is a trajectoryof the vector field, we write the associated linearized system of x = f (x , t) as

ξ = D f (γ (t), t) · ξ . (2.14)

Example 2.2. (The Duffing Oscillator and Associated Variational Equations) In prac-tice, the variational Eqs. (2.14) together with the base orbit, Eq. (2.12) should simply beintegrated together as a single coupled system. To make this clear, consider the specificexample of a Duffing equation in autonomous form, Eq. (1.40).18 Also review Figs. 1.12-1.16. The Jacobian matrix of the vector valued function f is,

D f =⎛⎜⎝

∂ f1∂x1

∂ f1∂x2

∂ f1∂x3

∂ f2∂x1

∂ f2∂x2

∂ f2∂x3

∂ f3∂x1

∂ f3∂x2

∂ f3∂x3

⎞⎟⎠=⎛⎝ 0 1 0−1−3x1 −a −ωb sinωx3

0 0 0

⎞⎠ . (2.15)

We see that the variables x1 and x3, but not x2 in this particular case, appear explicitly in thederivative matrix. These are evaluated along an orbit solution, γ (t) = (x1(t), x2(t), x3(t)).Therefore, to solve the variational equation, Eq. (2.14), we must solve simultaneously,

x = f (x , t), (2.16)

ξ = D f (x(t), t) · ξ . (2.17)

Notice the forward-only coupling from Eq. (2.16) to Eq. (2.17). In this Duffing example,

x =⎛⎝ x1

x2x3

⎞⎠=⎛⎝ x2

−ax2− x1− x31 +b cosωx3

1

⎞⎠ ,

⎛⎝ x1(0)x2(0)x3(0)

⎞⎠=⎛⎝ x1,0

x2,0x3,0

⎞⎠ (2.18)

17Clauraut’s theorem asserts symmetry of the partial derivatives with sufficient continuity, [276].18 x1 = f1(x1, x2, x3)= x2, x2 = f1(x1, x2, x3)=−ax2− x1− x3

1 +bcosωx3, x3 = f3(x1, x2, x3)= 1.

Page 44: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

36 Chapter 2. Dynamical Systems Terminology and Definitions

and,

ξ =⎛⎝ ξ1

ξ2ξ3

⎞⎠=⎛⎝ 0 1 0−1−3x1 −a −ωb sinωx3

0 0 0

⎞⎠ ·⎛⎝ ξ1

ξ2ξ3

⎞⎠ (2.19)

An orthonormal set of initial conditions for Eq. (2.19) allows exploring the complete fun-damental set of solutions. As such, it is convenient to choose,⎛⎝ ξ1(0)

ξ2(0)ξ3(0)

⎞⎠=⎛⎝ 1

00

⎞⎠ ,

⎛⎝ 010

⎞⎠ , or

⎛⎝ 001

⎞⎠ . (2.20)

Counting the size of the problem, we see that, x(t) ∈ Rn , n = 3 for Eqs. (2.16), (2.19).

The variations, ξ (t) ∈ Rn of the variational equations Eqs. (2.17), (2.19) are also n = 3 di-mensional, but exploring a complete basis set Eq. (2.20) requires n2 = 32 = 9 dimensions.Therefore a complete “tangent bundle" [276] which is the base flow together with the fun-damental solution of variations requires n+n2 = 3+32 = 12 dimensions. See Fig. 2.1. Weemphasize that computationally these equations should in practice all be integrated simul-taneously in one numerical integrator subroutine. In terms of the flow, �T (x(0))= x(t), thevariational equations evolved for a basis set as discussed, produces a matrix M = D�T |x(0)which is often called a “monodromy" matrix, [276]. Thus treating the time-T evolution ofthe flow and the derivative matrix of variations together form a discrete time mapping andits derivative matrix. �

Figure 2.1. The variational equations for an orthogonal set of initial conditionsare evolved together with the base flow x(t). See Eqs. (2.16)-(2.17).

2.3 HyperbolicityPart of the importance of hyperbolicity in applied dynamical system is that it allows signif-icant simplifications for the analysis of complex and even chaotic dynamics. This section isintended to provide a sufficient background of hyperbolic dynamics for our discussion. We

Page 45: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2.3. Hyperbolicity 37

will limit ourselves to the notion of uniform hyperbolicity but note that a thorough treat-ment of hyperbolic dynamics and non-uniform hyperbolicity and its relation to ergodictheory can be found in Pollicott and Yuri [272]. In Chapter 8, we will see how the relatedFinite Time Lyapunov Exponent (FTLE) theory has recently developed a central role in theempirical study of transport mechanism and partitioning the phase space of a complex sys-tem into dynamically related segments. The ideas behind that theory begin with an analogyto the linear theory. See Fig. 2.2 and compare to the many figures in Chapter 8.

In this section, we are interested in the (discrete time) map f : M→ M where M is acompact n-dimensional C∞ Riemannian manifold. For our purposes here, it is sufficient todescribe a Riemannian manifold as a differentiable topological space (Definition 4.3) withan inner product on tangent spaces and which is locally isomorphic to a Euclidean metricspace.

Figure 2.2. Stable and Unstable Invariant Manifolds of a fixed point p serve acritical role in organizing global behaviors such as invariant and almost invariant sets.

The Anosov diffeomorphism is a special prototypical example of hyperbolic behav-ior.

Definition 2.5. (Anosov diffeomorphism [329]) A diffeomorphism f : M→ M is Anosovif there exist constants c > 0,0 < λ < 1 and a decomposition of continuous tangent space,

Tx M = Esx ⊕ Eu

x , (2.21)

at each x ∈ M satisfying the following properties:(a) Es

x and Eux are invariant with respect to the derivative D fx :

D fx Esx = Es

f (x)

D fx Eux = Eu

f (x)(2.22)

(b) The expansion rate under the forward and backward iteration are uniformly bounded:∥∥D f nx vs∥∥= cλn ‖vs‖ for vs ∈ Es

x∥∥D f −nx vu

∥∥= cλn ‖vu‖ for vu ∈ Eux

(2.23)

Page 46: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

38 Chapter 2. Dynamical Systems Terminology and Definitions

Since the constants c and λ given in the above definition are the same for eachpoint x ∈ M , Anosov systems are uniformly hyperbolic systems. By uniformly hyper-bolic, it can be said roughly that at every point in the phase space, the hyperbolic splittingEqs. (2.22)-(2.23) holds.

In the following, we define

Definition 2.6. A closed set A ⊂ M is an invariant set with respect to a transformation fif if f (A)⊆ A.

Definition 2.7. If f is uniformly hyperbolic at each point x ∈ A, then A is called a uni-formly hyperbolic invariant set.

Definition 2.8. If the invariant set is a single point, A = {x}, that is f (x) = x , then x iscalled a hyperbolic fixed point.

Hyperbolic periodic points are similarly defined. Part of the importance of hyper-bolicity is that the dynamics in a (small) neighborhood of the hyperbolic points exhibitexpanding and contracting behavior in a manner which is homeomorphic to the associ-ated linearized system; this statement is made precise by the Hartman-Grobmann theo-rem [156, 279, 329].

A central concept in global analysis of dynamical systems in the presence of hyper-bolicity is due to the presence of stable and unstable manifolds. Part of their importance isthat these are codimension-one invariant sets, and as such, they can serve to partition thephase space. Furthermore stable and unstable manifolds tend to organize transport activityas we will see developed in the sequel. See Figs. 2.3-2.4. Their definition follows from thefollowing existence theorems.

Definition 2.9. (Local Stable and Unstable Manifold [279]) Let A ⊂ M be a hyperbolicinvariant set, x ∈ A and a neighborhood N(x ,ε) for ε > 0 sufficiently small. The localstable manifold, Ws

ε (x), and the local unstable manifold, Wuε (x), are given by

Wsε (x)= {y ∈ N(x ,ε)|d( f n(y), x)→ 0 as n→∞}

Wuε (x)= {y ∈ N(x ,ε)|d( f −n(y), x)→ 0 as n→∞} (2.24)

Note that the local unstable manifold can still be defined even if f is not invertibleby abusing the notation f −n(x) as an iterated preimage of x .

Theorem 2.1. (Stable manifold theorem [329]) Let f : M→ M , A be a hyperbolic invari-ant set with respect to f , and x ∈ A. Then, there exist local stable and unstable manifoldsWs

ε (x) and Wuε (x) that satisfy the following properties:

(i) Wsε (x) and Wu

ε (x) are differentiable submanifolds and depend continuously on x .(ii) Speaking of the tangent spaces, T Ws

ε (x) = Esx and T Wu

ε (x) = Eux , i.e., Ws

ε (x) andWu

ε (x) are of the same dimension and tangent to the subspace Esx and Eu

x , respectively.

The previous result can be extended to the entire domain M by the methods of theHadamard-Perron Manifold Theorem (See Wiggins [329]). Therefore, we can discuss thelocal stable and unstable manifold at each point in the domain of an Anosov diffeomor-phism.

Construction of the global stable and unstable manifolds follows by taking unions ofbackward and forward iterates of local stable and unstable manifolds respectively. This isdetailed by:

Page 47: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2.3. Hyperbolicity 39

Definition 2.10. (Global Stable and Unstable Manifolds [279]) The global stable and un-stable manifolds at each x ∈ M , Ws (x) and Wu (x) respectively, are given by

Ws (x)≡ {y ∈ M|d( f j (y), f j (x))→ 0 as j→∞}Ws (x)=

⋃n≥0

f −n(Wsε ( f n(x))) and

Wu (x)≡ {y ∈ M|d( f − j (y), f − j (x))→ 0 as j→∞}Wu (x)=

⋃n≥0

f n(Wuε ( f −n(x))).

(2.25)

It is clear from the above definition that the stable and unstable manifolds are in-variant since they are a union of trajectories. Also, it is worth mentioning that a stable(unstable) manifold can neither cross itself nor other stable (unstable) manifolds of anotherpoint; otherwise, the crossing point must be iterated to two different points by definitionbut this is not permitted by uniqueness of solutions. By contrast, an unstable manifold is al-lowed to intersect a stable manifold which simply implies a coincidence in future outcomesfrom prior histories. Very interesting behavior can result as we discuss below.

The following definitions are some types of special orbits that are particularly im-portant to transport and mixing theory since they often form the transport barrier betweenregions of qualitatively different dynamics.

Definition 2.11. (Homoclinic orbit [279]) Let p be a hyperbolic periodic point of period nfor a diffeomorphism f and O(p) be the orbit of p. Let

Ws,u(O(p))=n−1⋃j=0

Ws,u( f j (p)) and

W s,u(O(p))=Ws,u(O(p)) \O(p).

(2.26)

A point q ∈ W s (O(p))∩ Wu (O(p)), if it exists, is called a homoclinic point for p.The orbit of q is then called the homoclinic orbit for p. See Figs. 2.3-2.4.

It follows from the above definition that a point in the homoclinic orbit will asymp-totically approach the same hyperbolic periodic point p both in forward and backwardtime. The Smale-Birkhoff homoclinic theorem provides that the existence of a transversehomoclinic point induces a Smale horseshoe in the homoclinic tangle [279, 329, 327]. Thisimportant concept in the theory of transport mechanism will be detailed in Sec. 6.1.5.

Definition 2.12. (Heteroclinic orbit [279]) Let {pi},i=1. . . n, be a collection of hyperbolicperiodic orbits for a diffeomorphism f . A point q ∈ W s (O(pi))∩ Wu (O(pj )) for somei �= j , if it exists, is called a heteroclinic point. Similarly, the orbit of q is called theheteroclinic orbit.

In contrast to a homoclinic orbit, the orbit of a heteroclinic orbit asymptotically ap-proaches a hyperbolic periodic orbit forward in time and a different periodic orbit backwardin time, see Figure 2.4.

Page 48: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

40 Chapter 2. Dynamical Systems Terminology and Definitions

Figure 2.3. An example of a homoclinic orbit of a hyperbolic fixed point p. Hereq is the homoclinic point and the map f is assumed to be an orientation-preserving map,in which there must be at least one homoclinic point between q and f (q). Note also thateach and every of the infinitely many points { f i q}∞i=−∞ on the orbit of q are homoclinic.

Figure 2.4. An example of a heteroclinic orbit connecting two hyperbolic fixedpoints p1 and p2. Here q is the heteroclinic point and the map f is assumed to be anorientation-preserving map as in Figure 2.3

2.4 Hyperbolicity: Nonautonomous vector fieldsIn the preceding section, the linearization of a fixed point or an invariant set was used todiscuss the hyperbolicity property. However, for a general time-dependent system, we haveto define the notion of the hyperbolicity for the time-dependent trajectory for which the lin-earization along the trajectory does not necessarily give rise to a constant matrix associatedwith the linearized vector field. The standard notion to characterize the hyperbolicity of a

Page 49: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2.4. Hyperbolicity: Nonautonomous vector fields 41

time-dependent trajectory is that of the exponential dichotomy described below. We alsosummarize below the concept of stable and unstable manifolds of the hyperbolic trajectoryand their significance and relation to the Lagrangian coherent structure (LCS), which willbe discussed in Sec. 6.1.5.

As in the case of an autonomous system, we begin by defining an appropriate decom-position into stable and unstable subspaces along a trajectory of a nonautonomous system.This straight forward extension of the hyperbolicity of the autonomous vector field. Firstconsider the linearized vector field of the form (2.14) and keep in mind that the coefficientmatrix is now time-dependent.

Definition 2.13. (Exponential Dichotomy [329]) Consider a time-dependent linear differ-ential equation

ξ = A(t)ξ ξ ∈ Rn , (2.27)

where A(t) ∈ Rn×n is a time-dependent coefficient and continuous in time, t ∈ R. LetX(t) ∈Rn×n be the solution matrix such that ξ (t)= X(t)ξ (0) and X(0)= I . Then (2.27) issaid to possess an exponential dichotomy if there exists a projection operator P , P2 = P ,and constants K1, K2,λ1,λ2 > 0 such that∥∥∥X(t)P X−1(t)

∥∥∥≤ K1 exp(−λ1(t− τ )), t ≥ τ∥∥∥X(t)(I − P)X−1(t)∥∥∥≤ K2 exp(λ2(t− τ )), t ≤ τ

(2.28)

A generalization of a hyperbolic fixed point in autonomous systems is that of thehyperbolic trajectory of a time-dependent vector field, which can be defined via the expo-nential dichotomy.

Definition 2.14. (Hyperbolic Trajectory [329]) Let γ (t) be a trajectory of the vector fieldx = f (x , t). Then γ (t) is called a hyperbolic trajectory if the associated linearized systemξ = Dx ( f (γ (t), t))ξ has an exponential dichotomy.

The geometry of a hyperbolic trajectory can be understood in the extended phasespace:

E ≡ {(x , t) ∈ Rn×R}. (2.29)

The nonautonomous vector field can then be viewed as an autonomous one as follows

x = f (x , t)

t = 1.(2.30)

In the extended phase space E , we denote the hyperbolic trajectory by �(t) = (γ (t), t) anddefine a time slice of the extended phase space E by

�τ ≡ {(x , t) ∈ E |t = τ }. (2.31)

Then the condition (2.28) requires that there exists a projection onto a subspace of �τ ,called Es(τ ), so that a trajectory of an initial vector at time t = 0 projected onto Es(τ ) bythe associated linearized vector field (2.27) will have to decay to zero at an exponentialrate, determined by λ1 as t →∞. Similarly, the complimentary projection (I − P) ontoa subspace of �τ , called Eu(τ ), exists so that the initial vector projected onto Eu(τ ) will

Page 50: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

42 Chapter 2. Dynamical Systems Terminology and Definitions

decay at exponential rate, determined by λ2 as t→−∞. Moreover, �τ = Es (τ )⊕ Eu(τ ).Therefore, the exponential dichotomy guarantees the existence of a stable (unstable) sub-space for which the initial conditions on these spaces asymptotically approach a hyperbolictrajectory at an exponential rate at a forward (backward) time, see Figure 2.5. This isanalogous to that of the autonomous vector field where the initial conditions on the stable(unstable) manifold converge to some critical points, e.g., fixed points or periodic orbits,forward (backward) in time. However, the difference is that the stable (unstable) manifoldsfor an autonomous system are time-independent whereas those of a nonautnomous casevary in time.

In particular, for a 2-D vector field, the stable (unstable) manifolds in the autonomouscase is an invariant curve, but the stable (unstable) manifold in the nonautonomous casebecomes a time-varying curve or an invariant surface in the extended space E . With the

Figure 2.5. The time slice �τ = Es(τ )⊕ Eu(τ ). A trajectory of an initial seg-ment projected onto Es (τ ) by the associated linearized vector field (2.27) approaches thehyperbolic trajectory �(t) at an exponential rate as t →∞. Similarly, the initial vectorprojected onto Eu(τ ) approaches �(t) at an exponential rate as t→−∞.

geometrical explanation of the exponential dichotomy in mind, we can now state the the-orem that describes the existence of local stable and unstable manifolds of a hyperbolictrajectory.

Theorem 2.2. (Local Stable and Unstable Manifolds [329] and also see similarly [279])Let Dρ (τ )�τ denote the ball of radius ρ centered at γ (τ ) and define the tubular neighbor-hood of �(t) in E by Nρ(�(t)) ≡ ∪τ∈R(Dρ(τ ),τ ). There exists (k+ 1) dimensional Cr

manifold Wsloc(�(t))⊂ E , and an (n− k+1) dimensional Cr manifold Wu

loc(�(t))⊂ E andρ0 sufficiently small such that for ρ ∈ (0,ρ0)

(i) Wsloc(�(t)), the local stable manifold of �(t), is invariant under the forward time

evolution generated by (2.30); Wuloc(�(t)), the local unstable manifold of �(t), is invariant

under the backward time evolution generated by (2.30).(ii) Ws

loc(�(t)) and Wuloc(�(t)) intersect along �(t), and the angle between the mani-

folds is bounded away from zero uniformly for all t ∈ R.(iii) Every trajectory on Wu

loc(�(t)) can be continued to the boundary of Nρ(�(t))

Page 51: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

2.4. Hyperbolicity: Nonautonomous vector fields 43

backward in time, and every trajectory on Wuloc(�(t)) can be continued to the boundary of

Nρ(�(t)) forward in time.(iv) Trajectories starting on Ws

loc(�(t)) at time t = τ approach �(t) at an exponential

rate e−λ′(t−τ ) as t→∞ and trajectories starting on Wuloc(�(t)) at time t = τ approach �(t)

at an exponential rate e−λ′|t−τ | as t→∞, for some constant λ′ > 0.(v) Any trajectory in Nρ(�(t)) not on either Ws

loc(�(t)) or Wuloc(�(t)) will leave

Nρ(�(t)) in both forward and backward in time.

The above theorem suggests a way to determine global stable and unstable manifoldsof a hyperbolic trajectory [220, 221, 222]. In short, it allows us to extend an initial segmentof a local unstable manifold, Wu (γ (t1)), of a hyperbolic trajectory γ (t1) on the time slice�t1 for some t1 < τ by evolving this segment forward in time. Similarly, evolving the initialsegment of a stable manifold Ws (γ (t2)) of a hyperbolic trajectory γ (t2) on the time slice�t2 for some t2 > τ backward in time yields a global stable manifold. Figure 2.6 illustratesthis concept.

One rigorous method to determine a meaningful hyperbolic trajectory, called distin-guished hyperbolic trajectory (DHT), and its stable and unstable manifolds can be foundin [220, 221, 222]. Since a dynamical system can possess infinitely many hyperbolic trajec-tories, e.g., all trajectories in the stable or unstable manifolds, the bounded hyperbolic tra-jectory with the special properties is used as the reference hyperbolic trajectory for “grow-ing" the stable and unstable manifolds [220, 221]. It is demanded in [183] that a DHThas to remain in a bounded region for all time and that there exists a neighborhood N ofthe DHT such that all other hyperbolic trajectories within N have to leave N in a finitetime, either by forward or backward time manner. If one attempts to grow the stable orunstable manifold from a hyperbolic trajectory different from the DHT, it is possible toobserve a “drifting" phenomenon in the flow due to a slow expansion rate as demonstratedin a number of numerical examples in [222].

A rigorous method to determine the DHT in general is still an active research area. Inmost cases, DHT has to be selected with a careful observation and knowledge of the system.In another approach, Haller and Poje [166] determined a finite-time uniformly hyperbolictrajectory based on the local maxima of the time duration in which an initial point remainshyperbolic. The contour plot of such hyperbolicity time for each grid point of initial con-ditions reveals the uniformly hyperbolic invariant set, which can be used as a “seed" togradually construct the corresponding global stable and unstable manifolds by a traditionaltechnique, e.g., the straddling technique of You, Kostelich and Yorke [334]. Nevertheless,this technique demands the deformation of the uniformly hyperbolic trajectory to be at aslower speed than the speed of individual particles.

In a recent approach, following the work in a series of papers by Haller [166, 163,164], the search of DHT or hyperbolic invariant sets can be circumvented by computingthe finite-time Lyapunov exponent field [300, 299], which directly captures the region ofa large expansion rate. See Chapter 8. These regions corresponds to the stable (unstable)manifold of the hyperbolic trajectory when computing in forward (backward) time man-ner. This technique is based on the fact that the stable (unstable) manifold of a hyperbolictrajectory is repelling (attracting) in the normal direction to the manifold and hence initialpoints straddling the stable (unstable) manifold will eventually be separated from (attractedto) each other at some exponential rate. However, this approach gives us merely a scalarfield indicating the expansion rate of the grid point in a domain of a nonautonomous sys-

Page 52: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

44 Chapter 2. Dynamical Systems Terminology and Definitions

Figure 2.6. The forward-time evolution of a segment of Wu (γ (t1)) from the timeslice �t1 to the time slice �τ yields the unstable manifold of Wu (γ (τ )). Similarly, thebackward-time evolution of a small segment of Ws (γ (t2)) from the time slice �t2 to thetime slice �τ yields the unstable manifold Ws (γ (τ )).

tem whereas the preceding techniques yield the parameterized curve representing invariantmanifolds. Nevertheless, the second-derivative ridge of the finite-time Lyapunov exponentfield can be used as a curve that represents the global stable (or unstable) manifold of anonautonomous dynamical system by using ridge detection method.

Page 53: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 3

Frobenius-Perron Operator andInfinitesimal Generator

This chapter is devoted to a review of some classical tools and techniques for studying theevolution of density under a continuous time process. Specifically, we concentrate on theconcept of the Frobenius-Perron operator and its infinitesimal generator. The continuoustime problem will also be cast in terms of the related discrete time problem. The discussionherein expands a more mathematical discussion of these transfer operators to which wealready introduced in Chapter 1

Here we present the technical details of the Frobenius-Perron operator - the evolutionoperator point of view alluded to in Chapter 1. A good further review of this material isfound in [208]. In principle, this chapter could be skipped if only a more computationalperspective is required.

3.1 Frobenius-Perron OperatorWe define a continuous process in a topological Hausdorff space X by a family of mappings

St : X → X , t ≥ 0. (3.1)

For example, a continuous time process generated by an autonomous d-dimensional systemof differential equations

dx

dt= F(x) (3.2)

ordxi

dt= Fi (x), i = 1, . . . ,d , (3.3)

where x = (x1, . . . , xd ) and F : Rd → Rd is sufficiently smooth to guarantee the ex-istence and uniqueness of solutions. Then, Eq. (3.2) defines a transformation St (x0)= x(t)where x(t) is the solution of Eq. (3.2) at time t starting from the initial condition x0 at time0.

Definition 3.1. A semidynamical system {St }t≥0 on X is a family of transformations St :X → X , t ∈R, satisfying(a) S0(x)= x for all x ∈ X ;

45

Page 54: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

46 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

(b) St (St ′(x))= St+t ′(x) for all x ∈ X with t , t ′ ∈ R+; and(c) The mapping (x , t)→ St (x) from X ×R

+ into X is continuous

See also Definitions 2.1-2.2.

Remark 3.1. The reason to call {St }t≥0 in the above definition a semidynamical systeminstead of a dynamical system is to allow for the possibility for lack of invertibility. Byrestricting t ∈R+ a family of transformation that satisfies the above conditions possesses anAbelian semigroup properties and hence it may be called a semigroup of transformations.

Now we revisit the density transfer operator known as the Frobenius-Perron oper-ator, already discussed in the simplest case in Eqs. (1.10-1.18) of discrete time mappingsof the interval. Now we present this important tool for studying a propagation of densitiesin the more general settings.

Let (X ,�,μ) be a σ−finite measure space, where � denotes the σ−algebra of Borelsets. Assume that each transformation of a semidynamical system {St }t≥0 is a nonsingularmeasurable transformation on (X ,�,μ), that is,

μ(S−1t (A))= 0 for each A ∈� such that μ(A)= 0. (3.4)

Therefore, measure preserving transformations {St } is necessarily nonsingular with respectto μ. 19 The Frobenius Perron operator, Pt : L1(X)→ L1(X) with respect to the trans-formation St is defined [208] by the condition of conservation of mass,∫

S−1t (A)

f (x)dμ=∫

APt f (x)dμ, for each A ∈�. (3.5)

In the sequel, we will consider only the action of Pt on the space D(X ,�,μ) defined by,

D ≡ D(X ,�,μ)= { f ∈ L1(X ,�,μ) : f ≥ 0 and ‖ f ‖ = 1}. (3.6)

So, D is a set of a probability density function (PDF) of L1(X). Therefore, Pt f (x) is alsoa probability density function, which is unique a.e.,20 and depends on the transformations{St } and the initial probability density function f (x). It is straightforward to show [208]that Pt satisfies the following properties:

(a) Pt (λ1 f1+λ2 f2)= λ1 Pt f1+λ2 Pt f2, for all f1, f2 ∈ L1,λ1,λ2 ∈ R;

(b) Pt f ≥ 0 if f ≥ 0;

(c)∫

Xf (x)dμ=

∫X

Pt f (x)dμ, for all f ∈ L1.

(3.7)

By using the above properties, one may prove that Pt satisfies properties (a) and (b) of thedefinition of a semidynamical system.

19Transformations{St } are measure preserving with respect to μ if μ(S−1t (A)) = μ(A) for all A ∈�

20The uniqueness can be established by applying the Radon-Nikodym theorem. For a given functionf ∈ L1(X ), we may define the left hand side of (3.5) as a real measure μ f (A). Since {St } are nonsingularfor every t , μ f are absolutely continuous with respect to μ. By the Radon-Nikodym theorem, there existsa unique function (the so-called Radon-Nikodym derivative), denoted by Pt f ∈ L1(X ), such that μ f (A) =∫

A Pt f dμ for every A ∈�.

Page 55: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.2. Infinitesimal Operators 47

Consider in addition to the differential equation (3.2) an observable Kt f defined by

Kt f (x)= f (St (x)), (3.8)

where f , Kt f ∈ L∞(X). Hence, the operator Kt : L∞(X)→ L∞(X) as defined in Eq. (3.8),for every t ≥ 0, can be interpreted as the operator that evolves an observable f (St (x)) ofa semidynamical system {St }t≥0 given by Eq. (3.2). The operator Kt is known as theKoopman operator associated with the transformation St . It is easy to check that {Kt }t≥0is a semigroup. An important mathematical relation between the Frobenius-Perron andKoopman operators is that they are adjoint. That is,

〈Pt f , g〉 = 〈 f , Kt g〉 (3.9)

for all f ∈ L1(X), g ∈ L∞(X), and t ≥ 0. 21

Note that although the Frobenius-Perron operator preserves the L1-norm ‖ · ‖L1 forf ≥ 0, (Recall the discrete continuity equation, Eq. (1.12),

‖Pt f ‖L1 = ‖ f ‖L1 , (3.10)

the Koopman operator on the other hand satisfies the inequality

‖Kt f ‖L∞ ≤ ‖ f ‖L∞ . (3.11)

In a subsequent section we will derive the infinitesimal operators of the Frobenius-Perronand Koopman operators. Therefore, we will concentrate on the semigroup of contractinglinear operators defined in the following.

Definition 3.2. For L = L p , 1≤ p ≤ 1, a family {Tt }t≥0 of operators, Tt : L→ L is calleda semigroup of contracting operators if Tt has the following properties:(a) Tt (λ1 f1+λ2 f2)= λ1Tt f1+λ2Tt f2;(b) ‖Tt f ‖L ≤ ‖ f ‖L ;(c) T0 f = f ; and(d) Tt+t ′ f = Tt (Tt ′ f ),for f , f1, f2 ∈ L and λ1,λ2 ∈ R.

The semigroup is called a continuous semigroup, if it satisfies

limt→t0‖Tt f − Tt0 f ‖L = 0 for f ∈ L, t0 ≥ 0. (3.12)

3.2 Infinitesimal OperatorsFor a continuous semigroup of contractions {Tt }t≥0 we define D(A) by the set of all f (x)∈L p(X), 1≤ p ≤∞, such that the limit

A f = limt→0

Tt f − f

t, (3.13)

21The notation, 〈q,r〉, q ∈ L1(X ),r ∈ L∞(X ), denotes a bi-linear form, and here takes the form of integra-tion. Often to be explicit we emphasize the spaces from which the functions are drawn, 〈q,r〉L1(X )×L∞(X ) =∫

X q(x)r(x)dx .

Page 56: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

48 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

exists in the sense of strong convergence. That is,

limt→0

∥∥∥∥A f − Tt f − f

t

∥∥∥∥L p= 0. (3.14)

The operator,A : D(A)→ L, for L ≡ L p , (3.15)

is called the infinitesimal generator. Let,

I (t)≡ I (x , t)= Tt f (x) for fixed f (x) ∈ D(A). (3.16)

The function I ′(t) ≡ I ′(t)(x) ∈ L is said to be the strong derivative of I (t) if it satisfiesthe following condition:

limt→0

∥∥∥∥I ′(t)− I (t)− f (x)

t

∥∥∥∥L= 0. (3.17)

In this sense, I ′(t) describes the derivative of the ensemble of points with respect to timet . The following theorem demonstrates an important relationship between the infinitesimalgenerator and the strong derivative

Theorem 3.1. (See [208] ) Let {Tt }t≥0 be a continuous semigroup of contractions andA : D(A)→ L the corresponding infinitesimal generator. Then for each fixed f ∈ D(A)and t ≥ 0, the function I (t)= Tf t has the properties:(a) I (t) ∈ D(A);(b) I ′(t) exists; and(c) I (t) satisfies

I ′(t)= AI (t) (3.18)

with the initial condition I (0)= f (x).

Example 3.1. Consider the family of operators {Tt }t≥0 defined by

Tt f = f (x− ct) for x ∈ R, t ≥ 0. (3.19)

Under this operation a function f (x) is translated in a positive direction of x by the length ofct . By using the “change of variable" formula we can see that the L p norm is preserved for1 ≤ p ≤∞. The conditions (a),(c), and (d) of the definition 3.2 straightforwardly followfrom Eq. (3.19). Thus, {Tt } is a semigroup of contracting operators. It is slightly morecomplicated to show that {Tt } is also continuous, see [208]. Now assume that f is boundedand at least C1(R), then by the mean value theorem we have

f (x− ct)− f (x)

t=−c f ′(x− θct), (3.20)

where |θ | ≤ 1. This implies that

A f = limt→0

Tt f − f

t=−c f ′ (3.21)

Page 57: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.2. Infinitesimal Operators 49

and the limit is strong in L p , 1 ≤ p ≤∞ if f has compact support. Therefore, it followsfrom Eq. (3.18) that at each point in the (x , t)−plane u(t , x) satisfies the partial differentialequation

∂u

∂ t+ c

∂u

∂x= 0 with u(0, x)= f (x), (3.22)

where u(t , x) is in D(A) for each fixed t ≥ 0.

Remark 3.2. This example offers some insight into the relationship between the semi-group of continuous operators, strong derivatives, and the corresponding partial differen-tial equations. It is well-known that the solution of Eq. (3.22) at time t is Tt f as definedin Eq. (3.19). See [280] for a discussion of PDE theory as related to infinite-dimensionaldynamical systems, and see [115] for the modern functional analysis formulation PDE the-ory.

Now consider a calculation of the infinitesimal generator of the semigroup of theFrobenius-Perron operators {Pt }t≥0 as defined in Eq. (3.5) and the evolution of the time-dependent density function I (x , t) under an action of the Frobenius-Perron operator. Thiswill be done indirectly through the adjoint property of the Frobenius-Perron and Koopmanoperators.

It follows directly from the definition of the Koopman operator Eq. (3.8) that theinfinitesimal of the Koopman operator denoted by AK is

AK g(x)= limt→0

g(St (x0))− g(x0)

t= lim

t→0

g(x(t))− g(x0)

t. (3.23)

If g is continuously differentiable with compact support, we can apply the mean valuetheorem to obtain

AK g(x)= limt→0

d∑i=1

∂g

∂xi(x(θ t))x ′i (θ t)=

d∑i=1

∂g

∂xiFi (x), (3.24)

where 0 < θ < 1. Combining equations (3.18) and (3.24) we conclude that the function

I (x , t)= Kt f (x) (3.25)

satisfies the first-order partial differential equation

∂ I

∂ t−

d∑i=1

∂ I

∂xiFi (x)= 0. (3.26)

This leads to a derivation of the infinitesimal generator for the semigroup of Frobenius-Perron operators generated by the family {St }t≥0 defined in Eq. (3.1).

Let f ∈ D(AF P ) and g ∈ D(AK ), where AF P and AK denote the infinitesimal op-erators of the semigroups of the Frobenius-Perron and Koopman operators, respectively.Using the adjoint property of the two operators it can be shown that

〈(Pt f − f )/t , g〉 = 〈 f , (Kt g− g)/t〉. (3.27)

Page 58: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

50 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

Taking the limit as t→ 0 we obtain

〈AF P f , g〉 = 〈 f , AK g〉. (3.28)

Provided that g and f are continuously differentiable, and g has compact support it followsthat [208]

〈AF P f , g〉 = 〈−d∑

i=1

∂ f Fi

∂xi, g〉. (3.29)

Hence, we conclude that

AF P f =−d∑

i=1

∂ f Fi

∂xi. (3.30)

Again, using Eqs. (3.18) and (3.30) to conclude that the function

I (x , t)= Pt f (x) (3.31)

satisfies the partial differential equation (continuity equation)

∂ I

∂ t+

d∑i=1

∂ I Fi

∂xi= 0, (3.32)

or symbolically,∂ I

∂ t+ F ·∇ I + I∇ · F = 0. (3.33)

Note that this equation is actually the same as the well-known continuity equation in fluidmechanics and many fields, but now it is a statement of conservation of density functionof ensembles of trajectories.22 In the case when F is a divergence-free vector field, i.e.∇ · F = 0, Eq. (3.32) corresponds to incompressible fluids such as water and it can besimplified to

d I

dt= ∂ I

∂ t+ F ·∇ I = 0. (3.34)

A comparison of Eqs. (3.32) and (3.34) to the classical optical flow problem was discussedin [287].

Example 3.2. Now consider a Duffing’s oscillator in the domain [0,1]× [0,1] given by thefollowing differential equation:

dx

dt= 4y−2

dy

dt= 4x−2−8(2x−1)3.

(3.35)

22Compare the continuous continuity equation from the infinitesimal generator of the Frobenius-Perronoperator, Eq. (3.32) to the discrete time continuity equation noted in Eq. (1.12). Said simply, both essentiallystate that existence and uniqueness implies that orbits from all initial conditions are conserved, so histogramsand hence densities of many initial conditions must change in time only to the movement or advection of theorbits.

Page 59: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.3. Frobenius-Perron Operator of Discrete Stochastic Systems 51

According to Eq. (3.32), given the initial density u(x ,0)= f (x) the flow of the density(u(x , t)=Pt f (x)) under the Duffing’s oscillator is given by

∂u

∂ t+ (4y−2)

∂u

∂x+ (4x−2−8(2x−1)3)

∂u

∂y= 0. (3.36)

A numerical simulation of this continuity equation of the density propagation by Eq. (3.36)along with the vector field of this ODE is shown in Figure 3.1 using an initial densityu(x ,0) illustrated in Figure 3.1(a). Notice that the density is stretched and folded due to thehyperbolic structure of the system.

3.3 Frobenius-Perron Operator of Discrete StochasticSystems

In more realistic situations, we consider a stochastic differential equation, in which casethe evolution of densities is described by the well-known Fokker-Planck equation whosea special case is the continuity equation as derived in Eq. (3.32). However, a study ofa continuous time system with a continuous stochastic perturbation requires a great dealof preliminary concepts in stochastic differntial equations and more advance concepts insemigroup theory that are beyond the scope of this book. Nevertheless, the discrete timeanalog of this problem is less involved and can be readily developed.

In sequel, we will consider a stochastic perturbation as a random variable.

Definition 3.2. A random variable,

X : → R, (3.37)

is a measurable function 23 from a measure space (,F ,μ) to a measurable space (R,B(R)),where B(R) denotes the Borel σ−algebra on R. The realization of such a selection, X(ω)is often called a “random experiment".

We may interpret the random variable as a measurement device that returns a realnumber, or the random experiment in our language, for a given subset of . Recall thatfor a measure space (,F ,μ), a measure μ is called a probability measure if μ : F →[0,1] and μ()= 1; hence a measure space (,F ,μ) will also be accordingly referred toas a probability space. With a probabilistic viewpoint in mind, the random variable tells usthat the probability to observe a measurement outcome in some set A ∈ B(R) based on aprobability measure μ is precisely μ(X−1(A)), which makes sense only if X is measurable.

We now extend the well-established formulation of the deterministic Frobenius-Perronoperator in the preceding section to study the phase space transport of a discrete systemswith constantly applied stochastic perturbation. In particular, let S, T : X → X be (non-singular) measurable functions acting on X ⊂ Rd .

We consider a process with both additive and multiplicative stochastic perturbationdefined by

xn+1 = νnT (xn)+ S(xn) (3.38)

23Given a measurable space (,F ) and a mesurable space (S,S), an “(F -)measurable function”, or sim-ply “measurable function", is a function f : → S such that f −1(A) ≡ {ω ∈ : f (ω) ∈ A} ∈ F for everyA ∈ S.

Page 60: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

52 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

Figure 3.1. A simulation of the flow of a density function according to the continu-ity equation (3.36) with the velocity field given by Eq. (3.35). The time increment betweeneach snapshot is t = 0.08. The range of high to low density is plotted from red blue.

Page 61: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.3. Frobenius-Perron Operator of Discrete Stochastic Systems 53

where νn are identically and independently distributed (i.i.d.) random variables each hav-ing the same density g. Note that if we set S ≡ 0, we would have a process with a multi-plicative perturbation, whereas when T ≡ 1, we have a process with an additive stochasticperturbation. Suppose the density of xn given by ρn . Such as system can be considereddescriptive of both parametric noise and additive noise terms. We desire to show therelationship of ρn and ρn+1 of the above stochastic process analogous to Eq. (3.47) assumethat S(xn), T (xn), and νn are independent, [288].

Let h : X → X be an arbitrary, bounded, measurable function, and recall that theexpectation of h(xn+1) is given by

E[h(xn+1)]=∫

Xh(x)ρn+1(x)dx . (3.39)

Then, using Eq. (3.38) we also obtain

E[h(xn+1)]=∫

X

∫X

h(zT (y)+ S(y))ρn(y)g(z)dydz. (3.40)

By a change of variable, it follows that

E[h(xn+1)]=∫

X

∫X

h(x)ρn(y)g((x− S(y))T−1(y)

)|J |dxdy, (3.41)

where |J | is the Wronskian derivative which is equivalently stated as the determinant ofthe Jacobian derivative matrix of the transformation,

x = zT (y)+ S(y). (3.42)

Since h was an arbitrary, bounded, measurable function, we can equate Eqs. 3.39 and 3.41to conclude that

ρn+1(x)=∫

Xρn(y)g

((x− S(y))T−1(y)

)|J |dy. (3.43)

Based on the above expression, the (stochastic) Frobenius-Perron operator for this generalform of a stochastic system with both parametric and additive terms may be defined by

Pνρ(x)=∫

Xρ(y)g

((x− S(y))T−1(y)

)|J |dy. (3.44)

It is interesting to consider special cases. Specifically, for the case of the multiplicativeperturbation, where S(x)≡ 0, we have

Pνρ(x)=∫

Xρ(y)g

(xT−1(y)

)T−1(y)dy. (3.45)

Similarly, the stochastic Frobenius-Perron operator for the additive perturbation, whereT (x)≡ 1, is

Pνρ(x)=∫

Xρ(y)g

(x− S(y)

)dy. (3.46)

Finally, the process becomes deterministic when setting the density g to a delta func-tion δ(x − S(y)). The Frobenius-Perron operator associated with the map S can then bedefined by [208],

Pρ(x)=∫

Xδ(x− S(y))ρ(y)dy. (3.47)

Thus Pρ(x) gives us a new probability density function.

Page 62: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

54 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

3.4 Invariant Density is a “Fixed Point" of theFrobenius-Perron Operator

In Eq. (1.19), we already observed that an invariant density is a solution of the Frobenius-Perron fixed point equation, which we repeat,

ρ∗(x)= Pf ρ∗(x). (3.48)

This equation may be taken as defining the term invariant density. However, the invariantdensities may exhibit singularities, which may be dense in a given subspace, and hence itis not absolutely continuous w.r.t. the Lebesgue measure. Therefore, in the situations whenit is impossible to sensibly define an invariant density function, we will alternatively dealwith the corresponding invariant measure, which can still be defined in general. Also, insuch situations, the term invariant densities will be replaced by invariant measures instead.This requires a definition of invariant measure,

Definition 3.3. Invariant measure assumes a transformation T : X → X on a measurespace (X ,�,μ) and requires μ(T−1(B)) = μ(B) for each B ∈ � - that is the “weight"ensembles of initial conditions of each B is the same before and after application of thetransformation T . Here T−1 denotes “pre-image" rather than inverse in the case the pre-image may be multiply branched. In the case of a flow, a measure μ is a φt invariantmeasure if for every measurable set B , then μ(φ−1

t (B))= μ(B).

Observe that Eq. (3.48) is a functional equation; solutions are functions, ρ∗(x). No-tice the plural use of functions. Generally, a unique solution is not expected. In particular,

• If the dynamical system f has a fixed point, x , x = f (x) then there will be an in-variant measure which is atomic (delta function) supported over this fixed point.ρ∗(x)= δ(x). See for example Fig. 3.2(Middle).

• If the dynamical system f has a periodic orbit, {x1, x2, ..., xp} then there will be aninvariant measure which is atomic supported of this periodic orbit that is a sum ofdelta functions, ρ∗(x)=∑p

i=1 δ(xi ).

• A chaotic dynamical system is characterized by having infinitely many periodic or-bits, [97], and from the above there follows infinitely many atomic invariant mea-sures.

• Linearity of the Frobenius-Perron operator allows convex combinations of any in-variant densities to be invariant densities. I.e., if

ρ∗i (x)= Pf ρ∗i (x), for i = 1, ..,q , (3.49)

then

ρ∗(x)=q∑

i=1

αiρ∗i (x), (3.50)

is invariant, choosing∑q

i=1 αi = 1 to enforce the convex combination statement.From the above, there are infinitely many convex combinations of the infinitely manyperiodic orbits.

Page 63: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.5. Invariant Sets and Ergodic Measure 55

• There can be other exotic invariant sets, such as unstable chaotic saddles which aregenerally Cantor sets. See the discussion of these sets in Sec. 7.4.3 and Def. 7.8.Each of these supports a typically atomic invariant measure, (atomic if this saddlehas Lebesgue measure zero). A picture of one for the Henon map can be found inFig. 7.28. Illustrations of such atomic invariant measures supported over unstablechaotic saddles can be found in [25], and Fig. 3.2(Right).

• There can be invariant sets which are not Lebesgue measure zero, such as in Figs. 3.9and 5.1, and each of these can support an invariant density, and likewise convexcombinations of these will be invariant densities.

We often interested in the natural invariant measure, see Eq. (3.78) for the definitionof the natural measure. Numerical estimates popularly resort to Ulam’s method [319, 106],as discussed in Chapter 4 in Secs. 4.3.1-4.4. Roughly stated for now, the invariant measuremay be estimated by computing the dominant eigenvector of the stochastic matrix estimateof the Frobenius-Perron operator.

3.5 Invariant Sets and Ergodic MeasureThe topological dynamical feature of an invariant set, and the measure theoretic conceptof ergodicity are related in spirit since invariance does have measurable consequences. Aninvariant set is a set that evolves (in)to itself under the dynamics. The general situation maynot be one of invariance, but rather one such as shown in Figure 3.3, where a set C is shownmapping across C . We sharpen the notion of invariance with the following definitions andexamples.

Definition 3.4. (Invariant set of a dynamical system) A set C is invariant with respectto a dynamical system φt if φt (C)= C , for all t ∈ R in the case of a flow, or t ∈ Z in thecase of a mapping. A set C is positively invariant if φt (C)⊂ C for all t > 0.

However, invariance does not require that points are stationary. In the case of asemidynamical system, a slightly different definition is needed, since there may be multiplepre-histories to each trajectory.

Definition 3.5. (Invariant set of a semidynamical system) A set C is invariant withrespect to a semidynamical system, (See Definitions 2.1-2.3), φt if φ−1

t (C) = C , for allt ∈ R. Note that while a semiflow is not invertible, φ−1

t denotes the preimage of a set,which may be multiply branched. 24

Example 3.3. (An Invariant Set in a Flow) The linear equation,

x1 = 2x2

x2 = x1− x2 (3.51)

has an invariant set which is the line, C = {x1, x2 ∈ R : x1+ x2 = 0}, and furthermore theorigin (x1, x2)= (0,0) is an invariant subset, but it is the only stationary point in the set C .

24The logistic map xn+1 = axn(1− xn ) is a quadratic function, and hence has up to two preimages at each

point, xn = a±√

a2−42a following the usual quadratic formula.

Page 64: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

56 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

Figure 3.2. Invariant densities of the Henon map. (Left) Density corresponding tothe apparent natural measure. (Middle) Atomic invariant density supported over one of thefixed points. (Right) An invariant density supported over an unstable chaotic saddle whichis here a Cantor set which avoids the strip labelled "Removed". Compare to Figs. 7.29-??and discussion of unstable chaotic saddles in Sec. 7.4.3. [25]

Page 65: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.5. Invariant Sets and Ergodic Measure 57

Figure 3.3. A set C is shown mapping across C, since C ∩ T (C) �= ∅, but C �=F(C). C as shown is therefore not invariant with respect to T , and points in T more so arenot stationary.

Example 3.4. (An Invariant Set in a Map) The logistic map may be the most popularexamples, for pedogogical presentation in beginning texts, as a one dimensional map whichboth presents chaotic oscillations, and admits to many of the standard methods of analysis.Already introduced here, Eq. (1.2), the logistic map may be presented as a map of the realline,

f : R→ R, (3.52)

x �→ f (x)= ax(1− x),

but the map is often presented more so in an iterating form, and as an initial value map,

xn+1 = f (xn)= axn(1− xn), x0 ∈ R. (3.53)

When a is chosen to be a=4, the logistic map can be proven [12, 97] to display fullydeveloped chaos, in the sense of a full shift symbolic dynamics. Details of this statementwill be given in Chapter 6. For the purposes of this example, for any value of a, in the moststudied parameter range a ∈ [0,4], the logistic map has an invariant set of [0,1]. That is, inthe standard function statement,

f : Domain→ Range, (3.54)

the range of the mapping is contained within the domain of the mapping which allows forrepeated iteration, and in this case they are equal.

f : [0,1]→ [0,1]. (3.55)

As for the rest of the domain, all initial conditions x0 ∈ [0,1] are in the basin of −∞.

Page 66: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

58 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

Figure 3.4. A cobweb diagram of the tent map, Eq. (3.56), when b = 2.

The same story is reflected in the tent map,

xn+1 ={

bxn, for xn < 1/2b(1− xn), for xn ≥ 1/2

}, (3.56)

which when b = 2 has fully developed chaos, also as a full shift. This tent map is infact conjugate (Definition 1.1) to the Logistic map, a = 4; this map and its invariant setsare shown in Fig. 3.4 as a cobweb diagram suggesting the invariant set, [0,1], and thatBasin(−∞)= ¯[0,1].

Example 3.5. (A Stable Invariant Set in a Flow, Slow Manifold) The following illustra-tive example found in [17] of a fast-slow system, also called a singularly perturbed system,illustrates a common scenario: strong dissipation often leads to a nontrivial invariant set,and correspondingly two time scales,

ε x =−x+ sin(y)

y = 1. (3.57)

The slow manifold here is a curve given in the singular limit. That is, choosing ε = 0,

x = h0(y)= sin(y), (3.58)

and the dynamics restricted to the slow manifold are solved to be,

x(t)= sin(y(t)), and y(t)= y0+ t . (3.59)

Page 67: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.5. Invariant Sets and Ergodic Measure 59

The so called associated system, for change of time variable to long time scales,

s = t/ε, (3.60)

changes the system Eq. (3.57),

x ′ = −x+ sin(y), y = constant, where, ′ ≡ d

ds, (3.61)

has solution,x(s)= (x0− sin(y))e−s+ sin(y), (3.62)

from which we see all solutions must converge to the slow manifold x = sin(y), as s→∞.The direct solution of the system, Eq. (3.57), can be found by the method of variation ofparameters [114], written

x(t)= [x0− x(0,ε)]e−t/ε+ x(t ,ε), y(t)= y0+ t , (3.63)

where,

x(t ,ε)= sin(y0+ t)− εcos(y0+ t)

1+ ε2 . (3.64)

For small t ,x(t)= [x0− sin(y0)]e−t/ε+ sin(y0)+O(ε)+O(t), (3.65)

by approximation of,x(t ,ε)= sin(y0)+O(ε)+O(t), (3.66)

which well approximates the associated solution, Eq (3.62). However, for larger time, theremay be drift, but e−t/ε decreases quickly. For times scales,

t = kε|logε|, (3.67)

the exponential term goes to zero faster than any power of ε, and solution Eq (3.63) be-comes close to,

x(t)= sin(y(t))− εcos(y(t))

1+ ε2, y(t)= y0+ t . (3.68)

This solution suggests solutions stay within O(ε) from the slow manifold, Eq. (3.58), andthis slow manifold approximation approximation is good for time scales,

t >> ε|logε|. (3.69)

This behavior is illustrated in Figure 3.5, where we see that solutions remain near the slowmanifold, since there is an invariant manifold x = hε (y), which further lies within O(ε) ofthe ε = 0 slow manifold, x = h0(y)= sin(y), Eq. (3.58). Furthermore, this invariant man-ifold is stable. The concepts seen here are illustrative of the sufficient conditions providedby the Tikhonov theorem [318], and reviewed in [17, 320, 94], the idea being a decomposedsystem with at least two time scales,

ε x = f1(x , y)

y = f2(x , y), (3.70)

in which one can search an invariant solution x = hε (y), nearby the slow manifold x =h0(y).

Page 68: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

60 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

Figure 3.5. A slopefield from a fast-slow system Eq. (3.57) together with severalsolutions for various initial conditions. Notice that the long time behavior is close to theslow manifold x = h0(y)= sin(y) as can be seen explicitly by comparing Eqs. (3.64)-(3.65).

3.5.1 Ergodic Measure

In our “Ergodic Preamble" and “The Ensembles Perspective", Sections 1.1 and 1.2, wehave described that considering many initial conditions leads naturally to considering theevolution of histograms of orbits. Ergodic theory may be thought of as a sister topic totopological dynamical systems, concerning itself not just with how initial conditions move,but how measured ensembles of initial conditions evolve. So again referring to Figure 3.3,we understand a measure space which describes the relative “weight" of the ensemble inC , and how that measure moves under the dynamical system. As we have already alludedin Sections 1.1 and 1.2, ergodic relates the evolution of measure under a dynamical sys-tem, by the Koopman operator Eq. (3.8), and this is dual to evolution of density by theFrobenius-Perron operator by Eq. (3.5). Therefore, in this section we will describe some ofthe simplest key terms carefully in the mathematical language of the rich and related fieldof Ergodic theory.

Page 69: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.5. Invariant Sets and Ergodic Measure 61

Definition 3.6. (Ergodic) If there exists a measure μ of a nonsingular 25 measure space,(X ,�,μ) and a nonsingular semidynamical system φt : X → X satisfying the followingproperties, then the measure is called an ergodic invariant measure corresponding to thedynamical system which is simply called ergodic since from this point on we will be dis-cussing ergodicity mostly in a situation of using invariant measures to find average26. Theseproperties are,

• μ must be a φt invariant measure.27

• Every measurable φt invariant set A measures either zero, μ(A)= 0, or its comple-ment measures zero, μ(X − A)= 0

An ergodic dynamical system is often associated with complicated behavior, andeven with chaotic behavior. However, it is not strictly correct to associate the ideas, asthese notions are separate concepts. As the following example demonstrates, the dynamicsof and ergodic transformation can in fact be quite simple.

Example 3.6. (Irrational Rotation is Ergodic) Perhaps the simplest example of an er-godic dynamical system is an irrational rotation on a circle. The circle map is defined,

f (x) : [0,1]→ [0,1]

x �→ f (x)= x+ r mod 1. (3.71)

It is easy to show the following behavior.

1. If r is rational, r = p/q , p,q ∈ Z , then every initial condition x ∈ [0,1] correspondsto an at least q-periodic orbit.

2. If r is irrational, then there are no periodic orbits, and the Lebesgue measure isergodic.

Notice that in the second case, that while the transformation is ergodic, none of the chaoticproperties are satisfied and indeed it is hard to image a simpler transformation that is nottrivial.

Example 3.7. (The Logistic Map, a = 4 is Ergodic) The much studied Logistic map,Eq. (1.2), f (x) : [0,1]→ [0,1], xn+1= axn(1− xn), when a = 4 can be shown to be ergodicsince its invariant density,

ρ(x)= 1

π√

x(1− x), (3.72)

Eq. (1.22) generates the invariant measure,

μ(A)=∫

Aρ(x)dx ,∀measurable A ⊂ X . (3.73)

25We are referring to mutually nonsingular measures. A measure μ(·) is called nonsingular with respectto another measure, usually Lebesgue m(·) measure if no other qualifier is mentioned, if there are no twodisjoint sets A and B such that all subsets of A μ-measure 0, and all subsets of B m-measure zero, butA∩ B = ∅ and A∪ B = X .

26without specifying with respect to which other measure it is ergodicU then the ergodic transformationwith respect to Lebesgue measure is meant

27The issue of existence of an invariant measure is a problem which gathers a great deal of effort. Considerfor example the Kryloff-Bogoliouboff Theorem - any continuous mapping of a metrizable compact space toitself has an invariant Borel measure [305].

Page 70: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

62 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

This example is perhaps more stereotypical of the concept of ergodic since here the propertyof ergodic coincides with a complicated behavior which is even chaotic in this case. Whena < 4 there are many specific transformations corresponding to what is believed to be bothchaotic and also ergodic behavior, at least for those accumulation points in the Feigenbaumdiagram

Example 3.8. (An example where the invariant set is Lebesgue measure zero) Consideragain the tent map f (x) : [0,1]→ [0,1], xn+1= axn if xn < 1/2 and a−axn if x ≥ 1/2. It isstraight forward to show that if a > 2, then there is an interval of points of initial conditionsx0 ∈ [1/a,1−1/a] such that the orbit of those points leave [0,1] in one iteration; consider-ing the cobweb diagram in Fig. 3.4, and also comparing a similar scenario shown in Fig. 5.2exemplifies this statement. Likewise, (up to two) preimages of this set f −1[1/a,1− 1/a]is a set of points which leave [0,1] in two iterates. And f −2[1/a,1− 1/a] is up to foursubintervals which leave [0,1] in three iterates. Following this discussion indefinitely con-structs a Cantor set which is the invariant set � of [0,1]. The measure of this invariantset is m(�) = limn→(1− 2/a)n = 0. This set � supports an ergodic measure μ which isnot absolutely continuous to Lebesgue measure and topologically the description of thedynamics on � is chaotic in the sense that it is conjugate to a shift map. Such sets are oftencalled chaotic saddles, [253, 28]. Perhaps less exotic, an atomic invariant measure existsfor each periodic orbit, and these correspond to delta functions supported over the periodicpoints. There are infinitely many of these.

A central idea behind the importance of ergodic transformations is that time averagesexchange with space averages in the long time limit. A historical origin of the ergodichypothesis come from statistical physics and the study of the thermodynamics and states ofgases.28 See the Birkhoff ergodic theorem Eq. (1.5) and Examples 1.3 and 1.4.

The question of unique ergodicity is nontrivial. The following simple example hastwo invariant components and it is therefore definitely not ergodic.

Example 3.9. (Map with Two Invariant Components)The map of the interval, f : [0,2]→ [0,2] shown in Fig. 3.9 has an invariant measure

which is uniform, corresponding to the density ρ(x) = 1/2 on [0,2]. However, it alsohas two other nontrivial invariant measures generated by each of the following densitiesrespectively,

ρ1(x)=(

1 if 0≤ x ≤ 10 else

),ρ2(x)=

(1 if 1≤ x ≤ 20 else

). (3.74)

Since the invariant set X = [0,2] measures with each of these invariant densities,

μρ1 (X)=∫ 2

0ρ1(x)dx = μρ2 (X)=

∫ 2

0ρ2(x)dx = 1

2, (3.75)

28In statistical physics, it is said that in the long time study, passing to mean field, particles are saidto spend time in microstates of the phase space with energy which is proportional to the volume, thuscorresponding to equal probability. Boltzmann’s ergodic hypothesis developed in the 1870’s in his studiesof statistical mechanics, initially intuitively by simply disregarding classical dynamics in studies of gastheory of many particles [44]. A statement of the Boltzmann’s ergodic hypothesis may be taken as, Largesystems of dynamically interacting particles display a state in the long run where time averages approximateensemble equilibrium averages, and he called such systems “ergode." Gibbs [205] later called such statesthe canonical ensemble in his development of what is nearly a modern classical thermodynamics theory. TheBirkhoff ergodic theorem encompasses an essentially modern formulation of this idea [21].

Page 71: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.5. Invariant Sets and Ergodic Measure 63

Figure 3.6. A tent-like map with two invariant components in Example 3.9. Con-trast to the example in Fig. 5.2 in which this systems develops into one with a single invari-ant measure but with long transients, also known as weakly transitive in Sec. 5.2. Compareto Figs. 5.1-5.2 in Example 5.1.

this contradicts that these may be ergodic since ergodic measures by definition must mea-sure invariant sets to be either zero or one.

3.5.2 Almost Invariant Sets

While invariance is purely a topological concept, with ergodic consequence when measurestructure is assumed, the concept of almost invariance becomes essentially a measurableconcept, since it asks the question, relatively how much of a set may remain in place un-der the action of the dynamical system. Any question with the phrase, “how much" in itrequires a measure structure to discuss. To quantify almost-invariant sets we will formallydefine a notion of the almost-invariant set based on the Markov model (4.6).

Definition 3.7. For a dynamical system, consider a flow St : X → X , a Borel measurableset A⊂ X is called μ-almost invariant [89] if

μ(A∩ S−1t (A))

μ(A)≈ 1, (3.76)

where μ is a probability measure and S−1t denotes the preimage when an inverse does not

exist.

Page 72: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

64 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

While we see that the judgement of almost invariance depends on the flow period tand the choice of measure, that is, how we weight sets, there are two measures in particularthat may be most sensible to choose in most discussions of almost invariance. One choiceof course is to let μ to be Lebesgue measure,

μ(A)≡ m(A)=∫

Adx , (3.77)

when appropriate. In such case, simple Monte Carlo simulations often lead to useful resultswhich are easy to interpret. Another sensible choice that is true to the behavior of thedynamical system is the so-called natural measure, which may be described as the invariantmeasure on the attractor. These words that need some explanation.

Roughly speaking, the natural measure μ(B) of a measurable set B is high if theamount of time the orbit of Lebesgue almost-all points x ∈ X spends in B is large. As such,μ(B)=0 if no orbits enter B or revisit B after a certain number of iterations. When the limit,

μ(B)= limN→∞

#{Fn(x) ∈ B : 0≤ n ≤ N}N

. (3.78)

exists29 for a mapping F , we call it the natural measure of the set B . More precisely, itcan be called a natural invariant measure as it is evident from its definition that if a naturalmeasure exists, it must also be invariant, but we will, however, refer to it just a naturalmeasure for short. This is also called a rain gauge measure [2] and, despite the notoriouslydifficult theoretical nuances involved to prove existence, the idea is quite simple to usein practice quite simple to use, but the result can be quite misleading in the scenario ofvery long transients. In practice, an initial point is selected uniformly at random. Thena trajectory through the space is computed by iterating the system some large number oftimes, and then estimate the limiting occupancy fraction behavior. In fact, this is also areason that a natural measure is meant to be carried by an attractor where the fractionof iterates falling into a set in the attractor is the same for almost initial points (w.r.t theLebesgue measure) in the basin of attraction.

When the above measure is used to define the almost invariant set, as expressed inEq. (3.76), the interpretation is that if A is an almost invariant set then the probability thata point in A remains in A after one iteration is close to 1. Generally, the definition of thenatural measure is related to the Birkhoff ergodic theorem highlighted in Eq. (1.5).

Remark 3.3. What if a system has several nontrivial invariant sets and how does this relateto ergodicity? Recall that a measure is called invariant under St if μ(S−1

t (A))=μ(A) for allmeasurable sets A. But an invariant measure is defined to be ergodic if μ(A)= 0 or μ(A)=1 for all measurable sets A such that St (A) = A. This notion of ergodicity emphasizesindecomposibility of a space to be studied. In short, if a space comprises of more than oneinvariant set of positive measure, then one could study them separately. Several variousdescriptions of ergodicity and its relation to the notion of topological transitivity can befound in [315].

29We are avoiding here the detailed discussion as to when such measures do in fact exist, and for whichkinds of systems there are attractors on which such measures exist. See [336, 181] for discussion regardingexistence of SRB measures and the beginning of discussion of construction of “rain guage" measures in [2]such as Eq. (3.78).

Page 73: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.5. Invariant Sets and Ergodic Measure 65

3.5.3 Almost invariant sets as approximated by stochastic matrices

The fact is that for many dynamical systems, the asymptotic notion of the natural measurein Eq. (3.78) can be difficult to compute for several reasons. First, it is possible that a dy-namical system has a very long transient state and consequently we need to compute a longsequence of iterations to observe eventual behavior of such a system. Therefore, we maynot be able to observe the equilibrium distribution unless t in (3.78) is sufficiently large.Furthermore, in an extreme case, the round-off problem can prevent us from discoveringthe equilibrium distribution; for example, a sequence of iterations of the tent map Eq. (3.56)when b = 2 will send almost any initial point to negative infinity as is easy to confirm bysimulations. A way to circumvent these problems is to find the invariant measure basedon the left eigenvector with the eigenvalue one from the matrix P defined in Eq. (4.6) anddiscussed further in the next chapter.

More precisely, the invariant measure of the transition matrix Eq. (4.7) is a good ap-proximation of the natural measure of the Frobenius-Perron operator PSt .

30 Note that adeterministic chaotic dynamical system will have infinitely many invariant measures, onefor each invariant set such as periodic orbits and convex combinations of these measures.However, the question is if there is only one natural measure. In general, there is no guar-antee that the invariant density we discovered is the natural measure. In this regard, it wasshown [130, 232] that the supports of the approximate invariant measures contain the sup-port of the natural measure. So the consequence of such result is that at least we capture allthe regions with the positive natural measure.

Now consider invariant density as found from the transition matrix Markov modelapproximating the action of the Frobenius-Perron operator. Let p be the left eigenvector ofP with eigenvalue one. After normalizing p so that,

n∑i=1

pi = 1, (3.79)

one may approximate the natural measure μN for {Bi }Ni=1 by,

μN (Bi )= pi . (3.80)

Then the approximate invariant measure of a measurable set B can be defined by

μN (B) :=N∑

i=1

μN (Bi )m(Bi ∩B)

m(Bi ). (3.81)

Comparing Eq. (4.6) with Eq. (3.80) we see that it is just a restatement of the matrix timesvector statement pA. Note that the quality of this approximation with respect to a refin-ing topological partition {Bi }Ni=1 depends upon the number and size of partition elements.Results concerning the convergence of such approximations can be found in [89].

By combining Eqs. (4.6) and (3.81), one can alternatively define almost invariancebased on the estimated invariant density p and the transition matrix P , as in the followingLemma [131].

30Approximation of the action of the Frobenius-Perron operator by a sequence of refinements and cor-responding stochastic matrices is the subject of Chapter 4. In particular, the Ulam’s method [319] is theconjecture that this process can be used to compute the invariant density.

Page 74: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

66 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

Lemma 3.1. [131] Given a box partition set {Bi }Ni=1 with μN (Bi )= pi , a Borel measurableset P=⋃i∈I Bi for some set of box indices I is called almost invariant if∑

i, j∈Ipi Pi j∑

i∈Ipi

≈ 1. (3.82)

Proof. It is easy to see from (3.81) that μN (Bi )= pi and hence we have

μN (B)=∑i∈I

pi . (3.83)

Now observe that for each fixed i ∈ I, we have

μN (Bi ∩ S−1t (Bj ))=

∑j∈I

m(Bi ∩ S−1t (Bj ))

m(Bi )pi

=∑j∈I

Pij pi .(3.84)

The desired result follows after summing the above equation over all i .

3.6 Relation between the Frobenius-Perron and KoopmanOperators

In this section we will show the connections between the Frobenius-Perron and Koopmanoperators beyond their adjointness, which was already developed. First, we will demon-strate that while the Frobenius-Perron operator acts as a "push forward" operator for adensity function, the Koopman operator may be intuitively thought of as a "pull back"of a function. To make this description more precise, let consider a semidynamical sys-tem {St }t>0 on (X ,�,μ) and some A ⊂ � such that St (A) ∈ �. If we take f (x) = 0 forx ∈ X− A and g(x)= χX−St (A), then we have f ∈ L1 and g ∈ L∞. By the adjoint propertyof the two operator, we have∫

XχX/St (A) Pt f (x)dμ=

∫X

f (x)χX/St(A)(St f (x))dμ

=∫

AχX/St (A)(St f (x))dμ= 0.

(3.85)

This means that the integrand on the left hand side of the above equation must be zeroalmost everywhere, which happen if and only if

Pt f (x)= 0 for x ∈ X/St (A). (3.86)

In other words, Pt "pushes forward" a density function supported on A to a function sup-ported on St (A).

Conversely, if we take f (x)= 0 for x ∈ X − A again but consider it as f ∈ L∞, wehave

Kt f (x)= f (St (x))= 0 if St (x) ∈ X − A. (3.87)

Page 75: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

3.6. Relation between the Frobenius-Perron and Koopman Operators 67

This means that whenever f is zero outside a set A, then Kt f (x) is zero outside a setS−1

t (A). Therefore, Kt “pulls back" the function supported on A to a function supportedon S−1

t (A). 31

Now we contrast the dynamical description exhibited by Pt and Kt . Recall that thefunction f (t , x)= Pt f (x) satisfies the continuity equation, see (3.32), which describes theevolution of a density function. It is well know that first-order PDE of the form (3.32) canbe solved by the method of characteristics. This methods gives a solution of the densityalong the solution of the initial value problem (3.2). Let x(t) denote a solution of (3.2) withx(0)= x0 and define ρ(t)= f (t , x(t)) by parameterizing x by t . By applying the chain rule,the function ρ(t) must satisfies

dt= ∂ f (t , x(t)

∂ t+

d∑i=1

∂ f (t , x(t))

∂xi

dxi

dt

= ∂ f (t , x(t)

∂ t+∇( f ) · F(x).

(3.88)

Also, f (t , x) obeys the continuity equation:

∂ f

∂ t=−∇ · ( f F)=−∇(F) f −∇( f ) · F . (3.89)

Comparing the above two equations suggests that

dt=−∇(F) f . (3.90)

This means that for a given X(0) = x0 and ρ(0) = f0(x0), the continuity equation can besolved pointwise by solving the initial value problem of d+1 dimensional ODE system

dx

dt= F(x)

dt=−∇ · (F) f .

(3.91)

The negative sign in the above equation should be intuitive for if we start with an infinitesi-mal parcel of particles, as the parcel moves along with the flow its volume will be expanded,say ∇ · (F) > 0, which results in a lower density parcel.

Furthermore, the right hand side in the expression (3.88) can be expressed anotherway as

dt= ∂ f (t , x(t)

∂ t+ AK . (3.92)

31It is important to recall that the Pt acts on a density, which is defined in L1, but Kt is defined on L∞.It is not always legitimate to think of Kt as an operator that generally push a density backward in time. Thiscan be easily noticed by comparing Kt f (x)= f (St (x)) with Pt f (x)= f (S−1

t (x))|Dx S−1t (x)|. Clearly, the

Jacobian term is not included in the definition of Kt and hence it can be regarded as an operator that pullsback a density only if the Jacobian equals to unity everywhere, that is, the semidynamical system is area-preserving (∇ · (F) = 0), in which case the the Koopman operator becomes isometric and it can be definedon L P for 1≤ P ≤∞.

Page 76: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

68 Chapter 3. Frobenius-Perron Operator and Infinitesimal Generator

The quantity dρdt in the above expression is called the “Lagrangian" derivative (it is also

called variously the “material", the “substantial", or the “total" derivative). It has a physicalmeaning of the time rate of change of some characteristics (represented by ρ) of a partic-ular element of fluid (i.e. x0 in the above context). 32 At this point, we may think of theFrobenius-Perron operator as a “Eulerian PDF" method since its corresponding continu-ity equation has to be solved on a fixed coordinate. In contrast, the Lagrangian derivativederived from the Koopman operator by using a reverse time−t to obtain−Ak yields a “La-grangian PDF" method, which has to be solved simultaneously with the flow for a specificparticle x0.

Finally, it is worth to mention again that we introduce the concept of the Frobenius-Perron operator in order to utilize it to identify almost invariant sets in subsequent chapters.In particular, it will be shown that the information embedded in the spectrum of Frobenius-Perron operator is useful for our purpose. Nevertheless, it is not surprising that the Koop-man operator, as an adjoint operator of the Frobenius-Perron operator, can also be usedfor the same purpose and it yields similar results. There have been extensive works onusing the Koopman operator to identify invariant sets, which can be primarily referred toMezic [231, 230]

32If the the flow is area-preserving, it is easy to see that dρdt = 0. This can be interpreted that the physical

quantity described by ρ is conserved along the flow.

Page 77: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 4

Graph Theoretic Methods andMarkov Models of DynamicalTransport

The topic of this chapter stems from the idea that Frobenius Perron operators can be un-derstood as if it were infinitely large stochastic matrices acting on an infinite dimensionallinear space. Furthermore, there are finite rank (corresponding to finite sized matrices)representations that can give excellent results. Such is the story of compact operators,33

and this leads not only to better understanding of the operator, but most importantly thisleads to computable methods for carrying forward a practical basis for numerics on digitalcomputers.

We use what may be called the Ulam-Galerkin method - a specialized case of theGalerkin’s method [211] - to approximate the (perhaps stochastic) Frobenius-Perron opera-tor (3.44). In this chapter, we flesh out the discussion mentioned early that the approximateaction of dynamical system on density looks like a directed graph, and that Ulam’s methodis a form of Galerkin’s method. To hopefully offer some insight, we again refer the readerto the caricature partitioning of the action of the Henon mapping in Figure 1.1.

The approximation by Galerkin’s method is based on the projection of the infinitedimensional linear space L1(M) with basis functions,

{φi (x)∞i=1} ⊂ L1(M), (4.1)

onto a finite-dimensional linear subspace with a subset of the basis functions,

�N = span{φi (x)}Ni=1. (4.2)

For the Galerkin method, the projection

�N : L1(M)→�N , (4.3)

maps an operator from the infinite-dimensional space to an operator of finite rank, an N×N

33A compact operator is defined in the field of functional analysis in terms of having an atomic spectrum[8]. Compact operators are most easily described in a Hilbert space [201] (a complete inner product space)as these are the operators that are the closure of operators of finite rank. In other words, their action is “wellapproximated" by matrices, approximated in the appropriate sense.

69

Page 78: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

70 Chapter 4. Graph Theory and Markov Models of Transport

matrix, by using the inner product,34

Pi, j = 〈PFν (φi ),φj 〉 =∫

MPFν (φi (x))φj (x)dx . (4.4)

The advantage of such a projection is that the action of the Markov operator whichis initially a transfer operator in infinite dimensions reduces approximately to a Markovmatrix on a finite dimensional vector space. Such is the usual goal of a Galerkin’s methodin PDE’s, and similarly it is used here in the setting of transfer operators.

Historically, “Ulam’s conjecture" was proposed by S. Ulam, [319], in a broad col-lection of interesting open problems from applied mathematics including the problem ofapproximating Frobenius-Perron operators. His conjecture has been essentially proved ascited below, but referred to both,

Ulam’s Conjecture [319]

1. A finite rank approximation of the Frobenius-Perron operator by Eq. (4.6) and,

2. The conjecture that the dominant eigenvector (corresponding to eigenvalue equalto 1 as is necessary for stochastic matrices) weakly approximates35 the invariantdistribution of the Frobenius-Perron operator.

Ulam did not write his conjecture in the formal language of a Galerkin projection Eqs. (4.1)-(4.4), but due to the equivalence to such, we will use the phrase Ulam-Galerkin matrix torefer to any projection of the Frobenius-Perron operator by an inner product as in Eq. (4.4),not necessarily including the infinite time limit part of the conjecture regarding steady state,item 2.

Ulam’s method is often used to describe the process of using Ulam’s conjecture,by developing what we call here the Galerkin-Ulam matrix, and then using the dominanteigenvector of this stochastic matrix to estimate the invariant density. Sometimes however,the phrase Ulam’s method is used to simply describe what we call here the developing ofGalerkin-Ulam’s matrix. Some discussion of computional aspects of the Galerkin-Ulammatrix, and the Ulam’s method can be found in Appendix, A. As we will see, the one stepaction of the transfer of the operator is well approximated by Ulam-Galerkin matrices. Theanalysis to describe the approximation of the one step action is much simpler than that ofthe infinite limit referred to in the Ulam’s method, and other issues such as decompositionof the space into almost invariant sets is naturally approximated as well by the short timerepresentation. Also, in a special case the approximation is in fact exact and such approx-imations can be shown to be exact, as discussed in Section 4.4. This exact representationmay occur when the dynamical system is Markov. Ulam conjecture [319] has been provenin the special case of one-dimensional maps by methods of bounded variation, [211]. Inhigher dimensional dynamical systems, a rigorous footing of Ulam’s conjecture is incom-plete, except for special cases, [51, 103, 105, 122, 126, 127] and it remains an active area of

34Our use of the inner product structure requires the further assumption that the density functions are inthe Hilbert space L2(M), rather than just the Banach space L1(M), which using the embedding L2(M) ↪→L1(M), provided M is of finite measure.

35Weak approximation by functions may be defined as convergence of the functions under the integral

relative to test functions. That is, if { fn}∞n=1 ∈ L1(M), it is defined that fnw→ f ∗ if limn→∞

∫M | f ∗(x)−

fn(x)|h(x)dx = 0 for all h ∈ L∞(M) which is referred to as the test function space.

Page 79: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.1. Finite-rank approximation of the Frobenius-Perron operator 71

research, and we should also point out recent developments in a non uniformly expandingsetting, [246]. Nonetheless in practice it is easy and common to simply proceed to use thedominant eigenvector of the stochastic matrix and then to refer to it as an approximation ofthe invariant density. See Section 4.2, and compare to Fig. 1.1.

4.1 Finite-rank approximation of the Frobenius-Perronoperator

The quality of the Ulam-Galerkin approximation is discussed in several references, as isthe convergence of the Ulam’s method, [125, 123, 60, 104, 180, 52, 35] amongst many. Itis straight forward to cast the Ulam’s method as a Galerkin’s method with a special choiceof basis functions as follows. For the Ulam’s method, the basis functions are chosen to bea family of characteristic functions,

φi (x)= χBi (x)= 1 for x ∈ Bi and zero otherwise. (4.5)

Generally, Bi is chosen to be a simple tiling of the region of interest in the phase space,meaning some region covering a stable invariant set such as an attractor. For convenience,Bi may be chosen as a simple covering of rectangle boxes, but we have also had success inusing triangular tesselations using software packages often affiliated with PDE based finiteelement methods technology. In the deterministic case, using the inner product Eq. (4.4)the matrix approximation of the Frobenius-Perron operator has the form of

Pi, j = m(Bi ∩ F−1(Bj ))

m(Bi ). (4.6)

where m denotes the normalized Lebesgue measure on M and {Bi}Ni=1 is a finite family ofconnected sets with nonempty and disjoint interiors that covers M. That is M = ∪N

i=1 Bi ,and indexed in terms of nested refinements [319]. These Pi, j can be interpreted as the ratioof the fraction of the box Bi that will be mapped inside the box Bj after one iteration of amap to the measure of Bi .

Note that one may consider this matrix approximation of the Frobenius-Perron oper-ator as a finite Markov chain, where the partition set {Bi}Ni=1 represents a set of “states" andPij characterizes transition probabilities between states. It is well-known that the matrix Pin (4.6) is stochastic and has a left eigenvector with the eigenvalue one. Simply put, thiseigenvector characterizes the equilibrium distribution of the Frobenius-Perron operator. Infact, it can be proven, [53] that if the partition {Bi}Ni=1 is a Markov partition, then the(unique) left eigenvector of the matrix P defines a good approximation of the equilibriumdistribution, a statement that will be made precise in the next subsection. This leads to astraight forward way to understand the approximation theory for the generic non-Markovcase, by approximation using Markov representations. Definition of the phrase Markovpartition will be the subject of the very next section, Sec. 4.2.

However first we note a more readily computable experimental perspective to standin for Eq. (4.6), essentially by a Monte-Carlo sampling.

Remark 4.1. A key observation is that the kernel form of the operator in Eq.(3.44) allowsus to generally approximate the action of the operator with test orbits as follows. If we

Page 80: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

72 Chapter 4. Graph Theory and Markov Models of Transport

Figure 4.1. An Ulam-Galerkin’s method approximation of the Frobenius-Perronoperator is described by the action of a graph, as estimated by Eq. (4.7). Here we see abox i maps roughly across j , j + 1, and j + 2. As such a graph G A generated by matrixA would have an edge between vertex i and each of j , j + 1, and j + 2. The matrix A isformally described by the inner product, Eq. (4.4) where the basis functions are chosen tobe characteristic functions χk(x) supported over each of the boxes in the covering, includ-ing j , j + 1, j + 2, and i . As shown, the T ( j ) does not cover the i boxes in a way thatallows a Markov partition, and thus the lost measure causes the finite rank to be only anapproximation.

only have a test orbit {xj }Nj=1, which is actually the main interest of this work, the Lebesguemeasure can be approximated by

Pi, j ≈ #({xk|xk ∈ Bi and F(xk) ∈ Bj })#({xk ∈ Bi }) . (4.7)

This statement can be made precise in terms of quality of the approximation and numberof sample points by Monte-Carlo sampling theory for integration of the inner product,Eq. (4.4). See code A.1.1 and demo in Appendix, A for a Matlab implementation of thisorbit sampling method of developing a Galerkin-Ulam matrix.

See Fig. 4.1 for a graphical description of the situation of the approximations de-scribed in the previous paragraphs. Compare this to Fig. 4.2 which is developed usingEq. (4.7) in the case of a Henon map, and a rather course partition for sake of illustra-tion. Then recording the relative transition numbers according to Eq. (4.6) as estimatedby Eq. (4.7) leads to stochastic matrices as presented by the transition matrix shown inFig. 1.1. See Figs. 4.2-4.3 for illustrations of the set oriented methods reflected by theUlam-Galerkin approach in realistic systems, the Henon map and the flow of the Gulf re-spectively.

Page 81: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.2. The Markov Partition: How it Relates to the Frobenius-Perron Operator 73

Figure 4.2. The Ulam-Galerkin’s approximation of the Frobenius-Perron op-erator in its simplest terms is a “set oriented method" in that how points in sets map toother sets is recorded and this is used as an approximation of the action of the dynamicalsystem on points. Compare this figure showing a covering of boxes Bi which will be usedto estimate the action of the Henon map dynamical system in those cells Bi in which theattractor is embedded (yellow). Compare this covering to the action of a Henon mappingso estimated and shown in Fig. 1.1. The underlying estimates of the stochastic matrix aresummarized by the computations shown in Eqs. (4.6) and (4.7). Compare to Figs. 4.1, 4.3.

4.2 The Markov Partition: How it Relates to theFrobenius-Perron Operator

While generically most dynamical systems are not Markov, meaning that they do not admita Markov partition, and most partitions for such systems will not be Markov partitions,when we do have such, then the corresponding Frobenius-Perron operator can be exactlyrepresented by an operator of finite rank. We will give the necessary technical details andinterpretations here. In the next section, we will show how this perspective of Markov par-titions can be used to formulate a notion of approximation in the non Markov case. So aswe will see, besides having an important roll in a symbolic dynamics (from topological dy-namics) of a dynamical system where the concept of a generating partition of the symbols issignificantly simplified when there is a Markov partition, in measurable dynamics, Markovpartitions allow for a greatly simplified finite rank description of the Frobenius-Perron op-

Page 82: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

74 Chapter 4. Graph Theory and Markov Models of Transport

Figure 4.3. The Ulam-Galerkin’s approximation of the Frobenius-Perron startswith study of evolution of an ensemble of initial conditions from a single sell. Here a(rather large for artistic reasons) box Bi in a flow developed from a oceanographic modelof the fluid flow in the Gulf reveals how a single square progressively becomes swept in theGulf stream. The useful time scale is one that reveals some no trivial dynamic evolution,stretching across several image cells, but not so long that a loss of correlation may occur.Compare this image to a similar image developed for the Henon mapping, Figs. 4.1-4.2,and stochastic matrix estimates in Eqs. (4.6) and (4.7). operator in its simplest terms isa “set oriented method" in that how points in sets map to other sets is recorded and thisis used as an approximation of the action of the dynamical system on points. Comparethis figure showing a covering of boxes Bi which will be used to estimate the action of theHenon map dynamical system in those cells Bi in which the attractor is embedded (yel-low). Compare this covering to the action of a Henon mapping so estimated and shownin Fig. 1.1. The underlying estimates of the stochastic matrix are summarized by the com-putations shown in Eqs. (4.6) and (4.7). Compare to a similar presentation in the Duffingoscillator, Fig. 1.16.

erator. Hence the computation of relevant statistical measures is greatly simplified.

4.2.1 More Explicitly, Why a Markov Partition?

To simplify analysis of a dynamical system, we often study a topologically equivalent sys-tem using symbol dynamics, representing trajectories by infinite length sequences using afinite number of symbols. (An example of this idea is that we often write real numbers assequences of digits, a finite collection of symbols.) Symbolic dynamics will be discussedin some detail in the next chapter.

To represent the state space of a dynamical system with a finite number of symbols,

Page 83: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.2. The Markov Partition: How it Relates to the Frobenius-Perron Operator 75

we must partition the space into a finite number of elements and assign a symbol to eachone. In probability theory, the term “Markov" denotes a “finite memory" property. Inother words the probability of each outcome conditioned on all previous history is equalto conditioning on only the current state; no previous history is necessary. The same ideahas been adapted to dynamical systems theory to denote a partitioning of the state space sothat all of the past information in the symbol sequence is contained in the current symbol,giving rise to the idea of a Markov transformation.

4.2.2 Markov Property of One-Dimensional Transformations

In the special, but important case that a transformation of the interval is Markov, the symboldynamic is simply presented as a finite directed graph. A Markov transformation in R1 isdefined as follows: [53].

Definition 4.1. Let I = [c,d] and let τ : I→ I . Let P be a partition of I given by the pointsc = c0 < c1 < .. . < cp = d . For i = 1, . . . , p, let Ii = (ci−1,ci ) and denote the restrictionof τ to Ii by τi . If τi is a homeomorphism from Ii onto a union of intervals of P , then τ

is said to be Markov. The partition P is said to be a Markov partition with respect to thefunction τ .

Figure 4.4. (a) A Markov map with partition shown. (b) The transition graph formap a. (c) The partition is not Markov, because the image of I2 is not equal to a union ofintervals of the partition.

Example 4.1. One-Dimensional Example Map 1, (Fig 4.4a) is a Markov map with theassociated partition {I1, I2, I3, I4}. The symbol dynamics are captured by the transitiongraph (Fig 4.4b). Although Map 2 (Fig 4.4c) is piecewise linear and is logically partitionedby the same intervals as map 1, the partition is not Markov because interval I2 does not maponto (in the mathematical sense) a union of any of the intervals of the partition. However,we are not able to say that the map 2 is not Markov. There may be some other partition thatsatisfies the Markov condition. [40]

4.2.3 Markov Property In Higher Dimensions

Definition 4.2. A Topological Partition of a topological space (M ,τ ) is a finite collectionP = {P1, P2, ..., Pr } of disjoint open sets whose closures cover M in the sense that M =

Page 84: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

76 Chapter 4. Graph Theory and Markov Models of Transport

P1∪ ...∪ Pr .

Definition 4.3. A Topological Space (M ,τ ) is a set M together with the set of subsetsτ ⊂ 2M 36 that are defined to be open; as such τ must include the empty set ∅ and all of M ,and τ must be closed under arbitrary unions and finite intersections, [245].

Any topological partitioning of the state space will create a symbol dynamics for themap. In the special case where the partition is Markov, the symbol dynamics capture theessential dynamics of the original system.

Definition 4.4. Given a metric space M and a map f : M→ M , a Markov Partition of Mis a topological partition of M into rectangles {R1, ..., Rm} such that whenever x ∈ Ri andf (x) ∈ Rj , thenf [Wu (x)∩ Ri ]⊃Wu[ f (x)]∩ Rj ) and f [Ws (x)∩ Ri ]⊂Ws [ f (x)]∩ Rj . [48, 46].

In simplified terms, this definition says that whenever an image rectangle intersects apartition element, the image must stretch completely across that element in the expandingdirections and must be inside that partition element in the contracting direction. (See Fig4.5.) [40]

Figure 4.5. In the unstable (expanding) direction, the image rectangle muststretch completely across any of the partition rectangles that it intersects.

4.2.4 Generating Partition

It is important to use a “good" partition so that the resulting symbolic dynamics of orbitsthrough the partition well represents the dynamical system. If the partition is Markov,then “goodness" is most easily ensured. However, a broader notion, called generatingpartition, may be necessary to capture the dynamics. A Markov partition is generating, butthe converse is not generally true. See [284, 41] for a thorough discussion of the role ofpartitions in representing dynamical systems.

362M denotes the “power-set" of M meaning it is the set of all subsets.

Page 85: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.2. The Markov Partition: How it Relates to the Frobenius-Perron Operator 77

Definition 4.5. Given a topological space (M ,τ ) (Definition 4.3) and a Topological Parti-tion P = {P1, P2, ..., Pr } is a Topological Generating Partition for a mapping T : M→ Mif,

τ = ∨∞i=0T−iP , (4.8)

(or require τ =∨∞i=−∞T−iP , for invertible dynamical systems).

As usual, T−i denotes the i th pre-image (possibly with many branches), but it is thei th composition of the inverse map if the map is invertible. This definition is in terms ofthe join of partitions, defined recursively starting from

Definition 4.6. The join of a two partitions, P and P ′, is defined,

P ∨P ′ = {Pk ∩ P ′l : 0≤ k ≤ |P |−1,0≤ l ≤ |P ′|−1}, (4.9)

Thus terms such as T−iP in the definition of the generating partition are joined withother iterates of the original partition, T− j P . The idea of a generating partition is that thisprocess of joining many iterates of open sets creates collections of open sets. Proceedinginfinitely in this manner creates infinitely many open sets, and the question is if all of theopen sets in the topology are generated.

If furthermore, a measure space is assumed (M ,A,μ), where A is defined as theBorel sigma algebra of sets which are μ-measurable, then the question a generating parti-tion becomes,

Definition 4.7. Given a measure space (M ,A,μ) , then a Topological Partition P ={P1, P2, ..., Pr } of measurable sets is a Measurable Generating Partition for a mappingT : M→ M if,

A= ∨∞i=0T−iP , (4.10)

(or require τ = ∨∞i=−∞T−iP , for invertible dynamical systems.) We require that all themeasurable sets are generated.

Example 4.2. Two-Dimensional Example - Toral AutomorphismThe Cat Map, defined by

x= (Ax) mod 1 (4.11)

where A=[

2 11 1

](4.12)

yields a map from the unit square onto itself. This map is said to be on the toral space T2

because the mod1 operation causes the coordinate 1+ z to be equivalent to z. A Markovpartition for this map is shown in Fig 4.6. The Cat Map is part of a larger class of functionscalled toral Anosov diffeomorphisms, and [279] provides a detailed description of how toconstruct Markov partitions for this class of maps. [40]

Example 4.3. Generating Partition of the Henon-Map. Consider again the Henon map,T (x , y)= (1−ax2+ y,bx), (a,b)= (1.4,0.3), the prototypical diffeomorphism of the planewith the famous strange attractor. See Fig. 6.31 in which a piecewise linear curve C pro-duced as connecting tangencies of stable and unstable manifolds, according to the wellregarded conjecture [73, 154], produces what is apparently a generating partition but not aMarkov partition. See also the generating partition discussion of the Ikeda map, Eq. (9.87)in Example 9.3 shown in Fig. 9.7.

Page 86: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

78 Chapter 4. Graph Theory and Markov Models of Transport

Figure 4.6. The Cat Map is a toral automorphism. (a) The operation of the linearmap on the unit square. (b) Under the mod operation, the image is exactly the unit square.(c) Tesselation by rectangles R1 and R2 forms an infinite partition on R2. However, sincethe map is defined on the toral space T2, only two rectangles are required to cover thespace. The filled gray boxes illustrate that R1 and R2 are mapped completely across aunion of rectangles.

4.3 The Approximate Action of Dynamical System onDensity looks like a Directed Graph: Ulam’s Methodis a form of Galerkin’s Method

The title of this section says that the approximate action of dynamical system on a densitylooks like a directed graph: Ulam’s method is a form of Galerkin’s method. This is aperspective to which we already discussed theory of Galerkin’s method in Sec. 4.1. In fact,as stated above, when the dynamical system is Markov and using the Markov partition andthe corresponding basis functions which are supported over the elements of that Markovpartition, the action of the dynamical system is exactly represented by a directed graph.In this case, the inner product form, Eq. (4.4) becomes exactly, Eq. (4.6), resulting ina stochastic matrix A whose action and steady state are both discussed by the Perron-Frobenius theorem. In this section, we will summarize these statements more carefully, butfirst we motivate with the following examples.

Example 4.4. (Finite Rank Transfer Operator of a One-Dimensional Transformation)The map shown in Fig. 4.4(a) was already discussed to be a Markov map in the interval,with a Markov partition, {I1.I2, I3, I4} as shown, and according to the definition of Markovpartition shown in Fig. 4.4.

In this piecewise linear case, it is easy to directly integrate the Galerkin method,Eq. (4.4) integrals when choosing the basis functions to be one of each of the four charac-teristic functions37, {χI1 (x),χI2 (x),χI3 (x),χI4 (x)}. Let,

φi (x)= χIi (x), i = 1,2,3,4. (4.13)

For sake of example, we will explicitly write one such integral here. From the drawing,writing the function in Fig. 4.4(a) explicitly, for simplicity assuming the uniform partition

37A characteristic function, also called an indicator function, takes on the value one inside the argumentset, χA(x) : M→ {0,1}, χA(x)= i f (x ∈ A,1,0)

Page 87: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 79

shown,Ii = [i −1, i ], i = 1,2,3,4, (4.14)

and that∪4

i=1 Ii = [0,4], (4.15)

then F : [0,4]→ [0,4] may be written,

F(x)=

⎧⎪⎨⎪⎩3x if x < 1−x+4 if 1≤ x < 22x−2 if 2≤ x < 3−4x+16 else x ≥ 3

⎫⎪⎬⎪⎭ (4.16)

Now we may write and evaluate the inner-product Eq. (4.4), to derive each of theelements of the 4×4 matrix A which describes the relative movement of ensemble densitywith each iteration of the map. Substitution into Eq. (4.4) gives,

Pi, j = 〈PF (φi ),φj 〉 =∫ 4

x=0PFν (χi (x))χj (x)dx , i , j = 1,2,3,4. (4.17)

Continuing to produce A requires 16 = 42 such integrations for each χi (x),χj (x) basisfunction pairing, but in this simplified case where we have piecewise linear functions, wewill iilustrate with just one pairing, i = 1 and j = 2 to produce P1,2 using the basis functionsχ1(x) and χ2(x) in the Galerkin method’s inner product computation.

P1,2 =∫ 4

x=0PFν (χ1(x))χ2(x)dx ,

=∫ 2

x=1PFν (χ1(x))dx , (4.18)

by considering the nonzero support of χ2(x). Evaluating the Frobenius-Perron image ofthe initial density, χ1(x) for the integration by referring to the definition and specializing tothis example,

PF [χ1(x)]=∑

y∈F−1(x)

χ1(x)∣∣det Dy F∣∣ = 1

3χ1(x), (4.19)

Again we recall the region of nonzero support required to be considered in evaluating theintegral in Eq. (4.18). In this region, colored as shown in Fig. 4.7, each x has exactly onepre-image, still in I1, and thus, |det Dy F

∣∣= 3 there. Hence continuing,

P1,2 =∫ 2

x=1

1

3χ1(x)dx = 1

3. (4.20)

Similarly repeating this computation in kind for each i , j pairing results in the rest ofthe transition matrix,

A =⎡⎢⎣ 1/3 1/3 1/3 0

0 0 1 00 0 1/2 1/2

1/4 1/4 1/4 1/4

⎤⎥⎦ (4.21)

Page 88: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

80 Chapter 4. Graph Theory and Markov Models of Transport

Figure 4.7. The Markov map shown in Fig. 4.4(a) is again shown here, but theyellow colored region here is shaded to highlight the set which is active when computingthe integrations necessary to produce the matrix element P1,2 as described in Eqs. (4.13)-(4.20).

As a side note, observe that when computing PF [χi (x)], there is a potential for multiplepreimages, but then considering the inner product, 〈PF (φi ),φj 〉 means that we will alwaysselect exactly one of those branches since the other function in the inner product, χj (x)selects just one branch and no more. This follows from the fact that we have specializedto characteristic functions supported over the elements of a Markov partition; thus we im-plicitly rely on the monotonicity which follows the required homeomorphism for each leg.�

Thus the transfer operator becomes an operator of finite rank in terms of a Galerkin’smethod with finitely many terms. Furthermore, the approximation is exact when a Markovpartition and associated basis functions are used. The resulting A may be interpreted as atransfer matrix, for which it is easy to check that all row sums are 1,∑

j

Pi, j = 1,∀ j . (4.22)

As such, it is a stochastic matrix. Correspondingly, this generates a weighted directed graphfor a representing Markov process, as shown in Fig. 4.4(b).

It is general that matrices produced in this way have positive elements, and they willbe stochastic matrices when the Markov partition property exists and is used.38 As such, thetheory of Perron-Frobenius matrices39, applies and specializes to stochastic matrices. Some

38Without using the existing Markov partition, or if one does not exist, then a stochastic matrix does notresult, but it may be “almost" stochastic in that row sums are almost one, reflecting some leak of measure inthe representation, while not reflecting the reality of the dynamical system.

39Contrast that phrase to “Frobenius-Perron operators" already discussed here. While the names o thesemathematicians associated with the operator are found ordered both ways in the literature, we will take herea common convention that FP to denote the operators, and PF to denote the finite rank situation and therelated linear algebra.

Page 89: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 81

elements of this theory will be reviewed in the next section, but we will illustrate a mainidea continuing with the example of Fig. 4.4(a). That is, the dominant eigenvalue equals1, and uniquely if the dynamical system is ergodic or correspondingly the resulting matrixis irreducible. The dominant eigenvector corresponds to the steady state, and thereforeassociates to the steady state measure.

Example 4.5. (Steady State of a Finite Rank Transfer Operator )Continuing with example, Example 4.4 from Fig. 4.4(a), we compute the steady state

distribution. The steady state distribution of the Markov process Fig. 4.4(b) can be shownfrom the Perron-Frobenius theory of positive matrices (See Sec. 4.3.3) to equal the eigen-vector of the transition matrix corresponding to the largest eigenvalue, which is 1. Thiseigenvector of matrix Eq. (4.21) is,

v = 1√226

⎡⎢⎣ 33

128

⎤⎥⎦ . (4.23)

Now this eigenvector represents the steady state distribution of ensembles of initial condi-tions over long times, where each partition element of the Markov partition should see arelative number of points land in them, according to the relative proportions of this vector.To punctuate this point, we illustrate the histogram from a long orbit of 105 iterates, inFig. 4.8. Note that by showing the histogram of a long orbit, rather than the histogram pro-duced by an ensemble of many initial conditions, we have again appealed to the notion thattime averages and spatial averages can be exchanged. Such correspondence is the premiseof the Birkhoff ergodic theorem, which is valid as this example is an ergodic system. Theresulting matrix is irreducible and the system is Markov. Thus the finite rank representationexactly describes the steady state measure. �

4.3.1 Graph Theory, Reducible Matrices

A broader theme of this book is the analysis of global transport, which can be definedin terms of a partition to discuss how ensembles may propagate from one element of thepartition to others. The discrete counter part of this discussion for a directed graph alsocorresponds to notions of partitions of a graph into subgraphs as discussed here.

In this section we briefly review some basic concepts of reducible matrices and corre-sponding relationships to the graph induced by the matrices due to discrete approximationsof a Frobenius-Perron operator. In particular, we introduce the question of how to discovera permutation of the transition matrix (4.7) that reveals the basin structure of a dynamicalsystem revealed to us in the form of a test orbit segment {xj }Nj=1.

A graph G A associated to a matrix A consists of a set of vertices V and a set of edgesE ⊂ V ×V . The entry Pij is nonzero when there exists an edge (i , j )∈ E that connects thevertices i and j .

As we will describe, borrowing a term from the recent activity in complex networks,it is also useful to consider structure within a graph. First we have components, and thengeneralizing that we have communities. If a disjoint collection {Si }ki=1 of subsets Si ⊂ Vconsists of vertices such that there is a higher density of edges within each Si than betweenthem, then we say roughly that {Si }ki=1 forms a community structure for G A [250, 251].

Page 90: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

82 Chapter 4. Graph Theory and Markov Models of Transport

Figure 4.8. Steady state of the system in Fig. 4.4(a,b) and Fig. 4.7 and defined byEq. (4.16) was noted to be Markov, and as such the corresponding stochastic matrix fromUlam-Galerkin’s method gives the dominant eigenvector Eq. (4.23). (Top) The map shownagain to help match the regions of the partition relevant for the states in the density. Notethat each piecewise linear section of the map matches piecewise constant sections of theinvariant density. (Bottom) The histogram for a 105 iterate test orbit agrees quite well withthe density predicted by the eigenvector Eq. (4.23).

For the discussion of this section, it is only necessary to consider the sign of theentries of a transition matrix A, or perhaps of a stochastic matrix A. We shall define,

B = sign(A), (4.24)

which when A is a nonegative matrix, Bi, j = 0 or 1, and comparing to the directed graph

Page 91: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 83

G A = (E , V ) generated by A implies that there is an edge from j to i ,

Bi, j =(

1 if j→ i0 else

). (4.25)

Such is called an adjacency matrix when it encodes just possible transitions withoutweights. That is, a one step path is on an edge from a vertex labeled j to vertex labelled i .

When the graph represents the action of the Frobenius-Perron operator from a dy-namical system, a one step path means that there exists a direct transfer from mode φj (x)to φi (x); when those modes are characteristic functions. This in turn implies existence ofat least a one step orbit of the dynamical system between the cells when basis functions arecharacteristic functions. If an n-step path in the graph is possible this does not imply thatan n-step orbit through the corresponding cells covering the phase space of the dynamicalsystem is possible. Instead the correct statement is that an epsilon-chain pseudo-orbit40 ispossible, but perhaps not a true orbit segment.

Figure 4.9. A Path in the graph corresponding to the transfer matrix producedfrom orbits in a dynamical system may not imply an orbit in the dynamical system passingthrough the same corresponding boxes, but rather an epsilon-chain pseudo orbit. Hereshown the graph has an orbit from boxes labelled i → j → k, but x ∈ i , T (x) ∈ j andperhaps T 2(x) ∈ k+1 but perhaps there is an x ′ such that T (x ′) ∈ k.

An essential question then is if a given path41 in the graph is possible, which hassignificance to the distribution of measure in the corresponding Markov chain. A fewissues relevant to transition matrices are captured by the following language.

40An epsilon-chain pseudo-orbit is a sequence of points, {xi }, such that ‖T (xi )−xi+1‖ ≤ ε for each i , andthis specializes to a true orbit if ε = 0. There are theorems from the shadowing literature, [47, 48] especiallyfor hyperbolic system that describe that nearby an epsilon chain there may be true orbits, but in the contextused here, those true orbits may not pass through the boxes corresponding to a particular refinement.

41An n-step path from vertex i to vertex j is the existence of n edges in order “end-to-end" (stated roughlyto mean the obvious that each edge ends at a vertex where the next begins) beginning at i and ending at j .

Page 92: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

84 Chapter 4. Graph Theory and Markov Models of Transport

Definition 4.1. A square matrix An×n is called primitive if there exists a (time) k suchthat [An×n]k

i, j > 0,∀i , j .

Remark: A sufficient condition for a matrix to be a primitive matrix is for the matrix tobe a nonnegative, irreducible matrix with a positive element on the main diagonal.

The primitive property may be interpreted as follows. Since a direct path betweenj to i exists if and only if Pi, j > 0, and P2 describes the two step action of the graph,then a two step path exists between j to i if and only if P2

i, j > 0. In these terms, theprimitive property demands that there is a time-k when there is a path from everywhereto everywhere else in the graph. Conversely, if there is a i , j , pair such that Pk

i, j = 0 forall times k, then as a graph structure, the two vertices must occupy different componentsof the graph. Component is then a word which describes much the same information asprimitive-ness for the corresponding graph.

Definition 4.2. A component of a graph G is a subgraph G′ of G such that between everypair of vertices, there is a path between every vertex in the subgraph. A graph is calledconnected if it consists of exactly one component.

As a matter of example, we see in Fig. 4.10, the component including i , i + 1, andj is apparently not connected to the component including vertices k, k+ 1. As such, theadjacency matrix A which generates this graph cannot be primitive.

Figure 4.10. A segment of a larger directed graph. We see that a one step walkfrom vertex i to i+1 is possible but a walk to j from i requires two steps. No walk from thecomponent containing vertices i , i +1, and j to the k, k+1 component is possible at leastalong the edges shown.

Let I = {1 . . . N} be an index set. For our specific application to the transition matrixP generated by the Ulam-Galerkin method, we define the set of vertices, V = {υi }i∈I to

Page 93: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 85

label the original boxes {Bi }i∈I used to generate the matrix P, and define the edges to bethe set of ordered pairs of integers E = {(i , j ) : i , j ∈ I } which label the vertices as theirstarting and ending points.

Another concept which is useful in discussion of partitioning a graph into dynami-cally relevant components is to ask if the graph is reducible.

Definition 4.3. A graph is said to be reducible if there exists a subset Io ⊂ I such that thereare no edges (i , j ) for i ∈ Io and j ∈ I \ Io , otherwise it is said to be irreducible [22, 321].

This condition implies that the graph is irreducible if and only if there exists only oneconnected component of a graph, which is G P itself. In term of the transition matrix P, G Pis reducible if and only if there exist a subset Io ⊂ I such that Pij = 0 whenever i ∈ Io andj ∈ I \ Io. Furthermore, P is said to be a reducible matrix if and only if there exists somepermutation matrix S such that the result of the similarity transformation,

R= S−1PS (4.26)

is block upper triangular

R =(

R1,1 R1,20 R2,2

). (4.27)

This means that G P has a decomposition into a partition,

V = V1

⋃V2, (4.28)

such that V1 connects with V1 and V2, but V2 connects only with itself. When R1,2 = 0, Pis said to be completely reducible [321],

R =(

R1,1 00 R2,2

). (4.29)

An instructive observation when relating these concepts back to dynamical systemsis that in the case that G P is generated from a bi-stable dynamical system as from Example7.1, the transition matrix A will be completely reducible. Therefore, R1,1 and R2,2 cor-responds to the two basins of attractions of the system. The off diagonal elements R1,2and R2,1 give information regarding transport between these partition elements if there areany nonzero off diagonal elements. Also, in a general multi-stable dynamical systems thetransition matrix A has a similarity transformation into a block (upper) triangular form -emphasizing many components which may not communicate with each other.

A key point relevant to the theme of this writing is that most (randomly or arbitrar-ily realized) indexing of the configuration resulting from a Ulam-Galerkin matrix makes itdifficult to observe even simple structures like community structure or reducibility of thegraph and the corresponding transition matrix. That is the indexing that may come froma sensible and suitable covering of a phase space by rectangles, or perhaps of triangles,will not generally be expected to directly reveal a useful structure in the associated Ulam-Galerkin matrix. The goal is then to reveal reducibility or at least community structurewhen it is there. See Figs. 4.12-4.11 illustrating the kind of graph that results from a bi-stable system coming from a bi-stable system, when usefully sorting to reveal the structure.Notice that there are two disjoint components that do not communicate between each other.

Page 94: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

86 Chapter 4. Graph Theory and Markov Models of Transport

Correspondingly this informs that the bi-stable system has two basins that do not communi-cate with each other. A more difficult problem of sorting a transition matrix comes from thescenario when there may be some off diagonal elements even in a suitable sorting, as wouldarise when there may be some “leaking" or small transport between two basins which mayno longer be invariant sets, but rather simply almost invariant.

In Figures 5.3(a) and 5.3(b) we show an example of a transition matrix for a 30-vertex random community-structured graph with three communities of which membersare placed haphazardly, but after a proper permutation this matrix is transformed into an“almost" block diagonal form. This is the sort of Ulam-Galerkin matrix which wouldbe expected from a bistable dynamical system. Figures 5.4(a) and 5.4(b) illustrate theassociated graph of the matrix in the above example. Before sorting the vertices into threeseparate communities the graph looks like a random graph that has no community structure.However, after sorting the community structure of the graph becomes obvious. This iscentral to the problem of this topic, to find the sorting that reveals the useful partitionillustrative of transport.

There are several techniques to find an appropriate permutation if the matrix is re-ducible. In the language of graph theory, we would like to discover all connected com-ponents of a graph. The betweeness methods in [251] and the local method in [9] areexamples of numerous methods as reviewed in [75, 251] that can be successfully used touncover various kinds of community structures.

Remark: Lack of reducibility of a graph relates to the measure theoretic concept ofergodicity (Definition 3.6), and the topological dynamical concept of transitivity (see inDefinition 6.1), in that when the graph representations is sufficient, they all describe thereduction of the system either into noncommunicating components, or not in the case ofthe complement.Remark: An essential question is whether the represented dynamical system is wellcoarse grained by the graph and the action of the matrices. The question can be settled interms of discussion of those dynamical systems which are exactly represented by a coarsegrain, which are those called Markov dynamical systems, and then the density of these.When a coarse grained representation is used, the associated transition matrix may revealsome off diagonal elements when sorted suggesting some transitivity in the dynamical sys-tem, but this could just arise as the error of the coarse-grain estimation.

4.3.2 Convergence rates, The Power Method and Stochastic Matrices

Analysis of the result of iterating a matrix is very instructive regarding the long time be-havior of the transfer operators, and the matrix discussion is a simplifying discussion. Itis a well known feature in matrix theory that compositions of certain matrices in somesense have a tendency that large powers tend to point arbitrary vectors toward the it dom-inant eigenvector. This feature is the heart of the power method in numerical analysis[310] which has become crucial in computing eigenvalues and eigenvectors of very largematrices - too large to consider a computation directly by the characteristic polynomialdet(A−λI )= 0. This discussion is meant to give some geometric notion of the action ofthe Galerkin-Ulam matrices in the vector spaces upon which they act.

We will begin this discussion with a simple illustrative example,

Page 95: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 87

(a) The transition matrix before sorting, in arbitraryconfiguration.

(b) the transition matrix after sorting to reveal a blockdiagonal form which informs that there are two com-ponents that do not communicate with each other.When arising from a dynamical system, this is typi-cal of a bi-stable system meaning two basins as twoinvariant sets and no transport between them.

Figure 4.11. The matrix in this figure is reducible. However, the members of theoriginal transition matrix before sorting are placed haphazardly. After using a diagonal-izing similarity transformation Eq. (4.26) to the form Eq. (4.29) (when it exists) then theblock diagonal form is revealed. The corresponding graph description of the same is seenin Fig. 4.12.

Page 96: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

88 Chapter 4. Graph Theory and Markov Models of Transport

(a) The original graph. Embedding in the plane thevertices in a arbitrary configuration conceals simplestructure.

(b) Sorted graph. The same graph subject to appro-priate permutation reveals reducible structure commu-nity between components.

Figure 4.12. Graph representation of the adjecency matrices shown in Fig. 4.11.

Example 4.6. (Power Method) Let,

A =⎡⎢⎣ 1/3 1/3 1/3 0

0 0 1 00 0 1/2 1/2

1/4 1/4 1/4 1/4

⎤⎥⎦ (4.30)

Page 97: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 89

which is again matrix from the Ulam-Galerkin method example, Eq. (4.21), derived fromFig. 4.4(a). Here we discuss it simply in terms of its matrix properties. Checking,

A

⎡⎢⎣ 1111

⎤⎥⎦=⎡⎢⎣ 1

111

⎤⎥⎦ (4.31)

confirms that λ = 1 is a right eigenvalue with a corresponding right eigenvector, v =[1,1,1,1]′, (“ ′ " denotes transpose), in the usual definition of an eigenvalue-eigenvectorpair,

Av = λv. (4.32)

Likewise, a substitution, [1,1,4, 8

3

]A = [ 1,1,4, 8

3

](4.33)

which as an instance of the left eigenvalue/eigenvector equation,

u A = τu, (4.34)

wherein τ = 1 is a left eigenvalue with corresponding left eigenvector, u = [1,1,4, 83 ]. �

These particular eigenvalue/eigenvector pairs have special significance. It is easy tocheck that A row sums to one, ∑

j

Pi, j ,∀i , (4.35)

which defines A to be a stochastic matrix with probabilistic implications to be defined laterin this section. In particular, the left side of Eq. (4.31) may be directly interpreted as arow-by-row summation, and matching the right side of the equation means that the resultmust be one. Thus we summarize,

Remark: (λ= 1,v = [1,1,1,1]′) will be a right eigenvector/eigenvector pair if and only ifA is row-wise stochastic.

That (τ = 1,u = [1,1,4, 83 ]) is a left eigenvalue/eigenvector pair may be derived from

the characteristic polynomial,det(A′ − τ I )= 0. (4.36)

That we may directly solve this 4th degree polynomial, with up to four roots due to the fun-damental theorem of algebra, is possible primarily because the matrix is small. In principlethese roots may be found directly by a direct computation akin to the quadratic formula[213], or at least by numerical root solvers. However, numerically approaching the spec-trum of large matrices is only feasible through the “power-method" cited in many texts innumerical analysis, [310]. Furthermore, specifically considering the power method in thissimple case allows us an easy presentation of the behavior of the evolution of the corre-sponding Markov chain as we shall see.

Consider an arbitrary vector w ∈ En , where En denotes the vector space which servesas the domain of A, which in this case we generally choose to be Rn . By “arbitrary", wegenerally mean “almost any". A full measure of those “good" vectors w shall suffice, andthose which are not good not shall become clear momentarily. Let A act on w from the left,

wA = (c1u1+ c2u2+ ...+ cnun)A, (4.37)

Page 98: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

90 Chapter 4. Graph Theory and Markov Models of Transport

where we have written as if we have a canonical situation that there is a spanning set ofn-eigenvectors {ui }ni=1, but all we require is that,

• one eigenvalue is unique and largest,

• the subspace corresponding to the rest of the spectrum may be more degenerate.

Of course in general there can be both algebraic and geometric multiplicities, but for now,for sake of presentation we describe the simplest situation of unique eigenvectors, andfurther assume that the corresponding eigenvalues are such that one eigenvector is uniqueand largest,

τ1 > τi ,∀i > 1. (4.38)

By linearity of A, and further by resorting to the definition of a left eigenvector for each ofthe ui , Eq. (4.37) becomes,

wA = c1 Au1 A+ c2u2 A+ ...+ cnun A == c1τ1u1+ c2τ2u2+ ...+ cnτnun . (4.39)

Then proceeding similarly, applying A m-times, we get,

wAm = c1τm1 u1+ c2τ

m2 u2+ ...+ cnτ

mn un . (4.40)

Therefore, we see roughly that,42

τ1 > τi =⇒ τm1 τm

i , (4.41)

for large m, from which follows,

wAm ≈ c1τm1 u1. (4.42)

This says that repeated application of the matrix rotates arbitrary vectors toward the domi-nant (u1) direction.

The general power method from numerical analysis does not proceed in this wayalone because while it may be true that, τm

1 τmi , for large m both (all) of these numbers

become large, and the computation becomes impractical on a computer. In general it isbetter to renormalize at each application of A step, [310] at each step in the followingform. Let, s0 be chosen aribtrarily as was stated in Eq. (4.37), but then,

sk+1 = sk A

‖sk A‖ , (4.43)

stated in terms of left multiplication of left eigenvectors and where ‖wk A‖ is a renormal-ization factor at each step in terms of a vector norm. Similar arguments to those stated inEqs. (4.37)-(4.42) can be adjusted to show that,

sk →w1, as k→∞, (4.44)

42The notation “ " denotes “exceedingly larger than," which can be defined formally by the statement,

τm1 τm

i ≡ limm→∞τmiτm1= 0.

Page 99: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 91

in the vector norm, ‖sk −w1‖ → 0. The general statement is that a subsequence of skconverges to w1 because a general scenario is that the largest eigenvalue in magnitudemay be complex. Rotations upon application of A may occur complicating discussion ofconvergence. This will not be an issue for our specific problem of interest which involvesPerron-Frobenius matrices and in particular stochastic matrices, since such matrices have apositive real eigenvalue as we will expound upon below.

The three most common vector norms we will be interested in here are,

1. ‖v‖1 =∑ni=1 |vi | the sum of the absolute values of the entries of a vector v.

2. ‖v‖2 =√∑n

i=1,v2i =√v′v, the Euclidean norm, and

3. ‖v‖∞ = maxi=1,2,...,n|vi |, the infinity norm or also known as the max norm,

in terms of a n× 1 column vector. We shall discuss the power-method in terms of theEuclidean norm ‖ · ‖2. Further by similar arguments, the (left) spectral radius43 followsfrom the Raleigh-quotient,

sk As′ksks′k

→ τ1, as k→∞, (4.45)

Also from the discussion of the power method, we get an idea regarding the rate ofconvergence. Again we start,

wA = c1 Au1 A+ c2u2 A+ ...+ cnun A == c1τ1(u1+ c2

c1

τ2

τ1u2+ ...+ cn

c1

τn

τ1un), (4.46)

from which,

wAn = c1τn1 (u1+ c2

c1[τ2

τ1]nu2+o([

τ2

τ1]n). (4.47)

Thus follows geometric convergence as,

r = |λ2

λ1|. (4.48)

See a geometric presentation of the power method in Fig. 4.13. Really, the onlypart of the terms after the first two are that they form proj⊥[w,span(u1,u2)] whereproj⊥[w,span(u1,u2)] denotes the orthogonal projection of w onto the subspace spannedby u1 and u2, meaning w− c1u1− c2u2. The details of the multiplicities and degeneraciesof that main subspace is irrelevant to our discussion since we are interested in Perron-Frobenius matrices. See a characture representation of the power method in Fig. 4.13 andits spectrum in Fig. 4.14.

Note that usually this discussion of the power method is carried out regarding righteigenvectors in the numerical analysis literature, which we adapted here for our interestin row stochastic matrices. In particular, in the case of a stochastic matrix when τ1 = 1,then the renormalization step will not be zero, and the convergence rate will be geometricwith rate, r = |λ2|. While the expectation of the projection of Frobenius-Perron operators

43The spectral radius is the largest eigenvalue in complex modulus.

Page 100: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

92 Chapter 4. Graph Theory and Markov Models of Transport

Figure 4.13. Geometric representation of the power method representation ofEqs. (4.46)-(4.47).

may be a row stochastic matrix, projection errors (due to truncation of the would-be infiniteset of basis functions), and computational errors due to finite precision in the computers,typically the computation behaves as if τ ≈< 1 as if there were a small mass leak. Therewill be more on this leak issue in Sec. 4.3.4 by Eq. (4.58).

4.3.3 Perron-Frobenius Theory of Nonegative Matrices

There are several important and succinct statements that can be proven regarding stochasticmatrices, since they are nonegative. There are many useful statements which are part of thePerron-Frobenius theory, but we will highlight only a few here, those that are most rele-vant to the stochastic matrices which result from Ulam-Galerkin’s projection of Frobenius-Perron operators. Those matrices that are nonegative. As such, they have special propertiesthat complement the description in the previous subsection, regarding the (non)possibilityof multiplicity and complex value of the largest eigenvalue.

Where the previous section on the power method was meant to give geometry of theaction of a Galerkin-Ulam matrix, this section should be considered complementary sinceit is meant to give some algebraic information regarding the same action. We will start bystatement of a particular form of the Perron-Frobenius Theorem, following the necessarydefinitions to interpret it, and then examples. We omit the proofs of this theorem since itis standard from matrix theory, and refer to [13]. To intrepret this theorem, we state thefollowing definitions with brief examples.

Definition 4.4. An n× n square matrix An×n is called a positive matrix if An×n > 0meaning that each and every term in the matrix is nonegative, [An×n]i, j > 0,∀i , j .

We present the definition in terms of square matrices since we are only interested inspectral properties. However, the theory of positive matrices is interesting more broadly.

Definition 4.5. An n× n square matrix An×n is called a nonegative matrix if An×n ≥ 0meaning that each and every term in the matrix is positive, [An×n]i, j ≥ 0,∀i , j .

Page 101: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 93

Figure 4.14. Spectrum of the stochastic matrices in the complex plane. We expectthe dominant eigenvalue to be λ1 = 1, and the second eigenvalue, |λ2|< λ1 to describe thegeometric convergence of the power method according to Eqs. (4.47)-(4.48), r = |λ2

λ1|.

Example 4.7. (Positive and Nonegative Matrices) Inspecting,

A=⎛⎝ 1 2 3

4 5 67 8 9

⎞⎠ , B=⎛⎝ 1 2 0

4 5 67 8 9

⎞⎠ ,C =⎛⎝ 1 −1 3

4 5 67 8 9

⎞⎠ , D=⎛⎝ 1 2 3 3.1

4 5 6 6.27 8 9 9.2

⎞⎠ ,

(4.49)A is a positive matrix, but B is nonegative because of the single nonzero entry. C is neitherbecause of the single negative entry, and D is not considered in the discussion of positivitysince it is non-square and therefore irrelevant when discussing spectral properties.44

Remark: A nonegative matrix has a unique positive and real dominant eigenvalue dueto the Perron-Frobenius theory. Therefore the power method proceeds as described inEqs. (4.37)-(4.42) without the possible complications regarding multiplicties either geo-metric or algebraic when discussing convergence and convergence rate.

Remark: A stochastic matrix is nonegative. We have already noted that τ1 = 1is an eigenvalue since it corresponds to the definition that the matrix must row sum toone. Further, any larger eigenvalue would correspond to a larger row sum, so it must belargest. Therefore, by the Perron-Frobenius theory, this is the unique largest eigenvalue ofa stochastic matrix.

44 D has no eigenvalues, but there is a related discussion of the singular spectrum through the SVD de-composition.

Page 102: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

94 Chapter 4. Graph Theory and Markov Models of Transport

Theorem 4.8. (See [13] ) If An×n is nonegative and irreducible, then

1. The spectrum of eigenvalues of An×n includes one real positive eigenvalue which isuniquely the largest in absolute value - meaning that it is the only eigenvalue on theoutermost spectral circle. See Fig. 4.14.

2. The eigenvector corresponding to that eigenvalue has entries which are all positivereal numbers.

3. Further, the largest eigenvalue λ is algebraically unique, meaning that it is a simpleroot of the characteristic polynomial, det(A−λ)= 0.

4.3.4 Stochastic Matrices

It is easy to see that a stochastic matrix is a Perron-Frobenius matrix.

Definition 4.9. A (row) stochastic matrix is a square matrix An×n such that

1. Each row sums to one,∑n

j=1 Ai, j = 1,∀i = 1,2, ..,n, and

2. 0≤ Ai, j ≤ 1,∀i , j = 1,2, ..,n.

A convenient way to state the row-sums to one property is through a matrix naturalnorm,

Definition 4.10. Given a square matrix An×n , its matrix natural norm is defined,

‖An×n‖ = supv‖Av‖‖v‖ = supw:‖w‖=1‖Aw‖, (4.50)

which is also often called an induced norm since the matrix norm is induced (inherited)by the vector norm ‖ · ‖ : E → R+, where E is the vector space domain of An×n . OftenE = Rn .

In terms of the popular matrix norms listed in List 4.3.2, the matrix natural norms areespecially conveniently computed,

1. ‖A‖1 is the matrix 1-norm induced by the vector 1-norm, ‖A‖1=maxi=1,..,n∑n

j=1 |Ai, j |,which is the maximum row sum,

2. ‖A‖∞ is the matrix infinity-norm induced by the vector sup-norm, ‖A‖∞ =maxj=1,..,n

∑ni=1 |Ai, j |, which is the maximum column sum,

3. however, ‖A‖2 is conveniently but not as conveniently computed as, ‖A‖2=√ρ(A′A),where ρ(A′A) is the spectral radius of A′A.

Page 103: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.3. Action on Densities as a Directed Graph 95

In this notation, a stochastic matrix must satisfy the properties,

‖A‖1 = 1, 0≤ Ai, j ≤ 1, ∀i , j , (4.51)

The purpose of defining a stochastic matrix in such a manner, is that it may be in-terpreted as reflecting the transition probabilities of a finite state Markov chain, which is aspecial case of a discrete time stochastic process.

Definition 4.11. A Discrete Time Stochastic Process is a sequence of random variables,X1, X2, X3, ....

Definition 4.12. A Markov chain is a discrete time stochastic process of random variables,X1, X2, X3, ... such that the conditional probability of each next state is independent of theprior history,

P(Xm+1 = x |X1 = x1, X2 = x2, ..., Xm = xm)= P(Xm+1 = x |Xm = xm), (4.52)

where P(·) denotes the probability of enunciated state of the random variable.

In each of these, we have referred to the concept of random variable, Defined 3.2.We may consider the random variable as a measurement device that returns a real

number, or the random experiment in our language, for a given subset of . Recall thatfor a measure space (,F ,μ), a measure μ is called a probability measure if μ : F →[0,1] and μ()= 1; hence a measure space (,F ,μ) will also be accordingly referred toas a probability space. With a probabilistic viewpoint in mind, the random variable tells usthat the probability to observe a measurement outcome in some set A ∈ B(R) based on aprobability measure μ is precisely μ(X−1(A)), which makes sense only if X is measurable.

When the state space is a finite set, ={x1, x2, ....xn} or simply write ={1,2, ...,n},we have a finite state Markov chain, from which the set of all transition probabilities inEq. (4.52) form a finite sized table of possibilities, which may be recorded in a matrix,

Pi, j = P(Xm+1 = xj |Xm = xi ),∀i , j ∈ 1,2, ...,n. (4.53)

To draw the connection to our larger theme, in some sense any Ulam-Galerkin’smethod is simply an accounting of these probabilities for transitions between states iden-tified by energy in each mode - as represented by a given chosen basis function. Whena full (countable) basis set is used, then the resulting “matrix" would be infinite, but thetruncation involved in Galerkin’s method corresponds ignoring the “less important states"to be discussed further shortly. Generally, this ignoring of less important states leads toleak of measure in the states.

Further, evaluation of probability of events is a measure of the relative volume of astate in the set of outcomes. That is, given a probability space (,F ,μ), then formally,the assignment of probability to an event ω ∈ corresponds to a measure which may bedescribed by integration,

P(·) : → [0,1] (4.54)

P(B)=∫

χB(x)dμ(x)=∫

Bdμ(x), (4.55)

in terms of the indicator function χB (x) of a Borel set B . Here we follow a conventionalabuse of a notation so that P(B) is technically equivalent to P({ω ∈ : X(ω) ∈ B}), wherein the above case, the random variable X is χB .

Page 104: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

96 Chapter 4. Graph Theory and Markov Models of Transport

That a finite state Markov chain results in a stochastic matrix is a direct consequenceof independence,

P(Xm+1 = xj )=n∑

i=1

P(Xm+1 = xj |Xm = xi )P(xi )= 1,∀ j ∈ 1,2, ...,n, (4.56)

simply because the union of all the states forms the full set of possibilities which is the fullprobability,

= ∪nj=1Bj . (4.57)

Here is where we see the consequence of truncation in the Galerkin’s method inregard to leak of measure in terms of the probabilities. If the Eq. (4.56) is a finite, and thetrue state space is infinite, but a small fraction of the states account for the majority of theprobability, then we may write,

P(xj ) =n∑

i=1

P(Xm+1 = xj |Xm = xi )

<∼ 1=∞∑

i=1

P(Xm+1 = xj |Xm = xi )

=n∑

i=1

P(Xm+1 = xj |Xm = xi )

+∞∑

i=n

P(Xm+1 = xj |Xm = xi ), (4.58)

when,∞∑

i=n

P(Xm+1 = xj |Xm = xi ) << 1, (4.59)

which will happen when,

∪nj=1 Bj is a proper subset of , (4.60)

but measures most of the set. Formally, the way to extend the notions of finite state Markovchains even to uncountable state chains, is through Harris chains [168]. We already noted inSec. 4.2 how the Markov property of a map results in the Galerkin method yielding a finitetruncation Ulam-Galerkin matrix, which we now describe as a representing the transitionsof a finite state Markov chain.

We now revert to discussion of uniqueness of dominant states, in terms of matrixissues such as reducible matrices, primitive matrices and the relationship to the dynamicalsystems analogue of the dominant state(s) to (unique) ergodicity.

4.4 Exact Representations are Dense and theUlam-Galerkin’s Method

We close the chapter with presentations of some results from [19] of how Markov mapsare dense. In this manner, analysis of approximation of the matrix approximations of the

Page 105: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.4. Exact Representations are Dense and the Ulam-Galerkin’s Method 97

Frobenius-Perron operator can be developed. The error in an approximation of the infinitedimensional operator can be described in two ways. The more common way is by dis-cussion of an approximation due to finite truncation of the series in the Galerkin’s methodEq. (4.4). As lesser known but also useful description of the same question is as follows.Each finite series approximation can be understood as an exact presentation of another map,a map which is Markov on exactly the finite partition used. As such, the approximation canbe analyzed by description of how far the approximating Markov map is from the non-Markov map under description. Finally, the expectation is that there is always a nearbyMarkov map for such a description - The concept we will discuss in this section is thatMarkov maps are expected to be dense in the space of maps, and therefore an error analy-sis of a finite truncation can be considered by the nearby map where the representation isexact. We can prove this statement at least for a certain class of systems. The results hereare drawn from our work, [19] where the fuller details of the brief presentation here can befound.

Consider a family of chaotic skew tent maps. The skew tent map is a two-parameter,piecewise-linear, unimodal map of the interval. We show that these maps are Markov for adense set of parameters and we exactly find the probability density function (pdf), for anyof these maps.

Remark: The central concept to this approach is the following: It is well known, [53], thatwhen a sequence of transformations has a uniform limit F , and the corresponding sequenceof invariant pdf’s has a weak limit, then that invariant pdf must be F invariant.

As case in point [19], we show in the case of a family of skew tent maps that notonly does a suitable sequence of convergent sequence exist, but these can be constructedentirely within the family of skew tent maps. Furthermore, such a sequence can be foundamongst the set of Markov transformations, for which pdf’s are easily and exactly calcu-lated. While the theorems in the following sections assume the family of one dimensionaltransformations, we will discuss extensions to higher dimensions at the end of the chapter.

Before we begin the main theme of this section, note that the discussion requiressharpening our analysis so that the underlying topology is clearer as to what we mean bydense as well as in what sense functions and invariant densities may converge.

4.4.1 Background Analysis and Topology

In discussing convergence, there is an implication of an underlying topology with respectto which the convergence may be defined. Analysis [283], topology [245], and functionalanalysis [201, 8] are the background fields here, and of course these are major areas ofmathematics for which there are many excellent textbooks and we have just listed some ofour favorites for each. For a quick start relevant to this discussion we can recommend, [23].This becomes an appropriate place to sketch in brief details the different kinds of conver-gence relevant here, for sake of completeness of discussion. We avoid an overly structuredpresentation of this would-be deep branch into extensive and interesting mathematics, thatundoubtedly would require a large volume. For a complete introduction we recommendone of the cited texts. When discussing convergence of sequences of functions, a uniformlimit is a “stronger" form of convergence than simply convergence which is often said inshort form for the synonymous pointwise convergence. Therefore we define both, andin terms of sequences of functions. Given a sequence of functions, { fn(x)}∞n=0 for eachfn : D→ R defined on a domain x ∈ D and range R, then fn converges (pointwise) to f if

Page 106: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

98 Chapter 4. Graph Theory and Markov Models of Transport

each and every sequence of values fn(x) ∈ R (you get values by plugging x into each fn)converges to a value which we shall label f (x) one for each x . Stated formally, yn = fn(x)is a sequence of numbers that converge yn → y ≡ f (x) if for every ε > 0 there existsN(x) > 0 such that ‖y− yn‖< ε if n > N which is in terms of the metric topology of ‖ · ‖on the range space yn ∈ R. Note that as stated, there is allowed an ε > 0 for each x , sopointwise, ε(x) implies N is a function of x , N(x) > 0.

In terms of uniform convergence, the change relative to pointwise convergence iswhen we identify that there exists an N > 0 to discuss pointwise convergence, but that Nmay be found independently of x . N is not a function of x but still allows the convergencethroughout the domain and range. A primary theorem regarding the notion of uniformconvergence is that a sequence of continuous functions that converge uniformly must havea limit function which is also continuous. Pointwise limits of continuous functions neednot be continuous functions, as exemplified by the following common example,

fn(x)= xn , on [0,1]. (4.61)

from whichfn(x)→ f (x) if (0, x < 1,else,1) as n→∞, (4.62)

but only pointwise and correspondingly f (x) is not continuous, specifically at the pointx = 1. However, the limit function f (x) is continuous on a restricted domain, [0, L] for any0 < L < 1 and correspondingly the convergence fn(x)→ f (x) is uniform there.

A weak limit of a sequence of functions fn(x) is a separate notion of convergenceof functions which is distinct from either pointwise convergence or uniform convergence.Written in terms of test functions h(x), in a Hilbert space H(D) (defined as a completeinner product space),

fn(x)→w f (x), (4.63)

(note the superscript w) if,

< fn(x)− f (x),h(x) >:=∫

D| fn(x)− f (x)|h(x)dx→ 0, for all h(x) ∈H(D). (4.64)

Since the convergence is in terms of integration, it is possible for a set of zero measurepoints not to converge despite weak convergence but (strong) convergence implies conver-gence

4.4.2 Example Tent Maps for Dense Markov Transformations

As case in point, we present here a family of skew tent maps for which it can be explicitlyproven that Markov maps are dense. Let Fa,b denote the two-parameter piecewise-linearmap on the interval [0,1] satisfying

Fa,b(x)={

b+ 1−ba x if 0≤ x < a

1−x1−a if a ≤ x ≤ 1

(4.65)

with 0 < a < 1 and 0 ≤ b ≤ 1. It has been shown that [14] in the following region of theparameter space,

D = {(a,b) : b < 1/(2−a) and ((b < 1−a) or (b ≥ 1−a and b > a))} , (4.66)

Page 107: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.4. Exact Representations are Dense and the Ulam-Galerkin’s Method 99

Fa,b has chaotic dynamics, and in the parameter subset

D0 = {(a,b) : b < 1/(2−a) and (b < 1−a)} , (4.67)

chaotic dynamics occur on the entire interval 0 ≤ x ≤ 1. See Bassein [14] for a completeclassification of the dynamics in parameter space, (a,b) ∈ (0,1)× [0,1].

Specifically, Fa,b is Markov for a dense set of (a,b) in D0. It follows that the proba-bility density function for any one of the Markov transformations in this set can be foundexactly to be a piecewise constant function, 4.15 from [53]. Therefore, these exact resultsapproximate the probability density function for any other transformation in D0.

Sufficient conditions for a piecewise-linear map on the interval to be Markov canbe developed partly through symbolic dynamics. (See Chapter 6 for discussion of sym-bolic dynamics). Then proof that Markov maps are dense in the parameter set D0 follow.Probability density functions computed by stochastic matrix methods quickly follow.

The form of Fa,b from [14], allows a concise discussion of the full range of topologi-cal dynamics in the parameter space (a,b)∈ (0,1)× [0,1]. Piecewise linear transformationson the interval have been widely studied, and under different names as well: broken lin-ear transformations [146], weak unimodal maps [237]. These maps also allow explicitthe study the chaotic behavior including the stability of the associated Frobenius-Perronoperator [208], the topological entropy [238], and kneading sequences [65]. [19]

4.4.3 Markov Skew Map Transformations

In the special, but important, case that a transformation of the interval are Markov (Defini-tion of Markov transformations are given in 4.4), the symbol dynamic are simply presentedas a finite directed graph. The following result describes a set in D0 for which Fa,b isMarkov.

Theorem 4.13. [19] For a given (a,b)∈ D0, if x0 = 1 is a member of a periodic orbit, thenFa,b is Markov.

Proof. Set F = Fa,b. Assume x0 = 1 is a member of a period n orbit (n > 1). Next, forma partition of [0,1] using the n members of the periodic orbit. The two endpoints of theinterval are included since ∀(a,b) ∈ D0, F(1)= 0. Order these n points so 0 = c0 < c1 <

.. . < cn−1 = 1, regardless of the iteration order. For i = 1, . . . ,n−1, let Ii = (ci−1,ci ) anddenote the restriction of F to Ii by Fi .

For a given Ii = (ci−1,ci ), the endpoints ci−1 and ci will map exactly to two mem-bers of the partition endpoints by definition of the periodic orbit. Let these points be cjand ck , with cj < ck and j ,k ∈ {0, . . . ,n−1}. The only turning point of the map is x = a,and ∀(a,b) ∈ D0, F(a)= 1. Therefore x = a must always be part of the period orbit anda member of the partition endpoints, implying each Fi is linear and hence, a homeomor-phism. Also Fi (Ii )= (cj ,ck), a connected union of intervals of the partition. By definition,F is Markov.

At this point, symbolic notation becomes useful. In particular, some elements ofkneading theory from the study of symbolic dynamics is useful. Details on kneading theorycan be found in original form, [233] and nicely presented in [97]. The point x = a is thecritical point at the “center” of the interval, denoted by the letter C. All a < x ≤ 1 is rightof a, represented by R, and all 0 ≤ x < a is left of a, represented by L. Represent each

Page 108: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

100 Chapter 4. Graph Theory and Markov Models of Transport

step of the iteration map by one of these three symbols. All parameter sets for whichF(x) is Markov must have a period-n orbit containing the point x = a and be of the form{a, . . . , Fn−1(a), . . .

}. For example, the period three orbit has the form

{a,1,0, . . .

}and

occurs for any parameter set (a,b) on the curve b = a. It repeats the pattern CRL, whichwe call the kneading sequence K (Fa,b)= (C RL)∞.

If a periodic orbit contains the point x = a and a �= b, the point x = b will eitherbe greater or less than a. For a period four orbit with a > b, the symbolic sequence mustrepeat CRLL, or if a < b, CRLR. Therefore, a period four is found two ways. See

Figure 4.15. PDF (histogram bars) compared to the exact solution (solid line).The calculation used 50,000 iterations and a 50 intervals. [19]

Repeating this method for period five and higher, we see that there are 2n−3 possiblecombinations of the C-L-R sequences for a period n orbit which includes the critical point.The exponent n− 3 reflects the necessary 3-step prefix CRL. A full binary tree with 2n−3

leaves on each tier is possible, each implying a condition on the parameters and forminga countable set of curves in parameter space [238]. We can now restate Theorem 4.13 interms of kneading sequences.

Corollary 4.14. If K (Fa,b) is periodic, then Fa,b is Markov.

From this follows that the functions Fa,b(x) (4.65) which are Markov occur on adense set of curves in the (a,b) parameter space D0 (4.67). Taking �2 as the space ofsymbol sequences containing the full family of kneading sequences for two symbols andthe kneading sequence σ = σ0σ1σ2 . . ., with the metric d(σ , σ ) =∑∞i=0

|σi−σi |2i , and the

Page 109: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.4. Exact Representations are Dense and the Ulam-Galerkin’s Method 101

norm ||σ || =∑∞i=0σi2i = σ0+ σ1

2 + σ222 + . . ., we have

Lemma 4.1. Periodic σ are dense in �2.

Proof. ∀ε > 0 and given WOLOG φ = φ0φ1φ2 . . . which is not periodic, ∃N > 0 largeenough so that σi = φi , i ≤ N and σ = (φ0φ1 . . .φN )∞ for ||σ −φ||�2 < ε.

The result follows from combining Lemma 4.1 and Corollary 4.14. [19]

4.4.4 Probability density functions of Markov Maps are PiecewiseConstant

A useful property of Markov chaotic functions is that the probability density functions canbe determined quite easily. In fact, expanding piecewise linear Markov transformationshave piecewise constant invariant probability density functions. [19]

Theorem 4.15 ([53], Theorem 9.4.2). Let τ : I → I be a piecewise linear Markov trans-formation such that for some k ≥ 1,

|(τ k)′|> 1,

wherever the derivative exists. Then τ admits an invariant [probability] density functionwhich is piecewise constant on the partition P on which τ is Markov.

4.4.5 Invariant Measures of Non-Markov Transformations as WeakLimits from Markov Maps

First we must prove that Markov maps are dense in D0. Then, recalling results concern-ing weak limits of invariant measures of a sequence of transformations, we note that theMarkov techniques can be applied, in a limiting sense, to describe statistical properties forall Fa,b with a,b ∈ D0. The map fλ,μ studied in [238] is conjugate to the piecewise linear,interval map Fa,b we are studying.

fλ,μ(x)={

1+λx if x ≤ 01−μx if x ≥ 0.

with λ≤ 1, μ > 1, 0 < a < 1, and 0≤ b ≤ 1.Using the conjugacy,

ϑ(x)= (x−a)/(1−a), (4.68)

show that,ϑ−1 ◦ fλ,μ ◦ϑ(x)= Fa,b(x). (4.69)

Since ϑ(x) is a linear function of x , it is a homeomorphism and ϑ−1 is uniquely defined asϑ−1(x)= (1−a)x+a.

ϑ−1( f (ϑ(x)))={

(1−aλ)+λx if x ≤ a(1+aμ)−μx if x ≥ a.

(4.70)

Page 110: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

102 Chapter 4. Graph Theory and Markov Models of Transport

Let λ= (1−b)/a and μ= 1/(1−a). Then

1−aλ= 1−a

(1−b

a

)= b (4.71)

1+aμ= 1−μ

(1

1−a

)= 1

1−a. (4.72)

Therefore, ϑ−1 ◦ fλ,μ ◦ϑ(x) = Fa,b(x) by the homeomorphism ϑ and the two maps Fa,band fλ,μ are topologically conjugate.

Call M the class of sequences M which occur as kneading sequences of Fa, 1−2a1−a

for

0 < a ≤ 1/2, also known as the primary sequences. Misiurewicz and Visinescu proved thefollowing theorems:

Theorem 4.16 ([238], Theorem A). If (a,b), (a′,b′)∈ D and (a,b)> (a′,b′) then K (a,b)>K (a′,b′).

Theorem 4.17 ([238], Theorem B). If (a,b) ∈ D then K (a,b) ∈M.

Theorem 4.18 ([237], Intermediate Value Theorem for Kneading Sequences). If a one-parameter family Gt of continuous unimodal maps depends continuously on t and h(Gt ) >0 for all t then if K (Gt0) < K < K (Gt1 ) and K ∈M then there exists t between t0 and t1with K (Gt )= K .

Using these previous results, we can prove the following,

Theorem 4.19. ∀(a,b) ∈ D0, one of the following is true:

1. Fa,b is Markov.

2. ∃(a∗,b∗)D0 such that Fa∗,b∗ uniformly approximates Fa,b.

Proof. If Fa,b is Markov, we are done. Otherwise, choose Fa0,b0 such that K (Fa0,b0) isnonperiodic. Given a small ε > 0, let a0 = a∗ and b1 = b0 + ε. By Theorem 4.16,K (Fa∗,b0 ) < K (Fa∗,b1) and by Theorem 4.17, K (Fa∗,b0 ), K (Fa∗,b1) ∈M. WOLOG, wechoose indices of b0 and b1 to create this ordering. Recall by Lemma 4.1, that periodicsequences are dense in �2. Therefore, we may choose a sequence M ∈ M such thatK (Fa∗,b0 ) < M < K (Fa∗,b1). Since Fa,b does vary continuously with parameters a andb, then Theorem 4.18 implies an intermediate value b∗ such that b0 < b∗ < b1, and thisintermediate map has the decimal kneading K (Fa∗,b∗)= M . Therefore, in any given neigh-borhood of a non-Markov map in D0, there exists a Markov map M .

Hence, we can construct a sequence, in D0, of Markov maps that converges to anyFa,b with (a,b) ∈ D0. Considering Theorem 4.19, and the following result [53], we con-clude that any transformation Fa,b with (a,b) ∈ D0 is either a member of the Markov setwhich we constructed, and the invariant density function can be calculated directly as de-scribed earlier in this paper, or if Fa,b is not in that set, then a sequence of uniformlyconvergent Markov transformations, Fai ,bi → Fa,b and (ai ,bi )→ (a,b), each have easilycalculated invariant densities which converge to the invariant density of Fa,b.

Define Q be the set{c0,c1, . . . ,cp−1

}and P be the partition of I into closed intervals

with endpoints belonging to Q : I1 = [c0,c1], . . . , Ip−1 = [cp−2,cp−1]. [19]

Page 111: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

4.4. Exact Representations are Dense and the Ulam-Galerkin’s Method 103

Theorem 4.20 ([53], Theorem 10.3.2). Let f : I → I be a piecewise expanding transfor-mation, and let { fn}n≥1 be a family of Markov transformations associated with f . NoteQ(0) = Q, and

Q(k) =k⋃

j=0

f − j (Q(0)), k = 1,2, . . .

Moreover, we assume that fn → f uniformly on the set

I \⋃k≥0

Q(k)

and f ′n → f ′ in L1 as n →∞. Any fn has an invariant density hn and {hn}n≥1 is aprecompact set in L1. We claim that any limit point of {hn}n≥1 is an invariant density of f .

4.4.6 Density of Markov Maps Beyond One Dimensions

In the previous sections we have presented a rigorous treatment proving density, that nearbythe one dimensional skew tent maps, we can find a Markov map and therefore as such, anearby map has an exact finite representation of the Frobenius-Perron operator as a stochas-tic matrix. Further, by theorem summarized in Theorem 4.19, we know that the dominanteigenvectors of these resulting stochastic matrices can be used as a sequence of estimatesto the invariant density, which converge weakly. While we cannot back the generalizationswith theorem, intuition suggests that the idea is more general. For specific generalizationsof piecewise affine maps of the plane, the proofs will generalize in straightforward even iftedious manner. For maps of the interval that are not piecewise constant, such maps canbe continuously approximated by piecewise constant maps, from which a generalizing canagain be made. For several variable maps a similar strategy can be imagined but we leaveit as conjecture that the idea generalizes.

Page 112: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure
Page 113: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 5

Graph Partition Methods andTheir Relationship to Transportin Dynamical Systems

5.1 Graphs and PartitionsA major theme throughout this book is that appropriately partitioning the phase space is acritical aspect of analyzing transport. The Ulam-Garkin’s method discussed in the previouschapter, Chapter 4, gives a way to represent the transfer operator of a dynamical systemas a directed graph (either exactly in the case of a Markov representation in a sense ofSection 4.3 or otherwise approximately in the case of a fine set of basis functions). There-fore, the topic of this chapter will be how we can analyze questions of transport in termsof transport on a graph. The major aspect of transport analysis, partitioning into invariantsets, if they exist, or almost invariant sets allows us to borrow well developed theories forpartitioning graphs as we will review here. We will include discussion of strengths of eachapproach for our interest to the dynamical systems which they represent.

5.2 Weakly TransitiveA common scenario in advective turbulent dynamical systems is for the long time behaviorto be transitive (see in Definition 6.1), or correspondingly ergodic in terms of measurabledynamics, but the short time behavior may reveal only slow transport between almost in-variant sets. Such behavior has been called weakly transitive, [132, 135]. The followingexample is meant to emphasize almost-invariance and related graphs.

Example 5.1. (Map with Two Invariant Components) In Fig. 5.1, we see a one-dimen-sional dynamical system which has two invariant components that are the intervals labelledA and B . Using this partition, we get a representation by the directed graph also shownin Fig. 5.1(Bottom). This graph is completely reducible, as in Eq. (4.27). Compare to theexample in Example 3.9 studied for different tools.

By contrast, the dynamical system shown in Fig. 5.2 has two almost invariant compo-nents, labelled A and B . In fact, the escape set from A to B after one iteration is the intervallabelled y and colored yellow in a way which is representative of a cobweb diagram for theentire interval. That is,

eA := {x ∈ A : T (x) ∈ B} (5.1)

105

Page 114: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

106 Chapter 5. Graph Partition Methods and Transport

Figure 5.1. A one-dimensional dynamical system with two invariant components.As such this dynamical system lacks transitivity. Each of the of the intervals shown labelledA, and B are invariant under the this mapping. (Bottom) A directed graph representationrelative to the partition shown. The resulting graph is completely reducible. Compare tothe form of the generating matrix, Eq. (4.27). Contrast this example to that in Fig. 5.2.

where T is the dynamical system. In the above example, we have eA = y and furthercolored is the pre-image T−1(eA), of eA, which is the two thinner yellow subintervalsshown, again in a manner suggestive of a cobweb diagram. As we usually denote here, T−1

for these one-dimensional maps denotes the pre-images since the inverse does not exist; ithas the two branches shown and is thus multi-valued. Clearly, for any point x ∈ T−n(eA),we have T n+1(x)∈ B . Remark that it is not necessary true that T−n(eA)⊂ A in general, thatis, the orbit might escape from A to B and then return to A again before the nth iteration.Nevertheless, we are interested in defining a subset of A that escapes to B in term of thepreimage T−n . To achieve this, we may define the (first) escape time ξ (x) to be the smallestpositive integer such that T ξ (x) ∈ B . Then we could consider the set of points remain in Aup to the nth iteration of T , called An ,

An = {x ∈ A : ξ (x) > n}. (5.2)

So, the definition of An does not care about the fate of x after it escape from A. It is easyto check that

An = T−n(A)∩T−n+1(A)∩ . . .∩ A. (5.3)

In this notation, we can also define the set of points remain in A indefinitely by

A ⊃ A1 ⊃ ·· · ⊃ An , A∞ :=⋂n≥0

An . (5.4)

Page 115: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.2. Weakly Transitive 107

Figure 5.2. A one-dimensional dynamical system with two almost-invariantcomponents. Each of the of the intervals shown labelled A, and B are almost-invariantunder the mapping shown. The yellow region shown labelled y, and its pre-image(s) is theset of points which leave A in one, and two iterates respectively. Likewise, p is the subsetof points which leave B. (Bottom) The directed graph representation includes the weaktransport between A and B as the yellow and purple edges respectively. Compare to theform of the generating matrix, Eq. (5.9). Contrast this example to that in Fig. 5.1.

Generally, A∞ is a Cantor set of measure zero descriptive of the invariant set of A, butdepending on the details of the of the map, there can be missing branches complicating itscharacter [29]. The decay rate of a probability measure of An is of importance in under-standing stability of the open systems with holes, see [92, 64] for a review. Such a quantityis referred to as the escape rate of a reference measure m from A and it is given by

Em(A) :=− limn→∞

1

nlog(m(An)) (5.5)

Clearly, the escape rate lies between 0 and∞. For discussion here, the action of the map ony and p are depicted in Fig. 5.2(Bottom) by the yellow and purple edges respectively. Thefact that y is (depicted to be) relatively small compared to A, stated in terms of Lebesguemeasure m(·),

m(y)

m(A)<< 1, (5.6)

and likewise,m(p)

m(B)<< 1, (5.7)

denotes the weak transport descriptive of the almost-invariance of each of A and B .

Page 116: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

108 Chapter 5. Graph Partition Methods and Transport

As can be checked by the Definition 4.1, {A, B} is a Markov partition for Fig. 5.1, andas such the directed graph and corresponding matrix representations are exact in a senseof Section 4.3. Consequently the corresponding symbolic dynamics will be conjugate.However, by the same definition, {A, B} is not a Markov partition for Fig. 5.2. To make agood approximating representation by a directed graph, a fine grid covering may be used.As previously discussed, this is what we call the Galerkin-Ulam method. �

A dynamical system may be such that no partition exists that would reduce any graphrepresentation to a completely reducible form; as shown in Figs. 5.3(a)-5.4(b), and inSec. 4.3.1, a graph may be such that it can be brought “nearly" to a reducible form, allexcept for, “a few stray edges." A graph which is “almost reducible" is one which appearsas shown in Fig. 5.4(b). The corresponding transition matrix is seen in Fig. 5.3(b). Com-paring Figs. 5.3(a) and 5.4(a) to the Definition 4.3 of completely reducible yields a formEq. (4.27) by an appropriate sorting. This is another way of saying that there is a diag-onalizing permutation similarity transformation, Eq. (4.26). The major problem here isthat a graph in arbitrary configuration that may result from an arbitrary inspection45 of thedynamical system cannot be expected to present the resulting graph approximation, andits corresponding transition matrices, in the convenient almost-block-diagonal form as il-lustrated in Figs. 5.3(b)-5.4(b). Bringing the representation to such a convenient form isessentially equivalent to the problem of analyzing and understanding the transport mecha-nisms of the corresponding dynamical system.

We realize therefore the need for a notion of “almost reducible."

Definition 5.1. A graph shall be called almost reducible if there exists a subset Io ⊂ Isuch that there are “relatively few edges" (i , j ) for i ∈ Io and j ∈ I \ Io.

The above condition of “relative few edges" can be made more pricesly in the fol-lowing sense. In terms of the matrix representation of a graph by its transition matrix A,we can sharpen this notion by stating that this definition is equivalent to the statement thatan almost reducible matrix A must have a permutation matrix P such that,

R = P−1 AP , (5.8)

into a matrix R which is the form,

R =(

R1,1 E1,2E2,1 R2,2

). (5.9)

where E1,2 and E2,1 are relatively “small matrices," where small is to be measured in termsof the Frobenius-norm. The Frobenius-norm we remind is a matrix norm, which definedfor a matrix B is defined, [150],

‖B‖F =√

Tr (B · BT )=√√√√ n∑

i=1

n∑j=1

|bi, j |2. (5.10)

45We use the phrase “inspection" of a dynamical system as the choice and ordering of basis functionsused in sampling the action of the dynamical system through the Ulam-Galerkin method, Eqs. (4.2)-(4.4).Notice that even a simple permutation of the basis functions, Eq. (4.2) could easily change a configurationFig. 5.4(b) in the "dark star" appearance Fig. 5.4(a). In fact, a random embedding of a graph of any largenumber of vertices, say larger than N = 50, from basis functions in arbitrary configuration is expected tolook “all mixed up" like the dark star Fig. 5.4(a). In this type of Galerkin-Ulam graph analysis, we will oftenwork with N on the scale of many thousands and more.

Page 117: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.2. Weakly Transitive 109

(a) A transition matrix before sorting

(b) The same transition matrix after sorting

Figure 5.3. The matrix in this figure is irreducible. Members of the originaltransition matrix before sorting are generally expected to be placed haphazardly and thestructure is not obvious. After using the modularity method discussed in Section 5.10.2the “almost" block diagonal form is revealed. In this case, the sorted matrix indicatesthe existence of three communities and the few off block diagonal elements correspond tointra-basin diffusion.

Therefore, to make precise the notion of “relatively small" we mean that ‖E1,2‖F and‖E2,1‖F are small relative to ‖R1,2‖F and ‖R1,2‖F .

We summarize this introductory section and lead into the algorithmic work of thechapter with the statement that the problem of partition is central to the study of transport

Page 118: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

110 Chapter 5. Graph Partition Methods and Transport

(a) The original graph. Embedding vertices into theplane in a random configuration conceals simple struc-ture

(b) Sorted graph. The same graph subject to appro-priate permutation reveals “almost" block-diagonalstructure community between components

Figure 5.4. An undirected graph before and after sorting. The arrows point at theinter-community edges.

mechanism, flux and information flow in dynamical systems. Partitioning is squarely aproblem of associating sets together, and as such, the matrix representation amounts todefining appropriate permutations.

Page 119: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.3. Partition by Signs of The Second Eigenvector 111

5.3 Partition by Signs of The Second EigenvectorThere is a pleasingly simple but powerful method to partition an Ulam-Galerkin graph froma chaotic dynamical system to associate partition elements which are strongly transitivefrom other regions to which they may be only weakly transitive. Following the work of [93,96] for Markov chains, and similarly [135] for stochastic matrices from Frobenius-Perronoperators, we can inspect and threshold the eigenfunctions, and corresponding discreteapproximation of eigenvectors associated with secondary (not the largest) eigenvalues. Thediscussion here also has motivations from the early work with these methods in [95, 93].

We present here a simplified discussion of the thresholding concept in terms of acompletely reducible stochastic system. Assume a stochastic matrix,

P =(

P1,1 00 P2,2

). (5.11)

As such, it can be confirmed that the submatrices P1,1, and P2,2 must each be stochastic,albeit smaller of size N1× N1 and N2× N2, N = N1+ N2 where P is N × N . Then eachof P1,1, and P2,2 have dominant eigenvalues λ1 = 1 and λ2 = 1 respectively, and for sakeof presentation simplicity, we assume that each matrix is algebraically and geometricallysimple; it is sufficient to assume that they are each irreducible. So we write,

v1 P1 = v1, and v2 P2 = v2. (5.12)

Therefore,

Lemma 5.1. A stochastic matrix P of the reducible form Eq. (5.11) with irreducible com-ponents P1,1 and P2,2 has a multiplicity-2 dominant eigenvalue, λ = 1, and correspondingeigenvector spanning a two dimensional linear space, with eigenvector equations,(

v10

)· P =

(v10

), and,

(0v2

)· P =

(0v2

). (5.13)

The proof of this statement is immediate, keeping in mind the appropriate size of the0 vectors are N2×1 and N1×1 respectively.

In the span of {(

v10

),

(0v2

)} is included the vector, w =

( |v1|−|v2|

). Note that

Eq. (5.13) allows us to write,

(

( |v1|0

)−(

0|v2|

)) · S =

( |v1|−|v2|

)· S =

( |v1|−|v2|

). (5.14)

Regarding the question then of given a stochastic matrix P which may not be in blockdiagonal form, given a permutation,46 similarity transformation,47

P = ST · P · S, (5.15)

that brings it there, we can make the following statement regarding signs and a bi-partition.

46A permutation S is a square matrix that is simply a “resorting" of the identity matrix. An “elementarypermutation" S[i, j ] starts with the identity I, and we exchange the i’th and j’th rows. Therefore, appliedas a similarity transformation, ST

[i, j ] · P · S[i, j ] results in trading the positions of the i’th and j’th rows andcolumns of S. More complicated permutations can be formed as products (compositions) of such operations.S =�i, j P[ki ,kj ]

47Similarity transformations are especially relevant here since sprectral properties are preserved.

Page 120: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

112 Chapter 5. Graph Partition Methods and Transport

Proposition 5.2. Suppose P is a stochastic matrix that is a permutation of a stochasticmatrix P , P = ST · P · S, and P has components of the form Eq. (5.11). Then a dominanteigenvector v of P corresponding to λ = 1, which has opposite signs in its components,can be used to formulate a permutation S which diagonalizes P . Let,

s = sor t(v), 48 (5.16)

and let,S = perm(s), (5.17)

be a corresponding permutation matrix,49 then ST P S has a block diagonal form.

This Proposition describes the situation of a completely reducible system. The fol-lowing discussion is motivated in part from [95, 93] in what was called a, “Gedankenex-periment" for the basic idea that the matrix representations of almost invariant sets can beunderstood as perturbations of invariant sets. What if we perturb this completely reducibleparadigm slightly? In considering,

P �→ P+ E , (5.18)

as in comparing Eq. (5.11) to Eq. (5.9), we consider continuity results together with simplelinear algebraic considerations. The weakly transitive case Eq. (5.9) of a stochastic matrixR and its partitioning problem can be well understood in terms of Proposition 5.2, andnotions of continuity.

Theorem 5.3. (Continuity Eigenvalues The eigenvalues of a matrix P are continuouswith respect to variations of its matrix entries ai, j .

The proof of this statement is standard as can be found in [150], but is easily under-stood in terms of the statement that eigenvalues are roots of the polynomial det(P−λI )=0, and these are a continuous with respect to the coefficients. Alternatively and more typi-cally it can be understood in terms of the Gersgorin disc theory.

As for our problem Eq. (5.9), continuity results in the statement,

Corollary 5.4. If a block diagonalizable stochastic matrix is stochastically perturbed,50

P =(

P1,1 00 P2,2

)�→ P =

(P1,1+ εE1,1 εE1,2

εE2,1 P2,2+ εE2,2

). (5.19)

is continuous with respect to ε, or alternatively stated, in terms of each ‖εEi, j‖F . Further-more, since P is also assumed stochastic, ε: 0 < ε << 1 =⇒ 0≤ λ1−λ2 << 1.

48Let sort(v) be the vector of re-ordered elements of v from largest to smallest, and let sort(v) be thevector of integers whose i’th element records the original position of the i’th element of sort(v). E.g., ifv =< 1.6,3.2,−2.2 >, then sort(v) =< 3.2,1.6,−2.2 > and sort(v) =< 2,1,3 >

49If s =< 2,1,3 >, then S = perm(s) = [0,1,0;1,0,0;0,0,1] is a row restructuring of the 3× 3 identitymatrix, and the similarity transformation exchanges rows and columns of other matrices as specified.

50To remain stochastic, either P must be column rescaled, or the matrices εE1,2 and εE2,1 must includethe appropriate negative entries so that any variations to the total matrix S will be column-wise zero-sumchanges.

Page 121: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.3. Partition by Signs of The Second Eigenvector 113

By assumption that the perturbed matrix P remains stochastic, the statement is thatthe dominant eigenvalue remains λ1 = 1. However, while the second eigenvalue of P mayalso be one, generally the second eigenvalue of P may deviate from one. Nonetheless bycontinuity, it will remain close to one.

Example 5.2. Consider the matrix,

P =(

1 00 1

). (5.20)

Then, λ1 = λ2 = 1. The eigenspace of P is any spanning set of R2. It is con-

venient to use the orthonormal basis, {(

10

),

(01

)}, but it is just as suitable to use

{(

10

),

(11

)} as the eigenvectors of 1. A stochastically perturbed matrix may be cho-

sen for example,

P =(

1+ ε −ε−ε 1+ ε

), (5.21)

where −1 < ε < 0 to remain stochastic. As usual, P diagonalizes by a similarity transfor-mation in terms of the matrix whose columns are the (new) eigenvectors,(

1+ ε −ε−ε 1+ ε

)=(

1√2

1√2

1√2− 1√

2

)·(

1+2ε 00 1

)·(

1√2

1√2

1√2− 1√

2

), (5.22)

Notice that the diagonalizing similarity transformation is special,

P = V DV−1 = V DV t , (5.23)

since V whose columns are the eigenvectors happens to be unitary matrix51.There are certain points which may be interpreted from this example,

Remark: The eigenvalues are continuous with respect to variations in the matrix entries, asis promised by the general theory. We see specifically in this example that ‖P− P‖F = ε,and correspondingly,

λ1(P)= λ1(P)= 1, (5.24)

but,|λ2(P)−λ2(P)| = 2ε. (5.25)

However, the general theory only promises that eigenvectors must vary continuously if theeigenvalues are bounded away from a nontrivial multiplicity. In this case, since P has aλ1 = λ2 = 1 multiplicity-2, eigenvectors are not guaranteed to vary continuously. We seethat while

v1 =(

11

),v2 =

(1−1

)(5.26)

are eigenvectors of P (and also of P), these eigenvectors are not close to those eigenvectorsof P ,

v1 =(

10

),v2 =

(01

), (5.27)

51A matrix is called unitary if V t V = V V t = I .

Page 122: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

114 Chapter 5. Graph Partition Methods and Transport

which are (also) eigenvectors of P . This is the “wobble" problem [150] of lack of continuityassociated with geometrically nonsimple eigenspaces due to geometric multiplicity.

Remark The second eigenvalue of P is λ2 = 1+ 2ε, and its corresponding eigenvector

Eq. (5.27), v2 =(

1−1

), is notable in that the opposite signs are indicative of the par-

titions in more complicated almost reducible problems. The important part of v2 can beinterpreted as,

Signs(v2)=( +−)

. (5.28)

The generalized statement of the signs of the second eigenvector can be used to revealalmost invariance, and associate the appropriate partition even if the partition is not obviousas the vector is as good as scrambled with obvious pattern. The following is a variation onthe statements in [93, 96],

Proposition 5.5. The signs vector of the second eigenvector,

Signs(v2)=

⎛⎜⎜⎜⎜⎜⎜⎜⎜⎜⎜⎜⎜⎜⎝

+++..+−−−...−

⎞⎟⎟⎟⎟⎟⎟⎟⎟⎟⎟⎟⎟⎟⎠, (5.29)

with m, +’s, and n, -’s, is associated with

P =(

P1,1+ εE1,1 εE1,2εE2,1 P2,2+ εE2,2

)(5.30)

from Eq. (5.19) in Corollary 5.4, when P1,1 is m×m and P2,2 is n×n and ε << 1.

Now this proposition can be used to uncover a permutation which can be used toreveal a form P in Eq. (5.30). In other words, suppose P is a matrix presented in arbitraryconfiguration. That means that while we do not have the nice almost block diagonal formof P , there exists a permutation matrix S which re-arranges the rows and columns,

P = SPSt . (5.31)

The difficulty of the problem is that when presented with a P , but without an S, we need amethod to know if a re-arrangement to an almost block diagonal form P exists, or not; theessence of the partition problem then is to construct a permutation S52 if it exists.53 Thefollowing constructs a useful partition, as consequence of Proposition 5.5.

52Such a partition is not unique in that elements in a component may be permuted.53Note that such a P is not unique.

Page 123: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.3. Partition by Signs of The Second Eigenvector 115

Corollary 5.6. Let,[Y , J ]= Sort(Signs(v2)), 54 (5.32)

where v2 is the second eigenvector of a stochastic matrix P in arbitrary configuration. Thenif there exists a permutation S such that P has an almost block diagonal form, Eq. (5.30)where ε << 1, then a suitable permutation matrix may be formed by by rearranging therows of the identity matrix I according to the sort-indices J from Eq. (5.32),

S = IJ ,:, (5.33)

and this permutation can unfold the partition,

P = SPSt . (5.34)

In summary, the statement is simple. While the first eigenvector may encode the steadystate of the stochastic process, the second eigenvector reveals a partition associated withalmost invariance. If a block diagonal form of the stochastic matrix in general configurationexists, the sort on the second eigenvector associates the partition which reveals the almostinvariant sets.

Example 5.3. Let,

P =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.1430 0.4293 0.7359 0 0 0 00.3007 0.4084 0.1111 0 0 0 00.5563 0.1623 0.1530 0 0 0 0

0 0 0 0.0886 0.3202 0.2036 0.30450 0 0 0.4557 0.0916 0.3345 0.32700 0 0 0.3333 0.0327 0.2510 0.36840 0 0 0.1224 0.5555 0.2109 0.0001

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.35)

in the form of Eq. (5.19) and likewise, let

P =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.1428 0.4290 0.7359 0 0 0 0.00010.3007 0.4084 0.1110 0.0001 0.0001 0.0001 00.5563 0.1623 0.1530 0 0 0 0

0 0 0 0.0886 0.3201 0.2035 0.30440.0001 0 0 0.4556 0.0916 0.3345 0.3270

0 0.0002 0 0.3333 0.0327 0.2510 0.36840.0001 0.0001 0.0001 0.1224 0.5555 0.2109 0.0001

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.36)

also in the form of Eq. (5.19), the stochastically perturbed block almost diagonal variationand scaled to remain stochastic. Then corresponding to P is the spectrum,

eigs(P)={1.0000,1.0000,−0.5219,−0.2786−0.1294i ,−0.2786+0.1294i ,0.2263,−0.0115}(5.37)

54We interpret the Sort function similarly to the sort function stated in Eq. (5.16). While sort results ina vector of entries listed from largest to smallest, we define Sort to have an output whose values in Y arethose from sort listed from largest to smallest, and J is the corresponding rearranged indices.

Page 124: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

116 Chapter 5. Graph Partition Methods and Transport

in which we see the multiplicity-2, λ1 = 1 which agrees with the prediction of Eq. (5.24).The spectrum of the perturbed matrix P is close to that of P , but the repeated eigenvalueshave separated, as promised by Corollary 5.4 with one of them remaining λ1 = 1.

eigs(P)={1.0000,0.9997,−0.5220,−0.2786−0.1293i ,−0.2786+0.1293i ,0.2264,−0.0115}(5.38)

The second eigenvector, v2 of P has signs which reveal with the block partitioning of P ,

v1 =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

−0.3596−0.2358−0.2814−0.4008−0.4963−0.4015−0.4096

⎞⎟⎟⎟⎟⎟⎟⎟⎠,v2 =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.53150.34820.4159−0.3040−0.3765−0.3045−0.3106

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.39)

. Again we remind that the scale and sign of a given vector is not important since a scaledfactor avi of an eigenvector vi is also an eigenvector. Many software packages normalizethe vectors before they are presented but they may be given as all negative values for ex-ample as we see v1 here. To interpret as probabilities, we would need to make sure to useit with positive entries and normalized.

If we hide the block almost diagonal structure by scrambling P

SPS′ =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.0886 0 0.3201 0.2035 0.3044 0 00 0.1428 0 0 0.0001 0.4290 0.7359

0.4556 0.0001 0.0916 0.3345 0.3270 0 00.3333 0 0.0327 0.2510 0.3684 0.0002 00.1224 0.0001 0.5555 0.2109 0.0001 0.0001 0.00010.0001 0.3007 0.0001 0.0001 0 0.4084 0.1110

0 0.5563 0 0 0 0.1623 0.1530

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.40)

as would be expected in a general configuration, in this case due to a similarity transforma-tion by the permutation matrix,

S =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0 0 0 1 0 0 01 0 0 0 0 0 00 0 0 0 1 0 00 0 0 0 0 1 00 0 0 0 0 0 10 1 0 0 0 0 00 0 1 0 0 0 0

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.41)

then the second eigenvector can still be used to reveal the structure. S here is a permutationof the rows of the identity matrix. The second eigenvector of the scrambled matrix, SPS′is

v2 =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

−0.30400.5315−0.3765−0.3045−0.31060.34820.4159

⎞⎟⎟⎟⎟⎟⎟⎟⎠,sign(v2)=

⎛⎜⎜⎜⎜⎜⎜⎜⎝

−11−1−1−111

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.42)

Page 125: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.4. Graph Laplacian and Almost invariance 117

Using Eq. (5.32), [Y , J ]= Sort(Signs(v2)), yields

J =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

1345267

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.43)

Developing a new permutation matrix S from this permutation as a permutation,

S =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

1 0 0 0 0 0 00 0 1 0 0 0 00 0 0 1 0 0 00 0 0 0 1 0 00 1 0 0 0 0 00 0 0 0 0 1 00 0 0 0 0 0 1

⎞⎟⎟⎟⎟⎟⎟⎟⎠, (5.44)

unscrambles the matrix in Eq. (5.40) and yields,

S(SPS′)S′ =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.0886 0.3201 0.2035 0.3044 0 0 00.4556 0.0916 0.3345 0.3270 0.0001 0 00.3333 0.0327 0.2510 0.3684 0 0.0002 00.1224 0.5555 0.2109 0.0001 0.0001 0.0001 0.0001

0 0 0 0.0001 0.1428 0.4290 0.73590.0001 0.0001 0.0001 0 0.3007 0.4084 0.1110

0 0 0 0 0.5563 0.1623 0.1530

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.45)

which is again in a block almost diagonal form, albeit not identically to the original P ,Eq. (5.36), due to nonuniqueness of almost diagonalizing permutations. �

Example 5.4. Graphical Presentation of Almost Block Diagonal Decomposition Thematrix shown as a spy plot in Fig. 5.3(a)(Left) in an arbitrary configuration, and an arbitraryembedding of the corresponding graph in the plane, Fig. 5.4(a) both belie whether or notthere may be any useful structure. The second eigenvector signs method of this sectioncan be used to reveal the structure. The structure becomes obvious in the permuted versionof the same matrix, Fig. 5.3(b) since it exists. Likewise, an appropriate embedding in theplane reveals the same fact, Fig. 5.4(b).

While we have specialized the discussion of partitioning methods by inspection ofthe second eigenvector to stochastic martices, as did [132, 135], the work in [93, 96] ismore general in its application to transition matrices and Markov models. See also, [121]for discussion of an SVD algorithm to almost block diagonal matrices and see Section 5.7for discussion of some SVD connection.

5.4 Graph Laplacian and Almost invarianceIt has been previously demonstrated that the sign structure of the second eigenvector of thetransition matrix suggests a bi-partition of a graph. In this section, we will relate the opti-

Page 126: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

118 Chapter 5. Graph Partition Methods and Transport

mal partitioning by the eigenvectors of the Laplacian matrix with the (measure-theoretic)almost-invariance.

We will begin by assuming that the transition matrix P ∈ Rn×n is irreducible andsymmetric with Pi, j = Pj ,i for i , j ∈ {1, . . . ,n}. In the language of the graph network, thistransition matrix presents a weighted, undirected and symmetric graph. This implies that e,a constant vector with a value 1/n on each component, is the stationary distribution. Thisimplies there is a uniform probability to be in any state of Markov chain represented byP . By symmetry, P is also doubly-stochastic55, so the Markov chain represented by P isreversible56. Throughout this section, it will be assumed that there is no “isolated" statein the Markov chain (i.e. Pi,i < 1 for all i ) and the matrix is row-stochastic so that eachcomponent represents a conditional probability of the corresponding Markov chain.

A Laplacian matrix is defined as,

L = I − P , (5.46)

where I is the identity matrix with the dimension of P . Then, for any y ∈Rn , we can showthat

yT Ly = yT (I − P)y =∑

i

y2i −∑i, j

yi yj Pi, j

=∑

i

(∑

j

Pi, j )y2i −∑i, j

yi yj Pi, j

= 1

2

∑i, j

(yi − yj )2 Pi, j . (5.47)

In going from the first to the second line, we use the the fact that P is row-stochastic andthe symmetry of P allows us to write the quadratic form in the last line. Therefore, we canwrite the so-called “Rayleigh quotient" as:

yT Ly

yT y= 1

2

∑i, j (yi − yj )2 Pi, j∑

i y2i

. (5.48)

How is this related to almost-invariance? Supposed that we want to divide the n states ofMarkov chain governed by P into two non-empty disjoint subsets I1 and I2. To describethis 2-partition, we consider a binary vector y ∈±1 such that yi = 1 if i ∈ I1 while if i ∈ I2,yi =−1. Substituting this binary vector y into (5.48) yields

yT Ly

yT y= 2

n

⎛⎝∑I1→I2

Pi, j +∑

I2→I1

Pi, j

⎞⎠= 4

n

∑I1→I2

Pi, j . (5.49)

55Doubly-stochastic is defined to be simultaneously row and column stochastic.56A Markov chain represented by a transition matrix P is called reversible if pi Pi, j = pj Pj ,i for all i , j

and pi is a probability to be in the i−th state. In particular, we will use the stationary distribution of P tobe our probability measure p, i.e., pP = p. The reversibility condition implies that the probability to movefrom state i to j is the same as from j to i .

Page 127: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.4. Graph Laplacian and Almost invariance 119

The subscript I1 → I2 indicates that the sum is taken over the index i ∈ I1 and j ∈ I2 andsimilarly for I2 → I1. Also, the constant term can be excluded in the optimization. It isnow clear that minimizing the above Rayleigh quotient for y ∈ ±1 is equivalent to findingthe optimal 2-partition that gives the minimal probability to move from one set to the other(from I1 to I2 and vice versa).

Remark 5.1. Optimal Invariance. A pair of sets I1 and I2 that minimizes the Rayleighquotient is maximally almost-invariant, with respect to the uniform probability measure asassumed for the present discussion.

The task now is to find such an optimal vector y ∈ ±1. Unfortunately, this problemis a difficult combinatorial problem if we wish to tackle it by brute force. Therefore, as aheuristic means of finding a good solution, we relax the binary restriction on y and allowcontinuous real values. Furthermore, we add a constraint that,∑

i

yi = 0, (i.e., yT e = 0). (5.50)

This is equivalent to requiring that the partition I1 and I2 have the same “weight sum",i.e., the sum of positive part equals to the negative part. The only reason of adding thisconstraint is not that we want to balance the size of I1 and I2

57, but rather, the constraintallows an application of the well-known Courant-Fisher Theorem, which will be presentedsubsequently. Based on the above discussion, our relaxed problem can be expressed by:

MinimizeyT Ly

yT yy ∈ Rn

Subject to: yT e = 0. (5.51)

To this end, we may apply the well-know Courant-Fischer Theorem [150] to the min-imization problem (5.51).

Theorem 5.7 (Courant-Fischer Theorem). Let A be a real symmetric matrix with eigenval-ues 0≤ λ1 ≤ λ2 ≤ ·· · ≤ λn . Let Xk be a k−1 dimensional subspace of Rn and x ⊥ Xk (i.e.x ⊥ y for all y ∈ Xk ). Then

λi = minXn−i+1

maxx⊥Xn−i+1,x �=0

xT Ax

xT x=max

Ximin

x⊥Xi ,x �=0

xT Ax

xT x. (5.52)

The condition that A is real symmetric ensures the existence of a full set of orthogo-nal eigenvectors vi corresponding to λi ≥ 0 for i = 1, . . . ,n. Remark that the Xn−i+1 andx that result in λi in the first expression are indeed the span of the last n− i +1 eigenvec-tors and the n− i -th eigenvector, respectively, while in the second expression λi is attainedwhen Xi is the span of the first i eigenvectors and x is the i+1-th eigenvector. These factsare normally shown in the standard proof of the Courant-Fischer theorem for which we

57Nevertheless, this extra constraint will usually be relaxed since we will not strictly use the sign structureof the eigenvector for partitioning, but use it as a “level-set" in the thresholding algorithm, some sort of 1Doptimization.

Page 128: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

120 Chapter 5. Graph Partition Methods and Transport

refer the popular graduate textbook in linear algebra [150].

To apply the Courant-Fischer Theorem for our partitioning problem, we have to ver-ify that λ1 = 0, the smallest eigenvalue, and the eigenvector corresponding to λ1 = 0 is aconstant vector; this can be easily checked as,

eL = e(I − P)= e− e = 0, (5.53)

where we use the fact that P is doubly stochastic. Thus, based on the above comments, wecan calculate λ2 by

λ2 = minyT e=0,y �=0

yT Ly

yT y. (5.54)

It follows immediately from the above equation that the minimizing vector y that solvesour relaxed problem is the eigenvector v2 corresponding to λ2. It should be noted thatv2 must have both positive and negative parts in order to satisfy the orthogonality with e.Consequently, the positivity of λ2 > 0 is guaranteed by (5.47) and the presumption thatP is irreducible; hence there will be at least one pair of i and j with opposite signs tomake the summation in (5.47) positive. We close the case of a symmetric P here by brieflymentioning that the eigenvector v2 can be used to create our partition by interpreting thevalue of v2 as “fuzzy inclusion"; if the i−th component of v2 is strongly positive, the i−thstate of the Markov chain will be very likely to be in I1 and likewise for those with strongnegativity, they are likely to be in I2. For those near zero, the inclusion is less certain andan optimization algorithm is required to be the partition, see, for example, [131, 138].

It is fitting at this point to introduce an extension of the preceding technique to ageneral non-symmetric transition matrix, which corresponds to a weighted, directed graphor a non-reversible Markov chain. By observing the steps in the previous analysis, it isevident that the conditions that P is reversible and symmetric are the key properties; theformer allows us to write the Rayleigh quotient in the form of a quadratic cost functionwhile the latter is used to invoke the Courant-Fischer Theorem. Therefore, a natural wayto extend the previous technique is to first “reversibilize" the transition matrix P so that weobtain a new matrix representing a reversible Markov chain (the symmetry issue will bedealt with later on).

Before proceeding further, we need to introduce a number of variables required in thefollowing analysis. Since P is non-symmetric, in general e is not necessarily a stationarydensity of P; in fact, e is the stationary density only if P is still doubly stochastic. There-fore, we will use the eigenvector p such that pP = p to be our measure, i.e., the probabilityto move from state i to state j is pi Pi j .

Consider the transition matrix P of the time-reversed chain defined by

Pi, j = pj Pj ,i/pi (5.55)

or, in a matrix form,P =�−1 PT �, (5.56)

where � is a diagonal matrix with �i,i = pi . It is easy to check that pP = p. The additivereversibilization of P is defined by

R = 1

2(P+ P). (5.57)

Page 129: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.4. Graph Laplacian and Almost invariance 121

The Markov chain represented by the additively reversibilized transition matrix R is re-versible and has p as the stationary distribution as well, i.e., under the dynamic governingby the Markov chain R, the probability to go from states i to j is the same as going fromstates j to i (i.e. pi Ri, j = pj Rj ,i ). This has also been called “detail balance" in the lan-guage of statistical mechanics

How does the matrix R relate to the our minimization? Similar to what we have donepreviously, our aim is to minimize the cost function

Minimize

∑i, j (yi − yj )2 pi Pi, j∑

i y2i pi

y ∈ Rn

Subject to:∑

i

yi pi = 0. (5.58)

The main difference between this cost function and the preceding one for the symmetriccase is that the measure is now p instead of the uniform distribution e. To relate the aboveminimization problem to R, we rewrite the above cost function as∑

i, j

(yi − yj )2 pi Pi, j =

∑i, j

(yi − yj )2 (pi Pi, j + pj Pj ,i )

2

=∑i, j

(yi − yj )2 pi Ri, j (5.59)

Therefore, we can minimize the quadratic form related to R instead.Next, we want to rewrite the above summation in the form similar to the Laplacian

matrix we have used before. To achieve this, we can utilize the reversibility of R and thestationary measure p to show that∑

i, j

(yi − yj )2 pi Ri, j =

∑i

y2i pi −

∑i, j

yi yj pi Ri, j

=∑

i

x2i −∑i, j

xi xj (p1/2i Ri, j p−1/2

j ), (5.60)

where we substitute,x =�1/2y. (5.61)

For such x , it follows that

xT (I −�−1/2 R�1/2)x

xT x=∑

i, j (yi − yj )2 pi Pi, j∑i y2

i pi. (5.62)

We now define the normalized Laplacian by

L := I −�−1/2R�1/2 = I − �1/2 P�−1/2+�−1/2PT �1/2

2.58 (5.63)

58�1/2 denotes the “square root" of the matrix � in the sense that the square root of a square matrix

Page 130: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

122 Chapter 5. Graph Partition Methods and Transport

An extensive study of the graph spectrum and their relations to properties and structuresof a graph is covered in [61]. It is clear that L is symmetric, so we can now follow thesame heuristic as in the case of the symmetric P in order to find a partition by applyingthe Courant-Fischer Theorem to L. More precisely, according to (5.62) and the defini-tion (5.63), the optimization problem (5.58) can be rewritten as

MinimizexT (L)x

xT x

Subject to:∑

i

xi p1/2i = 0. (5.64)

Notice that due to the equation x = �1/2y the optimal vector x has to be constrained tobe orthogonal to p1/2 for y to be orthogonal to p. Also, we note that after obtaining theoptimal vector x , we have to revert it to y. To apply the Courant-Fischer Theorem like whatwe did previously, we have to check that p1/2 is indeed the eigenvector corresponding toλ1 = 0. This can be easily verified as the following:

p1/2L= p1/2(

I − (�1/2 P�−1/2+�−1/2 PT �1/2)

2

)= p1/2− pP�−1/2+ ePT�1/2

2

= p1/2− p�−1/2+ e�1/2

2

= p1/2− p1/2+ p1/2

2= 0. (5.65)

To derive the third line from the second line, we use the fact that pP = p and Pe = e.Therefore, the optimal value of the problem (5.58) is

λ2 = minxT p1/2=0,x �=0

xT Lx

xT x= min

yT p=0,y �=0

∑i, j (yi − yj )2 pi Pi, j∑

i y2i pi

(5.66)

attained by the eigenvector x = v2 (i.e. y =�−1/2v2) corresponding to λ2 of L.

Example 5.5. Let,

P =

⎛⎜⎜⎜⎝0.1 0.5 0.2 0.1 0.10.6 0.3 0.1 0.0 0.00.2 0.2 0.2 0.2 0.20.1 0.2 0.0 0.3 0.40.1 0.2 0.0 0.4 0.3

⎞⎟⎟⎟⎠ (5.67)

A can be taken to be a matrix B such that A = B2. Generally the existence and computation of sucha B requires conditions and an algorithm, but in the case that A is diagonal, as in fact � is diagonal,then the square root of the matrix is especially simple to compute by the square root of all of its entries:A1/2 = diag(

√A1,1, ..,

√An,n ). Likewise the inverse of a diagonal matrix is particularly simple to compute

by inverting the numbers on the diagonal. Therefore, �−1/2 := (�1/2)−1 = diag(1/√

A1,1, ..,1/√

An,n ),assuming none of the diagonal elements are zero as the condition for nonsingularity.

Page 131: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.4. Graph Laplacian and Almost invariance 123

It is quite clear that we have two separate groups {1,2} and {4,5}, where a group is takento mean as a highly connected subset of vertices. However, the group to which we shouldput {3} in order to obtain the minimal transport is ambiguous. The stationary distribution pcan be obtained by computing the left eigenvector of corresponding to the unit eigenvalueof P as:

p = {0.5558,0.6497,0.2202,0.3321,0.3321}T .

Based on the measure p, we can calculate the normalized Laplacian as per Eq. (5.63) to be

L=

⎛⎜⎜⎜⎝0.9000 −0.5556 −0.2218 −0.1033 −0.1033−0.5556 0.7000 −0.1441 −0.0715 −0.0715−0.2218 −0.1441 0.8000 −0.0814 −0.0814−0.1033 −0.0715 −0.0814 0.7000 −0.4000−0.1033 −0.0715 −0.0814 −0.4000 0.7000

⎞⎟⎟⎟⎠ (5.68)

The eigenvalues of L are λ1 = 0,λ2 = 0.4320,λ3= 0.8907,λ4= 1.1000,λ5 = 1.3773. Theoptimal vector can then be obtained by calculating y =�−1/2v2 as

y = [0.4486,0.5571,0.2577,−1.0059,−1.0059]. (5.69)

It is worthwhile to check that∑

i yi pi = 0 as required in the constrain optimization (5.58).Therefore, the 2-partition based on the parity of optimal vector y in this example is I1 ={1,2,3} and I2 = {4,5}, which gives∑

I1→I2

pi Pi, j +∑

I2→I1

pi Pi, j = 0.1992+0.1992= 0.3984; (5.70)

It can checked that the above partition does result in a lower value in cost function than analternative 2-partition I1 = {1,2} and I2 = {3,4,5}, which gives∑

I1→I2

pi Pi, j +∑

I2→I1

pi Pi, j = 0.2873+0.2873= 0.5746. (5.71)

So, this partition has a greater “leak".

Example 5.6. Consider the matrix P in (5.36). We will alternatively attempt to find thepartition based on the eigenvector of the generalized Laplacian matrix. Since the setup ofa stochastic matrix in this section is row-stochastic, we will use PT instead of P .

PT =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.1428 0.3007 0.5563 0 0.0001 0 0.00010.4290 0.4084 0.1623 0 0 0.0002 0.00010.7359 0.1110 0.1530 0 0 0 0.0001

0 0.0001 0 0.0886 0.4556 0.3333 0.12240 0.0001 0 0.3201 0.0916 0.0327 0.55550 0.0001 0 0.2035 0.3345 0.2510 0.2109

0.0001 0 0 0.3044 0.3270 0.3684 0.0001

⎞⎟⎟⎟⎟⎟⎟⎟⎠(5.72)

The stationary distribution p of the above matrix is calculated to be

p = [0.3596,0.2358,0.2814,0.4008,0.4963,0.4015,0.4096]T .

Page 132: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

124 Chapter 5. Graph Partition Methods and Transport

Figure 5.5. An example of a coherent set in a time-dependent dynamical system.The equation of the underlying flow is not important here, but the distinct behavior of thecoherent sets (showing in red) comparing to other sets is clear; they move coherently alongwith the flow and rarely disperse.

Based on the vector p, the normalized Laplacian is

L=

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0.8572 −0.3594 −0.6399 0 0 0 −0.0001−0.3594 0.5916 −0.1349 −0.0001 −0.0001 −0.0001 0−0.6399 −0.1349 0.8470 0 0 0 0

0 −0.0001 0 0.9114 −0.3828 −0.2683 −0.21440 −0.0001 0 −0.3828 0.9084 −0.1686 −0.45430 −0.0001 0 −0.2683 −0.1686 0.7490 −0.2904

−0.0001 0 0 −0.2144 −0.4543 −0.2904 0.9999

⎞⎟⎟⎟⎟⎟⎟⎟⎠The eigenvalues of L are λi = {0,0.0003,0.7730,0.9453,1.1543,1.4690,1.5226}. That λ2is nearly zero suggests a potential of almost-invariant partitioning. The optimal vector canthen be calculated from the equation y =�−1/2v2 as

y = [−0.8682,−0.8677,−0.8683,0.4457,0.4456,0.4456,0.4456]. (5.73)

The sign structure of y clearly reveals the optimal 2-partition I1={1,2,3} and I2={4,5,6,7}.

5.5 Finite-time coherent setsIntensive study has been dedicated to understand how to define and/or detect coherent sets(or structures). Many of these definitions are based on a physical quantities (e.g. usingpotential vorticity to define a polar vortex as a coherent set), which may not always beoptimal. Alternatively, the approach reviewed here is to identify optimal coherent structuresbased only on the underlying flow [138]. For a general time-dependent or nonautonomousdynamical systems, it can be imagined, in analogy to the almost invariant set, that a setmoving collectively in a minimally dispersive manner is the subject of our interest. Forinstance, Figure 5.5 shows a time sequence of a set that self evidently moved coherentlywith the underlying flow. A set with this kind of behavior will be called here as a “coherent"set; a formalism will be given in what follows.

Page 133: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.5. Finite-time coherent sets 125

The concept of coherent sets may be deemed as a generalization of the almost invari-ant set. Given a time-dependent flow map,

�(z, t;τ ) : M ×R×R→ R, (5.74)

which described the terminal point time t + τ of an initial point z at time t , we desire todiscover “coherent pairs" of subsets At and At+τ ⊂ M such that,

�(At , t;τ )≈ At+τ , (5.75)

i.e., At is mapped “mostly" to At+τ by the flow �. This approximate statement is madeprecise in terms of measure.

Definition 5.2. We will call At , At+τ a (ρ0, t ,τ )−coherent pair if

1.

ρμ(At , At+τ ) := μ(At ∩�(At+τ , t+ τ ;−τ ))

μ(At )≥ ρ0 (5.76)

2. μ(At )= μ(At+τ )

3. At and At+τ are "robust" to small perturbation

In the above definition, the measure μ is called a "reference" probability measure attime t , which usually describes the mass distribution of the quantity we wish to study thetransport over the interval [t , t + τ ]. Note that the reference measure does not have to beinvariant under the flow �. It is clear that the first two conditions in the above definitionprescribe the coherency. However, a (1, t ,τ )−coherent pair can be trivially produced bychoosing an arbitrary At and At+τ = �(At , t;τ ). To eliminate this pathology, the thirdcondition is used to select only those pairs that remain coherent under small diffusive per-turbation of the flow.

In chaotic systems, it is usually the case that an arbitrarily chosen coherent pair willexperience stretching and folding and after a sufficiently long time τ they will become ge-ometrically thin and irregular. Consequently, a small amount of diffusion is able to ejectmany particles from At+τ and hence reduce the coherence ratio ρμ(At , At+τ ). Therefore,the requirement of robustness under perturbation favors a large-scale coherent set with ageometrically regular structure, which is of interest in most applications. Since the trans-fer operator involves a spatial discretization, hence discretization-induced diffusion is in-evitable. This explains why the transfer operator approach tends to capture a large-scaleregular structure. Our basic tool for identifying the coherent pair will again be the transfer(Perron-Frobenius) operator. For a given t and τ ,

Pt ,τ : L1(M) �, (5.77)

defined byPt ,τ f (z) := f (�(z, t+ τ ;−τ )) · |det D�(z, t+ τ ;−τ )|. (5.78)

If f (z) is a density of passive tracers at time t , then Pt ,τ f (z) is the tracer density at timet+ τ induced by the flow �.

Page 134: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

126 Chapter 5. Graph Partition Methods and Transport

The above calculations involves constructing a Perron-Frobenius operator for the ac-tion of � on the entire domain M . In the time-dependent setting, we wish to study transportfrom X ⊂ M to a small neighborhood Y of �(X , t ;τ )⊂ M . A global analysis would meanthat X = Y = M and a transfer operator would be constructed for all of M . However, oftenone is interested in the situation where the domain of interest X is “open”, in contrast to aclosed system where X = Y . By open, we mean that trajectories may leave X in a finitetime. Moreover, the subset X may be very small in comparison to M . In such instances,there are great computational savings if the analysis can be carried out using a non-globalPerron-Frobenius operator defined on X rather than M . Therefore, we want to construct anumerical approximation of Frobenius-Perron operator to allow precisely this.

We now describe a numerical approximation of the action of Pt ,τ from a space offunctions supported on X to a space of functions supported on Y . We subdivide the subsetsX and Y into collections of sets {B1, . . . , Bm} and {C1, . . . ,Cn} respectively. We constructa finite-dimensional numerical approximation of the transfer operator Pt ,τ , using a modifi-cation of Ulam’s method [319]:

P(τ )(t)i, j = �(Bi ∩�(Cj , t+ τ ;−τ ))

�(Bi ), (5.79)

where � is a normalized volume measure on M .

Note especially that P is a non-square matrix. Clearly, P(τ )(t) is row-stochastic byits construction. The value P(τ )(t)i, j may be interpreted as the probability that a randomlychosen point in Bi has its image in Cj . We numerically estimate P(τ )(t)i, j by

P(τ )(t)i, j ≈ #{r : zi,r ∈ Bi ,�(zi,r , t ;τ ) ∈ Cj }/Q, (5.80)

where zi,r , r = 1, . . . , Q are uniformly distributed test points in Bi (t) and �(zi,r , t ;τ ) isobtained via a numerical integration. Figure 5.6 shows a schematic diagram of the con-struction of P(τ )(t)i, j . Probabilistically, P(τ )(t)i, j ≈ μ(Bi ∩�(Cj , t + τ ;−τ ))/μ(Bi ). Infact, if μ is absolutely continuous with a positive density that is Lipschitz on the interiorof each Bi , then the error of this estimate goes to zero with decreasing diameter of Bi andCj ; see Lemma 3.6 [128]. Also, the numerical discretization has the useful side-benefitof producing a discretization-induced effective diffusion with magnitude the order of theimage of box diameters (see Lemma 2.2 [124]).

Recall that the coherent pair is a generalization of almost invariant sets in autonomoussystems to nonautonomous systems and has properties as given in 5.2. For the remainderof this section we set P(i , j )= P(τ )(t)i, j , fixing t and τ .

Suppose that we have constructed the m×n row-stochastic transition matrix P(τ )(t)via the Ulam’s method. Let �1 be a diagonal matrix such that

�1(i , i )= π1(i ), i = 1 . . .m, (5.81)

for an arbitrary vector π1. In practices, we conventionally set,

π1(i )= μ(Bi ), i = 1, . . . ,m, (5.82)

and assume that,πi > 0 for all i = 1, . . . ,n. (5.83)

Page 135: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.5. Finite-time coherent sets 127

Figure 5.6. A schematic demonstrating the construction of the transfer operatorapproximation. Notice that X �= Y .

If πi = 0 for some i , then we remove those from the partition on X , i.e., we choose areference probability measure that gives a mass πi > 0 to the set Bi , i = 1, . . . ,m (if somesets have zero reference measure, we remove them from our collection). This referenceprobability measure describes the mass we wish to use to define our discrete version ofcoherence in (5.76). Now define,

π2 = π1 P , (5.84)

to be the image probability vector on Y and,

�2(i , i )= π2(i ). (5.85)

We also assume q > 0 (again, if not, we remove sets Cj with qj = 0). The probabilityvector π2 defines a probability measure ν on Y via,

ν(Y ′)=n∑

j=1

π2( j )�(Y ′ ∩Cj ), (5.86)

for measurable Y ′ ⊂ Y . We may think of the probability measure ν as the discretized imageof μ.

In analogous to what we did in preceding sections, we define a “Laplacian matrix"L by

L=(

�1 −�1 P−PT �1 �2

). (5.87)

Let I1, I2 partition {1, . . . ,m} and J1, J2 partition {1, . . . ,n} and set Xk = ∪i∈Ik Bi andYk = ∪j∈JkCj , k = 1,2. To describe 2-partitions of X and Y we consider vectors x ∈{±1}m , y ∈ {±1}n and define X1 = ⋃i:xi=1 Bi , X2 = ⋃i:xi=−1 Bi ,Y1 = ⋃i:yi=1 Ci ,Y2 =⋃

i:yi=−1 Ci . Thus, the partitions I1, I2 and J1, J2 are described by the parity of x and yrespectively.

Page 136: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

128 Chapter 5. Graph Partition Methods and Transport

Based on the above notation, we can show the following estimate:

α := xT �1 Py =∑

i∈I1, j∈J1

π1(i )P(i , j )+∑

i∈I2, j∈J2

π1(i )P(i , j )

−∑

i∈I1, j∈J2

π1(i )P(i , j )+∑

i∈I2, j∈J1

π1(i )P(i , j )

≈ (μ(X1∩�(Y1, t+ τ ;−τ ))+μ(X2∩�(Y2, t+ τ ;−τ )))

− (μ(X1∩�(Y2, t + τ ;−τ ))+μ(X2∩�(Y1, t+ τ ;−τ )))

= μ(X1)ρμ(X1,Y1)+μ(X2)ρμ(X2,Y2)−μ(X1)ρμ(X1,Y2)−μ(X2)ρμ(X2,Y1).(5.88)

Thus, maximizing α is a natural way to achieve our aim of finding partitions so thatρμ(Xk ,Yk) ≈ 1,k = 1,2. However, for consistency with our previous formulation, wewould like to develop a cost function in term of the generalized Laplacian L. To do this,let p = [x , y]T . We can show that

pT Lp = xT �1x+ yT�2y− (xT�1 Py+ yT PT �1x)

= π1(X)+π2(Y )−2α

= 2(1−α),

(5.89)

where the last line follows from a normalization of the reference measure. Therefore, tofind the most coherent pair, we want to approximate a pair of binary vectors x and y thatminimizes pT Lp.

5.6 Spectral partitioning for the coherent pairSome standard graph-theoritical results for approximating balanced cut via spectral parti-tioning can be usefully applied to derive a cost function via the Rayleigh Quotient. Thefollowing results are modified from [137] and will only be roughly sketched herein. Thefinal result of this section will provide a heuristic mean to find the optimal pair of vectorsthat maximizes α, hence the coherency.

First, we define a block diagonal matrix W to be

W =(�1 00 �2

). (5.90)

Lemma 5.1. Let L and W be defined as above. Let η1 =∑i∈I1

π1(i )+∑i∈J1

π2(i ) and η2 =∑i∈I2

π1(i )+∑i∈J2

π2(i ). Then the generalized partition vector q with elements

q(i )={+√η2/η1, i ∈ [1, . . . ,m] (called I )

−√η1/η2, i ∈ [m+1, . . . ,m+n] (called J )(5.91)

satisfies,qT We = 0, (5.92)

Page 137: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.6. Spectral partitioning for the coherent pair 129

where e is a vector with ei = 1 for all i , and

qT Wq = η1+η2. (5.93)

Proof. The orthogonality between q and e with respect to the W -weight inner product canbe readily seen from the expansion

qT We =√

η2

η1

∑i∈I

W (i , i )−√

η1

η2

∑i∈J

W (i , i ).

It is straightforward to show that

qT Wq =∑

i

W (i , i )q2(i )= η1+η2.

Lemma 5.2. [137] Suppose that∑

i=1...m π1(i ) =∑i=1...n π2(i ) = 1. The generalizedRayleigh quotient is given by

qT Lq

qT Wq= 1

η1η2−α, (5.94)

where α is defined in (5.88)

Proof. First, show that the generalized partition vector q can be rewritten as

q = η1+η2

2√η1η2

p+ η2−η1

2√η1η2

e,

Then, using the fact that, which can be readily derived,

Le = eT L= 0, (5.95)

we can show that

qT Lq = (η1+η2)2

4η1η2pT Lp, (5.96)

where p = [x , y] is the usual sign partition vector as used in (5.89). Finally, after substitut-ing the values of pT Lp and qT Wq , from Lemma 5.1, the desired result is achieved.

Note that the normalization in the above Lemma gives the relation η1+η2 = 2. Thus,if it happens that the measures of X1, X2, Y1 and Y2 are all equal, then (5.94) becomes

qT Lq

qT Wq= 1−α. (5.97)

Therefore, in such case, the pair of X1 and Y1 minimizing the quotient in (5.97) will beour optimal pair, which also maximizes α. In general, the mass equivalency of the partitionis, however, not directly achieved by minimizing the quotient in (5.94) and for this reasonminimizing (5.94) will be called the “relaxation" problem. The relaxation to the optimalgeneralized partition vector can then be achieved by the following theorem.

Page 138: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

130 Chapter 5. Graph Partition Methods and Transport

Theorem 5.8. The problem

minq �=0

qT Lq

qT Wq, subject to qT We = 0, (5.98)

is solved when q is the eigenvector corresponding to the second smallest eigenvalue of thegeneralized eigenvalue problem,

Lz = λWz. (5.99)

Proof. This follows from a standard result from linear algebra, [175].

5.7 The SVD ConnectionIn the previous section, it was shown that a real valued relaxation to the discrete opti-mization problem allows solution by the second eigenvector of the problem Lz = λWz.However, given that P is m×n and L is m+n×m+n, it would be preferable to solve thesame problem based on P instead. Particularly, it will be shown in the following that thecoherent partition can obtained from the second largest singular vector of �

1/21 P�

−1/22 .

This result can also be derived from another heuristic approach based on the similaritytransform of P presented in [289, 138].

To see this, we rewrite the generalized eigenvalue problem Lq = λ2Wq as(�1 −�1 P

−PT �1 �2

)(xy

)= λ2

(�1 00 �2

)(xy

)(5.100)

where,

q =(

xy

), (5.101)

(x , y ∈RN×1) denotes the eigenvector associated with the second smallest eigenvalue λ2.Assume that � is nonsingular, we may rewrite the above equation as

�1/21 x−�

−1/21 (�1 P)y = λ2�

1/21 x

−�−1/22 (PT �1)x+�

1/22 y = λ2�

1/22 y.

(5.102)

Letting,u =�

1/21 x and v =�

1/22 y, (5.103)

and by algebraic manipulation, we obtain

�−1/21 (�1 P)�−1/2

2 v = (1−λ2)u

�−1/22 (PT �1)�−1/2

1 u = (1−λ2)v.(5.104)

Let,P0 =�

1/21 P�

−1/22 , (5.105)

we may write the above equation as

P0v = (1−λ2)u

PT0 u = (1−λ2)v

(5.106)

Page 139: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.8. Example 1: Idealized Stratospheric flow 131

0 5 10 15 20

−2

0

2

t=5 days

0 5 10 15 20

−2

0

2

t=10 days

Figure 5.7. Snapshots of velocity fields at t = 5 and t = 10. The key features ofthe flow is the a time-varying jet core oscillating in a band around y = 0 and three Rossbywaves in each of the regions above and below the jet core. [61]

The above expression is precisely the SVD of the matrix P0. Thus, instead of computingthe eigenvalue of L in Theorem 5.8, we can compute the singular vectors of P0 associatedwith the second largest singular value and then retrieve,

q =(

xy

)from u and v by q =

(�−1/21 u

�−1/22 v

). (5.107)

Furthermore, the largest singular value is σ1(P0) = 1 with the corresponding left singularvector e�1/2

1 and right singular vector �1/22 e.

5.8 Example 1: Idealized Stratospheric flowFollowing [138], we consider the Hamiltonian system,

dx

dt=−∂ψ

∂y,dy

dt= ∂ψ

∂x, (5.108)

where

ψ(x , y, t)= c3y−U0L tanh(y/L)+ A3U0Lsech2(y/L)cos(k1x)

+ A2U0Lsech2(y/L)cos(k2x−σ2t)+ A1U0Lsech2(y/L)cos(k1x−σ1t).(5.109)

Snapshots of the velocity field are shown in Figure 5.8. This quasi-periodic systemrepresents an idealized stratospheric flow in either the northern or southern hemisphere.Rypina et al. [286] show that there is a time-varying jet core oscillating in a band aroundy = 0 and three Rossby waves in each of the regions above and below the jet core. Theparameters studied in [286] are chosen so that the jet core forms a complete transport bar-rier between the two Rossby wave regimes above and below it. We modify some of theparameters to remove the jet core band and allow transport between the two Rossby waveregimes [134]. We expect that the two Rossby wave regimes will form time-dependent co-herent sets because transport between the two regimes is considerably less than the trans-port within regimes. We set the parameters as follows: c2/U0 = 0.205, c3/U0 = 0.700,A3 = 0.2, A2 = 0.4 and A1 = 0.075, with the remaining parameters as stated in Rypina etal. [286].

Page 140: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

132 Chapter 5. Graph Partition Methods and Transport

Our initial time is t = 20 days and our final time is t + τ = 30 days. At our initialtime we set X = S1× [−2.5,2.5] Mm, where S1 is a circle parameterized from 0 to 6.371πMm, and subdivide X into a grid of m = 28200 identical boxes X = {B1, . . . , Bm}. Thischoice of m is sufficiently large to represent the dynamics to a good resolution. We com-pute an approximation of �(X ,20;30) by uniformly distributing Q = 400 sample points ineach grid box and numerically calculating �(zi,r ,20;30) using the standard Runge-Kuttamethod. The choice of Q is made so that over the flow duration, the image of boxes is wellrepresented by the Q sample points per box. These Q×m image points are then coveredby a grid of n = 34332 boxes {C1, . . . ,Cn} of the same size as the Bi , i = 1, . . . ,m. We setY =⋃n

j=1 Cj , covering the approximate image of X . The transition matrix P = P(30)20 is

computed using (5.80).As the flow is area preserving, a natural reference measure μ is Lebesgue measure,

which we normalize so that μ(X) = 1. Thus, μ(Bi ) = pi = 1/m, i = 1, . . . ,m and so(�p)ii = 1/m, i = 1, . . . ,m. The vector q is constructed as q = pP . We compute the second

largest singular value of �1/2p P�

−1/2q and the corresponding left and right singular vectors

and thus determine x and y from Eq. ( 5.107). The top two singular values were computedto be σ1 = 1.0 and σ2 ≈ 0.996. We expect x to determine coherent sets at time t = 20 daysand y to determine coherent sets at time t+τ = 30 days. Figure 5.8(a) and 5.8(b) illustratethe vectors x and y, which provide clear separations into red (positive) and green (mostlynegative) regions.

We apply the thresholding Algorithm from [138] to the vectors x and y to obtainthe pairs (X1,Y1) and (X2,Y2)59 shown in Figures 5.9(a) and 5.9(b). To demonstrate thatY1 ≈ �(X1,20;10), we plot the latter set in Figure 5.9(c). When compared with Figure5.9(b) we see that there is very little leakage from Y1, just a few thin filaments. Similarly,Figures 5.9(d) and 5.9(b) compares Y2 and �(X2,20;10), again showing a small amount ofleakage. This leakage is quantified by computing ρ(X1,Y1)≈ ρ(X2,Y2)≈ 0.98.

We compare our results with the attracting and repelling material lines computed viathe finite-time Lyapunov exponent (FTLE) field [163] with the flow time τ = 10. The ridgesof the FTLE fields are commonly used to identify barriers to transport, as discussed indetail in Chapter 8. Figures 5.8(c) and 5.8(d) present an overlay of forward- and backward-time FTLEs at t = 20 and t = 30, respectively. In this example, there are several FTLEridges in the vicinity of the dominant transport barrier across the middle of the domain,and also several ridges far away from this barrier. The FTLE ridges do not crisply andunambiguously identify the dominant transport barrier shown in Figures 5.8(a) and 5.8(b).

5.9 Example 2: Stratospheric polar vortex as coherentsets

In our second example, we use velocity fields obtained from the ECMWF Interim dataset (http://data.ecmwf.int/data/index.html). We focus on the stratosphere over the southernhemisphere south of 30 degrees latitude. In this region, there are strong persistent trans-port barriers to midlatitude mixing during the austral winter; these barriers give rise tothe Antarctic polar vortex. We will apply the methodology reviewed here to the ECMWFvector fields in two dimensions to resolve the polar vortex as a coherent set.

59When determining X1 and Y1, Algorithm 1 in [138] produced values b∗ ≈ 0.0077 and η(b∗)≈ 0.0005.

Page 141: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.9. Example 2: Stratospheric polar vortex as coherent sets 133

(a)

(b)

(c)

(d)

Figure 5.8. (a) The optimal vector x; (b) the optimal vector y; (c) Backward-time(blue) and Forward-time (red) FTLEs at t = 20 days computed with the flow time τ = 10days; (d) Backward-time (blue) and Forward-time (red) FTLEs at t = 30 days computedwith the flow time τ = 10 days (from [138]).

Page 142: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

134 Chapter 5. Graph Partition Methods and Transport

(a)

(b)

(c)

(d)

Figure 5.9. (a) The sets X1 (red) and X2 (blue); (b) the sets Y1 (red) and Y2(blue); (c) the set �(X1,20;10); (d) the set �(X2,20;10) (from [138]).

Page 143: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.9. Example 2: Stratospheric polar vortex as coherent sets 135

5.9.1 Two dimensions

Our input data consists of two-dimensional velocity fields on a 121× 240 element gridin the longitude and latitude directions, respectively. The ECMWF data provides updatedvelocity fields every 6 hours. The flow is initialised at September 1, 2008 on a 475Kisentropic surface and we follow the flow until September 14. To a good approximationisentropic surfaces are close to invariant over a period about two weeks [188].

We set X = S1× [−90◦,−30◦], where S1 is a circle parameterized from 0◦ to 360◦.The domain X is initially subdivided into the grid boxes Bi , i = 1, . . . ,m, where m = 13471in this example. Based on the hydrostatic balance and the ideal gas law, we set the referencemeasure pi = Pr5/7

i ai for all i = 1, . . . ,m, where Pri is the pressure at the center point of Biand ai is the area of box Bi .

Using Q= 100 sample points zi,r , r = 1, . . . , Q uniformly distributed in each grid boxBi , i = 1, . . . ,m we calculate an approximate image �(X , t ;τ )60 and cover this approximateimage with m = 14395 boxes {C1, . . . ,Cn} to produce the image domain Y . We constructP = P(t )

τ as described earlier using the same Q×m sample points.We compute x and y as described in Eq. (5.107); graphs of these vectors are shown in

Figure 5.10 (Upper left and Upper right). Figure 5.10 (Lower left and Lower right) showsthe result of Algorithm 1 in [138], extracting coherent sets At and At+τ from the vectors xand y. We calculate the coherent ratio ρμ(At , At+τ ) ≈ 0.991, which means that 99.1% ofthe mass in At (September 1, 2008) flows into At+τ (September 14, 2008), demonstratinga very high level of coherence.

To benchmark the methodology as was done in [138], we will compare our resultwith a method commonly used in the atmospheric sciences to delimit the “edge” of thevortex. It has been recognized that during the winter a strong gradient of potential vorticity(PV) in the polar stratosphere is developed due to (1) strong mixing in the mid-latitudes(resulting from the breaking of Rossby waves emerging from the troposphere and breakingin the stratospheric "surf zone" [226]) and (2) weak mixing in the vortex region. Whilepotential vorticity depends only on the instantaneous vector field, potential vorticity is ma-terially conserved for adiabatic, inviscid flow (both of which are good approximations instratospheric flow over timescales of a week or two). Thus, PV may be viewed as a quantityderived from the Lagrangian specification of the flow and is therefore a meaningful com-parator for these nonautonomous experiments who also use Lagrangian information. Weused the method of Sobel et al. [308] to calculate a PV-based estimate of the vortex edge.The result is shown by the green curve in Figure 5.10 (Lower left and Lower right). Notat-ing the area enclosed by the green curve at September 1, 2008 by APV

t and at September14, 2008 by APV

t+τ , we compute ρμ(APVt , APV

t+τ )≈ 0.984; 98.4% of the mass in APVt flows

into APVt+τ over the 13 day period.We see that a transfer operator methodology is clearly consistent with the accepted

potential vorticity approach and in fact identifies a region that experiences slightly greatertransport barriers across its boundary, indicated by the slightly larger coherence ratio:99.1% versus 98.4%.

60We use the standard Runge-Kutta method with step size of 3/4 hours. Linear interpolation is used toevaluate the velocity vector of a tracer lying between the data grid points in the longitude-latitude coordi-nates. In the temporal direction the data is independently affinely interpolated.

Page 144: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

136 Chapter 5. Graph Partition Methods and Transport

Figure 5.10. [Upper left]: Graph of x (September 1, 2008). [Upper right]: Graph of y(September 14, 2008). [Lower left]: The red set represents the coherent set At (September 1, 2008) obtainedfrom Algorithm 1 in [138]. The green curve illustrates the vortex edge as estimated using PV. [Lower right]:As for [Lower left] at September 14, 2008.

5.10 Community MethodsWe close this chapter by relating two seemingly disparate fields, being community findingto our interest here of phase space partitioning. A topic of current interest in the area ofnetwork sciences has been the idea of communities and their detection. See the review arti-cle, [273]. The applications of community finding range from social science, to biologicalscience, to chemical interactions to name a few. The betweeness methods in [251] and thelocal method in [9] are examples of numerous methods as reviewed in [75, 251] that canbe successfully used to uncover various kinds of community structures.

A community could be loosely described as a collection of vertices within a graphthat are densely connected amongst themselves while being loosely connected to the restof the graph [323, 119, 277, 10]. This description, however, is somewhat vague and opento interpretation. This leads to the possibility that different techniques for detecting thesecommunities may lead to slightly different yet equally interesting results.

The point of our discussion of community finding methods is that usually a searchof the network structure (a graph is often described as a network in social sciences). Manytechniques have been proposed to detect community structure inside of a network. Cluster-ing according to community structure in network science has many parallels to the problem

Page 145: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.10. Community Methods 137

of interest here of decomposing a dynamical system into almost invariant sets when we re-alize that an Ulam-Galerkin matrix describes a graph structure.

The breadth of methods might be described as either agglomerative or as divisive.Rather than provide a survey of community finding [273], we will only touch on thosemethods which are relevant to our theme in this book which regards partitioning the transferoperators so as to associate regions of phase space descriptive of transport in phase space.While most community methods were not designed explicitly to produce such a partition,some effectively do so. Only a true optimization cut method as discussed in previoussections addresses this directly, but these other methods can give comparable and sensiblepartitions.

5.10.1 Betweeness

Centrality measures serve a sensible starting point to partitioning a network. Betweenesscentrality has become a most popular partitioning tool, [251]. Betweenness centrality isdefined as the number of shortest (geodesic) paths passing through the edge.

Definition 5.9. Betweenness Centrality Assume a (directed) graph G = (E , V ), withvertices V = {v1,v2, ...,vn} and edges E = {e1,e2, ...,em}, and each edge can be writtenek = (vi ,vj ),k = 1, ...,m, i , j = 1, ...,n. A shortest path pi, j between any pair of vertices, viand vj exists and is computable. Let P = {pi, j }ni, j=1 be the set of all shortest paths betweeneach pair of vertices. Therefore the betweenness centrality of an edge ek is the number ofthose shortest paths from P that pass through ek .

The shortest path between a pair of vertices vi and vj can be computed by the BreadthFirst Search algorithm, [22]. Betweenness centrality can be used to partition a network bythe following process, [252], Given the graph, G = (E , V ), remove from the edge set E the(an) edge with maximal betweenness centrality. Say it is, ek . Thus,

G = (E , V )→ G′ = (E ′, V ), where E ′ = E− ek . (5.110)

This yields a new graph G′. This process is repeated until an initially irreducible graph isrendered reducible. The idea behind this algorithm closely mirrors the notion of congestionand bottle necks.

Example 5.7. Consider the synthetic graphs illustrated in Fig. 5.11, [10]. Both of thesesynthetic examples in the figure illustrate the notion that betweenness serves as a measureof congestion in an unweighted graph. Thus in this sense as a measurement of traffic flowbetween the components so identified we have a network analogous concept to transport ina dynamical system.

Example 5.8. Consider the graph illustrated in Fig. 5.12 showing the social connectionsamongst of group of children in a karate club, called Zachary’s Karate Club, [252]. Thisnetwork has become a paradigm for testing graph partition algorithms, and also partitionswell by the edge betweenness methods. See also the Amazon book example in Fig. 5.13,that naturally finds a politically left leaning and right leaning partition.

In terms of the betweenness centrality connection to the graph representation of dy-namical systems, since two almost invariant sets will communicate through a small number

Page 146: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

138 Chapter 5. Graph Partition Methods and Transport

Figure 5.11. Graphs with idealized betweenness centrality, [10]. (Left) Idealgraph 1: two complete subgraphs of size 15 bridged by a common vertex. Initially thisgraph’s highest betweenness centrality is carried by the edges (v15,v16), and likewise tiedwith (v16,v17) is this bridge serves as a point of congestion/bottleneck in the graph. Re-moval of these renders a graph with two large components that can no longer communicate.(Right) Ideal graph 3: three complete subgraphs joined together by multiple edges with sin-gletons attached bridged by a single vertex to another group of similar, but not identical,structure. Initially, maximal betweenness is carried by edges (v33,v45) tied with (v45,v54).Once these edges are removed the resulting graph in this example has further hierarchicalstructure.

of connecting edges, the betweeness of those edges will be very high. Thus cutting edgeswith the highest betweeness will render the graph partitioned into invariant sets if they wereinitially almost invariant. The issue of both computing and presenting in an understandableformat the huge data sets of a large network is a question of data mining.

The edge betweeness method [250], based on computing shortest paths, is a good ro-bust candidate algorithm for detecting communities, but its run time is of the order O(m2n)for a graph with m edges and n vertices, or O(n3) for a sparse graph(m ∼ n). Such expen-sive demands on computational resources make it infeasible for a large graph. Furthermore,a d-dimensional dynamical system with an hd -element box (for box of order length h on aside) or triangle covering of the invariant set makes therefore an O(h3d ) algorithm, whichis quite expensive for a fine grid, h # 1, in high dimension-d . In the next section we

Page 147: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.10. Community Methods 139

Figure 5.12. Actual breakdown of the karate club from Newman. [251]

will describe a method based on a modularity measure of a graph partition developed byNewmann [248]. This method is far less expensive than the betweeness method and givesuseful results in many applications.

5.10.2 Modularity Method

First, we discuss the concept of the modularity as a measure of quality of a proposed com-munity structure found by any graph partition algorithm. The modularity is a cost functionassociated with a partitioning of a given graph, G(E , V ),

Q : CG → R, (5.111)

where CG is the collection of all subpartitions C of a given graph G. Given a graph G anda (test) partition, C ∈PG , C = {Ck}, each C is a set of subsets Ck , and Ck is a collection ofvertices of G such that

⋃Ck = G. We will refer to Ck as “community"-k. Newman [248]

proposed that a good division of a graph network into communities could be justified onlyif the number of the within-communities edges is significantly larger than the “expectednumber" of such edges when edges are randomly placed into a network. This requires thatwe must formally define what we mean by the expected number of edges for each vertex.

For a given community partition C , let us denote by ci the community that the vertexi belongs to and Pij be an element of the transition matrix P ∈ �m×m of the network:

Q = 1

m

∑i j

(Pij −

kouti kin

j

m

)δ(ci ,cj ), (5.112)

where and the delta-function δ(ci ,cj ) = 1 if ci = cj and 0 otherwise, m =∑i j

Pi j , kouti =

Page 148: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

140 Chapter 5. Graph Partition Methods and Transport

Figure 5.13. The network of copurchased books on American politics as shownin [10]. Here a link is drawn between two vertices if those books were purchased togetherfrom a major online retailer. V. Krebs, Political Patterns on the WWWÑDivided We Standhomepage: http://www.orgnet.com/divided2.html

∑j Pi j and kin

j =∑

i Pi j . Intuitively, this is the difference between the actual number ofthe within-communities weighted edges and the expected number of such weighted edges.In the above definition, to see how the second term (the expected number of edges) comesfrom is to consider a weighted random graph model in which we give the vertex i the in-and out- degrees kin

i and kouti , respectively, so the probability to find the weighted i → j

Page 149: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.10. Community Methods 141

edge is (kouti kin

j )/m. When a strong community partition C exists, Q will be large. Put itin another way, Q measures how randomly the actual edges are distributed. If randomlydistributed, Q will be small, otherwise it will be large.

For a particular case of undirected, non-weighted graph, i.e. Pij = Aij , where theAij is an entry of the adjacency matrix given by

Aij ={

1 if (i , j ) ∈ E

0 otherwise(5.113)

The modularity Q can be expressed by

Q = 1

2m

∑i j

(Aij − kikj

2m

)δ(ci ,cj ). (5.114)

In this simplified case, the edges are counted twice, hence using 2m in the denominator,and there is no distinction between the in- and out- degrees, i.e, the probability to find theedge connecting i and j is

ki kj2m . Since δ(ci ,cj )=∑k δ(ci ,k)δ(cj ,k), the expression (5.114)

can be rewritten as

Q = 1

2m

∑i j

(Aij − Rij

)∑k

δ(ci ,k)δ(cj ,k)

=∑

k

(1

2m

∑i j

Ai j δ(ci ,k)δ(cj ,k)− 1

2m

∑i

kiδ(ci ,k)1

2m

∑j

kjδ(cj ,k)

).

(5.115)

Now we define ek to be the fraction of edges that connect vertices within the communityPk ,

ek = 1

2m

∑i j

Ai j δ(ci ,k)δ(cj ,k), (5.116)

and ak to be the fraction of ends of edges that are attached to vertices in community Pk ,

ak = 1

2m

∑i

kiδ(ci ,k). (5.117)

Therefore, an alternative expression of Eq. (5.114) is

Q =∑

k

(ek −a2

k

). (5.118)

Example 5.9. Consider a graph network consisting of K identical communities withoutany inter-communities edges. Let m be the total number of edges within the graph network.If we choose a test partition to be a collection of these K communities, it is simple to seethat ek = ak = 1

K . So, we have Q = 1− 1K .

Example 5.10. Consider again the graph network in the preceding example but allow anedge to connect between community i and i +1. Thus we now have 2m+ K −1 edges in

Page 150: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

142 Chapter 5. Graph Partition Methods and Transport

the network. Again, we choose the test partition to be the K communities. It follows that

Q ≤ K (1

K−( 2m

K +1

2m+ K −1

))2

= 1− 1

K

( 2mK +1

2mK +1− 1

K

)2

< 1− 1

K.

(5.119)

Recall that a good partition of a graph network into communities could be foundby optimizing the modularity function Q. The true optimization is, however, very costly toimplement in practice for very large networks. Clauset, Newman, and Moore [63] proposedan approximate algorithm based on greedy optimization. Suppose that we have a graph withn vertices and m edges. This algorithm starts with each vertex being the only member ofone of n communities. At each step, the change in Q(P) is computed after joining a pair ofcommunities together and then choosing the pair that gives the greatest increase or smallestdecrease in Q(P). The algorithm stops after n−1 such joins in which a single community isleft. This algorithm therefore runs in time O((m+n)n), or O(n2) on a sparse graph (n ∼m).Note that using more sophisticated data structure as introduced in [63] can reduce a run timeto O(n log2 n). Furthermore, the algorithm builds a dendrogram which represents nestedhierarchy of all possible community partition of the network. Then we select the partitioncorresponding to local peaks of the objective function Q for our satisfactory communitydivision. Figure 5.14 shows an example of the dendrogram of the graph network shownin Figure 5.4(a). We can see that the peak in the modularity corresponds to the correctidentification of the community structure and the modularity is dropped from the peakwhen two of the three communities are joined together at the final step. We note that thequality of the partition can be determined by the value of Q. A typical value of Q for agood division in many real-world network is around 0.2− 0.3. If the partition obtainedfrom an optimization scheme shows a considerably smaller value of Q, say Q < .1, weassume that there is no significant partitions of the given network into communities.

5.10.3 Matrix Form of Modularity Method

Another approach to optimize the Q functional reveals a spectral method. An idea ofoptimization schemes based on spectral methods has been traditionally used to determine adivision of a network so that the minimum cut is achieved. In this sense, both in spirit andin some details, the community finding methods from the network literature has a great dealof coincidence with the spectral methods already highlighted here from the almost invariantset work in Eq. (5.57) and the coherent set methods from Sec. 5.6 which also at the roothave a notion of balanced cut and there is also a notion of reversible Markov chains.

To minimum cut methods in the graph theory literature, the eigenvalue spectrum ofthe Laplacian matrix is utilized to divide the network. A review of this method as applied tothe minimum cut problem may be found in [249]. Min-cut problems have become a classicproblem in graph theory [197, 148, 193, 311] and specifically there is a large literature byspectral methods, [186, 187, 160] but see especially the text on spectral graph theory, [62].In particular, the second smallest eigenvector of the graph Laplacian and correspondingeigenvector called the Fiedler vector [117], and this gives rise to a classic bisection method

Page 151: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.10. Community Methods 143

Figure 5.14. Plot of dendrogram for a graph network in Figure 5.4(a). As we cansee the peak in the modularity (dashed line) occurs where the network is divided into threecommunities.

of the graph into only two communities based on the sign of the corresponding vector entry,[304] and these graph methods are known as algebraic connectivity.

For our purpose, we first consider the partition of a network into two communities.Let a vector s of n elements, where n is the number of vertices, be defined as

si ={

1 if i is in community 1−1 if i is in community 2

(5.120)

Then we can rewrite Eq. (5.112) as

Q = 1

2msT Bs, (5.121)

where B is called the modularity matrix defined by

Bij = Pij −kout

i kinj

m. (5.122)

Therefore, we would like to determine the vector s that maximizes Q under the constraintthat sT s = n. In general, the modularity matrix B for a directed network is not symmetricand so the spectral method cannot be directly applied. Instead, we consider the modularityQ of the following form:

Q = 1

4msT (B+ BT )s. (5.123)

Page 152: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

144 Chapter 5. Graph Partition Methods and Transport

Note that the matrix B+ BT is not the same as the modularity matrix of the same networkbut ignores the directions. Also note the similarity in spirit to the reversible Markov chaindiscussed in Eq. (5.57). The idea of the spectral scheme is to write s as a linear combinationof the eigenvectors vi of B+ BT . Then we can express Q in term of the eigenvectors viand their corresponding eigenvalues λi as

Q =∑

i

λi (vTi s)2. (5.124)

Then we can choose s to be as close to parallel with the leading eigenvector, say v1, of thelargest eigenvalue to maximize Q, i.e., maximizing the quantity

|vT1 s| = |

∑i

siv(i)1 | ≤ |

∑i

v(i)1 |. (5.125)

Under the constraint si =±1, this can be achieved by choosing the sign of si to be the sameas that of v(i)

1 .We have only determined the optimal partition of the network into two communities. If the network comprises of more

than two communities, we have to further divide these two communities. Thus we repeatthe bisection process to each community of the network using the spectral partitioning asdescribed above and continue dividing the result further until the modularity of the networkcannot be increased. In the course of this procedure, we have to compute the change inmodularity denoted by �Q to decide whether or not the community k can be subdividedfurther. We define a vector s as before except that only vertices within the community gare taken into account and note that �Q has to be computed based on the entire network.Then we have

�Q = 1

2m

[ ∑i, j∈cg

(Bij + Bji)si sj +1

2−∑

i, j∈cg

(Bij + Bji)

]

= 1

4m

∑i, j∈cg

[(Bij + Bji )− δi j

∑k∈cg

(Bik+ Bki )

]si sj

= 1

4msT (B(g)+ B(g)T

)s,

(5.126)

where we set

B(k)i j = Bij − δi j

∑k∈g

Bik . (5.127)

To obtain the first line in the above derivation we subtract the modularity Q after subdi-vision from Q before subdivision and notice that the term concerning vertices outside thecommunity g can be canceled out. Then we use the fact that s2

i = 1 to derive the secondline. Now, the spectral bisection scheme can be applied to determine the partition of thecommunity g that maximizes �Q of the entire network. If �Q happens to be negative orsmaller than some threshold value of �Q, it means that the community k should not besubdivided further.

Page 153: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.11. Open Systems 145

5.11 Open SystemsIn the previous sections of this chapter, the implication has been that the dynamical systemis closed relative to the window of interest. It is natural that the window of interest may besmaller than the full phase space and it may not be an invariant set. While the discussionof coherent sets in Sec. 5.5 also includes notions and application to open systems, theemphasis is designed for “tracking" moving objects by moving the “windows" along withthem. However, discussion of transport properties confined within a fixed window, we takethe perspective of open system, which is the main subject of this section.

5.11.1 Open Systems and Holes

When the phase space of a dynamical system is covered, say by rectangles, or a tesselationof triangles, as the support for characteristic functions as the basis set, then representing theFrobenius-Perron operator by Ulam-Galerkin’s method requires that we decide how muchof the phase space to inspect; we must choose a window of interest W . We can take awindow W to be an open set relative to which we will consider the orbits of a dynamicalsystem as they may enter or leave W , but generally it is simpler to discuss W that aretopological rectangles. While ideally, we would cover the entire phase space, that is notalways possible for various reasons, either engineering, technical or mathematical.

See Fig. 4.1, where some of the boxes (triangles) will partly or completely map out-side the region of focus W . In Fig. 4.3, it is apparent that to study the Gulf of Mexico,that data availability issues may be quite separate from the dynamical issue of invariance.The window focusing on the Gulf of Mexico naturally results in an open system accord-ing to the action of this dynamical system. The general the issue is if W maps into itself,T (W )⊆W then the system is closed.

For example, each of the maps shown in Fig. 5.1 and Fig. 5.2 are closed relative toW = [0,2]= A∪ B . But only the map shown in Fig. 5.1 is closed relative to the windowchoice, W = [0,1]= A, whereas the map shown in Fig. 5.2 is open. That is,

Definition 5.10. If T (W )− T (W )∩W = ∅, then the dynamical system is dynamicallyclosed relative to a window (set) W of interest. Otherwise the dynamical system is calleddynamically open relative to a window (set) of interest W .

Dynamically closed relative to W is synonymous then with the statement that Wis an “invariant set" but the word “closed" emphasizes the possibility of open or leakingwindows W to a dynamical systems. Note that the use of the words dynamically open andclosed should not to be confused with open and closed as used in point set topology.

Example 5.11. Compare the smaller window W in Figs. 5.15, with those shown in Figs. 5.1-5.2. Here we choose a smaller window of inspection, W = [0,1] as compared to previouslyW = [0,2]. Otherwise the maps shown in Figs. 5.15 and 5.2 are identical in [0,1]; it isirrelevant if they are different maps or not if we are inspecting a subregion and consideringescape and invariance in W = [0,1]. In Fig. 5.15 we see that points that have orbits that arein the yellow strips, the pre-images, escape W .

Those points that do not stay in W may be describe as if they disappear into a hole.Such systems have been studied in many papers, including [267, 15, 216, 139, 28, 29] toname a few. It is not surprising that since the pre-image of the middle yellow strip shown

Page 154: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

146 Chapter 5. Graph Partition Methods and Transport

Figure 5.15. In contrast to the map shown in Fig. 5.2 where the window of interestwas chosen to be W = [0,2], here the window of interest is chosen to be the smaller W =[0,1] and otherwise the maps are the same in this smaller window. Relative to W, thedynamical system behaves as if there is a hole. Dynamically, this is easy to see due to thefact that any orbit that falls in the center yellow colored interval, or any pre-image, willleave W. As far as W is concerned, these points have simply fallen into a hole outside thesystem, even if a larger windowed description in Fig. 5.2 shows further behavior of thesepoints not discussed; relative to this W, they fell into the hole. Correspondingly a coarsegraphical representation of the transfer operator is the directed graph shown below themap. Some orbits are invariant to W, but some leave W transitioning to what is called ahole, labelled H .

is a Cantor removal process, then those points that do not eventually escape form a Cantorset, and the dynamics restricted to this set is called an unstable chaotic saddle. More willbe said on this topic regarding escape and unstable chaotic saddles in Chapter 7 and inparticular in Sec. 7.4.3. See in particular Fig. 7.29 that shows an unstable chaotic saddle ina Henon map example.

In the subsequent of this section we will study the spectral properties of the corre-sponding transition matrix in systems such as these with holes. An extension discussion ofspectral properties of open systems can be found in [139, 271].

5.11.2 The Transition Matrix of an Open System

We will now mirror the analysis of Eqs. (5.11)-(5.19) in Sec. 5.19, particularly regardinga perturbation of a form P �→ P+ E as Eq. (5.18) which we previously specialized for a

Page 155: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.11. Open Systems 147

closed system of the form Eq. (5.19) in Corollary 5.4. Now we will adjust the analysis foropen systems. Specifically we will perturb a state labeled 1 to describe the leaking windowof interest W and 2 to describe the hole into which states evolve, as described by the leakytent map dynamical system shown in Fig. 5.15.

In Sec. 5.3 we discussed that signs of the second eigenvector give some useful par-tition information when the first eigenvector is uniform. The analysis assumed that thematrix is almost block diagonal of the form Eq. (5.11), and it is a closed system of the formEq. (5.19).

Here akin to Eq. (5.19), assume that an open system with a hole can be written as ablock upper diagonal system nearby a block diagonal system, of the form,

P =(

P1,1 00 P2,2

)�→ P =

(P1,1+ εE1,1 εE1,2

0 P2,2+ εE2,2

). (5.128)

Recalling Lemma 5.1, if the unperturbed block diagonal form has block matrices whichare algebraically and geometrically simple, then the dominant eigenvalue λ = 1 has amultiplicity-2 and partitioned eigenvectors as described by Eq. (5.13),

u = (vT1 |0T )T and w = (0T |vT

2 )T . (5.129)

As such we recall Eq. (5.14) that in the span of these two eigenvectors is the partitionedsigned vector,

v = (vT1 |− vT

2 )T , (5.130)

such that a permutation on sorting the signs of a random permutation of P recovers thepartition, analogous to Proposition 5.2.

We now describe how Proposition 5.5 and Corollary 5.6 that a sort on the signs ofthe second eigenvector of the perturbed matrix, Eq. (5.32), also reveals a useful partitiongeneralizes to open systems.

Proposition 5.11. If in Eq. (5.128), P is stochastic, E1,2 ≥ 0 and the perturbations E1,1,E1,2, and E2,2 are such P is also stochastic, then P has a dominant eigenvector,

v = (0Tn1×1|v2

T )T . (5.131)

where 0n1×1 is a vector of zero’s of the same size as v1 and v2 is a vector of the same sizeand nearby v2.

This simply states that the dominant behavior describes the measure leaking out ofthe partition elements labelled 1 into those labelled 2.

Corollary 5.12. If P2,2 = [1]1×1 is the 1×1 identity matrix, then v2 = v2 = [1]1×1.

These statements can be compared to the leaky tent map dynamical system shownin Fig. 5.15, choosing a single element for H as per Corollary 5.12, P2,2 = [1]1×1. Thedominant eigenvector labeling those states labeled 1 covering “W" has 0’s in the dominanteigenvector and the state labeled 2 describing “H" has a 1 in the eigenvector. This does notmean that all orbits fall into the hole, but only almost all orbits. This agrees with that weknow there is an invariant Cantor set in W which we have already described as the unstablechaotic saddle.

Page 156: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

148 Chapter 5. Graph Partition Methods and Transport

Note that to maintain that P1,1→ P1,1+ εE1,1 may describe a region W which leaksinto a hole, but as P1,1 is stochastic, then εE1,1 should have negative entries so that P1,1→P1,1+ εE1,1 will have a dominant eigenvalue less than 1 but the full system covers W andthe hole H should still be stochastic. Therefore, to meet the hypothesis of Proposition 5.11,some entries of E1,1, E1,2, and E2,2 will likely be negative.

Other interesting analysis regarding systems with hole come come in the theory ofMarkov processes [309], and practice in statistical physics for diffusive systems and ran-dom walkers [321], most notably in the first passage time and the mean first passage timeproblems. These address the question, “How long does it take a random walker to reach agiven target?" The mean first passage time Mi, j from the state i to state j is the expectednumber of steps to arrive at destination j when the origin is i , and the derivations involvedeliberately adding an extra hole vertex as a trap. Some related analysis for our dynami-cal systems interest here can be found in Sec. 7.4.2. All of the discussion in this sectionregarding holes and their Markov models should be compared to the traditional Markovprocess theory pertaining to absorbing sets, [185] and also hitting times and absorptionprobabilities.

5.11.3 Connecting Escape rate and the hole

Based on Eq. (5.128) it is certain that every transient states in P1,1 will be eventually ab-sorbed to the absorbing states and it is important to estimate the rate at which the massescapes from the set of transient states to the set of absorbing states. With out loss ofgenerality, we can “lump" all the absorbing states and assume that we have the absorbingMarkov chain with transition matrix

P =(

Q R0′ 1

), (5.132)

where Q is (N −1)× (N−1) and R,0 are both (N −1)×1. Then

Pn =(

Qn (I + Q+ . . .Qn−1)R0′ 1

). (5.133)

We denote the index set of transient states, corresponding to the block Q in Eq. (5.132),by A. We will also assume that Q is irreducible. The probability that the transient statesremain in the set of transient states after n steps, called M (n), is then given by

M (n) :=∑

i∈A, j∈A

Qn(i , j )π(i ) (5.134)

for a given finite reference measure π and Qn(i , j ) is the (i , j ) component of the matrixQn . Alternatively, we can write M (n) = π ′Qn1, where 1 is a N − 1× 1 vector with allelements equals to unity. We assume that Q is irreducible and substochastic. By the Perron-Frobenius theorem, we may arrange the eigenvalues of Q in the following order, 1 > λ1 >

|λ2| > |λ3| ≥ |λ4| ≥ · · · ≥ |λN−1| and their corresponding normalized (right) eigenvectorsare denoted by vj for j = 1, . . . , N − 1, (i.e. |vj |1 = 1). We assume the strict inequalitybetween |λ2| and |λ3| to simplify the below analysis. Then it can be readily shown, bywriting 1 as a linear combination of eigenvectors, that

Qn1∝ λn1v1+O(

∣∣λ2

λ1

∣∣n). (5.135)

Page 157: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.12. Relative Measure and Finite Time Relative Coherence 149

In the limit of n→∞, the second term in the above equation vanishes and we can showthat

M (n) = π ′Qn1

∝ λn1

∑j∈A

π( j )v1( j )≤ λn1 max[π( j )]

∑j∈A

v1( j )

≤ λn1 |v1|1 = λn

1.

(5.136)

In other word, M (n) ∝ λn1 in the limit and this is independent of the reference probability

measure π . It is clear that M (n) is decreasing as n goes to infinity. The question is how fastit is decreasing as n increasing. This motivates a notion of the escape rate w.r.t a referenceprobability measureπ , which is defined by

E(A) :=− limn→∞

1

nlog(M (n)). (5.137)

Therefore, the escape rate is E(A)=−logλ. The larger the escape rate, the faster the lossof the probability mass from the set of transient states A to the absorbing state.

Remark that the Ulam’s matrix is usually not in the form of Eq. (5.132). However,when consider a set of some states as a hole, we could eliminate the rows and columnsassociated those states labeled as holes. This will effectively give rise the Q matrix as inEq. (5.132). More formally, the modified Ulam’s matrix for an open system T with a holewould be

P := m(Bi ∩T−1(Bj )∩H )

m(Bi ), (5.138)

where H is the collection of Bi constituting the hole. The matrix P , as compared toEq. (4.6) will be sub-stochastic and it is exactly the same as Q except for the zero rowsand columns.

As mention earlier, this idea can be made analogous to the escape rate of the massinto the holes. In the case of general maps that does not admit a Markov partition, sim-ilar relation between the escape rate and the leading eigenvalue can also be obtained butthe discussion requires some technicalities beyond the scope of this book, so we refer to[92, 139, 267] for more details and reviews of this topic. Also, if the rate of escape is veryslow, it could be of more relevance to consider the so-called absolutely continuous condi-tional invariant measure (ACCIM), see [267, 92], which is analogous to the quasi-stationarydistribution in absorbing Markov chains, see [224].

5.12 Relative Measure and Finite Time RelativeCoherence

The Definition 5.76 of finite time coherent pairs in Sec. 5.5 from [138] leads to a notionof optimal bi-partition as computed by the spectral methods discussed in Sec. 5.6. See forexample the optimal partition computed and displayed in Fig. 5.8 for the Rossby wave sys-tem, Example 5.8. However, it may be relevant to consider finer scaled coherent structuresand even a hierarchy of coherent structures.

With the motivation of fine scaled coherent structures, we will recast as in [39] astraight forward adjustment of definitions of the original measure based notion of finite

Page 158: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

150 Chapter 5. Graph Partition Methods and Transport

time coherent structures, using relative measure instead of the global measure. Consider ameasure space (,A,μ) on (the window on) , the phase space for which the finite timecoherent sets are defined in Definition 5.2; notice that the coherence function ρμ(At , At+τ )in Eq. (5.76) explicitly includes the measure μ .

A standard definition of relative measure will be used in the following.

Definition 5.13. Given a measure space, (,A,μ), and a μ-measurable subset ω⊂, thenthe normalized inherited relative measure is defined on a measure space (ω,B,μω) by,

μω(B)= μ(B ∩ω)

μ(ω), for any μ−measurable set B∩ω ⊂A, (5.139)

where B is the μω-measurable sets.

If is an invariant set then the windowing is considered a closed system, but other-wise we are considering an open system. In any case, it is convenient for the following toassume that μ : →�+ is a probability measure and as such μ()= 1.

With relative measure immediately follows a notion of finite time relative coher-ence, by slightly adjusting the definition Definition 5.76 to use μω in the restricted windowω instead of μ on the larger window .

Definition 5.3. We will call At , At+τ a (ρ0, t ,τ )−μω-relative coherent pair if

1.

ρμω (At , At+τ ) := μω(At ∩�(At+τ , t+ τ ;−τ ))

μω(At )≥ ρ0 (5.140)

2. μω(At )= μω(At+τ )

3. At and At+τ are “robust" to small perturbations.

We take the phrase “robust" to mean simply that the observed coherent set variescontinuous with respect to perturbations.

Relative measure suggests a straight forward refinement of the spectral based bal-anced cut method from Sec. 5.6 to optimize coherence, Definition 5.76. Suppose that wehave already optimized coherence to produce a bipartition, X∪Y =. Then we may repeatthe process in each subset. Let ω = X and then use the same spectral method from Sec. 5.6to optimize coherence ρμω := ρμX . This produces what we shall label X1∪Y1 = X . Like-wise, repeat; let ω = Y to produce a ρμω := ρμY -coherent partition, X2 ∪Y2 = Y . Thusthe bipartition X ∪Y produces a partition in four parts, X1 ∪Y1 ∪ X2 ∪Y2. Then repeat,choosing successively ρμω := ρμX1

, ρμY1, ρμX2

, and then ρμY2, to produce an eight-part

partition, etc, in a deepening tree structured partitioning of the phase space into relativelycoherent sets. See Fig. 5.12 for a depiction of the repetition aspect of this algorithm. Anatural stopping criterion for the algorithm is to terminate any given branch of the treewhen the the spectrally computed optimal relative measure on the given branch produces apartition where the objective function ρμω which even when optimized is nonetheless notlarge. This indicates that the “test" subpartition is not relatively coherent, in the sense ofEq. (5.3).

Example 5.12. Consider again the Rossby wave system already discussed in Example 5.8,with the finite time coherent partition shown as a bipartition X , Y colored red and blue

Page 159: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

5.12. Relative Measure and Finite Time Relative Coherence 151

Figure 5.16. Algorithm Tree toward finite time relatively coherent sets. UsingDefinition 5.76 of finite time coherent sets from [138] and optimizing leads to a partitionX ∪Y = . Further refinements using relative measure on successively smaller elementsof the partition leads to finite time relatively coherent sets according to Definition 5.3 from[39].

in Fig. 5.8. In Fig. 5.12, we illustrate from [39] several subdivisions toward finite timerelatively coherent sets as described in Fig. 5.12.

Again, we write in terms of the Hamiltonian,

dx/dt =−∂�/∂y

dy/dt = ∂�/∂x , (5.141)

where

�(x , y, t)= c3y−U0Ltanh(y/L) (5.142)

+ A3U0Lsech2(y/L)cos(k1x)

+ A2U0Lsech2(y/L)cos(k2x−σ2t)

+ A1U0Lsech2(y/L)cos(k1x−σ1t)

This is a quasiperiodic system that represents an idealized zonal stratospheric flow

Page 160: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

152 Chapter 5. Graph Partition Methods and Transport

Figure 5.17. Compare to Fig. 5.8. Tree of finite time relatively coherent setsaccording to algorithm described in Fig. 5.12.

[138]. There are two known Rossby wave regimes in this system. Let U0 = 63.66,c2 =0.205U0,c3= 0.7U0, A3= 0.2, A2= 0.4, A1= 0.075 and the other parameters.

Building a 32640 by 39694 matrix we choose 20,000,000 points in the domain X =[0, 6.371π ∗ 106]× [−2.5 ∗ 106, 2.5 ∗ 106] of the flow and uses 32,640 triangles as thepartition {Bi}32640

i=1 for the initial status points and 39,694 triangles as the partition {Cj }39694j=1

for the final status of the points. Note that this system is “open" relative to the domain Xchosen, though it is a area-preserving flow. The two coherent pairs are colored blue and redwhich are defined as (X1,Y1) and (X2,Y2) in the first level of Figure 5.12. Again, we nowbuild the relative measures and tree of relatively coherent pairs. By applying the method aswe have done with the previous two examples, we develop four and eight different coherentstructures for the second level and the third level, respectively.

Page 161: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 6

The Topological DynamicsPerspective of SymbolDynamics

Why include a chapter on symbol dynamics in a book about measurable dynamics? Af-ter all, symbol dynamics can rightly be described as part of topological dynamics. Thatis the study of a dynamical system in a topological space (the set of open sets), mean-ing the action of the map absent considering any measure structure. So how does thisrelate to measurable dynamics? The answer is in the shared tools of analysis and hope-fully broader perspective and common methods. Specifically, we have already describedthe Frobenius-Perron transfer operator is well approximated stochastic matrices when it iscompact (Recall Secs. 4.2-4.3), and there is a corresponding description by weighted di-rected graphs. Likewise, symbol dynamics as we shall describe here is well understood asapproximated by an adjacency matrix (a matrix of 0’s and 1’s rather than weights summingto one as in the stochastic case), and the corresponding description by an unweighted di-rected graph. In fact, for a given dynamical system, the same directed graph can be used,if we ignore the weights. Furthermore, the discussion regarding exactness of the represen-tation by finite size directed graphs which occurs when the dynamical system is Markovand the Markov partition is used, is identical between both the measurable and topologicaldynamical systems analysis.

6.1 SymbolizationSome of the most basic questions in dynamical systems theory do not require us to evenconsider measure structure. These are questions of loss of precision regarding state andinformation rather than description of how much and how big. Symbolic dynamics is aframework by which a number of theoretical and also practical global descriptions of achaotic dynamical system may be analyzed, and often quite rigorously. Symbolic dynamicsas defined in this chapter and the related topic of lobe dynamics defined in the next chaptercan help lead to a better understanding of transport mechanisms. Then considering thecorresponding measure structure can lead to escape rates, loss of correlation and a partitionspecific description of the Frobenius-Perron operator.

We take a naive perspective here for sake of narration and inclusion of the topologicalissues of lobe dynamics relevant to the the measurable dynamics discussion of the rest ofthis book. For more in depth discussion of these topological issues, we refer to the excellent

153

Page 162: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

154 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

texts [281, 327].Formally, the topic of symbolic dynamics is descriptive of topological dynamics

whereas a great deal of the theme of this book regards measurable dynamics. Withoutincluding measure structure, only the evolution of open sets and points therein are consid-ered; topological dynamics does not worry about the size/weight of sets.

Symbolic dynamics may seem at first blush to be an abstract and even complicatedtopic without relevance to applied dynamical systems and certainly not useful to the ex-perimental scientist. To the contrary, symbolic dynamics is in many ways a simplifyingdescription of chaotic dynamical systems, that allows us to lay bare many of the fundamen-tal issues of otherwise complicated behavior. These include,

• the complementary roles of information evolution (See further in Chapter 9), mea-surement precision, entropy and state.

• the role of partitions.

• a better understanding of discrete time and discrete space descriptions of a dynamicalsystem - leading naturally to the graph theoretic tools already discussed.

• mechanisms of chaos.

Even applied topics such as the widely popular topic of controlling chaos [256, 120, 302]61

becomes approachable within the language of symbolic dynamics, simply by forcing de-sirable itineraries [265, 68, 266, 26].

The next section of this chapter will serve as an opener, quick-start tutorial describ-ing the link between symbolic dynamics as a description of a chaotic attractor, and how adynamical system in some sense constantly transmits information, and may even be ma-nipulated to be an information bearing signal. Some of this description is drawn roughlyfrom our review [26]. A more detailed treatment will be given in the sequel sections.

6.1.1 Those “Wild" Symbols

The “wild" is a standard chaos toy that can be purchased from many science educationoutlets. One such realization of the toy is shown in the photograph in Fig. 6.1(Left). Thedevice has a two-degree of freedom pendulum suspended from above which is free to oscil-late along both angular axis, θ and φ. While we could certainly write equations of motionby the Lagrangian method, [149], we will simply state the differential equation is availableand their detail is beside the point of this discussion. Perhaps more realistically, it is possi-ble that the data shown is collected from the actual physical experiment. The point is, wecan refer to output of said variables as seen as time-series in Fig. 6.1(Right).

A natural guess occurs to us as a partitioning from this system. There are 6 magnetsplaced symmetrically under the colored circular segments shown, which are placed to repelthe pendulum arm. Over the top of each magnet, a colored segment shown as labelled withone of the following words,

{“Go to Lunch", “Maybe", “Yes", “Quit", “Try Again", “No"}, (6.1)

61Control of chaos can be described briefly as a different perspective on sensitive dependence to initialconditions, and parameter variations - small feedback control inputs can be designed to yield dramatic outputvariations.

Page 163: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 155

Figure 6.1. (Left) Photograph of a chaotic “wild", which here is a realizationof a chaotic desktop toy. Six magnets are in the base, one under each colored segmentcorresponding to symbols in Eqs. (6.1)-(6.3). Each magnet in the base has an opposingpolarity and so repels to a magnet that is in the silver pendulum tip. (Middle) Time seriesin each of the free angles, θ (t) and φ(t), as well as angular velocities, θ (t) and φ(t). Thetop panel shows that the colored segments in the shown in the wild - left corespond to apartition in the time series easily seen in θ (t) as shown. (Right) The phase space of thechaotic trajectories in (θ ,φ, θ , φ) suppress the time as parameterized curves, here show topand bottom as projections in the (θ ,φ), and (θ , φ) respectively .

over the segments colored in order,

{orange, red, green, yellow, purple, white}. (6.2)

Either way, these serve as a perfectly good symbol sets, and no better than the simpleindexed set,

{1,2,3,4,5,6}. (6.3)

Given these symbolic labels, one can ask the following simple question: “Whichcolored segment is the pendulum to be found over when it reaches its maximum angulardisplacement and reverses direction?" Thus the labels become immediately connected tothe dynamics from which results a symbolic dynamics. In one run during this writing,we found, 1− 5− 1− 5− 1− 5− 4− 3− 5− ...62 which is equivalent to the labels, “Go

62Recall that in standard dynamical systems language, an orbit is an infinite sequence describing the longterm behavior of the system. In symbolic dynamics that means the orbit must be an infinite itinerary, andthe ellipsis, “..." denotes “keep going forever" but the lack of such precision is encoded in the distancefunction of the symbolic dynamics Eq. (6.11) which rewards for more precision near the beginning, and lossof precision as time goes on. Furthermore, the representation as short time segments is in many senses the

Page 164: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

156 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

to Lunch"-“Try Again"-“Go to Lunch"-“Try Again"-“Go to Lunch"-“Try Again"-“Quit"-“Yes"-“Try Again"-... This in a primitive form is a of symbolic dynamics.

The amazing story behind symbolic dynamics is that with an appropriately chosenpartition, the symbols alone are sufficient to completely classify all the possible motions ofthe dynamical system.63 That is there may be a conjugacy between the symbolic dynami-cal system and the dynamical system in natural physical variables. The Smale homoclinictheorem and horeshoe theory is the historical beginning of this correspondance [306, 307].Defining these concepts and determining their validity is the work of the subject as dis-cussed in the rest of this chapter. As it turns out, the partition shown here is not likely to beone of the special (generating) partitions which allow such an equivalence, but nonethelessinteresting symbolic streams result, and such is the issue of the “misplaced" partition whichwe shall also discuss [43] in Section 6.4.6. Now we shall segway to the more detailed topicwith a bit more analytic introductory example.

6.1.2 Lorenz Differential Equations and Successive Maxima Map

As another introductory example to symbolic dynamics, consider the now famous and fa-vorite64 Lorenz system [217],

x = 10(y− x),

y = x(28− z)− y,

z = xy− (8/3)z, (6.4)

as a benchmark example of chaotic oscillations. Those interested in optical transmissioncarriers may recall the Lorenz-like infrared NH3 laser data [324, 325, 178]. A time-seriesof the z(t) variable, of a “typical" trajectory is seen in Fig. 6.2.

Edward Lorenz showed that his equations have the property that the successive localmaxima of the z-time-series can be described by a one-dimensional, one-hump map,

zn+1 = f (zn), (6.5)

where we let zn be the nth local maximum of the state variable z(t). The chaotic attractor inthe phase space (x(t), y(t), z(t)) shown in Fig. 6.3 corresponds to a one-dimensional chaoticattractor in the phase space of the discrete map f (z), and hence the symbol dynamics areparticularly simple to analyze. See Fig. 6.4. The generating partition for defining a goodsymbolic dynamics is now simple as this is a one dimensional map; it is the critical pointzc of the f (z) function. A trajectory point with

z < zc (z > zc) bears the symbol 0 (1). (6.6)

The partition of this one-dimensional map, of successive z(t) maxima corresponds to a tra-ditional Poincare’ surface mapping, as the two leaves of the surface of section can be seen

theme of this book that short time representations of a dynamical system are also useful. In this case, thelong time behavior of this dissipative system is that the pendulum stops in the straight down position.

63If an “inappropriate" partition is used, many orbits may give the same symbolic stream, [43], and so thesymbolic dynamics representation may not be faithful.

64The Lorenz system was originally discussed as a toy model for modelling convection rolls in the atmo-sphere - a phenomenological weather model. However, following a now famous discovery regarding loss ofprecision and sensitive dependence, it has become a pedagogy favorite for introducing chaos.

Page 165: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 157

Figure 6.2. A z(t) time-series from the Lorenz Equations, Eqs. (6.4). This par-ticular z(t) time-series is from the (x(t), y(t), z(t)) trajectory shown in Fig. 6.3. The un-derlined bits denote non-information bearing buffer bits which are necessary either dueto nonmaximal topological entropy of the underlying attractor, or further code restrictionswhich were added for noise resistance, as discussed in [28, 26]. In Sec. 9.4 we will discussfurther this example in the context of information theory in dynamical systems, and thatthese underlined bits correspond to a sub-maximal entropy. [26]

Figure 6.3. Successive maxima map zn+1 = f (zn) from the measure z(t) variable,of the Lorenz flow (x(t), y(t), z(t)) from Eqs. (6.4). [26]

in Fig. 6.3. Each bit roughly corresponds to a rotation of the (x(t), y(t), z(t)) flow aroundthe left or the right wings of the Lorenz butterfly-shaped attractor. However, the Lorenzattractor does not allow arbitrary permutations of rotations around one and then the otherwing; the translates to the statement that corresponding symbolic dynamics has a somewhatrestricted grammar which must be learned. The grammar of the corresponding symbolicdynamics is a statement that completely characterizes the allowed trajectories of the map,

Page 166: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

158 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.4. Lorenz’s butterfly attractor. This particular“typical" trajectory ofEqs. (6.4) shown in Fig. 6.2 when interpreted relative to the generating partition Eq. (6.6)of the one-dimensional successive maxima map Eq. (6.5) shown in Fig. 6.4 is marked bythe red points which relates to the Poincare’ section manifold M. [26]

and hence the flow or equivalently classifies all periodic orbits. In fact, understanding howa chaotic oscillator can be forced to carry a message in its symbolic dynamics is not onlyinstructive and interesting but it became a subject of study for control and transmissionpurposes [170, 28, 26, 70].

6.1.3 One Dimensional Maps With A Single Critical Point

In this subsection we begin a bit more detailed presentation of symbolic dynamics corre-sponding to a dynamical system, starting with the simplest case, a one-humped intervalmap, such as the situation of Lorenz’s successive maxima map. Successively more compli-cated case, multi-humped and then multivariate will be handled in sequel subsections.

f : [a,b]→ [a,b]. (6.7)

Such a map “has" symbolic dynamics [233, 91] relative to a partition at the critical point xc.Choosing a two symbol partition, labeled I={0,1}, naming iterates of an initial conditionx0 according to,

σi (x0)=(

0 if f i (x0) < xc

1 if f i (x0) > xc

). (6.8)

The function h which labels each initial condition x0 and corresponding orbit {x0, x1, x2, ..}by an infinite symbol sequence,

h(x0)≡ σ (x0)= σ0(x0).σ1(x0)σ2(x0)... (6.9)

Defining the “fullshift",

�2 = {σ = σ0.σ1σ2... where σi = 0 or 1}, (6.10)

Page 167: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 159

to be the set of all possible infinite symbolic strings of 0′s and 1′s, then any given infinitesymbolic sequence is a singleton (a point) in the fullshift space, σ (x0) ∈�2. 65

The usual topology of open sets in the shift space �2 follows the metric,

d�2(σ ,σ )=∞∑

i=0

|σi −σ i |2i

, (6.11)

which defines two symbol sequences to be close if they agree in the first several bits.Eq. (6.8) is a good “change of coordinates," or more precisely a homeomorphism,66

h : [a,b]−∪∞i=0 f −i (xc)→�′2, 67 (6.12)

under conditions on f ,68 such as piecewise | f ′|> 1.69 The Bernoulli shift map moves thedecimal point in Eq. (6.32) to the right, and “eliminates" the leading symbol,

(s(σ ))i = σi+1. (6.13)

All of those itineraries from the map f Eq. (6.7) in the interval correspond by Eq. (6.8) tothe Bernoulli shift map restricted to a subshift,

s : �′2→�′2. (6.14)

Furthermore, the change of coordindates h respects the action of the map, meaning it com-mutes, and furthermore it is a conjugacy70.

In summary, the above simply says that corresponding to the orbit of each initialcondition of the map Eq. (6.7), there is an infinite itinerary of 0′s and 1′s, describing eachiterate’s position relative the partition in a natural way, which acts like a “change of coor-dinates" such that the dynamical description is equivalent in either space whether it be inthe interval or in the symbol space.

65 We are using the notation, σi (x0) to denote the symbol of the ith of the initial condition according toEq. (6.8) but σ (x0) is a function which assigns a full infinite symbolic sequence corresponding to the fullorbit from the initial condition x0 by σ : [a,b]→�2.

66A homeomorphism between two topological spaces A and B is a one-one and onto continuous functionh : A→ B , which is formerly the equivalence relationship between two topological spaces.

67A subshift �′2 is a closed and Bernoulli shift map invariant subset of the fullshift, �′2 ⊂�2. A subshiftdescribes the subset of the fullshift which is those infinite symbol sequence which actually do occur in thedynamical system as some periodic orbits may not exist, and correspondingly so must the correspondingsymbolic sequences be absent from the representing subshift �′2.

68Conditions are needed on the set to guarantee uniqueness of the symbolic representation. Some kind offixed point theorem is needed to guarantee that contraction to a unique point in R occurs when describinga symbol sequence of increasing length. In one dimensional dynamics, the contraction mapping theorem isoften used. In diffeomorphisms of the plane, often homology theory is used.

69Note that pre-images of the critical point are removed from [a,b] for the homeomorphism. Thisleaves a Cantor subset of the interval [a,b]. This is necessary since a shift space is also closed andperfect, whereas the real line is a continuum. This is an often over-looked technicality, which is actu-ally similar to the well known problem when constructing the real line in the decimal system (the ten-shift �10) which requires identifying repeating decimal expansions of repeating 9’s such as for example1/5 = 0.199 ≡ 0.2. The corresponding operation to the shift maps [12] is to identify the repeating binaryexpressions σ0.σ1..σn0N = 11 ≡ σ0.σ1..σn1N = 11, thus “closing the holes" of the shift space Cantor setcorresponding to the critical point of the map, and its preimages.

70A conjugacy between two maps is a homeomorphism h between each phase space as topological spacesA and B , which commutes the maps on those two spaces, α : A→ A, β : B → B , then h ◦ α = β ◦ h.Conjugacy can be considered as the major notion and the gold standard of equivalence used in dynamicalsystems theory when comparing two dynamical systems.

Page 168: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

160 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

6.1.4 One Dimensional Maps With Several Critical Points

In general, an interval map Eq. (6.7) may have n critical points,

xc, j , j = 1,2, ..,n, (6.15)

and hence there may be points x ∈ [a,b] with up to n + 1-preimages. See Fig. 6.5 forsuch an example one-dimensional map with many critical points. Therefore, the symboldynamics of such is naturally generalized [214] by expanding the symbol set,

I= {0,1, ..,n}, (6.16)

to define the shift space �n+1. The subshift,

�′n+1 ⊂�n+1, (6.17)

of itineraries corresponding to orbits of the map Eq. (6.7) follows the obvious generaliza-tion of Eq. (6.8),

σi (x0)= j if xc, j < f i (x0) < xc, j+1, j = 0,1, ..,n+1, (6.18)

and taking xc,0 = a and xc,n+1 = b. The characterization of the grammar of the resultingsubshift �′n+1 corresponding to a map with n-turning points is well developed followingthe kneading theory of Milnor and Thurston [233]. See also [91].

Figure 6.5. A one-dimensional map with many critical points requires manysymbols for its symbolic dynamical representation according to Eqs. (6.18)-(6.18).

6.1.5 The Topological Smale Horseshoe

In the 1960’s an American topologist named Stephen Smale developed [306] a surprisinglysimplified model of the complex dynamics apparent in oscillators such as the Van der Pol

Page 169: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 161

oscillator. Smale’s ingenious approach was one of simplification, stripping away the detailsand using the tools of topology, then often called “rubber sheet geometry". He did so duringan impressive summer at the beaches in Rio, about which he writes in his own words, [307]leading to the results we review here, as well as separately a huge break through leadingto his eventual Field’s medal on the topic of the Poincare’ conjecture. Despite the amazingsuccess of the summer, the very location of his work efforts apparently lead to a publicinquiry over doubts that there was work being done rather than a vacation. The continuingimportance even half a century later of the results of that summer leave little doubt as tothe strength of the summer’s work.

The discussion of symbolic dynamics for one dimensional transformations in theprevious subsections appears first in this writing for the simpler issues involved with theone dimensional phase space, and that noninvertible systems require singly infinite shifts.Historically, the more general case came first of bi-infinite shifts capable of discussing dif-feomorphisms which arise naturally from describing differential equations. The mantra oftopology/ “rubber sheet geometry" is to ignore scale and size and reveal structures whichare independent of such coordinate dependent issues. The approach is well presented pic-torally, and as such we can see here. In subsequent sections, in particular Sec. 7.1.3, wewill discuss how it relates to applied dynamical systems.

The basic horseshoe mapping is a diffeomorphism of the plane71,

H : R2→R2. (6.19)

However, we will be most interested in the invariant set of the (unit) box (square) whichwe will denote B ⊂ R2, See Fig. 6.6. The basic action of the horseshoe map shall be astretching phase, that we denote,

S : R2→R2, (6.20)

and a folding phase,F : R2→ R

2. (6.21)

The detailed specifics of each of these two phases are not as important as that we wish theresults to “appear" as drawn in Fig. 6.6 an subsequently Fig. 6.7. That is stated loosely,H (B) is meant to map across itself twice, and in such a manner that there is everywherehyperbolicity (stretching) in the invariant set, and the no-dangling ends property referredto in Fig. 4.5 and considered in the definition of Markov partition, Sec. 4.2.3. The de-tails are not important, but it is easy to see for example that S may be a linear mappingS(x , y) = (ax ,by), a > 1, 0 < b < 1, and certain quadratic functions may be used for aspecific realization of F . When describe we the Horseshoe mapping as having elements ofstretch+fold, we are referring to the definition,

H = F ◦ S. (6.22)

In fact, J. Moser discussed such a decomposition of Henon-type maps [244], the composi-tion of a rotation and a quadratic shear. We will review a Henon-type mapping [98] at theend of this section for specificity.

71A diffeomorphism, F : M→ M is a homeomorphism which is also differentiable. That is, in the topicof smooth manifolds, a diffeomorphism is an invertible function that maps one differentiable manifold to an-other and furthermore, both the function and its inverse are smooth. It serves as the equivalence relationshipin smooth topological manifolds.

Page 170: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

162 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

The form of the horseshoe map decomposition, Eq. (6.23) simplifies the inverse map-ping,

H−1 = S−1 ◦ F−1, (6.23)

which is shown simply by,

S−1 ◦ F−1 ◦ F ◦ S = S−1 ◦ I ◦ S = S−1 ◦ S = I . (6.24)

where I is the identity function. Hence, starting with Eq. (6.23) and left operating (multi-plying) by S−1 ◦ F−1,

S−1 ◦ F−1 ◦ H = S−1 ◦ F−1 ◦ F ◦ S = I , (6.25)

and therefore S−1 ◦ F−1 has the property of being the (left) inverse of H . Likewise, itcan be shown to be the right inverse. Such is shown in Fig. 6.8, that the inverse of thehorseshoe reverses the fold and then the stretch of the forward mapping. That H−1 is theinverse of the horseshoe mapping is itself a horseshoe mapping is argued geometrically bythe pictures in Fig. 6.8.

We shall be concerned with the invariant set � of the box B . As such, define,

�i = {z : z ∈ B , H j (z) ∈ B ∀ 0≤ j ≤ i}, (6.26)

where the notation z = (x , y) ∈R2 denotes an arbitrary point in the plane. Thus �i denotesall those points in B which remain in B for each and every iterate of H at least through thei th iterate. Likewise define,

�−i = {z : z ∈ B , H− j (z) ∈ B ∀ 0≤ j ≤ i}. (6.27)

The the invariant set is written simply as,

�= ∩∞i=−∞�i . (6.28)

In Figs. 6.9-6.14, we see various stages of the eventual invariant set, �1, �−1 ∩�1, �2,�−2, and �−2 ∩�2, respectively, labeled according to itinerary relative to the partitionshown. This leads naturally to symbolic dynamics, which in some sense is just a methodfor book keeping itineraries of each initial condition relative to the partition.

Analogous to the development in Sec. 6.1.3, the horseshoe map admits a symbolicdynamics. A key difference here as compared to the one-dimensional case is that the one-dimensional mappings got their “‘fold" step from a two-to-one-ness almost everywhere72,which is just another way to state the noninvertibility. This noninvertibility is reflected inthe one-sided shift of the symbolic dynamics. However, the horseshoe map is invertible.The invertibility correspondingly must be reflected in the symbolic dynamics which is whya two sided shift must be used. See Sec. 6.2.1 for an analogy of the logistic map to actualcard shuffling.

Define a symbolic partition P to be any curve in between the two legs of the fold -as shown in Fig. 6.9. Then a symbolic dynamics may be defined in terms of itineraries,

σi (z0)=(

0 if H i (x0) < P1 if H i (x0) > P

), for any −∞< i <∞. (6.29)

72Recall that is the logistic map xn+1 = λxn(1− xn) when λ= 4.

Page 171: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 163

Now function h labels each initial z0 according to its bi-infinite orbit,

{..., z−2, z−1, z0, z1, z2, ..} (6.30)

where,zi = H i (z0), (6.31)

for any initial z0 ∈R2 and −∞< i <∞. Analogous to Eq. (6.32) we have,

h(z0)≡ σ (z0)= ...σ−2(z0)σ−1(z0)σ0(z0).σ1(z0)σ2(z0)... (6.32)

For this reason, the symbol space must be bi-infinite in the following sense,

�2 = {σ = ...σ−2σ−1σ0.σ1σ2... where σ0 = 0 or 1}, (6.33)

since there is an infinite number of symbols before the decimal point, “.", representing theprehistory of a trajectory, and another infinity following representing the future. The set �2is the set of all possible such bi-infinite symbol sequences. In the sequel we explore severalproperties of �2, (uncountable, chaotic, infinitely many periodic orbits, positive Lyapunovexponent, etc.)

Open sets in this shift space �2 again follows the metric topology, but this time by abi-infinite symbol comparison,

d�2 (σ ,σ )=∞∑

i=−∞

|σi −σ i |2i

, (6.34)

rewarding agreement for the first several bits near the decimal point. This defines so-calledcylinder sets which are the symbolic version of open sets in �2.

Under a Markov partition property for hyperbolic diffeomorphisms, stated in Sec. 4.2.3,again a good “change of coordinates" exists,[48] but this time by,

h : �→ �2,

z �→ h(z)= σ . (6.35)

The resulting conjugacy describes an equivalence between the dynamical systems,

s ◦ h(z)= h ◦ H (z), (6.36)

meaning each point z ∈ λ ⊂ B ⊂ R2 can allow the change of coordinates and mapped sin �2, or alternatively this gives the same thing as mapping in R by the horseshoe H andthen changing coordinates to symbols by h. This is equivalently stated by the commutingdiagram,

H : � → �

h ↓ ↓s : � → �

. (6.37)

We remind that s is the Bernoulli shift which simply moves the decimal,

s(σi )= σi+1, (6.38)

Page 172: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

164 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

but as a bi-infinite shift, no symbols are dropped. Instead, symbols are simply forgotteninto a fading past in terms of the metric topology inherited from Eq. (6.34). This meansthe focus on the m+ n+ 1 bits before and after the decimal point is shifted to the right.Remembering that those symbols near the point define the neighborhood (the N-cylinder)in which the symbol sequence is found, this implies a loss of precision, or alternativelya loss of initial information regarding initial state; the rate at which this happens shallbe explored in subsequent sections concerning entropy and Lyapunov exponents, 9.4-9.5.Also see further discussion in Example 6.1.

Figure 6.6. The basic topological horseshoe mapping consists of the compositionof two basic mappings, S which is basically a stretching operation, and F which is basicallya folding operation.

Figure 6.7. The Horseshoe mapping is designed to map the two “legs" of H (B)over the original box B, since we are interested in the invariant set, �.

Example 6.1. Example: (Shift of focus and loss of precision - the symbolic metrictopology and the Bernoulli shift map). Consider a specific point σ from the set of allpossible bi-infinite symbol sequences �2, which we write so as to emphasize only m-bits

Page 173: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 165

Figure 6.8. The inverse of the horeshoe map simply reverses the fold then thestretch, according to Eq. (6.25). (Bottom) Thus the folded horseshoe is first straightened(F−1), and then unstretched/shortened (S−1). (Top) Meanwhile, considering what thesetwo operations must do to the original square, a horizontally folded u-bended version ofthe box must result.

Figure 6.9. (Right) The symbol dynamics generating partition P is shown. (Left)Those points which will next land in 0. are labelled .0, and likewise .1 to 1.. Thus on theright we see �1, the one-step invariant set colored yellow, and its pre-iterate on the right,also in yellow.

Page 174: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

166 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.10. The iterate and pre-iterate of B, which are the vertically orientedand horizontally oriented folded sets H (B) and H−1 are shown, and the one-step forward-backward set �−1∩�1 is labelled according to itinerary, and colored yellow.

Figure 6.11. Vertical strips shown describe �−1 and �−2 respectively.

before the current bit ahead of the decimal point, descriptive of the prehistory, and n-bitsafter the decimal point, descriptive of the future of the orbit of σ relative to the Bernoulli-shift. We choose the specific point to be,

σ = ...??σ−6

N=m+1+n=5+1+5=11︷ ︸︸ ︷σ−5σ−4σ−3σ−2σ−1σ0.σ1σ2σ3σ4σ5 σ6??...=

= ...??1

N=11︷ ︸︸ ︷010011.10010 ??... (6.39)

The overbrace is shown to indicate precision of N = m+ n+ 1 points, which is in fact a

Page 175: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 167

Figure 6.12. Horizontal stripes shown represents �2 and labels show the fourpossible symbolic states of itineraries through the partition P .

neighborhood in symbol space. That is, any other point that agrees to that many bits,

σ ′ = ...σ−7σ′−6

N=11︷ ︸︸ ︷010011.10010 σ6σ7... (6.40)

has a distance from σ within the symbol norm Eq. (6.34),

d�2(σ ,σ ′)≤6∑−∞

1

2i+∞∑6

1

2i= 2

26

1

1− 12

= 1

16, (6.41)

assuming the worst case scenario in the inequality, that all the other bits are opposite outsidethe bracketed window. The question marks outside of the braces are shown to emphasizethat to a given N-bit precision. Those bits not included are essentially unknown - just asunknown as in the physical world where claimed precision beyond the capabilities of anyexperiment should be considered fiction.

The Bernoulli-shift map of the decimal point σ , gives a new point in �2 which wewill call σ ,

σ = s(σ )= ...??10

N=11︷ ︸︸ ︷100111.0010? ??..., (6.42)

and likewise the second iterate is a point which we shall call ˆσ ,

ˆσ = s2(σ )= ...??101

N=11︷ ︸︸ ︷001110.010?? ??... (6.43)

Our use of the notation, “??", in these annotation of cylinder sets is meant to emphasizethe unspecified-unknown information corresponding to focusing on just the few bits cor-responding to a neighborhood in symbol space. Really however, when measuring to a

Page 176: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

168 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.13. Intersection of vertical and horizontal strips, ∩2i=−2�i , yield the

small yellow squares which would be labelled with the 2-step future symbols of the verticalstrips �−2, and 2-bit and pre-histories of the horizontal strips. For example, one of therectangles is labelled 01.01.

precision of m = 5 bits of pre-history, and n = 5 bits of future fate, it is unfair to maintainthe bits outside of the braces as we have shown in Eqs. (6.42)-(6.43). Instead, one shouldwrite,

σ = s(σ )= ...??

N=11︷ ︸︸ ︷100111.0010? ??..., ˆσ = s2(σ )= ...??

N=11︷ ︸︸ ︷001110.010?? ??... (6.44)

The resulting distance between the iterates of σ and σ ′ increases as the unspecified digitsbecome prominent, and the specified digits become “forgotten".

d�2(s(σ ),s(σ ′))≤6∑−∞

1

2i+∞∑5

1

2i= 2

26

1

1− 12

= 3

64

d�2(s(σ ),s(σ ′))≤6∑−∞

1

2i+∞∑4

1

2i= 2

26

1

1− 12

= 5

64.

The loss of precision occurs at a exponential rate definitive of the concepts of entropy, andalso Lyapunov exponents which will be explored in the subsequent. �

This example illustrates that in the symbol space metric, “close" means matching thefirst many bits, and this precision is lost with iteration of the Bernoulli shift. The continuityof the homeomorphism73 h : R2→�2 is reflected by the ever decreasing width and nesting

73As it is used here comparing dynamical systems, the homeomorphisms should be thought of as a changeof coordinates to compare to dynamical systems

Page 177: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.1. Symbolization 169

Figure 6.14. The Smale Horseshoe Map in summary. Collecting the geometryalready shown in Figs. 6.6-6.12 give the following summary of the stretch+fold process.

of the vertical strips, and likewise the horizontal strips in Figs. 6.9-6.14, whose intersectionsform small squares. Where continuity means small sets correspond to small sets under themapping, said succinctly as a topologist, the preimage under h of any open set in �2 mustbe an open set in R2.

The vertical stripes shown in Figs. 6.7, 6.9 and 6.11 are �−1 and �−2 respectively.These represent the future fate of the orbit of those z-values therein through the symbolicpartition. Inspection of these pictures and Eqs. (6.26)-(6.27) suggests that

�i ⊂�j , when 0 < j < i , or i < j < 0, (6.45)

Likewise, the horizontal stripes shown in Fig. 6.12 represent pre-histories. The intersectionof horizontal and vertical strips Eq. (6.28) makes for the rectangles shown. The nesting ofeach rectangle of

�−i ∩�i ⊂�− j ∩�j , when 0 < j < i , (6.46)

reflects the corresponding symbolic nesting that j -bit neighborhoods are finer and moreprecise than i -bit neighborhoods when 0 < j < i definitive of the continuity of h.

The limit set, � in Eq. (6.28) is a Cantor set, whose properties are shared by C∞ inFig. 7.10, but let us mention here that as such � is

• Closed,

• Perfect,

Page 178: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

170 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

• Totally disconnected,

• Has cardinality of the continuum,

Also, this Cantor set has zero measure. In this sense there is an oxymoron; it can be saidthat the set “counts big," but “measures small."

Figure 6.15. The Middle Thirds Cantor Set C∞ = ∩n→∞Cn is useful modelset since invariant sets in chaotic dynamical systems, and in particular of horseshoes, aregeneralized Cantor sets, sharing many of the topoligcal properties as well as many of themetric properties such as self similarity and fractal nature.

The power of this symbolic dynamics description is that we can actually prove manyof the basic tenants of chaos theory. That is, this symbolic dynamical system has infinitelymany periodic orbits, sensitive dependence to initial conditions, and dense orbits. Theseshall be included in Sec. 6.2. The trick then to make a proof regarding the original dynamicsis a proof that the representation is correct, and that is often the difficult part, even while atleast numerical evidence may be excellent and easily obtainable.

For now, we shall present the central theorem due to Smale. With this theorem, theseemingly abstract topic of symbolic dynamics becomes highly relevant to even certaindifferential equations that may be found in practical engineering applications.

6.2 Chaos“Chaos Theory" is the face of dynamical systems theory which has appeared most popu-larly in the media and cultural outlets representing dynamical systems [147] including abroad list applications74. This popularity has been a two edged sword, in that while it at-tracts interest, attention and eager students, the very same phrase “chaos" has encouragedmisunderstanding. A pervasive popular misconception is that the mathematical property

74Applications of chaos theory range from biology, epidemiology, astronomy, chemistry, medicine, andeven finance and economics to name a few. It is hardly an exaggeration to say that any science whichincludes a time evolving component and oscillation has the propensity to display chaos as one of its naturalbehaviors.

Page 179: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.2. Chaos 171

called chaos is somehow a philosophical pondering and even that it means the same as itdoes in its English definition; in English chaos denotes disorder. In mathematics, chaosdenotes a kind of order behind what seems as irregularity. Further, chaos is a property likeany other mathematical property in that there must be a specific mathematical definition.

Specifically, there are two popular definitions of chaos, which we highlight and con-trast here. Both are approachable and checkable when we have horseshoes or otherwisesymbolic systems.

The popular Devaney definition of chaos [97] is equivalent to,75,

Definition 6.1. Devaney chaos. Let X be a metric space. A continuous map T : X → Xon a metric space X is chaotic on X if,

1. T is transitive,76

2. the set of periodic points of T are dense in X ,

3. T has sensitive dependence to initial conditions.77

These properties are easy to check by constructive proof for a symbolic system, andhence by conjugacy, for the models which are equivalent to a symbolic shift map, such asthe horseshoe or similarly for certain one dimensional mappings such as the logistic mapwith r = 4.

Property 1: Transitive To show the transitive property, it is sufficient to construct adense orbit. Certainly not all points are dense in �2. For example, each of the (countably)infinitely many periodic points are not dense. E.g.,

σ = 0.000..., and, σ = 1.111..., (6.47)

are the two fixed points. Likewise,

σ = 0.1010..., (6.48)

represents the period-2 orbit, and any other periodic orbit can be written simply with arepeating symbol sequence in the obvious way. However, the point,

σ = 0. 11 00 01 10 112 000 001 ... 1113 0000 0001 ... 11114 ...

(6.49)

is an example of a dense orbit. Each overbar groups symbols of one, then two, then three,...n, etc, symbols of all permutations on those n symbols. There are 2n , n-bit strings in

75The original Devaney definition includes the requirement that periodic orbits of T are dense in X , butseveral papers eventually showed that two of the three original requirements are sufficient to yield the third[12].

76It is sufficient to state that a map T is transitive if there exists a dense orbit. That is, a set A (the orbitof a point is a set - let A = orbit(x0)) is dense in another B (in this case X if the closure of A includes B ,B ⊂ cl(A). For example, the rational numbers are dense in the set of real numbers because any real numberis approximated to arbitrary precision by an appropriately chosen fraction.

77A map T on a metric space is said to have sensitive dependence on initial conditions if there is an r > 0such that, given a point x and arbitrary ε > 0, there is a point y such that d(x, y) < ε and a time k whend(T k (x), T k (y))≥ r [279]. That is, here is always a nearby point whose orbit will end up far away.

Page 180: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

172 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

each grouping. Appended together in any order, including the specific one shown allowsfor a dense orbit. Why? Any other σ ′ ∈�2 is approached to any arbitrary precision, sinceprecision in the symbolic norm means agreeing to m bits for some (possibly large) m, andthose specific m bits are found somewhere amongst the grouping of all the possible m-bitpermutations encoded in σ . In fact, the set of all such dense orbits can be shown to be anuncountable set, as they are formed by all possible orderings of these permutations and soa variation on Cantor’s diagonal argument applies [200].

If instead we are discussing a bi-infinite shift arising from a horseshoe, the construc-tion is very similar, as repeating the same permutations in both bi-infinite directions issufficient.

Property 2: Dense Periodic Orbits Any periodic orbit in �2 is a symbol sequenceof m-bits which repeats. For example, the period-4 orbits include the 4-bit sequence σ =0.001000100010001.... Therefore, given any other symbol sequence σ ′, periodic or not,it is sufficient to identify a periodic point which agrees to the first m bits, no matter whatm > 0 may be chosen. All we need to do is select the first m-bits of σ ′ and then constructσ as a repetition of those bits. For example, if

σ ′ = 0.0011001010100011..., (6.50)

and m = 4 is selected, then the same,

σ = 0.001000100010001..., (6.51)

will suffice, repeating the first 4 symbols of σ ′.

Property 3: Sensitive Dependence to Initial Conditions: Consider a point σ andany arbitrary (close) precision, meaning a specific (large) number of symbols m. Then it isnecessary to demonstrate that we can always find some point σ ′, such that σ and σ ′ agreeto at least the first m-bits. But no matter how far we wish, r > 0, there is a time when theythey iterate at least that far apart. For notation, given σ , let us define σi to be the oppositesymbol of σi in the i th position of σ . That is, if σi = 0, then σi = 1, and otherwise σi = 0.Therefore, for the m-bit precision, if,

σ = σ0.σ1σ2...σmσm+1σm+2.., (6.52)

then choose,σ = σ0.σ1σ2...σmσm+1σm+2.... (6.53)

That is, make the first m-bits agree, and reverse the rest of the bits. It is a direct check bythe geometric series and Eq. (6.11) that such construction yields an r > 0, within diameterof the space �2, at least by the k = m iteration of the shift map. �

These constructions are symbolic, but that is the simplifying point and so with aconjugacy to the horseshoe recall that each symbolic sequence addresses specific points orneighborhoods as suggested in Figs. 6.10 and 6.14.

Another well accepted definition of chaos is found in the popular book by Alligood-Sauer-Yorke, [2] (ASY). It defines a chaotic orbit, (although it is stated therein for mapsof the interval, T : R→R), rather than a chaotic map,

Page 181: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.2. Chaos 173

Definition 6.2. ASY Chaos. A bounded orbit, orbi t(x0) = {x0, x1, x2, ...} ={x0, T (x0), T 2(x0), ...} of a map T is a chaotic orbit if,

1. orbi t(x0) is not asymptotically stable,

2. the Lyapunov exponent of T from x0 is strictly positive.

We will not confirm these here for the symbolic dynamical case since we will needto refer to Lyapunov exponents to be defined in Chapter 8, but it is easy to state roughlythat Lyapunov exponents are descriptive of stretching rates, or in some sense, the rate ofthe sensitive dependence found in the Devaney definition. So it allows for stretching, butin a bounded region by assumption. Further assuming the orbit is not converging to someperiodic orbit leaves only that it must be wandering in a way that may be thought of asanalogous to the transitivity of the Devaney definition.

On the other hand as descriptive contrast, where the Devaney definition states that themap is chaotic or not on some invariant set, the ASY definition is stated for single orbits.Whereas the Devaney definition is popular for the mathematical proof it allows for certainsystems78. By contrast, the ASY definition allows for a single long trajectory perhapseven from an orbit measured from an experimental system to be checked - it is a bit closerto the popular physicists’ notion of chaos based on positive Lyapunov exponents and thequantities are more directly estimated with numerics.

6.2.1 Stretch + Fold Yields Chaos: Shuffling Cards

While mathematical chaos may be misunderstood by many in the popular public to mean“random", the mathematician refers to a deterministic dynamical system which satisfiesthe definition of chaos (one of the favorite definitions at least, Definition 6.1 or Definition6.2). Many so called random processes such as coin flipping or shuffling cards are alsodeterministic in the sense that a robot can be built to flip a coin in a repeatable manner [99]and a machine can be made to shuffle cards in a repeatable manner, and even purchasedas standard equipment in the croupier’s79 arsenal. These are illustrated in Figs. 6.16-6.17.While these devices are deterministic in the sense that identical initial conditions yieldidentical results, theoretically, it is practically impossible to repeat initial conditions iden-tically, and they have such a high degree of sensitivity to initial conditions that such errorquickly grows to swamp the signal. Thus the randomness is not in the dynamical system,so goes the argument. Rather, the randomness is in the small initial imprecision - inabilityto specify initial conditions exactly.

Focusing on shuffling cards as an analogy, the dynamics of the logistic map, Eq. (1.2)which we repeat,

xn+1 = λxn(1− xn), (6.54)

when λ = 4, can be described literally as a card shuffling-like dynamics. This is a usefulanalogy to understand the interplay between chaos, determinism and randomness. SeeFig. 6.18. If we imagine 52 playing cards layed uniformly along the unit interval so thathalf are before the partition point xc = 1

2 , and half after, then the action on the first half is

78Usually for symbolic systems, namely the horseshoe, the Devaney definition is quite strong, and es-pecially popular since the Melnikov method [156, 329, 327] can be used to show that certain periodicallyforced integrable flows have embedded horseshoes

79A croupier is a professional card dealer such as employed in Las Vegas blackjack tables.

Page 182: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

174 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.16. Deterministic Random-like motion: chaos: A card-mixing machinewhich may be purchased from “Jako-O". Reprinted with permission from Noris-Spiele.

Figure 6.17. Deterministic Random-like motion: chaos: A coin-flipping robot,this figure having appeared as Fig. 1a in [99].

to lay them along their range. range([0, 12 ])= [0,1]. Thus the cards are shown spread out

vertically along the entire unit interval. Likewise, range([ 12,1]) = [0,1]. It has the same

range, and those “cards" are also laid out along the same unit interval. (But this is not agood shuffling since the cards on the right are placed right side up, whereas those on the leftremain upside down. This is denoted by the orientations of the arrows. Any croupier whoshuffles like this would not keep their job for long. Other than that, the double coveringreflects the two-one-ness of the map. Then the cut-deck is pushed together due to the

Page 183: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.2. Chaos 175

mapping, since upon the next iteration, the origin of left side or right side is“forgotten."

Figure 6.18. The Logistic Map has Dynamics analogous to shuffling cards. Thestretch plus folding process mechanism underlying chaos is illustrated here as if we areshuffling cards which consists of cutting and then recombining a deck of cards.

Continuing with the analogy between the logistic map and the card shuffling: Asto the role of sensitive dependence to initial conditions, following the two “blue cards"shown in [0, 1

2 ] in Fig. 6.18, we see them spread vertically under the action of the map.This is related to the unstable (positive) Lyapunov exponent. Then the shuffle pushes theextra card between them. Then as it turns out these two cards may end up falling onopposite sides of the partition xc, which means they will be cut to opposite sides of futureshuffles. Their initial closeness will be lost and forgotten. The determinism is nonethelessdescriptive of a stretching and folding process which naturally forgets initial imprecision.Thus the mechanism behind the chaos, which is described rigorously and mathematicallyby symbolic dynamics, is nothing other than an analogy to the stretch plus fold process ofa card shuffle.

6.2.2 Horseshoes in Henon

According to previous sections, when a horseshoe topology can be proven, just by con-sidering a single iterate of a carefully chosen subset, then chaos has been proven in anembedded subset by preceeding theory. Devaney and Nitecki have shown that exactly thisstrategy is straightforward for the Henon map. We will also discuss a similar strategy for aPoincare’ map derived from a continuous flow - the Duffing oscillator.

The Henon map [173] is an extremely famous example both for pedagogy and re-search of simple mapping of the plane, T :R2→R2, that gives rise to complicated behavior

Page 184: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

176 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

and which apparently has a “strange attractor". This mapping written in the form,

T (x , y)= (1−ax2+ y,bx), (6.55)

gives rise to a chaotic attractor80, as shown in Fig. 6.19 using the parameters a = 1.4,b =0.3. These are the most common parameter values and give rise to this familiar form shownand seen in so many other presentations.

Figure 6.19. The Henon attractor, from a long term trajectory of the Henon map,Eq. (6.55), with the usual parameter values a = 1.4,b = 0.3, gives rise to the form shownusing 100,000 iterates. The stretch and folding nature of the resulting attractor is high-lighted by a few blow-up insets

In a slightly different form, Devaney and Nitecki showed explicitly that the Henonequations can give rise to a Smale horseshoe [98]. They proved existence of an embed-ded horseshoe dynamics for a form of the Henon mapping which may be obtained by acoordinate transformation from the usual Eq. (6.20),

T (x , y)= (a−by− x2, x), (6.56)

which as a diffeomorphism of the plane has an inverse,

T−1(x , y)= (y, (a− x− y2)/b) (6.57)80The attractor set is apparently chaotic, but this turns out to be difficult to prove, and remains open for

specific typical parameter values [340, 16]

Page 185: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.2. Chaos 177

Specifically, their result may be stated81:

Theorem 6.3. If b �= 0 and a ≥ (5+ 2√

5)(1+ |b|)2/4, then there exists an embeddedhorseshoe which may be found by considering the invariant set of T in Eq. (6.56) in thesquare,

S = {(x , y) :−s ≤ x , y ≤ s} where s = (1+|b|+√

(1+|b|)2+4a)/2. (6.58)

It can be checked that a = 4,b= 0.3 are parameters that admit a horseshoe accordingto the Devaney-Nitecki theorem. Simply stated, the proof consists of a direct check that thestated square maps across itself in the geometric way described by the topological horeshoe,Sec. 6.1.5 and Fig. 6.6. Rather than repeat that analysis, we will simply refer to the pictureof the numerical iteration, and pre-iterate of the square shown in Fig. 6.20. This folding issufficient to cause an invariant set � which is a Cantor set on which T |� is conjugate tothe Bernoulli shift map on symbol space, s|�2 as promised by Smale’s horseshoe theory.Furthermore on that set with the stated geometry, we have already repeated the proof thatthere is chaos. Of course, as usual this specific set is a Cantor set of measure zero. A muchlarger set, the attractor if it exists is not addressed by the horseshoe, or in this case, theattractor includes that almost every point diverges to infinity. This leads to the followingremark.

Remark: Where as it turns out that the invariant set � for the Henon map Eq. (6.56) isan unstable Cantor set, chaos is proven, but the proof is only relevant almost nowhere ina measure theoretic sense for Fig. 6.20! It could be in some systems that either a measurezero set is the only invariant set, or on the other hand, there may be a larger invariant seton which there is also chaos, but the chaos may not be that of a full-shift82 In the case ofthe Henon map Eq. (6.56), almost every initial condition does not have a fate related to thechaotic set; these points have orbits which behaves quite simply. They diverge to infinity.The chaotic set of the Henon map Eq. (6.56) is called a chaotic saddle since the chaoticset � has a stable manifold Ws (�) which does not contain any open discs in R2 - it hasdimensionality less than 2. �

On the other hand, the Henon map Eq. (6.55) as shown in Fig. 6.19 has the attractorA shown. This follows from demonstrating a trapping region T which in this case may bedemonstrated by a trapezoid shown in blue in Fig. 6.21, where we demonstrate that T (T )in red properly maps into T . This means that every initial condition from T maps into T .For relevance, we also show the attractor set in blue, A⊂ T (T )⊂ T .

6.2.3 Horseshoes in the Duffing Oscillator

The direct construction method of Sec. 6.2.2 of finding a topological horseshoe in theHenon map by finding a topological rectangle which maps across itself in the appropriategeometry described in Sec. 6.1.5 may be attempted for other maps, and even for Poincare’maps derived from a flow. Here we show the example of the Duffing map derived from aDuffing flow of the differential equations, Eq. (1.36)-(1.39) with an attractor already shownin Fig. 1.12.

81We have abbreviated a more detailed form found in [279] to highlight the feature of the embeddedhorseshoe

82The restricted grammar of a subshift will be discussed in Sec. 6.4.

Page 186: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

178 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.20. A numerical iteration of the square S in Eq. (6.58) (black) by aspecial form of the Henon map T and its pre-iterate T−1, Eqs. (6.56)-(6.56). (Left) Theforward iterate (red) and preiterate (blue) reveal the topological horeshoe geometry.

In Fig. 6.2.3, we show an oriented disc which maps across itself in a manner sugges-tive of the topological horseshoe. The orientation is relevant for the topological horeshoe,and we have taken liberty to demonstrate the orientation with a happy face. This figurewhile indicative of embedded complex behavior, it also shows some short falls.

• The figure hides the fact since we have shown a disc instead of a rectangle, but the

Page 187: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.2. Chaos 179

Figure 6.21. That there is an invariant set of the Henon map Eq. (6.55), followsby demonstrating a set that maps entirely into itself. The trapezoid set shown in black,T maps to T (T ) which is the red set shown and which can be confirmed to be properlycontained. Also shown is the attractor set A in blue already seen in Fig. 6.19.

happy face does not demonstrate a fullshift horeshoe. Further investigation revealsthat only an incomplete horseshoe is described by this set and its iterations, suchas the incomplete horseshoes shown in Fig. 6.23. Such incomplete horseshoes stillindicate some stretch and folding behavior, but a fullshift does not result, as explainedby the picture and subtext of Fig. 6.24. Rather some missing words in the grammarare typical of a subshift, as discussed further in Sec. 6.4. The rigorous pruning theorycan be found in [73, 154, 83] as it applies to certain special cases.

• No matter how complex may be the invariant set of the oriented disc shown, only twosymbols can be accounted for for by a simple horseshoe. The Duffing map howeverhas many folds and pleats which requires partial use of many symbols at least. SeeFig. 6.23(Right). More discussion toward learning the symbolic dynamics (the gram-mar) of the attractor, rather than simply of a measure zero unstable chaotic saddle,is discussed in Sec. 6.4 which follows. While it is exceedingly difficult to prove therepresentation of such a complex symbolic dynamics is faithful, it is computationallyquite feasible and straight forward to use empirically at least.

Page 188: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

180 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.22. A Duffing map derived by the stroboscopic method of Poincare’mapping can be used as an example to investigate an embedded horeshoe. (Inset) ThePoincare’ section shown as successive 2π and copies of the attractor are painted on eachsuch surface between which each orbit flights. (Stretch and fold in the Duffing attractor).The oriented disc shown (happy face) maps across itself in a manner suggestive of a hore-shoe. However, as discussed the covering is not complete to reveal a fullshift horeshoe. SeeFig. 6.24 for further discussion of the incompleteness of this horseshoe.

6.3 Horseshoe Chaos by Melnikov Function AnalysisMelnikov’s method is a powerful analytic tool which has a role both in verifying existenceof chaos as well as sometimes characterizing transport activity in certain dynamical systemsfor which the appropriate setup can be achieved. For sake of noting it as an alternate to themore computational based methods discussed as a centerpiece in most of this book. Ratherthan getting too side-tracked into another large field, we will give a very short presentationof this rich method here, and then cite several references to excellent sources for Melnikovanalysis.

Direct construction of a horseshoe as described above in Secs. 6.2.2-6.2.3 for the

Page 189: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.3. Horseshoe Chaos by Melnikov Function Analysis 181

Figure 6.23. Incomplete horseshoes. The first horseshoe is complete, but thenext two “horseshoes" are incomplete in that they do not fold completely across the region.See also Fig. 6.24 for discussion of how such missing words could arise from tangencybifurcation, and 6.31 for illustration of the consequence of missing words when a tangencybifurcation occurs. See [83] for a full discussing of pruning horeshoes. (Right) The fourthpicture suggests that at least a 2-bit symbol space �2 would be required, but incompletefolding suggests a submaximal entropy, hT < ln2.

Figure 6.24. (Right) A charicature of a of homoclinic tangle due to a hyperbolicfixed point p with stable and unstable manifolds Ws ( p) and Wu ( p), causing a p.i.p. ho-moclinic point h. A full folding horseshoe allows for the Markov partition property - nodangling ends - suggested in Sec. 4.2.3 an Fig. 4.5. A fullshift on all symbols of �2 follow.(Left) Some of those words are lost when the Markov partition property is lost, here shownby an unfolding of a tangency bifurcation. The consequence of which is 2 of the 28 8-bitwords draw so close that they annihilate at t, and likewise, their iterates annihilate at f (t)and f −1(t). Thus the entropy decreases accordingly, hT < ln(2) which may be calculatedby the ln of the largest eigenvalue of a 28 deBroijn graph with 6 missing transitions. Asfurther tangencies bifurcations unfold, generally more words will be lost. [43]

Page 190: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

182 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Henon mapping and then the Duffing oscillator Poincare’ mapping Figs. 6.20, 6.21 and6.2.3 cannot be a general method due to the difficulty to produce the appropriate regions.However, the Melnikov analysis applies more broadly to certain differential equations andit also allows us to check for existence of horseshoes, and further it allows for parametericstudy to discuss homoclinic (and heteroclinic) bifurcations which produce such chaos. Re-call that Smale proved Theorem 7.3, [306], that a transverse homoclinic point of a hyper-bolic periodic point w of a Cr diffeomorphism, r ≥ 2, implies an embedded horseshoe.It is well known and straightforward to prove that a horseshoe is chaotic. The Melnikovfunction gives a measure of the distance between stable and unstable manifold, Ws (w) andWu (w) with respect to a parameterization of these curves when this distance is small. Inthis way, the Melnikov function can be used to decide the existence of a transverse inter-section. We follow most closely, [?].

For sake of simplicity of presentation, we will restrict to the a most straight forwardversion of the Melnikov method. We will assume here an autonomous Hamiltonian systemof the plane H (q , p), under the influence of a small time periodic perturbation,

g(q , p, t)= g(q , p, t+ T ), for some T > 0. (6.59)

The Melnikov analysis we use assumes a dynamical system of the form,

dq

dt= ∂ H

∂p+ εg1(q , p, t) (6.60)

dp

dt=−∂ H

∂q+ εg2(q , p, t), (6.61)

or,z= J ·∇ H (z)+ εg(z, t), (6.62)

where,

J =(

0 1−1 0

),∇ H =< ∂ H

∂q,∂ H

∂p>t ,g=< g1, g2 >t ,z=< q , p >t , (6.63)

[329]. Furthermore, the unperturbed system,

z= J ·∇ H(z), (6.64)

must have a hyperbolic fixed point w with a homoclinic connection orbit, which we callz∗(t), which surrounds a continuous family of nested periodic orbits. Note that in the ex-tended system, the hyperbolic fixed point of the unperturbed vector field becomes a periodpoint and under a sufficiently small perturbation of period T , we may assume the existenceof unique hyperbolic period orbit w∗ of period T. Under these assumptions, the Melnikovfunction,

M(t0)=∫ ∞−∞

g(z∗(t), t+ t0) ·∇ H(z∗(t))dt (6.65)

measures distance between the stable and unstable manifolds of w∗ along the direction ofalong the ∇ H in the time T -stroboscopic Poincare’ section phase plane, where t0 param-eterizes the unperturbed homoclinic orbit z∗(t). The Melnikov function M(t0) is propor-tional to distance between stable and unstable manifolds of w∗ at z∗(−t0). Under the above

Page 191: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.3. Horseshoe Chaos by Melnikov Function Analysis 183

assumptions, the result is that existence of a zero, (a t0 such that M(t0) = 0), which issimple, ( ∂M

∂t |t0 �= 0), implies that the dynamical system Eq. (6.60) has a transverse homo-clinic point and hence posses an embedded horseshoe. The depiction in Fig. 6.25 suggestsno intersection between stable and unstable manifolds and therefore the resulting M(t0)would have no roots. On the other hand, a tangency between stable and unstable manifoldswould result in non-simple roots. Finally, transverse intersections result in simple roots, asfocused on by the Melnikov’s theorem.

Figure 6.25. The Melnikov function M(t0), Eq. (6.65), as applied to flows ofthe form, Eq. (6.60), identifies the distance between stable and unstable manifolds of w∗of the perturbed system (in red) at a reference time t0. This perturbative analysis startswith a homoclinic connection of w of the autonomous system ε = 0, shown in black. Thisdepiction suggests that this Melnikov method will never reach zero since the stable andunstable manifolds do not cross.

For sake of brevity, we only mention a specific example here, but without details.Perhaps the most standard example is for a Duffing oscillator of the form x = y, y = x −x3+ ε(γ cosωt − δy), [262, 212, 327, 156]. This is a very special (rare) example in thatit results in a Melnikov integral which can actually be computed in closed form, and thestate of existence of simple roots can be decided explicitly and with respect to variationof the system parameters. See also pendulum examples in [?]. In this way, a completeglobal bifurcation analysis can be performed. However, Melnikov integrals are generallynontrivial to evaluate and one must resort to numerical evaluation, as in [36]. The difficultyis due to an infinity of oscillations of the integrand in the finite space, the length of theparameterized curve of the homoclinic connection of the autonomous system,; this shouldbe expected since the integrand is a description of a homoclinic tangle.

There are many generalizations of this basic Melnikov analysis which include, allow-ing for higher dimensional problems [333], stochastic problems [303], subharmonic analy-sis [199], as well as analysis of the area of lobes, turnstiles and thus discussion of transport[?]. The power of the Melnikov method is that it is a more analytically based approach forglobal analysis, and it is capable of including parametric study, but it does require the dy-namical system to be presented in some variation of the basic form, Eqs. (6.59)-(6.62). Weassert that the transfer operator methods and the FTLE methods are both more empirically

Page 192: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

184 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

oriented capable of handling dynamical systems known only through observations. Whenthe Melnikov method setup is possible it is very possible. The empirical methods of thisbook and the Melnikov methods have complementary information to offer.

6.4 Learning Symbolic Grammar in Practice6.4.1 Symbol Dynamics Theory

Symbol dynamics is a detailed and significant theory unto itself, with important connec-tions to dynamical systems as discussed here, but also importantly to coding theory andalso information theory. We will summarize here these concepts in a manner meant to al-low computational approximation relevant to empirically investigated dynamical systems.

One might ask what the discrete mathematical notions of grammar and subshiftshave to do with a flow such as the Lorenz equations, Eq. (6.4). In terms of the Lorenzbutterfly, and corresponding one-dimensional map Eq. (6.7), if a “0" bit represents a flightaround one of the two butterfly wings, and a “1" represents a flight around the other wing,before piercing the Poincare’ surface, then a specific infinite string of these symbols in-dicates a corresponding trajectory flight around the attractor. On the other hand, supposethere is a a missing word. For example, suppose no infinite word has the string “00" in it.This would indicate that no initial condition (x(0), y(0), z(0)) of the Lorenz equations existswhich makes two successive laps around the left wing. In fact, on the Lorenz attractor cor-responding to standard parameter values p0 = (10,28,8/3), there do exist initial conditionswhich permit both a “00" string and a “11" string, but in fact not all finite length stringsdo occur. The missing minimal forbidden words typically tend to be somewhat longer thantwo bits, but not necessarily. That point is that we generally expect only a “subshift" of thefullshift of all possible words.

Definition 6.4. Given a �′n a subset of an n-bit symbol space �n , �′n is a subshift if it is,

1. Topologically a closed set, (contains all of its limit points),

2. Closed with respect to the action of the Bernoulli shift map. That is, if σ ∈�′n , thens(σ ) ∈�′n .

In a physical experiment, corresponding to the one-dimensional map such as Eq. (6.7),it is possible to approximately deduce the grammar of the corresponding symbolic dynam-ics by systematic recording of the measured variables. Note that any real measurementof an experiment consists of a necessarily finite data set. Therefore, in practice, it can beargued that there is no such thing as a grammar of infinite type in the laboratory.83 Sowithout loss of generality, we may consider only grammars of finite type for our purposes.Such a subshift is a special case of a sofic shift [198, 214]. In other words, there exists a

83It could be furthermore argued that there is no such thing as measuring chaos in the laboratory, sincemost popular definitions of chaos [97, 2] are asymptotic in that they require sensitive dependance and topo-logical transitivity, both of which would require time going to infinity to confirm. In fact, it is valid to arguethat a mathematical dynamical systems is defined in terms of orbits as infinite trajectories, and therefore nosuch thing exists in physical experiments. In this sense, the premise of many points in this book regard whatcan be learned from finite samples of orbits.

Page 193: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.4. Learning Symbolic Grammar in Practice 185

finite digraph which completely describes the grammar. All allowed words of the subshift,�′2 corresponding to itineraries of orbits of the map correspond to some walk through thegraph.

Figure 6.26. The full 2-shift �2 is generated by all possible infinite walks throughthe in (Left) digraph. (Right) All walks through the graph describes a subshift �′2 whichis the set of all symbol sequences which can be described by a “grammar" in which thesymbol “0" never occurs more than once in a row. Contrast this presentation of these shiftsto lifted versions of the same to larger graph presentations in Fig. 6.27, but the largergraphs would also allow more nuanced and finer restrictions of the grammar.

Definition 6.5. [214, 198] A sofic shift is a subshift �′n which is generated by a digraph(directed graph) G.84 If each edge from each vertex is labelled by one of the symbolsfrom the symbol set {1,2, ...,n}, then an infinite walk through the graph generates a symbolsequence σ ∈ �n . Let S be the set of all possible symbol sequences so generated by thegraph G. If S =�′n then the subshift is generated by G. A sofic shift is a subshift generatedby some finite graph G.

Not all subshits are sofic, but any subshift can be well approximated by a sofic shift,which is convenient for computational reasons. For example, the full 2-shift shift is gener-ated by the graph in Fig. 6.27a, but this is not the minimal graph generating �2. The graphin Fig. 6.26(Left) also generates �2. Likewise, the “No two zero’s in a row" subshift �′2 isgenerated by all possible infinite walks through the digraph in Fig. 6.27b, in which the onlytwo vertices corresponding to “00" words have been eliminated, together with their inputand output edges.

Entropy is a way of comparing different subshifts. While subshifts are generallyuncountable, we may describe the entropy of a subshift as the growth rate of the numberof words of length n. There are strong connections to information theory, which will beexplored in the subsequent Chapter 9. In the meanwhile for our discussion here, we willstate simply that the topological entropy of a subshift �′k on k-symbols �′k , can be definedby cardinality (counting when finite), [198, 214, 281],

hT (�′k)= limsupn→∞

ln Nn

n, (6.66)

where Nn ≤ 2n is the number of ({0}, {1}) binary sequences (words) of length n.

Definition 6.6. A sofic shift is defined as right resolvent if each “out edge" from a vertexof the graph presentation is uniquely labelled.

84 A directed graph G = (E , V ) is a set of edges E = {e1,e2, ...,eM } and vertices consisting of specificedge pairs, {ei ,ej }.

Page 194: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

186 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.27. The full 2-shift �2 is generated by all possible infinite walks throughthe in a) above digraph. b) The “No two zero’s in a row" subshift grammar �′2 is generatedby all possible infinite walks through the above digraph, in which the only two verticescorresponding to “00" words have been eliminated, together with their input and outputedges. Contrast this figure to a minimal graph presentation of the same subshift in Fig. 6.26.[26]

By theorem [198, 214, 281], when the subshift is a right resolvent shift, then a spec-tral computation can be used to compute entropy of shift spaces of many symbols.

hT (�′k)= lnρ(A), (6.67)

where ρ(A) is the spectral radius85 of an adjacency matrix A86, and A generates the soficshift �′k .

Concepts of entropy are useful in considering evolution with respect to the bifur-cations which occur when a parameter is changed. For example, consider the entropy ofthe symbolic dynamics from the logistic map, Eq. 6.54. When λ = 4, we have alreadydiscussed that a fullshift results. Since,

A1 =(

1 11 1

)(6.68)

is the adjacency matrix for the graph G shown in Fig. 6.26(Left) generates the fullshift�2, and ρ(A) = 2, then hT (λ) = ln2 when λ = 4. Notice that now we are writing theargument of hT (·) to be the parameter of the logistic map. Comparing also to the graph andcorresponding adjacency matrix Fig. 6.29a, it can be confirmed that ρ(A1) = ρ(A4) = 2and in fact the generated shift spaces are the same fullshift �2. The graphs G A1 and G A4

are both called deBroijn graphs.

85Largest eigenvalue.86 A is an adjacency matrix of a graph A if A is a matrix of entries of 0’s and 1′ where Ai, j = 1 if there is

an edge from vertex j to i and otherwise Ai, j = 0.

Page 195: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.4. Learning Symbolic Grammar in Practice 187

6.4.2 Symbolic Dynamics Representation on a ComputerRepresentation of a Dynamical System

In practice, beyond the horseshoe map and beyond a one dimensional one hump map somework must be done to learn the grammar of a symbolic dynamics corresponding to thedynamical system which generated the data. We will discuss here some of the book keep-ing involved to develop a useful symbolic model of the underlying dynamical system, butwe do not claim that this representation is exactly correct. Several possible errors of do-ing this work on the computer can easily creep into the computations including, incorrectpartition or at least inexact computer representation of a partition even if known exactly,finite word book keeping when longer or infinite representations of the grammar should beused, non Markov partitions used which is related to the errors already stated. A great dealof excellent work has been carried out to produce computer assisted proofs of a symbolicdynamics representation of a dynamical system, and underlying periodic orbit structure, inpart using interval arithmetic, such as the work in [143, 141, 145, 234, 340] and includingproof of chaos. Our discussion here will simply be heuristic in that we will not discusscomputer proof but rather simply implicate that the methods presented suggest refining andimproving symbolic models of a dynamical system rather than exact representations.

6.4.3 Approximating a Symbolic Grammar in a One-DimensionalMap

Given a finite measured data set {xi}Ni=0, simply by recording all observed words of lengthn corresponding to observed orbits, and recording this list amongst all possible 2n suchwords, the appropriate digraph can be constructed as in Fig. 6.27. One should choose nto be the length of the observed minimal forbidden word. Sometimes, n is easy to deduceby inspection as would be the case if it were “00" as in Fig. 6.27b, but difficult to deducefor larger n. The minimal forbidden word length follows a subgroup construction [43, 42]related to the “follower-set" construction [198, 214].

Since, the data set {xi}Ni=0 is finite, then if the true minimal forbidden word corre-sponding to the dynamical system is longer than data sample size, n > N , only an ap-proximation of the grammar is possible. Therefore the corresponding observed subshift isexpected to be a subset of whatever might be the true subshift of the model map Eq. (6.7)or experiment. This is generally not a serious problem for our purposes since some course-graining always results in an experiment, and this sort of error will be small as long asthe word length is chosen to be reasonably large without any observed inconsistencies. Asa technical note of practical importance, we have found link-lists to be the most efficientmethod to record a directed graph together with its allowed transitions.

6.4.4 As a Parameter Varies, the Grammar also Varies as Told by theKneading Theory Story

Now it is both interesting and instructive to consider what happens to the shift spaces andthe corresponding entropy, as λ is decreased from 4. First we will discuss this in terms ofchanges to the graph representations of the grammar of the symbolic dynamics, and thenwe will relate that to the elegant kneading theory of Milnor and Thurston, [233].

There is a well known bifurcation sequence of the logistic map, Eq. 6.54, as we vary

Page 196: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

188 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

0≤ λ ≤ 4, gives rise to the famous Feigenbaum diagram shown in Fig. 6.28. The bifurca-tion analysis summarized in this figure has been written about extensively elsewhere, suchas in [97, 281, 2]. This is the story of the period doubling bifurcation route to chaos, in-cluding the period three implies chaos story [268] and the detailed elegant theory of theSharkovskii’s theorem [?] which completely characterizes the order in which the periodicorbits appear as the parameter λ is increased. Therefore we will not focus in these featureshere. Here, we will focus on just the changes to symbolic dynamics as learned from a finitegraph as this keeps the focus on the general topic of this writing.

In [29] we considered changes to a finite representation of the grammar of a logistic-like map (one hump map) as a parameter is varied equivalent87 to lowering the peak, suchas reducing λ from 4 in the logistic map. Consider a 4-bit representation of the sym-bolic dynamics on two symbols regarding �2 as illustrated by the deBroijn graph shown inFig. 6.29(top) which is an unrestricted 2-shift and as such the generated grammar is equiva-lent to the simpler 2-vertex graph presentation in Fig. 6.26 of the same grammar. Comparealso to 6.27.

Inherently, reducing λ from 4 results successively in loosing access to words in theshift space. The key into making such a statement computation is to know the order inwhich words are lost, and this is with respect to the so-called lexographic order, otherwisecalled the Gray-code order. This order is depicted to 4-bit representation in the graphs inFig. 6.26 and 6.29 as these words are laid out left to right from 0.000 to 1.000 in the orderwhich is monotonic with the standard order of the unit interval. Formerly, this is one ofthe major results of kneading theory that these two orders are monotonic with respect toeach other, [233]. That is, if symbol sequences are ordered σ (x ′) ≺ σ (x) implies x ′ < xwhere σ (x)�2 is the symbol sequence by Eq. (6.29) and (6.32) corresponding to real valuesx ∈ [0,1] and ≺ is the Gray-code order in the symbol space while < is the standard orderin R. Furthermore in the kneading theory are conditions under which as the parameter λvaries in a “full family" of maps (such as the logistic map since it gives the fullshift), thenthe corresponding grammar reduces continuously with respect to the order topology fromthe Gray code order. We can see the result of removing words in order with respect to ≺ inFig. 6.30 in terms of decreasing topological entropy. The computation is in terms of a finiterepresentation using the graph of size N = 2n vertices, n = 14 bits, and the correspondingspectral computation by Eq. (6.67) where words are removed one by one from the graphin the same order by known monotonicity by ≺ that they disappear by reducing λ from4. The coarser presentation of the same idea is depicted in Fig. 6.29(a-d)(Left Column)for 24 = 16 bit deBroijn graph presentations of the grammars resulting in removing 4-bit words in order monotone by ≺ to reducing λ. The resulting transition matrices areshown in Fig. 6.29(Right Column). Removing a word from the graph (a vertex) must resultin removing all the ingoing edges and outgoing edges to that vertex (for the shift-closedproperty in the definition of a subshift), and therefore sometimes some words not directlyremoved (x’ed) effectively disappear. Correspondingly all the row and column from aremoved word must be zero’ed. Thus the spectral radius and hence topological entropymust be monotone nonincreasing. This is what is seen in Fig. 6.30 where the entropy isindeed monotone nonincreasing with respect to words removed and hence also with respectto λ. Another enticing feature of the figure is that entropy has flat spots - apparently the

87We were interested in communication with chaos, and the formulation in [29] resulted in increasednoise resistance in the transmission of messages on a chaotic carrier by avoiding signals that wander nearthe symbol partition.

Page 197: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.4. Learning Symbolic Grammar in Practice 189

function reminds us of a devils’ straircase function, which indeed is the result of the limit ofthe process to the continuum of a fine grammar representation as n→∞. As discussed in[29], the flat spots result because those words which are removed by loosing paths to themby having direclty removing some other word cannot be removed directly when it might betheir turn in order to be removed; such is a common mechanism as we see resulting in theflat spots and in fact the width of the flat spot represents the order n of the presentation ofthe grammar as a finite graph of a given scale. The smaller the n where this phenomenonmay occur the wider the flat spot.

Figure 6.28. The Feigenbaum diagram illustrating bifurcations of the Logisticmap, Eq. 6.54, as we vary 0≤ λ≤ 4.

6.4.5 Approximating a Symbolic Grammar and Including aMultivariate Setting

Empirically, the multivariate setting of learning a symbolic grammar has a great deal incommon with the single variate map case of the previous subsection above. At its simplest,the problem reduces to good book keeping. For a given sample orbit segment, {zn}Nn=0,record observed m-bit words relative to a given partition, and the transitions between them.This is just a matter of keeping the observed generating graph as a link list since a link listis particularly efficient; as such, the book keeping work here has almost no issues that aredependent on dimension, other than that dimensionality effects how long must be the orbitto usefully saturate toward a full set of observations (fill the space). However, the main

Page 198: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

190 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.29. Restricting the grammar either by reducing the height of the onehump map (for example by reducing r in the logistic map xn+1 = r xn(1− xn) correspondsto removing words in this diadic graph presentation, shown here ordered to the Gray codeorder [29] as dictated by the kneading theory [233]. Compare also to 6.27. [29]

Page 199: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.4. Learning Symbolic Grammar in Practice 191

Figure 6.30. As the grammar becomes more restrictive, the topological entropydecreases. The devil’s staircase aspect of this function is described in [29]. Grammar re-strictions describe removed words from the grammar as illustrated in Fig. 6.29, but herefor a fine representation of the grammar with a graph of size 2n for a large n and a cor-responding transition matrix A2n×2n where entropy is computed by the spectral formulaEq. (6.67). n = 14. [26]

difficulty is to use an appropriate generating partition and this becomes nontrivial in morethan one dimension. The issue of partition is discussed in the following in Sec. 6.4.6. In anycase, given a partition there is a corresponding symbolic dynamics relative to that partitionwith a grammar that may be useful to learn, regardless of whether the appropriate partitionis used from which the dynamical system may be conjugate to the symbolic dynamics.

Example 6.2. Symbol Dynamics of the Henon Map Approximation and Recorded asa Link List. Consider the Henon map, Eq. (1.1). The attractor is shown in Fig. 6.31together with a symbolic partition that is useful for our discussion. A well consideredconjecture [154, 73] regarding the symbol partition of the Henon map is that a generatingpartition must be a curve that passes through homoclinic tangencies. As we see in Fig. 6.24,the darkened “horizontal" w-shaped curve near y = 0 is just such a curve passing throughprimary tangencies. Call this curve C . Above C we label the region “0” and below we label“1". Iterates and pre-iterates of C define a refining partition on the phase space with labelspartially shown in Fig. 6.31 up to two (pre)iterates. As an aside, it is easy to see that thepartition is not Markov in two iterates and preiterates at least as is most clear by inspectingthe example of 11.10 labeled branch as it has the “dangling ends" property forbidden of aMarkov partition. Compare to Definition 4.4 of higher dimensional Markov partitions and

Page 200: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

192 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Fig. 4.5 and Fig. 6.23.

ObservedWord → Shi f t“0"Observed Shi f t“1"Observed

00.11 → 01.1010.11 → 01.1010.10 → 01.00 01.0100.10 → 01.00 01.0100.00 → 00.00 00.0110.00 → 00.00 00.0100.01 → 00.10 00.1111.00 → 10.00 10.0101.01 → 10.10 10.1101.00 → 10.00 10.0111.10 → 11.00 11.0101.10 → 11.00 11.0101.11 → 11.10

(6.69)

Notice that there are only 13 4-bit words out of the 24 = 16 feasible words. Likewise thereare missing transitions. As we can see from Fig. 6.31 the branches of attractor fail to extendto would be so-labelled regions. For example, immediately above the region labelled 01.11is a region that would be labelled 11.11 if it were occupied but since the attractor does notintersect that region, no points on the attractor have orbits who symbolic words have four1’s in a row. This is reflected in the table of allowed words.

Similarly, consider the same map under a somewhat arbitrary partition. We shouldnot a priori expect that a partition of arbitrary rectangles is a Markov or even a generatingpartition. Consider the coarse partition of rectangles that was used to produce the directedgraph in Fig. 1.1. Nonetheless, a symbolic dynamics is induced, this one on 15 symbols in�15. Reading directly from the directed graph in Fig. 1.1, the corresponding link list couldbe written,

From → T o

1 → 132 → 10 13 153 → 104 → 7 85 → 3 4 56 → 7 87 → 5 6 78 → 1 2 39 → 6 710 → 2 3 4 611 → 1 2 312 → 9 12 1313 → 9 1014 → 715 → 2 3 4 6

(6.70)

Page 201: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.4. Learning Symbolic Grammar in Practice 193

Likewise, the a link list of the 2-bit words on these 15 symbols could be formed, to include1.1, 1.2, 1.3, etc.

These observations beg the question as to which partition is correct? Why do we needa generating partition? Can an arbitrary partition be used, or at least a very fine arbitrarypartition? These questions are addressed in Sec. 6.4.6. At least the homoclinic tangencyconjecture is satisfied by the partition as shown in Fig. 6.31 is believed to be generating.As such, the resulting symbolic dynamics generated by “typical orbits" are believed to wellrepresent the dynamical system as approximated by the 4-bit representation in Eq. (6.69),but of course this is a coarse representation here for illustration only and in principle manysymbols would be used as far as the length of test orbit would support.88

As an aside, using link lists as an efficient structure in the case of symbolic dynamicswe need only record that a transition occurs or not which is boolean information. In contrastto a transition matrix, or to a stochastic matrix, link lists are useful memory structure andeasy for programming the systematic recording of transitions while scanning through asampled orbit whether it be for symbolic dynamcs or measurable dynamics applications.

This empirical discussion of symbolic dynamics is useful and straightforward in sofar as we have a useful partition. Otherwise irrespective of dimension other than the diffi-culty in filling the space with a good test orbit. A more analytic analysis to considerationsof allowable symbolic dynamics is as already addressed in the case of one dimensionalmaps by the kneading theory in the case of one-dimensional maps, [233], discussed above.A semi rigorous “pruning theory" can be found in [73, 154, 83] for higher dimensionalcases for a few special cases which includes the Henon map. It indicates a partial order ina symbol plane representation of the symbol space to indicate which words occur withina grammar from the dynamical system, much in analogy to the order based description ofthe symbolic grammar from the kneading theory.

6.4.6 Is a Generating Partition Really Necessary?

We studied in [43, 42] the detailed consequences of using an arbitrarily chosen partition,which is generally not generating. Our work was motivated by the many recent experimen-talists’ studies, who with measured time-series data in hand, but in the absence of a theoryleading to a known generating partition, simply choose a threshold crossing value of thetime-series to serve as a partition of the phase space.

On the experimental side, there appears an increasing interest in chaotic symbolicdynamics [79, 26]. A common practice is to apply the threshold-crossing method, i .e., todefine a rather arbitrary partition, so that distinct symbols can be defined from measuredtime series. There are two reasons for the popularity of the threshold-crossing method:

1. it is extremely difficult to locate the generating partition from chaotic data, and

2. threshold-crossing is a physically intuitive and natural idea.

Consider, for instance, a time series of temperature T (t) recorded from a turbulent flow. Byreplacing the real-valued data with symbolic data relative to some threshold Tc, say a {0}

88A long enough orbit should be used so that there is high probability relative to the invariant measure thateach symbolic bin of the n=bit representation will have an iterate in it so that the observed symbolic link listsuch as Eq. (6.69) as well each of the actual allowed transitions will be observed when it should be observedfor a long enough orbit.

Page 202: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

194 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

Figure 6.31. Henon Map Eq. (1.1) Symbolic Dynamics. (Left) The zig-zag dark“horizontal" piecewise linear curve near y = 0 is the generating partition constructeddirectly according to the popularly believed conjecture [73, 154] that the generating par-tition must past through points of homo/heteroclinic tangencies between stable and un-stable manifolds. Calling this dark curve C, two images and preimages are also shown,T i (C), i = −2,−1,0,1,2. Notice that this cannot be 4-bit representation cannot be com-plete by definition of Markov partition Defn. 4.4, considering as an example the 11.10“rectangle", since the attractor does not stretch all the way across the 11.10 region. Fur-ther images and preimages T i (C) have the possibility of closing all the conditional depen-dencies (conditional probabilities when associated with probability measure) to producea Markov partition, but no such completion is known for this Henon map with parametervalues (a,b) = (1.4,0.3). Thus the partition may be generating but does not produce aMarkov partition. Compare to Definition 4.4 of higher dimensional Markov partitions andFig. 4.5. (Right) The action of trajectories through the few symbols shown describes thedirected graph here.

if T (t) < Tc and a {1} if T (t) > Tc, the problem of data analysis can be simplified. A wellchosen partition is clearly important: for instance, Tc cannot be outside the range of T (t)because, otherwise, the symbolic sequence will be trivial and carry no information aboutthe underlying dynamics. Similarly, an arbitrary covering of a dynamical system with smallrectangles leads to a directed graph representation as a Markov model of the system, andcorrespondingly a symbolic dynamics as was already seen for the Henon map for examplein Fig. 1.1 with symbolic transitions show in Eq. (6.70). It is thus of paramount interest,from both the theoretical and experimental points of view, to understand how misplacedpartitions affect the goodness of the symbolic dynamics such as the amount of informationthat can be extracted from the data.

As a model problem, we chose to analyze the tent map

f : [0,1]→ [0,1], x �→ 1−2|x−1/2|, (6.71)

for which most of our proofs applied. Our numerical experiments indicated that the resultswere indicative of a much wider class of dynamical systems, including the Henon map, andexperimental data from a chemical reaction.

Page 203: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.4. Learning Symbolic Grammar in Practice 195

The tent map is a one-humped map, and it is known that the symbolic dynamicsindicated by the generating partition at xc = 1/2, by Eq. (6.8) gives the full 2-shift �2,on symbols {0,1}. The topological entropy of �

0,12

89 is ln(2) since it is a fullshift. Now

Figure 6.32. Tent map and a misplaced partition at x = p. [26]

misplace the partition at

p = xc+d , where d ∈ [−1/2,1/2], (6.72)

is the misplacement parameter. In this case, the symbolic sequence corresponding to apoint x ∈ [0,1] becomes:

φ = φ0.φ1φ2 . . . , where φi (x)= a(b) if f i (x) < p(> p), (6.73)

as shown in Fig. 6.32. The shift so obtained: �{a,b}2 , will no longer be a full shift because

not every binary symbolic sequence is possible. Thus, �{a,b}2 will be a subshift on two

symbols a and b when d �= 0 (p �= xc). The topological entropy of the subshift �{a,b}2 ,

denoted by hT (d), will typically be less than hT (0) = ln2. Numerically, hT (d) can becomputed by using the formula [279]:

hT (d)= limsupn→∞

ln Nn

n, (6.74)

where Nn ≤ 2n is the number of (a,b) binary sequences (words) of length n. In our compu-tation, we choose 1024 values of d uniformly in the interval [−1/2,1/2]. For each value of

89We will denote the symbol sequence space resulting from the generating partition to be �0,12 whereas

the symbol space resulting from misplaced partition to be �a,b2 , both being 2-bit symbol sequences so both

share the notation �2 and the superscript just reminds us of which shift space is being discussed.

Page 204: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

196 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

d , we count Nn in the range 4≤ n ≤ 18 from a trajectory of 220 points generated by the tentmap. The slopes of the plots of ln Nn versus n approximates hT . Fig. 6.33 shows hT (d) ver-sus d for the tent map, where we observe a complicated, devil’s staircase-like, but clearlynonmonotone behavior. For d = 0, we have hT (0)≈ ln2, as expected. For d =−1/2 (1/2),from Fig. 6.32, we see that the grammar forbids the letter a (b) and, hence, �{a,b}

2 (−1/2)

[�{a,b}2 (1/2)] has only one sequence: φ = b.bb (φ = a.aa). Hence, hT (±1/2)= 0.

Figure 6.33. For the tent map: numerically computed hT (d) function by follow-ing sequences of a chaotic orbit. [26]

Many of our techniques were somewhat combinatorial, relying on a simple idea thea dense set of misplacement values (in particular if d is “dyadic", of the form d = p/2n),allows us to study the related problem of counting distinctly colored paths through an ap-propriate graphic presentation of the shift, in which vertices have been relabeled accordingto where the misplacement occurs. See Fig. 6.34

One of our principal results was a theorem that the entropy can be a nonmonotone anddevil’s staircase-like function of the misplacement parameter. As such, the consequence ofa misplaced partition can be severe, including significantly reduced topological entropiesand a high degree of nonuniqueness. Of importance to the experimentalist who wishes tocharacterize a dynamical system by observation of a bit stream generated from the mea-sured time-series, we showed that interpreting any results obtained from threshold-crossingtype of analysis should be exercised with extreme caution.

Specifically, we proved that the splitting properties of a generating partition are lost ina severe way. We defined a point x to be p-undistinguished if there exists a point y �= x suchthat the p-named a−b word according to Eq. (6.73) does not distinguish the points, φ(x)=φ(y). We defined a point x to be uncountably p-undistinguished, if there exists uncountablemany such y. We proved a theorem in [42] that states that if p = q/2n �= 1/2, then the setof uncountably p-undistinguished initial conditions is dense in [0,1]. In other words, theinability of symbolic dynamics from the “bad" nongenerating partition to distinguish thedynamics of points is severe. We described the situation as being similar to that of trying tointerpret the the dynamical system by watching a projection of the true dynamical system.In this senerio, some “shadow," or projection of the points correspond to uncountably manysuspension points. In our studies, [43, 42] we also gave many further results both describingthe mechanism behind the indistinguishability, and further elucidating the problem.

Page 205: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.5. Stochasticity, Symbolic Dynamics and Finest Scale 197

Figure 6.34. Graphic presentation for the Bernoulli fullshift and some dyadicmisplacements. [26]

6.5 Stochasticity, Symbolic Dynamics and Finest ScaleIn the tables in Eq. 6.69 we used 4-bits to approximate the symbolic dynamics representingthe dynamical system and we suggested that more bits should be used since more is better.Is more always better? It is natural to ask what is the appropriate resolution scale to estimatea symbolic dynamics, and correspondingly any discrete coarse grained representation of adynamical system, wether it be the a topological perspective of symbol dynamics or ameasurable dynamics perspective of Ulam’s method.

In the presence of noise, there is a blurring effect of the randomized input whichmakes it unreasonable to discuss arbitrary precision of an evolution. Likewise an infinitesymbolic stream corresponding to an initial condition would imply infinite precision re-garding knowledge an infinite future itinerary. That is beyond measurement possibilityin the case of a stochastic influence, and similarly the story is the same when limited byany finite precision arithmetic representation on a computer where round-off error acts asa noise-like influence. The question of appropriate scale in the presence of noise was thesubject of the recent work by Lippolis and Cvitanovic [215] who ask (as entitled by theirpaper), “How well can one resolve the state space of a chaotic map?" which we review

Page 206: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

198 Chapter 6. The Topological Dynamics Perspective of Symbol Dynamics

briefly in this section.

Figure 6.35. Periodic orbits of the stochastically perturbed map Eq. (6.78) aresymbolized up to a period such that neighborhoods about them blur to the point that theyoverlap other periodic orbits which also blur in both forward and backward time histories,according to the Ôthe best possible of all partitions hypothesis. In red is shown f0 and f1branches of the deterministic map which is stochastically perturbed with Gaussian noise,D = 0.001 of the form, Eq. (6.76). Following the effect of noise on the points precedingpoints on periodic orbits by the backward operator Eq. (6.77), from intervals [xa−σa , xa+σa] gives overlapping regions as shown at the symbolization shown with the seven regionsshown. These according to the hypothesis is the optimal stochastic partition. [215]

In Lippolis and Cvitanovic [215], the stochastically perturbed dynamical system inthe form of an additive Gaussian noise term was discussed,

xn+1 = f (xn)+ ξn, (6.75)

for normal ξn with mean 0 and variance 2D.90

This gives a Frobenius-Perron operator whose form we have written Eq. (3.46) as isa special case of that found in [208] of the general multiplicative and additive stochasticform Eq. (3.44). In [215], a Gaussian stochastic kernel was chosen,

g(y− f (x))= 1√4πD

e−(y− f (x))2

4D . (6.76)

The smallest noise-resolvable state space partition along the trajectory of a point xa , wasdetermined by the effect of noise on the points preceding xa . This is achieved by the

90In [215] this stochastic form was called a discrete Langevin equation, whereas following [208], we callthis form a discrete time with constantly applied stochastic perturbation system. Likewise, its correspond-ing transfer operator was called a Fokker-Planck operator whereas we call it a stochastic Frobenius-Perronoperator.

Page 207: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

6.5. Stochasticity, Symbolic Dynamics and Finest Scale 199

backward transfer operator PFν which describes the density preceding the current state,ρn−1 = P†

Fνρn . The Gaussian form allows the form,

ρn−1(x)= P†Fν

(ρn(y))= 1√4πD

∫Mρn(y)e−

(y− f (x))2

4D dy. (6.77)

Following evolution of densities along orbits, and specifically along periodic orbitsallows the discussion of dispersion of our ability to measure the state with confidence. In[215] the following partition optimality condition was stated,

• “... the best possible of all partitions hypothesis, as an algorithm: assign to eachperiodic point xa a neighborhood of finite width [xa −σa , xa +σa]. Consider peri-odic orbits of increasing period np , and stop the process of refining the state spacepartition as soon as the adjacent neighborhoods overlap."

Here, σa was taken to correspond to variance from the Gaussian computed as, σa =2D

1−�−2p

( 1( f ′a)2 + ...+ 1

�2p), where �p = f

npa is the Floquet multiplier of the of an unstable

periodic orbit xa ∈ p of period np , |�p| > 1. In Fig. 6.35 we reprise the main result from[215], which was for an example of a stochastically perturbed one dimensional map,

f (x)=�0x(1− x)(1−bx), (6.78)

with parameters chosen, �0 = 8,b = 0.6 and branches of the deterministic map f0 andf1 are shown in red. Some further description of a D = 0.001 is given in Fig. 6.35 lead-ing to a seven element partition before further attempted resolution does not lead to moreknowledge but rather just ambiguity. This is a confidence based description of the Markovchain model where we have used the word confidence in analogy to a statistics confidenceinterval.

A major note made in [215] is that the stochastic symbolic dynamics coming fromthis hypothesis is a finite state Markov chain whereas the no noise version may be infinite.We further emphasize here that this formalism is equally applicable to a measurable dynam-ics Markov chain description of the dynamical system and correspondingly an Ulam-likemethod would be by a finite rank matrix computation.

We close this section by pointing the reader to a very nice alternate method of as-cribing meaning to a stochastic version of symbolic dynamics, which is due to Kennel andBuhl, [196] called symbolic false nearest neighbors which in details is quite different fromwhat is described here. However in spirit there is similarity. A statistic is introduced toaccept or reject a hypothesis of false near neighbors which should be respected by a goodsymbolic partition according to the notion of a generating partition. In this sense, the workreviewed in this section which demands no overlapping smeared symbolic neighborhoodsis comparable.

Page 208: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure
Page 209: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 7

Transport Mechanism, LobeDynamics, Flux Rates andEscape

7.1 Transport MechanismA good understanding of transport mechanism will inform us regarding global analysisof topological dynamics and symbolic dynamics as well as measurable dynamics questionssuch as escape rates. This understanding can eventually lead to control strategies to developdynamical systems to yield results as engineered and desirable.

7.1.1 Preliminaries of Basic Transport Mechanism

Any discussion of transport must include a definition of “inside" and “outside" and thus abarrier in between, relative to which the transport can be referenced. To understand trans-port, we restrict the discussion to continuous dynamical systems of orientable manifolds.91

The reason for this will become clear shortly. In addition, we will discuss orientation pre-serving maps, which are so defined, in terms of the tangent map DT |z ,

det(DT |z) > 0,∀z. (7.1)

We can also easily develop transport mechanisms for other types of maps, such as orienta-tion reversing maps (det(DT |z) < 0,∀z), as long as the property is consistent.

Two-dimensional transport is particularly well understood and will be the subject ofmost of this discussion. Furthermore the drawings used for our presentation will be muchclearer in the two dimensional setting. A two-dimensional discrete time map can, as usual,result from a flow on a three-dimensional manifold. Consider a Jordan curve C enclosing aregion A,92. We will investigate the relative orientation of forward and backward iterationsof these sets (see Fig. (7.1)). There are four basic cases of iterations that may result:

1. T (A)∩ A is empty.

91Two-dimensional, orientable manifolds include the plane, the sphere, and the torus, but not the Kleinbottle or the Mobius strip.

92The Jordan curve theorem is a basic theorem in topology that states that every Jordan curve (a non-selfintersecting loop) divides the plane into an "inside" and an "outside" meaning that any path connecting apoint of one region to a point of the other intersects that loop somewhere.

201

Page 210: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

202 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

2. T (A) is completely contained in A.

3. A is completely contained in T (A).

4. T (A)∩ A is nonempty and neither set is completely contained in the other.

Proceeding in a manner to emphasize an exploration of consequences of scenarios, we willemphasize the final case because it is typical of the “nice" barriers we will define in thenext section and descriptive of the horseshoe dynamics from the previous Chapter in 6.1.5.Consider for example that C is chosen so that there exists a fixed point z∗ on C . ThenT (C)∩C �= ∅. Similarly, a fixed point z∗ ∈ A is sufficient for a nontrivial intersection, butnot necessary.

Figure 7.1. a: Jordan curve C enclosing a region A. b: The first iterates of C andA, intersect C and A respectively. c: The region B = T (A)− T (A)∩ A contains all pointswhich will enter A on one application of the inverse map. d: The region Ex = T−1(B)contains all the points in A which will leave A upon one application of the map, and hencewill be called the “exit region." e: The “entrance region", En = T−1(A)− T−1(A)∩ A,contains all points which will enter A upon one iteration of the map. Compare this figurewhich discusses arbitrary enclosing curves C to the special case in Fig. 7.6 which uses acarefully chosen enclosing curve of stable and unstable manifold segments.

To define the subset of A that leaves A on one iteration of the map, consider thefirst iterate of the curve, T (C), enclosing T (A) as an illustrative for instance. Notice inparticular the region,

B = T (A)− T (A)∩ A, (7.2)

shaded in Fig. (7.1c). The set B contains all those points that left A after one iteration.Alternately, it is the set that will enter A in one iteration of the inverse map T−1. Thus,

Page 211: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.1. Transport Mechanism 203

B defines the entrance set of T−1. In this sense, we can say that the points in B cross thebarrier C , thinking of C as a barrier.

A tyical inverse iterate scenario of B is shown in Fig. (7.1d), as labeled,

Ex = T−1(B). (7.3)

Ex is the subset of A that will leave A on one iteration of T and hence is called the exitset. The only way for an orbit initially contained in A to leave A is for an iterate of theorbit to land in Ex . The map moves all the contents of Ex outside the closed curve on eachiteration.

Repeating this discussion in reverse, we may similarly construct the entrance set Enoutside of C , which is defined as T−1(A)−T−1(A)∩ A. It is the set outside of C which ismoved inside of C on each iteration. It is the only way in, across C .

In summary, the entrance and exit sets are defined

En = T−1(A)− T−1(A)∩ A,

Ex = T−1[T (A)− T (A)∩ A]. (7.4)

These definitions apply to all four of the intersection types listed. The fourth is shown inFig. (7.1d), but the other three are just as valid. In the first case, for example, if T (A) isdisjoint from A, then Ex = A. In the second case where T (A) is completely contained inA, we see that T (A)− T (A)∩ A= T (A)− T (A)= ∅, and therefore Ex = ∅.

Figure 7.2. a: A possible, more complicated iterate (and back-iterate) of theregion A. The implied entrance and exit regions in this example can intersect, which simplymeans that some subset of points entering A will immediately exit A on the next iterate. b:This configuration of T (C) is not possible because it violates assumptions of continuity. (Acontinuous function must map a connected set to a connected set.) c: This configuration ofT (C) overlaps itself, and so violates single valuedness of T−1(T (C)).

There are certainly more complicated configurations possible for T (A), relative toa general set A, than are implied by the previous figures. Some of these are indicatedby Fig. (7.2). Nonetheless, we can uniquely define En as the set that enters A in oneiteration, and Ex as the set that leaves A in one iteration. Eq. (7.4) defines the entrance and

Page 212: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

204 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

exit lobes with no limitations on the amount of folding possible. A configuration such asFig. (7.2a) presents no contradictions despite the overlap of T (C) and T−1(C); it simplyimplies that once leaving A, the subset T (Ex )∩ En ∩ A will immediately re-enter A onthe next iteration. Eq. (7.4) implies no statement regarding two iterations. Configurationssuch as in Figs. (7.2b) and (7.2c), which may present problems, are not possible due toviolations of assumptions of continuity and single valuedness respectively.

As a matter of discussion, we ask what it takes for a barrier to divide a manifold,even if the manifold may not be homeomorphic to the plane. The key property for dis-cussing transport is the topological partition already defined formally in Defn. 4.2. Thequestion of transport is to ask how and where, in each of the open sets defining the topolog-ical partition, points move from one element to the other, and this description is the sameregardless of dimensionality. A most interesting special case is when the elements of thetopological partition are connected open sets. Specifically in two dimensions, the Jordancurve case discussed above consists of a two element topological partition, labelled “in-side" and “outside" and the boundary curve is so named Jordan. Barriers are generally theboundary points of the topological partition. Of course the whole discussion can be carriedforward in more than two dimensions but we have limited the presentation for simplicityof discussion as the main idea becomes clear.

On certain manifolds, it is possible to describe transport across a barrier C which isnot a Jordan curve. The role which the closed curve serves in the above discussion is thatit divides the space in two - an inside and an outside. On the other hand, if a curve does notcompletely divide the space, transport can occur “across" the barrier by going around it (orby going the “other way" around the cylinder S1×R to avoid an infinite line “barrier" in thecase of a cylinder for example). Without becoming involved in an out of scope discussionin defining the genus of a manifold,93 and ways to partition such manifolds, consider theexample of a cylinder can be divided in two (a top part and bottom part) by a closed curve(a “belt" wrapped around the “waist"); See Fig. 7.3, and notice in particular in Fig. 7.3cthat C2 divides the top from the bottom.

The description of transport across any barrier is made by forward and backwarditerating the barrier, then finding the regions bounded by C and T−1(C) (or C and T (C))and asking “Which region crosses the barrier on the next iterate (back-iterate)?" We found“lobe-like" structures in Figs. (7.1)-(7.2) because we illustrated the situation where T (A)∩A �= ∅ and neither set is completely contained in the other. We will see this situation in thenext section where there will typically be a fixed point z∗ on C .

7.1.2 Chaotic Transport Mechanism and Lobe Dynamics

In the discussion of the previous section, we have described that questions of transport areonly relevant to a topological partition and its barrier, Defn. 4.2. In this section, we askthe related question, “What are the most natural barriers in chaotic transport?" (And thetopological partition whose closure includes these barriers). The arbitrarily chosen barriersin the previous section typically evolve and deform upon iteration. The situation is typicallyexponentially exaggerated with continued iterations, continuing this discussion in terms ofmaps for sake of specificity. A similar discussion could be stated for flows.

A natural barrier can be constructed of segments of stable and unstable manifolds on

93Genus is a notion from algebraic topology which in the case of two dimensional manifolds may beroughly described as the number of hoops attached to the manifold.

Page 213: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.1. Transport Mechanism 205

Figure 7.3. a: A Jordan Curve C in the plane has an inside and an outsideand any path connection from inside to outside must intersect the curve according to theJordan curve theorem. b: C is not a Jordan curve, and likewise there is a path around it.b: A cylinder S1×R has the possibility of a Jordan curve C1, or a “belt"-like curve C2that induces a partition between those points above and below. In the cylinder, C3 does notpermit a topological partition.

a homoclinic orbit. Given a period-n point z of a map T , recall (Section 2.3) the stablesubspace Ws (z), and Wu (z), the unstable subspace are defined as follows:

Ws (z)≡ {x : T jn(x)→ z as j →∞},Wu (z)≡ {x : T jn(x)→ z as j →−∞}. (7.5)

We repeat that a point z is defined to be hyperbolic when the tangent space at that point isdecomposable as the direct sum

M = Es (z)⊕ Eu(z), (7.6)

where Es (z) (or Eu(z)) is the linear subspace of the tangent space at z, spanned by theeigenvectors corresponding to eigenvalues with modulus strictly different than one.

Summarizing 2.3 some standard hyperbolicity results from the mathematical theory,under suitable hypothesis requiring regularity of the vector field and hyperbolicity follow-ing either expanding or contracting eigenvalues,

• The stable manifold theorem [279] implies that these eigenvectors can be continuedto the global stable (unstable) manifolds.

• The Hartman-Grobman theorem [262, 156] states that for a diffeomorphism T n anda small enough neighborhood N(z) of z, there is a homeomorphism between thedynamics of the linearized mapping DT n on its the corresponding tangent spaceEs(z)⊕ Eu(z) and T n |N(z) on U which is a neighborhood of z in the original phasespace. See Fig. 7.4(Right).

Page 214: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

206 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

• A saddle point, a special case of a hyperbolic point, is categorized by having all ofthe eigenvalues λi of the tangent map at z such that λi ∈ C, |λi | �= 1,∀i , using thecomplex modulus | · |.

Figure 7.4. (Left) A hyperbolic linear (linearized) mapping DT has a stablespace Es spanned by eigenvectors with eigenvalues of modulus less than one and an un-stable space Eu spanned by eigenvectors with eigenvalues of modulus less than one, heredrawn in the scenario that all eigevalues are real and positive, 0 < λs < 1 < λu sinceit is a saddle. Iterates of zn successively jump pointing progressively along Eu. (Right)The Grobman-Hartman theorem provides that a hyperbolic fixed point z∗ has the propertythat there is a neighborhood N(z∗) such that in that neighborhood, the linearized map isconjugate to the nonlinear map, and furthermore the corresponding stable and unstablemanifolds Ws (z∗) and Wu (z∗) become tangent to the manifolds of linearized map, Es andEu.

A hyperbolic saddle fixed point of a two-dimensional map is shown in Fig. 7.4. Itshould be stressed that in the case of a discrete time map, the smooth curves shown are notflows of a single point. Each point “jumps" upon application of the map to another locationon the curve, Fig. 7.4(Left). Continuity of T implies that a nearby point jumps nearby.94

Certain rules must be obeyed by such dynamics on such manifolds.

• By definition, these manifolds are invariant; a point on the stable (unstable) manifoldremains on the manifold.

• Single valuedness forbids that a stable (unstable) manifold intersects itself or thestable (unstable) manifold of another point.

94It may seem paradoxical that a chaotic dynamical system can nonetheless be continuous. Sensitivity toinitial conditions and exponential spreading of nearby points may seem to some to exclude continuity sinceit suggests that nearby points eventually map far away, but the key is the emphasis is on that word eventually.Continuity is a property of single applications of the map, and sensitivity to initial conditions describes theevolution of nearby points under many applications of the map.

Page 215: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.1. Transport Mechanism 207

• It is allowed, however, for the stable manifold to intersect the unstable manifold.

Recall from Definition 2.11

Definition 7.1. A point p on an intersection of Ws (zi ) and Wu (zj ) is called a homoclinicpoint if i = j or a heteroclinic point if i �= j .

Fig. 7.5 illustrates homoclinic points p of stable and unstable manifolds of a fixed point z.In fact, it is a transverse homoclinic point. By definition, the orbit of p accumulates on zi inforward time, and on z j in backward time. Thus, iterates of homoclinic (heteroclinic) pointsare homoclinic (heteroclinic) points: The existence of one intersection implies infinitelymany intersections.

Figure 7.5. About principle intersection points, p.i.p.s a: A transverse homo-clinic connection at point p, and a few of its iterates and pre-iterates. b: A single “lobe"between p and T ( p) causes an illegal (by assumption) orientation change from p, x, z toT ( p), T ( y), T (x). c: The “orientation of surface" (or “signed area") of the parallelo-gram described by the vectors p− y and p− x has opposite sign to that of parallelogramT ( p)− T ( y), T ( p)− T (x). d: Inserting one more transverse homoclinic point q yields alegally oriented image of p, x, y. e: Here we can see that the sign of the area of the nearbyparallelogram is preserved by T .

Fig. 7.5a shows a homoclinic orbit with transverse intersection.95 Also shown is partof the family of points corresponding to the orbit of p. The stable and unstable manifoldsmust intersect at each point in this family, but it is easy to show that there exists anotherhomoclinic point q between p and T ( p), following an assumption of orientation preser-vation. To see this, consider two arbitrary nearby points, x near p where x is on Wu (z)

95As one varies parameters of a parameter dependent dynamical system, zn+1 = Tλ(zn), the manifoldsW s (zi ) and W u(z j ) tend to move, and may intersect either transversally or tangentially. Tangent-type inter-sections are not structurally stable (stable to arbitrarily small perturbations to the mapping itself perhaps byvarying a parameter, versus the usual notion of stability which is a discussion concerning perturbation to theinitial condition), unlike the transverse-type which will be the subject of the discussion to follow.

Page 216: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

208 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

“farther" along Wu (z) before T ( p),96 and y, also near p, but on Ws (z) closer to z, butagain before T ( p). The relative configurations of x, y and p are drawn in Fig. (7.5b).Reading clockwise around, the order is p, x, and y. T (x) must still be farther along thanT ( p), and likewise so must T ( y) occur after T ( p). Again, reading clockwise around, weget T ( p), T ( y), and T (x), which is in violation of orientation preservation. We can seethis in Fig. (7.5c), where the area of the parallelogram, described by the vectors p− y andp− x, has opposite to the parallelogram T ( p)− T ( y), T ( p)− T (x). However, we can seein Fig. (7.5d) that inserting an additional transverse homoclinic intersection at q preservesthe orientation, shown in Fig. (7.5e). Hence, there must be at least one more homoclinicpoint q.

It is convenient to choose p to be what Wiggins [327] defines as a principle inter-section point (or p.i.p.). Any point on Ws (zi )∩Wu(z j ) is a heteroclinic (or homoclinic)point.

Definition 7.2. Using the ordering implicit along these invariant manifolds, we can definea principle intersection point (or p.i.p.) as a heteroclinic (homoclinic) point for which thestable manifold segment between zi and p has no previous intersections with the unstablemanifold segment between z j and p.

Figure 7.6. Definition of a homoclinic barrier in the orientation preserving case,a: Defining the barrier by initial segments of the stable and unstable manifold between thefixed point z and p.i.p. p. b: The iterate C lies largely on top of C, as much of the curvestretches over itself. c: The exit and entrance lobes Ex and En, which together are calledthe “turnstile". Compare to Fig. 7.1 where we use a carefully chosen enclosing curve interms of stable and unstable manifold segments.

These segments of the stable and unstable manifolds have also been called “initial seg-ments" [111].

In Fig. 7.6c, the shaded regions are labeled “Ex" and “En", describing their trans-port roles. These “lobes" have infinitely many (pre)images, whose end points are the(pre)images of p and q. We may now define a Jordan curve C using the unstable man-ifold initial segment between z and p, and the stable manifold initial segment between zand p, for any p.i.p. p97. See Fig. 7.6a. There is a well defined inside and outside, for this

96An ordering on W u(z) is possible since the invariant manifold is one-dimensional. A point is defined asfarther away from z than another parameterized in the sense of the arc length along the unstable manifold.An ordering on W s (z) can be similarly defined in terms of the arc length closeness to z.

97In fact, as long as p is a p.i.p., any of its iterates are just as legitimate, and the resulting entrance andexit lobes can be used to define transport.

Page 217: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.1. Transport Mechanism 209

barrier C . Eq. (7.4), defining transport across an arbitrary barrier, applies to this specialchoice of C .

The commonly chosen barrier C of principle segments of stable and unstable man-ifold segments illustrated in Fig. 7.6 is quite natural because orbits on the manifolds stayon the manifolds. Following the discussion in the previous section, in Fig. (7.6b) we drawT (C), and in Fig. (7.6c), we draw T−1(C). In terms of the original barrier C , we see thatthe shaded region En in Fig. (7.6c) iterates to the region T (En) inside C , which we easilysee by following the iterates if p, q, and the manifold segments in between.

Figure 7.7. Principle intersection points, p.i.p.’s are not unique. Note that bothp and q are p.i.p.’s but so are all the iterates orbits shown and reiterates as well. Thestretching along unstable manifolds, and compression along stable manifolds causes thestretching of the turnstiles, which under considerations of growth of measure eventuallycauses further non-p.i.p. intersections such as r and s (nontangential) as shown - an thenall the points on their orbits are homoclinic.

The only alteration in the overall form of C is the “growth" of the lobes En and Exupon application of T−1. In this sense the choice of the barrier C is “minimal." Mackey,Meiss, and Percival coined the term “turnstile" [219] to describe the two lobes En andEx in that they act like the rotating doors of the same name used in underground railroadtransportation, transporting area across C .

Remark: P.i.p.s are not unique. Iterates of p.i.p.’s are also p.i.p.’s. Both familiesof points, shown in Fig. 7.5d, are examples of p.i.p.’s. Starting with p.i.p.’s, non-principleintersection points arise from the stretching and folding typical with transverse heteroclinic(homoclinic) intersections. See Fig. 7.7. The resulting “tangle" quickly generates infinitelymany other families of heteroclinic (homoclinic) points which are not p.i.p.’s. More will besaid about the tangling process in the next section since the generation of infinitely manyfurther families results due to issues of measure evolution.

In Fig. 7.1.3, we present a homoclinic tangle generated by an area preserving Henon

Page 218: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

210 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

map; the stable and unstable manifold reveal the story so described. See further discussionof entrance and exit steps for this map in Sec. 7.2.

In closing remark for this subsection, we can describe transport as in some sense justan illusion which is clarified only by the choice of barrier. Studying Fig. (7.6), there isanother perspective on “transport" to be made. Forgetting our barrier C for a moment, letus focus on a point in the entrance set En “outside"98 of the manifold segment of Ws (zi )between p and q. The role of iterating the map is to cause that segment of Ws (z) to pushin (relative to C). Points outside that segment may be viewed as still outside. In thisperspective, there is no transport at all; it is just an illusion of the outside punching infurther and further upon iteration. This description only makes use of the stable manifold.Of course, only in terms of the full barrier C can we truly describe transport across thebarrier.

7.1.3 Lobes, The Homoclinic Tangle and Embedded Horseshoes

The discussion in the previous sections about turnstiles concerned the one-step action ofthe map relative to a barrier. In this section, we ask what is the long-term fate of

• The barrier C in forward and backward time,

• The points inside and outside of C ,

• What is the fate of the points in En and Ex ?

These are the fundamental questions that lead to understanding of chaotic transport acrossbarriers, and the notion of the homoclinic tangle. The answers will also lead us to findhorseshoes embedded within the tangle, the prototypical example of chaos.

Taking a simple two p.i.p. family generated by q and p above as a case study(E.g., again in Fig. 7.6), we see that the arc length between q and p along Ws (z), la-beled �(Ws

[q, p]), must eventually shrink upon repeated applications of the map, as the twopoints eventually accumulate at the fixed point. The arc length at time n is the line integralof the nth iterate of Ws

[q , p]. Likewise, the arc Wu[q , p] iterates with q and p. Hence, the

curve Wu[q, p]∪Ws

[q , p] is a dynamically varying boundary of T (En).Through the recent sections, the symbolic dynamical descriptions including of the

topological horseshoe have been entirely topological, meaning no measure structure wasneeded or assumed. Only action of the mapping on the structure of the open sets wasassumed, definitive of the point set topology. To quantify the fate of lobes, we will nowresort to measure and distance.

Assume,T : X → X , occurs on a measure space (μ, X ,A), (7.7)

on measurable sets A, and there is a distance function,

d : X →R+. (7.8)

For narrative description, we will refer to a popular example, the area preserving mapswhich arise commonly when studying Hamiltonian systems [219, 229]. If the measure μ

is descriptive of area,Area(A)= μ(A), (7.9)

98Of course, a curve segment is not enough to define a barrier; a fully closed curve is required.

Page 219: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.1. Transport Mechanism 211

for each measurable A ∈A, then the map is called area preserving if,

μ(T (A))= μ(A), for each A ∈A. (7.10)

Each A may deform under T , but its area always remains the same. If |λu | > 1 is theexpaning exponent along unstable manifolds, and likewise, 0 < |λs |< 1 is the contractingexponent, then in the case of area preserving maps, it quickly follows that,

λu = λ−1s . (7.11)

Specializing to the lobes En and Ex , specifically T n(En) may stretch exponentially andtransversally contract, resulting in a long and narrow lobe for large n. See Fig. 7.7. Like-wise, T n(Ex ) becomes narrow and long. To make these statements rigorous, we wouldresort to the Lambda Lemma from hyperbolicity theory [279], but we will continue withour narrative description for now. It is half of the story that stretching is one of the maincomponents that can result in chaos, the other part of the story being the folding;99 thisstory is revealed in full by the horseshoe example.

Remark: In the area preserving case, it is easy to see that the finite area of region A,bounded by C, cannot completely contain all future iterates of En. There is a time m when

m∑i=1

μ(T i (En))≥ μ(A), (7.12)

i.e., the first time that,

m ≥ μ(A)

μ(En). (7.13)

In terms of transport, some of the points in En which enter A must leave A by the mth

iterate, implying that there exists an,

l ≤ m such that T l (En)∩ Ex �= ∅. (7.14)

This follows since the only way out is through an exit lobe and we already concluded byEqs. (7.12)-(7.13) that some points coming in by En must soon leave. Almost all of thepoints100 must eventually leave. Once this intersection occurs, a new family of homoclinicpoints is implied. See points r and s in Fig. 7.7. Considering the history of the lobe Ex ,which also becomes long and narrow (as n→∞), we see that a homoclinic point is impliedeach of the times m and −n that,

T m(En)∩T−n(Ex ) �= ∅. (7.15)

The segment Wu[q , p] accumulates at z as n→−∞, and Ws

[q , p] accumulates at z as n→∞. Of course, a “new" family of intersections implies infinitely more intersections as thehomoclinic point iterates in forward and backward time. And so forth further families. Thisis the “homoclinic tangle." �

As presented in Sec. 6.1.5 and Figs. 6.6-6.14, the horseshoe construction implies a setwhich remains trapped inside the region A for all time. S. Smale invented the topologicalhorseshoe, and also showed its relevance to applied dynamical systems, [306],

99A popularly described (necessary) “recipe" for chaos is “stretch+fold yields chaos".100“Almost all" and “almost every" are stated in the usual measure theoretic terms, meaning all but a set of

measure zero.

Page 220: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

212 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Theorem 7.3. (Smale, [306], see also [327]) A diffeomorphism T with a transverse homo-clinic point p, has a time m > 0 such that the composition map T m has an invariant Cantorset � ∈ A. Further, there exists a conjugacy

h : �→�2 (7.16)

such thath ◦ T m |� = α ◦ h. (7.17)

The conjugacy is with the dynamics of the Bernoulli shift map α on the space of bi-infinitesequences of countably many symbols σi . In the simple horseshoe, we let

σi = 0 or 1. (7.18)

This theorem tells us that a horseshoe and all of its implied complex and chaoticdynamics is relevant to a standard setting due to a transverse homoclinic point, a geometricconfiguration of stable and unstable manifolds that occurs even in real world and physicalsystems; see Fig. 7.8. The horseshoe may be constructed for Fig. (7.6) by drawing a thincurved strip S over Ws

[z, p] as shown in Fig. (7.8). As p iterates closer to z, it drags thestrip with it. Meanwhile, the point s, defined as the intersection Wu

[z, p]∩ S, marches awayfrom z upon repeated application of T . Define m as the first time that T m(s) is ordered onthe arc segment after p. By time m, the short side of the strip has stretched and folded overto the strip T m (S) along Wu

[z, p] which intersects S by construction. Here again we see thestretch and fold, which can be thought of as the ingredients for chaos.

Figure 7.8. Constructing a horseshoe on a homoclinic orbit, as per Theorem 7.16.The strip S contracts along the stable manifold, and expands along the unstable manifoldto the shorter, wider strip T (S). By the mth iterate, the point T (s) has passed p; the longand short sides of the strip T m (S) are reversed from the long and short sides of the originalstrip S. The invariant sets H0 and H1 are the first steps in generating the invariant Cantorset �.

DefineH0 = T m(S)∩ S (7.19)

at p andH1 = T m(S)∩ S (7.20)

Page 221: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.1. Transport Mechanism 213

at z, and defineH = H0∪H1. (7.21)

We see that the invariant set of T m is contained in T m(H )∩ H , which defines two verti-cal strips in H0, and the two vertical strips in H1. Similarly, the invariant set of T−m isT−m(H )∩H , which forms two horizontal strips in H0 and H1. Define

�=∞⋂

i=−∞T im (H ). (7.22)

� is the invariant set of the horseshoe,101 which we see is the Cartesian product of twoCantor sets, one in forward and one in backward time. For a thin enough strip S, theinvariant set is hyperbolic [279, 218].

The “address" of a point in “H" can be labeled “.0" if it is in H0 or “.1" if it is in H1.On iteration, the point (say it is .0) lands in either H0 or H1, and hence is labeled “.00" or“.01", that defines which vertical strip in H0 contains the point. Similarly, the address tothe left of the decimal determines in which square the point lands, H0 or H1. From thispoint, the theory of the topological horseshoe presented in Sec. 6.1.5, requires further rigorto prove that the representation is correct [281].

Horseshoes have been explicity constructed for a number of examples, separatelyfrom this Smale theorem that discusses transverse intersection of stable and unstable man-ifolds. As revealed in Sec. 6.2.2 it is possible to construct a horseshoe explicitly for theHenon map (Figs. 6.20 and 6.21), or for the standard map when k > 2π , (See [218]), andthe Poincare’ mapping from the Duffing equations (Fig. 6.2.3) in Sec. 6.2.3, to name afew. Horseshoes can also be constructed for heteroclinic cycles. As discussed further inSec. 6.4, other “grammars" besides the simple left shift102, on many symbols can also beuseful in the more general scenario of many folding and partially folded systems .

In terms of transport, the horseshoe serves only as an incomplete model of the dy-namical system, for two reasons:

1. Typically, one may be interested in the transport of more than a measure zero set ofpoints. The horseshoe set � is a topological Cantor set,103 and one of zero measure;

2. A serous deficiency for transport study is that the horseshoe models those points in-variant to the horseshoe, i.e., those points which never transport out of the horseshoeset. Transport within � is completely described by the horseshoe model, but no moreis said by the analysis of the action of broader map on the rest of its phase space.

The measure zero property is easy to show for the middle thirds Cantor set. Inspect-ing Fig. 7.10, since,

C∞ = ∩n→∞Cn , (7.23)

101The Smale horseshoe is so named because the horseshoe is constructed by stretching and folding asquare into horseshoe shapes (again and again and...). The process is perhaps more akin to building aJapanese Samurai sword whose building process includes thousands of stretches and folds.

102The Bernoulli left shift grammar on two symbols can be described by the directed graph 0↔ 1 whichis equivalent to the 2× 2 identity transition matrix. Other grammars on n symbols have directed graphsdescribable by more general n×n identity matrices.

103A generalized Cantor set is defined to be compact, nonempty, perfect (has no isolated points) and totallydisconnected and it will be uncountable. See Fig. 7.10 for a illustration of the famous middle thirds Cantorset which will remind us of the invariant sets of a horseshoe.

Page 222: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

214 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Figure 7.9. A “typical” homoclinic tangle. Compare to Figs. 7.11, 7.12 from thearea preserving Henon map Eq. (7.27).

then,

l(C0)= 1, l(C1)= 2

3l(C0), l(C2)= 2

3l(C1), (7.24)

and so,

l(C∞)= limn→∞

2

3

n

l(C0)= 0. (7.25)

A similar reasoning applies to the invariant set of the horseshoe as depicted in Figs. 6.6-6.14, and it has similar measure properties.

Given a complicated transport problem, say from near some point a to near a pointb, where only a long, convoluted heteroclinic connection may exist, one may be successfulin finding a complex grammatical rule on a long list of symbols, if a and b happen to bein some invariant set of the dynamics. But, in general, only heteroclinic cycles are homeo-morphic to the horseshoe, and hence have a reasonably easy-to-find symbol dynamics. Asmore of the phase space is to be described, then more complicated symbolic dynamics maybe required for each fattening of the corresponding unstable chaotic saddle Cantor set.

Page 223: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.2. Markov Model Dynamics for Lobe Dynamics, A Henon Map Example 215

Figure 7.10. The Middle Thirds Cantor Set C∞ = ∩n→∞Cn is useful modelset since invariant sets in chaotic dynamical systems, and in particular of horseshoes, aregeneralized Cantor sets, sharing many of the topoligcal properties as well as many of themetric properties such as self similarity and fractal nature.

7.2 Markov Model Dynamics for Lobe Dynamics, AHenon Map Example

Traditionally, following the Smale Horseshoe example, symbol dynamics has been incor-porated into dynamical systems theory in order to classify and investigate the behavior ofdynamics on invariant sets. As we have pointed out, these invariant sets are often of mea-sure zero. There is often a great deal of other behaviors that the dynamical system is capableof demonstrating. In particular, we will give example here of how symbolic dynamics canbe used to characterize transport behavior, and lobe dynamics. Such modern approachesare presented for more sophisticated and complete examples including homoclinic and her-toclinic transport, [229, 109, 227, 328, 330] and in particular symbolic characterization oflobe dynamics in [66] and [239].

7.2.1 An Example Area Preserving Henon Map

Take the area preserving Henon map [172],

zi+1 = T (zi )=(

xi cosα− yi sinα+ x2i sinα

xi sinα+ yi cosα− x2i cosα

)= (7.26)

=(

cosα −sinαsinα cosα

)(xi

yi − x2i

),

where α is a parameter, and as usual, we will consider this as an mapping of the plane,T : R2→R2, which is again a diffeomorphism. Furthermore, this is a special example of a

Page 224: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

216 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

quadratic map in that it is an area preserving104. Hamiltonian system105. We may interpretthis mapping according to Moser [244] as the composition of a rotation and a shear. Thismap is perhaps the prototypical example for its conjugacy to the horseshoe, whose stretchand fold dynamics are visibly apparent and proven in [218].

This map is in a form which is particularly elegant to invert. In Fig. (7.11c), we seethe orbits of several initial conditions under the influence of the altered Henon dynamics

zi+1 =(

T (S(zi )) if zi ∈ ExT (zi ) otherwise

). (7.27)

An interesting consequence of the symmetric nature of the dynamics, S : y = x ,106

as an aside, let us recall the definition (see [218]) that a map T has a symmetry S if S is anorientation reversing involution such that

S2 = (T S)2 = I . (7.28)

It follows thatT−1 = ST S−1. (7.29)

Those points which eventually escape the barrier outlined in Figs. 7.11-7.12 underthe Henon map Eq. (7.27). As shown in Fig. 7.11c, these are now periodic under Eq. 7.27.All those points which are forever bounded inside the barrier, including the quasi-periodicinvariant circles,107. the regions they bound, and other chaotic orbits including those whichare part of the horseshoe invariant set � which we know is embedded since we see atransverse homoclinic point at p as well as at q.

7.2.2 The Lobe Dynamics

The “typical" homoclinic tangle shown in Fig. 7.1.3 is from the area preserving Henon map,Eq. (7.27). A detailed barrier illustration is shown in Fig. (7.12) for the lobe dynamics ina manner analogous to Fig. 7.6, with the coloring and details regarding entrance and exitlobes, En and Ex . The particular details of this map, with parameters as chosen, reveal thatEx ∩T 5(En) is the first such intersection. This information yields a Markov model of theescape dynamics which may be symbolized by the graph(s) shown in Fig. 7.15.

Once the the “inside" and “outside" (Fig. 7.13) are defined, we have lobe dynamics asfollows. As shown in Fig. 7.12, the unstable manifold segments of Wu (z) from z to p.i.p.p and Ws (z) from z to p define a boundary curve C . This is a particularly convenientchoice of boundary curve to discuss the transport mechanism. The entrance set, En couldbe called in standard lobe dynamics notation,

L1,2 = {z|z ∈ R1, T (z) ∈ R2}, (7.30)

104Each Lebesgue measurable set A has the property that area(A) = area(T (A))105Formerly there is a Hamiltonian from which the dynamics follows, but we shall take area preserving

Hamiltonian here to be more loosely descriptive of “not dissipative".106If we fold the paper through the line y = x , we have still captured the whole orbit.107There is a interesting and deep story regarding KAM theory, invariant tori, and resonance pertaining to

this and many other examples of area preserving maps, but this story leads us too far away from the story ofthis book. See [229].

Page 225: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.2. Markov Model Dynamics for Lobe Dynamics, A Henon Map Example 217

Figure 7.11. Henon map turnstile and invariant sets. Compare to Figs. 7.1.3,7.12. a: The homoclinic tangle of the Henon map. Shown are the fixed point z, the twop.i.p.’s p and q, and the exit lobe Ex . b: The exit lobe Ex of the Henon map. c: Controlleddynamics. We reflect a point zi through the symmetry S : y = x into the entrance lobewhenever it enters the exit lobe. The area preserving Henon map is written Eq. (7.27).

where in the Fig. 7.13 we will name R1 the set of all points outside the curve C , and R2will be the set of those points inside C . Likewise, the exit lobe Ex can be denoted,

L2,1 = {z|z ∈ R2, T (z) ∈ R1}. (7.31)

In this notation for the lobe dynamics, the discussion of transport, flux and escape is all interms of the orbits of these lobes as sets,

{T i (L1,2)}∞i=−∞, and {T i (L2,1)}∞i=−∞. (7.32)

Especially interesting to the complex behaviors becomes the various higher order intersec-tions which may occur,

any i , j , such that T i (L1,2)∩T j (L2,1) �= ∅. (7.33)

Remembering that orbits are indexed infinitely in forward and backward time, we must ac-knowledge that if there is one such overlap, there must be infinitely many such intersections,at least in this family of intersections. In such case, it is possible that there is a set of points

Page 226: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

218 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Figure 7.12. Transport in a “typical" homoclinic tangle as was seen in Fig. 7.1.3related to discussion of horseshoe invariant sets is now illustrated to discuss lobe dynamicsand escape. The partition associated an “inside" and “outside" relative to the entranceand exit lobes corresponding to iterates and preiterates of the partitioned region. Hereshown are is a particularly convenient partition corresponding to segments of stable andunstable manifolds and the implied entrance and exit lobes labelled En and Ex . Compareto Figs. 7.11 and 7.13. The map used here as case example is the area preserving Henonmap in Eq. (7.27), with a fixed point z and p.i.p. homoclinic point p shown in Fig. 7.11.

whose fate is to enter the region R2 for some sojourn time. Then such points exit aftersome time, and then re-enter, repeating indefinitely and forever. Such scenario is anotherview of the homoclinc tangle which can be said to have been befuddlng to founders of thefield.108. By previous discussion regarding the possibility and likelihood of infinitely many

108Henri Poincare’ who was the founding father of the field of dynamical systems, understood hints of thehomoclinic tangle in his work, but was not able to draw this complicated structure [270], “If one seeks tovisualize the pattern formed by these two curves and their finite number of intersections, each correspondingto a doubly symptotic solution, these intersections for a kind of lattice-work, a weave, a chain-link network

Page 227: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.2. Markov Model Dynamics for Lobe Dynamics, A Henon Map Example 219

Figure 7.13. Entrance and exit lobes in the area preserving Henon map,Eq. (7.27). The darker green set together with the light green lobe labelled Ex whichwe call C the interior of the barrier C described by the stable and unstable manifold seg-ments Ws (z) and Wu (z) from z, meeting at p.i.p.’s p. The entrance set En is outside thisbarrier, but iterates to T (En) ⊂ C. Likewise Ex starts inside C, but leaves in one iterate.Compare to Fig. 7.6. Iterates of these lobes allow us to follow the transport activity of thismap relative to the green barrier. These are outlined by stable and unstable manifolds, andwe see that since Ex ∩T 5(En), then some points will escape the region after 5 iterates, butsome will remain. See a Markov model of the same in Fig. 7.15.

such families of intersections, one should expect the lobe dynamics to allow for pointswhich enter and exit the region in infinitely many different patterns, meaning infinitelymany different i , j pairings denoted Eq. (7.33). Not only are such complicated behaviorsexpected, under mild sufficient assumptions such as area preserving maps, we can provethe existence of this scenario as per Eqs. (7.12)-(7.14).109

The orbits of the L1,2 entrance lobe and the L2.1 exit lobe are illustrated as coloredblue and red respectively in Fig. 7.14. The discussion of transport rates then becomes amatter of considering the relative measure of these sets, the rate of change of their measure

of infinitely fine mesh; each of the two curves can never cross itself, but it must fold back on itself in a verycomplicated way so as to recross all the chain-links an infinite number of times. One will be struck by thecomplexity of this figure, which I am not even attempting to draw."

109A lobe has area and therefore, an infinite forward orbit of such a lobe cannot remain in a bounded regionforever at least for an area preserving map.

Page 228: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

220 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

with respect to iteration and the measure of nontrivial overlaps. See in particular the specialcase of resonance overlap in Sec. 7.3.

Figure 7.14. (Upper left) Entrance (blue) and exit (red) lobes labelled En andT (Ex ) in Fig. 7.11 are colored. By lobe dynamics notation in Eqs. (7.30)-(7.31), L1,2is colored blue in the lower left image, and L2.1 is colored red in the upper left figure.Iterations of these regions by the Henon map T are shown along the upper row where theaction of the map pushes those initial conditions inside the colored red set Ex . Compare tothe dark green colored region from Fig. 7.13. Those points inside the blue colored entranceset En are shown to enter the dark green colored region shown in Fig. 7.13 by the action ofT as seen here along the top row. Along the bottom row is shown the action of the inverseof the map T−1 successively on the sets En and Ex . Compare to Figs. 7.11-7.13

Markov Model Dynamics and the Symbol Dynamics of Lobes

The labeling of lobes, such as the example Eqs. (7.30)-(7.31), can be considered to imply aMarkov model and likewise generating a symbolic dynamics. Markov modelling of escapedynamics is a type of symbol dynamics where here the symbolization is relative to insideand outside a barrier. These states could just as well serve as a symbol set. Including

Page 229: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.3. On Lobe Dynamics of Resonance Overlap 221

Figure 7.15. The action of ensembles of initial conditions relative to the barrierC discussed in Fig. 7.12 can be summarized by the top graph, which is a Markov modeldescribing the first intersection, Ex ∩T 5(En). Summarizing further the quotient that if weare only interested that inside is inside, then a quotient vertex C gives the bottom graph.Implications for escape rate of measure is discussed in Eq. (7.12).

a symbol for escape to ∞ allows us the description shown Fig. 7.15. This leads to thediscussion of transport rates, flux, and resonance overlap who lobe dynamics are discussedin the next section, leading to interesting escape rate sets.

7.3 On Lobe Dynamics of Resonance OverlapThe lobe dynamics defined by Eqs. (7.30)-(7.31) can be readily generalized to discuss morecomplicated geometries of homoclinic and heteroclinic tangles corresponding to partitionswith many elements. We draw here from elements which can be found in Wiggins [328],and Meiss [219, 229] amongst other sources. Notably some of the more complicated sce-narios are discussed most recently in Mitchell [241, 239], including an enticing title, “Thetopology of nested homoclinic and heteroclinic tangles".

For sake of presentation, in Fig. 7.16 we illustrate a particular scenario of resonanceoverlap due to a pair of heteroclinic orbits of fixed points z1, z′1 and z2, z′2 respectively. Werepeat that the key to any transport study starts with the definition of a partition relative towhich the transport may be discussed. Now we need pairs of open sets, one pair for eachbarrier to be crossed. The complete barrier will be the meets,∧, of these four sets. Considerthe open regions 1,c

1 and also 2,c2 by segments of stable and unstable manifolds as

defined further in the figure caption of Fig. 7.16. In this example drawing, the curves shownto not form a Jordan curve to partition the space. In order to permit the partition as outlinedby these curves of stable and unstable manifold segments, we should allow that either thedynamics is on a cylinder, or otherwise the region not shown outside the figure also acts todynamically delineate the space.

The dynamics of entrance and exit lobes, is stated here in the same format as Eqs. (7.30)-(7.31),

L1,1c = {z|z ∈1, T (z) ∈c1}, (7.34)

and likewise,L1c ,1 = {z|z ∈c

1, T (z) ∈1}, (7.35)

and similarly so is defined the lobes L2,2c and L2,2c . These lobes, and the dynamics of theirfirst few iterates and pre-iterates are shown in Figs. 7.17-7.18. The most interesting feature

Page 230: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

222 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

in discussing transport between resonance layers is the overlap of these lobes. When itoccurs that,

T i (L1,1c)∩T j (L2c ,2) �= ∅, for some i , j , (7.36)

interesting orbits, and particularly interesting heteroclinic orbits occur. The example of thefate of this resonance overlap is illustrated in Fig. 7.19.

The orbit of these sets and their intersections explains the geometric description oftransport. Just as shown in Fig. 7.15, one can pursue a symbolic dynamics description oftransport following a Markov model. Some detailed topological discussion of lobe dynam-ics and symbolic dynamics can be found in [241, 239]. An elegant special case exampleof transport in a celestial mechanics system described in terms of symbolic dynamics canbe found in [202]. Discussion of the measure theoretic description of these evolving lobesaddressed questions of transport rates and flux across barriers to be discussed in the nextsection.

Figure 7.16. A “typical" resonance overlap scenario leading to transport andescape. A heteroclinic orbit of fixed points z1 and z′1 with stable manifold Ws (z1) andunstable manifold Wu (z′1) respectively shown. (Wu (z1) and Ws (z′1) not shown. Only prin-ciple segments from the fixed points to the principle intersection point, p.i.p., p1. Assumedynamics on a cylinder, so z1 = z′1 or z2 = z′2. Denote 1 and ′1 the two regions of thetopological partition such that 1 and c

1 form a open topological partition of the region.Likewise 2 and c

2 formed by the heteroclinic connection of segments of Wu (z2) andWs (z′2), to the p.i.p. p2, defined by the boundary curve of the open partition elements 2and ′2. See Figs. 7.17-7.19 for more narration of this lobe dynamics.

7.4 Transport RatesIn the previous section we discussed the transport in dynamical systems geometrically interms of watching the orbits of special sets called lobes and turnstiles consisting of stableand unstable manifold segments. We discussed how stable and unstable manifold segmentsbound partition elements across which transport can be discussed. Following sets alone isa geometric discussion of transport. On the other hand questions of flux, expected escape

Page 231: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 223

Figure 7.17. Lobes dynamics continuing the general resonance overlap scenarioof Fig. 7.16, as defined in Eqs. (7.34)-(7.36) . Here the lobe L1c ,1 consisting of those pointswhich cross the heteroclinic orbit segment curve defined of p.i.p. segments, crossing from1c to 1 is shown in blue. Also shown are love iterates and pre-iterates T i (L1c ,1), i =−2,−1,0,1. Similarly, the lobe dynamics of L2c ,2 and iterates for crossing from 2c to 2are shown. Contrast to lobe dynamics in Fig. 7.18 for crossing these pseudo-barriers inthe opposite direction.

time and so forth are inherently quantitative questions requiring a measure structure. Wesuggest in this section that these questions are a natural place to adjust our already usefultransfer operator methods. Particularly for area preserving maps, there has been detailedstudy of transport rates and flux by careful study of the stable and unstable manifold struc-tures and lobe dynamics [219] for example, and through resulting symbolic dynamics in[282, 331] by the so-called trellis structures of Easton [112, 110] descriptive of the topol-ogy of the homoclinic tangle, beginning with lobe areas calculated from action integralssuch as in [204]. See also by a Markov tree model in mixed resonance systems [257], andthe interesting discussion of area as a devil’s staircase, [57]. Rather we will focus here onuniform arbitrary fine partitions in the spirit of the Ulam’s method by use of the Frobenius-Perron operator methods which allow us to attack more general systems without the specialanalytic setup required for Melnikov’s method and action based methods. Questions suchas transport rates and flux, and high activity transport barriers have useful realizations inthe underlying Markov chain description implicit in the transfer operator approximated bythe graph theory discussion.

Page 232: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

224 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Figure 7.18. Lobes dynamics continuing the general resonance overlap scenarioof Fig. 7.16, and here shown for transport crossing in opposite directions relative to thetransport already shown in Fig. 7.17. L1,1c consists of those points which iterate from 1to 1c , and likewise the iterates of this lobe are shown. See Eqs. (7.34)-(7.36). See alsoshown the lobe dynamics of L2,2c .

7.4.1 Flux Across Barriers

When investigating transport, flux across barriers is the natural quantity to calculate. It isnatural to ask how flux may vary as system parameters vary; how does a barrier give wayto a partial barrier as an invariant set gives way to an almost invariant set? Our discussionof flux will be in terms of an ensemble density profile ρ. The idea is to measure how muchrelative quantity of the ensemble is moved from one identified region to another. We willmake this discussion in terms of transfer operators.

Discussion of transport only makes sense by first assuming a partition relative towhich the transport is described. Given a region, say an open set S = Si to be one of theelements of a topological partition {Si } of the phase space, it is possible to define massflux (or simply flux) in and out of the region, across the boundary ∂S of the set beingdiscussed[35]. Let,

F ±S [ρ]≡mass entering/exiting S from outside/inside S upon one (7.37)

application of the map, due to initial density profile ρ

where F ±S [ρ] can be read, “mass flux into/out-of S due to an initial density profile ρ." Tocalculate F +S [ρ], we appropriately restrict the region of integration of the transfer operatorwhere here we have assumed to be the form of the stochastic Frobenius-Perron operator for

Page 233: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 225

Figure 7.19. The resonance overlap scenario marked in green region, when lobesintersect, T i (L1,1c)∩T j (L2c ,2) �= ∅, for some i , j , from Eq. (7.36).

sake of discussion,

F +S [ρ]=∫

S

∫Sν(x− F(y))ρ(y)dydx . (7.38)

The symbol, S denotes the complement of the set S in order for discussion of transportaway from S. Observe that the inner integral,∫

Sν(x− F(y))ρ(y)dy,

gives the density at x which comes from a point y ∈ S, meaning not in S. The outerintegral accounts for the total mass of all such contributions in x . To calculate the flux intoS, F −S [ρ], we must simply reverse the regions of integration in Eq. (7.38). The obviousidentities follow immediately,

F +S [ρ]= F −S

[ρ], and F −S [ρ]= F +S

[ρ], (7.39)

due to conservation of mass.One has the option to calculate F ±S [ρ] by direct (numerical) application of Eq. (7.38).

On the other hand, projection onto basis elements corresponding to a fine grid, and corre-sponding integration quadrature, leads to matrix computations akin to the Ulam’s methoddiscussed in many other places in this book.

Page 234: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

226 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

The inner integral becomes,∫Sν(x− F(y))ρ(y)dy)

∫Sν(x− F(y))

N∑i=1

ciχBi (y)dy

=∑

i:Bi∩S �=∅ci

∫Bi

ν(x− F(y))dy. (7.40)

Substitution into Eq. (7.38) gives the approximation, Recognize that this last double inte-gral consists of a sum over those entries of a stochastic transition matrix Ai, j representingthe approximate full Frobenius-Perron operator such that Bi ∈ S, and Bj ∈ S. Hence, wedefine a flux matrix,

A+S ≡{

Ai, j if Bj ∈ S and Bi ∈ S,0 otherwise

}, (7.41)

which allows us to rewrite Eq. (7.39),

F +S [ρ]) ‖A+S · c‖1, (7.42)

where ‖.‖1 is defied by the absolute sum. In this form, we have flux into S terms of A+S ,which is a masked version of the full transfer matrix A times the coefficient weights vector,

c= (c1,c2, ...,cN )t , (7.43)

representing estimation of the density ρ on the grid. One can similarly form and interpretmasked transfer matrices A−S , A+

S, and A−

Srepresenting flux out of S, into S and out of S

respectively. Correspondingly the conservation statements Eq. (7.39) hold.Taking as a special choice the initial uniform density,

c= 1

N1, (7.44)

we can find stochastic “area flux." In this case, Eq. (7.42) reduces to the absolute sum ofmatrix entries,

F +S [1]=∑i, j

[A+S ]i, j . (7.45)

We highlight these quantities by the Example 7.1 and illustrated in Fig. 7.21. The partsof A±S , A±

Sare visible of each block shown. There also follows a discussion a stochastic

version of completing a heteroclinic connection in Example 7.1.In the previous section, the discussion of transport mechanism was in terms of lobe

dynamics. A lobe defined in terms of segments of stable and unstable manifolds leads to themechanism corresponding to transport across barriers defined also usually by segments ofstable and unstable manifolds. From these barriers we also saw there are partition elements.The discussion of this section on the other hand discusses transport in terms of transferoperators masked according to an a priori choice of partition. Both of these perspectivesare complementary. On the one hand we could simply choose S to be the “insides" of theregion defined by the homoclinic orbit labelled green in the Henon example in Fig. 7.13 forexample. As such, using an approximate transfer operator approach, we would expect theoff diagonal parts of A±S , A±

Sto correspond to the lobe dynamics illustrated geometrically

in Fig. 7.14 continuing with the example.

Page 235: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 227

7.4.2 Expected Escape Time

Another natural quantity to consider when investigating transport is expected escape timeacross a barrier, of the orbit starting from an initial condition from a point in a region.Again, suppose a region S as an element from a topological partition. Given a point x ∈Bi ⊂ S, let T (x)i be the actual time of escape for a particular sample path of the stochasticdynamical system of a randomly chosen initial condition in the i th box Bi . The expectedtime of escape is,

< T (x)i >=∞∑

n=1

n P(T (x)= n)=∞∑

i=1

n P(Fn(x) /∈ S and Fm(x) ∈ S, ∀m < n). (7.46)

The key issue to acknowledge is that we are interested in the mean time of firstescape. While,

Ani, j = P(Fn(x) ∈ Bj |x ∈ Bi ), (7.47)

this probability does not forbid multiple passages or recurrences. In particular, it accountsfor orbits which might leave S and return to S multiple times before finally landing inBj ⊂ S on the nth iterate.

We define an operator which addresses the probability of first escape from S, againby restricting (masking) the Galerkin matrix A. Let this “escape matrix" be defined,110

[E−S ]i, j ≡{

Ai, j if Bi ∈ S0 otherwise

}, (7.48)

Since E−S contains zero probability elements of transitions of the type, S → S, we nowhave,

[E−S ]ni, j = P(Fn(x) ∈ Bj ⊂ S and Fm (x) ∈ S, ∀m < n|x ∈ Bi ⊂ S), (7.49)

which is exactly the probability of first exit transition that we require to calculate the meanin Eq. (7.46).

Since the events described in the probability in Eq. (7.49) are disjoint events for twodifferent target boxes,

P(Fn(x) ∈ ∪j :Bj⊂S Bj and Fm(x) ∈ S, ∀m < n|x ∈ Bi ⊂ S)=∑

j :Bj⊂S

[E−S ]ni, j (7.50)

We now have the notation necessary to state the following Theorem:

Theorem 7.4. [35]. If the escape matrix defined in Eq. (7.48) is bounded,111 ‖[E−S ]‖< 1,then the expected mean time of escape from S for an orbit starting in box Bi ⊂ S to anybox Bj ⊂ S is,

< T (Bi , Bj ) >= 1

#{ j : Bj ⊂ S}∑

j :Bj⊂S

[[E−S ] · (I − [E−S ])−1 · (I − [E−S ])−1]i, j (7.51)

110This is an approximation similar to that in Eq. (7.41), except we have less severely restricted the regionof integration of the Frobenius-Perron operator:

∫S

∫M ν(x− F(y))ρ(y)dydx .

111The notation ‖.‖ used in this Section denotes the matrix natural norm, ‖A‖ = sup‖u‖=1 ‖A ·u‖, in termsof a vector norm, which we could choose here to be the sup-norm, in which case, ‖A‖∞ =maxi

∑j |Ai, j |,

[150], which is the maximum column sum.

Page 236: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

228 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

where #{ j : Bj ⊂ S} is the number of boxes Bj ⊂ S, and T (Bi , Bj ) is the time of first escapeof a single randomly sampled path starting at Bi , based on paths defined by the graph G Amodel of the Frobenius-Perron operator.

Note that this theorem as stated is closely related to the so-called fundamental matrixfrom the topic of absorbing Markov chains from the related theory of finite markov chainstheory [309, 185]. More on absorbing Markov chains can be found in Sec. 5.11 about opensystems. Proof: By Eq. (7.49),

< Time of first escape from Bi ⊂ S to Bj ⊂ S >=∞∑

n=1

n[E−S ]ni, j , (7.52)

By the assumed bound, ‖[E−S ]‖< 1, we have the matrix geometric series [23],

(I − [E−S ])−1 =∞∑

n=0

[E−S ]n , (7.53)

from which it is straightforward to derive,

∞∑n=1

n[E−S ]n = [E−S ] · (I − [E−S ])−1 · (I − [E−S ])−1, (7.54)

Hence, selecting the i , jth entry of this matrix on the right side of Eq. (7.54) gives themean escape time from S, starting at box Bi , and arriving after n-iterates at box Bj . Byindependence of the events of arriving at two different boxes Bj and Bj ′ , both in S, thetotal mean escape time from box Bi to any box Bj ⊂ S is the arithmetic mean of the meanescapes times to each individual box, which gives the formula Eq. (7.51). �

The restriction that ‖[E−S ]‖ < 1 is not severe, since in all but a trivial choices of S(when ‖[E−S ]‖ = 1, because the matrix A is stochastic), the inequality should hold. Ofcourse the formula Eq. (7.51) gives expectation in terms of the combinatorial model of theFrobenius-Perron operator, formed using a fine grid, and not the full operator. We expectthe calculation to be good for a fine grid.

Remark 7.1. By considering the disjointness of the covering by open elements in the grid{Bi} and also the topological cover {S, S}, then the probabilities involved in Eq. (7.51)are independent. Therefore the mean escape times between larger regions, for example< T (S, S) >, or any other disjoint set pairing, is easily computed by making the sum overthe larger number of boxes therein and likewise counting the larger number of boxes in thedenominator.

Example 7.1. Noise induced basin hopping in stochastically perturbed dynamical sys-tem. A model of childhood disease. Consider a model used to describe the dynamics of achildhood disease spread in a large population,

S′(t)= μ−μS(t)−β(t)I (t)S(t),

I ′(t)=(

α

μ+γ

)β(t)I (t)S(t)− (μ+α)I (t),

β(t)= β0(1+ δ cos2π t). (7.55)

Page 237: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 229

Figure 7.20. Noise induced basin hopping in stochastically perturbed dynamicalsystem. SEIR model with noise. (Left) The phase space depiction from the deterministicmodel Eq. (7.55) has two basins corresponding to a stable period-2 point (white region)and a stable period-3 point (pink region). With even a small amount of noise, points onthe stable manifold of a period-2 point can jump onto the stable manifold of the period-3point, across the basin boundary formed by the stable manifolds of the saddle-period-3orbit. (Right) The resulting almost invariant sets can be found by appropriately permutingthe Ulam-Galerkin matrices. The off diagonal elements carry the transport information.[35]

Here S(t) refers to the concentration number of susceptible individuals and I (t) refers tothe concentration number of infected individuals. This is a typical rate equation to be foundin mathematical epidemiology. These sort of equations are special cases of those equationsfound in population dynamics, which describes the growth and decay of each population.In this particular model, here we see quadratic interaction terms, and the contact rate β(t) istaken to be a periodic function to model the concept that childhood diseases amongst manyhave a seasonal aspect.

In [20] it was demonstrated that when normally distributed noise ν(x) with a suf-ficiently large standard deviation is added, then the behavior of the modified SI model[293, 294], was transformed from regular periodic cycles, to something completely differ-ent, which we may call stochastic chaos. In the deterministic case with parameter valuesat μ = 0.02, α = 1/0.0279, γ = 100, β0 = 1575, and δ = 0.095, the bifurcation diagram(not seen here but shown in [35]) reveals that there exist two stable periodic orbits, and twounstable periodic orbits. Furthermore, the deterministic system exhibits “bi-stability" sincethere are two competing basins of attraction. In Fig. 7.20 we show these two basins in whiteand pink, and corresponding stable and unstable manifolds from the pair of period-two sta-ble and then unstable orbits and likewise for the pair of period-three stable and unstableorbits. The stable manifold of the period-three orbit naturally outlines the two regionswhich are the basin of stable period-two and period-three respectively.

Page 238: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

230 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

In [35] we demonstrated that by methods reviewed in this section, that there is dif-fusion between the two basins due to the stochastic effect, a form of noise induced basinhopping. There is not yet a heteroclinic tangle, since as it turns out these parameters arenot adequately large where the global tangency bifurcation occurs which would give riseto chaos. Nonetheless, the stable and unstable manifolds serve an important role in noise-induced transport in this subcritical state. The noise facilitates jumping between thesemanifolds. The tools described herein identifies the flux between basins as noise is added.Multiplying these rates by the probability density function results in a measure of where atrajectory is most likely to escape to another basin. It was found [35] that the highest escaperates as described in Eq. (7.51) occur exactly where we previously conjectured, at the nearheteroclinic tangencies, thus creating a chaos-like orbit. In Figs. 7.21-7.22, we illustratethe effect of increasing noise for this system,. In Fig. 7.23 we illustrate the regions whichare most active in transport activity. As we see, the reddened area denoting greatest dif-fusion occurs just where the deterministic version of the system has the unstable manifoldof the period-two orbit most closely approaching to the stable manifold of the period-threeorbit. Effectively it is as if the stochastic effect is completing the heteroclinic tangle. Butthe analysis that finds this effect is inspection of the transport rates encoded in the maskedtransfer operators depicted in Fig. 7.20.

Figure 7.21. Invariant density (PDF) of the SI model for noise standard deviationσ = 0.03 due to direct simulation; there is strong mixing between what the deterministicsystem predicts is separate basins of the bistable system. [35]

Example 7.2. Transport in the Gulf. In [37] we studied the Deep Horizon oil spill inthe Gulf of Mexico, as already referenced at the end of Chapter 1 regarding Figs. 1.17-

Page 239: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 231

Figure 7.22. The dominant eigenvector of the corresponding Galerkin matrixapproximates invariant densities of the stochastic systems, for increasing noise, σ = 0.001and σ = 0.03. The essential feature is that when σ = 0, the main density spikes are at thedynamic centers of each respective basin. Initially, as σ increases, the density becomesless diffusely distributed around the stable fixed points, due to the predominantly smallstochastic diffusion added to the deterministic dynamics. There persist two stable fixedpoints and two distinct basins. As we continue to increase σ , a crossover effect occurs nearσ = 0.02, after which the density mass becomes mixed throughout a larger region, andpredominantly mixed between the originally separate basins. [35]

1.21. Using the resulting Ulam-Galerkin matrix mentioned in Fig. 4.3, we can discusstransport mechanisms in the Gulf of Mexico. Producing an Ulam-Galerkin matrix of thenon autonomous system progressively as time evolves leads to a set of transfer operatorsevolving in time. Escape rates by methods discussed in this section, and partitions asdiscussed in Chapter 5 have been presented in [37]. Further, by method of relative measureand relatively coherent sets, just as we illustrated for the Rossby system in Fig. 5.12 onrelative coherence, we produced in [39] a map of relative coherence shown in Fig. 7.24,including the use of coherence criterion inequality Eq. (5.140) of ρ0 used as a stoppingcriterion. We choose 20,000,000 points uniformly and randomly in the water region asthe initial status, on more details of data. See [38]. The final status is the positions ofthese points after 6 days. We use 32,867 triangles {Bi}32867

i=1 as a partition of X and 32,359triangles {Cj }32359

j=1 as a partition of Y . After applying our subdivision method on thesetriangles, the results are shown in Figure 7.24. In this example, we set the ρ0 = 0.9998 asthe threshold the stopping criterion. We find it to be particularly interesting to contrast thisoperator based perspective to the FTLE perspective shown in Fig. 7.25 in which we alsoshow tracers. This data is most enlightening as a movie which can be found at, [34].

7.4.3 Escape Time, Long Life, Scattering and Unstable Chaotic Saddles

Unstable invariant sets are important to understand mechanisms behind many dynamicallyimportant phenomenon such as chaotic transients. These phenomenon can be physicallyrelevant in experiments. Take for example the so-called “edge of chaos" scenario wherebya transient turbulence-like phenomenon in a plane Couette flow, or pipe flow, may appearbut gives way to linear stability of the laminar flow [291]. This was explained in Skufca-

Page 240: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

232 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

ln(S)

ln(I)

Figure 7.23. Transport PDF Flux. The conditional probability of transition fromsmall amplitudes to large outbreaks. The highest probability regions of transport (dark)point to a bull’s eye monitoring region for control. Overlaid are the stable and unstablemanifolds corresponding to the underlying deterministic model. Compare to Fig. 7.20 )[35]

Yorke-Bruno [297] as the presence of an unstable chaotic saddle (Defn 7.8).As a lead in to the discussion of unstable chaotic saddles, consider specifically from

Chapter 6 and refer to Fig. 6.29 wherein a sequence of embedded subshifts was used toapproximate the symbolic dynamics representation of the dynamical system. We reprisethis discussion, now in the context of unstable chaotic saddles and escape from these sets.We can now describe the embedded subshifts of Chapter 6 and Fig. 6.29 as unstable chaoticsaddles.

“Unstable chaotic saddles," may be Cantor sets embedded in some more regular at-tractor set [253, 6, 33, 29] as already referenced several times herein. Techniques suchas the PIM triple method [253] and simplex method variant [243], and even the step-and-stagger method [317] have been developed to compute long ε- chain pseudo-orbits nearsuch sets. The following relates to a long-life function [33] which was designed to bothdescribe the lifetime landscape function and to find points with long lived orbits beforeescape.

We use notation for a uniformly continuously differentiable discrete-time dynamicalsystem,

zn+1 = F(zn), (7.56)

The discussion of unstable sets must be for orbits relative to some reference set,which we will call B . We require that the orbit lies in some set B , or at least ask for sometime. See Fig. 7.28 for an example of such as set B chosen to be a disc. Then we say

Page 241: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 233

Figure 7.24. Hierarchical relative coherence in the Gulf of Mexico following theflow according to vector field from the HYCOM model, [182]. Tree structure and relativelycoherent pairs coloring as in (a)-(b). The inequality Eq. (5.140) of ρ0 is used here as astopping criterion.

that {zi }Ni=0 is a B-invariant orbit segment if zi ∈ B ,∀i = 0,1, .., N , and each zi satisfiesEq. (7.56). Exact orbits rarely exist in a finite precision computer and choosing ε = 10−15,slightly bigger than the order of machine precision, is the best that can be constructed. Auseful concept which is slightly weaker than the notion of orbits is,

Definition 7.5. [279]. An ε-chain segment {zi }Ni=0, also called a pseudo-orbit, satisfies,

‖zn+1−F(zn)‖< ε,∀i = 0...N . (7.57)

When discussing unstable invariant sets, it is useful to define,

Definition 7.6. [32] A discrete forward(backward) lifetime escape function,

L±B : B→ Z±, (7.58)

as follows,

L±B (z)= | j | if F±i (z) ∈ B , for 0≤ |i | ≤ | j |, but F±( j+1)(z) /∈ B . (7.59)

That is F±i denotes the i th forward or backward iterate, depending on the sign, and F0

denotes the identity map, F0(z)≡ z, for all z.

Example 7.3. Lifetime Escape Function of a Henon Map. In Figs. 7.26 and 7.27, we seelifetime functions plotted over the phase space, for the Henon map, Eq. (6.55), and where Bis a circle of radius 2 centered on the origin. Notice the central role of the chosen set B from

Page 242: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

234 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Figure 7.25. FTLE for the Gulf computed from a 3-day range beginning May 24,2010. Compare to Fig. 7.24. Also shown is the underlying vector field on May 24, 2010, aswell as black tracers representing the spread of oil. A month after the initial explosion, thetracer particles have dispersed significantly from the source. [37]

which the question of escape is posed. As an exercise in the notation, since the Henon maphas an attractor set which we may call A, consider then if we were to choose the escapeset B such that A ⊂ B and B in the trapping set. Then it is easy to see that L±B (z) =∞for all z ∈ B . Thus the interesting Cantor-like appearance of the invariant set suggested inFig. 7.28 and the tower-like appearance of the lifetime escape function [32, 240] shown inFigs. 7.26-7.27 is highly dependent on the fact that the chosen disc B intersects the Henoninvariant set A. �

To elucidate the tower-like appearance of these lifetime escape functions L±B (z) andthe relationship to invariant sets, we define.

Definition 7.7. Let,LB (i )± = {z : L±B (z)≥ |i |}, (7.60)

be the set of points with lifetime of at least i . We name i -lifetime escape set as denotedLB(i )±

Whereas L±B (z) is a function that measures lifetime from a given point z, a set withspecified escape properties is denoted LB (i )±. These definitions help to explain the tower-ing steps nature of the lifetime escape functions shown in Fig. 7.26-7.27. Now we see that

Page 243: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 235

Figure 7.26. (Left) A cross-section of the forward lifetime function of Henonmap, Eq. (6.55), shown in Fig. 7.27, where B is a circle of radius 2 centered on the origin.Notice the stepping nature of the towers of increasing lifetime. (Right) Layers show pointsinvariant in a box [−2,2]× [−2,2] for i = 1,2,3,4 steps successively. The spikes limit onthe invariant set L±(∞), approximated in Fig. 7.28. [32].

Figure 7.27. Forward (left) and backward (right) lifetime escape functionsEq.(7.59) of Henon map, Eq. (6.55), where B is a circle of radius 2 centered on the origin.[32].

the towers follow the natural nesting of steps of increasing heights for continuous maps F ,

L±B (i ±1)⊂L±B (i ),∀i . (7.61)

Further, the forward invariant set is,

I=L+B (∞). (7.62)

As another example, another set B in Fig. 7.29 is shown. Choosing B = {(x , y) : y >

ε} again intersects the attractor, A. By consequence, any invariant set of B ∩A must notlie in any pre-image of the gap shown, B , the complement of B .

It is a common scenario that when there is chaos on the attractor set for there to be

Page 244: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

236 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Figure 7.28. L+(60) (almost) invariant set of the Henon map, using B is the circleof radius 2 approximates the invariant set I = L+(∞) which is apparently an unstablechaotic saddle, Definition 7.8. [32].

chaos on an invariant set L+B (∞), and furthermore for these sets to be unstable. For thissituation, the following is a useful and commonly discussed concept. for these

Definition 7.8. An invariant set I⊂M of a map F : M→ M is an unstable chaotic saddleif,

• I is an invariant set,

• The map is chaotic even for the map restricted to I, F |I,

• I is unstable relative to M. That is, no open neighborhood112 N of I exists such thatall of its points remain forever. Fn(z) ∈ N ,∀n > 0 if z ∈ N ∩I.

The phrase “saddle" associated with “unstable chaotic saddle" refers to a commonscenario where points in I have both nonempty stable as well as unstable manifolds, butby practice the sets are still called “saddle" even if there is no stable direction as we see inthe one-dimensional scenario in Fig. 7.30.

Formally, these chaotic saddles can inherit the symbolic dynamical description assubshifts embedded in a larger shift grammer of A [28, 29, 206], a statement which ismade clear in the context of Sec. 6.4.4 and as seen in Fig. 6.29. Said symbolically, thewords corresponding to the set B cannot be in the subshift of the invariant set I. The

112 N is a open neighborhood of I means is N is open in the embedding set M and I ⊂ N . Generally thephrase neighborhood denotes that not only is N open but it “tightly wraps" in that no points in N are farfrom some points in I.

Page 245: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 237

analogy to the figures in Sec. 6.4.4 and symbolic dynamics described by Fig. 6.29 closelyrelated to the embedded unstable invariant sets by considering a tent map with a hole.

Example 7.4. Lifetime Escape Function and Lifetime Escape Set of a Tent Map. SeeFig. 7.30, where we choose a full tent map, and so the invariant set113 is chosen A= [0,1]and as in [339] we choose,

B = [0,1

2− ε]∪ [

1

2+ ε,1], (7.63)

which creates a hole B = ( 12−ε, 1

2+ε) relative to which we can discuss lifetime escape. Asshown, B = A10 is shaded darker gray. The first preimages A20∪A21= F−1(B) and secondpreimages A30∪ A31∪ A32∪ A33 = F−2(B) are shown as lighter gray strips. As a matterof example, according to the lifetime escape function, Eq. (7.59), for any z ∈ A20∪ A21 forexample, L+

B(z)= 1.

As an aside, we are prepared at this stage to make a remark concerning the interestingdevil’s staircase function as seen in Eq. (6.30) as studied in [29, 28, 206, 339]. As explainedin Fig. 6.29, increasing the size of the hole removes corresponding words from the grammarof the chaotic saddle. But what causes the flat spots in the entropy function? In Fig. 7.32we see that as the size of the hole B is increased, so is shown the gap B(s)= ( 1

2 − s, 12 + s)

of this v-shaped region in the shaped region in the x phase space by s parameter space.Also the preimages decrease, and for the simple piecewise linear tent map, they are alsov-shaped but narrower. The summary of the story is that when these tounge-like regionsintersect, the symbolic word corresponding to an interval disappears from the invariant setfor all further larger values of s. The key to the flat spots of the devil’s staircase topologicalentropy function in Eq. (6.30) is there are open intervals of s when there is no word toeliminate since the next word has already been eliminated. For example, notice the intervalof s centered on s = 0.2. Likewise, a Cantor set of these tongues114 of various openingangles emanates from the interval [0,1]×{0} with countable many such intersections andcorresponding flat spots. In between these, open intervals of s exist for which the symbolicdynamics of the unstable chaotic saddle does not change. We have even seen similar behav-ior in multivariate settings [339] such as the Henon map with a hole as a gap of increasingwidth as suggested by Fig. 7.30. �

Consider Fig. 6.29b which illustrates symbolic dynamics of just such a situation asthe hole in the tent map in Fig. 7.30. If ε is a diatic rational115, then the hole B has an exactrepresentation by finite symbolic words. The case of Fig. 6.29b corresponds to ε = 1

16 .The hole corresponds to the intervals labelled by the 4-bit words 0.100 and 1.100. In thisexample, we can represent the directed graph to emphasize the hole B.

113Notice we are speaking of the invariant set on this example rather than an attractor.114Which reminds us of the Arnold tongues from another setting in dynamical systems but which also gives

rise to a devil’s staircase function.115ε = p

2n for some integer n and 0 ≤ p ≤ 2n − 1. In other words it is a fraction whose denominator is apower of 2.

Page 246: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

238 Chapter 7. Transport Mechanism, Lobe Dynamics, Flux Rates and Escape

Figure 7.29. An invariant set unstable chaotic saddle, I built as Eq. 7.62 fromthe Henon attractor A with a set B is a horizontal strip removed as Eq. (7.62). A few pre-images are shown. A succession of holes are removed from the resulting unstable chaoticsaddle just as a classic construction of a Cantor set. Compare to Fig. 3.2. [206]

Figure 7.30. An invariant set unstable chaotic saddle, I built as Eq. 7.62 from theinvariant set of a full tent map A = [0,1] with a set B is an interval near 1

2 shaded darkgray. [339]

Page 247: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

7.4. Transport Rates 239

Figure 7.31. Considering a hole such as in Fig. 7.30 that is a diatic rationalallows us to describe the dynamics exactly as a symbolic dynamics whose grammar isa finite directed graph, such as was shown in Fig. 6.29b. Furthermore here we show areconfiguring of the embedding of this graph in the plane so as to emphasize the hole-likenature of this dynamics. [29]

Figure 7.32. The devil’s staircase like topological entropy function seen inFig. 6.29 is understood here for a tent map by considering a hole parameter s, forB(s) = ( 1

2 − s, 12 + s), and all of its preimages which creates tongues like structures that

overlap at countably many points and open sets of s for which the symbolic dynamics ofthe unstable chaotic saddle does not change. [29]

Page 248: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure
Page 249: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 8

Finite Time LyapunovExponents: FTLE

8.1 Lyapunov exponents: One-dimensional MapsA somewhat informal introduction to the Lyapunov exponent of one-dimensional differen-tiable map will be presented in this section. For simplicity, we start with a one-dimensionalmap. So, let X be a compact subset of R and suppose that T is a piecewise continuouslydifferentiable map on X defined by the rule:

x �→ T (x), x ∈ X . (8.1)

The iterated map T (n)(x) here will refer to a n−fold composition

T (n)(x)= T ◦ T ◦ · · · ◦ T (x)︸ ︷︷ ︸n times

m ∈ Z . (8.2)

This section is concerned with the determination of the separation rate of the derivative ofa map T (n) as n increases. In particular, we would like to study the rate at which the orbitsof infinitesimally nearby (initial) points are separating. If we choose |x− y|< ε for somesmall ε > 0, we would like to study the quantity

∣∣T n(x)− Tn(y)∣∣≈ n−1∏

i=0

∣∣∣T ′(T i (x))∣∣∣×|x− y| , (8.3)

which is readily derived from the chain rule. It then follows that

1

nlog∣∣T n(x)− Tn(y)

∣∣≈ 1

n

n−1∑i=0

log∣∣∣T ′(T i (x))

∣∣∣ . (8.4)

The right-hand side of the above equation represent the exponential growth rate of sep-aration. Observe that if the orbits of x and y converge as n increases, this quantity willbe negative; conversely, it will be positive if nearby orbits diverge. In general, as n goesto infinity, the limit of this quantity may not exist; for example, the point may come very

241

Page 250: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

242 Chapter 8. Finite Time Lyapunov Exponents: FTLE

close to each other infinitely many time. Nevertheless, it will always be bounded since weassume that X is compact. Therefore, for every (initial)point x ∈ X , we may define theLyapunov exponent of x , λ(x), as follows:

λ(x) := limsupn→∞

1

nlog∣∣∣T ′(T i (x))

∣∣∣ . (8.5)

Example 8.1. Let T (x)= 2x . Then λ(x)= log2 for all x ∈ X since T ′ = 2.

Example 8.2. Let X = [0,1] and consider a tent map defined by

T (x)={

2x if 0≤ x ≤ 0.5;

2(1− x) if 0.5≤ x ≤ 1.(8.6)

In this case, T (x) is piece-wise continuous with∣∣T ′(x)

∣∣ = 2 for all x except at x = 0.5,where T ′(x) does not exist. Therefore, λ(x) is not defined at x = 0.5 and all of x such thatT (i)(x)= 0.5 for some i . However, it can be shown that a set of those points is countable.Hence, we can say that λ(x)= log2 up to Lebesgue-measure of zero.

Example 8.3. Empirically estimating Lyapunov exponents for the logistic map, xn+1 =4xn(1− xn). Rather than as an efficient way to estimate λ(x), the following is presentedmerely as a numerical presentation of the idea behind Eq. (8.4) for the sake of intuition.See Fig. 8.1 where the orbits from two nearby initial conditions, and the growth of the errorbetween them is shown. Instead of averaging the derivative along an orbit according toEq. (8.4), this crude estimate reveals the idea behind the computation. Shown is a graphicaldescription of an average growth rate of error, log(|xn− yn|)/n for small |xn− yn|.

We close this example with the remark that an averaged quantity along an orbit alsocorresponds to a spatial average relative to an ergodic measure if it exists, which in this one-dimensional setting is according to the Birkhoff ergodic theorem. However, the Oseledet’sMultiplicative Ergodic Theorem is required more generally as discussed in Sec. 8.2. Com-paring the right hand side of Eq. (8.4), an average logarithmic estimate of the derivativealong an orbit is a special case of Eq. (1.5). As such, we can interpret that we are esti-mating quantity log |T ′(xn)| as the orbit xn visits various regions of phase space. Furtherthe orbit xn visits various regions of the phase space according to the ergodic measurewhich we know exists; since this chosen to be the logistic map, with parameter value 4,we know the ergodic invariant measure in closed form through its corresponding density,dμ(x)= 1

π√

x(1−x). We could directly compute that λ= 2.

The discussion in this chapter will lead us to not to consider the long time averagesas has been traditional practice, as fits a major theme of this book and recent practice; rel-atively short time estimates of traditional analysis may also give useful information. Here,even if the system is ergodic, it may not mix quickly everywhere. We have already calledthis concept “weakly transitive" due to almost invariant sets as discussed in Chapter 3. Insuch case, for short times, estimates of Lyapunov exponents can become highly spatiallydependent. This runs counter to the until recent almost universal folk-lore that since almostevery initial condition gives the same value, then there is no spatial information in the com-putation. Quite to the contrary, allowing for spatially dependent short time estimates canreveal almost invariant sets, leading to the finite time Lyapunov exponents as discussed inSec. 8.3.

Page 251: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.2. Lyapunov exponents: Diffeomorphism and flow 243

Figure 8.1. Exponential growth of error. (Upper Left) An empirical orbit of thelogistic map, xn+1 = 4xn(1− xn) starting from the initial condition x1 and also from anearby initial condition y1 = 0.4001. Errors |xn − yn| grow exponentially and thus thelog(|xn − yn|). (Right) A time series of both orbits xn and yn plotted as n vs xn, and wesee errors are initially small but grow quickly. (Lower Left) Errors grow exponentiallyuntil they saturate at close to the size O(1) of the phase space, xn ∈ [0,1], after whichthe since the orbits move chaotically, seemingly randomly relative to each other, the errorssometimes sporadically become small (recurrence) and then grow again. (Lower Right) Onaverage, the logarithmic error growth is linear for small errors until it saturates. The slopeof this average initial line estimates the Lyapunov exponent as described by Eqs. (8.3)-(8.5).

8.2 Lyapunov exponents: Diffeomorphism and flowThis section present a brief introduction of Lyapunov exponent for diffeomorphism andflows in multidimensional cases. To extend the previous treatment on one-dimensional maponto a multidimensional map, let us consider a manifold M ∈ Rm and a diffeomorphismf : M → M . Let ‖ · ‖ be the norm on the tangent vectors. Now for a point x ∈ M think ofa vector emanating from x on the tangent space, i.e., v ∈ Tx M . On the tangent space, theevolution is described by the linearized dynamic of f , i.e., Dxf. Like in the one-dimensionalcase, our interest is to determine the exponential growth rate in the direction of v under theiteration of Dxf. This leads us to definition of Lyapunov characteristic exponent (LCE) inthe direction v at x as

λ(x,v) := limk→∞

1

klog∣∣∣Dxfkv

∣∣∣ , (8.7)

Page 252: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

244 Chapter 8. Finite Time Lyapunov Exponents: FTLE

if the limit exists. In multi-dimensional case, a positive LCE at x in the direction of v im-plies that a infinitesimal line element emanating from x in the direction of v will experienceexponential expansion along the solution x(t; t0,x0) and likewise when LCE is negative, itwill see the exponential contraction. It is not difficult to see that λ(x,cv) = λ(x,v) for aconstant c, so the Lyapunov characteristic exponent depends on the initial point x and theorientation v but not on the length of v. Therefore, for a given x, it is interesting to askhow many distinct values of λ(x,v) for all v ∈ Tx M . The answer to this question was givenin the Oseledet’s Multiplicative Ergodic Theorem (MET) under very general conditions.Precise statements and proof are referred to [254, 3]. Roughly speaking, the MET statesthat:

1. there is a sequence of subspaces {0} = V r(x)+1x ⊂ V r(x)

x ⊂ ·· · ⊂ V 1x = Tx M such that

for any v ∈ V jx \V j−1

x , the limit in (8.7) exists and λ(x,v)= λj (x);

2. the exponents satisfy −∞≤ λr(x)(x) < · · ·< λ2(x) < λ1(x);

3. DxfV jx = Vf (x) j for all 1≤ j ≤ r (x);

4. the functions r (x) and λj (x) are both measurable and f−invariant, i.e., r ◦ f = r andλj ◦ f = λj .

The first two statements imply that there are at most r (x) of λ(x,v) depending on to whichsubspace,V j

x , v belong. Each V jx is invariant in a sense of (iii). The statement (iv) means

that for a given x and v, λ(x,v) and r (x) are constant along the orbits of f . These r (x)distinct values λj (x) are called the Lyapunov exponents of x.

We consider a geometric interpretation of the Lyapunov exponents for a multi-di-mensional map. It is clear in the one-dimensional case that there exists only one exponent,say λ1, and the length of a line will grow as exp(λ1). In the multi-dimensional cases,according to the Oseledet’s theorem there can be many Lyapunov exponents as dimensions,say λ1 > λ2 > .. . > λn . These exponent also have the same geometric interpretation as theone-dimensional case; an area will grow as exp(λ1+ λ2),a volume as exp(λ1+ λ2+ λ3).However, the area and volume will be distorted along the orbit. The direction v1 of thegreatest separation along the orbit of nearby points corresponds to λ1. Then, choosingamong all directions perpendicular to v1, the direction of greatest separation correspondsto v2 and etc. Figure 8.2 shows how an infinitesimal circle is transformed to an ellipse by athree-dimensional map, the sphere is stretched out in the expanding direction while shrankin the contraction directions. Also, the images (w1 and w2) of initially orthogonal vectors,v1 and v2 remain orthogonal.

In fact, some insight into the MET can be obtained by considering this geometric in-terpretation in view of SVD of the matrix Ak = Dxfk , say A ∈ Rm×n . Then the orthogonalsets obtained from SVD will represent the direction corresponding to the Lyapunov expo-nent of the dynamic of A. For example, in a two-dimensional, the vectors v1 and v2 must beparticularly the pre-images of the principle axes w1 and w2 respectively (i.e. Dxfkvi =wi ).Also, the vector vi will be expanded (or contracted) by a factor of σi , the singular vector ofAk corresponding to vi . The MET then says that the choice of vi will be independent of kfor a sufficiently large k and σi yields an approximation of Lyapunov exponent λi = logσifor a large k.

Our interest now turns to the calculation of λi (x,v). The MET states involves theexistence the Lyapunov exponents but not how to calculate them. Nevertheless, it suggests

Page 253: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.2. Lyapunov exponents: Diffeomorphism and flow 245

Figure 8.2. An infinitesimal circle is mapped to an ellipse by a two-dimensional map.

that a simple calculation of the largest Lyapunov exponent (i.e. λ1(x)) for a given x andarbitrary chosen v. When choosing arbitrarily, most vectors v will have all component inthe direction of the eigenvectors (i.e. most v will lie in V 1

x \V 2x ), hence

λ1(x)= limk→∞

1

klog(|Dxfkv

∣∣∣).So, as long as v does not lie in V 2

x , λ1(x) can be determined from the above limit. Onceλ1(x) is known, the calculation of the remaining Lyapunov exponents can be determinedbased on the condition (ii) but it is not easy to do. We refer the details of such calculationto [59].

We now discuss the concept of the Lyapunov exponent for the flow. Consider a vectorfield

x= f(x, t), x ∈ Rn , (8.8)

where f(x, t) is assumed to be Cr for some r ≥ 1. At time t , the notation x(t ; t0,x0) isconventionally used to denote the solution of the initial value problem (8.8) to emphasizethe explicit dependence on the initial conditions x0 and t0. However, we will occasionallyabbreviate the notation x(t; t0,x0) by x(t) when the initial conditions are understood. Theflow map associated with (8.8), which maps the initial position x0 of a trajectory beginningat time t0 into its position at time t , can be written in a form of two-parameter family oftransformations (called a flow map) defined as

φtt0 : U →U

: x(t0; t0,x0) �→ x(t ; t0,x0).(8.9)

This flow map is typically obtained by a numerical integration of the vector field in the firstequation of (8.8). Note that in the extended phase space U ×R, we can assume the exis-tence and uniqueness of the solution x(t; t0,x0) and hence it satisfies the so-called cocycleproperty:

x(t2; t1,x0)= x(t2, t1,x(t1; t0, x0)), and t0 ≤ t1 ≤ t2. (8.10)

The basic concept of Lyapunov exponents for the flows are similar to the diffeomorphismbut it requires a knowledge of the derivative of the flow map φt

t0 with respect to the initialpoint x0.

To study the Lyapunov exponent, we then need to describe the dynamics of the solu-tion near x(t , t0,x0). Denote by Dxf ≡ Dxft

t0 the linearization of f at x, represented by an

Page 254: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

246 Chapter 8. Finite Time Lyapunov Exponents: FTLE

n× n matrix. Note that this matrix varies with time as the point x evolves along its pathline from t0 to t . The time-dependence of Dxf arises through the explicit dependence off on t as well as through the time-dependent solution x(t ; t0,x0). Therefore, even if f istime-independent, Dxf may be time-dependent.

To find Dxφtt0 , the derivative of the flow map φt

t0 for some fixed t with respect tosome fiducial point x, we solve the variational equation

d

dtDxφ = Dxf

(φt

t0(x)), t) ·Dxφ, Dxft0

t0 = I. (8.11)

where Dxφ abbreviates Dxφtt0 and it is also called the fundamental matrix solution of (8.11).

Intuitively, the vector Dxφ(δx) is a linear operator that takes small variations tangent at timet0 to small variations tangent of the solution of (8.8) at time t :

Dxφtt0 : TxM→ Tφt

t0xM (8.12)

More precisely, we haveDxφ

tt0 (x) · f(x)= f(φt

t0(x)). (8.13)

The Lyapunov characteristic exponent (LCE) emanating from a point x0 with anorientation v ∈ TxM can then be defined as

λ(x0,v, t0) := limsupt→∞

1

t− t0log

( |Dxφtt0(x0)v||v|

). (8.14)

However, this asymptotic notion of the LCE does not lend itself well to a practicaltool for a study of transport and mixing of the flow. Observe that all LCEs are zero forany trajectory that eventually become regular in a sense that lim |λ(x0,v, t0)| convergesregardless of how much strong expansion takes place for a finite time. But such a finite-timephenomena can be practically important to identify transport and mixing regions. In thenext section, we will present a “direct" estimation of the maximal Lyapunov exponent anddemonstrate its potential as a tool to identify global time-dependent structures governingthe transport and mixing of a flow.

8.3 Finite-time Lyapunov exponents (FTLE) andLagrangian coherent structure (LCS)

8.3.1 Setup

It has been demonstrated that the stable and unstable manifolds of hyperbolic fixed pointsin two-dimensional autonomous systems separates regions of qualitatively different dy-namics. In the case of periodic or quasi-periodic systems, the dynamics of lobes betweeninterlacing stable and unstable manifolds governs the transport behavior of the systems.One salient behavior of initial points straddling stable manifolds of hyperbolic fixed pointsor periodic points is that they will typically experience exponential separation in forwardtime, which can be deduced from the Lambda lemma [156], and likewise for those pointsstraddling unstable manifolds under time-reversal system. This is illustrated in Figure 8.3;

Page 255: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 247

the expansion of the green neighborhood is a consequence of the Lambda Lemma. Also,one may recall that only those points on the stable manifold can converge to the fixed point.Together with the previous fact, this implies the separation of points initially straddle thestable manifold. Such characteristics play a pivotal role for approximating stable/unstablemanifolds in this chapter.

Figure 8.3. Any small neighborhood of a point q (except the fixed point itself) ona stable manifold will eventually expand exponentially in forward time, which, at least forthe homoclinic orbit, follows from the Lambda Lemma.

However, for general time-dependent systems, the stable/unstable manifolds may noteven exists neither with nor without a presence of instantaneous hyperbolic fixed points orperiodic points. Furthermore, the hyperbolic fixed points may not even be trajectories andmay vanish, emerge or lost its hyperbolic properties in time. So, how do we define thecore structures that organize the trajectory pattern around them in a similar fashion to thehyperbolic invariant manifold in the autonomous systems?

In the case of time-dependent systems, we will seek to locate the dynamically evolv-ing structures that form a skeleton of Lagrangian patterns. Such structure is termed La-grangian coherent structure (LCS), which was proposed in[163]. Therein, LCS is definedto be a material surface, a co-dimension one invariant surface in the extended phase space,and exhibit the strongest local attraction or repulsion in the flow. As a material surface,LCS is, therefore, a dynamical structure moving with the underlying flow.

Recalled from a discussion in the previous section that certain phenomena in mix-ing and transport can be identified only over a finite-time intervals and hence the abilityto locate LCSs assumed to exist over a finite-time interval is our goal here. Heuristically,LCS can be captured by the curves with locally relatively large separation rate. Therefore,one would need a way to estimate the separation rate of passive tracers for different initialconditions. A conventional tool to quantify tracer separation and sensitivity to the initialcondition of a time-dependent system is the finite-time Lyapunov exponent (FLTE). Theidea of using FTLE to locate LCS for time-dependent systems can be dated back sincethe work of Pierrehumbert and Yang [268], in which the spatial distribution of the finite-time Lyapunov exponent fields are used to identify the partial barrier to chaotic mixing onthe Isentropic surfaces of the Antarctic stratospheric circumpolar region. In[162, 298], itwas suggested that a repelling LCS over a finite-time interval may be captured through the

Page 256: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

248 Chapter 8. Finite Time Lyapunov Exponents: FTLE

“ridge" of the FTLE field. Likewise, the attracting LCS should the ridge of the backward-time FTLE field. However, the heuristic motivation of using the FTLE ridge to mark theLCSs can be problematic. Some counterexamples to show that LCSs may not be the FTLEridge and FTLE ridges may not locate LCSs have been presented in[163]. A discussion ofthese examples will be deferred till the next section. An introduction of the concept andcalculation of FTLE is given below.

Consider again the velocity fields of the form

x(t; t0,x0)= v(x, t), x ∈U ⊂ Rn

x(t0)= x0(8.15)

where v(x, t) : Rn ×R→ Rn is C2 in open set U ∈ Rn and C1 in time. In general, theFTLE can also be defined for a higher dimension and can be generalized to Riemanniandifferentiable manifolds [210].

Recall that we are interested in estimating the maximum stretching rate of trajecto-ries near a point x(t0). Let y(t) = x(t)+ δx(t) for all t . The amount of separation of theinfinitesimal perturbation δx(t0) (with arbitrary orientation) after the time τ is given by

δx(t0+ τ )= φτt0 (y(t0))−φτ

t0(x(t0))= Dφτt0 (x(t0))δx(t0)+O

(‖δx(t0)‖2), (8.16)

Therefore, the growth of a small linearized perturbation δx(t0) in the L2 norm is given by

‖δx(t0+ τ )‖ = ∥∥Dφτt0 (x(t0))δx(t0)

∥∥= ‖Dφτ

t0 (x(t0))δx(t0)‖‖δx(t0)‖ ‖δx(t0)‖< ‖Dφτ

t0(x(t0))‖ ‖δx(t0)‖ (8.17)

Notice that the separation of trajectories near a point x0 is controlled by ‖Dφt0+τt0 (x0)‖ and

it not only depends on position and time but also the integrating time τ . Depending onthe systems, the integrating time τ has to be properly chosen in order to reveal meaningfulcoherent structure and this will be demonstrated later on in the chapter.

The finite-time Lyapunov exponent (FTLE) is defined as

στ (x0)= 1

|τ | ln(

maxδx0

‖Dφt0+τt0 (x0)δx0‖‖δx0‖

). (8.18)

In the above definition, FTLE at x is taken to be the maximum Lyapunov exponent amongall orientations. So, how will we choose the orientation of the infinitesimal perturbation toobtain the maximum stretching? By considering the matrix Dφ

t0+τt0 as a linear operator, the

largest infinitesimal stretching along the solution of (8.15) starting at x0 is, therefore, givenby the largest singular value of Dφ

t0+τt0 . We introduce a symmetric matrix

Ct0+τt0 (x0)= (Dφ

t0+τt0 (x0))T Dφ

t0+τt0 (x0). (8.19)

Denote by ξ1(x0, t0,τ ) . . . ξn(x0, t0,τ ) an orthonormal eigenbasis of Ct0+τt0 (x0) with the cor-

responding eigenvalues

0 < λ1(x0, t0,τ ) . . .≤ λn−1(x0, t0,τ )≤ λn(x0, t0,τ ). (8.20)

Page 257: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 249

Note that in the language of fluid dynamics, the matrix Ct0+τt0 is called the Cauchy-Green

deformation tensor. The maximum expansion of infinitesimal perturbation at the point x0can then be obtained when its orientation is aligned with the eigenvector ξ1(x0, t0,τ ), i.e.,

maxδx(t0)‖δx(t0+ τ )‖ =√λn(x0, t0,τ )‖δx(t0)‖ (8.21)

Therefore, we may write the FTLE of the initial point x0 as

στ (x0)= 1

|2τ | lnλn(x0, t0,τ ). (8.22)

We want to stress that the leading singular vector as the solution of the optimization(8.21) has a close relationship to the true leading Lyapunov vector, denoted by ζ1(t), butthey are conceptually different. The latter has to be defined as the vector toward which allperturbation δx(t0− s) for any s > 0, any perturbation started at a long time s before t0,must converge, that is,

ζ1(t0)= lims→∞Dφ

t0t0−sδx(t0− s). (8.23)

The definition of the leading Lyapunov vector can be pictorially described as shown inFigure8.4.

Figure 8.4. A random pertubations are attracted toward the leading Lyapunovvector as they evolve.

The leading singular vector of Dφt0+τt0 (x0)), however, is merely a finite-time estimate

of the leading Lyapunov vector. In fact, any singular vectors, unless lying parallel exactlywith another Lyapunov vector, must also approach the leading Lyapunov vector under theevolution by the tangent linear dynamic, but they are initially off the leading Lyapunovvector and could have an initial growth that is higher than the maximum Lyapunov exponentof the system. As an example, consider a 2D linear map(

x1(t+ τ )x2(t+ τ )

)=(

2x1(t)+16x2(t)0.9x2(t)

), (8.24)

which has the constant tangent linear map for any t0 given by a 2×2 matrix:

Dφt0+τt0 =

(2 160 0.9

)(8.25)

Page 258: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

250 Chapter 8. Finite Time Lyapunov Exponents: FTLE

Since the tangent linear map is constant, the leading Lyapunov vector is, by the definition,the leading eigenvector of (8.25); thinking of the power method for approximating thelargest eigenvalue can give a clue to this claim. Therefore, the largest Lyapunov exponent islog(2). The singular values of (8.25) are

√λ1 = 4.02,

√λ2 = 0.33, that is, at time t = t0+τ

the leading singular vector grows around twice as fast as the leading Lyapunov vector.However, at t = t0 + nτ for a positive integer n > 1, the singular vectors approach theleading Lyapunov vector, losing its orthogonality, and bear a similar growth rate to that ofthe Lyapunov vector, see Figure8.5.

Figure 8.5. At the initial time t0, the leading singular vector ξ1 is far apart fromthe leading Lyapunov vector ζ1. At time t0+ τ , the singular vector grows at a faster ratethan the leading Lyapunov vector and the angle between the two vectors becomes smaller.At time t0+2τ the singular vector approach the leading Lyapunov vector and their growthrate are now similar.

Note that the argument in the above example is valid due to the lack of orthogonalityamong the eigvenvectors of the tangent linear map for if Dφ

t0+τt0 is symmetric, eigenvectors

and singular vectors would coincide and a growth rate higher than the Lyapunov numberwould not be allowed. We also want to remark that the analogue of the leading Lyapunovvector in the above example for a time-dependent flow is the time-dependent LCS, whichwill be formally defined later in this chapter by using a similar concept as (8.23).

Last but not least, a possibility of a fast growth rate of the leading singular vector inthe above example hints the fact that even though the dynamic may have a very low growthrate taking as an average over time to infinity, there could be some significant transient,but perhaps long-live, structures that might be of interest in term of transport and mixingproperty of the dynamics, which can be discovered only with the finite-time estimate ofthe leading Lyapunov vector. It is this observation that we try to utilize in order to learnabout core structures governing the transport mechanism of a given dynamic. The aim ofthe FTLE here is not to estimate the leading Lyapunov vector, which may or may not havean important role in transport mechanism.

8.3.2 Algorithm

For simplicity, we present the algorithm for the two-dimensional case in this section. Anapproximation in a higher dimension is similar, see [298]. Also, the cartesian grid willbe assumed throughout this section. For dynamical systems on a more general coordinate(e.g. on a sphere or S2) non-structure grids have been recently developed by Lieken andRoss [210].

Supposed that we want to approximate the FLTE field at time t on a bounded domainD ∈R2 and choose the flow period to be τ . The FTLE can be approximated in the followingsteps:

Page 259: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 251

1. Initialize cartesian grid points (xi j (t0)), yi j (t0)) for 1≤ i ≤ m and 1≤ i ≤ n on D.

2. Advect these grid points to time t + τ by some standard integration technique toobtain (xi j (t0+ τ ), yi j (t0)+ τ ))

3. Approximate the spatial gradient of the flow map Dφτ (·) by a finite-difference tech-nique (e.g. a central difference). In particular, we have

Dφt+τt (xi j (t), yi j (t))=

⎛⎝ xi+1, j (t+τ )−xi−1, j (t+τ )xi+1, j (t )−xi−1, j (t )

,xi, j+1(t+τ )−xi, j−1(t+τ )

yi, j+1(t )−yi, j−1(t )yi+1, j (t+τ )−yi−1, j (t+τ )

xi+1, j (t )−xi−1, j (t ),

yi, j+1(t+τ )−yi, j−1(t+τ )yi, j+1(t )−yi, j−1(t ) .

⎞⎠ (8.26)

4. Based on the above approximation of the spatial gradient at each grid point, thelargest singular value of Dφτ (·) is calculated and the FTLE can obtained accordingto (8.22).

Alternatively, instead of approximating the gradient from the neighboring grid points asabove, it is possible to approximate Dφt+τ

t (·) for each grid point by choosing nearby pointsspecifically for each individual point so that one can achieve a better approximation by thefinite-difference, as shown in blue dots in Figure 8.6. However, for the purpose of the LCSvisualization, the former implementation will always capture all LCS in D whereas thelatter may miss some meaningful LCS. As previously described, the points straddling LCSwill exponentially separate forward in time and so the finite-difference of these points willbe large. However, the FTLE of LCS can rapidly decrease in the direction perpendicularto the LCS. This implies that if the grid points used in the finite-difference do not straddleLCS, the true LCS could be invisible in the FTLE field.

Figure 8.6. The LCS (shown as the dotted line) could be missed if the off-grid testpoints (shown in blue) are used to estimate the FTLE. In contrast, the LCS will always becaptures when the grid points are used in the FTLE approximate.

In addition, the approximation of LCS requires a high resolution grid since a coarsegrid could underestimate FTLE due to the folding, which is typical in nonlinear systems.

Page 260: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

252 Chapter 8. Finite Time Lyapunov Exponents: FTLE

Figure 8.7. If the grid points are too coarse, there could be an underestimation ofthe FTLE. The actual stretching (shown in green) could be much larger than the apparentstretching (shown in blue) that are calculated on the finite grid.

Figure 8.7 depicts this situation. Note that the algorithm outlined above is suitable only forthe model flow from which the velocity vectors can be determined all along the trajectoryfrom t to t+ τ . However, in case of a finite data set where the velocity field is limited ontoa finite domain for some finite period of time, the algorithm is similar in principle but itrequires a treatment of an issue, in which the trajectory of a point could invariably leavethe domain of data and so integration cannot be continued.

8.3.3 Example 1: Duffing equation

Consider again the Duffing equation of the form

dx

dt=−y

dy

dt= x− x3

(8.27)

as our benchmark example. We compute the FTLEs both forward and backward in timeusing 22500 sample points taken from the domain [−1,1]× [−1,1] and with the time inte-gration of |τ | = 5. Visualization of FLTEs is shown in Figure 8.8.

To visualize the LCS, we need to approximate the ridge of the FTLE fields. In gen-eral, this may be a difficult task to extract smooth ridges from FLTE fields. However, it issufficient in this example to visualize the ridges by thresholding the FTLE field, see Fig-ure 8.9. Here we plot the attracting LCS computed from τ = 5 in red and the repelling LCScomputed from τ =−5 in blue.

Page 261: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 253

0 0.5 1

0

0.5

1

FTLE, τ = 5

0 0.5 1

0

0.5

1

FTLE, τ

0.5 1 1.5 2 2.5 3 3.5 4 4.5

Figure 8.8. [Left] The FTLE of Duffing equation with forward-time integration,τ = 5. The skeleton of this FTLE corresponds to the stable manifold of the hyperbolic fixedpoint at the origin. [Right] The FTLE calculated with τ = −5, which reveal the unstablemanifold emerging from the fixed point at the origin.

8.3.4 Example 2: Periodically-forcing Double Gyre

Consider the following autonomous system

dx

dt=−∂ψ

∂x= Aπ sin(πx)cos(πy)

dy

dt= ∂ψ

∂y= Aπ cos(πx) sin(πy),

(8.28)

where the flow trajectory is bounded in the domain [0,2]×[0,1]. The velocity field consistsof two gyres rotating in the opposite direction and divided by a separatrix in the middlex = 1, see Figure 8.10. The system consists of two hyperbolic fixed points; (1,0) and (1,1).A key feature of this system is that the unstable manifold of the fixed point (1,1) coincideswith the stable manifold of the fixed point (1,0) and they form the separatrix, which acts asa boundary of the two gyres. In other words, the initial particles on the left of the separatrixwill never transverse across the separatrix to the right and likewise for those on the right.

The forward and backward FTLE is shown in Figure 8.11 for various integratingtime. When τ is too small, the separatrix is not manifested since most sample points donot experience enough separation. On the other hand, when τ is too large, most samplepoints become decorrelated and their FTLEs approach their asymptotic value to becomeindistinguishable.

Page 262: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

254 Chapter 8. Finite Time Lyapunov Exponents: FTLE

0.5 0 0.5 1

0

0.2

0.4

0.6

0.8

1

Figure 8.9. Only the grid points with FTLEs above 85% of the maximum FTLEsare plotted and the FTLEs of the remaining grid points are suppressed to zero.

8.3.5 Example 3: Periodically-forcing Double Gyre

Although a periodic, Hamiltonian system can be thoroughly understood by some tradi-tional tools such as lobe dynamics, Melenikov methods or using Poincare’s map. For apedagogical reason, we extend the previous example to demonstrate the validity of theFTLE analysis when applying to a periodic system. The periodically-forcing double gyreflow is described by:

dx

dt=−∂ψ

∂x= Aπ sin(π f (x))cos(πy)

dy

dt= ∂ψ

∂y= Aπ cos(π f (x)) sin(πy)

d f

dx,

(8.29)

where the flow trajectory is bounded in the domain [0,2]× [0,1] and A determines thevelocity magnitude. The time-dependent stream function ψ(x , y, t) and the forcing function

Page 263: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 255

0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 20

0.2

0.4

0.6

0.8

1

Figure 8.10. The motion of the autonomous double gyre velocity field. The greenpoints are the hyperbolic fixed points. The unstable manifold of the fixed point at the topboundary coincides with the stable manifold of the fixed point the lower boundary, whichthe separatrix (the red line) to divide the two gyres.

0 0.5 1 1.5 20

0.5

1τ=1

0 0.5 1 1.5 20

0.5

1τ=5

0 0.5 1 1.5 20

0.5

1τ=20

0 0.5 1 1.5 20

0.5

1τ=1

0 0.5 1 1.5 20

0.5

1τ=5

0 0.5 1 1.5 20

0.5

1τ=20

BackwardFTLE

ForwardFTLE

Figure 8.11. The forward and backward FTLE field for various integrating time.

f (x , t) in the above equation are given by:

ψ(x , y, t)= A sin(π f (x , t)) sin(πy)

f (x , t)= ε sin(ωt)x2+ (1−2ε sin(ωt))x .(8.30)

At t = 0 the separatrix is in the middle of the two gyres and its periodic motion is governedby f (x , y, t), in which the maximum distance of the separatrix away from the middle x = 1is approximated by ε.

Since the separatrix is connected to the two (Eulerian) periodic points on the bound-ary y = 0 and y = 1 for each time t , we can track the motion of the separatrix via theconditions that dx

dt = 0 for all t and y and that at t = 0 the separatrix, denoted now by x , is

Page 264: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

256 Chapter 8. Finite Time Lyapunov Exponents: FTLE

at x = 1. These conditions are satisfied when

x = 1+√

1+4ε2 sin2(ωt)−1

2ε sin(ωt)

≈ ε sin(ωt) for small ε.

(8.31)

In this example, we set ω = 2π , A = 0.25 and ε = 0.25. Figure 8.12 demonstratethe periodic motion of the separatrix using the above parameters. To visualize the forwardand backward FTLE field at time t on the same picture, we superimpose them by plottinga function

F(x)={στ (x(t)) if στ (x(t)) > σ−τ (x(t))

−σ−τ (x(t)) if στ (x(t))≤ σ−τ (x(t))(8.32)

Approximation of LCS by FTLE ridges may be visualized by plotting F(x) as in 8.32, seeFigure 8.13 for the flow time τ = 5, which is deliberately chosen to be short enough toreveal the lobe dynamics. Increasing the flow time to τ = 40 reveals different aspect of thedynamics of the double gyre system, namely, the KAM islands, as shown in Figure 8.14 ina comparison with the stroboscopic map. The result in Figure 8.14 agree with the fact thatthe KAM islands are invariant region and enclosed by the invariant manifold.

0 0.5 1 1.5 20

0.5

1t = 0

0 0.5 1 1.5 20

0.5

1t = 0.25

0 0.5 1 1.5 20

0.5

1t = 0.5

0 0.5 1 1.5 20

0.5

1t = 0.75

Figure 8.12. The motion of the double gyre velocity field with the period of T = 1.The location of the separatrix can be approximated by (8.31).

Page 265: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 257

Figure 8.13. The approximation of the Lagrangian coherent structure by theFLTE field. The unstable material line (stable manifold) is represented in red and thestable material line (unstable manifold) is represented in blue. Observed that the initialpoint inside the lobe L5− L7 stretch much faster than the others when advecting forwardin time. In particular, as the lobe moving toward the (Eulerian) fixed point at the bottomboundary, it will undergo an exponential expanding in a similar manner of those pointsinitially straddling the stable manifold of a fixed point in the time-independent system.

8.3.6 Example 4: ABC flow

The ABC (Arnold-Beltrami-Childress) flow is given by:

dx

dt= A sin(z)+C cos(y)

dy

dt= B sin(x)+ A cos(z)

dz

dt= C sin(y)+ B cos(x),

(8.33)

where the domain is defined to be a torus 0≤ x , y, z≤ 2π . This flow is known for exhibitinga complicated geometrical structure of stable and unstable invariant manifold of hyperbolicperiodic orbits, which forms a barrier of the invariant sets. The FTLE for this flow hasbeen calculated in[136] for a choice of A =√3, B =√2,C = 1 for a comparison with theinvariant sets, see Figure 8.15. The result demonstrates that, for this flow, the FTLE fieldagree with the barriers of the invariant sets.

In fact, it has been shown that there exists exactly six invariant sets for this ABCflow [55, 155], although it is not clear by observing the 3D FTLE field. All of these sixinvariant sets can be approximated by using six dominated eigenvectors of the transitionmatrix, see 8.16.

Page 266: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

258 Chapter 8. Finite Time Lyapunov Exponents: FTLE

Figure 8.14. The areas enclosed by the stable/unstable material lines agree verywell with the invariant tori, which is manifested by the Poincare plot.

8.3.7 Example 5: Stratospheric coherent structures in SouthernHemisphere

The “edge" of the Antarctic polar vortex is known to behave as a barrier to the meridional(poleward) transport of ozone during the austral winter [269]. This chemical isolation ofthe polar vortex from the middle and low latitudes produces an ozone minimum in thevortex region, intensifying the ozone hole relative to that which would be produced byphotochemical processes alone. Observational determination of the vortex edge remainsan active field of research. In this example, we will apply the FTLE technique to EuropeanCentre for Medium-Range Weather Forecasts (ECMWF) Interim two-dimensional velocitydata on the isentropic surface of 850 Kelvin during September 2008 in order to identify the

Page 267: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 259

Figure 8.15. A comparison between the FTLE field, shown as the dark curve in(a), and some of the invariant sets in (b). A good agreement between the two is evident.Courtesy of [136]

Figure 8.16. All six invariant sets of the ABC flow

poleward transport of ozone. Traditionally, the contour curve where the potential vorticity(PV) decreases steepest is heuristically used to define the transport barrier. Figure 8.17(a)shows the PV obtained from ECMWF Interim data on the 475 Kelvin isentropic surface onSeptember 14 2008.

We compute backward and forward FTLE on September 14 2008 with the flow du-ration of 14 days and the results are shown in Figure 8.17(b), which may be comparedwith Figure 8.17(a) to see their similarity. In order to compare the FTLE result with thePV-determined coherent structures (the vortex “boundary"), we use a common approach,developed in [226, 247, 308] to subjectively define the vortex boundary as the PV contour

Page 268: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

260 Chapter 8. Finite Time Lyapunov Exponents: FTLE

14 September 2008

60oS

30oS

0

(a) Potential vorticity (PV)

14 September 2008

60oS

30oS

(b) The ridges of FLTEs

Figure 8.17. Comparison of PV boundary and the ridge of FTLEs. (a) The poten-tial vorticity in the PV units (PVUs) (1 PVU = 10−6K m2kg−1s−1) obtained from ECMWFInterim data on the isentropic surface of 475 Kelvin on September 14 2008. The “edge"of the PV, shown as the green curve where the gradient is largest, is conventionally usedto define the barrier to the poleward transport during the austral winter. (b) The ridge ofFTLE fields computed from ECMWF Interim velocity data on the isentropic surface of 475Kelvin on September 14 2008 with a flow duration of 14 days.

with the highest gradient w.r.t. the equivalent latitude. Briefly, the equivalent latitude,φe,is defined from the area A enclosed by a given PV contour as A = 2π R2(1− sinφe) [247].A transport boundary is then defined to be the location of the highest gradient in PV, asshown in Figure 8.18. The result of the vortex boundary is plotted in Figure 8.17(a) as the

Page 269: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 261

green curve.P

VU

Equivalent Latitude

0

1

2

3

4

5

Equivalent Latitude

PV

gra

dien

t

Maximum PVgradient (at ~66.9Equivalent Latitude)

Figure 8.18. [Top] Potential vorticity plotted in term of the equivalent latitude inthe unit of PVU. [Bottom] PV gradient w.r.t. the equivalent latitude in the unit PVU perdegree.

Note that the computational cost of the vortex boundary is relatively low, as the PVvalues have already been provided by the ECMWF Interim data.

Figure 8.17(b) compares the attracting and repelling LCSs with the PV boundary.Interestingly, the LCSs tightly wrap around the PV boundary and the lobes formed by theinterlacing of LCSs occurs mainly outside the PV boundary. This suggests that particles(ozone in this case) are completely trapped inside the PV boundary and transport occursonly in the vicinity of the mid-latitude region where the LCSs intricately interlace. In thestratosphere literatures, such region is sometimes referred to as a “stochastic" layer [188].

8.3.8 Relevance of FLTE as a tool for LCS detection

In previous examples, we have heuristically used the FTLE-maximizing curve to detect theLCS. The results were shown to be very reasonable on the physical ground. In this section,we will be concerned with their relations based on more rigorous definitions of LCSs andridges. This formality will raise an awareness of using FTLE ridges to locate LCSs.

We begin with a mathematical definition of the ridge. More formally, a second-derivative ridge of the FTLE field is a codimension-1 manifolds that maximizes FTLE inthe transverse direction along the ridge [298]. The (second-derivative) ridge here is defined

Page 270: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

262 Chapter 8. Finite Time Lyapunov Exponents: FTLE

as a parameterized curve c : (a,b)→M that satisfies two conditions. First, the derivativeof the curve c′(s) has to be parallel to the vector ∇στ (c(s)). This is to force the tangentline of the curve to be oriented in the direction of the largest variation of the FTLE field.Secondly, the direction of the steepest descent is that of the normal vector to the curve c(s),which can be expressed by

nT H (στ )n = min‖v‖=1vT H (στ )v < 0, (8.34)

where n is the unit normal vector to the ridge curve c(s) and H (στ ) is the Hessian matrixof the FTLE field:

H (στ )=⎡⎣ ∂2στ

∂x2∂2στ∂x∂y

∂2στ∂y∂x

∂2στ∂y2

⎤⎦. (8.35)

When computing in forward time (τ > 0), the ridge of the FTLE field reveals theso-called repelling Lagrangian coherent structure, which has a similar feature to the stablematerial manifold of a hyperbolic trajectory, see [163, 162, 165]. The same is true for thesimilar feature between attracting LCSs and unstable material manifold.

Roughly speaking, the repelling LCS is a structure that creates stretching in the for-ward time. As proposed in [165], a repelling LCS has two key properties:

1. It is a codimension-one invariant surface (called material surface) in the extendedphase space U ×R that “moves with the flow" and can be observed as evolvingLagrangian patterns, i.e., the LCS at an initial time t0 must evolve into the LCS atthe time t = t0 + τ under the flow φt

t0 . Formally speaking, if M(t0) ⊂ U denotesan codimension-one material surface defined at time t0, which may be viewed as anensemble of initial conditions at t0, the codimension-one material surface M(t) attime t must satisfy:

M(t)= φtt0 (M(t0)) (8.36)

Note that the “bundle" of M(t) generates an invariant manifold M of the ODE (8.15)in the extended phase space U ×R as schematically illustrated in Figure 8.19.

2. A repelling LCS should locally maximize the repulsion rate in the flow in order to bedistinguishable from all nearby trajectories that could also be material surfaces withsimilar repelling behavior due to the continuous dependence on the initial conditionsof the flow. If we denote, in the case of 2-dimensional flow, the one-dimensional tan-gent space of M(t) by Txt M(t) and the one-dimensional normal space by Nxt M(t),then the repulsion rate is defined as

ρtt0 (x0,n0)= 〈nt ,Dφt

t0 (x0)n0〉, (8.37)

which describes the growth of perturbation in the normal direction nt ∈ Nxt M(t) ofM(t), which is orthogonal to the tangent space Txt M(t). As illustrated in Figure 8.20,the tangent vector e0 ∈ Tx0 M(t0) based at a point x0 on M(t0) is mapped by the flow

Page 271: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 263

Figure 8.19. Material surface M(t)= φtt0(M(t0)) generated in the extended phase space

map Dφtt0 to a tangent vector Dφt

t0 (x0)e0 ∈ Txt M(t) at point xt on M(t). In otherwords, we have that

Tx0 M(t0)= Dφtt0(x0)Tx0 M(t0). (8.38)

However, the normal vector n0 ∈ Nx0 M(t0) is not necessarily mapped by the flowmap to a normal vector nt ∈ Nxt M(t). The repulsion rate as defined above measuresthe growth of perturbation via the orthogonal projection of the vector Dφt

t0 (x0)n0onto the normal space containing nt .

Figure 8.20. Geometry of the linearized flow map of a two-dimensional flow.

That is at any point x0 ∈ M(t0), the repulsion rate ρtt0 (x0,n0) > ρt

t0 (x0, n0) for any

t ∈ [t0, t0+ τ ], where x0 is a point on a nearby material surface M(t0) intersected bythe normal n0 and n0 is a normal vector associated with M(t0) and based at the pointx0, see Figure 8.21

Page 272: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

264 Chapter 8. Finite Time Lyapunov Exponents: FTLE

Figure 8.21. The LCS is locally defined as a material surface that maximizes therepulsion rate in the normal direction

A detailed discussion of these properties can be found in Haller [165].

Similarly, the ridge of the FTLE computed with negative integrating time (τ < 0)locates the attracting Lagrangian coherent structure, which acts as a core repelling struc-ture in the backward time; hence the previous LCS criterion are still applicable when re-garding the attracting LCS advected forward in time as the repelling LCS of backwardtime [163, 162]. It was analytically and experimentally verified by Shadden [298] that theLCS defined as the second-derivative ridge of the FTLE field is “nearly" Lagrangian in asense that the particle flux across the LCS becomes negligible as the integrating time τ

increases provided that the “true" LCS associated with the vector field is hyperbolic for alltime. However, the LCS obtained from a finite-time data set is likely to be hyperbolic onlyfor a finite time and so if the integrating time τ increases beyond the hyperbolic time of thetrajectory, which is not known a priori,the resulting LCS may not exhibit the Lagrangianproperty. In general, the ridges may not mark the true LCSs while many of published workwere using the FLTE in a non-rigorous way. The rigorous necessary sufficient conditionsfor the FTLE ridges to locate the LCSs have only recently been published in Haller [165].We refer the details of these conditions therein.

We are now concerned with some pathological examples in which the explicit ex-pression of the Cauchy-Green tensors can be explicitly established. It will be seen that theFTLE ridges do not necessarily mark the LCS.

Example 1: LCS exists but there is no FTLE ridges

Consider a decoupled, two-dimensional ODE

x = x

y =−y− y3.(8.39)

The vector field of this flow is shown in Figure 8.22 along with the stable and unstablemanifolds of the fixed point at (0,0). Clearly, the stable manifold (x = 0) here representsthe repelling LCS while the unstable manifold (y = 0) marks the attracting LCS. However,we will show that the forward FTLE field turns out to be a constant and hence there is no

Page 273: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 265

Figure 8.22. The vector field of the flow (8.39) with the attracting LCS at y = 0and repelling LCS at x = 0.

FTLE ridges. Since the vector field is autonomous and de-coupled, we may assume thatt0 = 0 and solve for the trajectory for a given initial condition (x0, y0) as

x(t)= x0e2t

y(t)= y0√(y2

0 +1)e2t− y20

(8.40)

For this example, the Cauchy-Green tensor as defined in (8.19) can be written as

�τ =⎛⎜⎝[∂x(t ,0,x0)

∂x0

]20

0[∂y(t ,0,y0)

∂y0

]2

⎞⎟⎠= ( e2t 0

0 e4t

[(y20+1)e2t−y2

0 ]3 .

)(8.41)

Therefore, for t > 0, the maximum eigenvalue is

λmax(t)= e2t ,

which, according to (8.22), leads to a constant FTLE field

σt (x(t))≡ 1. (8.42)

This implies that there is no ridges for all t > 0. An illustration of this misconception isshown in Figure8.23.

In contrast, for t < 0, the maximum eigenvalue becomes

λmax(t)= e4t

[(y20 +1)e2t− y2

0]3,

which depends both on the flow time t and the value of y0. Thus, the backward FTLE fielddepends only on y. To check the extremum of the backward FTLE we calculate

∂y0λmax(y0, t0)=−6

(e2t −1

)e4t y0(

e2t (y20 +1)− y2

0

)4∂2

∂y20

λmax(y0, t0)=−6(e2t −1

)e4t e2t −7y2

0e2t +7y20(

e2t (y20 +1)− y2

0

)5 .

Page 274: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

266 Chapter 8. Finite Time Lyapunov Exponents: FTLE

Figure 8.23. A saddle flow without a FTLE ridge that has a repelling LCS at x = 0

It is now easy to check that for t < 0 and y0 = 0, ∂∂y0

λmax(0, t0)= 0 and ∂2

∂y20λmax(0, t0) > 0.

It can then be concluded that the backward FTLE field has a minimum ridge (through) atthe x− axis, which coincides with the attracting LCS as demonstrated in Figure8.24. Notethat in this example we accepted the repelling and attracting LCSs based on our intuitionbut have not done a formal validation based on those LCSs conditions in [165]. In fact, itwas shown that the x− axis is not attracting LCS but only “weakly" LCS as defined therein.

Figure 8.24. An illustration of the scenario in which the attracting LCS at y = 0appears as the FTLE through instead of a ridge.

Example 2: FTLE ridges that is neither an attracting nor a repelling LCS

Consider the two-dimensional area-preserving system

x = 2+ tanh y,

y = 0.(8.43)

Since y = 0, the vector is just a flow parallel to the x-axis as illustrated in Figure. Next,we will show that the x−axis is indeed a ridge of both forward and backward FTLE in thisexample. Again, we may assume t0 = 0, and hence the trajectories of (8.43) are given by

x(t)= x0+ t(tanh y0+2)

y(t)= y0.(8.44)

Page 275: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

8.3. FTLE and LCS 267

The maximum and minimum eigenvalues of the Cauchy-Green tensor can then be calcu-lated as

λmax(t)= 1

2t2[sech(y0)]4+ 1

2

√(t4[sech(y0)]8+4+1

λmin(t)= 1

2t2[sech(y0)]4− 1

2

√(t4[sech(y0)]8+4+1.

(8.45)

The hyperbolicity of the trajectory can be readily seen from the above equation since wehave the relation

logλmin(t) < 0 < logλmax(t). (8.46)

Also, the sech(y0) attains a unique maximum at y0 = 0 and so does the quantityσt (x(0; t , x0, y0)), independent of x0. Therefore, for any t �= 0, the maximal ridge of theσt (x(0; t , x0, y0)) field is the x−axis with a constant height along the ridge.

However, since the flow in this example is area-preserving, the distance in the per-pendicular to the ridge (x− axis in this example) is constant for all times (which should alsobe intuitively clear by the geometry of the flow). Therefore, the ridge is neither a repellingnor an attracting LCS.

8.3.9 Conditions for using FTLE ridge to identify LCS

Below we present the sufficient and necessary conditions for a Hyperbolic LCS based onthe FTLE ridge, which has been proved in[165].

Theorem 8.1. Assume that M(t0)⊂U is a compact FTLE ridge. Then M(t)= φtt0 [M(t0)]

is a repelling LCS over the time interval [t0, t0+ T ] if and only if at each point x0 ∈ M(t0)the following conditions are satisfied:

1. un(x0, t0, T )⊥ Tx0 M(t0),

2. λn−1 �= λn(x0, t0, T ) > 1,

3. 〈∇λn(x0, t0, T ),un(x0, t0, T )〉 = 0,

4. The matrix L(x0, t0, T ) is positive definite, where

L=

⎛⎜⎜⎜⎜⎝∇2C−1[ξn ,ξn ,ξn ,ξn] 2 λn−λ1

λnλ1〈ξ1,∇ξnξn〉 · · · 2 λn−λn−1

λnλ1〈ξn−1,∇ξnξn〉

2 λn−λ1λnλ1

〈ξ1,∇ξnξn〉 2 λn−λ1λnλ1

· · · 0...

.... . .

...2 λn−λn−1

λnλ1〈ξn−1,∇ξnξn〉 0 · · · 2 λn−λ1

λnλ1

⎞⎟⎟⎟⎟⎠Proof. See[165].

The first diagonal term in L is the second derivative of the inverse Cauchy-Greentensor, which is equal to

∇2C−1[ξn ,ξn ,ξn ,ξn]=−1

2〈ξn ,∇2λnξn〉+2

n−1∑q=1

λn−λq

λqλn〈ξn ,∇ξnξn〉2 (8.47)

Page 276: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

268 Chapter 8. Finite Time Lyapunov Exponents: FTLE

The first condition implies that the normal vector n0 at x0 should align with the ξn(x0, t0, T ),the eigenvector corresponding to the largest eigenvalue λn . The second condition states thatthe largest eigenvalue must be greater than one and multiplicity one to ensures the domi-nance of the growth normal over the growth in tangent direction. Together with the secondcondition, the third condition, which can be interpreted as the directional derivative of λnfield in the direction of un , complete the stationary requirement for the normal repulsionrate. In the last condition, the required positive definiteness of L, which is not as intuitiveas the first two conditions, arises from the variational argument in the proof of this theoremin [165]. Nevertheless, it is shown in [194] that this positive definite condition is equiv-alent to the condition of 〈∇un(x0),λ2

n(x0, t0, T )un(x0, t0, T )〉 < 0. In words, this conditionensures that the repulsion rate, hence λn , has a non-degenerate maximum in the normaldirection of the hyperbolic LCS.

Page 277: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Chapter 9

Information Theory inDynamical Systems

In this Chapter, we outlined the strong connection between dynamical systems, and a sym-bolic representation through symbolic dynamics. The connection between dynamical sys-tems, and its sister topic of ergodic theory can also be emphasized through symbolizationby using the language inherent in information theory. “Information" as described by theShannon information theory begins with questions regarding code length necessary for aparticular message. Whether the message be a poem by Shakespeare, a raster scanned oreven a multi-scaled (wavelet) represented image, or even an initial condition and its trajec-tory describing the evolution under a given dynamical process, the language of informationtheory proves to be highly powerful and useful. In the first few sections of this chapter, wewill review just enough classical information theory in order to tie together some strongconnections to dynamical systems and ergodic theory in the later sections.

9.1 A Little Shannon Information on Coding by ExamplePutting the punchline first, we will roughly state that information is defined to describe adecrease in uncertainty. Further, less frequent outcomes confer more information about asystem. The Shannon entropy is a measure of average uncertainty.

Basic questions of data compression begins a story leading directly to Shannon en-tropy and information theory. We shall introduce this story in terms of representation ofa simple English phrase, for easy presentation. The discussion applies equally to phrasesin other human languages, representations of images, encodings of music, computer pro-grams, etc.

The basic idea behind entropy coding is the following simple principle,

• We assign short code-words to likely frequent source symbols and long codewordsto rare source symbols.

• Such source codes will therefore tend to be variable-length.

• Since long code words will be assigned to those sources which are less likely theyare therefore more “surprising", and conversely for short codes are less surprisingoutcomes.

269

Page 278: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

270 Chapter 9. Information Theory in Dynamical Systems

Consider a phrase such as the single English word, “Chaos"; we choose a singleword phrase only to make our presentation brief. This story would be the same if we wereto choose a whole book of words, such as this book in your hands. Encoded in standardASCII116,

Chaos→ 1000011 1101000 1100001 1101111 1110011. (9.1)

A space shown is indicated between each 7-bit block used to denote each individual letters,but the spacing here is just for convenience and clarity to the reader. Notice that in this7-bit version of standard ASCII coding, it takes 5× 7 = 35 bits to encode the 5 letters inthe word “Chaos", so stated including the uppercase beginning letter,

“C"→ 1000011, (9.2)

versus what would be the lowercase in ASCII,

“c"→ 1100011. (9.3)

ASCII is a useful code in that it is used universally on computers around the world.If a phrase is encoded in ASCII, then both the coder and the decoder at the other endwill understand how to translate back to standard English, using a standard ASCII table.The problem with ASCII however is that it is not very efficient. Consider that in ASCIIencoding,

“a"→ 1111010,“z"→ 1100001. (9.4)

Both the “a" and the “z" have reserved the same 7 bit allocation of space in the encoding,in this ASCII encoding that was designed specifically for English. Now if it were designedfor some language where the “a" and the the “z" occurred equally frequently, this would befine, but in English, it would be better if the more frequently used letters, such as vowelsincluding the “a" could be encoded with 1 or 2 bit words, say, and those which are rarelyused like the “z" (or even more so the speciality symbols such as $ or &, etc.), might bereasonably encoded with many bits. On average such an encoding would do well whenused for the English language for which it was designed.

Codes designed for specific information streams, or with assumed prior knowledgeregarding the information streams can be quite efficient. Amazingly, encoding efficienciesof better than 1-bit/letter may even be feasible. Consider the (nonsense) phrase with 20characters,

“chchocpohccchohhchco". (9.5)

In ASCII it would take 20∗7= 140 bits for a bit rate of 7 bits/letter. However, in a Huffmancode117 we can do much better. Huffman coding requires a statistical model regarding theexpected occurrence rate of each letter. We will take as our model,118

p1 = P(“c")= 0.4, p2 = P(“h")= 0.35, p3 = P(“o")= 0.2, p4 = P(“p")= 0.05, (9.6)

116American Standard Code for Information Interchange (ASCII) is a character-encoding of the Englishalphabet used commonly in computers. Each character gets the same length of 7-bit despite that somecharacters are not likely to be used.

117The Huffman code is a variable-length code algorithm that is in some sense optimal, as discussed inSec. 9.2, especially in Theorem 9.5. This breakthrough was developed by a then MIT student D.A. Huffmanpublished in his 1952 paper, [179].

118The notation P(A) denotes the probability of event A.

Page 279: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.1. A Little Shannon Information on Coding by Example 271

which we derive by simply counting occurrences119 of each letter to be, 8, 7, 4, and 1respectively and assuming stationarity120. With these probabilities, it is possible the fol-lowing Huffman code follows,

“c"→ 0,“h"→ 10,“o"→ 110,“p"→ 111, (9.7)

from which follows the Huffman encoding,

“chchocpohccchohhchc"→0 10 0 10 110 0 111 110 10 0 0 0 10 110 10 10 0 10 0 110. (9.8)

Again, spaces are used here to guide the eye separating bits related to each individualletter. The spaces are not actually part of the code. In this encoding, the 20 letter phrase“chchocpohccchohhchc" is encoded in 37 bits, for a bit rate of 37 bits/ 20 letters=1.85bits/letter which is a great deal better than what would have been 7 bits/letter in a 140 bitASCII encoding.

We have not given the details behind how to form a Huffman code from a givendiscrete probability distribution as this would be somewhat outside of the scope of thebook and beyond the need of our bring up the topic here. Our point is simply that there aremuch better ways to encode an information stream by a well chosen variable-length codeas exemplified by the well regarded Huffman code. The dramatic improvement of coursecomes from this “pretty"-good statistical model used. For example, zero bit resource isallocated for the letter “z", and the rest of the alphabet. Any message that requires thosesymbols would require a different encoding or the the encoding simply is not possible.

How shall the quality of a coding scheme such as the Huffman code be graded? Con-sidering the efficiency of the encoding is a great deal like playing the role of a bookie,121

hedging bets. Most of the bit resource is allocated for the letter “c" since it is most com-mon, and the least is allocated for the “o" and “p" since they are less common of those used.When a message is stored or transmitted in agreement with this statistical model, then highefficiency occurs and low bit rates are possible. When a message that is contrary to themodel is transmitted with this then ill fitted model, then less efficient coding occurs. Con-sider a phrase, “hhhhhhhhhhhhhhhhhhhh" (20 times “h").122 Then a 3 bit/letter efficiencyoccurs which is a worst case possible scenario with this particular Huffman coding. Thatis still better than ASCII, because the model still assumes only 4 different letters mightoccur. In other words, it still beats ASCII since at the outset we assumed that there is zeroprobability for all but those 4 symbols.

119Without speaking to the quality of the model, note that counting occurrences is surely the simplest wayto form a probabilistic model of the likelihood of character occurrences.

120Roughly stated, a stochastic process is stationary if the joint probabilities “do not change in time." Morewill be said precisely in Definition 9.19 in Sec. 9.5.

121A “bookie" is a person who handles bets and wagers on events, usually sporting events on which gam-blers place money in hopes that their favorite team will win and they will win money. The bookie needs agood probability model if they expect to win over many bets on average.

122Considering another coding method entirely, called run length encoding allows the “h" to be repeatedsome number of times, [274]. The longer the repeat, the more efficient such an encoding would be sincethe overhead of the annotations to repeat are amortized; the coding of “state 20 h’s" and “state 50 h’s" areessentially the same for example. Such codes can only be useful for very special and perhaps trivial phraseswith long runs of single characters. This discussion though relates to the notion of Komolgorov Com-plexity which is defined to be the length of the optimal algorithm which reproduces the data. KomolgorovComplexity is not generally computable.

Page 280: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

272 Chapter 9. Information Theory in Dynamical Systems

To make the notion of bit rate and efficiency more rigorous, note that the statisticalexpectation of the bit rate in units of bits/letter may be written,

Average bit rate= bits/letter

=∑

i

P(ith letter occurs in message) (Length used to encode ith letter)

= 0.4∗1+0.35∗2+0.2∗3+0.05∗3bits/letter

= 1.85bits/letter. (9.9)

The perfect coincidence in this example between expectation and actual encoding rate sim-ply reflects the fact that the toy message used matches the probabilities. The optimal bitlength of each encoding can be shown to be bounded,

(Length used to encode ith letter)≤− log2(pi ). (9.10)

See the relation of this statement to the optimal code rate implicit in Theorem 9.6.Considering an optimal encoding of a bit stream leads to what is called the Shannon

entropy, defined formally in 9.4, specialized for a coding with 2 outcomes (bits).

H2 :≡−∑

i

pi log2(pi ). (9.11)

Shannon entropy carries the units of bits/letter, or alternatively bits/time if the letters areread at a rate of letters/time. Comparing this to the question of how long is the coding inthe previous example, Eq. (9.6),

H2 =−0.4log2(0.4)−0.35log2(0.35)−0.2log2(0.2)−0.05log2(0.05)= 1.7394. (9.12)

Note that 1.7394 < 1.85 as the code used was not optimal. The degree to which it was notoptimal is the degree to which in Eq. (9.10),

− [(Length used to encode ith letter)+ log2(pi )] > 0. (9.13)

Specifically, notice that p3 = 0.2 > p4 = 0.05, but they each are “gambled" by the bit ratebookie who allocates 3 bits each. There are reasons for the suboptimality in this example,

• The probabities are not d-atic (d-atic with d = 2, also spelled dyatic in this case,means pi = r

2n for some r ,n integers and for every i ).

• A longer version of a Huffman coding would be required to differentiate these prob-abilities. Here only three bits were allowed at maximum. Huffman codes are devel-oped in a tree and depth is important too.

• Furthermore, the Huffman coding is a nonoverlapping code [179], meaning eachletter is encoded before the next letter can be encoded. Huffman code is a specialexample of so-called entropy coding within the problem of lossless123 compression.

123The object of lossless encoding to be able to recover the original “message" exactly, as opposed to lossyencoding in which some distortion is allowed. For example, consider a computer program as the source.A “zipped" file of a computer program must be decompressed exactly as the original in order that the de-compressed computer program might still work. On the other hand, lossless compression generally borrowsfrom representation theory of functions; a lossy scheme for compressing a digital photograph includes trun-cating a Fourier expansion. Some loss is acceptable for images in some applications but loss is entirelyunacceptable in other applications such the compression of a computer program.

Page 281: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.2. A Little More Shannon Information on Coding 273

Huffman codes are optimal within the class of entropy coding methods, beating theoriginal Shannon-Fano code for example, but it is not as efficient as the Lempel-Ziv(LZ)[338] or arithmetic coding methods, [207, 71]. Our purpose in using the Huff-man coding here was simply a matter if specificity and simplicity of presentation.

Important aspects of a proper coding scheme is at least that it must be,

• One-to-one and therefore invertible.124 Said otherwise, for each coded word, theremust be a way to recover the original letters so as to rebuild the word. Without thisrequirement, we could quite simply “compress" every message no matter how long,say the complete the works of Shakespeare, to the single bit “0", and not worry thatyou cannot go back based on that bit alone. The information is lost as such.

• Efficient. A poor coding scheme could make the message length longer than it mayhave been in the original letter coding. This is a legal feature of entropy encoding,but of course not useful.

The details of these several coding schemes are beyond the scope here, but simplyknowing of their existence as related to the general notions of coding theory is leading usto the strong connections of entropy in dynamical systems. Sharpening these statementsmathematically a bit further will allow us to discuss the connection.

9.2 A Little More Shannon Information on CodingThe Shannon entropy HD(X) defined 9.4 can be discussed in relationship to the questionof the possibility of an optimal code, as the example leading to Eq. (9.9) in the previoussection reveals. To this end, we require the following notation and definitions.

Definition 9.1. An Encoding c(x) for a random variable X (See Definition 3.2) is a func-tion from the countable set {x} of outcomes of the random variable to a string of symbolsfrom a finite alphabet, called a D-ary code.125

Remark 9.1. Commonly in digital computer applications which are based on binary bits,representing “0" and “1", or “on" and “off", generally D = 2.

Following the above discussion, it is easy to summarize with,

Definition 9.2. The expectation of the length L of a source encoding c(x) of the randomvariable X with an associated probability distribution function p(x) is given by,

L(C)=∑

x

p(x)l(c(x)), (9.14)

where l(c(x)) is the length of the encoding c(x) in units of bits. In the previous section,Eq. (9.9), we described the units of L to be bits/letter, however it can be interpreted alsosimply as length when a fixed positive number C of letters are coded.

124A code is called non-singular if for every outcome, there is a unique representation by a string from thesymbol set, ({0,1} if binary), and otherwise it is called singular.

125 D-ary refers to the use of D symbols, which may be taken from the symbol set, {0,1, .., D−1} or {0,1}the usual binary set if D = 2. Arithmetic occurs in base-D.

Page 282: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

274 Chapter 9. Information Theory in Dynamical Systems

Given an encoding, and repeated experiments from a random variable, we can get acode extension which is simply an appending of codes of each individual outcome.

Definition 9.3. Given a code c(x), an encoding extension is a mapping from orderedstrings of outcomes xi to an ordered string of symbols from the D-ary alphabet of the code,

C = c(x1x2...xn)≡ c(x1)c(x2)...c(xn). (9.15)

This is a concatenation of the alpabet representation of each outcome in the sequencex1, x2, ...., xn.

The formal definition of Shannon entropy can be stated,

Definition 9.4. The Shannon Entropy of a D-ary code for a random X with probabilitydistribution function p(x) is given by the following nonegative function,

HD(X)=−∑

x

p(x)logD p(x), (9.16)

in terms of the base-D logarithm.

There is a strong connection between the notion of optimal coding, and this definitionof entropy as revealed by the following classic theorems from information theory, as provenin [71]. Discussion of existence of optimal codes follows starting from the Kraft inequality,

Theorem 9.5. An instantaneous code126 C of a random variable X with code word lengthsl(x) satisfies the inequality ∑

x

D−l(x) ≤ 1. (9.17)

By converse, the Kraft inequality implies that such an instantaneous code exists.The proof of this second statement is by Lagrange multiplier optimization methods [71].Furthermore a statement relating Shannon entropy and expected code length L(C) can besummarized,

Theorem 9.6. Given a code C of a random variable X with probability distribution p(x),C is a minimal code if the code words lengths are given by

l(x)=−logD p(x). (9.18)

The following definition of Shannon information may be described as a pointwiseentropy in that it describes not only the length of a given optimally coded word, but it alsodescribes the entropy as if we know that the random variable takes on X = x and thereforethe corresponding code word is used.

126An instantaneous code, or synonymously a prefix code completes each code word before the nextword begins. No code word is a prefix of any other code word. Therefore the receiver does not requirea prefix before each word to know when to start reading the next word. The converse is that two wordsshare the same coding and therefore a prefix would be required to distinguish them. Since perhaps the mostcommonly used prefix code is the Huffman code, favored for its optimality properties, often by habit evenother source codes are called “Huffman" by some.

Page 283: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.3. Many Random Variables and Taxonomy of the Entropy Zoo 275

Definition 9.7. Shannon Information is defined by the quantity,

l(x)=−logD p(x). (9.19)

Shannon information in some sense describes a degree of “surprise" we should holdwhen an unlikely event comes to pass. A great deal more information is inferred whenthe unlikely occurs than when the usual, high probability outcomes x occur. ComparingEqs. (9.14), (9.16), and (9.19), we can describe Shannon Entropy HD(X) as an informationexpectation of the random variable.

Now we can state that the relationship of an optimal code C∗ to the entropy of theprocess which gives further meaning to the notion of Shannon entropy.

Theorem 9.8. Source Coding. Let C∗ be an optimal code of a random variable X , mean-ing the expected code length of any other code C is bounded, L(C) ≥ L(C∗), and if C∗ isan instantaneous D-ary code, then

HD(X)≤ L(C∗)≤ HD(X)+1. (9.20)

Following the example in the previous section, recall that in ASCII coded versionof the 20 character message “chchocpohccchohhchc" in Eq. (9.8) is L(C)= 140, whereaswith the Huffman coded version is, L(C∗)= 37 bits. Furthermore, with the statement thatHuffman is optimal, we know that this is a shortest possible encoding by an instantaneouscode. Since Huffman coding embodies an algorithm which provides optimality, this vali-dates existence.

Theorem 9.9. A Huffman code is an optimal instantaneous code.

That Huffman is an optimal code is enough for our discussion here regarding ex-istence of such codes, and the relationship between coding and entropy. The algorithmicdetails are not necessary for our purposes in this book, and so we skip the details for sake ofbrevity. Details on this and other codes, Lempel-Ziv (LZ) [338] or arithmetic coding meth-ods [207, 71] most notably can be found elsewhere. As it turns out, other not prefix codes,notably arithmetic coding, can yield even shorter codes, which is a play on the definitionsof optimal, code and encoding extension. In brief, arithmetic coding can be understood asa mapping from letters to base-D representations of real numbers in the unit interval [0,1],distributed according to the probabilities of X , and perhaps in levels by a stochastic processX1, X2, ... generating letters. In this perspective, a Huffman code is a special case encodingone letter at a time by the same mapping process, whereas arithmetic coding maps the en-tire message all together. Thus arithmetic coding allows the reading of letters in seeminglyoverlapping fashion.

9.3 Many Random Variables and Taxonomy of theEntropy Zoo

Given many random variables,{X1, X2, ..., Xn}, (9.21)

(or two when n = 2), in a product probabilities space,

{(1,A1, P1)× (2,A2, P2)×·· · (n,An , Pn)}, (9.22)

Page 284: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

276 Chapter 9. Information Theory in Dynamical Systems

one can form probability spaces associated with the many different intersections and unionsof outcomes. See Fig. 9.1. Likewise, the associate entropies we will review here give thedegree of “averaged surprise" one may infer from such compound events.

Figure 9.1. Given a compound event process, here n = 2 with 1 and 2 asso-ciated with Eqs. (9.21) and (9.22), we can discuss the various joint, conditional and indi-vidual probabilities, as well as the related entropies as each is shown with their associatedoutcomes in the Venn diagram.

Definition 9.10. The Joint Entropy associated with random variables, {X1, X2, ..., Xn}, is,

H (X1, X2, ..., Xn)=−∑

x1,x2,...,xn

p(x1, x2, ..., xn) log p(x1, x2, ..., xn), (9.23)

in terms of the joint probability density function, p(x1, x2, ..., xn) and with the sum takenover all possible joint outcomes (x1, x2, ..., xn).

Joint entropy is sometimes called the total entropy of the combined system. SeeFig. 9.1 where H (X1, X2) is presented as the uncertainty of the total colored regions.

Definition 9.11. The Conditional Entropy associated with two random variables, {X1, X2},is,

H (X1|X2)=−∑x2

p(x2)H (X1|X2 = x2), (9.24)

in terms of the probability density function, p2(x2).

Conditional entropy H (X1|X2) can be understood as the remaining entropy bits inthe uncertainty of the random variable X1 with the information bits already given regardingthe intersection events associated with X2. See the Venn diagram Fig. 9.1. In other words,measuring H (X1|X2) answers a question, “what does X2 not say about X1?" An alternativeformula for conditional entropy may be derived in terms of the joint probabilities p(x1, p2),

H (X1|X2)=−∑

(x1,x2)

p(x1, x2) logp(x1, x2)

p(x2), (9.25)

Page 285: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.3. Many Random Variables and Taxonomy of the Entropy Zoo 277

which is easy to see since the term H (X1|X2 = x2), in Definition 9.11

H (X1|X2 = x2)=∑x1

p(x1|x2) log p(x1|x2), (9.26)

Using the relationship for conditional probabilities,

p(x1|x2)= p(x1, x2)

p2(x2), (9.27)

substitution into Definition 9.11 yields,

H (X1|X2)=−∑x2

p2(x2)H (X1|X2 = x2) (9.28)

=−∑x1,x2

p2(x2)p(x1|x2) log(x1|x2) (9.29)

=−∑

p(x1, x2) log p(x1|x2), (9.30)

the last being a statement of a cross entropy. Finally again applying Eq. (9.27)

−∑

p(x1, x2) log p(x1|x2)=−∑

p(x1, x2) logp(x1, x2)

p2(x2)(9.31)

=−∑

p(x1, x2) log p(x1, x2) (9.32)

+∑

p(x1, x2) log p2(x2)127 (9.33)

From this follows the chain rule of entropies,

H (X1|X2)+ H (X2)= H (X1, X2). (9.34)

A few immediate statements regarding the relationships between these entropies can bemade,

Theorem 9.12. H (X1|X2)=0 iff X1 is a (deterministic) function of the random variableX2.

In other words, since X1 can be determined whenever X2 is known, the status of X1is “certain" for any given X2. Also,

Theorem 9.13. H (X1|X2)= H (X1) iff X1 and X2 are independent random variables.128

This can be understood as a statement that knowing X2 gives no further informationregarding X1 when the two random variables are independent.

Two useful further entropy-like measures comparing uncertainty between randomvariables are the mutual information and the Kullback-Leibler divergence.

128 X1 and X2 are defined to be independent if p(x1, x2) = p1(x1)p2(x2), or likewise by Eq. (9.27)p(x1|x2)= p1(x1) and p(x2|x1)= p2(x2).

Page 286: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

278 Chapter 9. Information Theory in Dynamical Systems

Definition 9.14. The Mutual Information associated with two random variables, {X1, X2},is,

I (X1; X2)=∑x1,x2

p(x1, x2) logp(x1, x2)

p1(x1)p2(x2)129 (9.35)

Alternatively, follows another useful form of the same,

I (X1; X2)= H (X1)− H (X1|X2). (9.36)

Mutual information may be understood as the amount of information that knowing thevalues of either X1 or X2 provides about the other random variable. Stated this way, mutualinformation should be symmetric, and indeed it is immediate to check that Eq. (9.35) isindeed so. Likewise, inspecting the intersection labelled I (X1; X2) in the Venn diagram,Fig. 9.1 also suggests the symmetric nature of the concept. An example application to thespatiotemporal system pertaining to global climate from [108] is reviewed in Sec. 9.9.2.

The Kullback-Leibler divergence on the other hand is a distance-like measure be-tween two random variables, which is decidedly asymmetric.

Definition 9.15. The Kullback-Leibler divergence between the probability density func-tions p1 and p2 associated with two random variables, X1 and X2 is,

DK L(p1||p2)=∑

x

p1(x) logp1(x)

p2(x). (9.37)

The DK L is often described as if it is a metric-like distance between two densityfunctions, but it is not technically a metric as it is not necessarily symmetric; generally,

DK L (p1||p2) �= DK L (p2||p1). (9.38)

Nonetheless, it is always non-negative as can be seen from (9.6) by considering−log(p2(x))as a length of encoding. Furthermore, DK L(p1||p2) can be understood as an entropy-likemeasure in that it measures the expected number of extra bits which would be requiredto code samples of X1 when using the wrong code as designed based on X2, instead ofpurpose designed for X1. This interpretation can be understood by writing,∑

x

p1(x) logp1(x)

p2(x)=∑

x

p1(x) log p1(x)−∑

x

p1(x) log p2(x) (9.39)

= Hc(X1, X2)− H (X1), (9.40)

where Hc(X1, X2) is the cross entropy,

Definition 9.16. The Cross Entropy associated with two random variables, {X1, X2}, withprobability density functions p1 and p2,

Hc(X1|X2)= H (X1)+ DK L(p1||p2), (9.41)

describes the inefficiency of using the wrong model p2 to build a code for X1 relative to acorrect model p1 to build an optimal code whose efficiency would be H (X1).

129It is useful to point at at this stage that p1(x1) and p2(x2) are the marginal distributions of p(x1, x2);p1(x1)=∑x2

p(x1, x2) and likewise, p2(x2)=∑x1p(x1, x2).

Page 287: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.3. Many Random Variables and Taxonomy of the Entropy Zoo 279

Thus when p1 = p2 and therefore DK L (p1||p2)= 0 the coding inefficiency as mea-sured by cross entropy Hc(X1|X2) becomes zero. Mutual information can then be written,

I (X1; X2)= DK L (p(x1, x2)||p1(x1)p(x2)). (9.42)

A stochastic process allows consideration of entropy rate.

Definition 9.17. Given a stochastic process, {X1, X2, ...}, the entropy rate is defined interms of a limit of joint entropies,

H = limn→∞

H (X1, X2, ..., Xn)

n, (9.43)

if this limit exists.

Assuming the special case that Xi are identically independently distributed (i.i.d.)and noting that independence130 gives,

p(x1, x2, ..., xn)=n∏

i=1

p(xi ), (9.44)

from which quickly follows,131

H (X1, X2, ..., Xn)=n∑

i=1

H (Xi )= nH (X1). (9.45)

The second statement in this equality only requires independence and the third follows theidentically distributed assumption. Now we are in a position to restate the result Eq. (9.20)of the Source Coding Theorem 9.8 in the case of a coding extension. If L(c(x1x2...xn) isthe length of the coding extension of n coded words of realizations of the random variables,X1, X2, ..., Xn, then Eq. (9.20) generalizes to a statement regarding the minimum code wordlength expected per symbol,

H (X1, X2, ..., Xn)≤ nL(C)≤ H (X1, X2, ..., Xn)+1, (9.46)

or in the case of a stationary process,

limn→∞ L = H , (9.47)

the entropy rate. Notice this length per symbol is just what was emphasized by examplenear Eq. (9.9). In the case of an i.i.d. stochastic process, Eq. (9.46) specializes,

H (X1)≤ L(C)≤ H (X1)+ 1

n, (9.48)

130Statistical independence of X1 from X2 is defined when given probabilities as follows, p(x1, x2) =p(x1) for each X1 = x1, and X2 = x2. And since p(x1, x2) = p(x1|x2)p(x2), then independence impliesp(x1, x2)= p(x1)p(x2). And likewise, P(x2|x1)= p(x2) since p(x2|x1)= p(x1 ,x2)

p(x1) = p(x1)p(x2)p(x1) = p(x2).

131 H (X1, X2, ..., Xn ) = ∑i p(x1, x2, ..., xn ) log p(x1, x2, ..., xn ) = ∑

i∏n

i=1 p(xi ) log∏n

i=1 p(xi ) =∑i∏n

i=1 p(xi )[∑n

i=1 log p(xi )]=∑i∏n

i=1 p(xi )[∑n

i=1 log p(xi )].

Page 288: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

280 Chapter 9. Information Theory in Dynamical Systems

This expression is symbolically similar to Eq. (9.20), but now we may interpret the entropyrate of the i.i.d. stochastic process to the expected code word length of a coded extensionfrom appending many coded words.

Finally in this section, coming back to our main purpose here to relate informationtheory to dynamical systems, it will be useful to introduce the notion of channel capacity.We recall the channel coding theorem, because as we will discuss in the next section, achaotic oscillator can be described as such a channel. Channel capacity is the answer tothe question, “how much information can be transmitted or processed in a given time?" Onmay intuit that a communication system may degrade in the sense of increasing error rateas the transmission rate increases. However, this is not at all the case as the true answer ismore nuanced. The part of the channel coding theorem for which we are interested herecan be stated,

Theorem 9.18. Channel Coding Theorem A “transmission" rate R is achievable withvanishingly small probability if R < C where C is the information channel capacity,

C =maxp(x)

I (X ;Y ), (9.49)

where the maximum is taken over all possible distributions p(x) on the input process, andY is the output process.

Now to interpret this theorem, a given communication system has a maximum rateof information C known as the channel capacity. If the information rate R is less than C,then one can approach arbitrarily small error probabilities by careful coding techniques,meaning cleverly designed codes, even in the presence of noise. Said alternatively, lowerror probabilities may require the encoder to work on long data blocks to encode the signal.This results longer transmission times and higher computational requirements. The usualtransmission system as a box diagram is shown in Fig. 9.2 including some description inthe caption of standard interpretations of the inputs and outputs X and Y . Fano’s converseabout error rate relates a lower bound on the error probability of a decoder,

H (X |Y )≤ H (e)+ p(e) log(r ), (9.50)

where,

H (X |Y )=−∑i, j

p(xi , yi ) log p(xi |yj ), and, p(e)= supi

∑i �= j

p(yj |xi ). (9.51)

To interpret the relevance of the Channel Coding theorem 9.18 in the context of dynamicalsystems theory requires only the work of properly casting the roles of X and Y as this isthe subject discussed in Sec. 9.4 as illustrated in Fig. 9.5.

9.4 Information theory in Dynamical SystemsIn Chapter 6 we highlighted the description of a dynamical system as an underlying sym-bolic dynamics. Here we will further cement this connection by describing a complemen-tary information theoretic perspective of the symbolic dynamics.

Perhaps there is no better way of emphasizing the information theoretic aspects ofa chaotic dynamical systems with an underlying symbolic dynamics than by explicitly

Page 289: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.4. Information theory in Dynamical Systems 281

Figure 9.2. Channel Capacity, Eq. (9.52) box diagram. Random variable Xdescribes the random variable descriptive of the possibilities of the input states message.Coming out of the channel is the output states whose values are according to the randomvariable Y which one wishes to be closely related (desired to be identical in many appli-cations). Alternatively, Y be understood as states of a decoder. Such will be the case iftransmission rate is slow enough R < C according to Theorem 9.18.

.

Figure 9.3. The Lorenz attractor has chaotic orbits and an underlying symbolicdynamics, usually presented according to the successive maxima map of the z(t) time series,was already discussed in Eq. (6.4) according to Eqs. (6.5)-(6.6) and the resulting mapwas already illustrated in Figs. 6.3-6.4. In this case, the property of chaos that there areinfinitely many periodic orbits has been leveraged against the symbolic dynamics to chooseperiodic orbits whose symbolic dynamics encode the word “Chaos" in ASCII form, subjectto the non-information bearing bits shown in red. Compare to the embedding form of theattractor shown in Fig. 9.4.

demonstrating orbits bearing messages as we wish. In Fig. 6.2, a time series of an orbitsegment from a Lorenz equation (Eq. 6.4) flow is plotted along with its phase space repre-sentation of the attractor, in Fig. 6.3. We repeat similar Figs. 9.4-9.3. Again, we can readsymbols from the z(t) time series, by the symbol partition Eq. (6.6); zeros and ones can beread by the position of the local maxima of the z(t) coordinate from the differential equationrelative to the cusp-maximum of this value zn in a one-dimensional ansatz zn+1 = f (zn);a “0" bit in the message as indicated in the time series Fig. 9.4 corresponds to a relativelysmaller local maximum as corresponding to the left side of the cusp in Fig. 9.4, meaning

Page 290: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

282 Chapter 9. Information Theory in Dynamical Systems

Figure 9.4. The time series shown in Fig. 9.3 of an orbit segment spelling“Chaos" shown in its phase space presentation of (x , y, z) coordinates on the chaotic at-tractor. This is a carefully chosen initial condition

local maximum (a red point) occurs on the left “wing" of the attractor, and likewise “1" bitsare encoded in the orbit. Now however, we choose to read the symbol and interpret themas if they are coded words in ASCII.

With the Lorenz parameters set to be the usual famous values as they were in inthe previous chapter, (10,28,8/3) we may observe that the underlying symbolic dynamicsgrammar never allows two zero’s in a row in a symbolic word string, as is described inFig. 6.26. Therefore a suitable source code can be built on a prefix coding of the transitionsin the graph shown in Fig. 9.6(Right). There are two possible outcomes from a previoussymbol “1"; either a “0" or a “1" can follow. However, there is only one outcome possiblefrom a symbol “0"; simply stated, there is no surprise from the transition, by observing the“1" bit which follows a “0" bit - it was the only possibility. So this transititon can be saidto be noninformation bearing, or a zero entropy state. To emphasize this, we labelled thistransition by a “*" in the directed graph since it serves only a role as a pause required totransition back to an information bearing state - that is a state where either a “0" or a “1"might follow.

Specifically, in terms of this source coding and using the ASCII prefix code, thephrase Chaos has been encoded into the the chaotic oscillations. All the noninformationbearing 1′s denoting the ∗-transition in directed graph Fig. 9.6 are those 1’s which arecolored red. Counting the ASCII code length to carry the phrase Chaos is 7×5= 35 bits,

Page 291: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.4. Information theory in Dynamical Systems 283

but including the noninformation bearing bits has further required 16+35= 51 bits. Theseextra bits represent a nonmaximal channel capacity of the message carrier, which in thiscase is the dynamical system itself.

The dynamical system encoding the word “Chaos," or any other word, phrase, orarbitrary information stream including sounds or images, etc [170, 26, 264, 58] has a fun-damental associated channel capacity C , Eq. (9.52). So by Theorem 9.18, transmision ratesR less than C are achievable. This imposes a rate of those noninformation bearing bits wedepicted as red in Fig. 9.3. See Fig. 9.5 where we illustrate the reinterpretation of the stan-

Figure 9.5. Channel Capacity, Eq. (9.52) box diagram interpreted as a dynami-cal system. Compare to Fig. 9.2 and Theorem 9.18.

.

dard box diagram Fig. 9.2 in the setting where the channel is taken to be the dynamicalsystem. In such case, we reinterpret Eq. (9.52) as,

C =maxp(x)

I (Xn; Xn+1), (9.52)

The entropy which is most relevant uses the so-called maximum entropy measure(Theorem 9.20) corresponding to the topological entropy htop(�′) which in Eq. (6.74) wesee is descriptive of the number of allowable words Nn of a given length n. Perhaps mostrevealing is the spectral description Eq. (6.67) stating, htop(�′k) = lnρ(A). The trade offbetween channel capacity, transmission rates and noise resistance was revealed in [29] andthe corresponding devil’s staircase graph is shown in Fig. 6.30 by a computation illustratedby Fig. 6.29 by the spectral method.

Before closing with this example, note that feedback control can be used to stabilizeorbits with arbitrary symbolic dynamics [170, 26, 29, 107] starting from control of chaosmethods [256, 301]. Specifically, electronic circuitry can be and has been built whereinsmall control actuations may be use to cause the chaotic oscillator to transmit informationby small energy perturbations with simple circuitry to otherwise high powered devices.This has been the research engineering emphasis of [69, 70] toward useful radar devices.

Otherwise stated without the emphasis on practical devices, since it can be arguedthat there is information concerning states in all dynamical systems and a chaotic oscillatorcould be characterized as a system with positive entropy, then the evolution of the systemthrough these states corresponds to an information generating system. These systems havebeen called “information baths" [161].

What does this example tell us in summary?

• Realizing chaotic oscillators as information sources that forgets current measurementstates and allow information bits to be inserted at a characteristic rate as the system

Page 292: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

284 Chapter 9. Information Theory in Dynamical Systems

Figure 9.6. Transitions with “No Surpise" carry no information. (Left) This ex-ample full tent map on the Markov partition shown has orbits that at all possible timesn, xn ∈ [0,1], there exist nearby initial conditions whose evolution will quickly evolve toa state with an opposite symbol “0" or “1" as the case may be; this unrestricted sym-bolic grammar is generated by the simple (diadic) two state directed graph shown below.Each state allows transition to a “0" or “1" outcome and so the surprise of observing theoutcome is the information born by random walks through the graph corresponding to it-erating the map. (Right) This piecewise linear map drawn on its Markov transition has asymbolic dynamics generated by the graph shown below. This grammar does not allow a“0" symbol to follow a “0" symbol. Thus when at the “0" labelled state (x < c), only a “1"can possibly follow; this transition bears no surprise. Said equivalently it is not informa-tion bearing. Thus the required “1" transition serves nothing other than a delay or pause inthe transmission of a message to a information bearing state - the grammar allows no two“0"s in a row. From “1" either a “0" or a “1" may follow. Compare to Fig. 9.3 where sucha grammar is used to encode oscillations in a Lorenz attractor to transmit a real message.

evolves (or is forced to evolve by feedback control) into new states summarizes theinterpretation of a dynamical system as an information channel, Fig. 9.5.

• The uncertainty in dynamical systems is in the choice of initial condition even if theevolution rule may be deterministic132. The uncertainty in the symbolic outcome isdescribed as the random variable defining the probability of states, corresponding tosymbols. That is realized corresponding in a deterministic dynamical system as the

132This is the difference between a dynamical system (deterministic) and a random dynamical system (non-deterministic). See for example the stochastic process in Eq. (3.38) which nonetheless has a deterministicevolution of densities rule by the random dynamical system’s Frobenius-Perron operator Eq. (3.43).

Page 293: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.5. Deterministic Dynamical Systems in Information Theory Terms 285

unknown precision of state which is amplified upon iteration of an unstable system.Even though the evolution of the states in the phase space of the dynamical systemis deterministic, the exact position in phase space is practically never known exactly.We will make this idea mathematically precise in Sec. 9.5.

• Small feedback control can be used to steer orbits so that those orbits may bear asymbolic dynamics corresponding to desired information.

9.5 Formally Interpreting a Deterministic DynamicalSystem in the Language of Information Theory

In the previous section, we illustrated by example that through symbolic dynamics, it isquite natural to think of a dynamical system with such a representation in terms of infor-mation theory. Here we will make this analogy more formal, thus describing the connectionwhich is some foundational aspects of ergodic theory. In this section we will show how tounderstand a deterministic dynamical system as an information bearing stochastic process.Here we will describe a fundamental information theoretic quantity called Komolgorov-Sinai entropy (KS entropy), hK S(T ) which gives a concept of information content of orbitsin measurable dynamics. By contrast, in the next section we will discuss in greater detailthe topological entropy, htop(T )

Assume a dynamical system,

T : M→ M , (9.53)

on a manifold M , with an invariant measure μ. For sake of simplicity of presentation, wewill assume a symbol space of just two symbols133,

= {0,1}. (9.54)

Symbolizing as discussed earlier,

s : M→

s(x) = χA0 (x)+χA1(x), (9.55)

but with an arbitrary open topological cover,

A0∪ A1 = M , but A0∩ A1 = ∅, (9.56)

and χA : M→ [0,1] is the indicator function on sets A ⊂ M as usual. So further assumingprobability measure using μ and a corresponding random variable (definition 3.2).

X : →R, (9.57)

for any randomly chosen initial condition x ∈ M and therefore random symbol s(x). Thuswith μ, let,

p0 = P(X = 0)= μ(A0), p1 = P(X = 1)= μ(A1), (9.58)

133The symbol partition need not be generating, and in which case the resulting symbol dynamics willperhaps be a positive entropy process, but not nesc fully descriptive of the maximal entropy of the dynamicalsystem, as discussed in Sec. 6.4.6 and [42].

Page 294: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

286 Chapter 9. Information Theory in Dynamical Systems

In this notation, a dynamical system describes a discrete time stochastic process (definition4.11) by the sequence of random variables as follows,

Xk(ω)= X(s(T k(x))). (9.59)

Now using natural invariant measure μ when it exists, we may write,

P(Xk = σ )= μ(Aσ ), (9.60)

whereσ = 0 or 1. (9.61)

A stochastic process has entropy rate which we describe. A probability space(P ,A,) and associated stochastic process defined by a sequence of random variables,{X1(ω), X2(ω), ...}. The stochastic process is defined to be stationary in terms of the thejoint probabilities as follows, from [54], and specialized for a discrete outcome space,

Definition 9.19. A stochastic process, X1, X2, ..., is stationary if for all k > 0, the process,Xk+1, Xk+2, ..., has the same distribution as, X1, X2, .... In other words, for every B ∈B∞,

P(X1 = x1, X2 = x2, ...)=P(Xk+1 = x1, Xk+2 = x2, ...),∀k > 1, (9.62)

and for each possible experimental outcome (x1, x2, ...) of the random variables.

It follows, [54], that the stochastic process, X1, X2, ..., is stationary if the stochasticprocess, X2, X3, ..., has the same distribution as, X1, X2, ....

Now the entropy rate of such a stochastic process by Definition 9.17,

H = limn→∞

1

nH (n)(ω) (9.63)

in terms of joint entropies,

H (n)(ω)=∑

P(X1 = x1, X2 = x2, ..., Xn = xn) log P(X1 = x1, X2 = x2, ..., Xn = xn).

(9.64)It is straight forward to prove [71] that a sufficient condition for this limit to exist is i.i.d.random variables, in which case,

H = limn→∞

1

nH (n)(ω)= lim

n→∞1

nnH (ω1)= H (ω1). (9.65)

The Shannon-McMillan-Breiman theorem [71] states more generally that for a finite-valuedstationary stochastic process {Xn}, then this limit exists and converges to the entropy rateH .

If the stochastic system is really a dynamical system as described above, one in whicha natural invariant measure μ describes the behavior of typical trajectories, then we attaina direct correspondence of the information theoretic description of the dynamical systemin terms of its symbolization. We may develop the so-called metric entropy, also known

Page 295: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.5. Deterministic Dynamical Systems in Information Theory Terms 287

as Komolgorov-Sinai entropy (KS-entropy, hK S) [195]. Assuming a more general topo-logical partition,

P = {Ai }ki=0, (9.66)

of k+1 components, then the resulting entropy of the stochastic process is,

H (P )=−k∑

i=0

μ(Ai ) lnμ(Ai ). (9.67)

However, we wish to build the entropy of the stochastic process of the set theoretic join134

of successive refinements that occur by progressively evolving the dynamical system. Let,

h(μ, T ,P )= limn→∞

1

n(

n∨i=0

P (n)), (9.68)

where we define,

P (n) =n∨

i=0

T−n(P ), (9.69)

and T−1 denotes the possibly many branched pre-image if T is not invertible. So the join,∨ni=0 T−n(P ) is the set of all set intersections of the form,

Ai1 ∩T−1(Ai2 )∩ ...∩T−n(Ain+1 ),0≤ ik ≤ n, (9.70)

Now we should interpret the meaning of these quantities. H for a stochastic processis the limit of the Shannon entropy of the of the joint distributions. Literally, it is anaverage time density of the average information in a stochastic process. A related conceptof entropy rate is an average conditional entropy rate,

H ′(X)= limn→∞H (Xn|Xn−1, ..., X1). (9.71)

Whereas H (X) is an entropy per symbol, H ′(X) can be interpreted as the average entropyof seeing the next symbol conditioned on all the previous symbols. There is an importantconnection which occurs for a stationary process. Under the hypothesis of a stationarystochastic process, there is a theorem [71] that states,

H (X)= H ′(X), (9.72)

which further confirms the existence of the limit, Eq. (9.71).Thus by this connection between entropy rates in dynamical systems we can interpret

h(μ, T ,P (i)) in Eq. (9.68) as the information gained per iterate averaged over the limit oflong time intervals; call this hμ(T ) for short. The details of the entropy depend on thechosen partition P . As was already discussed in Sec. 6.4.6 and highlighted in Fig. 6.33,the value of entropy measured in a dynamical system depends on the chosen partition. This

134The join between two partitions P1 = {P11 , P2

1 , ..., Pm1 } and P2 = {P1

2 , P22 , ..., Pn

2 } is defined P =P1∨P2 = ({Pi

1 ∩ P j2 }mi=0)nj=0, and in general P =∨N

i=1 Pi = P1∨P2∨ ...∨PN is interpreted successively byrepeated application.

Page 296: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

288 Chapter 9. Information Theory in Dynamical Systems

is most obvious in the extreme case that P = P0 is defined to be a partition of a singleelement covering the whole space, in which case, all possible symbol sequences of allpossible orbits consist of one symbol stream, 0.000.... This would give zero entropy dueto zero surprise, and likewise because p0 = 1 =⇒ log(1)= 0. It is natural then to ask ifthere is a fundamental entropy of the dynamical system, rather than having entropy closelyassociated with the choice of the partition. From this question follows the quantity,

hK S(T )≡ hμ(T )= supP

h(μ, T ,P (i)). (9.73)

This KS-entropy is the supremum of entropies over all possible partitions. It describes theaverage bit rate of seeing symbols in terms of all possible partitions and weighted accordingto natural invariant measure μ. The interpretation of hK S(T ) is the description of precisionas a degree of surprise of a next prediction with respect to increasing n. When this quantityis positive, in some sense this relates to sensitive dependence to initial conditions.

There is another useful concept of entropy often used in dynamical systems, calledtopological entropy htop [1] which we have already mentioned in Sec. 6.4.1. We mayinterpret htop as directly connected to basic information theory and also to hK S . Oneinterpretation of htop is in terms of maximizing entropy by rebalancing the measures so asto make the resulting probabilities extremal. Choosing a simple example of just two statesfor ease of presentation, we can support this extremal statement by the simple observationthat, (stating Shannon entropy for two states),

(1/2,1/2)= argmaxp[−p log p− (1− p) log(1− p)]. (9.74)

This is easily generalized for finer finite partitions. Expanding the same idea to Eq. (9.73)allows us to better understand Parry’s maximal entropy measure when it exists. Stated inits simplest terms for a Markov chain, Parry’s measure is a rebalancing of probabilities oftransition between states so that the resulting entropy of the invariant measure becomesmaximal. See [259] and also [56, 152] and [159] for some discussion of how such maximalentropy measures need not generally exist but do exist at least for irreducible subshifts offinite type. Generally the connection between htop(T ) and hμ(T ) is formal through thefollowing variational theorem,

Theorem 9.20. (Variational Principle for Entropy - Connection Between MeasureTheoretic Entropy and Topological Entropy, [152, 101, 50] ) Given a continuous mapf : M→ M on a compact metric space M ,

htop( f )= supμ

hμ( f ), (9.75)

where the supremum is taken over those measures μ which are f -invariant Borel probabil-ity measures on M .

On the other hand, the direct Definition 9.21-9.22 of topological entropy [259] is interms of counting numbers of ε-separated sets, and how quickly these states of finite pre-cision become separated by iterating the dynamical system. See also [50, 279]. We findthe variational principle to be more descriptive than the original definition in terms of un-derstanding the meaning of topological entropy. Further discussion of htop(T ) connectionswith computational methods are made in the next section, 9.6.

Page 297: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.6. Computational Estimates of Topological Entropy and Symbolic Dynamics 289

9.6 Computational Estimates of Topological Entropy andSymbolic Dynamics

In the previous section of this chapter, 9.4, the connection between orbits of a dynami-cal system and a dynamical system as an entropy process is discussed by example withdemonstration of the information present in distinguishing orbits. Further, the connectionbetween measurable dynamics and topological dynamics can be understood in terms of thevariational principle for entropy theorem 9.20. In the discussion of symbolic dynamics inChapter 6, especially in Sec. 6.4.1, we discussed symbolic dynamics in depth, includingformulas describing entropy in terms of cardinality, Eq. (6.74) and also a related spectralformula Eq. (6.67). In this section, we will reprise this discussion of topological entropyassociated with a dynamical system in more detail and in the context of the formal infor-mation theory of this chapter.

9.6.1 Review of Topological Entropy Theory

Adler, Konheim, and McAndrew introduced topological entropy in 1965, [1], in terms ofcounting the growth rate of covers of open sets under the dynamical system. However, theBowen definition [46, 48] is in terms of ε-separated sets,

Definition 9.21. ((n,ε)-separated) [281]. Given a metric space (M ,d) and a dynamicalsystem on this space, f : M → M , then a subset S ⊂ M is (n,ε)-separated if,

dn, f (x,y) > ε, (9.76)

for each distinct x,y ∈ S, x �= y, where,

dn, f (x,y)= sup0≤ j<n

d( f j (x), f j (y)), (9.77)

In terms of the metric topology 135 this can be roughly described as enumerating thecoarsening associated with how open sets are maximally “cast across" open sets. By “castacross," we simply mean that iterates of the set have images intersecting other open sets.Topological entropy is defined in terms of counting the growth rate of the iteration time nof different orbits, where different is in terms of an ε-scale, as time grows, and then in thelimit as this ε-coarse grain decreases.

Definition 9.22. (Topological Entropy) [46, 281]

htop( f )= limε→0

limsupn→∞

log(s(n,ε, f ))

n(9.78)

where,log(s(n,ε, f ))=max{#(S) : S ⊂ M is (n,ε)− separated by f } (9.79)

In this sense, topological entropy is a counting of rate of new states which develop underthe dynamics in terms of states on each coarse scale, in the scale limit.

135 A metric topology is a topology (See topological space 4.3.) whose basis of open sets is simply openballs Nε (x) defined in terms of a metric d on M . Nε (x)= {y : d(x,y) < ε}.

Page 298: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

290 Chapter 9. Information Theory in Dynamical Systems

This definition cannot be considered as practical as a computable quantity. Othermethods must be deployed. When the dynamical system is a right resolvent sofic shift(See Definition6.6) we have the simple spectral formula Eq. (6.67) to simply and exactlycompute topological entropy, which we restate [48, 281],

htop(�′N )= lnρ(A). (9.80)

ρ(A) is the spectral radius of the associated transition matrix A, meaning the largest eigen-value of the matrix of 0’s and 1’s corresponding to allowed transitions on the generateddirected graph G A denoting allowed words in the subshift �′N . Recall that following defi-nition 6.5, there is a finite sized transition matrix A that generates a graph G A that presentsthe grammar of the subshift �′N of a fullshift �N on N-symbols with a correspondingBernoulli mapping s : �′N →�′N . This formula is practical in general dynamical systemsin the scenario of a Markov partition, Definitions 4.1, 4.2.3, from which often a finite graphmay be associated with transitions of the corresponding dynamical system meant to prop-erly represent transitions of the associated shift map, and using Theorems 9.24-9.25 for theequivalence. We have used this computational method for the logistic map in [29, 43] andthe Henon map in [206, 43] to name a few. We wish to highlight two perspectives on thisapproximation by a Markov model,

• A nested sequence of (imbedded chaotic saddle) subsets allow Markov models of theattractor as a whole, as in Example 9.1.

• In terms of the uniform norm, nearby a given dynamical system, a dynamical systemmay exist which is exactly Markov, and finite dimensional computations are exact.Therefore one way to discuss entropy approximations of a given dynamical systemis in terms of sequences of such Markov estimates and understanding the density ofthe representation.136

Another useful description of the topological entropy is the expression, with hypoth-esis in [49],

htop( f )= limsupn→∞

log(Pn)

n, (9.81)

where Pn denotes the number of fixed points of f n .137 Remembering the popular Devaneydefinition of chaos 6.1 includes the requirement that periodic orbits are dense, [12, 97],there is some sense in measuring the complexity of the dynamical system in terms of theseorbits. A practical perspective is an observation that unstable periodic orbits, (UPOs) formthe “skeleton" of nonlinear dynamics (Auerbach et al., 1987; Cvitanovic, 1988). To thisend, Auerbach-Cvitanovic [7, 5] has progressed in estimates of thermodynamic proper-ties by using just the short orbits. Numerous theoretical systems were shown to be welldescribed through this approach [5]. Furthermore, recent work using interval arithmetic

136For example, in the special case of the set of all skew tent maps, in [19] can be found a theorem thatstates that a given map is either Markov, or it may be uniformly (sup-norm) estimated to arbitrary accuracyby a map that is Markov; in this sense the family of Markov maps is dense in the uniform topology. Seefurther discussion in Sec. 4.4 and especially Theorem 4.13.

137It is important to count correctly: Given a map x ′ = f (x), a periodic orbit {x0x1, .., xn−1} is n-pointswhich are roots of f n(x) = x . However, not all roots of f n(x) = x are period n since the period of a pointis defined as the smallest m such that f m (x)= x . For example, a fixed point (n = 1) solves f (x)= x , and italso solves f 2(x)= x (and f n(x)= x ,∀n > 0) but it is called period-1 since 1 is the smallest such n.

Page 299: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.6. Computational Estimates of Topological Entropy and Symbolic Dynamics 291

[142, 141, 140] and computer assisted proof [144] has allowed for rigorous entropy esti-mates with this perspective. See also, [80] for discussion of computer validated proof ofsymbolic dynamics, and also [234, 235, 236] based on methods of computational homology[191]; at the forefront and quite impressive, these methods can even be used in infinite di-mensional PDE systems validating the finite dimensional Galerkin representation as ODEsand the persistence of their periodic orbits despite the finite dimensional approximation[341, 81, 82].

The use of formula Eq. (9.81) is generally to empirically check for convergence fora general mapping, and we emphasize that this formula does not require symbolization asshown in Example 9.1. However, under the assumption that the mapping is a shift mapping,this formula is similar to a mathematically well founded statement,

htop(�′N )= limsupn→∞

logwn

n, (9.82)

where wn is the number of words of length n in �′N . The principled use of both formulasEq. (9.80) and (9.82) are confirmed by the theorem,

Theorem 9.23. [281], A subshift with the Bernoulli shift map dynamical system, s : �′N →�′N , has topological entropy that may be computed generally by Eq. (9.80) or Eq. (9.82)when sofic and right resolvent. (See comment 6.6).

Remark 9.2. Understanding the difference between the periodic orbits estimate Eq. (9.81)and the word count formula Eq. (9.82) may be described in terms of symbolizing orbits witha generating partition. When the periodic orbit estimate Eq. (9.81) is used specifically fora system that is already symbolized, we interpret as counting symbolized periodic orbits.Let,

un = σ0.σ1σ2...σn−1,σi ∈ {0,1, ..., N−1}, (9.83)

be a word segment of length n of the N symbols associated with �N . Then by definition,wn is the number of such blocks of n bits that appear in points σ in the subshift, σ ∈ �′N .These word segments may be part of periodic orbit, in which case that word segment isrepeated,

σ = unun ...≡ σ0.σ1σ2...σn−1σ0.σ1σ2...σn−1... (9.84)

or it would not be repeated if the point is not part of a periodic orbit. So in the symbolizedcase, the difference between the two formulae is that generally we expect,

Pn ≤ wn , (9.85)

but the hope is that for large n, the difference becomes small.

Remark 9.3. The key to the practical and general use of the periodic orbit formula Eq. (9.81)is that symbolization is not necessary. This is a useful transformation of the problem of es-timating entropy since finding a generating partition is generally a difficult problem in itsown right for a general dynamical system [73, 154, 77, 43]. Details of the misrepresenta-tion of using a partition that is not generating are discussed in Sec. 6.4.6 as a review of [43].Whether or not we symbolize, the number of periodic orbits is the changed, which is whythe periodic orbit formula is robust in this sense, although it has the alternative difficulty ofbeing confident that all of the periodic orbits up to some large period have been found.

Page 300: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

292 Chapter 9. Information Theory in Dynamical Systems

Remark 9.4. The key to algorithmically use the periodic orbit formula Eq. (9.81) is thepossibility of reliably finding all of the periodic orbits of period n for a rather large n. Giventhat Pn is expected to grow exponentially, it is a daunting problem to solve,

g(z)= f n(z)− z = 0, (9.86)

for many roots. By saying many roots, we are not exaggerating since in our experience[77], to reliably estimate the limit ratio in Eq. (9.81), we have worked with on the orderof hundreds of thousands of periodic orbits which are “apparently" complete lists for therelatively large n ) 18. See Example 9.1 using the Ikeda map. When n = 1, the obviousapproach would be to use a Newton’s method or variant. And this works even for n = 2 or3 perhaps, but as n increases, and Pn grows exponentially, the seeding of initial conditionsfor the root finder becomes exponentially more difficult as the space between the basinsof attraction of each root becomes small. A great deal of progress has been made towardsurprisingly robust methods in this computationally challenging problem of pushing theroot finding to produce the many roots corresponding to seemingly complete lists of orbits,[290, 18, 100, 78].

Example 9.1. Entropy by Periodic Orbits of the Ikeda Map. Consider the Ikeda map[184, 167],

Ikeda : R2→ R2

(x , y) �→ (x ′, y ′)= (a+b[x cos(φ)− y sin(φ)],b[x sin(φ)+ y cos(φ)]) (9.87)

where,φ = k− ν/(1+ x2+ y2), (9.88)

and we choose parameters a = 1.0,b = 0.9,k = 0.4 and ν = 6.0. In [76], the authorsconjectured a claim construction of all the periodic orbits through period-22, whereas in[77] we used seemingly all of the roughly 373,000 periodic orbits through period-20 toestimate

htop(Ikeda)) 0.602 < ln2, (9.89)

by Eq. (9.81). In Fig. 9.8 we show those periodic orbits through period-18. Furthermore in[77], we noted that the requirement that a generating partition must uniquely differentiate,by the labeling, all of the iterates on all of the periodic orbits; we used this statementto develop a simple construction to successively symbolize (color) each of the periodicorbits in successively longer periods. As these numbers of periodic orbits swell to tens andhundreds of thousands, the attractor begins to fill out as seen in Fig. 9.8 to become a usefulrepresentation of the symbolic dynamics. Notice an interesting white shadow reminiscentof the stable manifolds that are tangent to the unstable manifolds believed to be associatedwith generating partitions [73, 154].138 A thorough study of periodic orbits together with

138 Interestingly as an aside, notice that the periodic orbits are expected to distribute roughly according tothe invariant measure and thus are more rare at regions of low measure; apparently these “white regions" cor-respond to just those regions associated by tangencies of stable and unstable manifolds. To see this observa-tion, note that the SRB measure is the invariant measure along the closure of all the unstable manifolds of theperiodic orbits. A clear shadow of “white" missing periodic orbits (up to the period-18’s found) can be seenas transverse curves through the attractor, punctuated at tangency points. This conjecture-observation agreesin principle with the well accepted conjecture [73, 154] that generating partitions must connect betweenhomoclinic tangencies. See Fig. 6.31 for an explicit construction demonstrating the generating partition forthe Henon map constructed directly by this conjecture.

Page 301: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.6. Computational Estimates of Topological Entropy and Symbolic Dynamics 293

rigorous computer assisted proof by interval arithmetic [242] is developed in [141] fromwhich we reprise the table of counted orbits, Table 9.1, including comparable estimates oftopological entropy commensurate with our own best htop ) 0.602.

Example 9.2. Entropy of the Henon Map. Comparably, for Henon mapping, h(x , y)=(1+ y−ax2,bx), with (a,b)= (1.4,0.3) in [141], and finding even more periodic orbits,

(n, Qn , Pn , Q≤n , P≤n ,hn)= (30,37936,1139275,109033,3065317), (9.90)

where rigorous interval arithmetic was used [242]. Compare to the tabular values for theIkeda mapping Table 9.1. Interestingly, but not needed in this symbolic dynamics freecomputation by periodic orbits, see Fig. 6.31 where a generating partition for this Henonmap is constructed directly by locating tangency points.

Example 9.3. Generating Partition and Markov Partitions of the Ikeda-Map. Con-sider the Ikeda map, Eq. (9.87). In Fig. 9.8, we show a partition from [77] consistent witha generating partition, this having been constructed by requiring uniqueness of representa-tion for each of the hundreds of thousands of periodic orbits through period-18, and a tableof periodic orbits in Table 9.1.

In Fig. 9.7, we show two candidate Markov partitions each using several symbolsfrom [142]. On the left we see a Markov partition in 4 symbols and in the right we seea Markov partition in 7 symbols. Further, in [142], a further refined Markov partition in18 symbols is shown. Generally, a strange attractor may have (infinitely) many embeddedMarkov partitions representing embedded subshift, where high ordered representations canhope to represent the symbolic dynamics of a greater and greater subset of the full attractor,such as in [19]. Compare to Definition 4.4 and Fig. 4.5. Further discussion of the entropyof this attractor is addressed in Example 9.1.

Example 9.4. Entropy by Markov Model of the Ikeda Map In Example 9.3, Fig. 9.7, werecall from [142] two Markov model refinements corresponding to two imbedded subshifts�′4 and �′7 in the Ikeda attractor, using α = 6 of Eq. (9.87). See also [141] and [140] fortechniques of computing enclosures of trajectories, finding and proving the existence ofsymbolic dynamics and obtaining rigorous bounds for the topological entropy. From theIkeda map, the imbedded Markov models yield associated transition matrices

A4 =⎛⎜⎝ 0 0 0 1

1 0 0 01 1 0 00 0 1 0

⎞⎟⎠ , A7 =

⎛⎜⎜⎜⎜⎜⎜⎜⎝

0 0 0 1 0 0 01 0 0 0 1 1 01 1 0 0 0 0 00 0 1 0 0 0 00 0 0 0 0 1 00 0 0 0 0 0 11 1 0 0 0 0 0

⎞⎟⎟⎟⎟⎟⎟⎟⎠, (9.91)

Using these Markov representations of the transition matrices of the grammars of theimbedded subshifts, together with Eq. (9.80),

htop(�′4)= lnρ(A4)= 0.19946, and, htop(�′7)= lnρ(A7)= 0.40181, (9.92)

and similarly in [142], a further refined 18 symbol Markov model produces,

htop(�′18)= lnρ(A18)= 0.48585. (9.93)

Page 302: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

294 Chapter 9. Information Theory in Dynamical Systems

Figure 9.7. Symbol Dynamics of the Ikeda-Hammel-Jones-Moloney attractorEq. (9.87) from [142]. (Left) Ikeda map, α = 6 sets Ni on which the symbolic dynamics on4 symbols exists and their images. (Right) Ikeda map, α = 6, sets Ni on which the symbolicdynamics on 7 symbols exists and their images. Compare to Definition 4.4 and Fig. 4.5.

We do not show A18 here for sake of space, and any case, the principle is to continue torefine and therefore to increase the size of the matrices to build ever increasing refinements.These estimates are each lower bounds of the full entropy, for example,

htop(A) > htop(�′18), (9.94)

where htop(A) denotes the topological entropy on the chaotic Ikeda attractor A meaningthe entropy of the dynamics Eq. (9.87) on this attractor set. We have often used the phrase“embedded subshift" �′N here, by which we mean that there is a subset A′ of the attractorA such that the subshift �′N is semiconjugate to the dynamics on that subset A′; imbeddedis the more accurate term. As such, by

Theorem 9.24. [198]Comparing Topological Entropy of Factors.139 Suppose two irre-ducible subshifts of finite type such that �B is a factor of �A, then,

htop(�B) ≤ htop(�A). (9.95)

From this theorem is the related statement,

Lemma 9.1. Topological Entropy Equivalence. Two conjugate dynamical systems haveequivalent topological entropy.

Also related is a slightly weaker than conjugacy condition for equivalence of topo-logical entropy,

139A subshift of finite type �B is a factor (synonym to semi-conjugate) of another subshift of finite type�A if there exists a continuous and onto mapping f : �A → �B that commutes; s ◦ f (σ ) = f ◦ s(σ ). Aconjugacy is a special case where f a homeomorphism.

Page 303: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.6. Computational Estimates of Topological Entropy and Symbolic Dynamics 295

Theorem 9.25. [280] Topological Entropy Compared by Finite-One Semi-conjugacy.If g1 : X → X and g2 : Y → Y are two continuous mappings on compact metric spaces Xand Y , and f : X → Y is a semi-conjgacy that is uniformly finite-one, then,

htop(g1)= htop(g2). (9.96)

It is this third theorem that permits us to compute the topological entropy of a dynam-ical system in terms of its symbolic dynamics140 The topological entropy on the attractor ismodeled htop(A′)= htop(�′N ). The phrase “embedded subshift" is rightly often used. Thecorrect word from topology is imbedding141

The first theorem explains that in these examples, it is not a surprise that the succes-sive approximations of this example,

�′4 ↪→�′7 ↪→�′18, (9.97)

leads to the estimates as found [142, 77],

htop(�′4)= 0.19946≤ htop(�′7)= 0.40181≤ htop(�′18)= 0.48585≤ htop(Ikeda)) 0.602.(9.98)

The hooked-arrow ↪→ denotes imbedding. While not all imbedded Markov models willbe comparable between each other, nor will be their entropies, in the case of these exam-ples the nesting explains the reduced entropies. Further discussion can be found of nestedimbeddings of horseshoes in homoclinic tangles [239] and chaotic saddles [28].

9.6.2 A Transfer Operator Method of Computing Topological Entropy

Finally, in this section, we briefly mention the possibility of using a transition matrix ver-sion of the Ulam-Galerkin matrix to estimate topological entropy, as was studied in detailin [133] and similarly to the computation in [29]. In brief, the idea is to use an outer-covering of the attractor by “rectangles" or any other regular topological partition such asthe successive refinements of Delaunay triangulations [209]. Such an outercovering of theHenon attractor is shown for example in Fig. 9.9. As we will discuss, using the transferoperator seems like an excellent and easy method to produce a transfer matrix based ap-proach toward entropy estimates for a fine grid. After all this follows the same theme as isdone in an Ulam method to compute invariant measure or the spectral partitioning methodshighlighted earlier in this writing. However, it turns out the approach is not so simple sincecare must be made regarding using the generating partition, and since having the generat-ing partition a priori makes any grid method obsolete, we will suggest that transfer matrixbased methods are rarely useful. Nonetheless the theory is interesting with some interestinglessons which we will highlight [133, 29].

The idea behind transfer matrix based computational method of computing topolog-ical entropy hinges on the following refinement theorem, relating upper bounds,

140Further note that the theorem that the entropy of a continuous dynamical system, on a compact set, isequal to the entropy of the map restricted to its nonwandering points [280].

141A mapping f : X → Y is defined as an imbedding X into Y if a restriction of the range to Z resultsin a homeomorphism f : X → Z [245]. Furthermore, when used in the dynamical systems sense and twomappings, when this restriction is with respect two mappings g1 : X → X and g2 : Y → Y , and the domainrestriction of Z results in a conjugacy, we still use the phrase imbedding. Often the word “embedding" maybe used in a context that more accurately denotes imbedding as we have defined it here.

Page 304: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

296 Chapter 9. Information Theory in Dynamical Systems

Figure 9.8. Periodic orbit points up to period-18 for the Ikeda-Hammel-Jones-Moloney attractor Eq. (9.87) from [77]. Furthermore the points on the periodic orbitsare colored according to their symbolic representation: green and red dots represent orbitpoints encoded with symbols 0 and 1, respectively. Compare to Table 9.1 and also Fig. 9.7.[290]

Theorem 9.26. [133] Assuming a partition P , then let,

h∗(T ,P ) := limN→∞

log |wN (T ,P )

N= inf

N≥1

log |wN (T ,P )|N

, (9.99)

then the topological entropy htop(T ) of the map T is bounded:

If P is a generating partition, (see Definitions 4.5-4.7), then,

htop(T )≤ limdiam P→0

infh∗(T ,P )≤ h∗(T ,P ), (9.100)

This theorem provides an upper bound for the topological entropy and suggests asimple constructive algorithm but one which requires care as we point out here. We illus-trate a direct use of this theorem in Fig. 9.9, for which example in the lower right frame,N = 565 triangular cells cover the attractor. An outer covering of the attractor would result

Page 305: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.6. Computational Estimates of Topological Entropy and Symbolic Dynamics 297

Table 9.1. Periodic orbit counts for Ikeda-Hammel-Jones-Moloney attractorEq. (9.87) from [77]. Qn = the number of periodic orbits of period n. Pn is the num-ber of fixed points of the mapping, f n. Q≤n is the number of cycles of period less thanor equal to n. P≤n is the number of of fixed points of f i for i ≤ n. hn is the estimateof the topological entropy for n, using Eq. (9.81). For comparison, in [77], we estimatedh18 ) 0.602. [141]

n Qn Pn Q≤n P≤n hn1 2 2 2 2 0.69312 1 4 3 4 0.69313 2 8 5 10 0.69314 3 16 8 22 0.69315 4 22 12 42 0.61826 7 52 19 84 0.65857 10 72 29 154 0.61108 14 128 43 266 0.60659 26 242 69 500 0.609910 46 484 115 960 0.618211 76 838 191 1796 0.611912 110 1384 301 3116 0.602713 194 2524 495 5638 0.602614 317 4512 812 10076 0.601015 566 8518 1378 18566 0.6033

in h∗(T ,P ) ≤ log N , that we can see is divergent as N is refined. In any case, this is anextreme and not sharp estimate for typical grid refinement.

Example 9.5. Transfer Operator for Topological Entropy for Henon. Direct applica-tion of this theorem to data Fig. 9.9, building adjacency matrices on the fine Delaunaytriangulations,142

Ai, j = i f (1, Bi → Bj ,0,else)= ceil(Pi, j ). (9.101)

results in,h∗ = 1.0123,1.0271,1.0641,1.0245, (9.102)

for the specific grid refinements shown,

h = 0.5,0.2,0.1,0.05, (9.103)

yielding,N = 34,105,228,565, (9.104)

element coverings, respectively. We see that these are all poor upper bound estimates of awell regarded htop(T )) 0.4651 from [74] derived by methods discussed previously in thissection.

So what is wrong? An over-estimate is expected, but why does this seemingly ob-vious and easy approximation method, even with a relatively fine grid give such large

142The adjacency matrix A is easily derived from the stochastic matrix corresponding to the Ulam-Galerkinmatrix using the ceil function.

Page 306: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

298 Chapter 9. Information Theory in Dynamical Systems

Figure 9.9. An outercovering (red) of successive partitions by Delauney tri-angulations over a Henon attractor. Successively shown are coverings with edge lengthsh = 0.5,0.2,0.1,0.05 resulting in transition matrices AN×N , N = 34,105,228,565, whereN is the count of the size of the partition covering the strongly connected component (red)of the full covering.

over-estimates. The answer lies in the fact that the symbolization is wrong, and not evenclose. That is, the partition is wrong. P has 565 elements in our finest cover shown, andwe remind that log N is the upper bound of the entropy of any subshift in �N . While1.0245 << log565, it should not be a surprise that the estimate 1.0245 is not close to0.4651. Rather the generating partition must be used, and eigenvalues of the transfer ma-trix are exact if that finite partition happens to be Markov, or close by Eq. (9.100) if thepartition of the Markov representation is close to generating.

Since as we recall footnote138 and Fig. 6.31 that the generating partition for theHenon map connects primary homoclinic tangencies, which is a zig-zag line that turns

Page 307: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.6. Computational Estimates of Topological Entropy and Symbolic Dynamics 299

out to run roughly near y = 0. Therefore to correctly estimate htop(T ), it is necessary toassociate each cell Bi with a position relative to the generating partition. Clearly if thegenerating partition is,

P ′ = (P ′1, P ′2, .., P ′M ), M << N , and cell Bi ⊂ P ′j , (9.105)

then the new symbol j is associated with each such Bi . Similarly for cells Bi which are notentirely within a single partition element, then a decision must be made, perhaps to choosethe largest region of overlap, P ′j ∩ Bi . In this manner, a projection from the larger symbolspace is developed,

� : �N →�M . (9.106)

The corresponding projected transition matrix should produce the correct topological en-tropy, however, the arbitrarily presented graph cannot be expected to be right resolvent.(See Definition 6.6). The following theorem guarantees that a right resolvent presentationexists even if the arbitrary projection may not be right resolvent,

Theorem 9.27. (Lind and Markus) [214] Every sofic shift has a right-resolving presenta-tion.

The proof of this theorem is constructive, by the so-called follower method [214], asused in [43] in a context similar to the disucussion here, which is to associate new partitionswith transition matrices associated with arbitrary partitions. By this method, a new transfermatrix, one which is a right resolvent presentation of the grammar is developed. Thereforeits corresponding spectral radius is correctly the topological entropy.

However, proceeding in a manner as described above in two steps, a) develop a tran-sition matrix associated with a fine partition and then b) develop a projection to a rightresolvent presentation by the follower construction, may not be considered to be a usefulmethod, since it one still must already know the generating partition to properly associatelabels to the fine representation of the transfer matrix. As we have already stated, this is adifficult problem to find the generating partition. Further, if we already have the generat-ing partition, then simply counting words associated with long orbit segments is a usefuland fast converging method to estimate entropy without needing to resort to the transitionmatrix which skips the computational complexity associated with grid based methods. Fur-thermore, an exceedingly fine grid would be needed to properly represent the nuance of thew-shaped generating partition seen in Fig. 6.31. For the sake of simplicity, in [133] theauthors chose to associate a not generating partition as follows: let,

P ′ = (P1, P2) where P1 = {(x , y)|y < 0}, and P2 = {(x , y)|y > 0}. (9.107)

Clearly this partition is relatively close to the generating partition. As such, the right re-solvent presentation of the transfer matrix gives an estimate of 0.4628 which we see is lessthan that from [74],

0.4628 < htop(T )) 0.4651. (9.108)

That the estimate is close is a reflection of a continuity of the entropy with respect to degreeof misplacement discussed in [43, 42]. Furthermore, the theory detailed in [43, 42], showsthat using an arbitrary partition risks erratic large errors as emphasized by Figs. 6.32-6.33and Eqs. (6.72)-(6.73) from [43, 42], even if the result is a very interesting devils staircase-like function describing the consequences of using a nongenerating partition. It could be

Page 308: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

300 Chapter 9. Information Theory in Dynamical Systems

argued that the fact that y = 0 so easily guessed is in fact close to the generating partitionseen in Fig. 6.31 is just good luck to give a reasonable answer not to be relied upon ingeneral. In any case the form of the error is not known to be positive or negative, despitethe upper bounding statement, Eq. (9.100).

In summary, when estimating topological entropy, the use of transfer operator meth-ods still requires knowledge of the generating partition. Errors may likely be large asanalyzed in [42] if we do not use generating partition information, despite refining grids.However, if we do have the generating partition, then it is perhaps much simpler and moreaccurate to resort directly to counting words, Eq. (9.82).

9.7 Lyapunov Exponents, and Metric Entropy and theUlam’s Method Connection

In this section we will tighten the connection between metric entropy and how it canbe computed in terms of Ulam-Galerkin matrix approximations by consideration of theMarkov action on the corresponding directed graph. This continues in our general theme ofconnecting concepts from measurable dynamics to computational methods based on trans-fer operators. Further, we will discuss how this type of computation is exact in the case thata Markov partition is used. Thus, again referring to Sec. 4.4 and especially Theorem 4.13concerning density of Markov representations, then we can understand a way to analyzethe quality of the estimate. Finally, we will review the Pesin’s identity which provides abeautiful and deep connection between metric entropy hK S and Lyapunov exponents. Wewill discuss both estimation and interpretation of these exponents and their informationtheoretic implications. The main point here is that averaging on single trajectories versusensemble averages is again the Birkhoff ergodic theorem, Eq. (1.5), which here gives adoubly useful way to compute and understand the same quantities. Compare this section tothe introduction with description of two ways of computing Lyapunov exponents discussedin Example 1.4.

9.7.1 Piecewise Linear in an Interval

We start this discussion by specializing to piece-wise linear transformatons of the interval,specifically to Markov systems that are chaotic; such systems allow the probability densityfunctions to be computed exactly. It is well known that expanding piecewise linear Markovtransformations have piecewise constant invariant probability density functions, alreadyreferred to in Sec. 4.4.4,

Theorem 9.28 ([53], Piecewise Constant Invariant Density). Let τ : I → I be a piece-wise linear Markov transformation of an interval I = [a,b], such that for some k ≥ 1,

|(τ k)′|> 1,

where the derivative exists, which is assumed to be in the interiors of each partition seg-ment. Then τ admits an invariant [probability] density function which is piecewise constanton the partition P on which τ is Markov.

Using the Frobenius-Perron operator P , the fixed-point function ρ satisfies the def-inition Pτ ρ = ρ, implying that ρ is the probability density function for a measure that is

Page 309: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.7. Lyapunov Exponents, and Metric Entropy and the Ulam’s Method Connection301

invariant under τ . Since τ is assumed to be a piecewise monotone function, the action ofthe operator is simply

Pτ ρ(x)=∑

z∈{τ−1(x)}

ρ(z)

|τ ′(z)| .

The periodic orbit formed by the iteration of x = a forms a partition of the domain [0,1]on which ρ is piecewise-constant. On each interval Ii , call the corresponding constant,

ρi = ρ|Ii . (9.109)

The probability density function admits an absolutely continuous invariant measureon the Markov partition, the details of which can be found in [53]. For our discussion wenote that this measure can be used to find the Lyapunov exponent, and therefore quantifythe average rate of expansion or contraction for an interval under iteration. If we havea Markov partition P : 0 = c0 < c1 < .. . < cn−1 = 1, then the Lyapunov exponent � isexactly computed,

�=∫ 1

0ln∣∣τ ′(x)

∣∣ρ(x) dx (9.110)

=∫ c1

c0

ln∣∣τ ′(x)

∣∣ρ1 dx+ . . .+∫ cn−1

cn−2

ln∣∣τ ′(x)

∣∣ρn−1 dx

= ln[τ ′(c 12)]∫ c1

c0

ρ1 dx+ . . .+ [τ ′(cn− 12)]∫ cn−1

cn−2

lnρn−1 dx

=(n−1)∑i=1

ln |τ ′(ci− 12)|(ci − ci−1)ρi .

9.7.2 Nonlinear in an interval, as a limit of piecewise linear

Given a general transformation of the interval, τ : I → I which is not assumed to be eitherMarkov or piecewise linear, then we may estimate Lyapnov and other measurable andergodic quantities by refinement in terms of sequences of Markov transformations {τn}which uniformly estimate τ , and recalling that non-Markov transformations can be writtenas a weak limit of Markov transformations using Theorem 4.19, [19], at least in the scenarioproved for skew tent maps as discussed elsewhere in this text.

9.7.3 Pesin’s Identity Connects Lyapunov Exponents and MetricEntropy

The famous Pesin’s entropy identity, [263, 335, 203],

hμ(T )=∑

i:�i>0

�i (9.111)

provides a profound connection between entropy hK S and the (positive) Lyapunov expo-nents �i , under the hypothesis of ergodicity. In fact, a theorem of Ruelle, [285] established,

hμ(T )≤∑

i:�i>0

�i (9.112)

Page 310: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

302 Chapter 9. Information Theory in Dynamical Systems

under the hypothesis that T is differentiable and μ is an ergodic invariant measure on afinite-dimensional manifold with compact support.. In [113], Eckmann and Ruelle assertthat this inequality holds as equality often, but not always for natural measures. However,Pesin proved the equality holds at least if μ is a Lebesgue-absolutely continuous invariantmeasure for a diffeomorphism T , [263]. Since then a great deal of work has proceededin various settings including considerations of natural measure, of the infinite dimensionalsetting, and perhaps most interesting in the case of the nonhyperbolic setting of the presenceof zero Lyapunov exponents. See [335, 203] for further discussion.

A geometric interpretation of Pesin’s entropy formula may be stated as follows. Onthe one side of the formula, metric entropy describes growth rate of information states withrespect to evolution of the dynamics through partitions, as stated directly in Eq. (9.73).However, Lyapunov exponents describe an “average" growth rate of perturbations in char-acteristic directions of orthogonal successively maximal growth rate directions. Thus wecan understand the formula as stating that initial conditions with a given initial precisioncorresponding to initial hypercubes grow according to the positive exponents in time, thusspreading the initial states across elements of the partition implying new information gener-ated at a rate descriptive of these exponents. Considering this as an information productionprocess infinitesimally for small initial variations suggests the Pesin formula. Consideringfurther that Lyapunov exponents may be computed in two ways by the Birkhoff formula,either by averaging in time the differential information along “typical" (μ-almost every)initial conditions or alternatively by averaging amongst ensembles of initial conditions butweighting by the ergodic measure μ when it exists, this statement of the Birkhoff ergodictheorem provides two ways of computing and understanding metric entropy. See Eq. (1.5)and Example 1.4. Furthermore, often while sampling along a test orbit may provide thesimplest means to estimate Lyapunov exponents [332] and hence entropy hμ according tothe Pesin formula, alternatively computing Lyapunov exponents by Ulam’s method pro-vides another direct method for estimating hμ through Pesin’s identity.

9.8 Information Flow and Transfer EntropyA natural question in measurable dynamical systems is to ask which parts of a partitioneddynamical system influence other parts of the system. Detecting dependencies betweenvariables is a general statistical question and in a dynamical systems context this relates toquestions of causality. There are many ways one may interpret and computationally addressdependency. For example, familiar linear methods such as correlation have some relevanceto infer coupling from output signals from parts of a dynamical system, and these methodsare very popular especially for the simplicity of application [192]. A popular method is tocompute mutual information, I (X1; X2) in Eq. (9.35), as a method to consider dynamicalinfluence such as used in [108] in the context of global weather events as we review inSec. 9.9.2. However, both correlation and mutual information more so address overlap ofstates rather than information flow. Therefore time dependencies are also missed.

The transfer entropy TJ→I was recently developed by T. Schreiber [292] to be a sta-tistical measure of information flow, with respect to time, between states of a partitionedphase space in a dynamical system to other states in the dynamical system. Unlike othermethods that simply consider common histories, transfer entropy explicitly computes in-formation exchange in a dynamical signal. Here we will review the ideas behind transferentropy as a measurement of causality in a time evolving system. We present here our

Page 311: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.8. Information Flow and Transfer Entropy 303

work in [31] on this subject. Then we will show how this quantity can be computed us-ing estimates of the Frobenius-Perron transfer operator by carefully masking the resultingmatrices.

9.8.1 Definition and Interpretations of Transfer Entropy

To discuss transfer entropy in the setting of dynamical systems, suppose that we have apartitioned dynamical systems on a skew product space X ×Y ,

T : X ×Y → X ×Y . (9.113)

This notation of a single dynamical system with phase space written as a skew productspace allows a broad application as we will highlight in the examples and helps to clarifythe transfer of entropy between the X and Y states. For now, we will further write thissystem as if it is two coupled dynamical systems having x and y parts describing the actionon each component and perhaps with coupling between components,

T (x , y)= (Tx (x , y), Ty(x , y)). (9.114)

where,

Tx : X ×Y → X

xn �→ xn+1 = Tx (xn, yn), (9.115)

and likewise,

Ty : X ×Y → Y

yn �→ yn+1 = Ty(xn, yn). (9.116)

This notation allows that x ∈ X and y ∈ Y may each be vector (multivariate) quantities andeven of different dimensions from each other. See the caricature of this arrangement inFig. 9.10.

Let,x (k)

n = (xn , xn−1, xn−2, ..., xn−k+1), (9.117)

be the measurements of a dynamical system Tx , at times,

t (k) = (tn , tn−1, tn−2, ..., tn−k+1), (9.118)

sequentially. In this notation, the space X is partitioned into states {x} and hence xn denotesthe measured state at time tn . Note that we have chosen here not to index in any way thepartition {x}, which may be some numerical grid as shown in Fig. 9.10, since subindicesare already being used to denote time, and super indices denote time-depth of the sequencediscussed. So an index to denote space would be a bit of notation overload. We may denotesimply x , x ′ and x ′′ to distinguish states where needed. Likewise, y(k)

n denotes sequentialmeasurements of y at times t (k), and also Y may be partitioned into states {y} as seen inFig. 9.10.

Page 312: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

304 Chapter 9. Information Theory in Dynamical Systems

Figure 9.10. In a skew product space X ×Y , to discuss transfer entropy betweenstates {x} a partition of X and states {y} of Y , some of which are illustrated as x , x ′, x ′′ andy, y ′, y ′′, y ′′′, A coarser partition {a ,b} of X in symbols a and b and likewise {�0,�1} ofY in symbols 0 and 1 are also illustrated. [31]

• The main idea leading to transfer entropy will be to measure the deviation from theMarkov property, which would presume,

p(xn+1|x (k)n )= p(xn+1|x (k)

n , y(l)n ), (9.119)

that the state (xn+1|x (k)n ) does not include dependency on y(l)

n . When there is a depar-ture from this Markovian assumption, the suggestion is that there is no informationflow as conditional dependency in time from y to x . The measurement of this de-viation between these two distributions will be by a conditional Kullback-Leiblerdivergence which we will build toward in the following.

The joint entropy143 of a sequence of measurements written in the notation of Eqs. (9.117)-

143Definition 9.10.

Page 313: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.8. Information Flow and Transfer Entropy 305

(9.118) is,

H (x (k)n )=−

∑x (k)

n

p(x (k)n ) log p(x (k)

n ). (9.120)

A conditional entropy,144

H (xn+1|x (k)n )=−

∑p(xn+1, x (k)

n ) log p(xn+1|x (k)n )

= H (xn+1, x (k)n )− H (x (k)

n )

= H (x (k+1)n+1 )− H (x (k)

n ). (9.121)

is approximately an entropy rate145 which as it is written quantifies the amount of newinformation that a new measurement of xn+1 allows following the k-prior measurements,x (k)

n . Note that the second equality follows the probability chain rule,

p(xn+1|x (k)n )= p(x (k+1)

n+1 )

p(x (k)n )

, (9.122)

and the last equality from the notational convention for writing the states,

(xn+1, x (k)n )= (xn+1, xn , xn−1, ..., xn−k+1)= (x (k+1)

n+1 ). (9.123)

Transfer entropy is defined in terms of a Kullback-Leibler divergence, DK L (p1||p2)from Definition 9.15 but adapted for the conditional probabilities,146,

DK L (p1(A|B)||p2(A|B))=∑a,b

p1(a,b) logp1(a|b)

p2(a|b), (9.124)

The states are specifically designed to highlight transfer of entropy between the states Xto Y (or vice versa Y to X) of a dynamical system written as skew product, Eq. (9.113).Define [292],

Tx→y =∑

p(xn+1, x (k)n , y(l)

n ) logp(xn+1|x (k)

n , y(l)n )

p(xn+1|x (k)n )

, (9.125)

which we see may be equivalently written as a difference of entropy rates, like conditionalentropies,147,

Ty→x = H (xn+1|x (l)n )− H (xn+1|x (l)

n , y(k)n ). (9.126)

The key to computation is joint probabilities and conditional probabilities as they appearin Eqs. (9.126) and (9.129). There are two major ways we may make estimates of these

144Definition 9.11.145This is an entropy rate in the limit k→∞ according to Definition 9.17.146Recall that the Kullback-Leibler of a single random variable A with probability distribution is an er-

ror like quantity describing the entropy difference between the true entropy using the correct coding modellog p1(A) versus a coding model log p2(A) with a model distrubution p2(A) of A. This difference So condi-tional Kullback-Leibler is a direct application for conditional probability p1(A|B) with a model p2(A|B).

147Again, these become entropy rates as k, l→∞ as already discussed in Eq. (9.121)

Page 314: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

306 Chapter 9. Information Theory in Dynamical Systems

probabilities, but both involve course-graining the states. A direct application of formulasEqs. (9.120)-(9.121) and likewise for the joint conditional entropy to Eq. (9.125) allows,

Ty→x = [H (xn+1, xn)− H (xn)]− [H (xn+1, xn , yn)− H (xn, yn)], (9.127)

which serves as a useful method of direct computation.This may be a most useful form for computation, but for interpretation, a useful form

is in terms of a conditional Kullback-Leibler divergence,

Ty→x = DK L (p(xn+1|x (k)n , y(l)

n )||p(xn+1|x (k)n )), (9.128)

found by putting together Eqs. (9.124) and (9.125). In this form, as already noted inEq. (9.119), transfer entropy has the interpretation as a measurement of the deviation fromthe Markov property, that would be the truth of Eq. (9.119). That the state (xn+1|x (k)

n ) doesnot include dependency on y(l)

n suggesting that there is no information flow as a conditionaldependency in time from y to x causing an influence on transition probabilities of x . In thissense, the conditional Kullback-Leibler divergence Eq. (9.128) describes the deviation ofthe information content from the Markovian assumption. In this sense, Ty→x describes aninformation flow from the marginal subsystem y to marginal subsystem x . Likewise, andasymmetrically,

Tx→y = H (yn+1|y(l)n )− H (yn+1|x (l)

n , y(k)n ), (9.129)

and it is immediate to note that generally,

Tx→y �= Ty→x . (9.130)

This is not a surprise both on the grounds that it has already been stated that Kullback-Leibler divergence is not symmetric, but also there is no prior expectation that influencesshould be directionally equal.

A partition {z} serves as a symbolization which in projection by �x and �y is thegrid {x} and {y} respectively. It may be more useful to consider information transfer interms of a coarser statement of states. For example, see Fig. 9.10 where we represent apartition and � of X and Y respectively. For convenience of presentation we representtwo states in each partition,

= {a ,b}, and, � = {�0,�1}. (9.131)

In this case, then the estimates of all of the several probabilities can be summed in a mannerjust discussed above. Then the transfer entropy Tx→y becomes in terms of the states of thecoarse partitions. The question of how a coarse partition may represent the transfer entropyof a system relative to what would be computed with a finer partition has been discussedin [161] with the surprising result that the direction of information flow can be effectivelymeasured as not just a poor estimate by the coarse partition, but possibly even of the wrongsign. [31]

9.9 Examples of Transfer Entropy and Mutual Informationin Dynamical Systems

9.9.1 An Example of Transfer Entropy: Information Flow in Synchrony

In our recent paper [31], we chose the perspective that synchronization of oscillators is aprocess of transferring information between them. The phenomena of synchronization has

Page 315: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.9. Transfer Entropy and Mutual Information Example 307

been found in various aspects of nature and science[312]. It was initially perhaps a surprisewhen it was discovered that two, or many, oscillators can oscillate chaotically but if cou-pled appropriately they may synchronize and then oscillate identically even if all chaotictogether; this is a description of the simplest form of synchronization which is of identi-cal oscillators. See Fig. 9.11. Applications have ranged widely from biology[314, 151]to mathematical epidemiology[171], and chaotic oscillators[261], to communicational de-vices in engineering[72], etc. Generally the analysis of chaotic synchronization has fol-lowed a discussion of stability of the synchronization manifold, which is taken to be theidentity function when identical oscillators [260], or some perturbation thereof when non-identical oscillators, [316], often by some form of master stability function analysis.

Considering as we have reviewed in this text that chaotic oscillators have a corre-sponding symbolic dynamics description, then coupling must correspond to some form ofexchange of this information. Here we describe our perspective in [31] of coupled oscil-lators as sharing information, and then the process of synchronization is one where theshared information is an entrainment of the entropy production. In this perspective, whenoscillators synchronize, it can be understood that they must be sharing symbols in orderthat they may each express the same symbolic dynamics. Furthermore, depending on thedegree of co-coupling, or master-slave coupling or somewhere in between, the directional-ity of the information flow can be described by the transfer entropy. A study of anticipatingsynchronization with a transfer entropy perspective in the note of studying the appropriatescale necessary to infer directionality is found in [161].

Consider the following skew tent map system to use as an example coupling elementto highlight our discussion, [169] which is a full folding form [19], meaning two-one,

fa(x)={ x

a if 0≤ x ≤ a1−x1−a if a ≤ x ≤ 1

}, (9.132)

Let us couple these in the following nonlinear manner, [169],(xn+1yn+1

)=G

(xnyn

)=(

fa1 (xn)+ δ(yn− xn)fa2 (yn)+ ε(xn− yn)

). (9.133)

Note that written in this form, if a1 = a2 and ε=0 but δ > 0 we have a master slave systemof identical systems as illustrated in Fig. 9.11 where we see a stable synchronized identitymanifold with error decreasing exponentially to zero. On the other hand, if ε = δ but a1 �=a2 we can study symmetrically coupled but nonidentical systems in Fig. 9.12, where theidentity manifold is not exponentially stable but is apparently a Lyapunov-stable manifoldas the error, error (n)= |x(n)− y(n)| remains small for both scenarios shown in the figures,a1 = 0.63 but a2 = 0.65 and a2 = 0.7 respectively, with progressively larger but stableerrors. Our presentation here will be designed to introduce the perspective of transferentropy to understand the process of synchronization in terms of information flow, andfrom this perspective to gain not only an idea of when oscillators synchronize but perhapsif one or the other is acting as a master or a slave. Furthermore, the perspective is distinctfrom a master stability formalism.

With coupling resulting in various identical and nonidentical synchronization scenar-ios as illustrated in Fig. 9.12, we will may analyze the information transfer across a studyof both parameter matches and mismatches and across various coupling strengths and di-rectionalities. In Figs. 9.13-9.14, we see the results of transfer entropies, Tx→y and Ty→x

Page 316: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

308 Chapter 9. Information Theory in Dynamical Systems

Figure 9.11. In a nonlinearly coupled skew tent map system, Eq. (9.133), ofidentical oscillators, a1 = a2 = 0.63 and master slave configuration, δ = 0.6, ε = 0.0(parameters as in [169]). Note (above) how the signals entrain and (below) the error,error (n)= |x(n)− y(n)| decrease exponentially. [169]

respectively in the scenario of identical oscillators a1 = a2 = 0.63 for coupling parametersbeing swept 0 ≤ δ ≤ 0.8 and 0 ≤ ε ≤ 0.8. We see that due to the symmetry of the formof the coupled systems, Eq. (9.133), the mode of synchronization is opposite as expected.When Tx→y is relatively larger than Ty→x then the interpretation is then that relativelymore information is flowing from the x system to the y system, and vice versa. This sourceof communication is due to coupling the formulation of synchronization. Large changes inthis quantity signals the sharing of information leading to synchronization.

In the asymmetric case, 0.55≤ a1,a2 ≤ 0.65 we show a master-slave coupling ε = 0,δ = 0.6 in Fig. 9.15 and compare to Figs. 9.11-9.12. In the master-slave scenario chosen,the x oscillator is driving the y oscillator. As such, the x-oscillator is sending its statesin the form of bits to the y oscillator as should be measured that Tx→y > Ty→x whensynchronizing and more so when a great deal of information “effort" is required to maintainsynchronization. This we interpret as what is seen in Fig. 9.15 in that when the oscillatorsare identical, a1 = a2 shown on the diagonal, the transfer entropy difference Tx→y > Ty→xis smallest since the synchronization requires the smallest exchange of information oncestarted. In contrast, Tx→y > Ty→x is largest when the oscillators are most dissimilar, andwe see in Fig. 9.13 how “strained" the synchronization can be seen since the error cannot

Page 317: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.9. Transfer Entropy and Mutual Information Example 309

Figure 9.12. A nonlinearly coupled skew tent map system, Eq. (9.133), ofnonidentical oscillators, a1 = 0.63,a2 = 0.65 and master slave configuration, δ = 0.6,ε = 0.0. Note (above) how the signals approximately entrain and (below) the error,error (n)= |x(n)− y(n)| decrease close to zero, where it remains close to an identity man-ifold, x = y where it is stable in a Lyapunov stability sense. [31]

go to zero as the oscillators are only loosely bound.

9.9.2 An Example of Mutual Information: Information Sharing in aSpatiotemporal Dynamical System

Whereas transfer entropy is designed to determine direction of information flow, mutual in-formation I (X1; X2) in Eq. (9.35) is well suited to decide the simpler question as to whetherthere is simply a coupling in a large and complex dynamical system. The advantage of us-ing the simpler but perhaps less informative measure, as it does not give directionality, isthat it may require less data.

A recent and exciting application of mutual information comes from an importantspatiotemporal dynamical systems from analysis of global climate, [108] as seen in Fig. 9.16.The study in [108] used monthly averaged global SAT field since as they stated, to cap-ture the complex dynamics on the interface between ocean and atmosphere due to heatexchange and other local processes. This allowed the study of atmospheric and oceanicdynamics using the same climate network. They used data provided by the National Center

Page 318: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

310 Chapter 9. Information Theory in Dynamical Systems

Figure 9.13. Transfer entropy, Tx→y measured in bits, of the system Eq. (9.133),in the identical parameter scenario a1 = a2 = 0.63 which often results in synchronizationdepending on the coupling parameters swept, 0≤ δ ≤ 0.8 and 0≤ ε ≤ 0.8 as shown. Con-trast to Ty→x shown in Fig. 9.14 where the transfer entropy clearly has an opposite phaserelative to the coupling parameters, (ε,δ). [31]

for Environmental Prediction/National Center for Atmospheric Research (NCEP/NCAR)and model output from the World Climate Research ProgrammeÕs (WCRPÕs) CoupledModel Intercomparison Project phase 3 (CMIP3) multimodel data set.

This spatiotemporal data can be understood as a time-series, xi (t) at each spatial sitei modeled on the globe. Pairs of sites, i , j can be compared for the mutual information inthe measured values for states in the data xi (t) and xj (t) leading to I (Xi ; Xj ). Followinga thresholding decision leads to a matrix of couplings Ai, j descriptive of mutual informa-tion between sites on the globe. The interpretation is that the climate at sites with largevalues recorded in Ai, j are somehow dynamically linked. In Fig. 9.16, we illustrate whatwas shown in [108] which is a representation of the prominence of each site i on the globecolored according to that prominence. The measure of prominence shown is the vertex be-tweenness centrality, labelled BCi . Betweenness centrality is defined as the total numberof shortests paths148 in the corresponding undirected graph which pass through the vertex

148In a graph, G = (V , E) consisting of the set of vertices and edges which are simply vertex pairs, a a pathbetween i and j of steps along edges of the graph which connect a pair of vertices. A shortest path betweeni and j is a path which is no longer than any other path between i and j .

Page 319: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

9.9. Transfer Entropy and Mutual Information Example 311

Figure 9.14. Transfer entropy, Ty→x measured in bits, of the system Eq. (9.133),in the identical parameter scenario a1 = a2 = 0.63 which often results in synchronizationdepending on the coupling parameters swept, 0≤ δ ≤ 0.8 and 0≤ ε ≤ 0.8 as shown. Com-pare to Tx→y shown in Fig. 9.13. [31]

labelled i . BCi can be considered as descriptive as to how important is the vertex i to anyprocess running on the graph. Since the graph is built by considering mutual information,it may be inferred that a site i with high BCi is descriptive of a dynamically important sitein the spatiotemporal process, in this case the process being global climate. It is strikinghow this information theoretic quantification of global climate agrees with known oceano-graphic processes as shown in Fig. 9.16.

Page 320: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 9.15. Transfer entropy difference, Ty→x − Ty→x measured in bits, of thesystem Eq. (9.133), in the nonidentical parameter scenario sweep, 0.55≤ a1,a2≤ 0.65, andmaster-slave coupling ε = 0, δ = 0.6. Compare to Tx→y shown in Fig. 9.13. Contrast toTy→x shown in Fig. 9.14 where the transfer entropy clearly has an opposite phase relativeto the coupling parameters, (ε,δ). Also compare to Figs. 9.11-9.12. [31]

Page 321: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Figure 9.16. Mutual Information mapping of a global climate network from[108]. (Top) Underlying and known global oceanographic circulations. (Bottom) Between-ness centrality BCi from a network derived from mutual information I (Xi ; Xj ) between theglobal climate data from spatial sites i , j across the globe. Compare the mutual informa-tion theoretic quantification of the global climate is agreeable with the known underlyingoceanographic processes.

Page 322: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure
Page 323: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Appendix A

Computation, Codes, andComputational Complexity

A.1 Matlab Codes and Implementations ofUlam-Galerkin’s Matrix and Ulam’s Method

In Chapter 4, we discussed several different ways of developing the Galerkin-Ulam matrixPi, j approximation of the Frobenius-Perron matrix. We presented the theoretical construc-tion Eq. (4.4) in terms of projection onto a finite subspace of basis functions. In Eq. (4.6)and in Remark 4.1 we discussed the description of mapping sets across sets, which is moreclosely related to the original Ulam work, [319], and closely related to the exact descrip-tions using Markov partitions discussed in Sec. 4.2. Here we will highlight the samplingmethod using a test orbit, and Eq. (4.7), which we repeat,

Pi, j ≈ #({xk|xk ∈ Bi and F(xk) ∈ Bj })#({xk ∈ Bi }) . (A.1)

In the code which follows in Sec. A.1.1, we choose {Bi} to be a grid as a tesselation oftriangles to cover the domain of interest, which should contain the entire test orbit {xk}Nk=1.Theoretical discussion will be presented in Sec. A.5.2 of how the grid size, the test or-bit length N and system regularity parameters are related for good approximations of theFrobenius-Perron operator.

A.1.1 Galerkin-Ulam Code by Delauney Triangulation

The first Matlab code presented here assumes a test orbit {xk}Nk=1 ⊂ R2 and uses a Delau-

nay triangulation as the grid to cover the region of input interest uniformly by triangles.The output of the code is primarily the stochastic matrix, A and the arrays concerningthe grid. We present triangles which for technical computer science indexing reasons areeasy to work with. Delaunay triangulations are particularly popular in the numerical par-tial differential equations finite-element literature for this reason. For our concern here,this popularity of such methods allows us to save many steps toward writing compact andsimple Matlab code regarding indexing, identifying occupancy, plotting and so forth sincethere are already many Delaunay triangulation subroutines built into Matlab. See Matlabhelp pages concerning, “Tessellation and Interpolation of Scattered Data."

315

Page 324: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

316 Appendix A. Computation, Codes, and Computational Complexity

While our code below is specialized for a two-dimensional dynamical system, gen-eralizing it for a three-dimensional system is straight forward in part due to the strengthof tessellations which are much easier to work with in general domains than other populargrid geometries such as rectangles, but also the Matlab built in routines really shine in thisapplication. Also, while we have assumed a uniform grid, if a nonuniform grid is needed,such as for refining regions of high observed invariant density, and perhaps large Lipschitzconstants and discussed in Sec. A.5.2, then the tessellation structures are particularly wellsuited for refinement.

1 %%%%%%%%%%%%%%%2 %% by Erik Bollt3 %%%Build an Ulam-Galerkin’s Matrix Based on a Test Orbit Visiting Triangles4 %%%of a Tesselation5 %%%%%%%%%%%%%%%6 % Input:7 % z - Test orbit is n x 2 for an n-iterate orbit sample8 % h - is the side length of the triangle edges which share a right9 % angle

10 % ax, bx, ay, by - the low and high ends of a box bounding data11 % Output:12 % dt - a DelaunayTri13 %%14

15 function [dt,ll,A,zz]=TransitionMatrix(z,h,ax,bx,ay,by)16

17 %%18

19 %low=-2; high=2; low=0; high=1;20 [X1,X2] = ndgrid(ax:h:bx, ay:h:by);21 [m,n]=size(X1);22 x1=reshape(X1,m*n,1); x2=reshape(X2,m*n,1);23

24 %Formulate Delaunay Triangulation of region25 dt= DelaunayTri([x1 x2]); %See Matlab Subroutine DelauneyTri for26 %input/output information27 %dt is the triangulation class28 %29 % where30 %31 %dt.dt.Triangulation is an m1 by 3 array of32 %integers labeling the vertex corner numbers33 %of the triangles34 %35 % and36 %37 %dt.X is an m2 by 2 array of real numbers38 %defining positions39

40 %triplot(dt) %Optional plot command of this triangulation41

42 %%43

4445 %Count number of orbit points in z which cause a triangle to be counted as46 %occupied (and otherwise a triangle is not counted as it is empty until47 %observed occupied4849 nottrue=0;50 while(nottrue<1)51 nottrue=1;52 SI = pointLocation(dt,z); %Matlab command, Locate the simplex element in53 %dt containing the specified locations of each

Page 325: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.1. Matlab Codes: Ulam-Galerkin’s Method 317

54 %of the elements of the array z of orbit55 %samples56 %57 l=unique(SI); %Matlab subrouting: Count number of unique instances in SI58 k=1;59 while(isnan(l(k))<1&&k<length(l)) %Matlab subroutine True if Not-a-Number60 [ii,j]=find(SI==l(k)); %Collect locations corresponding to each unique l.61 cnt(k)=sum(j);62 k=k+1;63 end64 k=k-1; ll=l(1:k); cnt=reshape(cnt,size(ll));65 end66 %%67

68 % Plot those simplex elements of the dt which are occupied by an orbit69 % iterate of z - dt(ll,:) are those occupied simplex elements70 patch(’faces’,dt(ll,:), ’vertices’, dt.X, ’FaceColor’,’r’);71 N=length(ll);7273 %%74

75 %Translateback from ll back to phase space positions z by using the center76 %positions with Matlab subroutine "mean"77 zz=zeros(N,2);78 for i=1:N79 zz(i,:)=mean([dt.X(dt.Triangulation(ll(i),1),:)80 dt.X(dt.Triangulation(ll(i),2),:)81 dt.X(dt.Triangulation(ll(i),3),:)]);82 end8384 %%Now build the transition matrix A85 %86 %So that A(i,j)>0 iff there is an element in simplex element i such that87 % there is an iterate z(k,:) and that88 % the next iterate, z(k+1,:), transitions to simplex element j89 A=zeros(N,N);90 for i=1:(length(SI)-1)91 ii= find(ll==SI(i));92 jj=find(ll==SI(i+1));93 %[ii jj size(A)]94 A(ii,jj)=A(ii,jj)+1;95 end96 %Now make A into a stochastic matrix by row normalizing97 for i=1:N98 q=length(find(SI==ll(i)));99 A(i,:)=A(i,:)./q;

100 A(i,:)=A(i,:)./sum(A(i,:));101 end

A.1.2 Routine to test run subroutine TransitionMatrix.m

Here we give a sample test code to run the Galerkin-Ulam code TransitionMatrix.m above,with a test orbit produced using the standard map, from standard.m below, we param-eter k = 1.2 which is greater than the “magic" value k = 0.97.. such that the last torusdisappears, corresponding to the golden mean, thus allowing orbits to have unboundedconjugate-momentum, [229]. But still the phase space is highly compartmentalized intoalmost invariant sets as this system is only weakly transitive. See Figs. A.1-A.4 and com-ments therein describing the code and linking to the algorithms that they illustrate. TheFigure references are at the appropriate locations in the test matlab code.

Page 326: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

318 Appendix A. Computation, Codes, and Computational Complexity

1 close all; clear;2 nPlot = 100000; % nPlot-iterate test orbit many points after the transient.3 M=250; diam=10; a=0; b=1;4 h=0.025

6 %%7 %%%%%%%Make a Sample Orbit of the Standard Map8 k=1.2; %k is chosen >0.97... for the magic breakup of the "golden mean" torus9 z=[];

10 % Set some variables11 transientSteps = 100; %length of initial transient of test orbit to ignore12 initialX = rand(2,1);13 x = initialX; hold on;1415 %%16 % Throw away transients:17 for i=1:transientSteps18 x = standard(x,k); %Use as a test example, the standard map.19 end20

21 %%22 % Plot the next nPlot points visited:23 for i=1:nPlot24 x = standard(x,k);25 z=[z;x’];26 end27 z=mod(z,1);28 plot(z(:,1),z(:,2),’b.’,’markersize’,10); hold on;29 %30 %%31

32

33 %%%%%%%%%%%%%%%34 %%%Build an Ulam-Galerkin’s Matrix Based on a Test Orbit Visiting Triangles35 %%%of a Tesselation36 [dt,ll,A,zz]=TransitionMatrix(z,h,a,b,a,b);37 triplot(dt) %Matlab routine to draw the triangulation simplex38 drawnow;39

40 %%41 [v,d]=eigs(A’,2); w=abs(v(:,1)); %Compute 1st and send eigenvalues/vectors of42 %Galerkin-Ulam Matrix A43 %44 figure; stem3(zz(:,1),zz(:,2),w(:),’fill’); %Show the dominant eigenvector45 %(d(1)=1) meant to roughly46 %estimate invariant density47 %48 w2=v(:,1); [i,ww]=find(w2>0);49 figure; stem3(zz(i,1),zz(i,2),w2(i),’r’, ’fill’)50

51 %% Produce a reversible Markov Chain R52 P=A;53 [v,d]=eigs(A’,1);54 N=size(A,1);55 for i=1:N56 PI= v(:,1); Phat=57 spdiags((1./PI),0,length(PI),length(PI))*P’*spdiags(PI,0,length(PI),length(PI)58 );59 for j=1:N60 Phat(i,j)=v(j,1)*P(j,i)/v(i,1);61 end62 end63 R=(P+Phat)/2;64

Page 327: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.2. Ulam-Galerkin Code by Rectangles 319

65 %%66 %Partition!67 [w,l]=eigs(R,4);68 figure; plot(w(:,2))69 c=0; eps=0.005;70 [i,ww]=find(w(:,2)>c);71 [ii,ww]=find(w(:,2)<=c);72 [iii,ww]=find(abs(w(:,2))<eps);73

74 figure; hold on;75 patch(’faces’,dt(ll(i),:), ’vertices’, dt.X, ’FaceColor’,’r’);76 patch(’faces’,dt(ll(ii),:), ’vertices’, dt.X, ’FaceColor’,’b’);77 patch(’faces’,dt(ll(iii),:), ’vertices’, dt.X, ’FaceColor’,’k’);

149

Subroutine for the standard map function.

1 function xvecNew = standard(xvec,k)2 % Matlab coded in vectorized form of standard3 %Input:4 % xvec - vector of initial conditions5 % k - standard map parameter6 %7 %Output:8 % xvecNew - vector of their images9 %

10 xvecNew = zeros(size(xvec));11 xvecNew(1,:) = xvec(1,:)+xvec(2,:)-k*sin(2*pi*xvec(1,:))/(2*pi);12 xvecNew(2,:) = xvec(2,:)-k*sin(2*pi*xvec(1,:))/(2*pi);

A.2 Ulam-Galerkin Code by RectanglesIn the previous section we gave subroutines to build a stochastic matrix model of theFrobenius-Perron operator following covering the region by a Delaunay triangulation. Al-though we gave a two dimensional and uniform example, we argue that such a code is morerobust and flexible to difficulties of domain irregularities, dimensionality complexity suchas dimensionality greater than two, and it is more easily developed into a adaptive refine-ment method for nonuniform grid sizes to compensate for non homogeneity of the dynam-ical system on the domain. These are presented in subsequent sections. These strengths areall significantly difficult to implement as a practical code if the basic grid elements are rect-angles. Rectangles are nonetheless a very popular basic element often used when studying

149Matlab becomes faster and more efficient as a computing platform if we leverage its strength us-ing “array arithmetic." Without giving great details as to the programming intricacies of this language,we state simply that loops, and multiply nested loops especially, can often be replaced with array arith-metic. For example, lines 52 through 56 above can be replaced with a single line, P I = v(:,1); Phat =diag(1./P I )∗ P ′ ∗diag(P I ); which is both briefer and easier to read (once we are used to the techniques)and many times faster for the computer. An even more efficient method would be to leverage sparse ver-sions of matrix manipulations, (especially useful for huge matrices where creating a full diagonal matrixis prohibitive) by using for example, P I = v(:,1); Phat = spdiags((1./P I ),0, length(P I ), length(P I ))∗P ′ ∗spdiags(P I ,0, length(P I ), length(P I )); We will not point out any further specific lines of efficienciesto be gained by eliminating multiple loops. We emphasize that we have chosen the multiple loop style ofprogramming here for pedagogical reasons as this is likely the most commonly easily read format for mostreaders.

Page 328: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

320 Appendix A. Computation, Codes, and Computational Complexity

Figure A.1. A Delaunay triangulation in the domain [0,1]× [0,1], for a testorbit {xi }Ni=1 from a standard map with k = 1.2. Using the code TransitionMatrix.m, theinitial triangulation is pruned to a smaller collection of triangles to include only thosetriangles which are visited by the given test orbit. The test orbit is shown as blue dots, andvisited triangles are shown as filled red. This smaller number of triangles, M defines thedimensions of the M×M stochastic matrix A which is developed in the code according toEq. (4.7) as an Ulam-Galerkin estimate of the Frobenius-Perron operator.

a dynamical system in terms of cell-mapping methods, [176]. For this reason, we also givehere a simple rectangles-based code to develop an adjacency matrix, and to demo a fewstandard codes as an efficient coding example. An extremely robust code called GAIOwhich is based on rectangles has been quite popular amongst several groups, as it has beenused in both three dimensional problems and problems with grid refinement, [84], but wewill continue to suggest the strength of tesselations in each of these scenarios.

The code follows, together with Fig. A.5 which shows the resulting adjacency ma-trix of the operator estimate. Recall that the stochastic matrix is developed according toEq. (A.1) by relative occupancy in the grid elements of the sample orbit. In the code thatfollows, we develop instead the adjacency matrix for sake of discussion and presentation.Recall that an adjacency matrix B is a matrix of zero’s and ones and Bi, j = 1 is Pi, j > 0 andBi, j = 0 otherwise in terms of the stochastic matrix A from Eq. (A.1); it simply identifiestransitions without weights. The code below is easily adjusted to produce the stochasticmatrix if we wish.

Another useful memory structure to store a sparse matrix, which Ulam-Galerkin ma-trix are often quite sparse, is the link list. If A is a sparse stochastic matrix of size N × N ,with n nonzero values Pi, j > 0, then let, L be n×3. If Pi, j is the kth nonzero element of

Page 329: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.2. Ulam-Galerkin Code by Rectangles 321

Figure A.2. The dominant eigenvector of the stochastic matrix Ulam-Galerkinestimate of the Frobenius-Perron operator A is often used as an Ulam’s method estima-tor of the invariant density. Here, using the same {xi}Ni=1 test orbit from a standard mapwith k = 1.2 yields the following eigenvector which is reshaped to agree with the orig-inal domain. We emphasize with this example that while the Ulam’s method is popularfor estimating invariant density, the standard map satisfies none of the few theorems thatguarantee the success of the Ulam’s method, and furthermore it is not even known if thestandard map even has an invariant measure. Nonetheless, this illustration demonstratesthe use of the code and at least we see the support on the tessellation of the test orbit.Even with an incorrect computation of invariant density, some success can still be foundregarding estimates of almost-invariant sets shown in Figs. A.3-A.4.

the matrix (the specifics of this order are not important),

L(k,1)= i

L(k,2)= j

L(k,3)= Pi, j (A.2)

Such sparse representations matrix can be huge memory savers for large matrices. N scalesas (1/h)d for a domain of dimension d (here d = 2) and a rectangle size h. A scales as N2

since it is a square matrix, and L is of size

card(L)= 3n, where 0≤ n ≤ N2, (A.3)

Page 330: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

322 Appendix A. Computation, Codes, and Computational Complexity

Figure A.3. Second eigenvector v2 of the reversible Markov chain, as definedEq. (5.57) in Sec. 5.4.

so potentially 3n could be larger than the matrix size N2, 3n > N2. In practice n is muchsmaller. If each rectangle stretches across several rectangles under the action of the map(see Fig. 4.1), roughly according to the Lyapunov number l on average, then roughly weexpect the scaling,

n ∼ l N and l N << N2. (A.4)

1 clear; close all;2 %Initial value x0, which is taken from 0 to 13 x0=[0;0];4 %Set the grid size, ie. number of boxes are M by M5 M=100;6 %Set the size of the sampling points7 N=100000;89 %% Envolve the initial value x0 for N times, and record each point

10 x=zeros(N+1,2);11 x(1,:)=x0;12 a=1.2;13 b=0.4;14 % Iterate the Henon Map15 for i=1:N16 x(i+1,:) = henon(x(i,:),a,b);17 end

Page 331: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.2. Ulam-Galerkin Code by Rectangles 323

Figure A.4. Observed almost invariant sets of the standard map test orbit {xi}Ni=1,on the pruned tessellation shown in Fig. A.1. Following the methods in Sec. 5.4 from theCourant-Fischer Theorem 5.7 for almost invariant sets in Reversible Markov Chain de-veloped from the stochastic matrix, as defined Eq. (5.57), [134], the second eigenvector v2shown in Fig. A.3 can be thresholded to reveal those tessellation elements corresponding toweakly transitive components, colored red and blue, with those triangles on the boundarycolored black. Note that these components found agree with the expectation that the can-tori remnants of the golden mean frequency make for slow transport between the coloredregions, [229].

18

19 %% Now using the test orbit array x,20 %% find which unique boxes the iterates land in21 I0=FindBox(x,M); %Which boxes are hit - using FindBox Below22 I=Reorder(I0); %A sorting command - using subroutine below23

24 %Develop directed graph as a Link List structure25 L1=I(1:N); % from node26 L2=I(2:(N+1)); % to node27 %%2829 %% Convert Link List to Adjaceny Matrix30 A=zeros(max(I),max(I));31 % A(L1,L2)=1; this is wrong32 L_ind=sub2ind(size(A),L1,L2);33 A(L_ind)=1;

Page 332: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

324 Appendix A. Computation, Codes, and Computational Complexity

34

35 %View the adjacency matrix by spy-plot, built in Matlab polotting program36 spy(A);title(’Adjacent Matrix of Henon Map’); %See Fig. 11.53738 Asp=sparse(A); %turn Matrix A into Sparse Matrix39

40 %%%%%%%%%%%41 %%42 %This function determine points belong to which box43 %input x is a vector44 function y=FindBox(x,BoxSize)45

46 for nb = BoxSize47 % First convert coords to integers:48 ix = scaleToIntegers(x(:,1),nb);49 iy = scaleToIntegers(x(:,2),nb);50 ixy = ix + (iy-1)*nb; % Each box now has a unique integer address.51 end52 y=ixy;53

54

55 %%%%%%%%%%%56 %%57 % Scaling function58 function ix = scaleToIntegers(x,n)59 % Return x scaled into 1:n and rounded to integers.60 ix = 1+round((n-1)*(x-min(x))/(max(x)-min(x)));61

62

63 %%%%%%%%%%%64 %%65 % re-order vector I66 % for example if I=[2001 4010 1 44 1], then Reorder(I)=[4 5 1 3 1]67 function y=Reorder(vector)68

69 [b,m,n]=unique(vector);70 %From Matlab help: for the array A returns the same values as in A71 %but no repetitions.72 y=n;

A.3 Delauney Triangulation in a Three-DimensionalPhase Space

A Delauney tessellation really shines when we cover an attractor in a three-dimensionalphase space. The code in this section should be compared to the code “TransitionMa-trix.m" in Sec. A.1.1, which produces the analogous outer-cover by h-sized triangles, butnow in terms of h-sized tesselations. Not that a key difference in the array that stores thetesselations was dt.Triangulation that is an m1 by 3 array of integers labeling the vertex cor-ner numbers when a two-dimensional phase space, but m1 by 4 when a three-dimensionaltessellation, as expected by the dimensionality of the corresponding objects.

Note that in using a Delauney triangulation, the heavy lifting of the covering, in-dexing, and accounting of triangles is handled by the Delauney Matlab subroutines De-launayTri(.), and supporting routines such pointLocation(.) to decide which tessellationelement a given point is located. These are highly refined subroutines that are includedin the Maltab PDE toolbox suite. Delaunay tesselations are commonly used in the partialdifferential equations community, especially finite element methods, for their power in grid

Page 333: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.3. Delauney Triangulation in a Three-Dimensional Phase Space 325

Figure A.5. Spy plot of adjacency matrix from orbit from the Henon map.

Figure A.6. (Left) Test points of a sampled orbit of the Lorenz equations on (near)the Lorenz attaractor. (Right) A tessellation by Delaunay triangulation out-covering theLorenz attractor.

generation in complex domains, and for their use in grid refinement. We highlight both ofthese aspects in this section, and the next section respectively.

This code may be contrasted to a popular code suite called GAIO, [84], which per-forms a covering by rectangles rather than a tessellation which we offer as having geometricadvantages.

Here we do not include the code to produce a stochastic matrix approximation ofthe Frobenius-Perron operator, as once covered with indexed triangles, the same formulaEq. (A.1) applies using code as in the previous sections for the two-dimensional mappings.

1 clear; close all;2 %This should be any n x 3 array of real value numbers

Page 334: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

326 Appendix A. Computation, Codes, and Computational Complexity

3 %as a test orbit in a 3D phase space4 %in this case the test orbit is from a Lorenz attractor5 load ’LorenzDat.mat’6 z=X;7

8 %Produce a grid that covers the range n each coordinate9 xlow=floor(min(X)); xhigh=ceil(max(X)); h=2.5; %And choose a step size of h

10 [X1,X2,X3]=ndgrid(xlow(1):h:xhigh(1),xlow(2):h:xhigh(2),xlow(3):h:xhigh(3));11 m=size(X1);12 x1=reshape(X1,m(1)*m(2)*m(3),1);13 x2=reshape(X2,m(1)*m(2)*m(3),1);14 x3=reshape(X3,m(1)*m(2)*m(3),1);15

16 %Formulate Delaunay Triangulation of region17 dt= DelaunayTri([x1 x2, x3]);18 %See Matlab Subroutine DelauneyTri for input/output information19 %dt is the triangulation class20 %21 %dt.dt.Triangulation is an m1 by 4 array of22 %integers labeling the vertex corner numbers23 %of the triangles24 %25 %dt.X is an m2 by 3 array of real numbers26 %defining positions27

28 %Count number of orbit points in z which cause a triangle to be counted as29 %occupied (and otherwise a triangle is not counted as it is empty until30 %observed occupied31 nottrue=0;32 while(nottrue<1)33 nottrue=1;34 SI = pointLocation(dt,z); %Matlab command, Locate the simplex element in dt35 %containing the specified locations of each36 %of the elements of the array z of orbit37 %samples38 %39 l=unique(SI); %Matlab subrouting: Count number of unique instances in SI40 k=1;41 while(isnan(l(k))<1&&k<length(l)) %Matlab subroutine True for Not-a-Number42 [ii,j]=find(SI==l(k)); %Collect locations corresponding to unique l.43 cnt(k)=sum(j);44 k=k+1;45 end46 k=k-1; ll=l(1:k); cnt=reshape(cnt,size(ll));47 end48 %%49

50 % Plot those simplex elements of the dt which are occupied by an orbit51 % iterate of z - dt(ll,:) are those occupied simplex elements52 figure; subplot(1,2,1); plot3(X(:,1),X(:,2),X(:,3),’.’);53 xlabel(’x’); ylabel(’y’); zlabel(’z’);54 subplot(1,2,2); plot3(0,0,0);55 patch(’faces’,dt(ll,:), ’vertices’, dt.X, ’FaceColor’,’r’); xlabel(’x’); ylabel(’y’); zlabel(’

A.4 Delauney Triangulation and RefinementA major strength of the Delaunay tesselations is the flexibility for grid refinement, espe-cially in several variables. The codes here can be contrasted to the popular GAIO codes [84]which also have demonstrated adaptive refinement but for rectangular shaped elements.

The adaptive strategy used in this section in our Matlab codes implements the sim-plest of possible strategies:

Page 335: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.4. Delauney Triangulation and Refinement 327

Figure A.7. Successive refinements of an initially coarse h = 2.0 Delaunay tes-sellation grid covering 10,000 sample orbit points from the Henon map. In each finer gridshown, triangles have been split in half if there are more than thresh = 10 sample or-bit elements in that triangle. From the coursest initial grid in the upper left to the finestgrid shown after 8 steps in the lower right, refinement proceeds until the stopping criterionthresh = 10 is attained after 27 refinement steps as shown in Fig. A.8.

Page 336: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

328 Appendix A. Computation, Codes, and Computational Complexity

Figure A.8. From h = 2.0 shown in Fig. A.7, 27 refinement steps leaves an adap-tive Delaunay tessellation grid as shown that meets a stopping criterion that there are nomore than thresh = 10 orbit points in each triangle. (Left) the full grid covering the do-main, but those triangles covering the orbit are small of edge size h ∼ 1.7263∗10.−4. Thuswe have a fine cover with relatively fewer triangles than the uniform grid of tiny trianglesusing the codes of the previous section.

1. Starting with a coarse grid, of size h-triangles (on the triangles short side), thenprogressively split triangles in half (half with respect to area) if there is more orbitdata sampled in each triangle than a minimum threshold. See comments in Figs. A.7-A.8 for discussion regarding the refinements.

2. Furthermore, any empty triangles are removed from consideration.

The main point is that relatively few elements (triangles in our code) are needed to coverthe attractor, and furthermore a fine grid improves accuracy. This second issue issue canbe argued analytically by discussing Lipschitz constants as in Sec. A.5.2 and associatedliterature, but we give a novel discussion of both of these operations in terms of sharpeningthe variance of Monte-Carlo estimators of the integrals, as we argue in Sec. A.5.1.

Further, removing empty elements represents a particularly useful computational sav-ings when the box-counting fractal dimension db of the attractor [255] is relatively smallcompared to the embedding dimension d . Then the number of triangles for a total cov-ering grows exponentially with respect to the “dimension," but not as quickly if the box

Page 337: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.4. Delauney Triangulation and Refinement 329

dimension is used instead of the embedding dimension if db < d ,

h−dB ∼<< h−d . (A.5)

For example, taking the Henon mapping shown, an accepted value [153], db = 1.28±0.01 < d = 2. In the Figs. A.7-A.8 shown, there results n = 36439 triangles total coveringthe attractor with a fine grid to the same level of refinement as a fine uniform triangulation ofN = 1.0738∗109 triangles. Since developing a stochastic matrix of sizes n×n versus N ×N , we see that substantial quadratic memory and computational savings. See discussionin the next sections concerning estimation precision of the transfer operator and coveringprecision.

1 close all; clear;2 h=2; %Initial Grid size3 [X1,X2] = ndgrid(-2:h:2, -2:h:2); %Create uniform grid of triangles sized h;4 [m,n]=size(X1);5 x1=reshape(X1,m*n,1); x2=reshape(X2,m*n,1);67 %Initialize8 a=1.2; global b; b=0.4; %Henon map parameters9 z=[]; %collected orbit, initially empty

10 transientSteps = 10;11 initialX = [0;0];12 nPlot = 100000; % Plot this many points after the transient.1314 x = initialX;15 for i=1:transientSteps % Throw away transients:16 x = henon(x,a);17 end18 % Plot the next nPlot points visited:19 for i=1:nPlot20 x = henon(x,a); %Iteratre the Henon map21 z=[z;x’];22 end23 plot(z(:,1),z(:,2),’b.’,’markersize’,10); hold on;24

25 dt= DelaunayTri([x1 x2]); %dt= DelaunayTri(zsamp);26 triplot(dt)27

28

29 thresh=10; %The threshold number of samples we require in each triangle,30 %else split the triangle31 P=[1 2; 1 3; 2 3]; %Labels for triangle corners32 nottrue=0; qq=0; %Triggers used in deciding if refinement is done33 while(nottrue<1)34 nottrue=1;35 SI = pointLocation(dt,z);36 ln=unique(SI);37 n=hist(SI,ln);38

39 if(max(n)>thresh) %keep refining the grid as long as there are triangles40 %with more than thresh points filling41 nottrue=0;42 end43

44 xmk=[]; %Start with empty list of triangle corners which will be appended45 for i=1:length(ln)46 if(n(i)>thresh) %population in ln(i) triangle greater than thresh,47 %then split it48 r=ln(i);49 for l=1:3 %Choose the longest edge of triangle

Page 338: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

330 Appendix A. Computation, Codes, and Computational Complexity

50 %to divide that triangle - Note use of P - corners matrix51 d(l)=norm(dt.X(dt.Triangulation(r,P(l,1)),:)52 -dt.X(dt.Triangulation(r,P(l,2)),:));53 end54 [Y,I]=max(d); %Choose longest edge of triangle55 xm=mean(dt.X(dt.Triangulation(r,P(I(1),:)),:));56 %Place a point in the middle of that longest edge57 xmk=[xmk; xm]; %Keep this dividing point amongst collected list58 %of these midpoints to add to Delaunay triangulation59 end60 end61 pause62

63 %If there was some triangle splitting, then append to the triangulation64 if(length(xmk)>0)65 x1=[x1; xmk(:,1)]; x2=[x2; xmk(:,2)]; dt= DelaunayTri([x1 x2]);66 end67

68 figure; hold on; triplot(dt); plot(z(:,1),z(:,2),’r.’,’markersize’,10);69 %plot(x1,x2,’.r’); %plot(xmk(:,1),xmk(:,2),’.r’)70 end71 %Plot only those occupied triangles as filled triangle elements72 figure; hold on; ll = pointLocation(dt,z); triplot(dt);73 patch(’faces’,dt(ll,:), ’vertices’, dt.X, ’FaceColor’,’r’);74

75 function xvecNew = henon(xvec,a)76 % Vectorised henon, i.e. return images of lots of points.77 % xvec is a 2 by m matrix, representing m points. Returns a 2 by m matrix.78 global b7980 xvecNew = zeros(size(xvec));81 xvecNew(1,:) = a - xvec(1,:).^2 + b*xvec(2,:);82 xvecNew(2,:) = xvec(1,:);

A.5 Analysis of RefinementGrid refinement is generally performed when computing integrals for sake of improving ac-curacy, and our Ulam’s matrix estimation problem suffers some of the same computationalissues for any such problem. Generally finer grids should give better accuracy, but finergrids require more sample points, which in turns requires more computational time. In thissection we give three complementary discussions of the issues herein, one statistical basedon Monte-Carlo’s method and the usual random sampling method of Eq. (A.1), one ana-lytical based on Lipschitz continuity and stretching properties of the particular dynamicalsystem, and finally we discuss the ability to resolve fine scale structures when underlyingsets of interest tend to be Cantor sets in so many interesting dynamical systems.

A.5.1 Why Refine the Grid? A Monte-Carlo Discussion

Some theory of approximation in terms of coverings and Lipschitz constants leading to atheory for refinement, and even adaptive refinement can be found in Sec. A.5.2. Here wewill give a complementary description in terms of Monte-Carlo integration. As the gridrefines, that Eq. (4.6) produces a matrix whose dominant eigenvector converges in measureto the invariant measure is the statement of Ulam’s method [319]. Beyond that, betterestimates of the quantities can be understood for any fixed grid size in terms of integrationtheory. The finite size h discussion is most relevant for the many finite time questions, such

Page 339: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.5. Analysis of Refinement 331

as almost invariance and coherence, and many other issues discussed in this book.See [116] for a theory of Monte-Carlo integration in the context of Ulam’s method

with uniform grids and refinement, and then see [102] for the improvement to the (pseudo)-random Monte-Carlo integration by quasi-Monte-Carlo integration (using carefully de-signed uniform sampling, rather than random uniform sampling) and the improvementstherein. We assert here that adaptive refinement can be nicely understood in terms of im-portance sampling from Monte-Carlo theory.

The basic computed quantity for the action of the Frobenius-Perron operator on abox (or triangle) labelled Bi , is to estimate how much of Bi maps across box (triangle) Bj .Rewrite Pi, j in terms of Lebesgue measure as an integration, [86],

Pi, j = m(Bi ∩ F−1(Bj ))

m(Bi )= 1

area(Bi)

∫Bi

χF−1(Bj )(x)dx :=<< χF−1(Bj ) >>, (A.6)

where,χB (x)= i f (1, x ∈ B ,0,else), (A.7)

is the indicator function. Note the standard assumption for Monte-Carlo integration spe-cializes here,

x ∼U (Bi ), (A.8)

drawing from the uniform distribution on the grid element Bi Following [275], we haveused the notation << χF−1(Bj ) >> denotes the true average of the function χF−1(Bj ) on theregion Bi .

Refining the grid of {Bi} has two useful outcomes.

1. Improves the Ulam’s method [319], convergence [211]. This is an infinite time prob-lem, and many discussions and analysis have focused on convergence of the Ulam’smethod including refinement as well as adaptive refinement [85, 189, 157, 87]. Seenotably an analysis of the rate of convergence, [45].

2. Monte-Carlo estimate of the integrals on a the finite sized grid improves. In otherwords, how well do we estimate the integrals in Eq. (A.6).

Note that finite time analysis using the Ulam-Galerkin matrix such as almost invariancerely on point 2 more so than point 1.

The integral presentation of Ulam’s method problems may be estimated by Monte-Carlo’s methods [116, 102] by random samples for Eq. (A.6). In the adaptive grid appli-cation, we should not directly use the slightly easier approximation, Eq. (A.1), since equalsized units (triangles) is implicit in the form of the equation.

First we review the basic Monte-Carlo method, but we specialize notation for a givengrid element Bi and to the discussion of our integrand, χF−1(Bj )(x) in Eq. (A.6) but oth-erwise this is the usual beginning discussion. A collection of uniformly sampled points{xk}mk=1 ⊂ Bi leads to the standard Monte-Carlo integration formula [278, 275],

∫Bi

χF−1(Bj )(x)dx ≈ area(Bi)(< χF−1(Bj ) >±√

< [χF−1(Bj )]2 >−< χF−1(Bj ) >

2

m),

(A.9)

Page 340: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

332 Appendix A. Computation, Codes, and Computational Complexity

where < ·> denotes the sample mean, and likewise used for the second moment,

< χF−1(Bj ) >=1

m

m∑k=1

χF−1(Bj )(xk),< [χF−1(Bj )]2 >= 1

m

m∑k=1

[χF−1(Bj )(xk)]2. (A.10)

Also, the relationship between the true function variance V ar (χF−1(Bj )) and its estimatorV ar (< χF−1(Bj ) >) scales according to,

V ar (< χF−1(Bj )>)= V ar (χF−1(Bj ))

m. (A.11)

Variance refinements are designed to improve this denominator. Let Pi, j denote the truevalue of the integration in Eq. (A.6), and < χF−1(Bj ) > is the estimator. We write the set ofpoints {xk}mk=1 ⊂ Bi , and the subindex does not denote time, but it simply a list of uniformlyrandomly sampled points in the triangle element Bi .

A common scenario of sampling the behavior of the dynamical system is a large sam-ple of initial conditions, perhaps a uniform sample, that we may advect. On the other hand,if we have a single orbit sample instead of a uniform sample, then we write {xik }mk=1 ⊂ Bi asthe set of those points from a long orbit {xi }Ni=1 that fall in Bi , but not emphasize that theseare not likely consecutive in time. This sample is not expected to be uniform, but ratherfall roughly according to the invariant density (if it exists). Our discussion here proceedsas if this invariant density is absolutely continuous, which serves well we find for compu-tationally relevant problems where there is a bit of noise present. Great improvements byreduction of variance methods, leverage variable distributions by adaptive strategies.

Stratified sampling is a powerful method of variance refinement for non-uniformlydistributed samples, which further can be used recursively. We specialize the presentationin [275], but otherwise we closely adopt the notation. If we split Bi into equal disjointvolumes area(Bi)/2 each, that we label a and b. If furthermore m/2, half of of the sampledpoints {xik }mk=1 ⊂ Bi land in each of these sub-triangles. Now we have a new Monte-Carloestimator of Pi, j , which we denote < χF−1(Bj ) >

′,

< χF−1(Bj ) >′= < χF−1(Bj ) >a +< χF−1(Bj ) >b

2. (A.12)

The power of the stratified sampling method is the new estimator of variance.

V ar (< χF−1(Bj ) >′)= V ar (< χF−1(Bj ) >a)+Var (< χF−1(Bj ) >b)

4. (A.13)

Substituting each of the estimators on the right of this equation in Eq. (A.11) yields,

V ar (<χF−1(Bj ) >′)=

V ara(χF−1(Bj ))

m/2 +V arb(χF−1(Bj ))

m/2

4= V ara(χF−1(Bj ))+Varb(χF−1(Bj ))

2m(A.14)

Comparing Eqs. (A.11) and (A.14), together with the Huygens Ð Steiner parallel-axis the-orem,

V ar (χF−1(Bj ))= (A.15)

1

2[V ara(χF−1(Bj ))+Varb(χF−1(Bj ))]+

1

4(<< χF−1(Bj ) >>a −<< χF−1(Bj ) >>b)2.

Page 341: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.5. Analysis of Refinement 333

one can conclude that variance never increases when splitting the region equally and withequal samples, and the variance decreases when V ara(χF−1(Bj )) and V arb(χF−1(Bj )) aredifferent.

In reality if the triangles are split into equal areas, it is not expected that there willbe equal numbers of points in each half. Suppose in each half triangle a and b there areunequal numbers of points, ma , and mb and ma +mb = m. Following the discussion asabove, it follows,

V ar (< χF−1(Bj ) >′)= 1

4[V ara(χF−1(Bj ))

ma+ V arb(χF−1(Bj ))

mb]. (A.16)

This formula can be used to conclude that there will be an improvement in the stratifiedMonte-Carlo estimator due to reduced variance as long as we are not close to the scenariothat V ar (χF−1(Bj )) = V ara(χF−1(Bj )) = V arb(χF−1(Bj )). Since the estimate does not getworse, we can repeat the splitting as long as there are points in each split.

A.5.2 A Lipschitz Continuity Discussion of Computational Complexity

The phrase computational complexity as used here denotes the question of how much com-puter work is expected to fulfill the required task, and how does this computational workscale with changes of the parameters such as when developing the Ulam-Galerkin matrix.Generally a fine grid is desirable for more accurate representation of the dynamical system(unless a Markov partition is used), but a finer grid element such as a small triangle, orrectangle, requires integration of the integral in Eq. (A.6) and to do this requires more sam-ple orbit points as the grid is refined. Also more points are needed for dynamical systemsthat have stronger stretching properties so that the covered regions of F−1(Bj ) is prop-erly sampled for grid elements Bj . Here we discuss these trade-offs in terms of Lipschitzconstants of the mapping, and the Gronwall inequality [262] assuming a continuous timesystem with corresponding regularity. A related discussion of grid refinement in terms ofLipschitz continuity can be found in, [88, 90, 158, 190]. The discussion here is drawnclosely from [39].

Assume square boxes of length q . A triangulation can be handled similarly, butrectangles will simplify this discussion even if triangles allow for the powerful Delaunaytriangulation algorithms in practice. We will give arguments as to how many points arerequired so as to not “miss" any dynamics. We present these arguments as a list. Letf : X × [t0, t]→ X be a flow and X be compact.

1. To find an ε such that for ∀x1, x2 ∈ X , |x1(t0)− x2(t0)|< ε we have

|x1(t)− x2(t)|< q , (A.17)

where x1(t0) and x2(t0) are the initial status of x1, x2 and x1(t), x2(t) are the finalstatus of x1, x2. The purpose of this step is to avoid that the distance d betweenthe final positions of two closed initial points under the flow changing dramatically.Thus we require d ≥ q . See Figure A.9 case (a). So, we must control ε such that case(b) holds for all initial points.

Recall Gronwall’s inequality,

|x1(t)− x2(t)| ≤ |x1(t0)− x2(t0)|eM |t−t0|, (A.18)

Page 342: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

334 Appendix A. Computation, Codes, and Computational Complexity

Figure A.9. (Left)A failure of sampling. Two points in box i evolve to boxes j+1and j − 1 and without further samples, the resulting Ulam-Galerkin matrix would miss atransition i → j which is expected in a continuous transformation f . d > q. (Right) Thesuccessful sampling scenario. Points x1 and x2 are close enough that the f never caststhem further than across successive rectangles j−1 and j . Recall that a continuous f willmap a connected set to a connected set and this should be reflected in the images of thecoarse representation of boxes cast across boxes. d < q.

where M is the Lipschitz constant. Assuming f is uniformly continuously dif-ferentable,

M = max(x ,t )∈X×[t0,t ]

∣∣∣∣∂ f

∂x(x , t)

∣∣∣∣ . (A.19)

We must control the distance between x1(t) and x2(t) through controlling the dis-tance between x1(t0) and x2(t0). Let

ε = q

eM |t−t0| . (A.20)

Then we have,

|x1(t)− x(t)| ≤ |x1(t0)− x(t0)|eM |t−t0| < εeM |t−t0| = q , (A.21)

for all x ∈ X satisfying |x1(t0)− x(t0)| < ε. That is, for any x(t0) in an ε-ball ofx1(t0), x(t0) will keep its distance with x1(t0) less than q through time t− t0.

2. Next, we consider the more general case of the ε-ball of x1(t0); consider ε-ballscentered on each sample data x which must cover the whole X . Since X is compact,the ε-balls of a finite subset {xk}Pk=1 cover X. Let {xk}Pk=1 be initial conditions of theflow, with images under the flow {xk(t)}Pk=1 in a finite time t− t0. Define,

l = min{xk (t )}Pk=1

{distance between each of four boundaries of some j box and xk(t)}

Page 343: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.5. Analysis of Refinement 335

Figure A.10. The red arrow indicates the minimum distance between a point xk(t)for some k and four boundaries of a box j for some j that contains xk(t). We must check allxk in {xk(t)}Pk=1 with all the boxes of the partition of X(t) to develop l.

where the j box is as defined above, is the image of X which we denote X(t). SeeFigure A.10.

Here, again, we use Gronwall’s inequality as follows,

|x(t)− xk(t)| ≤ |x(t0)− xk(t0)|eM |t−t0| ≤ l,

where k = 1,2, ... P . Let

ε′ = l

eM |t−t0| , (A.22)

we have

|xk(t)− x(t)| ≤ |xk(t0)− x(t0)|eM |t−t0| < εeM |t−t0| = l, (A.23)

which is to say, all points x in an ε′-ball of xk in X for some k will map into thel-ball of xk(t) in X(t) for the same k.

3. In this step, we build a new cover for X(t) with l-balls by the following. Thereare two types of balls of the cover, Type I covers are those l-balls never touch theboundaries of the grid; Type II covers are those l-balls centered at boundaries andtheir centers never in the l-balls of Type I, see Figure A.11.

X(t) is compact, so we have a finite subset of this new cover still covering X(t).Let the centers of these balls be {xq (t)}Qq=1, for Q is some integer. Since the flowis assumed continuous as is its inverse is continuous, (see [?]), the pre-image of thethese (xq(t), l)-balls is still a cover of the pre-image X of X(t), which is also finite.By the above discussion, more specifically, the finite cover of X consists of ε′-ballsof some of {xq (t)}Qq=1. That is, the pre-image of Type I l-balls in X(t) and anothertype of covers whose image are Type II l-balls in X(t), the shape of these covers maybe vary. Thus, in X , we have the following case, see Figure A.12.

Page 344: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

336 Appendix A. Computation, Codes, and Computational Complexity

Figure A.11. Covering geometry scenarios.

Figure A.12. Pre-image of covers in two different scenarios. The ellipse denotesthe pre-image of some Type II l-ball, but the pre-image may be any shape which dependson the inverse of the flow.

Consider the areas which are not overlapping of the balls of both of X and X(t). Theflow is a bijection between these two areas. See Figure A.13, which is to say, thepoints in red ε′-balls in X will never go out of the l-balls in X(t).

Notice the important connection between the non-overlapping areas of X and X(t),we can control the non-overlapping area in X by shrink the size of l-balls in X(t).In other words, if l become smaller, so is the non-overlapping area in X , that is, thenon-overalpping area in X can be ignored for some small l. Then we can guaranteea sufficient sampling for a given grid with length q if we choose mesh-grid of pointswith distance ε′ or smaller. The number of points we need in each box of the partitionof the initial status X should be more than q

ε′ . To obtain a ε′, we just choose a smalll which depends on the Lipschitz constant of a specific problem. Note that thisdescription is in terms of a uniform ε′ grid. However, a uniformly random coverwhere this ε′ condition is satisfied is also sufficient.

Page 345: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

A.5. Analysis of Refinement 337

Figure A.13. Geometry between non-overlapping areas sample.

A.5.3 On Cantor-structures when developing Ulam-Galerkin matrices

A common scenario in chaotic dynamical systems and turbulent flow is that fine scalestructures develop, by some sort of Cantor set type construction. The following examplehighlights this propensity with a simple example, and discusses the implications to compu-tational complexity when developing an Ulam-Galerkin matrix.

Example A.1. (A Baker Map and Discussion of Computational Complexity)[288]Consider a Baker’s map as an example to describe the trade-off between the number

of box coverings and the Lyapunov exponent. A Baker’s map on the unit square [0,1]×[0,1] may be written,

xn+1 ={λ1xn if yn < α,(1−λ2)+λ2xn if yn > α,

yn+1 ={

yn/α if yn < α,

(yn−α)/(1−α) if yn > α,

(A.24)

where λ1+λ2 < 1 and 0 < α < 1. The forward-time invariant set of this map is a Cantorset of parallel vertical lines. Consider learning these structures with box covering methods,and what can be discovered by such finite computations. Suppose that we want to coverall parallel stripes left in the unit box after n iterations by square boxes of size ε so thattwo stripes are not contained in the same box, see Figure A.14. The size of the square boxrequired to do this is

ε(n)=min {λn1,λn

2}. (A.25)

Therefore, the required number of boxes to reveal the fractal structure of these parallelstripes is

N(ε) ≥ ε(n)−2, (A.26)

Page 346: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

338 Appendix A. Computation, Codes, and Computational Complexity

50 100 150 200 250 300 350 400 450

50

100

150

200

250

300

350

400

450

(a)

50 100 150 200 250 300 350 400 450

50

100

150

200

250

300

350

400

450

(b)

Figure A.14. Two iterations of the Baker map with λ1 = λ2. (a) All stripes aftern = 2. A square box of size ε1 cannot resolve a structure finer than ε1. But we see that asmaller square box of size ε2 is able to resolve a stripe of width up to ε2. (b) Here, the boxof size ε2 cannot resolve the stripe left in the unit box after n = 3.

and d = 2 is the dimension of the embedding phase space. Using more boxes than theminimum requirement (A.26) would only increase required computer storage and compu-tational work but does not give us a better knowledge of the stripes left in the unit box. Onelesson we see with this is that there is a trade-off between Lyapunov exponents (Chapter8) , λi , time, n, box size, ε, the size of the memory requirements and size of the resultingdigraph, N(ε), and the dimension of the underlying space, d .

Page 347: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography

[1] R.L. Adler, A.G. Konheim, and M.H. McAndrew. Topological entropy. Transactionsof the American Mathematical Society, 114(2):309–319, 1965. (Cited on pp. 288,289)

[2] K.T. Alligood, T. Sauer, and J.A. Yorke. Chaos: An Introduction to Dynamical Sys-tems. Springer, 1996. (Cited on pp. x, 8, 64, 172, 184, 188)

[3] L. Arnold. Random Dynamical Systems. Monographs in Mathematics. Springer,1998. (Cited on p. 244)

[4] L. Arnold. Random Dynamical Systems, 2nd ed. Springer, New York, 2003. (Citedon p. x)

[5] R. Artuso, E. Aurell, and P. Cvitanovic. Recycling of strange sets: I. Cycle expan-sions. Nonlinearity, 3:325, 1990. (Cited on p. 290)

[6] P. Ashwin, J. Buescu, and I. Stewart. From attractor to chaotic saddle: A tale oftransverse instability. Nonlinearity, 9:703–737, 1996. (Cited on p. 232)

[7] D. Auerbach, P. Cvitanovic, J.P. Eckmann, G. Gunaratne, and I. Procaccia. Explor-ing chaotic motion through periodic orbits. Physical Review Letters, 58(23):2387–2389, 1987. (Cited on p. 290)

[8] L.W. Bagget. Functional Analysis: A Primer, Volume 157. Marcel Dekker, 1992.(Cited on pp. 69, 97)

[9] J.P. Bagrow and E. M. Bollt. A local method for detecting communities. Phys. Rev.E, 72(046108), 2005. (Cited on pp. 86, 136)

[10] J.P. Bagrow and E.M. Bollt. Local method for detecting communities. Physical Re-view E, 72(4):46108, 2005. (Cited on pp. 136, 137, 138, 140)

[11] V. Baladi. Positive Transfer Operators and Decay of Correlations. Volume 16 ofAdvanced Series in Nonlinear Dynamics. World Scientific, River Edge, NJ, 2000.(Cited on p. x)

[12] J. Banks, J. Brooks, G. Cairns, G. Davis, and P. Stacey. On Devaney’s definition ofchaos. The American Mathematical Monthly, 99(4):332–334, 1992. (Cited on pp. 8,57, 159, 171, 290)

339

Page 348: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

340 Bibliography

[13] R.B. Bapat and T.E.S. Raghavan. Nonnegative Matrices and Applications. Vol-ume 64 of Encyclopedia of Mathematics and Its Applications. Cambridge UniversityPress, Cambridge, UK, 1997. (Cited on pp. 92, 94)

[14] S. Bassein. The dynamics of a family of one-dimensional maps. The American Math-ematical Monthly, 105(2):118–130, 1998. (Cited on pp. 98, 99)

[15] H. van den Bedem and N. Chernov. Expanding maps of an interval with holes. Er-godic Theory and Dynamical Systems, 12:637–654, 2002. (Cited on p. 145)

[16] M. Benedicks and L. Carleson. The dynamics of the Hénon map. The Annals ofMathematics, 133(1):73–169, 1991. (Cited on p. 176)

[17] N. Berglund. Perturbation Theory of Dynamical Systems. Arxiv preprintmath/0111178, 2001. (Cited on pp. 58, 59)

[18] O. Biham and W. Wenzel. Unstable periodic orbits and the symbolic dynamics of thecomplex Hénon map. Physical Review A, 42(8):4639–4646, 1990. (Cited on p. 292)

[19] L. Billings and E.M. Bollt. Probability density functions of some skew tent maps.Chaos, Solitons and Fractals, 12(2):365–376, 2001. (Cited on pp. 96, 97, 99, 100,101, 102, 290, 293, 301, 307)

[20] L. Billings and I.B. Schwartz. Exciting chaos with noise: Unexpected dynamics inepidemic outbreaks. Journal of Mathematical Biology, 44(1):31–48, 2002. (Citedon p. 229)

[21] G.D. Birkhoff. Dynamical Systems. AMS, Providence, RI, 1966. (Cited on p. 62)

[22] B. Bollobás. Modern Graph Theory. Springer, NewYork, 1998. (Cited on pp. 85,137)

[23] N. Boccara. Functional Analysis: An Introduction for Physicists. Academic Press,Boston, MA, 1990. (Cited on pp. 97, 228)

[24] E.M. Bollt. Communicating with chaos using high-dimensional symbolic dynamics.Physics Letters A, 255(1-2):75, 1999. (Not cited)

[25] E.M. Bollt. Controlling chaos and the inverse Frobenius-Perron problem: Globalstabilization of arbitrary invariant measures. International Journal of Bifurcationand Chaos, 10(5):1033–1050, 2000. (Cited on pp. 55, 56)

[26] E.M. Bollt. Review of chaos communication by feedback control of symbolic dy-namics. International Journal of Bifurcation and Chaos, 13(2):269–186, 2003.(Cited on pp. 154, 157, 158, 186, 191, 193, 195, 196, 197, 283)

[27] E.M. Bollt et al. Optimal targeting of chaos. Physics Letters A, 245(5):399–406,1998. (Cited on p. 3)

[28] E. Bollt, Y.C. Lai, and C. Grebogi. Coding, channel capacity, and noise resistancein communicating with chaos. Physical Review Letters, 79(19):3787–3790, 1997.(Cited on pp. 62, 145, 157, 158, 236, 237, 295)

Page 349: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 341

[29] E. Bollt and Y.C. Lai. Dynamics of coding in communicating with chaos. PhysicalReview E, 58(2):1724–1736, 1998. (Cited on pp. 107, 145, 188, 189, 190, 191, 232,236, 237, 239, 283, 290, 295)

[30] E. Bollt. Combinatorial control of global dynamics in a chaotic differential equation.International Journal of Bifurcations and Chaos, 11(8):2145-2162, 2001. (Cited onp. 20)

[31] E. Bollt. Synchronization as a process of sharing and transferring information. Inter-national Journal of Bifurcations and Chaos, 22, 1250261, 2012. (Cited on pp. 303,304, 306, 307, 309, 310, 311, 312)

[32] E.M. Bollt. The path towards a longer life: On invariant sets and the escape timelandscape. International Journal of Bifurcations and Chaos in Applied Sciences andEngineering, 15(5):1615–1624, 2005. (Cited on pp. 233, 234, 235, 236)

[33] E.M. Bollt. The path towards a longer life: On invariant sets and the escape timelandscape. International Journal of Bifurcation and Chaos, 15(5):1615–1624, 2005.(Cited on p. 232)

[34] E. M. Bollt. Gulf movie. http://people.clarkson.edu/˜ebollt/papers/gulfftle.mov.(Cited on p. 231)

[35] E. Bollt, L. Billings, and I. Schwartz. A manifold independent approach to under-standing transport in stochastic dynamical systems. Physica D, 173:153–177, 2002.(Cited on pp. 71, 224, 227, 229, 230, 231, 232)

[36] E.M. Bollt and A. Klebanoff. A new and simple chaos toy. The International Journalof Bifurcations and Chaos, 12:1843–1857, 2002. (Cited on p. 183)

[37] E.M. Bollt, A. Luttman, S. Kramer, and R. Basnayake. Measurable dynamics analy-sis of transport in the Gulf of Mexico during the oil spill. The International Journalof Bifurcations and Chaos, 22(03):1230012. (Cited on pp. 22, 29, 30, 230, 231, 234)

[38] E.M. Bollt, A. Luttman, S. Kramer, and R. Basnayake. Measurable dynamics anal-ysis of transport in the Gulf of Mexico during the oil spill. International Journal ofBifurcation and Chaos, 22(3):1230012, 2012. (Cited on p. 231)

[39] E. Bollt and T. Ma. Relatively coherent sets as a hierarchical partition method. In-ternational Journal of Bifurcation and Chaos, 2012. (Cited on pp. 149, 151, 231,333)

[40] E. Bollt and J. Skufca. Markov Partitions. Encyclopedia of Nonlinear Science. Edi-tor: Alwyn Scot, Routledge, New York, 2005. (Cited on pp. 75, 76, 77)

[41] E.M. Bollt, T. Stanford, and Y.C. Lai. What symbolic dynamics do we get with amisplaced partition? On the validity of threshold crossings analysis of chaotic time-series. Physica D, 154(3-4):259–286, 2001. (Cited on p. 76)

[42] E.M. Bollt, T. Stanford, Y.C. Lai, and K. Zyczkowski. Validity of threshold-crossinganalysis of symbolic dynamics from chaotic time series. Physical Review Letters,85(16):3524–3527, 2000. (Cited on pp. 187, 193, 196, 285, 299, 300)

Page 350: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

342 Bibliography

[43] E.M. Bollt, T. Stanford, Y.C. Lai, and K. Zyczkowski. What symbolic dynamics dowe get with a misplaced partition? On the validity of threshold crossings analysisof chaotic time-series. Physica D, 154(3-4):259–286, 2001. (Cited on pp. 156, 181,187, 193, 196, 290, 291, 299)

[44] L. Boltzmann and J.T. Blackmore. Ludwig Boltzmann: His Later Life and Philoso-phy, 1900–1906. A Documentary History. Springer, 1995. (Cited on p. 62)

[45] C.J. Bose, and R. Murray. The exact rate of approximation in Ulam’s method. Dis-crete and continuous dynamical systems. Discrete and Continuous Dynamical Sys-tems, 7(1):219–235, 2001. (Cited on p. 331)

[46] R. Bowen. Markov partitions for Axiom A diffeomorphisms. American Journal ofMathematics, 92(3):725–747, 1970. (Cited on pp. 76, 289)

[47] R. Bowen. ω-limit sets for axiom A diffeomorphisms. J. Differential Equations,2:333–339, 1975. (Cited on p. 83)

[48] R. Bowen. Equilibrium States and the Ergodic Theory of Anosov Diffeomorphisms.Volume 470 of Lecture Notes in Mathematics. Springer, Berlin, New York, 1975.(Cited on pp. 76, 83, 163, 289, 290)

[49] R. Bowen. Periodic points and measures for Axiom A diffeomorphisms. Transac-tions of the American Mathematical Society, 154:377–397, 1971. (Cited on p. 290)

[50] R. Bowen. Topological entropy for noncompact sets. Transactions of the AmericanMathematical Society, 184:125–136, 1973. (Cited on p. 288)

[51] A. Boyarsky and Y.-S. Lou. Approximating measures invariant under higher-dimensional chaotic transformations. Japan Research Society of Nonlinear Theoryand Its Applications (IEICE), 65(2):231–244, 1991. (Cited on p. 70)

[52] A. Boyarsky and Y. S. Lou. Approximating measure invariant under higher-dimensional chaotic transformation. Journal of Approximation Theory, 65(2):231–244, 1991. (Cited on p. 71)

[53] A. Boyarsky and P. Gora. Laws of Chaos: Invariant Measures and DynamicalSystems in One Dimension. Probability and Its Applications. Birkhäuser Boston,Boston, MA, 1997. (Cited on pp. x, 6, 71, 75, 97, 99, 101, 102, 103, 300, 301)

[54] L. Breiman. Probability. SIAM, Philadelphia, PA, 1992. (Cited on p. 286)

[55] M. Budišic and I. Mezic. Geometry of the ergodic quotient reveals coherent struc-tures in flows. Physica D, 15:1255–1269, 2012. (Cited on p. 257)

[56] H. Buljan and V. Paar. Parry measure and the topological entropy of chaotic repellersembedded within chaotic attractors. Physica D, 172(1-4):111–123, 2002. (Cited onp. 288)

[57] Q. Chen. Area as a devil’s staircase in twist maps. Physics Letters A, 123:444–450,1987. (Cited on p. 223)

Page 351: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 343

[58] G. Chen, Y. Mao, and C.K. Chui. A symmetric image encryption scheme based on3D chaotic cat maps. Chaos, Solitons and Fractals, 21(3):749–761, 2004. (Cited onp. 283)

[59] F. Christiansen and H.H. Rugh. Computing Lyapunov spectra with continuousGram–Schmidt orthonormalization. Nonlinearity, 10(5):1063–1072, 1997. (Citedon p. 245)

[60] C. Chui, Q. Du, and T. Li. Error estimates of the Markov finite approximation ofthe Frobenius Perron operator. Nonlinear Analysis, 19(4):291–308, 1992. (Cited onp. 71)

[61] F. Chung. Laplacians and the cheeger inequality for directed graphs. Annals of Com-binatorics, 9:1–19, 2005. (Cited on pp. 122, 131)

[62] F. Chung. Spectral Graph Theory. AMS, Providence, RI, 1997. (Cited on p. 142)

[63] A. Clauset, M.E.J. Newman, and C. Moore. Finding community structure in verylarge networks. Physical Review E, 70:066111, 2004. (Cited on p. 142)

[64] P. Collet, S. Martínez, and B. Schmitt. On the enhancement of diffusion by chaos,escape rates and stochastic stability. Transactions of the American MathematicalSociety, 351:7:2875–2897, 1999. (Cited on p. 107)

[65] P. Collet and J.P. Eckmann. Iterated Maps on the Interval as Dynamical Systems.Birkhäuser Boston, Boston, MA, 2009. (Cited on p. 99)

[66] P. Collins. Symbolic dynamics from homoclinic tangles. International Journal ofBifurcation and Chaos in Applied Sciences and Engineering, 12(3):605–618, 2002.(Cited on p. 215)

[67] P. Constantin. Integral Manifolds and Inertial Manifolds for Dissipative Partial Dif-ferential Equations. Springer, 1989. (Cited on p. 33)

[68] N.J. Corron and S.D. Pethel. Control of long-period orbits and arbitrary trajectoriesin chaotic systems using dynamic limiting. Chaos, 12:1, 2002. (Cited on p. 154)

[69] N.J. Corron, S.D. Pethel, and B.A. Hopper. Controlling chaos with simple limiters.Physical Review Letters, 84(17):3835–3838, 2000. (Cited on p. 283)

[70] N.J. Corron, S.D. Pethel, and K. Myneni. Synchronizing the information content ofa chaotic map and flow via symbolic dynamics. Physical Review E, 66(3):36204,2002. (Cited on pp. 158, 283)

[71] T.M. Cover and J.A. Thomas. Elements of Information Theory. John Wiley, 2006.(Cited on pp. 273, 274, 275, 286, 287)

[72] K.M. Cuomo and A.V. Oppenheim. Circuit implementation of synchronized chaoswith applications to communications. Physical Review Letters, 71(1):65–68, 1993.(Cited on p. 307)

Page 352: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

344 Bibliography

[73] P. Cvitanovic, G.H. Gunaratne, and I. Procaccia. Topological and metric propertiesof Hénon-type strange attractors. Physical Review A, 38(3):1503–1520, 1988. (Citedon pp. 77, 179, 191, 193, 194, 291, 292)

[74] G. D’Alessandro, P. Grassberger, S. Isola, and A. Politi. On the topology of theHénon Map. Journal of Physics A: Mathematical and General, 23:5285, 1990.(Cited on pp. 297, 299)

[75] L. Dannon, A. Diaz-Guilera, J. Duch, and A. Arenas. Comparing Community Struc-ture Identification. Preprint cond-mat/0505245, 2005. (Cited on pp. 86, 136)

[76] R.L. Davidchack and Y.C. Lai. Efficient algorithm for detecting unstable periodicorbits in chaotic systems. Physical Review E, 60(5):6172–6175, 1999. (Cited onp. 292)

[77] R.L. Davidchack, Y.C. Lai, E.M. Bollt, and M. Dhamala. Estimating generat-ing partitions of chaotic systems by unstable periodic orbits. Physical Review E,61(2):1353–1356, 2000. (Cited on pp. 291, 292, 293, 295, 296, 297)

[78] R.L. Davidchack, Y.C. Lai, A. Klebanoff, and E.M. Bollt. Towards complete detec-tion of unstable periodic orbits in chaotic systems. Physics Letters A, 287(1-2):99–104, 2001. (Cited on p. 292)

[79] C.S. Daw, C.E.A. Finney, and E.R. Tracy. A review of symbolic analysis of experi-mental data. Review of Scientific Instruments, 74:915, 2003. (Cited on p. 193)

[80] S. Day, R. Frongillo, and R. Treviño. Algorithms for rigorous entropy bounds andsymbolic dynamics. SIAM Journal on Applied Dynamical Systems, 7:1477–1506,2008. (Cited on p. 291)

[81] S. Day, O. Junge, and K. Mischaikow. A rigorous numerical method for the globalanalysis of infinite dimensional discrete dynamical systems. SIAM Journal on Ap-plied Dynamical Systems, 3(2):117–160, 2004. (Cited on p. 291)

[82] S. Day, J.P. Lessard, and K. Mischaikow. Validated continuation for equilibria ofPDEs. SIAM Journal on Numerical Analysis, 45(4):1398–1424, 2008. (Cited onp. 291)

[83] A. de Carvalho and T. Hall. How to prune a horseshoe. Nonlinearity, 15(3):19–68,2002. (Cited on pp. 179, 181, 193)

[84] M. Dellnitz, G. Froyland, and O. Junge. The algorithms behind GAIO-set orientednumerical methods for dynamical systems. In Ergodic Theory, Analysis, and Ef-ficient Simulation of Dynamical Systems, Springer, Berlin, 2001, pages 145–174.(Cited on pp. x, 320, 325, 326)

[85] M. Dellnitz, R. Guder, and E. Kreuzer. An adaptive method for the approximationof the generalized cell mapping. Chaos, Solitons and Fractals, 8(4):525–534, 1997.(Cited on p. 331)

[86] M. Dellnitz, A. Hohmann, O. Junge, and M. Rumpf. Exploring invariant sets andinvariant measures. Chaos, 7(2):221–228, 1997. (Cited on p. 331)

Page 353: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 345

[87] M. Dellnitz and O. Junge. An adaptive subdivision technique for the approxima-tion of attractors and invariant measures. Computing and Visualization in Science,1(2):63–68, 1998. (Cited on p. 331)

[88] M. Dellnitz and O. Junge. An adaptive subdivision technique for the approxima-tion of attractors and invariant measures. Computing and Visualization in Science,1(2):63–68, 1998. (Cited on p. 333)

[89] M. Dellnitz and O. Junge. On the approximation of complicated dynamical behavior.SIAM Journal on Numerical Analysis, 36(2):491–515, 1999. (Cited on pp. x, 63, 65)

[90] M. Dellnitz, R. Guder, and E. Kreuzer. An adaptive method for the approximationof the generalized cell mapping. Chaos, Solitons and Fractals, 8(4):525–534, 1997.(Cited on p. 333)

[91] W. De Melo and S. van Strien. One-Dimensional Dynamics. Springer, Berlin, Hei-delberg, 1993. (Cited on pp. x, 158, 160)

[92] M. Demers and L.-S. Young. Escape rates and conditionally invariant measures.Nonlinearity, 19:377–397, 2006. (Cited on pp. 107, 149)

[93] P. Deuflhard, W. Huisinga, A. Fischer, and C. Schütte. Identification of almost in-variant aggregates in reversible nearly uncoupled Markov chains. Linear Algebraand Its Applications, 315(1-3):39–59, 2000. (Cited on pp. 111, 112, 114, 117)

[94] P. Deuflhard and F. Bornemann. Scientific Computing with Ordinary DifferentialEquations. Springer, 2002. (Cited on p. 59)

[95] P. Deuflhard, M. Dellnitz, O. Junge, and C. Schütte. Computation of essential molec-ular dynamics by subdivision techniques. In Computational Molecular Dynamics:Challenges, Methods, Ideas. Springer-Verlag, Berlin, 1999, pages 98–115. (Citedon pp. 111, 112)

[96] P. Deuflhard and M. Weber. Robust Perron cluster analysis in conformation dynam-ics. Linear Algebra and Its Applications, 398:161–184, 2005. (Cited on pp. 111,114, 117)

[97] R.L. Devaney. An Introduction to Chaotic Dynamical Systems. Westview Press,Boulder, CO, 2003. (Cited on pp. x, 8, 54, 57, 99, 171, 184, 188, 290)

[98] R. Devaney and Z. Nitecki. Shift automorphisms in the Henon mapping. Communi-cations in Mathematical Physics, 67(2):137–146, 1979. (Cited on pp. 161, 176)

[99] P. Diaconis, S. Holmes, and R. Montgomery. Dynamical bias in the coin toss. SIAMReview, 49(2):211, 2007. (Cited on pp. 173, 174)

[100] F.K. Diakonos, P. Schmelcher, and O. Biham. Systematic computation of the leastunstable periodic orbits in chaotic attractors. Physical Review Letters, 81(20):4349–4352, 1998. (Cited on p. 292)

[101] E.I. Dinaburg. On the relations among various entropy characteristics of dynamicalsystems. Izvestiya: Mathematics, 5(2):337–378, 1971. (Cited on p. 288)

Page 354: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

346 Bibliography

[102] J. Ding and Z. Wang. Parallel computation of invariant measures. Annals of Opera-tions Research, 103(1-4):283–290, 2001. (Cited on p. 331)

[103] J. Ding and A.H. Zhou. The projection method for computing multidimensional ab-solutely continuous invariant measures. Journal of Statistical Physics, 77(3-4):899–908, 1994. (Cited on p. 70)

[104] J. Ding and A. Zhou. The projection method for computing multidimensional ab-solutely continuous invariant measures. Journal of Statistical Physics, 77(3/4):899–908, 1994. (Cited on p. 71)

[105] J. Ding and A.H. Zhou. Piecewise linear Markov approximations of Frobenius–Perron operators associated with multi-dimensional transformations, NonlinearAnalysis, 25(4):399–408, 1995. (Cited on p. 70)

[106] J. Ding and A. Zhou. Finite approximations of Frobenius-Perron operators. A so-lution of Ulam’s conjecture to multi-dimensional transformations. Physica D, 92(1-2):61–68, 1996. (Cited on p. 55)

[107] M. Dolnik and E.M. Bollt. Communication with chemical chaos in the presence ofnoise. Chaos, 8:702, 1998. (Cited on p. 283)

[108] J.F. Donges, Y. Zou, N. Marwan, and J. Kurths. The backbone of the climate net-work. Europhysics Letters, 87:48007, 2009. (Cited on pp. 278, 302, 309, 310, 313)

[109] R.W. Easton. Geometric Methods for Discrete Dynamical Systems. Oxford Univer-sity Press, New York, 1998. (Cited on p. 215)

[110] R.W. Easton. Transport through chaos. Nonlinearity, 4:583–590, 1991. (Cited onp. 223)

[111] R.W. Easton. Trellises formed by stable and unstable manifolds in the plane. Trans-actions of the American Mathematical Society, 294(2):719–732, 1986. (Cited onp. 208)

[112] R.W. Easton. Trellis formed by stable and unstable manifolds in the plane. Trans-actions of the American Mathematical Society, 294(2):719–732, 1986. (Cited onp. 223)

[113] J.P. Eckmann and D. Ruelle. Ergodic theory of chaos and strange attractors. Reviewsof Modern Physics, 57(3):617–656, 1985. (Cited on p. 302)

[114] C.H. Edwards, D.E. Penney, and D. Calvis. Elementary Differential Equations.Prentice–Hall, 2008. (Cited on p. 59)

[115] L.C. Evans. Partial Differential Equations. Volume 19 of Graduate Studies in Math-ematics. AMS, Providence, RI, 2010. (Cited on p. 49)

[116] E.Y. Hunt. A monte carlo approach to the approximation of invariant measures. Ran-dom Comput. Dynamics, 2(1):111–133, 1994. (Cited on p. 331)

[117] M. Fiedler. Algebraic connectivity of graphs. Czechoslovak Mathematical Journal,23(2):298–305, 1973. (Cited on p. 142)

Page 355: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 347

[118] B. Fiedler. Ergodic Theory, Analysis, and Efficient Simulation of Dynamical Systems.Springer, 2001. (Cited on p. x)

[119] G.W. Flake, S. Lawrence, C.L. Giles, and F.M. Coetzee. Self-organization of theWeb and identification of communities. Communities, 35(3):66–71, 2002. (Cited onp. 136)

[120] A.L. Fradkov and R.J. Evans. Control of chaos: Methods and applications in engi-neering. Annual Reviews in Control, 29(1):33–56, 2005. (Cited on p. 154)

[121] D. Fritzsche, V. Mehrmann, D.B. Szyld, and E. Virnik. An SVD approach to iden-tifying metastable states of Markov chains. Electronic Transactions on NumericalAnalysis, 29:46–69, 2008. (Cited on p. 117)

[122] G. Froyland. Finite approximation of Sinai-Bowen-Ruelle measures for Anosov sys-tems in two dimensions. Random Comput. Dynamics, 3(4):251–263, 1995. (Citedon p. 70)

[123] G. Froyland. Finite approximation of Sinai-Bowen-Ruelle measure for Anosov sys-tems in two dimension. Random Comput. Dynamics, 3(4):251–263, 1995. (Cited onp. 71)

[124] G. Froyland. Finite approximation of Sinai-Bowen-Ruelle measure for Anosov sys-tems in two dimensions. Random Comput. Dynamics, 3(251), 1995. (Cited onp. 126)

[125] G. Froyland. Estimating Physical Invariant Measures and Space Averages of Dy-namical Systems Indicators. Ph.D. thesis, University of Western Australia, 1996.(Cited on p. 71)

[126] G. Froyland. Computing physical invariant measures. In Proceedings of the 1997 In-ternational Symposium on Nonlinear Theory and Its Applications, Volume 2, 1997,pages 1129–1132. (Cited on p. 70)

[127] G. Froyland. Approximating physical invariant measures of mixing dynamical sys-tems in higher dimensions. Nonlinear Analysis, 32(7):831–860, 1998. (Cited onp. 70)

[128] G. Froyland. Approximating physical invariant measures of mixing dynamical sys-tems in higher dimensions. Nonlinear Analysis, 32(7):831–360, 1998. (Cited onp. 126)

[129] G. Froyland. Extracting dynamical behavior via Markov models. In Nonlinear Dy-namics and Statistics, Birkhäuser Boston, Boston, MA, 2001, pages 283–324. (Citedon p. x)

[130] G. Froyland. Extracting dynamical behaviour via Markov models. In Nonlinear Dy-namics and Statistics, A. Mees, ed., Birkhäuser Boston, Boston, MA, 2001, pages283–324. (Cited on p. 65)

[131] G. Froyland. Statistically optimal almost-invarint sets. Physica D, 200:205–219,2005. (Cited on pp. 65, 66, 120)

Page 356: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

348 Bibliography

[132] G. Froyland. Statistically optimal almost-invariant sets. Physica D, 200(3-4):205–219, 2005. (Cited on pp. 105, 117)

[133] G. Froyland, O. Junge, and G. Ochs. Rigorous computation of topological entropywith respect to a finite partition. Physica D, 154(1-2):68–84, 2001. (Cited on pp. 295,296, 299)

[134] G. Froyland, S. Lloyd, and N. Santitissadeekorn. Coherent sets for nonautonomousdynamical systems. Physica D, 239:1527–1541, 2010. (Cited on pp. 131, 323)

[135] G. Froyland and K. Padberg. Almost-invariant sets and invariant manifolds–Connecting probabilistic and geometric descriptions of coherent structures in flows.Physica D, 238(16):1507–1523, 2009. (Cited on pp. 105, 111, 117)

[136] G. Froyland and K. Padberg. Almost-invariant sets and invariant manifolds: Con-necting probabilistic and geometric descriptions of coherent structures in flows.Physica D, 238(16):1507–1523, 2009. (Cited on pp. 257, 259)

[137] G. Froyland, N. Santitissadeekorn, and A. Monahan. Normalized cut and imagesegmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence,22(8):888–905, 2000. (Cited on pp. 128, 129)

[138] G. Froyland, N. Santitissadeekorn, and A. Monahan. Transport in time-dependentdynamical systems: Finite-time coherent sets. Chaos, 20:043116, 2010. (Cited onpp. 120, 124, 130, 131, 132, 133, 134, 135, 136, 149, 151, 152)

[139] G. Froyland and O. Stancevic. Open and closed dynamical systems. Discrete andContinuous Dynamical Systems Series B, 14:457–472, 2010. (Cited on pp. 145,146, 149)

[140] Z. Galias. Computational methods for rigorous analysis of chaotic systems. In Intel-ligent Computing Based on Chaos, Springer, Berlin, 2009. pages 25–51. (Cited onpp. 291, 293)

[141] Z. Galias. Interval methods for rigorous investigations of periodic orbits. Interna-tional Journal of Bifurcation and Chaos, 11(9):2427–2450, 2001. (Cited on pp. 187,291, 293, 297)

[142] Z. Galias. Rigorous investigation of the Ikeda map by means of interval arithmetic.Nonlinearity, 15:1759, 2002. (Cited on pp. 291, 293, 294, 295)

[143] Z. Galias. Rigorous numerical studies of the existence of periodic orbits for theHénon map. Journal of Universal Computer Science, 4(2):114–124, 1998. (Citedon p. 187)

[144] Z. Galias and W. Tucker. Rigorous study of short periodic orbits for the Lorenzsystem. In IEEE International Symposium on Circuits and Systems (ISCAS 2008).IEEE, 2008, pages 764–767. (Cited on p. 291)

[145] Z. Galias and W. Tucker. Validated study of the existence of short cycles for chaoticsystems using symbolic dynamics and interval tools. International Journal of Bifur-cation and Chaos, 21(2):551–563, 2011. (Cited on p. 187)

Page 357: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 349

[146] A. Gervois and M.L. Mehta. Broken linear transformations. Journal of MathematicalPhysics, 18:1476, 1977. (Cited on p. 99)

[147] J. Gleick, J. Glazier, and G. Gunaratne. Chaos: Making a new science. PhysicsToday, 41:79, 1988. (Cited on p. 170)

[148] A.V. Goldberg and R.E. Tarjan. A new approach to the maximum flow problem.Journal of the ACM, 35:921–940, 1988. (Cited on p. 142)

[149] H. Goldstein, C. Poole, J. Safko, and S.R. Addison. Classical mechanics. AmericanJournal of Physics, 70:782, 2002. (Cited on p. 154)

[150] G.H. Golub and C.F. Van Loan. Matrix Computations, 3rd ed. Johns Hopkins Uni-versity Press, Baltimore, MD, 1996. (Cited on pp. 108, 112, 114, 119, 120, 227)

[151] M. Golubitsky, I. Stewart, P.L. Buono, and J.J. Collins. Symmetry in locomotor cen-tral pattern generators and animal gaits. Nature, 401(6754):693–695, 1999. (Citedon p. 307)

[152] T.N.T. Goodman. Relating topological entropy and measure entropy. Bulletin of theLondon Mathematical Society, 3(2):176, 1971. (Cited on p. 288)

[153] P. Grassberger. On the fractal dimension of the henon attractor. Physics Letters A,6(5):224–226, 1983. (Cited on p. 329)

[154] P. Grassberger, H. Kantz, and U. Moenig. On the symbolic dynamics of the Hénonmap. Journal of Physics A: Mathematical and General, 22:5217–5230, 1989. (Citedon pp. 77, 179, 191, 193, 194, 291, 292)

[155] J.M. Greene, M. Henon, A. Mehr, A.M. Soward, T. Dombre, and U. Frisch. Chaoticstreamlines in the ABC flows. Journal of Fluid Mechanics, 167:353–391, 1986.(Cited on p. 257)

[156] J. Guckenheimer and P. Holmes. Nonlinear Oscillators, Dynamical Systems, and Bi-fucations of Vector Fields. Volume 42 of Applied Mathematical Sciences. Springer,1983. (Cited on pp. x, 31, 34, 38, 173, 183, 205, 246)

[157] R. Guder and E. Kreuzer. Control of an adaptive refinement technique of generalizedcell mapping by system dynamics. Nonlinear Dynamics, 20(1):21–32, 1999. (Citedon p. 331)

[158] R. Guder and E. Kreuzer. Control of an adaptive refinement technique of generalizedcell mapping by system dynamics. Nonlinear Dynamics, 20(1):21–32, 1999. (Citedon p. 333)

[159] B.M. Gurevich and S.V. Savchenko. Thermodynamic formalism for countable sym-bolic Markov chains. Russian Mathematical Surveys, 53:245, 1998. (Cited onp. 288)

[160] L. Hagen and A.B. Kahng. New spectral methods for ratio cut partitioning and clus-tering. IEEE Transactions on Computer-Aided Design of Integrated Circuits andSystems, 11(9):1074–1085, 1992. (Cited on p. 142)

Page 358: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

350 Bibliography

[161] D.W. Hahs and S.D. Pethel. Distinguishing anticipation from causality: Anticipatorybias in the estimation of information flow. Physical Review Letters, 107(12):128701,2011. (Cited on pp. 283, 306, 307)

[162] G. Haller. Distinguished material surfaces and coherent structures in three-dimensional fluid flows. Physica D, 149:248–277, 2001. (Cited on pp. 247, 262,264)

[163] G. Haller. Finding finite-time invariant manifolds in two-dimensional velocity fields.Chaos, 10:99–108, 2000. (Cited on pp. 43, 132, 247, 248, 262, 264)

[164] G. Haller. Lagrangian coherent structures from approximate velocity data. Phys. Flu-ids A, 14:1851–1861, 2002. (Cited on p. 43)

[165] G. Haller. A variational theory of hyperbolic Lagrangian coherent structures. Phys-ica D, 240(7):574–598, 2011. (Cited on pp. 262, 264, 266, 267, 268)

[166] G. Haller and A. C. Poje. Finite-time transport in aperiodic flows. Physica D,119:352–380, 1998. (Cited on p. 43)

[167] S.M. Hammel, C. Jones, and J.V. Moloney. Global dynamical behavior of the opticalfield in a ring cavity. Optical Society of America, Journal B: Optical Physics, 2:552–564, 1985. (Cited on p. 292)

[168] T.E. Harris. The existence of stationary measures for certain markov processes. InProceedings of the Third Berkeley Symposium on Mathematical Statistics and Prob-ability: Held at the Statistical Laboratory, University of California, 1954–1955,Volume 1, University of California Press, Berkeley, Los Angeles, 1956, page 113.(Cited on p. 96)

[169] M. Hasler and Y.L. Maistrenko. An introduction to the synchronization of chaoticsystems: Coupled skew tent maps. IEEE Transactions on Circuits and Systems I:Fundamental Theory and Applications, 44(10):856–866, 1997. (Cited on pp. 307,308)

[170] S. Hayes, C. Grebogi, and E. Ott. Communicating with chaos. Physical Review Let-ters, 70(20):3031–3034, 1993. (Cited on pp. 158, 283)

[171] D. He and L. Stone. Spatio-temporal synchronization of recurrent epidemics.Proceedings of the Royal Society of London. Series B: Biological Sciences,270(1523):1519, 2003. (Cited on p. 307)

[172] M. Hénon. Numerical study of quadratic area-preserving mappings. Quarterly ofApplied Mathematics, 27:291–312, 1969. (Cited on p. 215)

[173] M. Henon. A two-dimensional mapping with a strange attractor. Communications inMathematical Physics, 50(1):69–77, 1976. (Cited on p. 175)

[174] M.W. Hirchh and S. Smale. Differential Eqations, Dynamical Systems, and LinearAlgebra. Academic Press, Orlando, FL, 1974. (Cited on p. 19)

Page 359: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 351

[175] B.K.P. Horn and B. G. Schunck. Determining optical flow. Artificial intelligence,17:185–203, 1981. (Cited on p. 130)

[176] C.S. Hsu. A theory of cell-to-cell mapping dynamical systems. Journal of AppliedMechanics, 47:931, 1980. (Cited on pp. x, 320)

[177] C.S. Hsu. Cell-to-Cell Mapping: A Method of Global Analysis for Nonlinear Sys-tems. Springer, 1987. (Cited on pp. ix, x)

[178] U. Huebner, NB Abraham, and CO Weiss. Dimensions and entropies of chaoticintensity pulsations in a single-mode far-infrared NH3 laser. Physical Review A,40(11):6354–6365, 1989. (Cited on p. 156)

[179] D.A. Huffman. A method for the construction of minimum-redundancy codes. Pro-ceedings of the IRE, 40(9):1098–1101, 1952. (Cited on pp. 270, 272)

[180] F. Hunt. Approximating the invariant measures of randomly perturbed dissipativemaps. Journal of Mathematical Analysis and Applications, 198(2):534–551, 1996.(Cited on p. 71)

[181] B.R. Hunt, J.A. Kennedy, T.Y. Li, and H.E. Nusse. SLYRB measures: Natural in-variant measures for chaotic systems. Physica D, 170(1):50–71, 2002. (Cited onpp. 6, 64)

[182] HYCOM. Hybrid Coordinate Ocean Model (HYCOM). http://www.hycom.org/,2010. (Cited on pp. 24, 27, 233)

[183] K. Ide, D. Small, and S. Wiggins. Distinguished hyperbolic trajectories in time-dependent fluid flows: analytical and computational approach for velocity fields de-fined as data sets. Nonlinear Processes in Geophysics, (9):237–263, 2002. (Cited onp. 43)

[184] K. Ikeda, H. Daido, and O. Akimoto. Optical turbulence: chaotic behavior of trans-mitted light from a ring cavity. Physical Review Letters, 45(9):709–712, 1980. (Citedon p. 292)

[185] M. Iosifescu. Finite Markov Processes and Their Applications. Dover, 1980. (Citedon pp. 148, 228)

[186] M. Jordan. Spectral clustering: Analysis and an algorithm. Machine Learning Jour-nal, 5, 2001. (Cited on p. 142)

[187] M.I. Jordan, A.Y. Ng, and Y. Weiss. On spectral clustering: Analysis and an algo-rithm. In Proceedings of the conference on Neural Information Processing Systems(NIPS), 2001. (Cited on p. 142)

[188] B. Joseph and B. Legras. On the relation between kinematic boundaries, stirring, andbarriers for the antarctic polar vortex. Journal of the Atmospheric Sciences, 59:1198–1212, 2002. (Cited on pp. 135, 261)

[189] O. Junge. An adaptive subdivision technique for the approximation of attractorsand invariant measures: Proof of convergence. Dynamical Systems. An InternationalJournal, 16(3):213–222, 2001. (Cited on p. 331)

Page 360: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

352 Bibliography

[190] O. Junge. An adaptive subdivision technique for the approximation of attractorsand invariant measures: Proof of convergence. Dynamical Systems. An InternationalJournal, 16(3):213–222, 2001. (Cited on p. 333)

[191] T. Kaczynski, K.M. Mischaikow, and M. Mrozek. Computational Homology.Springer, 2004. (Cited on p. 291)

[192] H. Kantz, T. Schreiber, and R.S. Mackay. Nonlinear Time Series Analysis. Volume2000. Cambridge University Press, Cambridge, UK, 1997. (Cited on p. 302)

[193] D.R. Karger and C. Stein. A new approach to the minimum cut problem. Journal ofthe ACM, 43(4):601–640, 1996. (Cited on p. 142)

[194] D. Karrasch. Comment on “A variational theory of hyperbolic Lagrangian coherentstructures, Physica D, 240 (2011), 574–598.” Physica D, 241(17):1470–1473, 2011.(Cited on p. 268)

[195] A. Katok. Fifty years of entropy in dynamics: 1958–2007. Journal of Modern Dy-namics, 1(4):689–718, 2007. (Cited on p. 287)

[196] M.B. Kennel and M. Buhl. Estimating good discrete partitions from observed data:Symbolic false nearest neighbors. Physical Review Letters, 91(8):84102, 2003.(Cited on p. 199)

[197] B.W. Kernighan and S. Lin. An efficient heuristic procedure for partitioning graphs.Bell Systems Technical Journal, 49:291–307, 1970. (Cited on p. 142)

[198] B. Kitchens. Symbolic Dynamics: One-Sided, Two-Sided, and Countable StateMarkov Shifts. Springer, 1998. (Cited on pp. 184, 185, 186, 187, 294)

[199] B.P. Koch and R.W. Leven. Subharmonic and homoclinic bifurcations in a paramet-rically forced pendulum. Physica D, 16(1):1–13, 1985. (Cited on p. 183)

[200] A.N. Kolmogorov and S.V. Fomin. Introductory Real Analysis. Dover, New York,1975. (Cited on pp. 6, 14, 33, 172)

[201] A.N. Kolmogorov and S.V. Fomin. Elements of the Theory of Functions and Func-tional Analysis. Dover, 1999. (Cited on pp. 69, 97)

[202] W.S. Koon, M.W. Lo, J.E. Marsden, and S.D. Ross. Dynamical systems, the three-body problem and space mission design. In International Conference on DifferentialEquations (Berlin, 1999), Volumes 1 and 2, World Scientific, River Edge, NJ, 2000,pages 1167–1181. (Cited on p. 222)

[203] N. Korabel and E. Barkai. Pesin-Type Identity for Weak Chaos. Arxiv preprintarXiv:0808.1398, 2008. (Cited on pp. 301, 302)

[204] G. Kovacic. Lobe area via action formalism in a class of Hamiltonian systems. Phys-ica D, 51(1-3):226–233, 1991. (Cited on p. 223)

[205] V. Kumaran. Josiah Willard Gibbs. Resonance, 12(7):4–11, 2007. (Cited on p. 62)

Page 361: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 353

[206] Y.C. Lai, E. Bollt, and C. Grebogi. Communicating with chaos using two-dimensional symbolic dynamics. Physics Letters A, 255(1-2):75–81, 1999. (Citedon pp. 236, 237, 238, 290)

[207] G.G. Langdon Jr. and J.J. Rissanen. Method and Means for Arithmetic Coding Uti-lizing a Reduced Number of Operations, August 25 1981. U.S. Patent 4,286,256.(Cited on pp. 273, 275)

[208] A. Lasota and M. C. Mackey. Chaos, Fractals, and Noise: Stochastic Aspects ofDynamics, 2nd ed. Springer, New York, 1994. (Cited on pp. x, 45, 46, 48, 50, 53,99, 198)

[209] D.T. Lee and B.J. Schachter. Two algorithms for constructing a Delaunay triangula-tion. International Journal of Parallel Programming, 9(3):219–242, 1980. (Cited onp. 295)

[210] F. Lekien and S.D. Ross. The computation of finite-time Lyapunov exponents onunstructured meshes and for non-Euclidean manifolds. Chaos, 20(1):1–20, 2010.(Cited on pp. 248, 250)

[211] T.Y. Li. Finite approximation for the Frobenius-Perron operator. A solution toUlam’s conjecture. Journal of Approximation Theory, 17(2):177–186, 1976. (Citedon pp. 69, 70, 331)

[212] A.J. Lichtenberg and M.A. Lieberman. Regular and stochastic motion. AppliedMathematics, 38, 1983. (Cited on p. 183)

[213] D.R. Lide. CRC Handbook of Chemistry and Physics. CRC, 2003. (Cited on p. 89)

[214] D.A. Lind and B. Marcus. An Introduction to Symbolic Dynamics and Coding. Cam-bridge University Press, Cambridge, UK, 1995. (Cited on pp. 160, 184, 185, 186,187, 299)

[215] D. Lippolis and P. Cvitanovic. How well can one resolve the state space of a chaoticmap? Physical Review Letters, 104(1):14101, 2010. (Cited on pp. 197, 198, 199)

[216] C. Liverani and V. Maume-Deschamps. Lasota-Yorke maps with holes: Condition-ally invariant probability measures and invariant probability measures on the sur-vivor set. Annales de l’Institut Henri Poincaré. Probabilités et Statistiques, 39:385–412, 2003. (Cited on p. 145)

[217] E.N. Lorenz. Deterministic Nonperiodic Flow. Journal of the Atmospheric Sciences,20:130–141, 1963. (Cited on p. 156)

[218] R.S. MacKay. Renormalisation in Area-Preserving Maps. World Scientific, RiverEdge, NJ, 1993. (Cited on pp. 213, 216)

[219] R.S. Mackey, J.D. Meiss, and I.C. Percival. Transport in Hamiltonian systems. Phys-ica D, 13(1-2):55–81, 1984. (Cited on pp. 209, 210, 221, 223)

[220] A.M. Mancho, D. Small, S. Wiggins, and K. Ide. Computation of stable and un-stable manifolds of hyperbolic trajectories in two-dimensional, aperiodically time-dependent vectors fields. Physica D, 182:188–222, 2003. (Cited on p. 43)

Page 362: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

354 Bibliography

[221] A.M. Mancho, D. Small, and S. Wiggins. Computation of hyperbolic trajectoriesand their stable and unstable manifolds for oceanographic flows represented as datasets. Nonlinear Processes in Geophysics, 11:17–33, 2004. (Cited on p. 43)

[222] A.M. Mancho, D. Small, and S. Wiggins. A tutorial on dynamical systems conceptsapplied to Lagrangian transport in oceanic flows defined as finite time data sets: The-oretical and computational issues. Physics Reports, 437(3-4):55–124, 2006. (Citedon p. 43)

[223] M. Martinelli, M. Dang, and T. Seph. Defining chaos. Mathematics Magazine,71:112–122, 1998. (Cited on p. 8)

[224] S. Martínez, P. Collet, and J. San Martín. Asymptotic laws for one-dimensionaldiffusions conditioned to nonabsorption. The Annals of Probability, 23:1300–1314,1995. (Cited on p. 149)

[225] J.L. McCauley. Chaos, Dynamics and Fractals: An Algorithmic Approach to Deter-ministic Chaos. Cambridge Nonlinear Science Series. Cambridge University Press,Cambridge, UK, 1993. (Cited on pp. x, 11)

[226] M.E. McIntyre and T.N. Palmer. The “surf zone" in the stratosphere. Journal ofAtmospheric and Terrstrial Physics, 46:825–849, 1983. (Cited on pp. 135, 259)

[227] F.A. McRobie and J.M.T. Thompson. Lobe dynamics and the escape from a potentialwell. Proceedings of the Royal Society. London. Series A: Mathematical, Physicaland Engineering Sciences, 435(1895):659–672, 1991. (Cited on p. 215)

[228] J.D. Meiss. Differential Dynamical Systems. Volume 14 of Monographs on Mathe-matical Modeling and Computation. SIAM, Philadelphia, 2007. (Cited on pp. x, 31,32)

[229] J.D. Meiss. Symplectic maps, variational principles, and transport. Reviews of Mod-ern Physics, 64(3):795–848, 1992. (Cited on pp. 210, 215, 216, 221, 317, 323)

[230] I. Mezic and F. Sotiropoulos. Ergodic theory and experimental visualization ofchaos. Physics of Fluids, 14:2235–2243, 2002. (Cited on p. 68)

[231] I. Mezic and S. Wiggins. A method for visualization of invariant sets of dynamicalsystems based on ergodic partition. Chaos, 9:213–218, 1999. (Cited on p. 68)

[232] W.M. Miller. Stability and approximation of invariant measures for a class of non-expanding transformations. Nonlinear Analysis, 23(8):1013–1015, 1994. (Cited onp. 65)

[233] J. Milnor and W. Thurston. On iterated maps of the interval. In Dynamical Systems.Volume 1342 of Lecture Notes in Mathematics. Springer, Berlin, 1988, pages 465–563. (Cited on pp. 99, 158, 160, 187, 188, 190, 193)

[234] K. Mischaikow and M. Mrozek. Chaos in the Lorenz equations: A computer assistedproof. American Mathematical Society. Bulletin. New Series, 32(1):66–72, 1995.(Cited on pp. 187, 291)

Page 363: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 355

[235] K. Mischaikow and M. Mrozek. Chaos in the Lorenz equations: A computer as-sisted proof. Part II: Details. Mathematics of Computation, 67(223):1023–1046,1998. (Cited on p. 291)

[236] K. Mischaikow, M. Mrozek, and A. Szymczak. Chaos in the Lorenz equations:A computer assisted proof. III. Classical parameter values. Journal of DifferentialEquations, 169(1):17–56, 2001. (Cited on p. 291)

[237] M. Misiurewicz. Possible jumps of entropy for interval maps. Qualitative Theory ofDynamical Systems, 2(2):289–306, 2001. (Cited on pp. 99, 102)

[238] M. Misiurewicz and E. Visinescu. Kneading sequences of skew tent maps. Annalesde l’Institut Henri Poincaré. Probabilités et Statistiques, 27:125–140, 1991. (Citedon pp. 99, 100, 101, 102)

[239] K.A. Mitchell. The topology of nested homoclinic and heteroclinic tangles. PhysicaD, 238(7):737–763, 2009. (Cited on pp. 215, 221, 222, 295)

[240] K.A. Mitchell and J.B. Delos. A new topological technique for characterizing ho-moclinic tangles. Physica D, 221(2):170–187, 2006. (Cited on p. 234)

[241] K.A. Mitchell, J.P. Handley, J.B. Delos, and S.K. Knudson. Geometry and topol-ogy of escape. II. Homotopic lobe dynamics. Chaos, 13:892–902, 2003. (Cited onpp. 221, 222)

[242] R.E. Moore and F. Bierbaum. Methods and Applications of Interval Analysis. SIAM,Philadelphia, PA, 1979. (Cited on p. 293)

[243] P. Moresco and S. Ponce Dawson. The PIM-simplex method: An extension of thePIM-triple method to saddles with an arbitrary number of expanding directions.Physica D, 126(1-2):38–48, 1999. (Cited on p. 232)

[244] J. Moser. On Quadratic Symplectic Mappings. Mathematische Zeitschrift, 216:417–430, 1994. (Cited on pp. 161, 216)

[245] J.R. Munkres. Toplogy, 2nd ed. Prentice–Hall, Upper Saddle River, NJ, 2000. (Citedon pp. 76, 97, 295)

[246] R. Murray. Ulam’s method for some non-uniformly expanding maps. Discrete andContinuous Dynamical Systems, 26(3):1007–1018, 2010. (Cited on p. 71)

[247] E.R. Nash, P.A. Newman, J.E. Rosenfield, and M.R. Schoeberl. An objective deter-mination of the polar vortex using Ertel’s potential vorticity. Journal of GeophysicalResearch: Atmospheres, 101:9471–9476, 1996. (Cited on pp. 259, 260)

[248] M.E.J. Newman. Fast algorithm for detecting community structure in networks.Physical Review E, 69:066133, 2004. (Cited on p. 139)

[249] M.E.J. Newman. Finding community structure in networks using the eigenvectors ofmatrices. Physical Review E, 74:036104, 2006. (Cited on p. 142)

[250] M.E.J. Newman and M. Girvan. Statistical Mechanics of Complex Networks.Springer, Berlin, 2004. (Cited on pp. 81, 138)

Page 364: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

356 Bibliography

[251] M.E.J. Newman and M. Girvan. Finding and evaluating community structure in net-works. Phys. Rev. E, 69(026113), 2004. (Cited on pp. 81, 86, 136, 137, 139)

[252] M.E.J. Newman and M. Girvan. Finding and evaluating community structure in net-works. Physical Review E, 69(2):26113, 2004. (Cited on p. 137)

[253] H.E. Nusse and J.A. Yorke. A procedure for finding numerical trajectories on chaoticsaddles. Physica D, 36(1-2):137–156, 1989. (Cited on pp. 62, 232)

[254] V.I. Oseledets. A multiplicative ergodic theorem. Lyapunov characteristic numbersfor dynamical systems. Transactions of the Moscow Mathematical Society, 19:197–231, 1968. (Cited on p. 244)

[255] E. Ott. Chaos in Dynamical Systems, 1st ed. Cambridge Universiy Press, Cambridge,UK, 1993. (Cited on pp. x, 328)

[256] E. Ott, C. Grebogi, and J.A. Yorke. Controlling chaos. Physical Review Letters,64(11):1196–1199, 1990. (Cited on pp. 154, 283)

[257] E. Ott and J.D. Meiss. Markov-tree model of intrinsic transport in Hamiltonian sys-tems. Physical Review Letters, 55:2741–2744, 1985. (Cited on p. 223)

[258] K. Padberg. Numerical Analysis of Transport in Dynamical Systems. 2005. (Citedon p. x)

[259] W. Parry. Symbolic dynamics and transformations of the unit interval. Transactionsof the American Mathematical Society, 122(2):368–378, 1966. (Cited on p. 288)

[260] L.M. Pecora and T.L. Carroll. Master stability functions for synchronized coupledsystems. Physical Review Letters, 80(10):2109–2112, 1998. (Cited on p. 307)

[261] L.M. Pecora and T.L. Carroll. Synchronization in chaotic systems. Physical ReviewLetters, 64(8):821–824, 1990. (Cited on p. 307)

[262] L. Perko. Differential Equations and Dynamical Systems. Springer, 2001. (Cited onpp. x, 19, 32, 33, 34, 183, 205, 333)

[263] Y.B. Pesin. Characteristic Lyapunov exponents and smooth ergodic theory. RussianMathematical Surveys, 32(4):55–114, 1977. (Cited on pp. 301, 302)

[264] S.D. Pethel, N.J. Corron, and E. Bollt. Deconstructing spatiotemporal chaos usinglocal symbolic dynamics. Physical Review Letters, 99(21):214101, 2007. (Cited onp. 283)

[265] S.D. Pethel, N.J. Corron, and E. Bollt. Symbolic dynamics of coupled map lattices.Physical Review Letters, 96(3):34105, 2006. (Cited on p. 154)

[266] S.D. Pethel, N.J. Corron, Q.R. Underwood, and K. Myneni. Information flow inchaos synchronization: Fundamental tradeoffs in precision, delay, and anticipation.Physical Review Letters, 90(25):254101, 2003. (Cited on p. 154)

Page 365: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 357

[267] G. Pianigiani and J. A. Yorke. Expanding maps on sets which are almost invariant.Decay and chaos. Transactions of the American Mathematical Society, 252:351–366, 1979. (Cited on pp. 145, 149)

[268] R.T. Pierrehumbert and H. Yang. Global chaotic mixing on isentropic surfaces. Jour-nal of the Atmospheric Sciences, 50:2462–2480, 1993. (Cited on pp. 188, 247)

[269] R.A. Plumb. Stratospheric transport. Journal of the Meteorological Society of Japan,80:793–809, 2002. (Cited on p. 258)

[270] H. Poincaré and R. Magini. Les méthodes nouvelles de la mécanique céleste. IlNuovo Cimento (1895–1900), 10(1):128–130, 1899. (Cited on p. 218)

[271] P.K. Pollett, G. Froyland, and R. M Stuart. A closing scheme for finding almost-invariant sets in open dynamical systems. Submitted, 2011. (Cited on p. 146)

[272] M. Pollicott and M. Yuri. Dynamical Systems and Ergodic Theory. Volume 40 ofLondon Mathematical Society Student Texts. Cambridge University Press, Cam-bridge, UK, 1998. (Cited on p. 37)

[273] M.A. Porter, J.P. Onnela, and P.J. Mucha. Communities in networks. Notices of theAmerican Mathematical Society, 56(9):1082–1097, 2009. (Cited on pp. 136, 137)

[274] D. Pountain. Run-length encoding. Byte, 12(6):317–319, 1987. (Cited on p. 271)

[275] W.H. Press, S.A. Teukolsky, W.T. Vetterling, and B.P. Flannery. Numerical Recipesin C: The Art of Scientific Computing, 2nd ed. Cambridge University Press, Cam-bridge, UK, 1992. (Cited on pp. 331, 332)

[276] A. Pressley. Elementary Differential Geometry. Springer, 2010. (Cited on pp. 35,36)

[277] F. Radicchi, C. Castellano, F. Cecconi, V. Loreto, and D. Parisi. Defining and iden-tifying communities in networks. Proceedings of the National Academy of Sciences,101(9):2658, 2004. (Cited on p. 136)

[278] C.P. Robert and G. Casella. Monte Carlo Statistical Methods. Springer, New York,1998. (Cited on p. 331)

[279] C. Robinson. Dynamical Systems: Stability, Symbolic Dynamics, and Chaos, 2nded. CRC, Boca Raton, FL, 1999. (Cited on pp. x, 31, 38, 39, 42, 77, 171, 177, 195,205, 211, 213, 233, 288)

[280] J.C. Robinson. Infinite-Dimensional Dynamical Systems: An Introduction to Dissi-pative Parabolic PDEs and the Theory of Global Attractors. Cambridge UniversityPress, Cambridge, UK, 2001. (Cited on pp. 33, 49, 295)

[281] R.C. Robinson and C. Robinson. Dynamical Systems: Stability, Symbolic Dynamics,and Chaos. CRC, Boca Raton, FL, 1999. (Cited on pp. 31, 154, 185, 186, 188, 213,289, 290, 291)

[282] V. Rom-Kedar. Transport rates of a class of two-dimensional maps and flows. Phys-ica D, 43:229–268, 1990. (Cited on p. 223)

Page 366: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

358 Bibliography

[283] H.L. Royden and P. Fitzpatrick. Real Analysis, Volume 2. Macmillan, New York,1968. (Cited on pp. 13, 97)

[284] D.J. Rudolph. Fundamentals of Measurable Dynamics: Ergodic Theory onLebesgue Spaces. Clarendon Press, Oxford University Press, New York, 1990.(Cited on p. 76)

[285] D. Ruelle. An inequality for the entropy of differentiable maps. Bulletin of the Brazil-ian Mathematical Society, 9(1):83–87, 1978. (Cited on p. 301)

[286] I.I. Rypina, M.G. Brown, F.J. Beron-Vera, H. Kocak, M.J. Olascoaga, and I.A.Udovydchenkov. On the Lagrangian dynamics of atmospheric zonal jets and thepermeability of the stratospheric polar vortex. Journal of the Atmospheric Sciences,64:3595–3610, 2007. (Cited on p. 131)

[287] N. Santitissadeekorn and E. Bollt. The infinitestimal operator for the semigroup ofthe frobenius-perron operator from image sequence data: Vector fields and com-putational measurable dynamics from movies. Chaos, 17:023126, 2007. (Cited onp. 50)

[288] N. Santitissadeekorn and E.M. Bollt. Identifying stochastic basin hopping and mech-anism by partitioning with graph modularity. Physica D, (231):95–107, 2007. (Citedon pp. 53, 337)

[289] N. Santitissadeekorn, G. Froyland, and A. Monahan. Optimally coherent sets in geo-physical flows: A new approach to delimiting the stratospheric polar vortex. PhysicalReview E, 82(056311), 2010. (Cited on p. 130)

[290] P. Schmelcher and F.K. Diakonos. Detecting unstable periodic orbits of chaoticdynamical systems. Physical Review Letters, 78(25):4733–4736, 1997. (Cited onpp. 292, 296)

[291] A. Schmiegel and B. Eckhardt. Fractal stability border in plane couette flow. Physi-cal Review Letters, 79:5250–5253, 1997. (Cited on p. 231)

[292] T. Schreiber. Measuring information transfer. Physical Review Letters, 85(2):461–464, 2000. (Cited on pp. 302, 305)

[293] I.B. Schwartz. Multiple stable recurrent outbreaks and predictability in seasonallyforced nonlinear epidemic models. Journal of Mathematical Biology, 21(3):347–361, 1985. (Cited on p. 229)

[294] I.B. Schwartz. Small amplitude, long period outbreaks in seasonally driven epi-demics. Journal of mathematical biology, 30(5):473–491, 1992. (Cited on p. 229)

[295] I. B. Schwartz, L. Billings, E. Bollt. Dynamical epidemic suppression using stochas-tic prediction and control. Physical Review E. 70:046220, 2004. (Not cited)

[296] J.D. Skufca and E.M. Bollt. A concept of homeomorphic defect for defining mostlyconjugate dynamical systems. Chaos, 18:013118, 2008. (Cited on p. 12)

Page 367: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 359

[297] J.D. Skufca, J.A. Yorke, and B. Eckhardt. Edge of chaos in a parallel shear flow.Physical Review Letters, 96:174101, 2006. (Cited on p. 232)

[298] S.C. Shadden. A Dynamical Systems Approach to Unsteady Systems. Dissertation,California Institute of Technology, Pasadena, CA, 2006. (Cited on pp. 247, 250,261, 264)

[299] S.C. Shadden, J.O. Dabiriand, and J.E. Marsden. Lagrangian analysis of fluid trans-port in empirical vortex ring flows. Physics of Fluids, 18(4):271–304, 2006. (Citedon p. 43)

[300] S.C. Shadden, F. Lekien, and J.E. Marsden. Definition and properties of Lagrangiancoherent structures from finite-time Lyapunov exponents in two-dimensional aperi-odic flows. Physica D, 212:271–304, 2005. (Cited on p. 43)

[301] T. Shinbrot, C. Grebogi, E. Ott, and J.A. Yorke. Using small perturbations to controlchaos. Nature, 363:411–417, 1993. (Cited on p. 283)

[302] T. Shinbrot. Progress in the control of chaos. Advances in Physics, 44(2):73–111,1995. (Cited on p. 154)

[303] E. Simiu. Chaotic Transistion in Deterministic and Stochastic Dynamcail Systems.Princeton University Press, Princeton, NJ, 2002. (Cited on p. 183)

[304] H.D. Simon, A. Pothen, and K.-P. Liou. Partitioning sparse matrices with eigenvec-tors of graphs. SIAM Journal on Matrix Analysis and Applications, 11(3):430–452,1990. (Cited on p. 143)

[305] Y.G. Sinai, editor. Dynamical Systems II. Ergodic Theory with Applications to Dy-namical Systems and Statistical Mechanics. Volume 2 of Encyclopaedia of Mathe-matical Sciences. Springer, Berlin, New York, 1989. (Cited on p. 61)

[306] S. Smale. Differentiable dynamical systems. Bulletin of the American MathematicalSociety, 73:747–817, 1967. (Cited on pp. 156, 160, 182, 211, 212)

[307] S. Smale. Finding a horseshoe on the beaches of Rio. The Mathematical Intelli-gencer, 20(1):39–44, 1998. (Cited on pp. 156, 161)

[308] A.H. Sobel, R.A. Plumb, and D.W. Waugh. On methods of calculating transportacross the polar vortex edge. Journal of the Atmospheric Sciences, 54:2241–2260,1997. (Cited on pp. 135, 259)

[309] J.L. Snell and J.G. Kemeny. Finite Markov Chains: With a New Appendix “Gener-alization of a Fundamental Matrix.” Undergraduate Texts in Mathematics. Springer,1976. (Cited on pp. 148, 228)

[310] J. Stoer and R. Bulirsch. Introduction to Numerical Analysis. Springer, 2002. (Citedon pp. 86, 89, 90)

[311] M. Stoer and F. Wagner. A simple min-cut algorithm. In Algorithms—ESA94. Vol-ume 855 of Lecture Notes in Computer Science. Springer, Berlin, 1994, pages 141–147. (Cited on p. 142)

Page 368: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

360 Bibliography

[312] S.H. Strogatz. Nonlinear Dynamics and Chaos: With Applications to Physics, Bi-ology, Chemistry, and Engineering. Westview Press, Boulder, CO, 1994. (Cited onp. 307)

[313] S.H. Strogatz. Nonlinear Dynamics and Chaos: With Applications to Physics, Bi-ology, Chemistry, and Engineering. Westview Press, Boulder, CO, 2000. (Cited onpp. x, 32)

[314] S.H. Strogatz and I. Stewart. Coupled oscillators and biological synchronization.Scientific American, 269(6):102–109, 1993. (Cited on p. 307)

[315] R. Sturman, J.M. Ottino, and S. Wiggins. The Mathematical Foundations of Mixing.The Linked Twist Map as a Paradigm in Applications: Micro to Macro, Fluids toSolids. Cambridge University Press, Cambridge, Uk, 2006. (Cited on p. 64)

[316] J. Sun, E.M. Bollt, and T. Nishikawa. Master stability functions for coupled nearlyidentical dynamical systems. Europhysics Letters, 85:60011, 2009. (Cited on p. 307)

[317] D. Sweet, H.E. Nusse, and J.A. Yorke. Stagger-and-step method: Detectingand computing chaotic saddles in higher dimensions. Physical Review Letters,86(11):2261–2264, 2001. (Cited on p. 232)

[318] A.N. Tikhonov. Systems of differential equations containing small parameters in thederivatives. Matematicheskii Sbornik, 73(3):575–586, 1952. (Cited on p. 59)

[319] S. M. Ulam. Problems in Modern Mathematics. Science Editions. John Wiley, NewYork, 1970. (Cited on pp. xi, 14, 55, 65, 70, 71, 126, 315, 330, 331)

[320] A.B. Vasil’eva and V.M. Volosov. The work of Tikhonov and his pupils on ordinarydifferential equations containing a small parameter. Russian Mathematical Surveys,22(2):124–142, 1967. (Cited on p. 59)

[321] N. G. Van Kampen. Stochastic Processes in Physics and Chemistry, 4th ed. North–Holland, 2003. (Cited on pp. 85, 148)

[322] J. Von Neumann. Various techniques used in connection with random digits. AppliedMathematics Series, 12(36-38):1, 1951. (Cited on p. 8)

[323] S. Wasserman and K. Faust. Social Network Analysis: Methods and Applications.Cambridge University Press, Cambridge, UK, 1994. (Cited on p. 136)

[324] C.O. Weiss and J. Brock. Evidence for Lorenz-type chaos in a laser. Physical ReviewLetters, 57(22):2804–2806, 1986. (Cited on p. 156)

[325] C.O. Weiss, W. Klische, N.B. Abraham, and U. Hübner. Comparison of NH3 laserdynamics with the extended Lorenz model. Applied Physics B: Lasers and Optics,49(3):211–215, 1989. (Cited on p. 156)

[326] C. Werndl. What are the new implications of chaos for unpredictability? The BritishJournal for the Philosophy of Science, 60(1):195, 2009. (Cited on p. 8)

[327] S. Wiggins. Chaotic Transport in Dynamical Systems. Springer, 1992. (Cited onpp. 39, 154, 173, 183, 208, 212)

Page 369: Contentsebollt/Box/Jan13.pdf · be found concerning connections of the theory of Frobenius-Perron operators and the ad-joint Koopman operator, as well as useful background in measure

Bibliography 361

[328] S. Wiggins. Chaotic Transport in Dynamical Systems. Springer, 1992. (Cited onpp. 215, 221)

[329] S. Wiggins. Introduction to Applied Nonlinear Dynamical Systems and Chaos, 2nded. Springer, 2003. (Cited on pp. x, 31, 37, 38, 39, 41, 42, 173, 182)

[330] S. Wiggins and J.M. Ottino. Foundations of chaotic mixing. Philosophical Transac-tions of the Royal Society of London. Series A: Mathematical, Physical and Engi-neering Sciences, 362(1818):937, 2004. (Cited on p. 215)

[331] S. Wiggins and V. Rom-Kedar. Transport in two-dimensional maps. Archive for Ra-tional Mechanics and Analysis, 109(3):239–298, 1990. (Cited on p. 223)

[332] A. Wolf, J.B. Swift, H.L. Swinney, and J.A. Vastano. Determining Lyapunov expo-nents from a time series. Physica D, 16(3):285–317, 1985. (Cited on p. 302)

[333] K. Yagasaki. The method of melnikov for perturbations of multi-degree-of-freedomHamiltonian systems. Nonlinearity, 12:799, 1999. (Cited on p. 183)

[334] Z.P. You, E.J. Kostelich, and J.A. Yorke. Calculating stable and unstable manifolds.International Journal of Bifurcation and Chaos, 1(3):605–623, 1991. (Cited onp. 43)

[335] L.S. Young. Dimension, entropy and Lyapunov exponents. Ergodic Theory and Dy-namical Systems, 2(01):109–124, 1982. (Cited on pp. 301, 302)

[336] L.S. Young. What are SRB measures, and which dynamical systems have them?Journal of Statistical Physics, 108(5):733–754, 2002. (Cited on pp. 20, 64)

[337] A. Zhou and J. Ding. Statistical Properties of Deterministic Systems. Springer, 2008.(Cited on p. x)

[338] J. Ziv and A. Lempel. A universal algorithm for sequential data compression. IEEETransactions on Information Theory, 23(3):337–343, 1977. (Cited on pp. 273, 275)

[339] K. Zyczkowski and E.M. Bollt. On the entropy devil’s staircase in a family of gap-tent maps. Physica D, 132(3):392–410, 1999. (Cited on pp. 13, 237, 238)

[340] P. Zgliczynski. Computer assisted proof of chaos in the Rössler equations and in theHénon map. Nonlinearity, 10:243, 1997. (Cited on pp. 176, 187)

[341] P. Zgliczynski. Rigorous numerics for dissipative partial differential equations II. Pe-riodic orbit for the Kuramoto–Sivashinsky PDE: A computer-assisted proof. Foun-dations of Computational Mathematics, 4(2):157–185, 2004. (Cited on p. 291)